mirror of
https://github.com/python/cpython.git
synced 2025-08-01 15:43:13 +00:00

The fix in gh-119561 introduced an assertion that doesn't hold true if any of the three new test extension modules are loaded more than once. This is fine normally but breaks if the new test_check_state_first() is run more than once, which happens for refleak checking and with the regrtest --forever flag. We fix that here by clearing each of the three modules after loading them. We also tweak a check in _modules_by_index_check().
(cherry picked from commit ae7b17673f
)
Co-authored-by: Eric Snow <ericsnowcurrently@gmail.com>
3135 lines
117 KiB
Python
3135 lines
117 KiB
Python
import builtins
|
|
import errno
|
|
import glob
|
|
import json
|
|
import importlib.util
|
|
from importlib._bootstrap_external import _get_sourcefile
|
|
from importlib.machinery import (
|
|
AppleFrameworkLoader,
|
|
BuiltinImporter,
|
|
ExtensionFileLoader,
|
|
FrozenImporter,
|
|
SourceFileLoader,
|
|
)
|
|
import marshal
|
|
import os
|
|
import py_compile
|
|
import random
|
|
import shutil
|
|
import stat
|
|
import subprocess
|
|
import sys
|
|
import textwrap
|
|
import threading
|
|
import time
|
|
import types
|
|
import unittest
|
|
from unittest import mock
|
|
import _imp
|
|
|
|
from test.support import os_helper
|
|
from test.support import (
|
|
STDLIB_DIR, swap_attr, swap_item, cpython_only, is_apple_mobile, is_emscripten,
|
|
is_wasi, run_in_subinterp, run_in_subinterp_with_config, Py_TRACE_REFS,
|
|
requires_gil_enabled, Py_GIL_DISABLED)
|
|
from test.support.import_helper import (
|
|
forget, make_legacy_pyc, unlink, unload, ready_to_import,
|
|
DirsOnSysPath, CleanImport, import_module)
|
|
from test.support.os_helper import (
|
|
TESTFN, rmtree, temp_umask, TESTFN_UNENCODABLE)
|
|
from test.support import script_helper
|
|
from test.support import threading_helper
|
|
from test.test_importlib.util import uncache
|
|
from types import ModuleType
|
|
try:
|
|
import _testsinglephase
|
|
except ImportError:
|
|
_testsinglephase = None
|
|
try:
|
|
import _testmultiphase
|
|
except ImportError:
|
|
_testmultiphase = None
|
|
try:
|
|
import _interpreters
|
|
except ModuleNotFoundError:
|
|
_interpreters = None
|
|
try:
|
|
import _testinternalcapi
|
|
except ImportError:
|
|
_testinternalcapi = None
|
|
|
|
|
|
skip_if_dont_write_bytecode = unittest.skipIf(
|
|
sys.dont_write_bytecode,
|
|
"test meaningful only when writing bytecode")
|
|
|
|
|
|
def _require_loader(module, loader, skip):
|
|
if isinstance(module, str):
|
|
module = __import__(module)
|
|
|
|
MODULE_KINDS = {
|
|
BuiltinImporter: 'built-in',
|
|
ExtensionFileLoader: 'extension',
|
|
AppleFrameworkLoader: 'framework extension',
|
|
FrozenImporter: 'frozen',
|
|
SourceFileLoader: 'pure Python',
|
|
}
|
|
|
|
expected = loader
|
|
assert isinstance(expected, type), expected
|
|
expected = MODULE_KINDS[expected]
|
|
|
|
actual = module.__spec__.loader
|
|
if not isinstance(actual, type):
|
|
actual = type(actual)
|
|
actual = MODULE_KINDS[actual]
|
|
|
|
if actual != expected:
|
|
err = f'expected module to be {expected}, got {module.__spec__}'
|
|
if skip:
|
|
raise unittest.SkipTest(err)
|
|
raise Exception(err)
|
|
return module
|
|
|
|
def require_builtin(module, *, skip=False):
|
|
module = _require_loader(module, BuiltinImporter, skip)
|
|
assert module.__spec__.origin == 'built-in', module.__spec__
|
|
|
|
def require_extension(module, *, skip=False):
|
|
# Apple extensions must be distributed as frameworks. This requires
|
|
# a specialist loader.
|
|
if is_apple_mobile:
|
|
_require_loader(module, AppleFrameworkLoader, skip)
|
|
else:
|
|
_require_loader(module, ExtensionFileLoader, skip)
|
|
|
|
def require_frozen(module, *, skip=True):
|
|
module = _require_loader(module, FrozenImporter, skip)
|
|
assert module.__spec__.origin == 'frozen', module.__spec__
|
|
|
|
def require_pure_python(module, *, skip=False):
|
|
_require_loader(module, SourceFileLoader, skip)
|
|
|
|
def remove_files(name):
|
|
for f in (name + ".py",
|
|
name + ".pyc",
|
|
name + ".pyw",
|
|
name + "$py.class"):
|
|
unlink(f)
|
|
rmtree('__pycache__')
|
|
|
|
|
|
def no_rerun(reason):
|
|
"""Skip rerunning for a particular test.
|
|
|
|
WARNING: Use this decorator with care; skipping rerunning makes it
|
|
impossible to find reference leaks. Provide a clear reason for skipping the
|
|
test using the 'reason' parameter.
|
|
"""
|
|
def deco(func):
|
|
_has_run = False
|
|
def wrapper(self):
|
|
nonlocal _has_run
|
|
if _has_run:
|
|
self.skipTest(reason)
|
|
func(self)
|
|
_has_run = True
|
|
return wrapper
|
|
return deco
|
|
|
|
|
|
if _testsinglephase is not None:
|
|
def restore__testsinglephase(*, _orig=_testsinglephase):
|
|
# We started with the module imported and want to restore
|
|
# it to its nominal state.
|
|
sys.modules.pop('_testsinglephase', None)
|
|
_orig._clear_globals()
|
|
origin = _orig.__spec__.origin
|
|
_testinternalcapi.clear_extension('_testsinglephase', origin)
|
|
import _testsinglephase
|
|
|
|
|
|
def requires_singlephase_init(meth):
|
|
"""Decorator to skip if single-phase init modules are not supported."""
|
|
if not isinstance(meth, type):
|
|
def meth(self, _meth=meth):
|
|
try:
|
|
return _meth(self)
|
|
finally:
|
|
restore__testsinglephase()
|
|
meth = cpython_only(meth)
|
|
msg = "gh-117694: free-threaded build does not currently support single-phase init modules in sub-interpreters"
|
|
meth = requires_gil_enabled(msg)(meth)
|
|
return unittest.skipIf(_testsinglephase is None,
|
|
'test requires _testsinglephase module')(meth)
|
|
|
|
|
|
def requires_subinterpreters(meth):
|
|
"""Decorator to skip a test if subinterpreters are not supported."""
|
|
return unittest.skipIf(_interpreters is None,
|
|
'subinterpreters required')(meth)
|
|
|
|
|
|
class ModuleSnapshot(types.SimpleNamespace):
|
|
"""A representation of a module for testing.
|
|
|
|
Fields:
|
|
|
|
* id - the module's object ID
|
|
* module - the actual module or an adequate substitute
|
|
* __file__
|
|
* __spec__
|
|
* name
|
|
* origin
|
|
* ns - a copy (dict) of the module's __dict__ (or None)
|
|
* ns_id - the object ID of the module's __dict__
|
|
* cached - the sys.modules[mod.__spec__.name] entry (or None)
|
|
* cached_id - the object ID of the sys.modules entry (or None)
|
|
|
|
In cases where the value is not available (e.g. due to serialization),
|
|
the value will be None.
|
|
"""
|
|
_fields = tuple('id module ns ns_id cached cached_id'.split())
|
|
|
|
@classmethod
|
|
def from_module(cls, mod):
|
|
name = mod.__spec__.name
|
|
cached = sys.modules.get(name)
|
|
return cls(
|
|
id=id(mod),
|
|
module=mod,
|
|
ns=types.SimpleNamespace(**mod.__dict__),
|
|
ns_id=id(mod.__dict__),
|
|
cached=cached,
|
|
cached_id=id(cached),
|
|
)
|
|
|
|
SCRIPT = textwrap.dedent('''
|
|
{imports}
|
|
|
|
name = {name!r}
|
|
|
|
{prescript}
|
|
|
|
mod = {name}
|
|
|
|
{body}
|
|
|
|
{postscript}
|
|
''')
|
|
IMPORTS = textwrap.dedent('''
|
|
import sys
|
|
''').strip()
|
|
SCRIPT_BODY = textwrap.dedent('''
|
|
# Capture the snapshot data.
|
|
cached = sys.modules.get(name)
|
|
snapshot = dict(
|
|
id=id(mod),
|
|
module=dict(
|
|
__file__=mod.__file__,
|
|
__spec__=dict(
|
|
name=mod.__spec__.name,
|
|
origin=mod.__spec__.origin,
|
|
),
|
|
),
|
|
ns=None,
|
|
ns_id=id(mod.__dict__),
|
|
cached=None,
|
|
cached_id=id(cached) if cached else None,
|
|
)
|
|
''').strip()
|
|
CLEANUP_SCRIPT = textwrap.dedent('''
|
|
# Clean up the module.
|
|
sys.modules.pop(name, None)
|
|
''').strip()
|
|
|
|
@classmethod
|
|
def build_script(cls, name, *,
|
|
prescript=None,
|
|
import_first=False,
|
|
postscript=None,
|
|
postcleanup=False,
|
|
):
|
|
if postcleanup is True:
|
|
postcleanup = cls.CLEANUP_SCRIPT
|
|
elif isinstance(postcleanup, str):
|
|
postcleanup = textwrap.dedent(postcleanup).strip()
|
|
postcleanup = cls.CLEANUP_SCRIPT + os.linesep + postcleanup
|
|
else:
|
|
postcleanup = ''
|
|
prescript = textwrap.dedent(prescript).strip() if prescript else ''
|
|
postscript = textwrap.dedent(postscript).strip() if postscript else ''
|
|
|
|
if postcleanup:
|
|
if postscript:
|
|
postscript = postscript + os.linesep * 2 + postcleanup
|
|
else:
|
|
postscript = postcleanup
|
|
|
|
if import_first:
|
|
prescript += textwrap.dedent(f'''
|
|
|
|
# Now import the module.
|
|
assert name not in sys.modules
|
|
import {name}''')
|
|
|
|
return cls.SCRIPT.format(
|
|
imports=cls.IMPORTS.strip(),
|
|
name=name,
|
|
prescript=prescript.strip(),
|
|
body=cls.SCRIPT_BODY.strip(),
|
|
postscript=postscript,
|
|
)
|
|
|
|
@classmethod
|
|
def parse(cls, text):
|
|
raw = json.loads(text)
|
|
mod = raw['module']
|
|
mod['__spec__'] = types.SimpleNamespace(**mod['__spec__'])
|
|
raw['module'] = types.SimpleNamespace(**mod)
|
|
return cls(**raw)
|
|
|
|
@classmethod
|
|
def from_subinterp(cls, name, interpid=None, *, pipe=None, **script_kwds):
|
|
if pipe is not None:
|
|
return cls._from_subinterp(name, interpid, pipe, script_kwds)
|
|
pipe = os.pipe()
|
|
try:
|
|
return cls._from_subinterp(name, interpid, pipe, script_kwds)
|
|
finally:
|
|
r, w = pipe
|
|
os.close(r)
|
|
os.close(w)
|
|
|
|
@classmethod
|
|
def _from_subinterp(cls, name, interpid, pipe, script_kwargs):
|
|
r, w = pipe
|
|
|
|
# Build the script.
|
|
postscript = textwrap.dedent(f'''
|
|
# Send the result over the pipe.
|
|
import json
|
|
import os
|
|
os.write({w}, json.dumps(snapshot).encode())
|
|
|
|
''')
|
|
_postscript = script_kwargs.get('postscript')
|
|
if _postscript:
|
|
_postscript = textwrap.dedent(_postscript).lstrip()
|
|
postscript += _postscript
|
|
script_kwargs['postscript'] = postscript.strip()
|
|
script = cls.build_script(name, **script_kwargs)
|
|
|
|
# Run the script.
|
|
if interpid is None:
|
|
ret = run_in_subinterp(script)
|
|
if ret != 0:
|
|
raise AssertionError(f'{ret} != 0')
|
|
else:
|
|
_interpreters.run_string(interpid, script)
|
|
|
|
# Parse the results.
|
|
text = os.read(r, 1000)
|
|
return cls.parse(text.decode())
|
|
|
|
|
|
class ImportTests(unittest.TestCase):
|
|
|
|
def setUp(self):
|
|
remove_files(TESTFN)
|
|
importlib.invalidate_caches()
|
|
|
|
def tearDown(self):
|
|
unload(TESTFN)
|
|
|
|
def test_import_raises_ModuleNotFoundError(self):
|
|
with self.assertRaises(ModuleNotFoundError):
|
|
import something_that_should_not_exist_anywhere
|
|
|
|
def test_from_import_missing_module_raises_ModuleNotFoundError(self):
|
|
with self.assertRaises(ModuleNotFoundError):
|
|
from something_that_should_not_exist_anywhere import blah
|
|
|
|
def test_from_import_missing_attr_raises_ImportError(self):
|
|
with self.assertRaises(ImportError):
|
|
from importlib import something_that_should_not_exist_anywhere
|
|
|
|
def test_from_import_missing_attr_has_name_and_path(self):
|
|
with CleanImport('os'):
|
|
import os
|
|
with self.assertRaises(ImportError) as cm:
|
|
from os import i_dont_exist
|
|
self.assertEqual(cm.exception.name, 'os')
|
|
self.assertEqual(cm.exception.path, os.__file__)
|
|
self.assertRegex(str(cm.exception), r"cannot import name 'i_dont_exist' from 'os' \(.*os.py\)")
|
|
|
|
@cpython_only
|
|
def test_from_import_missing_attr_has_name_and_so_path(self):
|
|
_testcapi = import_module("_testcapi")
|
|
with self.assertRaises(ImportError) as cm:
|
|
from _testcapi import i_dont_exist
|
|
self.assertEqual(cm.exception.name, '_testcapi')
|
|
if hasattr(_testcapi, "__file__"):
|
|
self.assertEqual(cm.exception.path, _testcapi.__file__)
|
|
self.assertRegex(
|
|
str(cm.exception),
|
|
r"cannot import name 'i_dont_exist' from '_testcapi' \(.*\.(so|fwork|pyd)\)"
|
|
)
|
|
else:
|
|
self.assertEqual(
|
|
str(cm.exception),
|
|
"cannot import name 'i_dont_exist' from '_testcapi' (unknown location)"
|
|
)
|
|
|
|
def test_from_import_missing_attr_has_name(self):
|
|
with self.assertRaises(ImportError) as cm:
|
|
# _warning has no path as it's a built-in module.
|
|
from _warning import i_dont_exist
|
|
self.assertEqual(cm.exception.name, '_warning')
|
|
self.assertIsNone(cm.exception.path)
|
|
|
|
def test_from_import_missing_attr_path_is_canonical(self):
|
|
with self.assertRaises(ImportError) as cm:
|
|
from os.path import i_dont_exist
|
|
self.assertIn(cm.exception.name, {'posixpath', 'ntpath'})
|
|
self.assertIsNotNone(cm.exception)
|
|
|
|
def test_from_import_star_invalid_type(self):
|
|
import re
|
|
with ready_to_import() as (name, path):
|
|
with open(path, 'w', encoding='utf-8') as f:
|
|
f.write("__all__ = [b'invalid_type']")
|
|
globals = {}
|
|
with self.assertRaisesRegex(
|
|
TypeError, f"{re.escape(name)}\\.__all__ must be str"
|
|
):
|
|
exec(f"from {name} import *", globals)
|
|
self.assertNotIn(b"invalid_type", globals)
|
|
with ready_to_import() as (name, path):
|
|
with open(path, 'w', encoding='utf-8') as f:
|
|
f.write("globals()[b'invalid_type'] = object()")
|
|
globals = {}
|
|
with self.assertRaisesRegex(
|
|
TypeError, f"{re.escape(name)}\\.__dict__ must be str"
|
|
):
|
|
exec(f"from {name} import *", globals)
|
|
self.assertNotIn(b"invalid_type", globals)
|
|
|
|
def test_case_sensitivity(self):
|
|
# Brief digression to test that import is case-sensitive: if we got
|
|
# this far, we know for sure that "random" exists.
|
|
with self.assertRaises(ImportError):
|
|
import RAnDoM
|
|
|
|
def test_double_const(self):
|
|
# Importing double_const checks that float constants
|
|
# serialiazed by marshal as PYC files don't lose precision
|
|
# (SF bug 422177).
|
|
from test.test_import.data import double_const
|
|
unload('test.test_import.data.double_const')
|
|
from test.test_import.data import double_const
|
|
|
|
def test_import(self):
|
|
def test_with_extension(ext):
|
|
# The extension is normally ".py", perhaps ".pyw".
|
|
source = TESTFN + ext
|
|
pyc = TESTFN + ".pyc"
|
|
|
|
with open(source, "w", encoding='utf-8') as f:
|
|
print("# This tests Python's ability to import a",
|
|
ext, "file.", file=f)
|
|
a = random.randrange(1000)
|
|
b = random.randrange(1000)
|
|
print("a =", a, file=f)
|
|
print("b =", b, file=f)
|
|
|
|
if TESTFN in sys.modules:
|
|
del sys.modules[TESTFN]
|
|
importlib.invalidate_caches()
|
|
try:
|
|
try:
|
|
mod = __import__(TESTFN)
|
|
except ImportError as err:
|
|
self.fail("import from %s failed: %s" % (ext, err))
|
|
|
|
self.assertEqual(mod.a, a,
|
|
"module loaded (%s) but contents invalid" % mod)
|
|
self.assertEqual(mod.b, b,
|
|
"module loaded (%s) but contents invalid" % mod)
|
|
finally:
|
|
forget(TESTFN)
|
|
unlink(source)
|
|
unlink(pyc)
|
|
|
|
sys.path.insert(0, os.curdir)
|
|
try:
|
|
test_with_extension(".py")
|
|
if sys.platform.startswith("win"):
|
|
for ext in [".PY", ".Py", ".pY", ".pyw", ".PYW", ".pYw"]:
|
|
test_with_extension(ext)
|
|
finally:
|
|
del sys.path[0]
|
|
|
|
def test_module_with_large_stack(self, module='longlist'):
|
|
# Regression test for http://bugs.python.org/issue561858.
|
|
filename = module + '.py'
|
|
|
|
# Create a file with a list of 65000 elements.
|
|
with open(filename, 'w', encoding='utf-8') as f:
|
|
f.write('d = [\n')
|
|
for i in range(65000):
|
|
f.write('"",\n')
|
|
f.write(']')
|
|
|
|
try:
|
|
# Compile & remove .py file; we only need .pyc.
|
|
# Bytecode must be relocated from the PEP 3147 bytecode-only location.
|
|
py_compile.compile(filename)
|
|
finally:
|
|
unlink(filename)
|
|
|
|
# Need to be able to load from current dir.
|
|
sys.path.append('')
|
|
importlib.invalidate_caches()
|
|
|
|
namespace = {}
|
|
try:
|
|
make_legacy_pyc(filename)
|
|
# This used to crash.
|
|
exec('import ' + module, None, namespace)
|
|
finally:
|
|
# Cleanup.
|
|
del sys.path[-1]
|
|
unlink(filename + 'c')
|
|
unlink(filename + 'o')
|
|
|
|
# Remove references to the module (unload the module)
|
|
namespace.clear()
|
|
try:
|
|
del sys.modules[module]
|
|
except KeyError:
|
|
pass
|
|
|
|
def test_failing_import_sticks(self):
|
|
source = TESTFN + ".py"
|
|
with open(source, "w", encoding='utf-8') as f:
|
|
print("a = 1/0", file=f)
|
|
|
|
# New in 2.4, we shouldn't be able to import that no matter how often
|
|
# we try.
|
|
sys.path.insert(0, os.curdir)
|
|
importlib.invalidate_caches()
|
|
if TESTFN in sys.modules:
|
|
del sys.modules[TESTFN]
|
|
try:
|
|
for i in [1, 2, 3]:
|
|
self.assertRaises(ZeroDivisionError, __import__, TESTFN)
|
|
self.assertNotIn(TESTFN, sys.modules,
|
|
"damaged module in sys.modules on %i try" % i)
|
|
finally:
|
|
del sys.path[0]
|
|
remove_files(TESTFN)
|
|
|
|
def test_import_name_binding(self):
|
|
# import x.y.z binds x in the current namespace
|
|
import test as x
|
|
import test.support
|
|
self.assertIs(x, test, x.__name__)
|
|
self.assertTrue(hasattr(test.support, "__file__"))
|
|
|
|
# import x.y.z as w binds z as w
|
|
import test.support as y
|
|
self.assertIs(y, test.support, y.__name__)
|
|
|
|
def test_issue31286(self):
|
|
# import in a 'finally' block resulted in SystemError
|
|
try:
|
|
x = ...
|
|
finally:
|
|
import test.support.script_helper as x
|
|
|
|
# import in a 'while' loop resulted in stack overflow
|
|
i = 0
|
|
while i < 10:
|
|
import test.support.script_helper as x
|
|
i += 1
|
|
|
|
# import in a 'for' loop resulted in segmentation fault
|
|
for i in range(2):
|
|
import test.support.script_helper as x
|
|
|
|
def test_failing_reload(self):
|
|
# A failing reload should leave the module object in sys.modules.
|
|
source = TESTFN + os.extsep + "py"
|
|
with open(source, "w", encoding='utf-8') as f:
|
|
f.write("a = 1\nb=2\n")
|
|
|
|
sys.path.insert(0, os.curdir)
|
|
try:
|
|
mod = __import__(TESTFN)
|
|
self.assertIn(TESTFN, sys.modules)
|
|
self.assertEqual(mod.a, 1, "module has wrong attribute values")
|
|
self.assertEqual(mod.b, 2, "module has wrong attribute values")
|
|
|
|
# On WinXP, just replacing the .py file wasn't enough to
|
|
# convince reload() to reparse it. Maybe the timestamp didn't
|
|
# move enough. We force it to get reparsed by removing the
|
|
# compiled file too.
|
|
remove_files(TESTFN)
|
|
|
|
# Now damage the module.
|
|
with open(source, "w", encoding='utf-8') as f:
|
|
f.write("a = 10\nb=20//0\n")
|
|
|
|
self.assertRaises(ZeroDivisionError, importlib.reload, mod)
|
|
# But we still expect the module to be in sys.modules.
|
|
mod = sys.modules.get(TESTFN)
|
|
self.assertIsNotNone(mod, "expected module to be in sys.modules")
|
|
|
|
# We should have replaced a w/ 10, but the old b value should
|
|
# stick.
|
|
self.assertEqual(mod.a, 10, "module has wrong attribute values")
|
|
self.assertEqual(mod.b, 2, "module has wrong attribute values")
|
|
|
|
finally:
|
|
del sys.path[0]
|
|
remove_files(TESTFN)
|
|
unload(TESTFN)
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_file_to_source(self):
|
|
# check if __file__ points to the source file where available
|
|
source = TESTFN + ".py"
|
|
with open(source, "w", encoding='utf-8') as f:
|
|
f.write("test = None\n")
|
|
|
|
sys.path.insert(0, os.curdir)
|
|
try:
|
|
mod = __import__(TESTFN)
|
|
self.assertTrue(mod.__file__.endswith('.py'))
|
|
os.remove(source)
|
|
del sys.modules[TESTFN]
|
|
make_legacy_pyc(source)
|
|
importlib.invalidate_caches()
|
|
mod = __import__(TESTFN)
|
|
base, ext = os.path.splitext(mod.__file__)
|
|
self.assertEqual(ext, '.pyc')
|
|
finally:
|
|
del sys.path[0]
|
|
remove_files(TESTFN)
|
|
if TESTFN in sys.modules:
|
|
del sys.modules[TESTFN]
|
|
|
|
def test_import_by_filename(self):
|
|
path = os.path.abspath(TESTFN)
|
|
encoding = sys.getfilesystemencoding()
|
|
try:
|
|
path.encode(encoding)
|
|
except UnicodeEncodeError:
|
|
self.skipTest('path is not encodable to {}'.format(encoding))
|
|
with self.assertRaises(ImportError) as c:
|
|
__import__(path)
|
|
|
|
def test_import_in_del_does_not_crash(self):
|
|
# Issue 4236
|
|
testfn = script_helper.make_script('', TESTFN, textwrap.dedent("""\
|
|
import sys
|
|
class C:
|
|
def __del__(self):
|
|
import importlib
|
|
sys.argv.insert(0, C())
|
|
"""))
|
|
script_helper.assert_python_ok(testfn)
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_timestamp_overflow(self):
|
|
# A modification timestamp larger than 2**32 should not be a problem
|
|
# when importing a module (issue #11235).
|
|
sys.path.insert(0, os.curdir)
|
|
try:
|
|
source = TESTFN + ".py"
|
|
compiled = importlib.util.cache_from_source(source)
|
|
with open(source, 'w', encoding='utf-8') as f:
|
|
pass
|
|
try:
|
|
os.utime(source, (2 ** 33 - 5, 2 ** 33 - 5))
|
|
except OverflowError:
|
|
self.skipTest("cannot set modification time to large integer")
|
|
except OSError as e:
|
|
if e.errno not in (getattr(errno, 'EOVERFLOW', None),
|
|
getattr(errno, 'EINVAL', None)):
|
|
raise
|
|
self.skipTest("cannot set modification time to large integer ({})".format(e))
|
|
__import__(TESTFN)
|
|
# The pyc file was created.
|
|
os.stat(compiled)
|
|
finally:
|
|
del sys.path[0]
|
|
remove_files(TESTFN)
|
|
|
|
def test_bogus_fromlist(self):
|
|
try:
|
|
__import__('http', fromlist=['blah'])
|
|
except ImportError:
|
|
self.fail("fromlist must allow bogus names")
|
|
|
|
@cpython_only
|
|
def test_delete_builtins_import(self):
|
|
args = ["-c", "del __builtins__.__import__; import os"]
|
|
popen = script_helper.spawn_python(*args)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertIn(b"ImportError", stdout)
|
|
|
|
def test_from_import_message_for_nonexistent_module(self):
|
|
with self.assertRaisesRegex(ImportError, "^No module named 'bogus'"):
|
|
from bogus import foo
|
|
|
|
def test_from_import_message_for_existing_module(self):
|
|
with self.assertRaisesRegex(ImportError, "^cannot import name 'bogus'"):
|
|
from re import bogus
|
|
|
|
def test_from_import_AttributeError(self):
|
|
# Issue #24492: trying to import an attribute that raises an
|
|
# AttributeError should lead to an ImportError.
|
|
class AlwaysAttributeError:
|
|
def __getattr__(self, _):
|
|
raise AttributeError
|
|
|
|
module_name = 'test_from_import_AttributeError'
|
|
self.addCleanup(unload, module_name)
|
|
sys.modules[module_name] = AlwaysAttributeError()
|
|
with self.assertRaises(ImportError) as cm:
|
|
from test_from_import_AttributeError import does_not_exist
|
|
|
|
self.assertEqual(str(cm.exception),
|
|
"cannot import name 'does_not_exist' from '<unknown module name>' (unknown location)")
|
|
|
|
@cpython_only
|
|
def test_issue31492(self):
|
|
# There shouldn't be an assertion failure in case of failing to import
|
|
# from a module with a bad __name__ attribute, or in case of failing
|
|
# to access an attribute of such a module.
|
|
with swap_attr(os, '__name__', None):
|
|
with self.assertRaises(ImportError):
|
|
from os import does_not_exist
|
|
|
|
with self.assertRaises(AttributeError):
|
|
os.does_not_exist
|
|
|
|
@threading_helper.requires_working_threading()
|
|
def test_concurrency(self):
|
|
# bpo 38091: this is a hack to slow down the code that calls
|
|
# has_deadlock(); the logic was itself sometimes deadlocking.
|
|
def delay_has_deadlock(frame, event, arg):
|
|
if event == 'call' and frame.f_code.co_name == 'has_deadlock':
|
|
time.sleep(0.1)
|
|
|
|
sys.path.insert(0, os.path.join(os.path.dirname(__file__), 'data'))
|
|
try:
|
|
exc = None
|
|
def run():
|
|
sys.settrace(delay_has_deadlock)
|
|
event.wait()
|
|
try:
|
|
import package
|
|
except BaseException as e:
|
|
nonlocal exc
|
|
exc = e
|
|
sys.settrace(None)
|
|
|
|
for i in range(10):
|
|
event = threading.Event()
|
|
threads = [threading.Thread(target=run) for x in range(2)]
|
|
try:
|
|
with threading_helper.start_threads(threads, event.set):
|
|
time.sleep(0)
|
|
finally:
|
|
sys.modules.pop('package', None)
|
|
sys.modules.pop('package.submodule', None)
|
|
if exc is not None:
|
|
raise exc
|
|
finally:
|
|
del sys.path[0]
|
|
|
|
@unittest.skipUnless(sys.platform == "win32", "Windows-specific")
|
|
def test_dll_dependency_import(self):
|
|
from _winapi import GetModuleFileName
|
|
dllname = GetModuleFileName(sys.dllhandle)
|
|
pydname = importlib.util.find_spec("_sqlite3").origin
|
|
depname = os.path.join(
|
|
os.path.dirname(pydname),
|
|
"sqlite3{}.dll".format("_d" if "_d" in pydname else ""))
|
|
|
|
with os_helper.temp_dir() as tmp:
|
|
tmp2 = os.path.join(tmp, "DLLs")
|
|
os.mkdir(tmp2)
|
|
|
|
pyexe = os.path.join(tmp, os.path.basename(sys.executable))
|
|
shutil.copy(sys.executable, pyexe)
|
|
shutil.copy(dllname, tmp)
|
|
for f in glob.glob(os.path.join(glob.escape(sys.prefix), "vcruntime*.dll")):
|
|
shutil.copy(f, tmp)
|
|
|
|
shutil.copy(pydname, tmp2)
|
|
|
|
env = None
|
|
env = {k.upper(): os.environ[k] for k in os.environ}
|
|
env["PYTHONPATH"] = tmp2 + ";" + STDLIB_DIR
|
|
|
|
# Test 1: import with added DLL directory
|
|
subprocess.check_call([
|
|
pyexe, "-Sc", ";".join([
|
|
"import os",
|
|
"p = os.add_dll_directory({!r})".format(
|
|
os.path.dirname(depname)),
|
|
"import _sqlite3",
|
|
"p.close"
|
|
])],
|
|
stderr=subprocess.STDOUT,
|
|
env=env,
|
|
cwd=os.path.dirname(pyexe))
|
|
|
|
# Test 2: import with DLL adjacent to PYD
|
|
shutil.copy(depname, tmp2)
|
|
subprocess.check_call([pyexe, "-Sc", "import _sqlite3"],
|
|
stderr=subprocess.STDOUT,
|
|
env=env,
|
|
cwd=os.path.dirname(pyexe))
|
|
|
|
def test_issue105979(self):
|
|
# this used to crash
|
|
with self.assertRaises(ImportError) as cm:
|
|
_imp.get_frozen_object("x", b"6\'\xd5Cu\x12")
|
|
self.assertIn("Frozen object named 'x' is invalid",
|
|
str(cm.exception))
|
|
|
|
def test_script_shadowing_stdlib(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
with open(os.path.join(tmp, "fractions.py"), "w", encoding='utf-8') as f:
|
|
f.write("import fractions\nfractions.Fraction")
|
|
|
|
expected_error = (
|
|
rb"AttributeError: module 'fractions' has no attribute 'Fraction' "
|
|
rb"\(consider renaming '.*fractions.py' since it has the "
|
|
rb"same name as the standard library module named 'fractions' "
|
|
rb"and the import system gives it precedence\)"
|
|
)
|
|
|
|
popen = script_helper.spawn_python(os.path.join(tmp, "fractions.py"), cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-m', 'fractions', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-c', 'import fractions', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
# and there's no error at all when using -P
|
|
popen = script_helper.spawn_python('-P', 'fractions.py', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(stdout, b'')
|
|
|
|
tmp_child = os.path.join(tmp, "child")
|
|
os.mkdir(tmp_child)
|
|
|
|
# test the logic with different cwd
|
|
popen = script_helper.spawn_python(os.path.join(tmp, "fractions.py"), cwd=tmp_child)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-m', 'fractions', cwd=tmp_child)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(stdout, b'') # no error
|
|
|
|
popen = script_helper.spawn_python('-c', 'import fractions', cwd=tmp_child)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(stdout, b'') # no error
|
|
|
|
def test_package_shadowing_stdlib_module(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
os.mkdir(os.path.join(tmp, "fractions"))
|
|
with open(os.path.join(tmp, "fractions", "__init__.py"), "w", encoding='utf-8') as f:
|
|
f.write("shadowing_module = True")
|
|
with open(os.path.join(tmp, "main.py"), "w", encoding='utf-8') as f:
|
|
f.write("""
|
|
import fractions
|
|
fractions.shadowing_module
|
|
fractions.Fraction
|
|
""")
|
|
|
|
expected_error = (
|
|
rb"AttributeError: module 'fractions' has no attribute 'Fraction' "
|
|
rb"\(consider renaming '.*fractions.__init__.py' since it has the "
|
|
rb"same name as the standard library module named 'fractions' "
|
|
rb"and the import system gives it precedence\)"
|
|
)
|
|
|
|
popen = script_helper.spawn_python(os.path.join(tmp, "main.py"), cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-m', 'main', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
# and there's no shadowing at all when using -P
|
|
popen = script_helper.spawn_python('-P', 'main.py', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, b"module 'fractions' has no attribute 'shadowing_module'")
|
|
|
|
def test_script_shadowing_third_party(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
with open(os.path.join(tmp, "numpy.py"), "w", encoding='utf-8') as f:
|
|
f.write("import numpy\nnumpy.array")
|
|
|
|
expected_error = (
|
|
rb"AttributeError: module 'numpy' has no attribute 'array' "
|
|
rb"\(consider renaming '.*numpy.py' if it has the "
|
|
rb"same name as a third-party module you intended to import\)\s+\Z"
|
|
)
|
|
|
|
popen = script_helper.spawn_python(os.path.join(tmp, "numpy.py"))
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-m', 'numpy', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
popen = script_helper.spawn_python('-c', 'import numpy', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
def test_script_maybe_not_shadowing_third_party(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
with open(os.path.join(tmp, "numpy.py"), "w", encoding='utf-8') as f:
|
|
f.write("this_script_does_not_attempt_to_import_numpy = True")
|
|
|
|
expected_error = (
|
|
rb"AttributeError: module 'numpy' has no attribute 'attr'\s+\Z"
|
|
)
|
|
|
|
popen = script_helper.spawn_python('-c', 'import numpy; numpy.attr', cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
def test_script_shadowing_stdlib_edge_cases(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
with open(os.path.join(tmp, "fractions.py"), "w", encoding='utf-8') as f:
|
|
f.write("shadowing_module = True")
|
|
with open(os.path.join(tmp, "main.py"), "w", encoding='utf-8') as f:
|
|
f.write("""
|
|
import fractions
|
|
fractions.shadowing_module
|
|
class substr(str):
|
|
__hash__ = None
|
|
fractions.__name__ = substr('fractions')
|
|
try:
|
|
fractions.Fraction
|
|
except TypeError as e:
|
|
print(str(e))
|
|
""")
|
|
|
|
popen = script_helper.spawn_python("main.py", cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(stdout.rstrip(), b"unhashable type: 'substr'")
|
|
|
|
with open(os.path.join(tmp, "main.py"), "w", encoding='utf-8') as f:
|
|
f.write("""
|
|
import fractions
|
|
fractions.shadowing_module
|
|
|
|
import sys
|
|
sys.stdlib_module_names = None
|
|
try:
|
|
fractions.Fraction
|
|
except AttributeError as e:
|
|
print(str(e))
|
|
|
|
del sys.stdlib_module_names
|
|
try:
|
|
fractions.Fraction
|
|
except AttributeError as e:
|
|
print(str(e))
|
|
|
|
sys.path = [0]
|
|
try:
|
|
fractions.Fraction
|
|
except AttributeError as e:
|
|
print(str(e))
|
|
""")
|
|
|
|
popen = script_helper.spawn_python("main.py", cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(
|
|
stdout.splitlines(),
|
|
[
|
|
b"module 'fractions' has no attribute 'Fraction'",
|
|
b"module 'fractions' has no attribute 'Fraction'",
|
|
b"module 'fractions' has no attribute 'Fraction'",
|
|
],
|
|
)
|
|
|
|
with open(os.path.join(tmp, "main.py"), "w", encoding='utf-8') as f:
|
|
f.write("""
|
|
import fractions
|
|
fractions.shadowing_module
|
|
del fractions.__spec__.origin
|
|
try:
|
|
fractions.Fraction
|
|
except AttributeError as e:
|
|
print(str(e))
|
|
|
|
fractions.__spec__.origin = 0
|
|
try:
|
|
fractions.Fraction
|
|
except AttributeError as e:
|
|
print(str(e))
|
|
""")
|
|
|
|
popen = script_helper.spawn_python("main.py", cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertEqual(
|
|
stdout.splitlines(),
|
|
[
|
|
b"module 'fractions' has no attribute 'Fraction'",
|
|
b"module 'fractions' has no attribute 'Fraction'"
|
|
],
|
|
)
|
|
|
|
def test_script_shadowing_stdlib_sys_path_modification(self):
|
|
with os_helper.temp_dir() as tmp:
|
|
with open(os.path.join(tmp, "fractions.py"), "w", encoding='utf-8') as f:
|
|
f.write("shadowing_module = True")
|
|
|
|
expected_error = (
|
|
rb"AttributeError: module 'fractions' has no attribute 'Fraction' "
|
|
rb"\(consider renaming '.*fractions.py' since it has the "
|
|
rb"same name as the standard library module named 'fractions' "
|
|
rb"and the import system gives it precedence\)"
|
|
)
|
|
|
|
with open(os.path.join(tmp, "main.py"), "w", encoding='utf-8') as f:
|
|
f.write("""
|
|
import sys
|
|
sys.path.insert(0, "this_folder_does_not_exist")
|
|
import fractions
|
|
fractions.Fraction
|
|
""")
|
|
|
|
popen = script_helper.spawn_python("main.py", cwd=tmp)
|
|
stdout, stderr = popen.communicate()
|
|
self.assertRegex(stdout, expected_error)
|
|
|
|
|
|
@skip_if_dont_write_bytecode
|
|
class FilePermissionTests(unittest.TestCase):
|
|
# tests for file mode on cached .pyc files
|
|
|
|
@unittest.skipUnless(os.name == 'posix',
|
|
"test meaningful only on posix systems")
|
|
@unittest.skipIf(
|
|
is_emscripten or is_wasi,
|
|
"Emscripten's/WASI's umask is a stub."
|
|
)
|
|
def test_creation_mode(self):
|
|
mask = 0o022
|
|
with temp_umask(mask), ready_to_import() as (name, path):
|
|
cached_path = importlib.util.cache_from_source(path)
|
|
module = __import__(name)
|
|
if not os.path.exists(cached_path):
|
|
self.fail("__import__ did not result in creation of "
|
|
"a .pyc file")
|
|
stat_info = os.stat(cached_path)
|
|
|
|
# Check that the umask is respected, and the executable bits
|
|
# aren't set.
|
|
self.assertEqual(oct(stat.S_IMODE(stat_info.st_mode)),
|
|
oct(0o666 & ~mask))
|
|
|
|
@unittest.skipUnless(os.name == 'posix',
|
|
"test meaningful only on posix systems")
|
|
@os_helper.skip_unless_working_chmod
|
|
def test_cached_mode_issue_2051(self):
|
|
# permissions of .pyc should match those of .py, regardless of mask
|
|
mode = 0o600
|
|
with temp_umask(0o022), ready_to_import() as (name, path):
|
|
cached_path = importlib.util.cache_from_source(path)
|
|
os.chmod(path, mode)
|
|
__import__(name)
|
|
if not os.path.exists(cached_path):
|
|
self.fail("__import__ did not result in creation of "
|
|
"a .pyc file")
|
|
stat_info = os.stat(cached_path)
|
|
|
|
self.assertEqual(oct(stat.S_IMODE(stat_info.st_mode)), oct(mode))
|
|
|
|
@unittest.skipUnless(os.name == 'posix',
|
|
"test meaningful only on posix systems")
|
|
@os_helper.skip_unless_working_chmod
|
|
def test_cached_readonly(self):
|
|
mode = 0o400
|
|
with temp_umask(0o022), ready_to_import() as (name, path):
|
|
cached_path = importlib.util.cache_from_source(path)
|
|
os.chmod(path, mode)
|
|
__import__(name)
|
|
if not os.path.exists(cached_path):
|
|
self.fail("__import__ did not result in creation of "
|
|
"a .pyc file")
|
|
stat_info = os.stat(cached_path)
|
|
|
|
expected = mode | 0o200 # Account for fix for issue #6074
|
|
self.assertEqual(oct(stat.S_IMODE(stat_info.st_mode)), oct(expected))
|
|
|
|
def test_pyc_always_writable(self):
|
|
# Initially read-only .pyc files on Windows used to cause problems
|
|
# with later updates, see issue #6074 for details
|
|
with ready_to_import() as (name, path):
|
|
# Write a Python file, make it read-only and import it
|
|
with open(path, 'w', encoding='utf-8') as f:
|
|
f.write("x = 'original'\n")
|
|
# Tweak the mtime of the source to ensure pyc gets updated later
|
|
s = os.stat(path)
|
|
os.utime(path, (s.st_atime, s.st_mtime-100000000))
|
|
os.chmod(path, 0o400)
|
|
m = __import__(name)
|
|
self.assertEqual(m.x, 'original')
|
|
# Change the file and then reimport it
|
|
os.chmod(path, 0o600)
|
|
with open(path, 'w', encoding='utf-8') as f:
|
|
f.write("x = 'rewritten'\n")
|
|
unload(name)
|
|
importlib.invalidate_caches()
|
|
m = __import__(name)
|
|
self.assertEqual(m.x, 'rewritten')
|
|
# Now delete the source file and check the pyc was rewritten
|
|
unlink(path)
|
|
unload(name)
|
|
importlib.invalidate_caches()
|
|
bytecode_only = path + "c"
|
|
os.rename(importlib.util.cache_from_source(path), bytecode_only)
|
|
m = __import__(name)
|
|
self.assertEqual(m.x, 'rewritten')
|
|
|
|
|
|
class PycRewritingTests(unittest.TestCase):
|
|
# Test that the `co_filename` attribute on code objects always points
|
|
# to the right file, even when various things happen (e.g. both the .py
|
|
# and the .pyc file are renamed).
|
|
|
|
module_name = "unlikely_module_name"
|
|
module_source = """
|
|
import sys
|
|
code_filename = sys._getframe().f_code.co_filename
|
|
module_filename = __file__
|
|
constant = 1
|
|
def func():
|
|
pass
|
|
func_filename = func.__code__.co_filename
|
|
"""
|
|
dir_name = os.path.abspath(TESTFN)
|
|
file_name = os.path.join(dir_name, module_name) + os.extsep + "py"
|
|
compiled_name = importlib.util.cache_from_source(file_name)
|
|
|
|
def setUp(self):
|
|
self.sys_path = sys.path[:]
|
|
self.orig_module = sys.modules.pop(self.module_name, None)
|
|
os.mkdir(self.dir_name)
|
|
with open(self.file_name, "w", encoding='utf-8') as f:
|
|
f.write(self.module_source)
|
|
sys.path.insert(0, self.dir_name)
|
|
importlib.invalidate_caches()
|
|
|
|
def tearDown(self):
|
|
sys.path[:] = self.sys_path
|
|
if self.orig_module is not None:
|
|
sys.modules[self.module_name] = self.orig_module
|
|
else:
|
|
unload(self.module_name)
|
|
unlink(self.file_name)
|
|
unlink(self.compiled_name)
|
|
rmtree(self.dir_name)
|
|
|
|
def import_module(self):
|
|
ns = globals()
|
|
__import__(self.module_name, ns, ns)
|
|
return sys.modules[self.module_name]
|
|
|
|
def test_basics(self):
|
|
mod = self.import_module()
|
|
self.assertEqual(mod.module_filename, self.file_name)
|
|
self.assertEqual(mod.code_filename, self.file_name)
|
|
self.assertEqual(mod.func_filename, self.file_name)
|
|
del sys.modules[self.module_name]
|
|
mod = self.import_module()
|
|
self.assertEqual(mod.module_filename, self.file_name)
|
|
self.assertEqual(mod.code_filename, self.file_name)
|
|
self.assertEqual(mod.func_filename, self.file_name)
|
|
|
|
def test_incorrect_code_name(self):
|
|
py_compile.compile(self.file_name, dfile="another_module.py")
|
|
mod = self.import_module()
|
|
self.assertEqual(mod.module_filename, self.file_name)
|
|
self.assertEqual(mod.code_filename, self.file_name)
|
|
self.assertEqual(mod.func_filename, self.file_name)
|
|
|
|
def test_module_without_source(self):
|
|
target = "another_module.py"
|
|
py_compile.compile(self.file_name, dfile=target)
|
|
os.remove(self.file_name)
|
|
pyc_file = make_legacy_pyc(self.file_name)
|
|
importlib.invalidate_caches()
|
|
mod = self.import_module()
|
|
self.assertEqual(mod.module_filename, pyc_file)
|
|
self.assertEqual(mod.code_filename, target)
|
|
self.assertEqual(mod.func_filename, target)
|
|
|
|
def test_foreign_code(self):
|
|
py_compile.compile(self.file_name)
|
|
with open(self.compiled_name, "rb") as f:
|
|
header = f.read(16)
|
|
code = marshal.load(f)
|
|
constants = list(code.co_consts)
|
|
foreign_code = importlib.import_module.__code__
|
|
pos = constants.index(1)
|
|
constants[pos] = foreign_code
|
|
code = code.replace(co_consts=tuple(constants))
|
|
with open(self.compiled_name, "wb") as f:
|
|
f.write(header)
|
|
marshal.dump(code, f)
|
|
mod = self.import_module()
|
|
self.assertEqual(mod.constant.co_filename, foreign_code.co_filename)
|
|
|
|
|
|
class PathsTests(unittest.TestCase):
|
|
SAMPLES = ('test', 'test\u00e4\u00f6\u00fc\u00df', 'test\u00e9\u00e8',
|
|
'test\u00b0\u00b3\u00b2')
|
|
path = TESTFN
|
|
|
|
def setUp(self):
|
|
os.mkdir(self.path)
|
|
self.syspath = sys.path[:]
|
|
|
|
def tearDown(self):
|
|
rmtree(self.path)
|
|
sys.path[:] = self.syspath
|
|
|
|
# Regression test for http://bugs.python.org/issue1293.
|
|
def test_trailing_slash(self):
|
|
with open(os.path.join(self.path, 'test_trailing_slash.py'),
|
|
'w', encoding='utf-8') as f:
|
|
f.write("testdata = 'test_trailing_slash'")
|
|
sys.path.append(self.path+'/')
|
|
mod = __import__("test_trailing_slash")
|
|
self.assertEqual(mod.testdata, 'test_trailing_slash')
|
|
unload("test_trailing_slash")
|
|
|
|
# Regression test for http://bugs.python.org/issue3677.
|
|
@unittest.skipUnless(sys.platform == 'win32', 'Windows-specific')
|
|
def test_UNC_path(self):
|
|
with open(os.path.join(self.path, 'test_unc_path.py'), 'w') as f:
|
|
f.write("testdata = 'test_unc_path'")
|
|
importlib.invalidate_caches()
|
|
# Create the UNC path, like \\myhost\c$\foo\bar.
|
|
path = os.path.abspath(self.path)
|
|
import socket
|
|
hn = socket.gethostname()
|
|
drive = path[0]
|
|
unc = "\\\\%s\\%s$"%(hn, drive)
|
|
unc += path[2:]
|
|
try:
|
|
os.listdir(unc)
|
|
except OSError as e:
|
|
if e.errno in (errno.EPERM, errno.EACCES, errno.ENOENT):
|
|
# See issue #15338
|
|
self.skipTest("cannot access administrative share %r" % (unc,))
|
|
raise
|
|
sys.path.insert(0, unc)
|
|
try:
|
|
mod = __import__("test_unc_path")
|
|
except ImportError as e:
|
|
self.fail("could not import 'test_unc_path' from %r: %r"
|
|
% (unc, e))
|
|
self.assertEqual(mod.testdata, 'test_unc_path')
|
|
self.assertTrue(mod.__file__.startswith(unc), mod.__file__)
|
|
unload("test_unc_path")
|
|
|
|
|
|
class RelativeImportTests(unittest.TestCase):
|
|
|
|
def tearDown(self):
|
|
unload("test.relimport")
|
|
setUp = tearDown
|
|
|
|
def test_relimport_star(self):
|
|
# This will import * from .test_import.
|
|
from .. import relimport
|
|
self.assertTrue(hasattr(relimport, "RelativeImportTests"))
|
|
|
|
def test_issue3221(self):
|
|
# Note for mergers: the 'absolute' tests from the 2.x branch
|
|
# are missing in Py3k because implicit relative imports are
|
|
# a thing of the past
|
|
#
|
|
# Regression test for http://bugs.python.org/issue3221.
|
|
def check_relative():
|
|
exec("from . import relimport", ns)
|
|
|
|
# Check relative import OK with __package__ and __name__ correct
|
|
ns = dict(__package__='test', __name__='test.notarealmodule')
|
|
check_relative()
|
|
|
|
# Check relative import OK with only __name__ wrong
|
|
ns = dict(__package__='test', __name__='notarealpkg.notarealmodule')
|
|
check_relative()
|
|
|
|
# Check relative import fails with only __package__ wrong
|
|
ns = dict(__package__='foo', __name__='test.notarealmodule')
|
|
self.assertRaises(ModuleNotFoundError, check_relative)
|
|
|
|
# Check relative import fails with __package__ and __name__ wrong
|
|
ns = dict(__package__='foo', __name__='notarealpkg.notarealmodule')
|
|
self.assertRaises(ModuleNotFoundError, check_relative)
|
|
|
|
# Check relative import fails with package set to a non-string
|
|
ns = dict(__package__=object())
|
|
self.assertRaises(TypeError, check_relative)
|
|
|
|
def test_parentless_import_shadowed_by_global(self):
|
|
# Test as if this were done from the REPL where this error most commonly occurs (bpo-37409).
|
|
script_helper.assert_python_failure('-W', 'ignore', '-c',
|
|
"foo = 1; from . import foo")
|
|
|
|
def test_absolute_import_without_future(self):
|
|
# If explicit relative import syntax is used, then do not try
|
|
# to perform an absolute import in the face of failure.
|
|
# Issue #7902.
|
|
with self.assertRaises(ImportError):
|
|
from .os import sep
|
|
self.fail("explicit relative import triggered an "
|
|
"implicit absolute import")
|
|
|
|
def test_import_from_non_package(self):
|
|
path = os.path.join(os.path.dirname(__file__), 'data', 'package2')
|
|
with uncache('submodule1', 'submodule2'), DirsOnSysPath(path):
|
|
with self.assertRaises(ImportError):
|
|
import submodule1
|
|
self.assertNotIn('submodule1', sys.modules)
|
|
self.assertNotIn('submodule2', sys.modules)
|
|
|
|
def test_import_from_unloaded_package(self):
|
|
with uncache('package2', 'package2.submodule1', 'package2.submodule2'), \
|
|
DirsOnSysPath(os.path.join(os.path.dirname(__file__), 'data')):
|
|
import package2.submodule1
|
|
package2.submodule1.submodule2
|
|
|
|
def test_rebinding(self):
|
|
# The same data is also used for testing pkgutil.resolve_name()
|
|
# in test_pkgutil and mock.patch in test_unittest.
|
|
path = os.path.join(os.path.dirname(__file__), 'data')
|
|
with uncache('package3', 'package3.submodule'), DirsOnSysPath(path):
|
|
from package3 import submodule
|
|
self.assertEqual(submodule.attr, 'rebound')
|
|
import package3.submodule as submodule
|
|
self.assertEqual(submodule.attr, 'rebound')
|
|
with uncache('package3', 'package3.submodule'), DirsOnSysPath(path):
|
|
import package3.submodule as submodule
|
|
self.assertEqual(submodule.attr, 'rebound')
|
|
from package3 import submodule
|
|
self.assertEqual(submodule.attr, 'rebound')
|
|
|
|
def test_rebinding2(self):
|
|
path = os.path.join(os.path.dirname(__file__), 'data')
|
|
with uncache('package4', 'package4.submodule'), DirsOnSysPath(path):
|
|
import package4.submodule as submodule
|
|
self.assertEqual(submodule.attr, 'submodule')
|
|
from package4 import submodule
|
|
self.assertEqual(submodule.attr, 'submodule')
|
|
with uncache('package4', 'package4.submodule'), DirsOnSysPath(path):
|
|
from package4 import submodule
|
|
self.assertEqual(submodule.attr, 'origin')
|
|
import package4.submodule as submodule
|
|
self.assertEqual(submodule.attr, 'submodule')
|
|
|
|
|
|
class OverridingImportBuiltinTests(unittest.TestCase):
|
|
def test_override_builtin(self):
|
|
# Test that overriding builtins.__import__ can bypass sys.modules.
|
|
import os
|
|
|
|
def foo():
|
|
import os
|
|
return os
|
|
self.assertEqual(foo(), os) # Quick sanity check.
|
|
|
|
with swap_attr(builtins, "__import__", lambda *x: 5):
|
|
self.assertEqual(foo(), 5)
|
|
|
|
# Test what happens when we shadow __import__ in globals(); this
|
|
# currently does not impact the import process, but if this changes,
|
|
# other code will need to change, so keep this test as a tripwire.
|
|
with swap_item(globals(), "__import__", lambda *x: 5):
|
|
self.assertEqual(foo(), os)
|
|
|
|
|
|
class PycacheTests(unittest.TestCase):
|
|
# Test the various PEP 3147/488-related behaviors.
|
|
|
|
def _clean(self):
|
|
forget(TESTFN)
|
|
rmtree('__pycache__')
|
|
unlink(self.source)
|
|
|
|
def setUp(self):
|
|
self.source = TESTFN + '.py'
|
|
self._clean()
|
|
with open(self.source, 'w', encoding='utf-8') as fp:
|
|
print('# This is a test file written by test_import.py', file=fp)
|
|
sys.path.insert(0, os.curdir)
|
|
importlib.invalidate_caches()
|
|
|
|
def tearDown(self):
|
|
assert sys.path[0] == os.curdir, 'Unexpected sys.path[0]'
|
|
del sys.path[0]
|
|
self._clean()
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_import_pyc_path(self):
|
|
self.assertFalse(os.path.exists('__pycache__'))
|
|
__import__(TESTFN)
|
|
self.assertTrue(os.path.exists('__pycache__'))
|
|
pyc_path = importlib.util.cache_from_source(self.source)
|
|
self.assertTrue(os.path.exists(pyc_path),
|
|
'bytecode file {!r} for {!r} does not '
|
|
'exist'.format(pyc_path, TESTFN))
|
|
|
|
@unittest.skipUnless(os.name == 'posix',
|
|
"test meaningful only on posix systems")
|
|
@skip_if_dont_write_bytecode
|
|
@os_helper.skip_unless_working_chmod
|
|
@os_helper.skip_if_dac_override
|
|
@unittest.skipIf(is_emscripten, "umask is a stub")
|
|
def test_unwritable_directory(self):
|
|
# When the umask causes the new __pycache__ directory to be
|
|
# unwritable, the import still succeeds but no .pyc file is written.
|
|
with temp_umask(0o222):
|
|
__import__(TESTFN)
|
|
self.assertTrue(os.path.exists('__pycache__'))
|
|
pyc_path = importlib.util.cache_from_source(self.source)
|
|
self.assertFalse(os.path.exists(pyc_path),
|
|
'bytecode file {!r} for {!r} '
|
|
'exists'.format(pyc_path, TESTFN))
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_missing_source(self):
|
|
# With PEP 3147 cache layout, removing the source but leaving the pyc
|
|
# file does not satisfy the import.
|
|
__import__(TESTFN)
|
|
pyc_file = importlib.util.cache_from_source(self.source)
|
|
self.assertTrue(os.path.exists(pyc_file))
|
|
os.remove(self.source)
|
|
forget(TESTFN)
|
|
importlib.invalidate_caches()
|
|
self.assertRaises(ImportError, __import__, TESTFN)
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_missing_source_legacy(self):
|
|
# Like test_missing_source() except that for backward compatibility,
|
|
# when the pyc file lives where the py file would have been (and named
|
|
# without the tag), it is importable. The __file__ of the imported
|
|
# module is the pyc location.
|
|
__import__(TESTFN)
|
|
# pyc_file gets removed in _clean() via tearDown().
|
|
pyc_file = make_legacy_pyc(self.source)
|
|
os.remove(self.source)
|
|
unload(TESTFN)
|
|
importlib.invalidate_caches()
|
|
m = __import__(TESTFN)
|
|
try:
|
|
self.assertEqual(m.__file__,
|
|
os.path.join(os.getcwd(), os.path.relpath(pyc_file)))
|
|
finally:
|
|
os.remove(pyc_file)
|
|
|
|
def test___cached__(self):
|
|
# Modules now also have an __cached__ that points to the pyc file.
|
|
m = __import__(TESTFN)
|
|
pyc_file = importlib.util.cache_from_source(TESTFN + '.py')
|
|
self.assertEqual(m.__cached__, os.path.join(os.getcwd(), pyc_file))
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test___cached___legacy_pyc(self):
|
|
# Like test___cached__() except that for backward compatibility,
|
|
# when the pyc file lives where the py file would have been (and named
|
|
# without the tag), it is importable. The __cached__ of the imported
|
|
# module is the pyc location.
|
|
__import__(TESTFN)
|
|
# pyc_file gets removed in _clean() via tearDown().
|
|
pyc_file = make_legacy_pyc(self.source)
|
|
os.remove(self.source)
|
|
unload(TESTFN)
|
|
importlib.invalidate_caches()
|
|
m = __import__(TESTFN)
|
|
self.assertEqual(m.__cached__,
|
|
os.path.join(os.getcwd(), os.path.relpath(pyc_file)))
|
|
|
|
@skip_if_dont_write_bytecode
|
|
def test_package___cached__(self):
|
|
# Like test___cached__ but for packages.
|
|
def cleanup():
|
|
rmtree('pep3147')
|
|
unload('pep3147.foo')
|
|
unload('pep3147')
|
|
os.mkdir('pep3147')
|
|
self.addCleanup(cleanup)
|
|
# Touch the __init__.py
|
|
with open(os.path.join('pep3147', '__init__.py'), 'wb'):
|
|
pass
|
|
with open(os.path.join('pep3147', 'foo.py'), 'wb'):
|
|
pass
|
|
importlib.invalidate_caches()
|
|
m = __import__('pep3147.foo')
|
|
init_pyc = importlib.util.cache_from_source(
|
|
os.path.join('pep3147', '__init__.py'))
|
|
self.assertEqual(m.__cached__, os.path.join(os.getcwd(), init_pyc))
|
|
foo_pyc = importlib.util.cache_from_source(os.path.join('pep3147', 'foo.py'))
|
|
self.assertEqual(sys.modules['pep3147.foo'].__cached__,
|
|
os.path.join(os.getcwd(), foo_pyc))
|
|
|
|
def test_package___cached___from_pyc(self):
|
|
# Like test___cached__ but ensuring __cached__ when imported from a
|
|
# PEP 3147 pyc file.
|
|
def cleanup():
|
|
rmtree('pep3147')
|
|
unload('pep3147.foo')
|
|
unload('pep3147')
|
|
os.mkdir('pep3147')
|
|
self.addCleanup(cleanup)
|
|
# Touch the __init__.py
|
|
with open(os.path.join('pep3147', '__init__.py'), 'wb'):
|
|
pass
|
|
with open(os.path.join('pep3147', 'foo.py'), 'wb'):
|
|
pass
|
|
importlib.invalidate_caches()
|
|
m = __import__('pep3147.foo')
|
|
unload('pep3147.foo')
|
|
unload('pep3147')
|
|
importlib.invalidate_caches()
|
|
m = __import__('pep3147.foo')
|
|
init_pyc = importlib.util.cache_from_source(
|
|
os.path.join('pep3147', '__init__.py'))
|
|
self.assertEqual(m.__cached__, os.path.join(os.getcwd(), init_pyc))
|
|
foo_pyc = importlib.util.cache_from_source(os.path.join('pep3147', 'foo.py'))
|
|
self.assertEqual(sys.modules['pep3147.foo'].__cached__,
|
|
os.path.join(os.getcwd(), foo_pyc))
|
|
|
|
def test_recompute_pyc_same_second(self):
|
|
# Even when the source file doesn't change timestamp, a change in
|
|
# source size is enough to trigger recomputation of the pyc file.
|
|
__import__(TESTFN)
|
|
unload(TESTFN)
|
|
with open(self.source, 'a', encoding='utf-8') as fp:
|
|
print("x = 5", file=fp)
|
|
m = __import__(TESTFN)
|
|
self.assertEqual(m.x, 5)
|
|
|
|
|
|
class TestSymbolicallyLinkedPackage(unittest.TestCase):
|
|
package_name = 'sample'
|
|
tagged = package_name + '-tagged'
|
|
|
|
def setUp(self):
|
|
os_helper.rmtree(self.tagged)
|
|
os_helper.rmtree(self.package_name)
|
|
self.orig_sys_path = sys.path[:]
|
|
|
|
# create a sample package; imagine you have a package with a tag and
|
|
# you want to symbolically link it from its untagged name.
|
|
os.mkdir(self.tagged)
|
|
self.addCleanup(os_helper.rmtree, self.tagged)
|
|
init_file = os.path.join(self.tagged, '__init__.py')
|
|
os_helper.create_empty_file(init_file)
|
|
assert os.path.exists(init_file)
|
|
|
|
# now create a symlink to the tagged package
|
|
# sample -> sample-tagged
|
|
os.symlink(self.tagged, self.package_name, target_is_directory=True)
|
|
self.addCleanup(os_helper.unlink, self.package_name)
|
|
importlib.invalidate_caches()
|
|
|
|
self.assertEqual(os.path.isdir(self.package_name), True)
|
|
|
|
assert os.path.isfile(os.path.join(self.package_name, '__init__.py'))
|
|
|
|
def tearDown(self):
|
|
sys.path[:] = self.orig_sys_path
|
|
|
|
# regression test for issue6727
|
|
@unittest.skipUnless(
|
|
not hasattr(sys, 'getwindowsversion')
|
|
or sys.getwindowsversion() >= (6, 0),
|
|
"Windows Vista or later required")
|
|
@os_helper.skip_unless_symlink
|
|
def test_symlinked_dir_importable(self):
|
|
# make sure sample can only be imported from the current directory.
|
|
sys.path[:] = ['.']
|
|
assert os.path.exists(self.package_name)
|
|
assert os.path.exists(os.path.join(self.package_name, '__init__.py'))
|
|
|
|
# Try to import the package
|
|
importlib.import_module(self.package_name)
|
|
|
|
|
|
@cpython_only
|
|
class ImportlibBootstrapTests(unittest.TestCase):
|
|
# These tests check that importlib is bootstrapped.
|
|
|
|
def test_frozen_importlib(self):
|
|
mod = sys.modules['_frozen_importlib']
|
|
self.assertTrue(mod)
|
|
|
|
def test_frozen_importlib_is_bootstrap(self):
|
|
from importlib import _bootstrap
|
|
mod = sys.modules['_frozen_importlib']
|
|
self.assertIs(mod, _bootstrap)
|
|
self.assertEqual(mod.__name__, 'importlib._bootstrap')
|
|
self.assertEqual(mod.__package__, 'importlib')
|
|
self.assertTrue(mod.__file__.endswith('_bootstrap.py'), mod.__file__)
|
|
|
|
def test_frozen_importlib_external_is_bootstrap_external(self):
|
|
from importlib import _bootstrap_external
|
|
mod = sys.modules['_frozen_importlib_external']
|
|
self.assertIs(mod, _bootstrap_external)
|
|
self.assertEqual(mod.__name__, 'importlib._bootstrap_external')
|
|
self.assertEqual(mod.__package__, 'importlib')
|
|
self.assertTrue(mod.__file__.endswith('_bootstrap_external.py'), mod.__file__)
|
|
|
|
def test_there_can_be_only_one(self):
|
|
# Issue #15386 revealed a tricky loophole in the bootstrapping
|
|
# This test is technically redundant, since the bug caused importing
|
|
# this test module to crash completely, but it helps prove the point
|
|
from importlib import machinery
|
|
mod = sys.modules['_frozen_importlib']
|
|
self.assertIs(machinery.ModuleSpec, mod.ModuleSpec)
|
|
|
|
|
|
@cpython_only
|
|
class GetSourcefileTests(unittest.TestCase):
|
|
|
|
"""Test importlib._bootstrap_external._get_sourcefile() as used by the C API.
|
|
|
|
Because of the peculiarities of the need of this function, the tests are
|
|
knowingly whitebox tests.
|
|
|
|
"""
|
|
|
|
def test_get_sourcefile(self):
|
|
# Given a valid bytecode path, return the path to the corresponding
|
|
# source file if it exists.
|
|
with mock.patch('importlib._bootstrap_external._path_isfile') as _path_isfile:
|
|
_path_isfile.return_value = True
|
|
path = TESTFN + '.pyc'
|
|
expect = TESTFN + '.py'
|
|
self.assertEqual(_get_sourcefile(path), expect)
|
|
|
|
def test_get_sourcefile_no_source(self):
|
|
# Given a valid bytecode path without a corresponding source path,
|
|
# return the original bytecode path.
|
|
with mock.patch('importlib._bootstrap_external._path_isfile') as _path_isfile:
|
|
_path_isfile.return_value = False
|
|
path = TESTFN + '.pyc'
|
|
self.assertEqual(_get_sourcefile(path), path)
|
|
|
|
def test_get_sourcefile_bad_ext(self):
|
|
# Given a path with an invalid bytecode extension, return the
|
|
# bytecode path passed as the argument.
|
|
path = TESTFN + '.bad_ext'
|
|
self.assertEqual(_get_sourcefile(path), path)
|
|
|
|
|
|
class ImportTracebackTests(unittest.TestCase):
|
|
|
|
def setUp(self):
|
|
os.mkdir(TESTFN)
|
|
self.old_path = sys.path[:]
|
|
sys.path.insert(0, TESTFN)
|
|
|
|
def tearDown(self):
|
|
sys.path[:] = self.old_path
|
|
rmtree(TESTFN)
|
|
|
|
def create_module(self, mod, contents, ext=".py"):
|
|
fname = os.path.join(TESTFN, mod + ext)
|
|
with open(fname, "w", encoding='utf-8') as f:
|
|
f.write(contents)
|
|
self.addCleanup(unload, mod)
|
|
importlib.invalidate_caches()
|
|
return fname
|
|
|
|
def assert_traceback(self, tb, files):
|
|
deduped_files = []
|
|
while tb:
|
|
code = tb.tb_frame.f_code
|
|
fn = code.co_filename
|
|
if not deduped_files or fn != deduped_files[-1]:
|
|
deduped_files.append(fn)
|
|
tb = tb.tb_next
|
|
self.assertEqual(len(deduped_files), len(files), deduped_files)
|
|
for fn, pat in zip(deduped_files, files):
|
|
self.assertIn(pat, fn)
|
|
|
|
def test_nonexistent_module(self):
|
|
try:
|
|
# assertRaises() clears __traceback__
|
|
import nonexistent_xyzzy
|
|
except ImportError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ImportError should have been raised")
|
|
self.assert_traceback(tb, [__file__])
|
|
|
|
def test_nonexistent_module_nested(self):
|
|
self.create_module("foo", "import nonexistent_xyzzy")
|
|
try:
|
|
import foo
|
|
except ImportError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ImportError should have been raised")
|
|
self.assert_traceback(tb, [__file__, 'foo.py'])
|
|
|
|
def test_exec_failure(self):
|
|
self.create_module("foo", "1/0")
|
|
try:
|
|
import foo
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, 'foo.py'])
|
|
|
|
def test_exec_failure_nested(self):
|
|
self.create_module("foo", "import bar")
|
|
self.create_module("bar", "1/0")
|
|
try:
|
|
import foo
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, 'foo.py', 'bar.py'])
|
|
|
|
# A few more examples from issue #15425
|
|
def test_syntax_error(self):
|
|
self.create_module("foo", "invalid syntax is invalid")
|
|
try:
|
|
import foo
|
|
except SyntaxError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("SyntaxError should have been raised")
|
|
self.assert_traceback(tb, [__file__])
|
|
|
|
def _setup_broken_package(self, parent, child):
|
|
pkg_name = "_parent_foo"
|
|
self.addCleanup(unload, pkg_name)
|
|
pkg_path = os.path.join(TESTFN, pkg_name)
|
|
os.mkdir(pkg_path)
|
|
# Touch the __init__.py
|
|
init_path = os.path.join(pkg_path, '__init__.py')
|
|
with open(init_path, 'w', encoding='utf-8') as f:
|
|
f.write(parent)
|
|
bar_path = os.path.join(pkg_path, 'bar.py')
|
|
with open(bar_path, 'w', encoding='utf-8') as f:
|
|
f.write(child)
|
|
importlib.invalidate_caches()
|
|
return init_path, bar_path
|
|
|
|
def test_broken_submodule(self):
|
|
init_path, bar_path = self._setup_broken_package("", "1/0")
|
|
try:
|
|
import _parent_foo.bar
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, bar_path])
|
|
|
|
def test_broken_from(self):
|
|
init_path, bar_path = self._setup_broken_package("", "1/0")
|
|
try:
|
|
from _parent_foo import bar
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ImportError should have been raised")
|
|
self.assert_traceback(tb, [__file__, bar_path])
|
|
|
|
def test_broken_parent(self):
|
|
init_path, bar_path = self._setup_broken_package("1/0", "")
|
|
try:
|
|
import _parent_foo.bar
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, init_path])
|
|
|
|
def test_broken_parent_from(self):
|
|
init_path, bar_path = self._setup_broken_package("1/0", "")
|
|
try:
|
|
from _parent_foo import bar
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, init_path])
|
|
|
|
@cpython_only
|
|
def test_import_bug(self):
|
|
# We simulate a bug in importlib and check that it's not stripped
|
|
# away from the traceback.
|
|
self.create_module("foo", "")
|
|
importlib = sys.modules['_frozen_importlib_external']
|
|
if 'load_module' in vars(importlib.SourceLoader):
|
|
old_exec_module = importlib.SourceLoader.exec_module
|
|
else:
|
|
old_exec_module = None
|
|
try:
|
|
def exec_module(*args):
|
|
1/0
|
|
importlib.SourceLoader.exec_module = exec_module
|
|
try:
|
|
import foo
|
|
except ZeroDivisionError as e:
|
|
tb = e.__traceback__
|
|
else:
|
|
self.fail("ZeroDivisionError should have been raised")
|
|
self.assert_traceback(tb, [__file__, '<frozen importlib', __file__])
|
|
finally:
|
|
if old_exec_module is None:
|
|
del importlib.SourceLoader.exec_module
|
|
else:
|
|
importlib.SourceLoader.exec_module = old_exec_module
|
|
|
|
@unittest.skipUnless(TESTFN_UNENCODABLE, 'need TESTFN_UNENCODABLE')
|
|
def test_unencodable_filename(self):
|
|
# Issue #11619: The Python parser and the import machinery must not
|
|
# encode filenames, especially on Windows
|
|
pyname = script_helper.make_script('', TESTFN_UNENCODABLE, 'pass')
|
|
self.addCleanup(unlink, pyname)
|
|
name = pyname[:-3]
|
|
script_helper.assert_python_ok("-c", "mod = __import__(%a)" % name,
|
|
__isolated=False)
|
|
|
|
|
|
class CircularImportTests(unittest.TestCase):
|
|
|
|
"""See the docstrings of the modules being imported for the purpose of the
|
|
test."""
|
|
|
|
def tearDown(self):
|
|
"""Make sure no modules pre-exist in sys.modules which are being used to
|
|
test."""
|
|
for key in list(sys.modules.keys()):
|
|
if key.startswith('test.test_import.data.circular_imports'):
|
|
del sys.modules[key]
|
|
|
|
def test_direct(self):
|
|
try:
|
|
import test.test_import.data.circular_imports.basic
|
|
except ImportError:
|
|
self.fail('circular import through relative imports failed')
|
|
|
|
def test_indirect(self):
|
|
try:
|
|
import test.test_import.data.circular_imports.indirect
|
|
except ImportError:
|
|
self.fail('relative import in module contributing to circular '
|
|
'import failed')
|
|
|
|
def test_subpackage(self):
|
|
try:
|
|
import test.test_import.data.circular_imports.subpackage
|
|
except ImportError:
|
|
self.fail('circular import involving a subpackage failed')
|
|
|
|
def test_rebinding(self):
|
|
try:
|
|
import test.test_import.data.circular_imports.rebinding as rebinding
|
|
except ImportError:
|
|
self.fail('circular import with rebinding of module attribute failed')
|
|
from test.test_import.data.circular_imports.subpkg import util
|
|
self.assertIs(util.util, rebinding.util)
|
|
|
|
def test_binding(self):
|
|
try:
|
|
import test.test_import.data.circular_imports.binding
|
|
except ImportError:
|
|
self.fail('circular import with binding a submodule to a name failed')
|
|
|
|
def test_crossreference1(self):
|
|
import test.test_import.data.circular_imports.use
|
|
import test.test_import.data.circular_imports.source
|
|
|
|
def test_crossreference2(self):
|
|
with self.assertRaises(AttributeError) as cm:
|
|
import test.test_import.data.circular_imports.source
|
|
errmsg = str(cm.exception)
|
|
self.assertIn('test.test_import.data.circular_imports.source', errmsg)
|
|
self.assertIn('spam', errmsg)
|
|
self.assertIn('partially initialized module', errmsg)
|
|
self.assertIn('circular import', errmsg)
|
|
|
|
def test_circular_from_import(self):
|
|
with self.assertRaises(ImportError) as cm:
|
|
import test.test_import.data.circular_imports.from_cycle1
|
|
self.assertIn(
|
|
"cannot import name 'b' from partially initialized module "
|
|
"'test.test_import.data.circular_imports.from_cycle1' "
|
|
"(most likely due to a circular import)",
|
|
str(cm.exception),
|
|
)
|
|
|
|
def test_circular_import(self):
|
|
with self.assertRaisesRegex(
|
|
AttributeError,
|
|
r"partially initialized module 'test.test_import.data.circular_imports.import_cycle' "
|
|
r"from '.*' has no attribute 'some_attribute' \(most likely due to a circular import\)"
|
|
):
|
|
import test.test_import.data.circular_imports.import_cycle
|
|
|
|
def test_absolute_circular_submodule(self):
|
|
with self.assertRaises(AttributeError) as cm:
|
|
import test.test_import.data.circular_imports.subpkg2.parent
|
|
self.assertIn(
|
|
"cannot access submodule 'parent' of module "
|
|
"'test.test_import.data.circular_imports.subpkg2' "
|
|
"(most likely due to a circular import)",
|
|
str(cm.exception),
|
|
)
|
|
|
|
def test_unwritable_module(self):
|
|
self.addCleanup(unload, "test.test_import.data.unwritable")
|
|
self.addCleanup(unload, "test.test_import.data.unwritable.x")
|
|
|
|
import test.test_import.data.unwritable as unwritable
|
|
with self.assertWarns(ImportWarning):
|
|
from test.test_import.data.unwritable import x
|
|
|
|
self.assertNotEqual(type(unwritable), ModuleType)
|
|
self.assertEqual(type(x), ModuleType)
|
|
with self.assertRaises(AttributeError):
|
|
unwritable.x = 42
|
|
|
|
|
|
class SubinterpImportTests(unittest.TestCase):
|
|
|
|
RUN_KWARGS = dict(
|
|
allow_fork=False,
|
|
allow_exec=False,
|
|
allow_threads=True,
|
|
allow_daemon_threads=False,
|
|
# Isolation-related config values aren't included here.
|
|
)
|
|
ISOLATED = dict(
|
|
use_main_obmalloc=False,
|
|
gil=2,
|
|
)
|
|
NOT_ISOLATED = {k: not v for k, v in ISOLATED.items()}
|
|
NOT_ISOLATED['gil'] = 1
|
|
|
|
@unittest.skipUnless(hasattr(os, "pipe"), "requires os.pipe()")
|
|
def pipe(self):
|
|
r, w = os.pipe()
|
|
self.addCleanup(os.close, r)
|
|
self.addCleanup(os.close, w)
|
|
if hasattr(os, 'set_blocking'):
|
|
os.set_blocking(r, False)
|
|
return (r, w)
|
|
|
|
def create_extension_loader(self, modname, filename):
|
|
# Apple extensions must be distributed as frameworks. This requires
|
|
# a specialist loader.
|
|
if is_apple_mobile:
|
|
return AppleFrameworkLoader(modname, filename)
|
|
else:
|
|
return ExtensionFileLoader(modname, filename)
|
|
|
|
def import_script(self, name, fd, filename=None, check_override=None):
|
|
override_text = ''
|
|
if check_override is not None:
|
|
override_text = f'''
|
|
import _imp
|
|
_imp._override_multi_interp_extensions_check({check_override})
|
|
'''
|
|
if filename:
|
|
# Apple extensions must be distributed as frameworks. This requires
|
|
# a specialist loader.
|
|
if is_apple_mobile:
|
|
loader = "AppleFrameworkLoader"
|
|
else:
|
|
loader = "ExtensionFileLoader"
|
|
|
|
return textwrap.dedent(f'''
|
|
from importlib.util import spec_from_loader, module_from_spec
|
|
from importlib.machinery import {loader}
|
|
import os, sys
|
|
{override_text}
|
|
loader = {loader}({name!r}, {filename!r})
|
|
spec = spec_from_loader({name!r}, loader)
|
|
try:
|
|
module = module_from_spec(spec)
|
|
loader.exec_module(module)
|
|
except ImportError as exc:
|
|
text = 'ImportError: ' + str(exc)
|
|
else:
|
|
text = 'okay'
|
|
os.write({fd}, text.encode('utf-8'))
|
|
''')
|
|
else:
|
|
return textwrap.dedent(f'''
|
|
import os, sys
|
|
{override_text}
|
|
try:
|
|
import {name}
|
|
except ImportError as exc:
|
|
text = 'ImportError: ' + str(exc)
|
|
else:
|
|
text = 'okay'
|
|
os.write({fd}, text.encode('utf-8'))
|
|
''')
|
|
|
|
def run_here(self, name, filename=None, *,
|
|
check_singlephase_setting=False,
|
|
check_singlephase_override=None,
|
|
isolated=False,
|
|
):
|
|
"""
|
|
Try importing the named module in a subinterpreter.
|
|
|
|
The subinterpreter will be in the current process.
|
|
The module will have already been imported in the main interpreter.
|
|
Thus, for extension/builtin modules, the module definition will
|
|
have been loaded already and cached globally.
|
|
|
|
"check_singlephase_setting" determines whether or not
|
|
the interpreter will be configured to check for modules
|
|
that are not compatible with use in multiple interpreters.
|
|
|
|
This should always return "okay" for all modules if the
|
|
setting is False (with no override).
|
|
"""
|
|
__import__(name)
|
|
|
|
kwargs = dict(
|
|
**self.RUN_KWARGS,
|
|
**(self.ISOLATED if isolated else self.NOT_ISOLATED),
|
|
check_multi_interp_extensions=check_singlephase_setting,
|
|
)
|
|
|
|
r, w = self.pipe()
|
|
script = self.import_script(name, w, filename,
|
|
check_singlephase_override)
|
|
|
|
ret = run_in_subinterp_with_config(script, **kwargs)
|
|
self.assertEqual(ret, 0)
|
|
return os.read(r, 100)
|
|
|
|
def check_compatible_here(self, name, filename=None, *,
|
|
strict=False,
|
|
isolated=False,
|
|
):
|
|
# Verify that the named module may be imported in a subinterpreter.
|
|
# (See run_here() for more info.)
|
|
out = self.run_here(name, filename,
|
|
check_singlephase_setting=strict,
|
|
isolated=isolated,
|
|
)
|
|
self.assertEqual(out, b'okay')
|
|
|
|
def check_incompatible_here(self, name, filename=None, *, isolated=False):
|
|
# Differences from check_compatible_here():
|
|
# * verify that import fails
|
|
# * "strict" is always True
|
|
out = self.run_here(name, filename,
|
|
check_singlephase_setting=True,
|
|
isolated=isolated,
|
|
)
|
|
self.assertEqual(
|
|
out.decode('utf-8'),
|
|
f'ImportError: module {name} does not support loading in subinterpreters',
|
|
)
|
|
|
|
def check_compatible_fresh(self, name, *, strict=False, isolated=False):
|
|
# Differences from check_compatible_here():
|
|
# * subinterpreter in a new process
|
|
# * module has never been imported before in that process
|
|
# * this tests importing the module for the first time
|
|
kwargs = dict(
|
|
**self.RUN_KWARGS,
|
|
**(self.ISOLATED if isolated else self.NOT_ISOLATED),
|
|
check_multi_interp_extensions=strict,
|
|
)
|
|
gil = kwargs['gil']
|
|
kwargs['gil'] = 'default' if gil == 0 else (
|
|
'shared' if gil == 1 else 'own' if gil == 2 else gil)
|
|
_, out, err = script_helper.assert_python_ok('-c', textwrap.dedent(f'''
|
|
import _testinternalcapi, sys
|
|
assert (
|
|
{name!r} in sys.builtin_module_names or
|
|
{name!r} not in sys.modules
|
|
), repr({name!r})
|
|
config = type(sys.implementation)(**{kwargs})
|
|
ret = _testinternalcapi.run_in_subinterp_with_config(
|
|
{self.import_script(name, "sys.stdout.fileno()")!r},
|
|
config,
|
|
)
|
|
assert ret == 0, ret
|
|
'''))
|
|
self.assertEqual(err, b'')
|
|
self.assertEqual(out, b'okay')
|
|
|
|
def check_incompatible_fresh(self, name, *, isolated=False):
|
|
# Differences from check_compatible_fresh():
|
|
# * verify that import fails
|
|
# * "strict" is always True
|
|
kwargs = dict(
|
|
**self.RUN_KWARGS,
|
|
**(self.ISOLATED if isolated else self.NOT_ISOLATED),
|
|
check_multi_interp_extensions=True,
|
|
)
|
|
gil = kwargs['gil']
|
|
kwargs['gil'] = 'default' if gil == 0 else (
|
|
'shared' if gil == 1 else 'own' if gil == 2 else gil)
|
|
_, out, err = script_helper.assert_python_ok('-c', textwrap.dedent(f'''
|
|
import _testinternalcapi, sys
|
|
assert {name!r} not in sys.modules, {name!r}
|
|
config = type(sys.implementation)(**{kwargs})
|
|
ret = _testinternalcapi.run_in_subinterp_with_config(
|
|
{self.import_script(name, "sys.stdout.fileno()")!r},
|
|
config,
|
|
)
|
|
assert ret == 0, ret
|
|
'''))
|
|
self.assertEqual(err, b'')
|
|
self.assertEqual(
|
|
out.decode('utf-8'),
|
|
f'ImportError: module {name} does not support loading in subinterpreters',
|
|
)
|
|
|
|
@unittest.skipIf(_testinternalcapi is None, "requires _testinternalcapi")
|
|
def test_builtin_compat(self):
|
|
# For now we avoid using sys or builtins
|
|
# since they still don't implement multi-phase init.
|
|
module = '_imp'
|
|
require_builtin(module)
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{module}: not strict'):
|
|
self.check_compatible_here(module, strict=False)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_compatible_here(module, strict=True)
|
|
|
|
@cpython_only
|
|
@unittest.skipIf(_testinternalcapi is None, "requires _testinternalcapi")
|
|
def test_frozen_compat(self):
|
|
module = '_frozen_importlib'
|
|
require_frozen(module, skip=True)
|
|
if __import__(module).__spec__.origin != 'frozen':
|
|
raise unittest.SkipTest(f'{module} is unexpectedly not frozen')
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{module}: not strict'):
|
|
self.check_compatible_here(module, strict=False)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_compatible_here(module, strict=True)
|
|
|
|
@requires_singlephase_init
|
|
def test_single_init_extension_compat(self):
|
|
module = '_testsinglephase'
|
|
require_extension(module)
|
|
with self.subTest(f'{module}: not strict'):
|
|
self.check_compatible_here(module, strict=False)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_incompatible_here(module)
|
|
with self.subTest(f'{module}: strict, fresh'):
|
|
self.check_incompatible_fresh(module)
|
|
|
|
@unittest.skipIf(_testmultiphase is None, "test requires _testmultiphase module")
|
|
def test_multi_init_extension_compat(self):
|
|
module = '_testmultiphase'
|
|
require_extension(module)
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{module}: not strict'):
|
|
self.check_compatible_here(module, strict=False)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_compatible_here(module, strict=True)
|
|
with self.subTest(f'{module}: strict, fresh'):
|
|
self.check_compatible_fresh(module, strict=True)
|
|
|
|
@unittest.skipIf(_testmultiphase is None, "test requires _testmultiphase module")
|
|
def test_multi_init_extension_non_isolated_compat(self):
|
|
modname = '_test_non_isolated'
|
|
filename = _testmultiphase.__file__
|
|
loader = self.create_extension_loader(modname, filename)
|
|
spec = importlib.util.spec_from_loader(modname, loader)
|
|
module = importlib.util.module_from_spec(spec)
|
|
loader.exec_module(module)
|
|
sys.modules[modname] = module
|
|
|
|
require_extension(module)
|
|
with self.subTest(f'{modname}: isolated'):
|
|
self.check_incompatible_here(modname, filename, isolated=True)
|
|
with self.subTest(f'{modname}: not isolated'):
|
|
self.check_incompatible_here(modname, filename, isolated=False)
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{modname}: not strict'):
|
|
self.check_compatible_here(modname, filename, strict=False)
|
|
|
|
@unittest.skipIf(_testmultiphase is None, "test requires _testmultiphase module")
|
|
def test_multi_init_extension_per_interpreter_gil_compat(self):
|
|
modname = '_test_shared_gil_only'
|
|
filename = _testmultiphase.__file__
|
|
loader = self.create_extension_loader(modname, filename)
|
|
spec = importlib.util.spec_from_loader(modname, loader)
|
|
module = importlib.util.module_from_spec(spec)
|
|
loader.exec_module(module)
|
|
sys.modules[modname] = module
|
|
|
|
require_extension(module)
|
|
with self.subTest(f'{modname}: isolated, strict'):
|
|
self.check_incompatible_here(modname, filename, isolated=True)
|
|
with self.subTest(f'{modname}: not isolated, strict'):
|
|
self.check_compatible_here(modname, filename,
|
|
strict=True, isolated=False)
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{modname}: not isolated, not strict'):
|
|
self.check_compatible_here(modname, filename,
|
|
strict=False, isolated=False)
|
|
|
|
@unittest.skipIf(_testinternalcapi is None, "requires _testinternalcapi")
|
|
def test_python_compat(self):
|
|
module = 'threading'
|
|
require_pure_python(module)
|
|
if not Py_GIL_DISABLED:
|
|
with self.subTest(f'{module}: not strict'):
|
|
self.check_compatible_here(module, strict=False)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_compatible_here(module, strict=True)
|
|
with self.subTest(f'{module}: strict, fresh'):
|
|
self.check_compatible_fresh(module, strict=True)
|
|
|
|
@requires_singlephase_init
|
|
def test_singlephase_check_with_setting_and_override(self):
|
|
module = '_testsinglephase'
|
|
require_extension(module)
|
|
|
|
def check_compatible(setting, override):
|
|
out = self.run_here(
|
|
module,
|
|
check_singlephase_setting=setting,
|
|
check_singlephase_override=override,
|
|
)
|
|
self.assertEqual(out, b'okay')
|
|
|
|
def check_incompatible(setting, override):
|
|
out = self.run_here(
|
|
module,
|
|
check_singlephase_setting=setting,
|
|
check_singlephase_override=override,
|
|
)
|
|
self.assertNotEqual(out, b'okay')
|
|
|
|
with self.subTest('config: check enabled; override: enabled'):
|
|
check_incompatible(True, 1)
|
|
with self.subTest('config: check enabled; override: use config'):
|
|
check_incompatible(True, 0)
|
|
with self.subTest('config: check enabled; override: disabled'):
|
|
check_compatible(True, -1)
|
|
|
|
with self.subTest('config: check disabled; override: enabled'):
|
|
check_incompatible(False, 1)
|
|
with self.subTest('config: check disabled; override: use config'):
|
|
check_compatible(False, 0)
|
|
with self.subTest('config: check disabled; override: disabled'):
|
|
check_compatible(False, -1)
|
|
|
|
@unittest.skipIf(_testinternalcapi is None, "requires _testinternalcapi")
|
|
def test_isolated_config(self):
|
|
module = 'threading'
|
|
require_pure_python(module)
|
|
with self.subTest(f'{module}: strict, not fresh'):
|
|
self.check_compatible_here(module, strict=True, isolated=True)
|
|
with self.subTest(f'{module}: strict, fresh'):
|
|
self.check_compatible_fresh(module, strict=True, isolated=True)
|
|
|
|
@requires_subinterpreters
|
|
@requires_singlephase_init
|
|
def test_disallowed_reimport(self):
|
|
# See https://github.com/python/cpython/issues/104621.
|
|
script = textwrap.dedent('''
|
|
import _testsinglephase
|
|
print(_testsinglephase)
|
|
''')
|
|
interpid = _interpreters.create()
|
|
self.addCleanup(lambda: _interpreters.destroy(interpid))
|
|
|
|
excsnap = _interpreters.run_string(interpid, script)
|
|
self.assertIsNot(excsnap, None)
|
|
|
|
excsnap = _interpreters.run_string(interpid, script)
|
|
self.assertIsNot(excsnap, None)
|
|
|
|
|
|
class TestSinglePhaseSnapshot(ModuleSnapshot):
|
|
"""A representation of a single-phase init module for testing.
|
|
|
|
Fields from ModuleSnapshot:
|
|
|
|
* id - id(mod)
|
|
* module - mod or a SimpleNamespace with __file__ & __spec__
|
|
* ns - a shallow copy of mod.__dict__
|
|
* ns_id - id(mod.__dict__)
|
|
* cached - sys.modules[name] (or None if not there or not snapshotable)
|
|
* cached_id - id(sys.modules[name]) (or None if not there)
|
|
|
|
Extra fields:
|
|
|
|
* summed - the result of calling "mod.sum(1, 2)"
|
|
* lookedup - the result of calling "mod.look_up_self()"
|
|
* lookedup_id - the object ID of self.lookedup
|
|
* state_initialized - the result of calling "mod.state_initialized()"
|
|
* init_count - (optional) the result of calling "mod.initialized_count()"
|
|
|
|
Overridden methods from ModuleSnapshot:
|
|
|
|
* from_module()
|
|
* parse()
|
|
|
|
Other methods from ModuleSnapshot:
|
|
|
|
* build_script()
|
|
* from_subinterp()
|
|
|
|
----
|
|
|
|
There are 5 modules in Modules/_testsinglephase.c:
|
|
|
|
* _testsinglephase
|
|
* has global state
|
|
* extra loads skip the init function, copy def.m_base.m_copy
|
|
* counts calls to init function
|
|
* _testsinglephase_basic_wrapper
|
|
* _testsinglephase by another name (and separate init function symbol)
|
|
* _testsinglephase_basic_copy
|
|
* same as _testsinglephase but with own def (and init func)
|
|
* _testsinglephase_with_reinit
|
|
* has no global or module state
|
|
* mod.state_initialized returns None
|
|
* an extra load in the main interpreter calls the cached init func
|
|
* an extra load in legacy subinterpreters does a full load
|
|
* _testsinglephase_with_state
|
|
* has module state
|
|
* an extra load in the main interpreter calls the cached init func
|
|
* an extra load in legacy subinterpreters does a full load
|
|
|
|
(See Modules/_testsinglephase.c for more info.)
|
|
|
|
For all those modules, the snapshot after the initial load (not in
|
|
the global extensions cache) would look like the following:
|
|
|
|
* initial load
|
|
* id: ID of nww module object
|
|
* ns: exactly what the module init put there
|
|
* ns_id: ID of new module's __dict__
|
|
* cached_id: same as self.id
|
|
* summed: 3 (never changes)
|
|
* lookedup_id: same as self.id
|
|
* state_initialized: a timestamp between the time of the load
|
|
and the time of the snapshot
|
|
* init_count: 1 (None for _testsinglephase_with_reinit)
|
|
|
|
For the other scenarios it varies.
|
|
|
|
For the _testsinglephase, _testsinglephase_basic_wrapper, and
|
|
_testsinglephase_basic_copy modules, the snapshot should look
|
|
like the following:
|
|
|
|
* reloaded
|
|
* id: no change
|
|
* ns: matches what the module init function put there,
|
|
including the IDs of all contained objects,
|
|
plus any extra attributes added before the reload
|
|
* ns_id: no change
|
|
* cached_id: no change
|
|
* lookedup_id: no change
|
|
* state_initialized: no change
|
|
* init_count: no change
|
|
* already loaded
|
|
* (same as initial load except for ns and state_initialized)
|
|
* ns: matches the initial load, incl. IDs of contained objects
|
|
* state_initialized: no change from initial load
|
|
|
|
For _testsinglephase_with_reinit:
|
|
|
|
* reloaded: same as initial load (old module & ns is discarded)
|
|
* already loaded: same as initial load (old module & ns is discarded)
|
|
|
|
For _testsinglephase_with_state:
|
|
|
|
* reloaded
|
|
* (same as initial load (old module & ns is discarded),
|
|
except init_count)
|
|
* init_count: increase by 1
|
|
* already loaded: same as reloaded
|
|
"""
|
|
|
|
@classmethod
|
|
def from_module(cls, mod):
|
|
self = super().from_module(mod)
|
|
self.summed = mod.sum(1, 2)
|
|
self.lookedup = mod.look_up_self()
|
|
self.lookedup_id = id(self.lookedup)
|
|
self.state_initialized = mod.state_initialized()
|
|
if hasattr(mod, 'initialized_count'):
|
|
self.init_count = mod.initialized_count()
|
|
return self
|
|
|
|
SCRIPT_BODY = ModuleSnapshot.SCRIPT_BODY + textwrap.dedent('''
|
|
snapshot['module'].update(dict(
|
|
int_const=mod.int_const,
|
|
str_const=mod.str_const,
|
|
_module_initialized=mod._module_initialized,
|
|
))
|
|
snapshot.update(dict(
|
|
summed=mod.sum(1, 2),
|
|
lookedup_id=id(mod.look_up_self()),
|
|
state_initialized=mod.state_initialized(),
|
|
init_count=mod.initialized_count(),
|
|
has_spam=hasattr(mod, 'spam'),
|
|
spam=getattr(mod, 'spam', None),
|
|
))
|
|
''').rstrip()
|
|
|
|
@classmethod
|
|
def parse(cls, text):
|
|
self = super().parse(text)
|
|
if not self.has_spam:
|
|
del self.spam
|
|
del self.has_spam
|
|
return self
|
|
|
|
|
|
@requires_singlephase_init
|
|
class SinglephaseInitTests(unittest.TestCase):
|
|
|
|
NAME = '_testsinglephase'
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
spec = importlib.util.find_spec(cls.NAME)
|
|
cls.LOADER = type(spec.loader)
|
|
|
|
# Apple extensions must be distributed as frameworks. This requires
|
|
# a specialist loader, and we need to differentiate between the
|
|
# spec.origin and the original file location.
|
|
if is_apple_mobile:
|
|
assert cls.LOADER is AppleFrameworkLoader
|
|
|
|
cls.ORIGIN = spec.origin
|
|
with open(spec.origin + ".origin", "r") as f:
|
|
cls.FILE = os.path.join(
|
|
os.path.dirname(sys.executable),
|
|
f.read().strip()
|
|
)
|
|
else:
|
|
assert cls.LOADER is ExtensionFileLoader
|
|
|
|
cls.ORIGIN = spec.origin
|
|
cls.FILE = spec.origin
|
|
|
|
# Start fresh.
|
|
cls.clean_up()
|
|
|
|
def tearDown(self):
|
|
# Clean up the module.
|
|
self.clean_up()
|
|
|
|
@classmethod
|
|
def clean_up(cls):
|
|
name = cls.NAME
|
|
if name in sys.modules:
|
|
if hasattr(sys.modules[name], '_clear_globals'):
|
|
assert sys.modules[name].__file__ == cls.FILE, \
|
|
f"{sys.modules[name].__file__} != {cls.FILE}"
|
|
|
|
sys.modules[name]._clear_globals()
|
|
del sys.modules[name]
|
|
# Clear all internally cached data for the extension.
|
|
_testinternalcapi.clear_extension(name, cls.ORIGIN)
|
|
|
|
#########################
|
|
# helpers
|
|
|
|
def add_module_cleanup(self, name):
|
|
def clean_up():
|
|
# Clear all internally cached data for the extension.
|
|
_testinternalcapi.clear_extension(name, self.ORIGIN)
|
|
self.addCleanup(clean_up)
|
|
|
|
def _load_dynamic(self, name, path):
|
|
"""
|
|
Load an extension module.
|
|
"""
|
|
# This is essentially copied from the old imp module.
|
|
from importlib._bootstrap import _load
|
|
loader = self.LOADER(name, path)
|
|
|
|
# Issue bpo-24748: Skip the sys.modules check in _load_module_shim;
|
|
# always load new extension.
|
|
spec = importlib.util.spec_from_file_location(name, path,
|
|
loader=loader)
|
|
return _load(spec)
|
|
|
|
def load(self, name):
|
|
try:
|
|
already_loaded = self.already_loaded
|
|
except AttributeError:
|
|
already_loaded = self.already_loaded = {}
|
|
assert name not in already_loaded
|
|
mod = self._load_dynamic(name, self.ORIGIN)
|
|
self.assertNotIn(mod, already_loaded.values())
|
|
already_loaded[name] = mod
|
|
return types.SimpleNamespace(
|
|
name=name,
|
|
module=mod,
|
|
snapshot=TestSinglePhaseSnapshot.from_module(mod),
|
|
)
|
|
|
|
def re_load(self, name, mod):
|
|
assert sys.modules[name] is mod
|
|
assert mod.__dict__ == mod.__dict__
|
|
reloaded = self._load_dynamic(name, self.ORIGIN)
|
|
return types.SimpleNamespace(
|
|
name=name,
|
|
module=reloaded,
|
|
snapshot=TestSinglePhaseSnapshot.from_module(reloaded),
|
|
)
|
|
|
|
# subinterpreters
|
|
|
|
def add_subinterpreter(self):
|
|
interpid = _interpreters.create('legacy')
|
|
def ensure_destroyed():
|
|
try:
|
|
_interpreters.destroy(interpid)
|
|
except _interpreters.InterpreterNotFoundError:
|
|
pass
|
|
self.addCleanup(ensure_destroyed)
|
|
_interpreters.exec(interpid, textwrap.dedent('''
|
|
import sys
|
|
import _testinternalcapi
|
|
'''))
|
|
def clean_up():
|
|
_interpreters.exec(interpid, textwrap.dedent(f'''
|
|
name = {self.NAME!r}
|
|
if name in sys.modules:
|
|
sys.modules.pop(name)._clear_globals()
|
|
_testinternalcapi.clear_extension(name, {self.ORIGIN!r})
|
|
'''))
|
|
_interpreters.destroy(interpid)
|
|
self.addCleanup(clean_up)
|
|
return interpid
|
|
|
|
def import_in_subinterp(self, interpid=None, *,
|
|
postscript=None,
|
|
postcleanup=False,
|
|
):
|
|
name = self.NAME
|
|
|
|
if postcleanup:
|
|
import_ = 'import _testinternalcapi' if interpid is None else ''
|
|
postcleanup = f'''
|
|
{import_}
|
|
mod._clear_globals()
|
|
_testinternalcapi.clear_extension(name, {self.ORIGIN!r})
|
|
'''
|
|
|
|
try:
|
|
pipe = self._pipe
|
|
except AttributeError:
|
|
r, w = pipe = self._pipe = os.pipe()
|
|
self.addCleanup(os.close, r)
|
|
self.addCleanup(os.close, w)
|
|
|
|
snapshot = TestSinglePhaseSnapshot.from_subinterp(
|
|
name,
|
|
interpid,
|
|
pipe=pipe,
|
|
import_first=True,
|
|
postscript=postscript,
|
|
postcleanup=postcleanup,
|
|
)
|
|
|
|
return types.SimpleNamespace(
|
|
name=name,
|
|
module=None,
|
|
snapshot=snapshot,
|
|
)
|
|
|
|
# checks
|
|
|
|
def check_common(self, loaded):
|
|
isolated = False
|
|
|
|
mod = loaded.module
|
|
if not mod:
|
|
# It came from a subinterpreter.
|
|
isolated = True
|
|
mod = loaded.snapshot.module
|
|
# mod.__name__ might not match, but the spec will.
|
|
self.assertEqual(mod.__spec__.name, loaded.name)
|
|
self.assertEqual(mod.__file__, self.FILE)
|
|
self.assertEqual(mod.__spec__.origin, self.ORIGIN)
|
|
if not isolated:
|
|
self.assertTrue(issubclass(mod.error, Exception))
|
|
self.assertEqual(mod.int_const, 1969)
|
|
self.assertEqual(mod.str_const, 'something different')
|
|
self.assertIsInstance(mod._module_initialized, float)
|
|
self.assertGreater(mod._module_initialized, 0)
|
|
|
|
snap = loaded.snapshot
|
|
self.assertEqual(snap.summed, 3)
|
|
if snap.state_initialized is not None:
|
|
self.assertIsInstance(snap.state_initialized, float)
|
|
self.assertGreater(snap.state_initialized, 0)
|
|
if isolated:
|
|
# The "looked up" module is interpreter-specific
|
|
# (interp->imports.modules_by_index was set for the module).
|
|
self.assertEqual(snap.lookedup_id, snap.id)
|
|
self.assertEqual(snap.cached_id, snap.id)
|
|
with self.assertRaises(AttributeError):
|
|
snap.spam
|
|
else:
|
|
self.assertIs(snap.lookedup, mod)
|
|
self.assertIs(snap.cached, mod)
|
|
|
|
def check_direct(self, loaded):
|
|
# The module has its own PyModuleDef, with a matching name.
|
|
self.assertEqual(loaded.module.__name__, loaded.name)
|
|
self.assertIs(loaded.snapshot.lookedup, loaded.module)
|
|
|
|
def check_indirect(self, loaded, orig):
|
|
# The module re-uses another's PyModuleDef, with a different name.
|
|
assert orig is not loaded.module
|
|
assert orig.__name__ != loaded.name
|
|
self.assertNotEqual(loaded.module.__name__, loaded.name)
|
|
self.assertIs(loaded.snapshot.lookedup, loaded.module)
|
|
|
|
def check_basic(self, loaded, expected_init_count):
|
|
# m_size == -1
|
|
# The module loads fresh the first time and copies m_copy after.
|
|
snap = loaded.snapshot
|
|
self.assertIsNot(snap.state_initialized, None)
|
|
self.assertIsInstance(snap.init_count, int)
|
|
self.assertGreater(snap.init_count, 0)
|
|
self.assertEqual(snap.init_count, expected_init_count)
|
|
|
|
def check_with_reinit(self, loaded):
|
|
# m_size >= 0
|
|
# The module loads fresh every time.
|
|
pass
|
|
|
|
def check_fresh(self, loaded):
|
|
"""
|
|
The module had not been loaded before (at least since fully reset).
|
|
"""
|
|
snap = loaded.snapshot
|
|
# The module's init func was run.
|
|
# A copy of the module's __dict__ was stored in def->m_base.m_copy.
|
|
# The previous m_copy was deleted first.
|
|
# _PyRuntime.imports.extensions was set.
|
|
self.assertEqual(snap.init_count, 1)
|
|
# The global state was initialized.
|
|
# The module attrs were initialized from that state.
|
|
self.assertEqual(snap.module._module_initialized,
|
|
snap.state_initialized)
|
|
|
|
def check_semi_fresh(self, loaded, base, prev):
|
|
"""
|
|
The module had been loaded before and then reset
|
|
(but the module global state wasn't).
|
|
"""
|
|
snap = loaded.snapshot
|
|
# The module's init func was run again.
|
|
# A copy of the module's __dict__ was stored in def->m_base.m_copy.
|
|
# The previous m_copy was deleted first.
|
|
# The module globals did not get reset.
|
|
self.assertNotEqual(snap.id, base.snapshot.id)
|
|
self.assertNotEqual(snap.id, prev.snapshot.id)
|
|
self.assertEqual(snap.init_count, prev.snapshot.init_count + 1)
|
|
# The global state was updated.
|
|
# The module attrs were initialized from that state.
|
|
self.assertEqual(snap.module._module_initialized,
|
|
snap.state_initialized)
|
|
self.assertNotEqual(snap.state_initialized,
|
|
base.snapshot.state_initialized)
|
|
self.assertNotEqual(snap.state_initialized,
|
|
prev.snapshot.state_initialized)
|
|
|
|
def check_copied(self, loaded, base):
|
|
"""
|
|
The module had been loaded before and never reset.
|
|
"""
|
|
snap = loaded.snapshot
|
|
# The module's init func was not run again.
|
|
# The interpreter copied m_copy, as set by the other interpreter,
|
|
# with objects owned by the other interpreter.
|
|
# The module globals did not get reset.
|
|
self.assertNotEqual(snap.id, base.snapshot.id)
|
|
self.assertEqual(snap.init_count, base.snapshot.init_count)
|
|
# The global state was not updated since the init func did not run.
|
|
# The module attrs were not directly initialized from that state.
|
|
# The state and module attrs still match the previous loading.
|
|
self.assertEqual(snap.module._module_initialized,
|
|
snap.state_initialized)
|
|
self.assertEqual(snap.state_initialized,
|
|
base.snapshot.state_initialized)
|
|
|
|
#########################
|
|
# the tests
|
|
|
|
def test_cleared_globals(self):
|
|
loaded = self.load(self.NAME)
|
|
_testsinglephase = loaded.module
|
|
init_before = _testsinglephase.state_initialized()
|
|
|
|
_testsinglephase._clear_globals()
|
|
init_after = _testsinglephase.state_initialized()
|
|
init_count = _testsinglephase.initialized_count()
|
|
|
|
self.assertGreater(init_before, 0)
|
|
self.assertEqual(init_after, 0)
|
|
self.assertEqual(init_count, -1)
|
|
|
|
def test_variants(self):
|
|
# Exercise the most meaningful variants described in Python/import.c.
|
|
self.maxDiff = None
|
|
|
|
# Check the "basic" module.
|
|
|
|
name = self.NAME
|
|
expected_init_count = 1
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
|
|
self.check_common(loaded)
|
|
self.check_direct(loaded)
|
|
self.check_basic(loaded, expected_init_count)
|
|
basic = loaded.module
|
|
|
|
# Check its indirect variants.
|
|
|
|
name = f'{self.NAME}_basic_wrapper'
|
|
self.add_module_cleanup(name)
|
|
expected_init_count += 1
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
|
|
self.check_common(loaded)
|
|
self.check_indirect(loaded, basic)
|
|
self.check_basic(loaded, expected_init_count)
|
|
|
|
# Currently PyState_AddModule() always replaces the cached module.
|
|
self.assertIs(basic.look_up_self(), loaded.module)
|
|
self.assertEqual(basic.initialized_count(), expected_init_count)
|
|
|
|
# The cached module shouldn't change after this point.
|
|
basic_lookedup = loaded.module
|
|
|
|
# Check its direct variant.
|
|
|
|
name = f'{self.NAME}_basic_copy'
|
|
self.add_module_cleanup(name)
|
|
expected_init_count += 1
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
|
|
self.check_common(loaded)
|
|
self.check_direct(loaded)
|
|
self.check_basic(loaded, expected_init_count)
|
|
|
|
# This should change the cached module for _testsinglephase.
|
|
self.assertIs(basic.look_up_self(), basic_lookedup)
|
|
self.assertEqual(basic.initialized_count(), expected_init_count)
|
|
|
|
# Check the non-basic variant that has no state.
|
|
|
|
name = f'{self.NAME}_with_reinit'
|
|
self.add_module_cleanup(name)
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
|
|
self.check_common(loaded)
|
|
self.assertIs(loaded.snapshot.state_initialized, None)
|
|
self.check_direct(loaded)
|
|
self.check_with_reinit(loaded)
|
|
|
|
# This should change the cached module for _testsinglephase.
|
|
self.assertIs(basic.look_up_self(), basic_lookedup)
|
|
self.assertEqual(basic.initialized_count(), expected_init_count)
|
|
|
|
# Check the basic variant that has state.
|
|
|
|
name = f'{self.NAME}_with_state'
|
|
self.add_module_cleanup(name)
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
self.addCleanup(loaded.module._clear_module_state)
|
|
|
|
self.check_common(loaded)
|
|
self.assertIsNot(loaded.snapshot.state_initialized, None)
|
|
self.check_direct(loaded)
|
|
self.check_with_reinit(loaded)
|
|
|
|
# This should change the cached module for _testsinglephase.
|
|
self.assertIs(basic.look_up_self(), basic_lookedup)
|
|
self.assertEqual(basic.initialized_count(), expected_init_count)
|
|
|
|
def test_basic_reloaded(self):
|
|
# m_copy is copied into the existing module object.
|
|
# Global state is not changed.
|
|
self.maxDiff = None
|
|
|
|
for name in [
|
|
self.NAME, # the "basic" module
|
|
f'{self.NAME}_basic_wrapper', # the indirect variant
|
|
f'{self.NAME}_basic_copy', # the direct variant
|
|
]:
|
|
self.add_module_cleanup(name)
|
|
with self.subTest(name):
|
|
loaded = self.load(name)
|
|
reloaded = self.re_load(name, loaded.module)
|
|
|
|
self.check_common(loaded)
|
|
self.check_common(reloaded)
|
|
|
|
# Make sure the original __dict__ did not get replaced.
|
|
self.assertEqual(id(loaded.module.__dict__),
|
|
loaded.snapshot.ns_id)
|
|
self.assertEqual(loaded.snapshot.ns.__dict__,
|
|
loaded.module.__dict__)
|
|
|
|
self.assertEqual(reloaded.module.__spec__.name, reloaded.name)
|
|
self.assertEqual(reloaded.module.__name__,
|
|
reloaded.snapshot.ns.__name__)
|
|
|
|
self.assertIs(reloaded.module, loaded.module)
|
|
self.assertIs(reloaded.module.__dict__, loaded.module.__dict__)
|
|
# It only happens to be the same but that's good enough here.
|
|
# We really just want to verify that the re-loaded attrs
|
|
# didn't change.
|
|
self.assertIs(reloaded.snapshot.lookedup,
|
|
loaded.snapshot.lookedup)
|
|
self.assertEqual(reloaded.snapshot.state_initialized,
|
|
loaded.snapshot.state_initialized)
|
|
self.assertEqual(reloaded.snapshot.init_count,
|
|
loaded.snapshot.init_count)
|
|
|
|
self.assertIs(reloaded.snapshot.cached, reloaded.module)
|
|
|
|
def test_with_reinit_reloaded(self):
|
|
# The module's m_init func is run again.
|
|
self.maxDiff = None
|
|
|
|
# Keep a reference around.
|
|
basic = self.load(self.NAME)
|
|
|
|
for name, has_state in [
|
|
(f'{self.NAME}_with_reinit', False), # m_size == 0
|
|
(f'{self.NAME}_with_state', True), # m_size > 0
|
|
]:
|
|
self.add_module_cleanup(name)
|
|
with self.subTest(name=name, has_state=has_state):
|
|
loaded = self.load(name)
|
|
if has_state:
|
|
self.addCleanup(loaded.module._clear_module_state)
|
|
|
|
reloaded = self.re_load(name, loaded.module)
|
|
if has_state:
|
|
self.addCleanup(reloaded.module._clear_module_state)
|
|
|
|
self.check_common(loaded)
|
|
self.check_common(reloaded)
|
|
|
|
# Make sure the original __dict__ did not get replaced.
|
|
self.assertEqual(id(loaded.module.__dict__),
|
|
loaded.snapshot.ns_id)
|
|
self.assertEqual(loaded.snapshot.ns.__dict__,
|
|
loaded.module.__dict__)
|
|
|
|
self.assertEqual(reloaded.module.__spec__.name, reloaded.name)
|
|
self.assertEqual(reloaded.module.__name__,
|
|
reloaded.snapshot.ns.__name__)
|
|
|
|
self.assertIsNot(reloaded.module, loaded.module)
|
|
self.assertNotEqual(reloaded.module.__dict__,
|
|
loaded.module.__dict__)
|
|
self.assertIs(reloaded.snapshot.lookedup, reloaded.module)
|
|
if loaded.snapshot.state_initialized is None:
|
|
self.assertIs(reloaded.snapshot.state_initialized, None)
|
|
else:
|
|
self.assertGreater(reloaded.snapshot.state_initialized,
|
|
loaded.snapshot.state_initialized)
|
|
|
|
self.assertIs(reloaded.snapshot.cached, reloaded.module)
|
|
|
|
@unittest.skipIf(_testinternalcapi is None, "requires _testinternalcapi")
|
|
def test_check_state_first(self):
|
|
for variant in ['', '_with_reinit', '_with_state']:
|
|
name = f'{self.NAME}{variant}_check_cache_first'
|
|
with self.subTest(name):
|
|
mod = self._load_dynamic(name, self.ORIGIN)
|
|
self.assertEqual(mod.__name__, name)
|
|
sys.modules.pop(name, None)
|
|
_testinternalcapi.clear_extension(name, self.ORIGIN)
|
|
|
|
# Currently, for every single-phrase init module loaded
|
|
# in multiple interpreters, those interpreters share a
|
|
# PyModuleDef for that object, which can be a problem.
|
|
# Also, we test with a single-phase module that has global state,
|
|
# which is shared by all interpreters.
|
|
|
|
@requires_subinterpreters
|
|
def test_basic_multiple_interpreters_main_no_reset(self):
|
|
# without resetting; already loaded in main interpreter
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def may or may not be loaded already
|
|
# * module def not in _PyRuntime.imports.extensions
|
|
# * mod init func has not run yet (since reset, at least)
|
|
# * m_copy not set (hasn't been loaded yet or already cleared)
|
|
# * module's global state has not been initialized yet
|
|
# (or already cleared)
|
|
|
|
main_loaded = self.load(self.NAME)
|
|
_testsinglephase = main_loaded.module
|
|
# Attrs set after loading are not in m_copy.
|
|
_testsinglephase.spam = 'spam, spam, spam, spam, eggs, and spam'
|
|
|
|
self.check_common(main_loaded)
|
|
self.check_fresh(main_loaded)
|
|
|
|
interpid1 = self.add_subinterpreter()
|
|
interpid2 = self.add_subinterpreter()
|
|
|
|
# At this point:
|
|
# * alive in 1 interpreter (main)
|
|
# * module def in _PyRuntime.imports.extensions
|
|
# * mod init func ran for the first time (since reset, at least)
|
|
# * m_copy was copied from the main interpreter (was NULL)
|
|
# * module's global state was initialized
|
|
|
|
# Use an interpreter that gets destroyed right away.
|
|
loaded = self.import_in_subinterp()
|
|
self.check_common(loaded)
|
|
self.check_copied(loaded, main_loaded)
|
|
|
|
# At this point:
|
|
# * alive in 1 interpreter (main)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func ran again
|
|
# * m_copy is NULL (claered when the interpreter was destroyed)
|
|
# (was from main interpreter)
|
|
# * module's global state was updated, not reset
|
|
|
|
# Use a subinterpreter that sticks around.
|
|
loaded = self.import_in_subinterp(interpid1)
|
|
self.check_common(loaded)
|
|
self.check_copied(loaded, main_loaded)
|
|
|
|
# At this point:
|
|
# * alive in 2 interpreters (main, interp1)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func ran again
|
|
# * m_copy was copied from interp1
|
|
# * module's global state was updated, not reset
|
|
|
|
# Use a subinterpreter while the previous one is still alive.
|
|
loaded = self.import_in_subinterp(interpid2)
|
|
self.check_common(loaded)
|
|
self.check_copied(loaded, main_loaded)
|
|
|
|
# At this point:
|
|
# * alive in 3 interpreters (main, interp1, interp2)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func ran again
|
|
# * m_copy was copied from interp2 (was from interp1)
|
|
# * module's global state was updated, not reset
|
|
|
|
@no_rerun(reason="rerun not possible; module state is never cleared (see gh-102251)")
|
|
@requires_subinterpreters
|
|
def test_basic_multiple_interpreters_deleted_no_reset(self):
|
|
# without resetting; already loaded in a deleted interpreter
|
|
|
|
if Py_TRACE_REFS:
|
|
# It's a Py_TRACE_REFS build.
|
|
# This test breaks interpreter isolation a little,
|
|
# which causes problems on Py_TRACE_REF builds.
|
|
raise unittest.SkipTest('crashes on Py_TRACE_REFS builds')
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def may or may not be loaded already
|
|
# * module def not in _PyRuntime.imports.extensions
|
|
# * mod init func has not run yet (since reset, at least)
|
|
# * m_copy not set (hasn't been loaded yet or already cleared)
|
|
# * module's global state has not been initialized yet
|
|
# (or already cleared)
|
|
|
|
interpid1 = self.add_subinterpreter()
|
|
interpid2 = self.add_subinterpreter()
|
|
|
|
# First, load in the main interpreter but then completely clear it.
|
|
loaded_main = self.load(self.NAME)
|
|
loaded_main.module._clear_globals()
|
|
_testinternalcapi.clear_extension(self.NAME, self.ORIGIN)
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def loaded already
|
|
# * module def was in _PyRuntime.imports.extensions, but cleared
|
|
# * mod init func ran for the first time (since reset, at least)
|
|
# * m_copy was set, but cleared (was NULL)
|
|
# * module's global state was initialized but cleared
|
|
|
|
# Start with an interpreter that gets destroyed right away.
|
|
base = self.import_in_subinterp(
|
|
postscript='''
|
|
# Attrs set after loading are not in m_copy.
|
|
mod.spam = 'spam, spam, mash, spam, eggs, and spam'
|
|
''')
|
|
self.check_common(base)
|
|
self.check_fresh(base)
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def in _PyRuntime.imports.extensions
|
|
# * mod init func ran for the first time (since reset)
|
|
# * m_copy is still set (owned by main interpreter)
|
|
# * module's global state was initialized, not reset
|
|
|
|
# Use a subinterpreter that sticks around.
|
|
loaded_interp1 = self.import_in_subinterp(interpid1)
|
|
self.check_common(loaded_interp1)
|
|
self.check_copied(loaded_interp1, base)
|
|
|
|
# At this point:
|
|
# * alive in 1 interpreter (interp1)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func did not run again
|
|
# * m_copy was not changed
|
|
# * module's global state was not touched
|
|
|
|
# Use a subinterpreter while the previous one is still alive.
|
|
loaded_interp2 = self.import_in_subinterp(interpid2)
|
|
self.check_common(loaded_interp2)
|
|
self.check_copied(loaded_interp2, loaded_interp1)
|
|
|
|
# At this point:
|
|
# * alive in 2 interpreters (interp1, interp2)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func did not run again
|
|
# * m_copy was not changed
|
|
# * module's global state was not touched
|
|
|
|
@requires_subinterpreters
|
|
def test_basic_multiple_interpreters_reset_each(self):
|
|
# resetting between each interpreter
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def may or may not be loaded already
|
|
# * module def not in _PyRuntime.imports.extensions
|
|
# * mod init func has not run yet (since reset, at least)
|
|
# * m_copy not set (hasn't been loaded yet or already cleared)
|
|
# * module's global state has not been initialized yet
|
|
# (or already cleared)
|
|
|
|
interpid1 = self.add_subinterpreter()
|
|
interpid2 = self.add_subinterpreter()
|
|
|
|
# Use an interpreter that gets destroyed right away.
|
|
loaded = self.import_in_subinterp(
|
|
postscript='''
|
|
# Attrs set after loading are not in m_copy.
|
|
mod.spam = 'spam, spam, mash, spam, eggs, and spam'
|
|
''',
|
|
postcleanup=True,
|
|
)
|
|
self.check_common(loaded)
|
|
self.check_fresh(loaded)
|
|
|
|
# At this point:
|
|
# * alive in 0 interpreters
|
|
# * module def in _PyRuntime.imports.extensions
|
|
# * mod init func ran for the first time (since reset, at least)
|
|
# * m_copy is NULL (claered when the interpreter was destroyed)
|
|
# * module's global state was initialized, not reset
|
|
|
|
# Use a subinterpreter that sticks around.
|
|
loaded = self.import_in_subinterp(interpid1, postcleanup=True)
|
|
self.check_common(loaded)
|
|
self.check_fresh(loaded)
|
|
|
|
# At this point:
|
|
# * alive in 1 interpreter (interp1)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func ran again
|
|
# * m_copy was copied from interp1 (was NULL)
|
|
# * module's global state was initialized, not reset
|
|
|
|
# Use a subinterpreter while the previous one is still alive.
|
|
loaded = self.import_in_subinterp(interpid2, postcleanup=True)
|
|
self.check_common(loaded)
|
|
self.check_fresh(loaded)
|
|
|
|
# At this point:
|
|
# * alive in 2 interpreters (interp2, interp2)
|
|
# * module def still in _PyRuntime.imports.extensions
|
|
# * mod init func ran again
|
|
# * m_copy was copied from interp2 (was from interp1)
|
|
# * module's global state was initialized, not reset
|
|
|
|
|
|
@cpython_only
|
|
class CAPITests(unittest.TestCase):
|
|
def test_pyimport_addmodule(self):
|
|
# gh-105922: Test PyImport_AddModuleRef(), PyImport_AddModule()
|
|
# and PyImport_AddModuleObject()
|
|
_testcapi = import_module("_testcapi")
|
|
for name in (
|
|
'sys', # frozen module
|
|
'test', # package
|
|
__name__, # package.module
|
|
):
|
|
_testcapi.check_pyimport_addmodule(name)
|
|
|
|
def test_pyimport_addmodule_create(self):
|
|
# gh-105922: Test PyImport_AddModuleRef(), create a new module
|
|
_testcapi = import_module("_testcapi")
|
|
name = 'dontexist'
|
|
self.assertNotIn(name, sys.modules)
|
|
self.addCleanup(unload, name)
|
|
|
|
mod = _testcapi.check_pyimport_addmodule(name)
|
|
self.assertIs(mod, sys.modules[name])
|
|
|
|
|
|
if __name__ == '__main__':
|
|
# Test needs to be a package, so we can do relative imports.
|
|
unittest.main()
|