mirror of
https://github.com/python/cpython.git
synced 2025-07-12 22:05:16 +00:00

* Revert "[3.11] gh-101634: regrtest reports decoding error as failed test (#106169) (#106175)" This reverts commitd5418e97fc
. * Revert "[3.11] bpo-46523: fix tests rerun when `setUp[Class|Module]` fails (GH-30895) (GH-103342)" This reverts commitecb09a8496
. * Revert "gh-95027: Fix regrtest stdout encoding on Windows (GH-98492)" This reverts commitb2aa28eec5
. * Revert "[3.11] gh-94026: Buffer regrtest worker stdout in temporary file (GH-94253) (GH-94408)" This reverts commit0122ab235b
. * Revert "Run Tools/scripts/reindent.py (GH-94225)" This reverts commitf0f3a424af
. * Revert "gh-94052: Don't re-run failed tests with --python option (GH-94054)" This reverts commit1347607db1
. * Revert "[3.11] gh-84461: Fix Emscripten umask and permission issues (GH-94002) (GH-94006)" This reverts commit1073184918
. * gh-93353: regrtest checks for leaked temporary files (#93776) When running tests with -jN, create a temporary directory per process and mark a test as "environment changed" if a test leaks a temporary file or directory. (cherry picked from commite566ce5496
) * gh-93353: Fix regrtest for -jN with N >= 2 (GH-93813) (cherry picked from commit36934a16e8
) * gh-93353: regrtest supports checking tmp files with -j2 (#93909) regrtest now also implements checking for leaked temporary files and directories when using -jN for N >= 2. Use tempfile.mkdtemp() to create the temporary directory. Skip this check on WASI. (cherry picked from commit4f85cec9e2
) * gh-84461: Fix Emscripten umask and permission issues (GH-94002) - Emscripten's default umask is too strict, see https://github.com/emscripten-core/emscripten/issues/17269 - getuid/getgid and geteuid/getegid are stubs that always return 0 (root). Disable effective uid/gid syscalls and fix tests that use chmod() current user. - Cannot drop X bit from directory. (cherry picked from commit2702e408fd
) * gh-94052: Don't re-run failed tests with --python option (#94054) (cherry picked from commit0ff7b996f5
) * Run Tools/scripts/reindent.py (#94225) Reindent files which were not properly formatted (PEP 8: 4 spaces). Remove also some trailing spaces. (cherry picked from commite87ada48a9
) * gh-94026: Buffer regrtest worker stdout in temporary file (GH-94253) Co-authored-by: Victor Stinner <vstinner@python.org> (cherry picked from commit199ba23324
) * gh-96465: Clear fractions hash lru_cache under refleak testing (GH-96689) Automerge-Triggered-By: GH:zware (cherry picked from commit9c8f379433
) * gh-95027: Fix regrtest stdout encoding on Windows (#98492) On Windows, when the Python test suite is run with the -jN option, the ANSI code page is now used as the encoding for the stdout temporary file, rather than using UTF-8 which can lead to decoding errors. (cherry picked from commitec1f6f5f13
) * gh-98903: Test suite fails with exit code 4 if no tests ran (#98904) The Python test suite now fails wit exit code 4 if no tests ran. It should help detecting typos in test names and test methods. * Add "EXITCODE_" constants to Lib/test/libregrtest/main.py. * Fix a typo: "NO TEST RUN" becomes "NO TESTS RAN" (cherry picked from commitc76db37c0d
) * gh-100086: Add build info to test.libregrtest (#100093) The Python test runner (libregrtest) now logs Python build information like "debug" vs "release" build, or LTO and PGO optimizations. (cherry picked from commit3c89202247
) * bpo-46523: fix tests rerun when `setUp[Class|Module]` fails (#30895) Co-authored-by: Jelle Zijlstra <jelle.zijlstra@gmail.com> Co-authored-by: Łukasz Langa <lukasz@langa.pl> (cherry picked from commit995386071f
) * gh-82054: allow test runner to split test_asyncio to execute in parallel by sharding. (#103927) This runs test_asyncio sub-tests in parallel using sharding from Cinder. This suite is typically the longest-pole in runs because it is a test package with a lot of further sub-tests otherwise run serially. By breaking out the sub-tests as independent modules we can run a lot more in parallel. After porting we can see the direct impact on a multicore system. Without this change: Running make test is 5 min 26 seconds With this change: Running make test takes 3 min 39 seconds That'll vary based on system and parallelism. On a `-j 4` run similar to what CI and buildbot systems often do, it reduced the overall test suite completion latency by 10%. The drawbacks are that this implementation is hacky and due to the sorting of the tests it obscures when the asyncio tests occur and involves changing CPython test infrastructure but, the wall time saved it is worth it, especially in low-core count CI runs as it pulls a long tail. The win for productivity and reserved CI resource usage is significant. Future tests that deserve to be refactored into split up suites to benefit from are test_concurrent_futures and the way the _test_multiprocessing suite gets run for all start methods. As exposed by passing the -o flag to python -m test to get a list of the 10 longest running tests. --------- Co-authored-by: Carl Meyer <carl@oddbird.net> Co-authored-by: Gregory P. Smith <greg@krypto.org> [Google, LLC] (cherry picked from commit9e011e7c77
) * Display the sanitizer config in the regrtest header. (#105301) Display the sanitizers present in libregrtest. Having this in the CI output for tests with the relevant environment variable displayed will help make it easier to do what we need to create an equivalent local test run. (cherry picked from commit852348ab65
) * gh-101634: regrtest reports decoding error as failed test (#106169) When running the Python test suite with -jN option, if a worker stdout cannot be decoded from the locale encoding report a failed testn so the exitcode is non-zero. (cherry picked from commit2ac3eec103
) * gh-108223: test.pythoninfo and libregrtest log Py_NOGIL (#108238) Enable with --disable-gil --without-pydebug: $ make pythoninfo|grep NOGIL sysconfig[Py_NOGIL]: 1 $ ./python -m test ... == Python build: nogil debug ... (cherry picked from commit5afe0c17ca
) * gh-90791: test.pythoninfo logs ASAN_OPTIONS env var (#108289) * Cleanup libregrtest code logging ASAN_OPTIONS. * Fix a typo on "ASAN_OPTIONS" vs "MSAN_OPTIONS". (cherry picked from commit3a1ac87f8f
) * gh-108388: regrtest splits test_asyncio package (#108393) Currently, test_asyncio package is only splitted into sub-tests when using command "./python -m test". With this change, it's also splitted when passing it on the command line: "./python -m test test_asyncio". Remove the concept of "STDTESTS". Python is now mature enough to not have to bother with that anymore. Removing STDTESTS simplify the code. (cherry picked from commit174e9da083
) * regrtest computes statistics (#108793) test_netrc, test_pep646_syntax and test_xml_etree now return results in the test_main() function. Changes: * Rewrite TestResult as a dataclass with a new State class. * Add test.support.TestStats class and Regrtest.stats_dict attribute. * libregrtest.runtest functions now modify a TestResult instance in-place. * libregrtest summary lists the number of run tests and skipped tests, and denied resources. * Add TestResult.has_meaningful_duration() method. * Compute TestResult duration in the upper function. * Use time.perf_counter() instead of time.monotonic(). * Regrtest: rename 'resource_denieds' attribute to 'resource_denied'. * Rename CHILD_ERROR to MULTIPROCESSING_ERROR. * Use match/case syntadx to have different code depending on the test state. Co-authored-by: Alex Waygood <Alex.Waygood@Gmail.com> (cherry picked from commitd4e534cbb3
) * gh-108822: Add Changelog entry for regrtest statistics (#108821) --------- Co-authored-by: Christian Heimes <christian@python.org> Co-authored-by: Zachary Ware <zach@python.org> Co-authored-by: Nikita Sobolev <mail@sobolevn.me> Co-authored-by: Joshua Herman <zitterbewegung@gmail.com> Co-authored-by: Gregory P. Smith <greg@krypto.org>
937 lines
24 KiB
Python
937 lines
24 KiB
Python
"""
|
|
Collect various information about Python to help debugging test failures.
|
|
"""
|
|
from __future__ import print_function
|
|
import errno
|
|
import re
|
|
import sys
|
|
import traceback
|
|
import unittest
|
|
import warnings
|
|
|
|
|
|
MS_WINDOWS = (sys.platform == 'win32')
|
|
|
|
|
|
def normalize_text(text):
|
|
if text is None:
|
|
return None
|
|
text = str(text)
|
|
text = re.sub(r'\s+', ' ', text)
|
|
return text.strip()
|
|
|
|
|
|
class PythonInfo:
|
|
def __init__(self):
|
|
self.info = {}
|
|
|
|
def add(self, key, value):
|
|
if key in self.info:
|
|
raise ValueError("duplicate key: %r" % key)
|
|
|
|
if value is None:
|
|
return
|
|
|
|
if not isinstance(value, int):
|
|
if not isinstance(value, str):
|
|
# convert other objects like sys.flags to string
|
|
value = str(value)
|
|
|
|
value = value.strip()
|
|
if not value:
|
|
return
|
|
|
|
self.info[key] = value
|
|
|
|
def get_infos(self):
|
|
"""
|
|
Get information as a key:value dictionary where values are strings.
|
|
"""
|
|
return {key: str(value) for key, value in self.info.items()}
|
|
|
|
|
|
def copy_attributes(info_add, obj, name_fmt, attributes, *, formatter=None):
|
|
for attr in attributes:
|
|
value = getattr(obj, attr, None)
|
|
if value is None:
|
|
continue
|
|
name = name_fmt % attr
|
|
if formatter is not None:
|
|
value = formatter(attr, value)
|
|
info_add(name, value)
|
|
|
|
|
|
def copy_attr(info_add, name, mod, attr_name):
|
|
try:
|
|
value = getattr(mod, attr_name)
|
|
except AttributeError:
|
|
return
|
|
info_add(name, value)
|
|
|
|
|
|
def call_func(info_add, name, mod, func_name, *, formatter=None):
|
|
try:
|
|
func = getattr(mod, func_name)
|
|
except AttributeError:
|
|
return
|
|
value = func()
|
|
if formatter is not None:
|
|
value = formatter(value)
|
|
info_add(name, value)
|
|
|
|
|
|
def collect_sys(info_add):
|
|
attributes = (
|
|
'_emscripten_info',
|
|
'_framework',
|
|
'abiflags',
|
|
'api_version',
|
|
'builtin_module_names',
|
|
'byteorder',
|
|
'dont_write_bytecode',
|
|
'executable',
|
|
'flags',
|
|
'float_info',
|
|
'float_repr_style',
|
|
'hash_info',
|
|
'hexversion',
|
|
'implementation',
|
|
'int_info',
|
|
'maxsize',
|
|
'maxunicode',
|
|
'path',
|
|
'platform',
|
|
'platlibdir',
|
|
'prefix',
|
|
'thread_info',
|
|
'version',
|
|
'version_info',
|
|
'winver',
|
|
)
|
|
copy_attributes(info_add, sys, 'sys.%s', attributes)
|
|
|
|
call_func(info_add, 'sys.androidapilevel', sys, 'getandroidapilevel')
|
|
call_func(info_add, 'sys.windowsversion', sys, 'getwindowsversion')
|
|
|
|
encoding = sys.getfilesystemencoding()
|
|
if hasattr(sys, 'getfilesystemencodeerrors'):
|
|
encoding = '%s/%s' % (encoding, sys.getfilesystemencodeerrors())
|
|
info_add('sys.filesystem_encoding', encoding)
|
|
|
|
for name in ('stdin', 'stdout', 'stderr'):
|
|
stream = getattr(sys, name)
|
|
if stream is None:
|
|
continue
|
|
encoding = getattr(stream, 'encoding', None)
|
|
if not encoding:
|
|
continue
|
|
errors = getattr(stream, 'errors', None)
|
|
if errors:
|
|
encoding = '%s/%s' % (encoding, errors)
|
|
info_add('sys.%s.encoding' % name, encoding)
|
|
|
|
# Were we compiled --with-pydebug?
|
|
Py_DEBUG = hasattr(sys, 'gettotalrefcount')
|
|
if Py_DEBUG:
|
|
text = 'Yes (sys.gettotalrefcount() present)'
|
|
else:
|
|
text = 'No (sys.gettotalrefcount() missing)'
|
|
info_add('build.Py_DEBUG', text)
|
|
|
|
# Were we compiled --with-trace-refs?
|
|
Py_TRACE_REFS = hasattr(sys, 'getobjects')
|
|
if Py_TRACE_REFS:
|
|
text = 'Yes (sys.getobjects() present)'
|
|
else:
|
|
text = 'No (sys.getobjects() missing)'
|
|
info_add('build.Py_TRACE_REFS', text)
|
|
|
|
|
|
def collect_platform(info_add):
|
|
import platform
|
|
|
|
arch = platform.architecture()
|
|
arch = ' '.join(filter(bool, arch))
|
|
info_add('platform.architecture', arch)
|
|
|
|
info_add('platform.python_implementation',
|
|
platform.python_implementation())
|
|
info_add('platform.platform',
|
|
platform.platform(aliased=True))
|
|
|
|
libc_ver = ('%s %s' % platform.libc_ver()).strip()
|
|
if libc_ver:
|
|
info_add('platform.libc_ver', libc_ver)
|
|
|
|
|
|
def collect_locale(info_add):
|
|
import locale
|
|
|
|
info_add('locale.getencoding', locale.getencoding())
|
|
|
|
|
|
def collect_builtins(info_add):
|
|
info_add('builtins.float.float_format', float.__getformat__("float"))
|
|
info_add('builtins.float.double_format', float.__getformat__("double"))
|
|
|
|
|
|
def collect_urandom(info_add):
|
|
import os
|
|
|
|
if hasattr(os, 'getrandom'):
|
|
# PEP 524: Check if system urandom is initialized
|
|
try:
|
|
try:
|
|
os.getrandom(1, os.GRND_NONBLOCK)
|
|
state = 'ready (initialized)'
|
|
except BlockingIOError as exc:
|
|
state = 'not seeded yet (%s)' % exc
|
|
info_add('os.getrandom', state)
|
|
except OSError as exc:
|
|
# Python was compiled on a more recent Linux version
|
|
# than the current Linux kernel: ignore OSError(ENOSYS)
|
|
if exc.errno != errno.ENOSYS:
|
|
raise
|
|
|
|
|
|
def collect_os(info_add):
|
|
import os
|
|
|
|
def format_attr(attr, value):
|
|
if attr in ('supports_follow_symlinks', 'supports_fd',
|
|
'supports_effective_ids'):
|
|
return str(sorted(func.__name__ for func in value))
|
|
else:
|
|
return value
|
|
|
|
attributes = (
|
|
'name',
|
|
'supports_bytes_environ',
|
|
'supports_effective_ids',
|
|
'supports_fd',
|
|
'supports_follow_symlinks',
|
|
)
|
|
copy_attributes(info_add, os, 'os.%s', attributes, formatter=format_attr)
|
|
|
|
for func in (
|
|
'cpu_count',
|
|
'getcwd',
|
|
'getegid',
|
|
'geteuid',
|
|
'getgid',
|
|
'getloadavg',
|
|
'getresgid',
|
|
'getresuid',
|
|
'getuid',
|
|
'uname',
|
|
):
|
|
call_func(info_add, 'os.%s' % func, os, func)
|
|
|
|
def format_groups(groups):
|
|
return ', '.join(map(str, groups))
|
|
|
|
call_func(info_add, 'os.getgroups', os, 'getgroups', formatter=format_groups)
|
|
|
|
if hasattr(os, 'getlogin'):
|
|
try:
|
|
login = os.getlogin()
|
|
except OSError:
|
|
# getlogin() fails with "OSError: [Errno 25] Inappropriate ioctl
|
|
# for device" on Travis CI
|
|
pass
|
|
else:
|
|
info_add("os.login", login)
|
|
|
|
# Environment variables used by the stdlib and tests. Don't log the full
|
|
# environment: filter to list to not leak sensitive information.
|
|
#
|
|
# HTTP_PROXY is not logged because it can contain a password.
|
|
ENV_VARS = frozenset((
|
|
"APPDATA",
|
|
"AR",
|
|
"ARCHFLAGS",
|
|
"ARFLAGS",
|
|
"AUDIODEV",
|
|
"CC",
|
|
"CFLAGS",
|
|
"COLUMNS",
|
|
"COMPUTERNAME",
|
|
"COMSPEC",
|
|
"CPP",
|
|
"CPPFLAGS",
|
|
"DISPLAY",
|
|
"DISTUTILS_DEBUG",
|
|
"DISTUTILS_USE_SDK",
|
|
"DYLD_LIBRARY_PATH",
|
|
"ENSUREPIP_OPTIONS",
|
|
"HISTORY_FILE",
|
|
"HOME",
|
|
"HOMEDRIVE",
|
|
"HOMEPATH",
|
|
"IDLESTARTUP",
|
|
"LANG",
|
|
"LDFLAGS",
|
|
"LDSHARED",
|
|
"LD_LIBRARY_PATH",
|
|
"LINES",
|
|
"MACOSX_DEPLOYMENT_TARGET",
|
|
"MAILCAPS",
|
|
"MAKEFLAGS",
|
|
"MIXERDEV",
|
|
"MSSDK",
|
|
"PATH",
|
|
"PATHEXT",
|
|
"PIP_CONFIG_FILE",
|
|
"PLAT",
|
|
"POSIXLY_CORRECT",
|
|
"PY_SAX_PARSER",
|
|
"ProgramFiles",
|
|
"ProgramFiles(x86)",
|
|
"RUNNING_ON_VALGRIND",
|
|
"SDK_TOOLS_BIN",
|
|
"SERVER_SOFTWARE",
|
|
"SHELL",
|
|
"SOURCE_DATE_EPOCH",
|
|
"SYSTEMROOT",
|
|
"TEMP",
|
|
"TERM",
|
|
"TILE_LIBRARY",
|
|
"TIX_LIBRARY",
|
|
"TMP",
|
|
"TMPDIR",
|
|
"TRAVIS",
|
|
"TZ",
|
|
"USERPROFILE",
|
|
"VIRTUAL_ENV",
|
|
"WAYLAND_DISPLAY",
|
|
"WINDIR",
|
|
"_PYTHON_HOST_PLATFORM",
|
|
"_PYTHON_PROJECT_BASE",
|
|
"_PYTHON_SYSCONFIGDATA_NAME",
|
|
"__PYVENV_LAUNCHER__",
|
|
|
|
# Sanitizer options
|
|
"ASAN_OPTIONS",
|
|
"LSAN_OPTIONS",
|
|
"MSAN_OPTIONS",
|
|
"TSAN_OPTIONS",
|
|
"UBSAN_OPTIONS",
|
|
))
|
|
for name, value in os.environ.items():
|
|
uname = name.upper()
|
|
if (uname in ENV_VARS
|
|
# Copy PYTHON* and LC_* variables
|
|
or uname.startswith(("PYTHON", "LC_"))
|
|
# Visual Studio: VS140COMNTOOLS
|
|
or (uname.startswith("VS") and uname.endswith("COMNTOOLS"))):
|
|
info_add('os.environ[%s]' % name, value)
|
|
|
|
if hasattr(os, 'umask'):
|
|
mask = os.umask(0)
|
|
os.umask(mask)
|
|
info_add("os.umask", '0o%03o' % mask)
|
|
|
|
|
|
def collect_pwd(info_add):
|
|
try:
|
|
import pwd
|
|
except ImportError:
|
|
return
|
|
import os
|
|
|
|
uid = os.getuid()
|
|
try:
|
|
entry = pwd.getpwuid(uid)
|
|
except KeyError:
|
|
entry = None
|
|
|
|
info_add('pwd.getpwuid(%s)'% uid,
|
|
entry if entry is not None else '<KeyError>')
|
|
|
|
if entry is None:
|
|
# there is nothing interesting to read if the current user identifier
|
|
# is not the password database
|
|
return
|
|
|
|
if hasattr(os, 'getgrouplist'):
|
|
groups = os.getgrouplist(entry.pw_name, entry.pw_gid)
|
|
groups = ', '.join(map(str, groups))
|
|
info_add('os.getgrouplist', groups)
|
|
|
|
|
|
def collect_readline(info_add):
|
|
try:
|
|
import readline
|
|
except ImportError:
|
|
return
|
|
|
|
def format_attr(attr, value):
|
|
if isinstance(value, int):
|
|
return "%#x" % value
|
|
else:
|
|
return value
|
|
|
|
attributes = (
|
|
"_READLINE_VERSION",
|
|
"_READLINE_RUNTIME_VERSION",
|
|
"_READLINE_LIBRARY_VERSION",
|
|
)
|
|
copy_attributes(info_add, readline, 'readline.%s', attributes,
|
|
formatter=format_attr)
|
|
|
|
if not hasattr(readline, "_READLINE_LIBRARY_VERSION"):
|
|
# _READLINE_LIBRARY_VERSION has been added to CPython 3.7
|
|
doc = getattr(readline, '__doc__', '')
|
|
if 'libedit readline' in doc:
|
|
info_add('readline.library', 'libedit readline')
|
|
elif 'GNU readline' in doc:
|
|
info_add('readline.library', 'GNU readline')
|
|
|
|
|
|
def collect_gdb(info_add):
|
|
import subprocess
|
|
|
|
try:
|
|
proc = subprocess.Popen(["gdb", "-nx", "--version"],
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
universal_newlines=True)
|
|
version = proc.communicate()[0]
|
|
if proc.returncode:
|
|
# ignore gdb failure: test_gdb will log the error
|
|
return
|
|
except OSError:
|
|
return
|
|
|
|
# Only keep the first line
|
|
version = version.splitlines()[0]
|
|
info_add('gdb_version', version)
|
|
|
|
|
|
def collect_tkinter(info_add):
|
|
try:
|
|
import _tkinter
|
|
except ImportError:
|
|
pass
|
|
else:
|
|
attributes = ('TK_VERSION', 'TCL_VERSION')
|
|
copy_attributes(info_add, _tkinter, 'tkinter.%s', attributes)
|
|
|
|
try:
|
|
import tkinter
|
|
except ImportError:
|
|
pass
|
|
else:
|
|
tcl = tkinter.Tcl()
|
|
patchlevel = tcl.call('info', 'patchlevel')
|
|
info_add('tkinter.info_patchlevel', patchlevel)
|
|
|
|
|
|
def collect_time(info_add):
|
|
import time
|
|
|
|
info_add('time.time', time.time())
|
|
|
|
attributes = (
|
|
'altzone',
|
|
'daylight',
|
|
'timezone',
|
|
'tzname',
|
|
)
|
|
copy_attributes(info_add, time, 'time.%s', attributes)
|
|
|
|
if hasattr(time, 'get_clock_info'):
|
|
for clock in ('clock', 'monotonic', 'perf_counter',
|
|
'process_time', 'thread_time', 'time'):
|
|
try:
|
|
# prevent DeprecatingWarning on get_clock_info('clock')
|
|
with warnings.catch_warnings(record=True):
|
|
clock_info = time.get_clock_info(clock)
|
|
except ValueError:
|
|
# missing clock like time.thread_time()
|
|
pass
|
|
else:
|
|
info_add('time.get_clock_info(%s)' % clock, clock_info)
|
|
|
|
|
|
def collect_curses(info_add):
|
|
try:
|
|
import curses
|
|
except ImportError:
|
|
return
|
|
|
|
copy_attr(info_add, 'curses.ncurses_version', curses, 'ncurses_version')
|
|
|
|
|
|
def collect_datetime(info_add):
|
|
try:
|
|
import datetime
|
|
except ImportError:
|
|
return
|
|
|
|
info_add('datetime.datetime.now', datetime.datetime.now())
|
|
|
|
|
|
def collect_sysconfig(info_add):
|
|
# On Windows, sysconfig is not reliable to get macros used
|
|
# to build Python
|
|
if MS_WINDOWS:
|
|
return
|
|
|
|
import sysconfig
|
|
|
|
for name in (
|
|
'ABIFLAGS',
|
|
'ANDROID_API_LEVEL',
|
|
'CC',
|
|
'CCSHARED',
|
|
'CFLAGS',
|
|
'CFLAGSFORSHARED',
|
|
'CONFIG_ARGS',
|
|
'HOST_GNU_TYPE',
|
|
'MACHDEP',
|
|
'MULTIARCH',
|
|
'OPT',
|
|
'PY_CFLAGS',
|
|
'PY_CFLAGS_NODIST',
|
|
'PY_CORE_LDFLAGS',
|
|
'PY_LDFLAGS',
|
|
'PY_LDFLAGS_NODIST',
|
|
'PY_STDMODULE_CFLAGS',
|
|
'Py_DEBUG',
|
|
'Py_ENABLE_SHARED',
|
|
'Py_NOGIL',
|
|
'SHELL',
|
|
'SOABI',
|
|
'prefix',
|
|
):
|
|
value = sysconfig.get_config_var(name)
|
|
if name == 'ANDROID_API_LEVEL' and not value:
|
|
# skip ANDROID_API_LEVEL=0
|
|
continue
|
|
value = normalize_text(value)
|
|
info_add('sysconfig[%s]' % name, value)
|
|
|
|
PY_CFLAGS = sysconfig.get_config_var('PY_CFLAGS')
|
|
NDEBUG = (PY_CFLAGS and '-DNDEBUG' in PY_CFLAGS)
|
|
if NDEBUG:
|
|
text = 'ignore assertions (macro defined)'
|
|
else:
|
|
text= 'build assertions (macro not defined)'
|
|
info_add('build.NDEBUG',text)
|
|
|
|
for name in (
|
|
'WITH_DOC_STRINGS',
|
|
'WITH_DTRACE',
|
|
'WITH_FREELISTS',
|
|
'WITH_PYMALLOC',
|
|
'WITH_VALGRIND',
|
|
):
|
|
value = sysconfig.get_config_var(name)
|
|
if value:
|
|
text = 'Yes'
|
|
else:
|
|
text = 'No'
|
|
info_add(f'build.{name}', text)
|
|
|
|
|
|
def collect_ssl(info_add):
|
|
import os
|
|
try:
|
|
import ssl
|
|
except ImportError:
|
|
return
|
|
try:
|
|
import _ssl
|
|
except ImportError:
|
|
_ssl = None
|
|
|
|
def format_attr(attr, value):
|
|
if attr.startswith('OP_'):
|
|
return '%#8x' % value
|
|
else:
|
|
return value
|
|
|
|
attributes = (
|
|
'OPENSSL_VERSION',
|
|
'OPENSSL_VERSION_INFO',
|
|
'HAS_SNI',
|
|
'OP_ALL',
|
|
'OP_NO_TLSv1_1',
|
|
)
|
|
copy_attributes(info_add, ssl, 'ssl.%s', attributes, formatter=format_attr)
|
|
|
|
for name, ctx in (
|
|
('SSLContext', ssl.SSLContext(ssl.PROTOCOL_TLS_CLIENT)),
|
|
('default_https_context', ssl._create_default_https_context()),
|
|
('stdlib_context', ssl._create_stdlib_context()),
|
|
):
|
|
attributes = (
|
|
'minimum_version',
|
|
'maximum_version',
|
|
'protocol',
|
|
'options',
|
|
'verify_mode',
|
|
)
|
|
copy_attributes(info_add, ctx, f'ssl.{name}.%s', attributes)
|
|
|
|
env_names = ["OPENSSL_CONF", "SSLKEYLOGFILE"]
|
|
if _ssl is not None and hasattr(_ssl, 'get_default_verify_paths'):
|
|
parts = _ssl.get_default_verify_paths()
|
|
env_names.extend((parts[0], parts[2]))
|
|
|
|
for name in env_names:
|
|
try:
|
|
value = os.environ[name]
|
|
except KeyError:
|
|
continue
|
|
info_add('ssl.environ[%s]' % name, value)
|
|
|
|
|
|
def collect_socket(info_add):
|
|
try:
|
|
import socket
|
|
except ImportError:
|
|
return
|
|
|
|
try:
|
|
hostname = socket.gethostname()
|
|
except (OSError, AttributeError):
|
|
# WASI SDK 16.0 does not have gethostname(2).
|
|
if sys.platform != "wasi":
|
|
raise
|
|
else:
|
|
info_add('socket.hostname', hostname)
|
|
|
|
|
|
def collect_sqlite(info_add):
|
|
try:
|
|
import sqlite3
|
|
except ImportError:
|
|
return
|
|
|
|
attributes = ('version', 'sqlite_version')
|
|
copy_attributes(info_add, sqlite3, 'sqlite3.%s', attributes)
|
|
|
|
|
|
def collect_zlib(info_add):
|
|
try:
|
|
import zlib
|
|
except ImportError:
|
|
return
|
|
|
|
attributes = ('ZLIB_VERSION', 'ZLIB_RUNTIME_VERSION')
|
|
copy_attributes(info_add, zlib, 'zlib.%s', attributes)
|
|
|
|
|
|
def collect_expat(info_add):
|
|
try:
|
|
from xml.parsers import expat
|
|
except ImportError:
|
|
return
|
|
|
|
attributes = ('EXPAT_VERSION',)
|
|
copy_attributes(info_add, expat, 'expat.%s', attributes)
|
|
|
|
|
|
def collect_decimal(info_add):
|
|
try:
|
|
import _decimal
|
|
except ImportError:
|
|
return
|
|
|
|
attributes = ('__libmpdec_version__',)
|
|
copy_attributes(info_add, _decimal, '_decimal.%s', attributes)
|
|
|
|
|
|
def collect_testcapi(info_add):
|
|
try:
|
|
import _testcapi
|
|
except ImportError:
|
|
return
|
|
|
|
call_func(info_add, 'pymem.allocator', _testcapi, 'pymem_getallocatorsname')
|
|
|
|
|
|
def collect_resource(info_add):
|
|
try:
|
|
import resource
|
|
except ImportError:
|
|
return
|
|
|
|
limits = [attr for attr in dir(resource) if attr.startswith('RLIMIT_')]
|
|
for name in limits:
|
|
key = getattr(resource, name)
|
|
value = resource.getrlimit(key)
|
|
info_add('resource.%s' % name, value)
|
|
|
|
call_func(info_add, 'resource.pagesize', resource, 'getpagesize')
|
|
|
|
|
|
def collect_test_socket(info_add):
|
|
try:
|
|
from test import test_socket
|
|
except (ImportError, unittest.SkipTest):
|
|
return
|
|
|
|
# all check attributes like HAVE_SOCKET_CAN
|
|
attributes = [name for name in dir(test_socket)
|
|
if name.startswith('HAVE_')]
|
|
copy_attributes(info_add, test_socket, 'test_socket.%s', attributes)
|
|
|
|
|
|
def collect_test_support(info_add):
|
|
try:
|
|
from test import support
|
|
except ImportError:
|
|
return
|
|
|
|
attributes = ('IPV6_ENABLED',)
|
|
copy_attributes(info_add, support, 'test_support.%s', attributes)
|
|
|
|
call_func(info_add, 'test_support._is_gui_available', support, '_is_gui_available')
|
|
call_func(info_add, 'test_support.python_is_optimized', support, 'python_is_optimized')
|
|
|
|
info_add('test_support.check_sanitizer(address=True)',
|
|
support.check_sanitizer(address=True))
|
|
info_add('test_support.check_sanitizer(memory=True)',
|
|
support.check_sanitizer(memory=True))
|
|
info_add('test_support.check_sanitizer(ub=True)',
|
|
support.check_sanitizer(ub=True))
|
|
|
|
|
|
def collect_cc(info_add):
|
|
import subprocess
|
|
import sysconfig
|
|
|
|
CC = sysconfig.get_config_var('CC')
|
|
if not CC:
|
|
return
|
|
|
|
try:
|
|
import shlex
|
|
args = shlex.split(CC)
|
|
except ImportError:
|
|
args = CC.split()
|
|
args.append('--version')
|
|
try:
|
|
proc = subprocess.Popen(args,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT,
|
|
universal_newlines=True)
|
|
except OSError:
|
|
# Cannot run the compiler, for example when Python has been
|
|
# cross-compiled and installed on the target platform where the
|
|
# compiler is missing.
|
|
return
|
|
|
|
stdout = proc.communicate()[0]
|
|
if proc.returncode:
|
|
# CC --version failed: ignore error
|
|
return
|
|
|
|
text = stdout.splitlines()[0]
|
|
text = normalize_text(text)
|
|
info_add('CC.version', text)
|
|
|
|
|
|
def collect_gdbm(info_add):
|
|
try:
|
|
from _gdbm import _GDBM_VERSION
|
|
except ImportError:
|
|
return
|
|
|
|
info_add('gdbm.GDBM_VERSION', '.'.join(map(str, _GDBM_VERSION)))
|
|
|
|
|
|
def collect_get_config(info_add):
|
|
# Get global configuration variables, _PyPreConfig and _PyCoreConfig
|
|
try:
|
|
from _testinternalcapi import get_configs
|
|
except ImportError:
|
|
return
|
|
|
|
all_configs = get_configs()
|
|
for config_type in sorted(all_configs):
|
|
config = all_configs[config_type]
|
|
for key in sorted(config):
|
|
info_add('%s[%s]' % (config_type, key), repr(config[key]))
|
|
|
|
|
|
def collect_subprocess(info_add):
|
|
import subprocess
|
|
copy_attributes(info_add, subprocess, 'subprocess.%s', ('_USE_POSIX_SPAWN',))
|
|
|
|
|
|
def collect_windows(info_add):
|
|
try:
|
|
import ctypes
|
|
except ImportError:
|
|
return
|
|
|
|
if not hasattr(ctypes, 'WinDLL'):
|
|
return
|
|
|
|
ntdll = ctypes.WinDLL('ntdll')
|
|
BOOLEAN = ctypes.c_ubyte
|
|
|
|
try:
|
|
RtlAreLongPathsEnabled = ntdll.RtlAreLongPathsEnabled
|
|
except AttributeError:
|
|
res = '<function not available>'
|
|
else:
|
|
RtlAreLongPathsEnabled.restype = BOOLEAN
|
|
RtlAreLongPathsEnabled.argtypes = ()
|
|
res = bool(RtlAreLongPathsEnabled())
|
|
info_add('windows.RtlAreLongPathsEnabled', res)
|
|
|
|
try:
|
|
import _winapi
|
|
dll_path = _winapi.GetModuleFileName(sys.dllhandle)
|
|
info_add('windows.dll_path', dll_path)
|
|
except (ImportError, AttributeError):
|
|
pass
|
|
|
|
import subprocess
|
|
try:
|
|
# When wmic.exe output is redirected to a pipe,
|
|
# it uses the OEM code page
|
|
proc = subprocess.Popen(["wmic", "os", "get", "Caption,Version", "/value"],
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
encoding="oem",
|
|
text=True)
|
|
output, stderr = proc.communicate()
|
|
if proc.returncode:
|
|
output = ""
|
|
except OSError:
|
|
pass
|
|
else:
|
|
for line in output.splitlines():
|
|
line = line.strip()
|
|
if line.startswith('Caption='):
|
|
line = line.removeprefix('Caption=').strip()
|
|
if line:
|
|
info_add('windows.version_caption', line)
|
|
elif line.startswith('Version='):
|
|
line = line.removeprefix('Version=').strip()
|
|
if line:
|
|
info_add('windows.version', line)
|
|
|
|
try:
|
|
proc = subprocess.Popen(["ver"], shell=True,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
text=True)
|
|
output = proc.communicate()[0]
|
|
if proc.returncode:
|
|
output = ""
|
|
except OSError:
|
|
return
|
|
else:
|
|
output = output.strip()
|
|
line = output.splitlines()[0]
|
|
if line:
|
|
info_add('windows.ver', line)
|
|
|
|
|
|
def collect_fips(info_add):
|
|
try:
|
|
import _hashlib
|
|
except ImportError:
|
|
_hashlib = None
|
|
|
|
if _hashlib is not None:
|
|
call_func(info_add, 'fips.openssl_fips_mode', _hashlib, 'get_fips_mode')
|
|
|
|
try:
|
|
with open("/proc/sys/crypto/fips_enabled", encoding="utf-8") as fp:
|
|
line = fp.readline().rstrip()
|
|
|
|
if line:
|
|
info_add('fips.linux_crypto_fips_enabled', line)
|
|
except OSError:
|
|
pass
|
|
|
|
|
|
def collect_info(info):
|
|
error = False
|
|
info_add = info.add
|
|
|
|
for collect_func in (
|
|
# collect_urandom() must be the first, to check the getrandom() status.
|
|
# Other functions may block on os.urandom() indirectly and so change
|
|
# its state.
|
|
collect_urandom,
|
|
|
|
collect_builtins,
|
|
collect_cc,
|
|
collect_curses,
|
|
collect_datetime,
|
|
collect_decimal,
|
|
collect_expat,
|
|
collect_fips,
|
|
collect_gdb,
|
|
collect_gdbm,
|
|
collect_get_config,
|
|
collect_locale,
|
|
collect_os,
|
|
collect_platform,
|
|
collect_pwd,
|
|
collect_readline,
|
|
collect_resource,
|
|
collect_socket,
|
|
collect_sqlite,
|
|
collect_ssl,
|
|
collect_subprocess,
|
|
collect_sys,
|
|
collect_sysconfig,
|
|
collect_testcapi,
|
|
collect_time,
|
|
collect_tkinter,
|
|
collect_windows,
|
|
collect_zlib,
|
|
|
|
# Collecting from tests should be last as they have side effects.
|
|
collect_test_socket,
|
|
collect_test_support,
|
|
):
|
|
try:
|
|
collect_func(info_add)
|
|
except Exception:
|
|
error = True
|
|
print("ERROR: %s() failed" % (collect_func.__name__),
|
|
file=sys.stderr)
|
|
traceback.print_exc(file=sys.stderr)
|
|
print(file=sys.stderr)
|
|
sys.stderr.flush()
|
|
|
|
return error
|
|
|
|
|
|
def dump_info(info, file=None):
|
|
title = "Python debug information"
|
|
print(title)
|
|
print("=" * len(title))
|
|
print()
|
|
|
|
infos = info.get_infos()
|
|
infos = sorted(infos.items())
|
|
for key, value in infos:
|
|
value = value.replace("\n", " ")
|
|
print("%s: %s" % (key, value))
|
|
print()
|
|
|
|
|
|
def main():
|
|
info = PythonInfo()
|
|
error = collect_info(info)
|
|
dump_info(info)
|
|
|
|
if error:
|
|
print("Collection failed: exit with error", file=sys.stderr)
|
|
sys.exit(1)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|