mirror of
https://github.com/python/cpython.git
synced 2025-08-04 08:59:19 +00:00
gh-122417: Implement per-thread heap type refcounts (#122418)
The free-threaded build partially stores heap type reference counts in distributed manner in per-thread arrays. This avoids reference count contention when creating or destroying instances. Co-authored-by: Ken Jin <kenjin@python.org>
This commit is contained in:
parent
1429651a06
commit
dc09301067
18 changed files with 427 additions and 69 deletions
|
@ -15,6 +15,7 @@
|
|||
#include "pycore_tstate.h" // _PyThreadStateImpl
|
||||
#include "pycore_weakref.h" // _PyWeakref_ClearRef()
|
||||
#include "pydtrace.h"
|
||||
#include "pycore_typeid.h" // _PyType_MergeThreadLocalRefcounts
|
||||
|
||||
#ifdef Py_GIL_DISABLED
|
||||
|
||||
|
@ -164,7 +165,15 @@ disable_deferred_refcounting(PyObject *op)
|
|||
{
|
||||
if (_PyObject_HasDeferredRefcount(op)) {
|
||||
op->ob_gc_bits &= ~_PyGC_BITS_DEFERRED;
|
||||
op->ob_ref_shared -= (1 << _Py_REF_SHARED_SHIFT);
|
||||
op->ob_ref_shared -= _Py_REF_SHARED(_Py_REF_DEFERRED, 0);
|
||||
|
||||
if (PyType_Check(op)) {
|
||||
// Disable thread-local refcounting for heap types
|
||||
PyTypeObject *type = (PyTypeObject *)op;
|
||||
if (PyType_HasFeature(type, Py_TPFLAGS_HEAPTYPE)) {
|
||||
_PyType_ReleaseId((PyHeapTypeObject *)op);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -328,16 +337,6 @@ merge_queued_objects(_PyThreadStateImpl *tstate, struct collection_state *state)
|
|||
}
|
||||
}
|
||||
|
||||
static void
|
||||
merge_all_queued_objects(PyInterpreterState *interp, struct collection_state *state)
|
||||
{
|
||||
HEAD_LOCK(&_PyRuntime);
|
||||
for (PyThreadState *p = interp->threads.head; p != NULL; p = p->next) {
|
||||
merge_queued_objects((_PyThreadStateImpl *)p, state);
|
||||
}
|
||||
HEAD_UNLOCK(&_PyRuntime);
|
||||
}
|
||||
|
||||
static void
|
||||
process_delayed_frees(PyInterpreterState *interp)
|
||||
{
|
||||
|
@ -389,7 +388,9 @@ update_refs(const mi_heap_t *heap, const mi_heap_area_t *area,
|
|||
}
|
||||
|
||||
Py_ssize_t refcount = Py_REFCNT(op);
|
||||
refcount -= _PyObject_HasDeferredRefcount(op);
|
||||
if (_PyObject_HasDeferredRefcount(op)) {
|
||||
refcount -= _Py_REF_DEFERRED;
|
||||
}
|
||||
_PyObject_ASSERT(op, refcount >= 0);
|
||||
|
||||
if (refcount > 0 && !_PyObject_HasDeferredRefcount(op)) {
|
||||
|
@ -754,10 +755,6 @@ _PyGC_Init(PyInterpreterState *interp)
|
|||
{
|
||||
GCState *gcstate = &interp->gc;
|
||||
|
||||
// gh-117783: immortalize objects that would use deferred refcounting
|
||||
// once the first non-main thread is created (but not in subinterpreters).
|
||||
gcstate->immortalize = _Py_IsMainInterpreter(interp) ? 0 : -1;
|
||||
|
||||
gcstate->garbage = PyList_New(0);
|
||||
if (gcstate->garbage == NULL) {
|
||||
return _PyStatus_NO_MEMORY();
|
||||
|
@ -1105,8 +1102,18 @@ gc_collect_internal(PyInterpreterState *interp, struct collection_state *state,
|
|||
state->gcstate->old[i-1].count = 0;
|
||||
}
|
||||
|
||||
// merge refcounts for all queued objects
|
||||
merge_all_queued_objects(interp, state);
|
||||
HEAD_LOCK(&_PyRuntime);
|
||||
for (PyThreadState *p = interp->threads.head; p != NULL; p = p->next) {
|
||||
_PyThreadStateImpl *tstate = (_PyThreadStateImpl *)p;
|
||||
|
||||
// merge per-thread refcount for types into the type's actual refcount
|
||||
_PyType_MergeThreadLocalRefcounts(tstate);
|
||||
|
||||
// merge refcounts for all queued objects
|
||||
merge_queued_objects(tstate, state);
|
||||
}
|
||||
HEAD_UNLOCK(&_PyRuntime);
|
||||
|
||||
process_delayed_frees(interp);
|
||||
|
||||
// Find unreachable objects
|
||||
|
@ -1835,32 +1842,6 @@ custom_visitor_wrapper(const mi_heap_t *heap, const mi_heap_area_t *area,
|
|||
return true;
|
||||
}
|
||||
|
||||
// gh-117783: Immortalize objects that use deferred reference counting to
|
||||
// temporarily work around scaling bottlenecks.
|
||||
static bool
|
||||
immortalize_visitor(const mi_heap_t *heap, const mi_heap_area_t *area,
|
||||
void *block, size_t block_size, void *args)
|
||||
{
|
||||
PyObject *op = op_from_block(block, args, false);
|
||||
if (op != NULL && _PyObject_HasDeferredRefcount(op)) {
|
||||
_Py_SetImmortal(op);
|
||||
op->ob_gc_bits &= ~_PyGC_BITS_DEFERRED;
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
void
|
||||
_PyGC_ImmortalizeDeferredObjects(PyInterpreterState *interp)
|
||||
{
|
||||
struct visitor_args args;
|
||||
_PyEval_StopTheWorld(interp);
|
||||
if (interp->gc.immortalize == 0) {
|
||||
gc_visit_heaps(interp, &immortalize_visitor, &args);
|
||||
interp->gc.immortalize = 1;
|
||||
}
|
||||
_PyEval_StartTheWorld(interp);
|
||||
}
|
||||
|
||||
void
|
||||
PyUnstable_GC_VisitObjects(gcvisitobjects_t callback, void *arg)
|
||||
{
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue