aboutsummaryrefslogtreecommitdiffstatshomepage
path: root/Python/gc_free_threading.c
diff options
context:
space:
mode:
authorSam Gross <colesbury@gmail.com>2025-01-29 10:40:51 -0500
committerGitHub <noreply@github.com>2025-01-29 10:40:51 -0500
commit5ff2fbc026b14eadd41b8c14d83bb1eb832292dd (patch)
tree24f154649f285275fd77f7c7c46470d5804eaac1 /Python/gc_free_threading.c
parenta29221675e7367608961c3484701ab2671ec6f3c (diff)
downloadcpython-5ff2fbc026b14eadd41b8c14d83bb1eb832292dd.tar.gz
cpython-5ff2fbc026b14eadd41b8c14d83bb1eb832292dd.zip
gh-129236: Use `stackpointer` in free threaded GC (#129240)
The stack pointers in interpreter frames are nearly always valid now, so use them when visiting each thread's frame. For now, don't collect objects with deferred references in the rare case that we see a frame with a NULL stack pointer.
Diffstat (limited to 'Python/gc_free_threading.c')
-rw-r--r--Python/gc_free_threading.c80
1 files changed, 52 insertions, 28 deletions
diff --git a/Python/gc_free_threading.c b/Python/gc_free_threading.c
index d1023d93510..211d52d6cc7 100644
--- a/Python/gc_free_threading.c
+++ b/Python/gc_free_threading.c
@@ -67,6 +67,10 @@ struct collection_state {
PyInterpreterState *interp;
GCState *gcstate;
_PyGC_Reason reason;
+ // GH-129236: If we see an active frame without a valid stack pointer,
+ // we can't collect objects with deferred references because we may not
+ // see all references.
+ int skip_deferred_objects;
Py_ssize_t collected;
Py_ssize_t uncollectable;
Py_ssize_t long_lived_total;
@@ -413,9 +417,6 @@ gc_visit_heaps(PyInterpreterState *interp, mi_block_visit_fun *visitor,
static inline void
gc_visit_stackref(_PyStackRef stackref)
{
- // Note: we MUST check that it is deferred before checking the rest.
- // Otherwise we might read into invalid memory due to non-deferred references
- // being dead already.
if (PyStackRef_IsDeferred(stackref) && !PyStackRef_IsNull(stackref)) {
PyObject *obj = PyStackRef_AsPyObjectBorrow(stackref);
if (_PyObject_GC_IS_TRACKED(obj) && !gc_is_frozen(obj)) {
@@ -426,20 +427,27 @@ gc_visit_stackref(_PyStackRef stackref)
// Add 1 to the gc_refs for every deferred reference on each thread's stack.
static void
-gc_visit_thread_stacks(PyInterpreterState *interp)
+gc_visit_thread_stacks(PyInterpreterState *interp, struct collection_state *state)
{
_Py_FOR_EACH_TSTATE_BEGIN(interp, p) {
for (_PyInterpreterFrame *f = p->current_frame; f != NULL; f = f->previous) {
- PyObject *executable = PyStackRef_AsPyObjectBorrow(f->f_executable);
- if (executable == NULL || !PyCode_Check(executable)) {
+ if (f->owner >= FRAME_OWNED_BY_INTERPRETER) {
+ continue;
+ }
+
+ _PyStackRef *top = f->stackpointer;
+ if (top == NULL) {
+ // GH-129236: The stackpointer may be NULL in cases where
+ // the GC is run during a PyStackRef_CLOSE() call. Skip this
+ // frame and don't collect objects with deferred references.
+ state->skip_deferred_objects = 1;
continue;
}
- PyCodeObject *co = (PyCodeObject *)executable;
- int max_stack = co->co_nlocalsplus + co->co_stacksize;
gc_visit_stackref(f->f_executable);
- for (int i = 0; i < max_stack; i++) {
- gc_visit_stackref(f->localsplus[i]);
+ while (top != f->localsplus) {
+ --top;
+ gc_visit_stackref(*top);
}
}
}
@@ -519,10 +527,7 @@ gc_abort_mark_alive(PyInterpreterState *interp,
static int
gc_visit_stackref_mark_alive(_PyObjectStack *stack, _PyStackRef stackref)
{
- // Note: we MUST check that it is deferred before checking the rest.
- // Otherwise we might read into invalid memory due to non-deferred references
- // being dead already.
- if (PyStackRef_IsDeferred(stackref) && !PyStackRef_IsNull(stackref)) {
+ if (!PyStackRef_IsNull(stackref)) {
PyObject *op = PyStackRef_AsPyObjectBorrow(stackref);
if (mark_alive_stack_push(op, stack) < 0) {
return -1;
@@ -534,27 +539,37 @@ gc_visit_stackref_mark_alive(_PyObjectStack *stack, _PyStackRef stackref)
static int
gc_visit_thread_stacks_mark_alive(PyInterpreterState *interp, _PyObjectStack *stack)
{
+ int err = 0;
_Py_FOR_EACH_TSTATE_BEGIN(interp, p) {
for (_PyInterpreterFrame *f = p->current_frame; f != NULL; f = f->previous) {
- PyObject *executable = PyStackRef_AsPyObjectBorrow(f->f_executable);
- if (executable == NULL || !PyCode_Check(executable)) {
+ if (f->owner >= FRAME_OWNED_BY_INTERPRETER) {
continue;
}
- PyCodeObject *co = (PyCodeObject *)executable;
- int max_stack = co->co_nlocalsplus + co->co_stacksize;
+ if (f->stackpointer == NULL) {
+ // GH-129236: The stackpointer may be NULL in cases where
+ // the GC is run during a PyStackRef_CLOSE() call. Skip this
+ // frame for now.
+ continue;
+ }
+
+ _PyStackRef *top = f->stackpointer;
if (gc_visit_stackref_mark_alive(stack, f->f_executable) < 0) {
- return -1;
+ err = -1;
+ goto exit;
}
- for (int i = 0; i < max_stack; i++) {
- if (gc_visit_stackref_mark_alive(stack, f->localsplus[i]) < 0) {
- return -1;
+ while (top != f->localsplus) {
+ --top;
+ if (gc_visit_stackref_mark_alive(stack, *top) < 0) {
+ err = -1;
+ goto exit;
}
}
}
}
+exit:
_Py_FOR_EACH_TSTATE_END(interp);
- return 0;
+ return err;
}
#endif // GC_MARK_ALIVE_STACKS
#endif // GC_ENABLE_MARK_ALIVE
@@ -789,14 +804,23 @@ mark_heap_visitor(const mi_heap_t *heap, const mi_heap_area_t *area,
return true;
}
- if (gc_is_alive(op)) {
+ _PyObject_ASSERT_WITH_MSG(op, gc_get_refs(op) >= 0,
+ "refcount is too small");
+
+ if (gc_is_alive(op) || !gc_is_unreachable(op)) {
+ // Object was already marked as reachable.
return true;
}
- _PyObject_ASSERT_WITH_MSG(op, gc_get_refs(op) >= 0,
- "refcount is too small");
+ // GH-129236: If we've seen an active frame without a valid stack pointer,
+ // then we can't collect objects with deferred references because we may
+ // have missed some reference to the object on the stack. In that case,
+ // treat the object as reachable even if gc_refs is zero.
+ struct collection_state *state = (struct collection_state *)args;
+ int keep_alive = (state->skip_deferred_objects &&
+ _PyObject_HasDeferredRefcount(op));
- if (gc_is_unreachable(op) && gc_get_refs(op) != 0) {
+ if (gc_get_refs(op) != 0 || keep_alive) {
// Object is reachable but currently marked as unreachable.
// Mark it as reachable and traverse its pointers to find
// any other object that may be directly reachable from it.
@@ -985,7 +1009,7 @@ deduce_unreachable_heap(PyInterpreterState *interp,
#endif
// Visit the thread stacks to account for any deferred references.
- gc_visit_thread_stacks(interp);
+ gc_visit_thread_stacks(interp, state);
// Transitively mark reachable objects by clearing the
// _PyGC_BITS_UNREACHABLE flag.