From 656a1c81083b76b9d998c983f4329348a65985d3 Mon Sep 17 00:00:00 2001 From: Xie Yanbo Date: Thu, 20 Jun 2024 11:16:14 +0800 Subject: [PATCH] Fix typos in comments (#120481) --- Include/internal/pycore_frame.h | 2 +- Include/internal/pycore_gc.h | 2 +- Python/qsbr.c | 4 ++-- Python/specialize.c | 2 +- 4 files changed, 5 insertions(+), 5 deletions(-) diff --git a/Include/internal/pycore_frame.h b/Include/internal/pycore_frame.h index 994900c007f..e4eb893263c 100644 --- a/Include/internal/pycore_frame.h +++ b/Include/internal/pycore_frame.h @@ -214,7 +214,7 @@ _PyFrame_MakeAndSetFrameObject(_PyInterpreterFrame *frame); /* Gets the PyFrameObject for this frame, lazily * creating it if necessary. - * Returns a borrowed referennce */ + * Returns a borrowed reference */ static inline PyFrameObject * _PyFrame_GetFrameObject(_PyInterpreterFrame *frame) { diff --git a/Include/internal/pycore_gc.h b/Include/internal/pycore_gc.h index ba8b8e1903f..28e34d38096 100644 --- a/Include/internal/pycore_gc.h +++ b/Include/internal/pycore_gc.h @@ -346,7 +346,7 @@ struct _gc_runtime_state { Py_ssize_t long_lived_pending; /* gh-117783: Deferred reference counting is not fully implemented yet, so - as a temporary measure we treat objects using deferred referenence + as a temporary measure we treat objects using deferred reference counting as immortal. The value may be zero, one, or a negative number: 0: immortalize deferred RC objects once the first thread is created 1: immortalize all deferred RC objects immediately diff --git a/Python/qsbr.c b/Python/qsbr.c index a7321154a62..a40219acfe2 100644 --- a/Python/qsbr.c +++ b/Python/qsbr.c @@ -2,7 +2,7 @@ * Implementation of safe memory reclamation scheme using * quiescent states. * - * This is dervied from the "GUS" safe memory reclamation technique + * This is derived from the "GUS" safe memory reclamation technique * in FreeBSD written by Jeffrey Roberson. It is heavily modified. Any bugs * in this code are likely due to the modifications. * @@ -238,7 +238,7 @@ _Py_qsbr_unregister(PyThreadState *tstate) struct _PyThreadStateImpl *tstate_imp = (_PyThreadStateImpl*) tstate; // gh-119369: GIL must be released (if held) to prevent deadlocks, because - // we might not have an active tstate, which means taht blocking on PyMutex + // we might not have an active tstate, which means that blocking on PyMutex // locks will not implicitly release the GIL. assert(!tstate->_status.holds_gil); diff --git a/Python/specialize.c b/Python/specialize.c index 5e22ebfe02d..ad2f74788b3 100644 --- a/Python/specialize.c +++ b/Python/specialize.c @@ -1301,7 +1301,7 @@ PyObject *descr, DescriptorClassification kind, bool is_method) } /* Cache entries must be unsigned values, so we offset the * dictoffset by MANAGED_DICT_OFFSET. - * We do the reverese offset in LOAD_ATTR_METHOD_LAZY_DICT */ + * We do the reverse offset in LOAD_ATTR_METHOD_LAZY_DICT */ dictoffset -= MANAGED_DICT_OFFSET; assert(((uint16_t)dictoffset) == dictoffset); cache->dict_offset = (uint16_t)dictoffset;