1
0
mirror of https://github.com/golang/go synced 2024-11-23 17:50:06 -07:00

runtime: avoid potential deadlock when tracing memory code

In reclaimChunk, the runtime is calling traceGCSweepDone() while holding the mheap
lock. traceGCSweepDone() can call traceEvent() and traceFlush(). These functions
not only can get various trace locks, but they may also do memory allocations
(runtime.newobject) that may end up getting the mheap lock. So, there may be
either a self-deadlock or a possible deadlock between multiple threads.

It seems better to release the mheap lock before calling traceGCSweepDone(). It is
fine to release the lock, since the operations to get the index of the chunk of
work to do are atomic. We already release the lock to call sweep, so there is no
new behavior for any of the callers of reclaimChunk.

With this change, mheap is a leaf lock (no other lock is ever acquired while it
is held).

Testing: besides normal all.bash, also ran all.bash with --long enabled, since
it does longer tests of runtime/trace.

Change-Id: I4f8cb66c24bb8d424f24d6c2305b4b8387409248
Reviewed-on: https://go-review.googlesource.com/c/go/+/207846
Reviewed-by: Austin Clements <austin@google.com>
Reviewed-by: Michael Knyszek <mknyszek@google.com>
This commit is contained in:
Dan Scales 2019-11-19 13:58:28 -08:00
parent 6b1a3f73ed
commit f266cce676

View File

@ -786,7 +786,9 @@ func (h *mheap) reclaim(npage uintptr) {
// reclaimChunk sweeps unmarked spans that start at page indexes [pageIdx, pageIdx+n). // reclaimChunk sweeps unmarked spans that start at page indexes [pageIdx, pageIdx+n).
// It returns the number of pages returned to the heap. // It returns the number of pages returned to the heap.
// //
// h.lock must be held and the caller must be non-preemptible. // h.lock must be held and the caller must be non-preemptible. Note: h.lock may be
// temporarily unlocked and re-locked in order to do sweeping or if tracing is
// enabled.
func (h *mheap) reclaimChunk(arenas []arenaIdx, pageIdx, n uintptr) uintptr { func (h *mheap) reclaimChunk(arenas []arenaIdx, pageIdx, n uintptr) uintptr {
// The heap lock must be held because this accesses the // The heap lock must be held because this accesses the
// heapArena.spans arrays using potentially non-live pointers. // heapArena.spans arrays using potentially non-live pointers.
@ -842,8 +844,10 @@ func (h *mheap) reclaimChunk(arenas []arenaIdx, pageIdx, n uintptr) uintptr {
n -= uintptr(len(inUse) * 8) n -= uintptr(len(inUse) * 8)
} }
if trace.enabled { if trace.enabled {
unlock(&h.lock)
// Account for pages scanned but not reclaimed. // Account for pages scanned but not reclaimed.
traceGCSweepSpan((n0 - nFreed) * pageSize) traceGCSweepSpan((n0 - nFreed) * pageSize)
lock(&h.lock)
} }
return nFreed return nFreed
} }