mirror of
https://github.com/golang/go
synced 2024-11-19 15:24:46 -07:00
b1d94c118f
Change-Id: Ic8c506289caaf6218494e5150d10002e0232feaa Reviewed-on: https://go-review.googlesource.com/85876 Run-TryBot: Austin Clements <austin@google.com> TryBot-Result: Gobot Gobot <gobot@golang.org> Reviewed-by: Rick Hudson <rlh@golang.org>
552 lines
14 KiB
Go
552 lines
14 KiB
Go
// Copyright 2009 The Go Authors. All rights reserved.
|
||
// Use of this source code is governed by a BSD-style
|
||
// license that can be found in the LICENSE file.
|
||
|
||
package runtime
|
||
|
||
import (
|
||
"runtime/internal/atomic"
|
||
"runtime/internal/sys"
|
||
"unsafe"
|
||
)
|
||
|
||
const (
|
||
_WorkbufSize = 2048 // in bytes; larger values result in less contention
|
||
|
||
// workbufAlloc is the number of bytes to allocate at a time
|
||
// for new workbufs. This must be a multiple of pageSize and
|
||
// should be a multiple of _WorkbufSize.
|
||
//
|
||
// Larger values reduce workbuf allocation overhead. Smaller
|
||
// values reduce heap fragmentation.
|
||
workbufAlloc = 32 << 10
|
||
)
|
||
|
||
func init() {
|
||
if workbufAlloc%pageSize != 0 || workbufAlloc%_WorkbufSize != 0 {
|
||
throw("bad workbufAlloc")
|
||
}
|
||
}
|
||
|
||
// Garbage collector work pool abstraction.
|
||
//
|
||
// This implements a producer/consumer model for pointers to grey
|
||
// objects. A grey object is one that is marked and on a work
|
||
// queue. A black object is marked and not on a work queue.
|
||
//
|
||
// Write barriers, root discovery, stack scanning, and object scanning
|
||
// produce pointers to grey objects. Scanning consumes pointers to
|
||
// grey objects, thus blackening them, and then scans them,
|
||
// potentially producing new pointers to grey objects.
|
||
|
||
// A gcWork provides the interface to produce and consume work for the
|
||
// garbage collector.
|
||
//
|
||
// A gcWork can be used on the stack as follows:
|
||
//
|
||
// (preemption must be disabled)
|
||
// gcw := &getg().m.p.ptr().gcw
|
||
// .. call gcw.put() to produce and gcw.get() to consume ..
|
||
// if gcBlackenPromptly {
|
||
// gcw.dispose()
|
||
// }
|
||
//
|
||
// It's important that any use of gcWork during the mark phase prevent
|
||
// the garbage collector from transitioning to mark termination since
|
||
// gcWork may locally hold GC work buffers. This can be done by
|
||
// disabling preemption (systemstack or acquirem).
|
||
type gcWork struct {
|
||
// wbuf1 and wbuf2 are the primary and secondary work buffers.
|
||
//
|
||
// This can be thought of as a stack of both work buffers'
|
||
// pointers concatenated. When we pop the last pointer, we
|
||
// shift the stack up by one work buffer by bringing in a new
|
||
// full buffer and discarding an empty one. When we fill both
|
||
// buffers, we shift the stack down by one work buffer by
|
||
// bringing in a new empty buffer and discarding a full one.
|
||
// This way we have one buffer's worth of hysteresis, which
|
||
// amortizes the cost of getting or putting a work buffer over
|
||
// at least one buffer of work and reduces contention on the
|
||
// global work lists.
|
||
//
|
||
// wbuf1 is always the buffer we're currently pushing to and
|
||
// popping from and wbuf2 is the buffer that will be discarded
|
||
// next.
|
||
//
|
||
// Invariant: Both wbuf1 and wbuf2 are nil or neither are.
|
||
wbuf1, wbuf2 *workbuf
|
||
|
||
// Bytes marked (blackened) on this gcWork. This is aggregated
|
||
// into work.bytesMarked by dispose.
|
||
bytesMarked uint64
|
||
|
||
// Scan work performed on this gcWork. This is aggregated into
|
||
// gcController by dispose and may also be flushed by callers.
|
||
scanWork int64
|
||
}
|
||
|
||
// Most of the methods of gcWork are go:nowritebarrierrec because the
|
||
// write barrier itself can invoke gcWork methods but the methods are
|
||
// not generally re-entrant. Hence, if a gcWork method invoked the
|
||
// write barrier while the gcWork was in an inconsistent state, and
|
||
// the write barrier in turn invoked a gcWork method, it could
|
||
// permanently corrupt the gcWork.
|
||
|
||
func (w *gcWork) init() {
|
||
w.wbuf1 = getempty()
|
||
wbuf2 := trygetfull()
|
||
if wbuf2 == nil {
|
||
wbuf2 = getempty()
|
||
}
|
||
w.wbuf2 = wbuf2
|
||
}
|
||
|
||
// put enqueues a pointer for the garbage collector to trace.
|
||
// obj must point to the beginning of a heap object or an oblet.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) put(obj uintptr) {
|
||
flushed := false
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
w.init()
|
||
wbuf = w.wbuf1
|
||
// wbuf is empty at this point.
|
||
} else if wbuf.nobj == len(wbuf.obj) {
|
||
w.wbuf1, w.wbuf2 = w.wbuf2, w.wbuf1
|
||
wbuf = w.wbuf1
|
||
if wbuf.nobj == len(wbuf.obj) {
|
||
putfull(wbuf)
|
||
wbuf = getempty()
|
||
w.wbuf1 = wbuf
|
||
flushed = true
|
||
}
|
||
}
|
||
|
||
wbuf.obj[wbuf.nobj] = obj
|
||
wbuf.nobj++
|
||
|
||
// If we put a buffer on full, let the GC controller know so
|
||
// it can encourage more workers to run. We delay this until
|
||
// the end of put so that w is in a consistent state, since
|
||
// enlistWorker may itself manipulate w.
|
||
if flushed && gcphase == _GCmark {
|
||
gcController.enlistWorker()
|
||
}
|
||
}
|
||
|
||
// putFast does a put and returns true if it can be done quickly
|
||
// otherwise it returns false and the caller needs to call put.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) putFast(obj uintptr) bool {
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
return false
|
||
} else if wbuf.nobj == len(wbuf.obj) {
|
||
return false
|
||
}
|
||
|
||
wbuf.obj[wbuf.nobj] = obj
|
||
wbuf.nobj++
|
||
return true
|
||
}
|
||
|
||
// putBatch performs a put on every pointer in obj. See put for
|
||
// constraints on these pointers.
|
||
//
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) putBatch(obj []uintptr) {
|
||
if len(obj) == 0 {
|
||
return
|
||
}
|
||
|
||
flushed := false
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
w.init()
|
||
wbuf = w.wbuf1
|
||
}
|
||
|
||
for len(obj) > 0 {
|
||
for wbuf.nobj == len(wbuf.obj) {
|
||
putfull(wbuf)
|
||
w.wbuf1, w.wbuf2 = w.wbuf2, getempty()
|
||
wbuf = w.wbuf1
|
||
flushed = true
|
||
}
|
||
n := copy(wbuf.obj[wbuf.nobj:], obj)
|
||
wbuf.nobj += n
|
||
obj = obj[n:]
|
||
}
|
||
|
||
if flushed && gcphase == _GCmark {
|
||
gcController.enlistWorker()
|
||
}
|
||
}
|
||
|
||
// tryGet dequeues a pointer for the garbage collector to trace.
|
||
//
|
||
// If there are no pointers remaining in this gcWork or in the global
|
||
// queue, tryGet returns 0. Note that there may still be pointers in
|
||
// other gcWork instances or other caches.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) tryGet() uintptr {
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
w.init()
|
||
wbuf = w.wbuf1
|
||
// wbuf is empty at this point.
|
||
}
|
||
if wbuf.nobj == 0 {
|
||
w.wbuf1, w.wbuf2 = w.wbuf2, w.wbuf1
|
||
wbuf = w.wbuf1
|
||
if wbuf.nobj == 0 {
|
||
owbuf := wbuf
|
||
wbuf = trygetfull()
|
||
if wbuf == nil {
|
||
return 0
|
||
}
|
||
putempty(owbuf)
|
||
w.wbuf1 = wbuf
|
||
}
|
||
}
|
||
|
||
wbuf.nobj--
|
||
return wbuf.obj[wbuf.nobj]
|
||
}
|
||
|
||
// tryGetFast dequeues a pointer for the garbage collector to trace
|
||
// if one is readily available. Otherwise it returns 0 and
|
||
// the caller is expected to call tryGet().
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) tryGetFast() uintptr {
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
return 0
|
||
}
|
||
if wbuf.nobj == 0 {
|
||
return 0
|
||
}
|
||
|
||
wbuf.nobj--
|
||
return wbuf.obj[wbuf.nobj]
|
||
}
|
||
|
||
// get dequeues a pointer for the garbage collector to trace, blocking
|
||
// if necessary to ensure all pointers from all queues and caches have
|
||
// been retrieved. get returns 0 if there are no pointers remaining.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) get() uintptr {
|
||
wbuf := w.wbuf1
|
||
if wbuf == nil {
|
||
w.init()
|
||
wbuf = w.wbuf1
|
||
// wbuf is empty at this point.
|
||
}
|
||
if wbuf.nobj == 0 {
|
||
w.wbuf1, w.wbuf2 = w.wbuf2, w.wbuf1
|
||
wbuf = w.wbuf1
|
||
if wbuf.nobj == 0 {
|
||
owbuf := wbuf
|
||
wbuf = getfull()
|
||
if wbuf == nil {
|
||
return 0
|
||
}
|
||
putempty(owbuf)
|
||
w.wbuf1 = wbuf
|
||
}
|
||
}
|
||
|
||
// TODO: This might be a good place to add prefetch code
|
||
|
||
wbuf.nobj--
|
||
return wbuf.obj[wbuf.nobj]
|
||
}
|
||
|
||
// dispose returns any cached pointers to the global queue.
|
||
// The buffers are being put on the full queue so that the
|
||
// write barriers will not simply reacquire them before the
|
||
// GC can inspect them. This helps reduce the mutator's
|
||
// ability to hide pointers during the concurrent mark phase.
|
||
//
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) dispose() {
|
||
if wbuf := w.wbuf1; wbuf != nil {
|
||
if wbuf.nobj == 0 {
|
||
putempty(wbuf)
|
||
} else {
|
||
putfull(wbuf)
|
||
}
|
||
w.wbuf1 = nil
|
||
|
||
wbuf = w.wbuf2
|
||
if wbuf.nobj == 0 {
|
||
putempty(wbuf)
|
||
} else {
|
||
putfull(wbuf)
|
||
}
|
||
w.wbuf2 = nil
|
||
}
|
||
if w.bytesMarked != 0 {
|
||
// dispose happens relatively infrequently. If this
|
||
// atomic becomes a problem, we should first try to
|
||
// dispose less and if necessary aggregate in a per-P
|
||
// counter.
|
||
atomic.Xadd64(&work.bytesMarked, int64(w.bytesMarked))
|
||
w.bytesMarked = 0
|
||
}
|
||
if w.scanWork != 0 {
|
||
atomic.Xaddint64(&gcController.scanWork, w.scanWork)
|
||
w.scanWork = 0
|
||
}
|
||
}
|
||
|
||
// balance moves some work that's cached in this gcWork back on the
|
||
// global queue.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) balance() {
|
||
if w.wbuf1 == nil {
|
||
return
|
||
}
|
||
if wbuf := w.wbuf2; wbuf.nobj != 0 {
|
||
putfull(wbuf)
|
||
w.wbuf2 = getempty()
|
||
} else if wbuf := w.wbuf1; wbuf.nobj > 4 {
|
||
w.wbuf1 = handoff(wbuf)
|
||
} else {
|
||
return
|
||
}
|
||
// We flushed a buffer to the full list, so wake a worker.
|
||
if gcphase == _GCmark {
|
||
gcController.enlistWorker()
|
||
}
|
||
}
|
||
|
||
// empty returns true if w has no mark work available.
|
||
//go:nowritebarrierrec
|
||
func (w *gcWork) empty() bool {
|
||
return w.wbuf1 == nil || (w.wbuf1.nobj == 0 && w.wbuf2.nobj == 0)
|
||
}
|
||
|
||
// Internally, the GC work pool is kept in arrays in work buffers.
|
||
// The gcWork interface caches a work buffer until full (or empty) to
|
||
// avoid contending on the global work buffer lists.
|
||
|
||
type workbufhdr struct {
|
||
node lfnode // must be first
|
||
nobj int
|
||
}
|
||
|
||
//go:notinheap
|
||
type workbuf struct {
|
||
workbufhdr
|
||
// account for the above fields
|
||
obj [(_WorkbufSize - unsafe.Sizeof(workbufhdr{})) / sys.PtrSize]uintptr
|
||
}
|
||
|
||
// workbuf factory routines. These funcs are used to manage the
|
||
// workbufs.
|
||
// If the GC asks for some work these are the only routines that
|
||
// make wbufs available to the GC.
|
||
|
||
func (b *workbuf) checknonempty() {
|
||
if b.nobj == 0 {
|
||
throw("workbuf is empty")
|
||
}
|
||
}
|
||
|
||
func (b *workbuf) checkempty() {
|
||
if b.nobj != 0 {
|
||
throw("workbuf is not empty")
|
||
}
|
||
}
|
||
|
||
// getempty pops an empty work buffer off the work.empty list,
|
||
// allocating new buffers if none are available.
|
||
//go:nowritebarrier
|
||
func getempty() *workbuf {
|
||
var b *workbuf
|
||
if work.empty != 0 {
|
||
b = (*workbuf)(work.empty.pop())
|
||
if b != nil {
|
||
b.checkempty()
|
||
}
|
||
}
|
||
if b == nil {
|
||
// Allocate more workbufs.
|
||
var s *mspan
|
||
if work.wbufSpans.free.first != nil {
|
||
lock(&work.wbufSpans.lock)
|
||
s = work.wbufSpans.free.first
|
||
if s != nil {
|
||
work.wbufSpans.free.remove(s)
|
||
work.wbufSpans.busy.insert(s)
|
||
}
|
||
unlock(&work.wbufSpans.lock)
|
||
}
|
||
if s == nil {
|
||
systemstack(func() {
|
||
s = mheap_.allocManual(workbufAlloc/pageSize, &memstats.gc_sys)
|
||
})
|
||
if s == nil {
|
||
throw("out of memory")
|
||
}
|
||
// Record the new span in the busy list.
|
||
lock(&work.wbufSpans.lock)
|
||
work.wbufSpans.busy.insert(s)
|
||
unlock(&work.wbufSpans.lock)
|
||
}
|
||
// Slice up the span into new workbufs. Return one and
|
||
// put the rest on the empty list.
|
||
for i := uintptr(0); i+_WorkbufSize <= workbufAlloc; i += _WorkbufSize {
|
||
newb := (*workbuf)(unsafe.Pointer(s.base() + i))
|
||
newb.nobj = 0
|
||
lfnodeValidate(&newb.node)
|
||
if i == 0 {
|
||
b = newb
|
||
} else {
|
||
putempty(newb)
|
||
}
|
||
}
|
||
}
|
||
return b
|
||
}
|
||
|
||
// putempty puts a workbuf onto the work.empty list.
|
||
// Upon entry this go routine owns b. The lfstack.push relinquishes ownership.
|
||
//go:nowritebarrier
|
||
func putempty(b *workbuf) {
|
||
b.checkempty()
|
||
work.empty.push(&b.node)
|
||
}
|
||
|
||
// putfull puts the workbuf on the work.full list for the GC.
|
||
// putfull accepts partially full buffers so the GC can avoid competing
|
||
// with the mutators for ownership of partially full buffers.
|
||
//go:nowritebarrier
|
||
func putfull(b *workbuf) {
|
||
b.checknonempty()
|
||
work.full.push(&b.node)
|
||
}
|
||
|
||
// trygetfull tries to get a full or partially empty workbuffer.
|
||
// If one is not immediately available return nil
|
||
//go:nowritebarrier
|
||
func trygetfull() *workbuf {
|
||
b := (*workbuf)(work.full.pop())
|
||
if b != nil {
|
||
b.checknonempty()
|
||
return b
|
||
}
|
||
return b
|
||
}
|
||
|
||
// Get a full work buffer off the work.full list.
|
||
// If nothing is available wait until all the other gc helpers have
|
||
// finished and then return nil.
|
||
// getfull acts as a barrier for work.nproc helpers. As long as one
|
||
// gchelper is actively marking objects it
|
||
// may create a workbuffer that the other helpers can work on.
|
||
// The for loop either exits when a work buffer is found
|
||
// or when _all_ of the work.nproc GC helpers are in the loop
|
||
// looking for work and thus not capable of creating new work.
|
||
// This is in fact the termination condition for the STW mark
|
||
// phase.
|
||
//go:nowritebarrier
|
||
func getfull() *workbuf {
|
||
b := (*workbuf)(work.full.pop())
|
||
if b != nil {
|
||
b.checknonempty()
|
||
return b
|
||
}
|
||
|
||
incnwait := atomic.Xadd(&work.nwait, +1)
|
||
if incnwait > work.nproc {
|
||
println("runtime: work.nwait=", incnwait, "work.nproc=", work.nproc)
|
||
throw("work.nwait > work.nproc")
|
||
}
|
||
for i := 0; ; i++ {
|
||
if work.full != 0 {
|
||
decnwait := atomic.Xadd(&work.nwait, -1)
|
||
if decnwait == work.nproc {
|
||
println("runtime: work.nwait=", decnwait, "work.nproc=", work.nproc)
|
||
throw("work.nwait > work.nproc")
|
||
}
|
||
b = (*workbuf)(work.full.pop())
|
||
if b != nil {
|
||
b.checknonempty()
|
||
return b
|
||
}
|
||
incnwait := atomic.Xadd(&work.nwait, +1)
|
||
if incnwait > work.nproc {
|
||
println("runtime: work.nwait=", incnwait, "work.nproc=", work.nproc)
|
||
throw("work.nwait > work.nproc")
|
||
}
|
||
}
|
||
if work.nwait == work.nproc && work.markrootNext >= work.markrootJobs {
|
||
return nil
|
||
}
|
||
if i < 10 {
|
||
procyield(20)
|
||
} else if i < 20 {
|
||
osyield()
|
||
} else {
|
||
usleep(100)
|
||
}
|
||
}
|
||
}
|
||
|
||
//go:nowritebarrier
|
||
func handoff(b *workbuf) *workbuf {
|
||
// Make new buffer with half of b's pointers.
|
||
b1 := getempty()
|
||
n := b.nobj / 2
|
||
b.nobj -= n
|
||
b1.nobj = n
|
||
memmove(unsafe.Pointer(&b1.obj[0]), unsafe.Pointer(&b.obj[b.nobj]), uintptr(n)*unsafe.Sizeof(b1.obj[0]))
|
||
|
||
// Put b on full list - let first half of b get stolen.
|
||
putfull(b)
|
||
return b1
|
||
}
|
||
|
||
// prepareFreeWorkbufs moves busy workbuf spans to free list so they
|
||
// can be freed to the heap. This must only be called when all
|
||
// workbufs are on the empty list.
|
||
func prepareFreeWorkbufs() {
|
||
lock(&work.wbufSpans.lock)
|
||
if work.full != 0 {
|
||
throw("cannot free workbufs when work.full != 0")
|
||
}
|
||
// Since all workbufs are on the empty list, we don't care
|
||
// which ones are in which spans. We can wipe the entire empty
|
||
// list and move all workbuf spans to the free list.
|
||
work.empty = 0
|
||
work.wbufSpans.free.takeAll(&work.wbufSpans.busy)
|
||
unlock(&work.wbufSpans.lock)
|
||
}
|
||
|
||
// freeSomeWbufs frees some workbufs back to the heap and returns
|
||
// true if it should be called again to free more.
|
||
func freeSomeWbufs(preemptible bool) bool {
|
||
const batchSize = 64 // ~1–2 µs per span.
|
||
lock(&work.wbufSpans.lock)
|
||
if gcphase != _GCoff || work.wbufSpans.free.isEmpty() {
|
||
unlock(&work.wbufSpans.lock)
|
||
return false
|
||
}
|
||
systemstack(func() {
|
||
gp := getg().m.curg
|
||
for i := 0; i < batchSize && !(preemptible && gp.preempt); i++ {
|
||
span := work.wbufSpans.free.first
|
||
if span == nil {
|
||
break
|
||
}
|
||
work.wbufSpans.free.remove(span)
|
||
mheap_.freeManual(span, &memstats.gc_sys)
|
||
}
|
||
})
|
||
more := !work.wbufSpans.free.isEmpty()
|
||
unlock(&work.wbufSpans.lock)
|
||
return more
|
||
}
|