// Copyright 2009 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Garbage collector: finalizers and block profiling. package runtime import "unsafe" var finlock mutex // protects the following variables var fing *g // goroutine that runs finalizers var finq *finblock // list of finalizers that are to be executed var finc *finblock // cache of free blocks var finptrmask [_FinBlockSize / typeBitmapScale]byte var fingwait bool var fingwake bool var allfin *finblock // list of all blocks var finalizer1 = [...]byte{ // Each Finalizer is 5 words, ptr ptr uintptr ptr ptr. // Each byte describes 4 words. // Need 4 Finalizers described by 5 bytes before pattern repeats: // ptr ptr uintptr ptr ptr // ptr ptr uintptr ptr ptr // ptr ptr uintptr ptr ptr // ptr ptr uintptr ptr ptr // aka // ptr ptr uintptr ptr // ptr ptr ptr uintptr // ptr ptr ptr ptr // uintptr ptr ptr ptr // ptr uintptr ptr ptr // Assumptions about Finalizer layout checked below. typePointer | typePointer<<2 | typeScalar<<4 | typePointer<<6, typePointer | typePointer<<2 | typePointer<<4 | typeScalar<<6, typePointer | typePointer<<2 | typePointer<<4 | typePointer<<6, typeScalar | typePointer<<2 | typePointer<<4 | typePointer<<6, typePointer | typeScalar<<2 | typePointer<<4 | typePointer<<6, } func queuefinalizer(p unsafe.Pointer, fn *funcval, nret uintptr, fint *_type, ot *ptrtype) { lock(&finlock) if finq == nil || finq.cnt == int32(len(finq.fin)) { if finc == nil { // Note: write barrier here, assigning to finc, but should be okay. finc = (*finblock)(persistentalloc(_FinBlockSize, 0, &memstats.gc_sys)) finc.alllink = allfin allfin = finc if finptrmask[0] == 0 { // Build pointer mask for Finalizer array in block. // Check assumptions made in finalizer1 array above. if (unsafe.Sizeof(finalizer{}) != 5*ptrSize || unsafe.Offsetof(finalizer{}.fn) != 0 || unsafe.Offsetof(finalizer{}.arg) != ptrSize || unsafe.Offsetof(finalizer{}.nret) != 2*ptrSize || unsafe.Offsetof(finalizer{}.fint) != 3*ptrSize || unsafe.Offsetof(finalizer{}.ot) != 4*ptrSize || typeBitsWidth != 2) { throw("finalizer out of sync") } for i := range finptrmask { finptrmask[i] = finalizer1[i%len(finalizer1)] } } } block := finc finc = block.next block.next = finq finq = block } f := &finq.fin[finq.cnt] finq.cnt++ f.fn = fn f.nret = nret f.fint = fint f.ot = ot f.arg = p fingwake = true unlock(&finlock) } //go:nowritebarrier func iterate_finq(callback func(*funcval, unsafe.Pointer, uintptr, *_type, *ptrtype)) { for fb := allfin; fb != nil; fb = fb.alllink { for i := int32(0); i < fb.cnt; i++ { f := &fb.fin[i] callback(f.fn, f.arg, f.nret, f.fint, f.ot) } } } func wakefing() *g { var res *g lock(&finlock) if fingwait && fingwake { fingwait = false fingwake = false res = fing } unlock(&finlock) return res } var fingCreate uint32 func createfing() { // start the finalizer goroutine exactly once if fingCreate == 0 && cas(&fingCreate, 0, 1) { go runfinq() } } // This is the goroutine that runs all of the finalizers func runfinq() { var ( frame unsafe.Pointer framecap uintptr ) for { lock(&finlock) fb := finq finq = nil if fb == nil { gp := getg() fing = gp fingwait = true gp.issystem = true goparkunlock(&finlock, "finalizer wait", traceEvGoBlock) gp.issystem = false continue } unlock(&finlock) if raceenabled { racefingo() } for fb != nil { for i := fb.cnt; i > 0; i-- { f := (*finalizer)(add(unsafe.Pointer(&fb.fin), uintptr(i-1)*unsafe.Sizeof(finalizer{}))) framesz := unsafe.Sizeof((interface{})(nil)) + uintptr(f.nret) if framecap < framesz { // The frame does not contain pointers interesting for GC, // all not yet finalized objects are stored in finq. // If we do not mark it as FlagNoScan, // the last finalized object is not collected. frame = mallocgc(framesz, nil, flagNoScan) framecap = framesz } if f.fint == nil { throw("missing type in runfinq") } switch f.fint.kind & kindMask { case kindPtr: // direct use of pointer *(*unsafe.Pointer)(frame) = f.arg case kindInterface: ityp := (*interfacetype)(unsafe.Pointer(f.fint)) // set up with empty interface (*eface)(frame)._type = &f.ot.typ (*eface)(frame).data = f.arg if len(ityp.mhdr) != 0 { // convert to interface with methods // this conversion is guaranteed to succeed - we checked in SetFinalizer assertE2I(ityp, *(*interface{})(frame), (*fInterface)(frame)) } default: throw("bad kind in runfinq") } reflectcall(nil, unsafe.Pointer(f.fn), frame, uint32(framesz), uint32(framesz)) // drop finalizer queue references to finalized object f.fn = nil f.arg = nil f.ot = nil fb.cnt = i - 1 } next := fb.next lock(&finlock) fb.next = finc finc = fb unlock(&finlock) fb = next } } } // SetFinalizer sets the finalizer associated with x to f. // When the garbage collector finds an unreachable block // with an associated finalizer, it clears the association and runs // f(x) in a separate goroutine. This makes x reachable again, but // now without an associated finalizer. Assuming that SetFinalizer // is not called again, the next time the garbage collector sees // that x is unreachable, it will free x. // // SetFinalizer(x, nil) clears any finalizer associated with x. // // The argument x must be a pointer to an object allocated by // calling new or by taking the address of a composite literal. // The argument f must be a function that takes a single argument // to which x's type can be assigned, and can have arbitrary ignored return // values. If either of these is not true, SetFinalizer aborts the // program. // // Finalizers are run in dependency order: if A points at B, both have // finalizers, and they are otherwise unreachable, only the finalizer // for A runs; once A is freed, the finalizer for B can run. // If a cyclic structure includes a block with a finalizer, that // cycle is not guaranteed to be garbage collected and the finalizer // is not guaranteed to run, because there is no ordering that // respects the dependencies. // // The finalizer for x is scheduled to run at some arbitrary time after // x becomes unreachable. // There is no guarantee that finalizers will run before a program exits, // so typically they are useful only for releasing non-memory resources // associated with an object during a long-running program. // For example, an os.File object could use a finalizer to close the // associated operating system file descriptor when a program discards // an os.File without calling Close, but it would be a mistake // to depend on a finalizer to flush an in-memory I/O buffer such as a // bufio.Writer, because the buffer would not be flushed at program exit. // // It is not guaranteed that a finalizer will run if the size of *x is // zero bytes. // // It is not guaranteed that a finalizer will run for objects allocated // in initializers for package-level variables. Such objects may be // linker-allocated, not heap-allocated. // // A single goroutine runs all finalizers for a program, sequentially. // If a finalizer must run for a long time, it should do so by starting // a new goroutine. func SetFinalizer(obj interface{}, finalizer interface{}) { e := (*eface)(unsafe.Pointer(&obj)) etyp := e._type if etyp == nil { throw("runtime.SetFinalizer: first argument is nil") } if etyp.kind&kindMask != kindPtr { throw("runtime.SetFinalizer: first argument is " + *etyp._string + ", not pointer") } ot := (*ptrtype)(unsafe.Pointer(etyp)) if ot.elem == nil { throw("nil elem type!") } // find the containing object _, base, _ := findObject(e.data) if base == nil { // 0-length objects are okay. if e.data == unsafe.Pointer(&zerobase) { return } // Global initializers might be linker-allocated. // var Foo = &Object{} // func main() { // runtime.SetFinalizer(Foo, nil) // } // The relevant segments are: noptrdata, data, bss, noptrbss. // We cannot assume they are in any order or even contiguous, // due to external linking. if uintptr(unsafe.Pointer(&noptrdata)) <= uintptr(e.data) && uintptr(e.data) < uintptr(unsafe.Pointer(&enoptrdata)) || uintptr(unsafe.Pointer(&data)) <= uintptr(e.data) && uintptr(e.data) < uintptr(unsafe.Pointer(&edata)) || uintptr(unsafe.Pointer(&bss)) <= uintptr(e.data) && uintptr(e.data) < uintptr(unsafe.Pointer(&ebss)) || uintptr(unsafe.Pointer(&noptrbss)) <= uintptr(e.data) && uintptr(e.data) < uintptr(unsafe.Pointer(&enoptrbss)) { return } throw("runtime.SetFinalizer: pointer not in allocated block") } if e.data != base { // As an implementation detail we allow to set finalizers for an inner byte // of an object if it could come from tiny alloc (see mallocgc for details). if ot.elem == nil || ot.elem.kind&kindNoPointers == 0 || ot.elem.size >= maxTinySize { throw("runtime.SetFinalizer: pointer not at beginning of allocated block") } } f := (*eface)(unsafe.Pointer(&finalizer)) ftyp := f._type if ftyp == nil { // switch to system stack and remove finalizer systemstack(func() { removefinalizer(e.data) }) return } if ftyp.kind&kindMask != kindFunc { throw("runtime.SetFinalizer: second argument is " + *ftyp._string + ", not a function") } ft := (*functype)(unsafe.Pointer(ftyp)) ins := *(*[]*_type)(unsafe.Pointer(&ft.in)) if ft.dotdotdot || len(ins) != 1 { throw("runtime.SetFinalizer: cannot pass " + *etyp._string + " to finalizer " + *ftyp._string) } fint := ins[0] switch { case fint == etyp: // ok - same type goto okarg case fint.kind&kindMask == kindPtr: if (fint.x == nil || fint.x.name == nil || etyp.x == nil || etyp.x.name == nil) && (*ptrtype)(unsafe.Pointer(fint)).elem == ot.elem { // ok - not same type, but both pointers, // one or the other is unnamed, and same element type, so assignable. goto okarg } case fint.kind&kindMask == kindInterface: ityp := (*interfacetype)(unsafe.Pointer(fint)) if len(ityp.mhdr) == 0 { // ok - satisfies empty interface goto okarg } if assertE2I2(ityp, obj, nil) { goto okarg } } throw("runtime.SetFinalizer: cannot pass " + *etyp._string + " to finalizer " + *ftyp._string) okarg: // compute size needed for return parameters nret := uintptr(0) for _, t := range *(*[]*_type)(unsafe.Pointer(&ft.out)) { nret = round(nret, uintptr(t.align)) + uintptr(t.size) } nret = round(nret, ptrSize) // make sure we have a finalizer goroutine createfing() systemstack(func() { if !addfinalizer(e.data, (*funcval)(f.data), nret, fint, ot) { throw("runtime.SetFinalizer: finalizer already set") } }) } // Look up pointer v in heap. Return the span containing the object, // the start of the object, and the size of the object. If the object // does not exist, return nil, nil, 0. func findObject(v unsafe.Pointer) (s *mspan, x unsafe.Pointer, n uintptr) { c := gomcache() c.local_nlookup++ if ptrSize == 4 && c.local_nlookup >= 1<<30 { // purge cache stats to prevent overflow lock(&mheap_.lock) purgecachedstats(c) unlock(&mheap_.lock) } // find span arena_start := uintptr(unsafe.Pointer(mheap_.arena_start)) arena_used := uintptr(unsafe.Pointer(mheap_.arena_used)) if uintptr(v) < arena_start || uintptr(v) >= arena_used { return } p := uintptr(v) >> pageShift q := p - arena_start>>pageShift s = *(**mspan)(add(unsafe.Pointer(mheap_.spans), q*ptrSize)) if s == nil { return } x = unsafe.Pointer(uintptr(s.start) << pageShift) if uintptr(v) < uintptr(x) || uintptr(v) >= uintptr(unsafe.Pointer(s.limit)) || s.state != mSpanInUse { s = nil x = nil return } n = uintptr(s.elemsize) if s.sizeclass != 0 { x = add(x, (uintptr(v)-uintptr(x))/n*n) } return }