2014-09-01 22:14:22 -06:00
|
|
|
// Copyright 2011 The Go Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
|
|
|
// CPU profiling.
|
|
|
|
// Based on algorithms and data structures used in
|
|
|
|
// http://code.google.com/p/google-perftools/.
|
|
|
|
//
|
|
|
|
// The main difference between this code and the google-perftools
|
|
|
|
// code is that this code is written to allow copying the profile data
|
|
|
|
// to an arbitrary io.Writer, while the google-perftools code always
|
|
|
|
// writes to an operating system file.
|
|
|
|
//
|
|
|
|
// The signal handler for the profiling clock tick adds a new stack trace
|
|
|
|
// to a hash table tracking counts for recent traces. Most clock ticks
|
|
|
|
// hit in the cache. In the event of a cache miss, an entry must be
|
|
|
|
// evicted from the hash table, copied to a log that will eventually be
|
|
|
|
// written as profile data. The google-perftools code flushed the
|
|
|
|
// log itself during the signal handler. This code cannot do that, because
|
|
|
|
// the io.Writer might block or need system calls or locks that are not
|
|
|
|
// safe to use from within the signal handler. Instead, we split the log
|
|
|
|
// into two halves and let the signal handler fill one half while a goroutine
|
|
|
|
// is writing out the other half. When the signal handler fills its half, it
|
|
|
|
// offers to swap with the goroutine. If the writer is not done with its half,
|
|
|
|
// we lose the stack trace for this clock tick (and record that loss).
|
|
|
|
// The goroutine interacts with the signal handler by calling getprofile() to
|
|
|
|
// get the next log piece to write, implicitly handing back the last log
|
|
|
|
// piece it obtained.
|
|
|
|
//
|
|
|
|
// The state of this dance between the signal handler and the goroutine
|
|
|
|
// is encoded in the Profile.handoff field. If handoff == 0, then the goroutine
|
|
|
|
// is not using either log half and is waiting (or will soon be waiting) for
|
2015-02-24 22:41:21 -07:00
|
|
|
// a new piece by calling notesleep(&p.wait). If the signal handler
|
|
|
|
// changes handoff from 0 to non-zero, it must call notewakeup(&p.wait)
|
2014-09-01 22:14:22 -06:00
|
|
|
// to wake the goroutine. The value indicates the number of entries in the
|
|
|
|
// log half being handed off. The goroutine leaves the non-zero value in
|
|
|
|
// place until it has finished processing the log half and then flips the number
|
|
|
|
// back to zero. Setting the high bit in handoff means that the profiling is over,
|
|
|
|
// and the goroutine is now in charge of flushing the data left in the hash table
|
|
|
|
// to the log and returning that data.
|
|
|
|
//
|
|
|
|
// The handoff field is manipulated using atomic operations.
|
|
|
|
// For the most part, the manipulation of handoff is orderly: if handoff == 0
|
|
|
|
// then the signal handler owns it and can change it to non-zero.
|
|
|
|
// If handoff != 0 then the goroutine owns it and can change it to zero.
|
|
|
|
// If that were the end of the story then we would not need to manipulate
|
|
|
|
// handoff using atomic operations. The operations are needed, however,
|
|
|
|
// in order to let the log closer set the high bit to indicate "EOF" safely
|
|
|
|
// in the situation when normally the goroutine "owns" handoff.
|
|
|
|
|
|
|
|
package runtime
|
|
|
|
|
2015-11-02 12:09:24 -07:00
|
|
|
import (
|
|
|
|
"runtime/internal/atomic"
|
|
|
|
"unsafe"
|
|
|
|
)
|
2014-09-01 22:14:22 -06:00
|
|
|
|
|
|
|
const (
|
|
|
|
numBuckets = 1 << 10
|
|
|
|
logSize = 1 << 17
|
|
|
|
assoc = 4
|
|
|
|
maxCPUProfStack = 64
|
|
|
|
)
|
|
|
|
|
|
|
|
type cpuprofEntry struct {
|
|
|
|
count uintptr
|
2015-02-24 22:41:21 -07:00
|
|
|
depth int
|
2014-09-01 22:14:22 -06:00
|
|
|
stack [maxCPUProfStack]uintptr
|
|
|
|
}
|
|
|
|
|
|
|
|
type cpuProfile struct {
|
|
|
|
on bool // profiling is on
|
|
|
|
wait note // goroutine waits here
|
|
|
|
count uintptr // tick count
|
|
|
|
evicts uintptr // eviction count
|
|
|
|
lost uintptr // lost ticks that need to be logged
|
|
|
|
|
|
|
|
// Active recent stack traces.
|
|
|
|
hash [numBuckets]struct {
|
|
|
|
entry [assoc]cpuprofEntry
|
|
|
|
}
|
|
|
|
|
|
|
|
// Log of traces evicted from hash.
|
|
|
|
// Signal handler has filled log[toggle][:nlog].
|
|
|
|
// Goroutine is writing log[1-toggle][:handoff].
|
|
|
|
log [2][logSize / 2]uintptr
|
2015-02-24 22:41:21 -07:00
|
|
|
nlog int
|
2014-09-01 22:14:22 -06:00
|
|
|
toggle int32
|
|
|
|
handoff uint32
|
|
|
|
|
|
|
|
// Writer state.
|
|
|
|
// Writer maintains its own toggle to avoid races
|
|
|
|
// looking at signal handler's toggle.
|
|
|
|
wtoggle uint32
|
|
|
|
wholding bool // holding & need to release a log half
|
|
|
|
flushing bool // flushing hash table - profile is over
|
|
|
|
eodSent bool // special end-of-data record sent; => flushing
|
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
cpuprofLock mutex
|
|
|
|
cpuprof *cpuProfile
|
|
|
|
|
|
|
|
eod = [3]uintptr{0, 1, 0}
|
|
|
|
)
|
|
|
|
|
2014-09-03 22:54:06 -06:00
|
|
|
func setcpuprofilerate(hz int32) {
|
[dev.cc] runtime: delete scalararg, ptrarg; rename onM to systemstack
Scalararg and ptrarg are not "signal safe".
Go code filling them out can be interrupted by a signal,
and then the signal handler runs, and if it also ends up
in Go code that uses scalararg or ptrarg, now the old
values have been smashed.
For the pieces of code that do need to run in a signal handler,
we introduced onM_signalok, which is really just onM
except that the _signalok is meant to convey that the caller
asserts that scalarg and ptrarg will be restored to their old
values after the call (instead of the usual behavior, zeroing them).
Scalararg and ptrarg are also untyped and therefore error-prone.
Go code can always pass a closure instead of using scalararg
and ptrarg; they were only really necessary for C code.
And there's no more C code.
For all these reasons, delete scalararg and ptrarg, converting
the few remaining references to use closures.
Once those are gone, there is no need for a distinction between
onM and onM_signalok, so replace both with a single function
equivalent to the current onM_signalok (that is, it can be called
on any of the curg, g0, and gsignal stacks).
The name onM and the phrase 'm stack' are misnomers,
because on most system an M has two system stacks:
the main thread stack and the signal handling stack.
Correct the misnomer by naming the replacement function systemstack.
Fix a few references to "M stack" in code.
The main motivation for this change is to eliminate scalararg/ptrarg.
Rick and I have already seen them cause problems because
the calling sequence m.ptrarg[0] = p is a heap pointer assignment,
so it gets a write barrier. The write barrier also uses onM, so it has
all the same problems as if it were being invoked by a signal handler.
We worked around this by saving and restoring the old values
and by calling onM_signalok, but there's no point in keeping this nice
home for bugs around any longer.
This CL also changes funcline to return the file name as a result
instead of filling in a passed-in *string. (The *string signature is
left over from when the code was written in and called from C.)
That's arguably an unrelated change, except that once I had done
the ptrarg/scalararg/onM cleanup I started getting false positives
about the *string argument escaping (not allowed in package runtime).
The compiler is wrong, but the easiest fix is to write the code like
Go code instead of like C code. I am a bit worried that the compiler
is wrong because of some use of uninitialized memory in the escape
analysis. If that's the reason, it will go away when we convert the
compiler to Go. (And if not, we'll debug it the next time.)
LGTM=khr
R=r, khr
CC=austin, golang-codereviews, iant, rlh
https://golang.org/cl/174950043
2014-11-12 12:54:31 -07:00
|
|
|
systemstack(func() {
|
|
|
|
setcpuprofilerate_m(hz)
|
|
|
|
})
|
2014-09-03 22:54:06 -06:00
|
|
|
}
|
2014-09-01 22:14:22 -06:00
|
|
|
|
|
|
|
// lostProfileData is a no-op function used in profiles
|
|
|
|
// to mark the number of profiling stack traces that were
|
|
|
|
// discarded due to slow data writers.
|
|
|
|
func lostProfileData() {}
|
|
|
|
|
|
|
|
// SetCPUProfileRate sets the CPU profiling rate to hz samples per second.
|
|
|
|
// If hz <= 0, SetCPUProfileRate turns off profiling.
|
|
|
|
// If the profiler is on, the rate cannot be changed without first turning it off.
|
|
|
|
//
|
|
|
|
// Most clients should use the runtime/pprof package or
|
|
|
|
// the testing package's -test.cpuprofile flag instead of calling
|
|
|
|
// SetCPUProfileRate directly.
|
|
|
|
func SetCPUProfileRate(hz int) {
|
|
|
|
// Clamp hz to something reasonable.
|
|
|
|
if hz < 0 {
|
|
|
|
hz = 0
|
|
|
|
}
|
|
|
|
if hz > 1000000 {
|
|
|
|
hz = 1000000
|
|
|
|
}
|
|
|
|
|
|
|
|
lock(&cpuprofLock)
|
|
|
|
if hz > 0 {
|
|
|
|
if cpuprof == nil {
|
|
|
|
cpuprof = (*cpuProfile)(sysAlloc(unsafe.Sizeof(cpuProfile{}), &memstats.other_sys))
|
|
|
|
if cpuprof == nil {
|
|
|
|
print("runtime: cpu profiling cannot allocate memory\n")
|
|
|
|
unlock(&cpuprofLock)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if cpuprof.on || cpuprof.handoff != 0 {
|
|
|
|
print("runtime: cannot set cpu profile rate until previous profile has finished.\n")
|
|
|
|
unlock(&cpuprofLock)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
cpuprof.on = true
|
|
|
|
// pprof binary header format.
|
|
|
|
// http://code.google.com/p/google-perftools/source/browse/trunk/src/profiledata.cc#117
|
|
|
|
p := &cpuprof.log[0]
|
|
|
|
p[0] = 0 // count for header
|
|
|
|
p[1] = 3 // depth for header
|
|
|
|
p[2] = 0 // version number
|
|
|
|
p[3] = uintptr(1e6 / hz) // period (microseconds)
|
|
|
|
p[4] = 0
|
|
|
|
cpuprof.nlog = 5
|
|
|
|
cpuprof.toggle = 0
|
|
|
|
cpuprof.wholding = false
|
|
|
|
cpuprof.wtoggle = 0
|
|
|
|
cpuprof.flushing = false
|
|
|
|
cpuprof.eodSent = false
|
|
|
|
noteclear(&cpuprof.wait)
|
|
|
|
|
|
|
|
setcpuprofilerate(int32(hz))
|
|
|
|
} else if cpuprof != nil && cpuprof.on {
|
|
|
|
setcpuprofilerate(0)
|
|
|
|
cpuprof.on = false
|
|
|
|
|
|
|
|
// Now add is not running anymore, and getprofile owns the entire log.
|
2015-02-24 22:41:21 -07:00
|
|
|
// Set the high bit in cpuprof.handoff to tell getprofile.
|
2014-09-01 22:14:22 -06:00
|
|
|
for {
|
|
|
|
n := cpuprof.handoff
|
|
|
|
if n&0x80000000 != 0 {
|
|
|
|
print("runtime: setcpuprofile(off) twice\n")
|
|
|
|
}
|
2015-11-02 12:09:24 -07:00
|
|
|
if atomic.Cas(&cpuprof.handoff, n, n|0x80000000) {
|
2014-09-01 22:14:22 -06:00
|
|
|
if n == 0 {
|
|
|
|
// we did the transition from 0 -> nonzero so we wake getprofile
|
|
|
|
notewakeup(&cpuprof.wait)
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
unlock(&cpuprofLock)
|
|
|
|
}
|
|
|
|
|
|
|
|
// add adds the stack trace to the profile.
|
|
|
|
// It is called from signal handlers and other limited environments
|
|
|
|
// and cannot allocate memory or acquire locks that might be
|
|
|
|
// held at the time of the signal, nor can it use substantial amounts
|
|
|
|
// of stack. It is allowed to call evict.
|
|
|
|
func (p *cpuProfile) add(pc []uintptr) {
|
2015-02-24 22:41:21 -07:00
|
|
|
if len(pc) > maxCPUProfStack {
|
|
|
|
pc = pc[:maxCPUProfStack]
|
|
|
|
}
|
|
|
|
|
2014-09-01 22:14:22 -06:00
|
|
|
// Compute hash.
|
|
|
|
h := uintptr(0)
|
|
|
|
for _, x := range pc {
|
|
|
|
h = h<<8 | (h >> (8 * (unsafe.Sizeof(h) - 1)))
|
2014-12-18 14:50:57 -07:00
|
|
|
h += x * 41
|
2014-09-01 22:14:22 -06:00
|
|
|
}
|
|
|
|
p.count++
|
|
|
|
|
|
|
|
// Add to entry count if already present in table.
|
|
|
|
b := &p.hash[h%numBuckets]
|
|
|
|
Assoc:
|
|
|
|
for i := range b.entry {
|
|
|
|
e := &b.entry[i]
|
2015-02-24 22:41:21 -07:00
|
|
|
if e.depth != len(pc) {
|
2014-09-01 22:14:22 -06:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
for j := range pc {
|
|
|
|
if e.stack[j] != pc[j] {
|
|
|
|
continue Assoc
|
|
|
|
}
|
|
|
|
}
|
|
|
|
e.count++
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Evict entry with smallest count.
|
|
|
|
var e *cpuprofEntry
|
|
|
|
for i := range b.entry {
|
|
|
|
if e == nil || b.entry[i].count < e.count {
|
|
|
|
e = &b.entry[i]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if e.count > 0 {
|
|
|
|
if !p.evict(e) {
|
|
|
|
// Could not evict entry. Record lost stack.
|
|
|
|
p.lost++
|
|
|
|
return
|
|
|
|
}
|
|
|
|
p.evicts++
|
|
|
|
}
|
|
|
|
|
|
|
|
// Reuse the newly evicted entry.
|
2015-02-24 22:41:21 -07:00
|
|
|
e.depth = len(pc)
|
2014-09-01 22:14:22 -06:00
|
|
|
e.count = 1
|
2014-09-03 10:47:30 -06:00
|
|
|
copy(e.stack[:], pc)
|
2014-09-01 22:14:22 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
// evict copies the given entry's data into the log, so that
|
|
|
|
// the entry can be reused. evict is called from add, which
|
|
|
|
// is called from the profiling signal handler, so it must not
|
|
|
|
// allocate memory or block. It is safe to call flushlog.
|
|
|
|
// evict returns true if the entry was copied to the log,
|
|
|
|
// false if there was no room available.
|
|
|
|
func (p *cpuProfile) evict(e *cpuprofEntry) bool {
|
|
|
|
d := e.depth
|
|
|
|
nslot := d + 2
|
|
|
|
log := &p.log[p.toggle]
|
2015-02-24 22:41:21 -07:00
|
|
|
if p.nlog+nslot > len(log) {
|
2014-09-01 22:14:22 -06:00
|
|
|
if !p.flushlog() {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
log = &p.log[p.toggle]
|
|
|
|
}
|
|
|
|
|
|
|
|
q := p.nlog
|
|
|
|
log[q] = e.count
|
|
|
|
q++
|
2015-02-24 22:41:21 -07:00
|
|
|
log[q] = uintptr(d)
|
2014-09-01 22:14:22 -06:00
|
|
|
q++
|
2014-09-03 10:47:30 -06:00
|
|
|
copy(log[q:], e.stack[:d])
|
|
|
|
q += d
|
2014-09-01 22:14:22 -06:00
|
|
|
p.nlog = q
|
|
|
|
e.count = 0
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// flushlog tries to flush the current log and switch to the other one.
|
|
|
|
// flushlog is called from evict, called from add, called from the signal handler,
|
|
|
|
// so it cannot allocate memory or block. It can try to swap logs with
|
|
|
|
// the writing goroutine, as explained in the comment at the top of this file.
|
|
|
|
func (p *cpuProfile) flushlog() bool {
|
2015-11-02 12:09:24 -07:00
|
|
|
if !atomic.Cas(&p.handoff, 0, uint32(p.nlog)) {
|
2014-09-01 22:14:22 -06:00
|
|
|
return false
|
|
|
|
}
|
|
|
|
notewakeup(&p.wait)
|
|
|
|
|
|
|
|
p.toggle = 1 - p.toggle
|
|
|
|
log := &p.log[p.toggle]
|
2015-02-24 22:41:21 -07:00
|
|
|
q := 0
|
2014-09-01 22:14:22 -06:00
|
|
|
if p.lost > 0 {
|
2014-09-03 09:10:38 -06:00
|
|
|
lostPC := funcPC(lostProfileData)
|
2014-09-01 22:14:22 -06:00
|
|
|
log[0] = p.lost
|
|
|
|
log[1] = 1
|
|
|
|
log[2] = lostPC
|
|
|
|
q = 3
|
|
|
|
p.lost = 0
|
|
|
|
}
|
|
|
|
p.nlog = q
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// getprofile blocks until the next block of profiling data is available
|
|
|
|
// and returns it as a []byte. It is called from the writing goroutine.
|
|
|
|
func (p *cpuProfile) getprofile() []byte {
|
|
|
|
if p == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.wholding {
|
|
|
|
// Release previous log to signal handling side.
|
|
|
|
// Loop because we are racing against SetCPUProfileRate(0).
|
|
|
|
for {
|
|
|
|
n := p.handoff
|
|
|
|
if n == 0 {
|
|
|
|
print("runtime: phase error during cpu profile handoff\n")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if n&0x80000000 != 0 {
|
|
|
|
p.wtoggle = 1 - p.wtoggle
|
|
|
|
p.wholding = false
|
|
|
|
p.flushing = true
|
|
|
|
goto Flush
|
|
|
|
}
|
2015-11-02 12:09:24 -07:00
|
|
|
if atomic.Cas(&p.handoff, n, 0) {
|
2014-09-01 22:14:22 -06:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
p.wtoggle = 1 - p.wtoggle
|
|
|
|
p.wholding = false
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.flushing {
|
|
|
|
goto Flush
|
|
|
|
}
|
|
|
|
|
|
|
|
if !p.on && p.handoff == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for new log.
|
|
|
|
notetsleepg(&p.wait, -1)
|
|
|
|
noteclear(&p.wait)
|
|
|
|
|
|
|
|
switch n := p.handoff; {
|
|
|
|
case n == 0:
|
|
|
|
print("runtime: phase error during cpu profile wait\n")
|
|
|
|
return nil
|
|
|
|
case n == 0x80000000:
|
|
|
|
p.flushing = true
|
|
|
|
goto Flush
|
|
|
|
default:
|
|
|
|
n &^= 0x80000000
|
|
|
|
|
|
|
|
// Return new log to caller.
|
|
|
|
p.wholding = true
|
|
|
|
|
|
|
|
return uintptrBytes(p.log[p.wtoggle][:n])
|
|
|
|
}
|
|
|
|
|
|
|
|
// In flush mode.
|
|
|
|
// Add is no longer being called. We own the log.
|
2015-02-24 22:41:21 -07:00
|
|
|
// Also, p.handoff is non-zero, so flushlog will return false.
|
2014-09-01 22:14:22 -06:00
|
|
|
// Evict the hash table into the log and return it.
|
|
|
|
Flush:
|
|
|
|
for i := range p.hash {
|
|
|
|
b := &p.hash[i]
|
|
|
|
for j := range b.entry {
|
|
|
|
e := &b.entry[j]
|
|
|
|
if e.count > 0 && !p.evict(e) {
|
|
|
|
// Filled the log. Stop the loop and return what we've got.
|
|
|
|
break Flush
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return pending log data.
|
|
|
|
if p.nlog > 0 {
|
|
|
|
// Note that we're using toggle now, not wtoggle,
|
|
|
|
// because we're working on the log directly.
|
|
|
|
n := p.nlog
|
|
|
|
p.nlog = 0
|
|
|
|
return uintptrBytes(p.log[p.toggle][:n])
|
|
|
|
}
|
|
|
|
|
|
|
|
// Made it through the table without finding anything to log.
|
|
|
|
if !p.eodSent {
|
|
|
|
// We may not have space to append this to the partial log buf,
|
|
|
|
// so we always return a new slice for the end-of-data marker.
|
|
|
|
p.eodSent = true
|
|
|
|
return uintptrBytes(eod[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally done. Clean up and return nil.
|
|
|
|
p.flushing = false
|
2015-11-02 12:09:24 -07:00
|
|
|
if !atomic.Cas(&p.handoff, p.handoff, 0) {
|
2014-09-01 22:14:22 -06:00
|
|
|
print("runtime: profile flush racing with something\n")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func uintptrBytes(p []uintptr) (ret []byte) {
|
2015-04-10 16:01:54 -06:00
|
|
|
pp := (*slice)(unsafe.Pointer(&p))
|
|
|
|
rp := (*slice)(unsafe.Pointer(&ret))
|
2014-09-01 22:14:22 -06:00
|
|
|
|
|
|
|
rp.array = pp.array
|
|
|
|
rp.len = pp.len * int(unsafe.Sizeof(p[0]))
|
|
|
|
rp.cap = rp.len
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// CPUProfile returns the next chunk of binary CPU profiling stack trace data,
|
|
|
|
// blocking until data is available. If profiling is turned off and all the profile
|
|
|
|
// data accumulated while it was on has been returned, CPUProfile returns nil.
|
|
|
|
// The caller must save the returned data before calling CPUProfile again.
|
|
|
|
//
|
|
|
|
// Most clients should use the runtime/pprof package or
|
|
|
|
// the testing package's -test.cpuprofile flag instead of calling
|
|
|
|
// CPUProfile directly.
|
|
|
|
func CPUProfile() []byte {
|
|
|
|
return cpuprof.getprofile()
|
|
|
|
}
|
2014-12-22 11:27:53 -07:00
|
|
|
|
|
|
|
//go:linkname runtime_pprof_runtime_cyclesPerSecond runtime/pprof.runtime_cyclesPerSecond
|
|
|
|
func runtime_pprof_runtime_cyclesPerSecond() int64 {
|
|
|
|
return tickspersecond()
|
|
|
|
}
|