2011-04-21 10:09:25 -06:00
|
|
|
// Copyright 2011 The Go Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
|
|
|
package runtime_test
|
|
|
|
|
|
|
|
import (
|
2013-02-28 15:41:45 -07:00
|
|
|
"math"
|
2011-04-21 10:09:25 -06:00
|
|
|
"runtime"
|
2015-01-19 12:59:58 -07:00
|
|
|
"sync"
|
2011-07-12 10:24:32 -06:00
|
|
|
"sync/atomic"
|
2013-06-15 06:06:28 -06:00
|
|
|
"syscall"
|
2011-04-21 10:09:25 -06:00
|
|
|
"testing"
|
2013-02-14 13:02:12 -07:00
|
|
|
"time"
|
2011-04-21 10:09:25 -06:00
|
|
|
)
|
|
|
|
|
2011-04-22 13:22:11 -06:00
|
|
|
var stop = make(chan bool, 1)
|
|
|
|
|
2011-04-21 10:09:25 -06:00
|
|
|
func perpetuumMobile() {
|
2011-04-22 13:22:11 -06:00
|
|
|
select {
|
|
|
|
case <-stop:
|
|
|
|
default:
|
|
|
|
go perpetuumMobile()
|
|
|
|
}
|
2011-04-21 10:09:25 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestStopTheWorldDeadlock(t *testing.T) {
|
2011-04-23 08:03:51 -06:00
|
|
|
if testing.Short() {
|
2013-01-23 23:32:10 -07:00
|
|
|
t.Skip("skipping during short test")
|
2011-04-23 08:03:51 -06:00
|
|
|
}
|
2011-05-31 08:38:51 -06:00
|
|
|
maxprocs := runtime.GOMAXPROCS(3)
|
|
|
|
compl := make(chan bool, 2)
|
2011-04-21 10:09:25 -06:00
|
|
|
go func() {
|
|
|
|
for i := 0; i != 1000; i += 1 {
|
|
|
|
runtime.GC()
|
|
|
|
}
|
2011-05-31 08:38:51 -06:00
|
|
|
compl <- true
|
2011-04-21 10:09:25 -06:00
|
|
|
}()
|
|
|
|
go func() {
|
|
|
|
for i := 0; i != 1000; i += 1 {
|
|
|
|
runtime.GOMAXPROCS(3)
|
|
|
|
}
|
2011-05-31 08:38:51 -06:00
|
|
|
compl <- true
|
2011-04-21 10:09:25 -06:00
|
|
|
}()
|
|
|
|
go perpetuumMobile()
|
|
|
|
<-compl
|
2011-05-31 08:38:51 -06:00
|
|
|
<-compl
|
2011-04-22 13:22:11 -06:00
|
|
|
stop <- true
|
2011-05-31 08:38:51 -06:00
|
|
|
runtime.GOMAXPROCS(maxprocs)
|
2011-04-21 10:09:25 -06:00
|
|
|
}
|
2011-07-12 10:24:32 -06:00
|
|
|
|
2013-02-20 01:13:04 -07:00
|
|
|
func TestYieldProgress(t *testing.T) {
|
|
|
|
testYieldProgress(t, false)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestYieldLockedProgress(t *testing.T) {
|
|
|
|
testYieldProgress(t, true)
|
|
|
|
}
|
|
|
|
|
|
|
|
func testYieldProgress(t *testing.T, locked bool) {
|
|
|
|
c := make(chan bool)
|
|
|
|
cack := make(chan bool)
|
|
|
|
go func() {
|
|
|
|
if locked {
|
|
|
|
runtime.LockOSThread()
|
|
|
|
}
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-c:
|
|
|
|
cack <- true
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
runtime.Gosched()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
c <- true
|
|
|
|
<-cack
|
|
|
|
}
|
|
|
|
|
2013-02-14 13:02:12 -07:00
|
|
|
func TestYieldLocked(t *testing.T) {
|
|
|
|
const N = 10
|
|
|
|
c := make(chan bool)
|
|
|
|
go func() {
|
|
|
|
runtime.LockOSThread()
|
|
|
|
for i := 0; i < N; i++ {
|
|
|
|
runtime.Gosched()
|
|
|
|
time.Sleep(time.Millisecond)
|
|
|
|
}
|
|
|
|
c <- true
|
|
|
|
// runtime.UnlockOSThread() is deliberately omitted
|
|
|
|
}()
|
|
|
|
<-c
|
|
|
|
}
|
|
|
|
|
2013-07-11 13:57:36 -06:00
|
|
|
func TestGoroutineParallelism(t *testing.T) {
|
2013-08-01 08:25:36 -06:00
|
|
|
P := 4
|
|
|
|
N := 10
|
|
|
|
if testing.Short() {
|
|
|
|
P = 3
|
|
|
|
N = 3
|
|
|
|
}
|
2013-07-11 13:57:36 -06:00
|
|
|
defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(P))
|
2014-07-15 00:30:12 -06:00
|
|
|
// If runtime triggers a forced GC during this test then it will deadlock,
|
|
|
|
// since the goroutines can't be stopped/preempted.
|
|
|
|
// So give this test as much time as possible.
|
|
|
|
runtime.GC()
|
2013-08-01 08:25:36 -06:00
|
|
|
for try := 0; try < N; try++ {
|
2013-07-11 13:57:36 -06:00
|
|
|
done := make(chan bool)
|
|
|
|
x := uint32(0)
|
|
|
|
for p := 0; p < P; p++ {
|
|
|
|
// Test that all P goroutines are scheduled at the same time
|
|
|
|
go func(p int) {
|
|
|
|
for i := 0; i < 3; i++ {
|
|
|
|
expected := uint32(P*i + p)
|
|
|
|
for atomic.LoadUint32(&x) != expected {
|
|
|
|
}
|
|
|
|
atomic.StoreUint32(&x, expected+1)
|
|
|
|
}
|
|
|
|
done <- true
|
|
|
|
}(p)
|
|
|
|
}
|
|
|
|
for p := 0; p < P; p++ {
|
|
|
|
<-done
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-14 13:02:12 -07:00
|
|
|
func TestBlockLocked(t *testing.T) {
|
|
|
|
const N = 10
|
|
|
|
c := make(chan bool)
|
|
|
|
go func() {
|
|
|
|
runtime.LockOSThread()
|
|
|
|
for i := 0; i < N; i++ {
|
|
|
|
c <- true
|
|
|
|
}
|
|
|
|
runtime.UnlockOSThread()
|
|
|
|
}()
|
|
|
|
for i := 0; i < N; i++ {
|
|
|
|
<-c
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-06-15 06:06:28 -06:00
|
|
|
func TestTimerFairness(t *testing.T) {
|
|
|
|
done := make(chan bool)
|
|
|
|
c := make(chan bool)
|
|
|
|
for i := 0; i < 2; i++ {
|
|
|
|
go func() {
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case c <- true:
|
|
|
|
case <-done:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
timer := time.After(20 * time.Millisecond)
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-c:
|
|
|
|
case <-timer:
|
|
|
|
close(done)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestTimerFairness2(t *testing.T) {
|
|
|
|
done := make(chan bool)
|
|
|
|
c := make(chan bool)
|
|
|
|
for i := 0; i < 2; i++ {
|
|
|
|
go func() {
|
|
|
|
timer := time.After(20 * time.Millisecond)
|
|
|
|
var buf [1]byte
|
|
|
|
for {
|
|
|
|
syscall.Read(0, buf[0:0])
|
|
|
|
select {
|
|
|
|
case c <- true:
|
|
|
|
case <-c:
|
|
|
|
case <-timer:
|
|
|
|
done <- true
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
<-done
|
|
|
|
<-done
|
|
|
|
}
|
|
|
|
|
2013-06-28 07:52:17 -06:00
|
|
|
// The function is used to test preemption at split stack checks.
|
|
|
|
// Declaring a var avoids inlining at the call site.
|
|
|
|
var preempt = func() int {
|
|
|
|
var a [128]int
|
|
|
|
sum := 0
|
|
|
|
for _, v := range a {
|
|
|
|
sum += v
|
|
|
|
}
|
|
|
|
return sum
|
|
|
|
}
|
|
|
|
|
2013-07-18 15:22:26 -06:00
|
|
|
func TestPreemption(t *testing.T) {
|
|
|
|
// Test that goroutines are preempted at function calls.
|
2013-08-01 08:25:36 -06:00
|
|
|
N := 5
|
|
|
|
if testing.Short() {
|
|
|
|
N = 2
|
|
|
|
}
|
2013-07-18 15:22:26 -06:00
|
|
|
c := make(chan bool)
|
|
|
|
var x uint32
|
|
|
|
for g := 0; g < 2; g++ {
|
|
|
|
go func(g int) {
|
|
|
|
for i := 0; i < N; i++ {
|
|
|
|
for atomic.LoadUint32(&x) != uint32(g) {
|
|
|
|
preempt()
|
|
|
|
}
|
|
|
|
atomic.StoreUint32(&x, uint32(1-g))
|
|
|
|
}
|
|
|
|
c <- true
|
|
|
|
}(g)
|
|
|
|
}
|
|
|
|
<-c
|
|
|
|
<-c
|
|
|
|
}
|
|
|
|
|
2013-06-28 07:52:17 -06:00
|
|
|
func TestPreemptionGC(t *testing.T) {
|
|
|
|
// Test that pending GC preempts running goroutines.
|
2013-08-01 08:25:36 -06:00
|
|
|
P := 5
|
|
|
|
N := 10
|
|
|
|
if testing.Short() {
|
|
|
|
P = 3
|
|
|
|
N = 2
|
|
|
|
}
|
2013-06-28 07:52:17 -06:00
|
|
|
defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(P + 1))
|
|
|
|
var stop uint32
|
|
|
|
for i := 0; i < P; i++ {
|
|
|
|
go func() {
|
|
|
|
for atomic.LoadUint32(&stop) == 0 {
|
|
|
|
preempt()
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
2013-08-01 08:25:36 -06:00
|
|
|
for i := 0; i < N; i++ {
|
2013-06-28 07:52:17 -06:00
|
|
|
runtime.Gosched()
|
|
|
|
runtime.GC()
|
|
|
|
}
|
|
|
|
atomic.StoreUint32(&stop, 1)
|
|
|
|
}
|
|
|
|
|
2014-01-20 23:24:42 -07:00
|
|
|
func TestGCFairness(t *testing.T) {
|
|
|
|
output := executeTest(t, testGCFairnessSource, nil)
|
|
|
|
want := "OK\n"
|
|
|
|
if output != want {
|
|
|
|
t.Fatalf("want %s, got %s\n", want, output)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
const testGCFairnessSource = `
|
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"runtime"
|
|
|
|
"time"
|
|
|
|
)
|
|
|
|
|
|
|
|
func main() {
|
|
|
|
runtime.GOMAXPROCS(1)
|
|
|
|
f, err := os.Open("/dev/null")
|
|
|
|
if os.IsNotExist(err) {
|
|
|
|
// This test tests what it is intended to test only if writes are fast.
|
|
|
|
// If there is no /dev/null, we just don't execute the test.
|
2014-01-20 23:44:08 -07:00
|
|
|
fmt.Println("OK")
|
2014-01-20 23:24:42 -07:00
|
|
|
return
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
fmt.Println(err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
|
|
|
for i := 0; i < 2; i++ {
|
|
|
|
go func() {
|
|
|
|
for {
|
|
|
|
f.Write([]byte("."))
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
fmt.Println("OK")
|
|
|
|
}
|
|
|
|
`
|
|
|
|
|
2011-07-12 10:24:32 -06:00
|
|
|
func stackGrowthRecursive(i int) {
|
|
|
|
var pad [128]uint64
|
|
|
|
if i != 0 && pad[0] == 0 {
|
|
|
|
stackGrowthRecursive(i - 1)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
cmd/ld: fix large stack split for preempt check
If the stack frame size is larger than the known-unmapped region at the
bottom of the address space, then the stack split prologue cannot use the usual
condition:
SP - size >= stackguard
because SP - size may wrap around to a very large number.
Instead, if the stack frame is large, the prologue tests:
SP - stackguard >= size
(This ends up being a few instructions more expensive, so we don't do it always.)
Preemption requests register by setting stackguard to a very large value, so
that the first test (SP - size >= stackguard) cannot possibly succeed.
Unfortunately, that same very large value causes a wraparound in the
second test (SP - stackguard >= size), making it succeed incorrectly.
To avoid *that* wraparound, we have to amend the test:
stackguard != StackPreempt && SP - stackguard >= size
This test is only used for functions with large frames, which essentially
always split the stack, so the cost of the few instructions is noise.
This CL and CL 11085043 together fix the known issues with preemption,
at the beginning of a function, so we will be able to try turning it on again.
R=ken2
CC=golang-dev
https://golang.org/cl/11205043
2013-07-12 10:12:56 -06:00
|
|
|
func TestPreemptSplitBig(t *testing.T) {
|
|
|
|
if testing.Short() {
|
|
|
|
t.Skip("skipping in -short mode")
|
|
|
|
}
|
|
|
|
defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(2))
|
|
|
|
stop := make(chan int)
|
|
|
|
go big(stop)
|
|
|
|
for i := 0; i < 3; i++ {
|
2013-07-15 15:02:42 -06:00
|
|
|
time.Sleep(10 * time.Microsecond) // let big start running
|
cmd/ld: fix large stack split for preempt check
If the stack frame size is larger than the known-unmapped region at the
bottom of the address space, then the stack split prologue cannot use the usual
condition:
SP - size >= stackguard
because SP - size may wrap around to a very large number.
Instead, if the stack frame is large, the prologue tests:
SP - stackguard >= size
(This ends up being a few instructions more expensive, so we don't do it always.)
Preemption requests register by setting stackguard to a very large value, so
that the first test (SP - size >= stackguard) cannot possibly succeed.
Unfortunately, that same very large value causes a wraparound in the
second test (SP - stackguard >= size), making it succeed incorrectly.
To avoid *that* wraparound, we have to amend the test:
stackguard != StackPreempt && SP - stackguard >= size
This test is only used for functions with large frames, which essentially
always split the stack, so the cost of the few instructions is noise.
This CL and CL 11085043 together fix the known issues with preemption,
at the beginning of a function, so we will be able to try turning it on again.
R=ken2
CC=golang-dev
https://golang.org/cl/11205043
2013-07-12 10:12:56 -06:00
|
|
|
runtime.GC()
|
|
|
|
}
|
|
|
|
close(stop)
|
|
|
|
}
|
|
|
|
|
|
|
|
func big(stop chan int) int {
|
|
|
|
n := 0
|
|
|
|
for {
|
|
|
|
// delay so that gc is sure to have asked for a preemption
|
2013-07-15 15:02:42 -06:00
|
|
|
for i := 0; i < 1e9; i++ {
|
cmd/ld: fix large stack split for preempt check
If the stack frame size is larger than the known-unmapped region at the
bottom of the address space, then the stack split prologue cannot use the usual
condition:
SP - size >= stackguard
because SP - size may wrap around to a very large number.
Instead, if the stack frame is large, the prologue tests:
SP - stackguard >= size
(This ends up being a few instructions more expensive, so we don't do it always.)
Preemption requests register by setting stackguard to a very large value, so
that the first test (SP - size >= stackguard) cannot possibly succeed.
Unfortunately, that same very large value causes a wraparound in the
second test (SP - stackguard >= size), making it succeed incorrectly.
To avoid *that* wraparound, we have to amend the test:
stackguard != StackPreempt && SP - stackguard >= size
This test is only used for functions with large frames, which essentially
always split the stack, so the cost of the few instructions is noise.
This CL and CL 11085043 together fix the known issues with preemption,
at the beginning of a function, so we will be able to try turning it on again.
R=ken2
CC=golang-dev
https://golang.org/cl/11205043
2013-07-12 10:12:56 -06:00
|
|
|
n++
|
|
|
|
}
|
|
|
|
|
|
|
|
// call bigframe, which used to miss the preemption in its prologue.
|
|
|
|
bigframe(stop)
|
|
|
|
|
|
|
|
// check if we've been asked to stop.
|
|
|
|
select {
|
|
|
|
case <-stop:
|
|
|
|
return n
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func bigframe(stop chan int) int {
|
|
|
|
// not splitting the stack will overflow.
|
|
|
|
// small will notice that it needs a stack split and will
|
|
|
|
// catch the overflow.
|
|
|
|
var x [8192]byte
|
|
|
|
return small(stop, &x)
|
|
|
|
}
|
|
|
|
|
|
|
|
func small(stop chan int, x *[8192]byte) int {
|
|
|
|
for i := range x {
|
|
|
|
x[i] = byte(i)
|
|
|
|
}
|
|
|
|
sum := 0
|
|
|
|
for i := range x {
|
|
|
|
sum += int(x[i])
|
|
|
|
}
|
|
|
|
|
|
|
|
// keep small from being a leaf function, which might
|
|
|
|
// make it not do any stack check at all.
|
|
|
|
nonleaf(stop)
|
|
|
|
|
|
|
|
return sum
|
|
|
|
}
|
|
|
|
|
|
|
|
func nonleaf(stop chan int) bool {
|
|
|
|
// do something that won't be inlined:
|
|
|
|
select {
|
|
|
|
case <-stop:
|
|
|
|
return true
|
|
|
|
default:
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-22 21:48:02 -07:00
|
|
|
func TestSchedLocalQueue(t *testing.T) {
|
2014-09-06 11:07:23 -06:00
|
|
|
runtime.RunSchedLocalQueueTest()
|
2013-02-22 21:48:02 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestSchedLocalQueueSteal(t *testing.T) {
|
2014-09-06 11:07:23 -06:00
|
|
|
runtime.RunSchedLocalQueueStealTest()
|
2013-02-22 21:48:02 -07:00
|
|
|
}
|
|
|
|
|
2013-01-09 22:57:06 -07:00
|
|
|
func benchmarkStackGrowth(b *testing.B, rec int) {
|
2014-02-24 09:50:12 -07:00
|
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
|
|
for pb.Next() {
|
|
|
|
stackGrowthRecursive(rec)
|
|
|
|
}
|
|
|
|
})
|
2011-07-12 10:24:32 -06:00
|
|
|
}
|
2011-07-19 09:01:17 -06:00
|
|
|
|
2013-01-09 22:57:06 -07:00
|
|
|
func BenchmarkStackGrowth(b *testing.B) {
|
|
|
|
benchmarkStackGrowth(b, 10)
|
|
|
|
}
|
|
|
|
|
|
|
|
func BenchmarkStackGrowthDeep(b *testing.B) {
|
|
|
|
benchmarkStackGrowth(b, 1024)
|
|
|
|
}
|
|
|
|
|
2012-06-27 11:57:49 -06:00
|
|
|
func BenchmarkCreateGoroutines(b *testing.B) {
|
|
|
|
benchmarkCreateGoroutines(b, 1)
|
|
|
|
}
|
|
|
|
|
|
|
|
func BenchmarkCreateGoroutinesParallel(b *testing.B) {
|
|
|
|
benchmarkCreateGoroutines(b, runtime.GOMAXPROCS(-1))
|
|
|
|
}
|
|
|
|
|
|
|
|
func benchmarkCreateGoroutines(b *testing.B, procs int) {
|
|
|
|
c := make(chan bool)
|
|
|
|
var f func(n int)
|
|
|
|
f = func(n int) {
|
|
|
|
if n == 0 {
|
|
|
|
c <- true
|
|
|
|
return
|
|
|
|
}
|
|
|
|
go f(n - 1)
|
|
|
|
}
|
|
|
|
for i := 0; i < procs; i++ {
|
|
|
|
go f(b.N / procs)
|
|
|
|
}
|
|
|
|
for i := 0; i < procs; i++ {
|
|
|
|
<-c
|
|
|
|
}
|
|
|
|
}
|
2013-02-28 15:41:45 -07:00
|
|
|
|
2015-01-19 12:59:58 -07:00
|
|
|
func BenchmarkCreateGoroutinesCapture(b *testing.B) {
|
|
|
|
b.ReportAllocs()
|
|
|
|
for i := 0; i < b.N; i++ {
|
|
|
|
const N = 4
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(N)
|
|
|
|
for i := 0; i < N; i++ {
|
|
|
|
i := i
|
|
|
|
go func() {
|
|
|
|
if i >= N {
|
|
|
|
b.Logf("bad") // just to capture b
|
|
|
|
}
|
|
|
|
wg.Done()
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
cmd/gc: transform closure calls to function calls
Currently we always create context objects for closures that capture variables.
However, it is completely unnecessary for direct calls of closures
(whether it is func()(), defer func()() or go func()()).
This change transforms any OCALLFUNC(OCLOSURE) to normal function call.
Closed variables become function arguments.
This transformation is especially beneficial for go func(),
because we do not need to allocate context object on heap.
But it makes direct closure calls a bit faster as well (see BenchmarkClosureCall).
On implementation level it required to introduce yet another compiler pass.
However, the pass iterates only over xtop, so it should not be an issue.
Transformation consists of two parts: closure transformation and call site
transformation. We can't run these parts on different sides of escape analysis,
because tree state is inconsistent. We can do both parts during typecheck,
we don't know how to capture variables and don't have call site.
We can't do both parts during walk of OCALLFUNC, because we can walk
OCLOSURE body earlier.
So now capturevars pass only decides how to capture variables
(this info is required for escape analysis). New transformclosure
pass, that runs just before order/walk, does all transformations
of a closure. And later walk of OCALLFUNC(OCLOSURE) transforms call site.
benchmark old ns/op new ns/op delta
BenchmarkClosureCall 4.89 3.09 -36.81%
BenchmarkCreateGoroutinesCapture 1634 1294 -20.81%
benchmark old allocs new allocs delta
BenchmarkCreateGoroutinesCapture 6 2 -66.67%
benchmark old bytes new bytes delta
BenchmarkCreateGoroutinesCapture 176 48 -72.73%
Change-Id: Ic85e1706e18c3235cc45b3c0c031a9c1cdb7a40e
Reviewed-on: https://go-review.googlesource.com/4050
Reviewed-by: Russ Cox <rsc@golang.org>
2015-02-06 05:09:46 -07:00
|
|
|
func BenchmarkClosureCall(b *testing.B) {
|
|
|
|
sum := 0
|
|
|
|
off1 := 1
|
|
|
|
for i := 0; i < b.N; i++ {
|
|
|
|
off2 := 2
|
|
|
|
func() {
|
|
|
|
sum += i + off1 + off2
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
_ = sum
|
|
|
|
}
|
|
|
|
|
2013-02-28 15:41:45 -07:00
|
|
|
type Matrix [][]float64
|
|
|
|
|
|
|
|
func BenchmarkMatmult(b *testing.B) {
|
|
|
|
b.StopTimer()
|
|
|
|
// matmult is O(N**3) but testing expects O(b.N),
|
|
|
|
// so we need to take cube root of b.N
|
|
|
|
n := int(math.Cbrt(float64(b.N))) + 1
|
|
|
|
A := makeMatrix(n)
|
|
|
|
B := makeMatrix(n)
|
|
|
|
C := makeMatrix(n)
|
|
|
|
b.StartTimer()
|
|
|
|
matmult(nil, A, B, C, 0, n, 0, n, 0, n, 8)
|
|
|
|
}
|
|
|
|
|
|
|
|
func makeMatrix(n int) Matrix {
|
|
|
|
m := make(Matrix, n)
|
|
|
|
for i := 0; i < n; i++ {
|
|
|
|
m[i] = make([]float64, n)
|
|
|
|
for j := 0; j < n; j++ {
|
|
|
|
m[i][j] = float64(i*n + j)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return m
|
|
|
|
}
|
|
|
|
|
|
|
|
func matmult(done chan<- struct{}, A, B, C Matrix, i0, i1, j0, j1, k0, k1, threshold int) {
|
|
|
|
di := i1 - i0
|
|
|
|
dj := j1 - j0
|
|
|
|
dk := k1 - k0
|
|
|
|
if di >= dj && di >= dk && di >= threshold {
|
|
|
|
// divide in two by y axis
|
|
|
|
mi := i0 + di/2
|
|
|
|
done1 := make(chan struct{}, 1)
|
|
|
|
go matmult(done1, A, B, C, i0, mi, j0, j1, k0, k1, threshold)
|
|
|
|
matmult(nil, A, B, C, mi, i1, j0, j1, k0, k1, threshold)
|
|
|
|
<-done1
|
|
|
|
} else if dj >= dk && dj >= threshold {
|
|
|
|
// divide in two by x axis
|
|
|
|
mj := j0 + dj/2
|
|
|
|
done1 := make(chan struct{}, 1)
|
|
|
|
go matmult(done1, A, B, C, i0, i1, j0, mj, k0, k1, threshold)
|
|
|
|
matmult(nil, A, B, C, i0, i1, mj, j1, k0, k1, threshold)
|
|
|
|
<-done1
|
|
|
|
} else if dk >= threshold {
|
|
|
|
// divide in two by "k" axis
|
|
|
|
// deliberately not parallel because of data races
|
|
|
|
mk := k0 + dk/2
|
|
|
|
matmult(nil, A, B, C, i0, i1, j0, j1, k0, mk, threshold)
|
|
|
|
matmult(nil, A, B, C, i0, i1, j0, j1, mk, k1, threshold)
|
|
|
|
} else {
|
|
|
|
// the matrices are small enough, compute directly
|
|
|
|
for i := i0; i < i1; i++ {
|
|
|
|
for j := j0; j < j1; j++ {
|
|
|
|
for k := k0; k < k1; k++ {
|
|
|
|
C[i][j] += A[i][k] * B[k][j]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if done != nil {
|
|
|
|
done <- struct{}{}
|
|
|
|
}
|
|
|
|
}
|