mirror of
https://github.com/golang/go
synced 2024-11-19 16:04:48 -07:00
99da8730b0
Went mainly for the ones that make no sense, such as the ones mid-sentence or after commas. Change-Id: Ie245d2c19cc7428a06295635cf6a9482ade25ff0 Reviewed-on: https://go-review.googlesource.com/57293 Reviewed-by: Ian Lance Taylor <iant@golang.org> Run-TryBot: Daniel Martí <mvdan@mvdan.cc> TryBot-Result: Gobot Gobot <gobot@golang.org>
285 lines
6.4 KiB
Go
285 lines
6.4 KiB
Go
// Copyright 2011 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// +build darwin nacl netbsd openbsd plan9 solaris windows
|
|
|
|
package runtime
|
|
|
|
import (
|
|
"runtime/internal/atomic"
|
|
"unsafe"
|
|
)
|
|
|
|
// This implementation depends on OS-specific implementations of
|
|
//
|
|
// func semacreate(mp *m)
|
|
// Create a semaphore for mp, if it does not already have one.
|
|
//
|
|
// func semasleep(ns int64) int32
|
|
// If ns < 0, acquire m's semaphore and return 0.
|
|
// If ns >= 0, try to acquire m's semaphore for at most ns nanoseconds.
|
|
// Return 0 if the semaphore was acquired, -1 if interrupted or timed out.
|
|
//
|
|
// func semawakeup(mp *m)
|
|
// Wake up mp, which is or will soon be sleeping on its semaphore.
|
|
//
|
|
const (
|
|
locked uintptr = 1
|
|
|
|
active_spin = 4
|
|
active_spin_cnt = 30
|
|
passive_spin = 1
|
|
)
|
|
|
|
func lock(l *mutex) {
|
|
gp := getg()
|
|
if gp.m.locks < 0 {
|
|
throw("runtime·lock: lock count")
|
|
}
|
|
gp.m.locks++
|
|
|
|
// Speculative grab for lock.
|
|
if atomic.Casuintptr(&l.key, 0, locked) {
|
|
return
|
|
}
|
|
semacreate(gp.m)
|
|
|
|
// On uniprocessor's, no point spinning.
|
|
// On multiprocessors, spin for ACTIVE_SPIN attempts.
|
|
spin := 0
|
|
if ncpu > 1 {
|
|
spin = active_spin
|
|
}
|
|
Loop:
|
|
for i := 0; ; i++ {
|
|
v := atomic.Loaduintptr(&l.key)
|
|
if v&locked == 0 {
|
|
// Unlocked. Try to lock.
|
|
if atomic.Casuintptr(&l.key, v, v|locked) {
|
|
return
|
|
}
|
|
i = 0
|
|
}
|
|
if i < spin {
|
|
procyield(active_spin_cnt)
|
|
} else if i < spin+passive_spin {
|
|
osyield()
|
|
} else {
|
|
// Someone else has it.
|
|
// l->waitm points to a linked list of M's waiting
|
|
// for this lock, chained through m->nextwaitm.
|
|
// Queue this M.
|
|
for {
|
|
gp.m.nextwaitm = v &^ locked
|
|
if atomic.Casuintptr(&l.key, v, uintptr(unsafe.Pointer(gp.m))|locked) {
|
|
break
|
|
}
|
|
v = atomic.Loaduintptr(&l.key)
|
|
if v&locked == 0 {
|
|
continue Loop
|
|
}
|
|
}
|
|
if v&locked != 0 {
|
|
// Queued. Wait.
|
|
semasleep(-1)
|
|
i = 0
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
//go:nowritebarrier
|
|
// We might not be holding a p in this code.
|
|
func unlock(l *mutex) {
|
|
gp := getg()
|
|
var mp *m
|
|
for {
|
|
v := atomic.Loaduintptr(&l.key)
|
|
if v == locked {
|
|
if atomic.Casuintptr(&l.key, locked, 0) {
|
|
break
|
|
}
|
|
} else {
|
|
// Other M's are waiting for the lock.
|
|
// Dequeue an M.
|
|
mp = (*m)(unsafe.Pointer(v &^ locked))
|
|
if atomic.Casuintptr(&l.key, v, mp.nextwaitm) {
|
|
// Dequeued an M. Wake it.
|
|
semawakeup(mp)
|
|
break
|
|
}
|
|
}
|
|
}
|
|
gp.m.locks--
|
|
if gp.m.locks < 0 {
|
|
throw("runtime·unlock: lock count")
|
|
}
|
|
if gp.m.locks == 0 && gp.preempt { // restore the preemption request in case we've cleared it in newstack
|
|
gp.stackguard0 = stackPreempt
|
|
}
|
|
}
|
|
|
|
// One-time notifications.
|
|
func noteclear(n *note) {
|
|
n.key = 0
|
|
}
|
|
|
|
func notewakeup(n *note) {
|
|
var v uintptr
|
|
for {
|
|
v = atomic.Loaduintptr(&n.key)
|
|
if atomic.Casuintptr(&n.key, v, locked) {
|
|
break
|
|
}
|
|
}
|
|
|
|
// Successfully set waitm to locked.
|
|
// What was it before?
|
|
switch {
|
|
case v == 0:
|
|
// Nothing was waiting. Done.
|
|
case v == locked:
|
|
// Two notewakeups! Not allowed.
|
|
throw("notewakeup - double wakeup")
|
|
default:
|
|
// Must be the waiting m. Wake it up.
|
|
semawakeup((*m)(unsafe.Pointer(v)))
|
|
}
|
|
}
|
|
|
|
func notesleep(n *note) {
|
|
gp := getg()
|
|
if gp != gp.m.g0 {
|
|
throw("notesleep not on g0")
|
|
}
|
|
semacreate(gp.m)
|
|
if !atomic.Casuintptr(&n.key, 0, uintptr(unsafe.Pointer(gp.m))) {
|
|
// Must be locked (got wakeup).
|
|
if n.key != locked {
|
|
throw("notesleep - waitm out of sync")
|
|
}
|
|
return
|
|
}
|
|
// Queued. Sleep.
|
|
gp.m.blocked = true
|
|
if *cgo_yield == nil {
|
|
semasleep(-1)
|
|
} else {
|
|
// Sleep for an arbitrary-but-moderate interval to poll libc interceptors.
|
|
const ns = 10e6
|
|
for atomic.Loaduintptr(&n.key) == 0 {
|
|
semasleep(ns)
|
|
asmcgocall(*cgo_yield, nil)
|
|
}
|
|
}
|
|
gp.m.blocked = false
|
|
}
|
|
|
|
//go:nosplit
|
|
func notetsleep_internal(n *note, ns int64, gp *g, deadline int64) bool {
|
|
// gp and deadline are logically local variables, but they are written
|
|
// as parameters so that the stack space they require is charged
|
|
// to the caller.
|
|
// This reduces the nosplit footprint of notetsleep_internal.
|
|
gp = getg()
|
|
|
|
// Register for wakeup on n->waitm.
|
|
if !atomic.Casuintptr(&n.key, 0, uintptr(unsafe.Pointer(gp.m))) {
|
|
// Must be locked (got wakeup).
|
|
if n.key != locked {
|
|
throw("notetsleep - waitm out of sync")
|
|
}
|
|
return true
|
|
}
|
|
if ns < 0 {
|
|
// Queued. Sleep.
|
|
gp.m.blocked = true
|
|
if *cgo_yield == nil {
|
|
semasleep(-1)
|
|
} else {
|
|
// Sleep in arbitrary-but-moderate intervals to poll libc interceptors.
|
|
const ns = 10e6
|
|
for semasleep(ns) < 0 {
|
|
asmcgocall(*cgo_yield, nil)
|
|
}
|
|
}
|
|
gp.m.blocked = false
|
|
return true
|
|
}
|
|
|
|
deadline = nanotime() + ns
|
|
for {
|
|
// Registered. Sleep.
|
|
gp.m.blocked = true
|
|
if *cgo_yield != nil && ns > 10e6 {
|
|
ns = 10e6
|
|
}
|
|
if semasleep(ns) >= 0 {
|
|
gp.m.blocked = false
|
|
// Acquired semaphore, semawakeup unregistered us.
|
|
// Done.
|
|
return true
|
|
}
|
|
if *cgo_yield != nil {
|
|
asmcgocall(*cgo_yield, nil)
|
|
}
|
|
gp.m.blocked = false
|
|
// Interrupted or timed out. Still registered. Semaphore not acquired.
|
|
ns = deadline - nanotime()
|
|
if ns <= 0 {
|
|
break
|
|
}
|
|
// Deadline hasn't arrived. Keep sleeping.
|
|
}
|
|
|
|
// Deadline arrived. Still registered. Semaphore not acquired.
|
|
// Want to give up and return, but have to unregister first,
|
|
// so that any notewakeup racing with the return does not
|
|
// try to grant us the semaphore when we don't expect it.
|
|
for {
|
|
v := atomic.Loaduintptr(&n.key)
|
|
switch v {
|
|
case uintptr(unsafe.Pointer(gp.m)):
|
|
// No wakeup yet; unregister if possible.
|
|
if atomic.Casuintptr(&n.key, v, 0) {
|
|
return false
|
|
}
|
|
case locked:
|
|
// Wakeup happened so semaphore is available.
|
|
// Grab it to avoid getting out of sync.
|
|
gp.m.blocked = true
|
|
if semasleep(-1) < 0 {
|
|
throw("runtime: unable to acquire - semaphore out of sync")
|
|
}
|
|
gp.m.blocked = false
|
|
return true
|
|
default:
|
|
throw("runtime: unexpected waitm - semaphore out of sync")
|
|
}
|
|
}
|
|
}
|
|
|
|
func notetsleep(n *note, ns int64) bool {
|
|
gp := getg()
|
|
if gp != gp.m.g0 && gp.m.preemptoff != "" {
|
|
throw("notetsleep not on g0")
|
|
}
|
|
semacreate(gp.m)
|
|
return notetsleep_internal(n, ns, nil, 0)
|
|
}
|
|
|
|
// same as runtime·notetsleep, but called on user g (not g0)
|
|
// calls only nosplit functions between entersyscallblock/exitsyscall
|
|
func notetsleepg(n *note, ns int64) bool {
|
|
gp := getg()
|
|
if gp == gp.m.g0 {
|
|
throw("notetsleepg on g0")
|
|
}
|
|
semacreate(gp.m)
|
|
entersyscallblock(0)
|
|
ok := notetsleep_internal(n, ns, nil, 0)
|
|
exitsyscall(0)
|
|
return ok
|
|
}
|