mirror of
https://github.com/golang/go
synced 2024-11-18 22:44:48 -07:00
sync: make WaitGroup more space-efficient
The struct stores its 64-bit state field in a 12-byte array to ensure that it can be 64-bit-aligned. This leaves 4 spare bytes, which we can reuse to store the sema field. (32-bit alignment is still guaranteed because the array type was changed to [3]uint32.) Fixes #19149. Change-Id: I9bc20e69e45e0e07fbf496080f3650e8be0d6e8d Reviewed-on: https://go-review.googlesource.com/100515 Reviewed-by: Dmitry Vyukov <dvyukov@google.com> Run-TryBot: Brad Fitzpatrick <bradfitz@golang.org> TryBot-Result: Gobot Gobot <gobot@golang.org>
This commit is contained in:
parent
672729ebbd
commit
9ff7df003d
@ -23,16 +23,17 @@ type WaitGroup struct {
|
|||||||
// 64-bit value: high 32 bits are counter, low 32 bits are waiter count.
|
// 64-bit value: high 32 bits are counter, low 32 bits are waiter count.
|
||||||
// 64-bit atomic operations require 64-bit alignment, but 32-bit
|
// 64-bit atomic operations require 64-bit alignment, but 32-bit
|
||||||
// compilers do not ensure it. So we allocate 12 bytes and then use
|
// compilers do not ensure it. So we allocate 12 bytes and then use
|
||||||
// the aligned 8 bytes in them as state.
|
// the aligned 8 bytes in them as state, and the other 4 as storage
|
||||||
state1 [12]byte
|
// for the sema.
|
||||||
sema uint32
|
state1 [3]uint32
|
||||||
}
|
}
|
||||||
|
|
||||||
func (wg *WaitGroup) state() *uint64 {
|
// state returns pointers to the state and sema fields stored within wg.state1.
|
||||||
|
func (wg *WaitGroup) state() (statep *uint64, semap *uint32) {
|
||||||
if uintptr(unsafe.Pointer(&wg.state1))%8 == 0 {
|
if uintptr(unsafe.Pointer(&wg.state1))%8 == 0 {
|
||||||
return (*uint64)(unsafe.Pointer(&wg.state1))
|
return (*uint64)(unsafe.Pointer(&wg.state1)), &wg.state1[2]
|
||||||
} else {
|
} else {
|
||||||
return (*uint64)(unsafe.Pointer(&wg.state1[4]))
|
return (*uint64)(unsafe.Pointer(&wg.state1[1])), &wg.state1[0]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -50,7 +51,7 @@ func (wg *WaitGroup) state() *uint64 {
|
|||||||
// new Add calls must happen after all previous Wait calls have returned.
|
// new Add calls must happen after all previous Wait calls have returned.
|
||||||
// See the WaitGroup example.
|
// See the WaitGroup example.
|
||||||
func (wg *WaitGroup) Add(delta int) {
|
func (wg *WaitGroup) Add(delta int) {
|
||||||
statep := wg.state()
|
statep, semap := wg.state()
|
||||||
if race.Enabled {
|
if race.Enabled {
|
||||||
_ = *statep // trigger nil deref early
|
_ = *statep // trigger nil deref early
|
||||||
if delta < 0 {
|
if delta < 0 {
|
||||||
@ -67,7 +68,7 @@ func (wg *WaitGroup) Add(delta int) {
|
|||||||
// The first increment must be synchronized with Wait.
|
// The first increment must be synchronized with Wait.
|
||||||
// Need to model this as a read, because there can be
|
// Need to model this as a read, because there can be
|
||||||
// several concurrent wg.counter transitions from 0.
|
// several concurrent wg.counter transitions from 0.
|
||||||
race.Read(unsafe.Pointer(&wg.sema))
|
race.Read(unsafe.Pointer(semap))
|
||||||
}
|
}
|
||||||
if v < 0 {
|
if v < 0 {
|
||||||
panic("sync: negative WaitGroup counter")
|
panic("sync: negative WaitGroup counter")
|
||||||
@ -89,7 +90,7 @@ func (wg *WaitGroup) Add(delta int) {
|
|||||||
// Reset waiters count to 0.
|
// Reset waiters count to 0.
|
||||||
*statep = 0
|
*statep = 0
|
||||||
for ; w != 0; w-- {
|
for ; w != 0; w-- {
|
||||||
runtime_Semrelease(&wg.sema, false)
|
runtime_Semrelease(semap, false)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -100,7 +101,7 @@ func (wg *WaitGroup) Done() {
|
|||||||
|
|
||||||
// Wait blocks until the WaitGroup counter is zero.
|
// Wait blocks until the WaitGroup counter is zero.
|
||||||
func (wg *WaitGroup) Wait() {
|
func (wg *WaitGroup) Wait() {
|
||||||
statep := wg.state()
|
statep, semap := wg.state()
|
||||||
if race.Enabled {
|
if race.Enabled {
|
||||||
_ = *statep // trigger nil deref early
|
_ = *statep // trigger nil deref early
|
||||||
race.Disable()
|
race.Disable()
|
||||||
@ -124,9 +125,9 @@ func (wg *WaitGroup) Wait() {
|
|||||||
// Need to model this is as a write to race with the read in Add.
|
// Need to model this is as a write to race with the read in Add.
|
||||||
// As a consequence, can do the write only for the first waiter,
|
// As a consequence, can do the write only for the first waiter,
|
||||||
// otherwise concurrent Waits will race with each other.
|
// otherwise concurrent Waits will race with each other.
|
||||||
race.Write(unsafe.Pointer(&wg.sema))
|
race.Write(unsafe.Pointer(semap))
|
||||||
}
|
}
|
||||||
runtime_Semacquire(&wg.sema)
|
runtime_Semacquire(semap)
|
||||||
if *statep != 0 {
|
if *statep != 0 {
|
||||||
panic("sync: WaitGroup is reused before previous Wait has returned")
|
panic("sync: WaitGroup is reused before previous Wait has returned")
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user