mirror of
https://github.com/golang/go
synced 2024-11-05 11:56:12 -07:00
cmd/compile/internal/gc: change bvfoo functions into bvec methods
plive.go (except for printeffects and livenessprintblock) and reflect.go changes were prepared mechanically with gofmt -r. Passes toolstash. name old alloc/op new alloc/op delta Template 44.3MB ± 0% 44.3MB ± 0% ~ (p=0.367 n=30+30) Unicode 37.4MB ± 0% 37.4MB ± 0% ~ (p=0.665 n=30+30) GoTypes 125MB ± 0% 125MB ± 0% ~ (p=0.067 n=30+30) Compiler 515MB ± 0% 515MB ± 0% ~ (p=0.542 n=30+28) name old allocs/op new allocs/op delta Template 434k ± 0% 434k ± 0% ~ (p=0.076 n=30+29) Unicode 367k ± 0% 367k ± 0% ~ (p=0.716 n=29+30) GoTypes 1.24M ± 0% 1.24M ± 0% ~ (p=0.428 n=29+29) Compiler 4.47M ± 0% 4.47M ± 0% ~ (p=0.225 n=28+30) Change-Id: Ibaf0668567b3f69fba06aa03b7997c8fb152113a Reviewed-on: https://go-review.googlesource.com/30356 Run-TryBot: Matthew Dempsky <mdempsky@google.com> TryBot-Result: Gobot Gobot <gobot@golang.org> Reviewed-by: Brad Fitzpatrick <bradfitz@golang.org>
This commit is contained in:
parent
7478ea5dba
commit
dce0df29dd
@ -4,8 +4,6 @@
|
||||
|
||||
package gc
|
||||
|
||||
import "fmt"
|
||||
|
||||
const (
|
||||
WORDBITS = 32
|
||||
WORDMASK = WORDBITS - 1
|
||||
@ -44,14 +42,7 @@ func (b *bulkBvec) next() bvec {
|
||||
return out
|
||||
}
|
||||
|
||||
// difference
|
||||
func bvandnot(dst bvec, src1 bvec, src2 bvec) {
|
||||
for i, x := range src1.b {
|
||||
dst.b[i] = x &^ src2.b[i]
|
||||
}
|
||||
}
|
||||
|
||||
func bveq(bv1 bvec, bv2 bvec) bool {
|
||||
func (bv1 bvec) Eq(bv2 bvec) bool {
|
||||
if bv1.n != bv2.n {
|
||||
Fatalf("bvequal: lengths %d and %d are not equal", bv1.n, bv2.n)
|
||||
}
|
||||
@ -63,22 +54,31 @@ func bveq(bv1 bvec, bv2 bvec) bool {
|
||||
return true
|
||||
}
|
||||
|
||||
func bvcopy(dst bvec, src bvec) {
|
||||
func (dst bvec) Copy(src bvec) {
|
||||
for i, x := range src.b {
|
||||
dst.b[i] = x
|
||||
}
|
||||
}
|
||||
|
||||
func bvget(bv bvec, i int32) int {
|
||||
func (bv bvec) Get(i int32) bool {
|
||||
if i < 0 || i >= bv.n {
|
||||
Fatalf("bvget: index %d is out of bounds with length %d\n", i, bv.n)
|
||||
}
|
||||
return int((bv.b[i>>WORDSHIFT] >> uint(i&WORDMASK)) & 1)
|
||||
mask := uint32(1 << uint(i%WORDBITS))
|
||||
return bv.b[i>>WORDSHIFT]&mask != 0
|
||||
}
|
||||
|
||||
func (bv bvec) Set(i int32) {
|
||||
if i < 0 || i >= bv.n {
|
||||
Fatalf("bvset: index %d is out of bounds with length %d\n", i, bv.n)
|
||||
}
|
||||
mask := uint32(1 << uint(i%WORDBITS))
|
||||
bv.b[i/WORDBITS] |= mask
|
||||
}
|
||||
|
||||
// bvnext returns the smallest index >= i for which bvget(bv, i) == 1.
|
||||
// If there is no such index, bvnext returns -1.
|
||||
func bvnext(bv bvec, i int32) int32 {
|
||||
func (bv bvec) Next(i int32) int32 {
|
||||
if i >= bv.n {
|
||||
return -1
|
||||
}
|
||||
@ -107,7 +107,7 @@ func bvnext(bv bvec, i int32) int32 {
|
||||
return i
|
||||
}
|
||||
|
||||
func bvisempty(bv bvec) bool {
|
||||
func (bv bvec) IsEmpty() bool {
|
||||
for i := int32(0); i < bv.n; i += WORDBITS {
|
||||
if bv.b[i>>WORDSHIFT] != 0 {
|
||||
return false
|
||||
@ -116,7 +116,7 @@ func bvisempty(bv bvec) bool {
|
||||
return true
|
||||
}
|
||||
|
||||
func bvnot(bv bvec) {
|
||||
func (bv bvec) Not() {
|
||||
i := int32(0)
|
||||
w := int32(0)
|
||||
for ; i < bv.n; i, w = i+WORDBITS, w+1 {
|
||||
@ -125,36 +125,41 @@ func bvnot(bv bvec) {
|
||||
}
|
||||
|
||||
// union
|
||||
func bvor(dst bvec, src1 bvec, src2 bvec) {
|
||||
func (dst bvec) Or(src1, src2 bvec) {
|
||||
for i, x := range src1.b {
|
||||
dst.b[i] = x | src2.b[i]
|
||||
}
|
||||
}
|
||||
|
||||
// intersection
|
||||
func bvand(dst bvec, src1 bvec, src2 bvec) {
|
||||
func (dst bvec) And(src1, src2 bvec) {
|
||||
for i, x := range src1.b {
|
||||
dst.b[i] = x & src2.b[i]
|
||||
}
|
||||
}
|
||||
|
||||
func bvprint(bv bvec) {
|
||||
fmt.Printf("#*")
|
||||
for i := int32(0); i < bv.n; i++ {
|
||||
fmt.Printf("%d", bvget(bv, i))
|
||||
// difference
|
||||
func (dst bvec) AndNot(src1, src2 bvec) {
|
||||
for i, x := range src1.b {
|
||||
dst.b[i] = x &^ src2.b[i]
|
||||
}
|
||||
}
|
||||
|
||||
func bvresetall(bv bvec) {
|
||||
func (bv bvec) String() string {
|
||||
s := make([]byte, 2+bv.n)
|
||||
copy(s, "#*")
|
||||
for i := int32(0); i < bv.n; i++ {
|
||||
ch := byte('0')
|
||||
if bv.Get(i) {
|
||||
ch = '1'
|
||||
}
|
||||
s[2+i] = ch
|
||||
}
|
||||
return string(s)
|
||||
}
|
||||
|
||||
func (bv bvec) Clear() {
|
||||
for i := range bv.b {
|
||||
bv.b[i] = 0
|
||||
}
|
||||
}
|
||||
|
||||
func bvset(bv bvec, i int32) {
|
||||
if i < 0 || i >= bv.n {
|
||||
Fatalf("bvset: index %d is out of bounds with length %d\n", i, bv.n)
|
||||
}
|
||||
mask := uint32(1 << uint(i%WORDBITS))
|
||||
bv.b[i/WORDBITS] |= mask
|
||||
}
|
||||
|
@ -545,9 +545,9 @@ func isfunny(n *Node) bool {
|
||||
// initialized, because any use of a variable must come after its
|
||||
// initialization.
|
||||
func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarinit bvec) {
|
||||
bvresetall(uevar)
|
||||
bvresetall(varkill)
|
||||
bvresetall(avarinit)
|
||||
uevar.Clear()
|
||||
varkill.Clear()
|
||||
avarinit.Clear()
|
||||
|
||||
// A return instruction with a p.to is a tail return, which brings
|
||||
// the stack pointer back up (if it ever went down) and then jumps
|
||||
@ -560,7 +560,7 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
// See issue 16016.
|
||||
for i, node := range vars {
|
||||
if node.Class == PPARAM {
|
||||
bvset(uevar, int32(i))
|
||||
uevar.Set(int32(i))
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -578,7 +578,7 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
// non-tail-call return instructions; see note below for details.
|
||||
case PPARAMOUT:
|
||||
if !node.Addrtaken && prog.To.Type == obj.TYPE_NONE {
|
||||
bvset(uevar, int32(i))
|
||||
uevar.Set(int32(i))
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -593,9 +593,9 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
switch node.Class {
|
||||
case PPARAM:
|
||||
if node.Addrtaken {
|
||||
bvset(avarinit, int32(i))
|
||||
avarinit.Set(int32(i))
|
||||
}
|
||||
bvset(varkill, int32(i))
|
||||
varkill.Set(int32(i))
|
||||
}
|
||||
}
|
||||
|
||||
@ -610,14 +610,14 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
n := from.Node.(*Node)
|
||||
if pos := liveIndex(n, vars); pos >= 0 {
|
||||
if n.Addrtaken {
|
||||
bvset(avarinit, pos)
|
||||
avarinit.Set(pos)
|
||||
} else {
|
||||
if info.Flags&(LeftRead|LeftAddr) != 0 {
|
||||
bvset(uevar, pos)
|
||||
uevar.Set(pos)
|
||||
}
|
||||
if info.Flags&LeftWrite != 0 {
|
||||
if !isfat(n.Type) {
|
||||
bvset(varkill, pos)
|
||||
varkill.Set(pos)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -631,9 +631,9 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
n := from.Node.(*Node)
|
||||
if pos := liveIndex(n, vars); pos >= 0 {
|
||||
if n.Addrtaken {
|
||||
bvset(avarinit, pos)
|
||||
avarinit.Set(pos)
|
||||
} else {
|
||||
bvset(uevar, pos)
|
||||
uevar.Set(pos)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -646,10 +646,10 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
if pos := liveIndex(n, vars); pos >= 0 {
|
||||
if n.Addrtaken {
|
||||
if prog.As != obj.AVARKILL {
|
||||
bvset(avarinit, pos)
|
||||
avarinit.Set(pos)
|
||||
}
|
||||
if prog.As == obj.AVARDEF || prog.As == obj.AVARKILL {
|
||||
bvset(varkill, pos)
|
||||
varkill.Set(pos)
|
||||
}
|
||||
} else {
|
||||
// RightRead is a read, obviously.
|
||||
@ -661,11 +661,11 @@ func progeffects(prog *obj.Prog, vars []*Node, uevar bvec, varkill bvec, avarini
|
||||
// having the RightAddr bit set keeps the registerizer from
|
||||
// trying to substitute a register for the memory location.
|
||||
if (info.Flags&RightRead != 0) || info.Flags&(RightAddr|RightWrite) == RightAddr {
|
||||
bvset(uevar, pos)
|
||||
uevar.Set(pos)
|
||||
}
|
||||
if info.Flags&RightWrite != 0 {
|
||||
if !isfat(n.Type) || prog.As == obj.AVARDEF {
|
||||
bvset(varkill, pos)
|
||||
varkill.Set(pos)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -719,14 +719,10 @@ func newliveness(fn *Node, ptxt *obj.Prog, cfg []*BasicBlock, vars []*Node) *Liv
|
||||
}
|
||||
|
||||
func printeffects(p *obj.Prog, uevar bvec, varkill bvec, avarinit bvec) {
|
||||
fmt.Printf("effects of %v", p)
|
||||
fmt.Printf("\nuevar: ")
|
||||
bvprint(uevar)
|
||||
fmt.Printf("\nvarkill: ")
|
||||
bvprint(varkill)
|
||||
fmt.Printf("\navarinit: ")
|
||||
bvprint(avarinit)
|
||||
fmt.Printf("\n")
|
||||
fmt.Printf("effects of %v\n", p)
|
||||
fmt.Println("uevar:", uevar)
|
||||
fmt.Println("varkill:", varkill)
|
||||
fmt.Println("avarinit:", avarinit)
|
||||
}
|
||||
|
||||
// Pretty print a variable node. Uses Pascal like conventions for pointers and
|
||||
@ -748,7 +744,7 @@ func printnode(node *Node) {
|
||||
func printvars(name string, bv bvec, vars []*Node) {
|
||||
fmt.Printf("%s:", name)
|
||||
for i, node := range vars {
|
||||
if bvget(bv, int32(i)) != 0 {
|
||||
if bv.Get(int32(i)) {
|
||||
printnode(node)
|
||||
}
|
||||
}
|
||||
@ -786,8 +782,7 @@ func livenessprintblock(lv *Liveness, bb *BasicBlock) {
|
||||
if prog.As == obj.APCDATA && prog.From.Offset == obj.PCDATA_StackMapIndex {
|
||||
pos := int32(prog.To.Offset)
|
||||
live := lv.livepointers[pos]
|
||||
fmt.Printf(" ")
|
||||
bvprint(live)
|
||||
fmt.Printf(" %s", live.String())
|
||||
}
|
||||
|
||||
fmt.Printf("\n")
|
||||
@ -913,7 +908,7 @@ func onebitwalktype1(t *Type, xoffset *int64, bv bvec) {
|
||||
if *xoffset&int64(Widthptr-1) != 0 {
|
||||
Fatalf("onebitwalktype1: invalid alignment, %v", t)
|
||||
}
|
||||
bvset(bv, int32(*xoffset/int64(Widthptr))) // pointer
|
||||
bv.Set(int32(*xoffset / int64(Widthptr))) // pointer
|
||||
*xoffset += t.Width
|
||||
|
||||
case TSTRING:
|
||||
@ -921,7 +916,7 @@ func onebitwalktype1(t *Type, xoffset *int64, bv bvec) {
|
||||
if *xoffset&int64(Widthptr-1) != 0 {
|
||||
Fatalf("onebitwalktype1: invalid alignment, %v", t)
|
||||
}
|
||||
bvset(bv, int32(*xoffset/int64(Widthptr))) //pointer in first slot
|
||||
bv.Set(int32(*xoffset / int64(Widthptr))) //pointer in first slot
|
||||
*xoffset += t.Width
|
||||
|
||||
case TINTER:
|
||||
@ -931,8 +926,8 @@ func onebitwalktype1(t *Type, xoffset *int64, bv bvec) {
|
||||
if *xoffset&int64(Widthptr-1) != 0 {
|
||||
Fatalf("onebitwalktype1: invalid alignment, %v", t)
|
||||
}
|
||||
bvset(bv, int32(*xoffset/int64(Widthptr))) // pointer in first slot
|
||||
bvset(bv, int32(*xoffset/int64(Widthptr)+1)) // pointer in second slot
|
||||
bv.Set(int32(*xoffset / int64(Widthptr))) // pointer in first slot
|
||||
bv.Set(int32(*xoffset/int64(Widthptr) + 1)) // pointer in second slot
|
||||
*xoffset += t.Width
|
||||
|
||||
case TSLICE:
|
||||
@ -940,7 +935,7 @@ func onebitwalktype1(t *Type, xoffset *int64, bv bvec) {
|
||||
if *xoffset&int64(Widthptr-1) != 0 {
|
||||
Fatalf("onebitwalktype1: invalid TARRAY alignment, %v", t)
|
||||
}
|
||||
bvset(bv, int32(*xoffset/int64(Widthptr))) // pointer in first slot (BitsPointer)
|
||||
bv.Set(int32(*xoffset / int64(Widthptr))) // pointer in first slot (BitsPointer)
|
||||
*xoffset += t.Width
|
||||
|
||||
case TARRAY:
|
||||
@ -981,7 +976,7 @@ func onebitlivepointermap(lv *Liveness, liveout bvec, vars []*Node, args bvec, l
|
||||
var xoffset int64
|
||||
|
||||
for i := int32(0); ; i++ {
|
||||
i = bvnext(liveout, i)
|
||||
i = liveout.Next(i)
|
||||
if i < 0 {
|
||||
break
|
||||
}
|
||||
@ -1040,22 +1035,22 @@ func livenessprologue(lv *Liveness) {
|
||||
if debuglive >= 3 {
|
||||
printeffects(p, uevar, varkill, avarinit)
|
||||
}
|
||||
bvor(bb.varkill, bb.varkill, varkill)
|
||||
bvandnot(bb.uevar, bb.uevar, varkill)
|
||||
bvor(bb.uevar, bb.uevar, uevar)
|
||||
bb.varkill.Or(bb.varkill, varkill)
|
||||
bb.uevar.AndNot(bb.uevar, varkill)
|
||||
bb.uevar.Or(bb.uevar, uevar)
|
||||
}
|
||||
|
||||
// Walk the block instructions forward to update avarinit bits.
|
||||
// avarinit describes the effect at the end of the block, not the beginning.
|
||||
bvresetall(varkill)
|
||||
varkill.Clear()
|
||||
|
||||
for p := bb.first; ; p = p.Link {
|
||||
progeffects(p, lv.vars, uevar, varkill, avarinit)
|
||||
if debuglive >= 3 {
|
||||
printeffects(p, uevar, varkill, avarinit)
|
||||
}
|
||||
bvandnot(bb.avarinit, bb.avarinit, varkill)
|
||||
bvor(bb.avarinit, bb.avarinit, avarinit)
|
||||
bb.avarinit.AndNot(bb.avarinit, varkill)
|
||||
bb.avarinit.Or(bb.avarinit, avarinit)
|
||||
if p == bb.last {
|
||||
break
|
||||
}
|
||||
@ -1078,41 +1073,41 @@ func livenesssolve(lv *Liveness) {
|
||||
// avarinitany says the addressed var is initialized along some path reaching the block exit.
|
||||
for i, bb := range lv.cfg {
|
||||
if i == 0 {
|
||||
bvcopy(bb.avarinitall, bb.avarinit)
|
||||
bb.avarinitall.Copy(bb.avarinit)
|
||||
} else {
|
||||
bvresetall(bb.avarinitall)
|
||||
bvnot(bb.avarinitall)
|
||||
bb.avarinitall.Clear()
|
||||
bb.avarinitall.Not()
|
||||
}
|
||||
bvcopy(bb.avarinitany, bb.avarinit)
|
||||
bb.avarinitany.Copy(bb.avarinit)
|
||||
}
|
||||
|
||||
for change := true; change; {
|
||||
change = false
|
||||
for _, bb := range lv.cfg {
|
||||
bvresetall(any)
|
||||
bvresetall(all)
|
||||
any.Clear()
|
||||
all.Clear()
|
||||
for j, pred := range bb.pred {
|
||||
if j == 0 {
|
||||
bvcopy(any, pred.avarinitany)
|
||||
bvcopy(all, pred.avarinitall)
|
||||
any.Copy(pred.avarinitany)
|
||||
all.Copy(pred.avarinitall)
|
||||
} else {
|
||||
bvor(any, any, pred.avarinitany)
|
||||
bvand(all, all, pred.avarinitall)
|
||||
any.Or(any, pred.avarinitany)
|
||||
all.And(all, pred.avarinitall)
|
||||
}
|
||||
}
|
||||
|
||||
bvandnot(any, any, bb.varkill)
|
||||
bvandnot(all, all, bb.varkill)
|
||||
bvor(any, any, bb.avarinit)
|
||||
bvor(all, all, bb.avarinit)
|
||||
if !bveq(any, bb.avarinitany) {
|
||||
any.AndNot(any, bb.varkill)
|
||||
all.AndNot(all, bb.varkill)
|
||||
any.Or(any, bb.avarinit)
|
||||
all.Or(all, bb.avarinit)
|
||||
if !any.Eq(bb.avarinitany) {
|
||||
change = true
|
||||
bvcopy(bb.avarinitany, any)
|
||||
bb.avarinitany.Copy(any)
|
||||
}
|
||||
|
||||
if !bveq(all, bb.avarinitall) {
|
||||
if !all.Eq(bb.avarinitall) {
|
||||
change = true
|
||||
bvcopy(bb.avarinitall, all)
|
||||
bb.avarinitall.Copy(all)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1133,14 +1128,14 @@ func livenesssolve(lv *Liveness) {
|
||||
// if it is live on input to some successor.
|
||||
//
|
||||
// out[b] = \bigcup_{s \in succ[b]} in[s]
|
||||
bvresetall(newliveout)
|
||||
newliveout.Clear()
|
||||
for _, succ := range bb.succ {
|
||||
bvor(newliveout, newliveout, succ.livein)
|
||||
newliveout.Or(newliveout, succ.livein)
|
||||
}
|
||||
|
||||
if !bveq(bb.liveout, newliveout) {
|
||||
if !bb.liveout.Eq(newliveout) {
|
||||
change = true
|
||||
bvcopy(bb.liveout, newliveout)
|
||||
bb.liveout.Copy(newliveout)
|
||||
}
|
||||
|
||||
// A variable is live on input to this block
|
||||
@ -1148,9 +1143,9 @@ func livenesssolve(lv *Liveness) {
|
||||
// not set by the code in this block.
|
||||
//
|
||||
// in[b] = uevar[b] \cup (out[b] \setminus varkill[b])
|
||||
bvandnot(newlivein, bb.liveout, bb.varkill)
|
||||
newlivein.AndNot(bb.liveout, bb.varkill)
|
||||
|
||||
bvor(bb.livein, newlivein, bb.uevar)
|
||||
bb.livein.Or(newlivein, bb.uevar)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1161,14 +1156,14 @@ func islive(n *Node, args bvec, locals bvec) bool {
|
||||
switch n.Class {
|
||||
case PPARAM, PPARAMOUT:
|
||||
for i := 0; int64(i) < n.Type.Width/int64(Widthptr); i++ {
|
||||
if bvget(args, int32(n.Xoffset/int64(Widthptr)+int64(i))) != 0 {
|
||||
if args.Get(int32(n.Xoffset/int64(Widthptr) + int64(i))) {
|
||||
return true
|
||||
}
|
||||
}
|
||||
|
||||
case PAUTO:
|
||||
for i := 0; int64(i) < n.Type.Width/int64(Widthptr); i++ {
|
||||
if bvget(locals, int32((n.Xoffset+stkptrsize)/int64(Widthptr)+int64(i))) != 0 {
|
||||
if locals.Get(int32((n.Xoffset+stkptrsize)/int64(Widthptr) + int64(i))) {
|
||||
return true
|
||||
}
|
||||
}
|
||||
@ -1208,17 +1203,17 @@ func livenessepilogue(lv *Liveness) {
|
||||
// Compute avarinitany and avarinitall for entry to block.
|
||||
// This duplicates information known during livenesssolve
|
||||
// but avoids storing two more vectors for each block.
|
||||
bvresetall(any)
|
||||
any.Clear()
|
||||
|
||||
bvresetall(all)
|
||||
all.Clear()
|
||||
for j := 0; j < len(bb.pred); j++ {
|
||||
pred := bb.pred[j]
|
||||
if j == 0 {
|
||||
bvcopy(any, pred.avarinitany)
|
||||
bvcopy(all, pred.avarinitall)
|
||||
any.Copy(pred.avarinitany)
|
||||
all.Copy(pred.avarinitall)
|
||||
} else {
|
||||
bvor(any, any, pred.avarinitany)
|
||||
bvand(all, all, pred.avarinitall)
|
||||
any.Or(any, pred.avarinitany)
|
||||
all.And(all, pred.avarinitall)
|
||||
}
|
||||
}
|
||||
|
||||
@ -1227,24 +1222,24 @@ func livenessepilogue(lv *Liveness) {
|
||||
// Seed the maps with information about the addrtaken variables.
|
||||
for p := bb.first; ; p = p.Link {
|
||||
progeffects(p, lv.vars, uevar, varkill, avarinit)
|
||||
bvandnot(any, any, varkill)
|
||||
bvandnot(all, all, varkill)
|
||||
bvor(any, any, avarinit)
|
||||
bvor(all, all, avarinit)
|
||||
any.AndNot(any, varkill)
|
||||
all.AndNot(all, varkill)
|
||||
any.Or(any, avarinit)
|
||||
all.Or(all, avarinit)
|
||||
|
||||
if issafepoint(p) {
|
||||
// Annotate ambiguously live variables so that they can
|
||||
// be zeroed at function entry.
|
||||
// livein and liveout are dead here and used as temporaries.
|
||||
bvresetall(livein)
|
||||
livein.Clear()
|
||||
|
||||
bvandnot(liveout, any, all)
|
||||
if !bvisempty(liveout) {
|
||||
liveout.AndNot(any, all)
|
||||
if !liveout.IsEmpty() {
|
||||
for pos := int32(0); pos < liveout.n; pos++ {
|
||||
if bvget(liveout, pos) == 0 {
|
||||
if !liveout.Get(pos) {
|
||||
continue
|
||||
}
|
||||
bvset(all, pos) // silence future warnings in this block
|
||||
all.Set(pos) // silence future warnings in this block
|
||||
n := lv.vars[pos]
|
||||
if !n.Name.Needzero {
|
||||
n.Name.Needzero = true
|
||||
@ -1307,7 +1302,7 @@ func livenessepilogue(lv *Liveness) {
|
||||
Fatalf("livenessepilogue")
|
||||
}
|
||||
|
||||
bvcopy(livein, bb.liveout)
|
||||
livein.Copy(bb.liveout)
|
||||
var next *obj.Prog
|
||||
for p := bb.last; p != nil; p = next {
|
||||
next = p.Opt.(*obj.Prog) // splicebefore modifies p.opt
|
||||
@ -1315,9 +1310,9 @@ func livenessepilogue(lv *Liveness) {
|
||||
// Propagate liveness information
|
||||
progeffects(p, lv.vars, uevar, varkill, avarinit)
|
||||
|
||||
bvcopy(liveout, livein)
|
||||
bvandnot(livein, liveout, varkill)
|
||||
bvor(livein, livein, uevar)
|
||||
liveout.Copy(livein)
|
||||
livein.AndNot(liveout, varkill)
|
||||
livein.Or(livein, uevar)
|
||||
if debuglive >= 3 && issafepoint(p) {
|
||||
fmt.Printf("%v\n", p)
|
||||
printvars("uevar", uevar, lv.vars)
|
||||
@ -1335,7 +1330,7 @@ func livenessepilogue(lv *Liveness) {
|
||||
// input parameters.
|
||||
if p.As == obj.ATEXT {
|
||||
for j := int32(0); j < liveout.n; j++ {
|
||||
if bvget(liveout, j) == 0 {
|
||||
if !liveout.Get(j) {
|
||||
continue
|
||||
}
|
||||
n := lv.vars[j]
|
||||
@ -1353,7 +1348,7 @@ func livenessepilogue(lv *Liveness) {
|
||||
|
||||
// Mark pparamout variables (as described above)
|
||||
if p.As == obj.ACALL {
|
||||
bvor(locals, locals, pparamout)
|
||||
locals.Or(locals, pparamout)
|
||||
}
|
||||
|
||||
// Show live pointer bitmaps.
|
||||
@ -1505,7 +1500,7 @@ func livenesscompact(lv *Liveness) {
|
||||
}
|
||||
jlocal := lv.livepointers[j]
|
||||
jarg := lv.argslivepointers[j]
|
||||
if bveq(local, jlocal) && bveq(arg, jarg) {
|
||||
if local.Eq(jlocal) && arg.Eq(jarg) {
|
||||
remap[i] = j
|
||||
goto Next
|
||||
}
|
||||
@ -1548,7 +1543,7 @@ func livenesscompact(lv *Liveness) {
|
||||
func printbitset(printed bool, name string, vars []*Node, bits bvec) bool {
|
||||
started := false
|
||||
for i, n := range vars {
|
||||
if bvget(bits, int32(i)) == 0 {
|
||||
if !bits.Get(int32(i)) {
|
||||
continue
|
||||
}
|
||||
if !started {
|
||||
|
@ -1639,7 +1639,7 @@ func fillptrmask(t *Type, ptrmask []byte) {
|
||||
|
||||
nptr := typeptrdata(t) / int64(Widthptr)
|
||||
for i := int64(0); i < nptr; i++ {
|
||||
if bvget(vec, int32(i)) == 1 {
|
||||
if vec.Get(int32(i)) {
|
||||
ptrmask[i/8] |= 1 << (uint(i) % 8)
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user