summaryrefslogtreecommitdiffstats
path: root/src/cmd/compile/internal/gc/range.go
diff options
context:
space:
mode:
Diffstat (limited to 'src/cmd/compile/internal/gc/range.go')
-rw-r--r--src/cmd/compile/internal/gc/range.go628
1 files changed, 628 insertions, 0 deletions
diff --git a/src/cmd/compile/internal/gc/range.go b/src/cmd/compile/internal/gc/range.go
new file mode 100644
index 0000000..1b4d765
--- /dev/null
+++ b/src/cmd/compile/internal/gc/range.go
@@ -0,0 +1,628 @@
+// Copyright 2009 The Go Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+package gc
+
+import (
+ "cmd/compile/internal/types"
+ "cmd/internal/sys"
+ "unicode/utf8"
+)
+
+// range
+func typecheckrange(n *Node) {
+ // Typechecking order is important here:
+ // 0. first typecheck range expression (slice/map/chan),
+ // it is evaluated only once and so logically it is not part of the loop.
+ // 1. typecheck produced values,
+ // this part can declare new vars and so it must be typechecked before body,
+ // because body can contain a closure that captures the vars.
+ // 2. decldepth++ to denote loop body.
+ // 3. typecheck body.
+ // 4. decldepth--.
+ typecheckrangeExpr(n)
+
+ // second half of dance, the first half being typecheckrangeExpr
+ n.SetTypecheck(1)
+ ls := n.List.Slice()
+ for i1, n1 := range ls {
+ if n1.Typecheck() == 0 {
+ ls[i1] = typecheck(ls[i1], ctxExpr|ctxAssign)
+ }
+ }
+
+ decldepth++
+ typecheckslice(n.Nbody.Slice(), ctxStmt)
+ decldepth--
+}
+
+func typecheckrangeExpr(n *Node) {
+ n.Right = typecheck(n.Right, ctxExpr)
+
+ t := n.Right.Type
+ if t == nil {
+ return
+ }
+ // delicate little dance. see typecheckas2
+ ls := n.List.Slice()
+ for i1, n1 := range ls {
+ if n1.Name == nil || n1.Name.Defn != n {
+ ls[i1] = typecheck(ls[i1], ctxExpr|ctxAssign)
+ }
+ }
+
+ if t.IsPtr() && t.Elem().IsArray() {
+ t = t.Elem()
+ }
+ n.Type = t
+
+ var t1, t2 *types.Type
+ toomany := false
+ switch t.Etype {
+ default:
+ yyerrorl(n.Pos, "cannot range over %L", n.Right)
+ return
+
+ case TARRAY, TSLICE:
+ t1 = types.Types[TINT]
+ t2 = t.Elem()
+
+ case TMAP:
+ t1 = t.Key()
+ t2 = t.Elem()
+
+ case TCHAN:
+ if !t.ChanDir().CanRecv() {
+ yyerrorl(n.Pos, "invalid operation: range %v (receive from send-only type %v)", n.Right, n.Right.Type)
+ return
+ }
+
+ t1 = t.Elem()
+ t2 = nil
+ if n.List.Len() == 2 {
+ toomany = true
+ }
+
+ case TSTRING:
+ t1 = types.Types[TINT]
+ t2 = types.Runetype
+ }
+
+ if n.List.Len() > 2 || toomany {
+ yyerrorl(n.Pos, "too many variables in range")
+ }
+
+ var v1, v2 *Node
+ if n.List.Len() != 0 {
+ v1 = n.List.First()
+ }
+ if n.List.Len() > 1 {
+ v2 = n.List.Second()
+ }
+
+ // this is not only an optimization but also a requirement in the spec.
+ // "if the second iteration variable is the blank identifier, the range
+ // clause is equivalent to the same clause with only the first variable
+ // present."
+ if v2.isBlank() {
+ if v1 != nil {
+ n.List.Set1(v1)
+ }
+ v2 = nil
+ }
+
+ if v1 != nil {
+ if v1.Name != nil && v1.Name.Defn == n {
+ v1.Type = t1
+ } else if v1.Type != nil {
+ if op, why := assignop(t1, v1.Type); op == OXXX {
+ yyerrorl(n.Pos, "cannot assign type %v to %L in range%s", t1, v1, why)
+ }
+ }
+ checkassign(n, v1)
+ }
+
+ if v2 != nil {
+ if v2.Name != nil && v2.Name.Defn == n {
+ v2.Type = t2
+ } else if v2.Type != nil {
+ if op, why := assignop(t2, v2.Type); op == OXXX {
+ yyerrorl(n.Pos, "cannot assign type %v to %L in range%s", t2, v2, why)
+ }
+ }
+ checkassign(n, v2)
+ }
+}
+
+func cheapComputableIndex(width int64) bool {
+ switch thearch.LinkArch.Family {
+ // MIPS does not have R+R addressing
+ // Arm64 may lack ability to generate this code in our assembler,
+ // but the architecture supports it.
+ case sys.PPC64, sys.S390X:
+ return width == 1
+ case sys.AMD64, sys.I386, sys.ARM64, sys.ARM:
+ switch width {
+ case 1, 2, 4, 8:
+ return true
+ }
+ }
+ return false
+}
+
+// walkrange transforms various forms of ORANGE into
+// simpler forms. The result must be assigned back to n.
+// Node n may also be modified in place, and may also be
+// the returned node.
+func walkrange(n *Node) *Node {
+ if isMapClear(n) {
+ m := n.Right
+ lno := setlineno(m)
+ n = mapClear(m)
+ lineno = lno
+ return n
+ }
+
+ // variable name conventions:
+ // ohv1, hv1, hv2: hidden (old) val 1, 2
+ // ha, hit: hidden aggregate, iterator
+ // hn, hp: hidden len, pointer
+ // hb: hidden bool
+ // a, v1, v2: not hidden aggregate, val 1, 2
+
+ t := n.Type
+
+ a := n.Right
+ lno := setlineno(a)
+ n.Right = nil
+
+ var v1, v2 *Node
+ l := n.List.Len()
+ if l > 0 {
+ v1 = n.List.First()
+ }
+
+ if l > 1 {
+ v2 = n.List.Second()
+ }
+
+ if v2.isBlank() {
+ v2 = nil
+ }
+
+ if v1.isBlank() && v2 == nil {
+ v1 = nil
+ }
+
+ if v1 == nil && v2 != nil {
+ Fatalf("walkrange: v2 != nil while v1 == nil")
+ }
+
+ // n.List has no meaning anymore, clear it
+ // to avoid erroneous processing by racewalk.
+ n.List.Set(nil)
+
+ var ifGuard *Node
+
+ translatedLoopOp := OFOR
+
+ var body []*Node
+ var init []*Node
+ switch t.Etype {
+ default:
+ Fatalf("walkrange")
+
+ case TARRAY, TSLICE:
+ if arrayClear(n, v1, v2, a) {
+ lineno = lno
+ return n
+ }
+
+ // order.stmt arranged for a copy of the array/slice variable if needed.
+ ha := a
+
+ hv1 := temp(types.Types[TINT])
+ hn := temp(types.Types[TINT])
+
+ init = append(init, nod(OAS, hv1, nil))
+ init = append(init, nod(OAS, hn, nod(OLEN, ha, nil)))
+
+ n.Left = nod(OLT, hv1, hn)
+ n.Right = nod(OAS, hv1, nod(OADD, hv1, nodintconst(1)))
+
+ // for range ha { body }
+ if v1 == nil {
+ break
+ }
+
+ // for v1 := range ha { body }
+ if v2 == nil {
+ body = []*Node{nod(OAS, v1, hv1)}
+ break
+ }
+
+ // for v1, v2 := range ha { body }
+ if cheapComputableIndex(n.Type.Elem().Width) {
+ // v1, v2 = hv1, ha[hv1]
+ tmp := nod(OINDEX, ha, hv1)
+ tmp.SetBounded(true)
+ // Use OAS2 to correctly handle assignments
+ // of the form "v1, a[v1] := range".
+ a := nod(OAS2, nil, nil)
+ a.List.Set2(v1, v2)
+ a.Rlist.Set2(hv1, tmp)
+ body = []*Node{a}
+ break
+ }
+
+ // TODO(austin): OFORUNTIL is a strange beast, but is
+ // necessary for expressing the control flow we need
+ // while also making "break" and "continue" work. It
+ // would be nice to just lower ORANGE during SSA, but
+ // racewalk needs to see many of the operations
+ // involved in ORANGE's implementation. If racewalk
+ // moves into SSA, consider moving ORANGE into SSA and
+ // eliminating OFORUNTIL.
+
+ // TODO(austin): OFORUNTIL inhibits bounds-check
+ // elimination on the index variable (see #20711).
+ // Enhance the prove pass to understand this.
+ ifGuard = nod(OIF, nil, nil)
+ ifGuard.Left = nod(OLT, hv1, hn)
+ translatedLoopOp = OFORUNTIL
+
+ hp := temp(types.NewPtr(n.Type.Elem()))
+ tmp := nod(OINDEX, ha, nodintconst(0))
+ tmp.SetBounded(true)
+ init = append(init, nod(OAS, hp, nod(OADDR, tmp, nil)))
+
+ // Use OAS2 to correctly handle assignments
+ // of the form "v1, a[v1] := range".
+ a := nod(OAS2, nil, nil)
+ a.List.Set2(v1, v2)
+ a.Rlist.Set2(hv1, nod(ODEREF, hp, nil))
+ body = append(body, a)
+
+ // Advance pointer as part of the late increment.
+ //
+ // This runs *after* the condition check, so we know
+ // advancing the pointer is safe and won't go past the
+ // end of the allocation.
+ a = nod(OAS, hp, addptr(hp, t.Elem().Width))
+ a = typecheck(a, ctxStmt)
+ n.List.Set1(a)
+
+ case TMAP:
+ // order.stmt allocated the iterator for us.
+ // we only use a once, so no copy needed.
+ ha := a
+
+ hit := prealloc[n]
+ th := hit.Type
+ n.Left = nil
+ keysym := th.Field(0).Sym // depends on layout of iterator struct. See reflect.go:hiter
+ elemsym := th.Field(1).Sym // ditto
+
+ fn := syslook("mapiterinit")
+
+ fn = substArgTypes(fn, t.Key(), t.Elem(), th)
+ init = append(init, mkcall1(fn, nil, nil, typename(t), ha, nod(OADDR, hit, nil)))
+ n.Left = nod(ONE, nodSym(ODOT, hit, keysym), nodnil())
+
+ fn = syslook("mapiternext")
+ fn = substArgTypes(fn, th)
+ n.Right = mkcall1(fn, nil, nil, nod(OADDR, hit, nil))
+
+ key := nodSym(ODOT, hit, keysym)
+ key = nod(ODEREF, key, nil)
+ if v1 == nil {
+ body = nil
+ } else if v2 == nil {
+ body = []*Node{nod(OAS, v1, key)}
+ } else {
+ elem := nodSym(ODOT, hit, elemsym)
+ elem = nod(ODEREF, elem, nil)
+ a := nod(OAS2, nil, nil)
+ a.List.Set2(v1, v2)
+ a.Rlist.Set2(key, elem)
+ body = []*Node{a}
+ }
+
+ case TCHAN:
+ // order.stmt arranged for a copy of the channel variable.
+ ha := a
+
+ n.Left = nil
+
+ hv1 := temp(t.Elem())
+ hv1.SetTypecheck(1)
+ if t.Elem().HasPointers() {
+ init = append(init, nod(OAS, hv1, nil))
+ }
+ hb := temp(types.Types[TBOOL])
+
+ n.Left = nod(ONE, hb, nodbool(false))
+ a := nod(OAS2RECV, nil, nil)
+ a.SetTypecheck(1)
+ a.List.Set2(hv1, hb)
+ a.Right = nod(ORECV, ha, nil)
+ n.Left.Ninit.Set1(a)
+ if v1 == nil {
+ body = nil
+ } else {
+ body = []*Node{nod(OAS, v1, hv1)}
+ }
+ // Zero hv1. This prevents hv1 from being the sole, inaccessible
+ // reference to an otherwise GC-able value during the next channel receive.
+ // See issue 15281.
+ body = append(body, nod(OAS, hv1, nil))
+
+ case TSTRING:
+ // Transform string range statements like "for v1, v2 = range a" into
+ //
+ // ha := a
+ // for hv1 := 0; hv1 < len(ha); {
+ // hv1t := hv1
+ // hv2 := rune(ha[hv1])
+ // if hv2 < utf8.RuneSelf {
+ // hv1++
+ // } else {
+ // hv2, hv1 = decoderune(ha, hv1)
+ // }
+ // v1, v2 = hv1t, hv2
+ // // original body
+ // }
+
+ // order.stmt arranged for a copy of the string variable.
+ ha := a
+
+ hv1 := temp(types.Types[TINT])
+ hv1t := temp(types.Types[TINT])
+ hv2 := temp(types.Runetype)
+
+ // hv1 := 0
+ init = append(init, nod(OAS, hv1, nil))
+
+ // hv1 < len(ha)
+ n.Left = nod(OLT, hv1, nod(OLEN, ha, nil))
+
+ if v1 != nil {
+ // hv1t = hv1
+ body = append(body, nod(OAS, hv1t, hv1))
+ }
+
+ // hv2 := rune(ha[hv1])
+ nind := nod(OINDEX, ha, hv1)
+ nind.SetBounded(true)
+ body = append(body, nod(OAS, hv2, conv(nind, types.Runetype)))
+
+ // if hv2 < utf8.RuneSelf
+ nif := nod(OIF, nil, nil)
+ nif.Left = nod(OLT, hv2, nodintconst(utf8.RuneSelf))
+
+ // hv1++
+ nif.Nbody.Set1(nod(OAS, hv1, nod(OADD, hv1, nodintconst(1))))
+
+ // } else {
+ eif := nod(OAS2, nil, nil)
+ nif.Rlist.Set1(eif)
+
+ // hv2, hv1 = decoderune(ha, hv1)
+ eif.List.Set2(hv2, hv1)
+ fn := syslook("decoderune")
+ eif.Rlist.Set1(mkcall1(fn, fn.Type.Results(), nil, ha, hv1))
+
+ body = append(body, nif)
+
+ if v1 != nil {
+ if v2 != nil {
+ // v1, v2 = hv1t, hv2
+ a := nod(OAS2, nil, nil)
+ a.List.Set2(v1, v2)
+ a.Rlist.Set2(hv1t, hv2)
+ body = append(body, a)
+ } else {
+ // v1 = hv1t
+ body = append(body, nod(OAS, v1, hv1t))
+ }
+ }
+ }
+
+ n.Op = translatedLoopOp
+ typecheckslice(init, ctxStmt)
+
+ if ifGuard != nil {
+ ifGuard.Ninit.Append(init...)
+ ifGuard = typecheck(ifGuard, ctxStmt)
+ } else {
+ n.Ninit.Append(init...)
+ }
+
+ typecheckslice(n.Left.Ninit.Slice(), ctxStmt)
+
+ n.Left = typecheck(n.Left, ctxExpr)
+ n.Left = defaultlit(n.Left, nil)
+ n.Right = typecheck(n.Right, ctxStmt)
+ typecheckslice(body, ctxStmt)
+ n.Nbody.Prepend(body...)
+
+ if ifGuard != nil {
+ ifGuard.Nbody.Set1(n)
+ n = ifGuard
+ }
+
+ n = walkstmt(n)
+
+ lineno = lno
+ return n
+}
+
+// isMapClear checks if n is of the form:
+//
+// for k := range m {
+// delete(m, k)
+// }
+//
+// where == for keys of map m is reflexive.
+func isMapClear(n *Node) bool {
+ if Debug.N != 0 || instrumenting {
+ return false
+ }
+
+ if n.Op != ORANGE || n.Type.Etype != TMAP || n.List.Len() != 1 {
+ return false
+ }
+
+ k := n.List.First()
+ if k == nil || k.isBlank() {
+ return false
+ }
+
+ // Require k to be a new variable name.
+ if k.Name == nil || k.Name.Defn != n {
+ return false
+ }
+
+ if n.Nbody.Len() != 1 {
+ return false
+ }
+
+ stmt := n.Nbody.First() // only stmt in body
+ if stmt == nil || stmt.Op != ODELETE {
+ return false
+ }
+
+ m := n.Right
+ if !samesafeexpr(stmt.List.First(), m) || !samesafeexpr(stmt.List.Second(), k) {
+ return false
+ }
+
+ // Keys where equality is not reflexive can not be deleted from maps.
+ if !isreflexive(m.Type.Key()) {
+ return false
+ }
+
+ return true
+}
+
+// mapClear constructs a call to runtime.mapclear for the map m.
+func mapClear(m *Node) *Node {
+ t := m.Type
+
+ // instantiate mapclear(typ *type, hmap map[any]any)
+ fn := syslook("mapclear")
+ fn = substArgTypes(fn, t.Key(), t.Elem())
+ n := mkcall1(fn, nil, nil, typename(t), m)
+
+ n = typecheck(n, ctxStmt)
+ n = walkstmt(n)
+
+ return n
+}
+
+// Lower n into runtime·memclr if possible, for
+// fast zeroing of slices and arrays (issue 5373).
+// Look for instances of
+//
+// for i := range a {
+// a[i] = zero
+// }
+//
+// in which the evaluation of a is side-effect-free.
+//
+// Parameters are as in walkrange: "for v1, v2 = range a".
+func arrayClear(n, v1, v2, a *Node) bool {
+ if Debug.N != 0 || instrumenting {
+ return false
+ }
+
+ if v1 == nil || v2 != nil {
+ return false
+ }
+
+ if n.Nbody.Len() != 1 || n.Nbody.First() == nil {
+ return false
+ }
+
+ stmt := n.Nbody.First() // only stmt in body
+ if stmt.Op != OAS || stmt.Left.Op != OINDEX {
+ return false
+ }
+
+ if !samesafeexpr(stmt.Left.Left, a) || !samesafeexpr(stmt.Left.Right, v1) {
+ return false
+ }
+
+ elemsize := n.Type.Elem().Width
+ if elemsize <= 0 || !isZero(stmt.Right) {
+ return false
+ }
+
+ // Convert to
+ // if len(a) != 0 {
+ // hp = &a[0]
+ // hn = len(a)*sizeof(elem(a))
+ // memclr{NoHeap,Has}Pointers(hp, hn)
+ // i = len(a) - 1
+ // }
+ n.Op = OIF
+
+ n.Nbody.Set(nil)
+ n.Left = nod(ONE, nod(OLEN, a, nil), nodintconst(0))
+
+ // hp = &a[0]
+ hp := temp(types.Types[TUNSAFEPTR])
+
+ tmp := nod(OINDEX, a, nodintconst(0))
+ tmp.SetBounded(true)
+ tmp = nod(OADDR, tmp, nil)
+ tmp = convnop(tmp, types.Types[TUNSAFEPTR])
+ n.Nbody.Append(nod(OAS, hp, tmp))
+
+ // hn = len(a) * sizeof(elem(a))
+ hn := temp(types.Types[TUINTPTR])
+
+ tmp = nod(OLEN, a, nil)
+ tmp = nod(OMUL, tmp, nodintconst(elemsize))
+ tmp = conv(tmp, types.Types[TUINTPTR])
+ n.Nbody.Append(nod(OAS, hn, tmp))
+
+ var fn *Node
+ if a.Type.Elem().HasPointers() {
+ // memclrHasPointers(hp, hn)
+ Curfn.Func.setWBPos(stmt.Pos)
+ fn = mkcall("memclrHasPointers", nil, nil, hp, hn)
+ } else {
+ // memclrNoHeapPointers(hp, hn)
+ fn = mkcall("memclrNoHeapPointers", nil, nil, hp, hn)
+ }
+
+ n.Nbody.Append(fn)
+
+ // i = len(a) - 1
+ v1 = nod(OAS, v1, nod(OSUB, nod(OLEN, a, nil), nodintconst(1)))
+
+ n.Nbody.Append(v1)
+
+ n.Left = typecheck(n.Left, ctxExpr)
+ n.Left = defaultlit(n.Left, nil)
+ typecheckslice(n.Nbody.Slice(), ctxStmt)
+ n = walkstmt(n)
+ return true
+}
+
+// addptr returns (*T)(uintptr(p) + n).
+func addptr(p *Node, n int64) *Node {
+ t := p.Type
+
+ p = nod(OCONVNOP, p, nil)
+ p.Type = types.Types[TUINTPTR]
+
+ p = nod(OADD, p, nodintconst(n))
+
+ p = nod(OCONVNOP, p, nil)
+ p.Type = t
+
+ return p
+}