diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-28 13:14:23 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-28 13:14:23 +0000 |
commit | 73df946d56c74384511a194dd01dbe099584fd1a (patch) | |
tree | fd0bcea490dd81327ddfbb31e215439672c9a068 /src/cmd/compile/internal/ssa/rewriteWasm.go | |
parent | Initial commit. (diff) | |
download | golang-1.16-upstream.tar.xz golang-1.16-upstream.zip |
Adding upstream version 1.16.10.upstream/1.16.10upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/cmd/compile/internal/ssa/rewriteWasm.go')
-rw-r--r-- | src/cmd/compile/internal/ssa/rewriteWasm.go | 4905 |
1 files changed, 4905 insertions, 0 deletions
diff --git a/src/cmd/compile/internal/ssa/rewriteWasm.go b/src/cmd/compile/internal/ssa/rewriteWasm.go new file mode 100644 index 0000000..c8ecefc --- /dev/null +++ b/src/cmd/compile/internal/ssa/rewriteWasm.go @@ -0,0 +1,4905 @@ +// Code generated from gen/Wasm.rules; DO NOT EDIT. +// generated with: cd gen; go run *.go + +package ssa + +import "math" +import "cmd/internal/objabi" +import "cmd/compile/internal/types" + +func rewriteValueWasm(v *Value) bool { + switch v.Op { + case OpAbs: + v.Op = OpWasmF64Abs + return true + case OpAdd16: + v.Op = OpWasmI64Add + return true + case OpAdd32: + v.Op = OpWasmI64Add + return true + case OpAdd32F: + v.Op = OpWasmF32Add + return true + case OpAdd64: + v.Op = OpWasmI64Add + return true + case OpAdd64F: + v.Op = OpWasmF64Add + return true + case OpAdd8: + v.Op = OpWasmI64Add + return true + case OpAddPtr: + v.Op = OpWasmI64Add + return true + case OpAddr: + return rewriteValueWasm_OpAddr(v) + case OpAnd16: + v.Op = OpWasmI64And + return true + case OpAnd32: + v.Op = OpWasmI64And + return true + case OpAnd64: + v.Op = OpWasmI64And + return true + case OpAnd8: + v.Op = OpWasmI64And + return true + case OpAndB: + v.Op = OpWasmI64And + return true + case OpBitLen64: + return rewriteValueWasm_OpBitLen64(v) + case OpCeil: + v.Op = OpWasmF64Ceil + return true + case OpClosureCall: + v.Op = OpWasmLoweredClosureCall + return true + case OpCom16: + return rewriteValueWasm_OpCom16(v) + case OpCom32: + return rewriteValueWasm_OpCom32(v) + case OpCom64: + return rewriteValueWasm_OpCom64(v) + case OpCom8: + return rewriteValueWasm_OpCom8(v) + case OpCondSelect: + v.Op = OpWasmSelect + return true + case OpConst16: + return rewriteValueWasm_OpConst16(v) + case OpConst32: + return rewriteValueWasm_OpConst32(v) + case OpConst32F: + v.Op = OpWasmF32Const + return true + case OpConst64: + v.Op = OpWasmI64Const + return true + case OpConst64F: + v.Op = OpWasmF64Const + return true + case OpConst8: + return rewriteValueWasm_OpConst8(v) + case OpConstBool: + return rewriteValueWasm_OpConstBool(v) + case OpConstNil: + return rewriteValueWasm_OpConstNil(v) + case OpConvert: + v.Op = OpWasmLoweredConvert + return true + case OpCopysign: + v.Op = OpWasmF64Copysign + return true + case OpCtz16: + return rewriteValueWasm_OpCtz16(v) + case OpCtz16NonZero: + v.Op = OpWasmI64Ctz + return true + case OpCtz32: + return rewriteValueWasm_OpCtz32(v) + case OpCtz32NonZero: + v.Op = OpWasmI64Ctz + return true + case OpCtz64: + v.Op = OpWasmI64Ctz + return true + case OpCtz64NonZero: + v.Op = OpWasmI64Ctz + return true + case OpCtz8: + return rewriteValueWasm_OpCtz8(v) + case OpCtz8NonZero: + v.Op = OpWasmI64Ctz + return true + case OpCvt32Fto32: + v.Op = OpWasmI64TruncSatF32S + return true + case OpCvt32Fto32U: + v.Op = OpWasmI64TruncSatF32U + return true + case OpCvt32Fto64: + v.Op = OpWasmI64TruncSatF32S + return true + case OpCvt32Fto64F: + v.Op = OpWasmF64PromoteF32 + return true + case OpCvt32Fto64U: + v.Op = OpWasmI64TruncSatF32U + return true + case OpCvt32Uto32F: + return rewriteValueWasm_OpCvt32Uto32F(v) + case OpCvt32Uto64F: + return rewriteValueWasm_OpCvt32Uto64F(v) + case OpCvt32to32F: + return rewriteValueWasm_OpCvt32to32F(v) + case OpCvt32to64F: + return rewriteValueWasm_OpCvt32to64F(v) + case OpCvt64Fto32: + v.Op = OpWasmI64TruncSatF64S + return true + case OpCvt64Fto32F: + v.Op = OpWasmF32DemoteF64 + return true + case OpCvt64Fto32U: + v.Op = OpWasmI64TruncSatF64U + return true + case OpCvt64Fto64: + v.Op = OpWasmI64TruncSatF64S + return true + case OpCvt64Fto64U: + v.Op = OpWasmI64TruncSatF64U + return true + case OpCvt64Uto32F: + v.Op = OpWasmF32ConvertI64U + return true + case OpCvt64Uto64F: + v.Op = OpWasmF64ConvertI64U + return true + case OpCvt64to32F: + v.Op = OpWasmF32ConvertI64S + return true + case OpCvt64to64F: + v.Op = OpWasmF64ConvertI64S + return true + case OpCvtBoolToUint8: + v.Op = OpCopy + return true + case OpDiv16: + return rewriteValueWasm_OpDiv16(v) + case OpDiv16u: + return rewriteValueWasm_OpDiv16u(v) + case OpDiv32: + return rewriteValueWasm_OpDiv32(v) + case OpDiv32F: + v.Op = OpWasmF32Div + return true + case OpDiv32u: + return rewriteValueWasm_OpDiv32u(v) + case OpDiv64: + return rewriteValueWasm_OpDiv64(v) + case OpDiv64F: + v.Op = OpWasmF64Div + return true + case OpDiv64u: + v.Op = OpWasmI64DivU + return true + case OpDiv8: + return rewriteValueWasm_OpDiv8(v) + case OpDiv8u: + return rewriteValueWasm_OpDiv8u(v) + case OpEq16: + return rewriteValueWasm_OpEq16(v) + case OpEq32: + return rewriteValueWasm_OpEq32(v) + case OpEq32F: + v.Op = OpWasmF32Eq + return true + case OpEq64: + v.Op = OpWasmI64Eq + return true + case OpEq64F: + v.Op = OpWasmF64Eq + return true + case OpEq8: + return rewriteValueWasm_OpEq8(v) + case OpEqB: + v.Op = OpWasmI64Eq + return true + case OpEqPtr: + v.Op = OpWasmI64Eq + return true + case OpFloor: + v.Op = OpWasmF64Floor + return true + case OpGetCallerPC: + v.Op = OpWasmLoweredGetCallerPC + return true + case OpGetCallerSP: + v.Op = OpWasmLoweredGetCallerSP + return true + case OpGetClosurePtr: + v.Op = OpWasmLoweredGetClosurePtr + return true + case OpInterCall: + v.Op = OpWasmLoweredInterCall + return true + case OpIsInBounds: + v.Op = OpWasmI64LtU + return true + case OpIsNonNil: + return rewriteValueWasm_OpIsNonNil(v) + case OpIsSliceInBounds: + v.Op = OpWasmI64LeU + return true + case OpLeq16: + return rewriteValueWasm_OpLeq16(v) + case OpLeq16U: + return rewriteValueWasm_OpLeq16U(v) + case OpLeq32: + return rewriteValueWasm_OpLeq32(v) + case OpLeq32F: + v.Op = OpWasmF32Le + return true + case OpLeq32U: + return rewriteValueWasm_OpLeq32U(v) + case OpLeq64: + v.Op = OpWasmI64LeS + return true + case OpLeq64F: + v.Op = OpWasmF64Le + return true + case OpLeq64U: + v.Op = OpWasmI64LeU + return true + case OpLeq8: + return rewriteValueWasm_OpLeq8(v) + case OpLeq8U: + return rewriteValueWasm_OpLeq8U(v) + case OpLess16: + return rewriteValueWasm_OpLess16(v) + case OpLess16U: + return rewriteValueWasm_OpLess16U(v) + case OpLess32: + return rewriteValueWasm_OpLess32(v) + case OpLess32F: + v.Op = OpWasmF32Lt + return true + case OpLess32U: + return rewriteValueWasm_OpLess32U(v) + case OpLess64: + v.Op = OpWasmI64LtS + return true + case OpLess64F: + v.Op = OpWasmF64Lt + return true + case OpLess64U: + v.Op = OpWasmI64LtU + return true + case OpLess8: + return rewriteValueWasm_OpLess8(v) + case OpLess8U: + return rewriteValueWasm_OpLess8U(v) + case OpLoad: + return rewriteValueWasm_OpLoad(v) + case OpLocalAddr: + return rewriteValueWasm_OpLocalAddr(v) + case OpLsh16x16: + return rewriteValueWasm_OpLsh16x16(v) + case OpLsh16x32: + return rewriteValueWasm_OpLsh16x32(v) + case OpLsh16x64: + v.Op = OpLsh64x64 + return true + case OpLsh16x8: + return rewriteValueWasm_OpLsh16x8(v) + case OpLsh32x16: + return rewriteValueWasm_OpLsh32x16(v) + case OpLsh32x32: + return rewriteValueWasm_OpLsh32x32(v) + case OpLsh32x64: + v.Op = OpLsh64x64 + return true + case OpLsh32x8: + return rewriteValueWasm_OpLsh32x8(v) + case OpLsh64x16: + return rewriteValueWasm_OpLsh64x16(v) + case OpLsh64x32: + return rewriteValueWasm_OpLsh64x32(v) + case OpLsh64x64: + return rewriteValueWasm_OpLsh64x64(v) + case OpLsh64x8: + return rewriteValueWasm_OpLsh64x8(v) + case OpLsh8x16: + return rewriteValueWasm_OpLsh8x16(v) + case OpLsh8x32: + return rewriteValueWasm_OpLsh8x32(v) + case OpLsh8x64: + v.Op = OpLsh64x64 + return true + case OpLsh8x8: + return rewriteValueWasm_OpLsh8x8(v) + case OpMod16: + return rewriteValueWasm_OpMod16(v) + case OpMod16u: + return rewriteValueWasm_OpMod16u(v) + case OpMod32: + return rewriteValueWasm_OpMod32(v) + case OpMod32u: + return rewriteValueWasm_OpMod32u(v) + case OpMod64: + return rewriteValueWasm_OpMod64(v) + case OpMod64u: + v.Op = OpWasmI64RemU + return true + case OpMod8: + return rewriteValueWasm_OpMod8(v) + case OpMod8u: + return rewriteValueWasm_OpMod8u(v) + case OpMove: + return rewriteValueWasm_OpMove(v) + case OpMul16: + v.Op = OpWasmI64Mul + return true + case OpMul32: + v.Op = OpWasmI64Mul + return true + case OpMul32F: + v.Op = OpWasmF32Mul + return true + case OpMul64: + v.Op = OpWasmI64Mul + return true + case OpMul64F: + v.Op = OpWasmF64Mul + return true + case OpMul8: + v.Op = OpWasmI64Mul + return true + case OpNeg16: + return rewriteValueWasm_OpNeg16(v) + case OpNeg32: + return rewriteValueWasm_OpNeg32(v) + case OpNeg32F: + v.Op = OpWasmF32Neg + return true + case OpNeg64: + return rewriteValueWasm_OpNeg64(v) + case OpNeg64F: + v.Op = OpWasmF64Neg + return true + case OpNeg8: + return rewriteValueWasm_OpNeg8(v) + case OpNeq16: + return rewriteValueWasm_OpNeq16(v) + case OpNeq32: + return rewriteValueWasm_OpNeq32(v) + case OpNeq32F: + v.Op = OpWasmF32Ne + return true + case OpNeq64: + v.Op = OpWasmI64Ne + return true + case OpNeq64F: + v.Op = OpWasmF64Ne + return true + case OpNeq8: + return rewriteValueWasm_OpNeq8(v) + case OpNeqB: + v.Op = OpWasmI64Ne + return true + case OpNeqPtr: + v.Op = OpWasmI64Ne + return true + case OpNilCheck: + v.Op = OpWasmLoweredNilCheck + return true + case OpNot: + v.Op = OpWasmI64Eqz + return true + case OpOffPtr: + v.Op = OpWasmI64AddConst + return true + case OpOr16: + v.Op = OpWasmI64Or + return true + case OpOr32: + v.Op = OpWasmI64Or + return true + case OpOr64: + v.Op = OpWasmI64Or + return true + case OpOr8: + v.Op = OpWasmI64Or + return true + case OpOrB: + v.Op = OpWasmI64Or + return true + case OpPopCount16: + return rewriteValueWasm_OpPopCount16(v) + case OpPopCount32: + return rewriteValueWasm_OpPopCount32(v) + case OpPopCount64: + v.Op = OpWasmI64Popcnt + return true + case OpPopCount8: + return rewriteValueWasm_OpPopCount8(v) + case OpRotateLeft16: + return rewriteValueWasm_OpRotateLeft16(v) + case OpRotateLeft32: + v.Op = OpWasmI32Rotl + return true + case OpRotateLeft64: + v.Op = OpWasmI64Rotl + return true + case OpRotateLeft8: + return rewriteValueWasm_OpRotateLeft8(v) + case OpRound32F: + v.Op = OpCopy + return true + case OpRound64F: + v.Op = OpCopy + return true + case OpRoundToEven: + v.Op = OpWasmF64Nearest + return true + case OpRsh16Ux16: + return rewriteValueWasm_OpRsh16Ux16(v) + case OpRsh16Ux32: + return rewriteValueWasm_OpRsh16Ux32(v) + case OpRsh16Ux64: + return rewriteValueWasm_OpRsh16Ux64(v) + case OpRsh16Ux8: + return rewriteValueWasm_OpRsh16Ux8(v) + case OpRsh16x16: + return rewriteValueWasm_OpRsh16x16(v) + case OpRsh16x32: + return rewriteValueWasm_OpRsh16x32(v) + case OpRsh16x64: + return rewriteValueWasm_OpRsh16x64(v) + case OpRsh16x8: + return rewriteValueWasm_OpRsh16x8(v) + case OpRsh32Ux16: + return rewriteValueWasm_OpRsh32Ux16(v) + case OpRsh32Ux32: + return rewriteValueWasm_OpRsh32Ux32(v) + case OpRsh32Ux64: + return rewriteValueWasm_OpRsh32Ux64(v) + case OpRsh32Ux8: + return rewriteValueWasm_OpRsh32Ux8(v) + case OpRsh32x16: + return rewriteValueWasm_OpRsh32x16(v) + case OpRsh32x32: + return rewriteValueWasm_OpRsh32x32(v) + case OpRsh32x64: + return rewriteValueWasm_OpRsh32x64(v) + case OpRsh32x8: + return rewriteValueWasm_OpRsh32x8(v) + case OpRsh64Ux16: + return rewriteValueWasm_OpRsh64Ux16(v) + case OpRsh64Ux32: + return rewriteValueWasm_OpRsh64Ux32(v) + case OpRsh64Ux64: + return rewriteValueWasm_OpRsh64Ux64(v) + case OpRsh64Ux8: + return rewriteValueWasm_OpRsh64Ux8(v) + case OpRsh64x16: + return rewriteValueWasm_OpRsh64x16(v) + case OpRsh64x32: + return rewriteValueWasm_OpRsh64x32(v) + case OpRsh64x64: + return rewriteValueWasm_OpRsh64x64(v) + case OpRsh64x8: + return rewriteValueWasm_OpRsh64x8(v) + case OpRsh8Ux16: + return rewriteValueWasm_OpRsh8Ux16(v) + case OpRsh8Ux32: + return rewriteValueWasm_OpRsh8Ux32(v) + case OpRsh8Ux64: + return rewriteValueWasm_OpRsh8Ux64(v) + case OpRsh8Ux8: + return rewriteValueWasm_OpRsh8Ux8(v) + case OpRsh8x16: + return rewriteValueWasm_OpRsh8x16(v) + case OpRsh8x32: + return rewriteValueWasm_OpRsh8x32(v) + case OpRsh8x64: + return rewriteValueWasm_OpRsh8x64(v) + case OpRsh8x8: + return rewriteValueWasm_OpRsh8x8(v) + case OpSignExt16to32: + return rewriteValueWasm_OpSignExt16to32(v) + case OpSignExt16to64: + return rewriteValueWasm_OpSignExt16to64(v) + case OpSignExt32to64: + return rewriteValueWasm_OpSignExt32to64(v) + case OpSignExt8to16: + return rewriteValueWasm_OpSignExt8to16(v) + case OpSignExt8to32: + return rewriteValueWasm_OpSignExt8to32(v) + case OpSignExt8to64: + return rewriteValueWasm_OpSignExt8to64(v) + case OpSlicemask: + return rewriteValueWasm_OpSlicemask(v) + case OpSqrt: + v.Op = OpWasmF64Sqrt + return true + case OpStaticCall: + v.Op = OpWasmLoweredStaticCall + return true + case OpStore: + return rewriteValueWasm_OpStore(v) + case OpSub16: + v.Op = OpWasmI64Sub + return true + case OpSub32: + v.Op = OpWasmI64Sub + return true + case OpSub32F: + v.Op = OpWasmF32Sub + return true + case OpSub64: + v.Op = OpWasmI64Sub + return true + case OpSub64F: + v.Op = OpWasmF64Sub + return true + case OpSub8: + v.Op = OpWasmI64Sub + return true + case OpSubPtr: + v.Op = OpWasmI64Sub + return true + case OpTrunc: + v.Op = OpWasmF64Trunc + return true + case OpTrunc16to8: + v.Op = OpCopy + return true + case OpTrunc32to16: + v.Op = OpCopy + return true + case OpTrunc32to8: + v.Op = OpCopy + return true + case OpTrunc64to16: + v.Op = OpCopy + return true + case OpTrunc64to32: + v.Op = OpCopy + return true + case OpTrunc64to8: + v.Op = OpCopy + return true + case OpWB: + v.Op = OpWasmLoweredWB + return true + case OpWasmF64Add: + return rewriteValueWasm_OpWasmF64Add(v) + case OpWasmF64Mul: + return rewriteValueWasm_OpWasmF64Mul(v) + case OpWasmI64Add: + return rewriteValueWasm_OpWasmI64Add(v) + case OpWasmI64AddConst: + return rewriteValueWasm_OpWasmI64AddConst(v) + case OpWasmI64And: + return rewriteValueWasm_OpWasmI64And(v) + case OpWasmI64Eq: + return rewriteValueWasm_OpWasmI64Eq(v) + case OpWasmI64Eqz: + return rewriteValueWasm_OpWasmI64Eqz(v) + case OpWasmI64LeU: + return rewriteValueWasm_OpWasmI64LeU(v) + case OpWasmI64Load: + return rewriteValueWasm_OpWasmI64Load(v) + case OpWasmI64Load16S: + return rewriteValueWasm_OpWasmI64Load16S(v) + case OpWasmI64Load16U: + return rewriteValueWasm_OpWasmI64Load16U(v) + case OpWasmI64Load32S: + return rewriteValueWasm_OpWasmI64Load32S(v) + case OpWasmI64Load32U: + return rewriteValueWasm_OpWasmI64Load32U(v) + case OpWasmI64Load8S: + return rewriteValueWasm_OpWasmI64Load8S(v) + case OpWasmI64Load8U: + return rewriteValueWasm_OpWasmI64Load8U(v) + case OpWasmI64LtU: + return rewriteValueWasm_OpWasmI64LtU(v) + case OpWasmI64Mul: + return rewriteValueWasm_OpWasmI64Mul(v) + case OpWasmI64Ne: + return rewriteValueWasm_OpWasmI64Ne(v) + case OpWasmI64Or: + return rewriteValueWasm_OpWasmI64Or(v) + case OpWasmI64Shl: + return rewriteValueWasm_OpWasmI64Shl(v) + case OpWasmI64ShrS: + return rewriteValueWasm_OpWasmI64ShrS(v) + case OpWasmI64ShrU: + return rewriteValueWasm_OpWasmI64ShrU(v) + case OpWasmI64Store: + return rewriteValueWasm_OpWasmI64Store(v) + case OpWasmI64Store16: + return rewriteValueWasm_OpWasmI64Store16(v) + case OpWasmI64Store32: + return rewriteValueWasm_OpWasmI64Store32(v) + case OpWasmI64Store8: + return rewriteValueWasm_OpWasmI64Store8(v) + case OpWasmI64Xor: + return rewriteValueWasm_OpWasmI64Xor(v) + case OpXor16: + v.Op = OpWasmI64Xor + return true + case OpXor32: + v.Op = OpWasmI64Xor + return true + case OpXor64: + v.Op = OpWasmI64Xor + return true + case OpXor8: + v.Op = OpWasmI64Xor + return true + case OpZero: + return rewriteValueWasm_OpZero(v) + case OpZeroExt16to32: + return rewriteValueWasm_OpZeroExt16to32(v) + case OpZeroExt16to64: + return rewriteValueWasm_OpZeroExt16to64(v) + case OpZeroExt32to64: + return rewriteValueWasm_OpZeroExt32to64(v) + case OpZeroExt8to16: + return rewriteValueWasm_OpZeroExt8to16(v) + case OpZeroExt8to32: + return rewriteValueWasm_OpZeroExt8to32(v) + case OpZeroExt8to64: + return rewriteValueWasm_OpZeroExt8to64(v) + } + return false +} +func rewriteValueWasm_OpAddr(v *Value) bool { + v_0 := v.Args[0] + // match: (Addr {sym} base) + // result: (LoweredAddr {sym} [0] base) + for { + sym := auxToSym(v.Aux) + base := v_0 + v.reset(OpWasmLoweredAddr) + v.AuxInt = int32ToAuxInt(0) + v.Aux = symToAux(sym) + v.AddArg(base) + return true + } +} +func rewriteValueWasm_OpBitLen64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (BitLen64 x) + // result: (I64Sub (I64Const [64]) (I64Clz x)) + for { + x := v_0 + v.reset(OpWasmI64Sub) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(64) + v1 := b.NewValue0(v.Pos, OpWasmI64Clz, typ.Int64) + v1.AddArg(x) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpCom16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Com16 x) + // result: (I64Xor x (I64Const [-1])) + for { + x := v_0 + v.reset(OpWasmI64Xor) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(-1) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpCom32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Com32 x) + // result: (I64Xor x (I64Const [-1])) + for { + x := v_0 + v.reset(OpWasmI64Xor) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(-1) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpCom64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Com64 x) + // result: (I64Xor x (I64Const [-1])) + for { + x := v_0 + v.reset(OpWasmI64Xor) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(-1) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpCom8(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Com8 x) + // result: (I64Xor x (I64Const [-1])) + for { + x := v_0 + v.reset(OpWasmI64Xor) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(-1) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpConst16(v *Value) bool { + // match: (Const16 [c]) + // result: (I64Const [int64(c)]) + for { + c := auxIntToInt16(v.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(c)) + return true + } +} +func rewriteValueWasm_OpConst32(v *Value) bool { + // match: (Const32 [c]) + // result: (I64Const [int64(c)]) + for { + c := auxIntToInt32(v.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(c)) + return true + } +} +func rewriteValueWasm_OpConst8(v *Value) bool { + // match: (Const8 [c]) + // result: (I64Const [int64(c)]) + for { + c := auxIntToInt8(v.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(c)) + return true + } +} +func rewriteValueWasm_OpConstBool(v *Value) bool { + // match: (ConstBool [c]) + // result: (I64Const [b2i(c)]) + for { + c := auxIntToBool(v.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(b2i(c)) + return true + } +} +func rewriteValueWasm_OpConstNil(v *Value) bool { + // match: (ConstNil) + // result: (I64Const [0]) + for { + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(0) + return true + } +} +func rewriteValueWasm_OpCtz16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Ctz16 x) + // result: (I64Ctz (I64Or x (I64Const [0x10000]))) + for { + x := v_0 + v.reset(OpWasmI64Ctz) + v0 := b.NewValue0(v.Pos, OpWasmI64Or, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0x10000) + v0.AddArg2(x, v1) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCtz32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Ctz32 x) + // result: (I64Ctz (I64Or x (I64Const [0x100000000]))) + for { + x := v_0 + v.reset(OpWasmI64Ctz) + v0 := b.NewValue0(v.Pos, OpWasmI64Or, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0x100000000) + v0.AddArg2(x, v1) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCtz8(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Ctz8 x) + // result: (I64Ctz (I64Or x (I64Const [0x100]))) + for { + x := v_0 + v.reset(OpWasmI64Ctz) + v0 := b.NewValue0(v.Pos, OpWasmI64Or, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0x100) + v0.AddArg2(x, v1) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCvt32Uto32F(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Cvt32Uto32F x) + // result: (F32ConvertI64U (ZeroExt32to64 x)) + for { + x := v_0 + v.reset(OpWasmF32ConvertI64U) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCvt32Uto64F(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Cvt32Uto64F x) + // result: (F64ConvertI64U (ZeroExt32to64 x)) + for { + x := v_0 + v.reset(OpWasmF64ConvertI64U) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCvt32to32F(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Cvt32to32F x) + // result: (F32ConvertI64S (SignExt32to64 x)) + for { + x := v_0 + v.reset(OpWasmF32ConvertI64S) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpCvt32to64F(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Cvt32to64F x) + // result: (F64ConvertI64S (SignExt32to64 x)) + for { + x := v_0 + v.reset(OpWasmF64ConvertI64S) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpDiv16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div16 [false] x y) + // result: (I64DivS (SignExt16to64 x) (SignExt16to64 y)) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64DivS) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } + return false +} +func rewriteValueWasm_OpDiv16u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div16u x y) + // result: (I64DivU (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64DivU) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpDiv32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div32 [false] x y) + // result: (I64DivS (SignExt32to64 x) (SignExt32to64 y)) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64DivS) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } + return false +} +func rewriteValueWasm_OpDiv32u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div32u x y) + // result: (I64DivU (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64DivU) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpDiv64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (Div64 [false] x y) + // result: (I64DivS x y) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64DivS) + v.AddArg2(x, y) + return true + } + return false +} +func rewriteValueWasm_OpDiv8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div8 x y) + // result: (I64DivS (SignExt8to64 x) (SignExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64DivS) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpDiv8u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Div8u x y) + // result: (I64DivU (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64DivU) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpEq16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Eq16 x y) + // result: (I64Eq (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Eq) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpEq32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Eq32 x y) + // result: (I64Eq (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Eq) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpEq8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Eq8 x y) + // result: (I64Eq (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Eq) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpIsNonNil(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (IsNonNil p) + // result: (I64Eqz (I64Eqz p)) + for { + p := v_0 + v.reset(OpWasmI64Eqz) + v0 := b.NewValue0(v.Pos, OpWasmI64Eqz, typ.Bool) + v0.AddArg(p) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpLeq16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq16 x y) + // result: (I64LeS (SignExt16to64 x) (SignExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeS) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLeq16U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq16U x y) + // result: (I64LeU (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeU) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLeq32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq32 x y) + // result: (I64LeS (SignExt32to64 x) (SignExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeS) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLeq32U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq32U x y) + // result: (I64LeU (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeU) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLeq8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq8 x y) + // result: (I64LeS (SignExt8to64 x) (SignExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeS) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLeq8U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Leq8U x y) + // result: (I64LeU (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LeU) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less16 x y) + // result: (I64LtS (SignExt16to64 x) (SignExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtS) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess16U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less16U x y) + // result: (I64LtU (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtU) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less32 x y) + // result: (I64LtS (SignExt32to64 x) (SignExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtS) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess32U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less32U x y) + // result: (I64LtU (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtU) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less8 x y) + // result: (I64LtS (SignExt8to64 x) (SignExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtS) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLess8U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Less8U x y) + // result: (I64LtU (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64LtU) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpLoad(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (Load <t> ptr mem) + // cond: is32BitFloat(t) + // result: (F32Load ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(is32BitFloat(t)) { + break + } + v.reset(OpWasmF32Load) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: is64BitFloat(t) + // result: (F64Load ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(is64BitFloat(t)) { + break + } + v.reset(OpWasmF64Load) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 8 + // result: (I64Load ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 8) { + break + } + v.reset(OpWasmI64Load) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 4 && !t.IsSigned() + // result: (I64Load32U ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 4 && !t.IsSigned()) { + break + } + v.reset(OpWasmI64Load32U) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 4 && t.IsSigned() + // result: (I64Load32S ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 4 && t.IsSigned()) { + break + } + v.reset(OpWasmI64Load32S) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 2 && !t.IsSigned() + // result: (I64Load16U ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 2 && !t.IsSigned()) { + break + } + v.reset(OpWasmI64Load16U) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 2 && t.IsSigned() + // result: (I64Load16S ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 2 && t.IsSigned()) { + break + } + v.reset(OpWasmI64Load16S) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 1 && !t.IsSigned() + // result: (I64Load8U ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 1 && !t.IsSigned()) { + break + } + v.reset(OpWasmI64Load8U) + v.AddArg2(ptr, mem) + return true + } + // match: (Load <t> ptr mem) + // cond: t.Size() == 1 && t.IsSigned() + // result: (I64Load8S ptr mem) + for { + t := v.Type + ptr := v_0 + mem := v_1 + if !(t.Size() == 1 && t.IsSigned()) { + break + } + v.reset(OpWasmI64Load8S) + v.AddArg2(ptr, mem) + return true + } + return false +} +func rewriteValueWasm_OpLocalAddr(v *Value) bool { + v_0 := v.Args[0] + // match: (LocalAddr {sym} base _) + // result: (LoweredAddr {sym} base) + for { + sym := auxToSym(v.Aux) + base := v_0 + v.reset(OpWasmLoweredAddr) + v.Aux = symToAux(sym) + v.AddArg(base) + return true + } +} +func rewriteValueWasm_OpLsh16x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh16x16 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh16x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh16x32 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh16x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh16x8 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh32x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh32x16 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh32x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh32x32 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh32x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh32x8 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh64x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh64x16 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh64x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh64x32 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh64x64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh64x64 x y) + // cond: shiftIsBounded(v) + // result: (I64Shl x y) + for { + x := v_0 + y := v_1 + if !(shiftIsBounded(v)) { + break + } + v.reset(OpWasmI64Shl) + v.AddArg2(x, y) + return true + } + // match: (Lsh64x64 x (I64Const [c])) + // cond: uint64(c) < 64 + // result: (I64Shl x (I64Const [c])) + for { + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) < 64) { + break + } + v.reset(OpWasmI64Shl) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(c) + v.AddArg2(x, v0) + return true + } + // match: (Lsh64x64 x (I64Const [c])) + // cond: uint64(c) >= 64 + // result: (I64Const [0]) + for { + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) >= 64) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(0) + return true + } + // match: (Lsh64x64 x y) + // result: (Select (I64Shl x y) (I64Const [0]) (I64LtU y (I64Const [64]))) + for { + x := v_0 + y := v_1 + v.reset(OpWasmSelect) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v0.AddArg2(x, y) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0) + v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool) + v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v3.AuxInt = int64ToAuxInt(64) + v2.AddArg2(y, v3) + v.AddArg3(v0, v1, v2) + return true + } +} +func rewriteValueWasm_OpLsh64x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh64x8 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh8x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh8x16 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh8x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh8x32 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpLsh8x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Lsh8x8 [c] x y) + // result: (Lsh64x64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpLsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpMod16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod16 [false] x y) + // result: (I64RemS (SignExt16to64 x) (SignExt16to64 y)) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64RemS) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } + return false +} +func rewriteValueWasm_OpMod16u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod16u x y) + // result: (I64RemU (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64RemU) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpMod32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod32 [false] x y) + // result: (I64RemS (SignExt32to64 x) (SignExt32to64 y)) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64RemS) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } + return false +} +func rewriteValueWasm_OpMod32u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod32u x y) + // result: (I64RemU (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64RemU) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpMod64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (Mod64 [false] x y) + // result: (I64RemS x y) + for { + if auxIntToBool(v.AuxInt) != false { + break + } + x := v_0 + y := v_1 + v.reset(OpWasmI64RemS) + v.AddArg2(x, y) + return true + } + return false +} +func rewriteValueWasm_OpMod8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod8 x y) + // result: (I64RemS (SignExt8to64 x) (SignExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64RemS) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpMod8u(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Mod8u x y) + // result: (I64RemU (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64RemU) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpMove(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Move [0] _ _ mem) + // result: mem + for { + if auxIntToInt64(v.AuxInt) != 0 { + break + } + mem := v_2 + v.copyOf(mem) + return true + } + // match: (Move [1] dst src mem) + // result: (I64Store8 dst (I64Load8U src mem) mem) + for { + if auxIntToInt64(v.AuxInt) != 1 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store8) + v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8) + v0.AddArg2(src, mem) + v.AddArg3(dst, v0, mem) + return true + } + // match: (Move [2] dst src mem) + // result: (I64Store16 dst (I64Load16U src mem) mem) + for { + if auxIntToInt64(v.AuxInt) != 2 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store16) + v0 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16) + v0.AddArg2(src, mem) + v.AddArg3(dst, v0, mem) + return true + } + // match: (Move [4] dst src mem) + // result: (I64Store32 dst (I64Load32U src mem) mem) + for { + if auxIntToInt64(v.AuxInt) != 4 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store32) + v0 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32) + v0.AddArg2(src, mem) + v.AddArg3(dst, v0, mem) + return true + } + // match: (Move [8] dst src mem) + // result: (I64Store dst (I64Load src mem) mem) + for { + if auxIntToInt64(v.AuxInt) != 8 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store) + v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v0.AddArg2(src, mem) + v.AddArg3(dst, v0, mem) + return true + } + // match: (Move [16] dst src mem) + // result: (I64Store [8] dst (I64Load [8] src mem) (I64Store dst (I64Load src mem) mem)) + for { + if auxIntToInt64(v.AuxInt) != 16 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(8) + v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v0.AuxInt = int64ToAuxInt(8) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [3] dst src mem) + // result: (I64Store8 [2] dst (I64Load8U [2] src mem) (I64Store16 dst (I64Load16U src mem) mem)) + for { + if auxIntToInt64(v.AuxInt) != 3 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store8) + v.AuxInt = int64ToAuxInt(2) + v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8) + v0.AuxInt = int64ToAuxInt(2) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store16, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [5] dst src mem) + // result: (I64Store8 [4] dst (I64Load8U [4] src mem) (I64Store32 dst (I64Load32U src mem) mem)) + for { + if auxIntToInt64(v.AuxInt) != 5 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store8) + v.AuxInt = int64ToAuxInt(4) + v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8) + v0.AuxInt = int64ToAuxInt(4) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [6] dst src mem) + // result: (I64Store16 [4] dst (I64Load16U [4] src mem) (I64Store32 dst (I64Load32U src mem) mem)) + for { + if auxIntToInt64(v.AuxInt) != 6 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store16) + v.AuxInt = int64ToAuxInt(4) + v0 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16) + v0.AuxInt = int64ToAuxInt(4) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [7] dst src mem) + // result: (I64Store32 [3] dst (I64Load32U [3] src mem) (I64Store32 dst (I64Load32U src mem) mem)) + for { + if auxIntToInt64(v.AuxInt) != 7 { + break + } + dst := v_0 + src := v_1 + mem := v_2 + v.reset(OpWasmI64Store32) + v.AuxInt = int64ToAuxInt(3) + v0 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32) + v0.AuxInt = int64ToAuxInt(3) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [s] dst src mem) + // cond: s > 8 && s < 16 + // result: (I64Store [s-8] dst (I64Load [s-8] src mem) (I64Store dst (I64Load src mem) mem)) + for { + s := auxIntToInt64(v.AuxInt) + dst := v_0 + src := v_1 + mem := v_2 + if !(s > 8 && s < 16) { + break + } + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(s - 8) + v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v0.AuxInt = int64ToAuxInt(s - 8) + v0.AddArg2(src, mem) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v2.AddArg2(src, mem) + v1.AddArg3(dst, v2, mem) + v.AddArg3(dst, v0, v1) + return true + } + // match: (Move [s] dst src mem) + // cond: s > 16 && s%16 != 0 && s%16 <= 8 + // result: (Move [s-s%16] (OffPtr <dst.Type> dst [s%16]) (OffPtr <src.Type> src [s%16]) (I64Store dst (I64Load src mem) mem)) + for { + s := auxIntToInt64(v.AuxInt) + dst := v_0 + src := v_1 + mem := v_2 + if !(s > 16 && s%16 != 0 && s%16 <= 8) { + break + } + v.reset(OpMove) + v.AuxInt = int64ToAuxInt(s - s%16) + v0 := b.NewValue0(v.Pos, OpOffPtr, dst.Type) + v0.AuxInt = int64ToAuxInt(s % 16) + v0.AddArg(dst) + v1 := b.NewValue0(v.Pos, OpOffPtr, src.Type) + v1.AuxInt = int64ToAuxInt(s % 16) + v1.AddArg(src) + v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v3 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v3.AddArg2(src, mem) + v2.AddArg3(dst, v3, mem) + v.AddArg3(v0, v1, v2) + return true + } + // match: (Move [s] dst src mem) + // cond: s > 16 && s%16 != 0 && s%16 > 8 + // result: (Move [s-s%16] (OffPtr <dst.Type> dst [s%16]) (OffPtr <src.Type> src [s%16]) (I64Store [8] dst (I64Load [8] src mem) (I64Store dst (I64Load src mem) mem))) + for { + s := auxIntToInt64(v.AuxInt) + dst := v_0 + src := v_1 + mem := v_2 + if !(s > 16 && s%16 != 0 && s%16 > 8) { + break + } + v.reset(OpMove) + v.AuxInt = int64ToAuxInt(s - s%16) + v0 := b.NewValue0(v.Pos, OpOffPtr, dst.Type) + v0.AuxInt = int64ToAuxInt(s % 16) + v0.AddArg(dst) + v1 := b.NewValue0(v.Pos, OpOffPtr, src.Type) + v1.AuxInt = int64ToAuxInt(s % 16) + v1.AddArg(src) + v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2.AuxInt = int64ToAuxInt(8) + v3 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v3.AuxInt = int64ToAuxInt(8) + v3.AddArg2(src, mem) + v4 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v5 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64) + v5.AddArg2(src, mem) + v4.AddArg3(dst, v5, mem) + v2.AddArg3(dst, v3, v4) + v.AddArg3(v0, v1, v2) + return true + } + // match: (Move [s] dst src mem) + // cond: s%8 == 0 && logLargeCopy(v, s) + // result: (LoweredMove [s/8] dst src mem) + for { + s := auxIntToInt64(v.AuxInt) + dst := v_0 + src := v_1 + mem := v_2 + if !(s%8 == 0 && logLargeCopy(v, s)) { + break + } + v.reset(OpWasmLoweredMove) + v.AuxInt = int64ToAuxInt(s / 8) + v.AddArg3(dst, src, mem) + return true + } + return false +} +func rewriteValueWasm_OpNeg16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neg16 x) + // result: (I64Sub (I64Const [0]) x) + for { + x := v_0 + v.reset(OpWasmI64Sub) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg2(v0, x) + return true + } +} +func rewriteValueWasm_OpNeg32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neg32 x) + // result: (I64Sub (I64Const [0]) x) + for { + x := v_0 + v.reset(OpWasmI64Sub) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg2(v0, x) + return true + } +} +func rewriteValueWasm_OpNeg64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neg64 x) + // result: (I64Sub (I64Const [0]) x) + for { + x := v_0 + v.reset(OpWasmI64Sub) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg2(v0, x) + return true + } +} +func rewriteValueWasm_OpNeg8(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neg8 x) + // result: (I64Sub (I64Const [0]) x) + for { + x := v_0 + v.reset(OpWasmI64Sub) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg2(v0, x) + return true + } +} +func rewriteValueWasm_OpNeq16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neq16 x y) + // result: (I64Ne (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Ne) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpNeq32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neq32 x y) + // result: (I64Ne (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Ne) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpNeq8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Neq8 x y) + // result: (I64Ne (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64Ne) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpPopCount16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (PopCount16 x) + // result: (I64Popcnt (ZeroExt16to64 x)) + for { + x := v_0 + v.reset(OpWasmI64Popcnt) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpPopCount32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (PopCount32 x) + // result: (I64Popcnt (ZeroExt32to64 x)) + for { + x := v_0 + v.reset(OpWasmI64Popcnt) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpPopCount8(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (PopCount8 x) + // result: (I64Popcnt (ZeroExt8to64 x)) + for { + x := v_0 + v.reset(OpWasmI64Popcnt) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v.AddArg(v0) + return true + } +} +func rewriteValueWasm_OpRotateLeft16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (RotateLeft16 <t> x (I64Const [c])) + // result: (Or16 (Lsh16x64 <t> x (I64Const [c&15])) (Rsh16Ux64 <t> x (I64Const [-c&15]))) + for { + t := v.Type + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + v.reset(OpOr16) + v0 := b.NewValue0(v.Pos, OpLsh16x64, t) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(c & 15) + v0.AddArg2(x, v1) + v2 := b.NewValue0(v.Pos, OpRsh16Ux64, t) + v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v3.AuxInt = int64ToAuxInt(-c & 15) + v2.AddArg2(x, v3) + v.AddArg2(v0, v2) + return true + } + return false +} +func rewriteValueWasm_OpRotateLeft8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (RotateLeft8 <t> x (I64Const [c])) + // result: (Or8 (Lsh8x64 <t> x (I64Const [c&7])) (Rsh8Ux64 <t> x (I64Const [-c&7]))) + for { + t := v.Type + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + v.reset(OpOr8) + v0 := b.NewValue0(v.Pos, OpLsh8x64, t) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(c & 7) + v0.AddArg2(x, v1) + v2 := b.NewValue0(v.Pos, OpRsh8Ux64, t) + v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v3.AuxInt = int64ToAuxInt(-c & 7) + v2.AddArg2(x, v3) + v.AddArg2(v0, v2) + return true + } + return false +} +func rewriteValueWasm_OpRsh16Ux16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16Ux16 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt16to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh16Ux32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16Ux32 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt16to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh16Ux64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16Ux64 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt16to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh16Ux8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16Ux8 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt16to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh16x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16x16 [c] x y) + // result: (Rsh64x64 [c] (SignExt16to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh16x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16x32 [c] x y) + // result: (Rsh64x64 [c] (SignExt16to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh16x64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16x64 [c] x y) + // result: (Rsh64x64 [c] (SignExt16to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh16x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh16x8 [c] x y) + // result: (Rsh64x64 [c] (SignExt16to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32Ux16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32Ux16 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt32to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32Ux32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32Ux32 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt32to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32Ux64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32Ux64 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt32to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh32Ux8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32Ux8 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt32to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32x16 [c] x y) + // result: (Rsh64x64 [c] (SignExt32to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32x32 [c] x y) + // result: (Rsh64x64 [c] (SignExt32to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh32x64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32x64 [c] x y) + // result: (Rsh64x64 [c] (SignExt32to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh32x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh32x8 [c] x y) + // result: (Rsh64x64 [c] (SignExt32to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh64Ux16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64Ux16 [c] x y) + // result: (Rsh64Ux64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64Ux32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64Ux32 [c] x y) + // result: (Rsh64Ux64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64Ux64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64Ux64 x y) + // cond: shiftIsBounded(v) + // result: (I64ShrU x y) + for { + x := v_0 + y := v_1 + if !(shiftIsBounded(v)) { + break + } + v.reset(OpWasmI64ShrU) + v.AddArg2(x, y) + return true + } + // match: (Rsh64Ux64 x (I64Const [c])) + // cond: uint64(c) < 64 + // result: (I64ShrU x (I64Const [c])) + for { + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) < 64) { + break + } + v.reset(OpWasmI64ShrU) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(c) + v.AddArg2(x, v0) + return true + } + // match: (Rsh64Ux64 x (I64Const [c])) + // cond: uint64(c) >= 64 + // result: (I64Const [0]) + for { + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) >= 64) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(0) + return true + } + // match: (Rsh64Ux64 x y) + // result: (Select (I64ShrU x y) (I64Const [0]) (I64LtU y (I64Const [64]))) + for { + x := v_0 + y := v_1 + v.reset(OpWasmSelect) + v0 := b.NewValue0(v.Pos, OpWasmI64ShrU, typ.Int64) + v0.AddArg2(x, y) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0) + v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool) + v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v3.AuxInt = int64ToAuxInt(64) + v2.AddArg2(y, v3) + v.AddArg3(v0, v1, v2) + return true + } +} +func rewriteValueWasm_OpRsh64Ux8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64Ux8 [c] x y) + // result: (Rsh64Ux64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64x16 [c] x y) + // result: (Rsh64x64 [c] x (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64x32 [c] x y) + // result: (Rsh64x64 [c] x (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64x64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64x64 x y) + // cond: shiftIsBounded(v) + // result: (I64ShrS x y) + for { + x := v_0 + y := v_1 + if !(shiftIsBounded(v)) { + break + } + v.reset(OpWasmI64ShrS) + v.AddArg2(x, y) + return true + } + // match: (Rsh64x64 x (I64Const [c])) + // cond: uint64(c) < 64 + // result: (I64ShrS x (I64Const [c])) + for { + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) < 64) { + break + } + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(c) + v.AddArg2(x, v0) + return true + } + // match: (Rsh64x64 x (I64Const [c])) + // cond: uint64(c) >= 64 + // result: (I64ShrS x (I64Const [63])) + for { + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + c := auxIntToInt64(v_1.AuxInt) + if !(uint64(c) >= 64) { + break + } + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(63) + v.AddArg2(x, v0) + return true + } + // match: (Rsh64x64 x y) + // result: (I64ShrS x (Select <typ.Int64> y (I64Const [63]) (I64LtU y (I64Const [64])))) + for { + x := v_0 + y := v_1 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmSelect, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(63) + v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool) + v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v3.AuxInt = int64ToAuxInt(64) + v2.AddArg2(y, v3) + v0.AddArg3(y, v1, v2) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh64x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh64x8 [c] x y) + // result: (Rsh64x64 [c] x (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(y) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpRsh8Ux16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8Ux16 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt8to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh8Ux32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8Ux32 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt8to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh8Ux64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8Ux64 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt8to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh8Ux8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8Ux8 [c] x y) + // result: (Rsh64Ux64 [c] (ZeroExt8to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64Ux64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh8x16(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8x16 [c] x y) + // result: (Rsh64x64 [c] (SignExt8to64 x) (ZeroExt16to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh8x32(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8x32 [c] x y) + // result: (Rsh64x64 [c] (SignExt8to64 x) (ZeroExt32to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpRsh8x64(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8x64 [c] x y) + // result: (Rsh64x64 [c] (SignExt8to64 x) y) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v.AddArg2(v0, y) + return true + } +} +func rewriteValueWasm_OpRsh8x8(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Rsh8x8 [c] x y) + // result: (Rsh64x64 [c] (SignExt8to64 x) (ZeroExt8to64 y)) + for { + c := auxIntToBool(v.AuxInt) + x := v_0 + y := v_1 + v.reset(OpRsh64x64) + v.AuxInt = boolToAuxInt(c) + v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64) + v0.AddArg(x) + v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64) + v1.AddArg(y) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt16to32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt16to32 x:(I64Load16S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load16S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt16to32 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend16S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend16S) + v.AddArg(x) + return true + } + // match: (SignExt16to32 x) + // result: (I64ShrS (I64Shl x (I64Const [48])) (I64Const [48])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(48) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt16to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt16to64 x:(I64Load16S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load16S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt16to64 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend16S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend16S) + v.AddArg(x) + return true + } + // match: (SignExt16to64 x) + // result: (I64ShrS (I64Shl x (I64Const [48])) (I64Const [48])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(48) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt32to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt32to64 x:(I64Load32S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load32S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt32to64 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend32S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend32S) + v.AddArg(x) + return true + } + // match: (SignExt32to64 x) + // result: (I64ShrS (I64Shl x (I64Const [32])) (I64Const [32])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(32) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt8to16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt8to16 x:(I64Load8S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt8to16 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend8S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend8S) + v.AddArg(x) + return true + } + // match: (SignExt8to16 x) + // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(56) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt8to32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt8to32 x:(I64Load8S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt8to32 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend8S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend8S) + v.AddArg(x) + return true + } + // match: (SignExt8to32 x) + // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(56) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSignExt8to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (SignExt8to64 x:(I64Load8S _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8S { + break + } + v.copyOf(x) + return true + } + // match: (SignExt8to64 x) + // cond: objabi.GOWASM.SignExt + // result: (I64Extend8S x) + for { + x := v_0 + if !(objabi.GOWASM.SignExt) { + break + } + v.reset(OpWasmI64Extend8S) + v.AddArg(x) + return true + } + // match: (SignExt8to64 x) + // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(56) + v0.AddArg2(x, v1) + v.AddArg2(v0, v1) + return true + } +} +func rewriteValueWasm_OpSlicemask(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Slicemask x) + // result: (I64ShrS (I64Sub (I64Const [0]) x) (I64Const [63])) + for { + x := v_0 + v.reset(OpWasmI64ShrS) + v0 := b.NewValue0(v.Pos, OpWasmI64Sub, typ.Int64) + v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v1.AuxInt = int64ToAuxInt(0) + v0.AddArg2(v1, x) + v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v2.AuxInt = int64ToAuxInt(63) + v.AddArg2(v0, v2) + return true + } +} +func rewriteValueWasm_OpStore(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (Store {t} ptr val mem) + // cond: is64BitFloat(t) + // result: (F64Store ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(is64BitFloat(t)) { + break + } + v.reset(OpWasmF64Store) + v.AddArg3(ptr, val, mem) + return true + } + // match: (Store {t} ptr val mem) + // cond: is32BitFloat(t) + // result: (F32Store ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(is32BitFloat(t)) { + break + } + v.reset(OpWasmF32Store) + v.AddArg3(ptr, val, mem) + return true + } + // match: (Store {t} ptr val mem) + // cond: t.Size() == 8 + // result: (I64Store ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(t.Size() == 8) { + break + } + v.reset(OpWasmI64Store) + v.AddArg3(ptr, val, mem) + return true + } + // match: (Store {t} ptr val mem) + // cond: t.Size() == 4 + // result: (I64Store32 ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(t.Size() == 4) { + break + } + v.reset(OpWasmI64Store32) + v.AddArg3(ptr, val, mem) + return true + } + // match: (Store {t} ptr val mem) + // cond: t.Size() == 2 + // result: (I64Store16 ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(t.Size() == 2) { + break + } + v.reset(OpWasmI64Store16) + v.AddArg3(ptr, val, mem) + return true + } + // match: (Store {t} ptr val mem) + // cond: t.Size() == 1 + // result: (I64Store8 ptr val mem) + for { + t := auxToType(v.Aux) + ptr := v_0 + val := v_1 + mem := v_2 + if !(t.Size() == 1) { + break + } + v.reset(OpWasmI64Store8) + v.AddArg3(ptr, val, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmF64Add(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (F64Add (F64Const [x]) (F64Const [y])) + // result: (F64Const [x + y]) + for { + if v_0.Op != OpWasmF64Const { + break + } + x := auxIntToFloat64(v_0.AuxInt) + if v_1.Op != OpWasmF64Const { + break + } + y := auxIntToFloat64(v_1.AuxInt) + v.reset(OpWasmF64Const) + v.AuxInt = float64ToAuxInt(x + y) + return true + } + // match: (F64Add (F64Const [x]) y) + // cond: y.Op != OpWasmF64Const + // result: (F64Add y (F64Const [x])) + for { + if v_0.Op != OpWasmF64Const { + break + } + x := auxIntToFloat64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmF64Const) { + break + } + v.reset(OpWasmF64Add) + v0 := b.NewValue0(v.Pos, OpWasmF64Const, typ.Float64) + v0.AuxInt = float64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmF64Mul(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (F64Mul (F64Const [x]) (F64Const [y])) + // cond: !math.IsNaN(x * y) + // result: (F64Const [x * y]) + for { + if v_0.Op != OpWasmF64Const { + break + } + x := auxIntToFloat64(v_0.AuxInt) + if v_1.Op != OpWasmF64Const { + break + } + y := auxIntToFloat64(v_1.AuxInt) + if !(!math.IsNaN(x * y)) { + break + } + v.reset(OpWasmF64Const) + v.AuxInt = float64ToAuxInt(x * y) + return true + } + // match: (F64Mul (F64Const [x]) y) + // cond: y.Op != OpWasmF64Const + // result: (F64Mul y (F64Const [x])) + for { + if v_0.Op != OpWasmF64Const { + break + } + x := auxIntToFloat64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmF64Const) { + break + } + v.reset(OpWasmF64Mul) + v0 := b.NewValue0(v.Pos, OpWasmF64Const, typ.Float64) + v0.AuxInt = float64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Add(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Add (I64Const [x]) (I64Const [y])) + // result: (I64Const [x + y]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x + y) + return true + } + // match: (I64Add (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Add y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Add) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + // match: (I64Add x (I64Const [y])) + // result: (I64AddConst [y] x) + for { + x := v_0 + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64AddConst) + v.AuxInt = int64ToAuxInt(y) + v.AddArg(x) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64AddConst(v *Value) bool { + v_0 := v.Args[0] + // match: (I64AddConst [0] x) + // result: x + for { + if auxIntToInt64(v.AuxInt) != 0 { + break + } + x := v_0 + v.copyOf(x) + return true + } + // match: (I64AddConst [off] (LoweredAddr {sym} [off2] base)) + // cond: isU32Bit(off+int64(off2)) + // result: (LoweredAddr {sym} [int32(off)+off2] base) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmLoweredAddr { + break + } + off2 := auxIntToInt32(v_0.AuxInt) + sym := auxToSym(v_0.Aux) + base := v_0.Args[0] + if !(isU32Bit(off + int64(off2))) { + break + } + v.reset(OpWasmLoweredAddr) + v.AuxInt = int32ToAuxInt(int32(off) + off2) + v.Aux = symToAux(sym) + v.AddArg(base) + return true + } + // match: (I64AddConst [off] x:(SP)) + // cond: isU32Bit(off) + // result: (LoweredAddr [int32(off)] x) + for { + off := auxIntToInt64(v.AuxInt) + x := v_0 + if x.Op != OpSP || !(isU32Bit(off)) { + break + } + v.reset(OpWasmLoweredAddr) + v.AuxInt = int32ToAuxInt(int32(off)) + v.AddArg(x) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64And(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64And (I64Const [x]) (I64Const [y])) + // result: (I64Const [x & y]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x & y) + return true + } + // match: (I64And (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64And y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Eq(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Eq (I64Const [x]) (I64Const [y])) + // cond: x == y + // result: (I64Const [1]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + if !(x == y) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(1) + return true + } + // match: (I64Eq (I64Const [x]) (I64Const [y])) + // cond: x != y + // result: (I64Const [0]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + if !(x != y) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(0) + return true + } + // match: (I64Eq (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Eq y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Eq) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + // match: (I64Eq x (I64Const [0])) + // result: (I64Eqz x) + for { + x := v_0 + if v_1.Op != OpWasmI64Const || auxIntToInt64(v_1.AuxInt) != 0 { + break + } + v.reset(OpWasmI64Eqz) + v.AddArg(x) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Eqz(v *Value) bool { + v_0 := v.Args[0] + // match: (I64Eqz (I64Eqz (I64Eqz x))) + // result: (I64Eqz x) + for { + if v_0.Op != OpWasmI64Eqz { + break + } + v_0_0 := v_0.Args[0] + if v_0_0.Op != OpWasmI64Eqz { + break + } + x := v_0_0.Args[0] + v.reset(OpWasmI64Eqz) + v.AddArg(x) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64LeU(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64LeU x (I64Const [0])) + // result: (I64Eqz x) + for { + x := v_0 + if v_1.Op != OpWasmI64Const || auxIntToInt64(v_1.AuxInt) != 0 { + break + } + v.reset(OpWasmI64Eqz) + v.AddArg(x) + return true + } + // match: (I64LeU (I64Const [1]) x) + // result: (I64Eqz (I64Eqz x)) + for { + if v_0.Op != OpWasmI64Const || auxIntToInt64(v_0.AuxInt) != 1 { + break + } + x := v_1 + v.reset(OpWasmI64Eqz) + v0 := b.NewValue0(v.Pos, OpWasmI64Eqz, typ.Bool) + v0.AddArg(x) + v.AddArg(v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + config := b.Func.Config + // match: (I64Load [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + // match: (I64Load [off] (LoweredAddr {sym} [off2] (SB)) _) + // cond: symIsRO(sym) && isU32Bit(off+int64(off2)) + // result: (I64Const [int64(read64(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))]) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmLoweredAddr { + break + } + off2 := auxIntToInt32(v_0.AuxInt) + sym := auxToSym(v_0.Aux) + v_0_0 := v_0.Args[0] + if v_0_0.Op != OpSB || !(symIsRO(sym) && isU32Bit(off+int64(off2))) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(read64(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load16S(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Load16S [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load16S [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load16S) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load16U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + config := b.Func.Config + // match: (I64Load16U [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load16U [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load16U) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + // match: (I64Load16U [off] (LoweredAddr {sym} [off2] (SB)) _) + // cond: symIsRO(sym) && isU32Bit(off+int64(off2)) + // result: (I64Const [int64(read16(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))]) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmLoweredAddr { + break + } + off2 := auxIntToInt32(v_0.AuxInt) + sym := auxToSym(v_0.Aux) + v_0_0 := v_0.Args[0] + if v_0_0.Op != OpSB || !(symIsRO(sym) && isU32Bit(off+int64(off2))) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(read16(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load32S(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Load32S [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load32S [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load32S) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load32U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + config := b.Func.Config + // match: (I64Load32U [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load32U [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load32U) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + // match: (I64Load32U [off] (LoweredAddr {sym} [off2] (SB)) _) + // cond: symIsRO(sym) && isU32Bit(off+int64(off2)) + // result: (I64Const [int64(read32(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))]) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmLoweredAddr { + break + } + off2 := auxIntToInt32(v_0.AuxInt) + sym := auxToSym(v_0.Aux) + v_0_0 := v_0.Args[0] + if v_0_0.Op != OpSB || !(symIsRO(sym) && isU32Bit(off+int64(off2))) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(read32(sym, off+int64(off2), config.ctxt.Arch.ByteOrder))) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load8S(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Load8S [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load8S [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load8S) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Load8U(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Load8U [off] (I64AddConst [off2] ptr) mem) + // cond: isU32Bit(off+off2) + // result: (I64Load8U [off+off2] ptr mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + mem := v_1 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Load8U) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg2(ptr, mem) + return true + } + // match: (I64Load8U [off] (LoweredAddr {sym} [off2] (SB)) _) + // cond: symIsRO(sym) && isU32Bit(off+int64(off2)) + // result: (I64Const [int64(read8(sym, off+int64(off2)))]) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmLoweredAddr { + break + } + off2 := auxIntToInt32(v_0.AuxInt) + sym := auxToSym(v_0.Aux) + v_0_0 := v_0.Args[0] + if v_0_0.Op != OpSB || !(symIsRO(sym) && isU32Bit(off+int64(off2))) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(read8(sym, off+int64(off2)))) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64LtU(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64LtU (I64Const [0]) x) + // result: (I64Eqz (I64Eqz x)) + for { + if v_0.Op != OpWasmI64Const || auxIntToInt64(v_0.AuxInt) != 0 { + break + } + x := v_1 + v.reset(OpWasmI64Eqz) + v0 := b.NewValue0(v.Pos, OpWasmI64Eqz, typ.Bool) + v0.AddArg(x) + v.AddArg(v0) + return true + } + // match: (I64LtU x (I64Const [1])) + // result: (I64Eqz x) + for { + x := v_0 + if v_1.Op != OpWasmI64Const || auxIntToInt64(v_1.AuxInt) != 1 { + break + } + v.reset(OpWasmI64Eqz) + v.AddArg(x) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Mul(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Mul (I64Const [x]) (I64Const [y])) + // result: (I64Const [x * y]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x * y) + return true + } + // match: (I64Mul (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Mul y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Mul) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Ne(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Ne (I64Const [x]) (I64Const [y])) + // cond: x == y + // result: (I64Const [0]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + if !(x == y) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(0) + return true + } + // match: (I64Ne (I64Const [x]) (I64Const [y])) + // cond: x != y + // result: (I64Const [1]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + if !(x != y) { + break + } + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(1) + return true + } + // match: (I64Ne (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Ne y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Ne) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + // match: (I64Ne x (I64Const [0])) + // result: (I64Eqz (I64Eqz x)) + for { + x := v_0 + if v_1.Op != OpWasmI64Const || auxIntToInt64(v_1.AuxInt) != 0 { + break + } + v.reset(OpWasmI64Eqz) + v0 := b.NewValue0(v.Pos, OpWasmI64Eqz, typ.Bool) + v0.AddArg(x) + v.AddArg(v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Or(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Or (I64Const [x]) (I64Const [y])) + // result: (I64Const [x | y]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x | y) + return true + } + // match: (I64Or (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Or y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Or) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Shl(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Shl (I64Const [x]) (I64Const [y])) + // result: (I64Const [x << uint64(y)]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x << uint64(y)) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64ShrS(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64ShrS (I64Const [x]) (I64Const [y])) + // result: (I64Const [x >> uint64(y)]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x >> uint64(y)) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64ShrU(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64ShrU (I64Const [x]) (I64Const [y])) + // result: (I64Const [int64(uint64(x) >> uint64(y))]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(int64(uint64(x) >> uint64(y))) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Store(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Store [off] (I64AddConst [off2] ptr) val mem) + // cond: isU32Bit(off+off2) + // result: (I64Store [off+off2] ptr val mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + val := v_1 + mem := v_2 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg3(ptr, val, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Store16(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Store16 [off] (I64AddConst [off2] ptr) val mem) + // cond: isU32Bit(off+off2) + // result: (I64Store16 [off+off2] ptr val mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + val := v_1 + mem := v_2 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Store16) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg3(ptr, val, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Store32(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Store32 [off] (I64AddConst [off2] ptr) val mem) + // cond: isU32Bit(off+off2) + // result: (I64Store32 [off+off2] ptr val mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + val := v_1 + mem := v_2 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Store32) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg3(ptr, val, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Store8(v *Value) bool { + v_2 := v.Args[2] + v_1 := v.Args[1] + v_0 := v.Args[0] + // match: (I64Store8 [off] (I64AddConst [off2] ptr) val mem) + // cond: isU32Bit(off+off2) + // result: (I64Store8 [off+off2] ptr val mem) + for { + off := auxIntToInt64(v.AuxInt) + if v_0.Op != OpWasmI64AddConst { + break + } + off2 := auxIntToInt64(v_0.AuxInt) + ptr := v_0.Args[0] + val := v_1 + mem := v_2 + if !(isU32Bit(off + off2)) { + break + } + v.reset(OpWasmI64Store8) + v.AuxInt = int64ToAuxInt(off + off2) + v.AddArg3(ptr, val, mem) + return true + } + return false +} +func rewriteValueWasm_OpWasmI64Xor(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (I64Xor (I64Const [x]) (I64Const [y])) + // result: (I64Const [x ^ y]) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + if v_1.Op != OpWasmI64Const { + break + } + y := auxIntToInt64(v_1.AuxInt) + v.reset(OpWasmI64Const) + v.AuxInt = int64ToAuxInt(x ^ y) + return true + } + // match: (I64Xor (I64Const [x]) y) + // cond: y.Op != OpWasmI64Const + // result: (I64Xor y (I64Const [x])) + for { + if v_0.Op != OpWasmI64Const { + break + } + x := auxIntToInt64(v_0.AuxInt) + y := v_1 + if !(y.Op != OpWasmI64Const) { + break + } + v.reset(OpWasmI64Xor) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(x) + v.AddArg2(y, v0) + return true + } + return false +} +func rewriteValueWasm_OpZero(v *Value) bool { + v_1 := v.Args[1] + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (Zero [0] _ mem) + // result: mem + for { + if auxIntToInt64(v.AuxInt) != 0 { + break + } + mem := v_1 + v.copyOf(mem) + return true + } + // match: (Zero [1] destptr mem) + // result: (I64Store8 destptr (I64Const [0]) mem) + for { + if auxIntToInt64(v.AuxInt) != 1 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store8) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg3(destptr, v0, mem) + return true + } + // match: (Zero [2] destptr mem) + // result: (I64Store16 destptr (I64Const [0]) mem) + for { + if auxIntToInt64(v.AuxInt) != 2 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store16) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg3(destptr, v0, mem) + return true + } + // match: (Zero [4] destptr mem) + // result: (I64Store32 destptr (I64Const [0]) mem) + for { + if auxIntToInt64(v.AuxInt) != 4 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store32) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg3(destptr, v0, mem) + return true + } + // match: (Zero [8] destptr mem) + // result: (I64Store destptr (I64Const [0]) mem) + for { + if auxIntToInt64(v.AuxInt) != 8 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v.AddArg3(destptr, v0, mem) + return true + } + // match: (Zero [3] destptr mem) + // result: (I64Store8 [2] destptr (I64Const [0]) (I64Store16 destptr (I64Const [0]) mem)) + for { + if auxIntToInt64(v.AuxInt) != 3 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store8) + v.AuxInt = int64ToAuxInt(2) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store16, types.TypeMem) + v1.AddArg3(destptr, v0, mem) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [5] destptr mem) + // result: (I64Store8 [4] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem)) + for { + if auxIntToInt64(v.AuxInt) != 5 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store8) + v.AuxInt = int64ToAuxInt(4) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v1.AddArg3(destptr, v0, mem) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [6] destptr mem) + // result: (I64Store16 [4] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem)) + for { + if auxIntToInt64(v.AuxInt) != 6 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store16) + v.AuxInt = int64ToAuxInt(4) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v1.AddArg3(destptr, v0, mem) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [7] destptr mem) + // result: (I64Store32 [3] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem)) + for { + if auxIntToInt64(v.AuxInt) != 7 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store32) + v.AuxInt = int64ToAuxInt(3) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem) + v1.AddArg3(destptr, v0, mem) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [s] destptr mem) + // cond: s%8 != 0 && s > 8 + // result: (Zero [s-s%8] (OffPtr <destptr.Type> destptr [s%8]) (I64Store destptr (I64Const [0]) mem)) + for { + s := auxIntToInt64(v.AuxInt) + destptr := v_0 + mem := v_1 + if !(s%8 != 0 && s > 8) { + break + } + v.reset(OpZero) + v.AuxInt = int64ToAuxInt(s - s%8) + v0 := b.NewValue0(v.Pos, OpOffPtr, destptr.Type) + v0.AuxInt = int64ToAuxInt(s % 8) + v0.AddArg(destptr) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v2.AuxInt = int64ToAuxInt(0) + v1.AddArg3(destptr, v2, mem) + v.AddArg2(v0, v1) + return true + } + // match: (Zero [16] destptr mem) + // result: (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem)) + for { + if auxIntToInt64(v.AuxInt) != 16 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(8) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v1.AddArg3(destptr, v0, mem) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [24] destptr mem) + // result: (I64Store [16] destptr (I64Const [0]) (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem))) + for { + if auxIntToInt64(v.AuxInt) != 24 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(16) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v1.AuxInt = int64ToAuxInt(8) + v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2.AddArg3(destptr, v0, mem) + v1.AddArg3(destptr, v0, v2) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [32] destptr mem) + // result: (I64Store [24] destptr (I64Const [0]) (I64Store [16] destptr (I64Const [0]) (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem)))) + for { + if auxIntToInt64(v.AuxInt) != 32 { + break + } + destptr := v_0 + mem := v_1 + v.reset(OpWasmI64Store) + v.AuxInt = int64ToAuxInt(24) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0) + v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v1.AuxInt = int64ToAuxInt(16) + v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v2.AuxInt = int64ToAuxInt(8) + v3 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem) + v3.AddArg3(destptr, v0, mem) + v2.AddArg3(destptr, v0, v3) + v1.AddArg3(destptr, v0, v2) + v.AddArg3(destptr, v0, v1) + return true + } + // match: (Zero [s] destptr mem) + // cond: s%8 == 0 && s > 32 + // result: (LoweredZero [s/8] destptr mem) + for { + s := auxIntToInt64(v.AuxInt) + destptr := v_0 + mem := v_1 + if !(s%8 == 0 && s > 32) { + break + } + v.reset(OpWasmLoweredZero) + v.AuxInt = int64ToAuxInt(s / 8) + v.AddArg2(destptr, mem) + return true + } + return false +} +func rewriteValueWasm_OpZeroExt16to32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt16to32 x:(I64Load16U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load16U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt16to32 x) + // result: (I64And x (I64Const [0xffff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xffff) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpZeroExt16to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt16to64 x:(I64Load16U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load16U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt16to64 x) + // result: (I64And x (I64Const [0xffff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xffff) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpZeroExt32to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt32to64 x:(I64Load32U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load32U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt32to64 x) + // result: (I64And x (I64Const [0xffffffff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xffffffff) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpZeroExt8to16(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt8to16 x:(I64Load8U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt8to16 x) + // result: (I64And x (I64Const [0xff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xff) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpZeroExt8to32(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt8to32 x:(I64Load8U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt8to32 x) + // result: (I64And x (I64Const [0xff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xff) + v.AddArg2(x, v0) + return true + } +} +func rewriteValueWasm_OpZeroExt8to64(v *Value) bool { + v_0 := v.Args[0] + b := v.Block + typ := &b.Func.Config.Types + // match: (ZeroExt8to64 x:(I64Load8U _ _)) + // result: x + for { + x := v_0 + if x.Op != OpWasmI64Load8U { + break + } + v.copyOf(x) + return true + } + // match: (ZeroExt8to64 x) + // result: (I64And x (I64Const [0xff])) + for { + x := v_0 + v.reset(OpWasmI64And) + v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64) + v0.AuxInt = int64ToAuxInt(0xff) + v.AddArg2(x, v0) + return true + } +} +func rewriteBlockWasm(b *Block) bool { + switch b.Kind { + } + return false +} |