aboutsummaryrefslogtreecommitdiff
path: root/src/cmd/compile/internal/ssa/rewriteWasm.go
diff options
context:
space:
mode:
authorRichard Musiol <mail@richard-musiol.de>2018-03-29 00:55:53 +0200
committerBrad Fitzpatrick <bradfitz@golang.org>2018-05-04 17:56:12 +0000
commit3b137dd2df19c261a007b8a620a2182cd679d700 (patch)
tree1008f11b278e1fc1dc6dc222c0c67654a6fee534 /src/cmd/compile/internal/ssa/rewriteWasm.go
parenta9fc37525891e47b4277cde040a06db585e1780d (diff)
downloadgo-3b137dd2df19c261a007b8a620a2182cd679d700.tar.gz
go-3b137dd2df19c261a007b8a620a2182cd679d700.zip
cmd/compile: add wasm architecture
This commit adds the wasm architecture to the compile command. A later commit will contain the corresponding linker changes. Design doc: https://docs.google.com/document/d/131vjr4DH6JFnb-blm_uRdaC0_Nv3OUwjEY5qVCxCup4 The following files are generated: - src/cmd/compile/internal/ssa/opGen.go - src/cmd/compile/internal/ssa/rewriteWasm.go - src/cmd/internal/obj/wasm/anames.go Updates #18892 Change-Id: Ifb4a96a3e427aac2362a1c97967d5667450fba3b Reviewed-on: https://go-review.googlesource.com/103295 Run-TryBot: Brad Fitzpatrick <bradfitz@golang.org> TryBot-Result: Gobot Gobot <gobot@golang.org> Reviewed-by: Cherry Zhang <cherryyz@google.com>
Diffstat (limited to 'src/cmd/compile/internal/ssa/rewriteWasm.go')
-rw-r--r--src/cmd/compile/internal/ssa/rewriteWasm.go6284
1 files changed, 6284 insertions, 0 deletions
diff --git a/src/cmd/compile/internal/ssa/rewriteWasm.go b/src/cmd/compile/internal/ssa/rewriteWasm.go
new file mode 100644
index 0000000000..f488a93522
--- /dev/null
+++ b/src/cmd/compile/internal/ssa/rewriteWasm.go
@@ -0,0 +1,6284 @@
+// Code generated from gen/Wasm.rules; DO NOT EDIT.
+// generated with: cd gen; go run *.go
+
+package ssa
+
+import "math"
+import "cmd/internal/obj"
+import "cmd/internal/objabi"
+import "cmd/compile/internal/types"
+
+var _ = math.MinInt8 // in case not otherwise used
+var _ = obj.ANOP // in case not otherwise used
+var _ = objabi.GOROOT // in case not otherwise used
+var _ = types.TypeMem // in case not otherwise used
+
+func rewriteValueWasm(v *Value) bool {
+ switch v.Op {
+ case OpAdd16:
+ return rewriteValueWasm_OpAdd16_0(v)
+ case OpAdd32:
+ return rewriteValueWasm_OpAdd32_0(v)
+ case OpAdd32F:
+ return rewriteValueWasm_OpAdd32F_0(v)
+ case OpAdd64:
+ return rewriteValueWasm_OpAdd64_0(v)
+ case OpAdd64F:
+ return rewriteValueWasm_OpAdd64F_0(v)
+ case OpAdd8:
+ return rewriteValueWasm_OpAdd8_0(v)
+ case OpAddPtr:
+ return rewriteValueWasm_OpAddPtr_0(v)
+ case OpAddr:
+ return rewriteValueWasm_OpAddr_0(v)
+ case OpAnd16:
+ return rewriteValueWasm_OpAnd16_0(v)
+ case OpAnd32:
+ return rewriteValueWasm_OpAnd32_0(v)
+ case OpAnd64:
+ return rewriteValueWasm_OpAnd64_0(v)
+ case OpAnd8:
+ return rewriteValueWasm_OpAnd8_0(v)
+ case OpAndB:
+ return rewriteValueWasm_OpAndB_0(v)
+ case OpClosureCall:
+ return rewriteValueWasm_OpClosureCall_0(v)
+ case OpCom16:
+ return rewriteValueWasm_OpCom16_0(v)
+ case OpCom32:
+ return rewriteValueWasm_OpCom32_0(v)
+ case OpCom64:
+ return rewriteValueWasm_OpCom64_0(v)
+ case OpCom8:
+ return rewriteValueWasm_OpCom8_0(v)
+ case OpConst16:
+ return rewriteValueWasm_OpConst16_0(v)
+ case OpConst32:
+ return rewriteValueWasm_OpConst32_0(v)
+ case OpConst32F:
+ return rewriteValueWasm_OpConst32F_0(v)
+ case OpConst64:
+ return rewriteValueWasm_OpConst64_0(v)
+ case OpConst64F:
+ return rewriteValueWasm_OpConst64F_0(v)
+ case OpConst8:
+ return rewriteValueWasm_OpConst8_0(v)
+ case OpConstBool:
+ return rewriteValueWasm_OpConstBool_0(v)
+ case OpConstNil:
+ return rewriteValueWasm_OpConstNil_0(v)
+ case OpConvert:
+ return rewriteValueWasm_OpConvert_0(v)
+ case OpCvt32Fto32:
+ return rewriteValueWasm_OpCvt32Fto32_0(v)
+ case OpCvt32Fto32U:
+ return rewriteValueWasm_OpCvt32Fto32U_0(v)
+ case OpCvt32Fto64:
+ return rewriteValueWasm_OpCvt32Fto64_0(v)
+ case OpCvt32Fto64F:
+ return rewriteValueWasm_OpCvt32Fto64F_0(v)
+ case OpCvt32Fto64U:
+ return rewriteValueWasm_OpCvt32Fto64U_0(v)
+ case OpCvt32Uto32F:
+ return rewriteValueWasm_OpCvt32Uto32F_0(v)
+ case OpCvt32Uto64F:
+ return rewriteValueWasm_OpCvt32Uto64F_0(v)
+ case OpCvt32to32F:
+ return rewriteValueWasm_OpCvt32to32F_0(v)
+ case OpCvt32to64F:
+ return rewriteValueWasm_OpCvt32to64F_0(v)
+ case OpCvt64Fto32:
+ return rewriteValueWasm_OpCvt64Fto32_0(v)
+ case OpCvt64Fto32F:
+ return rewriteValueWasm_OpCvt64Fto32F_0(v)
+ case OpCvt64Fto32U:
+ return rewriteValueWasm_OpCvt64Fto32U_0(v)
+ case OpCvt64Fto64:
+ return rewriteValueWasm_OpCvt64Fto64_0(v)
+ case OpCvt64Fto64U:
+ return rewriteValueWasm_OpCvt64Fto64U_0(v)
+ case OpCvt64Uto32F:
+ return rewriteValueWasm_OpCvt64Uto32F_0(v)
+ case OpCvt64Uto64F:
+ return rewriteValueWasm_OpCvt64Uto64F_0(v)
+ case OpCvt64to32F:
+ return rewriteValueWasm_OpCvt64to32F_0(v)
+ case OpCvt64to64F:
+ return rewriteValueWasm_OpCvt64to64F_0(v)
+ case OpDiv16:
+ return rewriteValueWasm_OpDiv16_0(v)
+ case OpDiv16u:
+ return rewriteValueWasm_OpDiv16u_0(v)
+ case OpDiv32:
+ return rewriteValueWasm_OpDiv32_0(v)
+ case OpDiv32F:
+ return rewriteValueWasm_OpDiv32F_0(v)
+ case OpDiv32u:
+ return rewriteValueWasm_OpDiv32u_0(v)
+ case OpDiv64:
+ return rewriteValueWasm_OpDiv64_0(v)
+ case OpDiv64F:
+ return rewriteValueWasm_OpDiv64F_0(v)
+ case OpDiv64u:
+ return rewriteValueWasm_OpDiv64u_0(v)
+ case OpDiv8:
+ return rewriteValueWasm_OpDiv8_0(v)
+ case OpDiv8u:
+ return rewriteValueWasm_OpDiv8u_0(v)
+ case OpEq16:
+ return rewriteValueWasm_OpEq16_0(v)
+ case OpEq32:
+ return rewriteValueWasm_OpEq32_0(v)
+ case OpEq32F:
+ return rewriteValueWasm_OpEq32F_0(v)
+ case OpEq64:
+ return rewriteValueWasm_OpEq64_0(v)
+ case OpEq64F:
+ return rewriteValueWasm_OpEq64F_0(v)
+ case OpEq8:
+ return rewriteValueWasm_OpEq8_0(v)
+ case OpEqB:
+ return rewriteValueWasm_OpEqB_0(v)
+ case OpEqPtr:
+ return rewriteValueWasm_OpEqPtr_0(v)
+ case OpGeq16:
+ return rewriteValueWasm_OpGeq16_0(v)
+ case OpGeq16U:
+ return rewriteValueWasm_OpGeq16U_0(v)
+ case OpGeq32:
+ return rewriteValueWasm_OpGeq32_0(v)
+ case OpGeq32F:
+ return rewriteValueWasm_OpGeq32F_0(v)
+ case OpGeq32U:
+ return rewriteValueWasm_OpGeq32U_0(v)
+ case OpGeq64:
+ return rewriteValueWasm_OpGeq64_0(v)
+ case OpGeq64F:
+ return rewriteValueWasm_OpGeq64F_0(v)
+ case OpGeq64U:
+ return rewriteValueWasm_OpGeq64U_0(v)
+ case OpGeq8:
+ return rewriteValueWasm_OpGeq8_0(v)
+ case OpGeq8U:
+ return rewriteValueWasm_OpGeq8U_0(v)
+ case OpGetCallerPC:
+ return rewriteValueWasm_OpGetCallerPC_0(v)
+ case OpGetCallerSP:
+ return rewriteValueWasm_OpGetCallerSP_0(v)
+ case OpGetClosurePtr:
+ return rewriteValueWasm_OpGetClosurePtr_0(v)
+ case OpGreater16:
+ return rewriteValueWasm_OpGreater16_0(v)
+ case OpGreater16U:
+ return rewriteValueWasm_OpGreater16U_0(v)
+ case OpGreater32:
+ return rewriteValueWasm_OpGreater32_0(v)
+ case OpGreater32F:
+ return rewriteValueWasm_OpGreater32F_0(v)
+ case OpGreater32U:
+ return rewriteValueWasm_OpGreater32U_0(v)
+ case OpGreater64:
+ return rewriteValueWasm_OpGreater64_0(v)
+ case OpGreater64F:
+ return rewriteValueWasm_OpGreater64F_0(v)
+ case OpGreater64U:
+ return rewriteValueWasm_OpGreater64U_0(v)
+ case OpGreater8:
+ return rewriteValueWasm_OpGreater8_0(v)
+ case OpGreater8U:
+ return rewriteValueWasm_OpGreater8U_0(v)
+ case OpInterCall:
+ return rewriteValueWasm_OpInterCall_0(v)
+ case OpIsInBounds:
+ return rewriteValueWasm_OpIsInBounds_0(v)
+ case OpIsNonNil:
+ return rewriteValueWasm_OpIsNonNil_0(v)
+ case OpIsSliceInBounds:
+ return rewriteValueWasm_OpIsSliceInBounds_0(v)
+ case OpLeq16:
+ return rewriteValueWasm_OpLeq16_0(v)
+ case OpLeq16U:
+ return rewriteValueWasm_OpLeq16U_0(v)
+ case OpLeq32:
+ return rewriteValueWasm_OpLeq32_0(v)
+ case OpLeq32F:
+ return rewriteValueWasm_OpLeq32F_0(v)
+ case OpLeq32U:
+ return rewriteValueWasm_OpLeq32U_0(v)
+ case OpLeq64:
+ return rewriteValueWasm_OpLeq64_0(v)
+ case OpLeq64F:
+ return rewriteValueWasm_OpLeq64F_0(v)
+ case OpLeq64U:
+ return rewriteValueWasm_OpLeq64U_0(v)
+ case OpLeq8:
+ return rewriteValueWasm_OpLeq8_0(v)
+ case OpLeq8U:
+ return rewriteValueWasm_OpLeq8U_0(v)
+ case OpLess16:
+ return rewriteValueWasm_OpLess16_0(v)
+ case OpLess16U:
+ return rewriteValueWasm_OpLess16U_0(v)
+ case OpLess32:
+ return rewriteValueWasm_OpLess32_0(v)
+ case OpLess32F:
+ return rewriteValueWasm_OpLess32F_0(v)
+ case OpLess32U:
+ return rewriteValueWasm_OpLess32U_0(v)
+ case OpLess64:
+ return rewriteValueWasm_OpLess64_0(v)
+ case OpLess64F:
+ return rewriteValueWasm_OpLess64F_0(v)
+ case OpLess64U:
+ return rewriteValueWasm_OpLess64U_0(v)
+ case OpLess8:
+ return rewriteValueWasm_OpLess8_0(v)
+ case OpLess8U:
+ return rewriteValueWasm_OpLess8U_0(v)
+ case OpLoad:
+ return rewriteValueWasm_OpLoad_0(v)
+ case OpLsh16x16:
+ return rewriteValueWasm_OpLsh16x16_0(v)
+ case OpLsh16x32:
+ return rewriteValueWasm_OpLsh16x32_0(v)
+ case OpLsh16x64:
+ return rewriteValueWasm_OpLsh16x64_0(v)
+ case OpLsh16x8:
+ return rewriteValueWasm_OpLsh16x8_0(v)
+ case OpLsh32x16:
+ return rewriteValueWasm_OpLsh32x16_0(v)
+ case OpLsh32x32:
+ return rewriteValueWasm_OpLsh32x32_0(v)
+ case OpLsh32x64:
+ return rewriteValueWasm_OpLsh32x64_0(v)
+ case OpLsh32x8:
+ return rewriteValueWasm_OpLsh32x8_0(v)
+ case OpLsh64x16:
+ return rewriteValueWasm_OpLsh64x16_0(v)
+ case OpLsh64x32:
+ return rewriteValueWasm_OpLsh64x32_0(v)
+ case OpLsh64x64:
+ return rewriteValueWasm_OpLsh64x64_0(v)
+ case OpLsh64x8:
+ return rewriteValueWasm_OpLsh64x8_0(v)
+ case OpLsh8x16:
+ return rewriteValueWasm_OpLsh8x16_0(v)
+ case OpLsh8x32:
+ return rewriteValueWasm_OpLsh8x32_0(v)
+ case OpLsh8x64:
+ return rewriteValueWasm_OpLsh8x64_0(v)
+ case OpLsh8x8:
+ return rewriteValueWasm_OpLsh8x8_0(v)
+ case OpMod16:
+ return rewriteValueWasm_OpMod16_0(v)
+ case OpMod16u:
+ return rewriteValueWasm_OpMod16u_0(v)
+ case OpMod32:
+ return rewriteValueWasm_OpMod32_0(v)
+ case OpMod32u:
+ return rewriteValueWasm_OpMod32u_0(v)
+ case OpMod64:
+ return rewriteValueWasm_OpMod64_0(v)
+ case OpMod64u:
+ return rewriteValueWasm_OpMod64u_0(v)
+ case OpMod8:
+ return rewriteValueWasm_OpMod8_0(v)
+ case OpMod8u:
+ return rewriteValueWasm_OpMod8u_0(v)
+ case OpMove:
+ return rewriteValueWasm_OpMove_0(v) || rewriteValueWasm_OpMove_10(v)
+ case OpMul16:
+ return rewriteValueWasm_OpMul16_0(v)
+ case OpMul32:
+ return rewriteValueWasm_OpMul32_0(v)
+ case OpMul32F:
+ return rewriteValueWasm_OpMul32F_0(v)
+ case OpMul64:
+ return rewriteValueWasm_OpMul64_0(v)
+ case OpMul64F:
+ return rewriteValueWasm_OpMul64F_0(v)
+ case OpMul8:
+ return rewriteValueWasm_OpMul8_0(v)
+ case OpNeg16:
+ return rewriteValueWasm_OpNeg16_0(v)
+ case OpNeg32:
+ return rewriteValueWasm_OpNeg32_0(v)
+ case OpNeg32F:
+ return rewriteValueWasm_OpNeg32F_0(v)
+ case OpNeg64:
+ return rewriteValueWasm_OpNeg64_0(v)
+ case OpNeg64F:
+ return rewriteValueWasm_OpNeg64F_0(v)
+ case OpNeg8:
+ return rewriteValueWasm_OpNeg8_0(v)
+ case OpNeq16:
+ return rewriteValueWasm_OpNeq16_0(v)
+ case OpNeq32:
+ return rewriteValueWasm_OpNeq32_0(v)
+ case OpNeq32F:
+ return rewriteValueWasm_OpNeq32F_0(v)
+ case OpNeq64:
+ return rewriteValueWasm_OpNeq64_0(v)
+ case OpNeq64F:
+ return rewriteValueWasm_OpNeq64F_0(v)
+ case OpNeq8:
+ return rewriteValueWasm_OpNeq8_0(v)
+ case OpNeqB:
+ return rewriteValueWasm_OpNeqB_0(v)
+ case OpNeqPtr:
+ return rewriteValueWasm_OpNeqPtr_0(v)
+ case OpNilCheck:
+ return rewriteValueWasm_OpNilCheck_0(v)
+ case OpNot:
+ return rewriteValueWasm_OpNot_0(v)
+ case OpOffPtr:
+ return rewriteValueWasm_OpOffPtr_0(v)
+ case OpOr16:
+ return rewriteValueWasm_OpOr16_0(v)
+ case OpOr32:
+ return rewriteValueWasm_OpOr32_0(v)
+ case OpOr64:
+ return rewriteValueWasm_OpOr64_0(v)
+ case OpOr8:
+ return rewriteValueWasm_OpOr8_0(v)
+ case OpOrB:
+ return rewriteValueWasm_OpOrB_0(v)
+ case OpRound32F:
+ return rewriteValueWasm_OpRound32F_0(v)
+ case OpRound64F:
+ return rewriteValueWasm_OpRound64F_0(v)
+ case OpRsh16Ux16:
+ return rewriteValueWasm_OpRsh16Ux16_0(v)
+ case OpRsh16Ux32:
+ return rewriteValueWasm_OpRsh16Ux32_0(v)
+ case OpRsh16Ux64:
+ return rewriteValueWasm_OpRsh16Ux64_0(v)
+ case OpRsh16Ux8:
+ return rewriteValueWasm_OpRsh16Ux8_0(v)
+ case OpRsh16x16:
+ return rewriteValueWasm_OpRsh16x16_0(v)
+ case OpRsh16x32:
+ return rewriteValueWasm_OpRsh16x32_0(v)
+ case OpRsh16x64:
+ return rewriteValueWasm_OpRsh16x64_0(v)
+ case OpRsh16x8:
+ return rewriteValueWasm_OpRsh16x8_0(v)
+ case OpRsh32Ux16:
+ return rewriteValueWasm_OpRsh32Ux16_0(v)
+ case OpRsh32Ux32:
+ return rewriteValueWasm_OpRsh32Ux32_0(v)
+ case OpRsh32Ux64:
+ return rewriteValueWasm_OpRsh32Ux64_0(v)
+ case OpRsh32Ux8:
+ return rewriteValueWasm_OpRsh32Ux8_0(v)
+ case OpRsh32x16:
+ return rewriteValueWasm_OpRsh32x16_0(v)
+ case OpRsh32x32:
+ return rewriteValueWasm_OpRsh32x32_0(v)
+ case OpRsh32x64:
+ return rewriteValueWasm_OpRsh32x64_0(v)
+ case OpRsh32x8:
+ return rewriteValueWasm_OpRsh32x8_0(v)
+ case OpRsh64Ux16:
+ return rewriteValueWasm_OpRsh64Ux16_0(v)
+ case OpRsh64Ux32:
+ return rewriteValueWasm_OpRsh64Ux32_0(v)
+ case OpRsh64Ux64:
+ return rewriteValueWasm_OpRsh64Ux64_0(v)
+ case OpRsh64Ux8:
+ return rewriteValueWasm_OpRsh64Ux8_0(v)
+ case OpRsh64x16:
+ return rewriteValueWasm_OpRsh64x16_0(v)
+ case OpRsh64x32:
+ return rewriteValueWasm_OpRsh64x32_0(v)
+ case OpRsh64x64:
+ return rewriteValueWasm_OpRsh64x64_0(v)
+ case OpRsh64x8:
+ return rewriteValueWasm_OpRsh64x8_0(v)
+ case OpRsh8Ux16:
+ return rewriteValueWasm_OpRsh8Ux16_0(v)
+ case OpRsh8Ux32:
+ return rewriteValueWasm_OpRsh8Ux32_0(v)
+ case OpRsh8Ux64:
+ return rewriteValueWasm_OpRsh8Ux64_0(v)
+ case OpRsh8Ux8:
+ return rewriteValueWasm_OpRsh8Ux8_0(v)
+ case OpRsh8x16:
+ return rewriteValueWasm_OpRsh8x16_0(v)
+ case OpRsh8x32:
+ return rewriteValueWasm_OpRsh8x32_0(v)
+ case OpRsh8x64:
+ return rewriteValueWasm_OpRsh8x64_0(v)
+ case OpRsh8x8:
+ return rewriteValueWasm_OpRsh8x8_0(v)
+ case OpSignExt16to32:
+ return rewriteValueWasm_OpSignExt16to32_0(v)
+ case OpSignExt16to64:
+ return rewriteValueWasm_OpSignExt16to64_0(v)
+ case OpSignExt32to64:
+ return rewriteValueWasm_OpSignExt32to64_0(v)
+ case OpSignExt8to16:
+ return rewriteValueWasm_OpSignExt8to16_0(v)
+ case OpSignExt8to32:
+ return rewriteValueWasm_OpSignExt8to32_0(v)
+ case OpSignExt8to64:
+ return rewriteValueWasm_OpSignExt8to64_0(v)
+ case OpSlicemask:
+ return rewriteValueWasm_OpSlicemask_0(v)
+ case OpStaticCall:
+ return rewriteValueWasm_OpStaticCall_0(v)
+ case OpStore:
+ return rewriteValueWasm_OpStore_0(v)
+ case OpSub16:
+ return rewriteValueWasm_OpSub16_0(v)
+ case OpSub32:
+ return rewriteValueWasm_OpSub32_0(v)
+ case OpSub32F:
+ return rewriteValueWasm_OpSub32F_0(v)
+ case OpSub64:
+ return rewriteValueWasm_OpSub64_0(v)
+ case OpSub64F:
+ return rewriteValueWasm_OpSub64F_0(v)
+ case OpSub8:
+ return rewriteValueWasm_OpSub8_0(v)
+ case OpSubPtr:
+ return rewriteValueWasm_OpSubPtr_0(v)
+ case OpTrunc16to8:
+ return rewriteValueWasm_OpTrunc16to8_0(v)
+ case OpTrunc32to16:
+ return rewriteValueWasm_OpTrunc32to16_0(v)
+ case OpTrunc32to8:
+ return rewriteValueWasm_OpTrunc32to8_0(v)
+ case OpTrunc64to16:
+ return rewriteValueWasm_OpTrunc64to16_0(v)
+ case OpTrunc64to32:
+ return rewriteValueWasm_OpTrunc64to32_0(v)
+ case OpTrunc64to8:
+ return rewriteValueWasm_OpTrunc64to8_0(v)
+ case OpWB:
+ return rewriteValueWasm_OpWB_0(v)
+ case OpWasmF64Add:
+ return rewriteValueWasm_OpWasmF64Add_0(v)
+ case OpWasmF64Mul:
+ return rewriteValueWasm_OpWasmF64Mul_0(v)
+ case OpWasmI64Add:
+ return rewriteValueWasm_OpWasmI64Add_0(v)
+ case OpWasmI64And:
+ return rewriteValueWasm_OpWasmI64And_0(v)
+ case OpWasmI64Eq:
+ return rewriteValueWasm_OpWasmI64Eq_0(v)
+ case OpWasmI64Eqz:
+ return rewriteValueWasm_OpWasmI64Eqz_0(v)
+ case OpWasmI64Load:
+ return rewriteValueWasm_OpWasmI64Load_0(v)
+ case OpWasmI64Load16S:
+ return rewriteValueWasm_OpWasmI64Load16S_0(v)
+ case OpWasmI64Load16U:
+ return rewriteValueWasm_OpWasmI64Load16U_0(v)
+ case OpWasmI64Load32S:
+ return rewriteValueWasm_OpWasmI64Load32S_0(v)
+ case OpWasmI64Load32U:
+ return rewriteValueWasm_OpWasmI64Load32U_0(v)
+ case OpWasmI64Load8S:
+ return rewriteValueWasm_OpWasmI64Load8S_0(v)
+ case OpWasmI64Load8U:
+ return rewriteValueWasm_OpWasmI64Load8U_0(v)
+ case OpWasmI64Mul:
+ return rewriteValueWasm_OpWasmI64Mul_0(v)
+ case OpWasmI64Ne:
+ return rewriteValueWasm_OpWasmI64Ne_0(v)
+ case OpWasmI64Or:
+ return rewriteValueWasm_OpWasmI64Or_0(v)
+ case OpWasmI64Store:
+ return rewriteValueWasm_OpWasmI64Store_0(v)
+ case OpWasmI64Store16:
+ return rewriteValueWasm_OpWasmI64Store16_0(v)
+ case OpWasmI64Store32:
+ return rewriteValueWasm_OpWasmI64Store32_0(v)
+ case OpWasmI64Store8:
+ return rewriteValueWasm_OpWasmI64Store8_0(v)
+ case OpWasmI64Xor:
+ return rewriteValueWasm_OpWasmI64Xor_0(v)
+ case OpXor16:
+ return rewriteValueWasm_OpXor16_0(v)
+ case OpXor32:
+ return rewriteValueWasm_OpXor32_0(v)
+ case OpXor64:
+ return rewriteValueWasm_OpXor64_0(v)
+ case OpXor8:
+ return rewriteValueWasm_OpXor8_0(v)
+ case OpZero:
+ return rewriteValueWasm_OpZero_0(v) || rewriteValueWasm_OpZero_10(v)
+ case OpZeroExt16to32:
+ return rewriteValueWasm_OpZeroExt16to32_0(v)
+ case OpZeroExt16to64:
+ return rewriteValueWasm_OpZeroExt16to64_0(v)
+ case OpZeroExt32to64:
+ return rewriteValueWasm_OpZeroExt32to64_0(v)
+ case OpZeroExt8to16:
+ return rewriteValueWasm_OpZeroExt8to16_0(v)
+ case OpZeroExt8to32:
+ return rewriteValueWasm_OpZeroExt8to32_0(v)
+ case OpZeroExt8to64:
+ return rewriteValueWasm_OpZeroExt8to64_0(v)
+ }
+ return false
+}
+func rewriteValueWasm_OpAdd16_0(v *Value) bool {
+ // match: (Add16 x y)
+ // cond:
+ // result: (I64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAdd32_0(v *Value) bool {
+ // match: (Add32 x y)
+ // cond:
+ // result: (I64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAdd32F_0(v *Value) bool {
+ // match: (Add32F x y)
+ // cond:
+ // result: (F64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAdd64_0(v *Value) bool {
+ // match: (Add64 x y)
+ // cond:
+ // result: (I64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAdd64F_0(v *Value) bool {
+ // match: (Add64F x y)
+ // cond:
+ // result: (F64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAdd8_0(v *Value) bool {
+ // match: (Add8 x y)
+ // cond:
+ // result: (I64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAddPtr_0(v *Value) bool {
+ // match: (AddPtr x y)
+ // cond:
+ // result: (I64Add x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAddr_0(v *Value) bool {
+ // match: (Addr {sym} base)
+ // cond:
+ // result: (LoweredAddr {sym} base)
+ for {
+ sym := v.Aux
+ base := v.Args[0]
+ v.reset(OpWasmLoweredAddr)
+ v.Aux = sym
+ v.AddArg(base)
+ return true
+ }
+}
+func rewriteValueWasm_OpAnd16_0(v *Value) bool {
+ // match: (And16 x y)
+ // cond:
+ // result: (I64And x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAnd32_0(v *Value) bool {
+ // match: (And32 x y)
+ // cond:
+ // result: (I64And x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAnd64_0(v *Value) bool {
+ // match: (And64 x y)
+ // cond:
+ // result: (I64And x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAnd8_0(v *Value) bool {
+ // match: (And8 x y)
+ // cond:
+ // result: (I64And x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpAndB_0(v *Value) bool {
+ // match: (AndB x y)
+ // cond:
+ // result: (I64And x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpClosureCall_0(v *Value) bool {
+ // match: (ClosureCall [argwid] entry closure mem)
+ // cond:
+ // result: (LoweredClosureCall [argwid] entry closure mem)
+ for {
+ argwid := v.AuxInt
+ _ = v.Args[2]
+ entry := v.Args[0]
+ closure := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmLoweredClosureCall)
+ v.AuxInt = argwid
+ v.AddArg(entry)
+ v.AddArg(closure)
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpCom16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Com16 x)
+ // cond:
+ // result: (I64Xor x (I64Const [-1]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = -1
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCom32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Com32 x)
+ // cond:
+ // result: (I64Xor x (I64Const [-1]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = -1
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCom64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Com64 x)
+ // cond:
+ // result: (I64Xor x (I64Const [-1]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = -1
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCom8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Com8 x)
+ // cond:
+ // result: (I64Xor x (I64Const [-1]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = -1
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpConst16_0(v *Value) bool {
+ // match: (Const16 [val])
+ // cond:
+ // result: (I64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConst32_0(v *Value) bool {
+ // match: (Const32 [val])
+ // cond:
+ // result: (I64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConst32F_0(v *Value) bool {
+ // match: (Const32F [val])
+ // cond:
+ // result: (F64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmF64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConst64_0(v *Value) bool {
+ // match: (Const64 [val])
+ // cond:
+ // result: (I64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConst64F_0(v *Value) bool {
+ // match: (Const64F [val])
+ // cond:
+ // result: (F64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmF64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConst8_0(v *Value) bool {
+ // match: (Const8 [val])
+ // cond:
+ // result: (I64Const [val])
+ for {
+ val := v.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = val
+ return true
+ }
+}
+func rewriteValueWasm_OpConstBool_0(v *Value) bool {
+ // match: (ConstBool [b])
+ // cond:
+ // result: (I64Const [b])
+ for {
+ b := v.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = b
+ return true
+ }
+}
+func rewriteValueWasm_OpConstNil_0(v *Value) bool {
+ // match: (ConstNil)
+ // cond:
+ // result: (I64Const [0])
+ for {
+ v.reset(OpWasmI64Const)
+ v.AuxInt = 0
+ return true
+ }
+}
+func rewriteValueWasm_OpConvert_0(v *Value) bool {
+ // match: (Convert <t> x mem)
+ // cond:
+ // result: (LoweredConvert <t> x mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ x := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmLoweredConvert)
+ v.Type = t
+ v.AddArg(x)
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Fto32_0(v *Value) bool {
+ // match: (Cvt32Fto32 x)
+ // cond:
+ // result: (I64TruncSF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncSF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Fto32U_0(v *Value) bool {
+ // match: (Cvt32Fto32U x)
+ // cond:
+ // result: (I64TruncUF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncUF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Fto64_0(v *Value) bool {
+ // match: (Cvt32Fto64 x)
+ // cond:
+ // result: (I64TruncSF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncSF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Fto64F_0(v *Value) bool {
+ // match: (Cvt32Fto64F x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Fto64U_0(v *Value) bool {
+ // match: (Cvt32Fto64U x)
+ // cond:
+ // result: (I64TruncUF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncUF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Uto32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt32Uto32F x)
+ // cond:
+ // result: (LoweredRound32F (F64ConvertUI64 (ZeroExt32to64 x)))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v0 := b.NewValue0(v.Pos, OpWasmF64ConvertUI64, typ.Float64)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(x)
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32Uto64F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt32Uto64F x)
+ // cond:
+ // result: (F64ConvertUI64 (ZeroExt32to64 x))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64ConvertUI64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32to32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt32to32F x)
+ // cond:
+ // result: (LoweredRound32F (F64ConvertSI64 (SignExt32to64 x)))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v0 := b.NewValue0(v.Pos, OpWasmF64ConvertSI64, typ.Float64)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(x)
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt32to64F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt32to64F x)
+ // cond:
+ // result: (F64ConvertSI64 (SignExt32to64 x))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64ConvertSI64)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Fto32_0(v *Value) bool {
+ // match: (Cvt64Fto32 x)
+ // cond:
+ // result: (I64TruncSF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncSF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Fto32F_0(v *Value) bool {
+ // match: (Cvt64Fto32F x)
+ // cond:
+ // result: (LoweredRound32F x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Fto32U_0(v *Value) bool {
+ // match: (Cvt64Fto32U x)
+ // cond:
+ // result: (I64TruncUF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncUF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Fto64_0(v *Value) bool {
+ // match: (Cvt64Fto64 x)
+ // cond:
+ // result: (I64TruncSF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncSF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Fto64U_0(v *Value) bool {
+ // match: (Cvt64Fto64U x)
+ // cond:
+ // result: (I64TruncUF64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64TruncUF64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Uto32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt64Uto32F x)
+ // cond:
+ // result: (LoweredRound32F (F64ConvertUI64 x))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v0 := b.NewValue0(v.Pos, OpWasmF64ConvertUI64, typ.Float64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64Uto64F_0(v *Value) bool {
+ // match: (Cvt64Uto64F x)
+ // cond:
+ // result: (F64ConvertUI64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64ConvertUI64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64to32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Cvt64to32F x)
+ // cond:
+ // result: (LoweredRound32F (F64ConvertSI64 x))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v0 := b.NewValue0(v.Pos, OpWasmF64ConvertSI64, typ.Float64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpCvt64to64F_0(v *Value) bool {
+ // match: (Cvt64to64F x)
+ // cond:
+ // result: (F64ConvertSI64 x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64ConvertSI64)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div16 x y)
+ // cond:
+ // result: (I64DivS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv16u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div16u x y)
+ // cond:
+ // result: (I64DivU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div32 x y)
+ // cond:
+ // result: (I64DivS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv32F_0(v *Value) bool {
+ // match: (Div32F x y)
+ // cond:
+ // result: (F64Div x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Div)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv32u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div32u x y)
+ // cond:
+ // result: (I64DivU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv64_0(v *Value) bool {
+ // match: (Div64 x y)
+ // cond:
+ // result: (I64DivS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv64F_0(v *Value) bool {
+ // match: (Div64F x y)
+ // cond:
+ // result: (F64Div x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Div)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv64u_0(v *Value) bool {
+ // match: (Div64u x y)
+ // cond:
+ // result: (I64DivU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div8 x y)
+ // cond:
+ // result: (I64DivS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpDiv8u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Div8u x y)
+ // cond:
+ // result: (I64DivU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64DivU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Eq16 x y)
+ // cond:
+ // result: (I64Eq (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Eq32 x y)
+ // cond:
+ // result: (I64Eq (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Eq32F x y)
+ // cond:
+ // result: (F64Eq (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Eq)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq64_0(v *Value) bool {
+ // match: (Eq64 x y)
+ // cond:
+ // result: (I64Eq x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq64F_0(v *Value) bool {
+ // match: (Eq64F x y)
+ // cond:
+ // result: (F64Eq x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Eq)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpEq8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Eq8 x y)
+ // cond:
+ // result: (I64Eq (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpEqB_0(v *Value) bool {
+ // match: (EqB x y)
+ // cond:
+ // result: (I64Eq x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpEqPtr_0(v *Value) bool {
+ // match: (EqPtr x y)
+ // cond:
+ // result: (I64Eq x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq16 x y)
+ // cond:
+ // result: (I64GeS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq16U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq16U x y)
+ // cond:
+ // result: (I64GeU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq32 x y)
+ // cond:
+ // result: (I64GeS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq32F x y)
+ // cond:
+ // result: (F64Ge (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Ge)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq32U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq32U x y)
+ // cond:
+ // result: (I64GeU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq64_0(v *Value) bool {
+ // match: (Geq64 x y)
+ // cond:
+ // result: (I64GeS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq64F_0(v *Value) bool {
+ // match: (Geq64F x y)
+ // cond:
+ // result: (F64Ge x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Ge)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq64U_0(v *Value) bool {
+ // match: (Geq64U x y)
+ // cond:
+ // result: (I64GeU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq8 x y)
+ // cond:
+ // result: (I64GeS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGeq8U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Geq8U x y)
+ // cond:
+ // result: (I64GeU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGetCallerPC_0(v *Value) bool {
+ // match: (GetCallerPC)
+ // cond:
+ // result: (LoweredGetCallerPC)
+ for {
+ v.reset(OpWasmLoweredGetCallerPC)
+ return true
+ }
+}
+func rewriteValueWasm_OpGetCallerSP_0(v *Value) bool {
+ // match: (GetCallerSP)
+ // cond:
+ // result: (LoweredGetCallerSP)
+ for {
+ v.reset(OpWasmLoweredGetCallerSP)
+ return true
+ }
+}
+func rewriteValueWasm_OpGetClosurePtr_0(v *Value) bool {
+ // match: (GetClosurePtr)
+ // cond:
+ // result: (LoweredGetClosurePtr)
+ for {
+ v.reset(OpWasmLoweredGetClosurePtr)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater16 x y)
+ // cond:
+ // result: (I64GtS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater16U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater16U x y)
+ // cond:
+ // result: (I64GtU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater32 x y)
+ // cond:
+ // result: (I64GtS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater32F x y)
+ // cond:
+ // result: (F64Gt (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Gt)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater32U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater32U x y)
+ // cond:
+ // result: (I64GtU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater64_0(v *Value) bool {
+ // match: (Greater64 x y)
+ // cond:
+ // result: (I64GtS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater64F_0(v *Value) bool {
+ // match: (Greater64F x y)
+ // cond:
+ // result: (F64Gt x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Gt)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater64U_0(v *Value) bool {
+ // match: (Greater64U x y)
+ // cond:
+ // result: (I64GtU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater8 x y)
+ // cond:
+ // result: (I64GtS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpGreater8U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Greater8U x y)
+ // cond:
+ // result: (I64GtU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64GtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpInterCall_0(v *Value) bool {
+ // match: (InterCall [argwid] entry mem)
+ // cond:
+ // result: (LoweredInterCall [argwid] entry mem)
+ for {
+ argwid := v.AuxInt
+ _ = v.Args[1]
+ entry := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmLoweredInterCall)
+ v.AuxInt = argwid
+ v.AddArg(entry)
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpIsInBounds_0(v *Value) bool {
+ // match: (IsInBounds idx len)
+ // cond:
+ // result: (I64LtU idx len)
+ for {
+ _ = v.Args[1]
+ idx := v.Args[0]
+ len := v.Args[1]
+ v.reset(OpWasmI64LtU)
+ v.AddArg(idx)
+ v.AddArg(len)
+ return true
+ }
+}
+func rewriteValueWasm_OpIsNonNil_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (IsNonNil p)
+ // cond:
+ // result: (I64Eqz (I64Eqz p))
+ for {
+ p := v.Args[0]
+ v.reset(OpWasmI64Eqz)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Eqz, typ.Bool)
+ v0.AddArg(p)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpIsSliceInBounds_0(v *Value) bool {
+ // match: (IsSliceInBounds idx len)
+ // cond:
+ // result: (I64LeU idx len)
+ for {
+ _ = v.Args[1]
+ idx := v.Args[0]
+ len := v.Args[1]
+ v.reset(OpWasmI64LeU)
+ v.AddArg(idx)
+ v.AddArg(len)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq16 x y)
+ // cond:
+ // result: (I64LeS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq16U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq16U x y)
+ // cond:
+ // result: (I64LeU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq32 x y)
+ // cond:
+ // result: (I64LeS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq32F x y)
+ // cond:
+ // result: (F64Le (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Le)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq32U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq32U x y)
+ // cond:
+ // result: (I64LeU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq64_0(v *Value) bool {
+ // match: (Leq64 x y)
+ // cond:
+ // result: (I64LeS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq64F_0(v *Value) bool {
+ // match: (Leq64F x y)
+ // cond:
+ // result: (F64Le x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Le)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq64U_0(v *Value) bool {
+ // match: (Leq64U x y)
+ // cond:
+ // result: (I64LeU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq8 x y)
+ // cond:
+ // result: (I64LeS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLeq8U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Leq8U x y)
+ // cond:
+ // result: (I64LeU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LeU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less16 x y)
+ // cond:
+ // result: (I64LtS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess16U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less16U x y)
+ // cond:
+ // result: (I64LtU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less32 x y)
+ // cond:
+ // result: (I64LtS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less32F x y)
+ // cond:
+ // result: (F64Lt (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Lt)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess32U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less32U x y)
+ // cond:
+ // result: (I64LtU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess64_0(v *Value) bool {
+ // match: (Less64 x y)
+ // cond:
+ // result: (I64LtS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess64F_0(v *Value) bool {
+ // match: (Less64F x y)
+ // cond:
+ // result: (F64Lt x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Lt)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess64U_0(v *Value) bool {
+ // match: (Less64U x y)
+ // cond:
+ // result: (I64LtU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less8 x y)
+ // cond:
+ // result: (I64LtS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLess8U_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Less8U x y)
+ // cond:
+ // result: (I64LtU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64LtU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpLoad_0(v *Value) bool {
+ // match: (Load <t> ptr mem)
+ // cond: is32BitFloat(t)
+ // result: (F32Load ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(is32BitFloat(t)) {
+ break
+ }
+ v.reset(OpWasmF32Load)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: is64BitFloat(t)
+ // result: (F64Load ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(is64BitFloat(t)) {
+ break
+ }
+ v.reset(OpWasmF64Load)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 8
+ // result: (I64Load ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 8) {
+ break
+ }
+ v.reset(OpWasmI64Load)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 4 && !t.IsSigned()
+ // result: (I64Load32U ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 4 && !t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load32U)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 4 && t.IsSigned()
+ // result: (I64Load32S ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 4 && t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load32S)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 2 && !t.IsSigned()
+ // result: (I64Load16U ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 2 && !t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load16U)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 2 && t.IsSigned()
+ // result: (I64Load16S ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 2 && t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load16S)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 1 && !t.IsSigned()
+ // result: (I64Load8U ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 1 && !t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load8U)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Load <t> ptr mem)
+ // cond: t.Size() == 1 && t.IsSigned()
+ // result: (I64Load8S ptr mem)
+ for {
+ t := v.Type
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ if !(t.Size() == 1 && t.IsSigned()) {
+ break
+ }
+ v.reset(OpWasmI64Load8S)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpLsh16x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh16x16 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh16x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh16x32 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh16x64_0(v *Value) bool {
+ // match: (Lsh16x64 x y)
+ // cond:
+ // result: (Lsh64x64 x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh16x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh16x8 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh32x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh32x16 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh32x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh32x32 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh32x64_0(v *Value) bool {
+ // match: (Lsh32x64 x y)
+ // cond:
+ // result: (Lsh64x64 x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh32x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh32x8 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh64x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh64x16 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh64x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh64x32 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh64x64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh64x64 x y)
+ // cond:
+ // result: (Select (I64Shl x y) (I64Const [0]) (I64LtU y (I64Const [64])))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmSelect)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 0
+ v.AddArg(v1)
+ v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool)
+ v2.AddArg(y)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v3.AuxInt = 64
+ v2.AddArg(v3)
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh64x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh64x8 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh8x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh8x16 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh8x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh8x32 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh8x64_0(v *Value) bool {
+ // match: (Lsh8x64 x y)
+ // cond:
+ // result: (Lsh64x64 x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpLsh8x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Lsh8x8 x y)
+ // cond:
+ // result: (Lsh64x64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpLsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod16 x y)
+ // cond:
+ // result: (I64RemS (SignExt16to64 x) (SignExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemS)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod16u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod16u x y)
+ // cond:
+ // result: (I64RemU (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod32 x y)
+ // cond:
+ // result: (I64RemS (SignExt32to64 x) (SignExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemS)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod32u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod32u x y)
+ // cond:
+ // result: (I64RemU (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod64_0(v *Value) bool {
+ // match: (Mod64 x y)
+ // cond:
+ // result: (I64RemS x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemS)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod64u_0(v *Value) bool {
+ // match: (Mod64u x y)
+ // cond:
+ // result: (I64RemU x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemU)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod8 x y)
+ // cond:
+ // result: (I64RemS (SignExt8to64 x) (SignExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemS)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMod8u_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Mod8u x y)
+ // cond:
+ // result: (I64RemU (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64RemU)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpMove_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Move [0] _ _ mem)
+ // cond:
+ // result: mem
+ for {
+ if v.AuxInt != 0 {
+ break
+ }
+ _ = v.Args[2]
+ mem := v.Args[2]
+ v.reset(OpCopy)
+ v.Type = mem.Type
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Move [1] dst src mem)
+ // cond:
+ // result: (I64Store8 dst (I64Load8U src mem) mem)
+ for {
+ if v.AuxInt != 1 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store8)
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8)
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Move [2] dst src mem)
+ // cond:
+ // result: (I64Store16 dst (I64Load16U src mem) mem)
+ for {
+ if v.AuxInt != 2 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store16)
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16)
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Move [4] dst src mem)
+ // cond:
+ // result: (I64Store32 dst (I64Load32U src mem) mem)
+ for {
+ if v.AuxInt != 4 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store32)
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32)
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Move [8] dst src mem)
+ // cond:
+ // result: (I64Store dst (I64Load src mem) mem)
+ for {
+ if v.AuxInt != 8 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store)
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Move [16] dst src mem)
+ // cond:
+ // result: (I64Store [8] dst (I64Load [8] src mem) (I64Store dst (I64Load src mem) mem))
+ for {
+ if v.AuxInt != 16 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store)
+ v.AuxInt = 8
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v0.AuxInt = 8
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Move [3] dst src mem)
+ // cond:
+ // result: (I64Store8 [2] dst (I64Load8U [2] src mem) (I64Store16 dst (I64Load16U src mem) mem))
+ for {
+ if v.AuxInt != 3 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store8)
+ v.AuxInt = 2
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8)
+ v0.AuxInt = 2
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store16, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Move [5] dst src mem)
+ // cond:
+ // result: (I64Store8 [4] dst (I64Load8U [4] src mem) (I64Store32 dst (I64Load32U src mem) mem))
+ for {
+ if v.AuxInt != 5 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store8)
+ v.AuxInt = 4
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load8U, typ.UInt8)
+ v0.AuxInt = 4
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Move [6] dst src mem)
+ // cond:
+ // result: (I64Store16 [4] dst (I64Load16U [4] src mem) (I64Store32 dst (I64Load32U src mem) mem))
+ for {
+ if v.AuxInt != 6 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store16)
+ v.AuxInt = 4
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load16U, typ.UInt16)
+ v0.AuxInt = 4
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Move [7] dst src mem)
+ // cond:
+ // result: (I64Store32 [3] dst (I64Load32U [3] src mem) (I64Store32 dst (I64Load32U src mem) mem))
+ for {
+ if v.AuxInt != 7 {
+ break
+ }
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmI64Store32)
+ v.AuxInt = 3
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32)
+ v0.AuxInt = 3
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load32U, typ.UInt32)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpMove_10(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Move [s] dst src mem)
+ // cond: s > 8 && s < 16
+ // result: (I64Store [s-8] dst (I64Load [s-8] src mem) (I64Store dst (I64Load src mem) mem))
+ for {
+ s := v.AuxInt
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ if !(s > 8 && s < 16) {
+ break
+ }
+ v.reset(OpWasmI64Store)
+ v.AuxInt = s - 8
+ v.AddArg(dst)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v0.AuxInt = s - 8
+ v0.AddArg(src)
+ v0.AddArg(mem)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AddArg(dst)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v2.AddArg(src)
+ v2.AddArg(mem)
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Move [s] dst src mem)
+ // cond: s > 16 && s%16 != 0 && s%16 <= 8
+ // result: (Move [s-s%16] (OffPtr <dst.Type> dst [s%16]) (OffPtr <src.Type> src [s%16]) (I64Store dst (I64Load src mem) mem))
+ for {
+ s := v.AuxInt
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ if !(s > 16 && s%16 != 0 && s%16 <= 8) {
+ break
+ }
+ v.reset(OpMove)
+ v.AuxInt = s - s%16
+ v0 := b.NewValue0(v.Pos, OpOffPtr, dst.Type)
+ v0.AuxInt = s % 16
+ v0.AddArg(dst)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpOffPtr, src.Type)
+ v1.AuxInt = s % 16
+ v1.AddArg(src)
+ v.AddArg(v1)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v2.AddArg(dst)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v3.AddArg(src)
+ v3.AddArg(mem)
+ v2.AddArg(v3)
+ v2.AddArg(mem)
+ v.AddArg(v2)
+ return true
+ }
+ // match: (Move [s] dst src mem)
+ // cond: s > 16 && s%16 != 0 && s%16 > 8
+ // result: (Move [s-s%16] (OffPtr <dst.Type> dst [s%16]) (OffPtr <src.Type> src [s%16]) (I64Store [8] dst (I64Load [8] src mem) (I64Store dst (I64Load src mem) mem)))
+ for {
+ s := v.AuxInt
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ if !(s > 16 && s%16 != 0 && s%16 > 8) {
+ break
+ }
+ v.reset(OpMove)
+ v.AuxInt = s - s%16
+ v0 := b.NewValue0(v.Pos, OpOffPtr, dst.Type)
+ v0.AuxInt = s % 16
+ v0.AddArg(dst)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpOffPtr, src.Type)
+ v1.AuxInt = s % 16
+ v1.AddArg(src)
+ v.AddArg(v1)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v2.AuxInt = 8
+ v2.AddArg(dst)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v3.AuxInt = 8
+ v3.AddArg(src)
+ v3.AddArg(mem)
+ v2.AddArg(v3)
+ v4 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v4.AddArg(dst)
+ v5 := b.NewValue0(v.Pos, OpWasmI64Load, typ.UInt64)
+ v5.AddArg(src)
+ v5.AddArg(mem)
+ v4.AddArg(v5)
+ v4.AddArg(mem)
+ v2.AddArg(v4)
+ v.AddArg(v2)
+ return true
+ }
+ // match: (Move [s] dst src mem)
+ // cond: s%8 == 0
+ // result: (LoweredMove [s/8] dst src mem)
+ for {
+ s := v.AuxInt
+ _ = v.Args[2]
+ dst := v.Args[0]
+ src := v.Args[1]
+ mem := v.Args[2]
+ if !(s%8 == 0) {
+ break
+ }
+ v.reset(OpWasmLoweredMove)
+ v.AuxInt = s / 8
+ v.AddArg(dst)
+ v.AddArg(src)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpMul16_0(v *Value) bool {
+ // match: (Mul16 x y)
+ // cond:
+ // result: (I64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMul32_0(v *Value) bool {
+ // match: (Mul32 x y)
+ // cond:
+ // result: (I64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMul32F_0(v *Value) bool {
+ // match: (Mul32F x y)
+ // cond:
+ // result: (F64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMul64_0(v *Value) bool {
+ // match: (Mul64 x y)
+ // cond:
+ // result: (I64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMul64F_0(v *Value) bool {
+ // match: (Mul64F x y)
+ // cond:
+ // result: (F64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpMul8_0(v *Value) bool {
+ // match: (Mul8 x y)
+ // cond:
+ // result: (I64Mul x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Mul)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neg16 x)
+ // cond:
+ // result: (I64Sub (I64Const [0]) x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Sub)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neg32 x)
+ // cond:
+ // result: (I64Sub (I64Const [0]) x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Sub)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg32F_0(v *Value) bool {
+ // match: (Neg32F x)
+ // cond:
+ // result: (F64Neg x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64Neg)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neg64 x)
+ // cond:
+ // result: (I64Sub (I64Const [0]) x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Sub)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg64F_0(v *Value) bool {
+ // match: (Neg64F x)
+ // cond:
+ // result: (F64Neg x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmF64Neg)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeg8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neg8 x)
+ // cond:
+ // result: (I64Sub (I64Const [0]) x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Sub)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neq16 x y)
+ // cond:
+ // result: (I64Ne (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neq32 x y)
+ // cond:
+ // result: (I64Ne (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq32F_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neq32F x y)
+ // cond:
+ // result: (F64Ne (LoweredRound32F x) (LoweredRound32F y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Ne)
+ v0 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmLoweredRound32F, typ.Float32)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq64_0(v *Value) bool {
+ // match: (Neq64 x y)
+ // cond:
+ // result: (I64Ne x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq64F_0(v *Value) bool {
+ // match: (Neq64F x y)
+ // cond:
+ // result: (F64Ne x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Ne)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeq8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Neq8 x y)
+ // cond:
+ // result: (I64Ne (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeqB_0(v *Value) bool {
+ // match: (NeqB x y)
+ // cond:
+ // result: (I64Ne x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpNeqPtr_0(v *Value) bool {
+ // match: (NeqPtr x y)
+ // cond:
+ // result: (I64Ne x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpNilCheck_0(v *Value) bool {
+ // match: (NilCheck ptr mem)
+ // cond:
+ // result: (LoweredNilCheck ptr mem)
+ for {
+ _ = v.Args[1]
+ ptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmLoweredNilCheck)
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpNot_0(v *Value) bool {
+ // match: (Not x)
+ // cond:
+ // result: (I64Eqz x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64Eqz)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpOffPtr_0(v *Value) bool {
+ // match: (OffPtr [0] ptr)
+ // cond:
+ // result: ptr
+ for {
+ if v.AuxInt != 0 {
+ break
+ }
+ ptr := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = ptr.Type
+ v.AddArg(ptr)
+ return true
+ }
+ // match: (OffPtr [off] ptr)
+ // cond: off > 0
+ // result: (I64AddConst [off] ptr)
+ for {
+ off := v.AuxInt
+ ptr := v.Args[0]
+ if !(off > 0) {
+ break
+ }
+ v.reset(OpWasmI64AddConst)
+ v.AuxInt = off
+ v.AddArg(ptr)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpOr16_0(v *Value) bool {
+ // match: (Or16 x y)
+ // cond:
+ // result: (I64Or x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpOr32_0(v *Value) bool {
+ // match: (Or32 x y)
+ // cond:
+ // result: (I64Or x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpOr64_0(v *Value) bool {
+ // match: (Or64 x y)
+ // cond:
+ // result: (I64Or x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpOr8_0(v *Value) bool {
+ // match: (Or8 x y)
+ // cond:
+ // result: (I64Or x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpOrB_0(v *Value) bool {
+ // match: (OrB x y)
+ // cond:
+ // result: (I64Or x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRound32F_0(v *Value) bool {
+ // match: (Round32F x)
+ // cond:
+ // result: (LoweredRound32F x)
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmLoweredRound32F)
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpRound64F_0(v *Value) bool {
+ // match: (Round64F x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16Ux16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16Ux16 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16Ux32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16Ux32 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt16to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16Ux64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16Ux64 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt16to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16Ux8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16Ux8 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt16to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16x16 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt16to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16x32 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt16to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16x64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16x64 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt16to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh16x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh16x8 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt16to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt16to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32Ux16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32Ux16 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt32to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32Ux32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32Ux32 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32Ux64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32Ux64 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt32to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32Ux8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32Ux8 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt32to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32x16 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt32to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32x32 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt32to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32x64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32x64 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt32to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh32x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh32x8 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt32to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt32to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64Ux16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64Ux16 x y)
+ // cond:
+ // result: (Rsh64Ux64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64Ux32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64Ux32 x y)
+ // cond:
+ // result: (Rsh64Ux64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64Ux64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64Ux64 x y)
+ // cond:
+ // result: (Select (I64ShrU x y) (I64Const [0]) (I64LtU y (I64Const [64])))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmSelect)
+ v0 := b.NewValue0(v.Pos, OpWasmI64ShrU, typ.Int64)
+ v0.AddArg(x)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 0
+ v.AddArg(v1)
+ v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool)
+ v2.AddArg(y)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v3.AuxInt = 64
+ v2.AddArg(v3)
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64Ux8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64Ux8 x y)
+ // cond:
+ // result: (Rsh64Ux64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64x16 x y)
+ // cond:
+ // result: (Rsh64x64 x (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64x32 x y)
+ // cond:
+ // result: (Rsh64x64 x (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64x64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64x64 x y)
+ // cond:
+ // result: (I64ShrS x (Select <typ.Int64> y (I64Const [63]) (I64LtU y (I64Const [64]))))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64ShrS)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpWasmSelect, typ.Int64)
+ v0.AddArg(y)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 63
+ v0.AddArg(v1)
+ v2 := b.NewValue0(v.Pos, OpWasmI64LtU, typ.Bool)
+ v2.AddArg(y)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v3.AuxInt = 64
+ v2.AddArg(v3)
+ v0.AddArg(v2)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh64x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh64x8 x y)
+ // cond:
+ // result: (Rsh64x64 x (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v.AddArg(x)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(y)
+ v.AddArg(v0)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8Ux16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8Ux16 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt8to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8Ux32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8Ux32 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt8to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8Ux64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8Ux64 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt8to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8Ux8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8Ux8 x y)
+ // cond:
+ // result: (Rsh64Ux64 (ZeroExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64Ux64)
+ v0 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8x16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8x16 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt8to64 x) (ZeroExt16to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt16to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8x32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8x32 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt8to64 x) (ZeroExt32to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt32to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8x64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8x64 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt8to64 x) y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpRsh8x8_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Rsh8x8 x y)
+ // cond:
+ // result: (Rsh64x64 (SignExt8to64 x) (ZeroExt8to64 y))
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpRsh64x64)
+ v0 := b.NewValue0(v.Pos, OpSignExt8to64, typ.Int64)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpZeroExt8to64, typ.UInt64)
+ v1.AddArg(y)
+ v.AddArg(v1)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt16to32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt16to32 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [48])) (I64Const [48]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 48
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 48
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt16to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt16to64 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [48])) (I64Const [48]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 48
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 48
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt32to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt32to64 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [32])) (I64Const [32]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 32
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 32
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt8to16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt8to16 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt8to32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt8to32 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSignExt8to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (SignExt8to64 x)
+ // cond:
+ // result: (I64ShrS (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpSlicemask_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Slicemask x)
+ // cond:
+ // result: (I64ShrS (I64Sub (I64Const [0]) x) (I64Const [63]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrS)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Sub, typ.Int64)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 0
+ v0.AddArg(v1)
+ v0.AddArg(x)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 63
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpStaticCall_0(v *Value) bool {
+ // match: (StaticCall [argwid] {target} mem)
+ // cond:
+ // result: (LoweredStaticCall [argwid] {target} mem)
+ for {
+ argwid := v.AuxInt
+ target := v.Aux
+ mem := v.Args[0]
+ v.reset(OpWasmLoweredStaticCall)
+ v.AuxInt = argwid
+ v.Aux = target
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpStore_0(v *Value) bool {
+ // match: (Store {t} ptr val mem)
+ // cond: is64BitFloat(t.(*types.Type))
+ // result: (F64Store ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(is64BitFloat(t.(*types.Type))) {
+ break
+ }
+ v.reset(OpWasmF64Store)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Store {t} ptr val mem)
+ // cond: is32BitFloat(t.(*types.Type))
+ // result: (F32Store ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(is32BitFloat(t.(*types.Type))) {
+ break
+ }
+ v.reset(OpWasmF32Store)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Store {t} ptr val mem)
+ // cond: t.(*types.Type).Size() == 8
+ // result: (I64Store ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(t.(*types.Type).Size() == 8) {
+ break
+ }
+ v.reset(OpWasmI64Store)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Store {t} ptr val mem)
+ // cond: t.(*types.Type).Size() == 4
+ // result: (I64Store32 ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(t.(*types.Type).Size() == 4) {
+ break
+ }
+ v.reset(OpWasmI64Store32)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Store {t} ptr val mem)
+ // cond: t.(*types.Type).Size() == 2
+ // result: (I64Store16 ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(t.(*types.Type).Size() == 2) {
+ break
+ }
+ v.reset(OpWasmI64Store16)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Store {t} ptr val mem)
+ // cond: t.(*types.Type).Size() == 1
+ // result: (I64Store8 ptr val mem)
+ for {
+ t := v.Aux
+ _ = v.Args[2]
+ ptr := v.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(t.(*types.Type).Size() == 1) {
+ break
+ }
+ v.reset(OpWasmI64Store8)
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpSub16_0(v *Value) bool {
+ // match: (Sub16 x y)
+ // cond:
+ // result: (I64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSub32_0(v *Value) bool {
+ // match: (Sub32 x y)
+ // cond:
+ // result: (I64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSub32F_0(v *Value) bool {
+ // match: (Sub32F x y)
+ // cond:
+ // result: (F64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSub64_0(v *Value) bool {
+ // match: (Sub64 x y)
+ // cond:
+ // result: (I64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSub64F_0(v *Value) bool {
+ // match: (Sub64F x y)
+ // cond:
+ // result: (F64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmF64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSub8_0(v *Value) bool {
+ // match: (Sub8 x y)
+ // cond:
+ // result: (I64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpSubPtr_0(v *Value) bool {
+ // match: (SubPtr x y)
+ // cond:
+ // result: (I64Sub x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Sub)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc16to8_0(v *Value) bool {
+ // match: (Trunc16to8 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc32to16_0(v *Value) bool {
+ // match: (Trunc32to16 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc32to8_0(v *Value) bool {
+ // match: (Trunc32to8 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc64to16_0(v *Value) bool {
+ // match: (Trunc64to16 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc64to32_0(v *Value) bool {
+ // match: (Trunc64to32 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpTrunc64to8_0(v *Value) bool {
+ // match: (Trunc64to8 x)
+ // cond:
+ // result: x
+ for {
+ x := v.Args[0]
+ v.reset(OpCopy)
+ v.Type = x.Type
+ v.AddArg(x)
+ return true
+ }
+}
+func rewriteValueWasm_OpWB_0(v *Value) bool {
+ // match: (WB {fn} destptr srcptr mem)
+ // cond:
+ // result: (LoweredWB {fn} destptr srcptr mem)
+ for {
+ fn := v.Aux
+ _ = v.Args[2]
+ destptr := v.Args[0]
+ srcptr := v.Args[1]
+ mem := v.Args[2]
+ v.reset(OpWasmLoweredWB)
+ v.Aux = fn
+ v.AddArg(destptr)
+ v.AddArg(srcptr)
+ v.AddArg(mem)
+ return true
+ }
+}
+func rewriteValueWasm_OpWasmF64Add_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (F64Add (F64Const [x]) (F64Const [y]))
+ // cond:
+ // result: (F64Const [f2i(i2f(x) + i2f(y))])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmF64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmF64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmF64Const)
+ v.AuxInt = f2i(i2f(x) + i2f(y))
+ return true
+ }
+ // match: (F64Add (F64Const [x]) y)
+ // cond:
+ // result: (F64Add y (F64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmF64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmF64Add)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmF64Const, typ.Float64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmF64Mul_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (F64Mul (F64Const [x]) (F64Const [y]))
+ // cond:
+ // result: (F64Const [f2i(i2f(x) * i2f(y))])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmF64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmF64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmF64Const)
+ v.AuxInt = f2i(i2f(x) * i2f(y))
+ return true
+ }
+ // match: (F64Mul (F64Const [x]) y)
+ // cond:
+ // result: (F64Mul y (F64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmF64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmF64Mul)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmF64Const, typ.Float64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Add_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Add (I64Const [x]) (I64Const [y]))
+ // cond:
+ // result: (I64Const [x + y])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = x + y
+ return true
+ }
+ // match: (I64Add (I64Const [x]) y)
+ // cond:
+ // result: (I64Add y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Add)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ // match: (I64Add x (I64Const [y]))
+ // cond:
+ // result: (I64AddConst [y] x)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64AddConst)
+ v.AuxInt = y
+ v.AddArg(x)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64And_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64And (I64Const [x]) (I64Const [y]))
+ // cond:
+ // result: (I64Const [x & y])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = x & y
+ return true
+ }
+ // match: (I64And (I64Const [x]) y)
+ // cond:
+ // result: (I64And y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64And)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Eq_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Eq (I64Const [x]) (I64Const [y]))
+ // cond: x == y
+ // result: (I64Const [1])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ if !(x == y) {
+ break
+ }
+ v.reset(OpWasmI64Const)
+ v.AuxInt = 1
+ return true
+ }
+ // match: (I64Eq (I64Const [x]) (I64Const [y]))
+ // cond: x != y
+ // result: (I64Const [0])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ if !(x != y) {
+ break
+ }
+ v.reset(OpWasmI64Const)
+ v.AuxInt = 0
+ return true
+ }
+ // match: (I64Eq (I64Const [x]) y)
+ // cond:
+ // result: (I64Eq y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Eq)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Eqz_0(v *Value) bool {
+ // match: (I64Eqz (I64Eqz (I64Eqz x)))
+ // cond:
+ // result: (I64Eqz x)
+ for {
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Eqz {
+ break
+ }
+ v_0_0 := v_0.Args[0]
+ if v_0_0.Op != OpWasmI64Eqz {
+ break
+ }
+ x := v_0_0.Args[0]
+ v.reset(OpWasmI64Eqz)
+ v.AddArg(x)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load_0(v *Value) bool {
+ // match: (I64Load [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load16S_0(v *Value) bool {
+ // match: (I64Load16S [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load16S [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load16S)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load16U_0(v *Value) bool {
+ // match: (I64Load16U [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load16U [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load16U)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load32S_0(v *Value) bool {
+ // match: (I64Load32S [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load32S [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load32S)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load32U_0(v *Value) bool {
+ // match: (I64Load32U [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load32U [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load32U)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load8S_0(v *Value) bool {
+ // match: (I64Load8S [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load8S [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load8S)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Load8U_0(v *Value) bool {
+ // match: (I64Load8U [off] (I64AddConst [off2] ptr) mem)
+ // cond: off+off2 >= 0
+ // result: (I64Load8U [off+off2] ptr mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ mem := v.Args[1]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Load8U)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Mul_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Mul (I64Const [x]) (I64Const [y]))
+ // cond:
+ // result: (I64Const [x * y])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = x * y
+ return true
+ }
+ // match: (I64Mul (I64Const [x]) y)
+ // cond:
+ // result: (I64Mul y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Mul)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Ne_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Ne (I64Const [x]) (I64Const [y]))
+ // cond: x == y
+ // result: (I64Const [0])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ if !(x == y) {
+ break
+ }
+ v.reset(OpWasmI64Const)
+ v.AuxInt = 0
+ return true
+ }
+ // match: (I64Ne (I64Const [x]) (I64Const [y]))
+ // cond: x != y
+ // result: (I64Const [1])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ if !(x != y) {
+ break
+ }
+ v.reset(OpWasmI64Const)
+ v.AuxInt = 1
+ return true
+ }
+ // match: (I64Ne (I64Const [x]) y)
+ // cond:
+ // result: (I64Ne y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Ne)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Or_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Or (I64Const [x]) (I64Const [y]))
+ // cond:
+ // result: (I64Const [x | y])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = x | y
+ return true
+ }
+ // match: (I64Or (I64Const [x]) y)
+ // cond:
+ // result: (I64Or y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Or)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Store_0(v *Value) bool {
+ // match: (I64Store [off] (I64AddConst [off2] ptr) val mem)
+ // cond: off+off2 >= 0
+ // result: (I64Store [off+off2] ptr val mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[2]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Store)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Store16_0(v *Value) bool {
+ // match: (I64Store16 [off] (I64AddConst [off2] ptr) val mem)
+ // cond: off+off2 >= 0
+ // result: (I64Store16 [off+off2] ptr val mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[2]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Store16)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Store32_0(v *Value) bool {
+ // match: (I64Store32 [off] (I64AddConst [off2] ptr) val mem)
+ // cond: off+off2 >= 0
+ // result: (I64Store32 [off+off2] ptr val mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[2]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Store32)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Store8_0(v *Value) bool {
+ // match: (I64Store8 [off] (I64AddConst [off2] ptr) val mem)
+ // cond: off+off2 >= 0
+ // result: (I64Store8 [off+off2] ptr val mem)
+ for {
+ off := v.AuxInt
+ _ = v.Args[2]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64AddConst {
+ break
+ }
+ off2 := v_0.AuxInt
+ ptr := v_0.Args[0]
+ val := v.Args[1]
+ mem := v.Args[2]
+ if !(off+off2 >= 0) {
+ break
+ }
+ v.reset(OpWasmI64Store8)
+ v.AuxInt = off + off2
+ v.AddArg(ptr)
+ v.AddArg(val)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpWasmI64Xor_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (I64Xor (I64Const [x]) (I64Const [y]))
+ // cond:
+ // result: (I64Const [x ^ y])
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ v_1 := v.Args[1]
+ if v_1.Op != OpWasmI64Const {
+ break
+ }
+ y := v_1.AuxInt
+ v.reset(OpWasmI64Const)
+ v.AuxInt = x ^ y
+ return true
+ }
+ // match: (I64Xor (I64Const [x]) y)
+ // cond:
+ // result: (I64Xor y (I64Const [x]))
+ for {
+ _ = v.Args[1]
+ v_0 := v.Args[0]
+ if v_0.Op != OpWasmI64Const {
+ break
+ }
+ x := v_0.AuxInt
+ y := v.Args[1]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(y)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = x
+ v.AddArg(v0)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpXor16_0(v *Value) bool {
+ // match: (Xor16 x y)
+ // cond:
+ // result: (I64Xor x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpXor32_0(v *Value) bool {
+ // match: (Xor32 x y)
+ // cond:
+ // result: (I64Xor x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpXor64_0(v *Value) bool {
+ // match: (Xor64 x y)
+ // cond:
+ // result: (I64Xor x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpXor8_0(v *Value) bool {
+ // match: (Xor8 x y)
+ // cond:
+ // result: (I64Xor x y)
+ for {
+ _ = v.Args[1]
+ x := v.Args[0]
+ y := v.Args[1]
+ v.reset(OpWasmI64Xor)
+ v.AddArg(x)
+ v.AddArg(y)
+ return true
+ }
+}
+func rewriteValueWasm_OpZero_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Zero [0] _ mem)
+ // cond:
+ // result: mem
+ for {
+ if v.AuxInt != 0 {
+ break
+ }
+ _ = v.Args[1]
+ mem := v.Args[1]
+ v.reset(OpCopy)
+ v.Type = mem.Type
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Zero [1] destptr mem)
+ // cond:
+ // result: (I64Store8 destptr (I64Const [0]) mem)
+ for {
+ if v.AuxInt != 1 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store8)
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Zero [2] destptr mem)
+ // cond:
+ // result: (I64Store16 destptr (I64Const [0]) mem)
+ for {
+ if v.AuxInt != 2 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store16)
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Zero [4] destptr mem)
+ // cond:
+ // result: (I64Store32 destptr (I64Const [0]) mem)
+ for {
+ if v.AuxInt != 4 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store32)
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Zero [8] destptr mem)
+ // cond:
+ // result: (I64Store destptr (I64Const [0]) mem)
+ for {
+ if v.AuxInt != 8 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store)
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v.AddArg(mem)
+ return true
+ }
+ // match: (Zero [3] destptr mem)
+ // cond:
+ // result: (I64Store8 [2] destptr (I64Const [0]) (I64Store16 destptr (I64Const [0]) mem))
+ for {
+ if v.AuxInt != 3 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store8)
+ v.AuxInt = 2
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store16, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [5] destptr mem)
+ // cond:
+ // result: (I64Store8 [4] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem))
+ for {
+ if v.AuxInt != 5 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store8)
+ v.AuxInt = 4
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [6] destptr mem)
+ // cond:
+ // result: (I64Store16 [4] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem))
+ for {
+ if v.AuxInt != 6 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store16)
+ v.AuxInt = 4
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [7] destptr mem)
+ // cond:
+ // result: (I64Store32 [3] destptr (I64Const [0]) (I64Store32 destptr (I64Const [0]) mem))
+ for {
+ if v.AuxInt != 7 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store32)
+ v.AuxInt = 3
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store32, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [s] destptr mem)
+ // cond: s%8 != 0 && s > 8
+ // result: (Zero [s-s%8] (OffPtr <destptr.Type> destptr [s%8]) (I64Store destptr (I64Const [0]) mem))
+ for {
+ s := v.AuxInt
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ if !(s%8 != 0 && s > 8) {
+ break
+ }
+ v.reset(OpZero)
+ v.AuxInt = s - s%8
+ v0 := b.NewValue0(v.Pos, OpOffPtr, destptr.Type)
+ v0.AuxInt = s % 8
+ v0.AddArg(destptr)
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpZero_10(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (Zero [16] destptr mem)
+ // cond:
+ // result: (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem))
+ for {
+ if v.AuxInt != 16 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store)
+ v.AuxInt = 8
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v1.AddArg(mem)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [24] destptr mem)
+ // cond:
+ // result: (I64Store [16] destptr (I64Const [0]) (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem)))
+ for {
+ if v.AuxInt != 24 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store)
+ v.AuxInt = 16
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AuxInt = 8
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v3.AddArg(destptr)
+ v4 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v4.AuxInt = 0
+ v3.AddArg(v4)
+ v3.AddArg(mem)
+ v1.AddArg(v3)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [32] destptr mem)
+ // cond:
+ // result: (I64Store [24] destptr (I64Const [0]) (I64Store [16] destptr (I64Const [0]) (I64Store [8] destptr (I64Const [0]) (I64Store destptr (I64Const [0]) mem))))
+ for {
+ if v.AuxInt != 32 {
+ break
+ }
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ v.reset(OpWasmI64Store)
+ v.AuxInt = 24
+ v.AddArg(destptr)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v0.AuxInt = 0
+ v.AddArg(v0)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v1.AuxInt = 16
+ v1.AddArg(destptr)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 0
+ v1.AddArg(v2)
+ v3 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v3.AuxInt = 8
+ v3.AddArg(destptr)
+ v4 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v4.AuxInt = 0
+ v3.AddArg(v4)
+ v5 := b.NewValue0(v.Pos, OpWasmI64Store, types.TypeMem)
+ v5.AddArg(destptr)
+ v6 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v6.AuxInt = 0
+ v5.AddArg(v6)
+ v5.AddArg(mem)
+ v3.AddArg(v5)
+ v1.AddArg(v3)
+ v.AddArg(v1)
+ return true
+ }
+ // match: (Zero [s] destptr mem)
+ // cond: s%8 == 0 && s > 32
+ // result: (LoweredZero [s/8] destptr mem)
+ for {
+ s := v.AuxInt
+ _ = v.Args[1]
+ destptr := v.Args[0]
+ mem := v.Args[1]
+ if !(s%8 == 0 && s > 32) {
+ break
+ }
+ v.reset(OpWasmLoweredZero)
+ v.AuxInt = s / 8
+ v.AddArg(destptr)
+ v.AddArg(mem)
+ return true
+ }
+ return false
+}
+func rewriteValueWasm_OpZeroExt16to32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt16to32 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [48])) (I64Const [48]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 48
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 48
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpZeroExt16to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt16to64 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [48])) (I64Const [48]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 48
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 48
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpZeroExt32to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt32to64 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [32])) (I64Const [32]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 32
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 32
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpZeroExt8to16_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt8to16 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpZeroExt8to32_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt8to32 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteValueWasm_OpZeroExt8to64_0(v *Value) bool {
+ b := v.Block
+ _ = b
+ typ := &b.Func.Config.Types
+ _ = typ
+ // match: (ZeroExt8to64 x)
+ // cond:
+ // result: (I64ShrU (I64Shl x (I64Const [56])) (I64Const [56]))
+ for {
+ x := v.Args[0]
+ v.reset(OpWasmI64ShrU)
+ v0 := b.NewValue0(v.Pos, OpWasmI64Shl, typ.Int64)
+ v0.AddArg(x)
+ v1 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v1.AuxInt = 56
+ v0.AddArg(v1)
+ v.AddArg(v0)
+ v2 := b.NewValue0(v.Pos, OpWasmI64Const, typ.Int64)
+ v2.AuxInt = 56
+ v.AddArg(v2)
+ return true
+ }
+}
+func rewriteBlockWasm(b *Block) bool {
+ config := b.Func.Config
+ _ = config
+ fe := b.Func.fe
+ _ = fe
+ typ := &config.Types
+ _ = typ
+ switch b.Kind {
+ }
+ return false
+}