diff options
author | fanzha02 <fannie.zhang@arm.com> | 2019-02-20 11:38:16 +0000 |
---|---|---|
committer | Cherry Zhang <cherryyz@google.com> | 2019-03-07 21:23:52 +0000 |
commit | 6efd51c6b768ecb55cd39b0dcb8a43d9a6c8e1b2 (patch) | |
tree | b1f5e0905b6d4718610e91b75a9c5322f438036d /src/cmd/compile/internal/arm64 | |
parent | a77f85a61874c05097a60f08d9dda71512d9dcc3 (diff) | |
download | go-6efd51c6b768ecb55cd39b0dcb8a43d9a6c8e1b2.tar.gz go-6efd51c6b768ecb55cd39b0dcb8a43d9a6c8e1b2.zip |
cmd/compile: change the condition flags of floating-point comparisons in arm64 backend
Current compiler reverses operands to work around NaN in
"less than" and "less equal than" comparisons. But if we
want to use "FCMPD/FCMPS $(0.0), Fn" to do some optimization,
the workaround way does not work. Because assembler does
not support instruction "FCMPD/FCMPS Fn, $(0.0)".
This CL sets condition flags for floating-point comparisons
to resolve this problem.
Change-Id: Ia48076a1da95da64596d6e68304018cb301ebe33
Reviewed-on: https://go-review.googlesource.com/c/go/+/164718
Run-TryBot: Cherry Zhang <cherryyz@google.com>
TryBot-Result: Gobot Gobot <gobot@golang.org>
Reviewed-by: Cherry Zhang <cherryyz@google.com>
Diffstat (limited to 'src/cmd/compile/internal/arm64')
-rw-r--r-- | src/cmd/compile/internal/arm64/ssa.go | 18 |
1 files changed, 16 insertions, 2 deletions
diff --git a/src/cmd/compile/internal/arm64/ssa.go b/src/cmd/compile/internal/arm64/ssa.go index 87703dd80d..0bc8f3a5ab 100644 --- a/src/cmd/compile/internal/arm64/ssa.go +++ b/src/cmd/compile/internal/arm64/ssa.go @@ -860,7 +860,11 @@ func ssaGenValue(s *gc.SSAGenState, v *ssa.Value) { ssa.OpARM64LessThanU, ssa.OpARM64LessEqualU, ssa.OpARM64GreaterThanU, - ssa.OpARM64GreaterEqualU: + ssa.OpARM64GreaterEqualU, + ssa.OpARM64LessThanF, + ssa.OpARM64LessEqualF, + ssa.OpARM64GreaterThanF, + ssa.OpARM64GreaterEqualF: // generate boolean values using CSET p := s.Prog(arm64.ACSET) p.From.Type = obj.TYPE_REG // assembler encodes conditional bits in Reg @@ -908,6 +912,10 @@ var condBits = map[ssa.Op]int16{ ssa.OpARM64GreaterThanU: arm64.COND_HI, ssa.OpARM64GreaterEqual: arm64.COND_GE, ssa.OpARM64GreaterEqualU: arm64.COND_HS, + ssa.OpARM64LessThanF: arm64.COND_MI, + ssa.OpARM64LessEqualF: arm64.COND_LS, + ssa.OpARM64GreaterThanF: arm64.COND_GT, + ssa.OpARM64GreaterEqualF: arm64.COND_GE, } var blockJump = map[ssa.BlockKind]struct { @@ -929,6 +937,10 @@ var blockJump = map[ssa.BlockKind]struct { ssa.BlockARM64NZW: {arm64.ACBNZW, arm64.ACBZW}, ssa.BlockARM64TBZ: {arm64.ATBZ, arm64.ATBNZ}, ssa.BlockARM64TBNZ: {arm64.ATBNZ, arm64.ATBZ}, + ssa.BlockARM64FLT: {arm64.ABMI, arm64.ABPL}, + ssa.BlockARM64FGE: {arm64.ABGE, arm64.ABLT}, + ssa.BlockARM64FLE: {arm64.ABLS, arm64.ABHI}, + ssa.BlockARM64FGT: {arm64.ABGT, arm64.ABLE}, } func ssaGenBlock(s *gc.SSAGenState, b, next *ssa.Block) { @@ -975,7 +987,9 @@ func ssaGenBlock(s *gc.SSAGenState, b, next *ssa.Block) { ssa.BlockARM64ULT, ssa.BlockARM64UGT, ssa.BlockARM64ULE, ssa.BlockARM64UGE, ssa.BlockARM64Z, ssa.BlockARM64NZ, - ssa.BlockARM64ZW, ssa.BlockARM64NZW: + ssa.BlockARM64ZW, ssa.BlockARM64NZW, + ssa.BlockARM64FLT, ssa.BlockARM64FGE, + ssa.BlockARM64FLE, ssa.BlockARM64FGT: jmp := blockJump[b.Kind] var p *obj.Prog switch next { |