mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-08-12 00:29:20 +00:00
e023141322
boolean flag to an enum: { Fast, Standard, Strict } (default = Standard). This option controls the creation by optimizations of fused FP ops that store intermediate results in higher precision than IEEE allows (E.g. FMAs). The behavior of this option is intended to match the behaviour specified by a soon-to-be-introduced frontend flag: '-ffuse-fp-ops'. Fast mode - allows formation of fused FP ops whenever they're profitable. Standard mode - allow fusion only for 'blessed' FP ops. At present the only blessed op is the fmuladd intrinsic. In the future more blessed ops may be added. Strict mode - allow fusion only if/when it can be proven that the excess precision won't effect the result. Note: This option only controls formation of fused ops by the optimizers. Fused operations that are explicitly requested (e.g. FMA via the llvm.fma.* intrinsic) will always be honored, regardless of the value of this option. Internally TargetOptions::AllowExcessFPPrecision has been replaced by TargetOptions::AllowFPOpFusion. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@158956 91177308-0d34-0410-b5e6-96231b3b80d8
213 lines
5.9 KiB
LLVM
213 lines
5.9 KiB
LLVM
; RUN: llc < %s -mtriple=armv7-eabi -mattr=+neon,+vfp4 -fuse-fp-ops=fast | FileCheck %s
|
|
; Check generated fused MAC and MLS.
|
|
|
|
define double @fusedMACTest1(double %d1, double %d2, double %d3) {
|
|
;CHECK: fusedMACTest1:
|
|
;CHECK: vfma.f64
|
|
%1 = fmul double %d1, %d2
|
|
%2 = fadd double %1, %d3
|
|
ret double %2
|
|
}
|
|
|
|
define float @fusedMACTest2(float %f1, float %f2, float %f3) {
|
|
;CHECK: fusedMACTest2:
|
|
;CHECK: vfma.f32
|
|
%1 = fmul float %f1, %f2
|
|
%2 = fadd float %1, %f3
|
|
ret float %2
|
|
}
|
|
|
|
define double @fusedMACTest3(double %d1, double %d2, double %d3) {
|
|
;CHECK: fusedMACTest3:
|
|
;CHECK: vfms.f64
|
|
%1 = fmul double %d2, %d3
|
|
%2 = fsub double %d1, %1
|
|
ret double %2
|
|
}
|
|
|
|
define float @fusedMACTest4(float %f1, float %f2, float %f3) {
|
|
;CHECK: fusedMACTest4:
|
|
;CHECK: vfms.f32
|
|
%1 = fmul float %f2, %f3
|
|
%2 = fsub float %f1, %1
|
|
ret float %2
|
|
}
|
|
|
|
define double @fusedMACTest5(double %d1, double %d2, double %d3) {
|
|
;CHECK: fusedMACTest5:
|
|
;CHECK: vfnma.f64
|
|
%1 = fmul double %d1, %d2
|
|
%2 = fsub double -0.0, %1
|
|
%3 = fsub double %2, %d3
|
|
ret double %3
|
|
}
|
|
|
|
define float @fusedMACTest6(float %f1, float %f2, float %f3) {
|
|
;CHECK: fusedMACTest6:
|
|
;CHECK: vfnma.f32
|
|
%1 = fmul float %f1, %f2
|
|
%2 = fsub float -0.0, %1
|
|
%3 = fsub float %2, %f3
|
|
ret float %3
|
|
}
|
|
|
|
define double @fusedMACTest7(double %d1, double %d2, double %d3) {
|
|
;CHECK: fusedMACTest7:
|
|
;CHECK: vfnms.f64
|
|
%1 = fmul double %d1, %d2
|
|
%2 = fsub double %1, %d3
|
|
ret double %2
|
|
}
|
|
|
|
define float @fusedMACTest8(float %f1, float %f2, float %f3) {
|
|
;CHECK: fusedMACTest8:
|
|
;CHECK: vfnms.f32
|
|
%1 = fmul float %f1, %f2
|
|
%2 = fsub float %1, %f3
|
|
ret float %2
|
|
}
|
|
|
|
define <2 x float> @fusedMACTest9(<2 x float> %a, <2 x float> %b) {
|
|
;CHECK: fusedMACTest9:
|
|
;CHECK: vfma.f32
|
|
%mul = fmul <2 x float> %a, %b
|
|
%add = fadd <2 x float> %mul, %a
|
|
ret <2 x float> %add
|
|
}
|
|
|
|
define <2 x float> @fusedMACTest10(<2 x float> %a, <2 x float> %b) {
|
|
;CHECK: fusedMACTest10:
|
|
;CHECK: vfms.f32
|
|
%mul = fmul <2 x float> %a, %b
|
|
%sub = fsub <2 x float> %a, %mul
|
|
ret <2 x float> %sub
|
|
}
|
|
|
|
define <4 x float> @fusedMACTest11(<4 x float> %a, <4 x float> %b) {
|
|
;CHECK: fusedMACTest11:
|
|
;CHECK: vfma.f32
|
|
%mul = fmul <4 x float> %a, %b
|
|
%add = fadd <4 x float> %mul, %a
|
|
ret <4 x float> %add
|
|
}
|
|
|
|
define <4 x float> @fusedMACTest12(<4 x float> %a, <4 x float> %b) {
|
|
;CHECK: fusedMACTest12:
|
|
;CHECK: vfms.f32
|
|
%mul = fmul <4 x float> %a, %b
|
|
%sub = fsub <4 x float> %a, %mul
|
|
ret <4 x float> %sub
|
|
}
|
|
|
|
define float @test_fma_f32(float %a, float %b, float %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fma_f32
|
|
; CHECK: vfma.f32
|
|
%tmp1 = tail call float @llvm.fma.f32(float %a, float %b, float %c) nounwind readnone
|
|
ret float %tmp1
|
|
}
|
|
|
|
define double @test_fma_f64(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fma_f64
|
|
; CHECK: vfma.f64
|
|
%tmp1 = tail call double @llvm.fma.f64(double %a, double %b, double %c) nounwind readnone
|
|
ret double %tmp1
|
|
}
|
|
|
|
define <2 x float> @test_fma_v2f32(<2 x float> %a, <2 x float> %b, <2 x float> %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fma_v2f32
|
|
; CHECK: vfma.f32
|
|
%tmp1 = tail call <2 x float> @llvm.fma.v2f32(<2 x float> %a, <2 x float> %b, <2 x float> %c) nounwind
|
|
ret <2 x float> %tmp1
|
|
}
|
|
|
|
define double @test_fms_f64(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fms_f64
|
|
; CHECK: vfms.f64
|
|
%tmp1 = fsub double -0.0, %a
|
|
%tmp2 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %c) nounwind readnone
|
|
ret double %tmp2
|
|
}
|
|
|
|
define double @test_fms_f64_2(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fms_f64_2
|
|
; CHECK: vfms.f64
|
|
%tmp1 = fsub double -0.0, %b
|
|
%tmp2 = tail call double @llvm.fma.f64(double %a, double %tmp1, double %c) nounwind readnone
|
|
ret double %tmp2
|
|
}
|
|
|
|
define float @test_fnms_f32(float %a, float %b, float* %c) nounwind readnone ssp {
|
|
; CHECK: test_fnms_f32
|
|
; CHECK: vfnms.f32
|
|
%tmp1 = load float* %c, align 4
|
|
%tmp2 = fsub float -0.0, %tmp1
|
|
%tmp3 = tail call float @llvm.fma.f32(float %a, float %b, float %tmp2) nounwind readnone
|
|
ret float %tmp3
|
|
}
|
|
|
|
define double @test_fnms_f64(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fnms_f64
|
|
; CHECK: vfnms.f64
|
|
%tmp1 = fsub double -0.0, %a
|
|
%tmp2 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %c) nounwind readnone
|
|
%tmp3 = fsub double -0.0, %tmp2
|
|
ret double %tmp3
|
|
}
|
|
|
|
define double @test_fnms_f64_2(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fnms_f64_2
|
|
; CHECK: vfnms.f64
|
|
%tmp1 = fsub double -0.0, %b
|
|
%tmp2 = tail call double @llvm.fma.f64(double %a, double %tmp1, double %c) nounwind readnone
|
|
%tmp3 = fsub double -0.0, %tmp2
|
|
ret double %tmp3
|
|
}
|
|
|
|
define double @test_fnma_f64(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fnma_f64
|
|
; CHECK: vfnma.f64
|
|
%tmp1 = tail call double @llvm.fma.f64(double %a, double %b, double %c) nounwind readnone
|
|
%tmp2 = fsub double -0.0, %tmp1
|
|
ret double %tmp2
|
|
}
|
|
|
|
define double @test_fnma_f64_2(double %a, double %b, double %c) nounwind readnone ssp {
|
|
entry:
|
|
; CHECK: test_fnma_f64_2
|
|
; CHECK: vfnma.f64
|
|
%tmp1 = fsub double -0.0, %a
|
|
%tmp2 = fsub double -0.0, %c
|
|
%tmp3 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %tmp2) nounwind readnone
|
|
ret double %tmp3
|
|
}
|
|
|
|
define float @test_fma_const_fold(float %a, float %b) nounwind {
|
|
; CHECK: test_fma_const_fold
|
|
; CHECK-NOT: vfma
|
|
; CHECK-NOT: vmul
|
|
; CHECK: vadd
|
|
%ret = call float @llvm.fma.f32(float %a, float 1.0, float %b)
|
|
ret float %ret
|
|
}
|
|
|
|
define float @test_fma_canonicalize(float %a, float %b) nounwind {
|
|
; CHECK: test_fma_canonicalize
|
|
; CHECK: vmov.f32 [[R1:s[0-9]+]], #2.000000e+00
|
|
; CHECK: vfma.f32 {{s[0-9]+}}, {{s[0-9]+}}, [[R1]]
|
|
%ret = call float @llvm.fma.f32(float 2.0, float %a, float %b)
|
|
ret float %ret
|
|
}
|
|
|
|
|
|
declare float @llvm.fma.f32(float, float, float) nounwind readnone
|
|
declare double @llvm.fma.f64(double, double, double) nounwind readnone
|
|
declare <2 x float> @llvm.fma.v2f32(<2 x float>, <2 x float>, <2 x float>) nounwind readnone
|