mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-14 13:07:31 +00:00
ae3a0be92e
integer and floating-point opcodes, introducing FAdd, FSub, and FMul. For now, the AsmParser, BitcodeReader, and IRBuilder all preserve backwards compatability, and the Core LLVM APIs preserve backwards compatibility for IR producers. Most front-ends won't need to change immediately. This implements the first step of the plan outlined here: http://nondot.org/sabre/LLVMNotes/IntegerOverflow.txt git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@72897 91177308-0d34-0410-b5e6-96231b3b80d8
145 lines
4.1 KiB
LLVM
145 lines
4.1 KiB
LLVM
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fabs | count 2
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fmscs | count 1
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fcvt | count 2
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fuito | count 2
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fto.i | count 4
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep bmi | count 1
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep bgt | count 1
|
|
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | \
|
|
; RUN: grep fcmpezs | count 1
|
|
|
|
define void @test(float* %P, double* %D) {
|
|
%A = load float* %P ; <float> [#uses=1]
|
|
%B = load double* %D ; <double> [#uses=1]
|
|
store float %A, float* %P
|
|
store double %B, double* %D
|
|
ret void
|
|
}
|
|
|
|
declare float @fabsf(float)
|
|
|
|
declare double @fabs(double)
|
|
|
|
define void @test_abs(float* %P, double* %D) {
|
|
%a = load float* %P ; <float> [#uses=1]
|
|
%b = call float @fabsf( float %a ) ; <float> [#uses=1]
|
|
store float %b, float* %P
|
|
%A = load double* %D ; <double> [#uses=1]
|
|
%B = call double @fabs( double %A ) ; <double> [#uses=1]
|
|
store double %B, double* %D
|
|
ret void
|
|
}
|
|
|
|
define void @test_add(float* %P, double* %D) {
|
|
%a = load float* %P ; <float> [#uses=2]
|
|
%b = fadd float %a, %a ; <float> [#uses=1]
|
|
store float %b, float* %P
|
|
%A = load double* %D ; <double> [#uses=2]
|
|
%B = fadd double %A, %A ; <double> [#uses=1]
|
|
store double %B, double* %D
|
|
ret void
|
|
}
|
|
|
|
define void @test_ext_round(float* %P, double* %D) {
|
|
%a = load float* %P ; <float> [#uses=1]
|
|
%b = fpext float %a to double ; <double> [#uses=1]
|
|
%A = load double* %D ; <double> [#uses=1]
|
|
%B = fptrunc double %A to float ; <float> [#uses=1]
|
|
store double %b, double* %D
|
|
store float %B, float* %P
|
|
ret void
|
|
}
|
|
|
|
define void @test_fma(float* %P1, float* %P2, float* %P3) {
|
|
%a1 = load float* %P1 ; <float> [#uses=1]
|
|
%a2 = load float* %P2 ; <float> [#uses=1]
|
|
%a3 = load float* %P3 ; <float> [#uses=1]
|
|
%X = fmul float %a1, %a2 ; <float> [#uses=1]
|
|
%Y = fsub float %X, %a3 ; <float> [#uses=1]
|
|
store float %Y, float* %P1
|
|
ret void
|
|
}
|
|
|
|
define i32 @test_ftoi(float* %P1) {
|
|
%a1 = load float* %P1 ; <float> [#uses=1]
|
|
%b1 = fptosi float %a1 to i32 ; <i32> [#uses=1]
|
|
ret i32 %b1
|
|
}
|
|
|
|
define i32 @test_ftou(float* %P1) {
|
|
%a1 = load float* %P1 ; <float> [#uses=1]
|
|
%b1 = fptoui float %a1 to i32 ; <i32> [#uses=1]
|
|
ret i32 %b1
|
|
}
|
|
|
|
define i32 @test_dtoi(double* %P1) {
|
|
%a1 = load double* %P1 ; <double> [#uses=1]
|
|
%b1 = fptosi double %a1 to i32 ; <i32> [#uses=1]
|
|
ret i32 %b1
|
|
}
|
|
|
|
define i32 @test_dtou(double* %P1) {
|
|
%a1 = load double* %P1 ; <double> [#uses=1]
|
|
%b1 = fptoui double %a1 to i32 ; <i32> [#uses=1]
|
|
ret i32 %b1
|
|
}
|
|
|
|
define void @test_utod(double* %P1, i32 %X) {
|
|
%b1 = uitofp i32 %X to double ; <double> [#uses=1]
|
|
store double %b1, double* %P1
|
|
ret void
|
|
}
|
|
|
|
define void @test_utod2(double* %P1, i8 %X) {
|
|
%b1 = uitofp i8 %X to double ; <double> [#uses=1]
|
|
store double %b1, double* %P1
|
|
ret void
|
|
}
|
|
|
|
define void @test_cmp(float* %glob, i32 %X) {
|
|
entry:
|
|
%tmp = load float* %glob ; <float> [#uses=2]
|
|
%tmp3 = getelementptr float* %glob, i32 2 ; <float*> [#uses=1]
|
|
%tmp4 = load float* %tmp3 ; <float> [#uses=2]
|
|
%tmp.upgrd.1 = fcmp oeq float %tmp, %tmp4 ; <i1> [#uses=1]
|
|
%tmp5 = fcmp uno float %tmp, %tmp4 ; <i1> [#uses=1]
|
|
%tmp6 = or i1 %tmp.upgrd.1, %tmp5 ; <i1> [#uses=1]
|
|
br i1 %tmp6, label %cond_true, label %cond_false
|
|
|
|
cond_true: ; preds = %entry
|
|
%tmp.upgrd.2 = tail call i32 (...)* @bar( ) ; <i32> [#uses=0]
|
|
ret void
|
|
|
|
cond_false: ; preds = %entry
|
|
%tmp7 = tail call i32 (...)* @baz( ) ; <i32> [#uses=0]
|
|
ret void
|
|
}
|
|
|
|
declare i1 @llvm.isunordered.f32(float, float)
|
|
|
|
declare i32 @bar(...)
|
|
|
|
declare i32 @baz(...)
|
|
|
|
define void @test_cmpfp0(float* %glob, i32 %X) {
|
|
entry:
|
|
%tmp = load float* %glob ; <float> [#uses=1]
|
|
%tmp.upgrd.3 = fcmp ogt float %tmp, 0.000000e+00 ; <i1> [#uses=1]
|
|
br i1 %tmp.upgrd.3, label %cond_true, label %cond_false
|
|
|
|
cond_true: ; preds = %entry
|
|
%tmp.upgrd.4 = tail call i32 (...)* @bar( ) ; <i32> [#uses=0]
|
|
ret void
|
|
|
|
cond_false: ; preds = %entry
|
|
%tmp1 = tail call i32 (...)* @baz( ) ; <i32> [#uses=0]
|
|
ret void
|
|
}
|