mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-22 07:32:48 +00:00
b3f912b510
r186399 aggressively used the RISBG instruction for immediate ANDs, both because it can handle some values that AND IMMEDIATE can't, and because it allows the destination register to be different from the source. I realized later while implementing the distinct-ops support that it would be better to leave the choice up to convertToThreeAddress() instead. The AND IMMEDIATE form is shorter and is less likely to be cracked. This is a problem for 32-bit ANDs because we assume that all 32-bit operations will leave the high word untouched, whereas RISBG used in this way will either clear the high word or copy it from the source register. The patch uses the z196 instruction RISBLG for this instead. This means that z10 will be restricted to NILL, NILH and NILF for 32-bit ANDs, but I think that should be OK for now. Although we're using z10 as the base architecture, the optimization work is going to be focused more on z196 and zEC12. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@187492 91177308-0d34-0410-b5e6-96231b3b80d8
393 lines
12 KiB
LLVM
393 lines
12 KiB
LLVM
; Test moves between FPRs and GPRs.
|
|
;
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
|
|
|
|
declare i64 @foo()
|
|
declare double @bar()
|
|
@dptr = external global double
|
|
@iptr = external global i64
|
|
|
|
; Test 32-bit moves from GPRs to FPRs. The GPR must be moved into the high
|
|
; 32 bits of the FPR.
|
|
define float @f1(i32 %a) {
|
|
; CHECK-LABEL: f1:
|
|
; CHECK: sllg [[REGISTER:%r[0-5]]], %r2, 32
|
|
; CHECK: ldgr %f0, [[REGISTER]]
|
|
%res = bitcast i32 %a to float
|
|
ret float %res
|
|
}
|
|
|
|
; Like f1, but create a situation where the shift can be folded with
|
|
; surrounding code.
|
|
define float @f2(i64 %big) {
|
|
; CHECK-LABEL: f2:
|
|
; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 31
|
|
; CHECK: ldgr %f0, [[REGISTER]]
|
|
%shift = lshr i64 %big, 1
|
|
%a = trunc i64 %shift to i32
|
|
%res = bitcast i32 %a to float
|
|
ret float %res
|
|
}
|
|
|
|
; Another example of the same thing.
|
|
define float @f3(i64 %big) {
|
|
; CHECK-LABEL: f3:
|
|
; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 2
|
|
; CHECK: ldgr %f0, [[REGISTER]]
|
|
%shift = ashr i64 %big, 30
|
|
%a = trunc i64 %shift to i32
|
|
%res = bitcast i32 %a to float
|
|
ret float %res
|
|
}
|
|
|
|
; Like f1, but the value to transfer is already in the high 32 bits.
|
|
define float @f4(i64 %big) {
|
|
; CHECK-LABEL: f4:
|
|
; CHECK-NOT: %r2
|
|
; CHECK: nilf %r2, 0
|
|
; CHECK-NOT: %r2
|
|
; CHECK: ldgr %f0, %r2
|
|
%shift = ashr i64 %big, 32
|
|
%a = trunc i64 %shift to i32
|
|
%res = bitcast i32 %a to float
|
|
ret float %res
|
|
}
|
|
|
|
; Test 64-bit moves from GPRs to FPRs.
|
|
define double @f5(i64 %a) {
|
|
; CHECK-LABEL: f5:
|
|
; CHECK: ldgr %f0, %r2
|
|
%res = bitcast i64 %a to double
|
|
ret double %res
|
|
}
|
|
|
|
; Test 128-bit moves from GPRs to FPRs. i128 isn't a legitimate type,
|
|
; so this goes through memory.
|
|
; FIXME: it would be better to use one MVC here.
|
|
define void @f6(fp128 *%a, i128 *%b) {
|
|
; CHECK-LABEL: f6:
|
|
; CHECK: lg
|
|
; CHECK: mvc
|
|
; CHECK: stg
|
|
; CHECK: br %r14
|
|
%val = load i128 *%b
|
|
%res = bitcast i128 %val to fp128
|
|
store fp128 %res, fp128 *%a
|
|
ret void
|
|
}
|
|
|
|
; Test 32-bit moves from FPRs to GPRs. The high 32 bits of the FPR should
|
|
; be moved into the low 32 bits of the GPR.
|
|
define i32 @f7(float %a) {
|
|
; CHECK-LABEL: f7:
|
|
; CHECK: lgdr [[REGISTER:%r[0-5]]], %f0
|
|
; CHECK: srlg %r2, [[REGISTER]], 32
|
|
%res = bitcast float %a to i32
|
|
ret i32 %res
|
|
}
|
|
|
|
; Test 64-bit moves from FPRs to GPRs.
|
|
define i64 @f8(double %a) {
|
|
; CHECK-LABEL: f8:
|
|
; CHECK: lgdr %r2, %f0
|
|
%res = bitcast double %a to i64
|
|
ret i64 %res
|
|
}
|
|
|
|
; Test 128-bit moves from FPRs to GPRs, with the same restriction as f6.
|
|
define void @f9(fp128 *%a, i128 *%b) {
|
|
; CHECK-LABEL: f9:
|
|
; CHECK: ld
|
|
; CHECK: ld
|
|
; CHECK: std
|
|
; CHECK: std
|
|
%val = load fp128 *%a
|
|
%res = bitcast fp128 %val to i128
|
|
store i128 %res, i128 *%b
|
|
ret void
|
|
}
|
|
|
|
; Test cases where the destination of an LGDR needs to be spilled.
|
|
; We shouldn't have any integer stack stores or floating-point loads.
|
|
define void @f10(double %extra) {
|
|
; CHECK-LABEL: f10:
|
|
; CHECK: dptr
|
|
; CHECK-NOT: stg {{.*}}(%r15)
|
|
; CHECK: %loop
|
|
; CHECK-NOT: ld {{.*}}(%r15)
|
|
; CHECK: %exit
|
|
; CHECK: br %r14
|
|
entry:
|
|
%double0 = load volatile double *@dptr
|
|
%biased0 = fadd double %double0, %extra
|
|
%int0 = bitcast double %biased0 to i64
|
|
%double1 = load volatile double *@dptr
|
|
%biased1 = fadd double %double1, %extra
|
|
%int1 = bitcast double %biased1 to i64
|
|
%double2 = load volatile double *@dptr
|
|
%biased2 = fadd double %double2, %extra
|
|
%int2 = bitcast double %biased2 to i64
|
|
%double3 = load volatile double *@dptr
|
|
%biased3 = fadd double %double3, %extra
|
|
%int3 = bitcast double %biased3 to i64
|
|
%double4 = load volatile double *@dptr
|
|
%biased4 = fadd double %double4, %extra
|
|
%int4 = bitcast double %biased4 to i64
|
|
%double5 = load volatile double *@dptr
|
|
%biased5 = fadd double %double5, %extra
|
|
%int5 = bitcast double %biased5 to i64
|
|
%double6 = load volatile double *@dptr
|
|
%biased6 = fadd double %double6, %extra
|
|
%int6 = bitcast double %biased6 to i64
|
|
%double7 = load volatile double *@dptr
|
|
%biased7 = fadd double %double7, %extra
|
|
%int7 = bitcast double %biased7 to i64
|
|
%double8 = load volatile double *@dptr
|
|
%biased8 = fadd double %double8, %extra
|
|
%int8 = bitcast double %biased8 to i64
|
|
%double9 = load volatile double *@dptr
|
|
%biased9 = fadd double %double9, %extra
|
|
%int9 = bitcast double %biased9 to i64
|
|
br label %loop
|
|
|
|
loop:
|
|
%start = call i64 @foo()
|
|
%or0 = or i64 %start, %int0
|
|
%or1 = or i64 %or0, %int1
|
|
%or2 = or i64 %or1, %int2
|
|
%or3 = or i64 %or2, %int3
|
|
%or4 = or i64 %or3, %int4
|
|
%or5 = or i64 %or4, %int5
|
|
%or6 = or i64 %or5, %int6
|
|
%or7 = or i64 %or6, %int7
|
|
%or8 = or i64 %or7, %int8
|
|
%or9 = or i64 %or8, %int9
|
|
store i64 %or9, i64 *@iptr
|
|
%cont = icmp ne i64 %start, 1
|
|
br i1 %cont, label %loop, label %exit
|
|
|
|
exit:
|
|
ret void
|
|
}
|
|
|
|
; ...likewise LDGR, with the requirements the other way around.
|
|
define void @f11(i64 %mask) {
|
|
; CHECK-LABEL: f11:
|
|
; CHECK: iptr
|
|
; CHECK-NOT: std {{.*}}(%r15)
|
|
; CHECK: %loop
|
|
; CHECK-NOT: lg {{.*}}(%r15)
|
|
; CHECK: %exit
|
|
; CHECK: br %r14
|
|
entry:
|
|
%int0 = load volatile i64 *@iptr
|
|
%masked0 = and i64 %int0, %mask
|
|
%double0 = bitcast i64 %masked0 to double
|
|
%int1 = load volatile i64 *@iptr
|
|
%masked1 = and i64 %int1, %mask
|
|
%double1 = bitcast i64 %masked1 to double
|
|
%int2 = load volatile i64 *@iptr
|
|
%masked2 = and i64 %int2, %mask
|
|
%double2 = bitcast i64 %masked2 to double
|
|
%int3 = load volatile i64 *@iptr
|
|
%masked3 = and i64 %int3, %mask
|
|
%double3 = bitcast i64 %masked3 to double
|
|
%int4 = load volatile i64 *@iptr
|
|
%masked4 = and i64 %int4, %mask
|
|
%double4 = bitcast i64 %masked4 to double
|
|
%int5 = load volatile i64 *@iptr
|
|
%masked5 = and i64 %int5, %mask
|
|
%double5 = bitcast i64 %masked5 to double
|
|
%int6 = load volatile i64 *@iptr
|
|
%masked6 = and i64 %int6, %mask
|
|
%double6 = bitcast i64 %masked6 to double
|
|
%int7 = load volatile i64 *@iptr
|
|
%masked7 = and i64 %int7, %mask
|
|
%double7 = bitcast i64 %masked7 to double
|
|
%int8 = load volatile i64 *@iptr
|
|
%masked8 = and i64 %int8, %mask
|
|
%double8 = bitcast i64 %masked8 to double
|
|
%int9 = load volatile i64 *@iptr
|
|
%masked9 = and i64 %int9, %mask
|
|
%double9 = bitcast i64 %masked9 to double
|
|
br label %loop
|
|
|
|
loop:
|
|
%start = call double @bar()
|
|
%add0 = fadd double %start, %double0
|
|
%add1 = fadd double %add0, %double1
|
|
%add2 = fadd double %add1, %double2
|
|
%add3 = fadd double %add2, %double3
|
|
%add4 = fadd double %add3, %double4
|
|
%add5 = fadd double %add4, %double5
|
|
%add6 = fadd double %add5, %double6
|
|
%add7 = fadd double %add6, %double7
|
|
%add8 = fadd double %add7, %double8
|
|
%add9 = fadd double %add8, %double9
|
|
store double %add9, double *@dptr
|
|
%cont = fcmp one double %start, 1.0
|
|
br i1 %cont, label %loop, label %exit
|
|
|
|
exit:
|
|
ret void
|
|
}
|
|
|
|
; Test cases where the source of an LDGR needs to be spilled.
|
|
; We shouldn't have any integer stack stores or floating-point loads.
|
|
define void @f12() {
|
|
; CHECK-LABEL: f12:
|
|
; CHECK: %loop
|
|
; CHECK-NOT: std {{.*}}(%r15)
|
|
; CHECK: %exit
|
|
; CHECK: foo@PLT
|
|
; CHECK-NOT: lg {{.*}}(%r15)
|
|
; CHECK: foo@PLT
|
|
; CHECK: br %r14
|
|
entry:
|
|
br label %loop
|
|
|
|
loop:
|
|
%int0 = phi i64 [ 0, %entry ], [ %add0, %loop ]
|
|
%int1 = phi i64 [ 0, %entry ], [ %add1, %loop ]
|
|
%int2 = phi i64 [ 0, %entry ], [ %add2, %loop ]
|
|
%int3 = phi i64 [ 0, %entry ], [ %add3, %loop ]
|
|
%int4 = phi i64 [ 0, %entry ], [ %add4, %loop ]
|
|
%int5 = phi i64 [ 0, %entry ], [ %add5, %loop ]
|
|
%int6 = phi i64 [ 0, %entry ], [ %add6, %loop ]
|
|
%int7 = phi i64 [ 0, %entry ], [ %add7, %loop ]
|
|
%int8 = phi i64 [ 0, %entry ], [ %add8, %loop ]
|
|
%int9 = phi i64 [ 0, %entry ], [ %add9, %loop ]
|
|
|
|
%bias = call i64 @foo()
|
|
%add0 = add i64 %int0, %bias
|
|
%add1 = add i64 %int1, %bias
|
|
%add2 = add i64 %int2, %bias
|
|
%add3 = add i64 %int3, %bias
|
|
%add4 = add i64 %int4, %bias
|
|
%add5 = add i64 %int5, %bias
|
|
%add6 = add i64 %int6, %bias
|
|
%add7 = add i64 %int7, %bias
|
|
%add8 = add i64 %int8, %bias
|
|
%add9 = add i64 %int9, %bias
|
|
%cont = icmp ne i64 %bias, 1
|
|
br i1 %cont, label %loop, label %exit
|
|
|
|
exit:
|
|
%unused1 = call i64 @foo()
|
|
%factor = load volatile double *@dptr
|
|
|
|
%conv0 = bitcast i64 %add0 to double
|
|
%mul0 = fmul double %conv0, %factor
|
|
store volatile double %mul0, double *@dptr
|
|
%conv1 = bitcast i64 %add1 to double
|
|
%mul1 = fmul double %conv1, %factor
|
|
store volatile double %mul1, double *@dptr
|
|
%conv2 = bitcast i64 %add2 to double
|
|
%mul2 = fmul double %conv2, %factor
|
|
store volatile double %mul2, double *@dptr
|
|
%conv3 = bitcast i64 %add3 to double
|
|
%mul3 = fmul double %conv3, %factor
|
|
store volatile double %mul3, double *@dptr
|
|
%conv4 = bitcast i64 %add4 to double
|
|
%mul4 = fmul double %conv4, %factor
|
|
store volatile double %mul4, double *@dptr
|
|
%conv5 = bitcast i64 %add5 to double
|
|
%mul5 = fmul double %conv5, %factor
|
|
store volatile double %mul5, double *@dptr
|
|
%conv6 = bitcast i64 %add6 to double
|
|
%mul6 = fmul double %conv6, %factor
|
|
store volatile double %mul6, double *@dptr
|
|
%conv7 = bitcast i64 %add7 to double
|
|
%mul7 = fmul double %conv7, %factor
|
|
store volatile double %mul7, double *@dptr
|
|
%conv8 = bitcast i64 %add8 to double
|
|
%mul8 = fmul double %conv8, %factor
|
|
store volatile double %mul8, double *@dptr
|
|
%conv9 = bitcast i64 %add9 to double
|
|
%mul9 = fmul double %conv9, %factor
|
|
store volatile double %mul9, double *@dptr
|
|
|
|
%unused2 = call i64 @foo()
|
|
|
|
ret void
|
|
}
|
|
|
|
; ...likewise LGDR, with the requirements the other way around.
|
|
define void @f13() {
|
|
; CHECK-LABEL: f13:
|
|
; CHECK: %loop
|
|
; CHECK-NOT: stg {{.*}}(%r15)
|
|
; CHECK: %exit
|
|
; CHECK: foo@PLT
|
|
; CHECK-NOT: ld {{.*}}(%r15)
|
|
; CHECK: foo@PLT
|
|
; CHECK: br %r14
|
|
entry:
|
|
br label %loop
|
|
|
|
loop:
|
|
%double0 = phi double [ 1.0, %entry ], [ %mul0, %loop ]
|
|
%double1 = phi double [ 1.0, %entry ], [ %mul1, %loop ]
|
|
%double2 = phi double [ 1.0, %entry ], [ %mul2, %loop ]
|
|
%double3 = phi double [ 1.0, %entry ], [ %mul3, %loop ]
|
|
%double4 = phi double [ 1.0, %entry ], [ %mul4, %loop ]
|
|
%double5 = phi double [ 1.0, %entry ], [ %mul5, %loop ]
|
|
%double6 = phi double [ 1.0, %entry ], [ %mul6, %loop ]
|
|
%double7 = phi double [ 1.0, %entry ], [ %mul7, %loop ]
|
|
%double8 = phi double [ 1.0, %entry ], [ %mul8, %loop ]
|
|
%double9 = phi double [ 1.0, %entry ], [ %mul9, %loop ]
|
|
|
|
%factor = call double @bar()
|
|
%mul0 = fmul double %double0, %factor
|
|
%mul1 = fmul double %double1, %factor
|
|
%mul2 = fmul double %double2, %factor
|
|
%mul3 = fmul double %double3, %factor
|
|
%mul4 = fmul double %double4, %factor
|
|
%mul5 = fmul double %double5, %factor
|
|
%mul6 = fmul double %double6, %factor
|
|
%mul7 = fmul double %double7, %factor
|
|
%mul8 = fmul double %double8, %factor
|
|
%mul9 = fmul double %double9, %factor
|
|
%cont = fcmp one double %factor, 1.0
|
|
br i1 %cont, label %loop, label %exit
|
|
|
|
exit:
|
|
%unused1 = call i64 @foo()
|
|
%bias = load volatile i64 *@iptr
|
|
|
|
%conv0 = bitcast double %mul0 to i64
|
|
%add0 = add i64 %conv0, %bias
|
|
store volatile i64 %add0, i64 *@iptr
|
|
%conv1 = bitcast double %mul1 to i64
|
|
%add1 = add i64 %conv1, %bias
|
|
store volatile i64 %add1, i64 *@iptr
|
|
%conv2 = bitcast double %mul2 to i64
|
|
%add2 = add i64 %conv2, %bias
|
|
store volatile i64 %add2, i64 *@iptr
|
|
%conv3 = bitcast double %mul3 to i64
|
|
%add3 = add i64 %conv3, %bias
|
|
store volatile i64 %add3, i64 *@iptr
|
|
%conv4 = bitcast double %mul4 to i64
|
|
%add4 = add i64 %conv4, %bias
|
|
store volatile i64 %add4, i64 *@iptr
|
|
%conv5 = bitcast double %mul5 to i64
|
|
%add5 = add i64 %conv5, %bias
|
|
store volatile i64 %add5, i64 *@iptr
|
|
%conv6 = bitcast double %mul6 to i64
|
|
%add6 = add i64 %conv6, %bias
|
|
store volatile i64 %add6, i64 *@iptr
|
|
%conv7 = bitcast double %mul7 to i64
|
|
%add7 = add i64 %conv7, %bias
|
|
store volatile i64 %add7, i64 *@iptr
|
|
%conv8 = bitcast double %mul8 to i64
|
|
%add8 = add i64 %conv8, %bias
|
|
store volatile i64 %add8, i64 *@iptr
|
|
%conv9 = bitcast double %mul9 to i64
|
|
%add9 = add i64 %conv9, %bias
|
|
store volatile i64 %add9, i64 *@iptr
|
|
|
|
%unused2 = call i64 @foo()
|
|
|
|
ret void
|
|
}
|