llvm-6502/test/CodeGen/SystemZ/fp-move-02.ll
Richard Sandiford 0548a5487a [SystemZ] Rein back the use of block operations
The backend tries to use block operations like MVC, NC, OC and XC for
simple scalar operations.  For correctness reasons, it rejects any case
in which the regions might partially overlap.  However, for performance
reasons, it should also reject cases where the regions might be equal,
since the instruction might then not use the fast path.

This fixes a performance regression seen in bzip2.  We may want to limit
the optimisation even more in future, or even remove it entirely, but I'll
try with this for now.


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@191525 91177308-0d34-0410-b5e6-96231b3b80d8
2013-09-27 15:29:20 +00:00

393 lines
12 KiB
LLVM

; Test moves between FPRs and GPRs.
;
; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
declare i64 @foo()
declare double @bar()
@dptr = external global double
@iptr = external global i64
; Test 32-bit moves from GPRs to FPRs. The GPR must be moved into the high
; 32 bits of the FPR.
define float @f1(i32 %a) {
; CHECK-LABEL: f1:
; CHECK: sllg [[REGISTER:%r[0-5]]], %r2, 32
; CHECK: ldgr %f0, [[REGISTER]]
%res = bitcast i32 %a to float
ret float %res
}
; Like f1, but create a situation where the shift can be folded with
; surrounding code.
define float @f2(i64 %big) {
; CHECK-LABEL: f2:
; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 31
; CHECK: ldgr %f0, [[REGISTER]]
%shift = lshr i64 %big, 1
%a = trunc i64 %shift to i32
%res = bitcast i32 %a to float
ret float %res
}
; Another example of the same thing.
define float @f3(i64 %big) {
; CHECK-LABEL: f3:
; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 2
; CHECK: ldgr %f0, [[REGISTER]]
%shift = ashr i64 %big, 30
%a = trunc i64 %shift to i32
%res = bitcast i32 %a to float
ret float %res
}
; Like f1, but the value to transfer is already in the high 32 bits.
define float @f4(i64 %big) {
; CHECK-LABEL: f4:
; CHECK-NOT: %r2
; CHECK: nilf %r2, 0
; CHECK-NOT: %r2
; CHECK: ldgr %f0, %r2
%shift = ashr i64 %big, 32
%a = trunc i64 %shift to i32
%res = bitcast i32 %a to float
ret float %res
}
; Test 64-bit moves from GPRs to FPRs.
define double @f5(i64 %a) {
; CHECK-LABEL: f5:
; CHECK: ldgr %f0, %r2
%res = bitcast i64 %a to double
ret double %res
}
; Test 128-bit moves from GPRs to FPRs. i128 isn't a legitimate type,
; so this goes through memory.
define void @f6(fp128 *%a, i128 *%b) {
; CHECK-LABEL: f6:
; CHECK: lg
; CHECK: lg
; CHECK: stg
; CHECK: stg
; CHECK: br %r14
%val = load i128 *%b
%res = bitcast i128 %val to fp128
store fp128 %res, fp128 *%a
ret void
}
; Test 32-bit moves from FPRs to GPRs. The high 32 bits of the FPR should
; be moved into the low 32 bits of the GPR.
define i32 @f7(float %a) {
; CHECK-LABEL: f7:
; CHECK: lgdr [[REGISTER:%r[0-5]]], %f0
; CHECK: srlg %r2, [[REGISTER]], 32
%res = bitcast float %a to i32
ret i32 %res
}
; Test 64-bit moves from FPRs to GPRs.
define i64 @f8(double %a) {
; CHECK-LABEL: f8:
; CHECK: lgdr %r2, %f0
%res = bitcast double %a to i64
ret i64 %res
}
; Test 128-bit moves from FPRs to GPRs, with the same restriction as f6.
define void @f9(fp128 *%a, i128 *%b) {
; CHECK-LABEL: f9:
; CHECK: ld
; CHECK: ld
; CHECK: std
; CHECK: std
%val = load fp128 *%a
%res = bitcast fp128 %val to i128
store i128 %res, i128 *%b
ret void
}
; Test cases where the destination of an LGDR needs to be spilled.
; We shouldn't have any integer stack stores or floating-point loads.
define void @f10(double %extra) {
; CHECK-LABEL: f10:
; CHECK: dptr
; CHECK-NOT: stg {{.*}}(%r15)
; CHECK: %loop
; CHECK-NOT: ld {{.*}}(%r15)
; CHECK: %exit
; CHECK: br %r14
entry:
%double0 = load volatile double *@dptr
%biased0 = fadd double %double0, %extra
%int0 = bitcast double %biased0 to i64
%double1 = load volatile double *@dptr
%biased1 = fadd double %double1, %extra
%int1 = bitcast double %biased1 to i64
%double2 = load volatile double *@dptr
%biased2 = fadd double %double2, %extra
%int2 = bitcast double %biased2 to i64
%double3 = load volatile double *@dptr
%biased3 = fadd double %double3, %extra
%int3 = bitcast double %biased3 to i64
%double4 = load volatile double *@dptr
%biased4 = fadd double %double4, %extra
%int4 = bitcast double %biased4 to i64
%double5 = load volatile double *@dptr
%biased5 = fadd double %double5, %extra
%int5 = bitcast double %biased5 to i64
%double6 = load volatile double *@dptr
%biased6 = fadd double %double6, %extra
%int6 = bitcast double %biased6 to i64
%double7 = load volatile double *@dptr
%biased7 = fadd double %double7, %extra
%int7 = bitcast double %biased7 to i64
%double8 = load volatile double *@dptr
%biased8 = fadd double %double8, %extra
%int8 = bitcast double %biased8 to i64
%double9 = load volatile double *@dptr
%biased9 = fadd double %double9, %extra
%int9 = bitcast double %biased9 to i64
br label %loop
loop:
%start = call i64 @foo()
%or0 = or i64 %start, %int0
%or1 = or i64 %or0, %int1
%or2 = or i64 %or1, %int2
%or3 = or i64 %or2, %int3
%or4 = or i64 %or3, %int4
%or5 = or i64 %or4, %int5
%or6 = or i64 %or5, %int6
%or7 = or i64 %or6, %int7
%or8 = or i64 %or7, %int8
%or9 = or i64 %or8, %int9
store i64 %or9, i64 *@iptr
%cont = icmp ne i64 %start, 1
br i1 %cont, label %loop, label %exit
exit:
ret void
}
; ...likewise LDGR, with the requirements the other way around.
define void @f11(i64 %mask) {
; CHECK-LABEL: f11:
; CHECK: iptr
; CHECK-NOT: std {{.*}}(%r15)
; CHECK: %loop
; CHECK-NOT: lg {{.*}}(%r15)
; CHECK: %exit
; CHECK: br %r14
entry:
%int0 = load volatile i64 *@iptr
%masked0 = and i64 %int0, %mask
%double0 = bitcast i64 %masked0 to double
%int1 = load volatile i64 *@iptr
%masked1 = and i64 %int1, %mask
%double1 = bitcast i64 %masked1 to double
%int2 = load volatile i64 *@iptr
%masked2 = and i64 %int2, %mask
%double2 = bitcast i64 %masked2 to double
%int3 = load volatile i64 *@iptr
%masked3 = and i64 %int3, %mask
%double3 = bitcast i64 %masked3 to double
%int4 = load volatile i64 *@iptr
%masked4 = and i64 %int4, %mask
%double4 = bitcast i64 %masked4 to double
%int5 = load volatile i64 *@iptr
%masked5 = and i64 %int5, %mask
%double5 = bitcast i64 %masked5 to double
%int6 = load volatile i64 *@iptr
%masked6 = and i64 %int6, %mask
%double6 = bitcast i64 %masked6 to double
%int7 = load volatile i64 *@iptr
%masked7 = and i64 %int7, %mask
%double7 = bitcast i64 %masked7 to double
%int8 = load volatile i64 *@iptr
%masked8 = and i64 %int8, %mask
%double8 = bitcast i64 %masked8 to double
%int9 = load volatile i64 *@iptr
%masked9 = and i64 %int9, %mask
%double9 = bitcast i64 %masked9 to double
br label %loop
loop:
%start = call double @bar()
%add0 = fadd double %start, %double0
%add1 = fadd double %add0, %double1
%add2 = fadd double %add1, %double2
%add3 = fadd double %add2, %double3
%add4 = fadd double %add3, %double4
%add5 = fadd double %add4, %double5
%add6 = fadd double %add5, %double6
%add7 = fadd double %add6, %double7
%add8 = fadd double %add7, %double8
%add9 = fadd double %add8, %double9
store double %add9, double *@dptr
%cont = fcmp one double %start, 1.0
br i1 %cont, label %loop, label %exit
exit:
ret void
}
; Test cases where the source of an LDGR needs to be spilled.
; We shouldn't have any integer stack stores or floating-point loads.
define void @f12() {
; CHECK-LABEL: f12:
; CHECK: %loop
; CHECK-NOT: std {{.*}}(%r15)
; CHECK: %exit
; CHECK: foo@PLT
; CHECK-NOT: lg {{.*}}(%r15)
; CHECK: foo@PLT
; CHECK: br %r14
entry:
br label %loop
loop:
%int0 = phi i64 [ 0, %entry ], [ %add0, %loop ]
%int1 = phi i64 [ 0, %entry ], [ %add1, %loop ]
%int2 = phi i64 [ 0, %entry ], [ %add2, %loop ]
%int3 = phi i64 [ 0, %entry ], [ %add3, %loop ]
%int4 = phi i64 [ 0, %entry ], [ %add4, %loop ]
%int5 = phi i64 [ 0, %entry ], [ %add5, %loop ]
%int6 = phi i64 [ 0, %entry ], [ %add6, %loop ]
%int7 = phi i64 [ 0, %entry ], [ %add7, %loop ]
%int8 = phi i64 [ 0, %entry ], [ %add8, %loop ]
%int9 = phi i64 [ 0, %entry ], [ %add9, %loop ]
%bias = call i64 @foo()
%add0 = add i64 %int0, %bias
%add1 = add i64 %int1, %bias
%add2 = add i64 %int2, %bias
%add3 = add i64 %int3, %bias
%add4 = add i64 %int4, %bias
%add5 = add i64 %int5, %bias
%add6 = add i64 %int6, %bias
%add7 = add i64 %int7, %bias
%add8 = add i64 %int8, %bias
%add9 = add i64 %int9, %bias
%cont = icmp ne i64 %bias, 1
br i1 %cont, label %loop, label %exit
exit:
%unused1 = call i64 @foo()
%factor = load volatile double *@dptr
%conv0 = bitcast i64 %add0 to double
%mul0 = fmul double %conv0, %factor
store volatile double %mul0, double *@dptr
%conv1 = bitcast i64 %add1 to double
%mul1 = fmul double %conv1, %factor
store volatile double %mul1, double *@dptr
%conv2 = bitcast i64 %add2 to double
%mul2 = fmul double %conv2, %factor
store volatile double %mul2, double *@dptr
%conv3 = bitcast i64 %add3 to double
%mul3 = fmul double %conv3, %factor
store volatile double %mul3, double *@dptr
%conv4 = bitcast i64 %add4 to double
%mul4 = fmul double %conv4, %factor
store volatile double %mul4, double *@dptr
%conv5 = bitcast i64 %add5 to double
%mul5 = fmul double %conv5, %factor
store volatile double %mul5, double *@dptr
%conv6 = bitcast i64 %add6 to double
%mul6 = fmul double %conv6, %factor
store volatile double %mul6, double *@dptr
%conv7 = bitcast i64 %add7 to double
%mul7 = fmul double %conv7, %factor
store volatile double %mul7, double *@dptr
%conv8 = bitcast i64 %add8 to double
%mul8 = fmul double %conv8, %factor
store volatile double %mul8, double *@dptr
%conv9 = bitcast i64 %add9 to double
%mul9 = fmul double %conv9, %factor
store volatile double %mul9, double *@dptr
%unused2 = call i64 @foo()
ret void
}
; ...likewise LGDR, with the requirements the other way around.
define void @f13() {
; CHECK-LABEL: f13:
; CHECK: %loop
; CHECK-NOT: stg {{.*}}(%r15)
; CHECK: %exit
; CHECK: foo@PLT
; CHECK-NOT: ld {{.*}}(%r15)
; CHECK: foo@PLT
; CHECK: br %r14
entry:
br label %loop
loop:
%double0 = phi double [ 1.0, %entry ], [ %mul0, %loop ]
%double1 = phi double [ 1.0, %entry ], [ %mul1, %loop ]
%double2 = phi double [ 1.0, %entry ], [ %mul2, %loop ]
%double3 = phi double [ 1.0, %entry ], [ %mul3, %loop ]
%double4 = phi double [ 1.0, %entry ], [ %mul4, %loop ]
%double5 = phi double [ 1.0, %entry ], [ %mul5, %loop ]
%double6 = phi double [ 1.0, %entry ], [ %mul6, %loop ]
%double7 = phi double [ 1.0, %entry ], [ %mul7, %loop ]
%double8 = phi double [ 1.0, %entry ], [ %mul8, %loop ]
%double9 = phi double [ 1.0, %entry ], [ %mul9, %loop ]
%factor = call double @bar()
%mul0 = fmul double %double0, %factor
%mul1 = fmul double %double1, %factor
%mul2 = fmul double %double2, %factor
%mul3 = fmul double %double3, %factor
%mul4 = fmul double %double4, %factor
%mul5 = fmul double %double5, %factor
%mul6 = fmul double %double6, %factor
%mul7 = fmul double %double7, %factor
%mul8 = fmul double %double8, %factor
%mul9 = fmul double %double9, %factor
%cont = fcmp one double %factor, 1.0
br i1 %cont, label %loop, label %exit
exit:
%unused1 = call i64 @foo()
%bias = load volatile i64 *@iptr
%conv0 = bitcast double %mul0 to i64
%add0 = add i64 %conv0, %bias
store volatile i64 %add0, i64 *@iptr
%conv1 = bitcast double %mul1 to i64
%add1 = add i64 %conv1, %bias
store volatile i64 %add1, i64 *@iptr
%conv2 = bitcast double %mul2 to i64
%add2 = add i64 %conv2, %bias
store volatile i64 %add2, i64 *@iptr
%conv3 = bitcast double %mul3 to i64
%add3 = add i64 %conv3, %bias
store volatile i64 %add3, i64 *@iptr
%conv4 = bitcast double %mul4 to i64
%add4 = add i64 %conv4, %bias
store volatile i64 %add4, i64 *@iptr
%conv5 = bitcast double %mul5 to i64
%add5 = add i64 %conv5, %bias
store volatile i64 %add5, i64 *@iptr
%conv6 = bitcast double %mul6 to i64
%add6 = add i64 %conv6, %bias
store volatile i64 %add6, i64 *@iptr
%conv7 = bitcast double %mul7 to i64
%add7 = add i64 %conv7, %bias
store volatile i64 %add7, i64 *@iptr
%conv8 = bitcast double %mul8 to i64
%add8 = add i64 %conv8, %bias
store volatile i64 %add8, i64 *@iptr
%conv9 = bitcast double %mul9 to i64
%add9 = add i64 %conv9, %bias
store volatile i64 %add9, i64 *@iptr
%unused2 = call i64 @foo()
ret void
}