mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 04:30:12 +00:00
3b87f6204f
misched is disabled by default. With -enable-misched, these heuristics balance the schedule to simultaneously avoid saturating processor resources, expose ILP, and minimize register pressure. I've been analyzing the performance of these heuristics on everything in the llvm test suite in addition to a few other benchmarks. I would like each heuristic check to be verified by a unit test, but I'm still trying to figure out the best way to do that. The heuristics are still in considerable flux, but as they are refined we should be rigorous about unit testing the improvements. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@167527 91177308-0d34-0410-b5e6-96231b3b80d8
231 lines
9.5 KiB
LLVM
231 lines
9.5 KiB
LLVM
; RUN: llc < %s -march=x86-64 -mcpu=core2 -pre-RA-sched=source -enable-misched \
|
|
; RUN: -verify-machineinstrs | FileCheck %s
|
|
;
|
|
; Verify that misched resource/latency balancy heuristics are sane.
|
|
|
|
define void @unrolled_mmult1(i32* %tmp55, i32* %tmp56, i32* %pre, i32* %pre94,
|
|
i32* %pre95, i32* %pre96, i32* %pre97, i32* %pre98, i32* %pre99,
|
|
i32* %pre100, i32* %pre101, i32* %pre102, i32* %pre103, i32* %pre104)
|
|
nounwind uwtable ssp {
|
|
entry:
|
|
br label %for.body
|
|
|
|
; imull folded loads should be in order and interleaved with addl, never
|
|
; adjacent. Also check that we have no spilling.
|
|
;
|
|
; Since mmult1 IR is already in good order, this effectively ensure
|
|
; the scheduler maintains source order.
|
|
;
|
|
; CHECK: %for.body
|
|
; CHECK-NOT: %rsp
|
|
; CHECK: imull 4
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 8
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 12
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 16
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 20
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 24
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 28
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 32
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 36
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: %end
|
|
for.body:
|
|
%indvars.iv42.i = phi i64 [ %indvars.iv.next43.i, %for.body ], [ 0, %entry ]
|
|
%tmp57 = load i32* %tmp56, align 4
|
|
%arrayidx12.us.i61 = getelementptr inbounds i32* %pre, i64 %indvars.iv42.i
|
|
%tmp58 = load i32* %arrayidx12.us.i61, align 4
|
|
%mul.us.i = mul nsw i32 %tmp58, %tmp57
|
|
%arrayidx8.us.i.1 = getelementptr inbounds i32* %tmp56, i64 1
|
|
%tmp59 = load i32* %arrayidx8.us.i.1, align 4
|
|
%arrayidx12.us.i61.1 = getelementptr inbounds i32* %pre94, i64 %indvars.iv42.i
|
|
%tmp60 = load i32* %arrayidx12.us.i61.1, align 4
|
|
%mul.us.i.1 = mul nsw i32 %tmp60, %tmp59
|
|
%add.us.i.1 = add nsw i32 %mul.us.i.1, %mul.us.i
|
|
%arrayidx8.us.i.2 = getelementptr inbounds i32* %tmp56, i64 2
|
|
%tmp61 = load i32* %arrayidx8.us.i.2, align 4
|
|
%arrayidx12.us.i61.2 = getelementptr inbounds i32* %pre95, i64 %indvars.iv42.i
|
|
%tmp62 = load i32* %arrayidx12.us.i61.2, align 4
|
|
%mul.us.i.2 = mul nsw i32 %tmp62, %tmp61
|
|
%add.us.i.2 = add nsw i32 %mul.us.i.2, %add.us.i.1
|
|
%arrayidx8.us.i.3 = getelementptr inbounds i32* %tmp56, i64 3
|
|
%tmp63 = load i32* %arrayidx8.us.i.3, align 4
|
|
%arrayidx12.us.i61.3 = getelementptr inbounds i32* %pre96, i64 %indvars.iv42.i
|
|
%tmp64 = load i32* %arrayidx12.us.i61.3, align 4
|
|
%mul.us.i.3 = mul nsw i32 %tmp64, %tmp63
|
|
%add.us.i.3 = add nsw i32 %mul.us.i.3, %add.us.i.2
|
|
%arrayidx8.us.i.4 = getelementptr inbounds i32* %tmp56, i64 4
|
|
%tmp65 = load i32* %arrayidx8.us.i.4, align 4
|
|
%arrayidx12.us.i61.4 = getelementptr inbounds i32* %pre97, i64 %indvars.iv42.i
|
|
%tmp66 = load i32* %arrayidx12.us.i61.4, align 4
|
|
%mul.us.i.4 = mul nsw i32 %tmp66, %tmp65
|
|
%add.us.i.4 = add nsw i32 %mul.us.i.4, %add.us.i.3
|
|
%arrayidx8.us.i.5 = getelementptr inbounds i32* %tmp56, i64 5
|
|
%tmp67 = load i32* %arrayidx8.us.i.5, align 4
|
|
%arrayidx12.us.i61.5 = getelementptr inbounds i32* %pre98, i64 %indvars.iv42.i
|
|
%tmp68 = load i32* %arrayidx12.us.i61.5, align 4
|
|
%mul.us.i.5 = mul nsw i32 %tmp68, %tmp67
|
|
%add.us.i.5 = add nsw i32 %mul.us.i.5, %add.us.i.4
|
|
%arrayidx8.us.i.6 = getelementptr inbounds i32* %tmp56, i64 6
|
|
%tmp69 = load i32* %arrayidx8.us.i.6, align 4
|
|
%arrayidx12.us.i61.6 = getelementptr inbounds i32* %pre99, i64 %indvars.iv42.i
|
|
%tmp70 = load i32* %arrayidx12.us.i61.6, align 4
|
|
%mul.us.i.6 = mul nsw i32 %tmp70, %tmp69
|
|
%add.us.i.6 = add nsw i32 %mul.us.i.6, %add.us.i.5
|
|
%arrayidx8.us.i.7 = getelementptr inbounds i32* %tmp56, i64 7
|
|
%tmp71 = load i32* %arrayidx8.us.i.7, align 4
|
|
%arrayidx12.us.i61.7 = getelementptr inbounds i32* %pre100, i64 %indvars.iv42.i
|
|
%tmp72 = load i32* %arrayidx12.us.i61.7, align 4
|
|
%mul.us.i.7 = mul nsw i32 %tmp72, %tmp71
|
|
%add.us.i.7 = add nsw i32 %mul.us.i.7, %add.us.i.6
|
|
%arrayidx8.us.i.8 = getelementptr inbounds i32* %tmp56, i64 8
|
|
%tmp73 = load i32* %arrayidx8.us.i.8, align 4
|
|
%arrayidx12.us.i61.8 = getelementptr inbounds i32* %pre101, i64 %indvars.iv42.i
|
|
%tmp74 = load i32* %arrayidx12.us.i61.8, align 4
|
|
%mul.us.i.8 = mul nsw i32 %tmp74, %tmp73
|
|
%add.us.i.8 = add nsw i32 %mul.us.i.8, %add.us.i.7
|
|
%arrayidx8.us.i.9 = getelementptr inbounds i32* %tmp56, i64 9
|
|
%tmp75 = load i32* %arrayidx8.us.i.9, align 4
|
|
%arrayidx12.us.i61.9 = getelementptr inbounds i32* %pre102, i64 %indvars.iv42.i
|
|
%tmp76 = load i32* %arrayidx12.us.i61.9, align 4
|
|
%mul.us.i.9 = mul nsw i32 %tmp76, %tmp75
|
|
%add.us.i.9 = add nsw i32 %mul.us.i.9, %add.us.i.8
|
|
%arrayidx16.us.i = getelementptr inbounds i32* %tmp55, i64 %indvars.iv42.i
|
|
store i32 %add.us.i.9, i32* %arrayidx16.us.i, align 4
|
|
%indvars.iv.next43.i = add i64 %indvars.iv42.i, 1
|
|
%lftr.wideiv = trunc i64 %indvars.iv.next43.i to i32
|
|
%exitcond = icmp eq i32 %lftr.wideiv, 10
|
|
br i1 %exitcond, label %end, label %for.body
|
|
|
|
end:
|
|
ret void
|
|
}
|
|
|
|
; Unlike the above loop, this IR starts out bad and must be
|
|
; rescheduled.
|
|
;
|
|
; CHECK: %for.body
|
|
; CHECK-NOT: %rsp
|
|
; CHECK: imull 4
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 8
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 12
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 16
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 20
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 24
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 28
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 32
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK: imull 36
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: addl
|
|
; CHECK-NOT: {{imull|rsp}}
|
|
; CHECK: %end
|
|
define void @unrolled_mmult2(i32* %tmp55, i32* %tmp56, i32* %pre, i32* %pre94,
|
|
i32* %pre95, i32* %pre96, i32* %pre97, i32* %pre98, i32* %pre99,
|
|
i32* %pre100, i32* %pre101, i32* %pre102, i32* %pre103, i32* %pre104)
|
|
nounwind uwtable ssp {
|
|
entry:
|
|
br label %for.body
|
|
for.body:
|
|
%indvars.iv42.i = phi i64 [ %indvars.iv.next43.i, %for.body ], [ 0, %entry ]
|
|
%tmp57 = load i32* %tmp56, align 4
|
|
%arrayidx12.us.i61 = getelementptr inbounds i32* %pre, i64 %indvars.iv42.i
|
|
%tmp58 = load i32* %arrayidx12.us.i61, align 4
|
|
%arrayidx8.us.i.1 = getelementptr inbounds i32* %tmp56, i64 1
|
|
%tmp59 = load i32* %arrayidx8.us.i.1, align 4
|
|
%arrayidx12.us.i61.1 = getelementptr inbounds i32* %pre94, i64 %indvars.iv42.i
|
|
%tmp60 = load i32* %arrayidx12.us.i61.1, align 4
|
|
%arrayidx8.us.i.2 = getelementptr inbounds i32* %tmp56, i64 2
|
|
%tmp61 = load i32* %arrayidx8.us.i.2, align 4
|
|
%arrayidx12.us.i61.2 = getelementptr inbounds i32* %pre95, i64 %indvars.iv42.i
|
|
%tmp62 = load i32* %arrayidx12.us.i61.2, align 4
|
|
%arrayidx8.us.i.3 = getelementptr inbounds i32* %tmp56, i64 3
|
|
%tmp63 = load i32* %arrayidx8.us.i.3, align 4
|
|
%arrayidx12.us.i61.3 = getelementptr inbounds i32* %pre96, i64 %indvars.iv42.i
|
|
%tmp64 = load i32* %arrayidx12.us.i61.3, align 4
|
|
%arrayidx8.us.i.4 = getelementptr inbounds i32* %tmp56, i64 4
|
|
%tmp65 = load i32* %arrayidx8.us.i.4, align 4
|
|
%arrayidx12.us.i61.4 = getelementptr inbounds i32* %pre97, i64 %indvars.iv42.i
|
|
%tmp66 = load i32* %arrayidx12.us.i61.4, align 4
|
|
%arrayidx8.us.i.5 = getelementptr inbounds i32* %tmp56, i64 5
|
|
%tmp67 = load i32* %arrayidx8.us.i.5, align 4
|
|
%arrayidx12.us.i61.5 = getelementptr inbounds i32* %pre98, i64 %indvars.iv42.i
|
|
%tmp68 = load i32* %arrayidx12.us.i61.5, align 4
|
|
%arrayidx8.us.i.6 = getelementptr inbounds i32* %tmp56, i64 6
|
|
%tmp69 = load i32* %arrayidx8.us.i.6, align 4
|
|
%arrayidx12.us.i61.6 = getelementptr inbounds i32* %pre99, i64 %indvars.iv42.i
|
|
%tmp70 = load i32* %arrayidx12.us.i61.6, align 4
|
|
%mul.us.i = mul nsw i32 %tmp58, %tmp57
|
|
%arrayidx8.us.i.7 = getelementptr inbounds i32* %tmp56, i64 7
|
|
%tmp71 = load i32* %arrayidx8.us.i.7, align 4
|
|
%arrayidx12.us.i61.7 = getelementptr inbounds i32* %pre100, i64 %indvars.iv42.i
|
|
%tmp72 = load i32* %arrayidx12.us.i61.7, align 4
|
|
%arrayidx8.us.i.8 = getelementptr inbounds i32* %tmp56, i64 8
|
|
%tmp73 = load i32* %arrayidx8.us.i.8, align 4
|
|
%arrayidx12.us.i61.8 = getelementptr inbounds i32* %pre101, i64 %indvars.iv42.i
|
|
%tmp74 = load i32* %arrayidx12.us.i61.8, align 4
|
|
%arrayidx8.us.i.9 = getelementptr inbounds i32* %tmp56, i64 9
|
|
%tmp75 = load i32* %arrayidx8.us.i.9, align 4
|
|
%arrayidx12.us.i61.9 = getelementptr inbounds i32* %pre102, i64 %indvars.iv42.i
|
|
%tmp76 = load i32* %arrayidx12.us.i61.9, align 4
|
|
%mul.us.i.1 = mul nsw i32 %tmp60, %tmp59
|
|
%add.us.i.1 = add nsw i32 %mul.us.i.1, %mul.us.i
|
|
%mul.us.i.2 = mul nsw i32 %tmp62, %tmp61
|
|
%add.us.i.2 = add nsw i32 %mul.us.i.2, %add.us.i.1
|
|
%mul.us.i.3 = mul nsw i32 %tmp64, %tmp63
|
|
%add.us.i.3 = add nsw i32 %mul.us.i.3, %add.us.i.2
|
|
%mul.us.i.4 = mul nsw i32 %tmp66, %tmp65
|
|
%add.us.i.4 = add nsw i32 %mul.us.i.4, %add.us.i.3
|
|
%mul.us.i.5 = mul nsw i32 %tmp68, %tmp67
|
|
%add.us.i.5 = add nsw i32 %mul.us.i.5, %add.us.i.4
|
|
%mul.us.i.6 = mul nsw i32 %tmp70, %tmp69
|
|
%add.us.i.6 = add nsw i32 %mul.us.i.6, %add.us.i.5
|
|
%mul.us.i.7 = mul nsw i32 %tmp72, %tmp71
|
|
%add.us.i.7 = add nsw i32 %mul.us.i.7, %add.us.i.6
|
|
%mul.us.i.8 = mul nsw i32 %tmp74, %tmp73
|
|
%add.us.i.8 = add nsw i32 %mul.us.i.8, %add.us.i.7
|
|
%mul.us.i.9 = mul nsw i32 %tmp76, %tmp75
|
|
%add.us.i.9 = add nsw i32 %mul.us.i.9, %add.us.i.8
|
|
%arrayidx16.us.i = getelementptr inbounds i32* %tmp55, i64 %indvars.iv42.i
|
|
store i32 %add.us.i.9, i32* %arrayidx16.us.i, align 4
|
|
%indvars.iv.next43.i = add i64 %indvars.iv42.i, 1
|
|
%lftr.wideiv = trunc i64 %indvars.iv.next43.i to i32
|
|
%exitcond = icmp eq i32 %lftr.wideiv, 10
|
|
br i1 %exitcond, label %end, label %for.body
|
|
|
|
end:
|
|
ret void
|
|
}
|