mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-29 10:32:47 +00:00
misched: Heuristics based on the machine model.
misched is disabled by default. With -enable-misched, these heuristics balance the schedule to simultaneously avoid saturating processor resources, expose ILP, and minimize register pressure. I've been analyzing the performance of these heuristics on everything in the llvm test suite in addition to a few other benchmarks. I would like each heuristic check to be verified by a unit test, but I'm still trying to figure out the best way to do that. The heuristics are still in considerable flux, but as they are refined we should be rigorous about unit testing the improvements. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@167527 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
881a05b46c
commit
3b87f6204f
@ -154,6 +154,8 @@ public:
|
||||
|
||||
bool empty() const { return Queue.empty(); }
|
||||
|
||||
void clear() { Queue.clear(); }
|
||||
|
||||
unsigned size() const { return Queue.size(); }
|
||||
|
||||
typedef std::vector<SUnit*>::iterator iterator;
|
||||
@ -171,10 +173,12 @@ public:
|
||||
SU->NodeQueueId |= ID;
|
||||
}
|
||||
|
||||
void remove(iterator I) {
|
||||
iterator remove(iterator I) {
|
||||
(*I)->NodeQueueId &= ~ID;
|
||||
*I = Queue.back();
|
||||
unsigned idx = I - Queue.begin();
|
||||
Queue.pop_back();
|
||||
return Queue.begin() + idx;
|
||||
}
|
||||
|
||||
#ifndef NDEBUG
|
||||
@ -306,6 +310,9 @@ protected:
|
||||
/// Reinsert debug_values recorded in ScheduleDAGInstrs::DbgValues.
|
||||
void placeDebugValues();
|
||||
|
||||
/// \brief dump the scheduled Sequence.
|
||||
void dumpSchedule() const;
|
||||
|
||||
// Lesser helpers...
|
||||
|
||||
void initRegPressure();
|
||||
|
File diff suppressed because it is too large
Load Diff
230
test/CodeGen/X86/misched-balance.ll
Normal file
230
test/CodeGen/X86/misched-balance.ll
Normal file
@ -0,0 +1,230 @@
|
||||
; RUN: llc < %s -march=x86-64 -mcpu=core2 -pre-RA-sched=source -enable-misched \
|
||||
; RUN: -verify-machineinstrs | FileCheck %s
|
||||
;
|
||||
; Verify that misched resource/latency balancy heuristics are sane.
|
||||
|
||||
define void @unrolled_mmult1(i32* %tmp55, i32* %tmp56, i32* %pre, i32* %pre94,
|
||||
i32* %pre95, i32* %pre96, i32* %pre97, i32* %pre98, i32* %pre99,
|
||||
i32* %pre100, i32* %pre101, i32* %pre102, i32* %pre103, i32* %pre104)
|
||||
nounwind uwtable ssp {
|
||||
entry:
|
||||
br label %for.body
|
||||
|
||||
; imull folded loads should be in order and interleaved with addl, never
|
||||
; adjacent. Also check that we have no spilling.
|
||||
;
|
||||
; Since mmult1 IR is already in good order, this effectively ensure
|
||||
; the scheduler maintains source order.
|
||||
;
|
||||
; CHECK: %for.body
|
||||
; CHECK-NOT: %rsp
|
||||
; CHECK: imull 4
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 8
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 12
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 16
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 20
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 24
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 28
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 32
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 36
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: %end
|
||||
for.body:
|
||||
%indvars.iv42.i = phi i64 [ %indvars.iv.next43.i, %for.body ], [ 0, %entry ]
|
||||
%tmp57 = load i32* %tmp56, align 4
|
||||
%arrayidx12.us.i61 = getelementptr inbounds i32* %pre, i64 %indvars.iv42.i
|
||||
%tmp58 = load i32* %arrayidx12.us.i61, align 4
|
||||
%mul.us.i = mul nsw i32 %tmp58, %tmp57
|
||||
%arrayidx8.us.i.1 = getelementptr inbounds i32* %tmp56, i64 1
|
||||
%tmp59 = load i32* %arrayidx8.us.i.1, align 4
|
||||
%arrayidx12.us.i61.1 = getelementptr inbounds i32* %pre94, i64 %indvars.iv42.i
|
||||
%tmp60 = load i32* %arrayidx12.us.i61.1, align 4
|
||||
%mul.us.i.1 = mul nsw i32 %tmp60, %tmp59
|
||||
%add.us.i.1 = add nsw i32 %mul.us.i.1, %mul.us.i
|
||||
%arrayidx8.us.i.2 = getelementptr inbounds i32* %tmp56, i64 2
|
||||
%tmp61 = load i32* %arrayidx8.us.i.2, align 4
|
||||
%arrayidx12.us.i61.2 = getelementptr inbounds i32* %pre95, i64 %indvars.iv42.i
|
||||
%tmp62 = load i32* %arrayidx12.us.i61.2, align 4
|
||||
%mul.us.i.2 = mul nsw i32 %tmp62, %tmp61
|
||||
%add.us.i.2 = add nsw i32 %mul.us.i.2, %add.us.i.1
|
||||
%arrayidx8.us.i.3 = getelementptr inbounds i32* %tmp56, i64 3
|
||||
%tmp63 = load i32* %arrayidx8.us.i.3, align 4
|
||||
%arrayidx12.us.i61.3 = getelementptr inbounds i32* %pre96, i64 %indvars.iv42.i
|
||||
%tmp64 = load i32* %arrayidx12.us.i61.3, align 4
|
||||
%mul.us.i.3 = mul nsw i32 %tmp64, %tmp63
|
||||
%add.us.i.3 = add nsw i32 %mul.us.i.3, %add.us.i.2
|
||||
%arrayidx8.us.i.4 = getelementptr inbounds i32* %tmp56, i64 4
|
||||
%tmp65 = load i32* %arrayidx8.us.i.4, align 4
|
||||
%arrayidx12.us.i61.4 = getelementptr inbounds i32* %pre97, i64 %indvars.iv42.i
|
||||
%tmp66 = load i32* %arrayidx12.us.i61.4, align 4
|
||||
%mul.us.i.4 = mul nsw i32 %tmp66, %tmp65
|
||||
%add.us.i.4 = add nsw i32 %mul.us.i.4, %add.us.i.3
|
||||
%arrayidx8.us.i.5 = getelementptr inbounds i32* %tmp56, i64 5
|
||||
%tmp67 = load i32* %arrayidx8.us.i.5, align 4
|
||||
%arrayidx12.us.i61.5 = getelementptr inbounds i32* %pre98, i64 %indvars.iv42.i
|
||||
%tmp68 = load i32* %arrayidx12.us.i61.5, align 4
|
||||
%mul.us.i.5 = mul nsw i32 %tmp68, %tmp67
|
||||
%add.us.i.5 = add nsw i32 %mul.us.i.5, %add.us.i.4
|
||||
%arrayidx8.us.i.6 = getelementptr inbounds i32* %tmp56, i64 6
|
||||
%tmp69 = load i32* %arrayidx8.us.i.6, align 4
|
||||
%arrayidx12.us.i61.6 = getelementptr inbounds i32* %pre99, i64 %indvars.iv42.i
|
||||
%tmp70 = load i32* %arrayidx12.us.i61.6, align 4
|
||||
%mul.us.i.6 = mul nsw i32 %tmp70, %tmp69
|
||||
%add.us.i.6 = add nsw i32 %mul.us.i.6, %add.us.i.5
|
||||
%arrayidx8.us.i.7 = getelementptr inbounds i32* %tmp56, i64 7
|
||||
%tmp71 = load i32* %arrayidx8.us.i.7, align 4
|
||||
%arrayidx12.us.i61.7 = getelementptr inbounds i32* %pre100, i64 %indvars.iv42.i
|
||||
%tmp72 = load i32* %arrayidx12.us.i61.7, align 4
|
||||
%mul.us.i.7 = mul nsw i32 %tmp72, %tmp71
|
||||
%add.us.i.7 = add nsw i32 %mul.us.i.7, %add.us.i.6
|
||||
%arrayidx8.us.i.8 = getelementptr inbounds i32* %tmp56, i64 8
|
||||
%tmp73 = load i32* %arrayidx8.us.i.8, align 4
|
||||
%arrayidx12.us.i61.8 = getelementptr inbounds i32* %pre101, i64 %indvars.iv42.i
|
||||
%tmp74 = load i32* %arrayidx12.us.i61.8, align 4
|
||||
%mul.us.i.8 = mul nsw i32 %tmp74, %tmp73
|
||||
%add.us.i.8 = add nsw i32 %mul.us.i.8, %add.us.i.7
|
||||
%arrayidx8.us.i.9 = getelementptr inbounds i32* %tmp56, i64 9
|
||||
%tmp75 = load i32* %arrayidx8.us.i.9, align 4
|
||||
%arrayidx12.us.i61.9 = getelementptr inbounds i32* %pre102, i64 %indvars.iv42.i
|
||||
%tmp76 = load i32* %arrayidx12.us.i61.9, align 4
|
||||
%mul.us.i.9 = mul nsw i32 %tmp76, %tmp75
|
||||
%add.us.i.9 = add nsw i32 %mul.us.i.9, %add.us.i.8
|
||||
%arrayidx16.us.i = getelementptr inbounds i32* %tmp55, i64 %indvars.iv42.i
|
||||
store i32 %add.us.i.9, i32* %arrayidx16.us.i, align 4
|
||||
%indvars.iv.next43.i = add i64 %indvars.iv42.i, 1
|
||||
%lftr.wideiv = trunc i64 %indvars.iv.next43.i to i32
|
||||
%exitcond = icmp eq i32 %lftr.wideiv, 10
|
||||
br i1 %exitcond, label %end, label %for.body
|
||||
|
||||
end:
|
||||
ret void
|
||||
}
|
||||
|
||||
; Unlike the above loop, this IR starts out bad and must be
|
||||
; rescheduled.
|
||||
;
|
||||
; CHECK: %for.body
|
||||
; CHECK-NOT: %rsp
|
||||
; CHECK: imull 4
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 8
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 12
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 16
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 20
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 24
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 28
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 32
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK: imull 36
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: addl
|
||||
; CHECK-NOT: {{imull|rsp}}
|
||||
; CHECK: %end
|
||||
define void @unrolled_mmult2(i32* %tmp55, i32* %tmp56, i32* %pre, i32* %pre94,
|
||||
i32* %pre95, i32* %pre96, i32* %pre97, i32* %pre98, i32* %pre99,
|
||||
i32* %pre100, i32* %pre101, i32* %pre102, i32* %pre103, i32* %pre104)
|
||||
nounwind uwtable ssp {
|
||||
entry:
|
||||
br label %for.body
|
||||
for.body:
|
||||
%indvars.iv42.i = phi i64 [ %indvars.iv.next43.i, %for.body ], [ 0, %entry ]
|
||||
%tmp57 = load i32* %tmp56, align 4
|
||||
%arrayidx12.us.i61 = getelementptr inbounds i32* %pre, i64 %indvars.iv42.i
|
||||
%tmp58 = load i32* %arrayidx12.us.i61, align 4
|
||||
%arrayidx8.us.i.1 = getelementptr inbounds i32* %tmp56, i64 1
|
||||
%tmp59 = load i32* %arrayidx8.us.i.1, align 4
|
||||
%arrayidx12.us.i61.1 = getelementptr inbounds i32* %pre94, i64 %indvars.iv42.i
|
||||
%tmp60 = load i32* %arrayidx12.us.i61.1, align 4
|
||||
%arrayidx8.us.i.2 = getelementptr inbounds i32* %tmp56, i64 2
|
||||
%tmp61 = load i32* %arrayidx8.us.i.2, align 4
|
||||
%arrayidx12.us.i61.2 = getelementptr inbounds i32* %pre95, i64 %indvars.iv42.i
|
||||
%tmp62 = load i32* %arrayidx12.us.i61.2, align 4
|
||||
%arrayidx8.us.i.3 = getelementptr inbounds i32* %tmp56, i64 3
|
||||
%tmp63 = load i32* %arrayidx8.us.i.3, align 4
|
||||
%arrayidx12.us.i61.3 = getelementptr inbounds i32* %pre96, i64 %indvars.iv42.i
|
||||
%tmp64 = load i32* %arrayidx12.us.i61.3, align 4
|
||||
%arrayidx8.us.i.4 = getelementptr inbounds i32* %tmp56, i64 4
|
||||
%tmp65 = load i32* %arrayidx8.us.i.4, align 4
|
||||
%arrayidx12.us.i61.4 = getelementptr inbounds i32* %pre97, i64 %indvars.iv42.i
|
||||
%tmp66 = load i32* %arrayidx12.us.i61.4, align 4
|
||||
%arrayidx8.us.i.5 = getelementptr inbounds i32* %tmp56, i64 5
|
||||
%tmp67 = load i32* %arrayidx8.us.i.5, align 4
|
||||
%arrayidx12.us.i61.5 = getelementptr inbounds i32* %pre98, i64 %indvars.iv42.i
|
||||
%tmp68 = load i32* %arrayidx12.us.i61.5, align 4
|
||||
%arrayidx8.us.i.6 = getelementptr inbounds i32* %tmp56, i64 6
|
||||
%tmp69 = load i32* %arrayidx8.us.i.6, align 4
|
||||
%arrayidx12.us.i61.6 = getelementptr inbounds i32* %pre99, i64 %indvars.iv42.i
|
||||
%tmp70 = load i32* %arrayidx12.us.i61.6, align 4
|
||||
%mul.us.i = mul nsw i32 %tmp58, %tmp57
|
||||
%arrayidx8.us.i.7 = getelementptr inbounds i32* %tmp56, i64 7
|
||||
%tmp71 = load i32* %arrayidx8.us.i.7, align 4
|
||||
%arrayidx12.us.i61.7 = getelementptr inbounds i32* %pre100, i64 %indvars.iv42.i
|
||||
%tmp72 = load i32* %arrayidx12.us.i61.7, align 4
|
||||
%arrayidx8.us.i.8 = getelementptr inbounds i32* %tmp56, i64 8
|
||||
%tmp73 = load i32* %arrayidx8.us.i.8, align 4
|
||||
%arrayidx12.us.i61.8 = getelementptr inbounds i32* %pre101, i64 %indvars.iv42.i
|
||||
%tmp74 = load i32* %arrayidx12.us.i61.8, align 4
|
||||
%arrayidx8.us.i.9 = getelementptr inbounds i32* %tmp56, i64 9
|
||||
%tmp75 = load i32* %arrayidx8.us.i.9, align 4
|
||||
%arrayidx12.us.i61.9 = getelementptr inbounds i32* %pre102, i64 %indvars.iv42.i
|
||||
%tmp76 = load i32* %arrayidx12.us.i61.9, align 4
|
||||
%mul.us.i.1 = mul nsw i32 %tmp60, %tmp59
|
||||
%add.us.i.1 = add nsw i32 %mul.us.i.1, %mul.us.i
|
||||
%mul.us.i.2 = mul nsw i32 %tmp62, %tmp61
|
||||
%add.us.i.2 = add nsw i32 %mul.us.i.2, %add.us.i.1
|
||||
%mul.us.i.3 = mul nsw i32 %tmp64, %tmp63
|
||||
%add.us.i.3 = add nsw i32 %mul.us.i.3, %add.us.i.2
|
||||
%mul.us.i.4 = mul nsw i32 %tmp66, %tmp65
|
||||
%add.us.i.4 = add nsw i32 %mul.us.i.4, %add.us.i.3
|
||||
%mul.us.i.5 = mul nsw i32 %tmp68, %tmp67
|
||||
%add.us.i.5 = add nsw i32 %mul.us.i.5, %add.us.i.4
|
||||
%mul.us.i.6 = mul nsw i32 %tmp70, %tmp69
|
||||
%add.us.i.6 = add nsw i32 %mul.us.i.6, %add.us.i.5
|
||||
%mul.us.i.7 = mul nsw i32 %tmp72, %tmp71
|
||||
%add.us.i.7 = add nsw i32 %mul.us.i.7, %add.us.i.6
|
||||
%mul.us.i.8 = mul nsw i32 %tmp74, %tmp73
|
||||
%add.us.i.8 = add nsw i32 %mul.us.i.8, %add.us.i.7
|
||||
%mul.us.i.9 = mul nsw i32 %tmp76, %tmp75
|
||||
%add.us.i.9 = add nsw i32 %mul.us.i.9, %add.us.i.8
|
||||
%arrayidx16.us.i = getelementptr inbounds i32* %tmp55, i64 %indvars.iv42.i
|
||||
store i32 %add.us.i.9, i32* %arrayidx16.us.i, align 4
|
||||
%indvars.iv.next43.i = add i64 %indvars.iv42.i, 1
|
||||
%lftr.wideiv = trunc i64 %indvars.iv.next43.i to i32
|
||||
%exitcond = icmp eq i32 %lftr.wideiv, 10
|
||||
br i1 %exitcond, label %end, label %for.body
|
||||
|
||||
end:
|
||||
ret void
|
||||
}
|
Loading…
Reference in New Issue
Block a user