mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 20:29:48 +00:00
12f0dc6bb5
This is done by pushing physical register definitions close to their use, which happens to handle flag definitions if they're not glued to the branch. This seems to be generally a good thing though, so I didn't need to add a target hook yet. The primary motivation is to generate code closer to what people expect and rule out missed opportunity from enabling macro-op fusion. As a side benefit, we get several 2-5% gains on x86 benchmarks. There is one regression: SingleSource/Benchmarks/Shootout/lists slows down be -10%. But this is an independent scheduler bug that will be tracked separately. See rdar://problem/9283108. Incidentally, pre-RA scheduling is only half the solution. Fixing the later passes is tracked by: <rdar://problem/8932804> [pre-RA-sched] on x86, attempt to schedule CMP/TEST adjacent with condition jump Fixes: <rdar://problem/9262453> Scheduler unnecessary break of cmp/jump fusion git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@129508 91177308-0d34-0410-b5e6-96231b3b80d8
138 lines
6.2 KiB
LLVM
138 lines
6.2 KiB
LLVM
; RUN: llc -march=x86-64 < %s | FileCheck %s
|
|
|
|
; CHECK: decq
|
|
; CHECK-NEXT: jne
|
|
|
|
@Te0 = external global [256 x i32] ; <[256 x i32]*> [#uses=5]
|
|
@Te1 = external global [256 x i32] ; <[256 x i32]*> [#uses=4]
|
|
@Te3 = external global [256 x i32] ; <[256 x i32]*> [#uses=2]
|
|
|
|
define void @t(i8* nocapture %in, i8* nocapture %out, i32* nocapture %rk, i32 %r) nounwind {
|
|
entry:
|
|
%0 = load i32* %rk, align 4 ; <i32> [#uses=1]
|
|
%1 = getelementptr i32* %rk, i64 1 ; <i32*> [#uses=1]
|
|
%2 = load i32* %1, align 4 ; <i32> [#uses=1]
|
|
%tmp15 = add i32 %r, -1 ; <i32> [#uses=1]
|
|
%tmp.16 = zext i32 %tmp15 to i64 ; <i64> [#uses=2]
|
|
br label %bb
|
|
|
|
bb: ; preds = %bb1, %entry
|
|
%indvar = phi i64 [ 0, %entry ], [ %indvar.next, %bb1 ] ; <i64> [#uses=3]
|
|
%s1.0 = phi i32 [ %2, %entry ], [ %56, %bb1 ] ; <i32> [#uses=2]
|
|
%s0.0 = phi i32 [ %0, %entry ], [ %43, %bb1 ] ; <i32> [#uses=2]
|
|
%tmp18 = shl i64 %indvar, 4 ; <i64> [#uses=4]
|
|
%rk26 = bitcast i32* %rk to i8* ; <i8*> [#uses=6]
|
|
%3 = lshr i32 %s0.0, 24 ; <i32> [#uses=1]
|
|
%4 = zext i32 %3 to i64 ; <i64> [#uses=1]
|
|
%5 = getelementptr [256 x i32]* @Te0, i64 0, i64 %4 ; <i32*> [#uses=1]
|
|
%6 = load i32* %5, align 4 ; <i32> [#uses=1]
|
|
%7 = lshr i32 %s1.0, 16 ; <i32> [#uses=1]
|
|
%8 = and i32 %7, 255 ; <i32> [#uses=1]
|
|
%9 = zext i32 %8 to i64 ; <i64> [#uses=1]
|
|
%10 = getelementptr [256 x i32]* @Te1, i64 0, i64 %9 ; <i32*> [#uses=1]
|
|
%11 = load i32* %10, align 4 ; <i32> [#uses=1]
|
|
%ctg2.sum2728 = or i64 %tmp18, 8 ; <i64> [#uses=1]
|
|
%12 = getelementptr i8* %rk26, i64 %ctg2.sum2728 ; <i8*> [#uses=1]
|
|
%13 = bitcast i8* %12 to i32* ; <i32*> [#uses=1]
|
|
%14 = load i32* %13, align 4 ; <i32> [#uses=1]
|
|
%15 = xor i32 %11, %6 ; <i32> [#uses=1]
|
|
%16 = xor i32 %15, %14 ; <i32> [#uses=3]
|
|
%17 = lshr i32 %s1.0, 24 ; <i32> [#uses=1]
|
|
%18 = zext i32 %17 to i64 ; <i64> [#uses=1]
|
|
%19 = getelementptr [256 x i32]* @Te0, i64 0, i64 %18 ; <i32*> [#uses=1]
|
|
%20 = load i32* %19, align 4 ; <i32> [#uses=1]
|
|
%21 = and i32 %s0.0, 255 ; <i32> [#uses=1]
|
|
%22 = zext i32 %21 to i64 ; <i64> [#uses=1]
|
|
%23 = getelementptr [256 x i32]* @Te3, i64 0, i64 %22 ; <i32*> [#uses=1]
|
|
%24 = load i32* %23, align 4 ; <i32> [#uses=1]
|
|
%ctg2.sum2930 = or i64 %tmp18, 12 ; <i64> [#uses=1]
|
|
%25 = getelementptr i8* %rk26, i64 %ctg2.sum2930 ; <i8*> [#uses=1]
|
|
%26 = bitcast i8* %25 to i32* ; <i32*> [#uses=1]
|
|
%27 = load i32* %26, align 4 ; <i32> [#uses=1]
|
|
%28 = xor i32 %24, %20 ; <i32> [#uses=1]
|
|
%29 = xor i32 %28, %27 ; <i32> [#uses=4]
|
|
%30 = lshr i32 %16, 24 ; <i32> [#uses=1]
|
|
%31 = zext i32 %30 to i64 ; <i64> [#uses=1]
|
|
%32 = getelementptr [256 x i32]* @Te0, i64 0, i64 %31 ; <i32*> [#uses=1]
|
|
%33 = load i32* %32, align 4 ; <i32> [#uses=2]
|
|
%exitcond = icmp eq i64 %indvar, %tmp.16 ; <i1> [#uses=1]
|
|
br i1 %exitcond, label %bb2, label %bb1
|
|
|
|
bb1: ; preds = %bb
|
|
%ctg2.sum31 = add i64 %tmp18, 16 ; <i64> [#uses=1]
|
|
%34 = getelementptr i8* %rk26, i64 %ctg2.sum31 ; <i8*> [#uses=1]
|
|
%35 = bitcast i8* %34 to i32* ; <i32*> [#uses=1]
|
|
%36 = lshr i32 %29, 16 ; <i32> [#uses=1]
|
|
%37 = and i32 %36, 255 ; <i32> [#uses=1]
|
|
%38 = zext i32 %37 to i64 ; <i64> [#uses=1]
|
|
%39 = getelementptr [256 x i32]* @Te1, i64 0, i64 %38 ; <i32*> [#uses=1]
|
|
%40 = load i32* %39, align 4 ; <i32> [#uses=1]
|
|
%41 = load i32* %35, align 4 ; <i32> [#uses=1]
|
|
%42 = xor i32 %40, %33 ; <i32> [#uses=1]
|
|
%43 = xor i32 %42, %41 ; <i32> [#uses=1]
|
|
%44 = lshr i32 %29, 24 ; <i32> [#uses=1]
|
|
%45 = zext i32 %44 to i64 ; <i64> [#uses=1]
|
|
%46 = getelementptr [256 x i32]* @Te0, i64 0, i64 %45 ; <i32*> [#uses=1]
|
|
%47 = load i32* %46, align 4 ; <i32> [#uses=1]
|
|
%48 = and i32 %16, 255 ; <i32> [#uses=1]
|
|
%49 = zext i32 %48 to i64 ; <i64> [#uses=1]
|
|
%50 = getelementptr [256 x i32]* @Te3, i64 0, i64 %49 ; <i32*> [#uses=1]
|
|
%51 = load i32* %50, align 4 ; <i32> [#uses=1]
|
|
%ctg2.sum32 = add i64 %tmp18, 20 ; <i64> [#uses=1]
|
|
%52 = getelementptr i8* %rk26, i64 %ctg2.sum32 ; <i8*> [#uses=1]
|
|
%53 = bitcast i8* %52 to i32* ; <i32*> [#uses=1]
|
|
%54 = load i32* %53, align 4 ; <i32> [#uses=1]
|
|
%55 = xor i32 %51, %47 ; <i32> [#uses=1]
|
|
%56 = xor i32 %55, %54 ; <i32> [#uses=1]
|
|
%indvar.next = add i64 %indvar, 1 ; <i64> [#uses=1]
|
|
br label %bb
|
|
|
|
bb2: ; preds = %bb
|
|
%tmp10 = shl i64 %tmp.16, 4 ; <i64> [#uses=2]
|
|
%ctg2.sum = add i64 %tmp10, 16 ; <i64> [#uses=1]
|
|
%tmp1213 = getelementptr i8* %rk26, i64 %ctg2.sum ; <i8*> [#uses=1]
|
|
%57 = bitcast i8* %tmp1213 to i32* ; <i32*> [#uses=1]
|
|
%58 = and i32 %33, -16777216 ; <i32> [#uses=1]
|
|
%59 = lshr i32 %29, 16 ; <i32> [#uses=1]
|
|
%60 = and i32 %59, 255 ; <i32> [#uses=1]
|
|
%61 = zext i32 %60 to i64 ; <i64> [#uses=1]
|
|
%62 = getelementptr [256 x i32]* @Te1, i64 0, i64 %61 ; <i32*> [#uses=1]
|
|
%63 = load i32* %62, align 4 ; <i32> [#uses=1]
|
|
%64 = and i32 %63, 16711680 ; <i32> [#uses=1]
|
|
%65 = or i32 %64, %58 ; <i32> [#uses=1]
|
|
%66 = load i32* %57, align 4 ; <i32> [#uses=1]
|
|
%67 = xor i32 %65, %66 ; <i32> [#uses=2]
|
|
%68 = lshr i32 %29, 8 ; <i32> [#uses=1]
|
|
%69 = zext i32 %68 to i64 ; <i64> [#uses=1]
|
|
%70 = getelementptr [256 x i32]* @Te0, i64 0, i64 %69 ; <i32*> [#uses=1]
|
|
%71 = load i32* %70, align 4 ; <i32> [#uses=1]
|
|
%72 = and i32 %71, -16777216 ; <i32> [#uses=1]
|
|
%73 = and i32 %16, 255 ; <i32> [#uses=1]
|
|
%74 = zext i32 %73 to i64 ; <i64> [#uses=1]
|
|
%75 = getelementptr [256 x i32]* @Te1, i64 0, i64 %74 ; <i32*> [#uses=1]
|
|
%76 = load i32* %75, align 4 ; <i32> [#uses=1]
|
|
%77 = and i32 %76, 16711680 ; <i32> [#uses=1]
|
|
%78 = or i32 %77, %72 ; <i32> [#uses=1]
|
|
%ctg2.sum25 = add i64 %tmp10, 20 ; <i64> [#uses=1]
|
|
%79 = getelementptr i8* %rk26, i64 %ctg2.sum25 ; <i8*> [#uses=1]
|
|
%80 = bitcast i8* %79 to i32* ; <i32*> [#uses=1]
|
|
%81 = load i32* %80, align 4 ; <i32> [#uses=1]
|
|
%82 = xor i32 %78, %81 ; <i32> [#uses=2]
|
|
%83 = lshr i32 %67, 24 ; <i32> [#uses=1]
|
|
%84 = trunc i32 %83 to i8 ; <i8> [#uses=1]
|
|
store i8 %84, i8* %out, align 1
|
|
%85 = lshr i32 %67, 16 ; <i32> [#uses=1]
|
|
%86 = trunc i32 %85 to i8 ; <i8> [#uses=1]
|
|
%87 = getelementptr i8* %out, i64 1 ; <i8*> [#uses=1]
|
|
store i8 %86, i8* %87, align 1
|
|
%88 = getelementptr i8* %out, i64 4 ; <i8*> [#uses=1]
|
|
%89 = lshr i32 %82, 24 ; <i32> [#uses=1]
|
|
%90 = trunc i32 %89 to i8 ; <i8> [#uses=1]
|
|
store i8 %90, i8* %88, align 1
|
|
%91 = lshr i32 %82, 16 ; <i32> [#uses=1]
|
|
%92 = trunc i32 %91 to i8 ; <i8> [#uses=1]
|
|
%93 = getelementptr i8* %out, i64 5 ; <i8*> [#uses=1]
|
|
store i8 %92, i8* %93, align 1
|
|
ret void
|
|
}
|