mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-22 07:32:48 +00:00
d336de318e
(movzx/movsx) because they give more information. Revert that part of the patch. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@129498 91177308-0d34-0410-b5e6-96231b3b80d8
110 lines
3.3 KiB
LLVM
110 lines
3.3 KiB
LLVM
; RUN: llc < %s -mtriple=x86_64-linux -O0 | FileCheck %s --check-prefix=X64
|
|
; RUN: llc < %s -mtriple=x86_64-win32 -O0 | FileCheck %s --check-prefix=X64
|
|
; RUN: llc < %s -march=x86 -O0 | FileCheck %s --check-prefix=X32
|
|
|
|
; GEP indices are interpreted as signed integers, so they
|
|
; should be sign-extended to 64 bits on 64-bit targets.
|
|
; PR3181
|
|
define i32 @test1(i32 %t3, i32* %t1) nounwind {
|
|
%t9 = getelementptr i32* %t1, i32 %t3 ; <i32*> [#uses=1]
|
|
%t15 = load i32* %t9 ; <i32> [#uses=1]
|
|
ret i32 %t15
|
|
; X32: test1:
|
|
; X32: movl (%eax,%ecx,4), %eax
|
|
; X32: ret
|
|
|
|
; X64: test1:
|
|
; X64: movslq %e[[A0:di|cx]], %rax
|
|
; X64: movl (%r[[A1:si|dx]],%rax,4), %eax
|
|
; X64: ret
|
|
|
|
}
|
|
define i32 @test2(i64 %t3, i32* %t1) nounwind {
|
|
%t9 = getelementptr i32* %t1, i64 %t3 ; <i32*> [#uses=1]
|
|
%t15 = load i32* %t9 ; <i32> [#uses=1]
|
|
ret i32 %t15
|
|
; X32: test2:
|
|
; X32: movl (%edx,%ecx,4), %eax
|
|
; X32: ret
|
|
|
|
; X64: test2:
|
|
; X64: movl (%r[[A1]],%r[[A0]],4), %eax
|
|
; X64: ret
|
|
}
|
|
|
|
|
|
|
|
; PR4984
|
|
define i8 @test3(i8* %start) nounwind {
|
|
entry:
|
|
%A = getelementptr i8* %start, i64 -2 ; <i8*> [#uses=1]
|
|
%B = load i8* %A, align 1 ; <i8> [#uses=1]
|
|
ret i8 %B
|
|
|
|
|
|
; X32: test3:
|
|
; X32: movl 4(%esp), %eax
|
|
; X32: movb -2(%eax), %al
|
|
; X32: ret
|
|
|
|
; X64: test3:
|
|
; X64: movb -2(%r[[A0]]), %al
|
|
; X64: ret
|
|
|
|
}
|
|
|
|
define double @test4(i64 %x, double* %p) nounwind {
|
|
entry:
|
|
%x.addr = alloca i64, align 8 ; <i64*> [#uses=2]
|
|
%p.addr = alloca double*, align 8 ; <double**> [#uses=2]
|
|
store i64 %x, i64* %x.addr
|
|
store double* %p, double** %p.addr
|
|
%tmp = load i64* %x.addr ; <i64> [#uses=1]
|
|
%add = add nsw i64 %tmp, 16 ; <i64> [#uses=1]
|
|
%tmp1 = load double** %p.addr ; <double*> [#uses=1]
|
|
%arrayidx = getelementptr inbounds double* %tmp1, i64 %add ; <double*> [#uses=1]
|
|
%tmp2 = load double* %arrayidx ; <double> [#uses=1]
|
|
ret double %tmp2
|
|
|
|
; X32: test4:
|
|
; X32: 128(%e{{.*}},%e{{.*}},8)
|
|
; X64: test4:
|
|
; X64: 128(%r{{.*}},%r{{.*}},8)
|
|
}
|
|
|
|
; PR8961 - Make sure the sext for the GEP addressing comes before the load that
|
|
; is folded.
|
|
define i64 @test5(i8* %A, i32 %I, i64 %B) nounwind {
|
|
%v8 = getelementptr i8* %A, i32 %I
|
|
%v9 = bitcast i8* %v8 to i64*
|
|
%v10 = load i64* %v9
|
|
%v11 = add i64 %B, %v10
|
|
ret i64 %v11
|
|
; X64: test5:
|
|
; X64: movslq %e[[A1]], %rax
|
|
; X64-NEXT: movq (%r[[A0]],%rax), %rax
|
|
; X64-NEXT: addq %{{rdx|r8}}, %rax
|
|
; X64-NEXT: ret
|
|
}
|
|
|
|
; PR9500, rdar://9156159 - Don't do non-local address mode folding,
|
|
; because it may require values which wouldn't otherwise be live out
|
|
; of their blocks.
|
|
define void @test6() {
|
|
if.end: ; preds = %if.then, %invoke.cont
|
|
%tmp15 = load i64* undef
|
|
%dec = add i64 %tmp15, 13
|
|
store i64 %dec, i64* undef
|
|
%call17 = invoke i8* @_ZNK18G__FastAllocString4dataEv()
|
|
to label %invoke.cont16 unwind label %lpad
|
|
|
|
invoke.cont16: ; preds = %if.then14
|
|
%arrayidx18 = getelementptr inbounds i8* %call17, i64 %dec
|
|
store i8 0, i8* %arrayidx18
|
|
unreachable
|
|
|
|
lpad: ; preds = %if.end19, %if.then14, %if.end, %entry
|
|
unreachable
|
|
}
|
|
declare i8* @_ZNK18G__FastAllocString4dataEv() nounwind
|