mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-16 11:05:54 +00:00
ed788b6283
ARM FastISel is currently only enabled for iOS non-Thumb1, and I'm working on enabling it for other targets. As a first step I've fixed some of the tests. Changes to ARM FastISel tests: - Different triples don't generate the same relocations (especially movw/movt versus constant pool loads). Use a regex to allow either. - Mangling is different. Use a regex to allow either. - The reserved registers are sometimes different, so registers get allocated in a different order. Capture the names only where this occurs. - Add -verify-machineinstrs to some tests where it works. It doesn't work everywhere it should yet. - Add -fast-isel-abort to many tests that didn't have it before. - Split out the VarArg test from fast-isel-call.ll into its own test. This simplifies test setup because of --check-prefix. Patch by JF Bastien git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@181801 91177308-0d34-0410-b5e6-96231b3b80d8
311 lines
6.9 KiB
LLVM
311 lines
6.9 KiB
LLVM
; RUN: llc < %s -O0 -fast-isel-abort -relocation-model=dynamic-no-pic -mtriple=armv7-apple-ios | FileCheck %s --check-prefix=ARM
|
|
; RUN: llc < %s -O0 -fast-isel-abort -relocation-model=dynamic-no-pic -mtriple=thumbv7-apple-ios | FileCheck %s --check-prefix=THUMB
|
|
; RUN: llc < %s -O0 -arm-strict-align -relocation-model=dynamic-no-pic -mtriple=armv7-apple-ios | FileCheck %s --check-prefix=ARM-STRICT-ALIGN
|
|
; RUN: llc < %s -O0 -arm-strict-align -relocation-model=dynamic-no-pic -mtriple=thumbv7-apple-ios | FileCheck %s --check-prefix=THUMB-STRICT-ALIGN
|
|
|
|
; Very basic fast-isel functionality.
|
|
define i32 @add(i32 %a, i32 %b) nounwind {
|
|
entry:
|
|
%a.addr = alloca i32, align 4
|
|
%b.addr = alloca i32, align 4
|
|
store i32 %a, i32* %a.addr
|
|
store i32 %b, i32* %b.addr
|
|
%tmp = load i32* %a.addr
|
|
%tmp1 = load i32* %b.addr
|
|
%add = add nsw i32 %tmp, %tmp1
|
|
ret i32 %add
|
|
}
|
|
|
|
; Check truncate to bool
|
|
define void @test1(i32 %tmp) nounwind {
|
|
entry:
|
|
%tobool = trunc i32 %tmp to i1
|
|
br i1 %tobool, label %if.then, label %if.end
|
|
|
|
if.then: ; preds = %entry
|
|
call void @test1(i32 0)
|
|
br label %if.end
|
|
|
|
if.end: ; preds = %if.then, %entry
|
|
ret void
|
|
; ARM: test1:
|
|
; ARM: tst r0, #1
|
|
; THUMB: test1:
|
|
; THUMB: tst.w r0, #1
|
|
}
|
|
|
|
; Check some simple operations with immediates
|
|
define void @test2(i32 %tmp, i32* %ptr) nounwind {
|
|
; THUMB: test2:
|
|
; ARM: test2:
|
|
|
|
b1:
|
|
%a = add i32 %tmp, 4096
|
|
store i32 %a, i32* %ptr
|
|
br label %b2
|
|
|
|
; THUMB: add.w {{.*}} #4096
|
|
; ARM: add {{.*}} #4096
|
|
|
|
b2:
|
|
%b = add i32 %tmp, 4095
|
|
store i32 %b, i32* %ptr
|
|
br label %b3
|
|
; THUMB: addw {{.*}} #4095
|
|
; ARM: movw {{.*}} #4095
|
|
; ARM: add
|
|
|
|
b3:
|
|
%c = or i32 %tmp, 4
|
|
store i32 %c, i32* %ptr
|
|
ret void
|
|
|
|
; THUMB: orr {{.*}} #4
|
|
; ARM: orr {{.*}} #4
|
|
}
|
|
|
|
define void @test3(i32 %tmp, i32* %ptr1, i16* %ptr2, i8* %ptr3) nounwind {
|
|
; THUMB: test3:
|
|
; ARM: test3:
|
|
|
|
bb1:
|
|
%a1 = trunc i32 %tmp to i16
|
|
%a2 = trunc i16 %a1 to i8
|
|
%a3 = trunc i8 %a2 to i1
|
|
%a4 = zext i1 %a3 to i8
|
|
store i8 %a4, i8* %ptr3
|
|
%a5 = zext i8 %a4 to i16
|
|
store i16 %a5, i16* %ptr2
|
|
%a6 = zext i16 %a5 to i32
|
|
store i32 %a6, i32* %ptr1
|
|
br label %bb2
|
|
|
|
; THUMB: and
|
|
; THUMB: strb
|
|
; THUMB: uxtb
|
|
; THUMB: strh
|
|
; THUMB: uxth
|
|
; ARM: and
|
|
; ARM: strb
|
|
; ARM: uxtb
|
|
; ARM: strh
|
|
; ARM: uxth
|
|
|
|
bb2:
|
|
%b1 = trunc i32 %tmp to i16
|
|
%b2 = trunc i16 %b1 to i8
|
|
store i8 %b2, i8* %ptr3
|
|
%b3 = sext i8 %b2 to i16
|
|
store i16 %b3, i16* %ptr2
|
|
%b4 = sext i16 %b3 to i32
|
|
store i32 %b4, i32* %ptr1
|
|
br label %bb3
|
|
|
|
; THUMB: strb
|
|
; THUMB: sxtb
|
|
; THUMB: strh
|
|
; THUMB: sxth
|
|
; ARM: strb
|
|
; ARM: sxtb
|
|
; ARM: strh
|
|
; ARM: sxth
|
|
|
|
bb3:
|
|
%c1 = load i8* %ptr3
|
|
%c2 = load i16* %ptr2
|
|
%c3 = load i32* %ptr1
|
|
%c4 = zext i8 %c1 to i32
|
|
%c5 = sext i16 %c2 to i32
|
|
%c6 = add i32 %c4, %c5
|
|
%c7 = sub i32 %c3, %c6
|
|
store i32 %c7, i32* %ptr1
|
|
ret void
|
|
|
|
; THUMB: ldrb
|
|
; THUMB: ldrh
|
|
; THUMB: uxtb
|
|
; THUMB: sxth
|
|
; THUMB: add
|
|
; THUMB: sub
|
|
; ARM: ldrb
|
|
; ARM: ldrh
|
|
; ARM: uxtb
|
|
; ARM: sxth
|
|
; ARM: add
|
|
; ARM: sub
|
|
}
|
|
|
|
; Check loads/stores with globals
|
|
@test4g = external global i32
|
|
|
|
define void @test4() {
|
|
%a = load i32* @test4g
|
|
%b = add i32 %a, 1
|
|
store i32 %b, i32* @test4g
|
|
ret void
|
|
|
|
|
|
; Note that relocations are either movw/movt or constant pool
|
|
; loads. Different platforms will select different approaches.
|
|
|
|
; THUMB: {{(movw r0, :lower16:L_test4g\$non_lazy_ptr)|(ldr.n r0, .LCPI)}}
|
|
; THUMB: {{(movt r0, :upper16:L_test4g\$non_lazy_ptr)?}}
|
|
; THUMB: ldr r0, [r0]
|
|
; THUMB: ldr r1, [r0]
|
|
; THUMB: adds r1, #1
|
|
; THUMB: str r1, [r0]
|
|
|
|
; ARM: {{(movw r0, :lower16:L_test4g\$non_lazy_ptr)|(ldr r0, .LCPI)}}
|
|
; ARM: {{(movt r0, :upper16:L_test4g\$non_lazy_ptr)?}}
|
|
; ARM: ldr r0, [r0]
|
|
; ARM: ldr r1, [r0]
|
|
; ARM: add r1, r1, #1
|
|
; ARM: str r1, [r0]
|
|
}
|
|
|
|
; Check unaligned stores
|
|
%struct.anon = type <{ float }>
|
|
|
|
@a = common global %struct.anon* null, align 4
|
|
|
|
define void @unaligned_store(float %x, float %y) nounwind {
|
|
entry:
|
|
; ARM: @unaligned_store
|
|
; ARM: vmov r1, s0
|
|
; ARM: str r1, [r0]
|
|
|
|
; THUMB: @unaligned_store
|
|
; THUMB: vmov r1, s0
|
|
; THUMB: str r1, [r0]
|
|
|
|
%add = fadd float %x, %y
|
|
%0 = load %struct.anon** @a, align 4
|
|
%x1 = getelementptr inbounds %struct.anon* %0, i32 0, i32 0
|
|
store float %add, float* %x1, align 1
|
|
ret void
|
|
}
|
|
|
|
; Doublewords require only word-alignment.
|
|
; rdar://10528060
|
|
%struct.anon.0 = type { double }
|
|
|
|
@foo_unpacked = common global %struct.anon.0 zeroinitializer, align 4
|
|
|
|
define void @test5(double %a, double %b) nounwind {
|
|
entry:
|
|
; ARM: @test5
|
|
; THUMB: @test5
|
|
%add = fadd double %a, %b
|
|
store double %add, double* getelementptr inbounds (%struct.anon.0* @foo_unpacked, i32 0, i32 0), align 4
|
|
; ARM: vstr d16, [r0]
|
|
; THUMB: vstr d16, [r0]
|
|
ret void
|
|
}
|
|
|
|
; Check unaligned loads of floats
|
|
%class.TAlignTest = type <{ i16, float }>
|
|
|
|
define zeroext i1 @test6(%class.TAlignTest* %this) nounwind align 2 {
|
|
entry:
|
|
; ARM: @test6
|
|
; THUMB: @test6
|
|
%0 = alloca %class.TAlignTest*, align 4
|
|
store %class.TAlignTest* %this, %class.TAlignTest** %0, align 4
|
|
%1 = load %class.TAlignTest** %0
|
|
%2 = getelementptr inbounds %class.TAlignTest* %1, i32 0, i32 1
|
|
%3 = load float* %2, align 1
|
|
%4 = fcmp une float %3, 0.000000e+00
|
|
; ARM: ldr r0, [r0, #2]
|
|
; ARM: vmov s0, r0
|
|
; ARM: vcmpe.f32 s0, #0
|
|
; THUMB: ldr.w r0, [r0, #2]
|
|
; THUMB: vmov s0, r0
|
|
; THUMB: vcmpe.f32 s0, #0
|
|
ret i1 %4
|
|
}
|
|
|
|
; ARM: @urem_fold
|
|
; THUMB: @urem_fold
|
|
; ARM: and r0, r0, #31
|
|
; THUMB: and r0, r0, #31
|
|
define i32 @urem_fold(i32 %a) nounwind {
|
|
%rem = urem i32 %a, 32
|
|
ret i32 %rem
|
|
}
|
|
|
|
define i32 @test7() noreturn nounwind {
|
|
entry:
|
|
; ARM: @test7
|
|
; THUMB: @test7
|
|
; ARM: trap
|
|
; THUMB: trap
|
|
tail call void @llvm.trap( )
|
|
unreachable
|
|
}
|
|
|
|
declare void @llvm.trap() nounwind
|
|
|
|
define void @unaligned_i16_store(i16 %x, i16* %y) nounwind {
|
|
entry:
|
|
; ARM-STRICT-ALIGN: @unaligned_i16_store
|
|
; ARM-STRICT-ALIGN: strb
|
|
; ARM-STRICT-ALIGN: strb
|
|
|
|
; THUMB-STRICT-ALIGN: @unaligned_i16_store
|
|
; THUMB-STRICT-ALIGN: strb
|
|
; THUMB-STRICT-ALIGN: strb
|
|
|
|
store i16 %x, i16* %y, align 1
|
|
ret void
|
|
}
|
|
|
|
define i16 @unaligned_i16_load(i16* %x) nounwind {
|
|
entry:
|
|
; ARM-STRICT-ALIGN: @unaligned_i16_load
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
|
|
; THUMB-STRICT-ALIGN: @unaligned_i16_load
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
|
|
%0 = load i16* %x, align 1
|
|
ret i16 %0
|
|
}
|
|
|
|
define void @unaligned_i32_store(i32 %x, i32* %y) nounwind {
|
|
entry:
|
|
; ARM-STRICT-ALIGN: @unaligned_i32_store
|
|
; ARM-STRICT-ALIGN: strb
|
|
; ARM-STRICT-ALIGN: strb
|
|
; ARM-STRICT-ALIGN: strb
|
|
; ARM-STRICT-ALIGN: strb
|
|
|
|
; THUMB-STRICT-ALIGN: @unaligned_i32_store
|
|
; THUMB-STRICT-ALIGN: strb
|
|
; THUMB-STRICT-ALIGN: strb
|
|
; THUMB-STRICT-ALIGN: strb
|
|
; THUMB-STRICT-ALIGN: strb
|
|
|
|
store i32 %x, i32* %y, align 1
|
|
ret void
|
|
}
|
|
|
|
define i32 @unaligned_i32_load(i32* %x) nounwind {
|
|
entry:
|
|
; ARM-STRICT-ALIGN: @unaligned_i32_load
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
; ARM-STRICT-ALIGN: ldrb
|
|
|
|
; THUMB-STRICT-ALIGN: @unaligned_i32_load
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
; THUMB-STRICT-ALIGN: ldrb
|
|
|
|
%0 = load i32* %x, align 1
|
|
ret i32 %0
|
|
}
|