mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-24 06:30:19 +00:00
32d728fbb9
There's no hard requirement on LLVM to align local variable to 32-bits, so the Thumb1 frame handling needs to be able to deal with variables that are only naturally aligned without falling over. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@219733 91177308-0d34-0410-b5e6-96231b3b80d8
374 lines
11 KiB
LLVM
374 lines
11 KiB
LLVM
; RUN: llc < %s -mtriple=armv7-apple-ios -verify-machineinstrs | FileCheck %s
|
|
; RUN: llc < %s -mtriple=thumbv7-apple-ios -verify-machineinstrs | FileCheck %s
|
|
; RUN: llc < %s -mtriple=thumbv6-apple-ios -verify-machineinstrs | FileCheck %s --check-prefix=CHECK-T1
|
|
; RUN: llc < %s -mtriple=thumbv6-apple-ios -verify-machineinstrs -mcpu=cortex-m0 | FileCheck %s --check-prefix=CHECK-M0
|
|
; RUN: llc < %s -mtriple=thumbv7--none-eabi -thread-model single -verify-machineinstrs | FileCheck %s --check-prefix=CHECK-BAREMETAL
|
|
|
|
target datalayout = "e-m:e-p:32:32-i64:64-v128:64:128-a:0:32-n32-S64"
|
|
|
|
define void @func(i32 %argc, i8** %argv) nounwind {
|
|
entry:
|
|
%argc.addr = alloca i32 ; <i32*> [#uses=1]
|
|
%argv.addr = alloca i8** ; <i8***> [#uses=1]
|
|
%val1 = alloca i32 ; <i32*> [#uses=2]
|
|
%val2 = alloca i32 ; <i32*> [#uses=15]
|
|
%andt = alloca i32 ; <i32*> [#uses=2]
|
|
%ort = alloca i32 ; <i32*> [#uses=2]
|
|
%xort = alloca i32 ; <i32*> [#uses=2]
|
|
%old = alloca i32 ; <i32*> [#uses=18]
|
|
%temp = alloca i32 ; <i32*> [#uses=2]
|
|
store i32 %argc, i32* %argc.addr
|
|
store i8** %argv, i8*** %argv.addr
|
|
store i32 0, i32* %val1
|
|
store i32 31, i32* %val2
|
|
store i32 3855, i32* %andt
|
|
store i32 3855, i32* %ort
|
|
store i32 3855, i32* %xort
|
|
store i32 4, i32* %temp
|
|
%tmp = load i32* %temp
|
|
; CHECK: ldrex
|
|
; CHECK: add
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_add_4
|
|
; CHECK-M0: bl ___sync_fetch_and_add_4
|
|
; CHECK-BAREMETAL: add
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%0 = atomicrmw add i32* %val1, i32 %tmp monotonic
|
|
store i32 %0, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: sub
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_sub_4
|
|
; CHECK-M0: bl ___sync_fetch_and_sub_4
|
|
; CHECK-BAREMETAL: sub
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%1 = atomicrmw sub i32* %val2, i32 30 monotonic
|
|
store i32 %1, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: add
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_add_4
|
|
; CHECK-M0: bl ___sync_fetch_and_add_4
|
|
; CHECK-BAREMETAL: add
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%2 = atomicrmw add i32* %val2, i32 1 monotonic
|
|
store i32 %2, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: sub
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_sub_4
|
|
; CHECK-M0: bl ___sync_fetch_and_sub_4
|
|
; CHECK-BAREMETAL: sub
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%3 = atomicrmw sub i32* %val2, i32 1 monotonic
|
|
store i32 %3, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: and
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_and_4
|
|
; CHECK-M0: bl ___sync_fetch_and_and_4
|
|
; CHECK-BAREMETAL: and
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%4 = atomicrmw and i32* %andt, i32 4080 monotonic
|
|
store i32 %4, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: or
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_or_4
|
|
; CHECK-M0: bl ___sync_fetch_and_or_4
|
|
; CHECK-BAREMETAL: or
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%5 = atomicrmw or i32* %ort, i32 4080 monotonic
|
|
store i32 %5, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: eor
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_xor_4
|
|
; CHECK-M0: bl ___sync_fetch_and_xor_4
|
|
; CHECK-BAREMETAL: eor
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%6 = atomicrmw xor i32* %xort, i32 4080 monotonic
|
|
store i32 %6, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_min_4
|
|
; CHECK-M0: bl ___sync_fetch_and_min_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%7 = atomicrmw min i32* %val2, i32 16 monotonic
|
|
store i32 %7, i32* %old
|
|
%neg = sub i32 0, 1
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_min_4
|
|
; CHECK-M0: bl ___sync_fetch_and_min_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%8 = atomicrmw min i32* %val2, i32 %neg monotonic
|
|
store i32 %8, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_max_4
|
|
; CHECK-M0: bl ___sync_fetch_and_max_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%9 = atomicrmw max i32* %val2, i32 1 monotonic
|
|
store i32 %9, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_max_4
|
|
; CHECK-M0: bl ___sync_fetch_and_max_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%10 = atomicrmw max i32* %val2, i32 0 monotonic
|
|
store i32 %10, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_4
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%11 = atomicrmw umin i32* %val2, i32 16 monotonic
|
|
store i32 %11, i32* %old
|
|
%uneg = sub i32 0, 1
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_4
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%12 = atomicrmw umin i32* %val2, i32 %uneg monotonic
|
|
store i32 %12, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_4
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%13 = atomicrmw umax i32* %val2, i32 1 monotonic
|
|
store i32 %13, i32* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_4
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_4
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%14 = atomicrmw umax i32* %val2, i32 0 monotonic
|
|
store i32 %14, i32* %old
|
|
|
|
ret void
|
|
}
|
|
|
|
define void @func2() nounwind {
|
|
entry:
|
|
%val = alloca i16
|
|
%old = alloca i16
|
|
store i16 31, i16* %val
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_2
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_2
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%0 = atomicrmw umin i16* %val, i16 16 monotonic
|
|
store i16 %0, i16* %old
|
|
%uneg = sub i16 0, 1
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_2
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_2
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%1 = atomicrmw umin i16* %val, i16 %uneg monotonic
|
|
store i16 %1, i16* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_2
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_2
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%2 = atomicrmw umax i16* %val, i16 1 monotonic
|
|
store i16 %2, i16* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_2
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_2
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%3 = atomicrmw umax i16* %val, i16 0 monotonic
|
|
store i16 %3, i16* %old
|
|
ret void
|
|
}
|
|
|
|
define void @func3() nounwind {
|
|
entry:
|
|
%val = alloca i8
|
|
%old = alloca i8
|
|
store i8 31, i8* %val
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_1
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_1
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%0 = atomicrmw umin i8* %val, i8 16 monotonic
|
|
store i8 %0, i8* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umin_1
|
|
; CHECK-M0: bl ___sync_fetch_and_umin_1
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%uneg = sub i8 0, 1
|
|
%1 = atomicrmw umin i8* %val, i8 %uneg monotonic
|
|
store i8 %1, i8* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_1
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_1
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%2 = atomicrmw umax i8* %val, i8 1 monotonic
|
|
store i8 %2, i8* %old
|
|
; CHECK: ldrex
|
|
; CHECK: cmp
|
|
; CHECK: strex
|
|
; CHECK-T1: blx ___sync_fetch_and_umax_1
|
|
; CHECK-M0: bl ___sync_fetch_and_umax_1
|
|
; CHECK-BAREMETAL: cmp
|
|
; CHECK-BAREMETAL-NOT: __sync
|
|
%3 = atomicrmw umax i8* %val, i8 0 monotonic
|
|
store i8 %3, i8* %old
|
|
ret void
|
|
}
|
|
|
|
; CHECK: func4
|
|
; This function should not need to use callee-saved registers.
|
|
; rdar://problem/12203728
|
|
; CHECK-NOT: r4
|
|
define i32 @func4(i32* %p) nounwind optsize ssp {
|
|
entry:
|
|
%0 = atomicrmw add i32* %p, i32 1 monotonic
|
|
ret i32 %0
|
|
}
|
|
|
|
define i32 @test_cmpxchg_fail_order(i32 *%addr, i32 %desired, i32 %new) {
|
|
; CHECK-LABEL: test_cmpxchg_fail_order:
|
|
|
|
%pair = cmpxchg i32* %addr, i32 %desired, i32 %new seq_cst monotonic
|
|
%oldval = extractvalue { i32, i1 } %pair, 0
|
|
; CHECK: dmb ish
|
|
; CHECK: [[LOOP_BB:\.?LBB[0-9]+_1]]:
|
|
; CHECK: ldrex [[OLDVAL:r[0-9]+]], [r[[ADDR:[0-9]+]]]
|
|
; CHECK: cmp [[OLDVAL]], r1
|
|
; CHECK: bxne lr
|
|
; CHECK: strex [[SUCCESS:r[0-9]+]], r2, [r[[ADDR]]]
|
|
; CHECK: cmp [[SUCCESS]], #0
|
|
; CHECK: bne [[LOOP_BB]]
|
|
; CHECK: dmb ish
|
|
; CHECK: bx lr
|
|
|
|
ret i32 %oldval
|
|
}
|
|
|
|
define i32 @test_cmpxchg_fail_order1(i32 *%addr, i32 %desired, i32 %new) {
|
|
; CHECK-LABEL: test_cmpxchg_fail_order1:
|
|
|
|
%pair = cmpxchg i32* %addr, i32 %desired, i32 %new acquire acquire
|
|
%oldval = extractvalue { i32, i1 } %pair, 0
|
|
; CHECK-NOT: dmb ish
|
|
; CHECK: [[LOOP_BB:\.?LBB[0-9]+_1]]:
|
|
; CHECK: ldrex [[OLDVAL:r[0-9]+]], [r[[ADDR:[0-9]+]]]
|
|
; CHECK: cmp [[OLDVAL]], r1
|
|
; CHECK: bne [[END_BB:\.?LBB[0-9]+_[0-9]+]]
|
|
; CHECK: strex [[SUCCESS:r[0-9]+]], r2, [r[[ADDR]]]
|
|
; CHECK: cmp [[SUCCESS]], #0
|
|
; CHECK: bne [[LOOP_BB]]
|
|
; CHECK: [[END_BB]]:
|
|
; CHECK: dmb ish
|
|
; CHECK: bx lr
|
|
|
|
ret i32 %oldval
|
|
}
|
|
|
|
define i32 @load_load_add_acquire(i32* %mem1, i32* %mem2) nounwind {
|
|
; CHECK-LABEL: load_load_add_acquire
|
|
%val1 = load atomic i32* %mem1 acquire, align 4
|
|
%val2 = load atomic i32* %mem2 acquire, align 4
|
|
%tmp = add i32 %val1, %val2
|
|
|
|
; CHECK: ldr {{r[0-9]}}, [r0]
|
|
; CHECK: dmb
|
|
; CHECK: ldr {{r[0-9]}}, [r1]
|
|
; CHECK: dmb
|
|
; CHECK: add r0,
|
|
|
|
; CHECK-M0: ___sync_val_compare_and_swap_4
|
|
; CHECK-M0: ___sync_val_compare_and_swap_4
|
|
|
|
; CHECK-BAREMETAL: ldr {{r[0-9]}}, [r0]
|
|
; CHECK-BAREMETAL-NOT: dmb
|
|
; CHECK-BAREMETAL: ldr {{r[0-9]}}, [r1]
|
|
; CHECK-BAREMETAL-NOT: dmb
|
|
; CHECK-BAREMETAL: add r0,
|
|
|
|
ret i32 %tmp
|
|
}
|
|
|
|
define void @store_store_release(i32* %mem1, i32 %val1, i32* %mem2, i32 %val2) {
|
|
; CHECK-LABEL: store_store_release
|
|
store atomic i32 %val1, i32* %mem1 release, align 4
|
|
store atomic i32 %val2, i32* %mem2 release, align 4
|
|
|
|
; CHECK: dmb
|
|
; CHECK: str r1, [r0]
|
|
; CHECK: dmb
|
|
; CHECK: str r3, [r2]
|
|
|
|
; CHECK-M0: ___sync_lock_test_and_set
|
|
; CHECK-M0: ___sync_lock_test_and_set
|
|
|
|
; CHECK-BAREMETAL-NOT: dmb
|
|
; CHECK-BAREMTEAL: str r1, [r0]
|
|
; CHECK-BAREMETAL-NOT: dmb
|
|
; CHECK-BAREMTEAL: str r3, [r2]
|
|
|
|
ret void
|
|
}
|
|
|
|
define void @load_fence_store_monotonic(i32* %mem1, i32* %mem2) {
|
|
; CHECK-LABEL: load_fence_store_monotonic
|
|
%val = load atomic i32* %mem1 monotonic, align 4
|
|
fence seq_cst
|
|
store atomic i32 %val, i32* %mem2 monotonic, align 4
|
|
|
|
; CHECK: ldr [[R0:r[0-9]]], [r0]
|
|
; CHECK: dmb
|
|
; CHECK: str [[R0]], [r1]
|
|
|
|
; CHECK-M0: ldr [[R0:r[0-9]]], [r0]
|
|
; CHECK-M0: dmb
|
|
; CHECK-M0: str [[R0]], [r1]
|
|
|
|
; CHECK-BAREMETAL: ldr [[R0:r[0-9]]], [r0]
|
|
; CHECK-BAREMETAL-NOT: dmb
|
|
; CHECK-BAREMETAL: str [[R0]], [r1]
|
|
|
|
ret void
|
|
}
|