mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-21 00:32:23 +00:00
2b1874cbd4
Summary: Atomic loads and store of up to the native size (32 bits, or 64 for PPC64) can be lowered to a simple load or store instruction (as the synchronization is already handled by AtomicExpand, and the atomicity is guaranteed thanks to the alignment requirements of atomic accesses). This is exactly what this patch does. Previously, these were implemented by complex load-linked/store-conditional loops.. an obvious performance problem. For example, this patch turns ``` define void @store_i8_unordered(i8* %mem) { store atomic i8 42, i8* %mem unordered, align 1 ret void } ``` from ``` _store_i8_unordered: ; @store_i8_unordered ; BB#0: rlwinm r2, r3, 3, 27, 28 li r4, 42 xori r5, r2, 24 rlwinm r2, r3, 0, 0, 29 li r3, 255 slw r4, r4, r5 slw r3, r3, r5 and r4, r4, r3 LBB4_1: ; =>This Inner Loop Header: Depth=1 lwarx r5, 0, r2 andc r5, r5, r3 or r5, r4, r5 stwcx. r5, 0, r2 bne cr0, LBB4_1 ; BB#2: blr ``` into ``` _store_i8_unordered: ; @store_i8_unordered ; BB#0: li r2, 42 stb r2, 0(r3) blr ``` which looks like a pretty clear win to me. Test Plan: fixed the tests + new test for indexed accesses + make check-all Reviewers: jfb, wschmidt, hfinkel Subscribers: llvm-commits Differential Revision: http://reviews.llvm.org/D5587 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@218922 91177308-0d34-0410-b5e6-96231b3b80d8
138 lines
3.9 KiB
LLVM
138 lines
3.9 KiB
LLVM
; RUN: llc < %s -mtriple=powerpc-apple-darwin -march=ppc32 -verify-machineinstrs | FileCheck %s --check-prefix=CHECK --check-prefix=PPC32
|
|
; FIXME: -verify-machineinstrs currently fail on ppc64 (mismatched register/instruction).
|
|
; This is already checked for in Atomics-64.ll
|
|
; RUN: llc < %s -mtriple=powerpc-apple-darwin -march=ppc64 | FileCheck %s --check-prefix=CHECK --check-prefix=PPC64
|
|
|
|
; FIXME: we don't currently check for the operations themselves with CHECK-NEXT,
|
|
; because they are implemented in a very messy way with lwarx/stwcx.
|
|
; It should be fixed soon in another patch.
|
|
|
|
; We first check loads, for all sizes from i8 to i64.
|
|
; We also vary orderings to check for barriers.
|
|
define i8 @load_i8_unordered(i8* %mem) {
|
|
; CHECK-LABEL: load_i8_unordered
|
|
; CHECK: lbz
|
|
; CHECK-NOT: sync
|
|
%val = load atomic i8* %mem unordered, align 1
|
|
ret i8 %val
|
|
}
|
|
define i16 @load_i16_monotonic(i16* %mem) {
|
|
; CHECK-LABEL: load_i16_monotonic
|
|
; CHECK: lhz
|
|
; CHECK-NOT: sync
|
|
%val = load atomic i16* %mem monotonic, align 2
|
|
ret i16 %val
|
|
}
|
|
define i32 @load_i32_acquire(i32* %mem) {
|
|
; CHECK-LABEL: load_i32_acquire
|
|
; CHECK: lwz
|
|
%val = load atomic i32* %mem acquire, align 4
|
|
; CHECK: sync 1
|
|
ret i32 %val
|
|
}
|
|
define i64 @load_i64_seq_cst(i64* %mem) {
|
|
; CHECK-LABEL: load_i64_seq_cst
|
|
; CHECK: sync 0
|
|
; PPC32: __sync_
|
|
; PPC64-NOT: __sync_
|
|
; PPC64: ld
|
|
%val = load atomic i64* %mem seq_cst, align 8
|
|
; CHECK: sync 1
|
|
ret i64 %val
|
|
}
|
|
|
|
; Stores
|
|
define void @store_i8_unordered(i8* %mem) {
|
|
; CHECK-LABEL: store_i8_unordered
|
|
; CHECK-NOT: sync
|
|
; CHECK: stb
|
|
store atomic i8 42, i8* %mem unordered, align 1
|
|
ret void
|
|
}
|
|
define void @store_i16_monotonic(i16* %mem) {
|
|
; CHECK-LABEL: store_i16_monotonic
|
|
; CHECK-NOT: sync
|
|
; CHECK: sth
|
|
store atomic i16 42, i16* %mem monotonic, align 2
|
|
ret void
|
|
}
|
|
define void @store_i32_release(i32* %mem) {
|
|
; CHECK-LABEL: store_i32_release
|
|
; CHECK: sync 1
|
|
; CHECK: stw
|
|
store atomic i32 42, i32* %mem release, align 4
|
|
ret void
|
|
}
|
|
define void @store_i64_seq_cst(i64* %mem) {
|
|
; CHECK-LABEL: store_i64_seq_cst
|
|
; CHECK: sync 0
|
|
; PPC32: __sync_
|
|
; PPC64-NOT: __sync_
|
|
; PPC64: std
|
|
store atomic i64 42, i64* %mem seq_cst, align 8
|
|
ret void
|
|
}
|
|
|
|
; Atomic CmpXchg
|
|
define i8 @cas_strong_i8_sc_sc(i8* %mem) {
|
|
; CHECK-LABEL: cas_strong_i8_sc_sc
|
|
; CHECK: sync 0
|
|
%val = cmpxchg i8* %mem, i8 0, i8 1 seq_cst seq_cst
|
|
; CHECK: sync 1
|
|
%loaded = extractvalue { i8, i1} %val, 0
|
|
ret i8 %loaded
|
|
}
|
|
define i16 @cas_weak_i16_acquire_acquire(i16* %mem) {
|
|
; CHECK-LABEL: cas_weak_i16_acquire_acquire
|
|
;CHECK-NOT: sync
|
|
%val = cmpxchg weak i16* %mem, i16 0, i16 1 acquire acquire
|
|
; CHECK: sync 1
|
|
%loaded = extractvalue { i16, i1} %val, 0
|
|
ret i16 %loaded
|
|
}
|
|
define i32 @cas_strong_i32_acqrel_acquire(i32* %mem) {
|
|
; CHECK-LABEL: cas_strong_i32_acqrel_acquire
|
|
; CHECK: sync 1
|
|
%val = cmpxchg i32* %mem, i32 0, i32 1 acq_rel acquire
|
|
; CHECK: sync 1
|
|
%loaded = extractvalue { i32, i1} %val, 0
|
|
ret i32 %loaded
|
|
}
|
|
define i64 @cas_weak_i64_release_monotonic(i64* %mem) {
|
|
; CHECK-LABEL: cas_weak_i64_release_monotonic
|
|
; CHECK: sync 1
|
|
%val = cmpxchg weak i64* %mem, i64 0, i64 1 release monotonic
|
|
; CHECK-NOT: [sync ]
|
|
%loaded = extractvalue { i64, i1} %val, 0
|
|
ret i64 %loaded
|
|
}
|
|
|
|
; AtomicRMW
|
|
define i8 @add_i8_monotonic(i8* %mem, i8 %operand) {
|
|
; CHECK-LABEL: add_i8_monotonic
|
|
; CHECK-NOT: sync
|
|
%val = atomicrmw add i8* %mem, i8 %operand monotonic
|
|
ret i8 %val
|
|
}
|
|
define i16 @xor_i16_seq_cst(i16* %mem, i16 %operand) {
|
|
; CHECK-LABEL: xor_i16_seq_cst
|
|
; CHECK: sync 0
|
|
%val = atomicrmw xor i16* %mem, i16 %operand seq_cst
|
|
; CHECK: sync 1
|
|
ret i16 %val
|
|
}
|
|
define i32 @xchg_i32_acq_rel(i32* %mem, i32 %operand) {
|
|
; CHECK-LABEL: xchg_i32_acq_rel
|
|
; CHECK: sync 1
|
|
%val = atomicrmw xchg i32* %mem, i32 %operand acq_rel
|
|
; CHECK: sync 1
|
|
ret i32 %val
|
|
}
|
|
define i64 @and_i64_release(i64* %mem, i64 %operand) {
|
|
; CHECK-LABEL: and_i64_release
|
|
; CHECK: sync 1
|
|
%val = atomicrmw and i64* %mem, i64 %operand release
|
|
; CHECK-NOT: [sync ]
|
|
ret i64 %val
|
|
}
|