mirror of
				https://github.com/c64scene-ar/llvm-6502.git
				synced 2025-11-04 05:17:07 +00:00 
			
		
		
		
	So long as the choice between printing msync and sync is not ambiguous, we can print 'sync 0' and just 'sync'. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@235663 91177308-0d34-0410-b5e6-96231b3b80d8
		
			
				
	
	
		
			138 lines
		
	
	
		
			3.9 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
			
		
		
	
	
			138 lines
		
	
	
		
			3.9 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
; RUN: llc < %s -mtriple=powerpc-apple-darwin -march=ppc32 -verify-machineinstrs | FileCheck %s --check-prefix=CHECK --check-prefix=PPC32
 | 
						|
; FIXME: -verify-machineinstrs currently fail on ppc64 (mismatched register/instruction).
 | 
						|
; This is already checked for in Atomics-64.ll
 | 
						|
; RUN: llc < %s -mtriple=powerpc-apple-darwin -march=ppc64 | FileCheck %s --check-prefix=CHECK --check-prefix=PPC64
 | 
						|
 | 
						|
; FIXME: we don't currently check for the operations themselves with CHECK-NEXT,
 | 
						|
;   because they are implemented in a very messy way with lwarx/stwcx.
 | 
						|
;   It should be fixed soon in another patch.
 | 
						|
 | 
						|
; We first check loads, for all sizes from i8 to i64.
 | 
						|
; We also vary orderings to check for barriers.
 | 
						|
define i8 @load_i8_unordered(i8* %mem) {
 | 
						|
; CHECK-LABEL: load_i8_unordered
 | 
						|
; CHECK: lbz
 | 
						|
; CHECK-NOT: sync
 | 
						|
  %val = load atomic i8, i8* %mem unordered, align 1
 | 
						|
  ret i8 %val
 | 
						|
}
 | 
						|
define i16 @load_i16_monotonic(i16* %mem) {
 | 
						|
; CHECK-LABEL: load_i16_monotonic
 | 
						|
; CHECK: lhz
 | 
						|
; CHECK-NOT: sync
 | 
						|
  %val = load atomic i16, i16* %mem monotonic, align 2
 | 
						|
  ret i16 %val
 | 
						|
}
 | 
						|
define i32 @load_i32_acquire(i32* %mem) {
 | 
						|
; CHECK-LABEL: load_i32_acquire
 | 
						|
; CHECK: lwz
 | 
						|
  %val = load atomic i32, i32* %mem acquire, align 4
 | 
						|
; CHECK: lwsync
 | 
						|
  ret i32 %val
 | 
						|
}
 | 
						|
define i64 @load_i64_seq_cst(i64* %mem) {
 | 
						|
; CHECK-LABEL: load_i64_seq_cst
 | 
						|
; CHECK: sync
 | 
						|
; PPC32: __sync_
 | 
						|
; PPC64-NOT: __sync_
 | 
						|
; PPC64: ld
 | 
						|
  %val = load atomic i64, i64* %mem seq_cst, align 8
 | 
						|
; CHECK: lwsync
 | 
						|
  ret i64 %val
 | 
						|
}
 | 
						|
 | 
						|
; Stores
 | 
						|
define void @store_i8_unordered(i8* %mem) {
 | 
						|
; CHECK-LABEL: store_i8_unordered
 | 
						|
; CHECK-NOT: sync
 | 
						|
; CHECK: stb
 | 
						|
  store atomic i8 42, i8* %mem unordered, align 1
 | 
						|
  ret void
 | 
						|
}
 | 
						|
define void @store_i16_monotonic(i16* %mem) {
 | 
						|
; CHECK-LABEL: store_i16_monotonic
 | 
						|
; CHECK-NOT: sync
 | 
						|
; CHECK: sth
 | 
						|
  store atomic i16 42, i16* %mem monotonic, align 2
 | 
						|
  ret void
 | 
						|
}
 | 
						|
define void @store_i32_release(i32* %mem) {
 | 
						|
; CHECK-LABEL: store_i32_release
 | 
						|
; CHECK: lwsync
 | 
						|
; CHECK: stw
 | 
						|
  store atomic i32 42, i32* %mem release, align 4
 | 
						|
  ret void
 | 
						|
}
 | 
						|
define void @store_i64_seq_cst(i64* %mem) {
 | 
						|
; CHECK-LABEL: store_i64_seq_cst
 | 
						|
; CHECK: sync
 | 
						|
; PPC32: __sync_
 | 
						|
; PPC64-NOT: __sync_
 | 
						|
; PPC64: std
 | 
						|
  store atomic i64 42, i64* %mem seq_cst, align 8
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
; Atomic CmpXchg
 | 
						|
define i8 @cas_strong_i8_sc_sc(i8* %mem) {
 | 
						|
; CHECK-LABEL: cas_strong_i8_sc_sc
 | 
						|
; CHECK: sync
 | 
						|
  %val = cmpxchg i8* %mem, i8 0, i8 1 seq_cst seq_cst
 | 
						|
; CHECK: lwsync
 | 
						|
  %loaded = extractvalue { i8, i1} %val, 0
 | 
						|
  ret i8 %loaded
 | 
						|
}
 | 
						|
define i16 @cas_weak_i16_acquire_acquire(i16* %mem) {
 | 
						|
; CHECK-LABEL: cas_weak_i16_acquire_acquire
 | 
						|
;CHECK-NOT: sync
 | 
						|
  %val = cmpxchg weak i16* %mem, i16 0, i16 1 acquire acquire
 | 
						|
; CHECK: lwsync
 | 
						|
  %loaded = extractvalue { i16, i1} %val, 0
 | 
						|
  ret i16 %loaded
 | 
						|
}
 | 
						|
define i32 @cas_strong_i32_acqrel_acquire(i32* %mem) {
 | 
						|
; CHECK-LABEL: cas_strong_i32_acqrel_acquire
 | 
						|
; CHECK: lwsync
 | 
						|
  %val = cmpxchg i32* %mem, i32 0, i32 1 acq_rel acquire
 | 
						|
; CHECK: lwsync
 | 
						|
  %loaded = extractvalue { i32, i1} %val, 0
 | 
						|
  ret i32 %loaded
 | 
						|
}
 | 
						|
define i64 @cas_weak_i64_release_monotonic(i64* %mem) {
 | 
						|
; CHECK-LABEL: cas_weak_i64_release_monotonic
 | 
						|
; CHECK: lwsync
 | 
						|
  %val = cmpxchg weak i64* %mem, i64 0, i64 1 release monotonic
 | 
						|
; CHECK-NOT: [sync ]
 | 
						|
  %loaded = extractvalue { i64, i1} %val, 0
 | 
						|
  ret i64 %loaded
 | 
						|
}
 | 
						|
 | 
						|
; AtomicRMW
 | 
						|
define i8 @add_i8_monotonic(i8* %mem, i8 %operand) {
 | 
						|
; CHECK-LABEL: add_i8_monotonic
 | 
						|
; CHECK-NOT: sync
 | 
						|
  %val = atomicrmw add i8* %mem, i8 %operand monotonic
 | 
						|
  ret i8 %val
 | 
						|
}
 | 
						|
define i16 @xor_i16_seq_cst(i16* %mem, i16 %operand) {
 | 
						|
; CHECK-LABEL: xor_i16_seq_cst
 | 
						|
; CHECK: sync
 | 
						|
  %val = atomicrmw xor i16* %mem, i16 %operand seq_cst
 | 
						|
; CHECK: lwsync
 | 
						|
  ret i16 %val
 | 
						|
}
 | 
						|
define i32 @xchg_i32_acq_rel(i32* %mem, i32 %operand) {
 | 
						|
; CHECK-LABEL: xchg_i32_acq_rel
 | 
						|
; CHECK: lwsync
 | 
						|
  %val = atomicrmw xchg i32* %mem, i32 %operand acq_rel
 | 
						|
; CHECK: lwsync
 | 
						|
  ret i32 %val
 | 
						|
}
 | 
						|
define i64 @and_i64_release(i64* %mem, i64 %operand) {
 | 
						|
; CHECK-LABEL: and_i64_release
 | 
						|
; CHECK: lwsync
 | 
						|
  %val = atomicrmw and i64* %mem, i64 %operand release
 | 
						|
; CHECK-NOT: [sync ]
 | 
						|
  ret i64 %val
 | 
						|
}
 |