mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 04:30:12 +00:00
d0dbe02fd2
The C and C++ semantics for compare_exchange require it to return a bool indicating success. This gets mapped to LLVM IR which follows each cmpxchg with an icmp of the value loaded against the desired value. When lowered to ldxr/stxr loops, this extra comparison is redundant: its results are implicit in the control-flow of the function. This commit makes two changes: it replaces that icmp with appropriate PHI nodes, and then makes sure earlyCSE is called after expansion to actually make use of the opportunities revealed. I've also added -{arm,aarch64}-enable-atomic-tidy options, so that existing fragile tests aren't perturbed too much by the change. Many of them either rely on undef/unreachable too pervasively to be restored to something well-defined (particularly while making sure they test the same obscure assert from many years ago), or depend on a particular CFG shape, which is disrupted by SimplifyCFG. rdar://problem/16227836 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@209883 91177308-0d34-0410-b5e6-96231b3b80d8
137 lines
4.1 KiB
LLVM
137 lines
4.1 KiB
LLVM
; RUN: llc -mtriple=arm-eabi -arm-atomic-cfg-tidy=0 %s -o - | FileCheck -check-prefix=ARM %s
|
|
; RUN: llc -mtriple=thumb-eabi -arm-atomic-cfg-tidy=0 %s -o - | FileCheck -check-prefix=THUMB %s
|
|
; RUN: llc -mtriple=thumb-eabi -arm-atomic-cfg-tidy=0 -mcpu=arm1156t2-s -mattr=+thumb2 %s -o - \
|
|
; RUN: | FileCheck -check-prefix=T2 %s
|
|
; RUN: llc -mtriple=thumbv8-eabi -arm-atomic-cfg-tidy=0 %s -o - | FileCheck -check-prefix=V8 %s
|
|
|
|
; FIXME: The -march=thumb test doesn't change if -disable-peephole is specified.
|
|
|
|
%struct.Foo = type { i8* }
|
|
|
|
; ARM: foo
|
|
; THUMB: foo
|
|
; T2: foo
|
|
define %struct.Foo* @foo(%struct.Foo* %this, i32 %acc) nounwind readonly align 2 {
|
|
entry:
|
|
%scevgep = getelementptr %struct.Foo* %this, i32 1
|
|
br label %tailrecurse
|
|
|
|
tailrecurse: ; preds = %sw.bb, %entry
|
|
%lsr.iv2 = phi %struct.Foo* [ %scevgep3, %sw.bb ], [ %scevgep, %entry ]
|
|
%lsr.iv = phi i32 [ %lsr.iv.next, %sw.bb ], [ 1, %entry ]
|
|
%acc.tr = phi i32 [ %or, %sw.bb ], [ %acc, %entry ]
|
|
%lsr.iv24 = bitcast %struct.Foo* %lsr.iv2 to i8**
|
|
%scevgep5 = getelementptr i8** %lsr.iv24, i32 -1
|
|
%tmp2 = load i8** %scevgep5
|
|
%0 = ptrtoint i8* %tmp2 to i32
|
|
|
|
; ARM: ands {{r[0-9]+}}, {{r[0-9]+}}, #3
|
|
; ARM-NEXT: beq
|
|
|
|
; THUMB: movs r[[R0:[0-9]+]], #3
|
|
; THUMB-NEXT: ands r[[R0]], r
|
|
; THUMB-NEXT: cmp r[[R0]], #0
|
|
; THUMB-NEXT: beq
|
|
|
|
; T2: ands {{r[0-9]+}}, {{r[0-9]+}}, #3
|
|
; T2-NEXT: beq
|
|
|
|
%and = and i32 %0, 3
|
|
%tst = icmp eq i32 %and, 0
|
|
br i1 %tst, label %sw.bb, label %tailrecurse.switch
|
|
|
|
tailrecurse.switch: ; preds = %tailrecurse
|
|
; V8-LABEL: %tailrecurse.switch
|
|
; V8: cmp
|
|
; V8-NEXT: beq
|
|
; V8-NEXT: %tailrecurse.switch
|
|
; V8: cmp
|
|
; V8-NEXT: beq
|
|
; V8-NEXT: %tailrecurse.switch
|
|
; V8: cmp
|
|
; V8-NEXT: beq
|
|
; V8-NEXT: b
|
|
; The trailing space in the last line checks that the branch is unconditional
|
|
switch i32 %and, label %sw.epilog [
|
|
i32 1, label %sw.bb
|
|
i32 3, label %sw.bb6
|
|
i32 2, label %sw.bb8
|
|
]
|
|
|
|
sw.bb: ; preds = %tailrecurse.switch, %tailrecurse
|
|
%shl = shl i32 %acc.tr, 1
|
|
%or = or i32 %and, %shl
|
|
%lsr.iv.next = add i32 %lsr.iv, 1
|
|
%scevgep3 = getelementptr %struct.Foo* %lsr.iv2, i32 1
|
|
br label %tailrecurse
|
|
|
|
sw.bb6: ; preds = %tailrecurse.switch
|
|
ret %struct.Foo* %lsr.iv2
|
|
|
|
sw.bb8: ; preds = %tailrecurse.switch
|
|
%tmp1 = add i32 %acc.tr, %lsr.iv
|
|
%add.ptr11 = getelementptr inbounds %struct.Foo* %this, i32 %tmp1
|
|
ret %struct.Foo* %add.ptr11
|
|
|
|
sw.epilog: ; preds = %tailrecurse.switch
|
|
ret %struct.Foo* undef
|
|
}
|
|
|
|
; Another test that exercises the AND/TST peephole optimization and also
|
|
; generates a predicated ANDS instruction. Check that the predicate is printed
|
|
; after the "S" modifier on the instruction.
|
|
|
|
%struct.S = type { i8* (i8*)*, [1 x i8] }
|
|
|
|
; ARM: bar
|
|
; THUMB: bar
|
|
; T2: bar
|
|
; V8-LABEL: bar:
|
|
define internal zeroext i8 @bar(%struct.S* %x, %struct.S* nocapture %y) nounwind readonly {
|
|
entry:
|
|
%0 = getelementptr inbounds %struct.S* %x, i32 0, i32 1, i32 0
|
|
%1 = load i8* %0, align 1
|
|
%2 = zext i8 %1 to i32
|
|
; ARM: ands
|
|
; THUMB: ands
|
|
; T2: ands
|
|
; V8: ands
|
|
; V8-NEXT: beq
|
|
%3 = and i32 %2, 112
|
|
%4 = icmp eq i32 %3, 0
|
|
br i1 %4, label %return, label %bb
|
|
|
|
bb: ; preds = %entry
|
|
; V8-NEXT: %bb
|
|
%5 = getelementptr inbounds %struct.S* %y, i32 0, i32 1, i32 0
|
|
%6 = load i8* %5, align 1
|
|
%7 = zext i8 %6 to i32
|
|
; ARM: andsne
|
|
; THUMB: ands
|
|
; T2: andsne
|
|
; V8: ands
|
|
; V8-NEXT: beq
|
|
%8 = and i32 %7, 112
|
|
%9 = icmp eq i32 %8, 0
|
|
br i1 %9, label %return, label %bb2
|
|
|
|
bb2: ; preds = %bb
|
|
; V8-NEXT: %bb2
|
|
; V8-NEXT: cmp
|
|
; V8-NEXT: it ne
|
|
; V8-NEXT: cmpne
|
|
; V8-NEXT: bne
|
|
%10 = icmp eq i32 %3, 16
|
|
%11 = icmp eq i32 %8, 16
|
|
%or.cond = or i1 %10, %11
|
|
br i1 %or.cond, label %bb4, label %return
|
|
|
|
bb4: ; preds = %bb2
|
|
%12 = ptrtoint %struct.S* %x to i32
|
|
%phitmp = trunc i32 %12 to i8
|
|
ret i8 %phitmp
|
|
|
|
return: ; preds = %bb2, %bb, %entry
|
|
ret i8 1
|
|
}
|