mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2025-03-06 20:33:54 +00:00
This adds a second implementation of the AArch64 architecture to LLVM, accessible in parallel via the "arm64" triple. The plan over the coming weeks & months is to merge the two into a single backend, during which time thorough code review should naturally occur. Everything will be easier with the target in-tree though, hence this commit. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@205090 91177308-0d34-0410-b5e6-96231b3b80d8
50 lines
1.7 KiB
LLVM
50 lines
1.7 KiB
LLVM
; RUN: llc -march=arm64 -arm64-neon-syntax=apple < %s | FileCheck %s
|
|
|
|
define <8 x i16> @testShiftRightArith_v8i16(<8 x i16> %a, <8 x i16> %b) #0 {
|
|
; CHECK-LABEL: testShiftRightArith_v8i16:
|
|
; CHECK: neg.8h [[REG1:v[0-9]+]], [[REG1]]
|
|
; CHECK-NEXT: sshl.8h [[REG2:v[0-9]+]], [[REG2]], [[REG1]]
|
|
|
|
entry:
|
|
%a.addr = alloca <8 x i16>, align 16
|
|
%b.addr = alloca <8 x i16>, align 16
|
|
store <8 x i16> %a, <8 x i16>* %a.addr, align 16
|
|
store <8 x i16> %b, <8 x i16>* %b.addr, align 16
|
|
%0 = load <8 x i16>* %a.addr, align 16
|
|
%1 = load <8 x i16>* %b.addr, align 16
|
|
%shr = ashr <8 x i16> %0, %1
|
|
ret <8 x i16> %shr
|
|
}
|
|
|
|
define <4 x i32> @testShiftRightArith_v4i32(<4 x i32> %a, <4 x i32> %b) #0 {
|
|
; CHECK-LABEL: testShiftRightArith_v4i32:
|
|
; CHECK: neg.4s [[REG3:v[0-9]+]], [[REG3]]
|
|
; CHECK-NEXT: sshl.4s [[REG4:v[0-9]+]], [[REG4]], [[REG3]]
|
|
entry:
|
|
%a.addr = alloca <4 x i32>, align 32
|
|
%b.addr = alloca <4 x i32>, align 32
|
|
store <4 x i32> %a, <4 x i32>* %a.addr, align 32
|
|
store <4 x i32> %b, <4 x i32>* %b.addr, align 32
|
|
%0 = load <4 x i32>* %a.addr, align 32
|
|
%1 = load <4 x i32>* %b.addr, align 32
|
|
%shr = ashr <4 x i32> %0, %1
|
|
ret <4 x i32> %shr
|
|
}
|
|
|
|
define <8 x i16> @testShiftRightLogical(<8 x i16> %a, <8 x i16> %b) #0 {
|
|
; CHECK: testShiftRightLogical
|
|
; CHECK: neg.8h [[REG5:v[0-9]+]], [[REG5]]
|
|
; CHECK-NEXT: ushl.8h [[REG6:v[0-9]+]], [[REG6]], [[REG5]]
|
|
entry:
|
|
%a.addr = alloca <8 x i16>, align 16
|
|
%b.addr = alloca <8 x i16>, align 16
|
|
store <8 x i16> %a, <8 x i16>* %a.addr, align 16
|
|
store <8 x i16> %b, <8 x i16>* %b.addr, align 16
|
|
%0 = load <8 x i16>* %a.addr, align 16
|
|
%1 = load <8 x i16>* %b.addr, align 16
|
|
%shr = lshr <8 x i16> %0, %1
|
|
ret <8 x i16> %shr
|
|
}
|
|
|
|
attributes #0 = { nounwind }
|