mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-26 21:32:10 +00:00
1835547ec1
While there is an encoding for it in VUZP, the result of that is undefined, so we should avoid it. Define the instruction as a pseudo for VTRN.32 instead, as the ARM ARM indicates. rdar://11222366 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@154511 91177308-0d34-0410-b5e6-96231b3b80d8
137 lines
5.4 KiB
ArmAsm
137 lines
5.4 KiB
ArmAsm
@ RUN: llvm-mc -mcpu=cortex-a8 -triple arm-unknown-unknown -show-encoding < %s | FileCheck %s
|
|
|
|
vext.8 d16, d17, d16, #3
|
|
vext.8 d16, d17, d16, #5
|
|
vext.8 q8, q9, q8, #3
|
|
vext.8 q8, q9, q8, #7
|
|
vext.16 d16, d17, d16, #3
|
|
vext.32 q8, q9, q8, #3
|
|
vext.64 q8, q9, q8, #1
|
|
|
|
vext.8 d17, d16, #3
|
|
vext.8 d7, d11, #5
|
|
vext.8 q3, q8, #3
|
|
vext.8 q9, q4, #7
|
|
vext.16 d1, d26, #3
|
|
vext.32 q5, q8, #3
|
|
vext.64 q5, q8, #1
|
|
|
|
|
|
@ CHECK: vext.8 d16, d17, d16, #3 @ encoding: [0xa0,0x03,0xf1,0xf2]
|
|
@ CHECK: vext.8 d16, d17, d16, #5 @ encoding: [0xa0,0x05,0xf1,0xf2]
|
|
@ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xe0,0x03,0xf2,0xf2]
|
|
@ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xe0,0x07,0xf2,0xf2]
|
|
@ CHECK: vext.16 d16, d17, d16, #3 @ encoding: [0xa0,0x06,0xf1,0xf2]
|
|
@ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xe0,0x0c,0xf2,0xf2]
|
|
@ CHECK: vext.64 q8, q9, q8, #1 @ encoding: [0xe0,0x08,0xf2,0xf2]
|
|
|
|
@ CHECK: vext.8 d17, d17, d16, #3 @ encoding: [0xa0,0x13,0xf1,0xf2]
|
|
@ CHECK: vext.8 d7, d7, d11, #5 @ encoding: [0x0b,0x75,0xb7,0xf2]
|
|
@ CHECK: vext.8 q3, q3, q8, #3 @ encoding: [0x60,0x63,0xb6,0xf2]
|
|
@ CHECK: vext.8 q9, q9, q4, #7 @ encoding: [0xc8,0x27,0xf2,0xf2]
|
|
@ CHECK: vext.16 d1, d1, d26, #3 @ encoding: [0x2a,0x16,0xb1,0xf2]
|
|
@ CHECK: vext.32 q5, q5, q8, #3 @ encoding: [0x60,0xac,0xba,0xf2]
|
|
@ CHECK: vext.64 q5, q5, q8, #1 @ encoding: [0x60,0xa8,0xba,0xf2]
|
|
|
|
|
|
vtrn.8 d17, d16
|
|
vtrn.16 d17, d16
|
|
vtrn.32 d17, d16
|
|
vtrn.8 q9, q8
|
|
vtrn.16 q9, q8
|
|
vtrn.32 q9, q8
|
|
|
|
@ CHECK: vtrn.8 d17, d16 @ encoding: [0xa0,0x10,0xf2,0xf3]
|
|
@ CHECK: vtrn.16 d17, d16 @ encoding: [0xa0,0x10,0xf6,0xf3]
|
|
@ CHECK: vtrn.32 d17, d16 @ encoding: [0xa0,0x10,0xfa,0xf3]
|
|
@ CHECK: vtrn.8 q9, q8 @ encoding: [0xe0,0x20,0xf2,0xf3]
|
|
@ CHECK: vtrn.16 q9, q8 @ encoding: [0xe0,0x20,0xf6,0xf3]
|
|
@ CHECK: vtrn.32 q9, q8 @ encoding: [0xe0,0x20,0xfa,0xf3]
|
|
|
|
|
|
vuzp.8 d17, d16
|
|
vuzp.16 d17, d16
|
|
vuzp.8 q9, q8
|
|
vuzp.16 q9, q8
|
|
vuzp.32 q9, q8
|
|
vzip.8 d17, d16
|
|
vzip.16 d17, d16
|
|
vzip.8 q9, q8
|
|
vzip.16 q9, q8
|
|
vzip.32 q9, q8
|
|
vzip.32 d2, d3
|
|
vuzp.32 d2, d3
|
|
|
|
@ CHECK: vuzp.8 d17, d16 @ encoding: [0x20,0x11,0xf2,0xf3]
|
|
@ CHECK: vuzp.16 d17, d16 @ encoding: [0x20,0x11,0xf6,0xf3]
|
|
@ CHECK: vuzp.8 q9, q8 @ encoding: [0x60,0x21,0xf2,0xf3]
|
|
@ CHECK: vuzp.16 q9, q8 @ encoding: [0x60,0x21,0xf6,0xf3]
|
|
@ CHECK: vuzp.32 q9, q8 @ encoding: [0x60,0x21,0xfa,0xf3]
|
|
@ CHECK: vzip.8 d17, d16 @ encoding: [0xa0,0x11,0xf2,0xf3]
|
|
@ CHECK: vzip.16 d17, d16 @ encoding: [0xa0,0x11,0xf6,0xf3]
|
|
@ CHECK: vzip.8 q9, q8 @ encoding: [0xe0,0x21,0xf2,0xf3]
|
|
@ CHECK: vzip.16 q9, q8 @ encoding: [0xe0,0x21,0xf6,0xf3]
|
|
@ CHECK: vzip.32 q9, q8 @ encoding: [0xe0,0x21,0xfa,0xf3]
|
|
@ CHECK: vtrn.32 d2, d3 @ encoding: [0x83,0x20,0xba,0xf3]
|
|
@ CHECK: vtrn.32 d2, d3 @ encoding: [0x83,0x20,0xba,0xf3]
|
|
|
|
|
|
@ VTRN alternate size suffices
|
|
|
|
vtrn.8 d3, d9
|
|
vtrn.i8 d3, d9
|
|
vtrn.u8 d3, d9
|
|
vtrn.p8 d3, d9
|
|
vtrn.16 d3, d9
|
|
vtrn.i16 d3, d9
|
|
vtrn.u16 d3, d9
|
|
vtrn.p16 d3, d9
|
|
vtrn.32 d3, d9
|
|
vtrn.i32 d3, d9
|
|
vtrn.u32 d3, d9
|
|
vtrn.f32 d3, d9
|
|
vtrn.f d3, d9
|
|
|
|
vtrn.8 q14, q6
|
|
vtrn.i8 q14, q6
|
|
vtrn.u8 q14, q6
|
|
vtrn.p8 q14, q6
|
|
vtrn.16 q14, q6
|
|
vtrn.i16 q14, q6
|
|
vtrn.u16 q14, q6
|
|
vtrn.p16 q14, q6
|
|
vtrn.32 q14, q6
|
|
vtrn.i32 q14, q6
|
|
vtrn.u32 q14, q6
|
|
vtrn.f32 q14, q6
|
|
vtrn.f q14, q6
|
|
|
|
@ CHECK: vtrn.8 d3, d9 @ encoding: [0x89,0x30,0xb2,0xf3]
|
|
@ CHECK: vtrn.8 d3, d9 @ encoding: [0x89,0x30,0xb2,0xf3]
|
|
@ CHECK: vtrn.8 d3, d9 @ encoding: [0x89,0x30,0xb2,0xf3]
|
|
@ CHECK: vtrn.8 d3, d9 @ encoding: [0x89,0x30,0xb2,0xf3]
|
|
@ CHECK: vtrn.16 d3, d9 @ encoding: [0x89,0x30,0xb6,0xf3]
|
|
@ CHECK: vtrn.16 d3, d9 @ encoding: [0x89,0x30,0xb6,0xf3]
|
|
@ CHECK: vtrn.16 d3, d9 @ encoding: [0x89,0x30,0xb6,0xf3]
|
|
@ CHECK: vtrn.16 d3, d9 @ encoding: [0x89,0x30,0xb6,0xf3]
|
|
@ CHECK: vtrn.32 d3, d9 @ encoding: [0x89,0x30,0xba,0xf3]
|
|
@ CHECK: vtrn.32 d3, d9 @ encoding: [0x89,0x30,0xba,0xf3]
|
|
@ CHECK: vtrn.32 d3, d9 @ encoding: [0x89,0x30,0xba,0xf3]
|
|
@ CHECK: vtrn.32 d3, d9 @ encoding: [0x89,0x30,0xba,0xf3]
|
|
@ CHECK: vtrn.32 d3, d9 @ encoding: [0x89,0x30,0xba,0xf3]
|
|
|
|
@ CHECK: vtrn.8 q14, q6 @ encoding: [0xcc,0xc0,0xf2,0xf3]
|
|
@ CHECK: vtrn.8 q14, q6 @ encoding: [0xcc,0xc0,0xf2,0xf3]
|
|
@ CHECK: vtrn.8 q14, q6 @ encoding: [0xcc,0xc0,0xf2,0xf3]
|
|
@ CHECK: vtrn.8 q14, q6 @ encoding: [0xcc,0xc0,0xf2,0xf3]
|
|
@ CHECK: vtrn.16 q14, q6 @ encoding: [0xcc,0xc0,0xf6,0xf3]
|
|
@ CHECK: vtrn.16 q14, q6 @ encoding: [0xcc,0xc0,0xf6,0xf3]
|
|
@ CHECK: vtrn.16 q14, q6 @ encoding: [0xcc,0xc0,0xf6,0xf3]
|
|
@ CHECK: vtrn.16 q14, q6 @ encoding: [0xcc,0xc0,0xf6,0xf3]
|
|
@ CHECK: vtrn.32 q14, q6 @ encoding: [0xcc,0xc0,0xfa,0xf3]
|
|
@ CHECK: vtrn.32 q14, q6 @ encoding: [0xcc,0xc0,0xfa,0xf3]
|
|
@ CHECK: vtrn.32 q14, q6 @ encoding: [0xcc,0xc0,0xfa,0xf3]
|
|
@ CHECK: vtrn.32 q14, q6 @ encoding: [0xcc,0xc0,0xfa,0xf3]
|
|
@ CHECK: vtrn.32 q14, q6 @ encoding: [0xcc,0xc0,0xfa,0xf3]
|
|
|