mirror of
				https://github.com/c64scene-ar/llvm-6502.git
				synced 2025-11-04 05:17:07 +00:00 
			
		
		
		
	halfword being emitted to the stream first. rdar://8728174 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@120848 91177308-0d34-0410-b5e6-96231b3b80d8
		
			
				
	
	
		
			139 lines
		
	
	
		
			5.9 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
			
		
		
	
	
			139 lines
		
	
	
		
			5.9 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
@ RUN: llvm-mc -mcpu=cortex-a8 -triple thumb-unknown-unknown -show-encoding < %s | FileCheck %s
 | 
						|
 | 
						|
.code 16
 | 
						|
 | 
						|
@ CHECK: vadd.i8	d16, d17, d16           @ encoding: [0x41,0xef,0xa0,0x08]
 | 
						|
	vadd.i8	d16, d17, d16
 | 
						|
@ CHECK: vadd.i16	d16, d17, d16   @ encoding: [0x51,0xef,0xa0,0x08]
 | 
						|
	vadd.i16	d16, d17, d16
 | 
						|
@ CHECK: vadd.i64	d16, d17, d16   @ encoding: [0x71,0xef,0xa0,0x08]
 | 
						|
	vadd.i64	d16, d17, d16
 | 
						|
@ CHECK: vadd.i32	d16, d17, d16   @ encoding: [0x61,0xef,0xa0,0x08]
 | 
						|
	vadd.i32	d16, d17, d16
 | 
						|
@ CHECK: vadd.f32	d16, d16, d17   @ encoding: [0x40,0xef,0xa1,0x0d]
 | 
						|
	vadd.f32	d16, d16, d17
 | 
						|
@ CHECK: vadd.f32	q8, q8, q9      @ encoding: [0x40,0xef,0xe2,0x0d]
 | 
						|
	vadd.f32	q8, q8, q9
 | 
						|
 | 
						|
@ CHECK: vaddl.s8	q8, d17, d16    @ encoding: [0xc1,0xef,0xa0,0x00]
 | 
						|
	vaddl.s8	q8, d17, d16
 | 
						|
@ CHECK: vaddl.s16	q8, d17, d16    @ encoding: [0xd1,0xef,0xa0,0x00]
 | 
						|
	vaddl.s16	q8, d17, d16
 | 
						|
@ CHECK: vaddl.s32	q8, d17, d16    @ encoding: [0xe1,0xef,0xa0,0x00]
 | 
						|
	vaddl.s32	q8, d17, d16
 | 
						|
@ CHECK: vaddl.u8	q8, d17, d16    @ encoding: [0xc1,0xff,0xa0,0x00]
 | 
						|
	vaddl.u8	q8, d17, d16
 | 
						|
@ CHECK: vaddl.u16	q8, d17, d16    @ encoding: [0xd1,0xff,0xa0,0x00]
 | 
						|
	vaddl.u16	q8, d17, d16
 | 
						|
@ CHECK: vaddl.u32	q8, d17, d16    @ encoding: [0xe1,0xff,0xa0,0x00]
 | 
						|
	vaddl.u32	q8, d17, d16
 | 
						|
 | 
						|
@ CHECK: vaddw.s8	q8, q8, d18     @ encoding: [0xc0,0xef,0xa2,0x01]
 | 
						|
	vaddw.s8	q8, q8, d18
 | 
						|
@ CHECK: vaddw.s16	q8, q8, d18     @ encoding: [0xd0,0xef,0xa2,0x01]
 | 
						|
	vaddw.s16	q8, q8, d18
 | 
						|
@ CHECK: vaddw.s32	q8, q8, d18     @ encoding: [0xe0,0xef,0xa2,0x01]
 | 
						|
	vaddw.s32	q8, q8, d18
 | 
						|
@ CHECK: vaddw.u8	q8, q8, d18     @ encoding: [0xc0,0xff,0xa2,0x01]
 | 
						|
	vaddw.u8	q8, q8, d18
 | 
						|
@ CHECK: vaddw.u16	q8, q8, d18     @ encoding: [0xd0,0xff,0xa2,0x01]
 | 
						|
	vaddw.u16	q8, q8, d18
 | 
						|
@ CHECK: vaddw.u32	q8, q8, d18     @ encoding: [0xe0,0xff,0xa2,0x01]
 | 
						|
	vaddw.u32	q8, q8, d18
 | 
						|
 | 
						|
@ CHECK: vhadd.s8	d16, d16, d17   @ encoding: [0x40,0xef,0xa1,0x00]
 | 
						|
	vhadd.s8	d16, d16, d17
 | 
						|
@ CHECK: vhadd.s16	d16, d16, d17   @ encoding: [0x50,0xef,0xa1,0x00]
 | 
						|
	vhadd.s16	d16, d16, d17
 | 
						|
@ CHECK: vhadd.s32	d16, d16, d17   @ encoding: [0x60,0xef,0xa1,0x00]
 | 
						|
	vhadd.s32	d16, d16, d17
 | 
						|
@ CHECK: vhadd.u8	d16, d16, d17   @ encoding: [0x40,0xff,0xa1,0x00]
 | 
						|
	vhadd.u8	d16, d16, d17
 | 
						|
@ CHECK: vhadd.u16	d16, d16, d17   @ encoding: [0x50,0xff,0xa1,0x00]
 | 
						|
	vhadd.u16	d16, d16, d17
 | 
						|
@ CHECK: vhadd.u32	d16, d16, d17   @ encoding: [0x60,0xff,0xa1,0x00]
 | 
						|
	vhadd.u32	d16, d16, d17
 | 
						|
@ CHECK: vhadd.s8	q8, q8, q9      @ encoding: [0x40,0xef,0xe2,0x00]
 | 
						|
	vhadd.s8	q8, q8, q9
 | 
						|
@ CHECK: vhadd.s16	q8, q8, q9      @ encoding: [0x50,0xef,0xe2,0x00]
 | 
						|
	vhadd.s16	q8, q8, q9
 | 
						|
@ CHECK: vhadd.s32	q8, q8, q9      @ encoding: [0x60,0xef,0xe2,0x00]
 | 
						|
	vhadd.s32	q8, q8, q9
 | 
						|
  @ CHECK: vhadd.u8	q8, q8, q9      @ encoding: [0x40,0xff,0xe2,0x00]
 | 
						|
	vhadd.u8	q8, q8, q9
 | 
						|
@ CHECK: vhadd.u16	q8, q8, q9      @ encoding: [0x50,0xff,0xe2,0x00]
 | 
						|
	vhadd.u16	q8, q8, q9
 | 
						|
@ CHECK: vhadd.u32	q8, q8, q9      @ encoding: [0x60,0xff,0xe2,0x00]
 | 
						|
	vhadd.u32	q8, q8, q9
 | 
						|
	
 | 
						|
@ CHECK: vrhadd.s8	d16, d16, d17   @ encoding: [0x40,0xef,0xa1,0x01]
 | 
						|
	vrhadd.s8	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.s16	d16, d16, d17   @ encoding: [0x50,0xef,0xa1,0x01]
 | 
						|
	vrhadd.s16	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.s32	d16, d16, d17   @ encoding: [0x60,0xef,0xa1,0x01]
 | 
						|
	vrhadd.s32	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.u8	d16, d16, d17   @ encoding: [0x40,0xff,0xa1,0x01]
 | 
						|
	vrhadd.u8	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.u16	d16, d16, d17   @ encoding: [0x50,0xff,0xa1,0x01]
 | 
						|
	vrhadd.u16	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.u32	d16, d16, d17   @ encoding: [0x60,0xff,0xa1,0x01]
 | 
						|
	vrhadd.u32	d16, d16, d17
 | 
						|
@ CHECK: vrhadd.s8	q8, q8, q9      @ encoding: [0x40,0xef,0xe2,0x01]
 | 
						|
	vrhadd.s8	q8, q8, q9
 | 
						|
@ CHECK: vrhadd.s16	q8, q8, q9      @ encoding: [0x50,0xef,0xe2,0x01]
 | 
						|
	vrhadd.s16	q8, q8, q9
 | 
						|
@ CHECK: vrhadd.s32	q8, q8, q9      @ encoding: [0x60,0xef,0xe2,0x01]
 | 
						|
	vrhadd.s32	q8, q8, q9
 | 
						|
@ CHECK: vrhadd.u8	q8, q8, q9      @ encoding: [0x40,0xff,0xe2,0x01]
 | 
						|
	vrhadd.u8	q8, q8, q9
 | 
						|
@ CHECK: vrhadd.u16	q8, q8, q9      @ encoding: [0x50,0xff,0xe2,0x01]
 | 
						|
	vrhadd.u16	q8, q8, q9
 | 
						|
@ CHECK: vrhadd.u32	q8, q8, q9      @ encoding: [0x60,0xff,0xe2,0x01]
 | 
						|
	vrhadd.u32	q8, q8, q9
 | 
						|
 | 
						|
@ CHECK: vqadd.s8	d16, d16, d17   @ encoding: [0x40,0xef,0xb1,0x00]
 | 
						|
	vqadd.s8	d16, d16, d17
 | 
						|
@ CHECK: vqadd.s16	d16, d16, d17   @ encoding: [0x50,0xef,0xb1,0x00]
 | 
						|
	vqadd.s16	d16, d16, d17
 | 
						|
@ CHECK: vqadd.s32	d16, d16, d17   @ encoding: [0x60,0xef,0xb1,0x00]
 | 
						|
	vqadd.s32	d16, d16, d17
 | 
						|
@ CHECK: vqadd.s64	d16, d16, d17   @ encoding: [0x70,0xef,0xb1,0x00]
 | 
						|
	vqadd.s64	d16, d16, d17
 | 
						|
@ CHECK: vqadd.u8	d16, d16, d17   @ encoding: [0x40,0xff,0xb1,0x00]
 | 
						|
	vqadd.u8	d16, d16, d17
 | 
						|
@ CHECK: vqadd.u16	d16, d16, d17   @ encoding: [0x50,0xff,0xb1,0x00]
 | 
						|
	vqadd.u16	d16, d16, d17
 | 
						|
@ CHECK: vqadd.u32	d16, d16, d17   @ encoding: [0x60,0xff,0xb1,0x00]
 | 
						|
	vqadd.u32	d16, d16, d17
 | 
						|
@ CHECK: vqadd.u64	d16, d16, d17   @ encoding: [0x70,0xff,0xb1,0x00]
 | 
						|
	vqadd.u64	d16, d16, d17
 | 
						|
@ CHECK: vqadd.s8	q8, q8, q9      @ encoding: [0x40,0xef,0xf2,0x00]
 | 
						|
	vqadd.s8	q8, q8, q9
 | 
						|
@ CHECK: vqadd.s16	q8, q8, q9      @ encoding: [0x50,0xef,0xf2,0x00]
 | 
						|
	vqadd.s16	q8, q8, q9
 | 
						|
@ CHECK: vqadd.s32	q8, q8, q9      @ encoding: [0x60,0xef,0xf2,0x00]
 | 
						|
	vqadd.s32	q8, q8, q9
 | 
						|
@ CHECK: vqadd.s64	q8, q8, q9      @ encoding: [0x70,0xef,0xf2,0x00]
 | 
						|
	vqadd.s64	q8, q8, q9
 | 
						|
@ CHECK: vqadd.u8	q8, q8, q9      @ encoding: [0x40,0xff,0xf2,0x00]
 | 
						|
	vqadd.u8	q8, q8, q9
 | 
						|
@ CHECK: vqadd.u16	q8, q8, q9      @ encoding: [0x50,0xff,0xf2,0x00]
 | 
						|
	vqadd.u16	q8, q8, q9
 | 
						|
@ CHECK: vqadd.u32	q8, q8, q9      @ encoding: [0x60,0xff,0xf2,0x00]
 | 
						|
	vqadd.u32	q8, q8, q9
 | 
						|
@ CHECK: vqadd.u64	q8, q8, q9      @ encoding: [0x70,0xff,0xf2,0x00]
 | 
						|
	vqadd.u64	q8, q8, q9
 | 
						|
 | 
						|
@ CHECK: vaddhn.i16	d16, q8, q9     @ encoding: [0xc0,0xef,0xa2,0x04]
 | 
						|
	vaddhn.i16	d16, q8, q9
 | 
						|
@ CHECK: vaddhn.i32	d16, q8, q9     @ encoding: [0xd0,0xef,0xa2,0x04]
 | 
						|
	vaddhn.i32	d16, q8, q9
 | 
						|
@ CHECK: vaddhn.i64	d16, q8, q9     @ encoding: [0xe0,0xef,0xa2,0x04]
 | 
						|
	vaddhn.i64	d16, q8, q9
 | 
						|
@ CHECK: vraddhn.i16	d16, q8, q9     @ encoding: [0xc0,0xff,0xa2,0x04]
 | 
						|
	vraddhn.i16	d16, q8, q9
 | 
						|
@ CHECK: vraddhn.i32	d16, q8, q9     @ encoding: [0xd0,0xff,0xa2,0x04]
 | 
						|
	vraddhn.i32	d16, q8, q9
 | 
						|
@ CHECK: vraddhn.i64	d16, q8, q9     @ encoding: [0xe0,0xff,0xa2,0x04]
 | 
						|
	vraddhn.i64	d16, q8, q9
 |