2013-09-17 02:21:02 +00:00
|
|
|
; RUN: llc < %s -verify-machineinstrs -mtriple=aarch64-none-linux-gnu -mattr=+neon -fp-contract=fast | FileCheck %s
|
|
|
|
|
|
|
|
|
|
|
|
define <16 x i8> @ins16bw(<16 x i8> %tmp1, i8 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.b[15], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <16 x i8> %tmp1, i8 %tmp2, i32 15
|
|
|
|
ret <16 x i8> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i16> @ins8hw(<8 x i16> %tmp1, i16 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.h[6], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <8 x i16> %tmp1, i16 %tmp2, i32 6
|
|
|
|
ret <8 x i16> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i32> @ins4sw(<4 x i32> %tmp1, i32 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.s[2], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <4 x i32> %tmp1, i32 %tmp2, i32 2
|
|
|
|
ret <4 x i32> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i64> @ins2dw(<2 x i64> %tmp1, i64 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.d[1], {{x[0-31]+}}
|
|
|
|
%tmp3 = insertelement <2 x i64> %tmp1, i64 %tmp2, i32 1
|
|
|
|
ret <2 x i64> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i8> @ins8bw(<8 x i8> %tmp1, i8 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.b[5], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <8 x i8> %tmp1, i8 %tmp2, i32 5
|
|
|
|
ret <8 x i8> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i16> @ins4hw(<4 x i16> %tmp1, i16 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.h[3], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <4 x i16> %tmp1, i16 %tmp2, i32 3
|
|
|
|
ret <4 x i16> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i32> @ins2sw(<2 x i32> %tmp1, i32 %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.s[1], {{w[0-31]+}}
|
|
|
|
%tmp3 = insertelement <2 x i32> %tmp1, i32 %tmp2, i32 1
|
|
|
|
ret <2 x i32> %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define <16 x i8> @ins16b16(<16 x i8> %tmp1, <16 x i8> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.b[15], {{v[0-31]+}}.b[2]
|
|
|
|
%tmp3 = extractelement <16 x i8> %tmp1, i32 2
|
|
|
|
%tmp4 = insertelement <16 x i8> %tmp2, i8 %tmp3, i32 15
|
|
|
|
ret <16 x i8> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i16> @ins8h8(<8 x i16> %tmp1, <8 x i16> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.h[7], {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <8 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = insertelement <8 x i16> %tmp2, i16 %tmp3, i32 7
|
|
|
|
ret <8 x i16> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i32> @ins4s4(<4 x i32> %tmp1, <4 x i32> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.s[1], {{v[0-31]+}}.s[2]
|
|
|
|
%tmp3 = extractelement <4 x i32> %tmp1, i32 2
|
|
|
|
%tmp4 = insertelement <4 x i32> %tmp2, i32 %tmp3, i32 1
|
|
|
|
ret <4 x i32> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i64> @ins2d2(<2 x i64> %tmp1, <2 x i64> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.d[1], {{v[0-31]+}}.d[0]
|
|
|
|
%tmp3 = extractelement <2 x i64> %tmp1, i32 0
|
|
|
|
%tmp4 = insertelement <2 x i64> %tmp2, i64 %tmp3, i32 1
|
|
|
|
ret <2 x i64> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i8> @ins8b8(<8 x i8> %tmp1, <8 x i8> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.b[4], {{v[0-31]+}}.b[2]
|
|
|
|
%tmp3 = extractelement <8 x i8> %tmp1, i32 2
|
|
|
|
%tmp4 = insertelement <8 x i8> %tmp2, i8 %tmp3, i32 4
|
|
|
|
ret <8 x i8> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i16> @ins4h4(<4 x i16> %tmp1, <4 x i16> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.h[3], {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <4 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = insertelement <4 x i16> %tmp2, i16 %tmp3, i32 3
|
|
|
|
ret <4 x i16> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i32> @ins2s2(<2 x i32> %tmp1, <2 x i32> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.s[1], {{v[0-31]+}}.s[0]
|
|
|
|
%tmp3 = extractelement <2 x i32> %tmp1, i32 0
|
|
|
|
%tmp4 = insertelement <2 x i32> %tmp2, i32 %tmp3, i32 1
|
|
|
|
ret <2 x i32> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define <1 x i64> @ins1d1(<1 x i64> %tmp1, <1 x i64> %tmp2) {
|
|
|
|
;CHECK: ins {{v[0-31]+}}.d[0], {{v[0-31]+}}.d[0]
|
|
|
|
%tmp3 = extractelement <1 x i64> %tmp1, i32 0
|
|
|
|
%tmp4 = insertelement <1 x i64> %tmp2, i64 %tmp3, i32 0
|
|
|
|
ret <1 x i64> %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw16b(<16 x i8> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.b[8]
|
|
|
|
%tmp3 = extractelement <16 x i8> %tmp1, i32 8
|
|
|
|
%tmp4 = zext i8 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw8h(<8 x i16> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <8 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = zext i16 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw4s(<4 x i32> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.s[2]
|
|
|
|
%tmp3 = extractelement <4 x i32> %tmp1, i32 2
|
|
|
|
ret i32 %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define i64 @umovx2d(<2 x i64> %tmp1) {
|
|
|
|
;CHECK: umov {{x[0-31]+}}, {{v[0-31]+}}.d[0]
|
|
|
|
%tmp3 = extractelement <2 x i64> %tmp1, i32 0
|
|
|
|
ret i64 %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw8b(<8 x i8> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.b[7]
|
|
|
|
%tmp3 = extractelement <8 x i8> %tmp1, i32 7
|
|
|
|
%tmp4 = zext i8 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw4h(<4 x i16> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <4 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = zext i16 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @umovw2s(<2 x i32> %tmp1) {
|
|
|
|
;CHECK: umov {{w[0-31]+}}, {{v[0-31]+}}.s[1]
|
|
|
|
%tmp3 = extractelement <2 x i32> %tmp1, i32 1
|
|
|
|
ret i32 %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define i64 @umovx1d(<1 x i64> %tmp1) {
|
2013-09-24 02:47:27 +00:00
|
|
|
;CHECK: fmov {{x[0-31]+}}, {{d[0-31]+}}
|
2013-09-17 02:21:02 +00:00
|
|
|
%tmp3 = extractelement <1 x i64> %tmp1, i32 0
|
|
|
|
ret i64 %tmp3
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovw16b(<16 x i8> %tmp1) {
|
|
|
|
;CHECK: smov {{w[0-31]+}}, {{v[0-31]+}}.b[8]
|
|
|
|
%tmp3 = extractelement <16 x i8> %tmp1, i32 8
|
|
|
|
%tmp4 = sext i8 %tmp3 to i32
|
|
|
|
%tmp5 = add i32 5, %tmp4
|
|
|
|
ret i32 %tmp5
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovw8h(<8 x i16> %tmp1) {
|
|
|
|
;CHECK: smov {{w[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <8 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = sext i16 %tmp3 to i32
|
|
|
|
%tmp5 = add i32 5, %tmp4
|
|
|
|
ret i32 %tmp5
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovx16b(<16 x i8> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.b[8]
|
|
|
|
%tmp3 = extractelement <16 x i8> %tmp1, i32 8
|
|
|
|
%tmp4 = sext i8 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovx8h(<8 x i16> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <8 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = sext i16 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i64 @smovx4s(<4 x i32> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.s[2]
|
|
|
|
%tmp3 = extractelement <4 x i32> %tmp1, i32 2
|
|
|
|
%tmp4 = sext i32 %tmp3 to i64
|
|
|
|
ret i64 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovw8b(<8 x i8> %tmp1) {
|
|
|
|
;CHECK: smov {{w[0-31]+}}, {{v[0-31]+}}.b[4]
|
|
|
|
%tmp3 = extractelement <8 x i8> %tmp1, i32 4
|
|
|
|
%tmp4 = sext i8 %tmp3 to i32
|
|
|
|
%tmp5 = add i32 5, %tmp4
|
|
|
|
ret i32 %tmp5
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovw4h(<4 x i16> %tmp1) {
|
|
|
|
;CHECK: smov {{w[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <4 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = sext i16 %tmp3 to i32
|
|
|
|
%tmp5 = add i32 5, %tmp4
|
|
|
|
ret i32 %tmp5
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovx8b(<8 x i8> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.b[6]
|
|
|
|
%tmp3 = extractelement <8 x i8> %tmp1, i32 6
|
|
|
|
%tmp4 = sext i8 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @smovx4h(<4 x i16> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.h[2]
|
|
|
|
%tmp3 = extractelement <4 x i16> %tmp1, i32 2
|
|
|
|
%tmp4 = sext i16 %tmp3 to i32
|
|
|
|
ret i32 %tmp4
|
|
|
|
}
|
|
|
|
|
|
|
|
define i64 @smovx2s(<2 x i32> %tmp1) {
|
|
|
|
;CHECK: smov {{x[0-31]+}}, {{v[0-31]+}}.s[1]
|
|
|
|
%tmp3 = extractelement <2 x i32> %tmp1, i32 1
|
|
|
|
%tmp4 = sext i32 %tmp3 to i64
|
|
|
|
ret i64 %tmp4
|
|
|
|
}
|
|
|
|
|
2013-10-11 02:33:55 +00:00
|
|
|
define <8 x i8> @test_vcopy_lane_s8(<8 x i8> %v1, <8 x i8> %v2) {
|
|
|
|
;CHECK: ins {{v[0-9]+}}.b[5], {{v[0-9]+}}.b[3]
|
|
|
|
%vset_lane = shufflevector <8 x i8> %v1, <8 x i8> %v2, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 11, i32 6, i32 7>
|
|
|
|
ret <8 x i8> %vset_lane
|
|
|
|
}
|
|
|
|
|
|
|
|
define <16 x i8> @test_vcopyq_laneq_s8(<16 x i8> %v1, <16 x i8> %v2) {
|
|
|
|
;CHECK: ins {{v[0-9]+}}.b[14], {{v[0-9]+}}.b[6]
|
|
|
|
%vset_lane = shufflevector <16 x i8> %v1, <16 x i8> %v2, <16 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7, i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 22, i32 15>
|
|
|
|
ret <16 x i8> %vset_lane
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i8> @test_vcopy_lane_swap_s8(<8 x i8> %v1, <8 x i8> %v2) {
|
|
|
|
;CHECK: ins {{v[0-9]+}}.b[7], {{v[0-9]+}}.b[0]
|
|
|
|
%vset_lane = shufflevector <8 x i8> %v1, <8 x i8> %v2, <8 x i32> <i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 0>
|
|
|
|
ret <8 x i8> %vset_lane
|
|
|
|
}
|
|
|
|
|
|
|
|
define <16 x i8> @test_vcopyq_laneq_swap_s8(<16 x i8> %v1, <16 x i8> %v2) {
|
|
|
|
;CHECK: ins {{v[0-9]+}}.b[0], {{v[0-9]+}}.b[15]
|
|
|
|
%vset_lane = shufflevector <16 x i8> %v1, <16 x i8> %v2, <16 x i32> <i32 15, i32 17, i32 18, i32 19, i32 20, i32 21, i32 22, i32 23, i32 24, i32 25, i32 26, i32 27, i32 28, i32 29, i32 30, i32 31>
|
|
|
|
ret <16 x i8> %vset_lane
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i8> @test_vdup_n_u8(i8 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.8b, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <8 x i8> undef, i8 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <8 x i8> %vecinit.i, i8 %v1, i32 1
|
|
|
|
%vecinit2.i = insertelement <8 x i8> %vecinit1.i, i8 %v1, i32 2
|
|
|
|
%vecinit3.i = insertelement <8 x i8> %vecinit2.i, i8 %v1, i32 3
|
|
|
|
%vecinit4.i = insertelement <8 x i8> %vecinit3.i, i8 %v1, i32 4
|
|
|
|
%vecinit5.i = insertelement <8 x i8> %vecinit4.i, i8 %v1, i32 5
|
|
|
|
%vecinit6.i = insertelement <8 x i8> %vecinit5.i, i8 %v1, i32 6
|
|
|
|
%vecinit7.i = insertelement <8 x i8> %vecinit6.i, i8 %v1, i32 7
|
|
|
|
ret <8 x i8> %vecinit7.i
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i16> @test_vdup_n_u16(i16 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.4h, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <4 x i16> undef, i16 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <4 x i16> %vecinit.i, i16 %v1, i32 1
|
|
|
|
%vecinit2.i = insertelement <4 x i16> %vecinit1.i, i16 %v1, i32 2
|
|
|
|
%vecinit3.i = insertelement <4 x i16> %vecinit2.i, i16 %v1, i32 3
|
|
|
|
ret <4 x i16> %vecinit3.i
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i32> @test_vdup_n_u32(i32 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.2s, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <2 x i32> undef, i32 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <2 x i32> %vecinit.i, i32 %v1, i32 1
|
|
|
|
ret <2 x i32> %vecinit1.i
|
|
|
|
}
|
|
|
|
|
|
|
|
define <1 x i64> @test_vdup_n_u64(i64 %v1) #0 {
|
|
|
|
;CHECK: fmov {{d[0-9]+}}, {{x[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <1 x i64> undef, i64 %v1, i32 0
|
|
|
|
ret <1 x i64> %vecinit.i
|
|
|
|
}
|
2013-09-17 02:21:02 +00:00
|
|
|
|
2013-10-11 02:33:55 +00:00
|
|
|
define <16 x i8> @test_vdupq_n_u8(i8 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.16b, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <16 x i8> undef, i8 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <16 x i8> %vecinit.i, i8 %v1, i32 1
|
|
|
|
%vecinit2.i = insertelement <16 x i8> %vecinit1.i, i8 %v1, i32 2
|
|
|
|
%vecinit3.i = insertelement <16 x i8> %vecinit2.i, i8 %v1, i32 3
|
|
|
|
%vecinit4.i = insertelement <16 x i8> %vecinit3.i, i8 %v1, i32 4
|
|
|
|
%vecinit5.i = insertelement <16 x i8> %vecinit4.i, i8 %v1, i32 5
|
|
|
|
%vecinit6.i = insertelement <16 x i8> %vecinit5.i, i8 %v1, i32 6
|
|
|
|
%vecinit7.i = insertelement <16 x i8> %vecinit6.i, i8 %v1, i32 7
|
|
|
|
%vecinit8.i = insertelement <16 x i8> %vecinit7.i, i8 %v1, i32 8
|
|
|
|
%vecinit9.i = insertelement <16 x i8> %vecinit8.i, i8 %v1, i32 9
|
|
|
|
%vecinit10.i = insertelement <16 x i8> %vecinit9.i, i8 %v1, i32 10
|
|
|
|
%vecinit11.i = insertelement <16 x i8> %vecinit10.i, i8 %v1, i32 11
|
|
|
|
%vecinit12.i = insertelement <16 x i8> %vecinit11.i, i8 %v1, i32 12
|
|
|
|
%vecinit13.i = insertelement <16 x i8> %vecinit12.i, i8 %v1, i32 13
|
|
|
|
%vecinit14.i = insertelement <16 x i8> %vecinit13.i, i8 %v1, i32 14
|
|
|
|
%vecinit15.i = insertelement <16 x i8> %vecinit14.i, i8 %v1, i32 15
|
|
|
|
ret <16 x i8> %vecinit15.i
|
|
|
|
}
|
2013-09-17 02:21:02 +00:00
|
|
|
|
2013-10-11 02:33:55 +00:00
|
|
|
define <8 x i16> @test_vdupq_n_u16(i16 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.8h, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <8 x i16> undef, i16 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <8 x i16> %vecinit.i, i16 %v1, i32 1
|
|
|
|
%vecinit2.i = insertelement <8 x i16> %vecinit1.i, i16 %v1, i32 2
|
|
|
|
%vecinit3.i = insertelement <8 x i16> %vecinit2.i, i16 %v1, i32 3
|
|
|
|
%vecinit4.i = insertelement <8 x i16> %vecinit3.i, i16 %v1, i32 4
|
|
|
|
%vecinit5.i = insertelement <8 x i16> %vecinit4.i, i16 %v1, i32 5
|
|
|
|
%vecinit6.i = insertelement <8 x i16> %vecinit5.i, i16 %v1, i32 6
|
|
|
|
%vecinit7.i = insertelement <8 x i16> %vecinit6.i, i16 %v1, i32 7
|
|
|
|
ret <8 x i16> %vecinit7.i
|
|
|
|
}
|
2013-09-17 02:21:02 +00:00
|
|
|
|
2013-10-11 02:33:55 +00:00
|
|
|
define <4 x i32> @test_vdupq_n_u32(i32 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.4s, {{w[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <4 x i32> undef, i32 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <4 x i32> %vecinit.i, i32 %v1, i32 1
|
|
|
|
%vecinit2.i = insertelement <4 x i32> %vecinit1.i, i32 %v1, i32 2
|
|
|
|
%vecinit3.i = insertelement <4 x i32> %vecinit2.i, i32 %v1, i32 3
|
|
|
|
ret <4 x i32> %vecinit3.i
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i64> @test_vdupq_n_u64(i64 %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.2d, {{x[0-9]+}}
|
|
|
|
%vecinit.i = insertelement <2 x i64> undef, i64 %v1, i32 0
|
|
|
|
%vecinit1.i = insertelement <2 x i64> %vecinit.i, i64 %v1, i32 1
|
|
|
|
ret <2 x i64> %vecinit1.i
|
|
|
|
}
|
2013-09-17 02:21:02 +00:00
|
|
|
|
2013-10-11 02:33:55 +00:00
|
|
|
define <8 x i8> @test_vdup_lane_s8(<8 x i8> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.8b, {{v[0-9]+}}.b[5]
|
|
|
|
%shuffle = shufflevector <8 x i8> %v1, <8 x i8> undef, <8 x i32> <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5>
|
|
|
|
ret <8 x i8> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i16> @test_vdup_lane_s16(<4 x i16> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.4h, {{v[0-9]+}}.h[2]
|
|
|
|
%shuffle = shufflevector <4 x i16> %v1, <4 x i16> undef, <4 x i32> <i32 2, i32 2, i32 2, i32 2>
|
|
|
|
ret <4 x i16> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i32> @test_vdup_lane_s32(<2 x i32> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.2s, {{v[0-9]+}}.s[1]
|
|
|
|
%shuffle = shufflevector <2 x i32> %v1, <2 x i32> undef, <2 x i32> <i32 1, i32 1>
|
|
|
|
ret <2 x i32> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <16 x i8> @test_vdupq_lane_s8(<8 x i8> %v1) #0 {
|
|
|
|
;CHECK: {{v[0-9]+}}.16b, {{v[0-9]+}}.b[5]
|
|
|
|
%shuffle = shufflevector <8 x i8> %v1, <8 x i8> undef, <16 x i32> <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5>
|
|
|
|
ret <16 x i8> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i16> @test_vdupq_lane_s16(<4 x i16> %v1) #0 {
|
|
|
|
;CHECK: {{v[0-9]+}}.8h, {{v[0-9]+}}.h[2]
|
|
|
|
%shuffle = shufflevector <4 x i16> %v1, <4 x i16> undef, <8 x i32> <i32 2, i32 2, i32 2, i32 2, i32 2, i32 2, i32 2, i32 2>
|
|
|
|
ret <8 x i16> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i32> @test_vdupq_lane_s32(<2 x i32> %v1) #0 {
|
|
|
|
;CHECK: {{v[0-9]+}}.4s, {{v[0-9]+}}.s[1]
|
|
|
|
%shuffle = shufflevector <2 x i32> %v1, <2 x i32> undef, <4 x i32> <i32 1, i32 1, i32 1, i32 1>
|
|
|
|
ret <4 x i32> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i64> @test_vdupq_lane_s64(<1 x i64> %v1) #0 {
|
|
|
|
;CHECK: {{v[0-9]+}}.2d, {{v[0-9]+}}.d[0]
|
|
|
|
%shuffle = shufflevector <1 x i64> %v1, <1 x i64> undef, <2 x i32> zeroinitializer
|
|
|
|
ret <2 x i64> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i8> @test_vdup_laneq_s8(<16 x i8> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.8b, {{v[0-9]+}}.b[5]
|
|
|
|
%shuffle = shufflevector <16 x i8> %v1, <16 x i8> undef, <8 x i32> <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5>
|
|
|
|
ret <8 x i8> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i16> @test_vdup_laneq_s16(<8 x i16> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.4h, {{v[0-9]+}}.h[2]
|
|
|
|
%shuffle = shufflevector <8 x i16> %v1, <8 x i16> undef, <4 x i32> <i32 2, i32 2, i32 2, i32 2>
|
|
|
|
ret <4 x i16> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i32> @test_vdup_laneq_s32(<4 x i32> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.2s, {{v[0-9]+}}.s[1]
|
|
|
|
%shuffle = shufflevector <4 x i32> %v1, <4 x i32> undef, <2 x i32> <i32 1, i32 1>
|
|
|
|
ret <2 x i32> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <16 x i8> @test_vdupq_laneq_s8(<16 x i8> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.16b, {{v[0-9]+}}.b[5]
|
|
|
|
%shuffle = shufflevector <16 x i8> %v1, <16 x i8> undef, <16 x i32> <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5>
|
|
|
|
ret <16 x i8> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <8 x i16> @test_vdupq_laneq_s16(<8 x i16> %v1) #0 {
|
|
|
|
;CHECK: {{v[0-9]+}}.8h, {{v[0-9]+}}.h[2]
|
|
|
|
%shuffle = shufflevector <8 x i16> %v1, <8 x i16> undef, <8 x i32> <i32 2, i32 2, i32 2, i32 2, i32 2, i32 2, i32 2, i32 2>
|
|
|
|
ret <8 x i16> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <4 x i32> @test_vdupq_laneq_s32(<4 x i32> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.4s, {{v[0-9]+}}.s[1]
|
|
|
|
%shuffle = shufflevector <4 x i32> %v1, <4 x i32> undef, <4 x i32> <i32 1, i32 1, i32 1, i32 1>
|
|
|
|
ret <4 x i32> %shuffle
|
|
|
|
}
|
|
|
|
|
|
|
|
define <2 x i64> @test_vdupq_laneq_s64(<2 x i64> %v1) #0 {
|
|
|
|
;CHECK: dup {{v[0-9]+}}.2d, {{v[0-9]+}}.d[0]
|
|
|
|
%shuffle = shufflevector <2 x i64> %v1, <2 x i64> undef, <2 x i32> zeroinitializer
|
|
|
|
ret <2 x i64> %shuffle
|
|
|
|
}
|
2013-09-17 02:21:02 +00:00
|
|
|
|