mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2025-11-02 07:17:36 +00:00
[X86][SSE] Vector shift test cleanup. NFC.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@241730 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
@@ -792,16 +792,16 @@ define <8 x i16> @constant_shift_v8i16(<8 x i16> %a) {
|
||||
; AVX1-LABEL: constant_shift_v8i16:
|
||||
; AVX1: # BB#0:
|
||||
; AVX1-NEXT: vpsraw $8, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,4112,8224,12336,16448,20560,24672,28784]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,4112,8224,12336,16448,20560,24672,28784]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsraw $4, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,8224,16448,24672,32896,41120,49344,57568]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,8224,16448,24672,32896,41120,49344,57568]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsraw $2, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,16448,32896,49344,256,16704,33152,49600]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,16448,32896,49344,256,16704,33152,49600]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsraw $1, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,32896,256,33152,512,33408,768,33664]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,32896,256,33152,512,33408,768,33664]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: retq
|
||||
;
|
||||
|
||||
@@ -614,16 +614,16 @@ define <8 x i16> @constant_shift_v8i16(<8 x i16> %a) {
|
||||
; AVX1-LABEL: constant_shift_v8i16:
|
||||
; AVX1: # BB#0:
|
||||
; AVX1-NEXT: vpsrlw $8, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,4112,8224,12336,16448,20560,24672,28784]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,4112,8224,12336,16448,20560,24672,28784]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsrlw $4, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,8224,16448,24672,32896,41120,49344,57568]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,8224,16448,24672,32896,41120,49344,57568]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsrlw $2, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,16448,32896,49344,256,16704,33152,49600]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,16448,32896,49344,256,16704,33152,49600]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: vpsrlw $1, %xmm0, %xmm1
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm2 # xmm2 = [0,32896,256,33152,512,33408,768,33664]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,32896,256,33152,512,33408,768,33664]
|
||||
; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
|
||||
; AVX1-NEXT: retq
|
||||
;
|
||||
@@ -770,8 +770,8 @@ define <16 x i8> @splatconstant_shift_v16i8(<16 x i8> %a) {
|
||||
;
|
||||
; AVX-LABEL: splatconstant_shift_v16i8:
|
||||
; AVX: # BB#0:
|
||||
; AVX-NEXT: vpsrlw $3, %xmm0
|
||||
; AVX-NEXT: vpand {{.*}}(%rip), %xmm0
|
||||
; AVX-NEXT: vpsrlw $3, %xmm0, %xmm0
|
||||
; AVX-NEXT: vpand {{.*}}(%rip), %xmm0, %xmm0
|
||||
; AVX-NEXT: retq
|
||||
%shift = lshr <16 x i8> %a, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
|
||||
ret <16 x i8> %shift
|
||||
|
||||
@@ -540,8 +540,8 @@ define <32 x i8> @splatconstant_shift_v32i8(<32 x i8> %a) {
|
||||
;
|
||||
; AVX2-LABEL: splatconstant_shift_v32i8:
|
||||
; AVX2: # BB#0:
|
||||
; AVX2-NEXT: vpsrlw $3, %ymm0
|
||||
; AVX2-NEXT: vpand {{.*}}(%rip), %ymm0
|
||||
; AVX2-NEXT: vpsrlw $3, %ymm0, %ymm0
|
||||
; AVX2-NEXT: vpand {{.*}}(%rip), %ymm0, %ymm0
|
||||
; AVX2-NEXT: retq
|
||||
%shift = lshr <32 x i8> %a, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
|
||||
ret <32 x i8> %shift
|
||||
|
||||
@@ -631,8 +631,8 @@ define <16 x i8> @splatconstant_shift_v16i8(<16 x i8> %a) {
|
||||
;
|
||||
; AVX-LABEL: splatconstant_shift_v16i8:
|
||||
; AVX: # BB#0:
|
||||
; AVX-NEXT: vpsllw $3, %xmm0
|
||||
; AVX-NEXT: vpand {{.*}}(%rip), %xmm0
|
||||
; AVX-NEXT: vpsllw $3, %xmm0, %xmm0
|
||||
; AVX-NEXT: vpand {{.*}}(%rip), %xmm0, %xmm0
|
||||
; AVX-NEXT: retq
|
||||
%shift = shl <16 x i8> %a, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
|
||||
ret <16 x i8> %shift
|
||||
|
||||
@@ -193,7 +193,7 @@ define <8 x i32> @splatvar_shift_v8i32(<8 x i32> %a, <8 x i32> %b) {
|
||||
; AVX2-LABEL: splatvar_shift_v8i32:
|
||||
; AVX2: # BB#0:
|
||||
; AVX2-NEXT: vpxor %xmm2, %xmm2, %xmm2
|
||||
; AVX2-NEXT: vpblendw $3, %xmm1, %xmm2, %xmm1 # xmm1 = xmm1[0,1],xmm2[2,3,4,5,6,7]
|
||||
; AVX2-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0,1],xmm2[2,3,4,5,6,7]
|
||||
; AVX2-NEXT: vpslld %xmm1, %ymm0, %ymm0
|
||||
; AVX2-NEXT: retq
|
||||
%splat = shufflevector <8 x i32> %b, <8 x i32> undef, <8 x i32> zeroinitializer
|
||||
@@ -341,7 +341,7 @@ define <32 x i8> @constant_shift_v32i8(<32 x i8> %a) {
|
||||
; AVX1-NEXT: vpsllw $4, %xmm1, %xmm2
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm3 = [240,240,240,240,240,240,240,240,240,240,240,240,240,240,240,240]
|
||||
; AVX1-NEXT: vpand %xmm3, %xmm2, %xmm2
|
||||
; AVX1-NEXT: vmovdqa {{.*}}(%rip), %xmm4 # xmm4 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0]
|
||||
; AVX1-NEXT: vmovdqa {{.*#+}} xmm4 = [0,1,2,3,4,5,6,7,7,6,5,4,3,2,1,0]
|
||||
; AVX1-NEXT: vpsllw $5, %xmm4, %xmm4
|
||||
; AVX1-NEXT: vpblendvb %xmm4, %xmm2, %xmm1, %xmm1
|
||||
; AVX1-NEXT: vpsllw $2, %xmm1, %xmm2
|
||||
|
||||
Reference in New Issue
Block a user