mirror of
				https://github.com/c64scene-ar/llvm-6502.git
				synced 2025-10-30 16:17:05 +00:00 
			
		
		
		
	shuffles using AVX and AVX2 instructions. This fixes PR21138, one of the few remaining regressions impacting benchmarks from the new vector shuffle lowering. You may note that it "regresses" many of the vperm2x128 test cases -- these were actually "improved" by the naive lowering that the new shuffle lowering previously did. This regression gave me fits. I had this patch ready-to-go about an hour after flipping the switch but wasn't sure how to have the best of both worlds here and thought the correct solution might be a completely different approach to lowering these vector shuffles. I'm now convinced this is the correct lowering and the missed optimizations shown in vperm2x128 are actually due to missing target-independent DAG combines. I've even written most of the needed DAG combine and will submit it shortly, but this part is ready and should help some real-world benchmarks out. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@219079 91177308-0d34-0410-b5e6-96231b3b80d8
		
			
				
	
	
		
			44 lines
		
	
	
		
			1.9 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
			
		
		
	
	
			44 lines
		
	
	
		
			1.9 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
| ; RUN: llc < %s -march=x86-64 -mtriple=x86_64-apple-darwin -mcpu=knl | FileCheck %s
 | |
| 
 | |
| define <16 x i32> @test1(i32* %x) {
 | |
| ; CHECK-LABEL: test1:
 | |
| ; CHECK:       ## BB#0:
 | |
| ; CHECK-NEXT:    vmovd (%rdi), %xmm0
 | |
| ; CHECK-NEXT:    vinserti128 $1, %xmm0, %ymm0, %ymm0
 | |
| ; CHECK-NEXT:    vpxor %ymm1, %ymm1, %ymm1
 | |
| ; CHECK-NEXT:    vpblendd {{.*#+}} ymm0 = ymm1[0,1,2,3],ymm0[4],ymm1[5,6,7]
 | |
| ; CHECK-NEXT:    vinserti64x4 $1, %ymm1, %zmm0, %zmm0
 | |
| ; CHECK-NEXT:    retq
 | |
|    %y = load i32* %x, align 4
 | |
|    %res = insertelement <16 x i32>zeroinitializer, i32 %y, i32 4
 | |
|    ret <16 x i32>%res
 | |
| }
 | |
| 
 | |
| define <16 x i32> @test2(<16 x i32> %x) {
 | |
| ; CHECK-LABEL: test2:
 | |
| ; CHECK:       ## BB#0:
 | |
| ; CHECK-NEXT:    vpaddd {{.*}}(%rip){1to16}, %zmm0, %zmm0
 | |
| ; CHECK-NEXT:    retq
 | |
|    %res = add <16 x i32><i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1, i32 -1>, %x
 | |
|    ret <16 x i32>%res
 | |
| }
 | |
| 
 | |
| define <16 x float> @test3(<4 x float> %a) {
 | |
| ; CHECK-LABEL: test3:
 | |
| ; CHECK:       ## BB#0:
 | |
| ; CHECK-NEXT:    vpermilpd {{.*#+}} xmm1 = xmm0[1,0]
 | |
| ; CHECK-NEXT:    vxorps %xmm2, %xmm2, %xmm2
 | |
| ; CHECK-NEXT:    vmovss %xmm0, %xmm2, %xmm0
 | |
| ; CHECK-NEXT:    vmovss %xmm1, %xmm2, %xmm1
 | |
| ; CHECK-NEXT:    vshufps {{.*#+}} xmm0 = xmm1[1,0],xmm0[0,1]
 | |
| ; CHECK-NEXT:    vinsertf128 $1, %xmm0, %ymm2, %ymm0
 | |
| ; CHECK-NEXT:    vxorps %ymm1, %ymm1, %ymm1
 | |
| ; CHECK-NEXT:    vinsertf64x4 $1, %ymm1, %zmm0, %zmm0
 | |
| ; CHECK-NEXT:    retq
 | |
|   %b = extractelement <4 x float> %a, i32 2
 | |
|   %c = insertelement <16 x float> <float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float undef, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00>, float %b, i32 5
 | |
|   %b1 = extractelement <4 x float> %a, i32 0
 | |
|   %c1 = insertelement <16 x float> %c, float %b1, i32 6
 | |
|   ret <16 x float>%c1
 | |
| }
 |