mirror of
				https://github.com/c64scene-ar/llvm-6502.git
				synced 2025-10-31 08:16:47 +00:00 
			
		
		
		
	The current approach to lower a vsetult is to flip the sign bit of the
operands, swap the operands and then use a (signed) pcmpgt.  psubus (unsigned
saturating subtract) can be used to emulate a vsetult more efficiently:
+    case ISD::SETULT: {
+      // If the comparison is against a constant we can turn this into a
+      // setule.  With psubus, setule does not require a swap.  This is
+      // beneficial because the constant in the register is no longer
+      // destructed as the destination so it can be hoisted out of a loop.
I also enable lowering via psubus in a few other cases where it's clearly
beneficial: setule and setuge if minu/maxu cannot be used.
    
rdar://problem/14338765
Patch by Adam Nemet <anemet@apple.com>.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@202301 91177308-0d34-0410-b5e6-96231b3b80d8
		
	
		
			
				
	
	
		
			182 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
			
		
		
	
	
			182 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
| ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=sse2 | FileCheck %s -check-prefix=SSE2
 | |
| ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=sse4.1 | FileCheck %s -check-prefix=SSE41
 | |
| ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=avx | FileCheck %s -check-prefix=AVX
 | |
| 
 | |
| define <16 x i8> @v16i8_icmp_uge(<16 x i8> %a, <16 x i8> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp uge <16 x i8> %a, %b
 | |
|   %2 = sext <16 x i1> %1 to <16 x i8>
 | |
|   ret <16 x i8> %2
 | |
| ; SSE2-LABEL: v16i8_icmp_uge:
 | |
| ; SSE2: pmaxub  %xmm0, %xmm1
 | |
| ; SSE2: pcmpeqb %xmm1, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v16i8_icmp_uge:
 | |
| ; SSE41: pmaxub  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqb %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v16i8_icmp_uge:
 | |
| ; AVX: vpmaxub  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| define <16 x i8> @v16i8_icmp_ule(<16 x i8> %a, <16 x i8> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp ule <16 x i8> %a, %b
 | |
|   %2 = sext <16 x i1> %1 to <16 x i8>
 | |
|   ret <16 x i8> %2
 | |
| ; SSE2-LABEL: v16i8_icmp_ule:
 | |
| ; SSE2: pminub  %xmm0, %xmm1
 | |
| ; SSE2: pcmpeqb %xmm1, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v16i8_icmp_ule:
 | |
| ; SSE41: pminub  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqb %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v16i8_icmp_ule:
 | |
| ; AVX: vpminub  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| 
 | |
| define <8 x i16> @v8i16_icmp_uge(<8 x i16> %a, <8 x i16> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp uge <8 x i16> %a, %b
 | |
|   %2 = sext <8 x i1> %1 to <8 x i16>
 | |
|   ret <8 x i16> %2
 | |
| ; SSE2-LABEL: v8i16_icmp_uge:
 | |
| ; SSE2: psubusw %xmm0, %xmm1
 | |
| ; SEE2: pxor    %xmm0, %xmm0
 | |
| ; SSE2: pcmpeqw %xmm1, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v8i16_icmp_uge:
 | |
| ; SSE41: pmaxuw  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqw %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v8i16_icmp_uge:
 | |
| ; AVX: vpmaxuw  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: vpcmpeqw %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| define <8 x i16> @v8i16_icmp_ule(<8 x i16> %a, <8 x i16> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp ule <8 x i16> %a, %b
 | |
|   %2 = sext <8 x i1> %1 to <8 x i16>
 | |
|   ret <8 x i16> %2
 | |
| ; SSE2-LABEL: v8i16_icmp_ule:
 | |
| ; SSE2: psubusw %xmm1, %xmm0
 | |
| ; SSE2: pxor    %xmm1, %xmm1
 | |
| ; SSE2: pcmpeqw %xmm0, %xmm1
 | |
| ; SSE2: movdqa  %xmm1, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v8i16_icmp_ule:
 | |
| ; SSE41: pminuw  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqw %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v8i16_icmp_ule:
 | |
| ; AVX: vpminuw  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: vpcmpeqw %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| 
 | |
| define <4 x i32> @v4i32_icmp_uge(<4 x i32> %a, <4 x i32> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp uge <4 x i32> %a, %b
 | |
|   %2 = sext <4 x i1> %1 to <4 x i32>
 | |
|   ret <4 x i32> %2
 | |
| ; SSE2-LABEL: v4i32_icmp_uge:
 | |
| ; SSE2: movdqa  {{.*}}(%rip), %xmm2
 | |
| ; SSE2: pxor    %xmm2, %xmm0
 | |
| ; SSE2: pxor    %xmm1, %xmm2
 | |
| ; SSE2: pcmpgtd %xmm0, %xmm2
 | |
| ; SSE2: pcmpeqd %xmm0, %xmm0
 | |
| ; SSE2: pxor    %xmm2, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v4i32_icmp_uge:
 | |
| ; SSE41: pmaxud  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqd %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v4i32_icmp_uge:
 | |
| ; AVX: vpmaxud  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: vpcmpeqd %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| define <4 x i32> @v4i32_icmp_ule(<4 x i32> %a, <4 x i32> %b) nounwind readnone ssp uwtable {
 | |
|   %1 = icmp ule <4 x i32> %a, %b
 | |
|   %2 = sext <4 x i1> %1 to <4 x i32>
 | |
|   ret <4 x i32> %2
 | |
| ; SSE2-LABEL: v4i32_icmp_ule:
 | |
| ; SSE2: movdqa  {{.*}}(%rip), %xmm2
 | |
| ; SSE2: pxor    %xmm2, %xmm1
 | |
| ; SSE2: pxor    %xmm2, %xmm0
 | |
| ; SSE2: pcmpgtd %xmm1, %xmm0
 | |
| ; SSE2: pcmpeqd %xmm1, %xmm1
 | |
| ; SSE2: pxor    %xmm0, %xmm1
 | |
| ; SSE2: movdqa  %xmm1, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: v4i32_icmp_ule:
 | |
| ; SSE41: pminud  %xmm0, %xmm1
 | |
| ; SSE41: pcmpeqd %xmm1, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: v4i32_icmp_ule:
 | |
| ; AVX: pminud  %xmm1, %xmm0, %xmm1
 | |
| ; AVX: pcmpeqd %xmm1, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| ; At one point we were incorrectly constant-folding a setcc to 0x1 instead of
 | |
| ; 0xff, leading to a constpool load. The instruction doesn't matter here, but it
 | |
| ; should set all bits to 1.
 | |
| define <16 x i8> @test_setcc_constfold_vi8(<16 x i8> %l, <16 x i8> %r) {
 | |
|   %test1 = icmp eq <16 x i8> %l, %r
 | |
|   %mask1 = sext <16 x i1> %test1 to <16 x i8>
 | |
| 
 | |
|   %test2 = icmp ne <16 x i8> %l, %r
 | |
|   %mask2 = sext <16 x i1> %test2 to <16 x i8>
 | |
| 
 | |
|   %res = or <16 x i8> %mask1, %mask2
 | |
|   ret <16 x i8> %res
 | |
| ; SSE2-LABEL: test_setcc_constfold_vi8:
 | |
| ; SSE2: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: test_setcc_constfold_vi8:
 | |
| ; SSE41: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: test_setcc_constfold_vi8:
 | |
| ; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| ; Make sure sensible results come from doing extension afterwards
 | |
| define <16 x i8> @test_setcc_constfold_vi1(<16 x i8> %l, <16 x i8> %r) {
 | |
|   %test1 = icmp eq <16 x i8> %l, %r
 | |
|   %test2 = icmp ne <16 x i8> %l, %r
 | |
| 
 | |
|   %res = or <16 x i1> %test1, %test2
 | |
|   %mask = sext <16 x i1> %res to <16 x i8>
 | |
|   ret <16 x i8> %mask
 | |
| ; SSE2-LABEL: test_setcc_constfold_vi1:
 | |
| ; SSE2: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: test_setcc_constfold_vi1:
 | |
| ; SSE41: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: test_setcc_constfold_vi1:
 | |
| ; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
 | |
| }
 | |
| 
 | |
| 
 | |
| ; 64-bit case is also particularly important, as the constant "-1" is probably
 | |
| ; just 32-bits wide.
 | |
| define <2 x i64> @test_setcc_constfold_vi64(<2 x i64> %l, <2 x i64> %r) {
 | |
|   %test1 = icmp eq <2 x i64> %l, %r
 | |
|   %mask1 = sext <2 x i1> %test1 to <2 x i64>
 | |
| 
 | |
|   %test2 = icmp ne <2 x i64> %l, %r
 | |
|   %mask2 = sext <2 x i1> %test2 to <2 x i64>
 | |
| 
 | |
|   %res = or <2 x i64> %mask1, %mask2
 | |
|   ret <2 x i64> %res
 | |
| ; SSE2-LABEL: test_setcc_constfold_vi64:
 | |
| ; SSE2: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; SSE41-LABEL: test_setcc_constfold_vi64:
 | |
| ; SSE41: pcmpeqd %xmm0, %xmm0
 | |
| 
 | |
| ; AVX-LABEL: test_setcc_constfold_vi64:
 | |
| ; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
 | |
| }
 |