mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-06 05:06:45 +00:00
685b0d9315
The current approach to lower a vsetult is to flip the sign bit of the operands, swap the operands and then use a (signed) pcmpgt. psubus (unsigned saturating subtract) can be used to emulate a vsetult more efficiently: + case ISD::SETULT: { + // If the comparison is against a constant we can turn this into a + // setule. With psubus, setule does not require a swap. This is + // beneficial because the constant in the register is no longer + // destructed as the destination so it can be hoisted out of a loop. I also enable lowering via psubus in a few other cases where it's clearly beneficial: setule and setuge if minu/maxu cannot be used. rdar://problem/14338765 Patch by Adam Nemet <anemet@apple.com>. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@202301 91177308-0d34-0410-b5e6-96231b3b80d8
182 lines
5.2 KiB
LLVM
182 lines
5.2 KiB
LLVM
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=sse2 | FileCheck %s -check-prefix=SSE2
|
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=sse4.1 | FileCheck %s -check-prefix=SSE41
|
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=x86-64 -mattr=avx | FileCheck %s -check-prefix=AVX
|
|
|
|
define <16 x i8> @v16i8_icmp_uge(<16 x i8> %a, <16 x i8> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp uge <16 x i8> %a, %b
|
|
%2 = sext <16 x i1> %1 to <16 x i8>
|
|
ret <16 x i8> %2
|
|
; SSE2-LABEL: v16i8_icmp_uge:
|
|
; SSE2: pmaxub %xmm0, %xmm1
|
|
; SSE2: pcmpeqb %xmm1, %xmm0
|
|
|
|
; SSE41-LABEL: v16i8_icmp_uge:
|
|
; SSE41: pmaxub %xmm0, %xmm1
|
|
; SSE41: pcmpeqb %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v16i8_icmp_uge:
|
|
; AVX: vpmaxub %xmm1, %xmm0, %xmm1
|
|
; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
define <16 x i8> @v16i8_icmp_ule(<16 x i8> %a, <16 x i8> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp ule <16 x i8> %a, %b
|
|
%2 = sext <16 x i1> %1 to <16 x i8>
|
|
ret <16 x i8> %2
|
|
; SSE2-LABEL: v16i8_icmp_ule:
|
|
; SSE2: pminub %xmm0, %xmm1
|
|
; SSE2: pcmpeqb %xmm1, %xmm0
|
|
|
|
; SSE41-LABEL: v16i8_icmp_ule:
|
|
; SSE41: pminub %xmm0, %xmm1
|
|
; SSE41: pcmpeqb %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v16i8_icmp_ule:
|
|
; AVX: vpminub %xmm1, %xmm0, %xmm1
|
|
; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
|
|
define <8 x i16> @v8i16_icmp_uge(<8 x i16> %a, <8 x i16> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp uge <8 x i16> %a, %b
|
|
%2 = sext <8 x i1> %1 to <8 x i16>
|
|
ret <8 x i16> %2
|
|
; SSE2-LABEL: v8i16_icmp_uge:
|
|
; SSE2: psubusw %xmm0, %xmm1
|
|
; SEE2: pxor %xmm0, %xmm0
|
|
; SSE2: pcmpeqw %xmm1, %xmm0
|
|
|
|
; SSE41-LABEL: v8i16_icmp_uge:
|
|
; SSE41: pmaxuw %xmm0, %xmm1
|
|
; SSE41: pcmpeqw %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v8i16_icmp_uge:
|
|
; AVX: vpmaxuw %xmm1, %xmm0, %xmm1
|
|
; AVX: vpcmpeqw %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
define <8 x i16> @v8i16_icmp_ule(<8 x i16> %a, <8 x i16> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp ule <8 x i16> %a, %b
|
|
%2 = sext <8 x i1> %1 to <8 x i16>
|
|
ret <8 x i16> %2
|
|
; SSE2-LABEL: v8i16_icmp_ule:
|
|
; SSE2: psubusw %xmm1, %xmm0
|
|
; SSE2: pxor %xmm1, %xmm1
|
|
; SSE2: pcmpeqw %xmm0, %xmm1
|
|
; SSE2: movdqa %xmm1, %xmm0
|
|
|
|
; SSE41-LABEL: v8i16_icmp_ule:
|
|
; SSE41: pminuw %xmm0, %xmm1
|
|
; SSE41: pcmpeqw %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v8i16_icmp_ule:
|
|
; AVX: vpminuw %xmm1, %xmm0, %xmm1
|
|
; AVX: vpcmpeqw %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
|
|
define <4 x i32> @v4i32_icmp_uge(<4 x i32> %a, <4 x i32> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp uge <4 x i32> %a, %b
|
|
%2 = sext <4 x i1> %1 to <4 x i32>
|
|
ret <4 x i32> %2
|
|
; SSE2-LABEL: v4i32_icmp_uge:
|
|
; SSE2: movdqa {{.*}}(%rip), %xmm2
|
|
; SSE2: pxor %xmm2, %xmm0
|
|
; SSE2: pxor %xmm1, %xmm2
|
|
; SSE2: pcmpgtd %xmm0, %xmm2
|
|
; SSE2: pcmpeqd %xmm0, %xmm0
|
|
; SSE2: pxor %xmm2, %xmm0
|
|
|
|
; SSE41-LABEL: v4i32_icmp_uge:
|
|
; SSE41: pmaxud %xmm0, %xmm1
|
|
; SSE41: pcmpeqd %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v4i32_icmp_uge:
|
|
; AVX: vpmaxud %xmm1, %xmm0, %xmm1
|
|
; AVX: vpcmpeqd %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
define <4 x i32> @v4i32_icmp_ule(<4 x i32> %a, <4 x i32> %b) nounwind readnone ssp uwtable {
|
|
%1 = icmp ule <4 x i32> %a, %b
|
|
%2 = sext <4 x i1> %1 to <4 x i32>
|
|
ret <4 x i32> %2
|
|
; SSE2-LABEL: v4i32_icmp_ule:
|
|
; SSE2: movdqa {{.*}}(%rip), %xmm2
|
|
; SSE2: pxor %xmm2, %xmm1
|
|
; SSE2: pxor %xmm2, %xmm0
|
|
; SSE2: pcmpgtd %xmm1, %xmm0
|
|
; SSE2: pcmpeqd %xmm1, %xmm1
|
|
; SSE2: pxor %xmm0, %xmm1
|
|
; SSE2: movdqa %xmm1, %xmm0
|
|
|
|
; SSE41-LABEL: v4i32_icmp_ule:
|
|
; SSE41: pminud %xmm0, %xmm1
|
|
; SSE41: pcmpeqd %xmm1, %xmm0
|
|
|
|
; AVX-LABEL: v4i32_icmp_ule:
|
|
; AVX: pminud %xmm1, %xmm0, %xmm1
|
|
; AVX: pcmpeqd %xmm1, %xmm0, %xmm0
|
|
}
|
|
|
|
; At one point we were incorrectly constant-folding a setcc to 0x1 instead of
|
|
; 0xff, leading to a constpool load. The instruction doesn't matter here, but it
|
|
; should set all bits to 1.
|
|
define <16 x i8> @test_setcc_constfold_vi8(<16 x i8> %l, <16 x i8> %r) {
|
|
%test1 = icmp eq <16 x i8> %l, %r
|
|
%mask1 = sext <16 x i1> %test1 to <16 x i8>
|
|
|
|
%test2 = icmp ne <16 x i8> %l, %r
|
|
%mask2 = sext <16 x i1> %test2 to <16 x i8>
|
|
|
|
%res = or <16 x i8> %mask1, %mask2
|
|
ret <16 x i8> %res
|
|
; SSE2-LABEL: test_setcc_constfold_vi8:
|
|
; SSE2: pcmpeqd %xmm0, %xmm0
|
|
|
|
; SSE41-LABEL: test_setcc_constfold_vi8:
|
|
; SSE41: pcmpeqd %xmm0, %xmm0
|
|
|
|
; AVX-LABEL: test_setcc_constfold_vi8:
|
|
; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
|
|
}
|
|
|
|
; Make sure sensible results come from doing extension afterwards
|
|
define <16 x i8> @test_setcc_constfold_vi1(<16 x i8> %l, <16 x i8> %r) {
|
|
%test1 = icmp eq <16 x i8> %l, %r
|
|
%test2 = icmp ne <16 x i8> %l, %r
|
|
|
|
%res = or <16 x i1> %test1, %test2
|
|
%mask = sext <16 x i1> %res to <16 x i8>
|
|
ret <16 x i8> %mask
|
|
; SSE2-LABEL: test_setcc_constfold_vi1:
|
|
; SSE2: pcmpeqd %xmm0, %xmm0
|
|
|
|
; SSE41-LABEL: test_setcc_constfold_vi1:
|
|
; SSE41: pcmpeqd %xmm0, %xmm0
|
|
|
|
; AVX-LABEL: test_setcc_constfold_vi1:
|
|
; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
|
|
}
|
|
|
|
|
|
; 64-bit case is also particularly important, as the constant "-1" is probably
|
|
; just 32-bits wide.
|
|
define <2 x i64> @test_setcc_constfold_vi64(<2 x i64> %l, <2 x i64> %r) {
|
|
%test1 = icmp eq <2 x i64> %l, %r
|
|
%mask1 = sext <2 x i1> %test1 to <2 x i64>
|
|
|
|
%test2 = icmp ne <2 x i64> %l, %r
|
|
%mask2 = sext <2 x i1> %test2 to <2 x i64>
|
|
|
|
%res = or <2 x i64> %mask1, %mask2
|
|
ret <2 x i64> %res
|
|
; SSE2-LABEL: test_setcc_constfold_vi64:
|
|
; SSE2: pcmpeqd %xmm0, %xmm0
|
|
|
|
; SSE41-LABEL: test_setcc_constfold_vi64:
|
|
; SSE41: pcmpeqd %xmm0, %xmm0
|
|
|
|
; AVX-LABEL: test_setcc_constfold_vi64:
|
|
; AVX: vpcmpeqd %xmm0, %xmm0, %xmm0
|
|
}
|