mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-27 13:30:05 +00:00
9836c47ea6
Added target specific combine rules to fold blend intrinsics according to the following rules: 1) fold(blend A, A, Mask) -> A; 2) fold(blend A, B, <allZeros>) -> A; 3) fold(blend A, B, <allOnes>) -> B. Added two new tests to verify that the new folding rules work for all the optimized blend intrinsics. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@208895 91177308-0d34-0410-b5e6-96231b3b80d8
165 lines
5.4 KiB
LLVM
165 lines
5.4 KiB
LLVM
; RUN: llc < %s -march=x86-64 -mcpu=core-avx2 | FileCheck %s
|
|
|
|
; Verify that the backend correctly combines AVX2 builtin intrinsics.
|
|
|
|
|
|
define <8 x i32> @test_psra_1(<8 x i32> %A) {
|
|
%1 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %A, i32 3)
|
|
%2 = tail call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %1, <4 x i32> <i32 3, i32 0, i32 7, i32 0>)
|
|
%3 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %2, i32 2)
|
|
ret <8 x i32> %3
|
|
}
|
|
; CHECK-LABEL: test_psra_1
|
|
; CHECK: vpsrad $8, %ymm0, %ymm0
|
|
; CHECK-NEXT: ret
|
|
|
|
define <16 x i16> @test_psra_2(<16 x i16> %A) {
|
|
%1 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %A, i32 3)
|
|
%2 = tail call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %1, <8 x i16> <i16 3, i16 0, i16 0, i16 0, i16 7, i16 0, i16 0, i16 0>)
|
|
%3 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %2, i32 2)
|
|
ret <16 x i16> %3
|
|
}
|
|
; CHECK-LABEL: test_psra_2
|
|
; CHECK: vpsraw $8, %ymm0, %ymm0
|
|
; CHECK-NEXT: ret
|
|
|
|
define <16 x i16> @test_psra_3(<16 x i16> %A) {
|
|
%1 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %A, i32 0)
|
|
%2 = tail call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %1, <8 x i16> <i16 0, i16 0, i16 0, i16 0, i16 7, i16 0, i16 0, i16 0>)
|
|
%3 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %2, i32 0)
|
|
ret <16 x i16> %3
|
|
}
|
|
; CHECK-LABEL: test_psra_3
|
|
; CHECK-NOT: vpsraw
|
|
; CHECK: ret
|
|
|
|
define <8 x i32> @test_psra_4(<8 x i32> %A) {
|
|
%1 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %A, i32 0)
|
|
%2 = tail call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %1, <4 x i32> <i32 0, i32 0, i32 7, i32 0>)
|
|
%3 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %2, i32 0)
|
|
ret <8 x i32> %3
|
|
}
|
|
; CHECK-LABEL: test_psra_4
|
|
; CHECK-NOT: vpsrad
|
|
; CHECK: ret
|
|
|
|
|
|
define <32 x i8> @test_x86_avx2_pblendvb(<32 x i8> %a0, <32 x i8> %a1) {
|
|
%res = call <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8> %a0, <32 x i8> %a0, <32 x i8> %a1)
|
|
ret <32 x i8> %res
|
|
}
|
|
; CHECK-LABEL: test_x86_avx2_pblendvb
|
|
; CHECK-NOT: vpblendvb
|
|
; CHECK: ret
|
|
|
|
|
|
define <16 x i16> @test_x86_avx2_pblendw(<16 x i16> %a0) {
|
|
%res = call <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16> %a0, <16 x i16> %a0, i32 7)
|
|
ret <16 x i16> %res
|
|
}
|
|
; CHECK-LABEL: test_x86_avx2_pblendw
|
|
; CHECK-NOT: vpblendw
|
|
; CHECK: ret
|
|
|
|
|
|
define <4 x i32> @test_x86_avx2_pblendd_128(<4 x i32> %a0) {
|
|
%res = call <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32> %a0, <4 x i32> %a0, i32 7)
|
|
ret <4 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test_x86_avx2_pblendd_128
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
define <8 x i32> @test_x86_avx2_pblendd_256(<8 x i32> %a0) {
|
|
%res = call <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32> %a0, <8 x i32> %a0, i32 7)
|
|
ret <8 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test_x86_avx2_pblendd_256
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
define <32 x i8> @test2_x86_avx2_pblendvb(<32 x i8> %a0, <32 x i8> %a1) {
|
|
%res = call <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8> %a0, <32 x i8> %a1, <32 x i8> zeroinitializer)
|
|
ret <32 x i8> %res
|
|
}
|
|
; CHECK-LABEL: test2_x86_avx2_pblendvb
|
|
; CHECK-NOT: vpblendvb
|
|
; CHECK: ret
|
|
|
|
|
|
define <16 x i16> @test2_x86_avx2_pblendw(<16 x i16> %a0, <16 x i16> %a1) {
|
|
%res = call <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16> %a0, <16 x i16> %a1, i32 0)
|
|
ret <16 x i16> %res
|
|
}
|
|
; CHECK-LABEL: test2_x86_avx2_pblendw
|
|
; CHECK-NOT: vpblendw
|
|
; CHECK: ret
|
|
|
|
|
|
define <4 x i32> @test2_x86_avx2_pblendd_128(<4 x i32> %a0, <4 x i32> %a1) {
|
|
%res = call <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32> %a0, <4 x i32> %a1, i32 0)
|
|
ret <4 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test2_x86_avx2_pblendd_128
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
define <8 x i32> @test2_x86_avx2_pblendd_256(<8 x i32> %a0, <8 x i32> %a1) {
|
|
%res = call <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32> %a0, <8 x i32> %a1, i32 0)
|
|
ret <8 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test2_x86_avx2_pblendd_256
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
define <32 x i8> @test3_x86_avx2_pblendvb(<32 x i8> %a0, <32 x i8> %a1) {
|
|
%1 = bitcast <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1> to <32 x i8>
|
|
%res = call <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8> %a0, <32 x i8> %a1, <32 x i8> %1)
|
|
ret <32 x i8> %res
|
|
}
|
|
; CHECK-LABEL: test3_x86_avx2_pblendvb
|
|
; CHECK-NOT: vpblendvb
|
|
; CHECK: ret
|
|
|
|
|
|
define <16 x i16> @test3_x86_avx2_pblendw(<16 x i16> %a0, <16 x i16> %a1) {
|
|
%res = call <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16> %a0, <16 x i16> %a1, i32 -1)
|
|
ret <16 x i16> %res
|
|
}
|
|
; CHECK-LABEL: test3_x86_avx2_pblendw
|
|
; CHECK-NOT: vpblendw
|
|
; CHECK: ret
|
|
|
|
|
|
define <4 x i32> @test3_x86_avx2_pblendd_128(<4 x i32> %a0, <4 x i32> %a1) {
|
|
%res = call <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32> %a0, <4 x i32> %a1, i32 -1)
|
|
ret <4 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test3_x86_avx2_pblendd_128
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
define <8 x i32> @test3_x86_avx2_pblendd_256(<8 x i32> %a0, <8 x i32> %a1) {
|
|
%res = call <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32> %a0, <8 x i32> %a1, i32 -1)
|
|
ret <8 x i32> %res
|
|
}
|
|
; CHECK-LABEL: test3_x86_avx2_pblendd_256
|
|
; CHECK-NOT: vpblendd
|
|
; CHECK: ret
|
|
|
|
|
|
declare <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8>, <32 x i8>, <32 x i8>)
|
|
declare <16 x i16> @llvm.x86.avx2.pblendw(<16 x i16>, <16 x i16>, i32)
|
|
declare <4 x i32> @llvm.x86.avx2.pblendd.128(<4 x i32>, <4 x i32>, i32)
|
|
declare <8 x i32> @llvm.x86.avx2.pblendd.256(<8 x i32>, <8 x i32>, i32)
|
|
declare <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16>, <8 x i16>)
|
|
declare <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16>, i32)
|
|
declare <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32>, <4 x i32>)
|
|
declare <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32>, i32)
|
|
|