; RUN: llc < %s -march=x86-64 -mcpu=core-avx2 | FileCheck %s ; Verify that the backend correctly combines AVX2 builtin intrinsics. define <8 x i32> @test_psra_1(<8 x i32> %A) { %1 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %A, i32 3) %2 = tail call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %1, <4 x i32> ) %3 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %2, i32 2) ret <8 x i32> %3 } ; CHECK-LABEL: test_psra_1 ; CHECK: vpsrad $8, %ymm0, %ymm0 ; CHECK-NEXT: ret define <16 x i16> @test_psra_2(<16 x i16> %A) { %1 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %A, i32 3) %2 = tail call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %1, <8 x i16> ) %3 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %2, i32 2) ret <16 x i16> %3 } ; CHECK-LABEL: test_psra_2 ; CHECK: vpsraw $8, %ymm0, %ymm0 ; CHECK-NEXT: ret define <16 x i16> @test_psra_3(<16 x i16> %A) { %1 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %A, i32 0) %2 = tail call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %1, <8 x i16> ) %3 = tail call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %2, i32 0) ret <16 x i16> %3 } ; CHECK-LABEL: test_psra_3 ; CHECK-NOT: vpsraw ; CHECK: ret define <8 x i32> @test_psra_4(<8 x i32> %A) { %1 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %A, i32 0) %2 = tail call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %1, <4 x i32> ) %3 = tail call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %2, i32 0) ret <8 x i32> %3 } ; CHECK-LABEL: test_psra_4 ; CHECK-NOT: vpsrad ; CHECK: ret declare <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16>, <8 x i16>) declare <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16>, i32) declare <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32>, <4 x i32>) declare <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32>, i32)