llvm-6502/test/CodeGen/X86/lower-bitcast.ll
Bruno Cardoso Lopes 807360ab08 [x86] Combine x86mmx/i64 to v2i64 conversion to use scalar_to_vector
Handle the poor codegen for i64/x86xmm->v2i64 (%mm -> %xmm) moves. Instead of
using stack store/load pair to do the job, use scalar_to_vector directly, which
in the MMX case can use movq2dq. This was the current behavior prior to
improvements for vector legalization of extloads in r213897.

This commit fixes the regression and as a side-effect also remove some
unnecessary shuffles.

In the new attached testcase, we go from:

pshufw  $-18, (%rdi), %mm0
movq    %mm0, -8(%rsp)
movq    -8(%rsp), %xmm0
pshufd  $-44, %xmm0, %xmm0
movd    %xmm0, %eax
...

To:

pshufw  $-18, (%rdi), %mm0
movq2dq %mm0, %xmm0
movd    %xmm0, %eax
...

Differential Revision: http://reviews.llvm.org/D7126
rdar://problem/19413324

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@226953 91177308-0d34-0410-b5e6-96231b3b80d8
2015-01-23 22:44:16 +00:00

190 lines
4.5 KiB
LLVM

; RUN: llc < %s -march=x86-64 -mcpu=core2 -mattr=+sse2 | FileCheck %s
; RUN: llc < %s -march=x86-64 -mcpu=core2 -mattr=+sse2 -x86-experimental-vector-widening-legalization | FileCheck %s --check-prefix=CHECK-WIDE
define double @test1(double %A) {
%1 = bitcast double %A to <2 x i32>
%add = add <2 x i32> %1, <i32 3, i32 5>
%2 = bitcast <2 x i32> %add to double
ret double %2
}
; FIXME: Ideally we should be able to fold the entire body of @test1 into a
; single paddd instruction. At the moment we produce the sequence
; pshufd+paddq+pshufd. This is fixed with the widening legalization.
;
; CHECK-LABEL: test1
; CHECK-NOT: movsd
; CHECK: pshufd
; CHECK-NEXT: paddd
; CHECK-NEXT: pshufd
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test1
; CHECK-WIDE-NOT: movsd
; CHECK-WIDE: paddd
; CHECK-WIDE-NEXT: ret
define double @test2(double %A, double %B) {
%1 = bitcast double %A to <2 x i32>
%2 = bitcast double %B to <2 x i32>
%add = add <2 x i32> %1, %2
%3 = bitcast <2 x i32> %add to double
ret double %3
}
; CHECK-LABEL: test2
; CHECK-NOT: movsd
; CHECK: paddd
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test2
; CHECK-WIDE-NOT: movsd
; CHECK-WIDE: paddd
; CHECK-WIDE-NEXT: ret
define i64 @test3(i64 %A) {
%1 = bitcast i64 %A to <2 x float>
%add = fadd <2 x float> %1, <float 3.0, float 5.0>
%2 = bitcast <2 x float> %add to i64
ret i64 %2
}
; CHECK-LABEL: test3
; CHECK-NOT: pshufd
; CHECK: addps
; CHECK-NOT: pshufd
; CHECK: ret
;
; CHECK-WIDE-LABEL: test3
; CHECK-WIDE-NOT: pshufd
; CHECK-WIDE: addps
; CHECK-WIDE-NOT: pshufd
; CHECK-WIDE: ret
define i64 @test4(i64 %A) {
%1 = bitcast i64 %A to <2 x i32>
%add = add <2 x i32> %1, <i32 3, i32 5>
%2 = bitcast <2 x i32> %add to i64
ret i64 %2
}
; FIXME: At the moment we still produce the sequence paddd+pshufd.
; Ideally, we should fold that sequence into a single paddd. This is fixed with
; the widening legalization.
;
; CHECK-LABEL: test4
; CHECK: movd
; CHECK-NOT: pshufd
; CHECK-NEXT: paddd
; CHECK-NEXT: pshufd
; CHECK: ret
;
; CHECK-WIDE-LABEL: test4
; CHECK-WIDE: movd %{{rdi|rcx}},
; CHECK-WIDE-NEXT: paddd
; CHECK-WIDE-NEXT: movd {{.*}}, %rax
; CHECK-WIDE: ret
define double @test5(double %A) {
%1 = bitcast double %A to <2 x float>
%add = fadd <2 x float> %1, <float 3.0, float 5.0>
%2 = bitcast <2 x float> %add to double
ret double %2
}
; CHECK-LABEL: test5
; CHECK: addps
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test5
; CHECK-WIDE: addps
; CHECK-WIDE-NEXT: ret
define double @test6(double %A) {
%1 = bitcast double %A to <4 x i16>
%add = add <4 x i16> %1, <i16 3, i16 4, i16 5, i16 6>
%2 = bitcast <4 x i16> %add to double
ret double %2
}
; FIXME: Ideally we should be able to fold the entire body of @test6 into a
; single paddw instruction. This is fixed with the widening legalization.
;
; CHECK-LABEL: test6
; CHECK-NOT: movsd
; CHECK: punpcklwd
; CHECK-NEXT: paddw
; CHECK-NEXT: pshufb
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test6
; CHECK-WIDE-NOT: mov
; CHECK-WIDE-NOT: punpcklwd
; CHECK-WIDE: paddw
; CHECK-WIDE-NEXT: ret
define double @test7(double %A, double %B) {
%1 = bitcast double %A to <4 x i16>
%2 = bitcast double %B to <4 x i16>
%add = add <4 x i16> %1, %2
%3 = bitcast <4 x i16> %add to double
ret double %3
}
; CHECK-LABEL: test7
; CHECK-NOT: movsd
; CHECK-NOT: punpcklwd
; CHECK: paddw
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test7
; CHECK-WIDE-NOT: movsd
; CHECK-WIDE-NOT: punpcklwd
; CHECK-WIDE: paddw
; CHECK-WIDE-NEXT: ret
define double @test8(double %A) {
%1 = bitcast double %A to <8 x i8>
%add = add <8 x i8> %1, <i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10>
%2 = bitcast <8 x i8> %add to double
ret double %2
}
; FIXME: Ideally we should be able to fold the entire body of @test8 into a
; single paddb instruction. At the moment we produce the sequence
; pshufd+paddw+pshufd. This is fixed with the widening legalization.
;
; CHECK-LABEL: test8
; CHECK-NOT: movsd
; CHECK: punpcklbw
; CHECK-NEXT: paddb
; CHECK-NEXT: pshufb
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test8
; CHECK-WIDE-NOT: movsd
; CHECK-WIDE-NOT: punpcklbw
; CHECK-WIDE: paddb
; CHECK-WIDE-NEXT: ret
define double @test9(double %A, double %B) {
%1 = bitcast double %A to <8 x i8>
%2 = bitcast double %B to <8 x i8>
%add = add <8 x i8> %1, %2
%3 = bitcast <8 x i8> %add to double
ret double %3
}
; CHECK-LABEL: test9
; CHECK-NOT: movsd
; CHECK-NOT: punpcklbw
; CHECK: paddb
; CHECK-NEXT: ret
;
; CHECK-WIDE-LABEL: test9
; CHECK-WIDE-NOT: movsd
; CHECK-WIDE-NOT: punpcklbw
; CHECK-WIDE: paddb
; CHECK-WIDE-NEXT: ret