mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-28 04:33:05 +00:00
4969310052
This required disabling a PowerPC optimization that did the following: input: x = BUILD_VECTOR <i32 16, i32 16, i32 16, i32 16> lowered to: tmp = BUILD_VECTOR <i32 8, i32 8, i32 8, i32 8> x = ADD tmp, tmp The add now gets folded immediately and we're back at the BUILD_VECTOR we started from. I don't see a way to fix this currently so I left it disabled for now. Fix some trivially foldable X86 tests too. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@174325 91177308-0d34-0410-b5e6-96231b3b80d8
26 lines
1.1 KiB
LLVM
26 lines
1.1 KiB
LLVM
; RUN: llc < %s -march=x86-64 -mcpu=corei7 | FileCheck %s
|
|
|
|
target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128-n8:16:32"
|
|
target triple = "x86_64-apple-macosx10.6.6"
|
|
|
|
; Test that the order of operands is correct
|
|
; CHECK: select_func
|
|
; CHECK: pblendvb %xmm1, %xmm2
|
|
; CHECK: ret
|
|
|
|
define void @select_func(<8 x i16> %in) {
|
|
entry:
|
|
%c.lobit.i.i.i = ashr <8 x i16> %in, <i16 15, i16 15, i16 15, i16 15, i16 15, i16 15, i16 15, i16 15>
|
|
%and.i56.i.i.i = and <8 x i16> %c.lobit.i.i.i, <i16 25, i16 8, i16 65, i16 25, i16 8, i16 95, i16 15, i16 45>
|
|
%and.i5.i.i.i = bitcast <8 x i16> %and.i56.i.i.i to <2 x i64>
|
|
%neg.i.i.i.i = xor <8 x i16> %c.lobit.i.i.i, <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1>
|
|
%and.i.i.i = and <8 x i16> %neg.i.i.i.i, <i16 45, i16 15, i16 95, i16 8, i16 25, i16 65, i16 8, i16 25>
|
|
%and.i2.i.i.i = bitcast <8 x i16> %and.i.i.i to <2 x i64>
|
|
%or.i.i.i.i = or <2 x i64> %and.i2.i.i.i, %and.i5.i.i.i
|
|
%a37 = bitcast <2 x i64> %or.i.i.i.i to <8 x i16>
|
|
store <8 x i16> %a37, <8 x i16> addrspace(1)* undef, align 4
|
|
ret void
|
|
}
|
|
|
|
|