mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-19 17:33:29 +00:00
8c05a850f4
term goal here is to be able to match enough of vector_shuffle and build_vector so all avx intrinsics which aren't mapped to their own built-ins but to shufflevector calls can be codegen'd. This is the first (baby) step, support building zeroed vectors. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@110897 91177308-0d34-0410-b5e6-96231b3b80d8
16 lines
452 B
LLVM
16 lines
452 B
LLVM
; RUN: llc < %s -mtriple=x86_64-apple-darwin -march=x86 -mcpu=corei7 -mattr=avx | FileCheck %s
|
|
|
|
@x = common global <8 x float> zeroinitializer, align 32
|
|
@y = common global <4 x double> zeroinitializer, align 32
|
|
|
|
define void @zero() nounwind ssp {
|
|
entry:
|
|
; CHECK: vxorps
|
|
; CHECK: vmovaps
|
|
; CHECK: vmovaps
|
|
store <8 x float> zeroinitializer, <8 x float>* @x, align 32
|
|
store <4 x double> zeroinitializer, <4 x double>* @y, align 32
|
|
ret void
|
|
}
|
|
|