mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 04:30:12 +00:00
7c9c6ed761
Essentially the same as the GEP change in r230786. A similar migration script can be used to update test cases, though a few more test case improvements/changes were required this time around: (r229269-r229278) import fileinput import sys import re pat = re.compile(r"((?:=|:|^)\s*load (?:atomic )?(?:volatile )?(.*?))(| addrspace\(\d+\) *)\*($| *(?:%|@|null|undef|blockaddress|getelementptr|addrspacecast|bitcast|inttoptr|\[\[[a-zA-Z]|\{\{).*$)") for line in sys.stdin: sys.stdout.write(re.sub(pat, r"\1, \2\3*\4", line)) Reviewers: rafael, dexonsmith, grosser Differential Revision: http://reviews.llvm.org/D7649 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@230794 91177308-0d34-0410-b5e6-96231b3b80d8
64 lines
2.0 KiB
LLVM
64 lines
2.0 KiB
LLVM
; RUN: llc < %s -relocation-model=static -mcpu=yonah | FileCheck %s
|
|
|
|
; The double argument is at 4(esp) which is 16-byte aligned, but we
|
|
; are required to read in extra bytes of memory in order to fold the
|
|
; load. Bad Things may happen when reading/processing undefined bytes,
|
|
; so don't fold the load.
|
|
; PR22371 / http://reviews.llvm.org/D7474
|
|
|
|
target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128"
|
|
target triple = "i686-apple-darwin8"
|
|
@G = external global double
|
|
|
|
define void @test({ double, double }* byval %z, double* %P) nounwind {
|
|
entry:
|
|
%tmp3 = load double, double* @G, align 16 ; <double> [#uses=1]
|
|
%tmp4 = tail call double @fabs( double %tmp3 ) readnone ; <double> [#uses=1]
|
|
store volatile double %tmp4, double* %P
|
|
%tmp = getelementptr { double, double }, { double, double }* %z, i32 0, i32 0 ; <double*> [#uses=1]
|
|
%tmp1 = load volatile double, double* %tmp, align 8 ; <double> [#uses=1]
|
|
%tmp2 = tail call double @fabs( double %tmp1 ) readnone ; <double> [#uses=1]
|
|
%tmp6 = fadd double %tmp4, %tmp2 ; <double> [#uses=1]
|
|
store volatile double %tmp6, double* %P, align 8
|
|
ret void
|
|
|
|
; CHECK-LABEL: test:
|
|
; CHECK: movsd {{.*}}G, %xmm{{.*}}
|
|
; CHECK: andpd %xmm{{.*}}, %xmm{{.*}}
|
|
; CHECK: movsd 4(%esp), %xmm{{.*}}
|
|
; CHECK: andpd %xmm{{.*}}, %xmm{{.*}}
|
|
|
|
|
|
}
|
|
|
|
define void @test2() alignstack(16) nounwind {
|
|
entry:
|
|
; CHECK-LABEL: test2:
|
|
; CHECK: andl{{.*}}$-16, %esp
|
|
ret void
|
|
}
|
|
|
|
; Use a call to force a spill.
|
|
define <2 x double> @test3(<2 x double> %x, <2 x double> %y) alignstack(32) nounwind {
|
|
entry:
|
|
; CHECK-LABEL: test3:
|
|
; CHECK: andl{{.*}}$-32, %esp
|
|
call void @test2()
|
|
%A = fmul <2 x double> %x, %y
|
|
ret <2 x double> %A
|
|
}
|
|
|
|
declare double @fabs(double)
|
|
|
|
; The pointer is already known aligned, so and x,-16 is eliminable.
|
|
define i32 @test4() nounwind {
|
|
entry:
|
|
%buffer = alloca [2048 x i8], align 16
|
|
%0 = ptrtoint [2048 x i8]* %buffer to i32
|
|
%and = and i32 %0, -16
|
|
ret i32 %and
|
|
; CHECK-LABEL: test4:
|
|
; CHECK-NOT: and
|
|
; CHECK: ret
|
|
}
|