mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-16 11:30:51 +00:00
1d142d4d47
Patch by Bill Seurer; some comment formatting changes by me. There are a few PowerPC test cases for FastISel support that currently fail with VSX support enabled. The temporary workaround under discussion in http://reviews.llvm.org/D5362 helps, but the tests still fail because they specify -fast-isel-abort, and the VSX workaround punts back to SelectionDAG. We have plans to fix FastISel permanently for VSX, but until that's in place these tests are preventing us from enabling VSX by default. Therefore we are adding -mattr=-vsx to these tests until the full support is ready. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@220172 91177308-0d34-0410-b5e6-96231b3b80d8
207 lines
4.3 KiB
LLVM
207 lines
4.3 KiB
LLVM
; FIXME: FastISel currently returns false if it hits code that uses VSX
|
|
; registers and with -fast-isel-abort turned on the test case will then fail.
|
|
; When fastisel better supports VSX fix up this test case.
|
|
;
|
|
; RUN: llc < %s -O0 -verify-machineinstrs -fast-isel -fast-isel-abort -mattr=-vsx -mtriple=powerpc64-unknown-linux-gnu -mcpu=pwr7 | FileCheck %s --check-prefix=ELF64
|
|
|
|
; This test verifies that load/store instructions are properly generated,
|
|
; and that they pass MI verification.
|
|
|
|
@a = global i8 1, align 1
|
|
@b = global i16 2, align 2
|
|
@c = global i32 4, align 4
|
|
@d = global i64 8, align 8
|
|
@e = global float 1.25, align 4
|
|
@f = global double 3.5, align 8
|
|
|
|
%struct.s = type<{ i8, i32 }>
|
|
%struct.t = type<{ i8, i64 }>
|
|
|
|
@g = global %struct.s <{ i8 1, i32 2 }>, align 1
|
|
@h = global %struct.t <{ i8 1, i64 2 }>, align 1
|
|
|
|
@i = common global [8192 x i64] zeroinitializer, align 8
|
|
|
|
; load
|
|
|
|
define i8 @t1() nounwind uwtable ssp {
|
|
; ELF64: t1
|
|
%1 = load i8* @a, align 1
|
|
; ELF64: lbz
|
|
%2 = add nsw i8 %1, 1
|
|
; ELF64: addi
|
|
ret i8 %2
|
|
}
|
|
|
|
define i16 @t2() nounwind uwtable ssp {
|
|
; ELF64: t2
|
|
%1 = load i16* @b, align 2
|
|
; ELF64: lhz
|
|
%2 = add nsw i16 %1, 1
|
|
; ELF64: addi
|
|
ret i16 %2
|
|
}
|
|
|
|
define i32 @t3() nounwind uwtable ssp {
|
|
; ELF64: t3
|
|
%1 = load i32* @c, align 4
|
|
; ELF64: lwz
|
|
%2 = add nsw i32 %1, 1
|
|
; ELF64: addi
|
|
ret i32 %2
|
|
}
|
|
|
|
define i64 @t4() nounwind uwtable ssp {
|
|
; ELF64: t4
|
|
%1 = load i64* @d, align 4
|
|
; ELF64: ld
|
|
%2 = add nsw i64 %1, 1
|
|
; ELF64: addi
|
|
ret i64 %2
|
|
}
|
|
|
|
define float @t5() nounwind uwtable ssp {
|
|
; ELF64: t5
|
|
%1 = load float* @e, align 4
|
|
; ELF64: lfs
|
|
%2 = fadd float %1, 1.0
|
|
; ELF64: fadds
|
|
ret float %2
|
|
}
|
|
|
|
define double @t6() nounwind uwtable ssp {
|
|
; ELF64: t6
|
|
%1 = load double* @f, align 8
|
|
; ELF64: lfd
|
|
%2 = fadd double %1, 1.0
|
|
; ELF64: fadd
|
|
ret double %2
|
|
}
|
|
|
|
; store
|
|
|
|
define void @t7(i8 %v) nounwind uwtable ssp {
|
|
; ELF64: t7
|
|
%1 = add nsw i8 %v, 1
|
|
store i8 %1, i8* @a, align 1
|
|
; ELF64: addis
|
|
; ELF64: addi
|
|
; ELF64: addi
|
|
; ELF64: stb
|
|
ret void
|
|
}
|
|
|
|
define void @t8(i16 %v) nounwind uwtable ssp {
|
|
; ELF64: t8
|
|
%1 = add nsw i16 %v, 1
|
|
store i16 %1, i16* @b, align 2
|
|
; ELF64: addis
|
|
; ELF64: addi
|
|
; ELF64: addi
|
|
; ELF64: sth
|
|
ret void
|
|
}
|
|
|
|
define void @t9(i32 %v) nounwind uwtable ssp {
|
|
; ELF64: t9
|
|
%1 = add nsw i32 %v, 1
|
|
store i32 %1, i32* @c, align 4
|
|
; ELF64: addis
|
|
; ELF64: addi
|
|
; ELF64: addi
|
|
; ELF64: stw
|
|
ret void
|
|
}
|
|
|
|
define void @t10(i64 %v) nounwind uwtable ssp {
|
|
; ELF64: t10
|
|
%1 = add nsw i64 %v, 1
|
|
store i64 %1, i64* @d, align 4
|
|
; ELF64: addis
|
|
; ELF64: addi
|
|
; ELF64: addi
|
|
; ELF64: std
|
|
ret void
|
|
}
|
|
|
|
define void @t11(float %v) nounwind uwtable ssp {
|
|
; ELF64: t11
|
|
%1 = fadd float %v, 1.0
|
|
store float %1, float* @e, align 4
|
|
; ELF64: fadds
|
|
; ELF64: stfs
|
|
ret void
|
|
}
|
|
|
|
define void @t12(double %v) nounwind uwtable ssp {
|
|
; ELF64: t12
|
|
%1 = fadd double %v, 1.0
|
|
store double %1, double* @f, align 8
|
|
; ELF64: fadd
|
|
; ELF64: stfd
|
|
ret void
|
|
}
|
|
|
|
;; lwa requires an offset divisible by 4, so we need lwax here.
|
|
define i64 @t13() nounwind uwtable ssp {
|
|
; ELF64: t13
|
|
%1 = load i32* getelementptr inbounds (%struct.s* @g, i32 0, i32 1), align 1
|
|
%2 = sext i32 %1 to i64
|
|
; ELF64: li
|
|
; ELF64: lwax
|
|
%3 = add nsw i64 %2, 1
|
|
; ELF64: addi
|
|
ret i64 %3
|
|
}
|
|
|
|
;; ld requires an offset divisible by 4, so we need ldx here.
|
|
define i64 @t14() nounwind uwtable ssp {
|
|
; ELF64: t14
|
|
%1 = load i64* getelementptr inbounds (%struct.t* @h, i32 0, i32 1), align 1
|
|
; ELF64: li
|
|
; ELF64: ldx
|
|
%2 = add nsw i64 %1, 1
|
|
; ELF64: addi
|
|
ret i64 %2
|
|
}
|
|
|
|
;; std requires an offset divisible by 4, so we need stdx here.
|
|
define void @t15(i64 %v) nounwind uwtable ssp {
|
|
; ELF64: t15
|
|
%1 = add nsw i64 %v, 1
|
|
store i64 %1, i64* getelementptr inbounds (%struct.t* @h, i32 0, i32 1), align 1
|
|
; ELF64: addis
|
|
; ELF64: addi
|
|
; ELF64: addi
|
|
; ELF64: li
|
|
; ELF64: stdx
|
|
ret void
|
|
}
|
|
|
|
;; ld requires an offset that fits in 16 bits, so we need ldx here.
|
|
define i64 @t16() nounwind uwtable ssp {
|
|
; ELF64: t16
|
|
%1 = load i64* getelementptr inbounds ([8192 x i64]* @i, i32 0, i64 5000), align 8
|
|
; ELF64: lis
|
|
; ELF64: ori
|
|
; ELF64: ldx
|
|
%2 = add nsw i64 %1, 1
|
|
; ELF64: addi
|
|
ret i64 %2
|
|
}
|
|
|
|
;; std requires an offset that fits in 16 bits, so we need stdx here.
|
|
define void @t17(i64 %v) nounwind uwtable ssp {
|
|
; ELF64: t17
|
|
%1 = add nsw i64 %v, 1
|
|
store i64 %1, i64* getelementptr inbounds ([8192 x i64]* @i, i32 0, i64 5000), align 8
|
|
; ELF64: addis
|
|
; ELF64: ld
|
|
; ELF64: addi
|
|
; ELF64: lis
|
|
; ELF64: ori
|
|
; ELF64: stdx
|
|
ret void
|
|
}
|
|
|