mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 04:30:12 +00:00
f0569be4a9
DAGcombine's ability to find reasons to remove truncates when they were not needed. Consequently, the CellSPU backend would produce correct, but _really slow and horrible_, code. Replaced with instruction sequences that do the equivalent truncation in SPUInstrInfo.td. - Re-examine how unaligned loads and stores work. Generated unaligned load code has been tested on the CellSPU hardware; see the i32operations.c and i64operations.c in CodeGen/CellSPU/useful-harnesses. (While they may be toy test code, it does prove that some real world code does compile correctly.) - Fix truncating stores in bug 3193 (note: unpack_df.ll will still make llc fault because i64 ult is not yet implemented.) - Added i64 eq and neq for setcc and select/setcc; started new instruction information file for them in SPU64InstrInfo.td. Additional i64 operations should be added to this file and not to SPUInstrInfo.td. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@61447 91177308-0d34-0410-b5e6-96231b3b80d8
50 lines
2.4 KiB
LLVM
50 lines
2.4 KiB
LLVM
; RUN: llvm-as -o - %s | llc -march=cellspu > %t1.s
|
|
; RUN: llvm-as -o - %s | llc -march=cellspu -mattr=large_mem > %t2.s
|
|
; RUN: grep bisl %t1.s | count 7
|
|
; RUN: grep ila %t1.s | count 1
|
|
; RUN: grep rotqby %t1.s | count 6
|
|
; RUN: grep lqa %t1.s | count 1
|
|
; RUN: grep lqd %t1.s | count 12
|
|
; RUN: grep dispatch_tab %t1.s | count 5
|
|
; RUN: grep bisl %t2.s | count 7
|
|
; RUN: grep ilhu %t2.s | count 2
|
|
; RUN: grep iohl %t2.s | count 2
|
|
; RUN: grep rotqby %t2.s | count 6
|
|
; RUN: grep lqd %t2.s | count 13
|
|
; RUN: grep ilhu %t2.s | count 2
|
|
; RUN: grep ai %t2.s | count 9
|
|
; RUN: grep dispatch_tab %t2.s | count 6
|
|
|
|
; ModuleID = 'call_indirect.bc'
|
|
target datalayout = "E-p:32:32:128-f64:64:128-f32:32:128-i64:32:128-i32:32:128-i16:16:128-i8:8:128-i1:8:128-a0:0:128-v128:128:128"
|
|
target triple = "spu-unknown-elf"
|
|
|
|
@dispatch_tab = global [6 x void (i32, float)*] zeroinitializer, align 16
|
|
|
|
define void @dispatcher(i32 %i_arg, float %f_arg) {
|
|
entry:
|
|
%tmp2 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 0), align 16
|
|
tail call void %tmp2( i32 %i_arg, float %f_arg )
|
|
%tmp2.1 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 1), align 4
|
|
tail call void %tmp2.1( i32 %i_arg, float %f_arg )
|
|
%tmp2.2 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 2), align 4
|
|
tail call void %tmp2.2( i32 %i_arg, float %f_arg )
|
|
%tmp2.3 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 3), align 4
|
|
tail call void %tmp2.3( i32 %i_arg, float %f_arg )
|
|
%tmp2.4 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 4), align 4
|
|
tail call void %tmp2.4( i32 %i_arg, float %f_arg )
|
|
%tmp2.5 = load void (i32, float)** getelementptr ([6 x void (i32, float)*]* @dispatch_tab, i32 0, i32 5), align 4
|
|
tail call void %tmp2.5( i32 %i_arg, float %f_arg )
|
|
ret void
|
|
}
|
|
|
|
@ptr_list = internal global [1 x void ()*] [ void ()* inttoptr (i64 4294967295 to void ()*) ], align 4
|
|
@ptr.a = internal global void ()** getelementptr ([1 x void ()*]* @ptr_list, i32 0, i32 1), align 16
|
|
|
|
define void @double_indirect_call() {
|
|
%a = load void ()*** @ptr.a, align 16
|
|
%b = load void ()** %a, align 4
|
|
tail call void %b()
|
|
ret void
|
|
}
|