mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-22 23:31:32 +00:00
7b837d8c75
This adds a second implementation of the AArch64 architecture to LLVM, accessible in parallel via the "arm64" triple. The plan over the coming weeks & months is to merge the two into a single backend, during which time thorough code review should naturally occur. Everything will be easier with the target in-tree though, hence this commit. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@205090 91177308-0d34-0410-b5e6-96231b3b80d8
144 lines
3.9 KiB
LLVM
144 lines
3.9 KiB
LLVM
; RUN: llc -verify-machineinstrs -mtriple=arm64-linux-gnu -pre-RA-sched=linearize -enable-misched=false < %s | FileCheck %s
|
|
|
|
%va_list = type {i8*, i8*, i8*, i32, i32}
|
|
|
|
@var = global %va_list zeroinitializer, align 8
|
|
|
|
declare void @llvm.va_start(i8*)
|
|
|
|
define void @test_simple(i32 %n, ...) {
|
|
; CHECK-LABEL: test_simple:
|
|
; CHECK: sub sp, sp, #[[STACKSIZE:[0-9]+]]
|
|
; CHECK: add [[STACK_TOP:x[0-9]+]], sp, #[[STACKSIZE]]
|
|
|
|
; CHECK: adrp x[[VA_LIST_HI:[0-9]+]], var
|
|
|
|
; CHECK: stp x1, x2, [sp, #[[GR_BASE:[0-9]+]]]
|
|
; ... omit middle ones ...
|
|
; CHECK: str x7, [sp, #
|
|
|
|
; CHECK: stp q0, q1, [sp]
|
|
; ... omit middle ones ...
|
|
; CHECK: stp q6, q7, [sp, #
|
|
|
|
; CHECK: str [[STACK_TOP]], [x[[VA_LIST_HI]], :lo12:var]
|
|
|
|
; CHECK: add [[GR_TOPTMP:x[0-9]+]], sp, #[[GR_BASE]]
|
|
; CHECK: add [[GR_TOP:x[0-9]+]], [[GR_TOPTMP]], #56
|
|
; CHECK: add x[[VA_LIST:[0-9]+]], {{x[0-9]+}}, :lo12:var
|
|
; CHECK: str [[GR_TOP]], [x[[VA_LIST]], #8]
|
|
|
|
; CHECK: mov [[VR_TOPTMP:x[0-9]+]], sp
|
|
; CHECK: add [[VR_TOP:x[0-9]+]], [[VR_TOPTMP]], #128
|
|
; CHECK: str [[VR_TOP]], [x[[VA_LIST]], #16]
|
|
|
|
; CHECK: movn [[GR_OFFS:w[0-9]+]], #55
|
|
; CHECK: str [[GR_OFFS]], [x[[VA_LIST]], #24]
|
|
|
|
; CHECK: orr [[VR_OFFS:w[0-9]+]], wzr, #0xffffff80
|
|
; CHECK: str [[VR_OFFS]], [x[[VA_LIST]], #28]
|
|
|
|
%addr = bitcast %va_list* @var to i8*
|
|
call void @llvm.va_start(i8* %addr)
|
|
|
|
ret void
|
|
}
|
|
|
|
define void @test_fewargs(i32 %n, i32 %n1, i32 %n2, float %m, ...) {
|
|
; CHECK-LABEL: test_fewargs:
|
|
; CHECK: sub sp, sp, #[[STACKSIZE:[0-9]+]]
|
|
; CHECK: add [[STACK_TOP:x[0-9]+]], sp, #[[STACKSIZE]]
|
|
|
|
; CHECK: adrp x[[VA_LIST_HI:[0-9]+]], var
|
|
|
|
; CHECK: stp x3, x4, [sp, #[[GR_BASE:[0-9]+]]]
|
|
; ... omit middle ones ...
|
|
; CHECK: str x7, [sp, #
|
|
|
|
; CHECK: stp q1, q2, [sp]
|
|
; ... omit middle ones ...
|
|
; CHECK: str q7, [sp, #
|
|
|
|
; CHECK: str [[STACK_TOP]], [x[[VA_LIST_HI]], :lo12:var]
|
|
|
|
; CHECK: add [[GR_TOPTMP:x[0-9]+]], sp, #[[GR_BASE]]
|
|
; CHECK: add [[GR_TOP:x[0-9]+]], [[GR_TOPTMP]], #40
|
|
; CHECK: add x[[VA_LIST:[0-9]+]], {{x[0-9]+}}, :lo12:var
|
|
; CHECK: str [[GR_TOP]], [x[[VA_LIST]], #8]
|
|
|
|
; CHECK: mov [[VR_TOPTMP:x[0-9]+]], sp
|
|
; CHECK: add [[VR_TOP:x[0-9]+]], [[VR_TOPTMP]], #112
|
|
; CHECK: str [[VR_TOP]], [x[[VA_LIST]], #16]
|
|
|
|
; CHECK: movn [[GR_OFFS:w[0-9]+]], #39
|
|
; CHECK: str [[GR_OFFS]], [x[[VA_LIST]], #24]
|
|
|
|
; CHECK: movn [[VR_OFFS:w[0-9]+]], #111
|
|
; CHECK: str [[VR_OFFS]], [x[[VA_LIST]], #28]
|
|
|
|
%addr = bitcast %va_list* @var to i8*
|
|
call void @llvm.va_start(i8* %addr)
|
|
|
|
ret void
|
|
}
|
|
|
|
define void @test_nospare([8 x i64], [8 x float], ...) {
|
|
; CHECK-LABEL: test_nospare:
|
|
|
|
%addr = bitcast %va_list* @var to i8*
|
|
call void @llvm.va_start(i8* %addr)
|
|
; CHECK-NOT: sub sp, sp
|
|
; CHECK: mov [[STACK:x[0-9]+]], sp
|
|
; CHECK: str [[STACK]], [{{x[0-9]+}}, :lo12:var]
|
|
|
|
ret void
|
|
}
|
|
|
|
; If there are non-variadic arguments on the stack (here two i64s) then the
|
|
; __stack field should point just past them.
|
|
define void @test_offsetstack([10 x i64], [3 x float], ...) {
|
|
; CHECK-LABEL: test_offsetstack:
|
|
; CHECK: sub sp, sp, #80
|
|
; CHECK: add [[STACK_TOP:x[0-9]+]], sp, #96
|
|
; CHECK: str [[STACK_TOP]], [{{x[0-9]+}}, :lo12:var]
|
|
|
|
%addr = bitcast %va_list* @var to i8*
|
|
call void @llvm.va_start(i8* %addr)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.va_end(i8*)
|
|
|
|
define void @test_va_end() nounwind {
|
|
; CHECK-LABEL: test_va_end:
|
|
; CHECK-NEXT: BB#0
|
|
|
|
%addr = bitcast %va_list* @var to i8*
|
|
call void @llvm.va_end(i8* %addr)
|
|
|
|
ret void
|
|
; CHECK-NEXT: ret
|
|
}
|
|
|
|
declare void @llvm.va_copy(i8* %dest, i8* %src)
|
|
|
|
@second_list = global %va_list zeroinitializer
|
|
|
|
define void @test_va_copy() {
|
|
; CHECK-LABEL: test_va_copy:
|
|
%srcaddr = bitcast %va_list* @var to i8*
|
|
%dstaddr = bitcast %va_list* @second_list to i8*
|
|
call void @llvm.va_copy(i8* %dstaddr, i8* %srcaddr)
|
|
|
|
; CHECK: add x[[SRC:[0-9]+]], {{x[0-9]+}}, :lo12:var
|
|
|
|
; CHECK: ldr [[BLOCK:q[0-9]+]], [x[[SRC]]]
|
|
; CHECK: add x[[DST:[0-9]+]], {{x[0-9]+}}, :lo12:second_list
|
|
; CHECK: str [[BLOCK]], [x[[DST]]]
|
|
|
|
; CHECK: ldr [[BLOCK:q[0-9]+]], [x[[SRC]], #16]
|
|
; CHECK: str [[BLOCK]], [x[[DST]], #16]
|
|
ret void
|
|
; CHECK: ret
|
|
}
|