mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-24 22:32:47 +00:00
38c6b58eec
Summary: AsmPrinter::EmitInlineAsm() will no longer use the EmitRawText() call for targets with mature MC support. Such targets will always parse the inline assembly (even when emitting assembly). Targets without mature MC support continue to use EmitRawText() for assembly output. The hasRawTextSupport() check in AsmPrinter::EmitInlineAsm() has been replaced with MCAsmInfo::UseIntegratedAs which when true, causes the integrated assembler to parse inline assembly (even when emitting assembly output). UseIntegratedAs is set to true for targets that consider any failure to parse valid assembly to be a bug. Target specific subclasses generally enable the integrated assembler in their constructor. The default value can be overridden with -no-integrated-as. All tests that rely on inline assembly supporting invalid assembly (for example, those that use mnemonics such as 'foo' or 'hello world') have been updated to disable the integrated assembler. Changes since review (and last commit attempt): - Fixed test failures that were missed due to configuration of local build. (fixes crash.ll and a couple others). - Fixed tests that happened to pass because the local build was on X86 (should fix 2007-12-17-InvokeAsm.ll) - mature-mc-support.ll's should no longer require all targets to be compiled. (should fix ARM and PPC buildbots) - Object output (-filetype=obj and similar) now forces the integrated assembler to be enabled regardless of default setting or -no-integrated-as. (should fix SystemZ buildbots) Reviewers: rafael Reviewed By: rafael CC: llvm-commits Differential Revision: http://llvm-reviews.chandlerc.com/D2686 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@201333 91177308-0d34-0410-b5e6-96231b3b80d8
74 lines
2.2 KiB
LLVM
74 lines
2.2 KiB
LLVM
; RUN: llc < %s -march=arm -mattr=+vfp2 -no-integrated-as | FileCheck %s
|
|
|
|
define i32 @foo(float %scale, float %scale2) nounwind {
|
|
entry:
|
|
%scale.addr = alloca float, align 4
|
|
%scale2.addr = alloca float, align 4
|
|
store float %scale, float* %scale.addr, align 4
|
|
store float %scale2, float* %scale2.addr, align 4
|
|
%tmp = load float* %scale.addr, align 4
|
|
%tmp1 = load float* %scale2.addr, align 4
|
|
call void asm sideeffect "vmul.f32 q0, q0, ${0:y} \0A\09vmul.f32 q1, q1, ${0:y} \0A\09vmul.f32 q1, q0, ${1:y} \0A\09", "w,w,~{q0},~{q1}"(float %tmp, float %tmp1) nounwind
|
|
ret i32 0
|
|
}
|
|
|
|
define void @f0() nounwind {
|
|
entry:
|
|
; CHECK: f0
|
|
; CHECK: .word -1
|
|
call void asm sideeffect ".word ${0:B} \0A\09", "i"(i32 0) nounwind
|
|
ret void
|
|
}
|
|
|
|
define void @f1() nounwind {
|
|
entry:
|
|
; CHECK: f1
|
|
; CHECK: .word 65535
|
|
call void asm sideeffect ".word ${0:L} \0A\09", "i"(i32 -1) nounwind
|
|
ret void
|
|
}
|
|
|
|
@f2_ptr = internal global i32* @f2_var, align 4
|
|
@f2_var = external global i32
|
|
|
|
define void @f2() nounwind {
|
|
entry:
|
|
; CHECK: f2
|
|
; CHECK: ldr r0, [r{{[0-9]+}}]
|
|
call void asm sideeffect "ldr r0, [${0:m}]\0A\09", "*m,~{r0}"(i32** @f2_ptr) nounwind
|
|
ret void
|
|
}
|
|
|
|
@f3_ptr = internal global i64* @f3_var, align 4
|
|
@f3_var = external global i64
|
|
@f3_var2 = external global i64
|
|
|
|
define void @f3() nounwind {
|
|
entry:
|
|
; CHECK: f3
|
|
; CHECK: stm {{lr|r[0-9]+}}, {[[REG1:(r[0-9]+)]], r{{[0-9]+}}}
|
|
; CHECK: adds {{lr|r[0-9]+}}, [[REG1]]
|
|
; CHECK: ldm {{lr|r[0-9]+}}, {r{{[0-9]+}}, r{{[0-9]+}}}
|
|
%tmp = load i64* @f3_var, align 4
|
|
%tmp1 = load i64* @f3_var2, align 4
|
|
%0 = call i64 asm sideeffect "stm ${0:m}, ${1:M}\0A\09adds $3, $1\0A\09", "=*m,=r,1,r"(i64** @f3_ptr, i64 %tmp, i64 %tmp1) nounwind
|
|
store i64 %0, i64* @f3_var, align 4
|
|
%1 = call i64 asm sideeffect "ldm ${1:m}, ${0:M}\0A\09", "=r,*m"(i64** @f3_ptr) nounwind
|
|
store i64 %1, i64* @f3_var, align 4
|
|
ret void
|
|
}
|
|
|
|
define i64 @f4(i64* %val) nounwind {
|
|
entry:
|
|
;CHECK-LABEL: f4:
|
|
;CHECK: ldrexd [[REG1:(r[0-9]?[02468])]], {{r[0-9]?[13579]}}, [r{{[0-9]+}}]
|
|
%0 = tail call i64 asm sideeffect "ldrexd $0, ${0:H}, [$1]", "=&r,r,*Qo"(i64* %val, i64* %val) nounwind
|
|
ret i64 %0
|
|
}
|
|
|
|
; PR16490
|
|
define void @f5(i64 %__pu_val) {
|
|
call void asm sideeffect "$1", "r,i"(i64 %__pu_val, i32 -14)
|
|
ret void
|
|
}
|