mirror of
				https://github.com/c64scene-ar/llvm-6502.git
				synced 2025-10-31 08:16:47 +00:00 
			
		
		
		
	LocalStackSlotPass assumes that isFrameOffsetLegal doesn't change its answer when the base register changes. Unfortunately this isn't true in thumb1, where SP-based loads allow a larger offset than non-SP-based loads, and this causes the base register reuse code to generate instructions that are unencodable, causing an assertion failure. Solve this by adding a BaseReg parameter to isFrameOffsetLegal, which ARMBaseRegisterInfo can then make use of to give the correct answer. Differential Revision: http://reviews.llvm.org/D8419 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@232825 91177308-0d34-0410-b5e6-96231b3b80d8
		
			
				
	
	
		
			128 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
			
		
		
	
	
			128 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
| ; RUN: llc -mtriple=thumb-eabi < %s -o - | FileCheck %s
 | |
| 
 | |
| ; Check that stack addresses are generated using a single ADD
 | |
| define void @test1(i8** %p) {
 | |
|   %x = alloca i8, align 1
 | |
|   %y = alloca i8, align 1
 | |
|   %z = alloca i8, align 1
 | |
| ; CHECK: add r1, sp, #8
 | |
| ; CHECK: str r1, [r0]
 | |
|   store i8* %x, i8** %p, align 4
 | |
| ; CHECK: add r1, sp, #4
 | |
| ; CHECK: str r1, [r0]
 | |
|   store i8* %y, i8** %p, align 4
 | |
| ; CHECK: mov r1, sp
 | |
| ; CHECK: str r1, [r0]
 | |
|   store i8* %z, i8** %p, align 4
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; Stack offsets larger than 1020 still need two ADDs
 | |
| define void @test2([1024 x i8]** %p) {
 | |
|   %arr1 = alloca [1024 x i8], align 1
 | |
|   %arr2 = alloca [1024 x i8], align 1
 | |
| ; CHECK: add r1, sp, #1020
 | |
| ; CHECK: adds r1, #4
 | |
| ; CHECK: str r1, [r0]
 | |
|   store [1024 x i8]* %arr1, [1024 x i8]** %p, align 4
 | |
| ; CHECK: mov r1, sp
 | |
| ; CHECK: str r1, [r0]
 | |
|   store [1024 x i8]* %arr2, [1024 x i8]** %p, align 4
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; If possible stack-based lrdb/ldrh are widened to use SP-based addressing
 | |
| define i32 @test3() #0 {
 | |
|   %x = alloca i8, align 1
 | |
|   %y = alloca i8, align 1
 | |
| ; CHECK: ldr r0, [sp]
 | |
|   %1 = load i8, i8* %x, align 1
 | |
| ; CHECK: ldr r1, [sp, #4]
 | |
|   %2 = load i8, i8* %y, align 1
 | |
|   %3 = add nsw i8 %1, %2
 | |
|   %4 = zext i8 %3 to i32
 | |
|   ret i32 %4
 | |
| }
 | |
| 
 | |
| define i32 @test4() #0 {
 | |
|   %x = alloca i16, align 2
 | |
|   %y = alloca i16, align 2
 | |
| ; CHECK: ldr r0, [sp]
 | |
|   %1 = load i16, i16* %x, align 2
 | |
| ; CHECK: ldr r1, [sp, #4]
 | |
|   %2 = load i16, i16* %y, align 2
 | |
|   %3 = add nsw i16 %1, %2
 | |
|   %4 = zext i16 %3 to i32
 | |
|   ret i32 %4
 | |
| }
 | |
| 
 | |
| ; Don't widen if the value needs to be zero-extended
 | |
| define zeroext i8 @test5() {
 | |
|   %x = alloca i8, align 1
 | |
| ; CHECK: mov r0, sp
 | |
| ; CHECK: ldrb r0, [r0]
 | |
|   %1 = load i8, i8* %x, align 1
 | |
|   ret i8 %1
 | |
| }
 | |
| 
 | |
| define zeroext i16 @test6() {
 | |
|   %x = alloca i16, align 2
 | |
| ; CHECK: mov r0, sp
 | |
| ; CHECK: ldrh r0, [r0]
 | |
|   %1 = load i16, i16* %x, align 2
 | |
|   ret i16 %1
 | |
| }
 | |
| 
 | |
| ; Accessing the bottom of a large array shouldn't require materializing a base
 | |
| define void @test7() {
 | |
|   %arr = alloca [200 x i32], align 4
 | |
| 
 | |
|   ; CHECK: movs [[REG:r[0-9]+]], #1
 | |
|   ; CHECK: str [[REG]], [sp, #4]
 | |
|   %arrayidx = getelementptr inbounds [200 x i32], [200 x i32]* %arr, i32 0, i32 1
 | |
|   store i32 1, i32* %arrayidx, align 4
 | |
| 
 | |
|   ; CHECK: str [[REG]], [sp, #16]
 | |
|   %arrayidx1 = getelementptr inbounds [200 x i32], [200 x i32]* %arr, i32 0, i32 4
 | |
|   store i32 1, i32* %arrayidx1, align 4
 | |
| 
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; Check that loads/stores with out-of-range offsets are handled correctly
 | |
| define void @test8() {
 | |
|   %arr3 = alloca [224 x i32], align 4
 | |
|   %arr2 = alloca [224 x i32], align 4
 | |
|   %arr1 = alloca [224 x i32], align 4
 | |
| 
 | |
| ; CHECK: movs [[REG:r[0-9]+]], #1
 | |
| ; CHECK: str [[REG]], [sp]
 | |
|   %arr1idx1 = getelementptr inbounds [224 x i32], [224 x i32]* %arr1, i32 0, i32 0
 | |
|   store i32 1, i32* %arr1idx1, align 4
 | |
| 
 | |
| ; Offset in range for sp-based store, but not for non-sp-based store
 | |
| ; CHECK: str [[REG]], [sp, #128]
 | |
|   %arr1idx2 = getelementptr inbounds [224 x i32], [224 x i32]* %arr1, i32 0, i32 32
 | |
|   store i32 1, i32* %arr1idx2, align 4
 | |
| 
 | |
| ; CHECK: str [[REG]], [sp, #896]
 | |
|   %arr2idx1 = getelementptr inbounds [224 x i32], [224 x i32]* %arr2, i32 0, i32 0
 | |
|   store i32 1, i32* %arr2idx1, align 4
 | |
| 
 | |
| ; %arr2 is in range, but this element of it is not
 | |
| ; CHECK: str [[REG]], [{{r[0-9]+}}]
 | |
|   %arr2idx2 = getelementptr inbounds [224 x i32], [224 x i32]* %arr2, i32 0, i32 32
 | |
|   store i32 1, i32* %arr2idx2, align 4
 | |
| 
 | |
| ; %arr3 is not in range
 | |
| ; CHECK: str [[REG]], [{{r[0-9]+}}]
 | |
|   %arr3idx1 = getelementptr inbounds [224 x i32], [224 x i32]* %arr3, i32 0, i32 0
 | |
|   store i32 1, i32* %arr3idx1, align 4
 | |
| 
 | |
| ; CHECK: str [[REG]], [{{r[0-9]+}}]
 | |
|   %arr3idx2 = getelementptr inbounds [224 x i32], [224 x i32]* %arr3, i32 0, i32 32
 | |
|   store i32 1, i32* %arr3idx2, align 4
 | |
| 
 | |
|   ret void
 | |
| }
 |