mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-15 20:06:46 +00:00
d336de318e
(movzx/movsx) because they give more information. Revert that part of the patch. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@129498 91177308-0d34-0410-b5e6-96231b3b80d8
107 lines
2.2 KiB
LLVM
107 lines
2.2 KiB
LLVM
; RUN: llc < %s -mtriple=x86_64-linux | FileCheck %s -check-prefix=X86-64
|
|
; RUN: llc < %s -mtriple=x86_64-win32 | FileCheck %s -check-prefix=WIN64
|
|
; RUN: llc < %s -march=x86 | FileCheck %s -check-prefix=X86-32
|
|
|
|
; Use h registers. On x86-64, codegen doesn't support general allocation
|
|
; of h registers yet, due to x86 encoding complications.
|
|
|
|
define void @bar64(i64 inreg %x, i8* inreg %p) nounwind {
|
|
; X86-64: bar64:
|
|
; X86-64: shrq $8, %rdi
|
|
; X86-64: incb %dil
|
|
|
|
; See FIXME: on regclass GR8.
|
|
; It could be optimally transformed like; incb %ch; movb %ch, (%rdx)
|
|
; WIN64: bar64:
|
|
; WIN64: shrq $8, %rcx
|
|
; WIN64: incb %cl
|
|
|
|
; X86-32: bar64:
|
|
; X86-32: incb %ah
|
|
%t0 = lshr i64 %x, 8
|
|
%t1 = trunc i64 %t0 to i8
|
|
%t2 = add i8 %t1, 1
|
|
store i8 %t2, i8* %p
|
|
ret void
|
|
}
|
|
|
|
define void @bar32(i32 inreg %x, i8* inreg %p) nounwind {
|
|
; X86-64: bar32:
|
|
; X86-64: shrl $8, %edi
|
|
; X86-64: incb %dil
|
|
|
|
; WIN64: bar32:
|
|
; WIN64: shrl $8, %ecx
|
|
; WIN64: incb %cl
|
|
|
|
; X86-32: bar32:
|
|
; X86-32: incb %ah
|
|
%t0 = lshr i32 %x, 8
|
|
%t1 = trunc i32 %t0 to i8
|
|
%t2 = add i8 %t1, 1
|
|
store i8 %t2, i8* %p
|
|
ret void
|
|
}
|
|
|
|
define void @bar16(i16 inreg %x, i8* inreg %p) nounwind {
|
|
; X86-64: bar16:
|
|
; X86-64: shrl $8, %edi
|
|
; X86-64: incb %dil
|
|
|
|
; WIN64: bar16:
|
|
; WIN64: shrl $8, %ecx
|
|
; WIN64: incb %cl
|
|
|
|
; X86-32: bar16:
|
|
; X86-32: incb %ah
|
|
%t0 = lshr i16 %x, 8
|
|
%t1 = trunc i16 %t0 to i8
|
|
%t2 = add i8 %t1, 1
|
|
store i8 %t2, i8* %p
|
|
ret void
|
|
}
|
|
|
|
define i64 @qux64(i64 inreg %x) nounwind {
|
|
; X86-64: qux64:
|
|
; X86-64: movq %rdi, %rax
|
|
; X86-64: movzbl %ah, %eax
|
|
|
|
; WIN64: qux64:
|
|
; WIN64: movzbl %ch, %eax
|
|
|
|
; X86-32: qux64:
|
|
; X86-32: movzbl %ah, %eax
|
|
%t0 = lshr i64 %x, 8
|
|
%t1 = and i64 %t0, 255
|
|
ret i64 %t1
|
|
}
|
|
|
|
define i32 @qux32(i32 inreg %x) nounwind {
|
|
; X86-64: qux32:
|
|
; X86-64: movl %edi, %eax
|
|
; X86-64: movzbl %ah, %eax
|
|
|
|
; WIN64: qux32:
|
|
; WIN64: movzbl %ch, %eax
|
|
|
|
; X86-32: qux32:
|
|
; X86-32: movzbl %ah, %eax
|
|
%t0 = lshr i32 %x, 8
|
|
%t1 = and i32 %t0, 255
|
|
ret i32 %t1
|
|
}
|
|
|
|
define i16 @qux16(i16 inreg %x) nounwind {
|
|
; X86-64: qux16:
|
|
; X86-64: movl %edi, %eax
|
|
; X86-64: movzbl %ah, %eax
|
|
|
|
; WIN64: qux16:
|
|
; WIN64: movzbl %ch, %eax
|
|
|
|
; X86-32: qux16:
|
|
; X86-32: movzbl %ah, %eax
|
|
%t0 = lshr i16 %x, 8
|
|
ret i16 %t0
|
|
}
|