mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-24 22:32:47 +00:00
5c9bb7119a
Moving these patterns from TableGen files to PerformDAGCombine() should allow us to generate better code by eliminating unnecessary shifts and extensions earlier. This also fixes a bug where the MAD pattern was calling SimplifyDemandedBits with a 24-bit mask on the first operand even when the full pattern wasn't being matched. This occasionally resulted in some instructions being incorrectly deleted from the program. v2: - Fix bug with 64-bit mul git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@205731 91177308-0d34-0410-b5e6-96231b3b80d8
76 lines
2.3 KiB
LLVM
76 lines
2.3 KiB
LLVM
; RUN: llc < %s -march=r600 -mcpu=redwood | FileCheck %s --check-prefix=EG --check-prefix=FUNC
|
|
; RUN: llc < %s -march=r600 -mcpu=cayman | FileCheck %s --check-prefix=EG --check-prefix=FUNC
|
|
; RUN: llc < %s -march=r600 -mcpu=SI -verify-machineinstrs | FileCheck %s --check-prefix=SI --check-prefix=FUNC
|
|
|
|
; FUNC-LABEL: @u32_mad24
|
|
; EG: MULADD_UINT24
|
|
; SI: V_MAD_U32_U24
|
|
|
|
define void @u32_mad24(i32 addrspace(1)* %out, i32 %a, i32 %b, i32 %c) {
|
|
entry:
|
|
%0 = shl i32 %a, 8
|
|
%a_24 = lshr i32 %0, 8
|
|
%1 = shl i32 %b, 8
|
|
%b_24 = lshr i32 %1, 8
|
|
%2 = mul i32 %a_24, %b_24
|
|
%3 = add i32 %2, %c
|
|
store i32 %3, i32 addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @i16_mad24
|
|
; The order of A and B does not matter.
|
|
; EG: MULADD_UINT24 {{[* ]*}}T{{[0-9]}}.[[MAD_CHAN:[XYZW]]]
|
|
; The result must be sign-extended
|
|
; EG: BFE_INT {{[* ]*}}T{{[0-9]\.[XYZW]}}, PV.[[MAD_CHAN]], 0.0, literal.x
|
|
; EG: 16
|
|
; SI: V_MAD_U32_U24 [[MAD:v[0-9]]], {{[sv][0-9], [sv][0-9]}}
|
|
; SI: V_BFE_I32 v{{[0-9]}}, [[MAD]], 0, 16
|
|
|
|
define void @i16_mad24(i32 addrspace(1)* %out, i16 %a, i16 %b, i16 %c) {
|
|
entry:
|
|
%0 = mul i16 %a, %b
|
|
%1 = add i16 %0, %c
|
|
%2 = sext i16 %1 to i32
|
|
store i32 %2, i32 addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @i8_mad24
|
|
; EG: MULADD_UINT24 {{[* ]*}}T{{[0-9]}}.[[MAD_CHAN:[XYZW]]]
|
|
; The result must be sign-extended
|
|
; EG: BFE_INT {{[* ]*}}T{{[0-9]\.[XYZW]}}, PV.[[MAD_CHAN]], 0.0, literal.x
|
|
; EG: 8
|
|
; SI: V_MAD_U32_U24 [[MUL:v[0-9]]], {{[sv][0-9], [sv][0-9]}}
|
|
; SI: V_BFE_I32 v{{[0-9]}}, [[MUL]], 0, 8
|
|
|
|
define void @i8_mad24(i32 addrspace(1)* %out, i8 %a, i8 %b, i8 %c) {
|
|
entry:
|
|
%0 = mul i8 %a, %b
|
|
%1 = add i8 %0, %c
|
|
%2 = sext i8 %1 to i32
|
|
store i32 %2, i32 addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; This tests for a bug where the mad_u24 pattern matcher would call
|
|
; SimplifyDemandedBits on the first operand of the mul instruction
|
|
; assuming that the pattern would be matched to a 24-bit mad. This
|
|
; led to some instructions being incorrectly erased when the entire
|
|
; 24-bit mad pattern wasn't being matched.
|
|
|
|
; Check that the select instruction is not deleted.
|
|
; FUNC-LABEL: @i24_i32_i32_mad
|
|
; EG: CNDE_INT
|
|
; SI: V_CNDMASK
|
|
define void @i24_i32_i32_mad(i32 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i32 %d) {
|
|
entry:
|
|
%0 = ashr i32 %a, 8
|
|
%1 = icmp ne i32 %c, 0
|
|
%2 = select i1 %1, i32 %0, i32 34
|
|
%3 = mul i32 %2, %c
|
|
%4 = add i32 %3, %d
|
|
store i32 %4, i32 addrspace(1)* %out
|
|
ret void
|
|
}
|