mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 04:30:12 +00:00
55a98b00c1
support constant extension. This patch doesn't introduce any functionality changes. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@175280 91177308-0d34-0410-b5e6-96231b3b80d8
4172 lines
154 KiB
TableGen
4172 lines
154 KiB
TableGen
//=- HexagonInstrInfoV4.td - Target Desc. for Hexagon Target -*- tablegen -*-=//
|
|
//
|
|
// The LLVM Compiler Infrastructure
|
|
//
|
|
// This file is distributed under the University of Illinois Open Source
|
|
// License. See LICENSE.TXT for details.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file describes the Hexagon V4 instructions in TableGen format.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
let neverHasSideEffects = 1 in
|
|
class T_Immext<dag ins> :
|
|
EXTENDERInst<(outs), ins, "immext(#$imm)", []>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def IMMEXT_b : T_Immext<(ins brtarget:$imm)>;
|
|
def IMMEXT_c : T_Immext<(ins calltarget:$imm)>;
|
|
def IMMEXT_g : T_Immext<(ins globaladdress:$imm)>;
|
|
def IMMEXT_i : T_Immext<(ins u26_6Imm:$imm)>;
|
|
|
|
// Fold (add (CONST32 tglobaladdr:$addr) <offset>) into a global address.
|
|
def FoldGlobalAddr : ComplexPattern<i32, 1, "foldGlobalAddress", [], []>;
|
|
|
|
// Fold (add (CONST32_GP tglobaladdr:$addr) <offset>) into a global address.
|
|
def FoldGlobalAddrGP : ComplexPattern<i32, 1, "foldGlobalAddressGP", [], []>;
|
|
|
|
def NumUsesBelowThresCONST32 : PatFrag<(ops node:$addr),
|
|
(HexagonCONST32 node:$addr), [{
|
|
return hasNumUsesBelowThresGA(N->getOperand(0).getNode());
|
|
}]>;
|
|
|
|
// Hexagon V4 Architecture spec defines 8 instruction classes:
|
|
// LD ST ALU32 XTYPE J JR MEMOP NV CR SYSTEM(system is not implemented in the
|
|
// compiler)
|
|
|
|
// LD Instructions:
|
|
// ========================================
|
|
// Loads (8/16/32/64 bit)
|
|
// Deallocframe
|
|
|
|
// ST Instructions:
|
|
// ========================================
|
|
// Stores (8/16/32/64 bit)
|
|
// Allocframe
|
|
|
|
// ALU32 Instructions:
|
|
// ========================================
|
|
// Arithmetic / Logical (32 bit)
|
|
// Vector Halfword
|
|
|
|
// XTYPE Instructions (32/64 bit):
|
|
// ========================================
|
|
// Arithmetic, Logical, Bit Manipulation
|
|
// Multiply (Integer, Fractional, Complex)
|
|
// Permute / Vector Permute Operations
|
|
// Predicate Operations
|
|
// Shift / Shift with Add/Sub/Logical
|
|
// Vector Byte ALU
|
|
// Vector Halfword (ALU, Shift, Multiply)
|
|
// Vector Word (ALU, Shift)
|
|
|
|
// J Instructions:
|
|
// ========================================
|
|
// Jump/Call PC-relative
|
|
|
|
// JR Instructions:
|
|
// ========================================
|
|
// Jump/Call Register
|
|
|
|
// MEMOP Instructions:
|
|
// ========================================
|
|
// Operation on memory (8/16/32 bit)
|
|
|
|
// NV Instructions:
|
|
// ========================================
|
|
// New-value Jumps
|
|
// New-value Stores
|
|
|
|
// CR Instructions:
|
|
// ========================================
|
|
// Control-Register Transfers
|
|
// Hardware Loop Setup
|
|
// Predicate Logicals & Reductions
|
|
|
|
// SYSTEM Instructions (not implemented in the compiler):
|
|
// ========================================
|
|
// Prefetch
|
|
// Cache Maintenance
|
|
// Bus Operations
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// ALU32 +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Shift halfword.
|
|
|
|
let isPredicated = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in {
|
|
def ASLH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = aslh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASLH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = aslh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASLH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = aslh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASLH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = aslh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASRH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = asrh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASRH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = asrh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASRH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = asrh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ASRH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = asrh($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Sign extend.
|
|
|
|
let isPredicated = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in {
|
|
def SXTB_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = sxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTB_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = sxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTB_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = sxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTB_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = sxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
def SXTH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = sxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = sxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = sxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def SXTH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = sxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Zero exten.
|
|
|
|
let neverHasSideEffects = 1, isPredicated = 1, validSubTargets = HasV4SubT in {
|
|
def ZXTB_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = zxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTB_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = zxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTB_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = zxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTB_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = zxtb($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1) $dst = zxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1) $dst = zxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if ($src1.new) $dst = zxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def ZXTH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2),
|
|
"if (!$src1.new) $dst = zxth($src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Generate frame index addresses.
|
|
let neverHasSideEffects = 1, isReMaterializable = 1,
|
|
isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in
|
|
def TFR_FI_immext_V4 : ALU32_ri<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, s32Imm:$offset),
|
|
"$dst = add($src1, ##$offset)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=cmp.eq(Rs,#s8)
|
|
let validSubTargets = HasV4SubT, isExtendable = 1, opExtendable = 2,
|
|
isExtentSigned = 1, opExtentBits = 8 in
|
|
def V4_A4_rcmpeqi : ALU32_ri<(outs IntRegs:$Rd),
|
|
(ins IntRegs:$Rs, s8Ext:$s8),
|
|
"$Rd = cmp.eq($Rs, #$s8)",
|
|
[(set (i32 IntRegs:$Rd),
|
|
(i32 (zext (i1 (seteq (i32 IntRegs:$Rs),
|
|
s8ExtPred:$s8)))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Preserve the TSTBIT generation
|
|
def : Pat <(i32 (zext (i1 (setne (i32 (and (i32 (shl 1, (i32 IntRegs:$src2))),
|
|
(i32 IntRegs:$src1))), 0)))),
|
|
(i32 (MUX_ii (i1 (TSTBIT_rr (i32 IntRegs:$src1), (i32 IntRegs:$src2))),
|
|
1, 0))>;
|
|
|
|
// Interfered with tstbit generation, above pattern preserves, see : tstbit.ll
|
|
// Rd=cmp.ne(Rs,#s8)
|
|
let validSubTargets = HasV4SubT, isExtendable = 1, opExtendable = 2,
|
|
isExtentSigned = 1, opExtentBits = 8 in
|
|
def V4_A4_rcmpneqi : ALU32_ri<(outs IntRegs:$Rd),
|
|
(ins IntRegs:$Rs, s8Ext:$s8),
|
|
"$Rd = !cmp.eq($Rs, #$s8)",
|
|
[(set (i32 IntRegs:$Rd),
|
|
(i32 (zext (i1 (setne (i32 IntRegs:$Rs),
|
|
s8ExtPred:$s8)))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=cmp.eq(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def V4_A4_rcmpeq : ALU32_ri<(outs IntRegs:$Rd),
|
|
(ins IntRegs:$Rs, IntRegs:$Rt),
|
|
"$Rd = cmp.eq($Rs, $Rt)",
|
|
[(set (i32 IntRegs:$Rd),
|
|
(i32 (zext (i1 (seteq (i32 IntRegs:$Rs),
|
|
IntRegs:$Rt)))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=cmp.ne(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def V4_A4_rcmpneq : ALU32_ri<(outs IntRegs:$Rd),
|
|
(ins IntRegs:$Rs, IntRegs:$Rt),
|
|
"$Rd = !cmp.eq($Rs, $Rt)",
|
|
[(set (i32 IntRegs:$Rd),
|
|
(i32 (zext (i1 (setne (i32 IntRegs:$Rs),
|
|
IntRegs:$Rt)))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// ALU32 -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// ALU32/PERM +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Combine
|
|
// Rdd=combine(Rs, #s8)
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8,
|
|
neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def COMBINE_rI_V4 : ALU32_ri<(outs DoubleRegs:$dst),
|
|
(ins IntRegs:$src1, s8Ext:$src2),
|
|
"$dst = combine($src1, #$src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rdd=combine(#s8, Rs)
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 1, opExtentBits = 8,
|
|
neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def COMBINE_Ir_V4 : ALU32_ir<(outs DoubleRegs:$dst),
|
|
(ins s8Ext:$src1, IntRegs:$src2),
|
|
"$dst = combine(#$src1, $src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def HexagonWrapperCombineRI_V4 :
|
|
SDNode<"HexagonISD::WrapperCombineRI_V4", SDTHexagonI64I32I32>;
|
|
def HexagonWrapperCombineIR_V4 :
|
|
SDNode<"HexagonISD::WrapperCombineIR_V4", SDTHexagonI64I32I32>;
|
|
|
|
def : Pat <(HexagonWrapperCombineRI_V4 IntRegs:$r, s8ExtPred:$i),
|
|
(COMBINE_rI_V4 IntRegs:$r, s8ExtPred:$i)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(HexagonWrapperCombineIR_V4 s8ExtPred:$i, IntRegs:$r),
|
|
(COMBINE_Ir_V4 s8ExtPred:$i, IntRegs:$r)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 6,
|
|
neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def COMBINE_iI_V4 : ALU32_ii<(outs DoubleRegs:$dst),
|
|
(ins s8Imm:$src1, u6Ext:$src2),
|
|
"$dst = combine(#$src1, #$src2)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// ALU32/PERM +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// LD +
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// These absolute set addressing mode instructions accept immediate as
|
|
// an operand. We have duplicated these patterns to take global address.
|
|
|
|
let isExtended = 1, opExtendable = 2, neverHasSideEffects = 1,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDrid_abs_setimm_V4 : LDInst2<(outs DoubleRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memd($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memb(Re=#U6)
|
|
def LDrib_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memb($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memh(Re=#U6)
|
|
def LDrih_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memh($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memub(Re=#U6)
|
|
def LDriub_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memub($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memuh(Re=#U6)
|
|
def LDriuh_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memuh($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memw(Re=#U6)
|
|
def LDriw_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins u0AlwaysExt:$addr),
|
|
"$dst1 = memw($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Following patterns are defined for absolute set addressing mode
|
|
// instruction which take global address as operand.
|
|
let isExtended = 1, opExtendable = 2, neverHasSideEffects = 1,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDrid_abs_set_V4 : LDInst2<(outs DoubleRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memd($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memb(Re=#U6)
|
|
def LDrib_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memb($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memh(Re=#U6)
|
|
def LDrih_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memh($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memub(Re=#U6)
|
|
def LDriub_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memub($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memuh(Re=#U6)
|
|
def LDriuh_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memuh($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=memw(Re=#U6)
|
|
def LDriw_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2),
|
|
(ins globaladdressExt:$addr),
|
|
"$dst1 = memw($dst2=##$addr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// multiclass for load instructions with base + register offset
|
|
// addressing mode
|
|
multiclass ld_idxd_shl_pbase<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME : LDInst2<(outs RC:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$offset),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#"$dst = "#mnemonic#"($src2+$src3<<#$offset)",
|
|
[]>, Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ld_idxd_shl_pred<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ld_idxd_shl_pbase<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ld_idxd_shl_pbase<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let neverHasSideEffects = 1 in
|
|
multiclass ld_idxd_shl<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in {
|
|
let isPredicable = 1 in
|
|
def NAME#_V4 : LDInst2<(outs RC:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2, u2Imm:$offset),
|
|
"$dst = "#mnemonic#"($src1+$src2<<#$offset)",
|
|
[]>, Requires<[HasV4T]>;
|
|
|
|
let isPredicated = 1 in {
|
|
defm Pt_V4 : ld_idxd_shl_pred<mnemonic, RC, 0 >;
|
|
defm NotPt_V4 : ld_idxd_shl_pred<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = BaseRegOffset in {
|
|
defm LDrib_indexed_shl: ld_idxd_shl<"memb", "LDrib", IntRegs>, AddrModeRel;
|
|
defm LDriub_indexed_shl: ld_idxd_shl<"memub", "LDriub", IntRegs>, AddrModeRel;
|
|
defm LDrih_indexed_shl: ld_idxd_shl<"memh", "LDrih", IntRegs>, AddrModeRel;
|
|
defm LDriuh_indexed_shl: ld_idxd_shl<"memuh", "LDriuh", IntRegs>, AddrModeRel;
|
|
defm LDriw_indexed_shl: ld_idxd_shl<"memw", "LDriw", IntRegs>, AddrModeRel;
|
|
defm LDrid_indexed_shl: ld_idxd_shl<"memd", "LDrid", DoubleRegs>, AddrModeRel;
|
|
}
|
|
|
|
// 'def pats' for load instructions with base + register offset and non-zero
|
|
// immediate value. Immediate value is used to left-shift the second
|
|
// register operand.
|
|
let AddedComplexity = 40 in {
|
|
def : Pat <(i32 (sextloadi8 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDrib_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (zextloadi8 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDriub_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (extloadi8 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDriub_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (sextloadi16 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDrih_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (zextloadi16 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDriuh_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (extloadi16 (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDriuh_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (load (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDriw_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i64 (load (add IntRegs:$src1,
|
|
(shl IntRegs:$src2, u2ImmPred:$offset)))),
|
|
(LDrid_indexed_shl_V4 IntRegs:$src1,
|
|
IntRegs:$src2, u2ImmPred:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
|
|
// 'def pats' for load instruction base + register offset and
|
|
// zero immediate value.
|
|
let AddedComplexity = 10 in {
|
|
def : Pat <(i64 (load (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDrid_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (sextloadi8 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDrib_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (zextloadi8 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (extloadi8 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (sextloadi16 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDrih_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (zextloadi16 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (extloadi16 (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(i32 (load (add IntRegs:$src1, IntRegs:$src2))),
|
|
(LDriw_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDd_GP_V4 : LDInst2<(outs DoubleRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memd(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rtt=memd(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDd_GP_cPt_V4 : LDInst2<(outs DoubleRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memd(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rtt=memd(##global)
|
|
def LDd_GP_cNotPt_V4 : LDInst2<(outs DoubleRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memd(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rtt=memd(##global)
|
|
def LDd_GP_cdnPt_V4 : LDInst2<(outs DoubleRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memd(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rtt=memd(##global)
|
|
def LDd_GP_cdnNotPt_V4 : LDInst2<(outs DoubleRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memd(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDb_GP_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memb(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memb(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDb_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memb(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memb(##global)
|
|
def LDb_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memb(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memb(##global)
|
|
def LDb_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memb(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memb(##global)
|
|
def LDb_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memb(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDub_GP_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memub(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memub(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDub_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memub(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rt=memub(##global)
|
|
def LDub_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memub(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memub(##global)
|
|
def LDub_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memub(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rt=memub(##global)
|
|
def LDub_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memub(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDh_GP_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memh(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memh(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDh_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memh(##global)
|
|
def LDh_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memh(##global)
|
|
def LDh_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memh(##global)
|
|
def LDh_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDuh_GP_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memuh(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memuh(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDuh_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memuh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memuh(##global)
|
|
def LDuh_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memuh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memuh(##global)
|
|
def LDuh_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memuh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) Rt=memuh(##global)
|
|
def LDuh_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memuh(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in
|
|
def LDw_GP_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins globaladdress:$global),
|
|
"$dst=memw(#$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memw(##global)
|
|
let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2,
|
|
validSubTargets = HasV4SubT in {
|
|
def LDw_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1) $dst=memw(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rt=memw(##global)
|
|
def LDw_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1) $dst=memw(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) Rt=memw(##global)
|
|
def LDw_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if ($src1.new) $dst=memw(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// if (!Pv) Rt=memw(##global)
|
|
def LDw_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$global),
|
|
"if (!$src1.new) $dst=memw(##$global)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
|
|
def : Pat <(atomic_load_64 (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(i64 (LDd_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(atomic_load_32 (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(i32 (LDw_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(atomic_load_16 (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(i32 (LDuh_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(atomic_load_8 (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(i32 (LDub_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memw(#foo + 0)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i64 (load (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i64 (LDd_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from Pd = load(globaladdress) -> Rd = memb(globaladdress), Pd = Rd
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i1 (load (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i1 (TFR_PdRs (i32 (LDb_GP_V4 tglobaladdr:$global))))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// When the Interprocedural Global Variable optimizer realizes that a certain
|
|
// global variable takes only two constant values, it shrinks the global to
|
|
// a boolean. Catch those loads here in the following 3 patterns.
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (extloadi1 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDb_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (sextloadi1 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDb_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memb(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (extloadi8 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDb_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memb(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (sextloadi8 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDb_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (zextloadi1 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDub_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memub(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (zextloadi8 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDub_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memh(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (extloadi16 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDh_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memh(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (sextloadi16 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDh_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memuh(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (zextloadi16 (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDuh_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress) -> memw(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(i32 (load (HexagonCONST32_GP tglobaladdr:$global))),
|
|
(i32 (LDw_GP_V4 tglobaladdr:$global))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// zext i1->i64
|
|
def : Pat <(i64 (zext (i1 PredRegs:$src1))),
|
|
(i64 (COMBINE_Ir_V4 0, (MUX_ii (i1 PredRegs:$src1), 1, 0)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// zext i32->i64
|
|
def : Pat <(i64 (zext (i32 IntRegs:$src1))),
|
|
(i64 (COMBINE_Ir_V4 0, (i32 IntRegs:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
// zext i8->i64
|
|
def: Pat <(i64 (zextloadi8 ADDRriS11_0:$src1)),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriub ADDRriS11_0:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 20 in
|
|
def: Pat <(i64 (zextloadi8 (add (i32 IntRegs:$src1),
|
|
s11_0ExtPred:$offset))),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriub_indexed IntRegs:$src1,
|
|
s11_0ExtPred:$offset)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// zext i16->i64
|
|
def: Pat <(i64 (zextloadi16 ADDRriS11_1:$src1)),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriuh ADDRriS11_1:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 20 in
|
|
def: Pat <(i64 (zextloadi16 (add (i32 IntRegs:$src1),
|
|
s11_1ExtPred:$offset))),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriuh_indexed IntRegs:$src1,
|
|
s11_1ExtPred:$offset)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// anyext i16->i64
|
|
def: Pat <(i64 (extloadi16 ADDRriS11_2:$src1)),
|
|
(i64 (COMBINE_Ir_V4 0, (LDrih ADDRriS11_2:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 20 in
|
|
def: Pat <(i64 (extloadi16 (add (i32 IntRegs:$src1),
|
|
s11_1ExtPred:$offset))),
|
|
(i64 (COMBINE_Ir_V4 0, (LDrih_indexed IntRegs:$src1,
|
|
s11_1ExtPred:$offset)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// zext i32->i64
|
|
def: Pat <(i64 (zextloadi32 ADDRriS11_2:$src1)),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriw ADDRriS11_2:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 100 in
|
|
def: Pat <(i64 (zextloadi32 (i32 (add IntRegs:$src1, s11_2ExtPred:$offset)))),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriw_indexed IntRegs:$src1,
|
|
s11_2ExtPred:$offset)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// anyext i32->i64
|
|
def: Pat <(i64 (extloadi32 ADDRriS11_2:$src1)),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriw ADDRriS11_2:$src1)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 100 in
|
|
def: Pat <(i64 (extloadi32 (i32 (add IntRegs:$src1, s11_2ExtPred:$offset)))),
|
|
(i64 (COMBINE_Ir_V4 0, (LDriw_indexed IntRegs:$src1,
|
|
s11_2ExtPred:$offset)))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// LD -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// ST +
|
|
//===----------------------------------------------------------------------===//
|
|
///
|
|
/// Assumptions::: ****** DO NOT IGNORE ********
|
|
/// 1. Make sure that in post increment store, the zero'th operand is always the
|
|
/// post increment operand.
|
|
/// 2. Make sure that the store value operand(Rt/Rtt) in a store is always the
|
|
/// last operand.
|
|
///
|
|
|
|
// memd(Re=#U)=Rtt
|
|
let isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in {
|
|
def STrid_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins DoubleRegs:$src1, u0AlwaysExt:$src2),
|
|
"memd($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memb(Re=#U)=Rs
|
|
def STrib_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, u0AlwaysExt:$src2),
|
|
"memb($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Re=#U)=Rs
|
|
def STrih_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, u0AlwaysExt:$src2),
|
|
"memh($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Re=#U)=Rs
|
|
def STriw_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, u0AlwaysExt:$src2),
|
|
"memw($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// memd(Re=#U)=Rtt
|
|
let isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in {
|
|
def STrid_abs_set_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins DoubleRegs:$src1, globaladdressExt:$src2),
|
|
"memd($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memb(Re=#U)=Rs
|
|
def STrib_abs_set_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, globaladdressExt:$src2),
|
|
"memb($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Re=#U)=Rs
|
|
def STrih_abs_set_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, globaladdressExt:$src2),
|
|
"memh($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Re=#U)=Rs
|
|
def STriw_abs_set_V4 : STInst2<(outs IntRegs:$dst1),
|
|
(ins IntRegs:$src1, globaladdressExt:$src2),
|
|
"memw($dst1=##$src2) = $src1",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// multiclass for store instructions with base + register offset addressing
|
|
// mode
|
|
multiclass ST_Idxd_shl_Pbase<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME : STInst2<(outs),
|
|
(ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$src4,
|
|
RC:$src5),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($src2+$src3<<#$src4) = $src5",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Idxd_shl_Pred<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Idxd_shl_Pbase<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Idxd_shl_Pbase<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let isNVStorable = 1 in
|
|
multiclass ST_Idxd_shl<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in {
|
|
let isPredicable = 1 in
|
|
def NAME#_V4 : STInst2<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, u2Imm:$src3, RC:$src4),
|
|
mnemonic#"($src1+$src2<<#$src3) = $src4",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isPredicated = 1 in {
|
|
defm Pt_V4 : ST_Idxd_shl_Pred<mnemonic, RC, 0 >;
|
|
defm NotPt_V4 : ST_Idxd_shl_Pred<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
// multiclass for new-value store instructions with base + register offset
|
|
// addressing mode.
|
|
multiclass ST_Idxd_shl_Pbase_nv<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$src4,
|
|
RC:$src5),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($src2+$src3<<#$src4) = $src5.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Idxd_shl_Pred_nv<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Idxd_shl_Pbase_nv<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Idxd_shl_Pbase_nv<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let mayStore = 1, isNVStore = 1 in
|
|
multiclass ST_Idxd_shl_nv<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in {
|
|
let isPredicable = 1 in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, u2Imm:$src3, RC:$src4),
|
|
mnemonic#"($src1+$src2<<#$src3) = $src4.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isPredicated = 1 in {
|
|
defm Pt : ST_Idxd_shl_Pred_nv<mnemonic, RC, 0 >;
|
|
defm NotPt : ST_Idxd_shl_Pred_nv<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = BaseRegOffset, neverHasSideEffects = 1,
|
|
validSubTargets = HasV4SubT in {
|
|
defm STrib_indexed_shl: ST_Idxd_shl<"memb", "STrib", IntRegs>,
|
|
ST_Idxd_shl_nv<"memb", "STrib", IntRegs>, AddrModeRel;
|
|
|
|
defm STrih_indexed_shl: ST_Idxd_shl<"memh", "STrih", IntRegs>,
|
|
ST_Idxd_shl_nv<"memh", "STrih", IntRegs>, AddrModeRel;
|
|
|
|
defm STriw_indexed_shl: ST_Idxd_shl<"memw", "STriw", IntRegs>,
|
|
ST_Idxd_shl_nv<"memw", "STriw", IntRegs>, AddrModeRel;
|
|
|
|
let isNVStorable = 0 in
|
|
defm STrid_indexed_shl: ST_Idxd_shl<"memd", "STrid", DoubleRegs>, AddrModeRel;
|
|
}
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 10 in {
|
|
def : Pat<(truncstorei8 (i32 IntRegs:$src4),
|
|
(add IntRegs:$src1, (shl IntRegs:$src2,
|
|
u2ImmPred:$src3))),
|
|
(STrib_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2,
|
|
u2ImmPred:$src3, IntRegs:$src4)>;
|
|
|
|
def : Pat<(truncstorei16 (i32 IntRegs:$src4),
|
|
(add IntRegs:$src1, (shl IntRegs:$src2,
|
|
u2ImmPred:$src3))),
|
|
(STrih_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2,
|
|
u2ImmPred:$src3, IntRegs:$src4)>;
|
|
|
|
def : Pat<(store (i32 IntRegs:$src4),
|
|
(add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))),
|
|
(STriw_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2,
|
|
u2ImmPred:$src3, IntRegs:$src4)>;
|
|
|
|
def : Pat<(store (i64 DoubleRegs:$src4),
|
|
(add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))),
|
|
(STrid_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2,
|
|
u2ImmPred:$src3, DoubleRegs:$src4)>;
|
|
}
|
|
|
|
// memd(Ru<<#u2+#U6)=Rtt
|
|
let isExtended = 1, opExtendable = 2, AddedComplexity = 10,
|
|
validSubTargets = HasV4SubT in
|
|
def STrid_shl_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, DoubleRegs:$src4),
|
|
"memd($src1<<#$src2+#$src3) = $src4",
|
|
[(store (i64 DoubleRegs:$src4),
|
|
(add (shl (i32 IntRegs:$src1), u2ImmPred:$src2),
|
|
u0AlwaysExtPred:$src3))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memd(Rx++#s4:3)=Rtt
|
|
// memd(Rx++#s4:3:circ(Mu))=Rtt
|
|
// memd(Rx++I:circ(Mu))=Rtt
|
|
// memd(Rx++Mu)=Rtt
|
|
// memd(Rx++Mu:brev)=Rtt
|
|
// memd(gp+#u16:3)=Rtt
|
|
|
|
// Store doubleword conditionally.
|
|
// if ([!]Pv[.new]) memd(#u6)=Rtt
|
|
// TODO: needs to be implemented.
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// multiclass for store instructions with base + immediate offset
|
|
// addressing mode and immediate stored value.
|
|
// mem[bhw](Rx++#s4:3)=#s8
|
|
// if ([!]Pv[.new]) mem[bhw](Rx++#s4:3)=#s6
|
|
//===----------------------------------------------------------------------===//
|
|
multiclass ST_Imm_Pbase<string mnemonic, Operand OffsetOp, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME : STInst2<(outs),
|
|
(ins PredRegs:$src1, IntRegs:$src2, OffsetOp:$src3, s6Ext:$src4),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($src2+#$src3) = #$src4",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Imm_Pred<string mnemonic, Operand OffsetOp, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Imm_Pbase<mnemonic, OffsetOp, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Imm_Pbase<mnemonic, OffsetOp, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let isExtendable = 1, isExtentSigned = 1, neverHasSideEffects = 1 in
|
|
multiclass ST_Imm<string mnemonic, string CextOp, Operand OffsetOp> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_imm in {
|
|
let opExtendable = 2, opExtentBits = 8, isPredicable = 1 in
|
|
def NAME#_V4 : STInst2<(outs),
|
|
(ins IntRegs:$src1, OffsetOp:$src2, s8Ext:$src3),
|
|
mnemonic#"($src1+#$src2) = #$src3",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 3, opExtentBits = 6, isPredicated = 1 in {
|
|
defm Pt_V4 : ST_Imm_Pred<mnemonic, OffsetOp, 0>;
|
|
defm NotPt_V4 : ST_Imm_Pred<mnemonic, OffsetOp, 1 >;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = BaseImmOffset, InputType = "imm",
|
|
validSubTargets = HasV4SubT in {
|
|
defm STrib_imm : ST_Imm<"memb", "STrib", u6_0Imm>, ImmRegRel, PredNewRel;
|
|
defm STrih_imm : ST_Imm<"memh", "STrih", u6_1Imm>, ImmRegRel, PredNewRel;
|
|
defm STriw_imm : ST_Imm<"memw", "STriw", u6_2Imm>, ImmRegRel, PredNewRel;
|
|
}
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 10 in {
|
|
def: Pat<(truncstorei8 s8ExtPred:$src3, (add IntRegs:$src1, u6_0ImmPred:$src2)),
|
|
(STrib_imm_V4 IntRegs:$src1, u6_0ImmPred:$src2, s8ExtPred:$src3)>;
|
|
|
|
def: Pat<(truncstorei16 s8ExtPred:$src3, (add IntRegs:$src1,
|
|
u6_1ImmPred:$src2)),
|
|
(STrih_imm_V4 IntRegs:$src1, u6_1ImmPred:$src2, s8ExtPred:$src3)>;
|
|
|
|
def: Pat<(store s8ExtPred:$src3, (add IntRegs:$src1, u6_2ImmPred:$src2)),
|
|
(STriw_imm_V4 IntRegs:$src1, u6_2ImmPred:$src2, s8ExtPred:$src3)>;
|
|
}
|
|
|
|
let AddedComplexity = 6 in
|
|
def : Pat <(truncstorei8 s8ExtPred:$src2, (i32 IntRegs:$src1)),
|
|
(STrib_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memb(Ru<<#u2+#U6)=Rt
|
|
let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STrib_shl_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memb($src1<<#$src2+#$src3) = $src4",
|
|
[(truncstorei8 (i32 IntRegs:$src4),
|
|
(add (shl (i32 IntRegs:$src1), u2ImmPred:$src2),
|
|
u0AlwaysExtPred:$src3))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memb(Rx++#s4:0:circ(Mu))=Rt
|
|
// memb(Rx++I:circ(Mu))=Rt
|
|
// memb(Rx++Mu)=Rt
|
|
// memb(Rx++Mu:brev)=Rt
|
|
// memb(gp+#u16:0)=Rt
|
|
|
|
|
|
// Store halfword.
|
|
// TODO: needs to be implemented
|
|
// memh(Re=#U6)=Rt.H
|
|
// memh(Rs+#s11:1)=Rt.H
|
|
let AddedComplexity = 6 in
|
|
def : Pat <(truncstorei16 s8ExtPred:$src2, (i32 IntRegs:$src1)),
|
|
(STrih_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Rs+Ru<<#u2)=Rt.H
|
|
// TODO: needs to be implemented.
|
|
|
|
// memh(Ru<<#u2+#U6)=Rt.H
|
|
// memh(Ru<<#u2+#U6)=Rt
|
|
let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STrih_shl_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memh($src1<<#$src2+#$src3) = $src4",
|
|
[(truncstorei16 (i32 IntRegs:$src4),
|
|
(add (shl (i32 IntRegs:$src1), u2ImmPred:$src2),
|
|
u0AlwaysExtPred:$src3))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Rx++#s4:1:circ(Mu))=Rt.H
|
|
// memh(Rx++#s4:1:circ(Mu))=Rt
|
|
// memh(Rx++I:circ(Mu))=Rt.H
|
|
// memh(Rx++I:circ(Mu))=Rt
|
|
// memh(Rx++Mu)=Rt.H
|
|
// memh(Rx++Mu)=Rt
|
|
// memh(Rx++Mu:brev)=Rt.H
|
|
// memh(Rx++Mu:brev)=Rt
|
|
// memh(gp+#u16:1)=Rt
|
|
// if ([!]Pv[.new]) memh(#u6)=Rt.H
|
|
// if ([!]Pv[.new]) memh(#u6)=Rt
|
|
|
|
|
|
// if ([!]Pv[.new]) memh(Rs+#u6:1)=Rt.H
|
|
// TODO: needs to be implemented.
|
|
|
|
// if ([!]Pv[.new]) memh(Rx++#s4:1)=Rt.H
|
|
// TODO: Needs to be implemented.
|
|
|
|
// Store word.
|
|
// memw(Re=#U6)=Rt
|
|
// TODO: Needs to be implemented.
|
|
|
|
// Store predicate:
|
|
let neverHasSideEffects = 1 in
|
|
def STriw_pred_V4 : STInst2<(outs),
|
|
(ins MEMri:$addr, PredRegs:$src1),
|
|
"Error; should not emit",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 6 in
|
|
def : Pat <(store s8ExtPred:$src2, (i32 IntRegs:$src1)),
|
|
(STriw_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Ru<<#u2+#U6)=Rt
|
|
let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STriw_shl_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memw($src1<<#$src2+#$src3) = $src4",
|
|
[(store (i32 IntRegs:$src4),
|
|
(add (shl (i32 IntRegs:$src1), u2ImmPred:$src2),
|
|
u0AlwaysExtPred:$src3))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Rx++#s4:2)=Rt
|
|
// memw(Rx++#s4:2:circ(Mu))=Rt
|
|
// memw(Rx++I:circ(Mu))=Rt
|
|
// memw(Rx++Mu)=Rt
|
|
// memw(Rx++Mu:brev)=Rt
|
|
// memw(gp+#u16:2)=Rt
|
|
|
|
|
|
// memd(#global)=Rtt
|
|
let isPredicable = 1, mayStore = 1, neverHasSideEffects = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STd_GP_V4 : STInst2<(outs),
|
|
(ins globaladdress:$global, DoubleRegs:$src),
|
|
"memd(#$global) = $src",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memd(##global) = Rtt
|
|
let mayStore = 1, neverHasSideEffects = 1, isPredicated = 1,
|
|
isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in {
|
|
def STd_GP_cPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2),
|
|
"if ($src1) memd(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memd(##global) = Rtt
|
|
def STd_GP_cNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2),
|
|
"if (!$src1) memd(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memd(##global) = Rtt
|
|
def STd_GP_cdnPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2),
|
|
"if ($src1.new) memd(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memd(##global) = Rtt
|
|
def STd_GP_cdnNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2),
|
|
"if (!$src1.new) memd(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// memb(#global)=Rt
|
|
let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STb_GP_V4 : STInst2<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memb(#$global) = $src",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memb(##global) = Rt
|
|
let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1,
|
|
isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in {
|
|
def STb_GP_cPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memb(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memb(##global) = Rt
|
|
def STb_GP_cNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memb(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memb(##global) = Rt
|
|
def STb_GP_cdnPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memb(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memb(##global) = Rt
|
|
def STb_GP_cdnNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memb(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// memh(#global)=Rt
|
|
let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STh_GP_V4 : STInst2<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memh(#$global) = $src",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memh(##global) = Rt
|
|
let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1,
|
|
isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in {
|
|
def STh_GP_cPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memh(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memh(##global) = Rt
|
|
def STh_GP_cNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memh(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memh(##global) = Rt
|
|
def STh_GP_cdnPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memh(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memh(##global) = Rt
|
|
def STh_GP_cdnNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memh(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// memw(#global)=Rt
|
|
let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STw_GP_V4 : STInst2<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memw(#$global) = $src",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memw(##global) = Rt
|
|
let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1,
|
|
isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in {
|
|
def STw_GP_cPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memw(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memw(##global) = Rt
|
|
def STw_GP_cNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memw(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memw(##global) = Rt
|
|
def STw_GP_cdnPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memw(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memw(##global) = Rt
|
|
def STw_GP_cdnNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memw(##$global) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// 64 bit atomic store
|
|
def : Pat <(atomic_store_64 (HexagonCONST32_GP tglobaladdr:$global),
|
|
(i64 DoubleRegs:$src1)),
|
|
(STd_GP_V4 tglobaladdr:$global, (i64 DoubleRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress) -> memd(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat <(store (i64 DoubleRegs:$src1),
|
|
(HexagonCONST32_GP tglobaladdr:$global)),
|
|
(STd_GP_V4 tglobaladdr:$global, (i64 DoubleRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// 8 bit atomic store
|
|
def : Pat < (atomic_store_8 (HexagonCONST32_GP tglobaladdr:$global),
|
|
(i32 IntRegs:$src1)),
|
|
(STb_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress) -> memb(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(truncstorei8 (i32 IntRegs:$src1),
|
|
(HexagonCONST32_GP tglobaladdr:$global)),
|
|
(STb_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from "i1 = constant<-1>; memw(CONST32(#foo)) = i1"
|
|
// to "r0 = 1; memw(#foo) = r0"
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(store (i1 -1), (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(STb_GP_V4 tglobaladdr:$global, (TFRI 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_store_16 (HexagonCONST32_GP tglobaladdr:$global),
|
|
(i32 IntRegs:$src1)),
|
|
(STh_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress) -> memh(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(truncstorei16 (i32 IntRegs:$src1),
|
|
(HexagonCONST32_GP tglobaladdr:$global)),
|
|
(STh_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// 32 bit atomic store
|
|
def : Pat<(atomic_store_32 (HexagonCONST32_GP tglobaladdr:$global),
|
|
(i32 IntRegs:$src1)),
|
|
(STw_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress) -> memw(#foo)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(store (i32 IntRegs:$src1), (HexagonCONST32_GP tglobaladdr:$global)),
|
|
(STw_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//===----------------------------------------------------------------------===
|
|
// ST -
|
|
//===----------------------------------------------------------------------===
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// NV/ST +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// multiclass for new-value store instructions with base + immediate offset.
|
|
//
|
|
multiclass ST_Idxd_Pbase_nv<string mnemonic, RegisterClass RC,
|
|
Operand predImmOp, bit isNot, bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, IntRegs:$src2, predImmOp:$src3, RC: $src4),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($src2+#$src3) = $src4.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Idxd_Pred_nv<string mnemonic, RegisterClass RC, Operand predImmOp,
|
|
bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Idxd_Pbase_nv<mnemonic, RC, predImmOp, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Idxd_Pbase_nv<mnemonic, RC, predImmOp, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let mayStore = 1, isNVStore = 1, neverHasSideEffects = 1, isExtendable = 1 in
|
|
multiclass ST_Idxd_nv<string mnemonic, string CextOp, RegisterClass RC,
|
|
Operand ImmOp, Operand predImmOp, bits<5> ImmBits,
|
|
bits<5> PredImmBits> {
|
|
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed in {
|
|
let opExtendable = 1, isExtentSigned = 1, opExtentBits = ImmBits,
|
|
isPredicable = 1 in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, ImmOp:$src2, RC:$src3),
|
|
mnemonic#"($src1+#$src2) = $src3.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 2, isExtentSigned = 0, opExtentBits = PredImmBits,
|
|
isPredicated = 1 in {
|
|
defm Pt : ST_Idxd_Pred_nv<mnemonic, RC, predImmOp, 0>;
|
|
defm NotPt : ST_Idxd_Pred_nv<mnemonic, RC, predImmOp, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = BaseImmOffset, validSubTargets = HasV4SubT in {
|
|
defm STrib_indexed: ST_Idxd_nv<"memb", "STrib", IntRegs, s11_0Ext,
|
|
u6_0Ext, 11, 6>, AddrModeRel;
|
|
defm STrih_indexed: ST_Idxd_nv<"memh", "STrih", IntRegs, s11_1Ext,
|
|
u6_1Ext, 12, 7>, AddrModeRel;
|
|
defm STriw_indexed: ST_Idxd_nv<"memw", "STriw", IntRegs, s11_2Ext,
|
|
u6_2Ext, 13, 8>, AddrModeRel;
|
|
}
|
|
|
|
// multiclass for new-value store instructions with base + immediate offset.
|
|
// and MEMri operand.
|
|
multiclass ST_MEMri_Pbase_nv<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, MEMri:$addr, RC: $src2),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($addr) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_MEMri_Pred_nv<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_MEMri_Pbase_nv<mnemonic, RC, PredNot, 0>;
|
|
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_MEMri_Pbase_nv<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let mayStore = 1, isNVStore = 1, isExtendable = 1, neverHasSideEffects = 1 in
|
|
multiclass ST_MEMri_nv<string mnemonic, string CextOp, RegisterClass RC,
|
|
bits<5> ImmBits, bits<5> PredImmBits> {
|
|
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp in {
|
|
let opExtendable = 1, isExtentSigned = 1, opExtentBits = ImmBits,
|
|
isPredicable = 1 in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins MEMri:$addr, RC:$src),
|
|
mnemonic#"($addr) = $src.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 2, isExtentSigned = 0, opExtentBits = PredImmBits,
|
|
neverHasSideEffects = 1, isPredicated = 1 in {
|
|
defm Pt : ST_MEMri_Pred_nv<mnemonic, RC, 0>;
|
|
defm NotPt : ST_MEMri_Pred_nv<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = BaseImmOffset, isMEMri = "true", validSubTargets = HasV4SubT,
|
|
mayStore = 1 in {
|
|
defm STrib: ST_MEMri_nv<"memb", "STrib", IntRegs, 11, 6>, AddrModeRel;
|
|
defm STrih: ST_MEMri_nv<"memh", "STrih", IntRegs, 12, 7>, AddrModeRel;
|
|
defm STriw: ST_MEMri_nv<"memw", "STriw", IntRegs, 13, 8>, AddrModeRel;
|
|
}
|
|
|
|
// memb(Ru<<#u2+#U6)=Nt.new
|
|
let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10,
|
|
isNVStore = 1, validSubTargets = HasV4SubT in
|
|
def STrib_shl_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memb($src1<<#$src2+#$src3) = $src4.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// Post increment store
|
|
// mem[bhwd](Rx++#s4:[0123])=Nt.new
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
multiclass ST_PostInc_Pbase_nv<string mnemonic, RegisterClass RC, Operand ImmOp,
|
|
bit isNot, bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_nv_V4 : NVInstPI_V4<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, IntRegs:$src2, ImmOp:$offset, RC:$src3),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"($src2++#$offset) = $src3.new",
|
|
[],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_PostInc_Pred_nv<string mnemonic, RegisterClass RC,
|
|
Operand ImmOp, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_PostInc_Pbase_nv<mnemonic, RC, ImmOp, PredNot, 0>;
|
|
// Predicate new
|
|
let Predicates = [HasV4T], validSubTargets = HasV4SubT in
|
|
defm _cdn#NAME : ST_PostInc_Pbase_nv<mnemonic, RC, ImmOp, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let hasCtrlDep = 1, isNVStore = 1, neverHasSideEffects = 1 in
|
|
multiclass ST_PostInc_nv<string mnemonic, string BaseOp, RegisterClass RC,
|
|
Operand ImmOp> {
|
|
|
|
let BaseOpcode = "POST_"#BaseOp in {
|
|
let isPredicable = 1 in
|
|
def NAME#_nv_V4 : NVInstPI_V4<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, ImmOp:$offset, RC:$src2),
|
|
mnemonic#"($src1++#$offset) = $src2.new",
|
|
[],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isPredicated = 1 in {
|
|
defm Pt : ST_PostInc_Pred_nv<mnemonic, RC, ImmOp, 0 >;
|
|
defm NotPt : ST_PostInc_Pred_nv<mnemonic, RC, ImmOp, 1 >;
|
|
}
|
|
}
|
|
}
|
|
|
|
let validSubTargets = HasV4SubT in {
|
|
defm POST_STbri: ST_PostInc_nv <"memb", "STrib", IntRegs, s4_0Imm>, AddrModeRel;
|
|
defm POST_SThri: ST_PostInc_nv <"memh", "STrih", IntRegs, s4_1Imm>, AddrModeRel;
|
|
defm POST_STwri: ST_PostInc_nv <"memw", "STriw", IntRegs, s4_2Imm>, AddrModeRel;
|
|
}
|
|
|
|
// memb(Rx++#s4:0:circ(Mu))=Nt.new
|
|
// memb(Rx++I:circ(Mu))=Nt.new
|
|
// memb(Rx++Mu)=Nt.new
|
|
// memb(Rx++Mu:brev)=Nt.new
|
|
|
|
// memb(#global)=Nt.new
|
|
let mayStore = 1, neverHasSideEffects = 1 in
|
|
def STb_GP_nv_V4 : NVInst_V4<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memb(#$global) = $src.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Ru<<#u2+#U6)=Nt.new
|
|
let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10,
|
|
isNVStore = 1, validSubTargets = HasV4SubT in
|
|
def STrih_shl_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memh($src1<<#$src2+#$src3) = $src4.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memh(Rx++#s4:1:circ(Mu))=Nt.new
|
|
// memh(Rx++I:circ(Mu))=Nt.new
|
|
// memh(Rx++Mu)=Nt.new
|
|
// memh(Rx++Mu:brev)=Nt.new
|
|
|
|
// memh(#global)=Nt.new
|
|
let mayStore = 1, neverHasSideEffects = 1 in
|
|
def STh_GP_nv_V4 : NVInst_V4<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memh(#$global) = $src.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Ru<<#u2+#U6)=Nt.new
|
|
let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10,
|
|
isNVStore = 1, validSubTargets = HasV4SubT in
|
|
def STriw_shl_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4),
|
|
"memw($src1<<#$src2+#$src3) = $src4.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// memw(Rx++#s4:2:circ(Mu))=Nt.new
|
|
// memw(Rx++I:circ(Mu))=Nt.new
|
|
// memw(Rx++Mu)=Nt.new
|
|
// memw(Rx++Mu:brev)=Nt.new
|
|
// memw(gp+#u16:2)=Nt.new
|
|
|
|
let mayStore = 1, neverHasSideEffects = 1, isNVStore = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def STw_GP_nv_V4 : NVInst_V4<(outs),
|
|
(ins globaladdress:$global, IntRegs:$src),
|
|
"memw(#$global) = $src.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memb(##global) = Rt
|
|
let mayStore = 1, neverHasSideEffects = 1, isNVStore = 1,
|
|
isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in {
|
|
def STb_GP_cPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memb(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memb(##global) = Rt
|
|
def STb_GP_cNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memb(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memb(##global) = Rt
|
|
def STb_GP_cdnPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memb(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memb(##global) = Rt
|
|
def STb_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memb(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memh(##global) = Rt
|
|
def STh_GP_cPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memh(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memh(##global) = Rt
|
|
def STh_GP_cNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memh(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memh(##global) = Rt
|
|
def STh_GP_cdnPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memh(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memh(##global) = Rt
|
|
def STh_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memh(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memw(##global) = Rt
|
|
def STw_GP_cPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1) memw(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memw(##global) = Rt
|
|
def STw_GP_cNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1) memw(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (Pv) memw(##global) = Rt
|
|
def STw_GP_cdnPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if ($src1.new) memw(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// if (!Pv) memw(##global) = Rt
|
|
def STw_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2),
|
|
"if (!$src1.new) memw(##$global) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// NV/ST -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// NV/J +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
multiclass NVJ_type_basic_reg<string NotStr, string OpcStr, string TakenStr> {
|
|
def _ie_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, $src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, $src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass NVJ_type_basic_2ndDotNew<string NotStr, string OpcStr,
|
|
string TakenStr> {
|
|
def _ie_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1, $src2.new)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1, $src2.new)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass NVJ_type_basic_imm<string NotStr, string OpcStr, string TakenStr> {
|
|
def _ie_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u5Imm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u5Imm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass NVJ_type_basic_neg<string NotStr, string OpcStr, string TakenStr> {
|
|
def _ie_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, nOneImm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, nOneImm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass NVJ_type_basic_tstbit<string NotStr, string OpcStr,
|
|
string TakenStr> {
|
|
def _ie_nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u1Imm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _nv_V4 : NVInst_V4<(outs),
|
|
(ins IntRegs:$src1, u1Imm:$src2, brtarget:$offset),
|
|
!strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr,
|
|
!strconcat("($src1.new, #$src2)) jump:",
|
|
!strconcat(TakenStr, " $offset"))))),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Multiclass for regular dot new of Ist operand register.
|
|
multiclass NVJ_type_br_pred_reg<string NotStr, string OpcStr> {
|
|
defm Pt : NVJ_type_basic_reg<NotStr, OpcStr, "t">;
|
|
defm Pnt : NVJ_type_basic_reg<NotStr, OpcStr, "nt">;
|
|
}
|
|
|
|
// Multiclass for dot new of 2nd operand register.
|
|
multiclass NVJ_type_br_pred_2ndDotNew<string NotStr, string OpcStr> {
|
|
defm Pt : NVJ_type_basic_2ndDotNew<NotStr, OpcStr, "t">;
|
|
defm Pnt : NVJ_type_basic_2ndDotNew<NotStr, OpcStr, "nt">;
|
|
}
|
|
|
|
// Multiclass for 2nd operand immediate, including -1.
|
|
multiclass NVJ_type_br_pred_imm<string NotStr, string OpcStr> {
|
|
defm Pt : NVJ_type_basic_imm<NotStr, OpcStr, "t">;
|
|
defm Pnt : NVJ_type_basic_imm<NotStr, OpcStr, "nt">;
|
|
defm Ptneg : NVJ_type_basic_neg<NotStr, OpcStr, "t">;
|
|
defm Pntneg : NVJ_type_basic_neg<NotStr, OpcStr, "nt">;
|
|
}
|
|
|
|
// Multiclass for 2nd operand immediate, excluding -1.
|
|
multiclass NVJ_type_br_pred_imm_only<string NotStr, string OpcStr> {
|
|
defm Pt : NVJ_type_basic_imm<NotStr, OpcStr, "t">;
|
|
defm Pnt : NVJ_type_basic_imm<NotStr, OpcStr, "nt">;
|
|
}
|
|
|
|
// Multiclass for tstbit, where 2nd operand is always #0.
|
|
multiclass NVJ_type_br_pred_tstbit<string NotStr, string OpcStr> {
|
|
defm Pt : NVJ_type_basic_tstbit<NotStr, OpcStr, "t">;
|
|
defm Pnt : NVJ_type_basic_tstbit<NotStr, OpcStr, "nt">;
|
|
}
|
|
|
|
// Multiclass for GT.
|
|
multiclass NVJ_type_rr_ri<string OpcStr> {
|
|
defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>;
|
|
defm rr : NVJ_type_br_pred_reg<"", OpcStr>;
|
|
defm rrdnNot : NVJ_type_br_pred_2ndDotNew<"!", OpcStr>;
|
|
defm rrdn : NVJ_type_br_pred_2ndDotNew<"", OpcStr>;
|
|
defm riNot : NVJ_type_br_pred_imm<"!", OpcStr>;
|
|
defm ri : NVJ_type_br_pred_imm<"", OpcStr>;
|
|
}
|
|
|
|
// Multiclass for EQ.
|
|
multiclass NVJ_type_rr_ri_no_2ndDotNew<string OpcStr> {
|
|
defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>;
|
|
defm rr : NVJ_type_br_pred_reg<"", OpcStr>;
|
|
defm riNot : NVJ_type_br_pred_imm<"!", OpcStr>;
|
|
defm ri : NVJ_type_br_pred_imm<"", OpcStr>;
|
|
}
|
|
|
|
// Multiclass for GTU.
|
|
multiclass NVJ_type_rr_ri_no_nOne<string OpcStr> {
|
|
defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>;
|
|
defm rr : NVJ_type_br_pred_reg<"", OpcStr>;
|
|
defm rrdnNot : NVJ_type_br_pred_2ndDotNew<"!", OpcStr>;
|
|
defm rrdn : NVJ_type_br_pred_2ndDotNew<"", OpcStr>;
|
|
defm riNot : NVJ_type_br_pred_imm_only<"!", OpcStr>;
|
|
defm ri : NVJ_type_br_pred_imm_only<"", OpcStr>;
|
|
}
|
|
|
|
// Multiclass for tstbit.
|
|
multiclass NVJ_type_r0<string OpcStr> {
|
|
defm r0Not : NVJ_type_br_pred_tstbit<"!", OpcStr>;
|
|
defm r0 : NVJ_type_br_pred_tstbit<"", OpcStr>;
|
|
}
|
|
|
|
// Base Multiclass for New Value Jump.
|
|
multiclass NVJ_type {
|
|
defm GT : NVJ_type_rr_ri<"cmp.gt">;
|
|
defm EQ : NVJ_type_rr_ri_no_2ndDotNew<"cmp.eq">;
|
|
defm GTU : NVJ_type_rr_ri_no_nOne<"cmp.gtu">;
|
|
defm TSTBIT : NVJ_type_r0<"tstbit">;
|
|
}
|
|
|
|
let isBranch = 1, isTerminator=1, neverHasSideEffects = 1, Defs = [PC] in {
|
|
defm JMP_ : NVJ_type;
|
|
}
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// NV/J -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/ALU +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Add and accumulate.
|
|
// Rd=add(Rs,add(Ru,#s6))
|
|
let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDr_ADDri_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2, s6Ext:$src3),
|
|
"$dst = add($src1, add($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (i32 IntRegs:$src1), (add (i32 IntRegs:$src2),
|
|
s6_16ExtPred:$src3)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=add(Rs,sub(#s6,Ru))
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDr_SUBri_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, s6Ext:$src2, IntRegs:$src3),
|
|
"$dst = add($src1, sub(#$src2, $src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (i32 IntRegs:$src1), (sub s6_10ExtPred:$src2,
|
|
(i32 IntRegs:$src3))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Generates the same instruction as ADDr_SUBri_V4 but matches different
|
|
// pattern.
|
|
// Rd=add(Rs,sub(#s6,Ru))
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDri_SUBr_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, s6Ext:$src2, IntRegs:$src3),
|
|
"$dst = add($src1, sub(#$src2, $src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(sub (add (i32 IntRegs:$src1), s6_10ExtPred:$src2),
|
|
(i32 IntRegs:$src3)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// Add or subtract doublewords with carry.
|
|
//TODO:
|
|
// Rdd=add(Rss,Rtt,Px):carry
|
|
//TODO:
|
|
// Rdd=sub(Rss,Rtt,Px):carry
|
|
|
|
|
|
// Logical doublewords.
|
|
// Rdd=and(Rtt,~Rss)
|
|
let validSubTargets = HasV4SubT in
|
|
def ANDd_NOTd_V4 : MInst<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2),
|
|
"$dst = and($src1, ~$src2)",
|
|
[(set (i64 DoubleRegs:$dst), (and (i64 DoubleRegs:$src1),
|
|
(not (i64 DoubleRegs:$src2))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rdd=or(Rtt,~Rss)
|
|
let validSubTargets = HasV4SubT in
|
|
def ORd_NOTd_V4 : MInst<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2),
|
|
"$dst = or($src1, ~$src2)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(or (i64 DoubleRegs:$src1), (not (i64 DoubleRegs:$src2))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// Logical-logical doublewords.
|
|
// Rxx^=xor(Rss,Rtt)
|
|
let validSubTargets = HasV4SubT in
|
|
def XORd_XORdd: MInst_acc<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2, DoubleRegs:$src3),
|
|
"$dst ^= xor($src2, $src3)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(xor (i64 DoubleRegs:$src1), (xor (i64 DoubleRegs:$src2),
|
|
(i64 DoubleRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// Logical-logical words.
|
|
// Rx=or(Ru,and(Rx,#s10))
|
|
let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10,
|
|
validSubTargets = HasV4SubT in
|
|
def ORr_ANDri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3),
|
|
"$dst = or($src1, and($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
s10ExtPred:$src3)))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx[&|^]=and(Rs,Rt)
|
|
// Rx&=and(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ANDr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst &= and($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx|=and(Rs,Rt)
|
|
let validSubTargets = HasV4SubT, CextOpcode = "ORr_ANDr", InputType = "reg" in
|
|
def ORr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst |= and($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Rx^=and(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def XORr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst ^= and($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(xor (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx[&|^]=and(Rs,~Rt)
|
|
// Rx&=and(Rs,~Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ANDr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst &= and($src2, ~$src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(not (i32 IntRegs:$src3)))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx|=and(Rs,~Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ORr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst |= and($src2, ~$src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(not (i32 IntRegs:$src3)))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx^=and(Rs,~Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def XORr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst ^= and($src2, ~$src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(xor (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
(not (i32 IntRegs:$src3)))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx[&|^]=or(Rs,Rt)
|
|
// Rx&=or(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ANDr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst &= or($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (or (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx|=or(Rs,Rt)
|
|
let validSubTargets = HasV4SubT, CextOpcode = "ORr_ORr", InputType = "reg" in
|
|
def ORr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst |= or($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (or (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Rx^=or(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def XORr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst ^= or($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(xor (i32 IntRegs:$src1), (or (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx[&|^]=xor(Rs,Rt)
|
|
// Rx&=xor(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ANDr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst &= xor($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx|=xor(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def ORr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst |= xor($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx^=xor(Rs,Rt)
|
|
let validSubTargets = HasV4SubT in
|
|
def XORr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3),
|
|
"$dst ^= xor($src2, $src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx|=and(Rs,#s10)
|
|
let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10,
|
|
validSubTargets = HasV4SubT, CextOpcode = "ORr_ANDr", InputType = "imm" in
|
|
def ORr_ANDri2_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3),
|
|
"$dst |= and($src2, #$src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
s10ExtPred:$src3)))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Rx|=or(Rs,#s10)
|
|
let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10,
|
|
validSubTargets = HasV4SubT, CextOpcode = "ORr_ORr", InputType = "imm" in
|
|
def ORr_ORri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3),
|
|
"$dst |= or($src2, #$src3)",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2),
|
|
s10ExtPred:$src3)))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
|
|
// Modulo wrap
|
|
// Rd=modwrap(Rs,Rt)
|
|
// Round
|
|
// Rd=cround(Rs,#u5)
|
|
// Rd=cround(Rs,Rt)
|
|
// Rd=round(Rs,#u5)[:sat]
|
|
// Rd=round(Rs,Rt)[:sat]
|
|
// Vector reduce add unsigned halfwords
|
|
// Rd=vraddh(Rss,Rtt)
|
|
// Vector add bytes
|
|
// Rdd=vaddb(Rss,Rtt)
|
|
// Vector conditional negate
|
|
// Rdd=vcnegh(Rss,Rt)
|
|
// Rxx+=vrcnegh(Rss,Rt)
|
|
// Vector maximum bytes
|
|
// Rdd=vmaxb(Rtt,Rss)
|
|
// Vector reduce maximum halfwords
|
|
// Rxx=vrmaxh(Rss,Ru)
|
|
// Rxx=vrmaxuh(Rss,Ru)
|
|
// Vector reduce maximum words
|
|
// Rxx=vrmaxuw(Rss,Ru)
|
|
// Rxx=vrmaxw(Rss,Ru)
|
|
// Vector minimum bytes
|
|
// Rdd=vminb(Rtt,Rss)
|
|
// Vector reduce minimum halfwords
|
|
// Rxx=vrminh(Rss,Ru)
|
|
// Rxx=vrminuh(Rss,Ru)
|
|
// Vector reduce minimum words
|
|
// Rxx=vrminuw(Rss,Ru)
|
|
// Rxx=vrminw(Rss,Ru)
|
|
// Vector subtract bytes
|
|
// Rdd=vsubb(Rss,Rtt)
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/ALU -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/MPY +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Multiply and user lower result.
|
|
// Rd=add(#u6,mpyi(Rs,#U6))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDi_MPYri_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins u6Ext:$src1, IntRegs:$src2, u6Imm:$src3),
|
|
"$dst = add(#$src1, mpyi($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (mul (i32 IntRegs:$src2), u6ImmPred:$src3),
|
|
u6ExtPred:$src1))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=add(##,mpyi(Rs,#U6))
|
|
def : Pat <(add (mul (i32 IntRegs:$src2), u6ImmPred:$src3),
|
|
(HexagonCONST32 tglobaladdr:$src1)),
|
|
(i32 (ADDi_MPYri_V4 tglobaladdr:$src1, IntRegs:$src2,
|
|
u6ImmPred:$src3))>;
|
|
|
|
// Rd=add(#u6,mpyi(Rs,Rt))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT, InputType = "imm", CextOpcode = "ADD_MPY" in
|
|
def ADDi_MPYrr_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins u6Ext:$src1, IntRegs:$src2, IntRegs:$src3),
|
|
"$dst = add(#$src1, mpyi($src2, $src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (mul (i32 IntRegs:$src2), (i32 IntRegs:$src3)),
|
|
u6ExtPred:$src1))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Rd=add(##,mpyi(Rs,Rt))
|
|
def : Pat <(add (mul (i32 IntRegs:$src2), (i32 IntRegs:$src3)),
|
|
(HexagonCONST32 tglobaladdr:$src1)),
|
|
(i32 (ADDi_MPYrr_V4 tglobaladdr:$src1, IntRegs:$src2,
|
|
IntRegs:$src3))>;
|
|
|
|
// Rd=add(Ru,mpyi(#u6:2,Rs))
|
|
let validSubTargets = HasV4SubT in
|
|
def ADDr_MPYir_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, u6Imm:$src2, IntRegs:$src3),
|
|
"$dst = add($src1, mpyi(#$src2, $src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src3),
|
|
u6_2ImmPred:$src2)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rd=add(Ru,mpyi(Rs,#u6))
|
|
let isExtendable = 1, opExtendable = 3, isExtentSigned = 0, opExtentBits = 6,
|
|
validSubTargets = HasV4SubT, InputType = "imm", CextOpcode = "ADD_MPY" in
|
|
def ADDr_MPYri_V4 : MInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2, u6Ext:$src3),
|
|
"$dst = add($src1, mpyi($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src2),
|
|
u6ExtPred:$src3)))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Rx=add(Ru,mpyi(Rx,Rs))
|
|
let validSubTargets = HasV4SubT, InputType = "reg", CextOpcode = "ADD_MPY" in
|
|
def ADDr_MPYrr_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2, IntRegs:$src3),
|
|
"$dst = add($src1, mpyi($src2, $src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
|
|
// Polynomial multiply words
|
|
// Rdd=pmpyw(Rs,Rt)
|
|
// Rxx^=pmpyw(Rs,Rt)
|
|
|
|
// Vector reduce multiply word by signed half (32x16)
|
|
// Rdd=vrmpyweh(Rss,Rtt)[:<<1]
|
|
// Rdd=vrmpywoh(Rss,Rtt)[:<<1]
|
|
// Rxx+=vrmpyweh(Rss,Rtt)[:<<1]
|
|
// Rxx+=vrmpywoh(Rss,Rtt)[:<<1]
|
|
|
|
// Multiply and use upper result
|
|
// Rd=mpy(Rs,Rt.H):<<1:sat
|
|
// Rd=mpy(Rs,Rt.L):<<1:sat
|
|
// Rd=mpy(Rs,Rt):<<1
|
|
// Rd=mpy(Rs,Rt):<<1:sat
|
|
// Rd=mpysu(Rs,Rt)
|
|
// Rx+=mpy(Rs,Rt):<<1:sat
|
|
// Rx-=mpy(Rs,Rt):<<1:sat
|
|
|
|
// Vector multiply bytes
|
|
// Rdd=vmpybsu(Rs,Rt)
|
|
// Rdd=vmpybu(Rs,Rt)
|
|
// Rxx+=vmpybsu(Rs,Rt)
|
|
// Rxx+=vmpybu(Rs,Rt)
|
|
|
|
// Vector polynomial multiply halfwords
|
|
// Rdd=vpmpyh(Rs,Rt)
|
|
// Rxx^=vpmpyh(Rs,Rt)
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/MPY -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/SHIFT +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Shift by immediate and accumulate.
|
|
// Rx=add(#u8,asl(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = add(#$src1, asl($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (shl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx=add(#u8,lsr(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def ADDi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = add(#$src1, lsr($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(add (srl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx=sub(#u8,asl(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def SUBi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = sub(#$src1, asl($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(sub (shl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Rx=sub(#u8,lsr(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def SUBi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = sub(#$src1, lsr($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(sub (srl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
//Shift by immediate and logical.
|
|
//Rx=and(#u8,asl(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def ANDi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = and(#$src1, asl($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (shl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rx=and(#u8,lsr(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
validSubTargets = HasV4SubT in
|
|
def ANDi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = and(#$src1, lsr($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(and (srl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rx=or(#u8,asl(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
AddedComplexity = 30, validSubTargets = HasV4SubT in
|
|
def ORi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = or(#$src1, asl($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (shl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rx=or(#u8,lsr(Rx,#U5))
|
|
let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8,
|
|
AddedComplexity = 30, validSubTargets = HasV4SubT in
|
|
def ORi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst),
|
|
(ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3),
|
|
"$dst = or(#$src1, lsr($src2, #$src3))",
|
|
[(set (i32 IntRegs:$dst),
|
|
(or (srl (i32 IntRegs:$src2), u5ImmPred:$src3),
|
|
u8ExtPred:$src1))],
|
|
"$src2 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
//Shift by register.
|
|
//Rd=lsl(#s6,Rt)
|
|
let validSubTargets = HasV4SubT in {
|
|
def LSLi_V4 : MInst<(outs IntRegs:$dst), (ins s6Imm:$src1, IntRegs:$src2),
|
|
"$dst = lsl(#$src1, $src2)",
|
|
[(set (i32 IntRegs:$dst), (shl s6ImmPred:$src1,
|
|
(i32 IntRegs:$src2)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
//Shift by register and logical.
|
|
//Rxx^=asl(Rss,Rt)
|
|
def ASLd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3),
|
|
"$dst ^= asl($src2, $src3)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(xor (i64 DoubleRegs:$src1), (shl (i64 DoubleRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rxx^=asr(Rss,Rt)
|
|
def ASRd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3),
|
|
"$dst ^= asr($src2, $src3)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(xor (i64 DoubleRegs:$src1), (sra (i64 DoubleRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rxx^=lsl(Rss,Rt)
|
|
def LSLd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3),
|
|
"$dst ^= lsl($src2, $src3)",
|
|
[(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1),
|
|
(shl (i64 DoubleRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
|
|
//Rxx^=lsr(Rss,Rt)
|
|
def LSRd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst),
|
|
(ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3),
|
|
"$dst ^= lsr($src2, $src3)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(xor (i64 DoubleRegs:$src1), (srl (i64 DoubleRegs:$src2),
|
|
(i32 IntRegs:$src3))))],
|
|
"$src1 = $dst">,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/SHIFT -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// MEMOP: Word, Half, Byte
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// MEMOP: Word
|
|
//
|
|
// Implemented:
|
|
// MEMw_ADDi_indexed_V4 : memw(Rs+#u6:2)+=#U5
|
|
// MEMw_SUBi_indexed_V4 : memw(Rs+#u6:2)-=#U5
|
|
// MEMw_ADDr_indexed_V4 : memw(Rs+#u6:2)+=Rt
|
|
// MEMw_SUBr_indexed_V4 : memw(Rs+#u6:2)-=Rt
|
|
// MEMw_CLRr_indexed_V4 : memw(Rs+#u6:2)&=Rt
|
|
// MEMw_SETr_indexed_V4 : memw(Rs+#u6:2)|=Rt
|
|
// MEMw_ADDi_V4 : memw(Rs+#u6:2)+=#U5
|
|
// MEMw_SUBi_V4 : memw(Rs+#u6:2)-=#U5
|
|
// MEMw_ADDr_V4 : memw(Rs+#u6:2)+=Rt
|
|
// MEMw_SUBr_V4 : memw(Rs+#u6:2)-=Rt
|
|
// MEMw_CLRr_V4 : memw(Rs+#u6:2)&=Rt
|
|
// MEMw_SETr_V4 : memw(Rs+#u6:2)|=Rt
|
|
//
|
|
// Not implemented:
|
|
// MEMw_CLRi_indexed_V4 : memw(Rs+#u6:2)=clrbit(#U5)
|
|
// MEMw_SETi_indexed_V4 : memw(Rs+#u6:2)=setbit(#U5)
|
|
// MEMw_CLRi_V4 : memw(Rs+#u6:2)=clrbit(#U5)
|
|
// MEMw_SETi_V4 : memw(Rs+#u6:2)=setbit(#U5)
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
// memw(Rs+#u6:2) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, u5Imm:$addend),
|
|
"memw($base+#$offset) += #$addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMw_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, u5Imm:$subend),
|
|
"memw($base+#$offset) -= #$subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$addend),
|
|
"memw($base+#$offset) += $addend",
|
|
[(store (add (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)),
|
|
(i32 IntRegs:$addend)),
|
|
(add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$subend),
|
|
"memw($base+#$offset) -= $subend",
|
|
[(store (sub (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)),
|
|
(i32 IntRegs:$subend)),
|
|
(add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$andend),
|
|
"memw($base+#$offset) &= $andend",
|
|
[(store (and (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)),
|
|
(i32 IntRegs:$andend)),
|
|
(add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ORr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$orend),
|
|
"memw($base+#$offset) |= $orend",
|
|
[(store (or (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)),
|
|
(i32 IntRegs:$orend)),
|
|
(add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ADDi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$addend),
|
|
"memw($addr) += $addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMw_SUBi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$subend),
|
|
"memw($addr) -= $subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ADDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$addend),
|
|
"memw($addr) += $addend",
|
|
[(store (add (load ADDRriU6_2:$addr), (i32 IntRegs:$addend)),
|
|
ADDRriU6_2:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_SUBr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$subend),
|
|
"memw($addr) -= $subend",
|
|
[(store (sub (load ADDRriU6_2:$addr), (i32 IntRegs:$subend)),
|
|
ADDRriU6_2:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ANDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$andend),
|
|
"memw($addr) &= $andend",
|
|
[(store (and (load ADDRriU6_2:$addr), (i32 IntRegs:$andend)),
|
|
ADDRriU6_2:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memw(Rs+#u6:2) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMw_ORr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$orend),
|
|
"memw($addr) |= $orend",
|
|
[(store (or (load ADDRriU6_2:$addr), (i32 IntRegs:$orend)),
|
|
ADDRriU6_2:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// MEMOP: Halfword
|
|
//
|
|
// Implemented:
|
|
// MEMh_ADDi_indexed_V4 : memw(Rs+#u6:2)+=#U5
|
|
// MEMh_SUBi_indexed_V4 : memw(Rs+#u6:2)-=#U5
|
|
// MEMh_ADDr_indexed_V4 : memw(Rs+#u6:2)+=Rt
|
|
// MEMh_SUBr_indexed_V4 : memw(Rs+#u6:2)-=Rt
|
|
// MEMh_CLRr_indexed_V4 : memw(Rs+#u6:2)&=Rt
|
|
// MEMh_SETr_indexed_V4 : memw(Rs+#u6:2)|=Rt
|
|
// MEMh_ADDi_V4 : memw(Rs+#u6:2)+=#U5
|
|
// MEMh_SUBi_V4 : memw(Rs+#u6:2)-=#U5
|
|
// MEMh_ADDr_V4 : memw(Rs+#u6:2)+=Rt
|
|
// MEMh_SUBr_V4 : memw(Rs+#u6:2)-=Rt
|
|
// MEMh_CLRr_V4 : memw(Rs+#u6:2)&=Rt
|
|
// MEMh_SETr_V4 : memw(Rs+#u6:2)|=Rt
|
|
//
|
|
// Not implemented:
|
|
// MEMh_CLRi_indexed_V4 : memw(Rs+#u6:2)=clrbit(#U5)
|
|
// MEMh_SETi_indexed_V4 : memw(Rs+#u6:2)=setbit(#U5)
|
|
// MEMh_CLRi_V4 : memw(Rs+#u6:2)=clrbit(#U5)
|
|
// MEMh_SETi_V4 : memw(Rs+#u6:2)=setbit(#U5)
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
// memh(Rs+#u6:1) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, u5Imm:$addend),
|
|
"memh($base+#$offset) += $addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMh_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, u5Imm:$subend),
|
|
"memh($base+#$offset) -= $subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$addend),
|
|
"memh($base+#$offset) += $addend",
|
|
[(truncstorei16 (add (sextloadi16 (add (i32 IntRegs:$base),
|
|
u6_1ImmPred:$offset)),
|
|
(i32 IntRegs:$addend)),
|
|
(add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$subend),
|
|
"memh($base+#$offset) -= $subend",
|
|
[(truncstorei16 (sub (sextloadi16 (add (i32 IntRegs:$base),
|
|
u6_1ImmPred:$offset)),
|
|
(i32 IntRegs:$subend)),
|
|
(add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$andend),
|
|
"memh($base+#$offset) += $andend",
|
|
[(truncstorei16 (and (sextloadi16 (add (i32 IntRegs:$base),
|
|
u6_1ImmPred:$offset)),
|
|
(i32 IntRegs:$andend)),
|
|
(add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ORr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$orend),
|
|
"memh($base+#$offset) |= $orend",
|
|
[(truncstorei16 (or (sextloadi16 (add (i32 IntRegs:$base),
|
|
u6_1ImmPred:$offset)),
|
|
(i32 IntRegs:$orend)),
|
|
(add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ADDi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$addend),
|
|
"memh($addr) += $addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMh_SUBi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$subend),
|
|
"memh($addr) -= $subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ADDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$addend),
|
|
"memh($addr) += $addend",
|
|
[(truncstorei16 (add (sextloadi16 ADDRriU6_1:$addr),
|
|
(i32 IntRegs:$addend)), ADDRriU6_1:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_SUBr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$subend),
|
|
"memh($addr) -= $subend",
|
|
[(truncstorei16 (sub (sextloadi16 ADDRriU6_1:$addr),
|
|
(i32 IntRegs:$subend)), ADDRriU6_1:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ANDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$andend),
|
|
"memh($addr) &= $andend",
|
|
[(truncstorei16 (and (sextloadi16 ADDRriU6_1:$addr),
|
|
(i32 IntRegs:$andend)), ADDRriU6_1:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memh(Rs+#u6:1) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMh_ORr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$orend),
|
|
"memh($addr) |= $orend",
|
|
[(truncstorei16 (or (sextloadi16 ADDRriU6_1:$addr),
|
|
(i32 IntRegs:$orend)), ADDRriU6_1:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// MEMOP: Byte
|
|
//
|
|
// Implemented:
|
|
// MEMb_ADDi_indexed_V4 : memb(Rs+#u6:0)+=#U5
|
|
// MEMb_SUBi_indexed_V4 : memb(Rs+#u6:0)-=#U5
|
|
// MEMb_ADDr_indexed_V4 : memb(Rs+#u6:0)+=Rt
|
|
// MEMb_SUBr_indexed_V4 : memb(Rs+#u6:0)-=Rt
|
|
// MEMb_CLRr_indexed_V4 : memb(Rs+#u6:0)&=Rt
|
|
// MEMb_SETr_indexed_V4 : memb(Rs+#u6:0)|=Rt
|
|
// MEMb_ADDi_V4 : memb(Rs+#u6:0)+=#U5
|
|
// MEMb_SUBi_V4 : memb(Rs+#u6:0)-=#U5
|
|
// MEMb_ADDr_V4 : memb(Rs+#u6:0)+=Rt
|
|
// MEMb_SUBr_V4 : memb(Rs+#u6:0)-=Rt
|
|
// MEMb_CLRr_V4 : memb(Rs+#u6:0)&=Rt
|
|
// MEMb_SETr_V4 : memb(Rs+#u6:0)|=Rt
|
|
//
|
|
// Not implemented:
|
|
// MEMb_CLRi_indexed_V4 : memb(Rs+#u6:0)=clrbit(#U5)
|
|
// MEMb_SETi_indexed_V4 : memb(Rs+#u6:0)=setbit(#U5)
|
|
// MEMb_CLRi_V4 : memb(Rs+#u6:0)=clrbit(#U5)
|
|
// MEMb_SETi_V4 : memb(Rs+#u6:0)=setbit(#U5)
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// memb(Rs+#u6:0) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, u5Imm:$addend),
|
|
"memb($base+#$offset) += $addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMb_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, u5Imm:$subend),
|
|
"memb($base+#$offset) -= $subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$addend),
|
|
"memb($base+#$offset) += $addend",
|
|
[(truncstorei8 (add (sextloadi8 (add (i32 IntRegs:$base),
|
|
u6_0ImmPred:$offset)),
|
|
(i32 IntRegs:$addend)),
|
|
(add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$subend),
|
|
"memb($base+#$offset) -= $subend",
|
|
[(truncstorei8 (sub (sextloadi8 (add (i32 IntRegs:$base),
|
|
u6_0ImmPred:$offset)),
|
|
(i32 IntRegs:$subend)),
|
|
(add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$andend),
|
|
"memb($base+#$offset) += $andend",
|
|
[(truncstorei8 (and (sextloadi8 (add (i32 IntRegs:$base),
|
|
u6_0ImmPred:$offset)),
|
|
(i32 IntRegs:$andend)),
|
|
(add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ORr_indexed_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$orend),
|
|
"memb($base+#$offset) |= $orend",
|
|
[(truncstorei8 (or (sextloadi8 (add (i32 IntRegs:$base),
|
|
u6_0ImmPred:$offset)),
|
|
(i32 IntRegs:$orend)),
|
|
(add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) += #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ADDi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$addend),
|
|
"memb($addr) += $addend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) -= #U5
|
|
let AddedComplexity = 30 in
|
|
def MEMb_SUBi_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, u5Imm:$subend),
|
|
"memb($addr) -= $subend",
|
|
[]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) += Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ADDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$addend),
|
|
"memb($addr) += $addend",
|
|
[(truncstorei8 (add (sextloadi8 ADDRriU6_0:$addr),
|
|
(i32 IntRegs:$addend)), ADDRriU6_0:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) -= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_SUBr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$subend),
|
|
"memb($addr) -= $subend",
|
|
[(truncstorei8 (sub (sextloadi8 ADDRriU6_0:$addr),
|
|
(i32 IntRegs:$subend)), ADDRriU6_0:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) &= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ANDr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$andend),
|
|
"memb($addr) &= $andend",
|
|
[(truncstorei8 (and (sextloadi8 ADDRriU6_0:$addr),
|
|
(i32 IntRegs:$andend)), ADDRriU6_0:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
// memb(Rs+#u6:0) |= Rt
|
|
let AddedComplexity = 30 in
|
|
def MEMb_ORr_MEM_V4 : MEMInst_V4<(outs),
|
|
(ins MEMri:$addr, IntRegs:$orend),
|
|
"memb($addr) |= $orend",
|
|
[(truncstorei8 (or (sextloadi8 ADDRriU6_0:$addr),
|
|
(i32 IntRegs:$orend)), ADDRriU6_0:$addr)]>,
|
|
Requires<[HasV4T, UseMEMOP]>;
|
|
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/PRED +
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Hexagon V4 only supports these flavors of byte/half compare instructions:
|
|
// EQ/GT/GTU. Other flavors like GE/GEU/LT/LTU/LE/LEU are not supported by
|
|
// hardware. However, compiler can still implement these patterns through
|
|
// appropriate patterns combinations based on current implemented patterns.
|
|
// The implemented patterns are: EQ/GT/GTU.
|
|
// Missing patterns are: GE/GEU/LT/LTU/LE/LEU.
|
|
|
|
// Following instruction is not being extended as it results into the
|
|
// incorrect code for negative numbers.
|
|
// Pd=cmpb.eq(Rs,#u8)
|
|
|
|
// p=!cmp.eq(r1,r2)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotEQ_rr : ALU32_rr<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = !cmp.eq($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setne (i32 IntRegs:$src1), (i32 IntRegs:$src2)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// p=!cmp.eq(r1,#s10)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotEQ_ri : ALU32_ri<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, s10Ext:$src2),
|
|
"$dst = !cmp.eq($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setne (i32 IntRegs:$src1), s10ImmPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// p=!cmp.gt(r1,r2)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotGT_rr : ALU32_rr<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = !cmp.gt($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(not (setgt (i32 IntRegs:$src1), (i32 IntRegs:$src2))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// p=!cmp.gt(r1,#s10)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotGT_ri : ALU32_ri<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, s10Ext:$src2),
|
|
"$dst = !cmp.gt($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(not (setgt (i32 IntRegs:$src1), s10ImmPred:$src2)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// p=!cmp.gtu(r1,r2)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotGTU_rr : ALU32_rr<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = !cmp.gtu($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(not (setugt (i32 IntRegs:$src1), (i32 IntRegs:$src2))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// p=!cmp.gtu(r1,#u9)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPnotGTU_ri : ALU32_ri<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, u9Ext:$src2),
|
|
"$dst = !cmp.gtu($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(not (setugt (i32 IntRegs:$src1), u9ImmPred:$src2)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPbEQri_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, u8Imm:$src2),
|
|
"$dst = cmpb.eq($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (and (i32 IntRegs:$src1), 255), u8ImmPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat <(brcond (i1 (setne (and (i32 IntRegs:$src1), 255), u8ImmPred:$src2)),
|
|
bb:$offset),
|
|
(JMP_cNot (CMPbEQri_V4 (i32 IntRegs:$src1), u8ImmPred:$src2),
|
|
bb:$offset)>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Pd=cmpb.eq(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPbEQrr_ubub_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmpb.eq($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (and (xor (i32 IntRegs:$src1),
|
|
(i32 IntRegs:$src2)), 255), 0))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Pd=cmpb.eq(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPbEQrr_sbsb_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmpb.eq($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (shl (i32 IntRegs:$src1), (i32 24)),
|
|
(shl (i32 IntRegs:$src2), (i32 24))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Pd=cmpb.gt(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPbGTrr_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmpb.gt($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setgt (shl (i32 IntRegs:$src1), (i32 24)),
|
|
(shl (i32 IntRegs:$src2), (i32 24))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Pd=cmpb.gtu(Rs,#u7)
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 7,
|
|
isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPbGTU", InputType = "imm" in
|
|
def CMPbGTUri_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, u7Ext:$src2),
|
|
"$dst = cmpb.gtu($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 255),
|
|
u7ExtPred:$src2))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// SDNode for converting immediate C to C-1.
|
|
def DEC_CONST_BYTE : SDNodeXForm<imm, [{
|
|
// Return the byte immediate const-1 as an SDNode.
|
|
int32_t imm = N->getSExtValue();
|
|
return XformU7ToU7M1Imm(imm);
|
|
}]>;
|
|
|
|
// For the sequence
|
|
// zext( seteq ( and(Rs, 255), u8))
|
|
// Generate
|
|
// Pd=cmpb.eq(Rs, #u8)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
def : Pat <(i32 (zext (i1 (seteq (i32 (and (i32 IntRegs:$Rs), 255)),
|
|
u8ExtPred:$u8)))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbEQri_V4 (i32 IntRegs:$Rs),
|
|
(u8ExtPred:$u8))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setne ( and(Rs, 255), u8))
|
|
// Generate
|
|
// Pd=cmpb.eq(Rs, #u8)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
def : Pat <(i32 (zext (i1 (setne (i32 (and (i32 IntRegs:$Rs), 255)),
|
|
u8ExtPred:$u8)))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbEQri_V4 (i32 IntRegs:$Rs),
|
|
(u8ExtPred:$u8))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( seteq (Rs, and(Rt, 255)))
|
|
// Generate
|
|
// Pd=cmpb.eq(Rs, Rt)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
def : Pat <(i32 (zext (i1 (seteq (i32 IntRegs:$Rt),
|
|
(i32 (and (i32 IntRegs:$Rs), 255)))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbEQrr_ubub_V4 (i32 IntRegs:$Rs),
|
|
(i32 IntRegs:$Rt))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setne (Rs, and(Rt, 255)))
|
|
// Generate
|
|
// Pd=cmpb.eq(Rs, Rt)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
def : Pat <(i32 (zext (i1 (setne (i32 IntRegs:$Rt),
|
|
(i32 (and (i32 IntRegs:$Rs), 255)))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbEQrr_ubub_V4 (i32 IntRegs:$Rs),
|
|
(i32 IntRegs:$Rt))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setugt ( and(Rs, 255), u8))
|
|
// Generate
|
|
// Pd=cmpb.gtu(Rs, #u8)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
def : Pat <(i32 (zext (i1 (setugt (i32 (and (i32 IntRegs:$Rs), 255)),
|
|
u8ExtPred:$u8)))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$Rs),
|
|
(u8ExtPred:$u8))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setugt ( and(Rs, 254), u8))
|
|
// Generate
|
|
// Pd=cmpb.gtu(Rs, #u8)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
def : Pat <(i32 (zext (i1 (setugt (i32 (and (i32 IntRegs:$Rs), 254)),
|
|
u8ExtPred:$u8)))),
|
|
(i32 (TFR_condset_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$Rs),
|
|
(u8ExtPred:$u8))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setult ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.ltu(Rs, Rt)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
// cmp.ltu(Rs, Rt) -> cmp.gtu(Rt, Rs)
|
|
def : Pat <(i32 (zext (i1 (setult (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rt),
|
|
(i32 IntRegs:$Rs))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setlt ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.lt(Rs, Rt)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
// cmp.lt(Rs, Rt) -> cmp.gt(Rt, Rs)
|
|
def : Pat <(i32 (zext (i1 (setlt (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rt),
|
|
(i32 IntRegs:$Rs))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setugt ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.gtu(Rs, Rt)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
def : Pat <(i32 (zext (i1 (setugt (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rs),
|
|
(i32 IntRegs:$Rt))),
|
|
1, 0))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// This pattern interefers with coremark performance, not implementing at this
|
|
// time.
|
|
// For the sequence
|
|
// zext( setgt ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.gt(Rs, Rt)
|
|
// if (Pd.new) Rd=#1
|
|
// if (!Pd.new) Rd=#0
|
|
|
|
// For the sequence
|
|
// zext( setuge ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.ltu(Rs, Rt)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
// cmp.ltu(Rs, Rt) -> cmp.gtu(Rt, Rs)
|
|
def : Pat <(i32 (zext (i1 (setuge (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rt),
|
|
(i32 IntRegs:$Rs))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setge ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.lt(Rs, Rt)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
// cmp.lt(Rs, Rt) -> cmp.gt(Rt, Rs)
|
|
def : Pat <(i32 (zext (i1 (setge (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rt),
|
|
(i32 IntRegs:$Rs))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setule ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.gtu(Rs, Rt)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
def : Pat <(i32 (zext (i1 (setule (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rs),
|
|
(i32 IntRegs:$Rt))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setle ( Rs, Rt))
|
|
// Generate
|
|
// Pd=cmp.gt(Rs, Rt)
|
|
// if (Pd.new) Rd=#0
|
|
// if (!Pd.new) Rd=#1
|
|
def : Pat <(i32 (zext (i1 (setle (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))),
|
|
(i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rs),
|
|
(i32 IntRegs:$Rt))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// For the sequence
|
|
// zext( setult ( and(Rs, 255), u8))
|
|
// Use the isdigit transformation below
|
|
|
|
// Generate code of the form 'mux_ii(cmpbgtu(Rdd, C-1),0,1)'
|
|
// for C code of the form r = ((c>='0') & (c<='9')) ? 1 : 0;.
|
|
// The isdigit transformation relies on two 'clever' aspects:
|
|
// 1) The data type is unsigned which allows us to eliminate a zero test after
|
|
// biasing the expression by 48. We are depending on the representation of
|
|
// the unsigned types, and semantics.
|
|
// 2) The front end has converted <= 9 into < 10 on entry to LLVM
|
|
//
|
|
// For the C code:
|
|
// retval = ((c>='0') & (c<='9')) ? 1 : 0;
|
|
// The code is transformed upstream of llvm into
|
|
// retval = (c-48) < 10 ? 1 : 0;
|
|
let AddedComplexity = 139 in
|
|
def : Pat <(i32 (zext (i1 (setult (i32 (and (i32 IntRegs:$src1), 255)),
|
|
u7StrictPosImmPred:$src2)))),
|
|
(i32 (MUX_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$src1),
|
|
(DEC_CONST_BYTE u7StrictPosImmPred:$src2))),
|
|
0, 1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Pd=cmpb.gtu(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPbGTU",
|
|
InputType = "reg" in
|
|
def CMPbGTUrr_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmpb.gtu($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 255),
|
|
(and (i32 IntRegs:$src2), 255)))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Following instruction is not being extended as it results into the incorrect
|
|
// code for negative numbers.
|
|
|
|
// Signed half compare(.eq) ri.
|
|
// Pd=cmph.eq(Rs,#s8)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPhEQri_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, s8Imm:$src2),
|
|
"$dst = cmph.eq($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst), (seteq (and (i32 IntRegs:$src1), 65535),
|
|
s8ImmPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Signed half compare(.eq) rr.
|
|
// Case 1: xor + and, then compare:
|
|
// r0=xor(r0,r1)
|
|
// r0=and(r0,#0xffff)
|
|
// p0=cmp.eq(r0,#0)
|
|
// Pd=cmph.eq(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPhEQrr_xor_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmph.eq($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst), (seteq (and (xor (i32 IntRegs:$src1),
|
|
(i32 IntRegs:$src2)),
|
|
65535), 0))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Signed half compare(.eq) rr.
|
|
// Case 2: shift left 16 bits then compare:
|
|
// r0=asl(r0,16)
|
|
// r1=asl(r1,16)
|
|
// p0=cmp.eq(r0,r1)
|
|
// Pd=cmph.eq(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPhEQrr_shl_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmph.eq($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (shl (i32 IntRegs:$src1), (i32 16)),
|
|
(shl (i32 IntRegs:$src2), (i32 16))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
/* Incorrect Pattern -- immediate should be right shifted before being
|
|
used in the cmph.gt instruction.
|
|
// Signed half compare(.gt) ri.
|
|
// Pd=cmph.gt(Rs,#s8)
|
|
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8,
|
|
isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPhGTri_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, s8Ext:$src2),
|
|
"$dst = cmph.gt($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setgt (shl (i32 IntRegs:$src1), (i32 16)),
|
|
s8ExtPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
*/
|
|
|
|
// Signed half compare(.gt) rr.
|
|
// Pd=cmph.gt(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT in
|
|
def CMPhGTrr_shl_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmph.gt($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setgt (shl (i32 IntRegs:$src1), (i32 16)),
|
|
(shl (i32 IntRegs:$src2), (i32 16))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Unsigned half compare rr (.gtu).
|
|
// Pd=cmph.gtu(Rs,Rt)
|
|
let isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPhGTU",
|
|
InputType = "reg" in
|
|
def CMPhGTUrr_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = cmph.gtu($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(setugt (and (i32 IntRegs:$src1), 65535),
|
|
(and (i32 IntRegs:$src2), 65535)))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
// Unsigned half compare ri (.gtu).
|
|
// Pd=cmph.gtu(Rs,#u7)
|
|
let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 7,
|
|
isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPhGTU",
|
|
InputType = "imm" in
|
|
def CMPhGTUri_V4 : MInst<(outs PredRegs:$dst),
|
|
(ins IntRegs:$src1, u7Ext:$src2),
|
|
"$dst = cmph.gtu($src1, #$src2)",
|
|
[(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 65535),
|
|
u7ExtPred:$src2))]>,
|
|
Requires<[HasV4T]>, ImmRegRel;
|
|
|
|
let validSubTargets = HasV4SubT in
|
|
def NTSTBIT_rr : SInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2),
|
|
"$dst = !tstbit($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (and (shl 1, (i32 IntRegs:$src2)), (i32 IntRegs:$src1)), 0))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let validSubTargets = HasV4SubT in
|
|
def NTSTBIT_ri : SInst<(outs PredRegs:$dst), (ins IntRegs:$src1, u5Imm:$src2),
|
|
"$dst = !tstbit($src1, $src2)",
|
|
[(set (i1 PredRegs:$dst),
|
|
(seteq (and (shl 1, u5ImmPred:$src2), (i32 IntRegs:$src1)), 0))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// XTYPE/PRED -
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//Deallocate frame and return.
|
|
// dealloc_return
|
|
let isReturn = 1, isTerminator = 1, isBarrier = 1, isPredicable = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1 in {
|
|
def DEALLOC_RET_V4 : NVInst_V4<(outs), (ins i32imm:$amt1),
|
|
"dealloc_return",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Restore registers and dealloc return function call.
|
|
let isCall = 1, isBarrier = 1, isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC] in {
|
|
def RESTORE_DEALLOC_RET_JMP_V4 : JInst<(outs),
|
|
(ins calltarget:$dst),
|
|
"jump $dst // Restore_and_dealloc_return",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Restore registers and dealloc frame before a tail call.
|
|
let isCall = 1, isBarrier = 1,
|
|
Defs = [R29, R30, R31, PC] in {
|
|
def RESTORE_DEALLOC_BEFORE_TAILCALL_V4 : JInst<(outs),
|
|
(ins calltarget:$dst),
|
|
"call $dst // Restore_and_dealloc_before_tailcall",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Save registers function call.
|
|
let isCall = 1, isBarrier = 1,
|
|
Uses = [R29, R31] in {
|
|
def SAVE_REGISTERS_CALL_V4 : JInst<(outs),
|
|
(ins calltarget:$dst),
|
|
"call $dst // Save_calle_saved_registers",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (Ps) dealloc_return
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cPt_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, i32imm:$amt1),
|
|
"if ($src1) dealloc_return",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (!Ps) dealloc_return
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cNotPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1,
|
|
i32imm:$amt1),
|
|
"if (!$src1) dealloc_return",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (Ps.new) dealloc_return:nt
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cdnPnt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1,
|
|
i32imm:$amt1),
|
|
"if ($src1.new) dealloc_return:nt",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (!Ps.new) dealloc_return:nt
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cNotdnPnt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1,
|
|
i32imm:$amt1),
|
|
"if (!$src1.new) dealloc_return:nt",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (Ps.new) dealloc_return:t
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cdnPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1,
|
|
i32imm:$amt1),
|
|
"if ($src1.new) dealloc_return:t",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// if (!Ps.new) dealloc_return:nt
|
|
let isReturn = 1, isTerminator = 1,
|
|
Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1,
|
|
isPredicated = 1 in {
|
|
def DEALLOC_RET_cNotdnPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1,
|
|
i32imm:$amt1),
|
|
"if (!$src1.new) dealloc_return:t",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
// Load/Store with absolute addressing mode
|
|
// memw(#u6)=Rt
|
|
|
|
multiclass ST_Abs_Predbase<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, globaladdressExt:$absaddr, RC: $src2),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"(##$absaddr) = $src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Abs_Pred<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Abs_Predbase<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Abs_Predbase<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let isNVStorable = 1, isExtended = 1, neverHasSideEffects = 1 in
|
|
multiclass ST_Abs<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in {
|
|
let opExtendable = 0, isPredicable = 1 in
|
|
def NAME#_V4 : STInst2<(outs),
|
|
(ins globaladdressExt:$absaddr, RC:$src),
|
|
mnemonic#"(##$absaddr) = $src",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 1, isPredicated = 1 in {
|
|
defm Pt : ST_Abs_Pred<mnemonic, RC, 0>;
|
|
defm NotPt : ST_Abs_Pred<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
multiclass ST_Abs_Predbase_nv<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, globaladdressExt:$absaddr, RC: $src2),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#mnemonic#"(##$absaddr) = $src2.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass ST_Abs_Pred_nv<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : ST_Abs_Predbase_nv<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : ST_Abs_Predbase_nv<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let mayStore = 1, isNVStore = 1, isExtended = 1, neverHasSideEffects = 1 in
|
|
multiclass ST_Abs_nv<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in {
|
|
let opExtendable = 0, isPredicable = 1 in
|
|
def NAME#_nv_V4 : NVInst_V4<(outs),
|
|
(ins globaladdressExt:$absaddr, RC:$src),
|
|
mnemonic#"(##$absaddr) = $src.new",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 1, isPredicated = 1 in {
|
|
defm Pt : ST_Abs_Pred_nv<mnemonic, RC, 0>;
|
|
defm NotPt : ST_Abs_Pred_nv<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = Absolute in {
|
|
defm STrib_abs : ST_Abs<"memb", "STrib", IntRegs>,
|
|
ST_Abs_nv<"memb", "STrib", IntRegs>, AddrModeRel;
|
|
|
|
defm STrih_abs : ST_Abs<"memh", "STrih", IntRegs>,
|
|
ST_Abs_nv<"memh", "STrih", IntRegs>, AddrModeRel;
|
|
|
|
defm STriw_abs : ST_Abs<"memw", "STriw", IntRegs>,
|
|
ST_Abs_nv<"memw", "STriw", IntRegs>, AddrModeRel;
|
|
|
|
let isNVStorable = 0 in
|
|
defm STrid_abs : ST_Abs<"memd", "STrid", DoubleRegs>, AddrModeRel;
|
|
}
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 30 in {
|
|
def : Pat<(truncstorei8 (i32 IntRegs:$src1),
|
|
(HexagonCONST32 tglobaladdr:$absaddr)),
|
|
(STrib_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>;
|
|
|
|
def : Pat<(truncstorei16 (i32 IntRegs:$src1),
|
|
(HexagonCONST32 tglobaladdr:$absaddr)),
|
|
(STrih_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>;
|
|
|
|
def : Pat<(store (i32 IntRegs:$src1), (HexagonCONST32 tglobaladdr:$absaddr)),
|
|
(STriw_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>;
|
|
|
|
def : Pat<(store (i64 DoubleRegs:$src1),
|
|
(HexagonCONST32 tglobaladdr:$absaddr)),
|
|
(STrid_abs_V4 tglobaladdr: $absaddr, DoubleRegs: $src1)>;
|
|
}
|
|
|
|
multiclass LD_Abs_Predbase<string mnemonic, RegisterClass RC, bit isNot,
|
|
bit isPredNew> {
|
|
let PNewValue = !if(isPredNew, "new", "") in
|
|
def NAME : LDInst2<(outs RC:$dst),
|
|
(ins PredRegs:$src1, globaladdressExt:$absaddr),
|
|
!if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ",
|
|
") ")#"$dst = "#mnemonic#"(##$absaddr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
multiclass LD_Abs_Pred<string mnemonic, RegisterClass RC, bit PredNot> {
|
|
let PredSense = !if(PredNot, "false", "true") in {
|
|
defm _c#NAME : LD_Abs_Predbase<mnemonic, RC, PredNot, 0>;
|
|
// Predicate new
|
|
defm _cdn#NAME : LD_Abs_Predbase<mnemonic, RC, PredNot, 1>;
|
|
}
|
|
}
|
|
|
|
let isExtended = 1, neverHasSideEffects = 1 in
|
|
multiclass LD_Abs<string mnemonic, string CextOp, RegisterClass RC> {
|
|
let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in {
|
|
let opExtendable = 1, isPredicable = 1 in
|
|
def NAME#_V4 : LDInst2<(outs RC:$dst),
|
|
(ins globaladdressExt:$absaddr),
|
|
"$dst = "#mnemonic#"(##$absaddr)",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let opExtendable = 2, isPredicated = 1 in {
|
|
defm Pt_V4 : LD_Abs_Pred<mnemonic, RC, 0>;
|
|
defm NotPt_V4 : LD_Abs_Pred<mnemonic, RC, 1>;
|
|
}
|
|
}
|
|
}
|
|
|
|
let addrMode = Absolute in {
|
|
defm LDrib_abs : LD_Abs<"memb", "LDrib", IntRegs>, AddrModeRel;
|
|
defm LDriub_abs : LD_Abs<"memub", "LDriub", IntRegs>, AddrModeRel;
|
|
defm LDrih_abs : LD_Abs<"memh", "LDrih", IntRegs>, AddrModeRel;
|
|
defm LDriuh_abs : LD_Abs<"memuh", "LDriuh", IntRegs>, AddrModeRel;
|
|
defm LDriw_abs : LD_Abs<"memw", "LDriw", IntRegs>, AddrModeRel;
|
|
defm LDrid_abs : LD_Abs<"memd", "LDrid", DoubleRegs>, AddrModeRel;
|
|
}
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 30 in
|
|
def : Pat<(i32 (load (HexagonCONST32 tglobaladdr:$absaddr))),
|
|
(LDriw_abs_V4 tglobaladdr: $absaddr)>;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity=30 in
|
|
def : Pat<(i32 (sextloadi8 (HexagonCONST32 tglobaladdr:$absaddr))),
|
|
(LDrib_abs_V4 tglobaladdr:$absaddr)>;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity=30 in
|
|
def : Pat<(i32 (zextloadi8 (HexagonCONST32 tglobaladdr:$absaddr))),
|
|
(LDriub_abs_V4 tglobaladdr:$absaddr)>;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity=30 in
|
|
def : Pat<(i32 (sextloadi16 (HexagonCONST32 tglobaladdr:$absaddr))),
|
|
(LDrih_abs_V4 tglobaladdr:$absaddr)>;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity=30 in
|
|
def : Pat<(i32 (zextloadi16 (HexagonCONST32 tglobaladdr:$absaddr))),
|
|
(LDriuh_abs_V4 tglobaladdr:$absaddr)>;
|
|
|
|
// Transfer global address into a register
|
|
let AddedComplexity=50, isMoveImm = 1, isReMaterializable = 1 in
|
|
def TFRI_V4 : ALU32_ri<(outs IntRegs:$dst), (ins globaladdress:$src1),
|
|
"$dst = ##$src1",
|
|
[(set IntRegs:$dst, (HexagonCONST32 tglobaladdr:$src1))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in
|
|
def TFRI_cPt_V4 : ALU32_ri<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$src2),
|
|
"if($src1) $dst = ##$src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in
|
|
def TFRI_cNotPt_V4 : ALU32_ri<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$src2),
|
|
"if(!$src1) $dst = ##$src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in
|
|
def TFRI_cdnPt_V4 : ALU32_ri<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$src2),
|
|
"if($src1.new) $dst = ##$src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in
|
|
def TFRI_cdnNotPt_V4 : ALU32_ri<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, globaladdress:$src2),
|
|
"if(!$src1.new) $dst = ##$src2",
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 50, Predicates = [HasV4T] in
|
|
def : Pat<(HexagonCONST32_GP tglobaladdr:$src1),
|
|
(TFRI_V4 tglobaladdr:$src1)>;
|
|
|
|
|
|
// Load - Indirect with long offset: These instructions take global address
|
|
// as an operand
|
|
let AddedComplexity = 10 in
|
|
def LDrid_ind_lo_V4 : LDInst<(outs DoubleRegs:$dst),
|
|
(ins IntRegs:$src1, u2Imm:$src2, globaladdress:$offset),
|
|
"$dst=memd($src1<<#$src2+##$offset)",
|
|
[(set (i64 DoubleRegs:$dst),
|
|
(load (add (shl IntRegs:$src1, u2ImmPred:$src2),
|
|
(HexagonCONST32 tglobaladdr:$offset))))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 10 in
|
|
multiclass LD_indirect_lo<string OpcStr, PatFrag OpNode> {
|
|
def _lo_V4 : LDInst<(outs IntRegs:$dst),
|
|
(ins IntRegs:$src1, u2Imm:$src2, globaladdress:$offset),
|
|
!strconcat("$dst = ",
|
|
!strconcat(OpcStr, "($src1<<#$src2+##$offset)")),
|
|
[(set IntRegs:$dst,
|
|
(i32 (OpNode (add (shl IntRegs:$src1, u2ImmPred:$src2),
|
|
(HexagonCONST32 tglobaladdr:$offset)))))]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
defm LDrib_ind : LD_indirect_lo<"memb", sextloadi8>;
|
|
defm LDriub_ind : LD_indirect_lo<"memub", zextloadi8>;
|
|
defm LDrih_ind : LD_indirect_lo<"memh", sextloadi16>;
|
|
defm LDriuh_ind : LD_indirect_lo<"memuh", zextloadi16>;
|
|
defm LDriw_ind : LD_indirect_lo<"memw", load>;
|
|
|
|
// Store - Indirect with long offset: These instructions take global address
|
|
// as an operand
|
|
let AddedComplexity = 10 in
|
|
def STrid_ind_lo_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, globaladdress:$src3,
|
|
DoubleRegs:$src4),
|
|
"memd($src1<<#$src2+#$src3) = $src4",
|
|
[(store (i64 DoubleRegs:$src4),
|
|
(add (shl IntRegs:$src1, u2ImmPred:$src2),
|
|
(HexagonCONST32 tglobaladdr:$src3)))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let AddedComplexity = 10 in
|
|
multiclass ST_indirect_lo<string OpcStr, PatFrag OpNode> {
|
|
def _lo_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u2Imm:$src2, globaladdress:$src3,
|
|
IntRegs:$src4),
|
|
!strconcat(OpcStr, "($src1<<#$src2+##$src3) = $src4"),
|
|
[(OpNode (i32 IntRegs:$src4),
|
|
(add (shl IntRegs:$src1, u2ImmPred:$src2),
|
|
(HexagonCONST32 tglobaladdr:$src3)))]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
defm STrib_ind : ST_indirect_lo<"memb", truncstorei8>;
|
|
defm STrih_ind : ST_indirect_lo<"memh", truncstorei16>;
|
|
defm STriw_ind : ST_indirect_lo<"memw", store>;
|
|
|
|
// Store - absolute addressing mode: These instruction take constant
|
|
// value as the extended operand.
|
|
multiclass ST_absimm<string OpcStr> {
|
|
let isExtended = 1, opExtendable = 0, isPredicable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def _abs_V4 : STInst2<(outs),
|
|
(ins u0AlwaysExt:$src1, IntRegs:$src2),
|
|
!strconcat(OpcStr, "(##$src1) = $src2"),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isExtended = 1, opExtendable = 1, isPredicated = 1,
|
|
validSubTargets = HasV4SubT in {
|
|
def _abs_cPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if ($src1)", !strconcat(OpcStr, "(##$src2) = $src3")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if (!$src1)", !strconcat(OpcStr, "(##$src2) = $src3")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if ($src1.new)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnNotPt_V4 : STInst2<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if (!$src1.new)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
|
|
let isExtended = 1, opExtendable = 0, mayStore = 1, isNVStore = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def _abs_nv_V4 : NVInst_V4<(outs),
|
|
(ins u0AlwaysExt:$src1, IntRegs:$src2),
|
|
!strconcat(OpcStr, "(##$src1) = $src2.new"),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isExtended = 1, opExtendable = 1, mayStore = 1, isPredicated = 1,
|
|
isNVStore = 1, validSubTargets = HasV4SubT in {
|
|
def _abs_cPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if ($src1)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3.new")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if (!$src1)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3.new")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if ($src1.new)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3.new")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnNotPt_nv_V4 : NVInst_V4<(outs),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3),
|
|
!strconcat("if (!$src1.new)",
|
|
!strconcat(OpcStr, "(##$src2) = $src3.new")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
}
|
|
|
|
defm STrib_imm : ST_absimm<"memb">;
|
|
defm STrih_imm : ST_absimm<"memh">;
|
|
defm STriw_imm : ST_absimm<"memw">;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 30 in {
|
|
def : Pat<(truncstorei8 (i32 IntRegs:$src1), u0AlwaysExtPred:$src2),
|
|
(STrib_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>;
|
|
|
|
def : Pat<(truncstorei16 (i32 IntRegs:$src1), u0AlwaysExtPred:$src2),
|
|
(STrih_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>;
|
|
|
|
def : Pat<(store (i32 IntRegs:$src1), u0AlwaysExtPred:$src2),
|
|
(STriw_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>;
|
|
}
|
|
|
|
// Load - absolute addressing mode: These instruction take constant
|
|
// value as the extended operand
|
|
|
|
multiclass LD_absimm<string OpcStr> {
|
|
let isExtended = 1, opExtendable = 1, isPredicable = 1,
|
|
validSubTargets = HasV4SubT in
|
|
def _abs_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins u0AlwaysExt:$src),
|
|
!strconcat("$dst = ",
|
|
!strconcat(OpcStr, "(##$src)")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
let isExtended = 1, opExtendable = 2, isPredicated = 1,
|
|
validSubTargets = HasV4SubT in {
|
|
def _abs_cPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2),
|
|
!strconcat("if ($src1) $dst = ",
|
|
!strconcat(OpcStr, "(##$src2)")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2),
|
|
!strconcat("if (!$src1) $dst = ",
|
|
!strconcat(OpcStr, "(##$src2)")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2),
|
|
!strconcat("if ($src1.new) $dst = ",
|
|
!strconcat(OpcStr, "(##$src2)")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def _abs_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst),
|
|
(ins PredRegs:$src1, u0AlwaysExt:$src2),
|
|
!strconcat("if (!$src1.new) $dst = ",
|
|
!strconcat(OpcStr, "(##$src2)")),
|
|
[]>,
|
|
Requires<[HasV4T]>;
|
|
}
|
|
}
|
|
|
|
defm LDrib_imm : LD_absimm<"memb">;
|
|
defm LDriub_imm : LD_absimm<"memub">;
|
|
defm LDrih_imm : LD_absimm<"memh">;
|
|
defm LDriuh_imm : LD_absimm<"memuh">;
|
|
defm LDriw_imm : LD_absimm<"memw">;
|
|
|
|
let Predicates = [HasV4T], AddedComplexity = 30 in {
|
|
def : Pat<(i32 (load u0AlwaysExtPred:$src)),
|
|
(LDriw_imm_abs_V4 u0AlwaysExtPred:$src)>;
|
|
|
|
def : Pat<(i32 (sextloadi8 u0AlwaysExtPred:$src)),
|
|
(LDrib_imm_abs_V4 u0AlwaysExtPred:$src)>;
|
|
|
|
def : Pat<(i32 (zextloadi8 u0AlwaysExtPred:$src)),
|
|
(LDriub_imm_abs_V4 u0AlwaysExtPred:$src)>;
|
|
|
|
def : Pat<(i32 (sextloadi16 u0AlwaysExtPred:$src)),
|
|
(LDrih_imm_abs_V4 u0AlwaysExtPred:$src)>;
|
|
|
|
def : Pat<(i32 (zextloadi16 u0AlwaysExtPred:$src)),
|
|
(LDriuh_imm_abs_V4 u0AlwaysExtPred:$src)>;
|
|
}
|
|
|
|
// Indexed store double word - global address.
|
|
// memw(Rs+#u6:2)=#S8
|
|
let AddedComplexity = 10 in
|
|
def STriw_offset_ext_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u6_2Imm:$src2, globaladdress:$src3),
|
|
"memw($src1+#$src2) = ##$src3",
|
|
[(store (HexagonCONST32 tglobaladdr:$src3),
|
|
(add IntRegs:$src1, u6_2ImmPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
|
|
|
|
// Indexed store double word - global address.
|
|
// memw(Rs+#u6:2)=#S8
|
|
let AddedComplexity = 10 in
|
|
def STrih_offset_ext_V4 : STInst<(outs),
|
|
(ins IntRegs:$src1, u6_1Imm:$src2, globaladdress:$src3),
|
|
"memh($src1+#$src2) = ##$src3",
|
|
[(truncstorei16 (HexagonCONST32 tglobaladdr:$src3),
|
|
(add IntRegs:$src1, u6_1ImmPred:$src2))]>,
|
|
Requires<[HasV4T]>;
|
|
// Map from store(globaladdress + x) -> memd(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(store (i64 DoubleRegs:$src1),
|
|
FoldGlobalAddrGP:$addr),
|
|
(STrid_abs_V4 FoldGlobalAddrGP:$addr, (i64 DoubleRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_store_64 FoldGlobalAddrGP:$addr,
|
|
(i64 DoubleRegs:$src1)),
|
|
(STrid_abs_V4 FoldGlobalAddrGP:$addr, (i64 DoubleRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress + x) -> memb(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(truncstorei8 (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr),
|
|
(STrib_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_store_8 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)),
|
|
(STrib_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress + x) -> memh(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(truncstorei16 (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr),
|
|
(STrih_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_store_16 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)),
|
|
(STrih_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from store(globaladdress + x) -> memw(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(store (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr),
|
|
(STriw_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_store_32 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)),
|
|
(STriw_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memd(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i64 (load FoldGlobalAddrGP:$addr)),
|
|
(i64 (LDrid_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_load_64 FoldGlobalAddrGP:$addr),
|
|
(i64 (LDrid_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memb(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (extloadi8 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDrib_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memb(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (sextloadi8 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDrib_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
//let AddedComplexity = 100 in
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (extloadi16 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDrih_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memh(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (sextloadi16 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDrih_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memuh(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (zextloadi16 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDriuh_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_load_16 FoldGlobalAddrGP:$addr),
|
|
(i32 (LDriuh_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memub(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (zextloadi8 FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDriub_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_load_8 FoldGlobalAddrGP:$addr),
|
|
(i32 (LDriub_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
// Map from load(globaladdress + x) -> memw(#foo + x)
|
|
let AddedComplexity = 100 in
|
|
def : Pat<(i32 (load FoldGlobalAddrGP:$addr)),
|
|
(i32 (LDriw_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|
|
def : Pat<(atomic_load_32 FoldGlobalAddrGP:$addr),
|
|
(i32 (LDriw_abs_V4 FoldGlobalAddrGP:$addr))>,
|
|
Requires<[HasV4T]>;
|
|
|