//=- HexagonInstrInfoV4.td - Target Desc. for Hexagon Target -*- tablegen -*-=// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // // This file describes the Hexagon V4 instructions in TableGen format. // //===----------------------------------------------------------------------===// let neverHasSideEffects = 1 in class T_Immext : EXTENDERInst<(outs), ins, "immext(#$imm)", []>, Requires<[HasV4T]>; def IMMEXT_b : T_Immext<(ins brtarget:$imm)>; def IMMEXT_c : T_Immext<(ins calltarget:$imm)>; def IMMEXT_g : T_Immext<(ins globaladdress:$imm)>; def IMMEXT_i : T_Immext<(ins u26_6Imm:$imm)>; // Fold (add (CONST32 tglobaladdr:$addr) ) into a global address. def FoldGlobalAddr : ComplexPattern; // Fold (add (CONST32_GP tglobaladdr:$addr) ) into a global address. def FoldGlobalAddrGP : ComplexPattern; def NumUsesBelowThresCONST32 : PatFrag<(ops node:$addr), (HexagonCONST32 node:$addr), [{ return hasNumUsesBelowThresGA(N->getOperand(0).getNode()); }]>; // Hexagon V4 Architecture spec defines 8 instruction classes: // LD ST ALU32 XTYPE J JR MEMOP NV CR SYSTEM(system is not implemented in the // compiler) // LD Instructions: // ======================================== // Loads (8/16/32/64 bit) // Deallocframe // ST Instructions: // ======================================== // Stores (8/16/32/64 bit) // Allocframe // ALU32 Instructions: // ======================================== // Arithmetic / Logical (32 bit) // Vector Halfword // XTYPE Instructions (32/64 bit): // ======================================== // Arithmetic, Logical, Bit Manipulation // Multiply (Integer, Fractional, Complex) // Permute / Vector Permute Operations // Predicate Operations // Shift / Shift with Add/Sub/Logical // Vector Byte ALU // Vector Halfword (ALU, Shift, Multiply) // Vector Word (ALU, Shift) // J Instructions: // ======================================== // Jump/Call PC-relative // JR Instructions: // ======================================== // Jump/Call Register // MEMOP Instructions: // ======================================== // Operation on memory (8/16/32 bit) // NV Instructions: // ======================================== // New-value Jumps // New-value Stores // CR Instructions: // ======================================== // Control-Register Transfers // Hardware Loop Setup // Predicate Logicals & Reductions // SYSTEM Instructions (not implemented in the compiler): // ======================================== // Prefetch // Cache Maintenance // Bus Operations //===----------------------------------------------------------------------===// // ALU32 + //===----------------------------------------------------------------------===// // Shift halfword. let isPredicated = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in { def ASLH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = aslh($src2)", []>, Requires<[HasV4T]>; def ASLH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = aslh($src2)", []>, Requires<[HasV4T]>; def ASLH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = aslh($src2)", []>, Requires<[HasV4T]>; def ASLH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = aslh($src2)", []>, Requires<[HasV4T]>; def ASRH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = asrh($src2)", []>, Requires<[HasV4T]>; def ASRH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = asrh($src2)", []>, Requires<[HasV4T]>; def ASRH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = asrh($src2)", []>, Requires<[HasV4T]>; def ASRH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = asrh($src2)", []>, Requires<[HasV4T]>; } // Sign extend. let isPredicated = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in { def SXTB_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = sxtb($src2)", []>, Requires<[HasV4T]>; def SXTB_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = sxtb($src2)", []>, Requires<[HasV4T]>; def SXTB_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = sxtb($src2)", []>, Requires<[HasV4T]>; def SXTB_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = sxtb($src2)", []>, Requires<[HasV4T]>; def SXTH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = sxth($src2)", []>, Requires<[HasV4T]>; def SXTH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = sxth($src2)", []>, Requires<[HasV4T]>; def SXTH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = sxth($src2)", []>, Requires<[HasV4T]>; def SXTH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = sxth($src2)", []>, Requires<[HasV4T]>; } // Zero exten. let neverHasSideEffects = 1, isPredicated = 1, validSubTargets = HasV4SubT in { def ZXTB_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = zxtb($src2)", []>, Requires<[HasV4T]>; def ZXTB_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = zxtb($src2)", []>, Requires<[HasV4T]>; def ZXTB_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = zxtb($src2)", []>, Requires<[HasV4T]>; def ZXTB_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = zxtb($src2)", []>, Requires<[HasV4T]>; def ZXTH_cPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1) $dst = zxth($src2)", []>, Requires<[HasV4T]>; def ZXTH_cNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1) $dst = zxth($src2)", []>, Requires<[HasV4T]>; def ZXTH_cdnPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if ($src1.new) $dst = zxth($src2)", []>, Requires<[HasV4T]>; def ZXTH_cdnNotPt_V4 : ALU32_rr<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2), "if (!$src1.new) $dst = zxth($src2)", []>, Requires<[HasV4T]>; } // Generate frame index addresses. let neverHasSideEffects = 1, isReMaterializable = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in def TFR_FI_immext_V4 : ALU32_ri<(outs IntRegs:$dst), (ins IntRegs:$src1, s32Imm:$offset), "$dst = add($src1, ##$offset)", []>, Requires<[HasV4T]>; // Rd=cmp.eq(Rs,#s8) let validSubTargets = HasV4SubT, isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8 in def V4_A4_rcmpeqi : ALU32_ri<(outs IntRegs:$Rd), (ins IntRegs:$Rs, s8Ext:$s8), "$Rd = cmp.eq($Rs, #$s8)", [(set (i32 IntRegs:$Rd), (i32 (zext (i1 (seteq (i32 IntRegs:$Rs), s8ExtPred:$s8)))))]>, Requires<[HasV4T]>; // Preserve the TSTBIT generation def : Pat <(i32 (zext (i1 (setne (i32 (and (i32 (shl 1, (i32 IntRegs:$src2))), (i32 IntRegs:$src1))), 0)))), (i32 (MUX_ii (i1 (TSTBIT_rr (i32 IntRegs:$src1), (i32 IntRegs:$src2))), 1, 0))>; // Interfered with tstbit generation, above pattern preserves, see : tstbit.ll // Rd=cmp.ne(Rs,#s8) let validSubTargets = HasV4SubT, isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8 in def V4_A4_rcmpneqi : ALU32_ri<(outs IntRegs:$Rd), (ins IntRegs:$Rs, s8Ext:$s8), "$Rd = !cmp.eq($Rs, #$s8)", [(set (i32 IntRegs:$Rd), (i32 (zext (i1 (setne (i32 IntRegs:$Rs), s8ExtPred:$s8)))))]>, Requires<[HasV4T]>; // Rd=cmp.eq(Rs,Rt) let validSubTargets = HasV4SubT in def V4_A4_rcmpeq : ALU32_ri<(outs IntRegs:$Rd), (ins IntRegs:$Rs, IntRegs:$Rt), "$Rd = cmp.eq($Rs, $Rt)", [(set (i32 IntRegs:$Rd), (i32 (zext (i1 (seteq (i32 IntRegs:$Rs), IntRegs:$Rt)))))]>, Requires<[HasV4T]>; // Rd=cmp.ne(Rs,Rt) let validSubTargets = HasV4SubT in def V4_A4_rcmpneq : ALU32_ri<(outs IntRegs:$Rd), (ins IntRegs:$Rs, IntRegs:$Rt), "$Rd = !cmp.eq($Rs, $Rt)", [(set (i32 IntRegs:$Rd), (i32 (zext (i1 (setne (i32 IntRegs:$Rs), IntRegs:$Rt)))))]>, Requires<[HasV4T]>; //===----------------------------------------------------------------------===// // ALU32 - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // ALU32/PERM + //===----------------------------------------------------------------------===// // Combine // Rdd=combine(Rs, #s8) let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def COMBINE_rI_V4 : ALU32_ri<(outs DoubleRegs:$dst), (ins IntRegs:$src1, s8Ext:$src2), "$dst = combine($src1, #$src2)", []>, Requires<[HasV4T]>; // Rdd=combine(#s8, Rs) let isExtendable = 1, opExtendable = 1, isExtentSigned = 1, opExtentBits = 8, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def COMBINE_Ir_V4 : ALU32_ir<(outs DoubleRegs:$dst), (ins s8Ext:$src1, IntRegs:$src2), "$dst = combine(#$src1, $src2)", []>, Requires<[HasV4T]>; def HexagonWrapperCombineRI_V4 : SDNode<"HexagonISD::WrapperCombineRI_V4", SDTHexagonI64I32I32>; def HexagonWrapperCombineIR_V4 : SDNode<"HexagonISD::WrapperCombineIR_V4", SDTHexagonI64I32I32>; def : Pat <(HexagonWrapperCombineRI_V4 IntRegs:$r, s8ExtPred:$i), (COMBINE_rI_V4 IntRegs:$r, s8ExtPred:$i)>, Requires<[HasV4T]>; def : Pat <(HexagonWrapperCombineIR_V4 s8ExtPred:$i, IntRegs:$r), (COMBINE_Ir_V4 s8ExtPred:$i, IntRegs:$r)>, Requires<[HasV4T]>; let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 6, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def COMBINE_iI_V4 : ALU32_ii<(outs DoubleRegs:$dst), (ins s8Imm:$src1, u6Ext:$src2), "$dst = combine(#$src1, #$src2)", []>, Requires<[HasV4T]>; //===----------------------------------------------------------------------===// // ALU32/PERM + //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // LD + //===----------------------------------------------------------------------===// // // These absolute set addressing mode instructions accept immediate as // an operand. We have duplicated these patterns to take global address. let isExtended = 1, opExtendable = 2, neverHasSideEffects = 1, validSubTargets = HasV4SubT in { def LDrid_abs_setimm_V4 : LDInst2<(outs DoubleRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memd($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memb(Re=#U6) def LDrib_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memb($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memh(Re=#U6) def LDrih_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memh($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memub(Re=#U6) def LDriub_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memub($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memuh(Re=#U6) def LDriuh_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memuh($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memw(Re=#U6) def LDriw_abs_setimm_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins u0AlwaysExt:$addr), "$dst1 = memw($dst2=##$addr)", []>, Requires<[HasV4T]>; } // Following patterns are defined for absolute set addressing mode // instruction which take global address as operand. let isExtended = 1, opExtendable = 2, neverHasSideEffects = 1, validSubTargets = HasV4SubT in { def LDrid_abs_set_V4 : LDInst2<(outs DoubleRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memd($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memb(Re=#U6) def LDrib_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memb($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memh(Re=#U6) def LDrih_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memh($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memub(Re=#U6) def LDriub_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memub($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memuh(Re=#U6) def LDriuh_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memuh($dst2=##$addr)", []>, Requires<[HasV4T]>; // Rd=memw(Re=#U6) def LDriw_abs_set_V4 : LDInst2<(outs IntRegs:$dst1, IntRegs:$dst2), (ins globaladdressExt:$addr), "$dst1 = memw($dst2=##$addr)", []>, Requires<[HasV4T]>; } // multiclass for load instructions with base + register offset // addressing mode multiclass ld_idxd_shl_pbase { let PNewValue = !if(isPredNew, "new", "") in def NAME : LDInst2<(outs RC:$dst), (ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$offset), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#"$dst = "#mnemonic#"($src2+$src3<<#$offset)", []>, Requires<[HasV4T]>; } multiclass ld_idxd_shl_pred { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ld_idxd_shl_pbase; // Predicate new defm _cdn#NAME : ld_idxd_shl_pbase; } } let neverHasSideEffects = 1 in multiclass ld_idxd_shl { let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in { let isPredicable = 1 in def NAME#_V4 : LDInst2<(outs RC:$dst), (ins IntRegs:$src1, IntRegs:$src2, u2Imm:$offset), "$dst = "#mnemonic#"($src1+$src2<<#$offset)", []>, Requires<[HasV4T]>; let isPredicated = 1 in { defm Pt_V4 : ld_idxd_shl_pred; defm NotPt_V4 : ld_idxd_shl_pred; } } } let addrMode = BaseRegOffset in { defm LDrib_indexed_shl: ld_idxd_shl<"memb", "LDrib", IntRegs>, AddrModeRel; defm LDriub_indexed_shl: ld_idxd_shl<"memub", "LDriub", IntRegs>, AddrModeRel; defm LDrih_indexed_shl: ld_idxd_shl<"memh", "LDrih", IntRegs>, AddrModeRel; defm LDriuh_indexed_shl: ld_idxd_shl<"memuh", "LDriuh", IntRegs>, AddrModeRel; defm LDriw_indexed_shl: ld_idxd_shl<"memw", "LDriw", IntRegs>, AddrModeRel; defm LDrid_indexed_shl: ld_idxd_shl<"memd", "LDrid", DoubleRegs>, AddrModeRel; } // 'def pats' for load instructions with base + register offset and non-zero // immediate value. Immediate value is used to left-shift the second // register operand. let AddedComplexity = 40 in { def : Pat <(i32 (sextloadi8 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDrib_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (zextloadi8 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (extloadi8 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (sextloadi16 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDrih_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (zextloadi16 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (extloadi16 (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i32 (load (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDriw_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; def : Pat <(i64 (load (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$offset)))), (LDrid_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$offset)>, Requires<[HasV4T]>; } // 'def pats' for load instruction base + register offset and // zero immediate value. let AddedComplexity = 10 in { def : Pat <(i64 (load (add IntRegs:$src1, IntRegs:$src2))), (LDrid_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (sextloadi8 (add IntRegs:$src1, IntRegs:$src2))), (LDrib_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (zextloadi8 (add IntRegs:$src1, IntRegs:$src2))), (LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (extloadi8 (add IntRegs:$src1, IntRegs:$src2))), (LDriub_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (sextloadi16 (add IntRegs:$src1, IntRegs:$src2))), (LDrih_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (zextloadi16 (add IntRegs:$src1, IntRegs:$src2))), (LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (extloadi16 (add IntRegs:$src1, IntRegs:$src2))), (LDriuh_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; def : Pat <(i32 (load (add IntRegs:$src1, IntRegs:$src2))), (LDriw_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, 0)>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDd_GP_V4 : LDInst2<(outs DoubleRegs:$dst), (ins globaladdress:$global), "$dst=memd(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rtt=memd(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDd_GP_cPt_V4 : LDInst2<(outs DoubleRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memd(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rtt=memd(##global) def LDd_GP_cNotPt_V4 : LDInst2<(outs DoubleRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memd(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rtt=memd(##global) def LDd_GP_cdnPt_V4 : LDInst2<(outs DoubleRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memd(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rtt=memd(##global) def LDd_GP_cdnNotPt_V4 : LDInst2<(outs DoubleRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memd(##$global)", []>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDb_GP_V4 : LDInst2<(outs IntRegs:$dst), (ins globaladdress:$global), "$dst=memb(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memb(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDb_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memb(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memb(##global) def LDb_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memb(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memb(##global) def LDb_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memb(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memb(##global) def LDb_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memb(##$global)", []>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDub_GP_V4 : LDInst2<(outs IntRegs:$dst), (ins globaladdress:$global), "$dst=memub(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memub(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDub_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memub(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memub(##global) def LDub_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memub(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memub(##global) def LDub_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memub(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memub(##global) def LDub_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memub(##$global)", []>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDh_GP_V4 : LDInst2<(outs IntRegs:$dst), (ins globaladdress:$global), "$dst=memh(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memh(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDh_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memh(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memh(##global) def LDh_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memh(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memh(##global) def LDh_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memh(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memh(##global) def LDh_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memh(##$global)", []>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDuh_GP_V4 : LDInst2<(outs IntRegs:$dst), (ins globaladdress:$global), "$dst=memuh(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memuh(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDuh_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memuh(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memuh(##global) def LDuh_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memuh(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memuh(##global) def LDuh_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memuh(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memuh(##global) def LDuh_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memuh(##$global)", []>, Requires<[HasV4T]>; } let isPredicable = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def LDw_GP_V4 : LDInst2<(outs IntRegs:$dst), (ins globaladdress:$global), "$dst=memw(#$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memw(##global) let neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def LDw_GP_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1) $dst=memw(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memw(##global) def LDw_GP_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1) $dst=memw(##$global)", []>, Requires<[HasV4T]>; // if (Pv) Rt=memw(##global) def LDw_GP_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if ($src1.new) $dst=memw(##$global)", []>, Requires<[HasV4T]>; // if (!Pv) Rt=memw(##global) def LDw_GP_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$global), "if (!$src1.new) $dst=memw(##$global)", []>, Requires<[HasV4T]>; } def : Pat <(atomic_load_64 (HexagonCONST32_GP tglobaladdr:$global)), (i64 (LDd_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; def : Pat <(atomic_load_32 (HexagonCONST32_GP tglobaladdr:$global)), (i32 (LDw_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; def : Pat <(atomic_load_16 (HexagonCONST32_GP tglobaladdr:$global)), (i32 (LDuh_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; def : Pat <(atomic_load_8 (HexagonCONST32_GP tglobaladdr:$global)), (i32 (LDub_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memw(#foo + 0) let AddedComplexity = 100 in def : Pat <(i64 (load (HexagonCONST32_GP tglobaladdr:$global))), (i64 (LDd_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from Pd = load(globaladdress) -> Rd = memb(globaladdress), Pd = Rd let AddedComplexity = 100 in def : Pat <(i1 (load (HexagonCONST32_GP tglobaladdr:$global))), (i1 (TFR_PdRs (i32 (LDb_GP_V4 tglobaladdr:$global))))>, Requires<[HasV4T]>; // When the Interprocedural Global Variable optimizer realizes that a certain // global variable takes only two constant values, it shrinks the global to // a boolean. Catch those loads here in the following 3 patterns. let AddedComplexity = 100 in def : Pat <(i32 (extloadi1 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDb_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; let AddedComplexity = 100 in def : Pat <(i32 (sextloadi1 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDb_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memb(#foo) let AddedComplexity = 100 in def : Pat <(i32 (extloadi8 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDb_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memb(#foo) let AddedComplexity = 100 in def : Pat <(i32 (sextloadi8 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDb_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; let AddedComplexity = 100 in def : Pat <(i32 (zextloadi1 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDub_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memub(#foo) let AddedComplexity = 100 in def : Pat <(i32 (zextloadi8 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDub_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memh(#foo) let AddedComplexity = 100 in def : Pat <(i32 (extloadi16 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDh_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memh(#foo) let AddedComplexity = 100 in def : Pat <(i32 (sextloadi16 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDh_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memuh(#foo) let AddedComplexity = 100 in def : Pat <(i32 (zextloadi16 (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDuh_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // Map from load(globaladdress) -> memw(#foo) let AddedComplexity = 100 in def : Pat <(i32 (load (HexagonCONST32_GP tglobaladdr:$global))), (i32 (LDw_GP_V4 tglobaladdr:$global))>, Requires<[HasV4T]>; // zext i1->i64 def : Pat <(i64 (zext (i1 PredRegs:$src1))), (i64 (COMBINE_Ir_V4 0, (MUX_ii (i1 PredRegs:$src1), 1, 0)))>, Requires<[HasV4T]>; // zext i32->i64 def : Pat <(i64 (zext (i32 IntRegs:$src1))), (i64 (COMBINE_Ir_V4 0, (i32 IntRegs:$src1)))>, Requires<[HasV4T]>; // zext i8->i64 def: Pat <(i64 (zextloadi8 ADDRriS11_0:$src1)), (i64 (COMBINE_Ir_V4 0, (LDriub ADDRriS11_0:$src1)))>, Requires<[HasV4T]>; let AddedComplexity = 20 in def: Pat <(i64 (zextloadi8 (add (i32 IntRegs:$src1), s11_0ExtPred:$offset))), (i64 (COMBINE_Ir_V4 0, (LDriub_indexed IntRegs:$src1, s11_0ExtPred:$offset)))>, Requires<[HasV4T]>; // zext i16->i64 def: Pat <(i64 (zextloadi16 ADDRriS11_1:$src1)), (i64 (COMBINE_Ir_V4 0, (LDriuh ADDRriS11_1:$src1)))>, Requires<[HasV4T]>; let AddedComplexity = 20 in def: Pat <(i64 (zextloadi16 (add (i32 IntRegs:$src1), s11_1ExtPred:$offset))), (i64 (COMBINE_Ir_V4 0, (LDriuh_indexed IntRegs:$src1, s11_1ExtPred:$offset)))>, Requires<[HasV4T]>; // anyext i16->i64 def: Pat <(i64 (extloadi16 ADDRriS11_2:$src1)), (i64 (COMBINE_Ir_V4 0, (LDrih ADDRriS11_2:$src1)))>, Requires<[HasV4T]>; let AddedComplexity = 20 in def: Pat <(i64 (extloadi16 (add (i32 IntRegs:$src1), s11_1ExtPred:$offset))), (i64 (COMBINE_Ir_V4 0, (LDrih_indexed IntRegs:$src1, s11_1ExtPred:$offset)))>, Requires<[HasV4T]>; // zext i32->i64 def: Pat <(i64 (zextloadi32 ADDRriS11_2:$src1)), (i64 (COMBINE_Ir_V4 0, (LDriw ADDRriS11_2:$src1)))>, Requires<[HasV4T]>; let AddedComplexity = 100 in def: Pat <(i64 (zextloadi32 (i32 (add IntRegs:$src1, s11_2ExtPred:$offset)))), (i64 (COMBINE_Ir_V4 0, (LDriw_indexed IntRegs:$src1, s11_2ExtPred:$offset)))>, Requires<[HasV4T]>; // anyext i32->i64 def: Pat <(i64 (extloadi32 ADDRriS11_2:$src1)), (i64 (COMBINE_Ir_V4 0, (LDriw ADDRriS11_2:$src1)))>, Requires<[HasV4T]>; let AddedComplexity = 100 in def: Pat <(i64 (extloadi32 (i32 (add IntRegs:$src1, s11_2ExtPred:$offset)))), (i64 (COMBINE_Ir_V4 0, (LDriw_indexed IntRegs:$src1, s11_2ExtPred:$offset)))>, Requires<[HasV4T]>; //===----------------------------------------------------------------------===// // LD - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // ST + //===----------------------------------------------------------------------===// /// /// Assumptions::: ****** DO NOT IGNORE ******** /// 1. Make sure that in post increment store, the zero'th operand is always the /// post increment operand. /// 2. Make sure that the store value operand(Rt/Rtt) in a store is always the /// last operand. /// // memd(Re=#U)=Rtt let isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def STrid_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1), (ins DoubleRegs:$src1, u0AlwaysExt:$src2), "memd($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memb(Re=#U)=Rs def STrib_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, u0AlwaysExt:$src2), "memb($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memh(Re=#U)=Rs def STrih_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, u0AlwaysExt:$src2), "memh($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memw(Re=#U)=Rs def STriw_abs_setimm_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, u0AlwaysExt:$src2), "memw($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; } // memd(Re=#U)=Rtt let isExtended = 1, opExtendable = 2, validSubTargets = HasV4SubT in { def STrid_abs_set_V4 : STInst2<(outs IntRegs:$dst1), (ins DoubleRegs:$src1, globaladdressExt:$src2), "memd($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memb(Re=#U)=Rs def STrib_abs_set_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, globaladdressExt:$src2), "memb($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memh(Re=#U)=Rs def STrih_abs_set_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, globaladdressExt:$src2), "memh($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; // memw(Re=#U)=Rs def STriw_abs_set_V4 : STInst2<(outs IntRegs:$dst1), (ins IntRegs:$src1, globaladdressExt:$src2), "memw($dst1=##$src2) = $src1", []>, Requires<[HasV4T]>; } // multiclass for store instructions with base + register offset addressing // mode multiclass ST_Idxd_shl_Pbase { let PNewValue = !if(isPredNew, "new", "") in def NAME : STInst2<(outs), (ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$src4, RC:$src5), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($src2+$src3<<#$src4) = $src5", []>, Requires<[HasV4T]>; } multiclass ST_Idxd_shl_Pred { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Idxd_shl_Pbase; // Predicate new defm _cdn#NAME : ST_Idxd_shl_Pbase; } } let isNVStorable = 1 in multiclass ST_Idxd_shl { let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in { let isPredicable = 1 in def NAME#_V4 : STInst2<(outs), (ins IntRegs:$src1, IntRegs:$src2, u2Imm:$src3, RC:$src4), mnemonic#"($src1+$src2<<#$src3) = $src4", []>, Requires<[HasV4T]>; let isPredicated = 1 in { defm Pt_V4 : ST_Idxd_shl_Pred; defm NotPt_V4 : ST_Idxd_shl_Pred; } } } // multiclass for new-value store instructions with base + register offset // addressing mode. multiclass ST_Idxd_shl_Pbase_nv { let PNewValue = !if(isPredNew, "new", "") in def NAME#_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, IntRegs:$src2, IntRegs:$src3, u2Imm:$src4, RC:$src5), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($src2+$src3<<#$src4) = $src5.new", []>, Requires<[HasV4T]>; } multiclass ST_Idxd_shl_Pred_nv { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Idxd_shl_Pbase_nv; // Predicate new defm _cdn#NAME : ST_Idxd_shl_Pbase_nv; } } let mayStore = 1, isNVStore = 1 in multiclass ST_Idxd_shl_nv { let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed_shl in { let isPredicable = 1 in def NAME#_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, IntRegs:$src2, u2Imm:$src3, RC:$src4), mnemonic#"($src1+$src2<<#$src3) = $src4.new", []>, Requires<[HasV4T]>; let isPredicated = 1 in { defm Pt : ST_Idxd_shl_Pred_nv; defm NotPt : ST_Idxd_shl_Pred_nv; } } } let addrMode = BaseRegOffset, neverHasSideEffects = 1, validSubTargets = HasV4SubT in { defm STrib_indexed_shl: ST_Idxd_shl<"memb", "STrib", IntRegs>, ST_Idxd_shl_nv<"memb", "STrib", IntRegs>, AddrModeRel; defm STrih_indexed_shl: ST_Idxd_shl<"memh", "STrih", IntRegs>, ST_Idxd_shl_nv<"memh", "STrih", IntRegs>, AddrModeRel; defm STriw_indexed_shl: ST_Idxd_shl<"memw", "STriw", IntRegs>, ST_Idxd_shl_nv<"memw", "STriw", IntRegs>, AddrModeRel; let isNVStorable = 0 in defm STrid_indexed_shl: ST_Idxd_shl<"memd", "STrid", DoubleRegs>, AddrModeRel; } let Predicates = [HasV4T], AddedComplexity = 10 in { def : Pat<(truncstorei8 (i32 IntRegs:$src4), (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))), (STrib_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$src3, IntRegs:$src4)>; def : Pat<(truncstorei16 (i32 IntRegs:$src4), (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))), (STrih_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$src3, IntRegs:$src4)>; def : Pat<(store (i32 IntRegs:$src4), (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))), (STriw_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$src3, IntRegs:$src4)>; def : Pat<(store (i64 DoubleRegs:$src4), (add IntRegs:$src1, (shl IntRegs:$src2, u2ImmPred:$src3))), (STrid_indexed_shl_V4 IntRegs:$src1, IntRegs:$src2, u2ImmPred:$src3, DoubleRegs:$src4)>; } // memd(Ru<<#u2+#U6)=Rtt let isExtended = 1, opExtendable = 2, AddedComplexity = 10, validSubTargets = HasV4SubT in def STrid_shl_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, DoubleRegs:$src4), "memd($src1<<#$src2+#$src3) = $src4", [(store (i64 DoubleRegs:$src4), (add (shl (i32 IntRegs:$src1), u2ImmPred:$src2), u0AlwaysExtPred:$src3))]>, Requires<[HasV4T]>; // memd(Rx++#s4:3)=Rtt // memd(Rx++#s4:3:circ(Mu))=Rtt // memd(Rx++I:circ(Mu))=Rtt // memd(Rx++Mu)=Rtt // memd(Rx++Mu:brev)=Rtt // memd(gp+#u16:3)=Rtt // Store doubleword conditionally. // if ([!]Pv[.new]) memd(#u6)=Rtt // TODO: needs to be implemented. //===----------------------------------------------------------------------===// // multiclass for store instructions with base + immediate offset // addressing mode and immediate stored value. // mem[bhw](Rx++#s4:3)=#s8 // if ([!]Pv[.new]) mem[bhw](Rx++#s4:3)=#s6 //===----------------------------------------------------------------------===// multiclass ST_Imm_Pbase { let PNewValue = !if(isPredNew, "new", "") in def NAME : STInst2<(outs), (ins PredRegs:$src1, IntRegs:$src2, OffsetOp:$src3, s6Ext:$src4), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($src2+#$src3) = #$src4", []>, Requires<[HasV4T]>; } multiclass ST_Imm_Pred { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Imm_Pbase; // Predicate new defm _cdn#NAME : ST_Imm_Pbase; } } let isExtendable = 1, isExtentSigned = 1, neverHasSideEffects = 1 in multiclass ST_Imm { let CextOpcode = CextOp, BaseOpcode = CextOp#_imm in { let opExtendable = 2, opExtentBits = 8, isPredicable = 1 in def NAME#_V4 : STInst2<(outs), (ins IntRegs:$src1, OffsetOp:$src2, s8Ext:$src3), mnemonic#"($src1+#$src2) = #$src3", []>, Requires<[HasV4T]>; let opExtendable = 3, opExtentBits = 6, isPredicated = 1 in { defm Pt_V4 : ST_Imm_Pred; defm NotPt_V4 : ST_Imm_Pred; } } } let addrMode = BaseImmOffset, InputType = "imm", validSubTargets = HasV4SubT in { defm STrib_imm : ST_Imm<"memb", "STrib", u6_0Imm>, ImmRegRel, PredNewRel; defm STrih_imm : ST_Imm<"memh", "STrih", u6_1Imm>, ImmRegRel, PredNewRel; defm STriw_imm : ST_Imm<"memw", "STriw", u6_2Imm>, ImmRegRel, PredNewRel; } let Predicates = [HasV4T], AddedComplexity = 10 in { def: Pat<(truncstorei8 s8ExtPred:$src3, (add IntRegs:$src1, u6_0ImmPred:$src2)), (STrib_imm_V4 IntRegs:$src1, u6_0ImmPred:$src2, s8ExtPred:$src3)>; def: Pat<(truncstorei16 s8ExtPred:$src3, (add IntRegs:$src1, u6_1ImmPred:$src2)), (STrih_imm_V4 IntRegs:$src1, u6_1ImmPred:$src2, s8ExtPred:$src3)>; def: Pat<(store s8ExtPred:$src3, (add IntRegs:$src1, u6_2ImmPred:$src2)), (STriw_imm_V4 IntRegs:$src1, u6_2ImmPred:$src2, s8ExtPred:$src3)>; } let AddedComplexity = 6 in def : Pat <(truncstorei8 s8ExtPred:$src2, (i32 IntRegs:$src1)), (STrib_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>, Requires<[HasV4T]>; // memb(Ru<<#u2+#U6)=Rt let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1, validSubTargets = HasV4SubT in def STrib_shl_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memb($src1<<#$src2+#$src3) = $src4", [(truncstorei8 (i32 IntRegs:$src4), (add (shl (i32 IntRegs:$src1), u2ImmPred:$src2), u0AlwaysExtPred:$src3))]>, Requires<[HasV4T]>; // memb(Rx++#s4:0:circ(Mu))=Rt // memb(Rx++I:circ(Mu))=Rt // memb(Rx++Mu)=Rt // memb(Rx++Mu:brev)=Rt // memb(gp+#u16:0)=Rt // Store halfword. // TODO: needs to be implemented // memh(Re=#U6)=Rt.H // memh(Rs+#s11:1)=Rt.H let AddedComplexity = 6 in def : Pat <(truncstorei16 s8ExtPred:$src2, (i32 IntRegs:$src1)), (STrih_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>, Requires<[HasV4T]>; // memh(Rs+Ru<<#u2)=Rt.H // TODO: needs to be implemented. // memh(Ru<<#u2+#U6)=Rt.H // memh(Ru<<#u2+#U6)=Rt let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1, validSubTargets = HasV4SubT in def STrih_shl_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memh($src1<<#$src2+#$src3) = $src4", [(truncstorei16 (i32 IntRegs:$src4), (add (shl (i32 IntRegs:$src1), u2ImmPred:$src2), u0AlwaysExtPred:$src3))]>, Requires<[HasV4T]>; // memh(Rx++#s4:1:circ(Mu))=Rt.H // memh(Rx++#s4:1:circ(Mu))=Rt // memh(Rx++I:circ(Mu))=Rt.H // memh(Rx++I:circ(Mu))=Rt // memh(Rx++Mu)=Rt.H // memh(Rx++Mu)=Rt // memh(Rx++Mu:brev)=Rt.H // memh(Rx++Mu:brev)=Rt // memh(gp+#u16:1)=Rt // if ([!]Pv[.new]) memh(#u6)=Rt.H // if ([!]Pv[.new]) memh(#u6)=Rt // if ([!]Pv[.new]) memh(Rs+#u6:1)=Rt.H // TODO: needs to be implemented. // if ([!]Pv[.new]) memh(Rx++#s4:1)=Rt.H // TODO: Needs to be implemented. // Store word. // memw(Re=#U6)=Rt // TODO: Needs to be implemented. // Store predicate: let neverHasSideEffects = 1 in def STriw_pred_V4 : STInst2<(outs), (ins MEMri:$addr, PredRegs:$src1), "Error; should not emit", []>, Requires<[HasV4T]>; let AddedComplexity = 6 in def : Pat <(store s8ExtPred:$src2, (i32 IntRegs:$src1)), (STriw_imm_V4 IntRegs:$src1, 0, s8ExtPred:$src2)>, Requires<[HasV4T]>; // memw(Ru<<#u2+#U6)=Rt let isExtended = 1, opExtendable = 2, AddedComplexity = 10, isNVStorable = 1, validSubTargets = HasV4SubT in def STriw_shl_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memw($src1<<#$src2+#$src3) = $src4", [(store (i32 IntRegs:$src4), (add (shl (i32 IntRegs:$src1), u2ImmPred:$src2), u0AlwaysExtPred:$src3))]>, Requires<[HasV4T]>; // memw(Rx++#s4:2)=Rt // memw(Rx++#s4:2:circ(Mu))=Rt // memw(Rx++I:circ(Mu))=Rt // memw(Rx++Mu)=Rt // memw(Rx++Mu:brev)=Rt // memw(gp+#u16:2)=Rt // memd(#global)=Rtt let isPredicable = 1, mayStore = 1, neverHasSideEffects = 1, validSubTargets = HasV4SubT in def STd_GP_V4 : STInst2<(outs), (ins globaladdress:$global, DoubleRegs:$src), "memd(#$global) = $src", []>, Requires<[HasV4T]>; // if (Pv) memd(##global) = Rtt let mayStore = 1, neverHasSideEffects = 1, isPredicated = 1, isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in { def STd_GP_cPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2), "if ($src1) memd(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memd(##global) = Rtt def STd_GP_cNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2), "if (!$src1) memd(##$global) = $src2", []>, Requires<[HasV4T]>; // if (Pv) memd(##global) = Rtt def STd_GP_cdnPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2), "if ($src1.new) memd(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memd(##global) = Rtt def STd_GP_cdnNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, DoubleRegs:$src2), "if (!$src1.new) memd(##$global) = $src2", []>, Requires<[HasV4T]>; } // memb(#global)=Rt let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1, validSubTargets = HasV4SubT in def STb_GP_V4 : STInst2<(outs), (ins globaladdress:$global, IntRegs:$src), "memb(#$global) = $src", []>, Requires<[HasV4T]>; // if (Pv) memb(##global) = Rt let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1, isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in { def STb_GP_cPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memb(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memb(##global) = Rt def STb_GP_cNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memb(##$global) = $src2", []>, Requires<[HasV4T]>; // if (Pv) memb(##global) = Rt def STb_GP_cdnPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memb(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memb(##global) = Rt def STb_GP_cdnNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memb(##$global) = $src2", []>, Requires<[HasV4T]>; } // memh(#global)=Rt let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1, validSubTargets = HasV4SubT in def STh_GP_V4 : STInst2<(outs), (ins globaladdress:$global, IntRegs:$src), "memh(#$global) = $src", []>, Requires<[HasV4T]>; // if (Pv) memh(##global) = Rt let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1, isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in { def STh_GP_cPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memh(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memh(##global) = Rt def STh_GP_cNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memh(##$global) = $src2", []>, Requires<[HasV4T]>; // if (Pv) memh(##global) = Rt def STh_GP_cdnPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memh(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memh(##global) = Rt def STh_GP_cdnNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memh(##$global) = $src2", []>, Requires<[HasV4T]>; } // memw(#global)=Rt let isPredicable = 1, neverHasSideEffects = 1, isNVStorable = 1, validSubTargets = HasV4SubT in def STw_GP_V4 : STInst2<(outs), (ins globaladdress:$global, IntRegs:$src), "memw(#$global) = $src", []>, Requires<[HasV4T]>; // if (Pv) memw(##global) = Rt let neverHasSideEffects = 1, isPredicated = 1, isNVStorable = 1, isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in { def STw_GP_cPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memw(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memw(##global) = Rt def STw_GP_cNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memw(##$global) = $src2", []>, Requires<[HasV4T]>; // if (Pv) memw(##global) = Rt def STw_GP_cdnPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memw(##$global) = $src2", []>, Requires<[HasV4T]>; // if (!Pv) memw(##global) = Rt def STw_GP_cdnNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memw(##$global) = $src2", []>, Requires<[HasV4T]>; } // 64 bit atomic store def : Pat <(atomic_store_64 (HexagonCONST32_GP tglobaladdr:$global), (i64 DoubleRegs:$src1)), (STd_GP_V4 tglobaladdr:$global, (i64 DoubleRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress) -> memd(#foo) let AddedComplexity = 100 in def : Pat <(store (i64 DoubleRegs:$src1), (HexagonCONST32_GP tglobaladdr:$global)), (STd_GP_V4 tglobaladdr:$global, (i64 DoubleRegs:$src1))>, Requires<[HasV4T]>; // 8 bit atomic store def : Pat < (atomic_store_8 (HexagonCONST32_GP tglobaladdr:$global), (i32 IntRegs:$src1)), (STb_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress) -> memb(#foo) let AddedComplexity = 100 in def : Pat<(truncstorei8 (i32 IntRegs:$src1), (HexagonCONST32_GP tglobaladdr:$global)), (STb_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from "i1 = constant<-1>; memw(CONST32(#foo)) = i1" // to "r0 = 1; memw(#foo) = r0" let AddedComplexity = 100 in def : Pat<(store (i1 -1), (HexagonCONST32_GP tglobaladdr:$global)), (STb_GP_V4 tglobaladdr:$global, (TFRI 1))>, Requires<[HasV4T]>; def : Pat<(atomic_store_16 (HexagonCONST32_GP tglobaladdr:$global), (i32 IntRegs:$src1)), (STh_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress) -> memh(#foo) let AddedComplexity = 100 in def : Pat<(truncstorei16 (i32 IntRegs:$src1), (HexagonCONST32_GP tglobaladdr:$global)), (STh_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // 32 bit atomic store def : Pat<(atomic_store_32 (HexagonCONST32_GP tglobaladdr:$global), (i32 IntRegs:$src1)), (STw_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress) -> memw(#foo) let AddedComplexity = 100 in def : Pat<(store (i32 IntRegs:$src1), (HexagonCONST32_GP tglobaladdr:$global)), (STw_GP_V4 tglobaladdr:$global, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; //===----------------------------------------------------------------------=== // ST - //===----------------------------------------------------------------------=== //===----------------------------------------------------------------------===// // NV/ST + //===----------------------------------------------------------------------===// // multiclass for new-value store instructions with base + immediate offset. // multiclass ST_Idxd_Pbase_nv { let PNewValue = !if(isPredNew, "new", "") in def NAME#_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, IntRegs:$src2, predImmOp:$src3, RC: $src4), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($src2+#$src3) = $src4.new", []>, Requires<[HasV4T]>; } multiclass ST_Idxd_Pred_nv { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Idxd_Pbase_nv; // Predicate new defm _cdn#NAME : ST_Idxd_Pbase_nv; } } let mayStore = 1, isNVStore = 1, neverHasSideEffects = 1, isExtendable = 1 in multiclass ST_Idxd_nv ImmBits, bits<5> PredImmBits> { let CextOpcode = CextOp, BaseOpcode = CextOp#_indexed in { let opExtendable = 1, isExtentSigned = 1, opExtentBits = ImmBits, isPredicable = 1 in def NAME#_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, ImmOp:$src2, RC:$src3), mnemonic#"($src1+#$src2) = $src3.new", []>, Requires<[HasV4T]>; let opExtendable = 2, isExtentSigned = 0, opExtentBits = PredImmBits, isPredicated = 1 in { defm Pt : ST_Idxd_Pred_nv; defm NotPt : ST_Idxd_Pred_nv; } } } let addrMode = BaseImmOffset, validSubTargets = HasV4SubT in { defm STrib_indexed: ST_Idxd_nv<"memb", "STrib", IntRegs, s11_0Ext, u6_0Ext, 11, 6>, AddrModeRel; defm STrih_indexed: ST_Idxd_nv<"memh", "STrih", IntRegs, s11_1Ext, u6_1Ext, 12, 7>, AddrModeRel; defm STriw_indexed: ST_Idxd_nv<"memw", "STriw", IntRegs, s11_2Ext, u6_2Ext, 13, 8>, AddrModeRel; } // multiclass for new-value store instructions with base + immediate offset. // and MEMri operand. multiclass ST_MEMri_Pbase_nv { let PNewValue = !if(isPredNew, "new", "") in def NAME#_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, MEMri:$addr, RC: $src2), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($addr) = $src2.new", []>, Requires<[HasV4T]>; } multiclass ST_MEMri_Pred_nv { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_MEMri_Pbase_nv; // Predicate new defm _cdn#NAME : ST_MEMri_Pbase_nv; } } let mayStore = 1, isNVStore = 1, isExtendable = 1, neverHasSideEffects = 1 in multiclass ST_MEMri_nv ImmBits, bits<5> PredImmBits> { let CextOpcode = CextOp, BaseOpcode = CextOp in { let opExtendable = 1, isExtentSigned = 1, opExtentBits = ImmBits, isPredicable = 1 in def NAME#_nv_V4 : NVInst_V4<(outs), (ins MEMri:$addr, RC:$src), mnemonic#"($addr) = $src.new", []>, Requires<[HasV4T]>; let opExtendable = 2, isExtentSigned = 0, opExtentBits = PredImmBits, neverHasSideEffects = 1, isPredicated = 1 in { defm Pt : ST_MEMri_Pred_nv; defm NotPt : ST_MEMri_Pred_nv; } } } let addrMode = BaseImmOffset, isMEMri = "true", validSubTargets = HasV4SubT, mayStore = 1 in { defm STrib: ST_MEMri_nv<"memb", "STrib", IntRegs, 11, 6>, AddrModeRel; defm STrih: ST_MEMri_nv<"memh", "STrih", IntRegs, 12, 7>, AddrModeRel; defm STriw: ST_MEMri_nv<"memw", "STriw", IntRegs, 13, 8>, AddrModeRel; } // memb(Ru<<#u2+#U6)=Nt.new let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10, isNVStore = 1, validSubTargets = HasV4SubT in def STrib_shl_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memb($src1<<#$src2+#$src3) = $src4.new", []>, Requires<[HasV4T]>; //===----------------------------------------------------------------------===// // Post increment store // mem[bhwd](Rx++#s4:[0123])=Nt.new //===----------------------------------------------------------------------===// multiclass ST_PostInc_Pbase_nv { let PNewValue = !if(isPredNew, "new", "") in def NAME#_nv_V4 : NVInstPI_V4<(outs IntRegs:$dst), (ins PredRegs:$src1, IntRegs:$src2, ImmOp:$offset, RC:$src3), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"($src2++#$offset) = $src3.new", [], "$src2 = $dst">, Requires<[HasV4T]>; } multiclass ST_PostInc_Pred_nv { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_PostInc_Pbase_nv; // Predicate new let Predicates = [HasV4T], validSubTargets = HasV4SubT in defm _cdn#NAME : ST_PostInc_Pbase_nv; } } let hasCtrlDep = 1, isNVStore = 1, neverHasSideEffects = 1 in multiclass ST_PostInc_nv { let BaseOpcode = "POST_"#BaseOp in { let isPredicable = 1 in def NAME#_nv_V4 : NVInstPI_V4<(outs IntRegs:$dst), (ins IntRegs:$src1, ImmOp:$offset, RC:$src2), mnemonic#"($src1++#$offset) = $src2.new", [], "$src1 = $dst">, Requires<[HasV4T]>; let isPredicated = 1 in { defm Pt : ST_PostInc_Pred_nv; defm NotPt : ST_PostInc_Pred_nv; } } } let validSubTargets = HasV4SubT in { defm POST_STbri: ST_PostInc_nv <"memb", "STrib", IntRegs, s4_0Imm>, AddrModeRel; defm POST_SThri: ST_PostInc_nv <"memh", "STrih", IntRegs, s4_1Imm>, AddrModeRel; defm POST_STwri: ST_PostInc_nv <"memw", "STriw", IntRegs, s4_2Imm>, AddrModeRel; } // memb(Rx++#s4:0:circ(Mu))=Nt.new // memb(Rx++I:circ(Mu))=Nt.new // memb(Rx++Mu)=Nt.new // memb(Rx++Mu:brev)=Nt.new // memb(#global)=Nt.new let mayStore = 1, neverHasSideEffects = 1 in def STb_GP_nv_V4 : NVInst_V4<(outs), (ins globaladdress:$global, IntRegs:$src), "memb(#$global) = $src.new", []>, Requires<[HasV4T]>; // memh(Ru<<#u2+#U6)=Nt.new let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10, isNVStore = 1, validSubTargets = HasV4SubT in def STrih_shl_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memh($src1<<#$src2+#$src3) = $src4.new", []>, Requires<[HasV4T]>; // memh(Rx++#s4:1:circ(Mu))=Nt.new // memh(Rx++I:circ(Mu))=Nt.new // memh(Rx++Mu)=Nt.new // memh(Rx++Mu:brev)=Nt.new // memh(#global)=Nt.new let mayStore = 1, neverHasSideEffects = 1 in def STh_GP_nv_V4 : NVInst_V4<(outs), (ins globaladdress:$global, IntRegs:$src), "memh(#$global) = $src.new", []>, Requires<[HasV4T]>; // memw(Ru<<#u2+#U6)=Nt.new let isExtended = 1, opExtendable = 2, mayStore = 1, AddedComplexity = 10, isNVStore = 1, validSubTargets = HasV4SubT in def STriw_shl_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u2Imm:$src2, u0AlwaysExt:$src3, IntRegs:$src4), "memw($src1<<#$src2+#$src3) = $src4.new", []>, Requires<[HasV4T]>; // memw(Rx++#s4:2:circ(Mu))=Nt.new // memw(Rx++I:circ(Mu))=Nt.new // memw(Rx++Mu)=Nt.new // memw(Rx++Mu:brev)=Nt.new // memw(gp+#u16:2)=Nt.new let mayStore = 1, neverHasSideEffects = 1, isNVStore = 1, validSubTargets = HasV4SubT in def STw_GP_nv_V4 : NVInst_V4<(outs), (ins globaladdress:$global, IntRegs:$src), "memw(#$global) = $src.new", []>, Requires<[HasV4T]>; // if (Pv) memb(##global) = Rt let mayStore = 1, neverHasSideEffects = 1, isNVStore = 1, isExtended = 1, opExtendable = 1, validSubTargets = HasV4SubT in { def STb_GP_cPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memb(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memb(##global) = Rt def STb_GP_cNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memb(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (Pv) memb(##global) = Rt def STb_GP_cdnPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memb(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memb(##global) = Rt def STb_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memb(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (Pv) memh(##global) = Rt def STh_GP_cPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memh(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memh(##global) = Rt def STh_GP_cNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memh(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (Pv) memh(##global) = Rt def STh_GP_cdnPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memh(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memh(##global) = Rt def STh_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memh(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (Pv) memw(##global) = Rt def STw_GP_cPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1) memw(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memw(##global) = Rt def STw_GP_cNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1) memw(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (Pv) memw(##global) = Rt def STw_GP_cdnPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if ($src1.new) memw(##$global) = $src2.new", []>, Requires<[HasV4T]>; // if (!Pv) memw(##global) = Rt def STw_GP_cdnNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdress:$global, IntRegs:$src2), "if (!$src1.new) memw(##$global) = $src2.new", []>, Requires<[HasV4T]>; } //===----------------------------------------------------------------------===// // NV/ST - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // NV/J + //===----------------------------------------------------------------------===// multiclass NVJ_type_basic_reg { def _ie_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, $src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; def _nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, $src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; } multiclass NVJ_type_basic_2ndDotNew { def _ie_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1, $src2.new)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; def _nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, IntRegs:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1, $src2.new)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; } multiclass NVJ_type_basic_imm { def _ie_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u5Imm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; def _nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u5Imm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; } multiclass NVJ_type_basic_neg { def _ie_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, nOneImm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; def _nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, nOneImm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; } multiclass NVJ_type_basic_tstbit { def _ie_nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u1Imm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; def _nv_V4 : NVInst_V4<(outs), (ins IntRegs:$src1, u1Imm:$src2, brtarget:$offset), !strconcat("if (", !strconcat(NotStr, !strconcat(OpcStr, !strconcat("($src1.new, #$src2)) jump:", !strconcat(TakenStr, " $offset"))))), []>, Requires<[HasV4T]>; } // Multiclass for regular dot new of Ist operand register. multiclass NVJ_type_br_pred_reg { defm Pt : NVJ_type_basic_reg; defm Pnt : NVJ_type_basic_reg; } // Multiclass for dot new of 2nd operand register. multiclass NVJ_type_br_pred_2ndDotNew { defm Pt : NVJ_type_basic_2ndDotNew; defm Pnt : NVJ_type_basic_2ndDotNew; } // Multiclass for 2nd operand immediate, including -1. multiclass NVJ_type_br_pred_imm { defm Pt : NVJ_type_basic_imm; defm Pnt : NVJ_type_basic_imm; defm Ptneg : NVJ_type_basic_neg; defm Pntneg : NVJ_type_basic_neg; } // Multiclass for 2nd operand immediate, excluding -1. multiclass NVJ_type_br_pred_imm_only { defm Pt : NVJ_type_basic_imm; defm Pnt : NVJ_type_basic_imm; } // Multiclass for tstbit, where 2nd operand is always #0. multiclass NVJ_type_br_pred_tstbit { defm Pt : NVJ_type_basic_tstbit; defm Pnt : NVJ_type_basic_tstbit; } // Multiclass for GT. multiclass NVJ_type_rr_ri { defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>; defm rr : NVJ_type_br_pred_reg<"", OpcStr>; defm rrdnNot : NVJ_type_br_pred_2ndDotNew<"!", OpcStr>; defm rrdn : NVJ_type_br_pred_2ndDotNew<"", OpcStr>; defm riNot : NVJ_type_br_pred_imm<"!", OpcStr>; defm ri : NVJ_type_br_pred_imm<"", OpcStr>; } // Multiclass for EQ. multiclass NVJ_type_rr_ri_no_2ndDotNew { defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>; defm rr : NVJ_type_br_pred_reg<"", OpcStr>; defm riNot : NVJ_type_br_pred_imm<"!", OpcStr>; defm ri : NVJ_type_br_pred_imm<"", OpcStr>; } // Multiclass for GTU. multiclass NVJ_type_rr_ri_no_nOne { defm rrNot : NVJ_type_br_pred_reg<"!", OpcStr>; defm rr : NVJ_type_br_pred_reg<"", OpcStr>; defm rrdnNot : NVJ_type_br_pred_2ndDotNew<"!", OpcStr>; defm rrdn : NVJ_type_br_pred_2ndDotNew<"", OpcStr>; defm riNot : NVJ_type_br_pred_imm_only<"!", OpcStr>; defm ri : NVJ_type_br_pred_imm_only<"", OpcStr>; } // Multiclass for tstbit. multiclass NVJ_type_r0 { defm r0Not : NVJ_type_br_pred_tstbit<"!", OpcStr>; defm r0 : NVJ_type_br_pred_tstbit<"", OpcStr>; } // Base Multiclass for New Value Jump. multiclass NVJ_type { defm GT : NVJ_type_rr_ri<"cmp.gt">; defm EQ : NVJ_type_rr_ri_no_2ndDotNew<"cmp.eq">; defm GTU : NVJ_type_rr_ri_no_nOne<"cmp.gtu">; defm TSTBIT : NVJ_type_r0<"tstbit">; } let isBranch = 1, isTerminator=1, neverHasSideEffects = 1, Defs = [PC] in { defm JMP_ : NVJ_type; } //===----------------------------------------------------------------------===// // NV/J - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // XTYPE/ALU + //===----------------------------------------------------------------------===// // Add and accumulate. // Rd=add(Rs,add(Ru,#s6)) let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 6, validSubTargets = HasV4SubT in def ADDr_ADDri_V4 : MInst<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2, s6Ext:$src3), "$dst = add($src1, add($src2, #$src3))", [(set (i32 IntRegs:$dst), (add (i32 IntRegs:$src1), (add (i32 IntRegs:$src2), s6_16ExtPred:$src3)))]>, Requires<[HasV4T]>; // Rd=add(Rs,sub(#s6,Ru)) let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 6, validSubTargets = HasV4SubT in def ADDr_SUBri_V4 : MInst<(outs IntRegs:$dst), (ins IntRegs:$src1, s6Ext:$src2, IntRegs:$src3), "$dst = add($src1, sub(#$src2, $src3))", [(set (i32 IntRegs:$dst), (add (i32 IntRegs:$src1), (sub s6_10ExtPred:$src2, (i32 IntRegs:$src3))))]>, Requires<[HasV4T]>; // Generates the same instruction as ADDr_SUBri_V4 but matches different // pattern. // Rd=add(Rs,sub(#s6,Ru)) let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 6, validSubTargets = HasV4SubT in def ADDri_SUBr_V4 : MInst<(outs IntRegs:$dst), (ins IntRegs:$src1, s6Ext:$src2, IntRegs:$src3), "$dst = add($src1, sub(#$src2, $src3))", [(set (i32 IntRegs:$dst), (sub (add (i32 IntRegs:$src1), s6_10ExtPred:$src2), (i32 IntRegs:$src3)))]>, Requires<[HasV4T]>; // Add or subtract doublewords with carry. //TODO: // Rdd=add(Rss,Rtt,Px):carry //TODO: // Rdd=sub(Rss,Rtt,Px):carry // Logical doublewords. // Rdd=and(Rtt,~Rss) let validSubTargets = HasV4SubT in def ANDd_NOTd_V4 : MInst<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2), "$dst = and($src1, ~$src2)", [(set (i64 DoubleRegs:$dst), (and (i64 DoubleRegs:$src1), (not (i64 DoubleRegs:$src2))))]>, Requires<[HasV4T]>; // Rdd=or(Rtt,~Rss) let validSubTargets = HasV4SubT in def ORd_NOTd_V4 : MInst<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2), "$dst = or($src1, ~$src2)", [(set (i64 DoubleRegs:$dst), (or (i64 DoubleRegs:$src1), (not (i64 DoubleRegs:$src2))))]>, Requires<[HasV4T]>; // Logical-logical doublewords. // Rxx^=xor(Rss,Rtt) let validSubTargets = HasV4SubT in def XORd_XORdd: MInst_acc<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2, DoubleRegs:$src3), "$dst ^= xor($src2, $src3)", [(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1), (xor (i64 DoubleRegs:$src2), (i64 DoubleRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Logical-logical words. // Rx=or(Ru,and(Rx,#s10)) let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10, validSubTargets = HasV4SubT in def ORr_ANDri_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3), "$dst = or($src1, and($src2, #$src3))", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), s10ExtPred:$src3)))], "$src2 = $dst">, Requires<[HasV4T]>; // Rx[&|^]=and(Rs,Rt) // Rx&=and(Rs,Rt) let validSubTargets = HasV4SubT in def ANDr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst &= and($src2, $src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx|=and(Rs,Rt) let validSubTargets = HasV4SubT, CextOpcode = "ORr_ANDr", InputType = "reg" in def ORr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst |= and($src2, $src3)", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>, ImmRegRel; // Rx^=and(Rs,Rt) let validSubTargets = HasV4SubT in def XORr_ANDrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst ^= and($src2, $src3)", [(set (i32 IntRegs:$dst), (xor (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx[&|^]=and(Rs,~Rt) // Rx&=and(Rs,~Rt) let validSubTargets = HasV4SubT in def ANDr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst &= and($src2, ~$src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (not (i32 IntRegs:$src3)))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx|=and(Rs,~Rt) let validSubTargets = HasV4SubT in def ORr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst |= and($src2, ~$src3)", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (not (i32 IntRegs:$src3)))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx^=and(Rs,~Rt) let validSubTargets = HasV4SubT in def XORr_ANDr_NOTr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst ^= and($src2, ~$src3)", [(set (i32 IntRegs:$dst), (xor (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), (not (i32 IntRegs:$src3)))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx[&|^]=or(Rs,Rt) // Rx&=or(Rs,Rt) let validSubTargets = HasV4SubT in def ANDr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst &= or($src2, $src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (or (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx|=or(Rs,Rt) let validSubTargets = HasV4SubT, CextOpcode = "ORr_ORr", InputType = "reg" in def ORr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst |= or($src2, $src3)", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (or (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>, ImmRegRel; // Rx^=or(Rs,Rt) let validSubTargets = HasV4SubT in def XORr_ORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst ^= or($src2, $src3)", [(set (i32 IntRegs:$dst), (xor (i32 IntRegs:$src1), (or (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx[&|^]=xor(Rs,Rt) // Rx&=xor(Rs,Rt) let validSubTargets = HasV4SubT in def ANDr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst &= xor($src2, $src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx|=xor(Rs,Rt) let validSubTargets = HasV4SubT in def ORr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst |= xor($src2, $src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx^=xor(Rs,Rt) let validSubTargets = HasV4SubT in def XORr_XORrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, IntRegs:$src3), "$dst ^= xor($src2, $src3)", [(set (i32 IntRegs:$dst), (and (i32 IntRegs:$src1), (xor (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; // Rx|=and(Rs,#s10) let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10, validSubTargets = HasV4SubT, CextOpcode = "ORr_ANDr", InputType = "imm" in def ORr_ANDri2_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3), "$dst |= and($src2, #$src3)", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), s10ExtPred:$src3)))], "$src1 = $dst">, Requires<[HasV4T]>, ImmRegRel; // Rx|=or(Rs,#s10) let isExtendable = 1, opExtendable = 3, isExtentSigned = 1, opExtentBits = 10, validSubTargets = HasV4SubT, CextOpcode = "ORr_ORr", InputType = "imm" in def ORr_ORri_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs: $src2, s10Ext:$src3), "$dst |= or($src2, #$src3)", [(set (i32 IntRegs:$dst), (or (i32 IntRegs:$src1), (and (i32 IntRegs:$src2), s10ExtPred:$src3)))], "$src1 = $dst">, Requires<[HasV4T]>, ImmRegRel; // Modulo wrap // Rd=modwrap(Rs,Rt) // Round // Rd=cround(Rs,#u5) // Rd=cround(Rs,Rt) // Rd=round(Rs,#u5)[:sat] // Rd=round(Rs,Rt)[:sat] // Vector reduce add unsigned halfwords // Rd=vraddh(Rss,Rtt) // Vector add bytes // Rdd=vaddb(Rss,Rtt) // Vector conditional negate // Rdd=vcnegh(Rss,Rt) // Rxx+=vrcnegh(Rss,Rt) // Vector maximum bytes // Rdd=vmaxb(Rtt,Rss) // Vector reduce maximum halfwords // Rxx=vrmaxh(Rss,Ru) // Rxx=vrmaxuh(Rss,Ru) // Vector reduce maximum words // Rxx=vrmaxuw(Rss,Ru) // Rxx=vrmaxw(Rss,Ru) // Vector minimum bytes // Rdd=vminb(Rtt,Rss) // Vector reduce minimum halfwords // Rxx=vrminh(Rss,Ru) // Rxx=vrminuh(Rss,Ru) // Vector reduce minimum words // Rxx=vrminuw(Rss,Ru) // Rxx=vrminw(Rss,Ru) // Vector subtract bytes // Rdd=vsubb(Rss,Rtt) //===----------------------------------------------------------------------===// // XTYPE/ALU - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // XTYPE/MPY + //===----------------------------------------------------------------------===// // Multiply and user lower result. // Rd=add(#u6,mpyi(Rs,#U6)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 6, validSubTargets = HasV4SubT in def ADDi_MPYri_V4 : MInst<(outs IntRegs:$dst), (ins u6Ext:$src1, IntRegs:$src2, u6Imm:$src3), "$dst = add(#$src1, mpyi($src2, #$src3))", [(set (i32 IntRegs:$dst), (add (mul (i32 IntRegs:$src2), u6ImmPred:$src3), u6ExtPred:$src1))]>, Requires<[HasV4T]>; // Rd=add(##,mpyi(Rs,#U6)) def : Pat <(add (mul (i32 IntRegs:$src2), u6ImmPred:$src3), (HexagonCONST32 tglobaladdr:$src1)), (i32 (ADDi_MPYri_V4 tglobaladdr:$src1, IntRegs:$src2, u6ImmPred:$src3))>; // Rd=add(#u6,mpyi(Rs,Rt)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 6, validSubTargets = HasV4SubT, InputType = "imm", CextOpcode = "ADD_MPY" in def ADDi_MPYrr_V4 : MInst<(outs IntRegs:$dst), (ins u6Ext:$src1, IntRegs:$src2, IntRegs:$src3), "$dst = add(#$src1, mpyi($src2, $src3))", [(set (i32 IntRegs:$dst), (add (mul (i32 IntRegs:$src2), (i32 IntRegs:$src3)), u6ExtPred:$src1))]>, Requires<[HasV4T]>, ImmRegRel; // Rd=add(##,mpyi(Rs,Rt)) def : Pat <(add (mul (i32 IntRegs:$src2), (i32 IntRegs:$src3)), (HexagonCONST32 tglobaladdr:$src1)), (i32 (ADDi_MPYrr_V4 tglobaladdr:$src1, IntRegs:$src2, IntRegs:$src3))>; // Rd=add(Ru,mpyi(#u6:2,Rs)) let validSubTargets = HasV4SubT in def ADDr_MPYir_V4 : MInst<(outs IntRegs:$dst), (ins IntRegs:$src1, u6Imm:$src2, IntRegs:$src3), "$dst = add($src1, mpyi(#$src2, $src3))", [(set (i32 IntRegs:$dst), (add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src3), u6_2ImmPred:$src2)))]>, Requires<[HasV4T]>; // Rd=add(Ru,mpyi(Rs,#u6)) let isExtendable = 1, opExtendable = 3, isExtentSigned = 0, opExtentBits = 6, validSubTargets = HasV4SubT, InputType = "imm", CextOpcode = "ADD_MPY" in def ADDr_MPYri_V4 : MInst<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2, u6Ext:$src3), "$dst = add($src1, mpyi($src2, #$src3))", [(set (i32 IntRegs:$dst), (add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src2), u6ExtPred:$src3)))]>, Requires<[HasV4T]>, ImmRegRel; // Rx=add(Ru,mpyi(Rx,Rs)) let validSubTargets = HasV4SubT, InputType = "reg", CextOpcode = "ADD_MPY" in def ADDr_MPYrr_V4 : MInst_acc<(outs IntRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2, IntRegs:$src3), "$dst = add($src1, mpyi($src2, $src3))", [(set (i32 IntRegs:$dst), (add (i32 IntRegs:$src1), (mul (i32 IntRegs:$src2), (i32 IntRegs:$src3))))], "$src2 = $dst">, Requires<[HasV4T]>, ImmRegRel; // Polynomial multiply words // Rdd=pmpyw(Rs,Rt) // Rxx^=pmpyw(Rs,Rt) // Vector reduce multiply word by signed half (32x16) // Rdd=vrmpyweh(Rss,Rtt)[:<<1] // Rdd=vrmpywoh(Rss,Rtt)[:<<1] // Rxx+=vrmpyweh(Rss,Rtt)[:<<1] // Rxx+=vrmpywoh(Rss,Rtt)[:<<1] // Multiply and use upper result // Rd=mpy(Rs,Rt.H):<<1:sat // Rd=mpy(Rs,Rt.L):<<1:sat // Rd=mpy(Rs,Rt):<<1 // Rd=mpy(Rs,Rt):<<1:sat // Rd=mpysu(Rs,Rt) // Rx+=mpy(Rs,Rt):<<1:sat // Rx-=mpy(Rs,Rt):<<1:sat // Vector multiply bytes // Rdd=vmpybsu(Rs,Rt) // Rdd=vmpybu(Rs,Rt) // Rxx+=vmpybsu(Rs,Rt) // Rxx+=vmpybu(Rs,Rt) // Vector polynomial multiply halfwords // Rdd=vpmpyh(Rs,Rt) // Rxx^=vpmpyh(Rs,Rt) //===----------------------------------------------------------------------===// // XTYPE/MPY - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // XTYPE/SHIFT + //===----------------------------------------------------------------------===// // Shift by immediate and accumulate. // Rx=add(#u8,asl(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def ADDi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = add(#$src1, asl($src2, #$src3))", [(set (i32 IntRegs:$dst), (add (shl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; // Rx=add(#u8,lsr(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def ADDi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = add(#$src1, lsr($src2, #$src3))", [(set (i32 IntRegs:$dst), (add (srl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; // Rx=sub(#u8,asl(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def SUBi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = sub(#$src1, asl($src2, #$src3))", [(set (i32 IntRegs:$dst), (sub (shl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; // Rx=sub(#u8,lsr(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def SUBi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = sub(#$src1, lsr($src2, #$src3))", [(set (i32 IntRegs:$dst), (sub (srl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; //Shift by immediate and logical. //Rx=and(#u8,asl(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def ANDi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = and(#$src1, asl($src2, #$src3))", [(set (i32 IntRegs:$dst), (and (shl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; //Rx=and(#u8,lsr(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, validSubTargets = HasV4SubT in def ANDi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = and(#$src1, lsr($src2, #$src3))", [(set (i32 IntRegs:$dst), (and (srl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; //Rx=or(#u8,asl(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, AddedComplexity = 30, validSubTargets = HasV4SubT in def ORi_ASLri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = or(#$src1, asl($src2, #$src3))", [(set (i32 IntRegs:$dst), (or (shl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; //Rx=or(#u8,lsr(Rx,#U5)) let isExtendable = 1, opExtendable = 1, isExtentSigned = 0, opExtentBits = 8, AddedComplexity = 30, validSubTargets = HasV4SubT in def ORi_LSRri_V4 : MInst_acc<(outs IntRegs:$dst), (ins u8Ext:$src1, IntRegs:$src2, u5Imm:$src3), "$dst = or(#$src1, lsr($src2, #$src3))", [(set (i32 IntRegs:$dst), (or (srl (i32 IntRegs:$src2), u5ImmPred:$src3), u8ExtPred:$src1))], "$src2 = $dst">, Requires<[HasV4T]>; //Shift by register. //Rd=lsl(#s6,Rt) let validSubTargets = HasV4SubT in { def LSLi_V4 : MInst<(outs IntRegs:$dst), (ins s6Imm:$src1, IntRegs:$src2), "$dst = lsl(#$src1, $src2)", [(set (i32 IntRegs:$dst), (shl s6ImmPred:$src1, (i32 IntRegs:$src2)))]>, Requires<[HasV4T]>; //Shift by register and logical. //Rxx^=asl(Rss,Rt) def ASLd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3), "$dst ^= asl($src2, $src3)", [(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1), (shl (i64 DoubleRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; //Rxx^=asr(Rss,Rt) def ASRd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3), "$dst ^= asr($src2, $src3)", [(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1), (sra (i64 DoubleRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; //Rxx^=lsl(Rss,Rt) def LSLd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3), "$dst ^= lsl($src2, $src3)", [(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1), (shl (i64 DoubleRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; //Rxx^=lsr(Rss,Rt) def LSRd_rr_xor_V4 : MInst_acc<(outs DoubleRegs:$dst), (ins DoubleRegs:$src1, DoubleRegs:$src2, IntRegs:$src3), "$dst ^= lsr($src2, $src3)", [(set (i64 DoubleRegs:$dst), (xor (i64 DoubleRegs:$src1), (srl (i64 DoubleRegs:$src2), (i32 IntRegs:$src3))))], "$src1 = $dst">, Requires<[HasV4T]>; } //===----------------------------------------------------------------------===// // XTYPE/SHIFT - //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // MEMOP: Word, Half, Byte //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // MEMOP: Word // // Implemented: // MEMw_ADDi_indexed_V4 : memw(Rs+#u6:2)+=#U5 // MEMw_SUBi_indexed_V4 : memw(Rs+#u6:2)-=#U5 // MEMw_ADDr_indexed_V4 : memw(Rs+#u6:2)+=Rt // MEMw_SUBr_indexed_V4 : memw(Rs+#u6:2)-=Rt // MEMw_CLRr_indexed_V4 : memw(Rs+#u6:2)&=Rt // MEMw_SETr_indexed_V4 : memw(Rs+#u6:2)|=Rt // MEMw_ADDi_V4 : memw(Rs+#u6:2)+=#U5 // MEMw_SUBi_V4 : memw(Rs+#u6:2)-=#U5 // MEMw_ADDr_V4 : memw(Rs+#u6:2)+=Rt // MEMw_SUBr_V4 : memw(Rs+#u6:2)-=Rt // MEMw_CLRr_V4 : memw(Rs+#u6:2)&=Rt // MEMw_SETr_V4 : memw(Rs+#u6:2)|=Rt // // Not implemented: // MEMw_CLRi_indexed_V4 : memw(Rs+#u6:2)=clrbit(#U5) // MEMw_SETi_indexed_V4 : memw(Rs+#u6:2)=setbit(#U5) // MEMw_CLRi_V4 : memw(Rs+#u6:2)=clrbit(#U5) // MEMw_SETi_V4 : memw(Rs+#u6:2)=setbit(#U5) //===----------------------------------------------------------------------===// // memw(Rs+#u6:2) += #U5 let AddedComplexity = 30 in def MEMw_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, u5Imm:$addend), "memw($base+#$offset) += #$addend", []>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) -= #U5 let AddedComplexity = 30 in def MEMw_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, u5Imm:$subend), "memw($base+#$offset) -= #$subend", []>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) += Rt let AddedComplexity = 30 in def MEMw_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$addend), "memw($base+#$offset) += $addend", [(store (add (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)), (i32 IntRegs:$addend)), (add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) -= Rt let AddedComplexity = 30 in def MEMw_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$subend), "memw($base+#$offset) -= $subend", [(store (sub (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)), (i32 IntRegs:$subend)), (add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) &= Rt let AddedComplexity = 30 in def MEMw_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$andend), "memw($base+#$offset) &= $andend", [(store (and (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)), (i32 IntRegs:$andend)), (add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) |= Rt let AddedComplexity = 30 in def MEMw_ORr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_2Imm:$offset, IntRegs:$orend), "memw($base+#$offset) |= $orend", [(store (or (load (add (i32 IntRegs:$base), u6_2ImmPred:$offset)), (i32 IntRegs:$orend)), (add (i32 IntRegs:$base), u6_2ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) += #U5 let AddedComplexity = 30 in def MEMw_ADDi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$addend), "memw($addr) += $addend", []>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) -= #U5 let AddedComplexity = 30 in def MEMw_SUBi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$subend), "memw($addr) -= $subend", []>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) += Rt let AddedComplexity = 30 in def MEMw_ADDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$addend), "memw($addr) += $addend", [(store (add (load ADDRriU6_2:$addr), (i32 IntRegs:$addend)), ADDRriU6_2:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) -= Rt let AddedComplexity = 30 in def MEMw_SUBr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$subend), "memw($addr) -= $subend", [(store (sub (load ADDRriU6_2:$addr), (i32 IntRegs:$subend)), ADDRriU6_2:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) &= Rt let AddedComplexity = 30 in def MEMw_ANDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$andend), "memw($addr) &= $andend", [(store (and (load ADDRriU6_2:$addr), (i32 IntRegs:$andend)), ADDRriU6_2:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memw(Rs+#u6:2) |= Rt let AddedComplexity = 30 in def MEMw_ORr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$orend), "memw($addr) |= $orend", [(store (or (load ADDRriU6_2:$addr), (i32 IntRegs:$orend)), ADDRriU6_2:$addr)]>, Requires<[HasV4T, UseMEMOP]>; //===----------------------------------------------------------------------===// // MEMOP: Halfword // // Implemented: // MEMh_ADDi_indexed_V4 : memw(Rs+#u6:2)+=#U5 // MEMh_SUBi_indexed_V4 : memw(Rs+#u6:2)-=#U5 // MEMh_ADDr_indexed_V4 : memw(Rs+#u6:2)+=Rt // MEMh_SUBr_indexed_V4 : memw(Rs+#u6:2)-=Rt // MEMh_CLRr_indexed_V4 : memw(Rs+#u6:2)&=Rt // MEMh_SETr_indexed_V4 : memw(Rs+#u6:2)|=Rt // MEMh_ADDi_V4 : memw(Rs+#u6:2)+=#U5 // MEMh_SUBi_V4 : memw(Rs+#u6:2)-=#U5 // MEMh_ADDr_V4 : memw(Rs+#u6:2)+=Rt // MEMh_SUBr_V4 : memw(Rs+#u6:2)-=Rt // MEMh_CLRr_V4 : memw(Rs+#u6:2)&=Rt // MEMh_SETr_V4 : memw(Rs+#u6:2)|=Rt // // Not implemented: // MEMh_CLRi_indexed_V4 : memw(Rs+#u6:2)=clrbit(#U5) // MEMh_SETi_indexed_V4 : memw(Rs+#u6:2)=setbit(#U5) // MEMh_CLRi_V4 : memw(Rs+#u6:2)=clrbit(#U5) // MEMh_SETi_V4 : memw(Rs+#u6:2)=setbit(#U5) //===----------------------------------------------------------------------===// // memh(Rs+#u6:1) += #U5 let AddedComplexity = 30 in def MEMh_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, u5Imm:$addend), "memh($base+#$offset) += $addend", []>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) -= #U5 let AddedComplexity = 30 in def MEMh_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, u5Imm:$subend), "memh($base+#$offset) -= $subend", []>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) += Rt let AddedComplexity = 30 in def MEMh_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$addend), "memh($base+#$offset) += $addend", [(truncstorei16 (add (sextloadi16 (add (i32 IntRegs:$base), u6_1ImmPred:$offset)), (i32 IntRegs:$addend)), (add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) -= Rt let AddedComplexity = 30 in def MEMh_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$subend), "memh($base+#$offset) -= $subend", [(truncstorei16 (sub (sextloadi16 (add (i32 IntRegs:$base), u6_1ImmPred:$offset)), (i32 IntRegs:$subend)), (add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) &= Rt let AddedComplexity = 30 in def MEMh_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$andend), "memh($base+#$offset) += $andend", [(truncstorei16 (and (sextloadi16 (add (i32 IntRegs:$base), u6_1ImmPred:$offset)), (i32 IntRegs:$andend)), (add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) |= Rt let AddedComplexity = 30 in def MEMh_ORr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_1Imm:$offset, IntRegs:$orend), "memh($base+#$offset) |= $orend", [(truncstorei16 (or (sextloadi16 (add (i32 IntRegs:$base), u6_1ImmPred:$offset)), (i32 IntRegs:$orend)), (add (i32 IntRegs:$base), u6_1ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) += #U5 let AddedComplexity = 30 in def MEMh_ADDi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$addend), "memh($addr) += $addend", []>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) -= #U5 let AddedComplexity = 30 in def MEMh_SUBi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$subend), "memh($addr) -= $subend", []>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) += Rt let AddedComplexity = 30 in def MEMh_ADDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$addend), "memh($addr) += $addend", [(truncstorei16 (add (sextloadi16 ADDRriU6_1:$addr), (i32 IntRegs:$addend)), ADDRriU6_1:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) -= Rt let AddedComplexity = 30 in def MEMh_SUBr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$subend), "memh($addr) -= $subend", [(truncstorei16 (sub (sextloadi16 ADDRriU6_1:$addr), (i32 IntRegs:$subend)), ADDRriU6_1:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) &= Rt let AddedComplexity = 30 in def MEMh_ANDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$andend), "memh($addr) &= $andend", [(truncstorei16 (and (sextloadi16 ADDRriU6_1:$addr), (i32 IntRegs:$andend)), ADDRriU6_1:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memh(Rs+#u6:1) |= Rt let AddedComplexity = 30 in def MEMh_ORr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$orend), "memh($addr) |= $orend", [(truncstorei16 (or (sextloadi16 ADDRriU6_1:$addr), (i32 IntRegs:$orend)), ADDRriU6_1:$addr)]>, Requires<[HasV4T, UseMEMOP]>; //===----------------------------------------------------------------------===// // MEMOP: Byte // // Implemented: // MEMb_ADDi_indexed_V4 : memb(Rs+#u6:0)+=#U5 // MEMb_SUBi_indexed_V4 : memb(Rs+#u6:0)-=#U5 // MEMb_ADDr_indexed_V4 : memb(Rs+#u6:0)+=Rt // MEMb_SUBr_indexed_V4 : memb(Rs+#u6:0)-=Rt // MEMb_CLRr_indexed_V4 : memb(Rs+#u6:0)&=Rt // MEMb_SETr_indexed_V4 : memb(Rs+#u6:0)|=Rt // MEMb_ADDi_V4 : memb(Rs+#u6:0)+=#U5 // MEMb_SUBi_V4 : memb(Rs+#u6:0)-=#U5 // MEMb_ADDr_V4 : memb(Rs+#u6:0)+=Rt // MEMb_SUBr_V4 : memb(Rs+#u6:0)-=Rt // MEMb_CLRr_V4 : memb(Rs+#u6:0)&=Rt // MEMb_SETr_V4 : memb(Rs+#u6:0)|=Rt // // Not implemented: // MEMb_CLRi_indexed_V4 : memb(Rs+#u6:0)=clrbit(#U5) // MEMb_SETi_indexed_V4 : memb(Rs+#u6:0)=setbit(#U5) // MEMb_CLRi_V4 : memb(Rs+#u6:0)=clrbit(#U5) // MEMb_SETi_V4 : memb(Rs+#u6:0)=setbit(#U5) //===----------------------------------------------------------------------===// // memb(Rs+#u6:0) += #U5 let AddedComplexity = 30 in def MEMb_ADDi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, u5Imm:$addend), "memb($base+#$offset) += $addend", []>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) -= #U5 let AddedComplexity = 30 in def MEMb_SUBi_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, u5Imm:$subend), "memb($base+#$offset) -= $subend", []>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) += Rt let AddedComplexity = 30 in def MEMb_ADDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$addend), "memb($base+#$offset) += $addend", [(truncstorei8 (add (sextloadi8 (add (i32 IntRegs:$base), u6_0ImmPred:$offset)), (i32 IntRegs:$addend)), (add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) -= Rt let AddedComplexity = 30 in def MEMb_SUBr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$subend), "memb($base+#$offset) -= $subend", [(truncstorei8 (sub (sextloadi8 (add (i32 IntRegs:$base), u6_0ImmPred:$offset)), (i32 IntRegs:$subend)), (add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) &= Rt let AddedComplexity = 30 in def MEMb_ANDr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$andend), "memb($base+#$offset) += $andend", [(truncstorei8 (and (sextloadi8 (add (i32 IntRegs:$base), u6_0ImmPred:$offset)), (i32 IntRegs:$andend)), (add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) |= Rt let AddedComplexity = 30 in def MEMb_ORr_indexed_MEM_V4 : MEMInst_V4<(outs), (ins IntRegs:$base, u6_0Imm:$offset, IntRegs:$orend), "memb($base+#$offset) |= $orend", [(truncstorei8 (or (sextloadi8 (add (i32 IntRegs:$base), u6_0ImmPred:$offset)), (i32 IntRegs:$orend)), (add (i32 IntRegs:$base), u6_0ImmPred:$offset))]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) += #U5 let AddedComplexity = 30 in def MEMb_ADDi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$addend), "memb($addr) += $addend", []>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) -= #U5 let AddedComplexity = 30 in def MEMb_SUBi_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, u5Imm:$subend), "memb($addr) -= $subend", []>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) += Rt let AddedComplexity = 30 in def MEMb_ADDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$addend), "memb($addr) += $addend", [(truncstorei8 (add (sextloadi8 ADDRriU6_0:$addr), (i32 IntRegs:$addend)), ADDRriU6_0:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) -= Rt let AddedComplexity = 30 in def MEMb_SUBr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$subend), "memb($addr) -= $subend", [(truncstorei8 (sub (sextloadi8 ADDRriU6_0:$addr), (i32 IntRegs:$subend)), ADDRriU6_0:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) &= Rt let AddedComplexity = 30 in def MEMb_ANDr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$andend), "memb($addr) &= $andend", [(truncstorei8 (and (sextloadi8 ADDRriU6_0:$addr), (i32 IntRegs:$andend)), ADDRriU6_0:$addr)]>, Requires<[HasV4T, UseMEMOP]>; // memb(Rs+#u6:0) |= Rt let AddedComplexity = 30 in def MEMb_ORr_MEM_V4 : MEMInst_V4<(outs), (ins MEMri:$addr, IntRegs:$orend), "memb($addr) |= $orend", [(truncstorei8 (or (sextloadi8 ADDRriU6_0:$addr), (i32 IntRegs:$orend)), ADDRriU6_0:$addr)]>, Requires<[HasV4T, UseMEMOP]>; //===----------------------------------------------------------------------===// // XTYPE/PRED + //===----------------------------------------------------------------------===// // Hexagon V4 only supports these flavors of byte/half compare instructions: // EQ/GT/GTU. Other flavors like GE/GEU/LT/LTU/LE/LEU are not supported by // hardware. However, compiler can still implement these patterns through // appropriate patterns combinations based on current implemented patterns. // The implemented patterns are: EQ/GT/GTU. // Missing patterns are: GE/GEU/LT/LTU/LE/LEU. // Following instruction is not being extended as it results into the // incorrect code for negative numbers. // Pd=cmpb.eq(Rs,#u8) // p=!cmp.eq(r1,r2) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotEQ_rr : ALU32_rr<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = !cmp.eq($src1, $src2)", [(set (i1 PredRegs:$dst), (setne (i32 IntRegs:$src1), (i32 IntRegs:$src2)))]>, Requires<[HasV4T]>; // p=!cmp.eq(r1,#s10) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotEQ_ri : ALU32_ri<(outs PredRegs:$dst), (ins IntRegs:$src1, s10Ext:$src2), "$dst = !cmp.eq($src1, #$src2)", [(set (i1 PredRegs:$dst), (setne (i32 IntRegs:$src1), s10ImmPred:$src2))]>, Requires<[HasV4T]>; // p=!cmp.gt(r1,r2) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotGT_rr : ALU32_rr<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = !cmp.gt($src1, $src2)", [(set (i1 PredRegs:$dst), (not (setgt (i32 IntRegs:$src1), (i32 IntRegs:$src2))))]>, Requires<[HasV4T]>; // p=!cmp.gt(r1,#s10) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotGT_ri : ALU32_ri<(outs PredRegs:$dst), (ins IntRegs:$src1, s10Ext:$src2), "$dst = !cmp.gt($src1, #$src2)", [(set (i1 PredRegs:$dst), (not (setgt (i32 IntRegs:$src1), s10ImmPred:$src2)))]>, Requires<[HasV4T]>; // p=!cmp.gtu(r1,r2) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotGTU_rr : ALU32_rr<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = !cmp.gtu($src1, $src2)", [(set (i1 PredRegs:$dst), (not (setugt (i32 IntRegs:$src1), (i32 IntRegs:$src2))))]>, Requires<[HasV4T]>; // p=!cmp.gtu(r1,#u9) let isCompare = 1, validSubTargets = HasV4SubT in def CMPnotGTU_ri : ALU32_ri<(outs PredRegs:$dst), (ins IntRegs:$src1, u9Ext:$src2), "$dst = !cmp.gtu($src1, #$src2)", [(set (i1 PredRegs:$dst), (not (setugt (i32 IntRegs:$src1), u9ImmPred:$src2)))]>, Requires<[HasV4T]>; let isCompare = 1, validSubTargets = HasV4SubT in def CMPbEQri_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, u8Imm:$src2), "$dst = cmpb.eq($src1, #$src2)", [(set (i1 PredRegs:$dst), (seteq (and (i32 IntRegs:$src1), 255), u8ImmPred:$src2))]>, Requires<[HasV4T]>; def : Pat <(brcond (i1 (setne (and (i32 IntRegs:$src1), 255), u8ImmPred:$src2)), bb:$offset), (JMP_cNot (CMPbEQri_V4 (i32 IntRegs:$src1), u8ImmPred:$src2), bb:$offset)>, Requires<[HasV4T]>; // Pd=cmpb.eq(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPbEQrr_ubub_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmpb.eq($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (and (xor (i32 IntRegs:$src1), (i32 IntRegs:$src2)), 255), 0))]>, Requires<[HasV4T]>; // Pd=cmpb.eq(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPbEQrr_sbsb_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmpb.eq($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (shl (i32 IntRegs:$src1), (i32 24)), (shl (i32 IntRegs:$src2), (i32 24))))]>, Requires<[HasV4T]>; // Pd=cmpb.gt(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPbGTrr_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmpb.gt($src1, $src2)", [(set (i1 PredRegs:$dst), (setgt (shl (i32 IntRegs:$src1), (i32 24)), (shl (i32 IntRegs:$src2), (i32 24))))]>, Requires<[HasV4T]>; // Pd=cmpb.gtu(Rs,#u7) let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 7, isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPbGTU", InputType = "imm" in def CMPbGTUri_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, u7Ext:$src2), "$dst = cmpb.gtu($src1, #$src2)", [(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 255), u7ExtPred:$src2))]>, Requires<[HasV4T]>, ImmRegRel; // SDNode for converting immediate C to C-1. def DEC_CONST_BYTE : SDNodeXFormgetSExtValue(); return XformU7ToU7M1Imm(imm); }]>; // For the sequence // zext( seteq ( and(Rs, 255), u8)) // Generate // Pd=cmpb.eq(Rs, #u8) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 def : Pat <(i32 (zext (i1 (seteq (i32 (and (i32 IntRegs:$Rs), 255)), u8ExtPred:$u8)))), (i32 (TFR_condset_ii (i1 (CMPbEQri_V4 (i32 IntRegs:$Rs), (u8ExtPred:$u8))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setne ( and(Rs, 255), u8)) // Generate // Pd=cmpb.eq(Rs, #u8) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 def : Pat <(i32 (zext (i1 (setne (i32 (and (i32 IntRegs:$Rs), 255)), u8ExtPred:$u8)))), (i32 (TFR_condset_ii (i1 (CMPbEQri_V4 (i32 IntRegs:$Rs), (u8ExtPred:$u8))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( seteq (Rs, and(Rt, 255))) // Generate // Pd=cmpb.eq(Rs, Rt) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 def : Pat <(i32 (zext (i1 (seteq (i32 IntRegs:$Rt), (i32 (and (i32 IntRegs:$Rs), 255)))))), (i32 (TFR_condset_ii (i1 (CMPbEQrr_ubub_V4 (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setne (Rs, and(Rt, 255))) // Generate // Pd=cmpb.eq(Rs, Rt) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 def : Pat <(i32 (zext (i1 (setne (i32 IntRegs:$Rt), (i32 (and (i32 IntRegs:$Rs), 255)))))), (i32 (TFR_condset_ii (i1 (CMPbEQrr_ubub_V4 (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( setugt ( and(Rs, 255), u8)) // Generate // Pd=cmpb.gtu(Rs, #u8) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 def : Pat <(i32 (zext (i1 (setugt (i32 (and (i32 IntRegs:$Rs), 255)), u8ExtPred:$u8)))), (i32 (TFR_condset_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$Rs), (u8ExtPred:$u8))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setugt ( and(Rs, 254), u8)) // Generate // Pd=cmpb.gtu(Rs, #u8) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 def : Pat <(i32 (zext (i1 (setugt (i32 (and (i32 IntRegs:$Rs), 254)), u8ExtPred:$u8)))), (i32 (TFR_condset_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$Rs), (u8ExtPred:$u8))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setult ( Rs, Rt)) // Generate // Pd=cmp.ltu(Rs, Rt) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 // cmp.ltu(Rs, Rt) -> cmp.gtu(Rt, Rs) def : Pat <(i32 (zext (i1 (setult (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rt), (i32 IntRegs:$Rs))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setlt ( Rs, Rt)) // Generate // Pd=cmp.lt(Rs, Rt) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 // cmp.lt(Rs, Rt) -> cmp.gt(Rt, Rs) def : Pat <(i32 (zext (i1 (setlt (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rt), (i32 IntRegs:$Rs))), 1, 0))>, Requires<[HasV4T]>; // For the sequence // zext( setugt ( Rs, Rt)) // Generate // Pd=cmp.gtu(Rs, Rt) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 def : Pat <(i32 (zext (i1 (setugt (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))), 1, 0))>, Requires<[HasV4T]>; // This pattern interefers with coremark performance, not implementing at this // time. // For the sequence // zext( setgt ( Rs, Rt)) // Generate // Pd=cmp.gt(Rs, Rt) // if (Pd.new) Rd=#1 // if (!Pd.new) Rd=#0 // For the sequence // zext( setuge ( Rs, Rt)) // Generate // Pd=cmp.ltu(Rs, Rt) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 // cmp.ltu(Rs, Rt) -> cmp.gtu(Rt, Rs) def : Pat <(i32 (zext (i1 (setuge (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rt), (i32 IntRegs:$Rs))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( setge ( Rs, Rt)) // Generate // Pd=cmp.lt(Rs, Rt) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 // cmp.lt(Rs, Rt) -> cmp.gt(Rt, Rs) def : Pat <(i32 (zext (i1 (setge (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rt), (i32 IntRegs:$Rs))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( setule ( Rs, Rt)) // Generate // Pd=cmp.gtu(Rs, Rt) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 def : Pat <(i32 (zext (i1 (setule (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTUrr (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( setle ( Rs, Rt)) // Generate // Pd=cmp.gt(Rs, Rt) // if (Pd.new) Rd=#0 // if (!Pd.new) Rd=#1 def : Pat <(i32 (zext (i1 (setle (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))))), (i32 (TFR_condset_ii (i1 (CMPGTrr (i32 IntRegs:$Rs), (i32 IntRegs:$Rt))), 0, 1))>, Requires<[HasV4T]>; // For the sequence // zext( setult ( and(Rs, 255), u8)) // Use the isdigit transformation below // Generate code of the form 'mux_ii(cmpbgtu(Rdd, C-1),0,1)' // for C code of the form r = ((c>='0') & (c<='9')) ? 1 : 0;. // The isdigit transformation relies on two 'clever' aspects: // 1) The data type is unsigned which allows us to eliminate a zero test after // biasing the expression by 48. We are depending on the representation of // the unsigned types, and semantics. // 2) The front end has converted <= 9 into < 10 on entry to LLVM // // For the C code: // retval = ((c>='0') & (c<='9')) ? 1 : 0; // The code is transformed upstream of llvm into // retval = (c-48) < 10 ? 1 : 0; let AddedComplexity = 139 in def : Pat <(i32 (zext (i1 (setult (i32 (and (i32 IntRegs:$src1), 255)), u7StrictPosImmPred:$src2)))), (i32 (MUX_ii (i1 (CMPbGTUri_V4 (i32 IntRegs:$src1), (DEC_CONST_BYTE u7StrictPosImmPred:$src2))), 0, 1))>, Requires<[HasV4T]>; // Pd=cmpb.gtu(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPbGTU", InputType = "reg" in def CMPbGTUrr_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmpb.gtu($src1, $src2)", [(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 255), (and (i32 IntRegs:$src2), 255)))]>, Requires<[HasV4T]>, ImmRegRel; // Following instruction is not being extended as it results into the incorrect // code for negative numbers. // Signed half compare(.eq) ri. // Pd=cmph.eq(Rs,#s8) let isCompare = 1, validSubTargets = HasV4SubT in def CMPhEQri_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, s8Imm:$src2), "$dst = cmph.eq($src1, #$src2)", [(set (i1 PredRegs:$dst), (seteq (and (i32 IntRegs:$src1), 65535), s8ImmPred:$src2))]>, Requires<[HasV4T]>; // Signed half compare(.eq) rr. // Case 1: xor + and, then compare: // r0=xor(r0,r1) // r0=and(r0,#0xffff) // p0=cmp.eq(r0,#0) // Pd=cmph.eq(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPhEQrr_xor_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmph.eq($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (and (xor (i32 IntRegs:$src1), (i32 IntRegs:$src2)), 65535), 0))]>, Requires<[HasV4T]>; // Signed half compare(.eq) rr. // Case 2: shift left 16 bits then compare: // r0=asl(r0,16) // r1=asl(r1,16) // p0=cmp.eq(r0,r1) // Pd=cmph.eq(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPhEQrr_shl_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmph.eq($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (shl (i32 IntRegs:$src1), (i32 16)), (shl (i32 IntRegs:$src2), (i32 16))))]>, Requires<[HasV4T]>; /* Incorrect Pattern -- immediate should be right shifted before being used in the cmph.gt instruction. // Signed half compare(.gt) ri. // Pd=cmph.gt(Rs,#s8) let isExtendable = 1, opExtendable = 2, isExtentSigned = 1, opExtentBits = 8, isCompare = 1, validSubTargets = HasV4SubT in def CMPhGTri_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, s8Ext:$src2), "$dst = cmph.gt($src1, #$src2)", [(set (i1 PredRegs:$dst), (setgt (shl (i32 IntRegs:$src1), (i32 16)), s8ExtPred:$src2))]>, Requires<[HasV4T]>; */ // Signed half compare(.gt) rr. // Pd=cmph.gt(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT in def CMPhGTrr_shl_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmph.gt($src1, $src2)", [(set (i1 PredRegs:$dst), (setgt (shl (i32 IntRegs:$src1), (i32 16)), (shl (i32 IntRegs:$src2), (i32 16))))]>, Requires<[HasV4T]>; // Unsigned half compare rr (.gtu). // Pd=cmph.gtu(Rs,Rt) let isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPhGTU", InputType = "reg" in def CMPhGTUrr_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = cmph.gtu($src1, $src2)", [(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 65535), (and (i32 IntRegs:$src2), 65535)))]>, Requires<[HasV4T]>, ImmRegRel; // Unsigned half compare ri (.gtu). // Pd=cmph.gtu(Rs,#u7) let isExtendable = 1, opExtendable = 2, isExtentSigned = 0, opExtentBits = 7, isCompare = 1, validSubTargets = HasV4SubT, CextOpcode = "CMPhGTU", InputType = "imm" in def CMPhGTUri_V4 : MInst<(outs PredRegs:$dst), (ins IntRegs:$src1, u7Ext:$src2), "$dst = cmph.gtu($src1, #$src2)", [(set (i1 PredRegs:$dst), (setugt (and (i32 IntRegs:$src1), 65535), u7ExtPred:$src2))]>, Requires<[HasV4T]>, ImmRegRel; let validSubTargets = HasV4SubT in def NTSTBIT_rr : SInst<(outs PredRegs:$dst), (ins IntRegs:$src1, IntRegs:$src2), "$dst = !tstbit($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (and (shl 1, (i32 IntRegs:$src2)), (i32 IntRegs:$src1)), 0))]>, Requires<[HasV4T]>; let validSubTargets = HasV4SubT in def NTSTBIT_ri : SInst<(outs PredRegs:$dst), (ins IntRegs:$src1, u5Imm:$src2), "$dst = !tstbit($src1, $src2)", [(set (i1 PredRegs:$dst), (seteq (and (shl 1, u5ImmPred:$src2), (i32 IntRegs:$src1)), 0))]>, Requires<[HasV4T]>; //===----------------------------------------------------------------------===// // XTYPE/PRED - //===----------------------------------------------------------------------===// //Deallocate frame and return. // dealloc_return let isReturn = 1, isTerminator = 1, isBarrier = 1, isPredicable = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1 in { def DEALLOC_RET_V4 : NVInst_V4<(outs), (ins i32imm:$amt1), "dealloc_return", []>, Requires<[HasV4T]>; } // Restore registers and dealloc return function call. let isCall = 1, isBarrier = 1, isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC] in { def RESTORE_DEALLOC_RET_JMP_V4 : JInst<(outs), (ins calltarget:$dst), "jump $dst // Restore_and_dealloc_return", []>, Requires<[HasV4T]>; } // Restore registers and dealloc frame before a tail call. let isCall = 1, isBarrier = 1, Defs = [R29, R30, R31, PC] in { def RESTORE_DEALLOC_BEFORE_TAILCALL_V4 : JInst<(outs), (ins calltarget:$dst), "call $dst // Restore_and_dealloc_before_tailcall", []>, Requires<[HasV4T]>; } // Save registers function call. let isCall = 1, isBarrier = 1, Uses = [R29, R31] in { def SAVE_REGISTERS_CALL_V4 : JInst<(outs), (ins calltarget:$dst), "call $dst // Save_calle_saved_registers", []>, Requires<[HasV4T]>; } // if (Ps) dealloc_return let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if ($src1) dealloc_return", []>, Requires<[HasV4T]>; } // if (!Ps) dealloc_return let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cNotPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if (!$src1) dealloc_return", []>, Requires<[HasV4T]>; } // if (Ps.new) dealloc_return:nt let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cdnPnt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if ($src1.new) dealloc_return:nt", []>, Requires<[HasV4T]>; } // if (!Ps.new) dealloc_return:nt let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cNotdnPnt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if (!$src1.new) dealloc_return:nt", []>, Requires<[HasV4T]>; } // if (Ps.new) dealloc_return:t let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cdnPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if ($src1.new) dealloc_return:t", []>, Requires<[HasV4T]>; } // if (!Ps.new) dealloc_return:nt let isReturn = 1, isTerminator = 1, Defs = [R29, R30, R31, PC], Uses = [R29, R31], neverHasSideEffects = 1, isPredicated = 1 in { def DEALLOC_RET_cNotdnPt_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, i32imm:$amt1), "if (!$src1.new) dealloc_return:t", []>, Requires<[HasV4T]>; } // Load/Store with absolute addressing mode // memw(#u6)=Rt multiclass ST_Abs_Predbase { let PNewValue = !if(isPredNew, "new", "") in def NAME#_V4 : STInst2<(outs), (ins PredRegs:$src1, globaladdressExt:$absaddr, RC: $src2), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"(##$absaddr) = $src2", []>, Requires<[HasV4T]>; } multiclass ST_Abs_Pred { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Abs_Predbase; // Predicate new defm _cdn#NAME : ST_Abs_Predbase; } } let isNVStorable = 1, isExtended = 1, neverHasSideEffects = 1 in multiclass ST_Abs { let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in { let opExtendable = 0, isPredicable = 1 in def NAME#_V4 : STInst2<(outs), (ins globaladdressExt:$absaddr, RC:$src), mnemonic#"(##$absaddr) = $src", []>, Requires<[HasV4T]>; let opExtendable = 1, isPredicated = 1 in { defm Pt : ST_Abs_Pred; defm NotPt : ST_Abs_Pred; } } } multiclass ST_Abs_Predbase_nv { let PNewValue = !if(isPredNew, "new", "") in def NAME#_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, globaladdressExt:$absaddr, RC: $src2), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#mnemonic#"(##$absaddr) = $src2.new", []>, Requires<[HasV4T]>; } multiclass ST_Abs_Pred_nv { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : ST_Abs_Predbase_nv; // Predicate new defm _cdn#NAME : ST_Abs_Predbase_nv; } } let mayStore = 1, isNVStore = 1, isExtended = 1, neverHasSideEffects = 1 in multiclass ST_Abs_nv { let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in { let opExtendable = 0, isPredicable = 1 in def NAME#_nv_V4 : NVInst_V4<(outs), (ins globaladdressExt:$absaddr, RC:$src), mnemonic#"(##$absaddr) = $src.new", []>, Requires<[HasV4T]>; let opExtendable = 1, isPredicated = 1 in { defm Pt : ST_Abs_Pred_nv; defm NotPt : ST_Abs_Pred_nv; } } } let addrMode = Absolute in { defm STrib_abs : ST_Abs<"memb", "STrib", IntRegs>, ST_Abs_nv<"memb", "STrib", IntRegs>, AddrModeRel; defm STrih_abs : ST_Abs<"memh", "STrih", IntRegs>, ST_Abs_nv<"memh", "STrih", IntRegs>, AddrModeRel; defm STriw_abs : ST_Abs<"memw", "STriw", IntRegs>, ST_Abs_nv<"memw", "STriw", IntRegs>, AddrModeRel; let isNVStorable = 0 in defm STrid_abs : ST_Abs<"memd", "STrid", DoubleRegs>, AddrModeRel; } let Predicates = [HasV4T], AddedComplexity = 30 in { def : Pat<(truncstorei8 (i32 IntRegs:$src1), (HexagonCONST32 tglobaladdr:$absaddr)), (STrib_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>; def : Pat<(truncstorei16 (i32 IntRegs:$src1), (HexagonCONST32 tglobaladdr:$absaddr)), (STrih_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>; def : Pat<(store (i32 IntRegs:$src1), (HexagonCONST32 tglobaladdr:$absaddr)), (STriw_abs_V4 tglobaladdr: $absaddr, IntRegs: $src1)>; def : Pat<(store (i64 DoubleRegs:$src1), (HexagonCONST32 tglobaladdr:$absaddr)), (STrid_abs_V4 tglobaladdr: $absaddr, DoubleRegs: $src1)>; } multiclass LD_Abs_Predbase { let PNewValue = !if(isPredNew, "new", "") in def NAME : LDInst2<(outs RC:$dst), (ins PredRegs:$src1, globaladdressExt:$absaddr), !if(isNot, "if (!$src1", "if ($src1")#!if(isPredNew, ".new) ", ") ")#"$dst = "#mnemonic#"(##$absaddr)", []>, Requires<[HasV4T]>; } multiclass LD_Abs_Pred { let PredSense = !if(PredNot, "false", "true") in { defm _c#NAME : LD_Abs_Predbase; // Predicate new defm _cdn#NAME : LD_Abs_Predbase; } } let isExtended = 1, neverHasSideEffects = 1 in multiclass LD_Abs { let CextOpcode = CextOp, BaseOpcode = CextOp#_abs in { let opExtendable = 1, isPredicable = 1 in def NAME#_V4 : LDInst2<(outs RC:$dst), (ins globaladdressExt:$absaddr), "$dst = "#mnemonic#"(##$absaddr)", []>, Requires<[HasV4T]>; let opExtendable = 2, isPredicated = 1 in { defm Pt_V4 : LD_Abs_Pred; defm NotPt_V4 : LD_Abs_Pred; } } } let addrMode = Absolute in { defm LDrib_abs : LD_Abs<"memb", "LDrib", IntRegs>, AddrModeRel; defm LDriub_abs : LD_Abs<"memub", "LDriub", IntRegs>, AddrModeRel; defm LDrih_abs : LD_Abs<"memh", "LDrih", IntRegs>, AddrModeRel; defm LDriuh_abs : LD_Abs<"memuh", "LDriuh", IntRegs>, AddrModeRel; defm LDriw_abs : LD_Abs<"memw", "LDriw", IntRegs>, AddrModeRel; defm LDrid_abs : LD_Abs<"memd", "LDrid", DoubleRegs>, AddrModeRel; } let Predicates = [HasV4T], AddedComplexity = 30 in def : Pat<(i32 (load (HexagonCONST32 tglobaladdr:$absaddr))), (LDriw_abs_V4 tglobaladdr: $absaddr)>; let Predicates = [HasV4T], AddedComplexity=30 in def : Pat<(i32 (sextloadi8 (HexagonCONST32 tglobaladdr:$absaddr))), (LDrib_abs_V4 tglobaladdr:$absaddr)>; let Predicates = [HasV4T], AddedComplexity=30 in def : Pat<(i32 (zextloadi8 (HexagonCONST32 tglobaladdr:$absaddr))), (LDriub_abs_V4 tglobaladdr:$absaddr)>; let Predicates = [HasV4T], AddedComplexity=30 in def : Pat<(i32 (sextloadi16 (HexagonCONST32 tglobaladdr:$absaddr))), (LDrih_abs_V4 tglobaladdr:$absaddr)>; let Predicates = [HasV4T], AddedComplexity=30 in def : Pat<(i32 (zextloadi16 (HexagonCONST32 tglobaladdr:$absaddr))), (LDriuh_abs_V4 tglobaladdr:$absaddr)>; // Transfer global address into a register let AddedComplexity=50, isMoveImm = 1, isReMaterializable = 1 in def TFRI_V4 : ALU32_ri<(outs IntRegs:$dst), (ins globaladdress:$src1), "$dst = ##$src1", [(set IntRegs:$dst, (HexagonCONST32 tglobaladdr:$src1))]>, Requires<[HasV4T]>; let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in def TFRI_cPt_V4 : ALU32_ri<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$src2), "if($src1) $dst = ##$src2", []>, Requires<[HasV4T]>; let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in def TFRI_cNotPt_V4 : ALU32_ri<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$src2), "if(!$src1) $dst = ##$src2", []>, Requires<[HasV4T]>; let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in def TFRI_cdnPt_V4 : ALU32_ri<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$src2), "if($src1.new) $dst = ##$src2", []>, Requires<[HasV4T]>; let AddedComplexity=50, neverHasSideEffects = 1, isPredicated = 1 in def TFRI_cdnNotPt_V4 : ALU32_ri<(outs IntRegs:$dst), (ins PredRegs:$src1, globaladdress:$src2), "if(!$src1.new) $dst = ##$src2", []>, Requires<[HasV4T]>; let AddedComplexity = 50, Predicates = [HasV4T] in def : Pat<(HexagonCONST32_GP tglobaladdr:$src1), (TFRI_V4 tglobaladdr:$src1)>; // Load - Indirect with long offset: These instructions take global address // as an operand let AddedComplexity = 10 in def LDrid_ind_lo_V4 : LDInst<(outs DoubleRegs:$dst), (ins IntRegs:$src1, u2Imm:$src2, globaladdress:$offset), "$dst=memd($src1<<#$src2+##$offset)", [(set (i64 DoubleRegs:$dst), (load (add (shl IntRegs:$src1, u2ImmPred:$src2), (HexagonCONST32 tglobaladdr:$offset))))]>, Requires<[HasV4T]>; let AddedComplexity = 10 in multiclass LD_indirect_lo { def _lo_V4 : LDInst<(outs IntRegs:$dst), (ins IntRegs:$src1, u2Imm:$src2, globaladdress:$offset), !strconcat("$dst = ", !strconcat(OpcStr, "($src1<<#$src2+##$offset)")), [(set IntRegs:$dst, (i32 (OpNode (add (shl IntRegs:$src1, u2ImmPred:$src2), (HexagonCONST32 tglobaladdr:$offset)))))]>, Requires<[HasV4T]>; } defm LDrib_ind : LD_indirect_lo<"memb", sextloadi8>; defm LDriub_ind : LD_indirect_lo<"memub", zextloadi8>; defm LDrih_ind : LD_indirect_lo<"memh", sextloadi16>; defm LDriuh_ind : LD_indirect_lo<"memuh", zextloadi16>; defm LDriw_ind : LD_indirect_lo<"memw", load>; // Store - Indirect with long offset: These instructions take global address // as an operand let AddedComplexity = 10 in def STrid_ind_lo_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, globaladdress:$src3, DoubleRegs:$src4), "memd($src1<<#$src2+#$src3) = $src4", [(store (i64 DoubleRegs:$src4), (add (shl IntRegs:$src1, u2ImmPred:$src2), (HexagonCONST32 tglobaladdr:$src3)))]>, Requires<[HasV4T]>; let AddedComplexity = 10 in multiclass ST_indirect_lo { def _lo_V4 : STInst<(outs), (ins IntRegs:$src1, u2Imm:$src2, globaladdress:$src3, IntRegs:$src4), !strconcat(OpcStr, "($src1<<#$src2+##$src3) = $src4"), [(OpNode (i32 IntRegs:$src4), (add (shl IntRegs:$src1, u2ImmPred:$src2), (HexagonCONST32 tglobaladdr:$src3)))]>, Requires<[HasV4T]>; } defm STrib_ind : ST_indirect_lo<"memb", truncstorei8>; defm STrih_ind : ST_indirect_lo<"memh", truncstorei16>; defm STriw_ind : ST_indirect_lo<"memw", store>; // Store - absolute addressing mode: These instruction take constant // value as the extended operand. multiclass ST_absimm { let isExtended = 1, opExtendable = 0, isPredicable = 1, validSubTargets = HasV4SubT in def _abs_V4 : STInst2<(outs), (ins u0AlwaysExt:$src1, IntRegs:$src2), !strconcat(OpcStr, "(##$src1) = $src2"), []>, Requires<[HasV4T]>; let isExtended = 1, opExtendable = 1, isPredicated = 1, validSubTargets = HasV4SubT in { def _abs_cPt_V4 : STInst2<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if ($src1)", !strconcat(OpcStr, "(##$src2) = $src3")), []>, Requires<[HasV4T]>; def _abs_cNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if (!$src1)", !strconcat(OpcStr, "(##$src2) = $src3")), []>, Requires<[HasV4T]>; def _abs_cdnPt_V4 : STInst2<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if ($src1.new)", !strconcat(OpcStr, "(##$src2) = $src3")), []>, Requires<[HasV4T]>; def _abs_cdnNotPt_V4 : STInst2<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if (!$src1.new)", !strconcat(OpcStr, "(##$src2) = $src3")), []>, Requires<[HasV4T]>; } let isExtended = 1, opExtendable = 0, mayStore = 1, isNVStore = 1, validSubTargets = HasV4SubT in def _abs_nv_V4 : NVInst_V4<(outs), (ins u0AlwaysExt:$src1, IntRegs:$src2), !strconcat(OpcStr, "(##$src1) = $src2.new"), []>, Requires<[HasV4T]>; let isExtended = 1, opExtendable = 1, mayStore = 1, isPredicated = 1, isNVStore = 1, validSubTargets = HasV4SubT in { def _abs_cPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if ($src1)", !strconcat(OpcStr, "(##$src2) = $src3.new")), []>, Requires<[HasV4T]>; def _abs_cNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if (!$src1)", !strconcat(OpcStr, "(##$src2) = $src3.new")), []>, Requires<[HasV4T]>; def _abs_cdnPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if ($src1.new)", !strconcat(OpcStr, "(##$src2) = $src3.new")), []>, Requires<[HasV4T]>; def _abs_cdnNotPt_nv_V4 : NVInst_V4<(outs), (ins PredRegs:$src1, u0AlwaysExt:$src2, IntRegs:$src3), !strconcat("if (!$src1.new)", !strconcat(OpcStr, "(##$src2) = $src3.new")), []>, Requires<[HasV4T]>; } } defm STrib_imm : ST_absimm<"memb">; defm STrih_imm : ST_absimm<"memh">; defm STriw_imm : ST_absimm<"memw">; let Predicates = [HasV4T], AddedComplexity = 30 in { def : Pat<(truncstorei8 (i32 IntRegs:$src1), u0AlwaysExtPred:$src2), (STrib_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>; def : Pat<(truncstorei16 (i32 IntRegs:$src1), u0AlwaysExtPred:$src2), (STrih_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>; def : Pat<(store (i32 IntRegs:$src1), u0AlwaysExtPred:$src2), (STriw_imm_abs_V4 u0AlwaysExtPred:$src2, IntRegs: $src1)>; } // Load - absolute addressing mode: These instruction take constant // value as the extended operand multiclass LD_absimm { let isExtended = 1, opExtendable = 1, isPredicable = 1, validSubTargets = HasV4SubT in def _abs_V4 : LDInst2<(outs IntRegs:$dst), (ins u0AlwaysExt:$src), !strconcat("$dst = ", !strconcat(OpcStr, "(##$src)")), []>, Requires<[HasV4T]>; let isExtended = 1, opExtendable = 2, isPredicated = 1, validSubTargets = HasV4SubT in { def _abs_cPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, u0AlwaysExt:$src2), !strconcat("if ($src1) $dst = ", !strconcat(OpcStr, "(##$src2)")), []>, Requires<[HasV4T]>; def _abs_cNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, u0AlwaysExt:$src2), !strconcat("if (!$src1) $dst = ", !strconcat(OpcStr, "(##$src2)")), []>, Requires<[HasV4T]>; def _abs_cdnPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, u0AlwaysExt:$src2), !strconcat("if ($src1.new) $dst = ", !strconcat(OpcStr, "(##$src2)")), []>, Requires<[HasV4T]>; def _abs_cdnNotPt_V4 : LDInst2<(outs IntRegs:$dst), (ins PredRegs:$src1, u0AlwaysExt:$src2), !strconcat("if (!$src1.new) $dst = ", !strconcat(OpcStr, "(##$src2)")), []>, Requires<[HasV4T]>; } } defm LDrib_imm : LD_absimm<"memb">; defm LDriub_imm : LD_absimm<"memub">; defm LDrih_imm : LD_absimm<"memh">; defm LDriuh_imm : LD_absimm<"memuh">; defm LDriw_imm : LD_absimm<"memw">; let Predicates = [HasV4T], AddedComplexity = 30 in { def : Pat<(i32 (load u0AlwaysExtPred:$src)), (LDriw_imm_abs_V4 u0AlwaysExtPred:$src)>; def : Pat<(i32 (sextloadi8 u0AlwaysExtPred:$src)), (LDrib_imm_abs_V4 u0AlwaysExtPred:$src)>; def : Pat<(i32 (zextloadi8 u0AlwaysExtPred:$src)), (LDriub_imm_abs_V4 u0AlwaysExtPred:$src)>; def : Pat<(i32 (sextloadi16 u0AlwaysExtPred:$src)), (LDrih_imm_abs_V4 u0AlwaysExtPred:$src)>; def : Pat<(i32 (zextloadi16 u0AlwaysExtPred:$src)), (LDriuh_imm_abs_V4 u0AlwaysExtPred:$src)>; } // Indexed store double word - global address. // memw(Rs+#u6:2)=#S8 let AddedComplexity = 10 in def STriw_offset_ext_V4 : STInst<(outs), (ins IntRegs:$src1, u6_2Imm:$src2, globaladdress:$src3), "memw($src1+#$src2) = ##$src3", [(store (HexagonCONST32 tglobaladdr:$src3), (add IntRegs:$src1, u6_2ImmPred:$src2))]>, Requires<[HasV4T]>; // Indexed store double word - global address. // memw(Rs+#u6:2)=#S8 let AddedComplexity = 10 in def STrih_offset_ext_V4 : STInst<(outs), (ins IntRegs:$src1, u6_1Imm:$src2, globaladdress:$src3), "memh($src1+#$src2) = ##$src3", [(truncstorei16 (HexagonCONST32 tglobaladdr:$src3), (add IntRegs:$src1, u6_1ImmPred:$src2))]>, Requires<[HasV4T]>; // Map from store(globaladdress + x) -> memd(#foo + x) let AddedComplexity = 100 in def : Pat<(store (i64 DoubleRegs:$src1), FoldGlobalAddrGP:$addr), (STrid_abs_V4 FoldGlobalAddrGP:$addr, (i64 DoubleRegs:$src1))>, Requires<[HasV4T]>; def : Pat<(atomic_store_64 FoldGlobalAddrGP:$addr, (i64 DoubleRegs:$src1)), (STrid_abs_V4 FoldGlobalAddrGP:$addr, (i64 DoubleRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress + x) -> memb(#foo + x) let AddedComplexity = 100 in def : Pat<(truncstorei8 (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr), (STrib_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; def : Pat<(atomic_store_8 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)), (STrib_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress + x) -> memh(#foo + x) let AddedComplexity = 100 in def : Pat<(truncstorei16 (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr), (STrih_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; def : Pat<(atomic_store_16 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)), (STrih_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from store(globaladdress + x) -> memw(#foo + x) let AddedComplexity = 100 in def : Pat<(store (i32 IntRegs:$src1), FoldGlobalAddrGP:$addr), (STriw_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; def : Pat<(atomic_store_32 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1)), (STriw_abs_V4 FoldGlobalAddrGP:$addr, (i32 IntRegs:$src1))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memd(#foo + x) let AddedComplexity = 100 in def : Pat<(i64 (load FoldGlobalAddrGP:$addr)), (i64 (LDrid_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; def : Pat<(atomic_load_64 FoldGlobalAddrGP:$addr), (i64 (LDrid_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memb(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (extloadi8 FoldGlobalAddrGP:$addr)), (i32 (LDrib_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memb(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (sextloadi8 FoldGlobalAddrGP:$addr)), (i32 (LDrib_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; //let AddedComplexity = 100 in let AddedComplexity = 100 in def : Pat<(i32 (extloadi16 FoldGlobalAddrGP:$addr)), (i32 (LDrih_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memh(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (sextloadi16 FoldGlobalAddrGP:$addr)), (i32 (LDrih_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memuh(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (zextloadi16 FoldGlobalAddrGP:$addr)), (i32 (LDriuh_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; def : Pat<(atomic_load_16 FoldGlobalAddrGP:$addr), (i32 (LDriuh_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memub(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (zextloadi8 FoldGlobalAddrGP:$addr)), (i32 (LDriub_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; def : Pat<(atomic_load_8 FoldGlobalAddrGP:$addr), (i32 (LDriub_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; // Map from load(globaladdress + x) -> memw(#foo + x) let AddedComplexity = 100 in def : Pat<(i32 (load FoldGlobalAddrGP:$addr)), (i32 (LDriw_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>; def : Pat<(atomic_load_32 FoldGlobalAddrGP:$addr), (i32 (LDriw_abs_V4 FoldGlobalAddrGP:$addr))>, Requires<[HasV4T]>;