mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-24 22:32:47 +00:00
Model unpack lower and interleave as vector_shuffle so we can lower the
intrinsics as such. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27200 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
ef8ef91634
commit
0038e59803
@ -303,14 +303,18 @@ X86TargetLowering::X86TargetLowering(TargetMachine &TM)
|
|||||||
setOperationAction(ISD::LOAD, MVT::v8i16, Legal);
|
setOperationAction(ISD::LOAD, MVT::v8i16, Legal);
|
||||||
setOperationAction(ISD::LOAD, MVT::v4i32, Legal);
|
setOperationAction(ISD::LOAD, MVT::v4i32, Legal);
|
||||||
setOperationAction(ISD::LOAD, MVT::v2i64, Legal);
|
setOperationAction(ISD::LOAD, MVT::v2i64, Legal);
|
||||||
|
setOperationAction(ISD::SCALAR_TO_VECTOR, MVT::v16i8, Custom);
|
||||||
|
setOperationAction(ISD::SCALAR_TO_VECTOR, MVT::v8i16, Custom);
|
||||||
setOperationAction(ISD::BUILD_VECTOR, MVT::v2f64, Custom);
|
setOperationAction(ISD::BUILD_VECTOR, MVT::v2f64, Custom);
|
||||||
setOperationAction(ISD::BUILD_VECTOR, MVT::v16i8, Custom);
|
setOperationAction(ISD::BUILD_VECTOR, MVT::v16i8, Custom);
|
||||||
setOperationAction(ISD::BUILD_VECTOR, MVT::v8i16, Custom);
|
setOperationAction(ISD::BUILD_VECTOR, MVT::v8i16, Custom);
|
||||||
setOperationAction(ISD::BUILD_VECTOR, MVT::v4i32, Custom);
|
setOperationAction(ISD::BUILD_VECTOR, MVT::v4i32, Custom);
|
||||||
setOperationAction(ISD::BUILD_VECTOR, MVT::v2i64, Custom);
|
setOperationAction(ISD::BUILD_VECTOR, MVT::v2i64, Custom);
|
||||||
setOperationAction(ISD::SCALAR_TO_VECTOR, MVT::v16i8, Custom);
|
|
||||||
setOperationAction(ISD::SCALAR_TO_VECTOR, MVT::v8i16, Custom);
|
|
||||||
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2f64, Custom);
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2f64, Custom);
|
||||||
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v16i8, Custom);
|
||||||
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v8i16, Custom);
|
||||||
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v4i32, Custom);
|
||||||
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2i64, Custom);
|
||||||
}
|
}
|
||||||
|
|
||||||
computeRegisterProperties();
|
computeRegisterProperties();
|
||||||
@ -1499,6 +1503,29 @@ bool X86::isUNPCKHPDMask(SDNode *N) {
|
|||||||
cast<ConstantSDNode>(Bit1)->getValue() == 3);
|
cast<ConstantSDNode>(Bit1)->getValue() == 3);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// isUNPCKLMask - Return true if the specified VECTOR_SHUFFLE operand
|
||||||
|
/// specifies a shuffle of elements that is suitable for input to UNPCKL.
|
||||||
|
bool X86::isUNPCKLMask(SDNode *N) {
|
||||||
|
assert(N->getOpcode() == ISD::BUILD_VECTOR);
|
||||||
|
|
||||||
|
unsigned NumElems = N->getNumOperands();
|
||||||
|
if (NumElems != 2 && NumElems != 4 && NumElems != 8 && NumElems != 16)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
for (unsigned i = 0, j = 0; i != NumElems; i += 2, ++j) {
|
||||||
|
SDOperand BitI = N->getOperand(i);
|
||||||
|
SDOperand BitI1 = N->getOperand(i+1);
|
||||||
|
assert(isa<ConstantSDNode>(BitI) && isa<ConstantSDNode>(BitI1) &&
|
||||||
|
"Invalid VECTOR_SHUFFLE mask!");
|
||||||
|
if (cast<ConstantSDNode>(BitI)->getValue() != j)
|
||||||
|
return false;
|
||||||
|
if (cast<ConstantSDNode>(BitI1)->getValue() != j + NumElems)
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
/// isSplatMask - Return true if the specified VECTOR_SHUFFLE operand specifies
|
/// isSplatMask - Return true if the specified VECTOR_SHUFFLE operand specifies
|
||||||
/// a splat of a single element.
|
/// a splat of a single element.
|
||||||
bool X86::isSplatMask(SDNode *N) {
|
bool X86::isSplatMask(SDNode *N) {
|
||||||
@ -2321,6 +2348,9 @@ SDOperand X86TargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG) {
|
|||||||
MVT::ValueType VT = Op.getValueType();
|
MVT::ValueType VT = Op.getValueType();
|
||||||
unsigned NumElems = PermMask.getNumOperands();
|
unsigned NumElems = PermMask.getNumOperands();
|
||||||
|
|
||||||
|
// All v2f64 cases are handled.
|
||||||
|
if (NumElems == 2) return SDOperand();
|
||||||
|
|
||||||
// Handle splat cases.
|
// Handle splat cases.
|
||||||
if (X86::isSplatMask(PermMask.Val)) {
|
if (X86::isSplatMask(PermMask.Val)) {
|
||||||
if (V2.getOpcode() == ISD::UNDEF)
|
if (V2.getOpcode() == ISD::UNDEF)
|
||||||
@ -2332,8 +2362,8 @@ SDOperand X86TargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG) {
|
|||||||
return DAG.getNode(ISD::VECTOR_SHUFFLE, VT, V1,
|
return DAG.getNode(ISD::VECTOR_SHUFFLE, VT, V1,
|
||||||
DAG.getNode(ISD::UNDEF, V1.getValueType()),
|
DAG.getNode(ISD::UNDEF, V1.getValueType()),
|
||||||
PermMask);
|
PermMask);
|
||||||
} else if (NumElems == 2) {
|
} else if (X86::isUNPCKLMask(PermMask.Val)) {
|
||||||
// All v2f64 cases are handled.
|
// Leave the VECTOR_SHUFFLE alone. It matches {P}UNPCKL*.
|
||||||
return SDOperand();
|
return SDOperand();
|
||||||
} else if (X86::isPSHUFDMask(PermMask.Val)) {
|
} else if (X86::isPSHUFDMask(PermMask.Val)) {
|
||||||
if (V2.getOpcode() == ISD::UNDEF)
|
if (V2.getOpcode() == ISD::UNDEF)
|
||||||
@ -2404,13 +2434,22 @@ SDOperand X86TargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG) {
|
|||||||
// : unpcklps 1, 3 ==> Y: <?, ?, 3, 1>
|
// : unpcklps 1, 3 ==> Y: <?, ?, 3, 1>
|
||||||
// Step 2: unpcklps X, Y ==> <3, 2, 1, 0>
|
// Step 2: unpcklps X, Y ==> <3, 2, 1, 0>
|
||||||
MVT::ValueType VT = Op.getValueType();
|
MVT::ValueType VT = Op.getValueType();
|
||||||
|
MVT::ValueType MaskVT = MVT::getIntVectorWithNumElements(NumElems);
|
||||||
|
MVT::ValueType BaseVT = MVT::getVectorBaseType(MaskVT);
|
||||||
|
std::vector<SDOperand> MaskVec;
|
||||||
|
for (unsigned i = 0, e = NumElems/2; i != e; ++i) {
|
||||||
|
MaskVec.push_back(DAG.getConstant(i, BaseVT));
|
||||||
|
MaskVec.push_back(DAG.getConstant(i + NumElems, BaseVT));
|
||||||
|
}
|
||||||
|
SDOperand PermMask = DAG.getNode(ISD::BUILD_VECTOR, MaskVT, MaskVec);
|
||||||
std::vector<SDOperand> V(NumElems);
|
std::vector<SDOperand> V(NumElems);
|
||||||
for (unsigned i = 0; i < NumElems; ++i)
|
for (unsigned i = 0; i < NumElems; ++i)
|
||||||
V[i] = DAG.getNode(ISD::SCALAR_TO_VECTOR, VT, Op.getOperand(i));
|
V[i] = DAG.getNode(ISD::SCALAR_TO_VECTOR, VT, Op.getOperand(i));
|
||||||
NumElems >>= 1;
|
NumElems >>= 1;
|
||||||
while (NumElems != 0) {
|
while (NumElems != 0) {
|
||||||
for (unsigned i = 0; i < NumElems; ++i)
|
for (unsigned i = 0; i < NumElems; ++i)
|
||||||
V[i] = DAG.getNode(X86ISD::UNPCKL, VT, V[i], V[i + NumElems]);
|
V[i] = DAG.getNode(ISD::VECTOR_SHUFFLE, VT, V[i], V[i + NumElems],
|
||||||
|
PermMask);
|
||||||
NumElems >>= 1;
|
NumElems >>= 1;
|
||||||
}
|
}
|
||||||
return V[0];
|
return V[0];
|
||||||
@ -2453,7 +2492,6 @@ const char *X86TargetLowering::getTargetNodeName(unsigned Opcode) const {
|
|||||||
case X86ISD::Wrapper: return "X86ISD::Wrapper";
|
case X86ISD::Wrapper: return "X86ISD::Wrapper";
|
||||||
case X86ISD::S2VEC: return "X86ISD::S2VEC";
|
case X86ISD::S2VEC: return "X86ISD::S2VEC";
|
||||||
case X86ISD::ZEXT_S2VEC: return "X86ISD::ZEXT_S2VEC";
|
case X86ISD::ZEXT_S2VEC: return "X86ISD::ZEXT_S2VEC";
|
||||||
case X86ISD::UNPCKL: return "X86ISD::UNPCKL";
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2543,5 +2581,6 @@ X86TargetLowering::isShuffleMaskLegal(SDOperand Mask, MVT::ValueType VT) const {
|
|||||||
return (Mask.Val->getNumOperands() == 2 ||
|
return (Mask.Val->getNumOperands() == 2 ||
|
||||||
X86::isSplatMask(Mask.Val) ||
|
X86::isSplatMask(Mask.Val) ||
|
||||||
X86::isPSHUFDMask(Mask.Val) ||
|
X86::isPSHUFDMask(Mask.Val) ||
|
||||||
X86::isSHUFPMask(Mask.Val));
|
X86::isSHUFPMask(Mask.Val) ||
|
||||||
|
X86::isUNPCKLMask(Mask.Val));
|
||||||
}
|
}
|
||||||
|
@ -153,10 +153,6 @@ namespace llvm {
|
|||||||
/// ZEXT_S2VEC - SCALAR_TO_VECTOR with zero extension. The destination base
|
/// ZEXT_S2VEC - SCALAR_TO_VECTOR with zero extension. The destination base
|
||||||
/// does not have to match the operand type.
|
/// does not have to match the operand type.
|
||||||
ZEXT_S2VEC,
|
ZEXT_S2VEC,
|
||||||
|
|
||||||
/// UNPCKL - Unpack and interleave low. This corresponds to X86::UNPCKLPS,
|
|
||||||
/// X86::PUNPCKL*.
|
|
||||||
UNPCKL,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// X86 specific condition code. These correspond to X86_*_COND in
|
// X86 specific condition code. These correspond to X86_*_COND in
|
||||||
@ -205,6 +201,10 @@ namespace llvm {
|
|||||||
/// specifies a shuffle of elements that is suitable for input to UNPCKHPD.
|
/// specifies a shuffle of elements that is suitable for input to UNPCKHPD.
|
||||||
bool isUNPCKHPDMask(SDNode *N);
|
bool isUNPCKHPDMask(SDNode *N);
|
||||||
|
|
||||||
|
/// isUNPCKLMask - Return true if the specified VECTOR_SHUFFLE operand
|
||||||
|
/// specifies a shuffle of elements that is suitable for input to UNPCKL.
|
||||||
|
bool isUNPCKLMask(SDNode *N);
|
||||||
|
|
||||||
/// isSplatMask - Return true if the specified VECTOR_SHUFFLE operand
|
/// isSplatMask - Return true if the specified VECTOR_SHUFFLE operand
|
||||||
/// specifies a splat of a single element.
|
/// specifies a splat of a single element.
|
||||||
bool isSplatMask(SDNode *N);
|
bool isSplatMask(SDNode *N);
|
||||||
|
@ -30,8 +30,6 @@ def X86zexts2vec : SDNode<"X86ISD::ZEXT_S2VEC",
|
|||||||
|
|
||||||
def SDTUnpckl : SDTypeProfile<1, 2,
|
def SDTUnpckl : SDTypeProfile<1, 2,
|
||||||
[SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>]>;
|
[SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>]>;
|
||||||
def X86unpckl : SDNode<"X86ISD::UNPCKL", SDTUnpckl,
|
|
||||||
[]>;
|
|
||||||
|
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
// SSE pattern fragments
|
// SSE pattern fragments
|
||||||
@ -77,6 +75,10 @@ def UNPCKHPD_shuffle_mask : PatLeaf<(build_vector), [{
|
|||||||
return X86::isUNPCKHPDMask(N);
|
return X86::isUNPCKHPDMask(N);
|
||||||
}], SHUFFLE_get_shuf_imm>;
|
}], SHUFFLE_get_shuf_imm>;
|
||||||
|
|
||||||
|
def UNPCKL_shuffle_mask : PatLeaf<(build_vector), [{
|
||||||
|
return X86::isUNPCKLMask(N);
|
||||||
|
}]>;
|
||||||
|
|
||||||
// Only use PSHUF if it is not a splat.
|
// Only use PSHUF if it is not a splat.
|
||||||
def PSHUFD_shuffle_mask : PatLeaf<(build_vector), [{
|
def PSHUFD_shuffle_mask : PatLeaf<(build_vector), [{
|
||||||
return !X86::isSplatMask(N) && X86::isPSHUFDMask(N);
|
return !X86::isSplatMask(N) && X86::isPSHUFDMask(N);
|
||||||
@ -756,14 +758,17 @@ def PSHUFDrm : PDIi8<0x70, MRMSrcMem,
|
|||||||
|
|
||||||
let isTwoAddress = 1 in {
|
let isTwoAddress = 1 in {
|
||||||
def SHUFPSrr : PSIi8<0xC6, MRMSrcReg,
|
def SHUFPSrr : PSIi8<0xC6, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2, i8imm:$src3),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2, i32i8imm:$src3),
|
||||||
"shufps {$src3, $src2, $dst|$dst, $src2, $src3}",
|
"shufps {$src3, $src2, $dst|$dst, $src2, $src3}",
|
||||||
[(set VR128:$dst, (vector_shuffle
|
[(set VR128:$dst, (vector_shuffle
|
||||||
(v4f32 VR128:$src1), (v4f32 VR128:$src2),
|
(v4f32 VR128:$src1), (v4f32 VR128:$src2),
|
||||||
SHUFP_shuffle_mask:$src3))]>;
|
SHUFP_shuffle_mask:$src3))]>;
|
||||||
def SHUFPSrm : PSIi8<0xC6, MRMSrcMem,
|
def SHUFPSrm : PSIi8<0xC6, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, f128mem:$src2, i8imm:$src3),
|
(ops VR128:$dst, VR128:$src1, f128mem:$src2, i32i8imm:$src3),
|
||||||
"shufps {$src3, $src2, $dst|$dst, $src2, $src3}", []>;
|
"shufps {$src3, $src2, $dst|$dst, $src2, $src3}",
|
||||||
|
[(set VR128:$dst, (vector_shuffle
|
||||||
|
(v4f32 VR128:$src1), (load addr:$src2),
|
||||||
|
SHUFP_shuffle_mask:$src3))]>;
|
||||||
def SHUFPDrr : PDIi8<0xC6, MRMSrcReg,
|
def SHUFPDrr : PDIi8<0xC6, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2, i8imm:$src3),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2, i8imm:$src3),
|
||||||
"shufpd {$src3, $src2, $dst|$dst, $src2, $src3}",
|
"shufpd {$src3, $src2, $dst|$dst, $src2, $src3}",
|
||||||
@ -772,7 +777,10 @@ def SHUFPDrr : PDIi8<0xC6, MRMSrcReg,
|
|||||||
SHUFP_shuffle_mask:$src3))]>;
|
SHUFP_shuffle_mask:$src3))]>;
|
||||||
def SHUFPDrm : PDIi8<0xC6, MRMSrcMem,
|
def SHUFPDrm : PDIi8<0xC6, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, f128mem:$src2, i8imm:$src3),
|
(ops VR128:$dst, VR128:$src1, f128mem:$src2, i8imm:$src3),
|
||||||
"shufpd {$src3, $src2, $dst|$dst, $src2, $src3}", []>;
|
"shufpd {$src3, $src2, $dst|$dst, $src2, $src3}",
|
||||||
|
[(set VR128:$dst, (vector_shuffle
|
||||||
|
(v2f64 VR128:$src1), (load addr:$src2),
|
||||||
|
SHUFP_shuffle_mask:$src3))]>;
|
||||||
|
|
||||||
def UNPCKHPSrr : PSI<0x15, MRMSrcReg,
|
def UNPCKHPSrr : PSI<0x15, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
@ -789,13 +797,15 @@ def UNPCKHPDrm : PDI<0x15, MRMSrcMem,
|
|||||||
def UNPCKLPSrr : PSI<0x14, MRMSrcReg,
|
def UNPCKLPSrr : PSI<0x14, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"unpcklps {$src2, $dst|$dst, $src2}",
|
"unpcklps {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v4f32 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
VR128:$src2)))]>;
|
(v4f32 (vector_shuffle VR128:$src1, VR128:$src2,
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def UNPCKLPSrm : PSI<0x14, MRMSrcMem,
|
def UNPCKLPSrm : PSI<0x14, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, f128mem:$src2),
|
(ops VR128:$dst, VR128:$src1, f128mem:$src2),
|
||||||
"unpcklps {$src2, $dst|$dst, $src2}",
|
"unpcklps {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v4f32 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
(load addr:$src2))))]>;
|
(v4f32 (vector_shuffle VR128:$src1, (load addr:$src2),
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def UNPCKLPDrr : PDI<0x14, MRMSrcReg,
|
def UNPCKLPDrr : PDI<0x14, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"unpcklpd {$src2, $dst|$dst, $src2}", []>;
|
"unpcklpd {$src2, $dst|$dst, $src2}", []>;
|
||||||
@ -895,33 +905,39 @@ def PSUBDrm : PDI<0xFA, MRMSrcMem, (ops VR128:$dst, VR128:$src1, f128mem:$src2),
|
|||||||
def PUNPCKLBWrr : PDI<0x60, MRMSrcReg,
|
def PUNPCKLBWrr : PDI<0x60, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"punpcklbw {$src2, $dst|$dst, $src2}",
|
"punpcklbw {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v16i8 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
VR128:$src2)))]>;
|
(v16i8 (vector_shuffle VR128:$src1, VR128:$src2,
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLBWrm : PDI<0x60, MRMSrcMem,
|
def PUNPCKLBWrm : PDI<0x60, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
||||||
"punpcklbw {$src2, $dst|$dst, $src2}",
|
"punpcklbw {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v16i8 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
(load addr:$src2))))]>;
|
(v16i8 (vector_shuffle VR128:$src1, (load addr:$src2),
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLWDrr : PDI<0x61, MRMSrcReg,
|
def PUNPCKLWDrr : PDI<0x61, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"punpcklwd {$src2, $dst|$dst, $src2}",
|
"punpcklwd {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v8i16 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
VR128:$src2)))]>;
|
(v8i16 (vector_shuffle VR128:$src1, VR128:$src2,
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLWDrm : PDI<0x61, MRMSrcMem,
|
def PUNPCKLWDrm : PDI<0x61, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
||||||
"punpcklwd {$src2, $dst|$dst, $src2}",
|
"punpcklwd {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v8i16 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
(load addr:$src2))))]>;
|
(v8i16 (vector_shuffle VR128:$src1, (load addr:$src2),
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLDQrr : PDI<0x62, MRMSrcReg,
|
def PUNPCKLDQrr : PDI<0x62, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"punpckldq {$src2, $dst|$dst, $src2}",
|
"punpckldq {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v4i32 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
VR128:$src2)))]>;
|
(v4i32 (vector_shuffle VR128:$src1, VR128:$src2,
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLDQrm : PDI<0x62, MRMSrcMem,
|
def PUNPCKLDQrm : PDI<0x62, MRMSrcMem,
|
||||||
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
(ops VR128:$dst, VR128:$src1, i128mem:$src2),
|
||||||
"punpckldq {$src2, $dst|$dst, $src2}",
|
"punpckldq {$src2, $dst|$dst, $src2}",
|
||||||
[(set VR128:$dst, (v4i32 (X86unpckl VR128:$src1,
|
[(set VR128:$dst,
|
||||||
(load addr:$src2))))]>;
|
(v4i32 (vector_shuffle VR128:$src1, (load addr:$src2),
|
||||||
|
UNPCKL_shuffle_mask)))]>;
|
||||||
def PUNPCKLQDQrr : PDI<0x6C, MRMSrcReg,
|
def PUNPCKLQDQrr : PDI<0x6C, MRMSrcReg,
|
||||||
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
(ops VR128:$dst, VR128:$src1, VR128:$src2),
|
||||||
"punpcklqdq {$src2, $dst|$dst, $src2}", []>;
|
"punpcklqdq {$src2, $dst|$dst, $src2}", []>;
|
||||||
|
Loading…
Reference in New Issue
Block a user