mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-01 15:11:24 +00:00
Add codegen support for NEON vld3 intrinsics with 128-bit vectors.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@83471 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
228c08b8dd
commit
ff8952e8a9
@ -444,7 +444,7 @@ bool ARMDAGToDAGISel::SelectAddrMode6(SDValue Op, SDValue N,
|
|||||||
SDValue &Addr, SDValue &Update,
|
SDValue &Addr, SDValue &Update,
|
||||||
SDValue &Opc) {
|
SDValue &Opc) {
|
||||||
Addr = N;
|
Addr = N;
|
||||||
// The optional writeback is handled in ARMLoadStoreOpt.
|
// Default to no writeback.
|
||||||
Update = CurDAG->getRegister(0, MVT::i32);
|
Update = CurDAG->getRegister(0, MVT::i32);
|
||||||
Opc = CurDAG->getTargetConstant(ARM_AM::getAM6Opc(false), MVT::i32);
|
Opc = CurDAG->getTargetConstant(ARM_AM::getAM6Opc(false), MVT::i32);
|
||||||
return true;
|
return true;
|
||||||
@ -1388,16 +1388,57 @@ SDNode *ARMDAGToDAGISel::Select(SDValue Op) {
|
|||||||
SDValue MemAddr, MemUpdate, MemOpc;
|
SDValue MemAddr, MemUpdate, MemOpc;
|
||||||
if (!SelectAddrMode6(Op, N->getOperand(2), MemAddr, MemUpdate, MemOpc))
|
if (!SelectAddrMode6(Op, N->getOperand(2), MemAddr, MemUpdate, MemOpc))
|
||||||
return NULL;
|
return NULL;
|
||||||
|
if (VT.is64BitVector()) {
|
||||||
|
switch (VT.getSimpleVT().SimpleTy) {
|
||||||
|
default: llvm_unreachable("unhandled vld3 type");
|
||||||
|
case MVT::v8i8: Opc = ARM::VLD3d8; break;
|
||||||
|
case MVT::v4i16: Opc = ARM::VLD3d16; break;
|
||||||
|
case MVT::v2f32:
|
||||||
|
case MVT::v2i32: Opc = ARM::VLD3d32; break;
|
||||||
|
}
|
||||||
|
SDValue Chain = N->getOperand(0);
|
||||||
|
const SDValue Ops[] = { MemAddr, MemUpdate, MemOpc, Chain };
|
||||||
|
return CurDAG->getMachineNode(Opc, dl, VT, VT, VT, MVT::Other, Ops, 4);
|
||||||
|
}
|
||||||
|
// Quad registers are loaded with two separate instructions, where one
|
||||||
|
// loads the even registers and the other loads the odd registers.
|
||||||
|
EVT RegVT = VT;
|
||||||
|
unsigned Opc2 = 0;
|
||||||
switch (VT.getSimpleVT().SimpleTy) {
|
switch (VT.getSimpleVT().SimpleTy) {
|
||||||
default: llvm_unreachable("unhandled vld3 type");
|
default: llvm_unreachable("unhandled vld3 type");
|
||||||
case MVT::v8i8: Opc = ARM::VLD3d8; break;
|
case MVT::v16i8:
|
||||||
case MVT::v4i16: Opc = ARM::VLD3d16; break;
|
Opc = ARM::VLD3q8a; Opc2 = ARM::VLD3q8b; RegVT = MVT::v8i8; break;
|
||||||
case MVT::v2f32:
|
case MVT::v8i16:
|
||||||
case MVT::v2i32: Opc = ARM::VLD3d32; break;
|
Opc = ARM::VLD3q16a; Opc2 = ARM::VLD3q16b; RegVT = MVT::v4i16; break;
|
||||||
|
case MVT::v4f32:
|
||||||
|
Opc = ARM::VLD3q32a; Opc2 = ARM::VLD3q32b; RegVT = MVT::v2f32; break;
|
||||||
|
case MVT::v4i32:
|
||||||
|
Opc = ARM::VLD3q32a; Opc2 = ARM::VLD3q32b; RegVT = MVT::v2i32; break;
|
||||||
}
|
}
|
||||||
SDValue Chain = N->getOperand(0);
|
SDValue Chain = N->getOperand(0);
|
||||||
const SDValue Ops[] = { MemAddr, MemUpdate, MemOpc, Chain };
|
// Enable writeback to the address register.
|
||||||
return CurDAG->getMachineNode(Opc, dl, VT, VT, VT, MVT::Other, Ops, 4);
|
MemOpc = CurDAG->getTargetConstant(ARM_AM::getAM6Opc(true), MVT::i32);
|
||||||
|
|
||||||
|
std::vector<EVT> ResTys(3, RegVT);
|
||||||
|
ResTys.push_back(MemAddr.getValueType());
|
||||||
|
ResTys.push_back(MVT::Other);
|
||||||
|
|
||||||
|
const SDValue OpsA[] = { MemAddr, MemUpdate, MemOpc, Chain };
|
||||||
|
SDNode *VLdA = CurDAG->getMachineNode(Opc, dl, ResTys, OpsA, 4);
|
||||||
|
Chain = SDValue(VLdA, 4);
|
||||||
|
|
||||||
|
const SDValue OpsB[] = { SDValue(VLdA, 3), MemUpdate, MemOpc, Chain };
|
||||||
|
SDNode *VLdB = CurDAG->getMachineNode(Opc2, dl, ResTys, OpsB, 4);
|
||||||
|
Chain = SDValue(VLdB, 4);
|
||||||
|
|
||||||
|
SDNode *Q0 = PairDRegs(VT, SDValue(VLdA, 0), SDValue(VLdB, 0));
|
||||||
|
SDNode *Q1 = PairDRegs(VT, SDValue(VLdA, 1), SDValue(VLdB, 1));
|
||||||
|
SDNode *Q2 = PairDRegs(VT, SDValue(VLdA, 2), SDValue(VLdB, 2));
|
||||||
|
ReplaceUses(SDValue(N, 0), SDValue(Q0, 0));
|
||||||
|
ReplaceUses(SDValue(N, 1), SDValue(Q1, 0));
|
||||||
|
ReplaceUses(SDValue(N, 2), SDValue(Q2, 0));
|
||||||
|
ReplaceUses(SDValue(N, 3), Chain);
|
||||||
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
case Intrinsic::arm_neon_vld4: {
|
case Intrinsic::arm_neon_vld4: {
|
||||||
|
@ -201,11 +201,26 @@ class VLD3D<string OpcodeStr>
|
|||||||
: NLdSt<(outs DPR:$dst1, DPR:$dst2, DPR:$dst3), (ins addrmode6:$addr),
|
: NLdSt<(outs DPR:$dst1, DPR:$dst2, DPR:$dst3), (ins addrmode6:$addr),
|
||||||
IIC_VLD3,
|
IIC_VLD3,
|
||||||
!strconcat(OpcodeStr, "\t\\{$dst1,$dst2,$dst3\\}, $addr"), "", []>;
|
!strconcat(OpcodeStr, "\t\\{$dst1,$dst2,$dst3\\}, $addr"), "", []>;
|
||||||
|
class VLD3WB<string OpcodeStr>
|
||||||
|
: NLdSt<(outs DPR:$dst1, DPR:$dst2, DPR:$dst3, GPR:$wb),
|
||||||
|
(ins addrmode6:$addr), IIC_VLD3,
|
||||||
|
!strconcat(OpcodeStr, "\t\\{$dst1,$dst2,$dst3\\}, $addr"),
|
||||||
|
"$addr.addr = $wb", []>;
|
||||||
|
|
||||||
def VLD3d8 : VLD3D<"vld3.8">;
|
def VLD3d8 : VLD3D<"vld3.8">;
|
||||||
def VLD3d16 : VLD3D<"vld3.16">;
|
def VLD3d16 : VLD3D<"vld3.16">;
|
||||||
def VLD3d32 : VLD3D<"vld3.32">;
|
def VLD3d32 : VLD3D<"vld3.32">;
|
||||||
|
|
||||||
|
// vld3 to double-spaced even registers.
|
||||||
|
def VLD3q8a : VLD3WB<"vld3.8">;
|
||||||
|
def VLD3q16a : VLD3WB<"vld3.16">;
|
||||||
|
def VLD3q32a : VLD3WB<"vld3.32">;
|
||||||
|
|
||||||
|
// vld3 to double-spaced odd registers.
|
||||||
|
def VLD3q8b : VLD3WB<"vld3.8">;
|
||||||
|
def VLD3q16b : VLD3WB<"vld3.16">;
|
||||||
|
def VLD3q32b : VLD3WB<"vld3.32">;
|
||||||
|
|
||||||
// VLD4 : Vector Load (multiple 4-element structures)
|
// VLD4 : Vector Load (multiple 4-element structures)
|
||||||
class VLD4D<string OpcodeStr>
|
class VLD4D<string OpcodeStr>
|
||||||
: NLdSt<(outs DPR:$dst1, DPR:$dst2, DPR:$dst3, DPR:$dst4),
|
: NLdSt<(outs DPR:$dst1, DPR:$dst2, DPR:$dst3, DPR:$dst4),
|
||||||
|
@ -36,8 +36,12 @@ namespace {
|
|||||||
char NEONPreAllocPass::ID = 0;
|
char NEONPreAllocPass::ID = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool isNEONMultiRegOp(int Opcode, unsigned &FirstOpnd,
|
static bool isNEONMultiRegOp(int Opcode, unsigned &FirstOpnd, unsigned &NumRegs,
|
||||||
unsigned &NumRegs) {
|
unsigned &Offset, unsigned &Stride) {
|
||||||
|
// Default to unit stride with no offset.
|
||||||
|
Stride = 1;
|
||||||
|
Offset = 0;
|
||||||
|
|
||||||
switch (Opcode) {
|
switch (Opcode) {
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
@ -69,6 +73,24 @@ static bool isNEONMultiRegOp(int Opcode, unsigned &FirstOpnd,
|
|||||||
NumRegs = 3;
|
NumRegs = 3;
|
||||||
return true;
|
return true;
|
||||||
|
|
||||||
|
case ARM::VLD3q8a:
|
||||||
|
case ARM::VLD3q16a:
|
||||||
|
case ARM::VLD3q32a:
|
||||||
|
FirstOpnd = 0;
|
||||||
|
NumRegs = 3;
|
||||||
|
Offset = 0;
|
||||||
|
Stride = 2;
|
||||||
|
return true;
|
||||||
|
|
||||||
|
case ARM::VLD3q8b:
|
||||||
|
case ARM::VLD3q16b:
|
||||||
|
case ARM::VLD3q32b:
|
||||||
|
FirstOpnd = 0;
|
||||||
|
NumRegs = 3;
|
||||||
|
Offset = 1;
|
||||||
|
Stride = 2;
|
||||||
|
return true;
|
||||||
|
|
||||||
case ARM::VLD4d8:
|
case ARM::VLD4d8:
|
||||||
case ARM::VLD4d16:
|
case ARM::VLD4d16:
|
||||||
case ARM::VLD4d32:
|
case ARM::VLD4d32:
|
||||||
@ -149,8 +171,8 @@ bool NEONPreAllocPass::PreAllocNEONRegisters(MachineBasicBlock &MBB) {
|
|||||||
MachineBasicBlock::iterator MBBI = MBB.begin(), E = MBB.end();
|
MachineBasicBlock::iterator MBBI = MBB.begin(), E = MBB.end();
|
||||||
for (; MBBI != E; ++MBBI) {
|
for (; MBBI != E; ++MBBI) {
|
||||||
MachineInstr *MI = &*MBBI;
|
MachineInstr *MI = &*MBBI;
|
||||||
unsigned FirstOpnd, NumRegs;
|
unsigned FirstOpnd, NumRegs, Offset, Stride;
|
||||||
if (!isNEONMultiRegOp(MI->getOpcode(), FirstOpnd, NumRegs))
|
if (!isNEONMultiRegOp(MI->getOpcode(), FirstOpnd, NumRegs, Offset, Stride))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
MachineBasicBlock::iterator NextI = next(MBBI);
|
MachineBasicBlock::iterator NextI = next(MBBI);
|
||||||
@ -164,9 +186,10 @@ bool NEONPreAllocPass::PreAllocNEONRegisters(MachineBasicBlock &MBB) {
|
|||||||
// For now, just assign a fixed set of adjacent registers.
|
// For now, just assign a fixed set of adjacent registers.
|
||||||
// This leaves plenty of room for future improvements.
|
// This leaves plenty of room for future improvements.
|
||||||
static const unsigned NEONDRegs[] = {
|
static const unsigned NEONDRegs[] = {
|
||||||
ARM::D0, ARM::D1, ARM::D2, ARM::D3
|
ARM::D0, ARM::D1, ARM::D2, ARM::D3,
|
||||||
|
ARM::D4, ARM::D5, ARM::D6, ARM::D7
|
||||||
};
|
};
|
||||||
MO.setReg(NEONDRegs[R]);
|
MO.setReg(NEONDRegs[Offset + R * Stride]);
|
||||||
|
|
||||||
if (MO.isUse()) {
|
if (MO.isUse()) {
|
||||||
// Insert a copy from VirtReg.
|
// Insert a copy from VirtReg.
|
||||||
|
@ -5,6 +5,11 @@
|
|||||||
%struct.__neon_int32x2x3_t = type { <2 x i32>, <2 x i32>, <2 x i32> }
|
%struct.__neon_int32x2x3_t = type { <2 x i32>, <2 x i32>, <2 x i32> }
|
||||||
%struct.__neon_float32x2x3_t = type { <2 x float>, <2 x float>, <2 x float> }
|
%struct.__neon_float32x2x3_t = type { <2 x float>, <2 x float>, <2 x float> }
|
||||||
|
|
||||||
|
%struct.__neon_int8x16x3_t = type { <16 x i8>, <16 x i8>, <16 x i8> }
|
||||||
|
%struct.__neon_int16x8x3_t = type { <8 x i16>, <8 x i16>, <8 x i16> }
|
||||||
|
%struct.__neon_int32x4x3_t = type { <4 x i32>, <4 x i32>, <4 x i32> }
|
||||||
|
%struct.__neon_float32x4x3_t = type { <4 x float>, <4 x float>, <4 x float> }
|
||||||
|
|
||||||
define <8 x i8> @vld3i8(i8* %A) nounwind {
|
define <8 x i8> @vld3i8(i8* %A) nounwind {
|
||||||
;CHECK: vld3i8:
|
;CHECK: vld3i8:
|
||||||
;CHECK: vld3.8
|
;CHECK: vld3.8
|
||||||
@ -45,7 +50,56 @@ define <2 x float> @vld3f(float* %A) nounwind {
|
|||||||
ret <2 x float> %tmp4
|
ret <2 x float> %tmp4
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define <16 x i8> @vld3Qi8(i8* %A) nounwind {
|
||||||
|
;CHECK: vld3Qi8:
|
||||||
|
;CHECK: vld3.8
|
||||||
|
;CHECK: vld3.8
|
||||||
|
%tmp1 = call %struct.__neon_int8x16x3_t @llvm.arm.neon.vld3.v16i8(i8* %A)
|
||||||
|
%tmp2 = extractvalue %struct.__neon_int8x16x3_t %tmp1, 0
|
||||||
|
%tmp3 = extractvalue %struct.__neon_int8x16x3_t %tmp1, 2
|
||||||
|
%tmp4 = add <16 x i8> %tmp2, %tmp3
|
||||||
|
ret <16 x i8> %tmp4
|
||||||
|
}
|
||||||
|
|
||||||
|
define <8 x i16> @vld3Qi16(i16* %A) nounwind {
|
||||||
|
;CHECK: vld3Qi16:
|
||||||
|
;CHECK: vld3.16
|
||||||
|
;CHECK: vld3.16
|
||||||
|
%tmp1 = call %struct.__neon_int16x8x3_t @llvm.arm.neon.vld3.v8i16(i16* %A)
|
||||||
|
%tmp2 = extractvalue %struct.__neon_int16x8x3_t %tmp1, 0
|
||||||
|
%tmp3 = extractvalue %struct.__neon_int16x8x3_t %tmp1, 2
|
||||||
|
%tmp4 = add <8 x i16> %tmp2, %tmp3
|
||||||
|
ret <8 x i16> %tmp4
|
||||||
|
}
|
||||||
|
|
||||||
|
define <4 x i32> @vld3Qi32(i32* %A) nounwind {
|
||||||
|
;CHECK: vld3Qi32:
|
||||||
|
;CHECK: vld3.32
|
||||||
|
;CHECK: vld3.32
|
||||||
|
%tmp1 = call %struct.__neon_int32x4x3_t @llvm.arm.neon.vld3.v4i32(i32* %A)
|
||||||
|
%tmp2 = extractvalue %struct.__neon_int32x4x3_t %tmp1, 0
|
||||||
|
%tmp3 = extractvalue %struct.__neon_int32x4x3_t %tmp1, 2
|
||||||
|
%tmp4 = add <4 x i32> %tmp2, %tmp3
|
||||||
|
ret <4 x i32> %tmp4
|
||||||
|
}
|
||||||
|
|
||||||
|
define <4 x float> @vld3Qf(float* %A) nounwind {
|
||||||
|
;CHECK: vld3Qf:
|
||||||
|
;CHECK: vld3.32
|
||||||
|
;CHECK: vld3.32
|
||||||
|
%tmp1 = call %struct.__neon_float32x4x3_t @llvm.arm.neon.vld3.v4f32(float* %A)
|
||||||
|
%tmp2 = extractvalue %struct.__neon_float32x4x3_t %tmp1, 0
|
||||||
|
%tmp3 = extractvalue %struct.__neon_float32x4x3_t %tmp1, 2
|
||||||
|
%tmp4 = add <4 x float> %tmp2, %tmp3
|
||||||
|
ret <4 x float> %tmp4
|
||||||
|
}
|
||||||
|
|
||||||
declare %struct.__neon_int8x8x3_t @llvm.arm.neon.vld3.v8i8(i8*) nounwind readonly
|
declare %struct.__neon_int8x8x3_t @llvm.arm.neon.vld3.v8i8(i8*) nounwind readonly
|
||||||
declare %struct.__neon_int16x4x3_t @llvm.arm.neon.vld3.v4i16(i8*) nounwind readonly
|
declare %struct.__neon_int16x4x3_t @llvm.arm.neon.vld3.v4i16(i8*) nounwind readonly
|
||||||
declare %struct.__neon_int32x2x3_t @llvm.arm.neon.vld3.v2i32(i8*) nounwind readonly
|
declare %struct.__neon_int32x2x3_t @llvm.arm.neon.vld3.v2i32(i8*) nounwind readonly
|
||||||
declare %struct.__neon_float32x2x3_t @llvm.arm.neon.vld3.v2f32(i8*) nounwind readonly
|
declare %struct.__neon_float32x2x3_t @llvm.arm.neon.vld3.v2f32(i8*) nounwind readonly
|
||||||
|
|
||||||
|
declare %struct.__neon_int8x16x3_t @llvm.arm.neon.vld3.v16i8(i8*) nounwind readonly
|
||||||
|
declare %struct.__neon_int16x8x3_t @llvm.arm.neon.vld3.v8i16(i8*) nounwind readonly
|
||||||
|
declare %struct.__neon_int32x4x3_t @llvm.arm.neon.vld3.v4i32(i8*) nounwind readonly
|
||||||
|
declare %struct.__neon_float32x4x3_t @llvm.arm.neon.vld3.v4f32(i8*) nounwind readonly
|
||||||
|
Loading…
Reference in New Issue
Block a user