mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-24 22:32:47 +00:00
ea28aafa83
Summary: When getConstant() is called for an expanded vector type, it is split into multiple scalar constants which are then combined using appropriate build_vector and bitcast operations. In addition to the usual big/little endian differences, the case where the element-order of the vector does not have the same endianness as the elements themselves is also accounted for. For example, for v4i32 on big-endian MIPS, the byte-order of the vector is <3210,7654,BA98,FEDC>. For little-endian, it is <0123,4567,89AB,CDEF>. Handling this case turns out to be a nop since getConstant() returns a splatted vector (so reversing the element order doesn't change the value) This fixes a number of cases in MIPS MSA where calling getConstant() during operation legalization introduces illegal types (e.g. to legalize v2i64 UNDEF into a v2i64 BUILD_VECTOR of illegal i64 zeros). It should also handle bigger differences between illegal and legal types such as legalizing v2i64 into v8i16. lowerMSASplatImm() in the MIPS backend no longer needs to avoid calling getConstant() so this function has been updated in the same patch. For the sake of transparency, the steps I've taken since the review are: * Added 'virtual' to isVectorEltOrderLittleEndian() as requested. This revealed that the MIPS tests were falsely passing because a polymorphic function was not actually polymorphic in the reviewed patch. * Fixed the tests that were now failing. This involved deleting the code to handle the MIPS MSA element-order (which was previously doing an byte-order swap instead of an element-order swap). This left isVectorEltOrderLittleEndian() unused and it was deleted. * Fixed build failures caused by rebasing beyond r194467-r194472. These build failures involved the bset, bneg, and bclr instructions added in these commits using lowerMSASplatImm() in a way that was no longer valid after this patch. Some of these were fixed by calling SelectionDAG::getConstant() instead, others were fixed by a new function getBuildVectorSplat() that provided the removed functionality of lowerMSASplatImm() in a more sensible way. Reviewers: bkramer Reviewed By: bkramer CC: llvm-commits Differential Revision: http://llvm-reviews.chandlerc.com/D1973 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@194811 91177308-0d34-0410-b5e6-96231b3b80d8
311 lines
10 KiB
LLVM
311 lines
10 KiB
LLVM
; Test the MSA intrinsics that are encoded with the I5 instruction format.
|
|
; There are lots of these so this covers those beginning with 'm'
|
|
|
|
; RUN: llc -march=mips -mattr=+msa,+fp64 < %s | FileCheck %s
|
|
; RUN: llc -march=mipsel -mattr=+msa,+fp64 < %s | FileCheck %s
|
|
|
|
@llvm_mips_maxi_s_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16
|
|
@llvm_mips_maxi_s_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_s_b_test() nounwind {
|
|
entry:
|
|
%0 = load <16 x i8>* @llvm_mips_maxi_s_b_ARG1
|
|
%1 = tail call <16 x i8> @llvm.mips.maxi.s.b(<16 x i8> %0, i32 14)
|
|
store <16 x i8> %1, <16 x i8>* @llvm_mips_maxi_s_b_RES
|
|
ret void
|
|
}
|
|
|
|
declare <16 x i8> @llvm.mips.maxi.s.b(<16 x i8>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_s_b_test:
|
|
; CHECK: ld.b
|
|
; CHECK: maxi_s.b
|
|
; CHECK: st.b
|
|
; CHECK: .size llvm_mips_maxi_s_b_test
|
|
;
|
|
@llvm_mips_maxi_s_h_ARG1 = global <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>, align 16
|
|
@llvm_mips_maxi_s_h_RES = global <8 x i16> <i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_s_h_test() nounwind {
|
|
entry:
|
|
%0 = load <8 x i16>* @llvm_mips_maxi_s_h_ARG1
|
|
%1 = tail call <8 x i16> @llvm.mips.maxi.s.h(<8 x i16> %0, i32 14)
|
|
store <8 x i16> %1, <8 x i16>* @llvm_mips_maxi_s_h_RES
|
|
ret void
|
|
}
|
|
|
|
declare <8 x i16> @llvm.mips.maxi.s.h(<8 x i16>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_s_h_test:
|
|
; CHECK: ld.h
|
|
; CHECK: maxi_s.h
|
|
; CHECK: st.h
|
|
; CHECK: .size llvm_mips_maxi_s_h_test
|
|
;
|
|
@llvm_mips_maxi_s_w_ARG1 = global <4 x i32> <i32 0, i32 1, i32 2, i32 3>, align 16
|
|
@llvm_mips_maxi_s_w_RES = global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_s_w_test() nounwind {
|
|
entry:
|
|
%0 = load <4 x i32>* @llvm_mips_maxi_s_w_ARG1
|
|
%1 = tail call <4 x i32> @llvm.mips.maxi.s.w(<4 x i32> %0, i32 14)
|
|
store <4 x i32> %1, <4 x i32>* @llvm_mips_maxi_s_w_RES
|
|
ret void
|
|
}
|
|
|
|
declare <4 x i32> @llvm.mips.maxi.s.w(<4 x i32>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_s_w_test:
|
|
; CHECK: ld.w
|
|
; CHECK: maxi_s.w
|
|
; CHECK: st.w
|
|
; CHECK: .size llvm_mips_maxi_s_w_test
|
|
;
|
|
@llvm_mips_maxi_s_d_ARG1 = global <2 x i64> <i64 0, i64 1>, align 16
|
|
@llvm_mips_maxi_s_d_RES = global <2 x i64> <i64 0, i64 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_s_d_test() nounwind {
|
|
entry:
|
|
%0 = load <2 x i64>* @llvm_mips_maxi_s_d_ARG1
|
|
%1 = tail call <2 x i64> @llvm.mips.maxi.s.d(<2 x i64> %0, i32 14)
|
|
store <2 x i64> %1, <2 x i64>* @llvm_mips_maxi_s_d_RES
|
|
ret void
|
|
}
|
|
|
|
declare <2 x i64> @llvm.mips.maxi.s.d(<2 x i64>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_s_d_test:
|
|
; CHECK: ld.d
|
|
; CHECK: maxi_s.d
|
|
; CHECK: st.d
|
|
; CHECK: .size llvm_mips_maxi_s_d_test
|
|
;
|
|
@llvm_mips_maxi_u_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16
|
|
@llvm_mips_maxi_u_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_u_b_test() nounwind {
|
|
entry:
|
|
%0 = load <16 x i8>* @llvm_mips_maxi_u_b_ARG1
|
|
%1 = tail call <16 x i8> @llvm.mips.maxi.u.b(<16 x i8> %0, i32 14)
|
|
store <16 x i8> %1, <16 x i8>* @llvm_mips_maxi_u_b_RES
|
|
ret void
|
|
}
|
|
|
|
declare <16 x i8> @llvm.mips.maxi.u.b(<16 x i8>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_u_b_test:
|
|
; CHECK: ld.b
|
|
; CHECK: maxi_u.b
|
|
; CHECK: st.b
|
|
; CHECK: .size llvm_mips_maxi_u_b_test
|
|
;
|
|
@llvm_mips_maxi_u_h_ARG1 = global <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>, align 16
|
|
@llvm_mips_maxi_u_h_RES = global <8 x i16> <i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_u_h_test() nounwind {
|
|
entry:
|
|
%0 = load <8 x i16>* @llvm_mips_maxi_u_h_ARG1
|
|
%1 = tail call <8 x i16> @llvm.mips.maxi.u.h(<8 x i16> %0, i32 14)
|
|
store <8 x i16> %1, <8 x i16>* @llvm_mips_maxi_u_h_RES
|
|
ret void
|
|
}
|
|
|
|
declare <8 x i16> @llvm.mips.maxi.u.h(<8 x i16>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_u_h_test:
|
|
; CHECK: ld.h
|
|
; CHECK: maxi_u.h
|
|
; CHECK: st.h
|
|
; CHECK: .size llvm_mips_maxi_u_h_test
|
|
;
|
|
@llvm_mips_maxi_u_w_ARG1 = global <4 x i32> <i32 0, i32 1, i32 2, i32 3>, align 16
|
|
@llvm_mips_maxi_u_w_RES = global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_u_w_test() nounwind {
|
|
entry:
|
|
%0 = load <4 x i32>* @llvm_mips_maxi_u_w_ARG1
|
|
%1 = tail call <4 x i32> @llvm.mips.maxi.u.w(<4 x i32> %0, i32 14)
|
|
store <4 x i32> %1, <4 x i32>* @llvm_mips_maxi_u_w_RES
|
|
ret void
|
|
}
|
|
|
|
declare <4 x i32> @llvm.mips.maxi.u.w(<4 x i32>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_u_w_test:
|
|
; CHECK: ld.w
|
|
; CHECK: maxi_u.w
|
|
; CHECK: st.w
|
|
; CHECK: .size llvm_mips_maxi_u_w_test
|
|
;
|
|
@llvm_mips_maxi_u_d_ARG1 = global <2 x i64> <i64 0, i64 1>, align 16
|
|
@llvm_mips_maxi_u_d_RES = global <2 x i64> <i64 0, i64 0>, align 16
|
|
|
|
define void @llvm_mips_maxi_u_d_test() nounwind {
|
|
entry:
|
|
%0 = load <2 x i64>* @llvm_mips_maxi_u_d_ARG1
|
|
%1 = tail call <2 x i64> @llvm.mips.maxi.u.d(<2 x i64> %0, i32 14)
|
|
store <2 x i64> %1, <2 x i64>* @llvm_mips_maxi_u_d_RES
|
|
ret void
|
|
}
|
|
|
|
declare <2 x i64> @llvm.mips.maxi.u.d(<2 x i64>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_maxi_u_d_test:
|
|
; CHECK: ld.d
|
|
; CHECK: maxi_u.d
|
|
; CHECK: st.d
|
|
; CHECK: .size llvm_mips_maxi_u_d_test
|
|
;
|
|
@llvm_mips_mini_s_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16
|
|
@llvm_mips_mini_s_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16
|
|
|
|
define void @llvm_mips_mini_s_b_test() nounwind {
|
|
entry:
|
|
%0 = load <16 x i8>* @llvm_mips_mini_s_b_ARG1
|
|
%1 = tail call <16 x i8> @llvm.mips.mini.s.b(<16 x i8> %0, i32 14)
|
|
store <16 x i8> %1, <16 x i8>* @llvm_mips_mini_s_b_RES
|
|
ret void
|
|
}
|
|
|
|
declare <16 x i8> @llvm.mips.mini.s.b(<16 x i8>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_s_b_test:
|
|
; CHECK: ld.b
|
|
; CHECK: mini_s.b
|
|
; CHECK: st.b
|
|
; CHECK: .size llvm_mips_mini_s_b_test
|
|
;
|
|
@llvm_mips_mini_s_h_ARG1 = global <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>, align 16
|
|
@llvm_mips_mini_s_h_RES = global <8 x i16> <i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0>, align 16
|
|
|
|
define void @llvm_mips_mini_s_h_test() nounwind {
|
|
entry:
|
|
%0 = load <8 x i16>* @llvm_mips_mini_s_h_ARG1
|
|
%1 = tail call <8 x i16> @llvm.mips.mini.s.h(<8 x i16> %0, i32 14)
|
|
store <8 x i16> %1, <8 x i16>* @llvm_mips_mini_s_h_RES
|
|
ret void
|
|
}
|
|
|
|
declare <8 x i16> @llvm.mips.mini.s.h(<8 x i16>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_s_h_test:
|
|
; CHECK: ld.h
|
|
; CHECK: mini_s.h
|
|
; CHECK: st.h
|
|
; CHECK: .size llvm_mips_mini_s_h_test
|
|
;
|
|
@llvm_mips_mini_s_w_ARG1 = global <4 x i32> <i32 0, i32 1, i32 2, i32 3>, align 16
|
|
@llvm_mips_mini_s_w_RES = global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 16
|
|
|
|
define void @llvm_mips_mini_s_w_test() nounwind {
|
|
entry:
|
|
%0 = load <4 x i32>* @llvm_mips_mini_s_w_ARG1
|
|
%1 = tail call <4 x i32> @llvm.mips.mini.s.w(<4 x i32> %0, i32 14)
|
|
store <4 x i32> %1, <4 x i32>* @llvm_mips_mini_s_w_RES
|
|
ret void
|
|
}
|
|
|
|
declare <4 x i32> @llvm.mips.mini.s.w(<4 x i32>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_s_w_test:
|
|
; CHECK: ld.w
|
|
; CHECK: mini_s.w
|
|
; CHECK: st.w
|
|
; CHECK: .size llvm_mips_mini_s_w_test
|
|
;
|
|
@llvm_mips_mini_s_d_ARG1 = global <2 x i64> <i64 0, i64 1>, align 16
|
|
@llvm_mips_mini_s_d_RES = global <2 x i64> <i64 0, i64 0>, align 16
|
|
|
|
define void @llvm_mips_mini_s_d_test() nounwind {
|
|
entry:
|
|
%0 = load <2 x i64>* @llvm_mips_mini_s_d_ARG1
|
|
%1 = tail call <2 x i64> @llvm.mips.mini.s.d(<2 x i64> %0, i32 14)
|
|
store <2 x i64> %1, <2 x i64>* @llvm_mips_mini_s_d_RES
|
|
ret void
|
|
}
|
|
|
|
declare <2 x i64> @llvm.mips.mini.s.d(<2 x i64>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_s_d_test:
|
|
; CHECK: ld.d
|
|
; CHECK: mini_s.d
|
|
; CHECK: st.d
|
|
; CHECK: .size llvm_mips_mini_s_d_test
|
|
;
|
|
@llvm_mips_mini_u_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16
|
|
@llvm_mips_mini_u_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16
|
|
|
|
define void @llvm_mips_mini_u_b_test() nounwind {
|
|
entry:
|
|
%0 = load <16 x i8>* @llvm_mips_mini_u_b_ARG1
|
|
%1 = tail call <16 x i8> @llvm.mips.mini.u.b(<16 x i8> %0, i32 14)
|
|
store <16 x i8> %1, <16 x i8>* @llvm_mips_mini_u_b_RES
|
|
ret void
|
|
}
|
|
|
|
declare <16 x i8> @llvm.mips.mini.u.b(<16 x i8>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_u_b_test:
|
|
; CHECK: ld.b
|
|
; CHECK: mini_u.b
|
|
; CHECK: st.b
|
|
; CHECK: .size llvm_mips_mini_u_b_test
|
|
;
|
|
@llvm_mips_mini_u_h_ARG1 = global <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>, align 16
|
|
@llvm_mips_mini_u_h_RES = global <8 x i16> <i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0>, align 16
|
|
|
|
define void @llvm_mips_mini_u_h_test() nounwind {
|
|
entry:
|
|
%0 = load <8 x i16>* @llvm_mips_mini_u_h_ARG1
|
|
%1 = tail call <8 x i16> @llvm.mips.mini.u.h(<8 x i16> %0, i32 14)
|
|
store <8 x i16> %1, <8 x i16>* @llvm_mips_mini_u_h_RES
|
|
ret void
|
|
}
|
|
|
|
declare <8 x i16> @llvm.mips.mini.u.h(<8 x i16>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_u_h_test:
|
|
; CHECK: ld.h
|
|
; CHECK: mini_u.h
|
|
; CHECK: st.h
|
|
; CHECK: .size llvm_mips_mini_u_h_test
|
|
;
|
|
@llvm_mips_mini_u_w_ARG1 = global <4 x i32> <i32 0, i32 1, i32 2, i32 3>, align 16
|
|
@llvm_mips_mini_u_w_RES = global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 16
|
|
|
|
define void @llvm_mips_mini_u_w_test() nounwind {
|
|
entry:
|
|
%0 = load <4 x i32>* @llvm_mips_mini_u_w_ARG1
|
|
%1 = tail call <4 x i32> @llvm.mips.mini.u.w(<4 x i32> %0, i32 14)
|
|
store <4 x i32> %1, <4 x i32>* @llvm_mips_mini_u_w_RES
|
|
ret void
|
|
}
|
|
|
|
declare <4 x i32> @llvm.mips.mini.u.w(<4 x i32>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_u_w_test:
|
|
; CHECK: ld.w
|
|
; CHECK: mini_u.w
|
|
; CHECK: st.w
|
|
; CHECK: .size llvm_mips_mini_u_w_test
|
|
;
|
|
@llvm_mips_mini_u_d_ARG1 = global <2 x i64> <i64 0, i64 1>, align 16
|
|
@llvm_mips_mini_u_d_RES = global <2 x i64> <i64 0, i64 0>, align 16
|
|
|
|
define void @llvm_mips_mini_u_d_test() nounwind {
|
|
entry:
|
|
%0 = load <2 x i64>* @llvm_mips_mini_u_d_ARG1
|
|
%1 = tail call <2 x i64> @llvm.mips.mini.u.d(<2 x i64> %0, i32 14)
|
|
store <2 x i64> %1, <2 x i64>* @llvm_mips_mini_u_d_RES
|
|
ret void
|
|
}
|
|
|
|
declare <2 x i64> @llvm.mips.mini.u.d(<2 x i64>, i32) nounwind
|
|
|
|
; CHECK: llvm_mips_mini_u_d_test:
|
|
; CHECK: ld.d
|
|
; CHECK: mini_u.d
|
|
; CHECK: st.d
|
|
; CHECK: .size llvm_mips_mini_u_d_test
|
|
;
|