mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-14 11:32:34 +00:00
dcc4f724cc
This patch adds a new SSA MI pass that runs on little-endian PPC64 code with VSX enabled. Loads and stores of 4x32 and 2x64 vectors without alignment constraints are accomplished for little-endian using lxvd2x/xxswapd and xxswapd/stxvd2x. The existence of the additional xxswapd instructions hurts performance in comparison with big-endian code, but they are necessary in the general case to support correct semantics. However, the general case does not apply to most vector code. Many vector instructions are lane-insensitive; they do not "care" which lanes the parallel computations are performed within, provided that the resulting data is stored into the correct locations. Thus this pass looks for computations that perform only lane-insensitive operations, and remove the unnecessary swaps from loads and stores in such computations. Future improvements will allow computations using certain lane-sensitive operations to also be optimized in this manner, by modifying the lane-sensitive operations to account for the permuted order of the lanes. However, this patch only adds the infrastructure to permit this; no lane-sensitive operations are optimized at this time. This code is heavily exercised by the various vectorizing applications in the projects/test-suite tree. For the time being, I have only added one simple test case to demonstrate what the pass is doing. Although it is quite simple, it provides coverage for much of the code, including the special case handling of copies and subreg-to-reg operations feeding the swaps. I plan to add additional tests in the future as I fill in more of the "special handling" code. Two existing tests were affected, because they expected the swaps to be present, but they are now removed. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@235910 91177308-0d34-0410-b5e6-96231b3b80d8
46 lines
2.3 KiB
LLVM
46 lines
2.3 KiB
LLVM
; RUN: llc -mcpu=pwr8 -mattr=+vsx -O2 -mtriple=powerpc64-unknown-linux-gnu < %s > %t
|
|
; RUN: grep lxvw4x < %t | count 3
|
|
; RUN: grep lxvd2x < %t | count 3
|
|
; RUN: grep stxvw4x < %t | count 3
|
|
; RUN: grep stxvd2x < %t | count 3
|
|
; RUN: llc -mcpu=pwr8 -mattr=+vsx -O0 -fast-isel=1 -mtriple=powerpc64-unknown-linux-gnu < %s > %t
|
|
; RUN: grep lxvw4x < %t | count 3
|
|
; RUN: grep lxvd2x < %t | count 3
|
|
; RUN: grep stxvw4x < %t | count 3
|
|
; RUN: grep stxvd2x < %t | count 3
|
|
|
|
; RUN: llc -mcpu=pwr8 -mattr=+vsx -O2 -mtriple=powerpc64le-unknown-linux-gnu < %s > %t
|
|
; RUN: grep lxvd2x < %t | count 6
|
|
; RUN: grep stxvd2x < %t | count 6
|
|
|
|
@vsi = global <4 x i32> <i32 -1, i32 2, i32 -3, i32 4>, align 16
|
|
@vui = global <4 x i32> <i32 0, i32 1, i32 2, i32 3>, align 16
|
|
@vf = global <4 x float> <float -1.500000e+00, float 2.500000e+00, float -3.500000e+00, float 4.500000e+00>, align 16
|
|
@vsll = global <2 x i64> <i64 255, i64 -937>, align 16
|
|
@vull = global <2 x i64> <i64 1447, i64 2894>, align 16
|
|
@vd = global <2 x double> <double 3.500000e+00, double -7.500000e+00>, align 16
|
|
@res_vsi = common global <4 x i32> zeroinitializer, align 16
|
|
@res_vui = common global <4 x i32> zeroinitializer, align 16
|
|
@res_vf = common global <4 x float> zeroinitializer, align 16
|
|
@res_vsll = common global <2 x i64> zeroinitializer, align 16
|
|
@res_vull = common global <2 x i64> zeroinitializer, align 16
|
|
@res_vd = common global <2 x double> zeroinitializer, align 16
|
|
|
|
; Function Attrs: nounwind
|
|
define void @test1() {
|
|
entry:
|
|
%0 = load <4 x i32>, <4 x i32>* @vsi, align 16
|
|
%1 = load <4 x i32>, <4 x i32>* @vui, align 16
|
|
%2 = load <4 x i32>, <4 x i32>* bitcast (<4 x float>* @vf to <4 x i32>*), align 16
|
|
%3 = load <2 x double>, <2 x double>* bitcast (<2 x i64>* @vsll to <2 x double>*), align 16
|
|
%4 = load <2 x double>, <2 x double>* bitcast (<2 x i64>* @vull to <2 x double>*), align 16
|
|
%5 = load <2 x double>, <2 x double>* @vd, align 16
|
|
store <4 x i32> %0, <4 x i32>* @res_vsi, align 16
|
|
store <4 x i32> %1, <4 x i32>* @res_vui, align 16
|
|
store <4 x i32> %2, <4 x i32>* bitcast (<4 x float>* @res_vf to <4 x i32>*), align 16
|
|
store <2 x double> %3, <2 x double>* bitcast (<2 x i64>* @res_vsll to <2 x double>*), align 16
|
|
store <2 x double> %4, <2 x double>* bitcast (<2 x i64>* @res_vull to <2 x double>*), align 16
|
|
store <2 x double> %5, <2 x double>* @res_vd, align 16
|
|
ret void
|
|
}
|