2005-09-01 23:24:04 +00:00
|
|
|
//===-- DAGCombiner.cpp - Implement a DAG node combiner -------------------===//
|
2005-09-01 00:19:25 +00:00
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
2007-12-29 20:36:04 +00:00
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
2005-09-01 00:19:25 +00:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This pass combines dag nodes to form fewer, simpler DAG nodes. It can be run
|
|
|
|
// both before and after the DAG is legalized.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#define DEBUG_TYPE "dagcombine"
|
|
|
|
#include "llvm/CodeGen/SelectionDAG.h"
|
2008-01-25 07:20:16 +00:00
|
|
|
#include "llvm/CodeGen/MachineFunction.h"
|
|
|
|
#include "llvm/CodeGen/MachineFrameInfo.h"
|
2007-05-16 06:37:59 +00:00
|
|
|
#include "llvm/Analysis/AliasAnalysis.h"
|
2007-05-07 21:27:48 +00:00
|
|
|
#include "llvm/Target/TargetData.h"
|
Infer alignment of loads and increase their alignment when we can tell they are
from the stack. This allows us to compile stack-align.ll to:
_test:
movsd LCPI1_0, %xmm0
movapd %xmm0, %xmm1
*** andpd 4(%esp), %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
instead of:
_test:
movsd LCPI1_0, %xmm0
** movsd 4(%esp), %xmm1
** andpd %xmm0, %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46401 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-26 19:45:50 +00:00
|
|
|
#include "llvm/Target/TargetFrameInfo.h"
|
2005-09-01 00:19:25 +00:00
|
|
|
#include "llvm/Target/TargetLowering.h"
|
2007-05-07 21:27:48 +00:00
|
|
|
#include "llvm/Target/TargetMachine.h"
|
2007-01-08 23:04:05 +00:00
|
|
|
#include "llvm/Target/TargetOptions.h"
|
2007-05-16 06:37:59 +00:00
|
|
|
#include "llvm/ADT/SmallPtrSet.h"
|
|
|
|
#include "llvm/ADT/Statistic.h"
|
2006-08-27 12:54:02 +00:00
|
|
|
#include "llvm/Support/Compiler.h"
|
2006-09-21 16:28:59 +00:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
2007-05-16 06:37:59 +00:00
|
|
|
#include "llvm/Support/Debug.h"
|
|
|
|
#include "llvm/Support/MathExtras.h"
|
2005-09-09 23:53:39 +00:00
|
|
|
#include <algorithm>
|
2008-05-23 20:40:06 +00:00
|
|
|
#include <set>
|
2005-09-01 00:19:25 +00:00
|
|
|
using namespace llvm;
|
|
|
|
|
2006-12-19 22:41:21 +00:00
|
|
|
STATISTIC(NodesCombined , "Number of dag nodes combined");
|
|
|
|
STATISTIC(PreIndexedNodes , "Number of pre-indexed nodes created");
|
|
|
|
STATISTIC(PostIndexedNodes, "Number of post-indexed nodes created");
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
namespace {
|
2007-01-16 04:55:25 +00:00
|
|
|
#ifndef NDEBUG
|
|
|
|
static cl::opt<bool>
|
|
|
|
ViewDAGCombine1("view-dag-combine1-dags", cl::Hidden,
|
|
|
|
cl::desc("Pop up a window to show dags before the first "
|
|
|
|
"dag combine pass"));
|
|
|
|
static cl::opt<bool>
|
|
|
|
ViewDAGCombine2("view-dag-combine2-dags", cl::Hidden,
|
|
|
|
cl::desc("Pop up a window to show dags before the second "
|
|
|
|
"dag combine pass"));
|
|
|
|
#else
|
|
|
|
static const bool ViewDAGCombine1 = false;
|
|
|
|
static const bool ViewDAGCombine2 = false;
|
|
|
|
#endif
|
|
|
|
|
2006-10-07 23:37:56 +00:00
|
|
|
static cl::opt<bool>
|
|
|
|
CombinerAA("combiner-alias-analysis", cl::Hidden,
|
2006-10-17 19:33:52 +00:00
|
|
|
cl::desc("Turn on alias analysis during testing"));
|
2006-10-12 15:22:24 +00:00
|
|
|
|
2006-10-18 19:08:31 +00:00
|
|
|
static cl::opt<bool>
|
|
|
|
CombinerGlobalAA("combiner-global-alias-analysis", cl::Hidden,
|
|
|
|
cl::desc("Include global information in alias analysis"));
|
|
|
|
|
2006-10-05 15:07:25 +00:00
|
|
|
//------------------------------ DAGCombiner ---------------------------------//
|
|
|
|
|
2006-10-07 23:37:56 +00:00
|
|
|
class VISIBILITY_HIDDEN DAGCombiner {
|
2005-09-01 00:19:25 +00:00
|
|
|
SelectionDAG &DAG;
|
|
|
|
TargetLowering &TLI;
|
2005-09-01 23:24:04 +00:00
|
|
|
bool AfterLegalize;
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// Worklist of all of the nodes that need to be simplified.
|
|
|
|
std::vector<SDNode*> WorkList;
|
|
|
|
|
2006-10-16 20:52:31 +00:00
|
|
|
// AA - Used for DAG load/store alias analysis.
|
|
|
|
AliasAnalysis &AA;
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
/// AddUsersToWorkList - When an instruction is simplified, add all users of
|
|
|
|
/// the instruction to the work lists because they might get more simplified
|
|
|
|
/// now.
|
|
|
|
///
|
|
|
|
void AddUsersToWorkList(SDNode *N) {
|
|
|
|
for (SDNode::use_iterator UI = N->use_begin(), UE = N->use_end();
|
2005-09-01 23:24:04 +00:00
|
|
|
UI != UE; ++UI)
|
2008-04-07 10:06:32 +00:00
|
|
|
AddToWorkList(UI->getUser());
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
/// visit - call the node-specific routine that knows how to fold each
|
|
|
|
/// particular type of node.
|
|
|
|
SDOperand visit(SDNode *N);
|
|
|
|
|
2006-03-01 04:53:38 +00:00
|
|
|
public:
|
2006-10-04 16:53:27 +00:00
|
|
|
/// AddToWorkList - Add to the work list making sure it's instance is at the
|
|
|
|
/// the back (next to be processed.)
|
2006-03-01 04:03:14 +00:00
|
|
|
void AddToWorkList(SDNode *N) {
|
2006-10-04 16:53:27 +00:00
|
|
|
removeFromWorkList(N);
|
2006-03-01 04:03:14 +00:00
|
|
|
WorkList.push_back(N);
|
|
|
|
}
|
2006-10-04 16:53:27 +00:00
|
|
|
|
2008-02-03 06:49:24 +00:00
|
|
|
/// removeFromWorkList - remove all instances of N from the worklist.
|
|
|
|
///
|
|
|
|
void removeFromWorkList(SDNode *N) {
|
|
|
|
WorkList.erase(std::remove(WorkList.begin(), WorkList.end(), N),
|
|
|
|
WorkList.end());
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
}
|
2006-02-16 21:11:51 +00:00
|
|
|
|
2008-02-03 06:49:24 +00:00
|
|
|
SDOperand CombineTo(SDNode *N, const SDOperand *To, unsigned NumTo,
|
|
|
|
bool AddTo = true);
|
|
|
|
|
2006-10-13 23:32:28 +00:00
|
|
|
SDOperand CombineTo(SDNode *N, SDOperand Res, bool AddTo = true) {
|
|
|
|
return CombineTo(N, &Res, 1, AddTo);
|
2006-03-01 04:53:38 +00:00
|
|
|
}
|
|
|
|
|
2006-10-13 23:32:28 +00:00
|
|
|
SDOperand CombineTo(SDNode *N, SDOperand Res0, SDOperand Res1,
|
|
|
|
bool AddTo = true) {
|
2006-08-11 17:56:38 +00:00
|
|
|
SDOperand To[] = { Res0, Res1 };
|
2006-10-13 23:32:28 +00:00
|
|
|
return CombineTo(N, To, 2, AddTo);
|
2006-03-01 04:53:38 +00:00
|
|
|
}
|
2008-01-17 07:00:52 +00:00
|
|
|
|
2006-03-01 04:53:38 +00:00
|
|
|
private:
|
|
|
|
|
2006-02-17 21:58:01 +00:00
|
|
|
/// SimplifyDemandedBits - Check the specified integer node value to see if
|
2006-03-01 19:55:35 +00:00
|
|
|
/// it can be simplified or if things it uses can be simplified by bit
|
2006-02-17 21:58:01 +00:00
|
|
|
/// propagation. If so, return true.
|
2008-02-27 00:25:32 +00:00
|
|
|
bool SimplifyDemandedBits(SDOperand Op) {
|
|
|
|
APInt Demanded = APInt::getAllOnesValue(Op.getValueSizeInBits());
|
|
|
|
return SimplifyDemandedBits(Op, Demanded);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool SimplifyDemandedBits(SDOperand Op, const APInt &Demanded);
|
2005-10-10 22:31:19 +00:00
|
|
|
|
2006-11-11 00:39:41 +00:00
|
|
|
bool CombineToPreIndexedLoadStore(SDNode *N);
|
|
|
|
bool CombineToPostIndexedLoadStore(SDNode *N);
|
|
|
|
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
/// combine - call the node-specific routine that knows how to fold each
|
|
|
|
/// particular type of node. If that doesn't do anything, try the
|
|
|
|
/// target-specific DAG combines.
|
|
|
|
SDOperand combine(SDNode *N);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// Visitation implementation - Implement dag node combining for different
|
|
|
|
// node types. The semantics are as follows:
|
|
|
|
// Return Value:
|
2005-09-07 00:15:36 +00:00
|
|
|
// SDOperand.Val == 0 - No change was made
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
// SDOperand.Val == N - N was replaced, is dead, and is already handled.
|
2005-09-07 00:15:36 +00:00
|
|
|
// otherwise - N should be replaced by the returned Operand.
|
2005-09-01 00:19:25 +00:00
|
|
|
//
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitTokenFactor(SDNode *N);
|
2008-02-13 07:25:05 +00:00
|
|
|
SDOperand visitMERGE_VALUES(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitADD(SDNode *N);
|
|
|
|
SDOperand visitSUB(SDNode *N);
|
2007-03-04 20:03:15 +00:00
|
|
|
SDOperand visitADDC(SDNode *N);
|
|
|
|
SDOperand visitADDE(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitMUL(SDNode *N);
|
|
|
|
SDOperand visitSDIV(SDNode *N);
|
|
|
|
SDOperand visitUDIV(SDNode *N);
|
|
|
|
SDOperand visitSREM(SDNode *N);
|
|
|
|
SDOperand visitUREM(SDNode *N);
|
|
|
|
SDOperand visitMULHU(SDNode *N);
|
|
|
|
SDOperand visitMULHS(SDNode *N);
|
2007-10-08 17:57:15 +00:00
|
|
|
SDOperand visitSMUL_LOHI(SDNode *N);
|
|
|
|
SDOperand visitUMUL_LOHI(SDNode *N);
|
|
|
|
SDOperand visitSDIVREM(SDNode *N);
|
|
|
|
SDOperand visitUDIVREM(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitAND(SDNode *N);
|
|
|
|
SDOperand visitOR(SDNode *N);
|
|
|
|
SDOperand visitXOR(SDNode *N);
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand SimplifyVBinOp(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitSHL(SDNode *N);
|
|
|
|
SDOperand visitSRA(SDNode *N);
|
|
|
|
SDOperand visitSRL(SDNode *N);
|
|
|
|
SDOperand visitCTLZ(SDNode *N);
|
|
|
|
SDOperand visitCTTZ(SDNode *N);
|
|
|
|
SDOperand visitCTPOP(SDNode *N);
|
2005-09-16 00:54:12 +00:00
|
|
|
SDOperand visitSELECT(SDNode *N);
|
|
|
|
SDOperand visitSELECT_CC(SDNode *N);
|
|
|
|
SDOperand visitSETCC(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitSIGN_EXTEND(SDNode *N);
|
|
|
|
SDOperand visitZERO_EXTEND(SDNode *N);
|
2006-05-05 05:58:59 +00:00
|
|
|
SDOperand visitANY_EXTEND(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitSIGN_EXTEND_INREG(SDNode *N);
|
|
|
|
SDOperand visitTRUNCATE(SDNode *N);
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
SDOperand visitBIT_CONVERT(SDNode *N);
|
2008-05-12 23:04:07 +00:00
|
|
|
SDOperand visitBUILD_PAIR(SDNode *N);
|
2005-09-28 22:28:18 +00:00
|
|
|
SDOperand visitFADD(SDNode *N);
|
|
|
|
SDOperand visitFSUB(SDNode *N);
|
|
|
|
SDOperand visitFMUL(SDNode *N);
|
|
|
|
SDOperand visitFDIV(SDNode *N);
|
|
|
|
SDOperand visitFREM(SDNode *N);
|
2006-03-05 05:30:57 +00:00
|
|
|
SDOperand visitFCOPYSIGN(SDNode *N);
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand visitSINT_TO_FP(SDNode *N);
|
|
|
|
SDOperand visitUINT_TO_FP(SDNode *N);
|
|
|
|
SDOperand visitFP_TO_SINT(SDNode *N);
|
|
|
|
SDOperand visitFP_TO_UINT(SDNode *N);
|
|
|
|
SDOperand visitFP_ROUND(SDNode *N);
|
|
|
|
SDOperand visitFP_ROUND_INREG(SDNode *N);
|
|
|
|
SDOperand visitFP_EXTEND(SDNode *N);
|
|
|
|
SDOperand visitFNEG(SDNode *N);
|
|
|
|
SDOperand visitFABS(SDNode *N);
|
2005-09-16 00:54:12 +00:00
|
|
|
SDOperand visitBRCOND(SDNode *N);
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand visitBR_CC(SDNode *N);
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
SDOperand visitLOAD(SDNode *N);
|
2005-10-10 22:31:19 +00:00
|
|
|
SDOperand visitSTORE(SDNode *N);
|
2006-03-19 01:27:56 +00:00
|
|
|
SDOperand visitINSERT_VECTOR_ELT(SDNode *N);
|
2007-10-06 08:19:55 +00:00
|
|
|
SDOperand visitEXTRACT_VECTOR_ELT(SDNode *N);
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand visitBUILD_VECTOR(SDNode *N);
|
|
|
|
SDOperand visitCONCAT_VECTORS(SDNode *N);
|
2006-03-28 22:11:53 +00:00
|
|
|
SDOperand visitVECTOR_SHUFFLE(SDNode *N);
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
|
2006-04-20 08:56:16 +00:00
|
|
|
SDOperand XformToShuffleWithZero(SDNode *N);
|
2006-02-03 06:46:56 +00:00
|
|
|
SDOperand ReassociateOps(unsigned Opc, SDOperand LHS, SDOperand RHS);
|
|
|
|
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
SDOperand visitShiftByConstant(SDNode *N, unsigned Amt);
|
|
|
|
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
bool SimplifySelectOps(SDNode *SELECT, SDOperand LHS, SDOperand RHS);
|
2006-05-05 05:51:50 +00:00
|
|
|
SDOperand SimplifyBinOpWithSameOpcodeHands(SDNode *N);
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand SimplifySelect(SDOperand N0, SDOperand N1, SDOperand N2);
|
|
|
|
SDOperand SimplifySelectCC(SDOperand N0, SDOperand N1, SDOperand N2,
|
2007-04-11 06:50:51 +00:00
|
|
|
SDOperand N3, ISD::CondCode CC,
|
|
|
|
bool NotExtCompare = false);
|
2008-06-06 12:08:01 +00:00
|
|
|
SDOperand SimplifySetCC(MVT VT, SDOperand N0, SDOperand N1,
|
2005-10-05 21:43:42 +00:00
|
|
|
ISD::CondCode Cond, bool foldBooleans = true);
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand SimplifyNodeWithTwoResults(SDNode *N, unsigned LoOp,
|
|
|
|
unsigned HiOp);
|
2008-06-06 12:08:01 +00:00
|
|
|
SDOperand CombineConsecutiveLoads(SDNode *N, MVT VT);
|
|
|
|
SDOperand ConstantFoldBIT_CONVERTofBUILD_VECTOR(SDNode *, MVT);
|
2005-10-20 02:15:44 +00:00
|
|
|
SDOperand BuildSDIV(SDNode *N);
|
2006-09-14 20:50:57 +00:00
|
|
|
SDOperand BuildUDIV(SDNode *N);
|
|
|
|
SDNode *MatchRotate(SDOperand LHS, SDOperand RHS);
|
2007-03-22 01:54:19 +00:00
|
|
|
SDOperand ReduceLoadWidth(SDNode *N);
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2008-02-25 21:11:39 +00:00
|
|
|
SDOperand GetDemandedBits(SDOperand V, const APInt &Mask);
|
2007-10-13 06:35:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
/// GatherAllAliases - Walk up chain skipping non-aliasing memory nodes,
|
|
|
|
/// looking for aliasing nodes and adding them to the Aliases vector.
|
2006-10-05 15:07:25 +00:00
|
|
|
void GatherAllAliases(SDNode *N, SDOperand OriginalChain,
|
2006-10-04 16:53:27 +00:00
|
|
|
SmallVector<SDOperand, 8> &Aliases);
|
|
|
|
|
2006-10-18 12:29:57 +00:00
|
|
|
/// isAlias - Return true if there is any possibility that the two addresses
|
|
|
|
/// overlap.
|
|
|
|
bool isAlias(SDOperand Ptr1, int64_t Size1,
|
|
|
|
const Value *SrcValue1, int SrcValueOffset1,
|
|
|
|
SDOperand Ptr2, int64_t Size2,
|
2006-11-05 19:31:28 +00:00
|
|
|
const Value *SrcValue2, int SrcValueOffset2);
|
2006-10-18 12:29:57 +00:00
|
|
|
|
2006-10-11 13:47:09 +00:00
|
|
|
/// FindAliasInfo - Extracts the relevant alias information from the memory
|
|
|
|
/// node. Returns true if the operand was a load.
|
|
|
|
bool FindAliasInfo(SDNode *N,
|
2006-10-18 12:29:57 +00:00
|
|
|
SDOperand &Ptr, int64_t &Size,
|
|
|
|
const Value *&SrcValue, int &SrcValueOffset);
|
2006-10-11 13:47:09 +00:00
|
|
|
|
2006-09-25 16:29:54 +00:00
|
|
|
/// FindBetterChain - Walk up chain skipping non-aliasing memory nodes,
|
2006-10-04 16:53:27 +00:00
|
|
|
/// looking for a better chain (aliasing node.)
|
2006-09-25 16:29:54 +00:00
|
|
|
SDOperand FindBetterChain(SDNode *N, SDOperand Chain);
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
public:
|
2006-10-16 20:52:31 +00:00
|
|
|
DAGCombiner(SelectionDAG &D, AliasAnalysis &A)
|
|
|
|
: DAG(D),
|
|
|
|
TLI(D.getTargetLoweringInfo()),
|
|
|
|
AfterLegalize(false),
|
|
|
|
AA(A) {}
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
/// Run - runs the dag combiner on all nodes in the work list
|
2005-09-01 23:24:04 +00:00
|
|
|
void Run(bool RunningAfterLegalize);
|
2005-09-01 00:19:25 +00:00
|
|
|
};
|
|
|
|
}
|
|
|
|
|
2008-02-03 06:49:24 +00:00
|
|
|
|
|
|
|
namespace {
|
|
|
|
/// WorkListRemover - This class is a DAGUpdateListener that removes any deleted
|
|
|
|
/// nodes from the worklist.
|
|
|
|
class VISIBILITY_HIDDEN WorkListRemover :
|
|
|
|
public SelectionDAG::DAGUpdateListener {
|
|
|
|
DAGCombiner &DC;
|
|
|
|
public:
|
2008-02-20 16:44:09 +00:00
|
|
|
explicit WorkListRemover(DAGCombiner &dc) : DC(dc) {}
|
2008-02-03 06:49:24 +00:00
|
|
|
|
2008-06-11 11:42:12 +00:00
|
|
|
virtual void NodeDeleted(SDNode *N, SDNode *E) {
|
2008-02-03 06:49:24 +00:00
|
|
|
DC.removeFromWorkList(N);
|
|
|
|
}
|
|
|
|
|
|
|
|
virtual void NodeUpdated(SDNode *N) {
|
|
|
|
// Ignore updates.
|
|
|
|
}
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
2006-03-01 04:53:38 +00:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// TargetLowering::DAGCombinerInfo implementation
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
void TargetLowering::DAGCombinerInfo::AddToWorklist(SDNode *N) {
|
|
|
|
((DAGCombiner*)DC)->AddToWorkList(N);
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand TargetLowering::DAGCombinerInfo::
|
|
|
|
CombineTo(SDNode *N, const std::vector<SDOperand> &To) {
|
2006-08-11 17:56:38 +00:00
|
|
|
return ((DAGCombiner*)DC)->CombineTo(N, &To[0], To.size());
|
2006-03-01 04:53:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand TargetLowering::DAGCombinerInfo::
|
|
|
|
CombineTo(SDNode *N, SDOperand Res) {
|
|
|
|
return ((DAGCombiner*)DC)->CombineTo(N, Res);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
SDOperand TargetLowering::DAGCombinerInfo::
|
|
|
|
CombineTo(SDNode *N, SDOperand Res0, SDOperand Res1) {
|
|
|
|
return ((DAGCombiner*)DC)->CombineTo(N, Res0, Res1);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Helper Functions
|
|
|
|
//===----------------------------------------------------------------------===//
|
2006-03-01 04:53:38 +00:00
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
/// isNegatibleForFree - Return 1 if we can compute the negated form of the
|
|
|
|
/// specified expression for the same cost as the expression itself, or 2 if we
|
|
|
|
/// can compute the negated form more cheaply than the expression itself.
|
2008-02-26 07:04:54 +00:00
|
|
|
static char isNegatibleForFree(SDOperand Op, bool AfterLegalize,
|
|
|
|
unsigned Depth = 0) {
|
2007-10-16 23:38:29 +00:00
|
|
|
// No compile time optimizations on this type.
|
|
|
|
if (Op.getValueType() == MVT::ppcf128)
|
|
|
|
return 0;
|
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
// fneg is removable even if it has multiple uses.
|
|
|
|
if (Op.getOpcode() == ISD::FNEG) return 2;
|
|
|
|
|
|
|
|
// Don't allow anything with multiple uses.
|
|
|
|
if (!Op.hasOneUse()) return 0;
|
|
|
|
|
2007-05-25 02:19:06 +00:00
|
|
|
// Don't recurse exponentially.
|
|
|
|
if (Depth > 6) return 0;
|
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
switch (Op.getOpcode()) {
|
|
|
|
default: return false;
|
|
|
|
case ISD::ConstantFP:
|
2008-02-26 07:04:54 +00:00
|
|
|
// Don't invert constant FP values after legalize. The negated constant
|
|
|
|
// isn't necessarily legal.
|
|
|
|
return AfterLegalize ? 0 : 1;
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
case ISD::FADD:
|
|
|
|
// FIXME: determine better conditions for this xform.
|
|
|
|
if (!UnsafeFPMath) return 0;
|
|
|
|
|
|
|
|
// -(A+B) -> -A - B
|
2008-02-26 07:04:54 +00:00
|
|
|
if (char V = isNegatibleForFree(Op.getOperand(0), AfterLegalize, Depth+1))
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
return V;
|
|
|
|
// -(A+B) -> -B - A
|
2008-02-26 07:04:54 +00:00
|
|
|
return isNegatibleForFree(Op.getOperand(1), AfterLegalize, Depth+1);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
case ISD::FSUB:
|
|
|
|
// We can't turn -(A-B) into B-A when we honor signed zeros.
|
|
|
|
if (!UnsafeFPMath) return 0;
|
|
|
|
|
|
|
|
// -(A-B) -> B-A
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
case ISD::FMUL:
|
|
|
|
case ISD::FDIV:
|
|
|
|
if (HonorSignDependentRoundingFPMath()) return 0;
|
|
|
|
|
|
|
|
// -(X*Y) -> (-X * Y) or (X*-Y)
|
2008-02-26 07:04:54 +00:00
|
|
|
if (char V = isNegatibleForFree(Op.getOperand(0), AfterLegalize, Depth+1))
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
return V;
|
|
|
|
|
2008-02-26 07:04:54 +00:00
|
|
|
return isNegatibleForFree(Op.getOperand(1), AfterLegalize, Depth+1);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
|
|
|
|
case ISD::FP_EXTEND:
|
|
|
|
case ISD::FP_ROUND:
|
|
|
|
case ISD::FSIN:
|
2008-02-26 07:04:54 +00:00
|
|
|
return isNegatibleForFree(Op.getOperand(0), AfterLegalize, Depth+1);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
}
|
|
|
|
}
|
2006-03-01 04:53:38 +00:00
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
/// GetNegatedExpression - If isNegatibleForFree returns true, this function
|
|
|
|
/// returns the newly negated expression.
|
2007-05-25 02:19:06 +00:00
|
|
|
static SDOperand GetNegatedExpression(SDOperand Op, SelectionDAG &DAG,
|
2008-02-26 07:04:54 +00:00
|
|
|
bool AfterLegalize, unsigned Depth = 0) {
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
// fneg is removable even if it has multiple uses.
|
|
|
|
if (Op.getOpcode() == ISD::FNEG) return Op.getOperand(0);
|
|
|
|
|
|
|
|
// Don't allow anything with multiple uses.
|
|
|
|
assert(Op.hasOneUse() && "Unknown reuse!");
|
|
|
|
|
2007-05-25 02:19:06 +00:00
|
|
|
assert(Depth <= 6 && "GetNegatedExpression doesn't match isNegatibleForFree");
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
switch (Op.getOpcode()) {
|
|
|
|
default: assert(0 && "Unknown code");
|
2007-08-31 23:34:27 +00:00
|
|
|
case ISD::ConstantFP: {
|
|
|
|
APFloat V = cast<ConstantFPSDNode>(Op)->getValueAPF();
|
|
|
|
V.changeSign();
|
|
|
|
return DAG.getConstantFP(V, Op.getValueType());
|
|
|
|
}
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
case ISD::FADD:
|
|
|
|
// FIXME: determine better conditions for this xform.
|
|
|
|
assert(UnsafeFPMath);
|
|
|
|
|
|
|
|
// -(A+B) -> -A - B
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(Op.getOperand(0), AfterLegalize, Depth+1))
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
return DAG.getNode(ISD::FSUB, Op.getValueType(),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(0), DAG,
|
|
|
|
AfterLegalize, Depth+1),
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
Op.getOperand(1));
|
|
|
|
// -(A+B) -> -B - A
|
|
|
|
return DAG.getNode(ISD::FSUB, Op.getValueType(),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(1), DAG,
|
|
|
|
AfterLegalize, Depth+1),
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
Op.getOperand(0));
|
|
|
|
case ISD::FSUB:
|
|
|
|
// We can't turn -(A-B) into B-A when we honor signed zeros.
|
|
|
|
assert(UnsafeFPMath);
|
2007-07-02 15:48:56 +00:00
|
|
|
|
|
|
|
// -(0-B) -> B
|
|
|
|
if (ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(Op.getOperand(0)))
|
2007-08-31 23:34:27 +00:00
|
|
|
if (N0CFP->getValueAPF().isZero())
|
2007-07-02 15:48:56 +00:00
|
|
|
return Op.getOperand(1);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
|
|
|
|
// -(A-B) -> B-A
|
|
|
|
return DAG.getNode(ISD::FSUB, Op.getValueType(), Op.getOperand(1),
|
|
|
|
Op.getOperand(0));
|
|
|
|
|
|
|
|
case ISD::FMUL:
|
|
|
|
case ISD::FDIV:
|
|
|
|
assert(!HonorSignDependentRoundingFPMath());
|
|
|
|
|
|
|
|
// -(X*Y) -> -X * Y
|
2008-02-26 17:09:59 +00:00
|
|
|
if (isNegatibleForFree(Op.getOperand(0), AfterLegalize, Depth+1))
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
return DAG.getNode(Op.getOpcode(), Op.getValueType(),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(0), DAG,
|
|
|
|
AfterLegalize, Depth+1),
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
Op.getOperand(1));
|
|
|
|
|
|
|
|
// -(X*Y) -> X * -Y
|
|
|
|
return DAG.getNode(Op.getOpcode(), Op.getValueType(),
|
|
|
|
Op.getOperand(0),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(1), DAG,
|
|
|
|
AfterLegalize, Depth+1));
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
|
|
|
|
case ISD::FP_EXTEND:
|
|
|
|
case ISD::FSIN:
|
|
|
|
return DAG.getNode(Op.getOpcode(), Op.getValueType(),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(0), DAG,
|
|
|
|
AfterLegalize, Depth+1));
|
2008-01-17 07:00:52 +00:00
|
|
|
case ISD::FP_ROUND:
|
|
|
|
return DAG.getNode(ISD::FP_ROUND, Op.getValueType(),
|
2008-02-26 07:04:54 +00:00
|
|
|
GetNegatedExpression(Op.getOperand(0), DAG,
|
|
|
|
AfterLegalize, Depth+1),
|
2008-01-17 07:00:52 +00:00
|
|
|
Op.getOperand(1));
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
}
|
|
|
|
}
|
2006-03-01 04:53:38 +00:00
|
|
|
|
|
|
|
|
2005-09-01 23:24:04 +00:00
|
|
|
// isSetCCEquivalent - Return true if this node is a setcc, or is a select_cc
|
|
|
|
// that selects between the values 1 and 0, making it equivalent to a setcc.
|
2005-09-02 21:18:40 +00:00
|
|
|
// Also, set the incoming LHS, RHS, and CC references to the appropriate
|
|
|
|
// nodes based on the type of node we are checking. This simplifies life a
|
|
|
|
// bit for the callers.
|
|
|
|
static bool isSetCCEquivalent(SDOperand N, SDOperand &LHS, SDOperand &RHS,
|
|
|
|
SDOperand &CC) {
|
|
|
|
if (N.getOpcode() == ISD::SETCC) {
|
|
|
|
LHS = N.getOperand(0);
|
|
|
|
RHS = N.getOperand(1);
|
|
|
|
CC = N.getOperand(2);
|
2005-09-01 23:24:04 +00:00
|
|
|
return true;
|
2005-09-02 21:18:40 +00:00
|
|
|
}
|
2005-09-01 00:19:25 +00:00
|
|
|
if (N.getOpcode() == ISD::SELECT_CC &&
|
|
|
|
N.getOperand(2).getOpcode() == ISD::Constant &&
|
|
|
|
N.getOperand(3).getOpcode() == ISD::Constant &&
|
2008-03-13 22:13:53 +00:00
|
|
|
cast<ConstantSDNode>(N.getOperand(2))->getAPIntValue() == 1 &&
|
2005-09-02 21:18:40 +00:00
|
|
|
cast<ConstantSDNode>(N.getOperand(3))->isNullValue()) {
|
|
|
|
LHS = N.getOperand(0);
|
|
|
|
RHS = N.getOperand(1);
|
|
|
|
CC = N.getOperand(4);
|
2005-09-01 00:19:25 +00:00
|
|
|
return true;
|
2005-09-02 21:18:40 +00:00
|
|
|
}
|
2005-09-01 00:19:25 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2005-09-07 23:25:52 +00:00
|
|
|
// isOneUseSetCC - Return true if this is a SetCC-equivalent operation with only
|
|
|
|
// one use. If this is true, it allows the users to invert the operation for
|
|
|
|
// free when it is profitable to do so.
|
|
|
|
static bool isOneUseSetCC(SDOperand N) {
|
2005-09-02 21:18:40 +00:00
|
|
|
SDOperand N0, N1, N2;
|
|
|
|
if (isSetCCEquivalent(N, N0, N1, N2) && N.Val->hasOneUse())
|
2005-09-01 23:24:04 +00:00
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2006-02-03 06:46:56 +00:00
|
|
|
SDOperand DAGCombiner::ReassociateOps(unsigned Opc, SDOperand N0, SDOperand N1){
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassoc. (op (op x, c1), y) -> (op (op x, y), c1) iff x+c1 has one use
|
|
|
|
// reassoc. (op (op x, c1), c2) -> (op x, (op c1, c2))
|
|
|
|
if (N0.getOpcode() == Opc && isa<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
if (isa<ConstantSDNode>(N1)) {
|
|
|
|
SDOperand OpNode = DAG.getNode(Opc, VT, N0.getOperand(1), N1);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(OpNode.Val);
|
2006-02-03 06:46:56 +00:00
|
|
|
return DAG.getNode(Opc, VT, OpNode, N0.getOperand(0));
|
|
|
|
} else if (N0.hasOneUse()) {
|
|
|
|
SDOperand OpNode = DAG.getNode(Opc, VT, N0.getOperand(0), N1);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(OpNode.Val);
|
2006-02-03 06:46:56 +00:00
|
|
|
return DAG.getNode(Opc, VT, OpNode, N0.getOperand(1));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// reassoc. (op y, (op x, c1)) -> (op (op x, y), c1) iff x+c1 has one use
|
|
|
|
// reassoc. (op c2, (op x, c1)) -> (op x, (op c1, c2))
|
|
|
|
if (N1.getOpcode() == Opc && isa<ConstantSDNode>(N1.getOperand(1))) {
|
|
|
|
if (isa<ConstantSDNode>(N0)) {
|
|
|
|
SDOperand OpNode = DAG.getNode(Opc, VT, N1.getOperand(1), N0);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(OpNode.Val);
|
2006-02-03 06:46:56 +00:00
|
|
|
return DAG.getNode(Opc, VT, OpNode, N1.getOperand(0));
|
|
|
|
} else if (N1.hasOneUse()) {
|
|
|
|
SDOperand OpNode = DAG.getNode(Opc, VT, N1.getOperand(0), N0);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(OpNode.Val);
|
2006-02-03 06:46:56 +00:00
|
|
|
return DAG.getNode(Opc, VT, OpNode, N1.getOperand(1));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2008-02-03 06:49:24 +00:00
|
|
|
SDOperand DAGCombiner::CombineTo(SDNode *N, const SDOperand *To, unsigned NumTo,
|
|
|
|
bool AddTo) {
|
|
|
|
assert(N->getNumValues() == NumTo && "Broken CombineTo call!");
|
|
|
|
++NodesCombined;
|
|
|
|
DOUT << "\nReplacing.1 "; DEBUG(N->dump(&DAG));
|
|
|
|
DOUT << "\nWith: "; DEBUG(To[0].Val->dump(&DAG));
|
|
|
|
DOUT << " and " << NumTo-1 << " other values\n";
|
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
DAG.ReplaceAllUsesWith(N, To, &DeadNodes);
|
|
|
|
|
|
|
|
if (AddTo) {
|
|
|
|
// Push the new nodes and any users onto the worklist
|
|
|
|
for (unsigned i = 0, e = NumTo; i != e; ++i) {
|
|
|
|
AddToWorkList(To[i].Val);
|
|
|
|
AddUsersToWorkList(To[i].Val);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Nodes can be reintroduced into the worklist. Make sure we do not
|
|
|
|
// process a node that has been replaced.
|
|
|
|
removeFromWorkList(N);
|
|
|
|
|
|
|
|
// Finally, since the node is now dead, remove it from the graph.
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
return SDOperand(N, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// SimplifyDemandedBits - Check the specified integer node value to see if
|
|
|
|
/// it can be simplified or if things it uses can be simplified by bit
|
|
|
|
/// propagation. If so, return true.
|
2008-02-27 00:25:32 +00:00
|
|
|
bool DAGCombiner::SimplifyDemandedBits(SDOperand Op, const APInt &Demanded) {
|
2008-02-03 06:49:24 +00:00
|
|
|
TargetLowering::TargetLoweringOpt TLO(DAG, AfterLegalize);
|
2008-02-27 00:25:32 +00:00
|
|
|
APInt KnownZero, KnownOne;
|
2008-02-03 06:49:24 +00:00
|
|
|
if (!TLI.SimplifyDemandedBits(Op, Demanded, KnownZero, KnownOne, TLO))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Revisit the node.
|
|
|
|
AddToWorkList(Op.Val);
|
|
|
|
|
|
|
|
// Replace the old value with the new one.
|
|
|
|
++NodesCombined;
|
|
|
|
DOUT << "\nReplacing.2 "; DEBUG(TLO.Old.Val->dump(&DAG));
|
|
|
|
DOUT << "\nWith: "; DEBUG(TLO.New.Val->dump(&DAG));
|
|
|
|
DOUT << '\n';
|
|
|
|
|
|
|
|
// Replace all uses. If any nodes become isomorphic to other nodes and
|
|
|
|
// are deleted, make sure to remove them from our worklist.
|
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(TLO.Old, TLO.New, &DeadNodes);
|
|
|
|
|
|
|
|
// Push the new node and any (possibly new) users onto the worklist.
|
|
|
|
AddToWorkList(TLO.New.Val);
|
|
|
|
AddUsersToWorkList(TLO.New.Val);
|
|
|
|
|
|
|
|
// Finally, if the node is now dead, remove it from the graph. The node
|
|
|
|
// may not be dead if the replacement process recursively simplified to
|
|
|
|
// something else needing this node.
|
|
|
|
if (TLO.Old.Val->use_empty()) {
|
|
|
|
removeFromWorkList(TLO.Old.Val);
|
|
|
|
|
|
|
|
// If the operands of this node are only used by the node, they will now
|
|
|
|
// be dead. Make sure to visit them first to delete dead nodes early.
|
|
|
|
for (unsigned i = 0, e = TLO.Old.Val->getNumOperands(); i != e; ++i)
|
|
|
|
if (TLO.Old.Val->getOperand(i).Val->hasOneUse())
|
|
|
|
AddToWorkList(TLO.Old.Val->getOperand(i).Val);
|
|
|
|
|
|
|
|
DAG.DeleteNode(TLO.Old.Val);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Main DAG Combiner implementation
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2005-09-01 23:24:04 +00:00
|
|
|
void DAGCombiner::Run(bool RunningAfterLegalize) {
|
|
|
|
// set the instance variable, so that the various visit routines may use it.
|
|
|
|
AfterLegalize = RunningAfterLegalize;
|
|
|
|
|
2005-09-02 21:18:40 +00:00
|
|
|
// Add all the dag nodes to the worklist.
|
2008-06-30 21:04:06 +00:00
|
|
|
WorkList.reserve(DAG.allnodes_size());
|
2005-11-09 23:47:37 +00:00
|
|
|
for (SelectionDAG::allnodes_iterator I = DAG.allnodes_begin(),
|
|
|
|
E = DAG.allnodes_end(); I != E; ++I)
|
|
|
|
WorkList.push_back(I);
|
2005-09-06 04:43:02 +00:00
|
|
|
|
2005-10-05 06:35:28 +00:00
|
|
|
// Create a dummy node (which is not added to allnodes), that adds a reference
|
|
|
|
// to the root node, preventing it from being deleted, and tracking any
|
|
|
|
// changes of the root.
|
|
|
|
HandleSDNode Dummy(DAG.getRoot());
|
|
|
|
|
2006-10-17 19:33:52 +00:00
|
|
|
// The root of the dag may dangle to deleted nodes until the dag combiner is
|
|
|
|
// done. Set it to null to avoid confusion.
|
|
|
|
DAG.setRoot(SDOperand());
|
2006-03-01 04:53:38 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// while the worklist isn't empty, inspect the node on the end of it and
|
|
|
|
// try and combine it.
|
|
|
|
while (!WorkList.empty()) {
|
|
|
|
SDNode *N = WorkList.back();
|
|
|
|
WorkList.pop_back();
|
|
|
|
|
|
|
|
// If N has no uses, it is dead. Make sure to revisit all N's operands once
|
2005-10-05 06:35:28 +00:00
|
|
|
// N is deleted from the DAG, since they too may now be dead or may have a
|
|
|
|
// reduced number of uses, allowing other xforms.
|
|
|
|
if (N->use_empty() && N != &Dummy) {
|
2005-09-01 00:19:25 +00:00
|
|
|
for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i)
|
2006-10-04 16:53:27 +00:00
|
|
|
AddToWorkList(N->getOperand(i).Val);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2005-10-05 06:35:28 +00:00
|
|
|
DAG.DeleteNode(N);
|
2005-09-01 00:19:25 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
SDOperand RV = combine(N);
|
2006-03-01 04:53:38 +00:00
|
|
|
|
2008-01-25 23:34:24 +00:00
|
|
|
if (RV.Val == 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
++NodesCombined;
|
2008-01-26 01:09:19 +00:00
|
|
|
|
2008-01-25 23:34:24 +00:00
|
|
|
// If we get back the same node we passed in, rather than a new node or
|
|
|
|
// zero, we know that the node must have defined multiple values and
|
|
|
|
// CombineTo was used. Since CombineTo takes care of the worklist
|
|
|
|
// mechanics for us, we have no work to do in this case.
|
|
|
|
if (RV.Val == N)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
assert(N->getOpcode() != ISD::DELETED_NODE &&
|
|
|
|
RV.Val->getOpcode() != ISD::DELETED_NODE &&
|
|
|
|
"Node was deleted but visit returned new node!");
|
|
|
|
|
|
|
|
DOUT << "\nReplacing.3 "; DEBUG(N->dump(&DAG));
|
|
|
|
DOUT << "\nWith: "; DEBUG(RV.Val->dump(&DAG));
|
|
|
|
DOUT << '\n';
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
2008-01-25 23:34:24 +00:00
|
|
|
if (N->getNumValues() == RV.Val->getNumValues())
|
2008-02-03 06:49:24 +00:00
|
|
|
DAG.ReplaceAllUsesWith(N, RV.Val, &DeadNodes);
|
2008-01-25 23:34:24 +00:00
|
|
|
else {
|
2008-01-26 01:09:19 +00:00
|
|
|
assert(N->getValueType(0) == RV.getValueType() &&
|
|
|
|
N->getNumValues() == 1 && "Type mismatch");
|
2008-01-25 23:34:24 +00:00
|
|
|
SDOperand OpV = RV;
|
2008-02-03 06:49:24 +00:00
|
|
|
DAG.ReplaceAllUsesWith(N, &OpV, &DeadNodes);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2008-01-25 23:34:24 +00:00
|
|
|
|
|
|
|
// Push the new node and any users onto the worklist
|
|
|
|
AddToWorkList(RV.Val);
|
|
|
|
AddUsersToWorkList(RV.Val);
|
|
|
|
|
|
|
|
// Add any uses of the old node to the worklist in case this node is the
|
|
|
|
// last one that uses them. They may become dead after this node is
|
|
|
|
// deleted.
|
|
|
|
for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i)
|
|
|
|
AddToWorkList(N->getOperand(i).Val);
|
|
|
|
|
|
|
|
// Nodes can be reintroduced into the worklist. Make sure we do not
|
|
|
|
// process a node that has been replaced.
|
|
|
|
removeFromWorkList(N);
|
|
|
|
|
|
|
|
// Finally, since the node is now dead, remove it from the graph.
|
|
|
|
DAG.DeleteNode(N);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2005-10-05 06:35:28 +00:00
|
|
|
|
|
|
|
// If the root changed (e.g. it was a dead load, update the root).
|
|
|
|
DAG.setRoot(Dummy.getValue());
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visit(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
switch(N->getOpcode()) {
|
|
|
|
default: break;
|
2005-09-01 00:33:32 +00:00
|
|
|
case ISD::TokenFactor: return visitTokenFactor(N);
|
2008-02-13 07:25:05 +00:00
|
|
|
case ISD::MERGE_VALUES: return visitMERGE_VALUES(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::ADD: return visitADD(N);
|
|
|
|
case ISD::SUB: return visitSUB(N);
|
2007-03-04 20:03:15 +00:00
|
|
|
case ISD::ADDC: return visitADDC(N);
|
|
|
|
case ISD::ADDE: return visitADDE(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::MUL: return visitMUL(N);
|
|
|
|
case ISD::SDIV: return visitSDIV(N);
|
|
|
|
case ISD::UDIV: return visitUDIV(N);
|
|
|
|
case ISD::SREM: return visitSREM(N);
|
|
|
|
case ISD::UREM: return visitUREM(N);
|
|
|
|
case ISD::MULHU: return visitMULHU(N);
|
|
|
|
case ISD::MULHS: return visitMULHS(N);
|
2007-10-08 17:57:15 +00:00
|
|
|
case ISD::SMUL_LOHI: return visitSMUL_LOHI(N);
|
|
|
|
case ISD::UMUL_LOHI: return visitUMUL_LOHI(N);
|
|
|
|
case ISD::SDIVREM: return visitSDIVREM(N);
|
|
|
|
case ISD::UDIVREM: return visitUDIVREM(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::AND: return visitAND(N);
|
|
|
|
case ISD::OR: return visitOR(N);
|
|
|
|
case ISD::XOR: return visitXOR(N);
|
|
|
|
case ISD::SHL: return visitSHL(N);
|
|
|
|
case ISD::SRA: return visitSRA(N);
|
|
|
|
case ISD::SRL: return visitSRL(N);
|
|
|
|
case ISD::CTLZ: return visitCTLZ(N);
|
|
|
|
case ISD::CTTZ: return visitCTTZ(N);
|
|
|
|
case ISD::CTPOP: return visitCTPOP(N);
|
2005-09-16 00:54:12 +00:00
|
|
|
case ISD::SELECT: return visitSELECT(N);
|
|
|
|
case ISD::SELECT_CC: return visitSELECT_CC(N);
|
|
|
|
case ISD::SETCC: return visitSETCC(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::SIGN_EXTEND: return visitSIGN_EXTEND(N);
|
|
|
|
case ISD::ZERO_EXTEND: return visitZERO_EXTEND(N);
|
2006-05-05 05:58:59 +00:00
|
|
|
case ISD::ANY_EXTEND: return visitANY_EXTEND(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::SIGN_EXTEND_INREG: return visitSIGN_EXTEND_INREG(N);
|
|
|
|
case ISD::TRUNCATE: return visitTRUNCATE(N);
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
case ISD::BIT_CONVERT: return visitBIT_CONVERT(N);
|
2008-05-12 23:04:07 +00:00
|
|
|
case ISD::BUILD_PAIR: return visitBUILD_PAIR(N);
|
2005-09-28 22:28:18 +00:00
|
|
|
case ISD::FADD: return visitFADD(N);
|
|
|
|
case ISD::FSUB: return visitFSUB(N);
|
|
|
|
case ISD::FMUL: return visitFMUL(N);
|
|
|
|
case ISD::FDIV: return visitFDIV(N);
|
|
|
|
case ISD::FREM: return visitFREM(N);
|
2006-03-05 05:30:57 +00:00
|
|
|
case ISD::FCOPYSIGN: return visitFCOPYSIGN(N);
|
2005-09-02 21:18:40 +00:00
|
|
|
case ISD::SINT_TO_FP: return visitSINT_TO_FP(N);
|
|
|
|
case ISD::UINT_TO_FP: return visitUINT_TO_FP(N);
|
|
|
|
case ISD::FP_TO_SINT: return visitFP_TO_SINT(N);
|
|
|
|
case ISD::FP_TO_UINT: return visitFP_TO_UINT(N);
|
|
|
|
case ISD::FP_ROUND: return visitFP_ROUND(N);
|
|
|
|
case ISD::FP_ROUND_INREG: return visitFP_ROUND_INREG(N);
|
|
|
|
case ISD::FP_EXTEND: return visitFP_EXTEND(N);
|
|
|
|
case ISD::FNEG: return visitFNEG(N);
|
|
|
|
case ISD::FABS: return visitFABS(N);
|
2005-09-19 22:34:01 +00:00
|
|
|
case ISD::BRCOND: return visitBRCOND(N);
|
|
|
|
case ISD::BR_CC: return visitBR_CC(N);
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
case ISD::LOAD: return visitLOAD(N);
|
2005-10-10 22:31:19 +00:00
|
|
|
case ISD::STORE: return visitSTORE(N);
|
2006-03-19 01:27:56 +00:00
|
|
|
case ISD::INSERT_VECTOR_ELT: return visitINSERT_VECTOR_ELT(N);
|
2007-10-06 08:19:55 +00:00
|
|
|
case ISD::EXTRACT_VECTOR_ELT: return visitEXTRACT_VECTOR_ELT(N);
|
2007-06-25 16:23:39 +00:00
|
|
|
case ISD::BUILD_VECTOR: return visitBUILD_VECTOR(N);
|
|
|
|
case ISD::CONCAT_VECTORS: return visitCONCAT_VECTORS(N);
|
2006-03-28 22:11:53 +00:00
|
|
|
case ISD::VECTOR_SHUFFLE: return visitVECTOR_SHUFFLE(N);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
SDOperand DAGCombiner::combine(SDNode *N) {
|
|
|
|
|
|
|
|
SDOperand RV = visit(N);
|
|
|
|
|
|
|
|
// If nothing happened, try a target-specific DAG combine.
|
|
|
|
if (RV.Val == 0) {
|
|
|
|
assert(N->getOpcode() != ISD::DELETED_NODE &&
|
|
|
|
"Node was deleted but visit returned NULL!");
|
|
|
|
|
|
|
|
if (N->getOpcode() >= ISD::BUILTIN_OP_END ||
|
|
|
|
TLI.hasTargetDAGCombine((ISD::NodeType)N->getOpcode())) {
|
|
|
|
|
|
|
|
// Expose the DAG combiner to the target combiner impls.
|
|
|
|
TargetLowering::DAGCombinerInfo
|
|
|
|
DagCombineInfo(DAG, !AfterLegalize, false, this);
|
|
|
|
|
|
|
|
RV = TLI.PerformDAGCombine(N, DagCombineInfo);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-03-22 01:55:50 +00:00
|
|
|
// If N is a commutative binary node, try commuting it to enable more
|
|
|
|
// sdisel CSE.
|
|
|
|
if (RV.Val == 0 &&
|
|
|
|
SelectionDAG::isCommutativeBinOp(N->getOpcode()) &&
|
|
|
|
N->getNumValues() == 1) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
// Constant operands are canonicalized to RHS.
|
|
|
|
if (isa<ConstantSDNode>(N0) || !isa<ConstantSDNode>(N1)) {
|
|
|
|
SDOperand Ops[] = { N1, N0 };
|
|
|
|
SDNode *CSENode = DAG.getNodeIfExists(N->getOpcode(), N->getVTList(),
|
|
|
|
Ops, 2);
|
2008-03-24 23:55:16 +00:00
|
|
|
if (CSENode)
|
2008-03-22 01:55:50 +00:00
|
|
|
return SDOperand(CSENode, 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
return RV;
|
|
|
|
}
|
|
|
|
|
Eliminate more token factors by taking advantage of transitivity:
if TF depends on A and B, and A depends on B, TF just needs to depend on
A. With Jim's alias-analysis stuff enabled, this compiles the testcase in
PR892 into:
__Z4test3Val:
subl $44, %esp
call L__Z3foov$stub
movl %edx, 28(%esp)
movl %eax, 32(%esp)
movl %eax, 24(%esp)
movl %edx, 36(%esp)
movl 52(%esp), %ecx
movl %ecx, 4(%esp)
movl %eax, 8(%esp)
movl %edx, 12(%esp)
movl 48(%esp), %eax
movl %eax, (%esp)
call L__Z3bar3ValS_$stub
addl $44, %esp
ret
instead of:
__Z4test3Val:
subl $44, %esp
call L__Z3foov$stub
movl %eax, 24(%esp)
movl %edx, 28(%esp)
movl 24(%esp), %eax
movl %eax, 32(%esp)
movl 28(%esp), %eax
movl %eax, 36(%esp)
movl 32(%esp), %eax
movl 36(%esp), %ecx
movl 52(%esp), %edx
movl %edx, 4(%esp)
movl %eax, 8(%esp)
movl %ecx, 12(%esp)
movl 48(%esp), %eax
movl %eax, (%esp)
call L__Z3bar3ValS_$stub
addl $44, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30821 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-08 22:57:01 +00:00
|
|
|
/// getInputChainForNode - Given a node, return its input chain if it has one,
|
|
|
|
/// otherwise return a null sd operand.
|
|
|
|
static SDOperand getInputChainForNode(SDNode *N) {
|
|
|
|
if (unsigned NumOps = N->getNumOperands()) {
|
|
|
|
if (N->getOperand(0).getValueType() == MVT::Other)
|
|
|
|
return N->getOperand(0);
|
|
|
|
else if (N->getOperand(NumOps-1).getValueType() == MVT::Other)
|
|
|
|
return N->getOperand(NumOps-1);
|
|
|
|
for (unsigned i = 1; i < NumOps-1; ++i)
|
|
|
|
if (N->getOperand(i).getValueType() == MVT::Other)
|
|
|
|
return N->getOperand(i);
|
|
|
|
}
|
|
|
|
return SDOperand(0, 0);
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitTokenFactor(SDNode *N) {
|
Eliminate more token factors by taking advantage of transitivity:
if TF depends on A and B, and A depends on B, TF just needs to depend on
A. With Jim's alias-analysis stuff enabled, this compiles the testcase in
PR892 into:
__Z4test3Val:
subl $44, %esp
call L__Z3foov$stub
movl %edx, 28(%esp)
movl %eax, 32(%esp)
movl %eax, 24(%esp)
movl %edx, 36(%esp)
movl 52(%esp), %ecx
movl %ecx, 4(%esp)
movl %eax, 8(%esp)
movl %edx, 12(%esp)
movl 48(%esp), %eax
movl %eax, (%esp)
call L__Z3bar3ValS_$stub
addl $44, %esp
ret
instead of:
__Z4test3Val:
subl $44, %esp
call L__Z3foov$stub
movl %eax, 24(%esp)
movl %edx, 28(%esp)
movl 24(%esp), %eax
movl %eax, 32(%esp)
movl 28(%esp), %eax
movl %eax, 36(%esp)
movl 32(%esp), %eax
movl 36(%esp), %ecx
movl 52(%esp), %edx
movl %edx, 4(%esp)
movl %eax, 8(%esp)
movl %ecx, 12(%esp)
movl 48(%esp), %eax
movl %eax, (%esp)
call L__Z3bar3ValS_$stub
addl $44, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30821 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-08 22:57:01 +00:00
|
|
|
// If N has two operands, where one has an input chain equal to the other,
|
|
|
|
// the 'other' chain is redundant.
|
|
|
|
if (N->getNumOperands() == 2) {
|
|
|
|
if (getInputChainForNode(N->getOperand(0).Val) == N->getOperand(1))
|
|
|
|
return N->getOperand(0);
|
|
|
|
if (getInputChainForNode(N->getOperand(1).Val) == N->getOperand(0))
|
|
|
|
return N->getOperand(1);
|
|
|
|
}
|
|
|
|
|
2007-05-16 06:37:59 +00:00
|
|
|
SmallVector<SDNode *, 8> TFs; // List of token factors to visit.
|
|
|
|
SmallVector<SDOperand, 8> Ops; // Ops for replacing token factor.
|
|
|
|
SmallPtrSet<SDNode*, 16> SeenOps;
|
|
|
|
bool Changed = false; // If we should replace this token factor.
|
2006-10-04 16:53:27 +00:00
|
|
|
|
|
|
|
// Start out with this token factor.
|
2006-09-25 16:29:54 +00:00
|
|
|
TFs.push_back(N);
|
|
|
|
|
2006-10-07 23:37:56 +00:00
|
|
|
// Iterate through token factors. The TFs grows when new token factors are
|
2006-10-05 15:07:25 +00:00
|
|
|
// encountered.
|
|
|
|
for (unsigned i = 0; i < TFs.size(); ++i) {
|
|
|
|
SDNode *TF = TFs[i];
|
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
// Check each of the operands.
|
|
|
|
for (unsigned i = 0, ie = TF->getNumOperands(); i != ie; ++i) {
|
|
|
|
SDOperand Op = TF->getOperand(i);
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
switch (Op.getOpcode()) {
|
|
|
|
case ISD::EntryToken:
|
2006-10-05 15:07:25 +00:00
|
|
|
// Entry tokens don't need to be added to the list. They are
|
|
|
|
// rededundant.
|
|
|
|
Changed = true;
|
2006-10-04 16:53:27 +00:00
|
|
|
break;
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
case ISD::TokenFactor:
|
2006-10-05 15:07:25 +00:00
|
|
|
if ((CombinerAA || Op.hasOneUse()) &&
|
|
|
|
std::find(TFs.begin(), TFs.end(), Op.Val) == TFs.end()) {
|
2006-10-04 16:53:27 +00:00
|
|
|
// Queue up for processing.
|
|
|
|
TFs.push_back(Op.Val);
|
|
|
|
// Clean up in case the token factor is removed.
|
|
|
|
AddToWorkList(Op.Val);
|
|
|
|
Changed = true;
|
|
|
|
break;
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
2006-10-04 16:53:27 +00:00
|
|
|
// Fall thru
|
|
|
|
|
|
|
|
default:
|
2007-05-16 06:37:59 +00:00
|
|
|
// Only add if it isn't already in the list.
|
|
|
|
if (SeenOps.insert(Op.Val))
|
2006-10-05 15:07:25 +00:00
|
|
|
Ops.push_back(Op);
|
2007-05-16 06:37:59 +00:00
|
|
|
else
|
|
|
|
Changed = true;
|
2006-10-04 16:53:27 +00:00
|
|
|
break;
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
|
|
|
}
|
2006-10-04 16:53:27 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand Result;
|
|
|
|
|
|
|
|
// If we've change things around then replace token factor.
|
|
|
|
if (Changed) {
|
2008-01-29 13:02:09 +00:00
|
|
|
if (Ops.empty()) {
|
2006-10-04 16:53:27 +00:00
|
|
|
// The entry token is the only possible outcome.
|
|
|
|
Result = DAG.getEntryNode();
|
|
|
|
} else {
|
|
|
|
// New and improved token factor.
|
|
|
|
Result = DAG.getNode(ISD::TokenFactor, MVT::Other, &Ops[0], Ops.size());
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
2006-10-13 23:32:28 +00:00
|
|
|
|
|
|
|
// Don't add users to work list.
|
|
|
|
return CombineTo(N, Result, false);
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
return Result;
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2008-02-13 07:25:05 +00:00
|
|
|
/// MERGE_VALUES can always be eliminated.
|
|
|
|
SDOperand DAGCombiner::visitMERGE_VALUES(SDNode *N) {
|
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i)
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, i), N->getOperand(i),
|
|
|
|
&DeadNodes);
|
|
|
|
removeFromWorkList(N);
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
|
|
|
|
|
|
|
|
Remove this xform:
(shl (add x, c1), c2) -> (add (shl x, c2), c1<<c2)
Replace it with:
(add (shl (add x, c1), c2), ) -> (add (add (shl x, c2), c1<<c2), )
This fixes test/CodeGen/ARM/smul.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@33361 91177308-0d34-0410-b5e6-96231b3b80d8
2007-01-19 17:51:44 +00:00
|
|
|
static
|
|
|
|
SDOperand combineShlAddConstant(SDOperand N0, SDOperand N1, SelectionDAG &DAG) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
Remove this xform:
(shl (add x, c1), c2) -> (add (shl x, c2), c1<<c2)
Replace it with:
(add (shl (add x, c1), c2), ) -> (add (add (shl x, c2), c1<<c2), )
This fixes test/CodeGen/ARM/smul.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@33361 91177308-0d34-0410-b5e6-96231b3b80d8
2007-01-19 17:51:44 +00:00
|
|
|
SDOperand N00 = N0.getOperand(0);
|
|
|
|
SDOperand N01 = N0.getOperand(1);
|
|
|
|
ConstantSDNode *N01C = dyn_cast<ConstantSDNode>(N01);
|
|
|
|
if (N01C && N00.getOpcode() == ISD::ADD && N00.Val->hasOneUse() &&
|
|
|
|
isa<ConstantSDNode>(N00.getOperand(1))) {
|
|
|
|
N0 = DAG.getNode(ISD::ADD, VT,
|
|
|
|
DAG.getNode(ISD::SHL, VT, N00.getOperand(0), N01),
|
|
|
|
DAG.getNode(ISD::SHL, VT, N00.getOperand(1), N01));
|
|
|
|
return DAG.getNode(ISD::ADD, VT, N0, N1);
|
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
Xforms:
(add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
(sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37685 91177308-0d34-0410-b5e6-96231b3b80d8
2007-06-21 07:39:16 +00:00
|
|
|
static
|
|
|
|
SDOperand combineSelectAndUse(SDNode *N, SDOperand Slct, SDOperand OtherOp,
|
|
|
|
SelectionDAG &DAG) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
Xforms:
(add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
(sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37685 91177308-0d34-0410-b5e6-96231b3b80d8
2007-06-21 07:39:16 +00:00
|
|
|
unsigned Opc = N->getOpcode();
|
|
|
|
bool isSlctCC = Slct.getOpcode() == ISD::SELECT_CC;
|
|
|
|
SDOperand LHS = isSlctCC ? Slct.getOperand(2) : Slct.getOperand(1);
|
|
|
|
SDOperand RHS = isSlctCC ? Slct.getOperand(3) : Slct.getOperand(2);
|
|
|
|
ISD::CondCode CC = ISD::SETCC_INVALID;
|
|
|
|
if (isSlctCC)
|
|
|
|
CC = cast<CondCodeSDNode>(Slct.getOperand(4))->get();
|
|
|
|
else {
|
|
|
|
SDOperand CCOp = Slct.getOperand(0);
|
|
|
|
if (CCOp.getOpcode() == ISD::SETCC)
|
|
|
|
CC = cast<CondCodeSDNode>(CCOp.getOperand(2))->get();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool DoXform = false;
|
|
|
|
bool InvCC = false;
|
|
|
|
assert ((Opc == ISD::ADD || (Opc == ISD::SUB && Slct == N->getOperand(1))) &&
|
|
|
|
"Bad input!");
|
|
|
|
if (LHS.getOpcode() == ISD::Constant &&
|
|
|
|
cast<ConstantSDNode>(LHS)->isNullValue())
|
|
|
|
DoXform = true;
|
|
|
|
else if (CC != ISD::SETCC_INVALID &&
|
|
|
|
RHS.getOpcode() == ISD::Constant &&
|
|
|
|
cast<ConstantSDNode>(RHS)->isNullValue()) {
|
|
|
|
std::swap(LHS, RHS);
|
2008-01-17 07:20:38 +00:00
|
|
|
SDOperand Op0 = Slct.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
bool isInt = (isSlctCC ? Op0.getValueType() :
|
|
|
|
Op0.getOperand(0).getValueType()).isInteger();
|
Xforms:
(add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
(sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37685 91177308-0d34-0410-b5e6-96231b3b80d8
2007-06-21 07:39:16 +00:00
|
|
|
CC = ISD::getSetCCInverse(CC, isInt);
|
|
|
|
DoXform = true;
|
|
|
|
InvCC = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (DoXform) {
|
|
|
|
SDOperand Result = DAG.getNode(Opc, VT, OtherOp, RHS);
|
|
|
|
if (isSlctCC)
|
|
|
|
return DAG.getSelectCC(OtherOp, Result,
|
|
|
|
Slct.getOperand(0), Slct.getOperand(1), CC);
|
|
|
|
SDOperand CCOp = Slct.getOperand(0);
|
|
|
|
if (InvCC)
|
|
|
|
CCOp = DAG.getSetCC(CCOp.getValueType(), CCOp.getOperand(0),
|
|
|
|
CCOp.getOperand(1), CC);
|
|
|
|
return DAG.getNode(ISD::SELECT, VT, CCOp, OtherOp, Result);
|
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitADD(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2007-06-25 16:23:39 +00:00
|
|
|
|
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (add x, undef) -> undef
|
2007-07-10 15:19:29 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return N0;
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return N1;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (add c1, c2) -> c1+c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2008-03-13 22:13:53 +00:00
|
|
|
return DAG.getConstant(N0C->getAPIntValue() + N1C->getAPIntValue(), VT);
|
2005-09-07 23:25:52 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0C && !N1C)
|
|
|
|
return DAG.getNode(ISD::ADD, VT, N1, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (add x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2006-01-12 20:22:43 +00:00
|
|
|
// fold ((c1-A)+c2) -> (c1+c2)-A
|
|
|
|
if (N1C && N0.getOpcode() == ISD::SUB)
|
|
|
|
if (ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0.getOperand(0)))
|
|
|
|
return DAG.getNode(ISD::SUB, VT,
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue()+
|
|
|
|
N0C->getAPIntValue(), VT),
|
2006-01-12 20:22:43 +00:00
|
|
|
N0.getOperand(1));
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassociate add
|
|
|
|
SDOperand RADD = ReassociateOps(ISD::ADD, N0, N1);
|
|
|
|
if (RADD.Val != 0)
|
|
|
|
return RADD;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold ((0-A) + B) -> B-A
|
|
|
|
if (N0.getOpcode() == ISD::SUB && isa<ConstantSDNode>(N0.getOperand(0)) &&
|
|
|
|
cast<ConstantSDNode>(N0.getOperand(0))->isNullValue())
|
2005-09-07 16:09:19 +00:00
|
|
|
return DAG.getNode(ISD::SUB, VT, N1, N0.getOperand(1));
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (A + (0-B)) -> A-B
|
|
|
|
if (N1.getOpcode() == ISD::SUB && isa<ConstantSDNode>(N1.getOperand(0)) &&
|
|
|
|
cast<ConstantSDNode>(N1.getOperand(0))->isNullValue())
|
2005-09-07 16:09:19 +00:00
|
|
|
return DAG.getNode(ISD::SUB, VT, N0, N1.getOperand(1));
|
2005-09-28 22:28:18 +00:00
|
|
|
// fold (A+(B-A)) -> B
|
|
|
|
if (N1.getOpcode() == ISD::SUB && N0 == N1.getOperand(1))
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1.getOperand(0);
|
2006-03-13 06:51:27 +00:00
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector() && SimplifyDemandedBits(SDOperand(N, 0)))
|
2006-04-21 15:32:26 +00:00
|
|
|
return SDOperand(N, 0);
|
2006-03-13 06:51:27 +00:00
|
|
|
|
|
|
|
// fold (a+b) -> (a|b) iff a and b share no bits.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isInteger() && !VT.isVector()) {
|
2008-02-20 16:33:30 +00:00
|
|
|
APInt LHSZero, LHSOne;
|
|
|
|
APInt RHSZero, RHSOne;
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt Mask = APInt::getAllOnesValue(VT.getSizeInBits());
|
2007-06-22 14:59:07 +00:00
|
|
|
DAG.ComputeMaskedBits(N0, Mask, LHSZero, LHSOne);
|
2008-02-20 16:33:30 +00:00
|
|
|
if (LHSZero.getBoolValue()) {
|
2007-06-22 14:59:07 +00:00
|
|
|
DAG.ComputeMaskedBits(N1, Mask, RHSZero, RHSOne);
|
2006-03-13 06:51:27 +00:00
|
|
|
|
|
|
|
// If all possibly-set bits on the LHS are clear on the RHS, return an OR.
|
|
|
|
// If all possibly-set bits on the RHS are clear on the LHS, return an OR.
|
|
|
|
if ((RHSZero & (~LHSZero & Mask)) == (~LHSZero & Mask) ||
|
|
|
|
(LHSZero & (~RHSZero & Mask)) == (~RHSZero & Mask))
|
|
|
|
return DAG.getNode(ISD::OR, VT, N0, N1);
|
|
|
|
}
|
|
|
|
}
|
2006-11-06 08:14:30 +00:00
|
|
|
|
Remove this xform:
(shl (add x, c1), c2) -> (add (shl x, c2), c1<<c2)
Replace it with:
(add (shl (add x, c1), c2), ) -> (add (add (shl x, c2), c1<<c2), )
This fixes test/CodeGen/ARM/smul.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@33361 91177308-0d34-0410-b5e6-96231b3b80d8
2007-01-19 17:51:44 +00:00
|
|
|
// fold (add (shl (add x, c1), c2), ) -> (add (add (shl x, c2), c1<<c2), )
|
|
|
|
if (N0.getOpcode() == ISD::SHL && N0.Val->hasOneUse()) {
|
|
|
|
SDOperand Result = combineShlAddConstant(N0, N1, DAG);
|
|
|
|
if (Result.Val) return Result;
|
|
|
|
}
|
|
|
|
if (N1.getOpcode() == ISD::SHL && N1.Val->hasOneUse()) {
|
|
|
|
SDOperand Result = combineShlAddConstant(N1, N0, DAG);
|
|
|
|
if (Result.Val) return Result;
|
|
|
|
}
|
|
|
|
|
Xforms:
(add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
(sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37685 91177308-0d34-0410-b5e6-96231b3b80d8
2007-06-21 07:39:16 +00:00
|
|
|
// fold (add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
|
|
|
|
if (N0.getOpcode() == ISD::SELECT && N0.Val->hasOneUse()) {
|
|
|
|
SDOperand Result = combineSelectAndUse(N, N0, N1, DAG);
|
|
|
|
if (Result.Val) return Result;
|
|
|
|
}
|
|
|
|
if (N1.getOpcode() == ISD::SELECT && N1.Val->hasOneUse()) {
|
|
|
|
SDOperand Result = combineSelectAndUse(N, N1, N0, DAG);
|
|
|
|
if (Result.Val) return Result;
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2007-03-04 20:03:15 +00:00
|
|
|
SDOperand DAGCombiner::visitADDC(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2007-03-04 20:03:15 +00:00
|
|
|
|
|
|
|
// If the flag result is dead, turn this into an ADD.
|
|
|
|
if (N->hasNUsesOfValue(0, 1))
|
|
|
|
return CombineTo(N, DAG.getNode(ISD::ADD, VT, N1, N0),
|
2007-03-04 20:40:38 +00:00
|
|
|
DAG.getNode(ISD::CARRY_FALSE, MVT::Flag));
|
2007-03-04 20:03:15 +00:00
|
|
|
|
|
|
|
// canonicalize constant to RHS.
|
2008-06-23 15:29:14 +00:00
|
|
|
if (N0C && !N1C)
|
2008-06-21 22:06:07 +00:00
|
|
|
return DAG.getNode(ISD::ADDC, N->getVTList(), N1, N0);
|
2007-03-04 20:03:15 +00:00
|
|
|
|
2007-03-04 20:40:38 +00:00
|
|
|
// fold (addc x, 0) -> x + no carry out
|
|
|
|
if (N1C && N1C->isNullValue())
|
|
|
|
return CombineTo(N, N0, DAG.getNode(ISD::CARRY_FALSE, MVT::Flag));
|
|
|
|
|
|
|
|
// fold (addc a, b) -> (or a, b), CARRY_FALSE iff a and b share no bits.
|
2008-02-20 16:33:30 +00:00
|
|
|
APInt LHSZero, LHSOne;
|
|
|
|
APInt RHSZero, RHSOne;
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt Mask = APInt::getAllOnesValue(VT.getSizeInBits());
|
2007-06-22 14:59:07 +00:00
|
|
|
DAG.ComputeMaskedBits(N0, Mask, LHSZero, LHSOne);
|
2008-02-20 16:33:30 +00:00
|
|
|
if (LHSZero.getBoolValue()) {
|
2007-06-22 14:59:07 +00:00
|
|
|
DAG.ComputeMaskedBits(N1, Mask, RHSZero, RHSOne);
|
2007-03-04 20:40:38 +00:00
|
|
|
|
|
|
|
// If all possibly-set bits on the LHS are clear on the RHS, return an OR.
|
|
|
|
// If all possibly-set bits on the RHS are clear on the LHS, return an OR.
|
|
|
|
if ((RHSZero & (~LHSZero & Mask)) == (~LHSZero & Mask) ||
|
|
|
|
(LHSZero & (~RHSZero & Mask)) == (~RHSZero & Mask))
|
|
|
|
return CombineTo(N, DAG.getNode(ISD::OR, VT, N0, N1),
|
|
|
|
DAG.getNode(ISD::CARRY_FALSE, MVT::Flag));
|
|
|
|
}
|
2007-03-04 20:03:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitADDE(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2007-03-04 20:40:38 +00:00
|
|
|
SDOperand CarryIn = N->getOperand(2);
|
2007-03-04 20:03:15 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
//MVT VT = N0.getValueType();
|
2007-03-04 20:03:15 +00:00
|
|
|
|
|
|
|
// canonicalize constant to RHS
|
2008-06-23 15:29:14 +00:00
|
|
|
if (N0C && !N1C)
|
2008-06-21 22:06:07 +00:00
|
|
|
return DAG.getNode(ISD::ADDE, N->getVTList(), N1, N0, CarryIn);
|
2007-03-04 20:03:15 +00:00
|
|
|
|
2007-03-04 20:40:38 +00:00
|
|
|
// fold (adde x, y, false) -> (addc x, y)
|
2008-06-23 15:29:14 +00:00
|
|
|
if (CarryIn.getOpcode() == ISD::CARRY_FALSE)
|
2008-06-21 22:06:07 +00:00
|
|
|
return DAG.getNode(ISD::ADDC, N->getVTList(), N1, N0);
|
2007-03-04 20:03:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSUB(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0.Val);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.Val);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-10-17 01:07:11 +00:00
|
|
|
// fold (sub x, x) -> 0
|
2008-03-12 07:02:50 +00:00
|
|
|
if (N0 == N1)
|
2005-10-17 01:07:11 +00:00
|
|
|
return DAG.getConstant(0, N->getValueType(0));
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sub c1, c2) -> c1-c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SUB, VT, N0, N1);
|
2005-10-11 06:07:15 +00:00
|
|
|
// fold (sub x, c) -> (add x, -c)
|
|
|
|
if (N1C)
|
2008-03-13 22:13:53 +00:00
|
|
|
return DAG.getNode(ISD::ADD, VT, N0,
|
|
|
|
DAG.getConstant(-N1C->getAPIntValue(), VT));
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (A+B)-A -> B
|
2005-09-28 22:28:18 +00:00
|
|
|
if (N0.getOpcode() == ISD::ADD && N0.getOperand(0) == N1)
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0.getOperand(1);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (A+B)-B -> A
|
2005-09-28 22:28:18 +00:00
|
|
|
if (N0.getOpcode() == ISD::ADD && N0.getOperand(1) == N1)
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0.getOperand(0);
|
Xforms:
(add (select cc, 0, c), x) -> (select cc, x, (add, x, c))
(sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37685 91177308-0d34-0410-b5e6-96231b3b80d8
2007-06-21 07:39:16 +00:00
|
|
|
// fold (sub x, (select cc, 0, c)) -> (select cc, x, (sub, x, c))
|
|
|
|
if (N1.getOpcode() == ISD::SELECT && N1.Val->hasOneUse()) {
|
|
|
|
SDOperand Result = combineSelectAndUse(N, N1, N0, DAG);
|
|
|
|
if (Result.Val) return Result;
|
|
|
|
}
|
2007-07-03 14:03:57 +00:00
|
|
|
// If either operand of a sub is undef, the result is undef
|
2007-07-10 15:19:29 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return N0;
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return N1;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitMUL(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (mul x, undef) -> 0
|
2007-07-10 14:20:37 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF || N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mul c1, c2) -> c1*c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::MUL, VT, N0, N1);
|
2005-09-07 23:25:52 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0C && !N1C)
|
|
|
|
return DAG.getNode(ISD::MUL, VT, N1, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mul x, 0) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mul x, -1) -> 0-x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isAllOnesValue())
|
2005-10-21 00:02:42 +00:00
|
|
|
return DAG.getNode(ISD::SUB, VT, DAG.getConstant(0, VT), N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mul x, (1 << c)) -> x << c
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue().isPowerOf2())
|
2005-10-30 06:41:49 +00:00
|
|
|
return DAG.getNode(ISD::SHL, VT, N0,
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue().logBase2(),
|
2005-09-06 04:43:02 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2005-10-30 06:41:49 +00:00
|
|
|
// fold (mul x, -(1 << c)) -> -(x << c) or (-x) << c
|
|
|
|
if (N1C && isPowerOf2_64(-N1C->getSignExtended())) {
|
|
|
|
// FIXME: If the input is something that is easily negated (e.g. a
|
|
|
|
// single-use add), we should put the negate there.
|
|
|
|
return DAG.getNode(ISD::SUB, VT, DAG.getConstant(0, VT),
|
|
|
|
DAG.getNode(ISD::SHL, VT, N0,
|
|
|
|
DAG.getConstant(Log2_64(-N1C->getSignExtended()),
|
|
|
|
TLI.getShiftAmountTy())));
|
|
|
|
}
|
2006-04-02 21:42:45 +00:00
|
|
|
|
2006-03-01 03:44:24 +00:00
|
|
|
// (mul (shl X, c1), c2) -> (mul X, c2 << c1)
|
|
|
|
if (N1C && N0.getOpcode() == ISD::SHL &&
|
|
|
|
isa<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
SDOperand C3 = DAG.getNode(ISD::SHL, VT, N1, N0.getOperand(1));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(C3.Val);
|
2006-03-01 03:44:24 +00:00
|
|
|
return DAG.getNode(ISD::MUL, VT, N0.getOperand(0), C3);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Change (mul (shl X, C), Y) -> (shl (mul X, Y), C) when the shift has one
|
|
|
|
// use.
|
|
|
|
{
|
|
|
|
SDOperand Sh(0,0), Y(0,0);
|
|
|
|
// Check for both (mul (shl X, C), Y) and (mul Y, (shl X, C)).
|
|
|
|
if (N0.getOpcode() == ISD::SHL && isa<ConstantSDNode>(N0.getOperand(1)) &&
|
|
|
|
N0.Val->hasOneUse()) {
|
|
|
|
Sh = N0; Y = N1;
|
|
|
|
} else if (N1.getOpcode() == ISD::SHL &&
|
|
|
|
isa<ConstantSDNode>(N1.getOperand(1)) && N1.Val->hasOneUse()) {
|
|
|
|
Sh = N1; Y = N0;
|
|
|
|
}
|
|
|
|
if (Sh.Val) {
|
|
|
|
SDOperand Mul = DAG.getNode(ISD::MUL, VT, Sh.getOperand(0), Y);
|
|
|
|
return DAG.getNode(ISD::SHL, VT, Mul, Sh.getOperand(1));
|
|
|
|
}
|
|
|
|
}
|
fold (mul (add x, c1), c2) -> (add (mul x, c2), c1*c2)
fold (shl (add x, c1), c2) -> (add (shl x, c2), c1<<c2)
This allows us to compile CodeGen/PowerPC/addi-reassoc.ll into:
_test1:
slwi r2, r4, 4
add r2, r2, r3
lwz r3, 36(r2)
blr
_test2:
mulli r2, r4, 5
add r2, r2, r3
lbz r2, 11(r2)
extsb r3, r2
blr
instead of:
_test1:
addi r2, r4, 2
slwi r2, r2, 4
add r2, r3, r2
lwz r3, 4(r2)
blr
_test2:
addi r2, r4, 2
mulli r2, r2, 5
add r2, r3, r2
lbz r2, 1(r2)
extsb r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@26535 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-04 23:33:26 +00:00
|
|
|
// fold (mul (add x, c1), c2) -> (add (mul x, c2), c1*c2)
|
|
|
|
if (N1C && N0.getOpcode() == ISD::ADD && N0.Val->hasOneUse() &&
|
|
|
|
isa<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
return DAG.getNode(ISD::ADD, VT,
|
|
|
|
DAG.getNode(ISD::MUL, VT, N0.getOperand(0), N1),
|
|
|
|
DAG.getNode(ISD::MUL, VT, N0.getOperand(1), N1));
|
|
|
|
}
|
2006-03-01 03:44:24 +00:00
|
|
|
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassociate mul
|
|
|
|
SDOperand RMUL = ReassociateOps(ISD::MUL, N0, N1);
|
|
|
|
if (RMUL.Val != 0)
|
|
|
|
return RMUL;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSDIV(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0.Val);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.Val);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sdiv c1, c2) -> c1/c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C && !N1C->isNullValue())
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SDIV, VT, N0, N1);
|
2005-10-21 00:02:42 +00:00
|
|
|
// fold (sdiv X, 1) -> X
|
|
|
|
if (N1C && N1C->getSignExtended() == 1LL)
|
|
|
|
return N0;
|
|
|
|
// fold (sdiv X, -1) -> 0-X
|
|
|
|
if (N1C && N1C->isAllOnesValue())
|
|
|
|
return DAG.getNode(ISD::SUB, VT, DAG.getConstant(0, VT), N0);
|
2005-10-07 06:10:46 +00:00
|
|
|
// If we know the sign bits of both operands are zero, strength reduce to a
|
|
|
|
// udiv instead. Handles (X&15) /s 4 -> X&15 >> 2
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector()) {
|
2008-02-25 21:11:39 +00:00
|
|
|
if (DAG.SignBitIsZero(N1) && DAG.SignBitIsZero(N0))
|
2008-01-27 23:32:17 +00:00
|
|
|
return DAG.getNode(ISD::UDIV, N1.getValueType(), N0, N1);
|
|
|
|
}
|
2006-02-17 07:26:20 +00:00
|
|
|
// fold (sdiv X, pow2) -> simple ops after legalize
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && !N1C->isNullValue() && !TLI.isIntDivCheap() &&
|
2005-10-21 00:02:42 +00:00
|
|
|
(isPowerOf2_64(N1C->getSignExtended()) ||
|
|
|
|
isPowerOf2_64(-N1C->getSignExtended()))) {
|
|
|
|
// If dividing by powers of two is cheap, then don't perform the following
|
|
|
|
// fold.
|
|
|
|
if (TLI.isPow2DivCheap())
|
|
|
|
return SDOperand();
|
|
|
|
int64_t pow2 = N1C->getSignExtended();
|
|
|
|
int64_t abs2 = pow2 > 0 ? pow2 : -pow2;
|
2006-02-16 08:02:36 +00:00
|
|
|
unsigned lg2 = Log2_64(abs2);
|
|
|
|
// Splat the sign bit into the register
|
|
|
|
SDOperand SGN = DAG.getNode(ISD::SRA, VT, N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(VT.getSizeInBits()-1,
|
2005-10-21 00:02:42 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(SGN.Val);
|
2006-02-16 08:02:36 +00:00
|
|
|
// Add (N0 < 0) ? abs2 - 1 : 0;
|
|
|
|
SDOperand SRL = DAG.getNode(ISD::SRL, VT, SGN,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(VT.getSizeInBits()-lg2,
|
2005-10-21 00:02:42 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2006-02-16 08:02:36 +00:00
|
|
|
SDOperand ADD = DAG.getNode(ISD::ADD, VT, N0, SRL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(SRL.Val);
|
|
|
|
AddToWorkList(ADD.Val); // Divide by pow2
|
2006-02-16 08:02:36 +00:00
|
|
|
SDOperand SRA = DAG.getNode(ISD::SRA, VT, ADD,
|
|
|
|
DAG.getConstant(lg2, TLI.getShiftAmountTy()));
|
2005-10-21 00:02:42 +00:00
|
|
|
// If we're dividing by a positive value, we're done. Otherwise, we must
|
|
|
|
// negate the result.
|
|
|
|
if (pow2 > 0)
|
|
|
|
return SRA;
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(SRA.Val);
|
2005-10-21 00:02:42 +00:00
|
|
|
return DAG.getNode(ISD::SUB, VT, DAG.getConstant(0, VT), SRA);
|
|
|
|
}
|
2005-10-20 02:15:44 +00:00
|
|
|
// if integer divide is expensive and we satisfy the requirements, emit an
|
|
|
|
// alternate sequence.
|
2005-10-21 00:02:42 +00:00
|
|
|
if (N1C && (N1C->getSignExtended() < -1 || N1C->getSignExtended() > 1) &&
|
2005-10-22 18:50:15 +00:00
|
|
|
!TLI.isIntDivCheap()) {
|
|
|
|
SDOperand Op = BuildSDIV(N);
|
|
|
|
if (Op.Val) return Op;
|
2005-10-20 02:15:44 +00:00
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// undef / X -> 0
|
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return DAG.getConstant(0, VT);
|
|
|
|
// X / undef -> undef
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return N1;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitUDIV(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0.Val);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.Val);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (udiv c1, c2) -> c1/c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C && !N1C->isNullValue())
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::UDIV, VT, N0, N1);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (udiv x, (1 << c)) -> x >>u c
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue().isPowerOf2())
|
2006-02-05 07:20:23 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, N0,
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue().logBase2(),
|
2005-09-06 04:43:02 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2006-02-05 07:20:23 +00:00
|
|
|
// fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) iff c is power of 2
|
|
|
|
if (N1.getOpcode() == ISD::SHL) {
|
|
|
|
if (ConstantSDNode *SHC = dyn_cast<ConstantSDNode>(N1.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SHC->getAPIntValue().isPowerOf2()) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT ADDVT = N1.getOperand(1).getValueType();
|
2006-02-05 07:36:48 +00:00
|
|
|
SDOperand Add = DAG.getNode(ISD::ADD, ADDVT, N1.getOperand(1),
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(SHC->getAPIntValue()
|
|
|
|
.logBase2(),
|
2006-02-05 07:36:48 +00:00
|
|
|
ADDVT));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Add.Val);
|
2006-02-05 07:36:48 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, N0, Add);
|
2006-02-05 07:20:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2005-10-20 02:15:44 +00:00
|
|
|
// fold (udiv x, c) -> alternate
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && !N1C->isNullValue() && !TLI.isIntDivCheap()) {
|
2005-10-22 18:50:15 +00:00
|
|
|
SDOperand Op = BuildUDIV(N);
|
|
|
|
if (Op.Val) return Op;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// undef / X -> 0
|
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return DAG.getConstant(0, VT);
|
|
|
|
// X / undef -> undef
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return N1;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSREM(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (srem c1, c2) -> c1%c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C && !N1C->isNullValue())
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SREM, VT, N0, N1);
|
2005-10-10 21:26:48 +00:00
|
|
|
// If we know the sign bits of both operands are zero, strength reduce to a
|
|
|
|
// urem instead. Handles (X & 0x0FFFFFFF) %s 16 -> X&15
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector()) {
|
2008-02-25 21:11:39 +00:00
|
|
|
if (DAG.SignBitIsZero(N1) && DAG.SignBitIsZero(N0))
|
2008-01-27 23:21:58 +00:00
|
|
|
return DAG.getNode(ISD::UREM, VT, N0, N1);
|
|
|
|
}
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
|
2007-11-26 23:46:11 +00:00
|
|
|
// If X/C can be simplified by the division-by-constant logic, lower
|
|
|
|
// X%C to the equivalent of X-X/C*C.
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
if (N1C && !N1C->isNullValue()) {
|
|
|
|
SDOperand Div = DAG.getNode(ISD::SDIV, VT, N0, N1);
|
2008-01-26 01:09:19 +00:00
|
|
|
AddToWorkList(Div.Val);
|
2007-11-26 23:46:11 +00:00
|
|
|
SDOperand OptimizedDiv = combine(Div.Val);
|
|
|
|
if (OptimizedDiv.Val && OptimizedDiv.Val != Div.Val) {
|
|
|
|
SDOperand Mul = DAG.getNode(ISD::MUL, VT, OptimizedDiv, N1);
|
|
|
|
SDOperand Sub = DAG.getNode(ISD::SUB, VT, N0, Mul);
|
|
|
|
AddToWorkList(Mul.Val);
|
|
|
|
return Sub;
|
|
|
|
}
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
}
|
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// undef % X -> 0
|
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return DAG.getConstant(0, VT);
|
|
|
|
// X % undef -> undef
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return N1;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitUREM(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (urem c1, c2) -> c1%c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C && !N1C->isNullValue())
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::UREM, VT, N0, N1);
|
2005-10-10 21:26:48 +00:00
|
|
|
// fold (urem x, pow2) -> (and x, pow2-1)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && !N1C->isNullValue() && N1C->getAPIntValue().isPowerOf2())
|
|
|
|
return DAG.getNode(ISD::AND, VT, N0,
|
|
|
|
DAG.getConstant(N1C->getAPIntValue()-1,VT));
|
2006-02-05 07:36:48 +00:00
|
|
|
// fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1))
|
|
|
|
if (N1.getOpcode() == ISD::SHL) {
|
|
|
|
if (ConstantSDNode *SHC = dyn_cast<ConstantSDNode>(N1.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SHC->getAPIntValue().isPowerOf2()) {
|
|
|
|
SDOperand Add =
|
|
|
|
DAG.getNode(ISD::ADD, VT, N1,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(APInt::getAllOnesValue(VT.getSizeInBits()),
|
2008-03-13 22:13:53 +00:00
|
|
|
VT));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Add.Val);
|
2006-02-05 07:36:48 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, N0, Add);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
|
2007-11-26 23:46:11 +00:00
|
|
|
// If X/C can be simplified by the division-by-constant logic, lower
|
|
|
|
// X%C to the equivalent of X-X/C*C.
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
if (N1C && !N1C->isNullValue()) {
|
|
|
|
SDOperand Div = DAG.getNode(ISD::UDIV, VT, N0, N1);
|
2007-11-26 23:46:11 +00:00
|
|
|
SDOperand OptimizedDiv = combine(Div.Val);
|
|
|
|
if (OptimizedDiv.Val && OptimizedDiv.Val != Div.Val) {
|
|
|
|
SDOperand Mul = DAG.getNode(ISD::MUL, VT, OptimizedDiv, N1);
|
|
|
|
SDOperand Sub = DAG.getNode(ISD::SUB, VT, N0, Mul);
|
|
|
|
AddToWorkList(Mul.Val);
|
|
|
|
return Sub;
|
|
|
|
}
|
Lower X%C into X/C+stuff. This allows the 'division by a constant' logic to
apply to rems as well as divs. This fixes PR945 and speeds up ReedSolomon
from 14.57s to 10.90s (which is now faster than gcc).
It compiles CodeGen/X86/rem.ll into:
_test1:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
imull %ecx
addl %esi, %edx
movl %edx, %eax
shrl $31, %eax
sarl $7, %edx
addl %eax, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
_test2:
movl 4(%esp), %eax
movl %eax, %ecx
sarl $31, %ecx
shrl $24, %ecx
addl %eax, %ecx
andl $4294967040, %ecx
subl %ecx, %eax
ret
_test3:
subl $4, %esp
movl %esi, (%esp)
movl $2155905153, %ecx
movl 8(%esp), %esi
movl %esi, %eax
mull %ecx
shrl $7, %edx
imull $255, %edx, %eax
subl %eax, %esi
movl %esi, %eax
movl (%esp), %esi
addl $4, %esp
ret
instead of div/idiv instructions.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30920 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-12 20:58:32 +00:00
|
|
|
}
|
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// undef % X -> 0
|
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return DAG.getConstant(0, VT);
|
|
|
|
// X % undef -> undef
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return N1;
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitMULHS(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (mulhs x, 0) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mulhs x, 1) -> (sra x, size(x)-1)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1)
|
2005-09-01 00:19:25 +00:00
|
|
|
return DAG.getNode(ISD::SRA, N0.getValueType(), N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(N0.getValueType().getSizeInBits()-1,
|
2005-09-06 04:43:02 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (mulhs x, undef) -> 0
|
2007-07-10 14:20:37 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF || N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitMULHU(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (mulhu x, 0) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (mulhu x, 1) -> 0
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, N0.getValueType());
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (mulhu x, undef) -> 0
|
2007-07-10 14:20:37 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF || N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2007-10-08 17:57:15 +00:00
|
|
|
/// SimplifyNodeWithTwoResults - Perform optimizations common to nodes that
|
|
|
|
/// compute two values. LoOp and HiOp give the opcodes for the two computations
|
|
|
|
/// that are being performed. Return true if a simplification was made.
|
|
|
|
///
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand DAGCombiner::SimplifyNodeWithTwoResults(SDNode *N, unsigned LoOp,
|
|
|
|
unsigned HiOp) {
|
2007-10-08 17:57:15 +00:00
|
|
|
// If the high half is not needed, just compute the low half.
|
2007-11-08 09:25:29 +00:00
|
|
|
bool HiExists = N->hasAnyUseOfValue(1);
|
|
|
|
if (!HiExists &&
|
2007-10-08 17:57:15 +00:00
|
|
|
(!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(LoOp, N->getValueType(0)))) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = DAG.getNode(LoOp, N->getValueType(0), N->op_begin(),
|
|
|
|
N->getNumOperands());
|
|
|
|
return CombineTo(N, Res, Res);
|
2007-10-08 17:57:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// If the low half is not needed, just compute the high half.
|
2007-11-08 09:25:29 +00:00
|
|
|
bool LoExists = N->hasAnyUseOfValue(0);
|
|
|
|
if (!LoExists &&
|
2007-10-08 17:57:15 +00:00
|
|
|
(!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(HiOp, N->getValueType(1)))) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = DAG.getNode(HiOp, N->getValueType(1), N->op_begin(),
|
|
|
|
N->getNumOperands());
|
|
|
|
return CombineTo(N, Res, Res);
|
2007-10-08 17:57:15 +00:00
|
|
|
}
|
|
|
|
|
2007-11-08 09:25:29 +00:00
|
|
|
// If both halves are used, return as it is.
|
|
|
|
if (LoExists && HiExists)
|
2008-01-26 01:09:19 +00:00
|
|
|
return SDOperand();
|
2007-11-08 09:25:29 +00:00
|
|
|
|
|
|
|
// If the two computed results can be simplified separately, separate them.
|
|
|
|
if (LoExists) {
|
|
|
|
SDOperand Lo = DAG.getNode(LoOp, N->getValueType(0),
|
|
|
|
N->op_begin(), N->getNumOperands());
|
2008-01-26 01:09:19 +00:00
|
|
|
AddToWorkList(Lo.Val);
|
2007-11-08 09:25:29 +00:00
|
|
|
SDOperand LoOpt = combine(Lo.Val);
|
2008-01-26 01:09:19 +00:00
|
|
|
if (LoOpt.Val && LoOpt.Val != Lo.Val &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
(!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(LoOpt.getOpcode(), LoOpt.getValueType())))
|
2008-01-26 01:09:19 +00:00
|
|
|
return CombineTo(N, LoOpt, LoOpt);
|
2007-10-08 17:57:15 +00:00
|
|
|
}
|
2007-11-08 09:25:29 +00:00
|
|
|
|
|
|
|
if (HiExists) {
|
|
|
|
SDOperand Hi = DAG.getNode(HiOp, N->getValueType(1),
|
|
|
|
N->op_begin(), N->getNumOperands());
|
2008-01-26 01:09:19 +00:00
|
|
|
AddToWorkList(Hi.Val);
|
2007-11-08 09:25:29 +00:00
|
|
|
SDOperand HiOpt = combine(Hi.Val);
|
|
|
|
if (HiOpt.Val && HiOpt != Hi &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
(!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(HiOpt.getOpcode(), HiOpt.getValueType())))
|
2008-01-26 01:09:19 +00:00
|
|
|
return CombineTo(N, HiOpt, HiOpt);
|
2007-10-08 17:57:15 +00:00
|
|
|
}
|
2008-01-26 01:09:19 +00:00
|
|
|
return SDOperand();
|
2007-10-08 17:57:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitSMUL_LOHI(SDNode *N) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = SimplifyNodeWithTwoResults(N, ISD::MUL, ISD::MULHS);
|
|
|
|
if (Res.Val) return Res;
|
2007-10-08 17:57:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitUMUL_LOHI(SDNode *N) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = SimplifyNodeWithTwoResults(N, ISD::MUL, ISD::MULHU);
|
|
|
|
if (Res.Val) return Res;
|
2007-10-08 17:57:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitSDIVREM(SDNode *N) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = SimplifyNodeWithTwoResults(N, ISD::SDIV, ISD::SREM);
|
|
|
|
if (Res.Val) return Res;
|
2007-10-08 17:57:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitUDIVREM(SDNode *N) {
|
2008-01-26 01:09:19 +00:00
|
|
|
SDOperand Res = SimplifyNodeWithTwoResults(N, ISD::UDIV, ISD::UREM);
|
|
|
|
if (Res.Val) return Res;
|
2007-10-08 17:57:15 +00:00
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-05-05 05:51:50 +00:00
|
|
|
/// SimplifyBinOpWithSameOpcodeHands - If this is a binary operator with
|
|
|
|
/// two operands of the same opcode, try to simplify it.
|
|
|
|
SDOperand DAGCombiner::SimplifyBinOpWithSameOpcodeHands(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0), N1 = N->getOperand(1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2006-05-05 05:51:50 +00:00
|
|
|
assert(N0.getOpcode() == N1.getOpcode() && "Bad input!");
|
|
|
|
|
Implement:
// fold (and (sext x), (sext y)) -> (sext (and x, y))
// fold (or (sext x), (sext y)) -> (sext (or x, y))
// fold (xor (sext x), (sext y)) -> (sext (xor x, y))
// fold (and (aext x), (aext y)) -> (aext (and x, y))
// fold (or (aext x), (aext y)) -> (aext (or x, y))
// fold (xor (aext x), (aext y)) -> (aext (xor x, y))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28123 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-05 06:31:05 +00:00
|
|
|
// For each of OP in AND/OR/XOR:
|
|
|
|
// fold (OP (zext x), (zext y)) -> (zext (OP x, y))
|
|
|
|
// fold (OP (sext x), (sext y)) -> (sext (OP x, y))
|
|
|
|
// fold (OP (aext x), (aext y)) -> (aext (OP x, y))
|
2006-05-05 06:32:04 +00:00
|
|
|
// fold (OP (trunc x), (trunc y)) -> (trunc (OP x, y))
|
Implement:
// fold (and (sext x), (sext y)) -> (sext (and x, y))
// fold (or (sext x), (sext y)) -> (sext (or x, y))
// fold (xor (sext x), (sext y)) -> (sext (xor x, y))
// fold (and (aext x), (aext y)) -> (aext (and x, y))
// fold (or (aext x), (aext y)) -> (aext (or x, y))
// fold (xor (aext x), (aext y)) -> (aext (xor x, y))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28123 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-05 06:31:05 +00:00
|
|
|
if ((N0.getOpcode() == ISD::ZERO_EXTEND || N0.getOpcode() == ISD::ANY_EXTEND||
|
2006-05-05 06:32:04 +00:00
|
|
|
N0.getOpcode() == ISD::SIGN_EXTEND || N0.getOpcode() == ISD::TRUNCATE) &&
|
2006-05-05 05:51:50 +00:00
|
|
|
N0.getOperand(0).getValueType() == N1.getOperand(0).getValueType()) {
|
|
|
|
SDOperand ORNode = DAG.getNode(N->getOpcode(),
|
|
|
|
N0.getOperand(0).getValueType(),
|
|
|
|
N0.getOperand(0), N1.getOperand(0));
|
|
|
|
AddToWorkList(ORNode.Val);
|
Implement:
// fold (and (sext x), (sext y)) -> (sext (and x, y))
// fold (or (sext x), (sext y)) -> (sext (or x, y))
// fold (xor (sext x), (sext y)) -> (sext (xor x, y))
// fold (and (aext x), (aext y)) -> (aext (and x, y))
// fold (or (aext x), (aext y)) -> (aext (or x, y))
// fold (xor (aext x), (aext y)) -> (aext (xor x, y))
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28123 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-05 06:31:05 +00:00
|
|
|
return DAG.getNode(N0.getOpcode(), VT, ORNode);
|
2006-05-05 05:51:50 +00:00
|
|
|
}
|
|
|
|
|
Pull and through and/or/xor. This compiles some bitfield code to:
mov EAX, DWORD PTR [ESP + 4]
mov ECX, DWORD PTR [EAX]
mov EDX, ECX
add EDX, EDX
or EDX, ECX
and EDX, -2147483648
and ECX, 2147483647
or EDX, ECX
mov DWORD PTR [EAX], EDX
ret
instead of:
sub ESP, 4
mov DWORD PTR [ESP], ESI
mov EAX, DWORD PTR [ESP + 8]
mov ECX, DWORD PTR [EAX]
mov EDX, ECX
add EDX, EDX
mov ESI, ECX
and ESI, -2147483648
and EDX, -2147483648
or EDX, ESI
and ECX, 2147483647
or EDX, ECX
mov DWORD PTR [EAX], EDX
mov ESI, DWORD PTR [ESP]
add ESP, 4
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28122 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-05 06:10:43 +00:00
|
|
|
// For each of OP in SHL/SRL/SRA/AND...
|
|
|
|
// fold (and (OP x, z), (OP y, z)) -> (OP (and x, y), z)
|
|
|
|
// fold (or (OP x, z), (OP y, z)) -> (OP (or x, y), z)
|
|
|
|
// fold (xor (OP x, z), (OP y, z)) -> (OP (xor x, y), z)
|
2006-05-05 05:51:50 +00:00
|
|
|
if ((N0.getOpcode() == ISD::SHL || N0.getOpcode() == ISD::SRL ||
|
Pull and through and/or/xor. This compiles some bitfield code to:
mov EAX, DWORD PTR [ESP + 4]
mov ECX, DWORD PTR [EAX]
mov EDX, ECX
add EDX, EDX
or EDX, ECX
and EDX, -2147483648
and ECX, 2147483647
or EDX, ECX
mov DWORD PTR [EAX], EDX
ret
instead of:
sub ESP, 4
mov DWORD PTR [ESP], ESI
mov EAX, DWORD PTR [ESP + 8]
mov ECX, DWORD PTR [EAX]
mov EDX, ECX
add EDX, EDX
mov ESI, ECX
and ESI, -2147483648
and EDX, -2147483648
or EDX, ESI
and ECX, 2147483647
or EDX, ECX
mov DWORD PTR [EAX], EDX
mov ESI, DWORD PTR [ESP]
add ESP, 4
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28122 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-05 06:10:43 +00:00
|
|
|
N0.getOpcode() == ISD::SRA || N0.getOpcode() == ISD::AND) &&
|
2006-05-05 05:51:50 +00:00
|
|
|
N0.getOperand(1) == N1.getOperand(1)) {
|
|
|
|
SDOperand ORNode = DAG.getNode(N->getOpcode(),
|
|
|
|
N0.getOperand(0).getValueType(),
|
|
|
|
N0.getOperand(0), N1.getOperand(0));
|
|
|
|
AddToWorkList(ORNode.Val);
|
|
|
|
return DAG.getNode(N0.getOpcode(), VT, ORNode, N0.getOperand(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitAND(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2006-02-17 19:54:08 +00:00
|
|
|
SDOperand LL, LR, RL, RR, CC0, CC1;
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N1.getValueType();
|
|
|
|
unsigned BitWidth = VT.getSizeInBits();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (and x, undef) -> 0
|
2007-07-10 14:20:37 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF || N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (and c1, c2) -> c1&c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, N0, N1);
|
2005-09-07 23:25:52 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0C && !N1C)
|
|
|
|
return DAG.getNode(ISD::AND, VT, N1, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (and x, -1) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isAllOnesValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
|
|
|
// if (and x, c) is known to be zero, return 0
|
2008-02-25 21:11:39 +00:00
|
|
|
if (N1C && DAG.MaskedValueIsZero(SDOperand(N, 0),
|
|
|
|
APInt::getAllOnesValue(BitWidth)))
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassociate and
|
|
|
|
SDOperand RAND = ReassociateOps(ISD::AND, N0, N1);
|
|
|
|
if (RAND.Val != 0)
|
|
|
|
return RAND;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (and (or x, 0xFFFF), 0xFF) -> 0xFF
|
2005-11-02 18:42:59 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::OR)
|
2005-09-01 00:19:25 +00:00
|
|
|
if (ConstantSDNode *ORI = dyn_cast<ConstantSDNode>(N0.getOperand(1)))
|
2008-03-13 22:13:53 +00:00
|
|
|
if ((ORI->getAPIntValue() & N1C->getAPIntValue()) == N1C->getAPIntValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
Turn any_extend nodes into zero_extend nodes when it allows us to remove an
and instruction. This allows us to compile stuff like this:
bool %X(int %X) {
%Y = add int %X, 14
%Z = setne int %Y, 12345
ret bool %Z
}
to this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
ret
instead of this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
andl $1, %eax
ret
This occurs quite a bit with the X86 backend. For example, 25 times in
lambda, 30 times in 177.mesa, 14 times in galgel, 70 times in fma3d,
25 times in vpr, several hundred times in gcc, ~45 times in crafty,
~60 times in parser, ~140 times in eon, 110 times in perlbmk, 55 on gap,
16 times on bzip2, 14 times on twolf, and 1-2 times in many other SPEC2K
programs.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@25901 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-02 07:17:31 +00:00
|
|
|
// fold (and (any_ext V), c) -> (zero_ext V) if 'and' only clears top bits.
|
|
|
|
if (N1C && N0.getOpcode() == ISD::ANY_EXTEND) {
|
2008-02-25 21:11:39 +00:00
|
|
|
SDOperand N0Op0 = N0.getOperand(0);
|
|
|
|
APInt Mask = ~N1C->getAPIntValue();
|
|
|
|
Mask.trunc(N0Op0.getValueSizeInBits());
|
|
|
|
if (DAG.MaskedValueIsZero(N0Op0, Mask)) {
|
2006-03-01 21:47:21 +00:00
|
|
|
SDOperand Zext = DAG.getNode(ISD::ZERO_EXTEND, N0.getValueType(),
|
2008-02-25 21:11:39 +00:00
|
|
|
N0Op0);
|
2006-03-01 21:47:21 +00:00
|
|
|
|
|
|
|
// Replace uses of the AND with uses of the Zero extend node.
|
|
|
|
CombineTo(N, Zext);
|
|
|
|
|
Turn any_extend nodes into zero_extend nodes when it allows us to remove an
and instruction. This allows us to compile stuff like this:
bool %X(int %X) {
%Y = add int %X, 14
%Z = setne int %Y, 12345
ret bool %Z
}
to this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
ret
instead of this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
andl $1, %eax
ret
This occurs quite a bit with the X86 backend. For example, 25 times in
lambda, 30 times in 177.mesa, 14 times in galgel, 70 times in fma3d,
25 times in vpr, several hundred times in gcc, ~45 times in crafty,
~60 times in parser, ~140 times in eon, 110 times in perlbmk, 55 on gap,
16 times on bzip2, 14 times on twolf, and 1-2 times in many other SPEC2K
programs.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@25901 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-02 07:17:31 +00:00
|
|
|
// We actually want to replace all uses of the any_extend with the
|
|
|
|
// zero_extend, to avoid duplicating things. This will later cause this
|
|
|
|
// AND to be folded.
|
2006-03-01 21:47:21 +00:00
|
|
|
CombineTo(N0.Val, Zext);
|
2006-04-20 23:55:59 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
Turn any_extend nodes into zero_extend nodes when it allows us to remove an
and instruction. This allows us to compile stuff like this:
bool %X(int %X) {
%Y = add int %X, 14
%Z = setne int %Y, 12345
ret bool %Z
}
to this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
ret
instead of this:
_X:
cmpl $12331, 4(%esp)
setne %al
movzbl %al, %eax
andl $1, %eax
ret
This occurs quite a bit with the X86 backend. For example, 25 times in
lambda, 30 times in 177.mesa, 14 times in galgel, 70 times in fma3d,
25 times in vpr, several hundred times in gcc, ~45 times in crafty,
~60 times in parser, ~140 times in eon, 110 times in perlbmk, 55 on gap,
16 times on bzip2, 14 times on twolf, and 1-2 times in many other SPEC2K
programs.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@25901 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-02 07:17:31 +00:00
|
|
|
}
|
|
|
|
}
|
2005-09-09 19:49:52 +00:00
|
|
|
// fold (and (setcc x), (setcc y)) -> (setcc (and x, y))
|
|
|
|
if (isSetCCEquivalent(N0, LL, LR, CC0) && isSetCCEquivalent(N1, RL, RR, CC1)){
|
|
|
|
ISD::CondCode Op0 = cast<CondCodeSDNode>(CC0)->get();
|
|
|
|
ISD::CondCode Op1 = cast<CondCodeSDNode>(CC1)->get();
|
|
|
|
|
|
|
|
if (LR == RR && isa<ConstantSDNode>(LR) && Op0 == Op1 &&
|
2008-06-06 12:08:01 +00:00
|
|
|
LL.getValueType().isInteger()) {
|
2005-09-09 19:49:52 +00:00
|
|
|
// fold (X == 0) & (Y == 0) -> (X|Y == 0)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (cast<ConstantSDNode>(LR)->isNullValue() && Op1 == ISD::SETEQ) {
|
2005-09-09 19:49:52 +00:00
|
|
|
SDOperand ORNode = DAG.getNode(ISD::OR, LR.getValueType(), LL, RL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(ORNode.Val);
|
2005-09-09 19:49:52 +00:00
|
|
|
return DAG.getSetCC(VT, ORNode, LR, Op1);
|
|
|
|
}
|
|
|
|
// fold (X == -1) & (Y == -1) -> (X&Y == -1)
|
|
|
|
if (cast<ConstantSDNode>(LR)->isAllOnesValue() && Op1 == ISD::SETEQ) {
|
|
|
|
SDOperand ANDNode = DAG.getNode(ISD::AND, LR.getValueType(), LL, RL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(ANDNode.Val);
|
2005-09-09 19:49:52 +00:00
|
|
|
return DAG.getSetCC(VT, ANDNode, LR, Op1);
|
|
|
|
}
|
|
|
|
// fold (X > -1) & (Y > -1) -> (X|Y > -1)
|
|
|
|
if (cast<ConstantSDNode>(LR)->isAllOnesValue() && Op1 == ISD::SETGT) {
|
|
|
|
SDOperand ORNode = DAG.getNode(ISD::OR, LR.getValueType(), LL, RL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(ORNode.Val);
|
2005-09-09 19:49:52 +00:00
|
|
|
return DAG.getSetCC(VT, ORNode, LR, Op1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// canonicalize equivalent to ll == rl
|
|
|
|
if (LL == RR && LR == RL) {
|
|
|
|
Op1 = ISD::getSetCCSwappedOperands(Op1);
|
|
|
|
std::swap(RL, RR);
|
|
|
|
}
|
|
|
|
if (LL == RL && LR == RR) {
|
2008-06-06 12:08:01 +00:00
|
|
|
bool isInteger = LL.getValueType().isInteger();
|
2005-09-09 19:49:52 +00:00
|
|
|
ISD::CondCode Result = ISD::getSetCCAndOperation(Op0, Op1, isInteger);
|
|
|
|
if (Result != ISD::SETCC_INVALID)
|
|
|
|
return DAG.getSetCC(N0.getValueType(), LL, LR, Result);
|
|
|
|
}
|
|
|
|
}
|
2006-05-05 05:51:50 +00:00
|
|
|
|
|
|
|
// Simplify: and (op x...), (op y...) -> (op (and x, y))
|
|
|
|
if (N0.getOpcode() == N1.getOpcode()) {
|
|
|
|
SDOperand Tmp = SimplifyBinOpWithSameOpcodeHands(N);
|
|
|
|
if (Tmp.Val) return Tmp;
|
2005-09-16 00:54:12 +00:00
|
|
|
}
|
2006-05-05 05:51:50 +00:00
|
|
|
|
2006-02-03 22:24:05 +00:00
|
|
|
// fold (and (sign_extend_inreg x, i16 to i32), 1) -> (and x, 1)
|
|
|
|
// fold (and (sra)) -> (and (srl)) when possible.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector() &&
|
2006-03-25 22:19:00 +00:00
|
|
|
SimplifyDemandedBits(SDOperand(N, 0)))
|
2006-04-21 15:32:26 +00:00
|
|
|
return SDOperand(N, 0);
|
2005-10-13 03:11:28 +00:00
|
|
|
// fold (zext_inreg (extload x)) -> (zextload x)
|
2007-03-07 08:07:03 +00:00
|
|
|
if (ISD::isEXTLoad(N0.Val) && ISD::isUNINDEXEDLoad(N0.Val)) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = LN0->getMemoryVT();
|
2005-10-13 18:34:58 +00:00
|
|
|
// If we zero all the possible extended bits, then we can turn this into
|
|
|
|
// a zextload if we are running before legalize or the operation is legal.
|
2008-02-25 21:11:39 +00:00
|
|
|
unsigned BitWidth = N1.getValueSizeInBits();
|
|
|
|
if (DAG.MaskedValueIsZero(N1, APInt::getHighBitsSet(BitWidth,
|
2008-06-06 12:08:01 +00:00
|
|
|
BitWidth - EVT.getSizeInBits())) &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !LN0->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::ZEXTLOAD, EVT))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::ZEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(N);
|
2005-10-13 18:16:34 +00:00
|
|
|
CombineTo(N0.Val, ExtLoad, ExtLoad.getValue(1));
|
2006-04-20 23:55:59 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
// fold (zext_inreg (sextload x)) -> (zextload x) iff load has one use
|
2007-03-07 08:07:03 +00:00
|
|
|
if (ISD::isSEXTLoad(N0.Val) && ISD::isUNINDEXEDLoad(N0.Val) &&
|
|
|
|
N0.hasOneUse()) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = LN0->getMemoryVT();
|
2005-10-13 18:34:58 +00:00
|
|
|
// If we zero all the possible extended bits, then we can turn this into
|
|
|
|
// a zextload if we are running before legalize or the operation is legal.
|
2008-02-25 21:11:39 +00:00
|
|
|
unsigned BitWidth = N1.getValueSizeInBits();
|
|
|
|
if (DAG.MaskedValueIsZero(N1, APInt::getHighBitsSet(BitWidth,
|
2008-06-06 12:08:01 +00:00
|
|
|
BitWidth - EVT.getSizeInBits())) &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !LN0->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::ZEXTLOAD, EVT))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::ZEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(N);
|
2005-10-13 18:16:34 +00:00
|
|
|
CombineTo(N0.Val, ExtLoad, ExtLoad.getValue(1));
|
2006-04-20 23:55:59 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
|
|
|
}
|
Fold "and (LOAD P), 255" -> zextload. This allows us to compile:
unsigned foo3(unsigned *P) { return *P & 255; }
as:
_foo3:
lbz r3, 3(r3)
blr
instead of:
_foo3:
lwz r2, 0(r3)
rlwinm r3, r2, 0, 24, 31
blr
and:
unsigned short foo2(float a) { return a; }
as:
_foo2:
fctiwz f0, f1
stfd f0, -8(r1)
lhz r3, -2(r1)
blr
instead of:
_foo2:
fctiwz f0, f1
stfd f0, -8(r1)
lwz r2, -4(r1)
rlwinm r3, r2, 0, 16, 31
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@26417 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-28 06:35:35 +00:00
|
|
|
|
Compile:
unsigned foo4(unsigned short *P) { return *P & 255; }
unsigned foo5(short *P) { return *P & 255; }
to:
_foo4:
lbz r3,1(r3)
blr
_foo5:
lbz r3,1(r3)
blr
not:
_foo4:
lhz r2, 0(r3)
rlwinm r3, r2, 0, 24, 31
blr
_foo5:
lhz r2, 0(r3)
rlwinm r3, r2, 0, 24, 31
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@26419 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-28 06:49:37 +00:00
|
|
|
// fold (and (load x), 255) -> (zextload x, i8)
|
|
|
|
// fold (and (extload x, i16), 255) -> (zextload x, i8)
|
2006-10-09 20:57:25 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::LOAD) {
|
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
if (LN0->getExtensionType() != ISD::SEXTLOAD &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
LN0->isUnindexed() && N0.hasOneUse() &&
|
|
|
|
// Do not change the width of a volatile load.
|
|
|
|
!LN0->isVolatile()) {
|
2008-06-09 11:32:28 +00:00
|
|
|
MVT EVT = MVT::Other;
|
|
|
|
uint32_t ActiveBits = N1C->getAPIntValue().getActiveBits();
|
|
|
|
if (ActiveBits > 0 && APIntOps::isMask(ActiveBits, N1C->getAPIntValue()))
|
|
|
|
EVT = MVT::getIntegerVT(ActiveBits);
|
|
|
|
|
|
|
|
MVT LoadedVT = LN0->getMemoryVT();
|
2008-06-16 08:14:38 +00:00
|
|
|
// Do not generate loads of non-round integer types since these can
|
|
|
|
// be expensive (and would be wrong if the type is not byte sized).
|
|
|
|
if (EVT != MVT::Other && LoadedVT.bitsGT(EVT) && EVT.isRound() &&
|
2006-10-09 20:57:25 +00:00
|
|
|
(!AfterLegalize || TLI.isLoadXLegal(ISD::ZEXTLOAD, EVT))) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT PtrType = N0.getOperand(1).getValueType();
|
2006-10-09 20:57:25 +00:00
|
|
|
// For big endian targets, we need to add an offset to the pointer to
|
|
|
|
// load the correct bytes. For little endian systems, we merely need to
|
|
|
|
// read fewer bytes from the same pointer.
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned LVTStoreBytes = LoadedVT.getStoreSizeInBits()/8;
|
|
|
|
unsigned EVTStoreBytes = EVT.getStoreSizeInBits()/8;
|
2007-11-09 08:57:19 +00:00
|
|
|
unsigned PtrOff = LVTStoreBytes - EVTStoreBytes;
|
2007-10-28 12:59:45 +00:00
|
|
|
unsigned Alignment = LN0->getAlignment();
|
2006-10-09 20:57:25 +00:00
|
|
|
SDOperand NewPtr = LN0->getBasePtr();
|
2008-02-11 10:37:04 +00:00
|
|
|
if (TLI.isBigEndian()) {
|
2006-10-09 20:57:25 +00:00
|
|
|
NewPtr = DAG.getNode(ISD::ADD, PtrType, NewPtr,
|
|
|
|
DAG.getConstant(PtrOff, PtrType));
|
2007-10-28 12:59:45 +00:00
|
|
|
Alignment = MinAlign(Alignment, PtrOff);
|
|
|
|
}
|
2006-10-09 20:57:25 +00:00
|
|
|
AddToWorkList(NewPtr.Val);
|
|
|
|
SDOperand Load =
|
|
|
|
DAG.getExtLoad(ISD::ZEXTLOAD, VT, LN0->getChain(), NewPtr,
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValue(), LN0->getSrcValueOffset(), EVT,
|
2007-10-28 12:59:45 +00:00
|
|
|
LN0->isVolatile(), Alignment);
|
2006-10-09 20:57:25 +00:00
|
|
|
AddToWorkList(N);
|
|
|
|
CombineTo(N0.Val, Load, Load.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
Fold "and (LOAD P), 255" -> zextload. This allows us to compile:
unsigned foo3(unsigned *P) { return *P & 255; }
as:
_foo3:
lbz r3, 3(r3)
blr
instead of:
_foo3:
lwz r2, 0(r3)
rlwinm r3, r2, 0, 24, 31
blr
and:
unsigned short foo2(float a) { return a; }
as:
_foo2:
fctiwz f0, f1
stfd f0, -8(r1)
lhz r3, -2(r1)
blr
instead of:
_foo2:
fctiwz f0, f1
stfd f0, -8(r1)
lwz r2, -4(r1)
rlwinm r3, r2, 0, 16, 31
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@26417 91177308-0d34-0410-b5e6-96231b3b80d8
2006-02-28 06:35:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitOR(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-09 19:49:52 +00:00
|
|
|
SDOperand LL, LR, RL, RR, CC0, CC1;
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N1.getValueType();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (or x, undef) -> -1
|
2007-07-10 14:20:37 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF || N1.getOpcode() == ISD::UNDEF)
|
2007-07-09 16:16:34 +00:00
|
|
|
return DAG.getConstant(~0ULL, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (or c1, c2) -> c1|c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::OR, VT, N0, N1);
|
2005-09-07 23:25:52 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0C && !N1C)
|
|
|
|
return DAG.getNode(ISD::OR, VT, N1, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (or x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (or x, -1) -> -1
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isAllOnesValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
|
|
|
// fold (or x, c) -> c iff (x & ~c) == 0
|
2008-02-25 21:11:39 +00:00
|
|
|
if (N1C && DAG.MaskedValueIsZero(N0, ~N1C->getAPIntValue()))
|
2005-09-06 04:43:02 +00:00
|
|
|
return N1;
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassociate or
|
|
|
|
SDOperand ROR = ReassociateOps(ISD::OR, N0, N1);
|
|
|
|
if (ROR.Val != 0)
|
|
|
|
return ROR;
|
|
|
|
// Canonicalize (or (and X, c1), c2) -> (and (or X, c2), c1|c2)
|
|
|
|
if (N1C && N0.getOpcode() == ISD::AND && N0.Val->hasOneUse() &&
|
2005-10-27 05:06:38 +00:00
|
|
|
isa<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
ConstantSDNode *C1 = cast<ConstantSDNode>(N0.getOperand(1));
|
|
|
|
return DAG.getNode(ISD::AND, VT, DAG.getNode(ISD::OR, VT, N0.getOperand(0),
|
|
|
|
N1),
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue() |
|
|
|
|
C1->getAPIntValue(), VT));
|
2005-09-08 20:18:10 +00:00
|
|
|
}
|
2005-09-09 19:49:52 +00:00
|
|
|
// fold (or (setcc x), (setcc y)) -> (setcc (or x, y))
|
|
|
|
if (isSetCCEquivalent(N0, LL, LR, CC0) && isSetCCEquivalent(N1, RL, RR, CC1)){
|
|
|
|
ISD::CondCode Op0 = cast<CondCodeSDNode>(CC0)->get();
|
|
|
|
ISD::CondCode Op1 = cast<CondCodeSDNode>(CC1)->get();
|
|
|
|
|
|
|
|
if (LR == RR && isa<ConstantSDNode>(LR) && Op0 == Op1 &&
|
2008-06-06 12:08:01 +00:00
|
|
|
LL.getValueType().isInteger()) {
|
2005-09-09 19:49:52 +00:00
|
|
|
// fold (X != 0) | (Y != 0) -> (X|Y != 0)
|
|
|
|
// fold (X < 0) | (Y < 0) -> (X|Y < 0)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (cast<ConstantSDNode>(LR)->isNullValue() &&
|
2005-09-09 19:49:52 +00:00
|
|
|
(Op1 == ISD::SETNE || Op1 == ISD::SETLT)) {
|
|
|
|
SDOperand ORNode = DAG.getNode(ISD::OR, LR.getValueType(), LL, RL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(ORNode.Val);
|
2005-09-09 19:49:52 +00:00
|
|
|
return DAG.getSetCC(VT, ORNode, LR, Op1);
|
|
|
|
}
|
|
|
|
// fold (X != -1) | (Y != -1) -> (X&Y != -1)
|
|
|
|
// fold (X > -1) | (Y > -1) -> (X&Y > -1)
|
|
|
|
if (cast<ConstantSDNode>(LR)->isAllOnesValue() &&
|
|
|
|
(Op1 == ISD::SETNE || Op1 == ISD::SETGT)) {
|
|
|
|
SDOperand ANDNode = DAG.getNode(ISD::AND, LR.getValueType(), LL, RL);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(ANDNode.Val);
|
2005-09-09 19:49:52 +00:00
|
|
|
return DAG.getSetCC(VT, ANDNode, LR, Op1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// canonicalize equivalent to ll == rl
|
|
|
|
if (LL == RR && LR == RL) {
|
|
|
|
Op1 = ISD::getSetCCSwappedOperands(Op1);
|
|
|
|
std::swap(RL, RR);
|
|
|
|
}
|
|
|
|
if (LL == RL && LR == RR) {
|
2008-06-06 12:08:01 +00:00
|
|
|
bool isInteger = LL.getValueType().isInteger();
|
2005-09-09 19:49:52 +00:00
|
|
|
ISD::CondCode Result = ISD::getSetCCOrOperation(Op0, Op1, isInteger);
|
|
|
|
if (Result != ISD::SETCC_INVALID)
|
|
|
|
return DAG.getSetCC(N0.getValueType(), LL, LR, Result);
|
|
|
|
}
|
|
|
|
}
|
2006-05-05 05:51:50 +00:00
|
|
|
|
|
|
|
// Simplify: or (op x...), (op y...) -> (op (or x, y))
|
|
|
|
if (N0.getOpcode() == N1.getOpcode()) {
|
|
|
|
SDOperand Tmp = SimplifyBinOpWithSameOpcodeHands(N);
|
|
|
|
if (Tmp.Val) return Tmp;
|
2006-02-01 07:19:44 +00:00
|
|
|
}
|
2006-09-14 20:50:57 +00:00
|
|
|
|
2006-09-14 21:11:37 +00:00
|
|
|
// (X & C1) | (Y & C2) -> (X|Y) & C3 if possible.
|
|
|
|
if (N0.getOpcode() == ISD::AND &&
|
|
|
|
N1.getOpcode() == ISD::AND &&
|
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant &&
|
|
|
|
N1.getOperand(1).getOpcode() == ISD::Constant &&
|
|
|
|
// Don't increase # computations.
|
|
|
|
(N0.Val->hasOneUse() || N1.Val->hasOneUse())) {
|
|
|
|
// We can only do this xform if we know that bits from X that are set in C2
|
|
|
|
// but not in C1 are already zero. Likewise for Y.
|
2008-02-25 21:11:39 +00:00
|
|
|
const APInt &LHSMask =
|
|
|
|
cast<ConstantSDNode>(N0.getOperand(1))->getAPIntValue();
|
|
|
|
const APInt &RHSMask =
|
|
|
|
cast<ConstantSDNode>(N1.getOperand(1))->getAPIntValue();
|
2006-09-14 21:11:37 +00:00
|
|
|
|
2007-06-22 14:59:07 +00:00
|
|
|
if (DAG.MaskedValueIsZero(N0.getOperand(0), RHSMask&~LHSMask) &&
|
|
|
|
DAG.MaskedValueIsZero(N1.getOperand(0), LHSMask&~RHSMask)) {
|
2006-09-14 21:11:37 +00:00
|
|
|
SDOperand X =DAG.getNode(ISD::OR, VT, N0.getOperand(0), N1.getOperand(0));
|
|
|
|
return DAG.getNode(ISD::AND, VT, X, DAG.getConstant(LHSMask|RHSMask, VT));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2006-09-14 20:50:57 +00:00
|
|
|
// See if this is some rotate idiom.
|
|
|
|
if (SDNode *Rot = MatchRotate(N0, N1))
|
|
|
|
return SDOperand(Rot, 0);
|
2006-05-05 05:51:50 +00:00
|
|
|
|
2006-09-14 20:50:57 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/// MatchRotateHalf - Match "(X shl/srl V1) & V2" where V2 may not be present.
|
|
|
|
static bool MatchRotateHalf(SDOperand Op, SDOperand &Shift, SDOperand &Mask) {
|
|
|
|
if (Op.getOpcode() == ISD::AND) {
|
2006-11-02 20:25:50 +00:00
|
|
|
if (isa<ConstantSDNode>(Op.getOperand(1))) {
|
2006-09-14 20:50:57 +00:00
|
|
|
Mask = Op.getOperand(1);
|
|
|
|
Op = Op.getOperand(0);
|
|
|
|
} else {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (Op.getOpcode() == ISD::SRL || Op.getOpcode() == ISD::SHL) {
|
|
|
|
Shift = Op;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// MatchRotate - Handle an 'or' of two operands. If this is one of the many
|
|
|
|
// idioms for rotate, and if the target supports rotation instructions, generate
|
|
|
|
// a rot[lr].
|
|
|
|
SDNode *DAGCombiner::MatchRotate(SDOperand LHS, SDOperand RHS) {
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// Must be a legal type. Expanded 'n promoted things won't work with rotates.
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = LHS.getValueType();
|
2006-09-14 20:50:57 +00:00
|
|
|
if (!TLI.isTypeLegal(VT)) return 0;
|
|
|
|
|
|
|
|
// The target must have at least one rotate flavor.
|
|
|
|
bool HasROTL = TLI.isOperationLegal(ISD::ROTL, VT);
|
|
|
|
bool HasROTR = TLI.isOperationLegal(ISD::ROTR, VT);
|
|
|
|
if (!HasROTL && !HasROTR) return 0;
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2006-09-14 20:50:57 +00:00
|
|
|
// Match "(X shl/srl V1) & V2" where V2 may not be present.
|
|
|
|
SDOperand LHSShift; // The shift.
|
|
|
|
SDOperand LHSMask; // AND value if any.
|
|
|
|
if (!MatchRotateHalf(LHS, LHSShift, LHSMask))
|
|
|
|
return 0; // Not part of a rotate.
|
|
|
|
|
|
|
|
SDOperand RHSShift; // The shift.
|
|
|
|
SDOperand RHSMask; // AND value if any.
|
|
|
|
if (!MatchRotateHalf(RHS, RHSShift, RHSMask))
|
|
|
|
return 0; // Not part of a rotate.
|
|
|
|
|
|
|
|
if (LHSShift.getOperand(0) != RHSShift.getOperand(0))
|
|
|
|
return 0; // Not shifting the same value.
|
|
|
|
|
|
|
|
if (LHSShift.getOpcode() == RHSShift.getOpcode())
|
|
|
|
return 0; // Shifts must disagree.
|
|
|
|
|
|
|
|
// Canonicalize shl to left side in a shl/srl pair.
|
|
|
|
if (RHSShift.getOpcode() == ISD::SHL) {
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
std::swap(LHSShift, RHSShift);
|
|
|
|
std::swap(LHSMask , RHSMask );
|
|
|
|
}
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned OpSizeInBits = VT.getSizeInBits();
|
2007-04-02 21:36:32 +00:00
|
|
|
SDOperand LHSShiftArg = LHSShift.getOperand(0);
|
|
|
|
SDOperand LHSShiftAmt = LHSShift.getOperand(1);
|
|
|
|
SDOperand RHSShiftAmt = RHSShift.getOperand(1);
|
2006-09-14 20:50:57 +00:00
|
|
|
|
|
|
|
// fold (or (shl x, C1), (srl x, C2)) -> (rotl x, C1)
|
|
|
|
// fold (or (shl x, C1), (srl x, C2)) -> (rotr x, C2)
|
2007-04-02 21:36:32 +00:00
|
|
|
if (LHSShiftAmt.getOpcode() == ISD::Constant &&
|
|
|
|
RHSShiftAmt.getOpcode() == ISD::Constant) {
|
|
|
|
uint64_t LShVal = cast<ConstantSDNode>(LHSShiftAmt)->getValue();
|
|
|
|
uint64_t RShVal = cast<ConstantSDNode>(RHSShiftAmt)->getValue();
|
2006-09-14 20:50:57 +00:00
|
|
|
if ((LShVal + RShVal) != OpSizeInBits)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
SDOperand Rot;
|
|
|
|
if (HasROTL)
|
2007-04-02 21:36:32 +00:00
|
|
|
Rot = DAG.getNode(ISD::ROTL, VT, LHSShiftArg, LHSShiftAmt);
|
2006-09-14 20:50:57 +00:00
|
|
|
else
|
2007-04-02 21:36:32 +00:00
|
|
|
Rot = DAG.getNode(ISD::ROTR, VT, LHSShiftArg, RHSShiftAmt);
|
2006-09-14 20:50:57 +00:00
|
|
|
|
|
|
|
// If there is an AND of either shifted operand, apply it to the result.
|
|
|
|
if (LHSMask.Val || RHSMask.Val) {
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt Mask = APInt::getAllOnesValue(OpSizeInBits);
|
2006-09-14 20:50:57 +00:00
|
|
|
|
|
|
|
if (LHSMask.Val) {
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt RHSBits = APInt::getLowBitsSet(OpSizeInBits, LShVal);
|
|
|
|
Mask &= cast<ConstantSDNode>(LHSMask)->getAPIntValue() | RHSBits;
|
2006-08-31 07:41:12 +00:00
|
|
|
}
|
2006-09-14 20:50:57 +00:00
|
|
|
if (RHSMask.Val) {
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt LHSBits = APInt::getHighBitsSet(OpSizeInBits, RShVal);
|
|
|
|
Mask &= cast<ConstantSDNode>(RHSMask)->getAPIntValue() | LHSBits;
|
2006-09-14 20:50:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
Rot = DAG.getNode(ISD::AND, VT, Rot, DAG.getConstant(Mask, VT));
|
|
|
|
}
|
|
|
|
|
|
|
|
return Rot.Val;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If there is a mask here, and we have a variable shift, we can't be sure
|
|
|
|
// that we're masking out the right stuff.
|
|
|
|
if (LHSMask.Val || RHSMask.Val)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
// fold (or (shl x, y), (srl x, (sub 32, y))) -> (rotl x, y)
|
|
|
|
// fold (or (shl x, y), (srl x, (sub 32, y))) -> (rotr x, (sub 32, y))
|
2007-04-02 21:36:32 +00:00
|
|
|
if (RHSShiftAmt.getOpcode() == ISD::SUB &&
|
|
|
|
LHSShiftAmt == RHSShiftAmt.getOperand(1)) {
|
2006-09-14 20:50:57 +00:00
|
|
|
if (ConstantSDNode *SUBC =
|
2007-04-02 21:36:32 +00:00
|
|
|
dyn_cast<ConstantSDNode>(RHSShiftAmt.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SUBC->getAPIntValue() == OpSizeInBits) {
|
2006-09-14 20:50:57 +00:00
|
|
|
if (HasROTL)
|
2007-04-02 21:36:32 +00:00
|
|
|
return DAG.getNode(ISD::ROTL, VT, LHSShiftArg, LHSShiftAmt).Val;
|
2006-09-14 20:50:57 +00:00
|
|
|
else
|
2007-04-02 21:36:32 +00:00
|
|
|
return DAG.getNode(ISD::ROTR, VT, LHSShiftArg, RHSShiftAmt).Val;
|
2008-02-20 11:10:28 +00:00
|
|
|
}
|
2006-08-31 07:41:12 +00:00
|
|
|
}
|
2006-01-11 21:21:00 +00:00
|
|
|
}
|
2006-09-14 20:50:57 +00:00
|
|
|
|
|
|
|
// fold (or (shl x, (sub 32, y)), (srl x, r)) -> (rotr x, y)
|
|
|
|
// fold (or (shl x, (sub 32, y)), (srl x, r)) -> (rotl x, (sub 32, y))
|
2007-04-02 21:36:32 +00:00
|
|
|
if (LHSShiftAmt.getOpcode() == ISD::SUB &&
|
|
|
|
RHSShiftAmt == LHSShiftAmt.getOperand(1)) {
|
2006-09-14 20:50:57 +00:00
|
|
|
if (ConstantSDNode *SUBC =
|
2007-04-02 21:36:32 +00:00
|
|
|
dyn_cast<ConstantSDNode>(LHSShiftAmt.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SUBC->getAPIntValue() == OpSizeInBits) {
|
2006-09-14 20:50:57 +00:00
|
|
|
if (HasROTL)
|
2007-04-02 21:36:32 +00:00
|
|
|
return DAG.getNode(ISD::ROTL, VT, LHSShiftArg, LHSShiftAmt).Val;
|
2006-09-14 20:50:57 +00:00
|
|
|
else
|
2007-04-02 21:36:32 +00:00
|
|
|
return DAG.getNode(ISD::ROTR, VT, LHSShiftArg, RHSShiftAmt).Val;
|
2008-02-20 11:10:28 +00:00
|
|
|
}
|
2007-04-02 21:36:32 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Look for sign/zext/any-extended cases:
|
|
|
|
if ((LHSShiftAmt.getOpcode() == ISD::SIGN_EXTEND
|
|
|
|
|| LHSShiftAmt.getOpcode() == ISD::ZERO_EXTEND
|
|
|
|
|| LHSShiftAmt.getOpcode() == ISD::ANY_EXTEND) &&
|
|
|
|
(RHSShiftAmt.getOpcode() == ISD::SIGN_EXTEND
|
|
|
|
|| RHSShiftAmt.getOpcode() == ISD::ZERO_EXTEND
|
|
|
|
|| RHSShiftAmt.getOpcode() == ISD::ANY_EXTEND)) {
|
|
|
|
SDOperand LExtOp0 = LHSShiftAmt.getOperand(0);
|
|
|
|
SDOperand RExtOp0 = RHSShiftAmt.getOperand(0);
|
|
|
|
if (RExtOp0.getOpcode() == ISD::SUB &&
|
|
|
|
RExtOp0.getOperand(1) == LExtOp0) {
|
|
|
|
// fold (or (shl x, (*ext y)), (srl x, (*ext (sub 32, y)))) ->
|
|
|
|
// (rotr x, y)
|
|
|
|
// fold (or (shl x, (*ext y)), (srl x, (*ext (sub 32, y)))) ->
|
|
|
|
// (rotl x, (sub 32, y))
|
|
|
|
if (ConstantSDNode *SUBC = cast<ConstantSDNode>(RExtOp0.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SUBC->getAPIntValue() == OpSizeInBits) {
|
2007-04-02 21:36:32 +00:00
|
|
|
if (HasROTL)
|
|
|
|
return DAG.getNode(ISD::ROTL, VT, LHSShiftArg, LHSShiftAmt).Val;
|
|
|
|
else
|
|
|
|
return DAG.getNode(ISD::ROTR, VT, LHSShiftArg, RHSShiftAmt).Val;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else if (LExtOp0.getOpcode() == ISD::SUB &&
|
|
|
|
RExtOp0 == LExtOp0.getOperand(1)) {
|
|
|
|
// fold (or (shl x, (*ext (sub 32, y))), (srl x, (*ext r))) ->
|
|
|
|
// (rotl x, y)
|
|
|
|
// fold (or (shl x, (*ext (sub 32, y))), (srl x, (*ext r))) ->
|
|
|
|
// (rotr x, (sub 32, y))
|
|
|
|
if (ConstantSDNode *SUBC = cast<ConstantSDNode>(LExtOp0.getOperand(0))) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SUBC->getAPIntValue() == OpSizeInBits) {
|
2007-04-02 21:36:32 +00:00
|
|
|
if (HasROTL)
|
|
|
|
return DAG.getNode(ISD::ROTL, VT, LHSShiftArg, RHSShiftAmt).Val;
|
|
|
|
else
|
|
|
|
return DAG.getNode(ISD::ROTL, VT, LHSShiftArg, LHSShiftAmt).Val;
|
|
|
|
}
|
|
|
|
}
|
2006-09-14 20:50:57 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2006-09-14 20:50:57 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitXOR(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
SDOperand LHS, RHS, CC;
|
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2008-03-25 20:08:07 +00:00
|
|
|
// fold (xor undef, undef) -> 0. This is a common idiom (misuse).
|
|
|
|
if (N0.getOpcode() == ISD::UNDEF && N1.getOpcode() == ISD::UNDEF)
|
|
|
|
return DAG.getConstant(0, VT);
|
2007-07-03 14:03:57 +00:00
|
|
|
// fold (xor x, undef) -> undef
|
2007-07-10 15:19:29 +00:00
|
|
|
if (N0.getOpcode() == ISD::UNDEF)
|
|
|
|
return N0;
|
|
|
|
if (N1.getOpcode() == ISD::UNDEF)
|
2007-07-03 14:03:57 +00:00
|
|
|
return N1;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (xor c1, c2) -> c1^c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::XOR, VT, N0, N1);
|
2005-09-07 23:25:52 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0C && !N1C)
|
|
|
|
return DAG.getNode(ISD::XOR, VT, N1, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (xor x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2006-02-03 06:46:56 +00:00
|
|
|
// reassociate xor
|
|
|
|
SDOperand RXOR = ReassociateOps(ISD::XOR, N0, N1);
|
|
|
|
if (RXOR.Val != 0)
|
|
|
|
return RXOR;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold !(x cc y) -> (x !cc y)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1 && isSetCCEquivalent(N0, LHS, RHS, CC)) {
|
2008-06-06 12:08:01 +00:00
|
|
|
bool isInt = LHS.getValueType().isInteger();
|
2005-09-02 21:18:40 +00:00
|
|
|
ISD::CondCode NotCC = ISD::getSetCCInverse(cast<CondCodeSDNode>(CC)->get(),
|
|
|
|
isInt);
|
|
|
|
if (N0.getOpcode() == ISD::SETCC)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getSetCC(VT, LHS, RHS, NotCC);
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0.getOpcode() == ISD::SELECT_CC)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getSelectCC(LHS, RHS, N0.getOperand(2),N0.getOperand(3),NotCC);
|
2005-09-02 21:18:40 +00:00
|
|
|
assert(0 && "Unhandled SetCC Equivalent!");
|
|
|
|
abort();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2007-09-10 21:39:07 +00:00
|
|
|
// fold (not (zext (setcc x, y))) -> (zext (not (setcc x, y)))
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1 && N0.getOpcode() == ISD::ZERO_EXTEND &&
|
2007-09-10 21:39:07 +00:00
|
|
|
N0.Val->hasOneUse() && isSetCCEquivalent(N0.getOperand(0), LHS, RHS, CC)){
|
|
|
|
SDOperand V = N0.getOperand(0);
|
|
|
|
V = DAG.getNode(ISD::XOR, V.getValueType(), V,
|
2007-10-10 09:54:50 +00:00
|
|
|
DAG.getConstant(1, V.getValueType()));
|
2007-09-10 21:39:07 +00:00
|
|
|
AddToWorkList(V.Val);
|
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, V);
|
|
|
|
}
|
|
|
|
|
2005-09-07 23:25:52 +00:00
|
|
|
// fold !(x or y) -> (!x and !y) iff x or y are setcc
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1 && VT == MVT::i1 &&
|
2005-09-07 23:25:52 +00:00
|
|
|
(N0.getOpcode() == ISD::OR || N0.getOpcode() == ISD::AND)) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand LHS = N0.getOperand(0), RHS = N0.getOperand(1);
|
2005-09-07 23:25:52 +00:00
|
|
|
if (isOneUseSetCC(RHS) || isOneUseSetCC(LHS)) {
|
|
|
|
unsigned NewOpcode = N0.getOpcode() == ISD::AND ? ISD::OR : ISD::AND;
|
2005-09-01 00:19:25 +00:00
|
|
|
LHS = DAG.getNode(ISD::XOR, VT, LHS, N1); // RHS = ~LHS
|
|
|
|
RHS = DAG.getNode(ISD::XOR, VT, RHS, N1); // RHS = ~RHS
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(LHS.Val); AddToWorkList(RHS.Val);
|
2005-09-07 23:25:52 +00:00
|
|
|
return DAG.getNode(NewOpcode, VT, LHS, RHS);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
}
|
2005-09-07 23:25:52 +00:00
|
|
|
// fold !(x or y) -> (!x and !y) iff x or y are constants
|
|
|
|
if (N1C && N1C->isAllOnesValue() &&
|
|
|
|
(N0.getOpcode() == ISD::OR || N0.getOpcode() == ISD::AND)) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand LHS = N0.getOperand(0), RHS = N0.getOperand(1);
|
2005-09-07 23:25:52 +00:00
|
|
|
if (isa<ConstantSDNode>(RHS) || isa<ConstantSDNode>(LHS)) {
|
|
|
|
unsigned NewOpcode = N0.getOpcode() == ISD::AND ? ISD::OR : ISD::AND;
|
2005-09-01 00:19:25 +00:00
|
|
|
LHS = DAG.getNode(ISD::XOR, VT, LHS, N1); // RHS = ~LHS
|
|
|
|
RHS = DAG.getNode(ISD::XOR, VT, RHS, N1); // RHS = ~RHS
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(LHS.Val); AddToWorkList(RHS.Val);
|
2005-09-07 23:25:52 +00:00
|
|
|
return DAG.getNode(NewOpcode, VT, LHS, RHS);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
}
|
2005-09-08 20:18:10 +00:00
|
|
|
// fold (xor (xor x, c1), c2) -> (xor x, c1^c2)
|
|
|
|
if (N1C && N0.getOpcode() == ISD::XOR) {
|
|
|
|
ConstantSDNode *N00C = dyn_cast<ConstantSDNode>(N0.getOperand(0));
|
|
|
|
ConstantSDNode *N01C = dyn_cast<ConstantSDNode>(N0.getOperand(1));
|
|
|
|
if (N00C)
|
|
|
|
return DAG.getNode(ISD::XOR, VT, N0.getOperand(1),
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue()^
|
|
|
|
N00C->getAPIntValue(), VT));
|
2005-09-08 20:18:10 +00:00
|
|
|
if (N01C)
|
|
|
|
return DAG.getNode(ISD::XOR, VT, N0.getOperand(0),
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N1C->getAPIntValue()^
|
|
|
|
N01C->getAPIntValue(), VT));
|
2005-09-08 20:18:10 +00:00
|
|
|
}
|
|
|
|
// fold (xor x, x) -> 0
|
2006-03-28 19:11:05 +00:00
|
|
|
if (N0 == N1) {
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector()) {
|
2006-03-28 19:11:05 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
|
|
|
} else if (!AfterLegalize || TLI.isOperationLegal(ISD::BUILD_VECTOR, VT)) {
|
|
|
|
// Produce a vector of zeros.
|
2008-06-06 12:08:01 +00:00
|
|
|
SDOperand El = DAG.getConstant(0, VT.getVectorElementType());
|
|
|
|
std::vector<SDOperand> Ops(VT.getVectorNumElements(), El);
|
2006-08-08 02:23:42 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, VT, &Ops[0], Ops.size());
|
2006-03-28 19:11:05 +00:00
|
|
|
}
|
|
|
|
}
|
2006-05-05 05:51:50 +00:00
|
|
|
|
|
|
|
// Simplify: xor (op x...), (op y...) -> (op (xor x, y))
|
|
|
|
if (N0.getOpcode() == N1.getOpcode()) {
|
|
|
|
SDOperand Tmp = SimplifyBinOpWithSameOpcodeHands(N);
|
|
|
|
if (Tmp.Val) return Tmp;
|
2006-02-01 07:19:44 +00:00
|
|
|
}
|
2006-05-05 05:51:50 +00:00
|
|
|
|
2006-04-08 04:15:24 +00:00
|
|
|
// Simplify the expression using non-local knowledge.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (!VT.isVector() &&
|
2006-04-08 04:15:24 +00:00
|
|
|
SimplifyDemandedBits(SDOperand(N, 0)))
|
2006-04-21 15:32:26 +00:00
|
|
|
return SDOperand(N, 0);
|
2006-04-08 04:15:24 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
/// visitShiftByConstant - Handle transforms common to the three shifts, when
|
|
|
|
/// the shift amount is a constant.
|
|
|
|
SDOperand DAGCombiner::visitShiftByConstant(SDNode *N, unsigned Amt) {
|
|
|
|
SDNode *LHS = N->getOperand(0).Val;
|
|
|
|
if (!LHS->hasOneUse()) return SDOperand();
|
|
|
|
|
|
|
|
// We want to pull some binops through shifts, so that we have (and (shift))
|
|
|
|
// instead of (shift (and)), likewise for add, or, xor, etc. This sort of
|
|
|
|
// thing happens with address calculations, so it's important to canonicalize
|
|
|
|
// it.
|
|
|
|
bool HighBitSet = false; // Can we transform this if the high bit is set?
|
|
|
|
|
|
|
|
switch (LHS->getOpcode()) {
|
|
|
|
default: return SDOperand();
|
|
|
|
case ISD::OR:
|
|
|
|
case ISD::XOR:
|
|
|
|
HighBitSet = false; // We can only transform sra if the high bit is clear.
|
|
|
|
break;
|
|
|
|
case ISD::AND:
|
|
|
|
HighBitSet = true; // We can only transform sra if the high bit is set.
|
|
|
|
break;
|
|
|
|
case ISD::ADD:
|
|
|
|
if (N->getOpcode() != ISD::SHL)
|
|
|
|
return SDOperand(); // only shl(add) not sr[al](add).
|
|
|
|
HighBitSet = false; // We can only transform sra if the high bit is clear.
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// We require the RHS of the binop to be a constant as well.
|
|
|
|
ConstantSDNode *BinOpCst = dyn_cast<ConstantSDNode>(LHS->getOperand(1));
|
|
|
|
if (!BinOpCst) return SDOperand();
|
|
|
|
|
2007-12-06 07:47:55 +00:00
|
|
|
|
|
|
|
// FIXME: disable this for unless the input to the binop is a shift by a
|
|
|
|
// constant. If it is not a shift, it pessimizes some common cases like:
|
|
|
|
//
|
|
|
|
//void foo(int *X, int i) { X[i & 1235] = 1; }
|
|
|
|
//int bar(int *X, int i) { return X[i & 255]; }
|
|
|
|
SDNode *BinOpLHSVal = LHS->getOperand(0).Val;
|
|
|
|
if ((BinOpLHSVal->getOpcode() != ISD::SHL &&
|
|
|
|
BinOpLHSVal->getOpcode() != ISD::SRA &&
|
|
|
|
BinOpLHSVal->getOpcode() != ISD::SRL) ||
|
|
|
|
!isa<ConstantSDNode>(BinOpLHSVal->getOperand(1)))
|
|
|
|
return SDOperand();
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
|
|
|
|
// If this is a signed shift right, and the high bit is modified
|
|
|
|
// by the logical operation, do not perform the transformation.
|
|
|
|
// The highBitSet boolean indicates the value of the high bit of
|
|
|
|
// the constant which would cause it to be modified for this
|
|
|
|
// operation.
|
|
|
|
if (N->getOpcode() == ISD::SRA) {
|
2008-03-03 23:51:38 +00:00
|
|
|
bool BinOpRHSSignSet = BinOpCst->getAPIntValue().isNegative();
|
|
|
|
if (BinOpRHSSignSet != HighBitSet)
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Fold the constants, shifting the binop RHS by the shift amount.
|
|
|
|
SDOperand NewRHS = DAG.getNode(N->getOpcode(), N->getValueType(0),
|
|
|
|
LHS->getOperand(1), N->getOperand(1));
|
|
|
|
|
|
|
|
// Create the new shift.
|
|
|
|
SDOperand NewShift = DAG.getNode(N->getOpcode(), VT, LHS->getOperand(0),
|
|
|
|
N->getOperand(1));
|
|
|
|
|
|
|
|
// Create the new binop.
|
|
|
|
return DAG.getNode(LHS->getOpcode(), VT, NewShift, NewRHS);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSHL(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
|
|
|
unsigned OpSizeInBits = VT.getSizeInBits();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (shl c1, c2) -> c1<<c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SHL, VT, N0, N1);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (shl 0, x) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N0C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (shl x, c >= size(x)) -> undef
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->getValue() >= OpSizeInBits)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::UNDEF, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (shl x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// if (shl x, c) is known to be zero, return 0
|
2008-02-25 21:11:39 +00:00
|
|
|
if (DAG.MaskedValueIsZero(SDOperand(N, 0),
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt::getAllOnesValue(VT.getSizeInBits())))
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2007-04-18 03:06:49 +00:00
|
|
|
if (N1C && SimplifyDemandedBits(SDOperand(N, 0)))
|
2006-04-21 15:32:26 +00:00
|
|
|
return SDOperand(N, 0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (shl (shl x, c1), c2) -> 0 or (shl x, c1+c2)
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::SHL &&
|
2005-09-01 00:19:25 +00:00
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant) {
|
|
|
|
uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getValue();
|
2005-09-02 21:18:40 +00:00
|
|
|
uint64_t c2 = N1C->getValue();
|
2005-09-01 00:19:25 +00:00
|
|
|
if (c1 + c2 > OpSizeInBits)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
return DAG.getNode(ISD::SHL, VT, N0.getOperand(0),
|
2005-09-06 04:43:02 +00:00
|
|
|
DAG.getConstant(c1 + c2, N1.getValueType()));
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
// fold (shl (srl x, c1), c2) -> (shl (and x, -1 << c1), c2-c1) or
|
|
|
|
// (srl (and x, -1 << c1), c1-c2)
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::SRL &&
|
2005-09-01 00:19:25 +00:00
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant) {
|
|
|
|
uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getValue();
|
2005-09-02 21:18:40 +00:00
|
|
|
uint64_t c2 = N1C->getValue();
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand Mask = DAG.getNode(ISD::AND, VT, N0.getOperand(0),
|
|
|
|
DAG.getConstant(~0ULL << c1, VT));
|
|
|
|
if (c2 > c1)
|
|
|
|
return DAG.getNode(ISD::SHL, VT, Mask,
|
2005-09-06 04:43:02 +00:00
|
|
|
DAG.getConstant(c2-c1, N1.getValueType()));
|
2005-09-01 00:19:25 +00:00
|
|
|
else
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, Mask,
|
|
|
|
DAG.getConstant(c1-c2, N1.getValueType()));
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
// fold (shl (sra x, c1), c1) -> (and x, -1 << c1)
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::SRA && N1 == N0.getOperand(1))
|
2005-09-01 23:24:04 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, N0.getOperand(0),
|
2005-09-06 04:43:02 +00:00
|
|
|
DAG.getConstant(~0ULL << N1C->getValue(), VT));
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
|
|
|
|
return N1C ? visitShiftByConstant(N, N1C->getValue()) : SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSRA(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (sra c1, c2) -> c1>>c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SRA, VT, N0, N1);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sra 0, x) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N0C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sra -1, x) -> -1
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N0C->isAllOnesValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sra x, c >= size(x)) -> undef
|
2008-06-06 12:08:01 +00:00
|
|
|
if (N1C && N1C->getValue() >= VT.getSizeInBits())
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::UNDEF, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sra x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2006-02-17 19:54:08 +00:00
|
|
|
// fold (sra (shl x, c1), c1) -> sext_inreg for some c1 and target supports
|
|
|
|
// sext_inreg.
|
|
|
|
if (N1C && N0.getOpcode() == ISD::SHL && N1 == N0.getOperand(1)) {
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned LowBits = VT.getSizeInBits() - (unsigned)N1C->getValue();
|
2008-06-09 11:32:28 +00:00
|
|
|
MVT EVT = MVT::getIntegerVT(LowBits);
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if (EVT.isSimple() && // TODO: remove when apint codegen support lands.
|
|
|
|
(!AfterLegalize || TLI.isOperationLegal(ISD::SIGN_EXTEND_INREG, EVT)))
|
2006-02-17 19:54:08 +00:00
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND_INREG, VT, N0.getOperand(0),
|
|
|
|
DAG.getValueType(EVT));
|
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2006-02-28 06:23:04 +00:00
|
|
|
// fold (sra (sra x, c1), c2) -> (sra x, c1+c2)
|
|
|
|
if (N1C && N0.getOpcode() == ISD::SRA) {
|
|
|
|
if (ConstantSDNode *C1 = dyn_cast<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
unsigned Sum = N1C->getValue() + C1->getValue();
|
2008-06-06 12:08:01 +00:00
|
|
|
if (Sum >= VT.getSizeInBits()) Sum = VT.getSizeInBits()-1;
|
2006-02-28 06:23:04 +00:00
|
|
|
return DAG.getNode(ISD::SRA, VT, N0.getOperand(0),
|
|
|
|
DAG.getConstant(Sum, N1C->getValueType(0)));
|
|
|
|
}
|
|
|
|
}
|
2008-03-19 08:30:06 +00:00
|
|
|
|
|
|
|
// fold sra (shl X, m), result_size - n
|
|
|
|
// -> (sign_extend (trunc (shl X, result_size - n - m))) for
|
2008-03-20 04:31:39 +00:00
|
|
|
// result_size - n != m.
|
|
|
|
// If truncate is free for the target sext(shl) is likely to result in better
|
|
|
|
// code.
|
2008-03-19 08:30:06 +00:00
|
|
|
if (N0.getOpcode() == ISD::SHL) {
|
|
|
|
// Get the two constanst of the shifts, CN0 = m, CN = n.
|
|
|
|
const ConstantSDNode *N01C = dyn_cast<ConstantSDNode>(N0.getOperand(1));
|
|
|
|
if (N01C && N1C) {
|
2008-03-20 04:31:39 +00:00
|
|
|
// Determine what the truncate's result bitsize and type would be.
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned VTValSize = VT.getSizeInBits();
|
|
|
|
MVT TruncVT =
|
|
|
|
MVT::getIntegerVT(VTValSize - N1C->getValue());
|
2008-03-20 04:31:39 +00:00
|
|
|
// Determine the residual right-shift amount.
|
2008-03-19 08:30:06 +00:00
|
|
|
unsigned ShiftAmt = N1C->getValue() - N01C->getValue();
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2008-03-20 04:31:39 +00:00
|
|
|
// If the shift is not a no-op (in which case this should be just a sign
|
|
|
|
// extend already), the truncated to type is legal, sign_extend is legal
|
|
|
|
// on that type, and the the truncate to that type is both legal and free,
|
|
|
|
// perform the transform.
|
|
|
|
if (ShiftAmt &&
|
|
|
|
TLI.isOperationLegal(ISD::SIGN_EXTEND, TruncVT) &&
|
|
|
|
TLI.isOperationLegal(ISD::TRUNCATE, VT) &&
|
2008-03-20 02:18:41 +00:00
|
|
|
TLI.isTruncateFree(VT, TruncVT)) {
|
2008-03-20 04:31:39 +00:00
|
|
|
|
|
|
|
SDOperand Amt = DAG.getConstant(ShiftAmt, TLI.getShiftAmountTy());
|
|
|
|
SDOperand Shift = DAG.getNode(ISD::SRL, VT, N0.getOperand(0), Amt);
|
|
|
|
SDOperand Trunc = DAG.getNode(ISD::TRUNCATE, TruncVT, Shift);
|
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND, N->getValueType(0), Trunc);
|
2008-03-19 08:30:06 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2006-02-28 06:23:04 +00:00
|
|
|
|
2006-05-08 20:51:54 +00:00
|
|
|
// Simplify, based on bits shifted out of the LHS.
|
|
|
|
if (N1C && SimplifyDemandedBits(SDOperand(N, 0)))
|
|
|
|
return SDOperand(N, 0);
|
|
|
|
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// If the sign bit is known to be zero, switch this to a SRL.
|
2008-02-25 21:11:39 +00:00
|
|
|
if (DAG.SignBitIsZero(N0))
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, N0, N1);
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
|
|
|
|
return N1C ? visitShiftByConstant(N, N1C->getValue()) : SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSRL(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N0.getValueType();
|
|
|
|
unsigned OpSizeInBits = VT.getSizeInBits();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (srl c1, c2) -> c1 >>u c2
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N1C)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, N0, N1);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (srl 0, x) -> 0
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0C && N0C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (srl x, c >= size(x)) -> undef
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->getValue() >= OpSizeInBits)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::UNDEF, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (srl x, 0) -> x
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N1C->isNullValue())
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// if (srl x, c) is known to be zero, return 0
|
2008-02-25 21:11:39 +00:00
|
|
|
if (N1C && DAG.MaskedValueIsZero(SDOperand(N, 0),
|
|
|
|
APInt::getAllOnesValue(OpSizeInBits)))
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2007-04-18 03:05:22 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (srl (srl x, c1), c2) -> 0 or (srl x, c1+c2)
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N1C && N0.getOpcode() == ISD::SRL &&
|
2005-09-01 00:19:25 +00:00
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant) {
|
|
|
|
uint64_t c1 = cast<ConstantSDNode>(N0.getOperand(1))->getValue();
|
2005-09-02 21:18:40 +00:00
|
|
|
uint64_t c2 = N1C->getValue();
|
2005-09-01 00:19:25 +00:00
|
|
|
if (c1 + c2 > OpSizeInBits)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getConstant(0, VT);
|
2005-09-01 00:19:25 +00:00
|
|
|
return DAG.getNode(ISD::SRL, VT, N0.getOperand(0),
|
2005-09-06 04:43:02 +00:00
|
|
|
DAG.getConstant(c1 + c2, N1.getValueType()));
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
|
2006-05-05 22:53:17 +00:00
|
|
|
// fold (srl (anyextend x), c) -> (anyextend (srl x, c))
|
|
|
|
if (N1C && N0.getOpcode() == ISD::ANY_EXTEND) {
|
|
|
|
// Shifting in all undef bits?
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT SmallVT = N0.getOperand(0).getValueType();
|
|
|
|
if (N1C->getValue() >= SmallVT.getSizeInBits())
|
2006-05-05 22:53:17 +00:00
|
|
|
return DAG.getNode(ISD::UNDEF, VT);
|
|
|
|
|
|
|
|
SDOperand SmallShift = DAG.getNode(ISD::SRL, SmallVT, N0.getOperand(0), N1);
|
|
|
|
AddToWorkList(SmallShift.Val);
|
|
|
|
return DAG.getNode(ISD::ANY_EXTEND, VT, SmallShift);
|
|
|
|
}
|
|
|
|
|
2006-10-12 20:23:19 +00:00
|
|
|
// fold (srl (sra X, Y), 31) -> (srl X, 31). This srl only looks at the sign
|
|
|
|
// bit, which is unmodified by sra.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (N1C && N1C->getValue()+1 == VT.getSizeInBits()) {
|
2006-10-12 20:23:19 +00:00
|
|
|
if (N0.getOpcode() == ISD::SRA)
|
|
|
|
return DAG.getNode(ISD::SRL, VT, N0.getOperand(0), N1);
|
|
|
|
}
|
|
|
|
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
// fold (srl (ctlz x), "5") -> x iff x has one bit set (the low bit).
|
|
|
|
if (N1C && N0.getOpcode() == ISD::CTLZ &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N1C->getAPIntValue() == Log2_32(VT.getSizeInBits())) {
|
2008-02-20 16:33:30 +00:00
|
|
|
APInt KnownZero, KnownOne;
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt Mask = APInt::getAllOnesValue(VT.getSizeInBits());
|
2007-06-22 14:59:07 +00:00
|
|
|
DAG.ComputeMaskedBits(N0.getOperand(0), Mask, KnownZero, KnownOne);
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
|
|
|
|
// If any of the input bits are KnownOne, then the input couldn't be all
|
|
|
|
// zeros, thus the result of the srl will always be zero.
|
2008-02-20 16:33:30 +00:00
|
|
|
if (KnownOne.getBoolValue()) return DAG.getConstant(0, VT);
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
|
|
|
|
// If all of the bits input the to ctlz node are known to be zero, then
|
|
|
|
// the result of the ctlz is "32" and the result of the shift is one.
|
2008-02-20 16:33:30 +00:00
|
|
|
APInt UnknownBits = ~KnownZero & Mask;
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
if (UnknownBits == 0) return DAG.getConstant(1, VT);
|
|
|
|
|
|
|
|
// Otherwise, check to see if there is exactly one bit input to the ctlz.
|
|
|
|
if ((UnknownBits & (UnknownBits-1)) == 0) {
|
|
|
|
// Okay, we know that only that the single bit specified by UnknownBits
|
|
|
|
// could be set on input to the CTLZ node. If this bit is set, the SRL
|
|
|
|
// will return 0, if it is clear, it returns 1. Change the CTLZ/SRL pair
|
|
|
|
// to an SRL,XOR pair, which is likely to simplify more.
|
2008-02-20 16:33:30 +00:00
|
|
|
unsigned ShAmt = UnknownBits.countTrailingZeros();
|
Add a little dag combine to compile this:
int %AreSecondAndThirdElementsBothNegative(<4 x float>* %in) {
entry:
%tmp1 = load <4 x float>* %in ; <<4 x float>> [#uses=1]
%tmp = tail call int %llvm.ppc.altivec.vcmpgefp.p( int 1, <4 x float> < float 0x7FF8000000000000, float 0.000000e+00, float 0.000000e+00, float 0x7FF8000000000000 >, <4 x float> %tmp1 ) ; <int> [#uses=1]
%tmp = seteq int %tmp, 0 ; <bool> [#uses=1]
%tmp3 = cast bool %tmp to int ; <int> [#uses=1]
ret int %tmp3
}
into this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
mtspr 256, r2
blr
instead of this:
_AreSecondAndThirdElementsBothNegative:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI1_0)
lis r5, ha16(LCPI1_0)
lvx v0, 0, r3
lvx v1, r5, r4
vcmpgefp. v0, v1, v0
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
xori r3, r3, 1
cntlzw r3, r3
srwi r3, r3, 5
mtspr 256, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27356 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 06:11:11 +00:00
|
|
|
SDOperand Op = N0.getOperand(0);
|
|
|
|
if (ShAmt) {
|
|
|
|
Op = DAG.getNode(ISD::SRL, VT, Op,
|
|
|
|
DAG.getConstant(ShAmt, TLI.getShiftAmountTy()));
|
|
|
|
AddToWorkList(Op.Val);
|
|
|
|
}
|
|
|
|
return DAG.getNode(ISD::XOR, VT, Op, DAG.getConstant(1, VT));
|
|
|
|
}
|
|
|
|
}
|
2007-04-18 03:06:49 +00:00
|
|
|
|
|
|
|
// fold operands of srl based on knowledge that the low bits are not
|
|
|
|
// demanded.
|
|
|
|
if (N1C && SimplifyDemandedBits(SDOperand(N, 0)))
|
|
|
|
return SDOperand(N, 0);
|
|
|
|
|
implement a readme entry, compiling the code into:
_foo:
movl $12, %eax
andl 4(%esp), %eax
movl _array(%eax), %eax
ret
instead of:
_foo:
movl 4(%esp), %eax
shrl $2, %eax
andl $3, %eax
movl _array(,%eax,4), %eax
ret
As it turns out, this triggers all the time, in a wide variety of
situations, for example, I see diffs like this in various programs:
- movl 8(%eax), %eax
- shll $2, %eax
- andl $1020, %eax
- movl (%esi,%eax), %eax
+ movzbl 8(%eax), %eax
+ movl (%esi,%eax,4), %eax
- shll $2, %edx
- andl $1020, %edx
- movl (%edi,%edx), %edx
+ andl $255, %edx
+ movl (%edi,%edx,4), %edx
Unfortunately, I also see stuff like this, which can be fixed in the
X86 backend:
- andl $85, %ebx
- addl _bit_count(,%ebx,4), %ebp
+ shll $2, %ebx
+ andl $340, %ebx
+ addl _bit_count(%ebx), %ebp
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44656 91177308-0d34-0410-b5e6-96231b3b80d8
2007-12-06 07:33:36 +00:00
|
|
|
return N1C ? visitShiftByConstant(N, N1C->getValue()) : SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitCTLZ(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (ctlz c1) -> c2
|
2006-05-06 23:06:26 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::CTLZ, VT, N0);
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitCTTZ(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (cttz c1) -> c2
|
2006-05-06 23:06:26 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::CTTZ, VT, N0);
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitCTPOP(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (ctpop c1) -> c2
|
2006-05-06 23:06:26 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::CTPOP, VT, N0);
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-16 00:54:12 +00:00
|
|
|
SDOperand DAGCombiner::visitSELECT(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
SDOperand N2 = N->getOperand(2);
|
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
|
|
|
ConstantSDNode *N2C = dyn_cast<ConstantSDNode>(N2);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
MVT VT0 = N0.getValueType();
|
2005-09-19 22:34:01 +00:00
|
|
|
|
2005-09-16 00:54:12 +00:00
|
|
|
// fold select C, X, X -> X
|
|
|
|
if (N1 == N2)
|
|
|
|
return N1;
|
|
|
|
// fold select true, X, Y -> X
|
|
|
|
if (N0C && !N0C->isNullValue())
|
|
|
|
return N1;
|
|
|
|
// fold select false, X, Y -> Y
|
|
|
|
if (N0C && N0C->isNullValue())
|
|
|
|
return N2;
|
|
|
|
// fold select C, 1, X -> C | X
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT == MVT::i1 && N1C && N1C->getAPIntValue() == 1)
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::OR, VT, N0, N2);
|
2007-08-18 05:57:05 +00:00
|
|
|
// fold select C, 0, 1 -> ~C
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isInteger() && VT0.isInteger() &&
|
2008-03-13 22:13:53 +00:00
|
|
|
N1C && N2C && N1C->isNullValue() && N2C->getAPIntValue() == 1) {
|
2007-08-18 05:57:05 +00:00
|
|
|
SDOperand XORNode = DAG.getNode(ISD::XOR, VT0, N0, DAG.getConstant(1, VT0));
|
|
|
|
if (VT == VT0)
|
|
|
|
return XORNode;
|
|
|
|
AddToWorkList(XORNode.Val);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (VT.bitsGT(VT0))
|
2007-08-18 05:57:05 +00:00
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, XORNode);
|
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, XORNode);
|
|
|
|
}
|
2005-09-16 00:54:12 +00:00
|
|
|
// fold select C, 0, X -> ~C & X
|
2007-12-06 17:53:31 +00:00
|
|
|
if (VT == VT0 && VT == MVT::i1 && N1C && N1C->isNullValue()) {
|
|
|
|
SDOperand XORNode = DAG.getNode(ISD::XOR, VT, N0, DAG.getConstant(1, VT));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(XORNode.Val);
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, XORNode, N2);
|
|
|
|
}
|
|
|
|
// fold select C, X, 1 -> ~C | X
|
2008-03-13 22:13:53 +00:00
|
|
|
if (VT == VT0 && VT == MVT::i1 && N2C && N2C->getAPIntValue() == 1) {
|
2007-12-06 17:53:31 +00:00
|
|
|
SDOperand XORNode = DAG.getNode(ISD::XOR, VT, N0, DAG.getConstant(1, VT));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(XORNode.Val);
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::OR, VT, XORNode, N1);
|
|
|
|
}
|
|
|
|
// fold select C, X, 0 -> C & X
|
|
|
|
// FIXME: this should check for C type == X type, not i1?
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT == MVT::i1 && N2C && N2C->isNullValue())
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, N0, N1);
|
|
|
|
// fold X ? X : Y --> X ? 1 : Y --> X | Y
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT == MVT::i1 && N0 == N1)
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::OR, VT, N0, N2);
|
|
|
|
// fold X ? Y : X --> X ? Y : 0 --> X & Y
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT == MVT::i1 && N0 == N2)
|
2005-09-16 00:54:12 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, N0, N1);
|
2006-05-27 00:43:02 +00:00
|
|
|
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
// If we can fold this based on the true/false value, do so.
|
|
|
|
if (SimplifySelectOps(N, N1, N2))
|
2006-05-27 00:43:02 +00:00
|
|
|
return SDOperand(N, 0); // Don't revisit N.
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
// fold selects based on a setcc into other things, such as min/max/abs
|
2008-02-20 11:10:28 +00:00
|
|
|
if (N0.getOpcode() == ISD::SETCC) {
|
2006-02-01 07:19:44 +00:00
|
|
|
// FIXME:
|
|
|
|
// Check against MVT::Other for SELECT_CC, which is a workaround for targets
|
|
|
|
// having to say they don't support SELECT_CC on every type the DAG knows
|
|
|
|
// about, since there is no way to mark an opcode illegal at all value types
|
|
|
|
if (TLI.isOperationLegal(ISD::SELECT_CC, MVT::Other))
|
|
|
|
return DAG.getNode(ISD::SELECT_CC, VT, N0.getOperand(0), N0.getOperand(1),
|
|
|
|
N1, N2, N0.getOperand(2));
|
|
|
|
else
|
|
|
|
return SimplifySelect(N0, N1, N2);
|
2008-02-20 11:10:28 +00:00
|
|
|
}
|
2005-09-16 00:54:12 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitSELECT_CC(SDNode *N) {
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
SDOperand N2 = N->getOperand(2);
|
|
|
|
SDOperand N3 = N->getOperand(3);
|
|
|
|
SDOperand N4 = N->getOperand(4);
|
|
|
|
ISD::CondCode CC = cast<CondCodeSDNode>(N4)->get();
|
|
|
|
|
|
|
|
// fold select_cc lhs, rhs, x, x, cc -> x
|
|
|
|
if (N2 == N3)
|
|
|
|
return N2;
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
|
Two things:
1. teach SimplifySetCC that '(srl (ctlz x), 5) == 0' is really x != 0.
2. Teach visitSELECT_CC to use SimplifySetCC instead of calling it and
ignoring the result. This allows us to compile:
bool %test(ulong %x) {
%tmp = setlt ulong %x, 4294967296
ret bool %tmp
}
to:
_test:
cntlzw r2, r3
cmplwi cr0, r3, 1
srwi r2, r2, 5
li r3, 0
beq cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
instead of:
_test:
addi r2, r3, -1
cntlzw r2, r2
cntlzw r3, r3
srwi r2, r2, 5
cmplwi cr0, r2, 0
srwi r2, r3, 5
li r3, 0
bne cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
This isn't wonderful, but it's an improvement.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30513 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-20 06:19:26 +00:00
|
|
|
// Determine if the condition we're dealing with is constant
|
2008-03-10 15:42:14 +00:00
|
|
|
SDOperand SCC = SimplifySetCC(TLI.getSetCCResultType(N0), N0, N1, CC, false);
|
2006-10-14 03:52:46 +00:00
|
|
|
if (SCC.Val) AddToWorkList(SCC.Val);
|
Two things:
1. teach SimplifySetCC that '(srl (ctlz x), 5) == 0' is really x != 0.
2. Teach visitSELECT_CC to use SimplifySetCC instead of calling it and
ignoring the result. This allows us to compile:
bool %test(ulong %x) {
%tmp = setlt ulong %x, 4294967296
ret bool %tmp
}
to:
_test:
cntlzw r2, r3
cmplwi cr0, r3, 1
srwi r2, r2, 5
li r3, 0
beq cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
instead of:
_test:
addi r2, r3, -1
cntlzw r2, r2
cntlzw r3, r3
srwi r2, r2, 5
cmplwi cr0, r2, 0
srwi r2, r3, 5
li r3, 0
bne cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
This isn't wonderful, but it's an improvement.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30513 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-20 06:19:26 +00:00
|
|
|
|
|
|
|
if (ConstantSDNode *SCCC = dyn_cast_or_null<ConstantSDNode>(SCC.Val)) {
|
2008-03-13 22:13:53 +00:00
|
|
|
if (!SCCC->isNullValue())
|
Two things:
1. teach SimplifySetCC that '(srl (ctlz x), 5) == 0' is really x != 0.
2. Teach visitSELECT_CC to use SimplifySetCC instead of calling it and
ignoring the result. This allows us to compile:
bool %test(ulong %x) {
%tmp = setlt ulong %x, 4294967296
ret bool %tmp
}
to:
_test:
cntlzw r2, r3
cmplwi cr0, r3, 1
srwi r2, r2, 5
li r3, 0
beq cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
instead of:
_test:
addi r2, r3, -1
cntlzw r2, r2
cntlzw r3, r3
srwi r2, r2, 5
cmplwi cr0, r2, 0
srwi r2, r3, 5
li r3, 0
bne cr0, LBB1_2 ;
LBB1_1: ;
mr r3, r2
LBB1_2: ;
blr
This isn't wonderful, but it's an improvement.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30513 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-20 06:19:26 +00:00
|
|
|
return N2; // cond always true -> true val
|
|
|
|
else
|
|
|
|
return N3; // cond always false -> false val
|
|
|
|
}
|
|
|
|
|
|
|
|
// Fold to a simpler select_cc
|
|
|
|
if (SCC.Val && SCC.getOpcode() == ISD::SETCC)
|
|
|
|
return DAG.getNode(ISD::SELECT_CC, N2.getValueType(),
|
|
|
|
SCC.getOperand(0), SCC.getOperand(1), N2, N3,
|
|
|
|
SCC.getOperand(2));
|
|
|
|
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
// If we can fold this based on the true/false value, do so.
|
|
|
|
if (SimplifySelectOps(N, N2, N3))
|
2006-05-27 00:43:02 +00:00
|
|
|
return SDOperand(N, 0); // Don't revisit N.
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
// fold select_cc into other things, such as min/max/abs
|
|
|
|
return SimplifySelectCC(N0, N1, N2, N3, CC);
|
2005-09-16 00:54:12 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitSETCC(SDNode *N) {
|
|
|
|
return SimplifySetCC(N->getValueType(0), N->getOperand(0), N->getOperand(1),
|
|
|
|
cast<CondCodeSDNode>(N->getOperand(2))->get());
|
|
|
|
}
|
|
|
|
|
2007-10-29 19:58:20 +00:00
|
|
|
// ExtendUsesToFormExtLoad - Trying to extend uses of a load to enable this:
|
|
|
|
// "fold ({s|z}ext (load x)) -> ({s|z}ext (truncate ({s|z}extload x)))"
|
|
|
|
// transformation. Returns true if extension are possible and the above
|
|
|
|
// mentioned transformation is profitable.
|
|
|
|
static bool ExtendUsesToFormExtLoad(SDNode *N, SDOperand N0,
|
|
|
|
unsigned ExtOpc,
|
|
|
|
SmallVector<SDNode*, 4> &ExtendNodes,
|
|
|
|
TargetLowering &TLI) {
|
|
|
|
bool HasCopyToRegUses = false;
|
|
|
|
bool isTruncFree = TLI.isTruncateFree(N->getValueType(0), N0.getValueType());
|
|
|
|
for (SDNode::use_iterator UI = N0.Val->use_begin(), UE = N0.Val->use_end();
|
|
|
|
UI != UE; ++UI) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *User = UI->getUser();
|
2007-10-29 19:58:20 +00:00
|
|
|
if (User == N)
|
|
|
|
continue;
|
|
|
|
// FIXME: Only extend SETCC N, N and SETCC N, c for now.
|
|
|
|
if (User->getOpcode() == ISD::SETCC) {
|
|
|
|
ISD::CondCode CC = cast<CondCodeSDNode>(User->getOperand(2))->get();
|
|
|
|
if (ExtOpc == ISD::ZERO_EXTEND && ISD::isSignedIntSetCC(CC))
|
|
|
|
// Sign bits will be lost after a zext.
|
|
|
|
return false;
|
|
|
|
bool Add = false;
|
|
|
|
for (unsigned i = 0; i != 2; ++i) {
|
|
|
|
SDOperand UseOp = User->getOperand(i);
|
|
|
|
if (UseOp == N0)
|
|
|
|
continue;
|
|
|
|
if (!isa<ConstantSDNode>(UseOp))
|
|
|
|
return false;
|
|
|
|
Add = true;
|
|
|
|
}
|
|
|
|
if (Add)
|
|
|
|
ExtendNodes.push_back(User);
|
|
|
|
} else {
|
|
|
|
for (unsigned i = 0, e = User->getNumOperands(); i != e; ++i) {
|
|
|
|
SDOperand UseOp = User->getOperand(i);
|
|
|
|
if (UseOp == N0) {
|
|
|
|
// If truncate from extended type to original load type is free
|
|
|
|
// on this target, then it's ok to extend a CopyToReg.
|
|
|
|
if (isTruncFree && User->getOpcode() == ISD::CopyToReg)
|
|
|
|
HasCopyToRegUses = true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (HasCopyToRegUses) {
|
|
|
|
bool BothLiveOut = false;
|
|
|
|
for (SDNode::use_iterator UI = N->use_begin(), UE = N->use_end();
|
|
|
|
UI != UE; ++UI) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *User = UI->getUser();
|
2007-10-29 19:58:20 +00:00
|
|
|
for (unsigned i = 0, e = User->getNumOperands(); i != e; ++i) {
|
|
|
|
SDOperand UseOp = User->getOperand(i);
|
|
|
|
if (UseOp.Val == N && UseOp.ResNo == 0) {
|
|
|
|
BothLiveOut = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (BothLiveOut)
|
|
|
|
// Both unextended and extended values are live out. There had better be
|
|
|
|
// good a reason for the transformation.
|
|
|
|
return ExtendNodes.size();
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSIGN_EXTEND(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (sext c1) -> c1
|
2006-11-02 20:25:50 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND, VT, N0);
|
2006-05-06 23:06:26 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sext (sext x)) -> (sext x)
|
2006-05-06 23:06:26 +00:00
|
|
|
// fold (sext (aext x)) -> (sext x)
|
|
|
|
if (N0.getOpcode() == ISD::SIGN_EXTEND || N0.getOpcode() == ISD::ANY_EXTEND)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND, VT, N0.getOperand(0));
|
2006-05-06 23:06:26 +00:00
|
|
|
|
2007-03-22 01:54:19 +00:00
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE) {
|
2008-05-20 20:56:33 +00:00
|
|
|
// fold (sext (truncate (load x))) -> (sext (smaller load x))
|
|
|
|
// fold (sext (truncate (srl (load x), c))) -> (sext (smaller load (x+c/n)))
|
2007-03-22 01:54:19 +00:00
|
|
|
SDOperand NarrowLoad = ReduceLoadWidth(N0.Val);
|
2007-03-23 02:16:52 +00:00
|
|
|
if (NarrowLoad.Val) {
|
|
|
|
if (NarrowLoad.Val != N0.Val)
|
|
|
|
CombineTo(N0.Val, NarrowLoad);
|
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND, VT, NarrowLoad);
|
|
|
|
}
|
2007-03-22 01:54:19 +00:00
|
|
|
|
2008-05-20 20:56:33 +00:00
|
|
|
// See if the value being truncated is already sign extended. If so, just
|
|
|
|
// eliminate the trunc/sext pair.
|
2006-09-21 06:00:20 +00:00
|
|
|
SDOperand Op = N0.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned OpBits = Op.getValueType().getSizeInBits();
|
|
|
|
unsigned MidBits = N0.getValueType().getSizeInBits();
|
|
|
|
unsigned DestBits = VT.getSizeInBits();
|
2007-06-22 14:59:07 +00:00
|
|
|
unsigned NumSignBits = DAG.ComputeNumSignBits(Op);
|
2007-02-26 03:13:59 +00:00
|
|
|
|
|
|
|
if (OpBits == DestBits) {
|
|
|
|
// Op is i32, Mid is i8, and Dest is i32. If Op has more than 24 sign
|
|
|
|
// bits, it is already ready.
|
|
|
|
if (NumSignBits > DestBits-MidBits)
|
|
|
|
return Op;
|
|
|
|
} else if (OpBits < DestBits) {
|
|
|
|
// Op is i32, Mid is i8, and Dest is i64. If Op has more than 24 sign
|
|
|
|
// bits, just sext from i32.
|
|
|
|
if (NumSignBits > OpBits-MidBits)
|
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND, VT, Op);
|
|
|
|
} else {
|
|
|
|
// Op is i64, Mid is i8, and Dest is i32. If Op has more than 56 sign
|
|
|
|
// bits, just truncate to i32.
|
|
|
|
if (NumSignBits > OpBits-MidBits)
|
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, Op);
|
|
|
|
}
|
|
|
|
|
|
|
|
// fold (sext (truncate x)) -> (sextinreg x).
|
|
|
|
if (!AfterLegalize || TLI.isOperationLegal(ISD::SIGN_EXTEND_INREG,
|
|
|
|
N0.getValueType())) {
|
2008-06-08 20:54:56 +00:00
|
|
|
if (Op.getValueType().bitsLT(VT))
|
2007-02-26 03:13:59 +00:00
|
|
|
Op = DAG.getNode(ISD::ANY_EXTEND, VT, Op);
|
2008-06-08 20:54:56 +00:00
|
|
|
else if (Op.getValueType().bitsGT(VT))
|
2007-02-26 03:13:59 +00:00
|
|
|
Op = DAG.getNode(ISD::TRUNCATE, VT, Op);
|
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND_INREG, VT, Op,
|
|
|
|
DAG.getValueType(N0.getValueType()));
|
2006-09-21 06:00:20 +00:00
|
|
|
}
|
|
|
|
}
|
2006-05-06 23:06:26 +00:00
|
|
|
|
2005-12-14 02:19:23 +00:00
|
|
|
// fold (sext (load x)) -> (sext (truncate (sextload x)))
|
2007-10-29 19:58:20 +00:00
|
|
|
if (ISD::isNON_EXTLoad(N0.Val) &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::SEXTLOAD, N0.getValueType()))) {
|
2007-10-29 19:58:20 +00:00
|
|
|
bool DoXform = true;
|
|
|
|
SmallVector<SDNode*, 4> SetCCs;
|
|
|
|
if (!N0.hasOneUse())
|
|
|
|
DoXform = ExtendUsesToFormExtLoad(N, N0, ISD::SIGN_EXTEND, SetCCs, TLI);
|
|
|
|
if (DoXform) {
|
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::SEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
|
|
|
LN0->getSrcValueOffset(),
|
|
|
|
N0.getValueType(),
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
SDOperand Trunc = DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad);
|
|
|
|
CombineTo(N0.Val, Trunc, ExtLoad.getValue(1));
|
|
|
|
// Extend SetCC uses if necessary.
|
|
|
|
for (unsigned i = 0, e = SetCCs.size(); i != e; ++i) {
|
|
|
|
SDNode *SetCC = SetCCs[i];
|
|
|
|
SmallVector<SDOperand, 4> Ops;
|
|
|
|
for (unsigned j = 0; j != 2; ++j) {
|
|
|
|
SDOperand SOp = SetCC->getOperand(j);
|
|
|
|
if (SOp == Trunc)
|
|
|
|
Ops.push_back(ExtLoad);
|
|
|
|
else
|
|
|
|
Ops.push_back(DAG.getNode(ISD::SIGN_EXTEND, VT, SOp));
|
|
|
|
}
|
|
|
|
Ops.push_back(SetCC->getOperand(2));
|
|
|
|
CombineTo(SetCC, DAG.getNode(ISD::SETCC, SetCC->getValueType(0),
|
|
|
|
&Ops[0], Ops.size()));
|
|
|
|
}
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
More cool stuff for the dag combiner. We can now finally handle things
like turning:
_foo:
fctiwz f0, f1
stfd f0, -8(r1)
lwz r2, -4(r1)
rlwinm r3, r2, 0, 16, 31
blr
into
_foo:
fctiwz f0,f1
stfd f0,-8(r1)
lhz r3,-2(r1)
blr
Also removed an unncessary constraint from sra -> srl conversion, which
should take care of hte only reason we would ever need to handle sra in
MaskedValueIsZero, AFAIK.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23703 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-12 20:40:40 +00:00
|
|
|
}
|
2005-12-14 19:05:06 +00:00
|
|
|
|
|
|
|
// fold (sext (sextload x)) -> (sext (truncate (sextload x)))
|
|
|
|
// fold (sext ( extload x)) -> (sext (truncate (sextload x)))
|
2007-03-07 08:07:03 +00:00
|
|
|
if ((ISD::isSEXTLoad(N0.Val) || ISD::isEXTLoad(N0.Val)) &&
|
|
|
|
ISD::isUNINDEXEDLoad(N0.Val) && N0.hasOneUse()) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = LN0->getMemoryVT();
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if ((!AfterLegalize && !LN0->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::SEXTLOAD, EVT)) {
|
2006-12-15 21:38:30 +00:00
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::SEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-12-15 21:38:30 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad),
|
|
|
|
ExtLoad.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
2005-12-14 19:05:06 +00:00
|
|
|
}
|
|
|
|
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
// sext(setcc x,y,cc) -> select_cc x, y, -1, 0, cc
|
|
|
|
if (N0.getOpcode() == ISD::SETCC) {
|
|
|
|
SDOperand SCC =
|
2007-04-11 06:50:51 +00:00
|
|
|
SimplifySelectCC(N0.getOperand(0), N0.getOperand(1),
|
|
|
|
DAG.getConstant(~0ULL, VT), DAG.getConstant(0, VT),
|
|
|
|
cast<CondCodeSDNode>(N0.getOperand(2))->get(), true);
|
|
|
|
if (SCC.Val) return SCC;
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
}
|
|
|
|
|
2008-04-28 16:58:24 +00:00
|
|
|
// fold (sext x) -> (zext x) if the sign bit is known zero.
|
2008-04-28 18:47:17 +00:00
|
|
|
if ((!AfterLegalize || TLI.isOperationLegal(ISD::ZERO_EXTEND, VT)) &&
|
|
|
|
DAG.SignBitIsZero(N0))
|
2008-04-28 16:58:24 +00:00
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, N0);
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitZERO_EXTEND(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (zext c1) -> c1
|
2006-11-02 20:25:50 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (zext (zext x)) -> (zext x)
|
2006-05-06 23:06:26 +00:00
|
|
|
// fold (zext (aext x)) -> (zext x)
|
|
|
|
if (N0.getOpcode() == ISD::ZERO_EXTEND || N0.getOpcode() == ISD::ANY_EXTEND)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, N0.getOperand(0));
|
2006-09-21 06:00:20 +00:00
|
|
|
|
2007-03-22 01:54:19 +00:00
|
|
|
// fold (zext (truncate (load x))) -> (zext (smaller load x))
|
|
|
|
// fold (zext (truncate (srl (load x), c))) -> (zext (small load (x+c/n)))
|
2007-03-30 21:38:07 +00:00
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE) {
|
2007-03-22 01:54:19 +00:00
|
|
|
SDOperand NarrowLoad = ReduceLoadWidth(N0.Val);
|
2007-03-23 02:16:52 +00:00
|
|
|
if (NarrowLoad.Val) {
|
|
|
|
if (NarrowLoad.Val != N0.Val)
|
|
|
|
CombineTo(N0.Val, NarrowLoad);
|
|
|
|
return DAG.getNode(ISD::ZERO_EXTEND, VT, NarrowLoad);
|
|
|
|
}
|
2007-03-22 01:54:19 +00:00
|
|
|
}
|
|
|
|
|
2006-09-21 06:00:20 +00:00
|
|
|
// fold (zext (truncate x)) -> (and x, mask)
|
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE &&
|
|
|
|
(!AfterLegalize || TLI.isOperationLegal(ISD::AND, VT))) {
|
|
|
|
SDOperand Op = N0.getOperand(0);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (Op.getValueType().bitsLT(VT)) {
|
2006-09-21 06:00:20 +00:00
|
|
|
Op = DAG.getNode(ISD::ANY_EXTEND, VT, Op);
|
2008-06-08 20:54:56 +00:00
|
|
|
} else if (Op.getValueType().bitsGT(VT)) {
|
2006-09-21 06:00:20 +00:00
|
|
|
Op = DAG.getNode(ISD::TRUNCATE, VT, Op);
|
|
|
|
}
|
|
|
|
return DAG.getZeroExtendInReg(Op, N0.getValueType());
|
|
|
|
}
|
|
|
|
|
Compile:
int %test(ulong *%tmp) {
%tmp = load ulong* %tmp ; <ulong> [#uses=1]
%tmp.mask = shr ulong %tmp, ubyte 50 ; <ulong> [#uses=1]
%tmp.mask = cast ulong %tmp.mask to ubyte
%tmp2 = and ubyte %tmp.mask, 3 ; <ubyte> [#uses=1]
%tmp2 = cast ubyte %tmp2 to int ; <int> [#uses=1]
ret int %tmp2
}
to:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
andl $3, %eax
ret
instead of:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
# TRUNCATE movb %al, %al
andb $3, %al
movzbl %al, %eax
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30558 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-21 06:14:31 +00:00
|
|
|
// fold (zext (and (trunc x), cst)) -> (and x, cst).
|
|
|
|
if (N0.getOpcode() == ISD::AND &&
|
|
|
|
N0.getOperand(0).getOpcode() == ISD::TRUNCATE &&
|
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant) {
|
|
|
|
SDOperand X = N0.getOperand(0).getOperand(0);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (X.getValueType().bitsLT(VT)) {
|
Compile:
int %test(ulong *%tmp) {
%tmp = load ulong* %tmp ; <ulong> [#uses=1]
%tmp.mask = shr ulong %tmp, ubyte 50 ; <ulong> [#uses=1]
%tmp.mask = cast ulong %tmp.mask to ubyte
%tmp2 = and ubyte %tmp.mask, 3 ; <ubyte> [#uses=1]
%tmp2 = cast ubyte %tmp2 to int ; <int> [#uses=1]
ret int %tmp2
}
to:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
andl $3, %eax
ret
instead of:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
# TRUNCATE movb %al, %al
andb $3, %al
movzbl %al, %eax
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30558 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-21 06:14:31 +00:00
|
|
|
X = DAG.getNode(ISD::ANY_EXTEND, VT, X);
|
2008-06-08 20:54:56 +00:00
|
|
|
} else if (X.getValueType().bitsGT(VT)) {
|
Compile:
int %test(ulong *%tmp) {
%tmp = load ulong* %tmp ; <ulong> [#uses=1]
%tmp.mask = shr ulong %tmp, ubyte 50 ; <ulong> [#uses=1]
%tmp.mask = cast ulong %tmp.mask to ubyte
%tmp2 = and ubyte %tmp.mask, 3 ; <ubyte> [#uses=1]
%tmp2 = cast ubyte %tmp2 to int ; <int> [#uses=1]
ret int %tmp2
}
to:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
andl $3, %eax
ret
instead of:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
# TRUNCATE movb %al, %al
andb $3, %al
movzbl %al, %eax
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30558 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-21 06:14:31 +00:00
|
|
|
X = DAG.getNode(ISD::TRUNCATE, VT, X);
|
|
|
|
}
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt Mask = cast<ConstantSDNode>(N0.getOperand(1))->getAPIntValue();
|
2008-06-06 12:08:01 +00:00
|
|
|
Mask.zext(VT.getSizeInBits());
|
Compile:
int %test(ulong *%tmp) {
%tmp = load ulong* %tmp ; <ulong> [#uses=1]
%tmp.mask = shr ulong %tmp, ubyte 50 ; <ulong> [#uses=1]
%tmp.mask = cast ulong %tmp.mask to ubyte
%tmp2 = and ubyte %tmp.mask, 3 ; <ubyte> [#uses=1]
%tmp2 = cast ubyte %tmp2 to int ; <int> [#uses=1]
ret int %tmp2
}
to:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
andl $3, %eax
ret
instead of:
_test:
movl 4(%esp), %eax
movl 4(%eax), %eax
shrl $18, %eax
# TRUNCATE movb %al, %al
andb $3, %al
movzbl %al, %eax
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30558 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-21 06:14:31 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, X, DAG.getConstant(Mask, VT));
|
|
|
|
}
|
|
|
|
|
2005-12-14 02:19:23 +00:00
|
|
|
// fold (zext (load x)) -> (zext (truncate (zextload x)))
|
2007-10-29 19:58:20 +00:00
|
|
|
if (ISD::isNON_EXTLoad(N0.Val) &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::ZEXTLOAD, N0.getValueType()))) {
|
2007-10-29 19:58:20 +00:00
|
|
|
bool DoXform = true;
|
|
|
|
SmallVector<SDNode*, 4> SetCCs;
|
|
|
|
if (!N0.hasOneUse())
|
|
|
|
DoXform = ExtendUsesToFormExtLoad(N, N0, ISD::ZERO_EXTEND, SetCCs, TLI);
|
|
|
|
if (DoXform) {
|
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::ZEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
|
|
|
LN0->getSrcValueOffset(),
|
|
|
|
N0.getValueType(),
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
SDOperand Trunc = DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad);
|
|
|
|
CombineTo(N0.Val, Trunc, ExtLoad.getValue(1));
|
|
|
|
// Extend SetCC uses if necessary.
|
|
|
|
for (unsigned i = 0, e = SetCCs.size(); i != e; ++i) {
|
|
|
|
SDNode *SetCC = SetCCs[i];
|
|
|
|
SmallVector<SDOperand, 4> Ops;
|
|
|
|
for (unsigned j = 0; j != 2; ++j) {
|
|
|
|
SDOperand SOp = SetCC->getOperand(j);
|
|
|
|
if (SOp == Trunc)
|
|
|
|
Ops.push_back(ExtLoad);
|
|
|
|
else
|
2007-10-30 20:11:21 +00:00
|
|
|
Ops.push_back(DAG.getNode(ISD::ZERO_EXTEND, VT, SOp));
|
2007-10-29 19:58:20 +00:00
|
|
|
}
|
|
|
|
Ops.push_back(SetCC->getOperand(2));
|
|
|
|
CombineTo(SetCC, DAG.getNode(ISD::SETCC, SetCC->getValueType(0),
|
|
|
|
&Ops[0], Ops.size()));
|
|
|
|
}
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
2005-12-14 02:19:23 +00:00
|
|
|
}
|
2005-12-14 19:05:06 +00:00
|
|
|
|
|
|
|
// fold (zext (zextload x)) -> (zext (truncate (zextload x)))
|
|
|
|
// fold (zext ( extload x)) -> (zext (truncate (zextload x)))
|
2007-03-07 08:07:03 +00:00
|
|
|
if ((ISD::isZEXTLoad(N0.Val) || ISD::isEXTLoad(N0.Val)) &&
|
|
|
|
ISD::isUNINDEXEDLoad(N0.Val) && N0.hasOneUse()) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = LN0->getMemoryVT();
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if ((!AfterLegalize && !LN0->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::ZEXTLOAD, EVT)) {
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::ZEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad),
|
|
|
|
ExtLoad.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
2005-12-14 19:05:06 +00:00
|
|
|
}
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
|
|
|
|
// zext(setcc x,y,cc) -> select_cc x, y, 1, 0, cc
|
|
|
|
if (N0.getOpcode() == ISD::SETCC) {
|
|
|
|
SDOperand SCC =
|
|
|
|
SimplifySelectCC(N0.getOperand(0), N0.getOperand(1),
|
|
|
|
DAG.getConstant(1, VT), DAG.getConstant(0, VT),
|
2007-04-11 06:50:51 +00:00
|
|
|
cast<CondCodeSDNode>(N0.getOperand(2))->get(), true);
|
|
|
|
if (SCC.Val) return SCC;
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2006-05-05 05:58:59 +00:00
|
|
|
SDOperand DAGCombiner::visitANY_EXTEND(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2006-05-05 05:58:59 +00:00
|
|
|
|
|
|
|
// fold (aext c1) -> c1
|
2006-05-06 23:06:26 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-05-05 05:58:59 +00:00
|
|
|
return DAG.getNode(ISD::ANY_EXTEND, VT, N0);
|
|
|
|
// fold (aext (aext x)) -> (aext x)
|
|
|
|
// fold (aext (zext x)) -> (zext x)
|
|
|
|
// fold (aext (sext x)) -> (sext x)
|
|
|
|
if (N0.getOpcode() == ISD::ANY_EXTEND ||
|
|
|
|
N0.getOpcode() == ISD::ZERO_EXTEND ||
|
|
|
|
N0.getOpcode() == ISD::SIGN_EXTEND)
|
|
|
|
return DAG.getNode(N0.getOpcode(), VT, N0.getOperand(0));
|
|
|
|
|
2007-03-22 01:54:19 +00:00
|
|
|
// fold (aext (truncate (load x))) -> (aext (smaller load x))
|
|
|
|
// fold (aext (truncate (srl (load x), c))) -> (aext (small load (x+c/n)))
|
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE) {
|
|
|
|
SDOperand NarrowLoad = ReduceLoadWidth(N0.Val);
|
2007-03-23 02:16:52 +00:00
|
|
|
if (NarrowLoad.Val) {
|
|
|
|
if (NarrowLoad.Val != N0.Val)
|
|
|
|
CombineTo(N0.Val, NarrowLoad);
|
|
|
|
return DAG.getNode(ISD::ANY_EXTEND, VT, NarrowLoad);
|
|
|
|
}
|
2007-03-22 01:54:19 +00:00
|
|
|
}
|
|
|
|
|
2006-09-20 06:29:17 +00:00
|
|
|
// fold (aext (truncate x))
|
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE) {
|
|
|
|
SDOperand TruncOp = N0.getOperand(0);
|
|
|
|
if (TruncOp.getValueType() == VT)
|
|
|
|
return TruncOp; // x iff x size == zext size.
|
2008-06-08 20:54:56 +00:00
|
|
|
if (TruncOp.getValueType().bitsGT(VT))
|
2006-09-20 06:29:17 +00:00
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, TruncOp);
|
|
|
|
return DAG.getNode(ISD::ANY_EXTEND, VT, TruncOp);
|
|
|
|
}
|
2006-09-21 06:40:43 +00:00
|
|
|
|
|
|
|
// fold (aext (and (trunc x), cst)) -> (and x, cst).
|
|
|
|
if (N0.getOpcode() == ISD::AND &&
|
|
|
|
N0.getOperand(0).getOpcode() == ISD::TRUNCATE &&
|
|
|
|
N0.getOperand(1).getOpcode() == ISD::Constant) {
|
|
|
|
SDOperand X = N0.getOperand(0).getOperand(0);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (X.getValueType().bitsLT(VT)) {
|
2006-09-21 06:40:43 +00:00
|
|
|
X = DAG.getNode(ISD::ANY_EXTEND, VT, X);
|
2008-06-08 20:54:56 +00:00
|
|
|
} else if (X.getValueType().bitsGT(VT)) {
|
2006-09-21 06:40:43 +00:00
|
|
|
X = DAG.getNode(ISD::TRUNCATE, VT, X);
|
|
|
|
}
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt Mask = cast<ConstantSDNode>(N0.getOperand(1))->getAPIntValue();
|
2008-06-06 12:08:01 +00:00
|
|
|
Mask.zext(VT.getSizeInBits());
|
2006-09-21 06:40:43 +00:00
|
|
|
return DAG.getNode(ISD::AND, VT, X, DAG.getConstant(Mask, VT));
|
|
|
|
}
|
|
|
|
|
2006-05-05 05:58:59 +00:00
|
|
|
// fold (aext (load x)) -> (aext (truncate (extload x)))
|
2006-10-09 20:57:25 +00:00
|
|
|
if (ISD::isNON_EXTLoad(N0.Val) && N0.hasOneUse() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::EXTLOAD, N0.getValueType()))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::EXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
|
|
|
LN0->getSrcValueOffset(),
|
2007-04-22 23:15:30 +00:00
|
|
|
N0.getValueType(),
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-05-05 05:58:59 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad),
|
|
|
|
ExtLoad.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
|
|
|
|
|
|
|
// fold (aext (zextload x)) -> (aext (truncate (zextload x)))
|
|
|
|
// fold (aext (sextload x)) -> (aext (truncate (sextload x)))
|
|
|
|
// fold (aext ( extload x)) -> (aext (truncate (extload x)))
|
2007-03-07 08:07:03 +00:00
|
|
|
if (N0.getOpcode() == ISD::LOAD &&
|
|
|
|
!ISD::isNON_EXTLoad(N0.Val) && ISD::isUNINDEXEDLoad(N0.Val) &&
|
2006-10-09 20:57:25 +00:00
|
|
|
N0.hasOneUse()) {
|
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = LN0->getMemoryVT();
|
2006-10-09 20:57:25 +00:00
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(LN0->getExtensionType(), VT,
|
|
|
|
LN0->getChain(), LN0->getBasePtr(),
|
|
|
|
LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-05-05 05:58:59 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::TRUNCATE, N0.getValueType(), ExtLoad),
|
|
|
|
ExtLoad.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
|
|
|
|
// aext(setcc x,y,cc) -> select_cc x, y, 1, 0, cc
|
|
|
|
if (N0.getOpcode() == ISD::SETCC) {
|
|
|
|
SDOperand SCC =
|
2007-04-11 06:50:51 +00:00
|
|
|
SimplifySelectCC(N0.getOperand(0), N0.getOperand(1),
|
|
|
|
DAG.getConstant(1, VT), DAG.getConstant(0, VT),
|
2007-04-11 16:51:53 +00:00
|
|
|
cast<CondCodeSDNode>(N0.getOperand(2))->get(), true);
|
2007-04-11 06:50:51 +00:00
|
|
|
if (SCC.Val)
|
2007-04-11 06:43:25 +00:00
|
|
|
return SCC;
|
Teach the codegen to turn [aez]ext (setcc) -> selectcc of 1/0, which often
allows other simplifications. For example, this compiles:
int isnegative(unsigned int X) {
return !(X < 2147483648U);
}
Into this code:
x86:
movl 4(%esp), %eax
shrl $31, %eax
ret
arm:
mov r0, r0, lsr #31
bx lr
thumb:
lsr r0, r0, #31
bx lr
instead of:
x86:
cmpl $0, 4(%esp)
sets %al
movzbl %al, %eax
ret
arm:
mov r3, #0
cmp r0, #0
movlt r3, #1
mov r0, r3
bx lr
thumb:
mov r2, #1
mov r1, #0
cmp r0, #0
blt LBB1_2 @entry
LBB1_1: @entry
cpy r2, r1
LBB1_2: @entry
cpy r0, r2
bx lr
Testcase here: test/CodeGen/Generic/ispositive.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35883 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:32:27 +00:00
|
|
|
}
|
|
|
|
|
2006-05-05 05:58:59 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-10-13 06:35:54 +00:00
|
|
|
/// GetDemandedBits - See if the specified operand can be simplified with the
|
|
|
|
/// knowledge that only the bits specified by Mask are used. If so, return the
|
|
|
|
/// simpler operand, otherwise return a null SDOperand.
|
2008-02-25 21:11:39 +00:00
|
|
|
SDOperand DAGCombiner::GetDemandedBits(SDOperand V, const APInt &Mask) {
|
2007-10-13 06:35:54 +00:00
|
|
|
switch (V.getOpcode()) {
|
|
|
|
default: break;
|
|
|
|
case ISD::OR:
|
|
|
|
case ISD::XOR:
|
|
|
|
// If the LHS or RHS don't contribute bits to the or, drop them.
|
|
|
|
if (DAG.MaskedValueIsZero(V.getOperand(0), Mask))
|
|
|
|
return V.getOperand(1);
|
|
|
|
if (DAG.MaskedValueIsZero(V.getOperand(1), Mask))
|
|
|
|
return V.getOperand(0);
|
|
|
|
break;
|
Enhance the truncstore optimization code to handle shifted
values and propagate demanded bits through them in simple cases.
This allows this code:
void foo(char *P) {
strcpy(P, "abc");
}
to compile to:
_foo:
ldrb r3, [r1]
ldrb r2, [r1, #+1]
ldrb r12, [r1, #+2]!
ldrb r1, [r1, #+1]
strb r1, [r0, #+3]
strb r2, [r0, #+1]
strb r12, [r0, #+2]
strb r3, [r0]
bx lr
instead of:
_foo:
ldrb r3, [r1, #+3]
ldrb r2, [r1, #+2]
orr r3, r2, r3, lsl #8
ldrb r2, [r1, #+1]
ldrb r1, [r1]
orr r2, r1, r2, lsl #8
orr r3, r2, r3, lsl #16
strb r3, [r0]
mov r2, r3, lsr #24
strb r2, [r0, #+3]
mov r2, r3, lsr #16
strb r2, [r0, #+2]
mov r3, r3, lsr #8
strb r3, [r0, #+1]
bx lr
testcase here: test/CodeGen/ARM/truncstore-dag-combine.ll
This also helps occasionally for X86 and other cases not involving
unaligned load/stores.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@42954 91177308-0d34-0410-b5e6-96231b3b80d8
2007-10-13 06:58:48 +00:00
|
|
|
case ISD::SRL:
|
|
|
|
// Only look at single-use SRLs.
|
|
|
|
if (!V.Val->hasOneUse())
|
|
|
|
break;
|
|
|
|
if (ConstantSDNode *RHSC = dyn_cast<ConstantSDNode>(V.getOperand(1))) {
|
|
|
|
// See if we can recursively simplify the LHS.
|
|
|
|
unsigned Amt = RHSC->getValue();
|
2008-02-25 21:11:39 +00:00
|
|
|
APInt NewMask = Mask << Amt;
|
|
|
|
SDOperand SimplifyLHS = GetDemandedBits(V.getOperand(0), NewMask);
|
Enhance the truncstore optimization code to handle shifted
values and propagate demanded bits through them in simple cases.
This allows this code:
void foo(char *P) {
strcpy(P, "abc");
}
to compile to:
_foo:
ldrb r3, [r1]
ldrb r2, [r1, #+1]
ldrb r12, [r1, #+2]!
ldrb r1, [r1, #+1]
strb r1, [r0, #+3]
strb r2, [r0, #+1]
strb r12, [r0, #+2]
strb r3, [r0]
bx lr
instead of:
_foo:
ldrb r3, [r1, #+3]
ldrb r2, [r1, #+2]
orr r3, r2, r3, lsl #8
ldrb r2, [r1, #+1]
ldrb r1, [r1]
orr r2, r1, r2, lsl #8
orr r3, r2, r3, lsl #16
strb r3, [r0]
mov r2, r3, lsr #24
strb r2, [r0, #+3]
mov r2, r3, lsr #16
strb r2, [r0, #+2]
mov r3, r3, lsr #8
strb r3, [r0, #+1]
bx lr
testcase here: test/CodeGen/ARM/truncstore-dag-combine.ll
This also helps occasionally for X86 and other cases not involving
unaligned load/stores.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@42954 91177308-0d34-0410-b5e6-96231b3b80d8
2007-10-13 06:58:48 +00:00
|
|
|
if (SimplifyLHS.Val) {
|
|
|
|
return DAG.getNode(ISD::SRL, V.getValueType(),
|
|
|
|
SimplifyLHS, V.getOperand(1));
|
|
|
|
}
|
|
|
|
}
|
2007-10-13 06:35:54 +00:00
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-03-22 01:54:19 +00:00
|
|
|
/// ReduceLoadWidth - If the result of a wider load is shifted to right of N
|
|
|
|
/// bits and then truncated to a narrower type and where N is a multiple
|
|
|
|
/// of number of bits of the narrower type, transform it to a narrower load
|
|
|
|
/// from address + N / num of bits of new type. If the result is to be
|
|
|
|
/// extended, also fold the extension to form a extending load.
|
|
|
|
SDOperand DAGCombiner::ReduceLoadWidth(SDNode *N) {
|
|
|
|
unsigned Opc = N->getOpcode();
|
|
|
|
ISD::LoadExtType ExtType = ISD::NON_EXTLOAD;
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
MVT EVT = N->getValueType(0);
|
2007-03-22 01:54:19 +00:00
|
|
|
|
2007-03-23 22:13:36 +00:00
|
|
|
// Special case: SIGN_EXTEND_INREG is basically truncating to EVT then
|
|
|
|
// extended to VT.
|
2007-03-22 01:54:19 +00:00
|
|
|
if (Opc == ISD::SIGN_EXTEND_INREG) {
|
|
|
|
ExtType = ISD::SEXTLOAD;
|
|
|
|
EVT = cast<VTSDNode>(N->getOperand(1))->getVT();
|
2007-03-23 22:13:36 +00:00
|
|
|
if (AfterLegalize && !TLI.isLoadXLegal(ISD::SEXTLOAD, EVT))
|
|
|
|
return SDOperand();
|
2007-03-22 01:54:19 +00:00
|
|
|
}
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned EVTBits = EVT.getSizeInBits();
|
2007-03-22 01:54:19 +00:00
|
|
|
unsigned ShAmt = 0;
|
2007-03-23 20:55:21 +00:00
|
|
|
bool CombineSRL = false;
|
2007-03-22 01:54:19 +00:00
|
|
|
if (N0.getOpcode() == ISD::SRL && N0.hasOneUse()) {
|
|
|
|
if (ConstantSDNode *N01 = dyn_cast<ConstantSDNode>(N0.getOperand(1))) {
|
|
|
|
ShAmt = N01->getValue();
|
|
|
|
// Is the shift amount a multiple of size of VT?
|
|
|
|
if ((ShAmt & (EVTBits-1)) == 0) {
|
|
|
|
N0 = N0.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
if (N0.getValueType().getSizeInBits() <= EVTBits)
|
2007-03-22 01:54:19 +00:00
|
|
|
return SDOperand();
|
2007-03-23 20:55:21 +00:00
|
|
|
CombineSRL = true;
|
2007-03-22 01:54:19 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-06-16 08:14:38 +00:00
|
|
|
// Do not generate loads of non-round integer types since these can
|
|
|
|
// be expensive (and would be wrong if the type is not byte sized).
|
|
|
|
if (ISD::isNON_EXTLoad(N0.Val) && N0.hasOneUse() && VT.isRound() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// Do not change the width of a volatile load.
|
|
|
|
!cast<LoadSDNode>(N0)->isVolatile()) {
|
2008-06-06 12:08:01 +00:00
|
|
|
assert(N0.getValueType().getSizeInBits() > EVTBits &&
|
2007-03-22 01:54:19 +00:00
|
|
|
"Cannot truncate to larger type!");
|
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT PtrType = N0.getOperand(1).getValueType();
|
2007-03-24 00:02:43 +00:00
|
|
|
// For big endian targets, we need to adjust the offset to the pointer to
|
|
|
|
// load the correct bytes.
|
2008-02-11 10:37:04 +00:00
|
|
|
if (TLI.isBigEndian()) {
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned LVTStoreBits = N0.getValueType().getStoreSizeInBits();
|
|
|
|
unsigned EVTStoreBits = EVT.getStoreSizeInBits();
|
2007-11-09 08:57:19 +00:00
|
|
|
ShAmt = LVTStoreBits - EVTStoreBits - ShAmt;
|
|
|
|
}
|
2007-03-24 00:02:43 +00:00
|
|
|
uint64_t PtrOff = ShAmt / 8;
|
2007-10-28 12:59:45 +00:00
|
|
|
unsigned NewAlign = MinAlign(LN0->getAlignment(), PtrOff);
|
2007-03-22 01:54:19 +00:00
|
|
|
SDOperand NewPtr = DAG.getNode(ISD::ADD, PtrType, LN0->getBasePtr(),
|
|
|
|
DAG.getConstant(PtrOff, PtrType));
|
|
|
|
AddToWorkList(NewPtr.Val);
|
|
|
|
SDOperand Load = (ExtType == ISD::NON_EXTLOAD)
|
|
|
|
? DAG.getLoad(VT, LN0->getChain(), NewPtr,
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValue(), LN0->getSrcValueOffset(),
|
2007-10-28 12:59:45 +00:00
|
|
|
LN0->isVolatile(), NewAlign)
|
2007-03-22 01:54:19 +00:00
|
|
|
: DAG.getExtLoad(ExtType, VT, LN0->getChain(), NewPtr,
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValue(), LN0->getSrcValueOffset(), EVT,
|
2007-10-28 12:59:45 +00:00
|
|
|
LN0->isVolatile(), NewAlign);
|
2007-03-22 01:54:19 +00:00
|
|
|
AddToWorkList(N);
|
2007-03-23 20:55:21 +00:00
|
|
|
if (CombineSRL) {
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(N0.getValue(1), Load.getValue(1),
|
|
|
|
&DeadNodes);
|
2007-03-23 20:55:21 +00:00
|
|
|
CombineTo(N->getOperand(0).Val, Load);
|
|
|
|
} else
|
|
|
|
CombineTo(N0.Val, Load, Load.getValue(1));
|
2007-03-26 07:12:51 +00:00
|
|
|
if (ShAmt) {
|
|
|
|
if (Opc == ISD::SIGN_EXTEND_INREG)
|
|
|
|
return DAG.getNode(Opc, VT, Load, N->getOperand(1));
|
|
|
|
else
|
|
|
|
return DAG.getNode(Opc, VT, Load);
|
|
|
|
}
|
2007-03-22 01:54:19 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-05-05 05:58:59 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSIGN_EXTEND_INREG(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2005-09-02 21:18:40 +00:00
|
|
|
SDOperand N1 = N->getOperand(1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
MVT EVT = cast<VTSDNode>(N1)->getVT();
|
|
|
|
unsigned VTBits = VT.getSizeInBits();
|
|
|
|
unsigned EVTBits = EVT.getSizeInBits();
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (sext_in_reg c1) -> c1
|
2006-05-08 20:59:41 +00:00
|
|
|
if (isa<ConstantSDNode>(N0) || N0.getOpcode() == ISD::UNDEF)
|
2006-05-06 23:06:26 +00:00
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND_INREG, VT, N0, N1);
|
Use the new TargetLowering::ComputeNumSignBits method to eliminate
sign_extend_inreg operations. Though ComputeNumSignBits is still rudimentary,
this is enough to compile this:
short test(short X, short x) {
int Y = X+x;
return (Y >> 1);
}
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
into:
_test:
add r2, r3, r4
srawi r3, r2, 1
blr
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test:
add r2, r3, r4
srawi r2, r2, 1
extsh r3, r2
blr
_test2:
add r2, r3, r4
extsh r2, r2
srawi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28146 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-06 09:30:03 +00:00
|
|
|
|
2006-05-06 22:43:44 +00:00
|
|
|
// If the input is already sign extended, just drop the extension.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (DAG.ComputeNumSignBits(N0) >= VT.getSizeInBits()-EVTBits+1)
|
Use the new TargetLowering::ComputeNumSignBits method to eliminate
sign_extend_inreg operations. Though ComputeNumSignBits is still rudimentary,
this is enough to compile this:
short test(short X, short x) {
int Y = X+x;
return (Y >> 1);
}
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
into:
_test:
add r2, r3, r4
srawi r3, r2, 1
blr
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test:
add r2, r3, r4
srawi r2, r2, 1
extsh r3, r2
blr
_test2:
add r2, r3, r4
extsh r2, r2
srawi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28146 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-06 09:30:03 +00:00
|
|
|
return N0;
|
|
|
|
|
2005-09-02 21:18:40 +00:00
|
|
|
// fold (sext_in_reg (sext_in_reg x, VT2), VT1) -> (sext_in_reg x, minVT) pt2
|
|
|
|
if (N0.getOpcode() == ISD::SIGN_EXTEND_INREG &&
|
2008-06-08 20:54:56 +00:00
|
|
|
EVT.bitsLT(cast<VTSDNode>(N0.getOperand(1))->getVT())) {
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::SIGN_EXTEND_INREG, VT, N0.getOperand(0), N1);
|
2005-09-02 21:18:40 +00:00
|
|
|
}
|
Make the case I just checked in stronger. Now we compile this:
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
to:
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test2:
add r2, r3, r4
extsh r2, r2
srwi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28175 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-08 21:18:59 +00:00
|
|
|
|
SIGN_EXTEND_INREG does not demand its top bits. Give SimplifyDemandedBits
a chance to hack on it. This compiles:
int baz(long long a) { return (short)(((int)(a >>24)) >> 9); }
into:
_baz:
slwi r2, r3, 8
srwi r2, r2, 9
extsh r3, r2
blr
instead of:
_baz:
srwi r2, r4, 24
rlwimi r2, r3, 8, 0, 23
srwi r2, r2, 9
extsh r3, r2
blr
This implements CodeGen/PowerPC/sign_ext_inreg1.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@36212 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-17 19:03:21 +00:00
|
|
|
// fold (sext_in_reg x) -> (zext_in_reg x) if the sign bit is known zero.
|
2008-02-25 21:11:39 +00:00
|
|
|
if (DAG.MaskedValueIsZero(N0, APInt::getBitsSet(VTBits, EVTBits-1, EVTBits)))
|
Make the case I just checked in stronger. Now we compile this:
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
to:
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test2:
add r2, r3, r4
extsh r2, r2
srwi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28175 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-08 21:18:59 +00:00
|
|
|
return DAG.getZeroExtendInReg(N0, EVT);
|
2006-05-06 22:43:44 +00:00
|
|
|
|
SIGN_EXTEND_INREG does not demand its top bits. Give SimplifyDemandedBits
a chance to hack on it. This compiles:
int baz(long long a) { return (short)(((int)(a >>24)) >> 9); }
into:
_baz:
slwi r2, r3, 8
srwi r2, r2, 9
extsh r3, r2
blr
instead of:
_baz:
srwi r2, r4, 24
rlwimi r2, r3, 8, 0, 23
srwi r2, r2, 9
extsh r3, r2
blr
This implements CodeGen/PowerPC/sign_ext_inreg1.ll
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@36212 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-17 19:03:21 +00:00
|
|
|
// fold operands of sext_in_reg based on knowledge that the top bits are not
|
|
|
|
// demanded.
|
|
|
|
if (SimplifyDemandedBits(SDOperand(N, 0)))
|
|
|
|
return SDOperand(N, 0);
|
|
|
|
|
2007-03-22 01:54:19 +00:00
|
|
|
// fold (sext_in_reg (load x)) -> (smaller sextload x)
|
|
|
|
// fold (sext_in_reg (srl (load x), c)) -> (smaller sextload (x+c/evtbits))
|
|
|
|
SDOperand NarrowLoad = ReduceLoadWidth(N);
|
|
|
|
if (NarrowLoad.Val)
|
|
|
|
return NarrowLoad;
|
|
|
|
|
2006-05-08 20:59:41 +00:00
|
|
|
// fold (sext_in_reg (srl X, 24), i8) -> sra X, 24
|
Make the case I just checked in stronger. Now we compile this:
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
to:
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test2:
add r2, r3, r4
extsh r2, r2
srwi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28175 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-08 21:18:59 +00:00
|
|
|
// fold (sext_in_reg (srl X, 23), i8) -> sra X, 23 iff possible.
|
|
|
|
// We already fold "(sext_in_reg (srl X, 25), i8) -> srl X, 25" above.
|
2006-05-08 20:59:41 +00:00
|
|
|
if (N0.getOpcode() == ISD::SRL) {
|
|
|
|
if (ConstantSDNode *ShAmt = dyn_cast<ConstantSDNode>(N0.getOperand(1)))
|
2008-06-06 12:08:01 +00:00
|
|
|
if (ShAmt->getValue()+EVTBits <= VT.getSizeInBits()) {
|
Make the case I just checked in stronger. Now we compile this:
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
to:
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test2:
add r2, r3, r4
extsh r2, r2
srwi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28175 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-08 21:18:59 +00:00
|
|
|
// We can turn this into an SRA iff the input to the SRL is already sign
|
|
|
|
// extended enough.
|
2007-06-22 14:59:07 +00:00
|
|
|
unsigned InSignBits = DAG.ComputeNumSignBits(N0.getOperand(0));
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.getSizeInBits()-(ShAmt->getValue()+EVTBits) < InSignBits)
|
Make the case I just checked in stronger. Now we compile this:
short test2(short X, short x) {
int Y = (short)(X+x);
return Y >> 1;
}
to:
_test2:
add r2, r3, r4
extsh r2, r2
srawi r3, r2, 1
blr
instead of:
_test2:
add r2, r3, r4
extsh r2, r2
srwi r2, r2, 1
extsh r3, r2
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@28175 91177308-0d34-0410-b5e6-96231b3b80d8
2006-05-08 21:18:59 +00:00
|
|
|
return DAG.getNode(ISD::SRA, VT, N0.getOperand(0), N0.getOperand(1));
|
|
|
|
}
|
2006-05-08 20:59:41 +00:00
|
|
|
}
|
2007-03-22 01:54:19 +00:00
|
|
|
|
2005-10-13 03:11:28 +00:00
|
|
|
// fold (sext_inreg (extload x)) -> (sextload x)
|
2006-10-04 00:56:09 +00:00
|
|
|
if (ISD::isEXTLoad(N0.Val) &&
|
2007-03-07 08:07:03 +00:00
|
|
|
ISD::isUNINDEXEDLoad(N0.Val) &&
|
2008-01-30 00:15:11 +00:00
|
|
|
EVT == cast<LoadSDNode>(N0)->getMemoryVT() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::SEXTLOAD, EVT))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::SEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2005-12-14 19:25:30 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
2005-10-13 18:34:58 +00:00
|
|
|
CombineTo(N0.Val, ExtLoad, ExtLoad.getValue(1));
|
2006-04-20 23:55:59 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
|
|
|
// fold (sext_inreg (zextload x)) -> (sextload x) iff load has one use
|
2007-03-07 08:07:03 +00:00
|
|
|
if (ISD::isZEXTLoad(N0.Val) && ISD::isUNINDEXEDLoad(N0.Val) &&
|
|
|
|
N0.hasOneUse() &&
|
2008-01-30 00:15:11 +00:00
|
|
|
EVT == cast<LoadSDNode>(N0)->getMemoryVT() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::SEXTLOAD, EVT))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::SEXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LN0->getSrcValueOffset(), EVT,
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2005-12-14 19:25:30 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
2005-10-13 18:34:58 +00:00
|
|
|
CombineTo(N0.Val, ExtLoad, ExtLoad.getValue(1));
|
2006-04-20 23:55:59 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
2005-10-13 03:11:28 +00:00
|
|
|
}
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitTRUNCATE(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// noop truncate
|
|
|
|
if (N0.getValueType() == N->getValueType(0))
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0;
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (truncate c1) -> c1
|
2006-05-06 23:06:26 +00:00
|
|
|
if (isa<ConstantSDNode>(N0))
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (truncate (truncate x)) -> (truncate x)
|
|
|
|
if (N0.getOpcode() == ISD::TRUNCATE)
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, N0.getOperand(0));
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (truncate (ext x)) -> (ext x) or (truncate x) or x
|
2006-05-05 22:56:26 +00:00
|
|
|
if (N0.getOpcode() == ISD::ZERO_EXTEND || N0.getOpcode() == ISD::SIGN_EXTEND||
|
|
|
|
N0.getOpcode() == ISD::ANY_EXTEND) {
|
2008-06-08 20:54:56 +00:00
|
|
|
if (N0.getOperand(0).getValueType().bitsLT(VT))
|
2005-09-01 00:19:25 +00:00
|
|
|
// if the source is smaller than the dest, we still need an extend
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(N0.getOpcode(), VT, N0.getOperand(0));
|
2008-06-08 20:54:56 +00:00
|
|
|
else if (N0.getOperand(0).getValueType().bitsGT(VT))
|
2005-09-01 00:19:25 +00:00
|
|
|
// if the source is larger than the dest, than we just need the truncate
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, N0.getOperand(0));
|
2005-09-01 00:19:25 +00:00
|
|
|
else
|
|
|
|
// if the source and dest are the same type, we can drop both the extend
|
|
|
|
// and the truncate
|
2005-09-06 04:43:02 +00:00
|
|
|
return N0.getOperand(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2007-03-21 20:14:05 +00:00
|
|
|
|
2007-10-13 06:35:54 +00:00
|
|
|
// See if we can simplify the input to this truncate through knowledge that
|
|
|
|
// only the low bits are being used. For example "trunc (or (shl x, 8), y)"
|
|
|
|
// -> trunc y
|
2008-02-25 21:11:39 +00:00
|
|
|
SDOperand Shorter =
|
|
|
|
GetDemandedBits(N0, APInt::getLowBitsSet(N0.getValueSizeInBits(),
|
2008-06-06 12:08:01 +00:00
|
|
|
VT.getSizeInBits()));
|
2007-10-13 06:35:54 +00:00
|
|
|
if (Shorter.Val)
|
|
|
|
return DAG.getNode(ISD::TRUNCATE, VT, Shorter);
|
|
|
|
|
More cool stuff for the dag combiner. We can now finally handle things
like turning:
_foo:
fctiwz f0, f1
stfd f0, -8(r1)
lwz r2, -4(r1)
rlwinm r3, r2, 0, 16, 31
blr
into
_foo:
fctiwz f0,f1
stfd f0,-8(r1)
lhz r3,-2(r1)
blr
Also removed an unncessary constraint from sra -> srl conversion, which
should take care of hte only reason we would ever need to handle sra in
MaskedValueIsZero, AFAIK.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23703 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-12 20:40:40 +00:00
|
|
|
// fold (truncate (load x)) -> (smaller load x)
|
2007-03-21 20:14:05 +00:00
|
|
|
// fold (truncate (srl (load x), c)) -> (smaller load (x+c/evtbits))
|
2007-03-22 01:54:19 +00:00
|
|
|
return ReduceLoadWidth(N);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2008-05-12 23:04:07 +00:00
|
|
|
static SDNode *getBuildPairElt(SDNode *N, unsigned i) {
|
|
|
|
SDOperand Elt = N->getOperand(i);
|
|
|
|
if (Elt.getOpcode() != ISD::MERGE_VALUES)
|
|
|
|
return Elt.Val;
|
|
|
|
return Elt.getOperand(Elt.ResNo).Val;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// CombineConsecutiveLoads - build_pair (load, load) -> load
|
|
|
|
/// if load locations are consecutive.
|
2008-06-06 12:08:01 +00:00
|
|
|
SDOperand DAGCombiner::CombineConsecutiveLoads(SDNode *N, MVT VT) {
|
2008-05-12 23:04:07 +00:00
|
|
|
assert(N->getOpcode() == ISD::BUILD_PAIR);
|
|
|
|
|
|
|
|
SDNode *LD1 = getBuildPairElt(N, 0);
|
|
|
|
if (!ISD::isNON_EXTLoad(LD1) || !LD1->hasOneUse())
|
|
|
|
return SDOperand();
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT LD1VT = LD1->getValueType(0);
|
2008-05-12 23:04:07 +00:00
|
|
|
SDNode *LD2 = getBuildPairElt(N, 1);
|
|
|
|
const MachineFrameInfo *MFI = DAG.getMachineFunction().getFrameInfo();
|
|
|
|
if (ISD::isNON_EXTLoad(LD2) &&
|
|
|
|
LD2->hasOneUse() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// If both are volatile this would reduce the number of volatile loads.
|
|
|
|
// If one is volatile it might be ok, but play conservative and bail out.
|
|
|
|
!cast<LoadSDNode>(LD1)->isVolatile() &&
|
|
|
|
!cast<LoadSDNode>(LD2)->isVolatile() &&
|
2008-06-06 12:08:01 +00:00
|
|
|
TLI.isConsecutiveLoad(LD2, LD1, LD1VT.getSizeInBits()/8, 1, MFI)) {
|
2008-05-12 23:04:07 +00:00
|
|
|
LoadSDNode *LD = cast<LoadSDNode>(LD1);
|
|
|
|
unsigned Align = LD->getAlignment();
|
|
|
|
unsigned NewAlign = TLI.getTargetMachine().getTargetData()->
|
2008-06-06 12:08:01 +00:00
|
|
|
getABITypeAlignment(VT.getTypeForMVT());
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if (NewAlign <= Align &&
|
|
|
|
(!AfterLegalize || TLI.isOperationLegal(ISD::LOAD, VT)))
|
2008-05-12 23:04:07 +00:00
|
|
|
return DAG.getLoad(VT, LD->getChain(), LD->getBasePtr(),
|
|
|
|
LD->getSrcValue(), LD->getSrcValueOffset(),
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
false, Align);
|
2008-05-12 23:04:07 +00:00
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
SDOperand DAGCombiner::visitBIT_CONVERT(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If the input is a BUILD_VECTOR with all constant elements, fold this now.
|
|
|
|
// Only do this before legalize, since afterward the target may be depending
|
|
|
|
// on the bitconvert.
|
|
|
|
// First check to see if this is all constant.
|
|
|
|
if (!AfterLegalize &&
|
|
|
|
N0.getOpcode() == ISD::BUILD_VECTOR && N0.Val->hasOneUse() &&
|
2008-06-06 12:08:01 +00:00
|
|
|
VT.isVector()) {
|
2007-06-25 16:23:39 +00:00
|
|
|
bool isSimple = true;
|
|
|
|
for (unsigned i = 0, e = N0.getNumOperands(); i != e; ++i)
|
|
|
|
if (N0.getOperand(i).getOpcode() != ISD::UNDEF &&
|
|
|
|
N0.getOperand(i).getOpcode() != ISD::Constant &&
|
|
|
|
N0.getOperand(i).getOpcode() != ISD::ConstantFP) {
|
|
|
|
isSimple = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT DestEltVT = N->getValueType(0).getVectorElementType();
|
|
|
|
assert(!DestEltVT.isVector() &&
|
2007-06-25 16:23:39 +00:00
|
|
|
"Element type of vector ValueType must not be vector!");
|
|
|
|
if (isSimple) {
|
|
|
|
return ConstantFoldBIT_CONVERTofBUILD_VECTOR(N0.Val, DestEltVT);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
// If the input is a constant, let getNode() fold it.
|
|
|
|
if (isa<ConstantSDNode>(N0) || isa<ConstantFPSDNode>(N0)) {
|
|
|
|
SDOperand Res = DAG.getNode(ISD::BIT_CONVERT, VT, N0);
|
|
|
|
if (Res.Val != N) return Res;
|
|
|
|
}
|
|
|
|
|
Fold bitconv(bitconv(x)) -> x. We now compile this:
void foo(double);
void bar(double X) { foo(X); }
to this:
bar:
save -96, %o6, %o6
or %g0, %i0, %o0
or %g0, %i1, %o1
call foo
nop
restore %g0, %g0, %g0
retl
nop
instead of this:
bar:
save -112, %o6, %o6
st %i1, [%i6+-4]
st %i0, [%i6+-8]
ldd [%i6+-8], %f0
std %f0, [%i6+-16]
ld [%i6+-12], %o1
ld [%i6+-16], %o0
call foo
nop
restore %g0, %g0, %g0
retl
nop
on V8.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24981 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:37:50 +00:00
|
|
|
if (N0.getOpcode() == ISD::BIT_CONVERT) // conv(conv(x,t1),t2) -> conv(x,t2)
|
|
|
|
return DAG.getNode(ISD::BIT_CONVERT, VT, N0.getOperand(0));
|
2006-04-02 02:53:43 +00:00
|
|
|
|
fold (conv (load x)) -> (load (conv*)x).
This allows us to compile this:
void foo(double);
void bar(double *X) { foo(*X); }
To this:
bar:
save -96, %o6, %o6
ld [%i0+4], %o1
ld [%i0], %o0
call foo
nop
restore %g0, %g0, %g0
retl
nop
instead of this:
bar:
save -104, %o6, %o6
ldd [%i0], %f0
std %f0, [%i6+-8]
ld [%i6+-4], %o1
ld [%i6+-8], %o0
call foo
nop
restore %g0, %g0, %g0
retl
nop
on SparcV8.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24982 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:44:41 +00:00
|
|
|
// fold (conv (load x)) -> (load (conv*)x)
|
2007-10-06 08:19:55 +00:00
|
|
|
// If the resultant load doesn't need a higher alignment than the original!
|
|
|
|
if (ISD::isNormalLoad(N0.Val) && N0.hasOneUse() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// Do not change the width of a volatile load.
|
|
|
|
!cast<LoadSDNode>(N0)->isVolatile() &&
|
|
|
|
(!AfterLegalize || TLI.isOperationLegal(ISD::LOAD, VT))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
2007-05-07 21:27:48 +00:00
|
|
|
unsigned Align = TLI.getTargetMachine().getTargetData()->
|
2008-06-06 12:08:01 +00:00
|
|
|
getABITypeAlignment(VT.getTypeForMVT());
|
2007-05-07 21:27:48 +00:00
|
|
|
unsigned OrigAlign = LN0->getAlignment();
|
|
|
|
if (Align <= OrigAlign) {
|
|
|
|
SDOperand Load = DAG.getLoad(VT, LN0->getChain(), LN0->getBasePtr(),
|
|
|
|
LN0->getSrcValue(), LN0->getSrcValueOffset(),
|
2008-06-28 00:45:22 +00:00
|
|
|
LN0->isVolatile(), OrigAlign);
|
2007-05-07 21:27:48 +00:00
|
|
|
AddToWorkList(N);
|
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::BIT_CONVERT, N0.getValueType(), Load),
|
|
|
|
Load.getValue(1));
|
|
|
|
return Load;
|
|
|
|
}
|
fold (conv (load x)) -> (load (conv*)x).
This allows us to compile this:
void foo(double);
void bar(double *X) { foo(*X); }
To this:
bar:
save -96, %o6, %o6
ld [%i0+4], %o1
ld [%i0], %o0
call foo
nop
restore %g0, %g0, %g0
retl
nop
instead of this:
bar:
save -104, %o6, %o6
ldd [%i0], %f0
std %f0, [%i6+-8]
ld [%i6+-4], %o1
ld [%i6+-8], %o0
call foo
nop
restore %g0, %g0, %g0
retl
nop
on SparcV8.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24982 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:44:41 +00:00
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
// Fold bitconvert(fneg(x)) -> xor(bitconvert(x), signbit)
|
|
|
|
// Fold bitconvert(fabs(x)) -> and(bitconvert(x), ~signbit)
|
|
|
|
// This often reduces constant pool loads.
|
|
|
|
if ((N0.getOpcode() == ISD::FNEG || N0.getOpcode() == ISD::FABS) &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N0.Val->hasOneUse() && VT.isInteger() && !VT.isVector()) {
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
SDOperand NewConv = DAG.getNode(ISD::BIT_CONVERT, VT, N0.getOperand(0));
|
|
|
|
AddToWorkList(NewConv.Val);
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt SignBit = APInt::getSignBit(VT.getSizeInBits());
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
if (N0.getOpcode() == ISD::FNEG)
|
|
|
|
return DAG.getNode(ISD::XOR, VT, NewConv, DAG.getConstant(SignBit, VT));
|
|
|
|
assert(N0.getOpcode() == ISD::FABS);
|
|
|
|
return DAG.getNode(ISD::AND, VT, NewConv, DAG.getConstant(~SignBit, VT));
|
|
|
|
}
|
|
|
|
|
|
|
|
// Fold bitconvert(fcopysign(cst, x)) -> bitconvert(x)&sign | cst&~sign'
|
|
|
|
// Note that we don't handle copysign(x,cst) because this can always be folded
|
|
|
|
// to an fneg or fabs.
|
|
|
|
if (N0.getOpcode() == ISD::FCOPYSIGN && N0.Val->hasOneUse() &&
|
2008-01-27 23:32:17 +00:00
|
|
|
isa<ConstantFPSDNode>(N0.getOperand(0)) &&
|
2008-06-06 12:08:01 +00:00
|
|
|
VT.isInteger() && !VT.isVector()) {
|
|
|
|
unsigned OrigXWidth = N0.getOperand(1).getValueType().getSizeInBits();
|
|
|
|
SDOperand X = DAG.getNode(ISD::BIT_CONVERT,
|
|
|
|
MVT::getIntegerVT(OrigXWidth),
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
N0.getOperand(1));
|
|
|
|
AddToWorkList(X.Val);
|
|
|
|
|
|
|
|
// If X has a different width than the result/lhs, sext it or truncate it.
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned VTWidth = VT.getSizeInBits();
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
if (OrigXWidth < VTWidth) {
|
|
|
|
X = DAG.getNode(ISD::SIGN_EXTEND, VT, X);
|
|
|
|
AddToWorkList(X.Val);
|
|
|
|
} else if (OrigXWidth > VTWidth) {
|
|
|
|
// To get the sign bit in the right place, we have to shift it right
|
|
|
|
// before truncating.
|
|
|
|
X = DAG.getNode(ISD::SRL, X.getValueType(), X,
|
|
|
|
DAG.getConstant(OrigXWidth-VTWidth, X.getValueType()));
|
|
|
|
AddToWorkList(X.Val);
|
|
|
|
X = DAG.getNode(ISD::TRUNCATE, VT, X);
|
|
|
|
AddToWorkList(X.Val);
|
|
|
|
}
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
APInt SignBit = APInt::getSignBit(VT.getSizeInBits());
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
X = DAG.getNode(ISD::AND, VT, X, DAG.getConstant(SignBit, VT));
|
|
|
|
AddToWorkList(X.Val);
|
|
|
|
|
|
|
|
SDOperand Cst = DAG.getNode(ISD::BIT_CONVERT, VT, N0.getOperand(0));
|
|
|
|
Cst = DAG.getNode(ISD::AND, VT, Cst, DAG.getConstant(~SignBit, VT));
|
|
|
|
AddToWorkList(Cst.Val);
|
|
|
|
|
|
|
|
return DAG.getNode(ISD::OR, VT, X, Cst);
|
|
|
|
}
|
2008-05-12 23:04:07 +00:00
|
|
|
|
|
|
|
// bitconvert(build_pair(ld, ld)) -> ld iff load locations are consecutive.
|
|
|
|
if (N0.getOpcode() == ISD::BUILD_PAIR) {
|
|
|
|
SDOperand CombineLD = CombineConsecutiveLoads(N0.Val, VT);
|
|
|
|
if (CombineLD.Val)
|
|
|
|
return CombineLD;
|
|
|
|
}
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
|
constant fold bits_convert in getNode and in the dag combiner for fp<->int
conversions. This allows V8 to compiles this:
void %test() {
call float %test2( float 1.000000e+00, float 2.000000e+00, double 3.000000e+00, double* null )
ret void
}
into:
test:
save -96, %o6, %o6
sethi 0, %o3
sethi 1049088, %o2
sethi 1048576, %o1
sethi 1040384, %o0
or %g0, %o3, %o4
call test2
nop
restore %g0, %g0, %g0
retl
nop
instead of:
test:
save -112, %o6, %o6
sethi 0, %o4
sethi 1049088, %l0
st %o4, [%i6+-12]
st %l0, [%i6+-16]
ld [%i6+-12], %o3
ld [%i6+-16], %o2
sethi 1048576, %o1
sethi 1040384, %o0
call test2
nop
restore %g0, %g0, %g0
retl
nop
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24980 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:30:37 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2008-05-12 23:04:07 +00:00
|
|
|
SDOperand DAGCombiner::visitBUILD_PAIR(SDNode *N) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2008-05-12 23:04:07 +00:00
|
|
|
return CombineConsecutiveLoads(N, VT);
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
/// ConstantFoldBIT_CONVERTofBUILD_VECTOR - We know that BV is a build_vector
|
2006-04-02 02:53:43 +00:00
|
|
|
/// node with Constant, ConstantFP or Undef operands. DstEltVT indicates the
|
|
|
|
/// destination element value type.
|
|
|
|
SDOperand DAGCombiner::
|
2008-06-06 12:08:01 +00:00
|
|
|
ConstantFoldBIT_CONVERTofBUILD_VECTOR(SDNode *BV, MVT DstEltVT) {
|
|
|
|
MVT SrcEltVT = BV->getOperand(0).getValueType();
|
2006-04-02 02:53:43 +00:00
|
|
|
|
|
|
|
// If this is already the right type, we're done.
|
|
|
|
if (SrcEltVT == DstEltVT) return SDOperand(BV, 0);
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned SrcBitSize = SrcEltVT.getSizeInBits();
|
|
|
|
unsigned DstBitSize = DstEltVT.getSizeInBits();
|
2006-04-02 02:53:43 +00:00
|
|
|
|
|
|
|
// If this is a conversion of N elements of one type to N elements of another
|
|
|
|
// type, convert each element. This handles FP<->INT cases.
|
|
|
|
if (SrcBitSize == DstBitSize) {
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> Ops;
|
2007-06-25 16:23:39 +00:00
|
|
|
for (unsigned i = 0, e = BV->getNumOperands(); i != e; ++i) {
|
2006-04-02 02:53:43 +00:00
|
|
|
Ops.push_back(DAG.getNode(ISD::BIT_CONVERT, DstEltVT, BV->getOperand(i)));
|
2006-04-08 04:15:24 +00:00
|
|
|
AddToWorkList(Ops.back().Val);
|
|
|
|
}
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = MVT::getVectorVT(DstEltVT,
|
|
|
|
BV->getValueType(0).getVectorNumElements());
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, VT, &Ops[0], Ops.size());
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we're growing or shrinking the elements. To avoid having to
|
|
|
|
// handle annoying details of growing/shrinking FP values, we convert them to
|
|
|
|
// int first.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (SrcEltVT.isFloatingPoint()) {
|
2006-04-02 02:53:43 +00:00
|
|
|
// Convert the input float vector to a int vector where the elements are the
|
|
|
|
// same sizes.
|
|
|
|
assert((SrcEltVT == MVT::f32 || SrcEltVT == MVT::f64) && "Unknown FP VT!");
|
2008-06-09 11:32:28 +00:00
|
|
|
MVT IntVT = MVT::getIntegerVT(SrcEltVT.getSizeInBits());
|
2007-06-25 16:23:39 +00:00
|
|
|
BV = ConstantFoldBIT_CONVERTofBUILD_VECTOR(BV, IntVT).Val;
|
2006-04-02 02:53:43 +00:00
|
|
|
SrcEltVT = IntVT;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now we know the input is an integer vector. If the output is a FP type,
|
|
|
|
// convert to integer first, then to FP of the right size.
|
2008-06-06 12:08:01 +00:00
|
|
|
if (DstEltVT.isFloatingPoint()) {
|
2006-04-02 02:53:43 +00:00
|
|
|
assert((DstEltVT == MVT::f32 || DstEltVT == MVT::f64) && "Unknown FP VT!");
|
2008-06-09 11:32:28 +00:00
|
|
|
MVT TmpVT = MVT::getIntegerVT(DstEltVT.getSizeInBits());
|
2007-06-25 16:23:39 +00:00
|
|
|
SDNode *Tmp = ConstantFoldBIT_CONVERTofBUILD_VECTOR(BV, TmpVT).Val;
|
2006-04-02 02:53:43 +00:00
|
|
|
|
|
|
|
// Next, convert to FP elements of the same size.
|
2007-06-25 16:23:39 +00:00
|
|
|
return ConstantFoldBIT_CONVERTofBUILD_VECTOR(Tmp, DstEltVT);
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Okay, we know the src/dst types are both integers of differing types.
|
|
|
|
// Handling growing first.
|
2008-06-06 12:08:01 +00:00
|
|
|
assert(SrcEltVT.isInteger() && DstEltVT.isInteger());
|
2006-04-02 02:53:43 +00:00
|
|
|
if (SrcBitSize < DstBitSize) {
|
|
|
|
unsigned NumInputsPerOutput = DstBitSize/SrcBitSize;
|
|
|
|
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> Ops;
|
2007-06-25 16:23:39 +00:00
|
|
|
for (unsigned i = 0, e = BV->getNumOperands(); i != e;
|
2006-04-02 02:53:43 +00:00
|
|
|
i += NumInputsPerOutput) {
|
|
|
|
bool isLE = TLI.isLittleEndian();
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt NewBits = APInt(DstBitSize, 0);
|
2006-04-02 02:53:43 +00:00
|
|
|
bool EltIsUndef = true;
|
|
|
|
for (unsigned j = 0; j != NumInputsPerOutput; ++j) {
|
|
|
|
// Shift the previously computed bits over.
|
|
|
|
NewBits <<= SrcBitSize;
|
|
|
|
SDOperand Op = BV->getOperand(i+ (isLE ? (NumInputsPerOutput-j-1) : j));
|
|
|
|
if (Op.getOpcode() == ISD::UNDEF) continue;
|
|
|
|
EltIsUndef = false;
|
|
|
|
|
2008-03-03 23:51:38 +00:00
|
|
|
NewBits |=
|
|
|
|
APInt(cast<ConstantSDNode>(Op)->getAPIntValue()).zext(DstBitSize);
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (EltIsUndef)
|
|
|
|
Ops.push_back(DAG.getNode(ISD::UNDEF, DstEltVT));
|
|
|
|
else
|
|
|
|
Ops.push_back(DAG.getConstant(NewBits, DstEltVT));
|
|
|
|
}
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = MVT::getVectorVT(DstEltVT, Ops.size());
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, VT, &Ops[0], Ops.size());
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, this must be the case where we are shrinking elements: each input
|
|
|
|
// turns into multiple outputs.
|
- When DAG combiner is folding a bit convert into a BUILD_VECTOR, it should check if it's essentially a SCALAR_TO_VECTOR. Avoid turning (v8i16) <10, u, u, u> to <10, 0, u, u, u, u, u, u>. Instead, simply convert it to a SCALAR_TO_VECTOR of the proper type.
- X86 now normalize SCALAR_TO_VECTOR to (BIT_CONVERT (v4i32 SCALAR_TO_VECTOR)). Get rid of X86ISD::S2VEC.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@47290 91177308-0d34-0410-b5e6-96231b3b80d8
2008-02-18 23:04:32 +00:00
|
|
|
bool isS2V = ISD::isScalarToVector(BV);
|
2006-04-02 02:53:43 +00:00
|
|
|
unsigned NumOutputsPerInput = SrcBitSize/DstBitSize;
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = MVT::getVectorVT(DstEltVT, NumOutputsPerInput*BV->getNumOperands());
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> Ops;
|
2007-06-25 16:23:39 +00:00
|
|
|
for (unsigned i = 0, e = BV->getNumOperands(); i != e; ++i) {
|
2006-04-02 02:53:43 +00:00
|
|
|
if (BV->getOperand(i).getOpcode() == ISD::UNDEF) {
|
|
|
|
for (unsigned j = 0; j != NumOutputsPerInput; ++j)
|
|
|
|
Ops.push_back(DAG.getNode(ISD::UNDEF, DstEltVT));
|
|
|
|
continue;
|
|
|
|
}
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt OpVal = cast<ConstantSDNode>(BV->getOperand(i))->getAPIntValue();
|
2006-04-02 02:53:43 +00:00
|
|
|
for (unsigned j = 0; j != NumOutputsPerInput; ++j) {
|
2008-03-03 23:51:38 +00:00
|
|
|
APInt ThisVal = APInt(OpVal).trunc(DstBitSize);
|
2006-04-02 02:53:43 +00:00
|
|
|
Ops.push_back(DAG.getConstant(ThisVal, DstEltVT));
|
2008-03-03 23:51:38 +00:00
|
|
|
if (isS2V && i == 0 && j == 0 && APInt(ThisVal).zext(SrcBitSize) == OpVal)
|
- When DAG combiner is folding a bit convert into a BUILD_VECTOR, it should check if it's essentially a SCALAR_TO_VECTOR. Avoid turning (v8i16) <10, u, u, u> to <10, 0, u, u, u, u, u, u>. Instead, simply convert it to a SCALAR_TO_VECTOR of the proper type.
- X86 now normalize SCALAR_TO_VECTOR to (BIT_CONVERT (v4i32 SCALAR_TO_VECTOR)). Get rid of X86ISD::S2VEC.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@47290 91177308-0d34-0410-b5e6-96231b3b80d8
2008-02-18 23:04:32 +00:00
|
|
|
// Simply turn this into a SCALAR_TO_VECTOR of the new type.
|
|
|
|
return DAG.getNode(ISD::SCALAR_TO_VECTOR, VT, Ops[0]);
|
2008-03-03 23:51:38 +00:00
|
|
|
OpVal = OpVal.lshr(DstBitSize);
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// For big endian targets, swap the order of the pieces of each element.
|
2008-02-11 10:37:04 +00:00
|
|
|
if (TLI.isBigEndian())
|
2006-04-02 02:53:43 +00:00
|
|
|
std::reverse(Ops.end()-NumOutputsPerInput, Ops.end());
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, VT, &Ops[0], Ops.size());
|
2006-04-02 02:53:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
SDOperand DAGCombiner::visitFADD(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-10-18 00:28:13 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-10-18 00:28:13 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-10-18 00:28:13 +00:00
|
|
|
// fold (fadd c1, c2) -> c1+c2
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FADD, VT, N0, N1);
|
2005-10-18 00:28:13 +00:00
|
|
|
// canonicalize constant to RHS
|
|
|
|
if (N0CFP && !N1CFP)
|
|
|
|
return DAG.getNode(ISD::FADD, VT, N1, N0);
|
2005-09-28 22:28:18 +00:00
|
|
|
// fold (A + (-B)) -> A-B
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(N1, AfterLegalize) == 2)
|
|
|
|
return DAG.getNode(ISD::FSUB, VT, N0,
|
|
|
|
GetNegatedExpression(N1, DAG, AfterLegalize));
|
2005-09-28 22:28:18 +00:00
|
|
|
// fold ((-A) + B) -> B-A
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(N0, AfterLegalize) == 2)
|
|
|
|
return DAG.getNode(ISD::FSUB, VT, N1,
|
|
|
|
GetNegatedExpression(N0, DAG, AfterLegalize));
|
2007-01-08 23:04:05 +00:00
|
|
|
|
|
|
|
// If allowed, fold (fadd (fadd x, c1), c2) -> (fadd x, (fadd c1, c2))
|
|
|
|
if (UnsafeFPMath && N1CFP && N0.getOpcode() == ISD::FADD &&
|
|
|
|
N0.Val->hasOneUse() && isa<ConstantFPSDNode>(N0.getOperand(1)))
|
|
|
|
return DAG.getNode(ISD::FADD, VT, N0.getOperand(0),
|
|
|
|
DAG.getNode(ISD::FADD, VT, N0.getOperand(1), N1));
|
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitFSUB(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-10-18 00:28:13 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-10-18 00:28:13 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-10-18 00:28:13 +00:00
|
|
|
// fold (fsub c1, c2) -> c1-c2
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FSUB, VT, N0, N1);
|
2007-07-02 15:48:56 +00:00
|
|
|
// fold (0-B) -> -B
|
2007-08-31 23:34:27 +00:00
|
|
|
if (UnsafeFPMath && N0CFP && N0CFP->getValueAPF().isZero()) {
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(N1, AfterLegalize))
|
|
|
|
return GetNegatedExpression(N1, DAG, AfterLegalize);
|
2007-07-02 15:48:56 +00:00
|
|
|
return DAG.getNode(ISD::FNEG, VT, N1);
|
|
|
|
}
|
2005-09-28 22:28:18 +00:00
|
|
|
// fold (A-(-B)) -> A+B
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(N1, AfterLegalize))
|
|
|
|
return DAG.getNode(ISD::FADD, VT, N0,
|
|
|
|
GetNegatedExpression(N1, DAG, AfterLegalize));
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitFMUL(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2005-10-17 20:40:11 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-28 22:28:18 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-10-17 20:40:11 +00:00
|
|
|
// fold (fmul c1, c2) -> c1*c2
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FMUL, VT, N0, N1);
|
2005-10-17 20:40:11 +00:00
|
|
|
// canonicalize constant to RHS
|
2005-10-18 00:28:13 +00:00
|
|
|
if (N0CFP && !N1CFP)
|
|
|
|
return DAG.getNode(ISD::FMUL, VT, N1, N0);
|
2005-10-17 20:40:11 +00:00
|
|
|
// fold (fmul X, 2.0) -> (fadd X, X)
|
|
|
|
if (N1CFP && N1CFP->isExactlyValue(+2.0))
|
|
|
|
return DAG.getNode(ISD::FADD, VT, N0, N0);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
// fold (fmul X, -1.0) -> (fneg X)
|
|
|
|
if (N1CFP && N1CFP->isExactlyValue(-1.0))
|
|
|
|
return DAG.getNode(ISD::FNEG, VT, N0);
|
|
|
|
|
|
|
|
// -X * -Y -> X*Y
|
2008-02-26 07:04:54 +00:00
|
|
|
if (char LHSNeg = isNegatibleForFree(N0, AfterLegalize)) {
|
|
|
|
if (char RHSNeg = isNegatibleForFree(N1, AfterLegalize)) {
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
// Both can be negated for free, check to see if at least one is cheaper
|
|
|
|
// negated.
|
|
|
|
if (LHSNeg == 2 || RHSNeg == 2)
|
2008-02-26 07:04:54 +00:00
|
|
|
return DAG.getNode(ISD::FMUL, VT,
|
|
|
|
GetNegatedExpression(N0, DAG, AfterLegalize),
|
|
|
|
GetNegatedExpression(N1, DAG, AfterLegalize));
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
}
|
|
|
|
}
|
2007-01-08 23:04:05 +00:00
|
|
|
|
|
|
|
// If allowed, fold (fmul (fmul x, c1), c2) -> (fmul x, (fmul c1, c2))
|
|
|
|
if (UnsafeFPMath && N1CFP && N0.getOpcode() == ISD::FMUL &&
|
|
|
|
N0.Val->hasOneUse() && isa<ConstantFPSDNode>(N0.getOperand(1)))
|
|
|
|
return DAG.getNode(ISD::FMUL, VT, N0.getOperand(0),
|
|
|
|
DAG.getNode(ISD::FMUL, VT, N0.getOperand(1), N1));
|
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitFDIV(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2006-01-18 22:35:16 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-28 22:28:18 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// fold vector ops
|
2008-06-06 12:08:01 +00:00
|
|
|
if (VT.isVector()) {
|
2007-07-13 20:03:40 +00:00
|
|
|
SDOperand FoldedVOp = SimplifyVBinOp(N);
|
|
|
|
if (FoldedVOp.Val) return FoldedVOp;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2006-01-18 22:35:16 +00:00
|
|
|
// fold (fdiv c1, c2) -> c1/c2
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FDIV, VT, N0, N1);
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
|
|
|
|
|
|
|
|
// -X / -Y -> X*Y
|
2008-02-26 07:04:54 +00:00
|
|
|
if (char LHSNeg = isNegatibleForFree(N0, AfterLegalize)) {
|
|
|
|
if (char RHSNeg = isNegatibleForFree(N1, AfterLegalize)) {
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
// Both can be negated for free, check to see if at least one is cheaper
|
|
|
|
// negated.
|
|
|
|
if (LHSNeg == 2 || RHSNeg == 2)
|
2008-02-26 07:04:54 +00:00
|
|
|
return DAG.getNode(ISD::FDIV, VT,
|
|
|
|
GetNegatedExpression(N0, DAG, AfterLegalize),
|
|
|
|
GetNegatedExpression(N1, DAG, AfterLegalize));
|
implement a simple fneg optimization/propagation thing. This compiles:
CodeGen/PowerPC/fneg.ll into:
_t4:
fmul f0, f3, f4
fmadd f1, f1, f2, f0
blr
instead of:
_t4:
fneg f0, f3
fmul f0, f0, f4
fmsub f1, f1, f2, f0
blr
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@37054 91177308-0d34-0410-b5e6-96231b3b80d8
2007-05-14 22:04:50 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand DAGCombiner::visitFREM(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
2006-01-18 22:35:16 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-28 22:28:18 +00:00
|
|
|
|
2006-01-18 22:35:16 +00:00
|
|
|
// fold (frem c1, c2) -> fmod(c1,c2)
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FREM, VT, N0, N1);
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-03-05 05:30:57 +00:00
|
|
|
SDOperand DAGCombiner::visitFCOPYSIGN(SDNode *N) {
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
|
|
|
ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2006-03-05 05:30:57 +00:00
|
|
|
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N1CFP && VT != MVT::ppcf128) // Constant fold
|
2006-03-05 05:30:57 +00:00
|
|
|
return DAG.getNode(ISD::FCOPYSIGN, VT, N0, N1);
|
|
|
|
|
|
|
|
if (N1CFP) {
|
2007-08-26 01:18:27 +00:00
|
|
|
const APFloat& V = N1CFP->getValueAPF();
|
2006-03-05 05:30:57 +00:00
|
|
|
// copysign(x, c1) -> fabs(x) iff ispos(c1)
|
|
|
|
// copysign(x, c1) -> fneg(fabs(x)) iff isneg(c1)
|
2007-08-25 22:10:57 +00:00
|
|
|
if (!V.isNegative())
|
2006-03-05 05:30:57 +00:00
|
|
|
return DAG.getNode(ISD::FABS, VT, N0);
|
|
|
|
else
|
|
|
|
return DAG.getNode(ISD::FNEG, VT, DAG.getNode(ISD::FABS, VT, N0));
|
|
|
|
}
|
|
|
|
|
|
|
|
// copysign(fabs(x), y) -> copysign(x, y)
|
|
|
|
// copysign(fneg(x), y) -> copysign(x, y)
|
|
|
|
// copysign(copysign(x,z), y) -> copysign(x, y)
|
|
|
|
if (N0.getOpcode() == ISD::FABS || N0.getOpcode() == ISD::FNEG ||
|
|
|
|
N0.getOpcode() == ISD::FCOPYSIGN)
|
|
|
|
return DAG.getNode(ISD::FCOPYSIGN, VT, N0.getOperand(0), N1);
|
|
|
|
|
|
|
|
// copysign(x, abs(y)) -> abs(x)
|
|
|
|
if (N1.getOpcode() == ISD::FABS)
|
|
|
|
return DAG.getNode(ISD::FABS, VT, N0);
|
|
|
|
|
|
|
|
// copysign(x, copysign(y,z)) -> copysign(x, z)
|
|
|
|
if (N1.getOpcode() == ISD::FCOPYSIGN)
|
|
|
|
return DAG.getNode(ISD::FCOPYSIGN, VT, N0, N1.getOperand(1));
|
|
|
|
|
|
|
|
// copysign(x, fp_extend(y)) -> copysign(x, y)
|
|
|
|
// copysign(x, fp_round(y)) -> copysign(x, y)
|
|
|
|
if (N1.getOpcode() == ISD::FP_EXTEND || N1.getOpcode() == ISD::FP_ROUND)
|
|
|
|
return DAG.getNode(ISD::FCOPYSIGN, VT, N0, N1.getOperand(0));
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-09-28 22:28:18 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitSINT_TO_FP(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
MVT OpVT = N0.getValueType();
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (sint_to_fp c1) -> c1fp
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
if (N0C && OpVT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::SINT_TO_FP, VT, N0);
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
|
|
|
|
// If the input is a legal type, and SINT_TO_FP is not legal on this target,
|
|
|
|
// but UINT_TO_FP is legal on this target, try to convert.
|
2008-06-26 17:16:00 +00:00
|
|
|
if (!TLI.isOperationLegal(ISD::SINT_TO_FP, OpVT) &&
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
TLI.isOperationLegal(ISD::UINT_TO_FP, OpVT)) {
|
|
|
|
// If the sign bit is known to be zero, we can change this to UINT_TO_FP.
|
|
|
|
if (DAG.SignBitIsZero(N0))
|
|
|
|
return DAG.getNode(ISD::UINT_TO_FP, VT, N0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitUINT_TO_FP(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantSDNode *N0C = dyn_cast<ConstantSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
MVT OpVT = N0.getValueType();
|
2006-01-18 22:35:16 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (uint_to_fp c1) -> c1fp
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
if (N0C && OpVT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::UINT_TO_FP, VT, N0);
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
|
|
|
|
// If the input is a legal type, and UINT_TO_FP is not legal on this target,
|
|
|
|
// but SINT_TO_FP is legal on this target, try to convert.
|
2008-06-26 17:16:00 +00:00
|
|
|
if (!TLI.isOperationLegal(ISD::UINT_TO_FP, OpVT) &&
|
when we know the signbit of an input to uint_to_fp is zero,
change it to sint_to_fp on targets where that is cheaper (and
visaversa of course). This allows us to compile uint_to_fp to:
_test:
movl 4(%esp), %eax
shrl $23, %eax
cvtsi2ss %eax, %xmm0
movl 8(%esp), %eax
movss %xmm0, (%eax)
ret
instead of:
.align 3
LCPI1_0: ## double
.long 0 ## double least significant word 4.5036e+15
.long 1127219200 ## double most significant word 4.5036e+15
.text
.align 4,0x90
.globl _test
_test:
subl $12, %esp
movl 16(%esp), %eax
shrl $23, %eax
movl %eax, (%esp)
movl $1127219200, 4(%esp)
movsd (%esp), %xmm0
subsd LCPI1_0, %xmm0
cvtsd2ss %xmm0, %xmm0
movl 20(%esp), %eax
movss %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52747 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-26 00:16:49 +00:00
|
|
|
TLI.isOperationLegal(ISD::SINT_TO_FP, OpVT)) {
|
|
|
|
// If the sign bit is known to be zero, we can change this to SINT_TO_FP.
|
|
|
|
if (DAG.SignBitIsZero(N0))
|
|
|
|
return DAG.getNode(ISD::SINT_TO_FP, VT, N0);
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFP_TO_SINT(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (fp_to_sint c1fp) -> c1
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0CFP)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FP_TO_SINT, VT, N0);
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFP_TO_UINT(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (fp_to_uint c1fp) -> c1
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FP_TO_UINT, VT, N0);
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFP_ROUND(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-01-17 07:00:52 +00:00
|
|
|
SDOperand N1 = N->getOperand(1);
|
2006-01-18 22:35:16 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (fp_round c1fp) -> c1fp
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && N0.getValueType() != MVT::ppcf128)
|
2008-01-17 07:00:52 +00:00
|
|
|
return DAG.getNode(ISD::FP_ROUND, VT, N0, N1);
|
2006-03-13 06:26:26 +00:00
|
|
|
|
|
|
|
// fold (fp_round (fp_extend x)) -> x
|
|
|
|
if (N0.getOpcode() == ISD::FP_EXTEND && VT == N0.getOperand(0).getValueType())
|
|
|
|
return N0.getOperand(0);
|
|
|
|
|
2008-01-24 06:45:35 +00:00
|
|
|
// fold (fp_round (fp_round x)) -> (fp_round x)
|
|
|
|
if (N0.getOpcode() == ISD::FP_ROUND) {
|
|
|
|
// This is a value preserving truncation if both round's are.
|
|
|
|
bool IsTrunc = N->getConstantOperandVal(1) == 1 &&
|
|
|
|
N0.Val->getConstantOperandVal(1) == 1;
|
|
|
|
return DAG.getNode(ISD::FP_ROUND, VT, N0.getOperand(0),
|
|
|
|
DAG.getIntPtrConstant(IsTrunc));
|
|
|
|
}
|
|
|
|
|
2006-03-13 06:26:26 +00:00
|
|
|
// fold (fp_round (copysign X, Y)) -> (copysign (fp_round X), Y)
|
|
|
|
if (N0.getOpcode() == ISD::FCOPYSIGN && N0.Val->hasOneUse()) {
|
2008-01-17 07:00:52 +00:00
|
|
|
SDOperand Tmp = DAG.getNode(ISD::FP_ROUND, VT, N0.getOperand(0), N1);
|
2006-03-13 06:26:26 +00:00
|
|
|
AddToWorkList(Tmp.Val);
|
|
|
|
return DAG.getNode(ISD::FCOPYSIGN, VT, Tmp, N0.getOperand(1));
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFP_ROUND_INREG(SDNode *N) {
|
2005-09-01 00:19:25 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
MVT EVT = cast<VTSDNode>(N->getOperand(1))->getVT();
|
2005-09-02 21:18:40 +00:00
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
|
|
|
// fold (fp_round_inreg c1fp) -> c1fp
|
2005-09-02 21:18:40 +00:00
|
|
|
if (N0CFP) {
|
2007-08-31 23:34:27 +00:00
|
|
|
SDOperand Round = DAG.getConstantFP(N0CFP->getValueAPF(), EVT);
|
2005-09-06 04:43:02 +00:00
|
|
|
return DAG.getNode(ISD::FP_EXTEND, VT, Round);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFP_EXTEND(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
|
2007-12-29 06:55:23 +00:00
|
|
|
// If this is fp_round(fpextend), don't fold it, allow ourselves to be folded.
|
2008-04-16 16:15:27 +00:00
|
|
|
if (N->hasOneUse() &&
|
|
|
|
N->use_begin()->getSDOperand().getOpcode() == ISD::FP_ROUND)
|
2007-12-29 06:55:23 +00:00
|
|
|
return SDOperand();
|
2008-01-17 07:00:52 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (fp_extend c1fp) -> c1fp
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FP_EXTEND, VT, N0);
|
2008-01-17 07:00:52 +00:00
|
|
|
|
|
|
|
// Turn fp_extend(fp_round(X, 1)) -> x since the fp_round doesn't affect the
|
|
|
|
// value of X.
|
|
|
|
if (N0.getOpcode() == ISD::FP_ROUND && N0.Val->getConstantOperandVal(1) == 1){
|
|
|
|
SDOperand In = N0.getOperand(0);
|
|
|
|
if (In.getValueType() == VT) return In;
|
2008-06-08 20:54:56 +00:00
|
|
|
if (VT.bitsLT(In.getValueType()))
|
2008-01-17 07:00:52 +00:00
|
|
|
return DAG.getNode(ISD::FP_ROUND, VT, In, N0.getOperand(1));
|
|
|
|
return DAG.getNode(ISD::FP_EXTEND, VT, In);
|
|
|
|
}
|
|
|
|
|
|
|
|
// fold (fpext (load x)) -> (fpext (fptrunc (extload x)))
|
2007-10-19 20:29:00 +00:00
|
|
|
if (ISD::isNON_EXTLoad(N0.Val) && N0.hasOneUse() &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
((!AfterLegalize && !cast<LoadSDNode>(N0)->isVolatile()) ||
|
|
|
|
TLI.isLoadXLegal(ISD::EXTLOAD, N0.getValueType()))) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LN0 = cast<LoadSDNode>(N0);
|
|
|
|
SDOperand ExtLoad = DAG.getExtLoad(ISD::EXTLOAD, VT, LN0->getChain(),
|
|
|
|
LN0->getBasePtr(), LN0->getSrcValue(),
|
|
|
|
LN0->getSrcValueOffset(),
|
2007-04-22 23:15:30 +00:00
|
|
|
N0.getValueType(),
|
|
|
|
LN0->isVolatile(),
|
|
|
|
LN0->getAlignment());
|
2006-05-05 21:34:35 +00:00
|
|
|
CombineTo(N, ExtLoad);
|
2008-01-17 07:00:52 +00:00
|
|
|
CombineTo(N0.Val, DAG.getNode(ISD::FP_ROUND, N0.getValueType(), ExtLoad,
|
|
|
|
DAG.getIntPtrConstant(1)),
|
2006-05-05 21:34:35 +00:00
|
|
|
ExtLoad.getValue(1));
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFNEG(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
|
2008-02-26 07:04:54 +00:00
|
|
|
if (isNegatibleForFree(N0, AfterLegalize))
|
|
|
|
return GetNegatedExpression(N0, DAG, AfterLegalize);
|
2007-07-02 15:48:56 +00:00
|
|
|
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
// Transform fneg(bitconvert(x)) -> bitconvert(x^sign) to avoid loading
|
|
|
|
// constant pool values.
|
2008-01-27 23:32:17 +00:00
|
|
|
if (N0.getOpcode() == ISD::BIT_CONVERT && N0.Val->hasOneUse() &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N0.getOperand(0).getValueType().isInteger() &&
|
|
|
|
!N0.getOperand(0).getValueType().isVector()) {
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
SDOperand Int = N0.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT IntVT = Int.getValueType();
|
|
|
|
if (IntVT.isInteger() && !IntVT.isVector()) {
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
Int = DAG.getNode(ISD::XOR, IntVT, Int,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(IntVT.getIntegerVTSignBit(), IntVT));
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
AddToWorkList(Int.Val);
|
|
|
|
return DAG.getNode(ISD::BIT_CONVERT, N->getValueType(0), Int);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
SDOperand DAGCombiner::visitFABS(SDNode *N) {
|
2006-01-18 22:35:16 +00:00
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
ConstantFPSDNode *N0CFP = dyn_cast<ConstantFPSDNode>(N0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
2006-01-18 22:35:16 +00:00
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (fabs c1) -> fabs(c1)
|
2007-10-16 23:38:29 +00:00
|
|
|
if (N0CFP && VT != MVT::ppcf128)
|
2006-01-18 22:35:16 +00:00
|
|
|
return DAG.getNode(ISD::FABS, VT, N0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (fabs (fabs x)) -> (fabs x)
|
2006-03-05 05:30:57 +00:00
|
|
|
if (N0.getOpcode() == ISD::FABS)
|
2005-09-06 04:43:02 +00:00
|
|
|
return N->getOperand(0);
|
2005-09-01 00:19:25 +00:00
|
|
|
// fold (fabs (fneg x)) -> (fabs x)
|
2006-03-05 05:30:57 +00:00
|
|
|
// fold (fabs (fcopysign x, y)) -> (fabs x)
|
|
|
|
if (N0.getOpcode() == ISD::FNEG || N0.getOpcode() == ISD::FCOPYSIGN)
|
|
|
|
return DAG.getNode(ISD::FABS, VT, N0.getOperand(0));
|
|
|
|
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
// Transform fabs(bitconvert(x)) -> bitconvert(x&~sign) to avoid loading
|
|
|
|
// constant pool values.
|
2008-01-27 23:32:17 +00:00
|
|
|
if (N0.getOpcode() == ISD::BIT_CONVERT && N0.Val->hasOneUse() &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N0.getOperand(0).getValueType().isInteger() &&
|
|
|
|
!N0.getOperand(0).getValueType().isVector()) {
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
SDOperand Int = N0.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT IntVT = Int.getValueType();
|
|
|
|
if (IntVT.isInteger() && !IntVT.isVector()) {
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
Int = DAG.getNode(ISD::AND, IntVT, Int,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(~IntVT.getIntegerVTSignBit(), IntVT));
|
Implement some dag combines that allow doing fneg/fabs/fcopysign in integer
registers if used by a bitconvert or using a bitconvert. This allows us to
avoid constant pool loads and use cheaper integer instructions when the
values come from or end up in integer regs anyway. For example, we now
compile CodeGen/X86/fp-in-intregs.ll to:
_test1:
movl $2147483648, %eax
xorl 4(%esp), %eax
ret
_test2:
movl $1065353216, %eax
orl 4(%esp), %eax
andl $3212836864, %eax
ret
Instead of:
_test1:
movss 4(%esp), %xmm0
xorps LCPI2_0, %xmm0
movd %xmm0, %eax
ret
_test2:
movss 4(%esp), %xmm0
andps LCPI3_0, %xmm0
movss LCPI3_1, %xmm1
andps LCPI3_2, %xmm1
orps %xmm0, %xmm1
movd %xmm1, %eax
ret
bitconverts can happen due to various calling conventions that require
fp values to passed in integer regs in some cases, e.g. when returning
a complex.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46414 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-27 17:42:27 +00:00
|
|
|
AddToWorkList(Int.Val);
|
|
|
|
return DAG.getNode(ISD::BIT_CONVERT, N->getValueType(0), Int);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-09-06 04:43:02 +00:00
|
|
|
return SDOperand();
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand DAGCombiner::visitBRCOND(SDNode *N) {
|
|
|
|
SDOperand Chain = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
SDOperand N2 = N->getOperand(2);
|
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1);
|
|
|
|
|
|
|
|
// never taken branch, fold to chain
|
|
|
|
if (N1C && N1C->isNullValue())
|
|
|
|
return Chain;
|
|
|
|
// unconditional branch
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N1C && N1C->getAPIntValue() == 1)
|
2005-09-19 22:34:01 +00:00
|
|
|
return DAG.getNode(ISD::BR, MVT::Other, Chain, N2);
|
2006-02-01 07:19:44 +00:00
|
|
|
// fold a brcond with a setcc condition into a BR_CC node if BR_CC is legal
|
|
|
|
// on the target.
|
|
|
|
if (N1.getOpcode() == ISD::SETCC &&
|
|
|
|
TLI.isOperationLegal(ISD::BR_CC, MVT::Other)) {
|
|
|
|
return DAG.getNode(ISD::BR_CC, MVT::Other, Chain, N1.getOperand(2),
|
|
|
|
N1.getOperand(0), N1.getOperand(1), N2);
|
|
|
|
}
|
2005-09-19 22:34:01 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2005-10-05 06:47:48 +00:00
|
|
|
// Operand List for BR_CC: Chain, CondCC, CondLHS, CondRHS, DestBB.
|
|
|
|
//
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand DAGCombiner::visitBR_CC(SDNode *N) {
|
2005-10-05 06:47:48 +00:00
|
|
|
CondCodeSDNode *CC = cast<CondCodeSDNode>(N->getOperand(1));
|
|
|
|
SDOperand CondLHS = N->getOperand(2), CondRHS = N->getOperand(3);
|
|
|
|
|
2008-06-09 11:32:28 +00:00
|
|
|
// Use SimplifySetCC to simplify SETCC's.
|
2005-10-05 21:43:42 +00:00
|
|
|
SDOperand Simp = SimplifySetCC(MVT::i1, CondLHS, CondRHS, CC->get(), false);
|
2006-10-14 03:52:46 +00:00
|
|
|
if (Simp.Val) AddToWorkList(Simp.Val);
|
|
|
|
|
2005-10-05 21:43:42 +00:00
|
|
|
ConstantSDNode *SCCC = dyn_cast_or_null<ConstantSDNode>(Simp.Val);
|
|
|
|
|
|
|
|
// fold br_cc true, dest -> br dest (unconditional branch)
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SCCC && !SCCC->isNullValue())
|
2005-10-05 21:43:42 +00:00
|
|
|
return DAG.getNode(ISD::BR, MVT::Other, N->getOperand(0),
|
|
|
|
N->getOperand(4));
|
|
|
|
// fold br_cc false, dest -> unconditional fall through
|
|
|
|
if (SCCC && SCCC->isNullValue())
|
|
|
|
return N->getOperand(0);
|
2006-10-14 03:52:46 +00:00
|
|
|
|
2005-10-05 21:43:42 +00:00
|
|
|
// fold to a simpler setcc
|
|
|
|
if (Simp.Val && Simp.getOpcode() == ISD::SETCC)
|
|
|
|
return DAG.getNode(ISD::BR_CC, MVT::Other, N->getOperand(0),
|
|
|
|
Simp.getOperand(2), Simp.getOperand(0),
|
|
|
|
Simp.getOperand(1), N->getOperand(4));
|
2005-09-19 22:34:01 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2008-06-15 20:12:31 +00:00
|
|
|
/// CombineToPreIndexedLoadStore - Try turning a load / store into a
|
|
|
|
/// pre-indexed load / store when the base pointer is an add or subtract
|
2006-11-11 00:39:41 +00:00
|
|
|
/// and it has other uses besides the load / store. After the
|
|
|
|
/// transformation, the new indexed load / store has effectively folded
|
|
|
|
/// the add / subtract in and all of its other uses are redirected to the
|
|
|
|
/// new load / store.
|
|
|
|
bool DAGCombiner::CombineToPreIndexedLoadStore(SDNode *N) {
|
|
|
|
if (!AfterLegalize)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
bool isLoad = true;
|
|
|
|
SDOperand Ptr;
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT;
|
2006-11-11 00:39:41 +00:00
|
|
|
if (LoadSDNode *LD = dyn_cast<LoadSDNode>(N)) {
|
2008-01-17 19:59:44 +00:00
|
|
|
if (LD->isIndexed())
|
2006-12-16 06:25:23 +00:00
|
|
|
return false;
|
2008-01-30 00:15:11 +00:00
|
|
|
VT = LD->getMemoryVT();
|
2007-03-07 08:07:03 +00:00
|
|
|
if (!TLI.isIndexedLoadLegal(ISD::PRE_INC, VT) &&
|
2006-11-11 00:39:41 +00:00
|
|
|
!TLI.isIndexedLoadLegal(ISD::PRE_DEC, VT))
|
|
|
|
return false;
|
|
|
|
Ptr = LD->getBasePtr();
|
|
|
|
} else if (StoreSDNode *ST = dyn_cast<StoreSDNode>(N)) {
|
2008-01-17 19:59:44 +00:00
|
|
|
if (ST->isIndexed())
|
2006-12-16 06:25:23 +00:00
|
|
|
return false;
|
2008-01-30 00:15:11 +00:00
|
|
|
VT = ST->getMemoryVT();
|
2006-11-11 00:39:41 +00:00
|
|
|
if (!TLI.isIndexedStoreLegal(ISD::PRE_INC, VT) &&
|
|
|
|
!TLI.isIndexedStoreLegal(ISD::PRE_DEC, VT))
|
|
|
|
return false;
|
|
|
|
Ptr = ST->getBasePtr();
|
|
|
|
isLoad = false;
|
|
|
|
} else
|
|
|
|
return false;
|
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
// If the pointer is not an add/sub, or if it doesn't have multiple uses, bail
|
|
|
|
// out. There is no reason to make this a preinc/predec.
|
|
|
|
if ((Ptr.getOpcode() != ISD::ADD && Ptr.getOpcode() != ISD::SUB) ||
|
|
|
|
Ptr.Val->hasOneUse())
|
|
|
|
return false;
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
// Ask the target to do addressing mode selection.
|
|
|
|
SDOperand BasePtr;
|
|
|
|
SDOperand Offset;
|
|
|
|
ISD::MemIndexedMode AM = ISD::UNINDEXED;
|
|
|
|
if (!TLI.getPreIndexedAddressParts(N, BasePtr, Offset, AM, DAG))
|
|
|
|
return false;
|
2007-05-03 23:52:19 +00:00
|
|
|
// Don't create a indexed load / store with zero offset.
|
|
|
|
if (isa<ConstantSDNode>(Offset) &&
|
2008-03-13 22:13:53 +00:00
|
|
|
cast<ConstantSDNode>(Offset)->isNullValue())
|
2007-05-03 23:52:19 +00:00
|
|
|
return false;
|
2006-11-11 00:56:29 +00:00
|
|
|
|
2006-11-11 01:00:15 +00:00
|
|
|
// Try turning it into a pre-indexed load / store except when:
|
2007-05-24 02:35:39 +00:00
|
|
|
// 1) The new base ptr is a frame index.
|
|
|
|
// 2) If N is a store and the new base ptr is either the same as or is a
|
2006-11-11 00:56:29 +00:00
|
|
|
// predecessor of the value being stored.
|
2007-05-24 02:35:39 +00:00
|
|
|
// 3) Another use of old base ptr is a predecessor of N. If ptr is folded
|
2006-11-11 00:56:29 +00:00
|
|
|
// that would create a cycle.
|
2007-05-24 02:35:39 +00:00
|
|
|
// 4) All uses are load / store ops that use it as old base ptr.
|
2006-11-11 00:56:29 +00:00
|
|
|
|
2006-11-11 01:00:15 +00:00
|
|
|
// Check #1. Preinc'ing a frame index would require copying the stack pointer
|
|
|
|
// (plus the implicit offset) to a register to preinc anyway.
|
|
|
|
if (isa<FrameIndexSDNode>(BasePtr))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Check #2.
|
2006-11-11 00:56:29 +00:00
|
|
|
if (!isLoad) {
|
|
|
|
SDOperand Val = cast<StoreSDNode>(N)->getValue();
|
2008-03-04 00:41:45 +00:00
|
|
|
if (Val == BasePtr || BasePtr.Val->isPredecessorOf(Val.Val))
|
2006-11-11 00:56:29 +00:00
|
|
|
return false;
|
|
|
|
}
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2007-05-24 02:35:39 +00:00
|
|
|
// Now check for #3 and #4.
|
2006-11-11 00:56:29 +00:00
|
|
|
bool RealUse = false;
|
|
|
|
for (SDNode::use_iterator I = Ptr.Val->use_begin(),
|
|
|
|
E = Ptr.Val->use_end(); I != E; ++I) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *Use = I->getUser();
|
2006-11-11 00:56:29 +00:00
|
|
|
if (Use == N)
|
|
|
|
continue;
|
2008-03-04 00:41:45 +00:00
|
|
|
if (Use->isPredecessorOf(N))
|
2006-11-11 00:56:29 +00:00
|
|
|
return false;
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
if (!((Use->getOpcode() == ISD::LOAD &&
|
|
|
|
cast<LoadSDNode>(Use)->getBasePtr() == Ptr) ||
|
2008-02-20 11:10:28 +00:00
|
|
|
(Use->getOpcode() == ISD::STORE &&
|
|
|
|
cast<StoreSDNode>(Use)->getBasePtr() == Ptr)))
|
2006-11-11 00:56:29 +00:00
|
|
|
RealUse = true;
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
2006-11-11 00:56:29 +00:00
|
|
|
if (!RealUse)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
SDOperand Result;
|
|
|
|
if (isLoad)
|
|
|
|
Result = DAG.getIndexedLoad(SDOperand(N,0), BasePtr, Offset, AM);
|
|
|
|
else
|
|
|
|
Result = DAG.getIndexedStore(SDOperand(N,0), BasePtr, Offset, AM);
|
|
|
|
++PreIndexedNodes;
|
|
|
|
++NodesCombined;
|
2007-06-19 14:13:56 +00:00
|
|
|
DOUT << "\nReplacing.4 "; DEBUG(N->dump(&DAG));
|
2006-12-07 20:04:42 +00:00
|
|
|
DOUT << "\nWith: "; DEBUG(Result.Val->dump(&DAG));
|
|
|
|
DOUT << '\n';
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
2006-11-11 00:56:29 +00:00
|
|
|
if (isLoad) {
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 0), Result.getValue(0),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 1), Result.getValue(2),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
} else {
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 0), Result.getValue(1),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, since the node is now dead, remove it from the graph.
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
|
|
|
|
// Replace the uses of Ptr with uses of the updated base value.
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(Ptr, Result.getValue(isLoad ? 1 : 0),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
removeFromWorkList(Ptr.Val);
|
|
|
|
DAG.DeleteNode(Ptr.Val);
|
|
|
|
|
|
|
|
return true;
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
|
|
|
|
2008-06-15 20:12:31 +00:00
|
|
|
/// CombineToPostIndexedLoadStore - Try to combine a load / store with a
|
2006-11-11 00:39:41 +00:00
|
|
|
/// add / sub of the base pointer node into a post-indexed load / store.
|
|
|
|
/// The transformation folded the add / subtract into the new indexed
|
|
|
|
/// load / store effectively and all of its uses are redirected to the
|
|
|
|
/// new load / store.
|
|
|
|
bool DAGCombiner::CombineToPostIndexedLoadStore(SDNode *N) {
|
|
|
|
if (!AfterLegalize)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
bool isLoad = true;
|
|
|
|
SDOperand Ptr;
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT;
|
2006-11-11 00:39:41 +00:00
|
|
|
if (LoadSDNode *LD = dyn_cast<LoadSDNode>(N)) {
|
2008-01-17 19:59:44 +00:00
|
|
|
if (LD->isIndexed())
|
2006-12-16 06:25:23 +00:00
|
|
|
return false;
|
2008-01-30 00:15:11 +00:00
|
|
|
VT = LD->getMemoryVT();
|
2006-11-11 00:39:41 +00:00
|
|
|
if (!TLI.isIndexedLoadLegal(ISD::POST_INC, VT) &&
|
|
|
|
!TLI.isIndexedLoadLegal(ISD::POST_DEC, VT))
|
|
|
|
return false;
|
|
|
|
Ptr = LD->getBasePtr();
|
|
|
|
} else if (StoreSDNode *ST = dyn_cast<StoreSDNode>(N)) {
|
2008-01-17 19:59:44 +00:00
|
|
|
if (ST->isIndexed())
|
2006-12-16 06:25:23 +00:00
|
|
|
return false;
|
2008-01-30 00:15:11 +00:00
|
|
|
VT = ST->getMemoryVT();
|
2006-11-11 00:39:41 +00:00
|
|
|
if (!TLI.isIndexedStoreLegal(ISD::POST_INC, VT) &&
|
|
|
|
!TLI.isIndexedStoreLegal(ISD::POST_DEC, VT))
|
|
|
|
return false;
|
|
|
|
Ptr = ST->getBasePtr();
|
|
|
|
isLoad = false;
|
|
|
|
} else
|
|
|
|
return false;
|
|
|
|
|
2006-11-16 00:08:20 +00:00
|
|
|
if (Ptr.Val->hasOneUse())
|
2006-11-11 00:56:29 +00:00
|
|
|
return false;
|
|
|
|
|
|
|
|
for (SDNode::use_iterator I = Ptr.Val->use_begin(),
|
|
|
|
E = Ptr.Val->use_end(); I != E; ++I) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *Op = I->getUser();
|
2006-11-11 00:56:29 +00:00
|
|
|
if (Op == N ||
|
|
|
|
(Op->getOpcode() != ISD::ADD && Op->getOpcode() != ISD::SUB))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
SDOperand BasePtr;
|
|
|
|
SDOperand Offset;
|
|
|
|
ISD::MemIndexedMode AM = ISD::UNINDEXED;
|
|
|
|
if (TLI.getPostIndexedAddressParts(N, Op, BasePtr, Offset, AM, DAG)) {
|
|
|
|
if (Ptr == Offset)
|
|
|
|
std::swap(BasePtr, Offset);
|
|
|
|
if (Ptr != BasePtr)
|
2006-11-11 00:39:41 +00:00
|
|
|
continue;
|
2007-05-03 23:52:19 +00:00
|
|
|
// Don't create a indexed load / store with zero offset.
|
|
|
|
if (isa<ConstantSDNode>(Offset) &&
|
2008-03-13 22:13:53 +00:00
|
|
|
cast<ConstantSDNode>(Offset)->isNullValue())
|
2007-05-03 23:52:19 +00:00
|
|
|
continue;
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
// Try turning it into a post-indexed load / store except when
|
|
|
|
// 1) All uses are load / store ops that use it as base ptr.
|
|
|
|
// 2) Op must be independent of N, i.e. Op is neither a predecessor
|
|
|
|
// nor a successor of N. Otherwise, if Op is folded that would
|
|
|
|
// create a cycle.
|
|
|
|
|
|
|
|
// Check for #1.
|
|
|
|
bool TryNext = false;
|
|
|
|
for (SDNode::use_iterator II = BasePtr.Val->use_begin(),
|
|
|
|
EE = BasePtr.Val->use_end(); II != EE; ++II) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *Use = II->getUser();
|
2006-11-11 00:56:29 +00:00
|
|
|
if (Use == Ptr.Val)
|
2006-11-11 00:39:41 +00:00
|
|
|
continue;
|
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
// If all the uses are load / store addresses, then don't do the
|
|
|
|
// transformation.
|
|
|
|
if (Use->getOpcode() == ISD::ADD || Use->getOpcode() == ISD::SUB){
|
|
|
|
bool RealUse = false;
|
|
|
|
for (SDNode::use_iterator III = Use->use_begin(),
|
|
|
|
EEE = Use->use_end(); III != EEE; ++III) {
|
2008-04-07 10:06:32 +00:00
|
|
|
SDNode *UseUse = III->getUser();
|
2006-11-11 00:56:29 +00:00
|
|
|
if (!((UseUse->getOpcode() == ISD::LOAD &&
|
|
|
|
cast<LoadSDNode>(UseUse)->getBasePtr().Val == Use) ||
|
2008-02-20 11:10:28 +00:00
|
|
|
(UseUse->getOpcode() == ISD::STORE &&
|
|
|
|
cast<StoreSDNode>(UseUse)->getBasePtr().Val == Use)))
|
2006-11-11 00:56:29 +00:00
|
|
|
RealUse = true;
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
if (!RealUse) {
|
|
|
|
TryNext = true;
|
|
|
|
break;
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
2006-11-11 00:56:29 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if (TryNext)
|
|
|
|
continue;
|
2006-11-11 00:39:41 +00:00
|
|
|
|
2006-11-11 00:56:29 +00:00
|
|
|
// Check for #2
|
2008-03-04 00:41:45 +00:00
|
|
|
if (!Op->isPredecessorOf(N) && !N->isPredecessorOf(Op)) {
|
2006-11-11 00:56:29 +00:00
|
|
|
SDOperand Result = isLoad
|
|
|
|
? DAG.getIndexedLoad(SDOperand(N,0), BasePtr, Offset, AM)
|
|
|
|
: DAG.getIndexedStore(SDOperand(N,0), BasePtr, Offset, AM);
|
|
|
|
++PostIndexedNodes;
|
|
|
|
++NodesCombined;
|
2007-06-19 14:13:56 +00:00
|
|
|
DOUT << "\nReplacing.5 "; DEBUG(N->dump(&DAG));
|
2006-12-07 20:04:42 +00:00
|
|
|
DOUT << "\nWith: "; DEBUG(Result.Val->dump(&DAG));
|
|
|
|
DOUT << '\n';
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
2006-11-11 00:56:29 +00:00
|
|
|
if (isLoad) {
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 0), Result.getValue(0),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 1), Result.getValue(2),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
} else {
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 0), Result.getValue(1),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
2006-11-11 00:56:29 +00:00
|
|
|
|
|
|
|
// Finally, since the node is now dead, remove it from the graph.
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
|
|
|
|
// Replace the uses of Use with uses of the updated base value.
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(Op, 0),
|
|
|
|
Result.getValue(isLoad ? 1 : 0),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
2006-11-11 00:56:29 +00:00
|
|
|
removeFromWorkList(Op);
|
|
|
|
DAG.DeleteNode(Op);
|
|
|
|
return true;
|
2006-11-11 00:39:41 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2008-01-25 07:20:16 +00:00
|
|
|
/// InferAlignment - If we can infer some alignment information from this
|
|
|
|
/// pointer, return it.
|
|
|
|
static unsigned InferAlignment(SDOperand Ptr, SelectionDAG &DAG) {
|
|
|
|
// If this is a direct reference to a stack slot, use information about the
|
|
|
|
// stack slot's alignment.
|
Infer alignment of loads and increase their alignment when we can tell they are
from the stack. This allows us to compile stack-align.ll to:
_test:
movsd LCPI1_0, %xmm0
movapd %xmm0, %xmm1
*** andpd 4(%esp), %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
instead of:
_test:
movsd LCPI1_0, %xmm0
** movsd 4(%esp), %xmm1
** andpd %xmm0, %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46401 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-26 19:45:50 +00:00
|
|
|
int FrameIdx = 1 << 31;
|
|
|
|
int64_t FrameOffset = 0;
|
2008-01-25 07:20:16 +00:00
|
|
|
if (FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Ptr)) {
|
Infer alignment of loads and increase their alignment when we can tell they are
from the stack. This allows us to compile stack-align.ll to:
_test:
movsd LCPI1_0, %xmm0
movapd %xmm0, %xmm1
*** andpd 4(%esp), %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
instead of:
_test:
movsd LCPI1_0, %xmm0
** movsd 4(%esp), %xmm1
** andpd %xmm0, %xmm1
andpd _G, %xmm0
addsd %xmm1, %xmm0
movl 20(%esp), %eax
movsd %xmm0, (%eax)
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46401 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-26 19:45:50 +00:00
|
|
|
FrameIdx = FI->getIndex();
|
|
|
|
} else if (Ptr.getOpcode() == ISD::ADD &&
|
|
|
|
isa<ConstantSDNode>(Ptr.getOperand(1)) &&
|
|
|
|
isa<FrameIndexSDNode>(Ptr.getOperand(0))) {
|
|
|
|
FrameIdx = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex();
|
|
|
|
FrameOffset = Ptr.getConstantOperandVal(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (FrameIdx != (1 << 31)) {
|
|
|
|
// FIXME: Handle FI+CST.
|
|
|
|
const MachineFrameInfo &MFI = *DAG.getMachineFunction().getFrameInfo();
|
|
|
|
if (MFI.isFixedObjectIndex(FrameIdx)) {
|
|
|
|
int64_t ObjectOffset = MFI.getObjectOffset(FrameIdx);
|
|
|
|
|
|
|
|
// The alignment of the frame index can be determined from its offset from
|
|
|
|
// the incoming frame position. If the frame object is at offset 32 and
|
|
|
|
// the stack is guaranteed to be 16-byte aligned, then we know that the
|
|
|
|
// object is 16-byte aligned.
|
|
|
|
unsigned StackAlign = DAG.getTarget().getFrameInfo()->getStackAlignment();
|
|
|
|
unsigned Align = MinAlign(ObjectOffset, StackAlign);
|
|
|
|
|
|
|
|
// Finally, the frame object itself may have a known alignment. Factor
|
|
|
|
// the alignment + offset into a new alignment. For example, if we know
|
|
|
|
// the FI is 8 byte aligned, but the pointer is 4 off, we really have a
|
|
|
|
// 4-byte alignment of the resultant pointer. Likewise align 4 + 4-byte
|
|
|
|
// offset = 4-byte alignment, align 4 + 1-byte offset = align 1, etc.
|
|
|
|
unsigned FIInfoAlign = MinAlign(MFI.getObjectAlignment(FrameIdx),
|
|
|
|
FrameOffset);
|
|
|
|
return std::max(Align, FIInfoAlign);
|
|
|
|
}
|
2008-01-25 07:20:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2006-11-11 00:39:41 +00:00
|
|
|
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
SDOperand DAGCombiner::visitLOAD(SDNode *N) {
|
2006-10-09 20:57:25 +00:00
|
|
|
LoadSDNode *LD = cast<LoadSDNode>(N);
|
|
|
|
SDOperand Chain = LD->getChain();
|
|
|
|
SDOperand Ptr = LD->getBasePtr();
|
2008-01-25 07:20:16 +00:00
|
|
|
|
|
|
|
// Try to infer better alignment information than the load already has.
|
|
|
|
if (LD->isUnindexed()) {
|
|
|
|
if (unsigned Align = InferAlignment(Ptr, DAG)) {
|
|
|
|
if (Align > LD->getAlignment())
|
|
|
|
return DAG.getExtLoad(LD->getExtensionType(), LD->getValueType(0),
|
|
|
|
Chain, Ptr, LD->getSrcValue(),
|
2008-01-30 00:15:11 +00:00
|
|
|
LD->getSrcValueOffset(), LD->getMemoryVT(),
|
2008-01-25 07:20:16 +00:00
|
|
|
LD->isVolatile(), Align);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-05-01 00:38:21 +00:00
|
|
|
|
|
|
|
// If load is not volatile and there are no uses of the loaded value (and
|
|
|
|
// the updated indexed value in case of indexed loads), change uses of the
|
|
|
|
// chain value into uses of the chain input (i.e. delete the dead load).
|
|
|
|
if (!LD->isVolatile()) {
|
2007-05-01 08:53:39 +00:00
|
|
|
if (N->getValueType(1) == MVT::Other) {
|
|
|
|
// Unindexed loads.
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
if (N->hasNUsesOfValue(0, 0)) {
|
|
|
|
// It's not safe to use the two value CombineTo variant here. e.g.
|
|
|
|
// v1, chain2 = load chain1, loc
|
|
|
|
// v2, chain3 = load chain2, loc
|
|
|
|
// v3 = add v2, c
|
2008-01-24 07:57:06 +00:00
|
|
|
// Now we replace use of chain2 with chain1. This makes the second load
|
|
|
|
// isomorphic to the one we are deleting, and thus makes this load live.
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
DOUT << "\nReplacing.6 "; DEBUG(N->dump(&DAG));
|
2008-01-24 07:57:06 +00:00
|
|
|
DOUT << "\nWith chain: "; DEBUG(Chain.Val->dump(&DAG));
|
|
|
|
DOUT << "\n";
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 1), Chain, &DeadNodes);
|
2008-01-24 07:57:06 +00:00
|
|
|
if (N->use_empty()) {
|
|
|
|
removeFromWorkList(N);
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
}
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
|
|
|
}
|
2007-05-01 08:53:39 +00:00
|
|
|
} else {
|
|
|
|
// Indexed loads.
|
|
|
|
assert(N->getValueType(2) == MVT::Other && "Malformed indexed loads?");
|
|
|
|
if (N->hasNUsesOfValue(0, 0) && N->hasNUsesOfValue(0, 1)) {
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
SDOperand Undef = DAG.getNode(ISD::UNDEF, N->getValueType(0));
|
|
|
|
DOUT << "\nReplacing.6 "; DEBUG(N->dump(&DAG));
|
|
|
|
DOUT << "\nWith: "; DEBUG(Undef.Val->dump(&DAG));
|
|
|
|
DOUT << " and 2 other values\n";
|
2008-02-03 06:49:24 +00:00
|
|
|
WorkListRemover DeadNodes(*this);
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 0), Undef, &DeadNodes);
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 1),
|
2008-01-17 07:20:38 +00:00
|
|
|
DAG.getNode(ISD::UNDEF, N->getValueType(1)),
|
2008-02-03 06:49:24 +00:00
|
|
|
&DeadNodes);
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(SDOperand(N, 2), Chain, &DeadNodes);
|
Fixes a nasty dag combiner bug that causes a bunch of tests to fail at -O0.
It's not safe to use the two value CombineTo variant to combine away a dead load.
e.g.
v1, chain2 = load chain1, loc
v2, chain3 = load chain2, loc
v3 = add v2, c
Now we replace use of v1 with undef, use of chain2 with chain1.
ReplaceAllUsesWith() will iterate through uses of the first load and update operands:
v1, chain2 = load chain1, loc
v2, chain3 = load chain1, loc
v3 = add v2, c
Now the second load is the same as the first load, SelectionDAG cse will ensure
the use of second load is replaced with the first load.
v1, chain2 = load chain1, loc
v3 = add v1, c
Then v1 is replaced with undef and bad things happen.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@46099 91177308-0d34-0410-b5e6-96231b3b80d8
2008-01-16 23:11:54 +00:00
|
|
|
removeFromWorkList(N);
|
|
|
|
DAG.DeleteNode(N);
|
|
|
|
return SDOperand(N, 0); // Return N so it doesn't get rechecked!
|
2007-05-01 00:38:21 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
|
|
|
|
// If this load is directly stored, replace the load value with the stored
|
|
|
|
// value.
|
|
|
|
// TODO: Handle store large -> read small portion.
|
2006-10-11 17:47:52 +00:00
|
|
|
// TODO: Handle TRUNCSTORE/LOADEXT
|
2008-03-31 20:32:52 +00:00
|
|
|
if (LD->getExtensionType() == ISD::NON_EXTLOAD &&
|
|
|
|
!LD->isVolatile()) {
|
2006-10-13 21:14:26 +00:00
|
|
|
if (ISD::isNON_TRUNCStore(Chain.Val)) {
|
|
|
|
StoreSDNode *PrevST = cast<StoreSDNode>(Chain);
|
|
|
|
if (PrevST->getBasePtr() == Ptr &&
|
|
|
|
PrevST->getValue().getValueType() == N->getValueType(0))
|
2006-10-11 17:47:52 +00:00
|
|
|
return CombineTo(N, Chain.getOperand(1), Chain);
|
2006-10-13 21:14:26 +00:00
|
|
|
}
|
2006-10-11 17:47:52 +00:00
|
|
|
}
|
2006-10-04 16:53:27 +00:00
|
|
|
|
2006-10-11 13:47:09 +00:00
|
|
|
if (CombinerAA) {
|
2006-09-25 16:29:54 +00:00
|
|
|
// Walk up chain skipping non-aliasing memory nodes.
|
|
|
|
SDOperand BetterChain = FindBetterChain(N, Chain);
|
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
// If there is a better chain.
|
2006-09-25 16:29:54 +00:00
|
|
|
if (Chain != BetterChain) {
|
2006-10-11 17:47:52 +00:00
|
|
|
SDOperand ReplLoad;
|
|
|
|
|
2006-09-25 16:29:54 +00:00
|
|
|
// Replace the chain to void dependency.
|
2006-10-11 17:47:52 +00:00
|
|
|
if (LD->getExtensionType() == ISD::NON_EXTLOAD) {
|
|
|
|
ReplLoad = DAG.getLoad(N->getValueType(0), BetterChain, Ptr,
|
2007-10-28 12:59:45 +00:00
|
|
|
LD->getSrcValue(), LD->getSrcValueOffset(),
|
|
|
|
LD->isVolatile(), LD->getAlignment());
|
2006-10-11 17:47:52 +00:00
|
|
|
} else {
|
|
|
|
ReplLoad = DAG.getExtLoad(LD->getExtensionType(),
|
|
|
|
LD->getValueType(0),
|
|
|
|
BetterChain, Ptr, LD->getSrcValue(),
|
|
|
|
LD->getSrcValueOffset(),
|
2008-01-30 00:15:11 +00:00
|
|
|
LD->getMemoryVT(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LD->isVolatile(),
|
|
|
|
LD->getAlignment());
|
2006-10-11 17:47:52 +00:00
|
|
|
}
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
// Create token factor to keep old chain connected.
|
2006-09-25 19:32:58 +00:00
|
|
|
SDOperand Token = DAG.getNode(ISD::TokenFactor, MVT::Other,
|
|
|
|
Chain, ReplLoad.getValue(1));
|
2006-10-04 16:53:27 +00:00
|
|
|
|
2006-10-13 23:32:28 +00:00
|
|
|
// Replace uses with load result and token factor. Don't add users
|
|
|
|
// to work list.
|
|
|
|
return CombineTo(N, ReplLoad.getValue(0), Token, false);
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-11-03 03:06:21 +00:00
|
|
|
// Try transforming N to an indexed load.
|
2006-11-07 09:03:05 +00:00
|
|
|
if (CombineToPreIndexedLoadStore(N) || CombineToPostIndexedLoadStore(N))
|
2006-11-03 03:06:21 +00:00
|
|
|
return SDOperand(N, 0);
|
|
|
|
|
Add support for CombineTo, allowing the dag combiner to replace nodes with
multiple results.
Use this support to implement trivial store->load forwarding, implementing
CodeGen/PowerPC/store-load-fwd.ll. Though this is the most simple case and
can be extended in the future, it is still useful. For example, it speeds
up 197.parser by 6.2% by avoiding an LSU reject in xalloc:
stw r6, lo16(l5_end_of_array)(r2)
addi r2, r5, -4
stwx r5, r4, r2
- lwzx r5, r4, r2
- rlwinm r5, r5, 0, 0, 30
stwx r5, r4, r2
lwz r2, -4(r4)
ori r2, r2, 1
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23690 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-10 22:04:48 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2008-01-08 23:08:06 +00:00
|
|
|
|
2005-10-10 22:31:19 +00:00
|
|
|
SDOperand DAGCombiner::visitSTORE(SDNode *N) {
|
2006-10-13 21:14:26 +00:00
|
|
|
StoreSDNode *ST = cast<StoreSDNode>(N);
|
|
|
|
SDOperand Chain = ST->getChain();
|
|
|
|
SDOperand Value = ST->getValue();
|
|
|
|
SDOperand Ptr = ST->getBasePtr();
|
2006-10-11 18:55:16 +00:00
|
|
|
|
2008-01-25 07:20:16 +00:00
|
|
|
// Try to infer better alignment information than the store already has.
|
|
|
|
if (ST->isUnindexed()) {
|
|
|
|
if (unsigned Align = InferAlignment(Ptr, DAG)) {
|
|
|
|
if (Align > ST->getAlignment())
|
|
|
|
return DAG.getTruncStore(Chain, Value, Ptr, ST->getSrcValue(),
|
2008-01-30 00:15:11 +00:00
|
|
|
ST->getSrcValueOffset(), ST->getMemoryVT(),
|
2008-01-25 07:20:16 +00:00
|
|
|
ST->isVolatile(), Align);
|
|
|
|
}
|
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2007-05-07 21:27:48 +00:00
|
|
|
// If this is a store of a bit convert, store the input value if the
|
2007-05-09 21:49:47 +00:00
|
|
|
// resultant store does not need a higher alignment than the original.
|
2007-05-16 22:45:30 +00:00
|
|
|
if (Value.getOpcode() == ISD::BIT_CONVERT && !ST->isTruncatingStore() &&
|
2008-01-17 19:59:44 +00:00
|
|
|
ST->isUnindexed()) {
|
2007-05-07 21:27:48 +00:00
|
|
|
unsigned Align = ST->getAlignment();
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT SVT = Value.getOperand(0).getValueType();
|
2007-05-07 21:27:48 +00:00
|
|
|
unsigned OrigAlign = TLI.getTargetMachine().getTargetData()->
|
2008-06-06 12:08:01 +00:00
|
|
|
getABITypeAlignment(SVT.getTypeForMVT());
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if (Align <= OrigAlign &&
|
|
|
|
((!AfterLegalize && !ST->isVolatile()) ||
|
|
|
|
TLI.isOperationLegal(ISD::STORE, SVT)))
|
2007-05-07 21:27:48 +00:00
|
|
|
return DAG.getStore(Chain, Value.getOperand(0), Ptr, ST->getSrcValue(),
|
2008-06-28 00:45:22 +00:00
|
|
|
ST->getSrcValueOffset(), ST->isVolatile(), OrigAlign);
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2006-12-11 02:23:46 +00:00
|
|
|
// Turn 'store float 1.0, Ptr' -> 'store int 0x12345678, Ptr'
|
|
|
|
if (ConstantFPSDNode *CFP = dyn_cast<ConstantFPSDNode>(Value)) {
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// NOTE: If the original store is volatile, this transform must not increase
|
|
|
|
// the number of stores. For example, on x86-32 an f64 can be stored in one
|
|
|
|
// processor operation but an i64 (which is not legal) requires two. So the
|
|
|
|
// transform should not be done in this case.
|
2006-12-11 17:25:19 +00:00
|
|
|
if (Value.getOpcode() != ISD::TargetConstantFP) {
|
|
|
|
SDOperand Tmp;
|
2008-06-06 12:08:01 +00:00
|
|
|
switch (CFP->getValueType(0).getSimpleVT()) {
|
2006-12-12 04:16:14 +00:00
|
|
|
default: assert(0 && "Unknown FP type");
|
2007-09-18 18:36:59 +00:00
|
|
|
case MVT::f80: // We don't do this for these yet.
|
|
|
|
case MVT::f128:
|
|
|
|
case MVT::ppcf128:
|
|
|
|
break;
|
2006-12-12 04:16:14 +00:00
|
|
|
case MVT::f32:
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if ((!AfterLegalize && !ST->isVolatile()) ||
|
|
|
|
TLI.isOperationLegal(ISD::STORE, MVT::i32)) {
|
2007-09-12 03:30:33 +00:00
|
|
|
Tmp = DAG.getConstant((uint32_t)CFP->getValueAPF().
|
|
|
|
convertToAPInt().getZExtValue(), MVT::i32);
|
2006-12-12 04:16:14 +00:00
|
|
|
return DAG.getStore(Chain, Tmp, Ptr, ST->getSrcValue(),
|
2007-07-09 22:18:38 +00:00
|
|
|
ST->getSrcValueOffset(), ST->isVolatile(),
|
|
|
|
ST->getAlignment());
|
2006-12-12 04:16:14 +00:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
case MVT::f64:
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
if ((!AfterLegalize && !ST->isVolatile()) ||
|
|
|
|
TLI.isOperationLegal(ISD::STORE, MVT::i64)) {
|
2007-09-12 03:30:33 +00:00
|
|
|
Tmp = DAG.getConstant(CFP->getValueAPF().convertToAPInt().
|
|
|
|
getZExtValue(), MVT::i64);
|
2006-12-12 04:16:14 +00:00
|
|
|
return DAG.getStore(Chain, Tmp, Ptr, ST->getSrcValue(),
|
2007-07-09 22:18:38 +00:00
|
|
|
ST->getSrcValueOffset(), ST->isVolatile(),
|
|
|
|
ST->getAlignment());
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
} else if (!ST->isVolatile() &&
|
|
|
|
TLI.isOperationLegal(ISD::STORE, MVT::i32)) {
|
2007-10-28 12:59:45 +00:00
|
|
|
// Many FP stores are not made apparent until after legalize, e.g. for
|
2006-12-12 04:16:14 +00:00
|
|
|
// argument passing. Since this is so common, custom legalize the
|
|
|
|
// 64-bit integer store into two 32-bit stores.
|
2007-09-12 03:30:33 +00:00
|
|
|
uint64_t Val = CFP->getValueAPF().convertToAPInt().getZExtValue();
|
2006-12-12 04:16:14 +00:00
|
|
|
SDOperand Lo = DAG.getConstant(Val & 0xFFFFFFFF, MVT::i32);
|
|
|
|
SDOperand Hi = DAG.getConstant(Val >> 32, MVT::i32);
|
2008-02-11 10:37:04 +00:00
|
|
|
if (TLI.isBigEndian()) std::swap(Lo, Hi);
|
2006-12-12 04:16:14 +00:00
|
|
|
|
2007-07-09 22:18:38 +00:00
|
|
|
int SVOffset = ST->getSrcValueOffset();
|
|
|
|
unsigned Alignment = ST->getAlignment();
|
|
|
|
bool isVolatile = ST->isVolatile();
|
|
|
|
|
2006-12-12 04:16:14 +00:00
|
|
|
SDOperand St0 = DAG.getStore(Chain, Lo, Ptr, ST->getSrcValue(),
|
2007-07-09 22:18:38 +00:00
|
|
|
ST->getSrcValueOffset(),
|
|
|
|
isVolatile, ST->getAlignment());
|
2006-12-12 04:16:14 +00:00
|
|
|
Ptr = DAG.getNode(ISD::ADD, Ptr.getValueType(), Ptr,
|
|
|
|
DAG.getConstant(4, Ptr.getValueType()));
|
2007-07-09 22:18:38 +00:00
|
|
|
SVOffset += 4;
|
2007-10-28 12:59:45 +00:00
|
|
|
Alignment = MinAlign(Alignment, 4U);
|
2006-12-12 04:16:14 +00:00
|
|
|
SDOperand St1 = DAG.getStore(Chain, Hi, Ptr, ST->getSrcValue(),
|
2007-07-09 22:18:38 +00:00
|
|
|
SVOffset, isVolatile, Alignment);
|
2006-12-12 04:16:14 +00:00
|
|
|
return DAG.getNode(ISD::TokenFactor, MVT::Other, St0, St1);
|
|
|
|
}
|
|
|
|
break;
|
2006-12-11 17:25:19 +00:00
|
|
|
}
|
2006-12-11 02:23:46 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-09-25 16:29:54 +00:00
|
|
|
if (CombinerAA) {
|
|
|
|
// Walk up chain skipping non-aliasing memory nodes.
|
|
|
|
SDOperand BetterChain = FindBetterChain(N, Chain);
|
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
// If there is a better chain.
|
2006-09-25 16:29:54 +00:00
|
|
|
if (Chain != BetterChain) {
|
2006-10-04 16:53:27 +00:00
|
|
|
// Replace the chain to avoid dependency.
|
2006-10-14 12:14:27 +00:00
|
|
|
SDOperand ReplStore;
|
|
|
|
if (ST->isTruncatingStore()) {
|
|
|
|
ReplStore = DAG.getTruncStore(BetterChain, Value, Ptr,
|
2008-01-17 07:20:38 +00:00
|
|
|
ST->getSrcValue(),ST->getSrcValueOffset(),
|
2008-01-30 00:15:11 +00:00
|
|
|
ST->getMemoryVT(),
|
2008-01-17 07:20:38 +00:00
|
|
|
ST->isVolatile(), ST->getAlignment());
|
2006-10-14 12:14:27 +00:00
|
|
|
} else {
|
|
|
|
ReplStore = DAG.getStore(BetterChain, Value, Ptr,
|
2008-01-17 07:20:38 +00:00
|
|
|
ST->getSrcValue(), ST->getSrcValueOffset(),
|
|
|
|
ST->isVolatile(), ST->getAlignment());
|
2006-10-14 12:14:27 +00:00
|
|
|
}
|
|
|
|
|
2006-09-25 16:29:54 +00:00
|
|
|
// Create token to keep both nodes around.
|
2006-10-13 23:32:28 +00:00
|
|
|
SDOperand Token =
|
|
|
|
DAG.getNode(ISD::TokenFactor, MVT::Other, Chain, ReplStore);
|
|
|
|
|
|
|
|
// Don't add users to work list.
|
|
|
|
return CombineTo(N, Token, false);
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
2006-09-21 16:28:59 +00:00
|
|
|
}
|
Simplify store(bitconv(x)) to store(x). This allows us to compile this:
void bar(double Y, double *X) {
*X = Y;
}
to this:
bar:
save -96, %o6, %o6
st %i1, [%i2+4]
st %i0, [%i2]
restore %g0, %g0, %g0
retl
nop
instead of this:
bar:
save -104, %o6, %o6
st %i1, [%i6+-4]
st %i0, [%i6+-8]
ldd [%i6+-8], %f0
std %f0, [%i2]
restore %g0, %g0, %g0
retl
nop
on sparcv8.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@24983 91177308-0d34-0410-b5e6-96231b3b80d8
2005-12-23 05:48:07 +00:00
|
|
|
|
2006-11-05 09:31:14 +00:00
|
|
|
// Try transforming N to an indexed store.
|
2006-11-07 09:03:05 +00:00
|
|
|
if (CombineToPreIndexedLoadStore(N) || CombineToPostIndexedLoadStore(N))
|
2006-11-05 09:31:14 +00:00
|
|
|
return SDOperand(N, 0);
|
|
|
|
|
2007-12-29 06:26:16 +00:00
|
|
|
// FIXME: is there such a thing as a truncating indexed store?
|
2008-01-17 19:59:44 +00:00
|
|
|
if (ST->isTruncatingStore() && ST->isUnindexed() &&
|
2008-06-06 12:08:01 +00:00
|
|
|
Value.getValueType().isInteger()) {
|
2007-10-13 06:35:54 +00:00
|
|
|
// See if we can simplify the input to this truncstore with knowledge that
|
|
|
|
// only the low bits are being used. For example:
|
|
|
|
// "truncstore (or (shl x, 8), y), i8" -> "truncstore y, i8"
|
|
|
|
SDOperand Shorter =
|
2008-02-25 21:11:39 +00:00
|
|
|
GetDemandedBits(Value,
|
|
|
|
APInt::getLowBitsSet(Value.getValueSizeInBits(),
|
2008-06-06 12:08:01 +00:00
|
|
|
ST->getMemoryVT().getSizeInBits()));
|
2007-10-13 06:35:54 +00:00
|
|
|
AddToWorkList(Value.Val);
|
|
|
|
if (Shorter.Val)
|
|
|
|
return DAG.getTruncStore(Chain, Shorter, Ptr, ST->getSrcValue(),
|
2008-01-30 00:15:11 +00:00
|
|
|
ST->getSrcValueOffset(), ST->getMemoryVT(),
|
2007-10-13 06:35:54 +00:00
|
|
|
ST->isVolatile(), ST->getAlignment());
|
Enhance the truncstore optimization code to handle shifted
values and propagate demanded bits through them in simple cases.
This allows this code:
void foo(char *P) {
strcpy(P, "abc");
}
to compile to:
_foo:
ldrb r3, [r1]
ldrb r2, [r1, #+1]
ldrb r12, [r1, #+2]!
ldrb r1, [r1, #+1]
strb r1, [r0, #+3]
strb r2, [r0, #+1]
strb r12, [r0, #+2]
strb r3, [r0]
bx lr
instead of:
_foo:
ldrb r3, [r1, #+3]
ldrb r2, [r1, #+2]
orr r3, r2, r3, lsl #8
ldrb r2, [r1, #+1]
ldrb r1, [r1]
orr r2, r1, r2, lsl #8
orr r3, r2, r3, lsl #16
strb r3, [r0]
mov r2, r3, lsr #24
strb r2, [r0, #+3]
mov r2, r3, lsr #16
strb r2, [r0, #+2]
mov r3, r3, lsr #8
strb r3, [r0, #+1]
bx lr
testcase here: test/CodeGen/ARM/truncstore-dag-combine.ll
This also helps occasionally for X86 and other cases not involving
unaligned load/stores.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@42954 91177308-0d34-0410-b5e6-96231b3b80d8
2007-10-13 06:58:48 +00:00
|
|
|
|
|
|
|
// Otherwise, see if we can simplify the operation with
|
|
|
|
// SimplifyDemandedBits, which only works if the value has a single use.
|
2008-02-27 00:25:32 +00:00
|
|
|
if (SimplifyDemandedBits(Value,
|
|
|
|
APInt::getLowBitsSet(
|
|
|
|
Value.getValueSizeInBits(),
|
2008-06-06 12:08:01 +00:00
|
|
|
ST->getMemoryVT().getSizeInBits())))
|
Enhance the truncstore optimization code to handle shifted
values and propagate demanded bits through them in simple cases.
This allows this code:
void foo(char *P) {
strcpy(P, "abc");
}
to compile to:
_foo:
ldrb r3, [r1]
ldrb r2, [r1, #+1]
ldrb r12, [r1, #+2]!
ldrb r1, [r1, #+1]
strb r1, [r0, #+3]
strb r2, [r0, #+1]
strb r12, [r0, #+2]
strb r3, [r0]
bx lr
instead of:
_foo:
ldrb r3, [r1, #+3]
ldrb r2, [r1, #+2]
orr r3, r2, r3, lsl #8
ldrb r2, [r1, #+1]
ldrb r1, [r1]
orr r2, r1, r2, lsl #8
orr r3, r2, r3, lsl #16
strb r3, [r0]
mov r2, r3, lsr #24
strb r2, [r0, #+3]
mov r2, r3, lsr #16
strb r2, [r0, #+2]
mov r3, r3, lsr #8
strb r3, [r0, #+1]
bx lr
testcase here: test/CodeGen/ARM/truncstore-dag-combine.ll
This also helps occasionally for X86 and other cases not involving
unaligned load/stores.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@42954 91177308-0d34-0410-b5e6-96231b3b80d8
2007-10-13 06:58:48 +00:00
|
|
|
return SDOperand(N, 0);
|
2007-10-13 06:35:54 +00:00
|
|
|
}
|
|
|
|
|
2007-12-29 06:26:16 +00:00
|
|
|
// If this is a load followed by a store to the same location, then the store
|
|
|
|
// is dead/noop.
|
|
|
|
if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(Value)) {
|
2008-01-30 00:15:11 +00:00
|
|
|
if (Ld->getBasePtr() == Ptr && ST->getMemoryVT() == Ld->getMemoryVT() &&
|
2008-01-17 19:59:44 +00:00
|
|
|
ST->isUnindexed() && !ST->isVolatile() &&
|
2008-01-08 23:08:06 +00:00
|
|
|
// There can't be any side effects between the load and store, such as
|
|
|
|
// a call or store.
|
2008-01-16 05:49:24 +00:00
|
|
|
Chain.reachesChainWithoutSideEffects(SDOperand(Ld, 1))) {
|
2007-12-29 06:26:16 +00:00
|
|
|
// The store is dead, remove it.
|
|
|
|
return Chain;
|
|
|
|
}
|
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2008-01-17 19:59:44 +00:00
|
|
|
// If this is an FP_ROUND or TRUNC followed by a store, fold this into a
|
|
|
|
// truncating store. We can do this even if this is already a truncstore.
|
|
|
|
if ((Value.getOpcode() == ISD::FP_ROUND || Value.getOpcode() == ISD::TRUNCATE)
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
&& Value.Val->hasOneUse() && ST->isUnindexed() &&
|
2008-01-17 19:59:44 +00:00
|
|
|
TLI.isTruncStoreLegal(Value.getOperand(0).getValueType(),
|
2008-01-30 00:15:11 +00:00
|
|
|
ST->getMemoryVT())) {
|
2008-01-17 19:59:44 +00:00
|
|
|
return DAG.getTruncStore(Chain, Value.getOperand(0), Ptr, ST->getSrcValue(),
|
2008-01-30 00:15:11 +00:00
|
|
|
ST->getSrcValueOffset(), ST->getMemoryVT(),
|
2008-01-17 19:59:44 +00:00
|
|
|
ST->isVolatile(), ST->getAlignment());
|
|
|
|
}
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
|
2005-10-10 22:31:19 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-03-19 01:27:56 +00:00
|
|
|
SDOperand DAGCombiner::visitINSERT_VECTOR_ELT(SDNode *N) {
|
|
|
|
SDOperand InVec = N->getOperand(0);
|
|
|
|
SDOperand InVal = N->getOperand(1);
|
|
|
|
SDOperand EltNo = N->getOperand(2);
|
|
|
|
|
|
|
|
// If the invec is a BUILD_VECTOR and if EltNo is a constant, build a new
|
|
|
|
// vector with the inserted element.
|
|
|
|
if (InVec.getOpcode() == ISD::BUILD_VECTOR && isa<ConstantSDNode>(EltNo)) {
|
|
|
|
unsigned Elt = cast<ConstantSDNode>(EltNo)->getValue();
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> Ops(InVec.Val->op_begin(), InVec.Val->op_end());
|
2006-03-19 01:27:56 +00:00
|
|
|
if (Elt < Ops.size())
|
|
|
|
Ops[Elt] = InVal;
|
2006-08-08 02:23:42 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, InVec.getValueType(),
|
|
|
|
&Ops[0], Ops.size());
|
2006-03-19 01:27:56 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-10-06 08:19:55 +00:00
|
|
|
SDOperand DAGCombiner::visitEXTRACT_VECTOR_ELT(SDNode *N) {
|
2008-05-13 08:35:03 +00:00
|
|
|
// (vextract (v4f32 load $addr), c) -> (f32 load $addr+c*size)
|
|
|
|
// (vextract (v4f32 s2v (f32 load $addr)), c) -> (f32 load $addr+c*size)
|
|
|
|
// (vextract (v4f32 shuffle (load $addr), <1,u,u,u>), 0) -> (f32 load $addr)
|
|
|
|
|
|
|
|
// Perform only after legalization to ensure build_vector / vector_shuffle
|
|
|
|
// optimizations have already been done.
|
|
|
|
if (!AfterLegalize) return SDOperand();
|
|
|
|
|
2007-10-06 08:19:55 +00:00
|
|
|
SDOperand InVec = N->getOperand(0);
|
|
|
|
SDOperand EltNo = N->getOperand(1);
|
|
|
|
|
|
|
|
if (isa<ConstantSDNode>(EltNo)) {
|
|
|
|
unsigned Elt = cast<ConstantSDNode>(EltNo)->getValue();
|
|
|
|
bool NewLoad = false;
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = InVec.getValueType();
|
|
|
|
MVT EVT = VT.getVectorElementType();
|
|
|
|
MVT LVT = EVT;
|
2008-05-13 08:35:03 +00:00
|
|
|
if (InVec.getOpcode() == ISD::BIT_CONVERT) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT BCVT = InVec.getOperand(0).getValueType();
|
2008-06-08 20:54:56 +00:00
|
|
|
if (!BCVT.isVector() || EVT.bitsGT(BCVT.getVectorElementType()))
|
2008-05-13 08:35:03 +00:00
|
|
|
return SDOperand();
|
|
|
|
InVec = InVec.getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
EVT = BCVT.getVectorElementType();
|
2008-05-13 08:35:03 +00:00
|
|
|
NewLoad = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
LoadSDNode *LN0 = NULL;
|
|
|
|
if (ISD::isNormalLoad(InVec.Val))
|
|
|
|
LN0 = cast<LoadSDNode>(InVec);
|
|
|
|
else if (InVec.getOpcode() == ISD::SCALAR_TO_VECTOR &&
|
|
|
|
InVec.getOperand(0).getValueType() == EVT &&
|
|
|
|
ISD::isNormalLoad(InVec.getOperand(0).Val)) {
|
|
|
|
LN0 = cast<LoadSDNode>(InVec.getOperand(0));
|
|
|
|
} else if (InVec.getOpcode() == ISD::VECTOR_SHUFFLE) {
|
|
|
|
// (vextract (vector_shuffle (load $addr), v2, <1, u, u, u>), 1)
|
|
|
|
// =>
|
|
|
|
// (load $addr+1*size)
|
|
|
|
unsigned Idx = cast<ConstantSDNode>(InVec.getOperand(2).
|
|
|
|
getOperand(Elt))->getValue();
|
|
|
|
unsigned NumElems = InVec.getOperand(2).getNumOperands();
|
|
|
|
InVec = (Idx < NumElems) ? InVec.getOperand(0) : InVec.getOperand(1);
|
|
|
|
if (InVec.getOpcode() == ISD::BIT_CONVERT)
|
2007-10-06 08:19:55 +00:00
|
|
|
InVec = InVec.getOperand(0);
|
2008-05-13 08:35:03 +00:00
|
|
|
if (ISD::isNormalLoad(InVec.Val)) {
|
|
|
|
LN0 = cast<LoadSDNode>(InVec);
|
|
|
|
Elt = (Idx < NumElems) ? Idx : Idx - NumElems;
|
2007-10-06 08:19:55 +00:00
|
|
|
}
|
2008-05-13 08:35:03 +00:00
|
|
|
}
|
2008-06-15 20:12:31 +00:00
|
|
|
if (!LN0 || !LN0->hasOneUse() || LN0->isVolatile())
|
2008-05-13 08:35:03 +00:00
|
|
|
return SDOperand();
|
2007-10-06 08:19:55 +00:00
|
|
|
|
2008-05-13 08:35:03 +00:00
|
|
|
unsigned Align = LN0->getAlignment();
|
|
|
|
if (NewLoad) {
|
|
|
|
// Check the resultant load doesn't need a higher alignment than the
|
|
|
|
// original load.
|
|
|
|
unsigned NewAlign = TLI.getTargetMachine().getTargetData()->
|
2008-06-06 12:08:01 +00:00
|
|
|
getABITypeAlignment(LVT.getTypeForMVT());
|
2008-06-14 17:48:34 +00:00
|
|
|
if (NewAlign > Align || !TLI.isOperationLegal(ISD::LOAD, LVT))
|
2008-05-13 08:35:03 +00:00
|
|
|
return SDOperand();
|
|
|
|
Align = NewAlign;
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand NewPtr = LN0->getBasePtr();
|
|
|
|
if (Elt) {
|
2008-06-06 12:08:01 +00:00
|
|
|
unsigned PtrOff = LVT.getSizeInBits() * Elt / 8;
|
|
|
|
MVT PtrType = NewPtr.getValueType();
|
2008-05-13 08:35:03 +00:00
|
|
|
if (TLI.isBigEndian())
|
2008-06-06 12:08:01 +00:00
|
|
|
PtrOff = VT.getSizeInBits() / 8 - PtrOff;
|
2008-05-13 08:35:03 +00:00
|
|
|
NewPtr = DAG.getNode(ISD::ADD, PtrType, NewPtr,
|
|
|
|
DAG.getConstant(PtrOff, PtrType));
|
2007-10-06 08:19:55 +00:00
|
|
|
}
|
2008-05-13 08:35:03 +00:00
|
|
|
return DAG.getLoad(LVT, LN0->getChain(), NewPtr,
|
|
|
|
LN0->getSrcValue(), LN0->getSrcValueOffset(),
|
|
|
|
LN0->isVolatile(), Align);
|
2007-10-06 08:19:55 +00:00
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand DAGCombiner::visitBUILD_VECTOR(SDNode *N) {
|
|
|
|
unsigned NumInScalars = N->getNumOperands();
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
unsigned NumElts = VT.getVectorNumElements();
|
|
|
|
MVT EltType = VT.getVectorElementType();
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// Check to see if this is a BUILD_VECTOR of a bunch of EXTRACT_VECTOR_ELT
|
|
|
|
// operations. If so, and if the EXTRACT_VECTOR_ELT vector inputs come from
|
|
|
|
// at most two distinct vectors, turn this into a shuffle node.
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
SDOperand VecIn1, VecIn2;
|
|
|
|
for (unsigned i = 0; i != NumInScalars; ++i) {
|
|
|
|
// Ignore undef inputs.
|
|
|
|
if (N->getOperand(i).getOpcode() == ISD::UNDEF) continue;
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If this input is something other than a EXTRACT_VECTOR_ELT with a
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
// constant index, bail out.
|
2007-06-25 16:23:39 +00:00
|
|
|
if (N->getOperand(i).getOpcode() != ISD::EXTRACT_VECTOR_ELT ||
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
!isa<ConstantSDNode>(N->getOperand(i).getOperand(1))) {
|
|
|
|
VecIn1 = VecIn2 = SDOperand(0, 0);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If the input vector type disagrees with the result of the build_vector,
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
// we can't make a shuffle.
|
|
|
|
SDOperand ExtractedFromVec = N->getOperand(i).getOperand(0);
|
2007-06-25 16:23:39 +00:00
|
|
|
if (ExtractedFromVec.getValueType() != VT) {
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
VecIn1 = VecIn2 = SDOperand(0, 0);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, remember this. We allow up to two distinct input vectors.
|
|
|
|
if (ExtractedFromVec == VecIn1 || ExtractedFromVec == VecIn2)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (VecIn1.Val == 0) {
|
|
|
|
VecIn1 = ExtractedFromVec;
|
|
|
|
} else if (VecIn2.Val == 0) {
|
|
|
|
VecIn2 = ExtractedFromVec;
|
|
|
|
} else {
|
|
|
|
// Too many inputs.
|
|
|
|
VecIn1 = VecIn2 = SDOperand(0, 0);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If everything is good, we can make a shuffle operation.
|
|
|
|
if (VecIn1.Val) {
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> BuildVecIndices;
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
for (unsigned i = 0; i != NumInScalars; ++i) {
|
|
|
|
if (N->getOperand(i).getOpcode() == ISD::UNDEF) {
|
2007-01-20 10:10:26 +00:00
|
|
|
BuildVecIndices.push_back(DAG.getNode(ISD::UNDEF, TLI.getPointerTy()));
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand Extract = N->getOperand(i);
|
|
|
|
|
|
|
|
// If extracting from the first vector, just use the index directly.
|
|
|
|
if (Extract.getOperand(0) == VecIn1) {
|
|
|
|
BuildVecIndices.push_back(Extract.getOperand(1));
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, use InIdx + VecSize
|
|
|
|
unsigned Idx = cast<ConstantSDNode>(Extract.getOperand(1))->getValue();
|
2008-01-17 07:00:52 +00:00
|
|
|
BuildVecIndices.push_back(DAG.getIntPtrConstant(Idx+NumInScalars));
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Add count and size info.
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT BuildVecVT = MVT::getVectorVT(TLI.getPointerTy(), NumElts);
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// Return the new VECTOR_SHUFFLE node.
|
2006-08-08 02:23:42 +00:00
|
|
|
SDOperand Ops[5];
|
|
|
|
Ops[0] = VecIn1;
|
When building a VVECTOR_SHUFFLE node from extract_element operations, make
sure to build it as SHUFFLE(X, undef, mask), not SHUFFLE(X, X, mask).
The later is not canonical form, and prevents the PPC splat pattern from
matching. For a particular splat, we go from generating this:
li r10, lo16(LCPI1_0)
lis r11, ha16(LCPI1_0)
lvx v3, r11, r10
vperm v3, v2, v2, v3
to generating:
vspltw v3, v2, 3
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27236 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 22:19:47 +00:00
|
|
|
if (VecIn2.Val) {
|
2006-08-08 02:23:42 +00:00
|
|
|
Ops[1] = VecIn2;
|
When building a VVECTOR_SHUFFLE node from extract_element operations, make
sure to build it as SHUFFLE(X, undef, mask), not SHUFFLE(X, X, mask).
The later is not canonical form, and prevents the PPC splat pattern from
matching. For a particular splat, we go from generating this:
li r10, lo16(LCPI1_0)
lis r11, ha16(LCPI1_0)
lvx v3, r11, r10
vperm v3, v2, v2, v3
to generating:
vspltw v3, v2, 3
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27236 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 22:19:47 +00:00
|
|
|
} else {
|
2007-06-25 16:23:39 +00:00
|
|
|
// Use an undef build_vector as input for the second operand.
|
When building a VVECTOR_SHUFFLE node from extract_element operations, make
sure to build it as SHUFFLE(X, undef, mask), not SHUFFLE(X, X, mask).
The later is not canonical form, and prevents the PPC splat pattern from
matching. For a particular splat, we go from generating this:
li r10, lo16(LCPI1_0)
lis r11, ha16(LCPI1_0)
lvx v3, r11, r10
vperm v3, v2, v2, v3
to generating:
vspltw v3, v2, 3
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27236 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 22:19:47 +00:00
|
|
|
std::vector<SDOperand> UnOps(NumInScalars,
|
|
|
|
DAG.getNode(ISD::UNDEF,
|
2007-06-25 16:23:39 +00:00
|
|
|
EltType));
|
|
|
|
Ops[1] = DAG.getNode(ISD::BUILD_VECTOR, VT,
|
2006-08-08 02:23:42 +00:00
|
|
|
&UnOps[0], UnOps.size());
|
|
|
|
AddToWorkList(Ops[1].Val);
|
When building a VVECTOR_SHUFFLE node from extract_element operations, make
sure to build it as SHUFFLE(X, undef, mask), not SHUFFLE(X, X, mask).
The later is not canonical form, and prevents the PPC splat pattern from
matching. For a particular splat, we go from generating this:
li r10, lo16(LCPI1_0)
lis r11, ha16(LCPI1_0)
lvx v3, r11, r10
vperm v3, v2, v2, v3
to generating:
vspltw v3, v2, 3
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27236 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 22:19:47 +00:00
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
Ops[2] = DAG.getNode(ISD::BUILD_VECTOR, BuildVecVT,
|
2006-08-08 02:23:42 +00:00
|
|
|
&BuildVecIndices[0], BuildVecIndices.size());
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::VECTOR_SHUFFLE, VT, Ops, 3);
|
Turn a series of extract_element's feeding a build_vector into a
vector_shuffle node. For this:
void test(__m128 *res, __m128 *A, __m128 *B) {
*res = _mm_unpacklo_ps(*A, *B);
}
we now produce this code:
_test:
movl 8(%esp), %eax
movaps (%eax), %xmm0
movl 12(%esp), %eax
unpcklps (%eax), %xmm0
movl 4(%esp), %eax
movaps %xmm0, (%eax)
ret
instead of this:
_test:
subl $76, %esp
movl 88(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, (%esp)
movaps %xmm0, 32(%esp)
movss 4(%esp), %xmm0
movss 32(%esp), %xmm1
unpcklps %xmm0, %xmm1
movl 84(%esp), %eax
movaps (%eax), %xmm0
movaps %xmm0, 16(%esp)
movaps %xmm0, 48(%esp)
movss 20(%esp), %xmm0
movss 48(%esp), %xmm2
unpcklps %xmm0, %xmm2
unpcklps %xmm1, %xmm2
movl 80(%esp), %eax
movaps %xmm2, (%eax)
addl $76, %esp
ret
GCC produces this (with -fomit-frame-pointer):
_test:
subl $12, %esp
movl 20(%esp), %eax
movaps (%eax), %xmm0
movl 24(%esp), %eax
unpcklps (%eax), %xmm0
movl 16(%esp), %eax
movaps %xmm0, (%eax)
addl $12, %esp
ret
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27233 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-28 20:28:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand DAGCombiner::visitCONCAT_VECTORS(SDNode *N) {
|
|
|
|
// TODO: Check to see if this is a CONCAT_VECTORS of a bunch of
|
|
|
|
// EXTRACT_SUBVECTOR operations. If so, and if the EXTRACT_SUBVECTOR vector
|
|
|
|
// inputs come from at most two distinct vectors, turn this into a shuffle
|
|
|
|
// node.
|
2006-03-31 22:16:43 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If we only have one input vector, we don't need to do any concatenation.
|
|
|
|
if (N->getNumOperands() == 1) {
|
|
|
|
return N->getOperand(0);
|
2006-07-21 08:25:53 +00:00
|
|
|
}
|
|
|
|
|
2006-03-28 22:11:53 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand DAGCombiner::visitVECTOR_SHUFFLE(SDNode *N) {
|
2006-03-31 22:16:43 +00:00
|
|
|
SDOperand ShufMask = N->getOperand(2);
|
2007-06-25 16:23:39 +00:00
|
|
|
unsigned NumElts = ShufMask.getNumOperands();
|
|
|
|
|
2006-03-31 22:16:43 +00:00
|
|
|
// If the shuffle mask is an identity operation on the LHS, return the LHS.
|
|
|
|
bool isIdentity = true;
|
|
|
|
for (unsigned i = 0; i != NumElts; ++i) {
|
|
|
|
if (ShufMask.getOperand(i).getOpcode() != ISD::UNDEF &&
|
|
|
|
cast<ConstantSDNode>(ShufMask.getOperand(i))->getValue() != i) {
|
|
|
|
isIdentity = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (isIdentity) return N->getOperand(0);
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2006-03-31 22:16:43 +00:00
|
|
|
// If the shuffle mask is an identity operation on the RHS, return the RHS.
|
|
|
|
isIdentity = true;
|
|
|
|
for (unsigned i = 0; i != NumElts; ++i) {
|
|
|
|
if (ShufMask.getOperand(i).getOpcode() != ISD::UNDEF &&
|
|
|
|
cast<ConstantSDNode>(ShufMask.getOperand(i))->getValue() != i+NumElts) {
|
|
|
|
isIdentity = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (isIdentity) return N->getOperand(1);
|
|
|
|
|
2006-07-20 22:44:41 +00:00
|
|
|
// Check if the shuffle is a unary shuffle, i.e. one of the vectors is not
|
|
|
|
// needed at all.
|
|
|
|
bool isUnary = true;
|
2006-07-21 08:25:53 +00:00
|
|
|
bool isSplat = true;
|
2006-07-20 22:44:41 +00:00
|
|
|
int VecNum = -1;
|
2006-07-25 20:44:41 +00:00
|
|
|
unsigned BaseIdx = 0;
|
2006-07-20 22:44:41 +00:00
|
|
|
for (unsigned i = 0; i != NumElts; ++i)
|
|
|
|
if (ShufMask.getOperand(i).getOpcode() != ISD::UNDEF) {
|
|
|
|
unsigned Idx = cast<ConstantSDNode>(ShufMask.getOperand(i))->getValue();
|
|
|
|
int V = (Idx < NumElts) ? 0 : 1;
|
2006-07-21 08:25:53 +00:00
|
|
|
if (VecNum == -1) {
|
2006-07-20 22:44:41 +00:00
|
|
|
VecNum = V;
|
2006-07-21 08:25:53 +00:00
|
|
|
BaseIdx = Idx;
|
|
|
|
} else {
|
|
|
|
if (BaseIdx != Idx)
|
|
|
|
isSplat = false;
|
|
|
|
if (VecNum != V) {
|
|
|
|
isUnary = false;
|
|
|
|
break;
|
|
|
|
}
|
2006-07-20 22:44:41 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
SDOperand N0 = N->getOperand(0);
|
|
|
|
SDOperand N1 = N->getOperand(1);
|
|
|
|
// Normalize unary shuffle so the RHS is undef.
|
|
|
|
if (isUnary && VecNum == 1)
|
|
|
|
std::swap(N0, N1);
|
|
|
|
|
2006-07-21 08:25:53 +00:00
|
|
|
// If it is a splat, check if the argument vector is a build_vector with
|
|
|
|
// all scalar elements the same.
|
|
|
|
if (isSplat) {
|
|
|
|
SDNode *V = N0.Val;
|
Be careful when looking through a vbit_convert. Optimizing this:
(vector_shuffle
(vbitconvert (vbuildvector (copyfromreg v4f32), 1, v4f32), 4, f32),
(undef, undef, undef, undef), (0, 0, 0, 0), 4, f32)
to the
vbitconvert
is a very bad idea.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30989 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-16 22:49:37 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If this is a bit convert that changes the element type of the vector but
|
Be careful when looking through a vbit_convert. Optimizing this:
(vector_shuffle
(vbitconvert (vbuildvector (copyfromreg v4f32), 1, v4f32), 4, f32),
(undef, undef, undef, undef), (0, 0, 0, 0), 4, f32)
to the
vbitconvert
is a very bad idea.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30989 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-16 22:49:37 +00:00
|
|
|
// not the number of vector elements, look through it. Be careful not to
|
|
|
|
// look though conversions that change things like v4f32 to v2f64.
|
2007-06-25 16:23:39 +00:00
|
|
|
if (V->getOpcode() == ISD::BIT_CONVERT) {
|
Be careful when looking through a vbit_convert. Optimizing this:
(vector_shuffle
(vbitconvert (vbuildvector (copyfromreg v4f32), 1, v4f32), 4, f32),
(undef, undef, undef, undef), (0, 0, 0, 0), 4, f32)
to the
vbitconvert
is a very bad idea.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30989 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-16 22:49:37 +00:00
|
|
|
SDOperand ConvInput = V->getOperand(0);
|
2008-06-06 12:08:01 +00:00
|
|
|
if (ConvInput.getValueType().getVectorNumElements() == NumElts)
|
Be careful when looking through a vbit_convert. Optimizing this:
(vector_shuffle
(vbitconvert (vbuildvector (copyfromreg v4f32), 1, v4f32), 4, f32),
(undef, undef, undef, undef), (0, 0, 0, 0), 4, f32)
to the
vbitconvert
is a very bad idea.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30989 91177308-0d34-0410-b5e6-96231b3b80d8
2006-10-16 22:49:37 +00:00
|
|
|
V = ConvInput.Val;
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
if (V->getOpcode() == ISD::BUILD_VECTOR) {
|
|
|
|
unsigned NumElems = V->getNumOperands();
|
2006-07-21 08:25:53 +00:00
|
|
|
if (NumElems > BaseIdx) {
|
|
|
|
SDOperand Base;
|
|
|
|
bool AllSame = true;
|
|
|
|
for (unsigned i = 0; i != NumElems; ++i) {
|
|
|
|
if (V->getOperand(i).getOpcode() != ISD::UNDEF) {
|
|
|
|
Base = V->getOperand(i);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Splat of <u, u, u, u>, return <u, u, u, u>
|
|
|
|
if (!Base.Val)
|
|
|
|
return N0;
|
|
|
|
for (unsigned i = 0; i != NumElems; ++i) {
|
Fix a bogus splat xform:
shuffle <undef, undef, x, undef>, <undef, undef, undef, undef>, <2, 2, 2, 2>
!=
<undef, undef, x, undef>
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@42111 91177308-0d34-0410-b5e6-96231b3b80d8
2007-09-18 21:54:37 +00:00
|
|
|
if (V->getOperand(i) != Base) {
|
2006-07-21 08:25:53 +00:00
|
|
|
AllSame = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Splat of <x, x, x, x>, return <x, x, x, x>
|
|
|
|
if (AllSame)
|
|
|
|
return N0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-07-20 22:44:41 +00:00
|
|
|
// If it is a unary or the LHS and the RHS are the same node, turn the RHS
|
|
|
|
// into an undef.
|
|
|
|
if (isUnary || N0 == N1) {
|
2006-04-08 05:34:25 +00:00
|
|
|
// Check the SHUFFLE mask, mapping any inputs from the 2nd operand into the
|
|
|
|
// first operand.
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> MappedOps;
|
2006-04-08 05:34:25 +00:00
|
|
|
for (unsigned i = 0; i != NumElts; ++i) {
|
|
|
|
if (ShufMask.getOperand(i).getOpcode() == ISD::UNDEF ||
|
|
|
|
cast<ConstantSDNode>(ShufMask.getOperand(i))->getValue() < NumElts) {
|
|
|
|
MappedOps.push_back(ShufMask.getOperand(i));
|
|
|
|
} else {
|
|
|
|
unsigned NewIdx =
|
|
|
|
cast<ConstantSDNode>(ShufMask.getOperand(i))->getValue() - NumElts;
|
|
|
|
MappedOps.push_back(DAG.getConstant(NewIdx, MVT::i32));
|
|
|
|
}
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
ShufMask = DAG.getNode(ISD::BUILD_VECTOR, ShufMask.getValueType(),
|
2006-08-08 02:23:42 +00:00
|
|
|
&MappedOps[0], MappedOps.size());
|
2006-04-08 05:34:25 +00:00
|
|
|
AddToWorkList(ShufMask.Val);
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::VECTOR_SHUFFLE, N->getValueType(0),
|
|
|
|
N0,
|
|
|
|
DAG.getNode(ISD::UNDEF, N->getValueType(0)),
|
|
|
|
ShufMask);
|
2006-04-08 05:34:25 +00:00
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2006-03-31 22:16:43 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2006-04-20 08:56:16 +00:00
|
|
|
/// XformToShuffleWithZero - Returns a vector_shuffle if it able to transform
|
2007-06-25 16:23:39 +00:00
|
|
|
/// an AND to a vector_shuffle with the destination vector and a zero vector.
|
|
|
|
/// e.g. AND V, <0xffffffff, 0, 0xffffffff, 0>. ==>
|
2006-04-20 08:56:16 +00:00
|
|
|
/// vector_shuffle V, Zero, <0, 4, 2, 4>
|
|
|
|
SDOperand DAGCombiner::XformToShuffleWithZero(SDNode *N) {
|
|
|
|
SDOperand LHS = N->getOperand(0);
|
|
|
|
SDOperand RHS = N->getOperand(1);
|
2007-06-25 16:23:39 +00:00
|
|
|
if (N->getOpcode() == ISD::AND) {
|
|
|
|
if (RHS.getOpcode() == ISD::BIT_CONVERT)
|
2006-04-20 08:56:16 +00:00
|
|
|
RHS = RHS.getOperand(0);
|
2007-06-25 16:23:39 +00:00
|
|
|
if (RHS.getOpcode() == ISD::BUILD_VECTOR) {
|
2006-04-20 08:56:16 +00:00
|
|
|
std::vector<SDOperand> IdxOps;
|
|
|
|
unsigned NumOps = RHS.getNumOperands();
|
2007-06-25 16:23:39 +00:00
|
|
|
unsigned NumElts = NumOps;
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EVT = RHS.getValueType().getVectorElementType();
|
2006-04-20 08:56:16 +00:00
|
|
|
for (unsigned i = 0; i != NumElts; ++i) {
|
|
|
|
SDOperand Elt = RHS.getOperand(i);
|
|
|
|
if (!isa<ConstantSDNode>(Elt))
|
|
|
|
return SDOperand();
|
|
|
|
else if (cast<ConstantSDNode>(Elt)->isAllOnesValue())
|
|
|
|
IdxOps.push_back(DAG.getConstant(i, EVT));
|
|
|
|
else if (cast<ConstantSDNode>(Elt)->isNullValue())
|
|
|
|
IdxOps.push_back(DAG.getConstant(NumElts, EVT));
|
|
|
|
else
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Let's see if the target supports this vector_shuffle.
|
|
|
|
if (!TLI.isVectorClearMaskLegal(IdxOps, EVT, DAG))
|
|
|
|
return SDOperand();
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// Return the new VECTOR_SHUFFLE node.
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = MVT::getVectorVT(EVT, NumElts);
|
2006-04-20 08:56:16 +00:00
|
|
|
std::vector<SDOperand> Ops;
|
2007-06-25 16:23:39 +00:00
|
|
|
LHS = DAG.getNode(ISD::BIT_CONVERT, VT, LHS);
|
2006-04-20 08:56:16 +00:00
|
|
|
Ops.push_back(LHS);
|
|
|
|
AddToWorkList(LHS.Val);
|
|
|
|
std::vector<SDOperand> ZeroOps(NumElts, DAG.getConstant(0, EVT));
|
2007-06-25 16:23:39 +00:00
|
|
|
Ops.push_back(DAG.getNode(ISD::BUILD_VECTOR, VT,
|
2006-08-08 02:23:42 +00:00
|
|
|
&ZeroOps[0], ZeroOps.size()));
|
2007-06-25 16:23:39 +00:00
|
|
|
Ops.push_back(DAG.getNode(ISD::BUILD_VECTOR, VT,
|
2006-08-08 02:23:42 +00:00
|
|
|
&IdxOps[0], IdxOps.size()));
|
2007-06-25 16:23:39 +00:00
|
|
|
SDOperand Result = DAG.getNode(ISD::VECTOR_SHUFFLE, VT,
|
2006-08-08 02:23:42 +00:00
|
|
|
&Ops[0], Ops.size());
|
2007-06-25 16:23:39 +00:00
|
|
|
if (VT != LHS.getValueType()) {
|
|
|
|
Result = DAG.getNode(ISD::BIT_CONVERT, LHS.getValueType(), Result);
|
2006-04-20 08:56:16 +00:00
|
|
|
}
|
|
|
|
return Result;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
/// SimplifyVBinOp - Visit a binary vector operation, like ADD.
|
|
|
|
SDOperand DAGCombiner::SimplifyVBinOp(SDNode *N) {
|
|
|
|
// After legalize, the target may be depending on adds and other
|
|
|
|
// binary ops to provide legal ways to construct constants or other
|
|
|
|
// things. Simplifying them may result in a loss of legality.
|
|
|
|
if (AfterLegalize) return SDOperand();
|
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N->getValueType(0);
|
|
|
|
assert(VT.isVector() && "SimplifyVBinOp only works on vectors!");
|
2007-06-25 16:23:39 +00:00
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT EltType = VT.getVectorElementType();
|
Constant fold all of the vector binops. This allows us to compile this:
"vector unsigned char mergeLowHigh = (vector unsigned char)
( 8, 9, 10, 11, 16, 17, 18, 19, 12, 13, 14, 15, 20, 21, 22, 23 );
vector unsigned char mergeHighLow = vec_xor( mergeLowHigh, vec_splat_u8(8));"
aka:
void %test2(<16 x sbyte>* %P) {
store <16 x sbyte> cast (<4 x int> xor (<4 x int> cast (<16 x ubyte> < ubyte 8, ubyte 9, ubyte 10, ubyte 11, ubyte 16, ubyte 17, ubyte 18, ubyte 19, ubyte 12, ubyte 13, ubyte 14, ubyte 15, ubyte 20, ubyte 21, ubyte 22, ubyte 23 > to <4 x int>), <4 x int> cast (<16 x sbyte> < sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8 > to <4 x int>)) to <16 x sbyte>), <16 x sbyte> * %P
ret void
}
into this:
_test2:
mfspr r2, 256
oris r4, r2, 32768
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
lvx v0, r5, r4
stvx v0, 0, r3
mtspr 256, r2
blr
instead of this:
_test2:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
vspltisb v0, 8
lvx v1, r5, r4
vxor v0, v1, v0
stvx v0, 0, r3
mtspr 256, r2
blr
... which occurs here:
http://developer.apple.com/hardware/ve/calcspeed.html
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27343 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 03:25:57 +00:00
|
|
|
SDOperand LHS = N->getOperand(0);
|
|
|
|
SDOperand RHS = N->getOperand(1);
|
2006-04-20 08:56:16 +00:00
|
|
|
SDOperand Shuffle = XformToShuffleWithZero(N);
|
|
|
|
if (Shuffle.Val) return Shuffle;
|
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
// If the LHS and RHS are BUILD_VECTOR nodes, see if we can constant fold
|
Constant fold all of the vector binops. This allows us to compile this:
"vector unsigned char mergeLowHigh = (vector unsigned char)
( 8, 9, 10, 11, 16, 17, 18, 19, 12, 13, 14, 15, 20, 21, 22, 23 );
vector unsigned char mergeHighLow = vec_xor( mergeLowHigh, vec_splat_u8(8));"
aka:
void %test2(<16 x sbyte>* %P) {
store <16 x sbyte> cast (<4 x int> xor (<4 x int> cast (<16 x ubyte> < ubyte 8, ubyte 9, ubyte 10, ubyte 11, ubyte 16, ubyte 17, ubyte 18, ubyte 19, ubyte 12, ubyte 13, ubyte 14, ubyte 15, ubyte 20, ubyte 21, ubyte 22, ubyte 23 > to <4 x int>), <4 x int> cast (<16 x sbyte> < sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8 > to <4 x int>)) to <16 x sbyte>), <16 x sbyte> * %P
ret void
}
into this:
_test2:
mfspr r2, 256
oris r4, r2, 32768
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
lvx v0, r5, r4
stvx v0, 0, r3
mtspr 256, r2
blr
instead of this:
_test2:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
vspltisb v0, 8
lvx v1, r5, r4
vxor v0, v1, v0
stvx v0, 0, r3
mtspr 256, r2
blr
... which occurs here:
http://developer.apple.com/hardware/ve/calcspeed.html
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27343 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 03:25:57 +00:00
|
|
|
// this operation.
|
2007-06-25 16:23:39 +00:00
|
|
|
if (LHS.getOpcode() == ISD::BUILD_VECTOR &&
|
|
|
|
RHS.getOpcode() == ISD::BUILD_VECTOR) {
|
2006-08-08 02:23:42 +00:00
|
|
|
SmallVector<SDOperand, 8> Ops;
|
2007-06-25 16:23:39 +00:00
|
|
|
for (unsigned i = 0, e = LHS.getNumOperands(); i != e; ++i) {
|
Constant fold all of the vector binops. This allows us to compile this:
"vector unsigned char mergeLowHigh = (vector unsigned char)
( 8, 9, 10, 11, 16, 17, 18, 19, 12, 13, 14, 15, 20, 21, 22, 23 );
vector unsigned char mergeHighLow = vec_xor( mergeLowHigh, vec_splat_u8(8));"
aka:
void %test2(<16 x sbyte>* %P) {
store <16 x sbyte> cast (<4 x int> xor (<4 x int> cast (<16 x ubyte> < ubyte 8, ubyte 9, ubyte 10, ubyte 11, ubyte 16, ubyte 17, ubyte 18, ubyte 19, ubyte 12, ubyte 13, ubyte 14, ubyte 15, ubyte 20, ubyte 21, ubyte 22, ubyte 23 > to <4 x int>), <4 x int> cast (<16 x sbyte> < sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8 > to <4 x int>)) to <16 x sbyte>), <16 x sbyte> * %P
ret void
}
into this:
_test2:
mfspr r2, 256
oris r4, r2, 32768
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
lvx v0, r5, r4
stvx v0, 0, r3
mtspr 256, r2
blr
instead of this:
_test2:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
vspltisb v0, 8
lvx v1, r5, r4
vxor v0, v1, v0
stvx v0, 0, r3
mtspr 256, r2
blr
... which occurs here:
http://developer.apple.com/hardware/ve/calcspeed.html
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27343 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 03:25:57 +00:00
|
|
|
SDOperand LHSOp = LHS.getOperand(i);
|
|
|
|
SDOperand RHSOp = RHS.getOperand(i);
|
|
|
|
// If these two elements can't be folded, bail out.
|
|
|
|
if ((LHSOp.getOpcode() != ISD::UNDEF &&
|
|
|
|
LHSOp.getOpcode() != ISD::Constant &&
|
|
|
|
LHSOp.getOpcode() != ISD::ConstantFP) ||
|
|
|
|
(RHSOp.getOpcode() != ISD::UNDEF &&
|
|
|
|
RHSOp.getOpcode() != ISD::Constant &&
|
|
|
|
RHSOp.getOpcode() != ISD::ConstantFP))
|
|
|
|
break;
|
2006-05-31 06:08:35 +00:00
|
|
|
// Can't fold divide by zero.
|
2007-06-25 16:23:39 +00:00
|
|
|
if (N->getOpcode() == ISD::SDIV || N->getOpcode() == ISD::UDIV ||
|
|
|
|
N->getOpcode() == ISD::FDIV) {
|
2006-05-31 06:08:35 +00:00
|
|
|
if ((RHSOp.getOpcode() == ISD::Constant &&
|
|
|
|
cast<ConstantSDNode>(RHSOp.Val)->isNullValue()) ||
|
|
|
|
(RHSOp.getOpcode() == ISD::ConstantFP &&
|
2007-08-31 23:34:27 +00:00
|
|
|
cast<ConstantFPSDNode>(RHSOp.Val)->getValueAPF().isZero()))
|
2006-05-31 06:08:35 +00:00
|
|
|
break;
|
|
|
|
}
|
2007-06-25 16:23:39 +00:00
|
|
|
Ops.push_back(DAG.getNode(N->getOpcode(), EltType, LHSOp, RHSOp));
|
2006-04-08 04:15:24 +00:00
|
|
|
AddToWorkList(Ops.back().Val);
|
Constant fold all of the vector binops. This allows us to compile this:
"vector unsigned char mergeLowHigh = (vector unsigned char)
( 8, 9, 10, 11, 16, 17, 18, 19, 12, 13, 14, 15, 20, 21, 22, 23 );
vector unsigned char mergeHighLow = vec_xor( mergeLowHigh, vec_splat_u8(8));"
aka:
void %test2(<16 x sbyte>* %P) {
store <16 x sbyte> cast (<4 x int> xor (<4 x int> cast (<16 x ubyte> < ubyte 8, ubyte 9, ubyte 10, ubyte 11, ubyte 16, ubyte 17, ubyte 18, ubyte 19, ubyte 12, ubyte 13, ubyte 14, ubyte 15, ubyte 20, ubyte 21, ubyte 22, ubyte 23 > to <4 x int>), <4 x int> cast (<16 x sbyte> < sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8 > to <4 x int>)) to <16 x sbyte>), <16 x sbyte> * %P
ret void
}
into this:
_test2:
mfspr r2, 256
oris r4, r2, 32768
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
lvx v0, r5, r4
stvx v0, 0, r3
mtspr 256, r2
blr
instead of this:
_test2:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
vspltisb v0, 8
lvx v1, r5, r4
vxor v0, v1, v0
stvx v0, 0, r3
mtspr 256, r2
blr
... which occurs here:
http://developer.apple.com/hardware/ve/calcspeed.html
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27343 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 03:25:57 +00:00
|
|
|
assert((Ops.back().getOpcode() == ISD::UNDEF ||
|
|
|
|
Ops.back().getOpcode() == ISD::Constant ||
|
|
|
|
Ops.back().getOpcode() == ISD::ConstantFP) &&
|
|
|
|
"Scalar binop didn't fold!");
|
|
|
|
}
|
2006-04-03 17:21:50 +00:00
|
|
|
|
2007-06-25 16:23:39 +00:00
|
|
|
if (Ops.size() == LHS.getNumOperands()) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = LHS.getValueType();
|
2007-06-25 16:23:39 +00:00
|
|
|
return DAG.getNode(ISD::BUILD_VECTOR, VT, &Ops[0], Ops.size());
|
2006-04-03 17:21:50 +00:00
|
|
|
}
|
Constant fold all of the vector binops. This allows us to compile this:
"vector unsigned char mergeLowHigh = (vector unsigned char)
( 8, 9, 10, 11, 16, 17, 18, 19, 12, 13, 14, 15, 20, 21, 22, 23 );
vector unsigned char mergeHighLow = vec_xor( mergeLowHigh, vec_splat_u8(8));"
aka:
void %test2(<16 x sbyte>* %P) {
store <16 x sbyte> cast (<4 x int> xor (<4 x int> cast (<16 x ubyte> < ubyte 8, ubyte 9, ubyte 10, ubyte 11, ubyte 16, ubyte 17, ubyte 18, ubyte 19, ubyte 12, ubyte 13, ubyte 14, ubyte 15, ubyte 20, ubyte 21, ubyte 22, ubyte 23 > to <4 x int>), <4 x int> cast (<16 x sbyte> < sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8, sbyte 8 > to <4 x int>)) to <16 x sbyte>), <16 x sbyte> * %P
ret void
}
into this:
_test2:
mfspr r2, 256
oris r4, r2, 32768
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
lvx v0, r5, r4
stvx v0, 0, r3
mtspr 256, r2
blr
instead of this:
_test2:
mfspr r2, 256
oris r4, r2, 49152
mtspr 256, r4
li r4, lo16(LCPI2_0)
lis r5, ha16(LCPI2_0)
vspltisb v0, 8
lvx v1, r5, r4
vxor v0, v1, v0
stvx v0, 0, r3
mtspr 256, r2
blr
... which occurs here:
http://developer.apple.com/hardware/ve/calcspeed.html
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27343 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-02 03:25:57 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand DAGCombiner::SimplifySelect(SDOperand N0, SDOperand N1, SDOperand N2){
|
2005-10-08 00:29:44 +00:00
|
|
|
assert(N0.getOpcode() ==ISD::SETCC && "First argument must be a SetCC node!");
|
|
|
|
|
|
|
|
SDOperand SCC = SimplifySelectCC(N0.getOperand(0), N0.getOperand(1), N1, N2,
|
|
|
|
cast<CondCodeSDNode>(N0.getOperand(2))->get());
|
|
|
|
// If we got a simplified select_cc node back from SimplifySelectCC, then
|
|
|
|
// break it down into a new SETCC node, and a new SELECT node, and then return
|
|
|
|
// the SELECT node, since we were called with a SELECT node.
|
|
|
|
if (SCC.Val) {
|
|
|
|
// Check to see if we got a select_cc back (to turn into setcc/select).
|
|
|
|
// Otherwise, just return whatever node we got back, like fabs.
|
|
|
|
if (SCC.getOpcode() == ISD::SELECT_CC) {
|
|
|
|
SDOperand SETCC = DAG.getNode(ISD::SETCC, N0.getValueType(),
|
|
|
|
SCC.getOperand(0), SCC.getOperand(1),
|
|
|
|
SCC.getOperand(4));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(SETCC.Val);
|
2005-10-08 00:29:44 +00:00
|
|
|
return DAG.getNode(ISD::SELECT, SCC.getValueType(), SCC.getOperand(2),
|
|
|
|
SCC.getOperand(3), SETCC);
|
|
|
|
}
|
|
|
|
return SCC;
|
|
|
|
}
|
2005-09-19 22:34:01 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
/// SimplifySelectOps - Given a SELECT or a SELECT_CC node, where LHS and RHS
|
|
|
|
/// are the two values being selected between, see if we can simplify the
|
2006-05-27 00:43:02 +00:00
|
|
|
/// select. Callers of this should assume that TheSelect is deleted if this
|
|
|
|
/// returns true. As such, they should return the appropriate thing (e.g. the
|
|
|
|
/// node) back to the top-level of the DAG combiner loop to avoid it being
|
|
|
|
/// looked at.
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
///
|
|
|
|
bool DAGCombiner::SimplifySelectOps(SDNode *TheSelect, SDOperand LHS,
|
|
|
|
SDOperand RHS) {
|
|
|
|
|
|
|
|
// If this is a select from two identical things, try to pull the operation
|
|
|
|
// through the select.
|
|
|
|
if (LHS.getOpcode() == RHS.getOpcode() && LHS.hasOneUse() && RHS.hasOneUse()){
|
|
|
|
// If this is a load and the token chain is identical, replace the select
|
|
|
|
// of two loads with a load through a select of the address to load from.
|
|
|
|
// This triggers in things like "select bool X, 10.0, 123.0" after the FP
|
|
|
|
// constants have been dropped into the constant pool.
|
2006-10-09 20:57:25 +00:00
|
|
|
if (LHS.getOpcode() == ISD::LOAD &&
|
Disable some DAG combiner optimizations that may be
wrong for volatile loads and stores. In fact this
is almost all of them! There are three types of
problems: (1) it is wrong to change the width of
a volatile memory access. These may be used to
do memory mapped i/o, in which case a load can have
an effect even if the result is not used. Consider
loading an i32 but only using the lower 8 bits. It
is wrong to change this into a load of an i8, because
you are no longer tickling the other three bytes. It
is also unwise to make a load/store wider. For
example, changing an i16 load into an i32 load is
wrong no matter how aligned things are, since the
fact of loading an additional 2 bytes can have
i/o side-effects. (2) it is wrong to change the
number of volatile load/stores: they may be counted
by the hardware. (3) it is wrong to change a volatile
load/store that requires one memory access into one
that requires several. For example on x86-32, you
can store a double in one processor operation, but to
store an i64 requires two (two i32 stores). In a
multi-threaded program you may want to bitcast an i64
to a double and store as a double because that will
occur atomically, and be indivisible to other threads.
So it would be wrong to convert the store-of-double
into a store of an i64, because this will become two
i32 stores - no longer atomic. My policy here is
to say that the number of processor operations for
an illegal operation is undefined. So it is alright
to change a store of an i64 (requires at least two
stores; but could be validly lowered to memcpy for
example) into a store of double (one processor op).
In short, if the new store is legal and has the same
size then I say that the transform is ok. It would
also be possible to say that transforms are always
ok if before they were illegal, whether after they
are illegal or not, but that's more awkward to do
and I doubt it buys us anything much.
However this exposed an interesting thing - on x86-32
a store of i64 is considered legal! That is because
operations are marked legal by default, regardless of
whether the type is legal or not. In some ways this
is clever: before type legalization this means that
operations on illegal types are considered legal;
after type legalization there are no illegal types
so now operations are only legal if they really are.
But I consider this to be too cunning for mere mortals.
Better to do things explicitly by testing AfterLegalize.
So I have changed things so that operations with illegal
types are considered illegal - indeed they can never
map to a machine operation. However this means that
the DAG combiner is more conservative because before
it was "accidentally" performing transforms where the
type was illegal because the operation was nonetheless
marked legal. So in a few such places I added a check
on AfterLegalize, which I suppose was actually just
forgotten before. This causes the DAG combiner to do
slightly more than it used to, which resulted in the X86
backend blowing up because it got a slightly surprising
node it wasn't expecting, so I tweaked it.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@52254 91177308-0d34-0410-b5e6-96231b3b80d8
2008-06-13 19:07:40 +00:00
|
|
|
// Do not let this transformation reduce the number of volatile loads.
|
|
|
|
!cast<LoadSDNode>(LHS)->isVolatile() &&
|
|
|
|
!cast<LoadSDNode>(RHS)->isVolatile() &&
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
// Token chains must be identical.
|
2006-10-09 20:57:25 +00:00
|
|
|
LHS.getOperand(0) == RHS.getOperand(0)) {
|
|
|
|
LoadSDNode *LLD = cast<LoadSDNode>(LHS);
|
|
|
|
LoadSDNode *RLD = cast<LoadSDNode>(RHS);
|
|
|
|
|
|
|
|
// If this is an EXTLOAD, the VT's must match.
|
2008-01-30 00:15:11 +00:00
|
|
|
if (LLD->getMemoryVT() == RLD->getMemoryVT()) {
|
2006-10-09 20:57:25 +00:00
|
|
|
// FIXME: this conflates two src values, discarding one. This is not
|
|
|
|
// the right thing to do, but nothing uses srcvalues now. When they do,
|
|
|
|
// turn SrcValue into a list of locations.
|
|
|
|
SDOperand Addr;
|
2007-01-16 05:59:59 +00:00
|
|
|
if (TheSelect->getOpcode() == ISD::SELECT) {
|
|
|
|
// Check that the condition doesn't reach either load. If so, folding
|
|
|
|
// this will induce a cycle into the DAG.
|
2008-03-04 00:41:45 +00:00
|
|
|
if (!LLD->isPredecessorOf(TheSelect->getOperand(0).Val) &&
|
|
|
|
!RLD->isPredecessorOf(TheSelect->getOperand(0).Val)) {
|
2007-01-16 05:59:59 +00:00
|
|
|
Addr = DAG.getNode(ISD::SELECT, LLD->getBasePtr().getValueType(),
|
|
|
|
TheSelect->getOperand(0), LLD->getBasePtr(),
|
|
|
|
RLD->getBasePtr());
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Check that the condition doesn't reach either load. If so, folding
|
|
|
|
// this will induce a cycle into the DAG.
|
2008-03-04 00:41:45 +00:00
|
|
|
if (!LLD->isPredecessorOf(TheSelect->getOperand(0).Val) &&
|
|
|
|
!RLD->isPredecessorOf(TheSelect->getOperand(0).Val) &&
|
|
|
|
!LLD->isPredecessorOf(TheSelect->getOperand(1).Val) &&
|
|
|
|
!RLD->isPredecessorOf(TheSelect->getOperand(1).Val)) {
|
2007-01-16 05:59:59 +00:00
|
|
|
Addr = DAG.getNode(ISD::SELECT_CC, LLD->getBasePtr().getValueType(),
|
2006-10-09 20:57:25 +00:00
|
|
|
TheSelect->getOperand(0),
|
|
|
|
TheSelect->getOperand(1),
|
|
|
|
LLD->getBasePtr(), RLD->getBasePtr(),
|
|
|
|
TheSelect->getOperand(4));
|
2007-01-16 05:59:59 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (Addr.Val) {
|
|
|
|
SDOperand Load;
|
|
|
|
if (LLD->getExtensionType() == ISD::NON_EXTLOAD)
|
|
|
|
Load = DAG.getLoad(TheSelect->getValueType(0), LLD->getChain(),
|
|
|
|
Addr,LLD->getSrcValue(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LLD->getSrcValueOffset(),
|
|
|
|
LLD->isVolatile(),
|
|
|
|
LLD->getAlignment());
|
2007-01-16 05:59:59 +00:00
|
|
|
else {
|
|
|
|
Load = DAG.getExtLoad(LLD->getExtensionType(),
|
|
|
|
TheSelect->getValueType(0),
|
|
|
|
LLD->getChain(), Addr, LLD->getSrcValue(),
|
|
|
|
LLD->getSrcValueOffset(),
|
2008-01-30 00:15:11 +00:00
|
|
|
LLD->getMemoryVT(),
|
2007-04-22 23:15:30 +00:00
|
|
|
LLD->isVolatile(),
|
|
|
|
LLD->getAlignment());
|
2007-01-16 05:59:59 +00:00
|
|
|
}
|
|
|
|
// Users of the select now use the result of the load.
|
|
|
|
CombineTo(TheSelect, Load);
|
|
|
|
|
|
|
|
// Users of the old loads now use the new load's chain. We know the
|
|
|
|
// old-load value is dead now.
|
|
|
|
CombineTo(LHS.Val, Load.getValue(0), Load.getValue(1));
|
|
|
|
CombineTo(RHS.Val, Load.getValue(0), Load.getValue(1));
|
|
|
|
return true;
|
2006-10-09 20:57:25 +00:00
|
|
|
}
|
|
|
|
}
|
Fold (select C, load A, load B) -> load (select C, A, B). This happens quite
a lot throughout many programs. In particular, specfp triggers it a bunch for
constant FP nodes when you have code like cond ? 1.0 : -1.0.
If the PPC ISel exposed the loads implicit in pic references to external globals,
we would be able to eliminate a load in cases like this as well:
%X = external global int
%Y = external global int
int* %test4(bool %C) {
%G = select bool %C, int* %X, int* %Y
ret int* %G
}
Note that this breaks things that use SrcValue's (see the fixme), but since nothing
uses them yet, this is ok.
Also, simplify some code to use hasOneUse() on an SDOperand instead of hasNUsesOfValue directly.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@23781 91177308-0d34-0410-b5e6-96231b3b80d8
2005-10-18 06:04:22 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
SDOperand DAGCombiner::SimplifySelectCC(SDOperand N0, SDOperand N1,
|
|
|
|
SDOperand N2, SDOperand N3,
|
2007-04-11 06:50:51 +00:00
|
|
|
ISD::CondCode CC, bool NotExtCompare) {
|
2005-10-08 00:29:44 +00:00
|
|
|
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT VT = N2.getValueType();
|
2005-10-08 00:29:44 +00:00
|
|
|
ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.Val);
|
|
|
|
ConstantSDNode *N2C = dyn_cast<ConstantSDNode>(N2.Val);
|
|
|
|
ConstantSDNode *N3C = dyn_cast<ConstantSDNode>(N3.Val);
|
|
|
|
|
|
|
|
// Determine if the condition we're dealing with is constant
|
2008-03-10 15:42:14 +00:00
|
|
|
SDOperand SCC = SimplifySetCC(TLI.getSetCCResultType(N0), N0, N1, CC, false);
|
2006-10-14 03:52:46 +00:00
|
|
|
if (SCC.Val) AddToWorkList(SCC.Val);
|
2005-10-08 00:29:44 +00:00
|
|
|
ConstantSDNode *SCCC = dyn_cast_or_null<ConstantSDNode>(SCC.Val);
|
|
|
|
|
|
|
|
// fold select_cc true, x, y -> x
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SCCC && !SCCC->isNullValue())
|
2005-10-08 00:29:44 +00:00
|
|
|
return N2;
|
|
|
|
// fold select_cc false, x, y -> y
|
2008-03-13 22:13:53 +00:00
|
|
|
if (SCCC && SCCC->isNullValue())
|
2005-10-08 00:29:44 +00:00
|
|
|
return N3;
|
|
|
|
|
|
|
|
// Check to see if we can simplify the select into an fabs node
|
|
|
|
if (ConstantFPSDNode *CFP = dyn_cast<ConstantFPSDNode>(N1)) {
|
|
|
|
// Allow either -0.0 or 0.0
|
2007-08-25 22:10:57 +00:00
|
|
|
if (CFP->getValueAPF().isZero()) {
|
2005-10-08 00:29:44 +00:00
|
|
|
// select (setg[te] X, +/-0.0), X, fneg(X) -> fabs
|
|
|
|
if ((CC == ISD::SETGE || CC == ISD::SETGT) &&
|
|
|
|
N0 == N2 && N3.getOpcode() == ISD::FNEG &&
|
|
|
|
N2 == N3.getOperand(0))
|
|
|
|
return DAG.getNode(ISD::FABS, VT, N0);
|
|
|
|
|
|
|
|
// select (setl[te] X, +/-0.0), fneg(X), X -> fabs
|
|
|
|
if ((CC == ISD::SETLT || CC == ISD::SETLE) &&
|
|
|
|
N0 == N3 && N2.getOpcode() == ISD::FNEG &&
|
|
|
|
N2.getOperand(0) == N3)
|
|
|
|
return DAG.getNode(ISD::FABS, VT, N3);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check to see if we can perform the "gzip trick", transforming
|
|
|
|
// select_cc setlt X, 0, A, 0 -> and (sra X, size(X)-1), A
|
Compile:
int test3(int a, int b) { return (a < 0) ? a : 0; }
to:
_test3:
srawi r2, r3, 31
and r3, r2, r3
blr
instead of:
_test3:
cmpwi cr0, r3, 1
li r2, 0
blt cr0, LBB2_2 ;entry
LBB2_1: ;entry
mr r3, r2
LBB2_2: ;entry
blr
This implements: PowerPC/select_lt0.ll:seli32_a_a
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30517 91177308-0d34-0410-b5e6-96231b3b80d8
2006-09-20 06:41:35 +00:00
|
|
|
if (N1C && N3C && N3C->isNullValue() && CC == ISD::SETLT &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N0.getValueType().isInteger() &&
|
|
|
|
N2.getValueType().isInteger() &&
|
2008-03-13 22:13:53 +00:00
|
|
|
(N1C->isNullValue() || // (a < 0) ? b : 0
|
|
|
|
(N1C->getAPIntValue() == 1 && N0 == N2))) { // (a < 1) ? a : 0
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT XType = N0.getValueType();
|
|
|
|
MVT AType = N2.getValueType();
|
2008-06-08 20:54:56 +00:00
|
|
|
if (XType.bitsGE(AType)) {
|
2005-10-08 00:29:44 +00:00
|
|
|
// and (sra X, size(X)-1, A) -> "and (srl X, C2), A" iff A is a
|
2005-10-10 21:26:48 +00:00
|
|
|
// single-bit constant.
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N2C && ((N2C->getAPIntValue() & (N2C->getAPIntValue()-1)) == 0)) {
|
|
|
|
unsigned ShCtV = N2C->getAPIntValue().logBase2();
|
2008-06-06 12:08:01 +00:00
|
|
|
ShCtV = XType.getSizeInBits()-ShCtV-1;
|
2005-10-08 00:29:44 +00:00
|
|
|
SDOperand ShCt = DAG.getConstant(ShCtV, TLI.getShiftAmountTy());
|
|
|
|
SDOperand Shift = DAG.getNode(ISD::SRL, XType, N0, ShCt);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Shift.Val);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (XType.bitsGT(AType)) {
|
2005-10-08 00:29:44 +00:00
|
|
|
Shift = DAG.getNode(ISD::TRUNCATE, AType, Shift);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Shift.Val);
|
2005-10-08 00:29:44 +00:00
|
|
|
}
|
|
|
|
return DAG.getNode(ISD::AND, AType, Shift, N2);
|
|
|
|
}
|
|
|
|
SDOperand Shift = DAG.getNode(ISD::SRA, XType, N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(XType.getSizeInBits()-1,
|
2005-10-08 00:29:44 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Shift.Val);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (XType.bitsGT(AType)) {
|
2005-10-08 00:29:44 +00:00
|
|
|
Shift = DAG.getNode(ISD::TRUNCATE, AType, Shift);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Shift.Val);
|
2005-10-08 00:29:44 +00:00
|
|
|
}
|
|
|
|
return DAG.getNode(ISD::AND, AType, Shift, N2);
|
|
|
|
}
|
|
|
|
}
|
2005-10-10 21:26:48 +00:00
|
|
|
|
|
|
|
// fold select C, 16, 0 -> shl C, 4
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N2C && N3C && N3C->isNullValue() && N2C->getAPIntValue().isPowerOf2() &&
|
2005-10-10 21:26:48 +00:00
|
|
|
TLI.getSetCCResultContents() == TargetLowering::ZeroOrOneSetCCResult) {
|
2007-04-11 06:50:51 +00:00
|
|
|
|
|
|
|
// If the caller doesn't want us to simplify this into a zext of a compare,
|
|
|
|
// don't do it.
|
2008-03-13 22:13:53 +00:00
|
|
|
if (NotExtCompare && N2C->getAPIntValue() == 1)
|
2007-04-11 06:50:51 +00:00
|
|
|
return SDOperand();
|
|
|
|
|
2005-10-10 21:26:48 +00:00
|
|
|
// Get a SetCC of the condition
|
|
|
|
// FIXME: Should probably make sure that setcc is legal if we ever have a
|
|
|
|
// target where it isn't.
|
2006-02-18 02:40:58 +00:00
|
|
|
SDOperand Temp, SCC;
|
2005-10-10 21:26:48 +00:00
|
|
|
// cast from setcc result type to select result type
|
2006-02-18 02:40:58 +00:00
|
|
|
if (AfterLegalize) {
|
2008-03-10 15:42:14 +00:00
|
|
|
SCC = DAG.getSetCC(TLI.getSetCCResultType(N0), N0, N1, CC);
|
2008-06-08 20:54:56 +00:00
|
|
|
if (N2.getValueType().bitsLT(SCC.getValueType()))
|
2006-12-07 22:36:47 +00:00
|
|
|
Temp = DAG.getZeroExtendInReg(SCC, N2.getValueType());
|
|
|
|
else
|
|
|
|
Temp = DAG.getNode(ISD::ZERO_EXTEND, N2.getValueType(), SCC);
|
2006-02-18 02:40:58 +00:00
|
|
|
} else {
|
|
|
|
SCC = DAG.getSetCC(MVT::i1, N0, N1, CC);
|
2005-10-10 21:26:48 +00:00
|
|
|
Temp = DAG.getNode(ISD::ZERO_EXTEND, N2.getValueType(), SCC);
|
2006-02-18 02:40:58 +00:00
|
|
|
}
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(SCC.Val);
|
|
|
|
AddToWorkList(Temp.Val);
|
2007-04-11 06:43:25 +00:00
|
|
|
|
2008-03-13 22:13:53 +00:00
|
|
|
if (N2C->getAPIntValue() == 1)
|
2007-04-11 06:43:25 +00:00
|
|
|
return Temp;
|
2005-10-10 21:26:48 +00:00
|
|
|
// shl setcc result by log2 n2c
|
|
|
|
return DAG.getNode(ISD::SHL, N2.getValueType(), Temp,
|
2008-03-13 22:13:53 +00:00
|
|
|
DAG.getConstant(N2C->getAPIntValue().logBase2(),
|
2005-10-10 21:26:48 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
|
|
|
}
|
|
|
|
|
2005-10-08 00:29:44 +00:00
|
|
|
// Check to see if this is the equivalent of setcc
|
|
|
|
// FIXME: Turn all of these into setcc if setcc if setcc is legal
|
|
|
|
// otherwise, go ahead with the folds.
|
2008-03-13 22:13:53 +00:00
|
|
|
if (0 && N3C && N3C->isNullValue() && N2C && (N2C->getAPIntValue() == 1ULL)) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT XType = N0.getValueType();
|
2008-06-14 17:48:34 +00:00
|
|
|
if (!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(ISD::SETCC, TLI.getSetCCResultType(N0))) {
|
2008-03-10 15:42:14 +00:00
|
|
|
SDOperand Res = DAG.getSetCC(TLI.getSetCCResultType(N0), N0, N1, CC);
|
2005-10-08 00:29:44 +00:00
|
|
|
if (Res.getValueType() != VT)
|
|
|
|
Res = DAG.getNode(ISD::ZERO_EXTEND, VT, Res);
|
|
|
|
return Res;
|
|
|
|
}
|
|
|
|
|
|
|
|
// seteq X, 0 -> srl (ctlz X, log2(size(X)))
|
|
|
|
if (N1C && N1C->isNullValue() && CC == ISD::SETEQ &&
|
2008-06-14 17:48:34 +00:00
|
|
|
(!AfterLegalize ||
|
|
|
|
TLI.isOperationLegal(ISD::CTLZ, XType))) {
|
2005-10-08 00:29:44 +00:00
|
|
|
SDOperand Ctlz = DAG.getNode(ISD::CTLZ, XType, N0);
|
|
|
|
return DAG.getNode(ISD::SRL, XType, Ctlz,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(Log2_32(XType.getSizeInBits()),
|
2005-10-08 00:29:44 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
|
|
|
}
|
|
|
|
// setgt X, 0 -> srl (and (-X, ~X), size(X)-1)
|
|
|
|
if (N1C && N1C->isNullValue() && CC == ISD::SETGT) {
|
|
|
|
SDOperand NegN0 = DAG.getNode(ISD::SUB, XType, DAG.getConstant(0, XType),
|
|
|
|
N0);
|
|
|
|
SDOperand NotN0 = DAG.getNode(ISD::XOR, XType, N0,
|
|
|
|
DAG.getConstant(~0ULL, XType));
|
|
|
|
return DAG.getNode(ISD::SRL, XType,
|
|
|
|
DAG.getNode(ISD::AND, XType, NegN0, NotN0),
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(XType.getSizeInBits()-1,
|
2005-10-08 00:29:44 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
|
|
|
}
|
|
|
|
// setgt X, -1 -> xor (srl (X, size(X)-1), 1)
|
|
|
|
if (N1C && N1C->isAllOnesValue() && CC == ISD::SETGT) {
|
|
|
|
SDOperand Sign = DAG.getNode(ISD::SRL, XType, N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(XType.getSizeInBits()-1,
|
2005-10-08 00:29:44 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
|
|
|
return DAG.getNode(ISD::XOR, XType, Sign, DAG.getConstant(1, XType));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check to see if this is an integer abs. select_cc setl[te] X, 0, -X, X ->
|
|
|
|
// Y = sra (X, size(X)-1); xor (add (X, Y), Y)
|
|
|
|
if (N1C && N1C->isNullValue() && (CC == ISD::SETLT || CC == ISD::SETLE) &&
|
Codegen integer abs more efficiently using the trick from the PPC CWG. This
improves codegen on many architectures. Tests committed as CodeGen/*/iabs.ll
X86 Old: X86 New:
_test: _test:
movl 4(%esp), %ecx movl 4(%esp), %eax
movl %ecx, %eax movl %eax, %ecx
negl %eax sarl $31, %ecx
testl %ecx, %ecx addl %ecx, %eax
cmovns %ecx, %eax xorl %ecx, %eax
ret ret
PPC Old: PPC New:
_test: _test:
cmpwi cr0, r3, -1 srawi r2, r3, 31
neg r2, r3 add r3, r3, r2
bgt cr0, LBB1_2 ; xor r3, r3, r2
LBB1_1: ; blr
mr r3, r2
LBB1_2: ;
blr
ARM Old: ARM New:
_test: _test:
rsb r3, r0, #0 add r3, r0, r0, asr #31
cmp r0, #0 eor r0, r3, r0, asr #31
movge r3, r0 bx lr
mov r0, r3
bx lr
Thumb Old: Thumb New:
_test: _test:
neg r2, r0 asr r2, r0, #31
cmp r0, #0 add r0, r0, r2
bge LBB1_2 eor r0, r2
LBB1_1: @ bx lr
cpy r0, r2
LBB1_2: @
bx lr
Sparc Old: Sparc New:
test: test:
save -96, %o6, %o6 save -96, %o6, %o6
sethi 0, %l0 sra %i0, 31, %l0
sub %l0, %i0, %l0 add %i0, %l0, %l1
subcc %i0, -1, %l1 xor %l1, %l0, %i0
bg .BB1_2 restore %g0, %g0, %g0
nop retl
.BB1_1: nop
or %g0, %l0, %i0
.BB1_2:
restore %g0, %g0, %g0
retl
nop
It also helps alpha/ia64 :)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35881 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:11:38 +00:00
|
|
|
N0 == N3 && N2.getOpcode() == ISD::SUB && N0 == N2.getOperand(1) &&
|
2008-06-06 12:08:01 +00:00
|
|
|
N2.getOperand(0) == N1 && N0.getValueType().isInteger()) {
|
|
|
|
MVT XType = N0.getValueType();
|
Codegen integer abs more efficiently using the trick from the PPC CWG. This
improves codegen on many architectures. Tests committed as CodeGen/*/iabs.ll
X86 Old: X86 New:
_test: _test:
movl 4(%esp), %ecx movl 4(%esp), %eax
movl %ecx, %eax movl %eax, %ecx
negl %eax sarl $31, %ecx
testl %ecx, %ecx addl %ecx, %eax
cmovns %ecx, %eax xorl %ecx, %eax
ret ret
PPC Old: PPC New:
_test: _test:
cmpwi cr0, r3, -1 srawi r2, r3, 31
neg r2, r3 add r3, r3, r2
bgt cr0, LBB1_2 ; xor r3, r3, r2
LBB1_1: ; blr
mr r3, r2
LBB1_2: ;
blr
ARM Old: ARM New:
_test: _test:
rsb r3, r0, #0 add r3, r0, r0, asr #31
cmp r0, #0 eor r0, r3, r0, asr #31
movge r3, r0 bx lr
mov r0, r3
bx lr
Thumb Old: Thumb New:
_test: _test:
neg r2, r0 asr r2, r0, #31
cmp r0, #0 add r0, r0, r2
bge LBB1_2 eor r0, r2
LBB1_1: @ bx lr
cpy r0, r2
LBB1_2: @
bx lr
Sparc Old: Sparc New:
test: test:
save -96, %o6, %o6 save -96, %o6, %o6
sethi 0, %l0 sra %i0, 31, %l0
sub %l0, %i0, %l0 add %i0, %l0, %l1
subcc %i0, -1, %l1 xor %l1, %l0, %i0
bg .BB1_2 restore %g0, %g0, %g0
nop retl
.BB1_1: nop
or %g0, %l0, %i0
.BB1_2:
restore %g0, %g0, %g0
retl
nop
It also helps alpha/ia64 :)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35881 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:11:38 +00:00
|
|
|
SDOperand Shift = DAG.getNode(ISD::SRA, XType, N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(XType.getSizeInBits()-1,
|
Codegen integer abs more efficiently using the trick from the PPC CWG. This
improves codegen on many architectures. Tests committed as CodeGen/*/iabs.ll
X86 Old: X86 New:
_test: _test:
movl 4(%esp), %ecx movl 4(%esp), %eax
movl %ecx, %eax movl %eax, %ecx
negl %eax sarl $31, %ecx
testl %ecx, %ecx addl %ecx, %eax
cmovns %ecx, %eax xorl %ecx, %eax
ret ret
PPC Old: PPC New:
_test: _test:
cmpwi cr0, r3, -1 srawi r2, r3, 31
neg r2, r3 add r3, r3, r2
bgt cr0, LBB1_2 ; xor r3, r3, r2
LBB1_1: ; blr
mr r3, r2
LBB1_2: ;
blr
ARM Old: ARM New:
_test: _test:
rsb r3, r0, #0 add r3, r0, r0, asr #31
cmp r0, #0 eor r0, r3, r0, asr #31
movge r3, r0 bx lr
mov r0, r3
bx lr
Thumb Old: Thumb New:
_test: _test:
neg r2, r0 asr r2, r0, #31
cmp r0, #0 add r0, r0, r2
bge LBB1_2 eor r0, r2
LBB1_1: @ bx lr
cpy r0, r2
LBB1_2: @
bx lr
Sparc Old: Sparc New:
test: test:
save -96, %o6, %o6 save -96, %o6, %o6
sethi 0, %l0 sra %i0, 31, %l0
sub %l0, %i0, %l0 add %i0, %l0, %l1
subcc %i0, -1, %l1 xor %l1, %l0, %i0
bg .BB1_2 restore %g0, %g0, %g0
nop retl
.BB1_1: nop
or %g0, %l0, %i0
.BB1_2:
restore %g0, %g0, %g0
retl
nop
It also helps alpha/ia64 :)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35881 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:11:38 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
|
|
|
SDOperand Add = DAG.getNode(ISD::ADD, XType, N0, Shift);
|
|
|
|
AddToWorkList(Shift.Val);
|
|
|
|
AddToWorkList(Add.Val);
|
|
|
|
return DAG.getNode(ISD::XOR, XType, Add, Shift);
|
|
|
|
}
|
|
|
|
// Check to see if this is an integer abs. select_cc setgt X, -1, X, -X ->
|
|
|
|
// Y = sra (X, size(X)-1); xor (add (X, Y), Y)
|
|
|
|
if (N1C && N1C->isAllOnesValue() && CC == ISD::SETGT &&
|
|
|
|
N0 == N2 && N3.getOpcode() == ISD::SUB && N0 == N3.getOperand(1)) {
|
|
|
|
if (ConstantSDNode *SubC = dyn_cast<ConstantSDNode>(N3.getOperand(0))) {
|
2008-06-06 12:08:01 +00:00
|
|
|
MVT XType = N0.getValueType();
|
|
|
|
if (SubC->isNullValue() && XType.isInteger()) {
|
2005-10-08 00:29:44 +00:00
|
|
|
SDOperand Shift = DAG.getNode(ISD::SRA, XType, N0,
|
2008-06-06 12:08:01 +00:00
|
|
|
DAG.getConstant(XType.getSizeInBits()-1,
|
Codegen integer abs more efficiently using the trick from the PPC CWG. This
improves codegen on many architectures. Tests committed as CodeGen/*/iabs.ll
X86 Old: X86 New:
_test: _test:
movl 4(%esp), %ecx movl 4(%esp), %eax
movl %ecx, %eax movl %eax, %ecx
negl %eax sarl $31, %ecx
testl %ecx, %ecx addl %ecx, %eax
cmovns %ecx, %eax xorl %ecx, %eax
ret ret
PPC Old: PPC New:
_test: _test:
cmpwi cr0, r3, -1 srawi r2, r3, 31
neg r2, r3 add r3, r3, r2
bgt cr0, LBB1_2 ; xor r3, r3, r2
LBB1_1: ; blr
mr r3, r2
LBB1_2: ;
blr
ARM Old: ARM New:
_test: _test:
rsb r3, r0, #0 add r3, r0, r0, asr #31
cmp r0, #0 eor r0, r3, r0, asr #31
movge r3, r0 bx lr
mov r0, r3
bx lr
Thumb Old: Thumb New:
_test: _test:
neg r2, r0 asr r2, r0, #31
cmp r0, #0 add r0, r0, r2
bge LBB1_2 eor r0, r2
LBB1_1: @ bx lr
cpy r0, r2
LBB1_2: @
bx lr
Sparc Old: Sparc New:
test: test:
save -96, %o6, %o6 save -96, %o6, %o6
sethi 0, %l0 sra %i0, 31, %l0
sub %l0, %i0, %l0 add %i0, %l0, %l1
subcc %i0, -1, %l1 xor %l1, %l0, %i0
bg .BB1_2 restore %g0, %g0, %g0
nop retl
.BB1_1: nop
or %g0, %l0, %i0
.BB1_2:
restore %g0, %g0, %g0
retl
nop
It also helps alpha/ia64 :)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35881 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:11:38 +00:00
|
|
|
TLI.getShiftAmountTy()));
|
2005-10-08 00:29:44 +00:00
|
|
|
SDOperand Add = DAG.getNode(ISD::ADD, XType, N0, Shift);
|
2006-03-01 04:03:14 +00:00
|
|
|
AddToWorkList(Shift.Val);
|
|
|
|
AddToWorkList(Add.Val);
|
2005-10-08 00:29:44 +00:00
|
|
|
return DAG.getNode(ISD::XOR, XType, Add, Shift);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
Codegen integer abs more efficiently using the trick from the PPC CWG. This
improves codegen on many architectures. Tests committed as CodeGen/*/iabs.ll
X86 Old: X86 New:
_test: _test:
movl 4(%esp), %ecx movl 4(%esp), %eax
movl %ecx, %eax movl %eax, %ecx
negl %eax sarl $31, %ecx
testl %ecx, %ecx addl %ecx, %eax
cmovns %ecx, %eax xorl %ecx, %eax
ret ret
PPC Old: PPC New:
_test: _test:
cmpwi cr0, r3, -1 srawi r2, r3, 31
neg r2, r3 add r3, r3, r2
bgt cr0, LBB1_2 ; xor r3, r3, r2
LBB1_1: ; blr
mr r3, r2
LBB1_2: ;
blr
ARM Old: ARM New:
_test: _test:
rsb r3, r0, #0 add r3, r0, r0, asr #31
cmp r0, #0 eor r0, r3, r0, asr #31
movge r3, r0 bx lr
mov r0, r3
bx lr
Thumb Old: Thumb New:
_test: _test:
neg r2, r0 asr r2, r0, #31
cmp r0, #0 add r0, r0, r2
bge LBB1_2 eor r0, r2
LBB1_1: @ bx lr
cpy r0, r2
LBB1_2: @
bx lr
Sparc Old: Sparc New:
test: test:
save -96, %o6, %o6 save -96, %o6, %o6
sethi 0, %l0 sra %i0, 31, %l0
sub %l0, %i0, %l0 add %i0, %l0, %l1
subcc %i0, -1, %l1 xor %l1, %l0, %i0
bg .BB1_2 restore %g0, %g0, %g0
nop retl
.BB1_1: nop
or %g0, %l0, %i0
.BB1_2:
restore %g0, %g0, %g0
retl
nop
It also helps alpha/ia64 :)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@35881 91177308-0d34-0410-b5e6-96231b3b80d8
2007-04-11 05:11:38 +00:00
|
|
|
|
2005-09-19 22:34:01 +00:00
|
|
|
return SDOperand();
|
|
|
|
}
|
|
|
|
|
2007-02-08 22:13:59 +00:00
|
|
|
/// SimplifySetCC - This is a stub for TargetLowering::SimplifySetCC.
|
2008-06-06 12:08:01 +00:00
|
|
|
SDOperand DAGCombiner::SimplifySetCC(MVT VT, SDOperand N0,
|
2005-10-05 21:43:42 +00:00
|
|
|
SDOperand N1, ISD::CondCode Cond,
|
|
|
|
bool foldBooleans) {
|
2007-02-08 22:13:59 +00:00
|
|
|
TargetLowering::DAGCombinerInfo
|
|
|
|
DagCombineInfo(DAG, !AfterLegalize, false, this);
|
|
|
|
return TLI.SimplifySetCC(VT, N0, N1, Cond, foldBooleans, DagCombineInfo);
|
2005-09-16 00:54:12 +00:00
|
|
|
}
|
|
|
|
|
2005-10-20 02:15:44 +00:00
|
|
|
/// BuildSDIVSequence - Given an ISD::SDIV node expressing a divide by constant,
|
|
|
|
/// return a DAG expression to select that will generate the same value by
|
|
|
|
/// multiplying by a magic number. See:
|
|
|
|
/// <http://the.wall.riscom.net/books/proc/ppc/cwg/code2.html>
|
|
|
|
SDOperand DAGCombiner::BuildSDIV(SDNode *N) {
|
2006-06-12 16:07:18 +00:00
|
|
|
std::vector<SDNode*> Built;
|
2006-05-16 17:42:15 +00:00
|
|
|
SDOperand S = TLI.BuildSDIV(N, DAG, &Built);
|
|
|
|
|
2006-06-12 16:07:18 +00:00
|
|
|
for (std::vector<SDNode*>::iterator ii = Built.begin(), ee = Built.end();
|
2006-05-16 17:42:15 +00:00
|
|
|
ii != ee; ++ii)
|
|
|
|
AddToWorkList(*ii);
|
|
|
|
return S;
|
2005-10-20 02:15:44 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// BuildUDIVSequence - Given an ISD::UDIV node expressing a divide by constant,
|
|
|
|
/// return a DAG expression to select that will generate the same value by
|
|
|
|
/// multiplying by a magic number. See:
|
|
|
|
/// <http://the.wall.riscom.net/books/proc/ppc/cwg/code2.html>
|
|
|
|
SDOperand DAGCombiner::BuildUDIV(SDNode *N) {
|
2006-06-12 16:07:18 +00:00
|
|
|
std::vector<SDNode*> Built;
|
2006-05-16 17:42:15 +00:00
|
|
|
SDOperand S = TLI.BuildUDIV(N, DAG, &Built);
|
|
|
|
|
2006-06-12 16:07:18 +00:00
|
|
|
for (std::vector<SDNode*>::iterator ii = Built.begin(), ee = Built.end();
|
2006-05-16 17:42:15 +00:00
|
|
|
ii != ee; ++ii)
|
|
|
|
AddToWorkList(*ii);
|
|
|
|
return S;
|
2005-10-20 02:15:44 +00:00
|
|
|
}
|
|
|
|
|
2006-10-07 23:37:56 +00:00
|
|
|
/// FindBaseOffset - Return true if base is known not to alias with anything
|
|
|
|
/// but itself. Provides base object and offset as results.
|
|
|
|
static bool FindBaseOffset(SDOperand Ptr, SDOperand &Base, int64_t &Offset) {
|
|
|
|
// Assume it is a primitive operation.
|
|
|
|
Base = Ptr; Offset = 0;
|
|
|
|
|
|
|
|
// If it's an adding a simple constant then integrate the offset.
|
|
|
|
if (Base.getOpcode() == ISD::ADD) {
|
|
|
|
if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Base.getOperand(1))) {
|
|
|
|
Base = Base.getOperand(0);
|
|
|
|
Offset += C->getValue();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If it's any of the following then it can't alias with anything but itself.
|
|
|
|
return isa<FrameIndexSDNode>(Base) ||
|
|
|
|
isa<ConstantPoolSDNode>(Base) ||
|
|
|
|
isa<GlobalAddressSDNode>(Base);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// isAlias - Return true if there is any possibility that the two addresses
|
|
|
|
/// overlap.
|
2006-10-18 12:29:57 +00:00
|
|
|
bool DAGCombiner::isAlias(SDOperand Ptr1, int64_t Size1,
|
|
|
|
const Value *SrcValue1, int SrcValueOffset1,
|
|
|
|
SDOperand Ptr2, int64_t Size2,
|
|
|
|
const Value *SrcValue2, int SrcValueOffset2)
|
|
|
|
{
|
2006-10-07 23:37:56 +00:00
|
|
|
// If they are the same then they must be aliases.
|
|
|
|
if (Ptr1 == Ptr2) return true;
|
|
|
|
|
|
|
|
// Gather base node and offset information.
|
|
|
|
SDOperand Base1, Base2;
|
|
|
|
int64_t Offset1, Offset2;
|
|
|
|
bool KnownBase1 = FindBaseOffset(Ptr1, Base1, Offset1);
|
|
|
|
bool KnownBase2 = FindBaseOffset(Ptr2, Base2, Offset2);
|
|
|
|
|
|
|
|
// If they have a same base address then...
|
|
|
|
if (Base1 == Base2) {
|
|
|
|
// Check to see if the addresses overlap.
|
|
|
|
return!((Offset1 + Size1) <= Offset2 || (Offset2 + Size2) <= Offset1);
|
|
|
|
}
|
|
|
|
|
2006-10-18 12:29:57 +00:00
|
|
|
// If we know both bases then they can't alias.
|
|
|
|
if (KnownBase1 && KnownBase2) return false;
|
|
|
|
|
2006-10-18 19:08:31 +00:00
|
|
|
if (CombinerGlobalAA) {
|
|
|
|
// Use alias analysis information.
|
2007-08-27 16:32:11 +00:00
|
|
|
int64_t MinOffset = std::min(SrcValueOffset1, SrcValueOffset2);
|
|
|
|
int64_t Overlap1 = Size1 + SrcValueOffset1 - MinOffset;
|
|
|
|
int64_t Overlap2 = Size2 + SrcValueOffset2 - MinOffset;
|
2006-10-18 19:08:31 +00:00
|
|
|
AliasAnalysis::AliasResult AAResult =
|
2006-10-18 12:29:57 +00:00
|
|
|
AA.alias(SrcValue1, Overlap1, SrcValue2, Overlap2);
|
2006-10-18 19:08:31 +00:00
|
|
|
if (AAResult == AliasAnalysis::NoAlias)
|
|
|
|
return false;
|
|
|
|
}
|
2006-10-18 12:29:57 +00:00
|
|
|
|
|
|
|
// Otherwise we have to assume they alias.
|
|
|
|
return true;
|
2006-10-07 23:37:56 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// FindAliasInfo - Extracts the relevant alias information from the memory
|
|
|
|
/// node. Returns true if the operand was a load.
|
2006-10-11 13:47:09 +00:00
|
|
|
bool DAGCombiner::FindAliasInfo(SDNode *N,
|
2006-10-18 12:29:57 +00:00
|
|
|
SDOperand &Ptr, int64_t &Size,
|
|
|
|
const Value *&SrcValue, int &SrcValueOffset) {
|
2006-10-11 13:47:09 +00:00
|
|
|
if (LoadSDNode *LD = dyn_cast<LoadSDNode>(N)) {
|
|
|
|
Ptr = LD->getBasePtr();
|
2008-06-06 12:08:01 +00:00
|
|
|
Size = LD->getMemoryVT().getSizeInBits() >> 3;
|
2006-10-11 13:47:09 +00:00
|
|
|
SrcValue = LD->getSrcValue();
|
2006-10-18 12:29:57 +00:00
|
|
|
SrcValueOffset = LD->getSrcValueOffset();
|
2006-10-07 23:37:56 +00:00
|
|
|
return true;
|
2006-10-11 13:47:09 +00:00
|
|
|
} else if (StoreSDNode *ST = dyn_cast<StoreSDNode>(N)) {
|
|
|
|
Ptr = ST->getBasePtr();
|
2008-06-06 12:08:01 +00:00
|
|
|
Size = ST->getMemoryVT().getSizeInBits() >> 3;
|
2006-10-11 13:47:09 +00:00
|
|
|
SrcValue = ST->getSrcValue();
|
2006-10-18 12:29:57 +00:00
|
|
|
SrcValueOffset = ST->getSrcValueOffset();
|
2006-10-11 13:47:09 +00:00
|
|
|
} else {
|
2006-10-07 23:37:56 +00:00
|
|
|
assert(0 && "FindAliasInfo expected a memory operand");
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
/// GatherAllAliases - Walk up chain skipping non-aliasing memory nodes,
|
|
|
|
/// looking for aliasing nodes and adding them to the Aliases vector.
|
2006-10-05 15:07:25 +00:00
|
|
|
void DAGCombiner::GatherAllAliases(SDNode *N, SDOperand OriginalChain,
|
2006-10-04 16:53:27 +00:00
|
|
|
SmallVector<SDOperand, 8> &Aliases) {
|
2006-10-05 15:07:25 +00:00
|
|
|
SmallVector<SDOperand, 8> Chains; // List of chains to visit.
|
2006-10-04 16:53:27 +00:00
|
|
|
std::set<SDNode *> Visited; // Visited node set.
|
|
|
|
|
2006-09-25 16:29:54 +00:00
|
|
|
// Get alias information for node.
|
|
|
|
SDOperand Ptr;
|
|
|
|
int64_t Size;
|
2006-10-11 13:47:09 +00:00
|
|
|
const Value *SrcValue;
|
2006-10-18 12:29:57 +00:00
|
|
|
int SrcValueOffset;
|
|
|
|
bool IsLoad = FindAliasInfo(N, Ptr, Size, SrcValue, SrcValueOffset);
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
// Starting off.
|
2006-10-05 15:07:25 +00:00
|
|
|
Chains.push_back(OriginalChain);
|
|
|
|
|
|
|
|
// Look at each chain and determine if it is an alias. If so, add it to the
|
|
|
|
// aliases list. If not, then continue up the chain looking for the next
|
|
|
|
// candidate.
|
|
|
|
while (!Chains.empty()) {
|
|
|
|
SDOperand Chain = Chains.back();
|
|
|
|
Chains.pop_back();
|
2006-10-04 16:53:27 +00:00
|
|
|
|
2006-10-05 15:07:25 +00:00
|
|
|
// Don't bother if we've been before.
|
|
|
|
if (Visited.find(Chain.Val) != Visited.end()) continue;
|
|
|
|
Visited.insert(Chain.Val);
|
|
|
|
|
|
|
|
switch (Chain.getOpcode()) {
|
|
|
|
case ISD::EntryToken:
|
|
|
|
// Entry token is ideal chain operand, but handled in FindBetterChain.
|
|
|
|
break;
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-05 15:07:25 +00:00
|
|
|
case ISD::LOAD:
|
|
|
|
case ISD::STORE: {
|
|
|
|
// Get alias information for Chain.
|
|
|
|
SDOperand OpPtr;
|
|
|
|
int64_t OpSize;
|
2006-10-11 13:47:09 +00:00
|
|
|
const Value *OpSrcValue;
|
2006-10-18 12:29:57 +00:00
|
|
|
int OpSrcValueOffset;
|
|
|
|
bool IsOpLoad = FindAliasInfo(Chain.Val, OpPtr, OpSize,
|
|
|
|
OpSrcValue, OpSrcValueOffset);
|
2006-10-05 15:07:25 +00:00
|
|
|
|
|
|
|
// If chain is alias then stop here.
|
|
|
|
if (!(IsLoad && IsOpLoad) &&
|
2006-10-18 12:29:57 +00:00
|
|
|
isAlias(Ptr, Size, SrcValue, SrcValueOffset,
|
|
|
|
OpPtr, OpSize, OpSrcValue, OpSrcValueOffset)) {
|
2006-10-05 15:07:25 +00:00
|
|
|
Aliases.push_back(Chain);
|
|
|
|
} else {
|
|
|
|
// Look further up the chain.
|
|
|
|
Chains.push_back(Chain.getOperand(0));
|
|
|
|
// Clean up old chain.
|
|
|
|
AddToWorkList(Chain.Val);
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
2006-10-05 15:07:25 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case ISD::TokenFactor:
|
|
|
|
// We have to check each of the operands of the token factor, so we queue
|
|
|
|
// then up. Adding the operands to the queue (stack) in reverse order
|
|
|
|
// maintains the original order and increases the likelihood that getNode
|
|
|
|
// will find a matching token factor (CSE.)
|
|
|
|
for (unsigned n = Chain.getNumOperands(); n;)
|
|
|
|
Chains.push_back(Chain.getOperand(--n));
|
|
|
|
// Eliminate the token factor if we can.
|
|
|
|
AddToWorkList(Chain.Val);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
// For all other instructions we will just have to take what we can get.
|
|
|
|
Aliases.push_back(Chain);
|
|
|
|
break;
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
|
|
|
}
|
2006-10-04 16:53:27 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// FindBetterChain - Walk up chain skipping non-aliasing memory nodes, looking
|
|
|
|
/// for a better chain (aliasing node.)
|
|
|
|
SDOperand DAGCombiner::FindBetterChain(SDNode *N, SDOperand OldChain) {
|
|
|
|
SmallVector<SDOperand, 8> Aliases; // Ops for replacing token factor.
|
|
|
|
|
|
|
|
// Accumulate all the aliases to this node.
|
|
|
|
GatherAllAliases(N, OldChain, Aliases);
|
|
|
|
|
|
|
|
if (Aliases.size() == 0) {
|
|
|
|
// If no operands then chain to entry token.
|
|
|
|
return DAG.getEntryNode();
|
|
|
|
} else if (Aliases.size() == 1) {
|
|
|
|
// If a single operand then chain to it. We don't need to revisit it.
|
|
|
|
return Aliases[0];
|
|
|
|
}
|
|
|
|
|
|
|
|
// Construct a custom tailored token factor.
|
|
|
|
SDOperand NewChain = DAG.getNode(ISD::TokenFactor, MVT::Other,
|
|
|
|
&Aliases[0], Aliases.size());
|
|
|
|
|
|
|
|
// Make sure the old chain gets cleaned up.
|
|
|
|
if (NewChain != OldChain) AddToWorkList(OldChain.Val);
|
2006-09-25 16:29:54 +00:00
|
|
|
|
2006-10-04 16:53:27 +00:00
|
|
|
return NewChain;
|
2006-09-25 16:29:54 +00:00
|
|
|
}
|
|
|
|
|
2005-09-01 00:19:25 +00:00
|
|
|
// SelectionDAG::Combine - This is the entry point for the file.
|
|
|
|
//
|
2006-10-16 20:52:31 +00:00
|
|
|
void SelectionDAG::Combine(bool RunningAfterLegalize, AliasAnalysis &AA) {
|
2007-01-16 04:55:25 +00:00
|
|
|
if (!RunningAfterLegalize && ViewDAGCombine1)
|
|
|
|
viewGraph();
|
|
|
|
if (RunningAfterLegalize && ViewDAGCombine2)
|
|
|
|
viewGraph();
|
2005-09-01 00:19:25 +00:00
|
|
|
/// run - This is the main entry point to this class.
|
|
|
|
///
|
2006-10-16 20:52:31 +00:00
|
|
|
DAGCombiner(*this, AA).Run(RunningAfterLegalize);
|
2005-09-01 00:19:25 +00:00
|
|
|
}
|