Revert 112442 and 112440 until the compile time problems introduced

by 112440 are resolved.


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@112692 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
Dan Gohman
2010-09-01 01:45:53 +00:00
parent 9cfad89a68
commit 191bd64a39
4 changed files with 133 additions and 203 deletions

View File

@ -27,7 +27,6 @@ class Value;
class IVUsers; class IVUsers;
class ScalarEvolution; class ScalarEvolution;
class SCEV; class SCEV;
class SCEVAddRecExpr;
class IVUsers; class IVUsers;
/// IVStrideUse - Keep track of one use of a strided induction variable. /// IVStrideUse - Keep track of one use of a strided induction variable.
@ -123,7 +122,7 @@ class IVUsers : public LoopPass {
LoopInfo *LI; LoopInfo *LI;
DominatorTree *DT; DominatorTree *DT;
ScalarEvolution *SE; ScalarEvolution *SE;
SmallPtrSet<Instruction *, 16> Processed; SmallPtrSet<Instruction*,16> Processed;
/// IVUses - A list of all tracked IV uses of induction variable expressions /// IVUses - A list of all tracked IV uses of induction variable expressions
/// we are interested in. /// we are interested in.
@ -135,16 +134,14 @@ class IVUsers : public LoopPass {
virtual void releaseMemory(); virtual void releaseMemory();
const SCEVAddRecExpr *findInterestingAddRec(const SCEV *S) const;
bool isInterestingUser(const Instruction *User) const;
public: public:
static char ID; // Pass ID, replacement for typeid static char ID; // Pass ID, replacement for typeid
IVUsers(); IVUsers();
/// AddUsersIfInteresting - Inspect the def-use graph starting at the /// AddUsersIfInteresting - Inspect the specified Instruction. If it is a
/// specified Instruction and add IVUsers. /// reducible SCEV, recursively add its users to the IVUsesByStride set and
void AddUsersIfInteresting(Instruction *I); /// return true. Otherwise, return false.
bool AddUsersIfInteresting(Instruction *I);
IVStrideUse &AddUser(Instruction *User, Value *Operand); IVStrideUse &AddUser(Instruction *User, Value *Operand);

View File

@ -35,123 +35,112 @@ Pass *llvm::createIVUsersPass() {
return new IVUsers(); return new IVUsers();
} }
/// findInterestingAddRec - Test whether the given expression is interesting. /// isInteresting - Test whether the given expression is "interesting" when
/// Return the addrec with the current loop which makes it interesting, or /// used by the given expression, within the context of analyzing the
/// null if it is not interesting. /// given loop.
const SCEVAddRecExpr *IVUsers::findInterestingAddRec(const SCEV *S) const { static bool isInteresting(const SCEV *S, const Instruction *I, const Loop *L,
ScalarEvolution *SE) {
// An addrec is interesting if it's affine or if it has an interesting start. // An addrec is interesting if it's affine or if it has an interesting start.
if (const SCEVAddRecExpr *AR = dyn_cast<SCEVAddRecExpr>(S)) { if (const SCEVAddRecExpr *AR = dyn_cast<SCEVAddRecExpr>(S)) {
// Keep things simple. Don't touch loop-variant strides. // Keep things simple. Don't touch loop-variant strides.
if (AR->getLoop() == L) if (AR->getLoop() == L)
return AR; return AR->isAffine() || !L->contains(I);
// We don't yet know how to do effective SCEV expansions for addrecs // Otherwise recurse to see if the start value is interesting, and that
// with interesting steps. // the step value is not interesting, since we don't yet know how to
if (findInterestingAddRec(AR->getStepRecurrence(*SE))) // do effective SCEV expansions for addrecs with interesting steps.
return 0; return isInteresting(AR->getStart(), I, L, SE) &&
// Otherwise recurse to see if the start value is interesting. !isInteresting(AR->getStepRecurrence(*SE), I, L, SE);
return findInterestingAddRec(AR->getStart());
} }
// An add is interesting if exactly one of its operands is interesting. // An add is interesting if exactly one of its operands is interesting.
if (const SCEVAddExpr *Add = dyn_cast<SCEVAddExpr>(S)) { if (const SCEVAddExpr *Add = dyn_cast<SCEVAddExpr>(S)) {
bool AnyInterestingYet = false;
for (SCEVAddExpr::op_iterator OI = Add->op_begin(), OE = Add->op_end(); for (SCEVAddExpr::op_iterator OI = Add->op_begin(), OE = Add->op_end();
OI != OE; ++OI) OI != OE; ++OI)
if (const SCEVAddRecExpr *AR = findInterestingAddRec(*OI)) if (isInteresting(*OI, I, L, SE)) {
return AR; if (AnyInterestingYet)
return 0; return false;
AnyInterestingYet = true;
}
return AnyInterestingYet;
} }
// Nothing else is interesting here. // Nothing else is interesting here.
return 0; return false;
}
bool IVUsers::isInterestingUser(const Instruction *User) const {
// Void and FP expressions cannot be reduced.
if (!SE->isSCEVable(User->getType()))
return false;
// LSR is not APInt clean, do not touch integers bigger than 64-bits.
if (SE->getTypeSizeInBits(User->getType()) > 64)
return false;
// Don't descend into PHI nodes outside the current loop.
if (LI->getLoopFor(User->getParent()) != L &&
isa<PHINode>(User))
return false;
// Otherwise, it may be interesting.
return true;
} }
/// AddUsersIfInteresting - Inspect the specified instruction. If it is a /// AddUsersIfInteresting - Inspect the specified instruction. If it is a
/// reducible SCEV, recursively add its users to the IVUsesByStride set and /// reducible SCEV, recursively add its users to the IVUsesByStride set and
/// return true. Otherwise, return false. /// return true. Otherwise, return false.
void IVUsers::AddUsersIfInteresting(Instruction *I) { bool IVUsers::AddUsersIfInteresting(Instruction *I) {
// Stop if we've seen this before.
if (!Processed.insert(I))
return;
// If this PHI node is not SCEVable, ignore it.
if (!SE->isSCEVable(I->getType())) if (!SE->isSCEVable(I->getType()))
return; return false; // Void and FP expressions cannot be reduced.
// If this PHI node is not an addrec for this loop, ignore it. // LSR is not APInt clean, do not touch integers bigger than 64-bits.
const SCEVAddRecExpr *Expr = findInterestingAddRec(SE->getSCEV(I)); if (SE->getTypeSizeInBits(I->getType()) > 64)
if (!Expr) return false;
return;
// Walk the def-use graph. if (!Processed.insert(I))
SmallVector<std::pair<Instruction *, const SCEVAddRecExpr *>, 16> Worklist; return true; // Instruction already handled.
Worklist.push_back(std::make_pair(I, Expr));
do {
std::pair<Instruction *, const SCEVAddRecExpr *> P =
Worklist.pop_back_val();
Instruction *Op = P.first;
const SCEVAddRecExpr *OpAR = P.second;
// Visit Op's users. // Get the symbolic expression for this instruction.
SmallPtrSet<Instruction *, 8> VisitedUsers; const SCEV *ISE = SE->getSCEV(I);
for (Value::use_iterator UI = Op->use_begin(), E = Op->use_end();
UI != E; ++UI) {
// Don't visit any individual user more than once.
Instruction *User = cast<Instruction>(*UI);
if (!VisitedUsers.insert(User))
continue;
// If it's an affine addrec (which we can pretty safely re-expand) inside // If we've come to an uninteresting expression, stop the traversal and
// the loop, or a potentially non-affine addrec outside the loop (which // call this a user.
// we can evaluate outside of the loop), follow it. if (!isInteresting(ISE, I, L, SE))
if (OpAR->isAffine() || !L->contains(User)) { return false;
if (isInterestingUser(User)) {
const SCEV *UserExpr = SE->getSCEV(User);
if (const SCEVAddRecExpr *AR = findInterestingAddRec(UserExpr)) { SmallPtrSet<Instruction *, 4> UniqueUsers;
// Interesting. Keep searching. for (Value::use_iterator UI = I->use_begin(), E = I->use_end();
if (Processed.insert(User)) UI != E; ++UI) {
Worklist.push_back(std::make_pair(User, AR)); Instruction *User = cast<Instruction>(*UI);
continue; if (!UniqueUsers.insert(User))
} continue;
}
// Do not infinitely recurse on PHI nodes.
if (isa<PHINode>(User) && Processed.count(User))
continue;
// Descend recursively, but not into PHI nodes outside the current loop.
// It's important to see the entire expression outside the loop to get
// choices that depend on addressing mode use right, although we won't
// consider references outside the loop in all cases.
// If User is already in Processed, we don't want to recurse into it again,
// but do want to record a second reference in the same instruction.
bool AddUserToIVUsers = false;
if (LI->getLoopFor(User->getParent()) != L) {
if (isa<PHINode>(User) || Processed.count(User) ||
!AddUsersIfInteresting(User)) {
DEBUG(dbgs() << "FOUND USER in other loop: " << *User << '\n'
<< " OF SCEV: " << *ISE << '\n');
AddUserToIVUsers = true;
} }
} else if (Processed.count(User) ||
// Otherwise, this is the point where the def-use chain !AddUsersIfInteresting(User)) {
// becomes uninteresting. Call it an IV User. DEBUG(dbgs() << "FOUND USER: " << *User << '\n'
AddUser(User, Op); << " OF SCEV: " << *ISE << '\n');
AddUserToIVUsers = true;
} }
} while (!Worklist.empty());
if (AddUserToIVUsers) {
// Okay, we found a user that we cannot reduce.
IVUses.push_back(new IVStrideUse(this, User, I));
IVStrideUse &NewUse = IVUses.back();
// Transform the expression into a normalized form.
ISE = TransformForPostIncUse(NormalizeAutodetect,
ISE, User, I,
NewUse.PostIncLoops,
*SE, *DT);
DEBUG(dbgs() << " NORMALIZED TO: " << *ISE << '\n');
}
}
return true;
} }
IVStrideUse &IVUsers::AddUser(Instruction *User, Value *Operand) { IVStrideUse &IVUsers::AddUser(Instruction *User, Value *Operand) {
IVUses.push_back(new IVStrideUse(this, User, Operand)); IVUses.push_back(new IVStrideUse(this, User, Operand));
IVStrideUse &NewUse = IVUses.back(); return IVUses.back();
// Auto-detect and remember post-inc loops for this expression.
const SCEV *S = SE->getSCEV(Operand);
(void)TransformForPostIncUse(NormalizeAutodetect,
S, User, Operand,
NewUse.PostIncLoops,
*SE, *DT);
return NewUse;
} }
IVUsers::IVUsers() IVUsers::IVUsers()
@ -176,7 +165,7 @@ bool IVUsers::runOnLoop(Loop *l, LPPassManager &LPM) {
// them by stride. Start by finding all of the PHI nodes in the header for // them by stride. Start by finding all of the PHI nodes in the header for
// this loop. If they are induction variables, inspect their uses. // this loop. If they are induction variables, inspect their uses.
for (BasicBlock::iterator I = L->getHeader()->begin(); isa<PHINode>(I); ++I) for (BasicBlock::iterator I = L->getHeader()->begin(); isa<PHINode>(I); ++I)
AddUsersIfInteresting(I); (void)AddUsersIfInteresting(I);
return false; return false;
} }

View File

@ -113,7 +113,6 @@ class RegUseTracker {
public: public:
void CountRegister(const SCEV *Reg, size_t LUIdx); void CountRegister(const SCEV *Reg, size_t LUIdx);
void DropRegister(const SCEV *Reg, size_t LUIdx); void DropRegister(const SCEV *Reg, size_t LUIdx);
void DropUse(size_t LUIdx, size_t NewLUIdx);
void DropUse(size_t LUIdx); void DropUse(size_t LUIdx);
bool isRegUsedByUsesOtherThan(const SCEV *Reg, size_t LUIdx) const; bool isRegUsedByUsesOtherThan(const SCEV *Reg, size_t LUIdx) const;
@ -152,24 +151,6 @@ RegUseTracker::DropRegister(const SCEV *Reg, size_t LUIdx) {
RSD.UsedByIndices.reset(LUIdx); RSD.UsedByIndices.reset(LUIdx);
} }
/// DropUse - Clear out reference by use LUIdx, and prepare for use NewLUIdx
/// to be swapped into LUIdx's position.
void
RegUseTracker::DropUse(size_t LUIdx, size_t NewLUIdx) {
// Remove the use index from every register's use list.
for (RegUsesTy::iterator I = RegUsesMap.begin(), E = RegUsesMap.end();
I != E; ++I) {
SmallBitVector &UsedByIndices = I->second.UsedByIndices;
UsedByIndices.resize(std::max(UsedByIndices.size(), NewLUIdx + 1));
if (LUIdx < UsedByIndices.size()) {
UsedByIndices[LUIdx] = UsedByIndices[NewLUIdx];
UsedByIndices.reset(NewLUIdx);
} else
UsedByIndices.reset(LUIdx);
}
}
/// DropUse - Clear out reference by use LUIdx.
void void
RegUseTracker::DropUse(size_t LUIdx) { RegUseTracker::DropUse(size_t LUIdx) {
// Remove the use index from every register's use list. // Remove the use index from every register's use list.
@ -1353,9 +1334,7 @@ class LSRInstance {
UseMapDenseMapInfo> UseMapTy; UseMapDenseMapInfo> UseMapTy;
UseMapTy UseMap; UseMapTy UseMap;
bool reconcileNewOffset(LSRUse &LU, bool reconcileNewOffset(LSRUse &LU, int64_t NewOffset, bool HasBaseReg,
int64_t NewMinOffset, int64_t NewMaxOffset,
bool HasBaseReg,
LSRUse::KindType Kind, const Type *AccessTy); LSRUse::KindType Kind, const Type *AccessTy);
std::pair<size_t, int64_t> getUse(const SCEV *&Expr, std::pair<size_t, int64_t> getUse(const SCEV *&Expr,
@ -1364,8 +1343,7 @@ class LSRInstance {
void DeleteUse(LSRUse &LU); void DeleteUse(LSRUse &LU);
LSRUse *FindUseWithSimilarFormula(const Formula &F, const LSRUse &OrigLU, LSRUse *FindUseWithSimilarFormula(const Formula &F, const LSRUse &OrigLU);
int64_t &NewBaseOffs);
public: public:
void InsertInitialFormula(const SCEV *S, LSRUse &LU, size_t LUIdx); void InsertInitialFormula(const SCEV *S, LSRUse &LU, size_t LUIdx);
@ -1866,13 +1844,11 @@ LSRInstance::OptimizeLoopTermCond() {
/// at the given offset and other details. If so, update the use and /// at the given offset and other details. If so, update the use and
/// return true. /// return true.
bool bool
LSRInstance::reconcileNewOffset(LSRUse &LU, LSRInstance::reconcileNewOffset(LSRUse &LU, int64_t NewOffset, bool HasBaseReg,
int64_t NewMinOffset, int64_t NewMaxOffset,
bool HasBaseReg,
LSRUse::KindType Kind, const Type *AccessTy) { LSRUse::KindType Kind, const Type *AccessTy) {
int64_t ResultMinOffset = LU.MinOffset; int64_t NewMinOffset = LU.MinOffset;
int64_t ResultMaxOffset = LU.MaxOffset; int64_t NewMaxOffset = LU.MaxOffset;
const Type *ResultAccessTy = AccessTy; const Type *NewAccessTy = AccessTy;
// Check for a mismatched kind. It's tempting to collapse mismatched kinds to // Check for a mismatched kind. It's tempting to collapse mismatched kinds to
// something conservative, however this can pessimize in the case that one of // something conservative, however this can pessimize in the case that one of
@ -1880,27 +1856,29 @@ LSRInstance::reconcileNewOffset(LSRUse &LU,
if (LU.Kind != Kind) if (LU.Kind != Kind)
return false; return false;
// Conservatively assume HasBaseReg is true for now. // Conservatively assume HasBaseReg is true for now.
if (NewMinOffset < LU.MinOffset) { if (NewOffset < LU.MinOffset) {
if (!isAlwaysFoldable(LU.MaxOffset - NewMinOffset, 0, HasBaseReg, if (!isAlwaysFoldable(LU.MaxOffset - NewOffset, 0, HasBaseReg,
Kind, AccessTy, TLI)) Kind, AccessTy, TLI))
return false; return false;
ResultMinOffset = NewMinOffset; NewMinOffset = NewOffset;
} else if (NewMaxOffset > LU.MaxOffset) { } else if (NewOffset > LU.MaxOffset) {
if (!isAlwaysFoldable(NewMaxOffset - LU.MinOffset, 0, HasBaseReg, if (!isAlwaysFoldable(NewOffset - LU.MinOffset, 0, HasBaseReg,
Kind, AccessTy, TLI)) Kind, AccessTy, TLI))
return false; return false;
ResultMaxOffset = NewMaxOffset; NewMaxOffset = NewOffset;
} }
// Check for a mismatched access type, and fall back conservatively as needed. // Check for a mismatched access type, and fall back conservatively as needed.
// TODO: Be less conservative when the type is similar and can use the same // TODO: Be less conservative when the type is similar and can use the same
// addressing modes. // addressing modes.
if (Kind == LSRUse::Address && AccessTy != LU.AccessTy) if (Kind == LSRUse::Address && AccessTy != LU.AccessTy)
ResultAccessTy = Type::getVoidTy(AccessTy->getContext()); NewAccessTy = Type::getVoidTy(AccessTy->getContext());
// Update the use. // Update the use.
LU.MinOffset = ResultMinOffset; LU.MinOffset = NewMinOffset;
LU.MaxOffset = ResultMaxOffset; LU.MaxOffset = NewMaxOffset;
LU.AccessTy = ResultAccessTy; LU.AccessTy = NewAccessTy;
if (NewOffset != LU.Offsets.back())
LU.Offsets.push_back(NewOffset);
return true; return true;
} }
@ -1925,12 +1903,9 @@ LSRInstance::getUse(const SCEV *&Expr,
// A use already existed with this base. // A use already existed with this base.
size_t LUIdx = P.first->second; size_t LUIdx = P.first->second;
LSRUse &LU = Uses[LUIdx]; LSRUse &LU = Uses[LUIdx];
if (reconcileNewOffset(LU, Offset, Offset, if (reconcileNewOffset(LU, Offset, /*HasBaseReg=*/true, Kind, AccessTy))
/*HasBaseReg=*/true, Kind, AccessTy)) {
LU.Offsets.push_back(Offset);
// Reuse this use. // Reuse this use.
return std::make_pair(LUIdx, Offset); return std::make_pair(LUIdx, Offset);
}
} }
// Create a new use. // Create a new use.
@ -1939,7 +1914,11 @@ LSRInstance::getUse(const SCEV *&Expr,
Uses.push_back(LSRUse(Kind, AccessTy)); Uses.push_back(LSRUse(Kind, AccessTy));
LSRUse &LU = Uses[LUIdx]; LSRUse &LU = Uses[LUIdx];
LU.Offsets.push_back(Offset); // We don't need to track redundant offsets, but we don't need to go out
// of our way here to avoid them.
if (LU.Offsets.empty() || Offset != LU.Offsets.back())
LU.Offsets.push_back(Offset);
LU.MinOffset = Offset; LU.MinOffset = Offset;
LU.MaxOffset = Offset; LU.MaxOffset = Offset;
return std::make_pair(LUIdx, Offset); return std::make_pair(LUIdx, Offset);
@ -1947,12 +1926,8 @@ LSRInstance::getUse(const SCEV *&Expr,
/// DeleteUse - Delete the given use from the Uses list. /// DeleteUse - Delete the given use from the Uses list.
void LSRInstance::DeleteUse(LSRUse &LU) { void LSRInstance::DeleteUse(LSRUse &LU) {
if (&LU != &Uses.back()) { if (&LU != &Uses.back())
std::swap(LU, Uses.back()); std::swap(LU, Uses.back());
RegUses.DropUse(&LU - Uses.begin(), Uses.size() - 1);
} else {
RegUses.DropUse(&LU - Uses.begin());
}
Uses.pop_back(); Uses.pop_back();
} }
@ -1960,9 +1935,8 @@ void LSRInstance::DeleteUse(LSRUse &LU) {
/// a formula that has the same registers as the given formula. /// a formula that has the same registers as the given formula.
LSRUse * LSRUse *
LSRInstance::FindUseWithSimilarFormula(const Formula &OrigF, LSRInstance::FindUseWithSimilarFormula(const Formula &OrigF,
const LSRUse &OrigLU, const LSRUse &OrigLU) {
int64_t &NewBaseOffs) { // Search all uses for the formula. This could be more clever.
// Search all uses for a formula similar to OrigF. This could be more clever.
for (size_t LUIdx = 0, NumUses = Uses.size(); LUIdx != NumUses; ++LUIdx) { for (size_t LUIdx = 0, NumUses = Uses.size(); LUIdx != NumUses; ++LUIdx) {
LSRUse &LU = Uses[LUIdx]; LSRUse &LU = Uses[LUIdx];
// Check whether this use is close enough to OrigLU, to see whether it's // Check whether this use is close enough to OrigLU, to see whether it's
@ -1985,15 +1959,8 @@ LSRInstance::FindUseWithSimilarFormula(const Formula &OrigF,
F.ScaledReg == OrigF.ScaledReg && F.ScaledReg == OrigF.ScaledReg &&
F.AM.BaseGV == OrigF.AM.BaseGV && F.AM.BaseGV == OrigF.AM.BaseGV &&
F.AM.Scale == OrigF.AM.Scale) { F.AM.Scale == OrigF.AM.Scale) {
// Ok, all the registers and symbols matched. Check to see if the if (F.AM.BaseOffs == 0)
// immediate looks nicer than our old one.
if (OrigF.AM.BaseOffs == INT64_MIN ||
(F.AM.BaseOffs != INT64_MIN &&
abs64(F.AM.BaseOffs) < abs64(OrigF.AM.BaseOffs))) {
// Looks good. Take it.
NewBaseOffs = F.AM.BaseOffs;
return &LU; return &LU;
}
// This is the formula where all the registers and symbols matched; // This is the formula where all the registers and symbols matched;
// there aren't going to be any others. Since we declined it, we // there aren't going to be any others. Since we declined it, we
// can skip the rest of the formulae and procede to the next LSRUse. // can skip the rest of the formulae and procede to the next LSRUse.
@ -2634,17 +2601,6 @@ struct WorkItem {
WorkItem(size_t LI, int64_t I, const SCEV *R) WorkItem(size_t LI, int64_t I, const SCEV *R)
: LUIdx(LI), Imm(I), OrigReg(R) {} : LUIdx(LI), Imm(I), OrigReg(R) {}
bool operator==(const WorkItem &that) const {
return LUIdx == that.LUIdx && Imm == that.Imm && OrigReg == that.OrigReg;
}
bool operator<(const WorkItem &that) const {
if (LUIdx != that.LUIdx)
return LUIdx < that.LUIdx;
if (Imm != that.Imm)
return Imm < that.Imm;
return OrigReg < that.OrigReg;
}
void print(raw_ostream &OS) const; void print(raw_ostream &OS) const;
void dump() const; void dump() const;
}; };
@ -2684,7 +2640,8 @@ void LSRInstance::GenerateCrossUseConstantOffsets() {
// Now examine each set of registers with the same base value. Build up // Now examine each set of registers with the same base value. Build up
// a list of work to do and do the work in a separate step so that we're // a list of work to do and do the work in a separate step so that we're
// not adding formulae and register counts while we're searching. // not adding formulae and register counts while we're searching.
SmallSetVector<WorkItem, 32> WorkItems; SmallVector<WorkItem, 32> WorkItems;
SmallSet<std::pair<size_t, int64_t>, 32> UniqueItems;
for (SmallVectorImpl<const SCEV *>::const_iterator I = Sequence.begin(), for (SmallVectorImpl<const SCEV *>::const_iterator I = Sequence.begin(),
E = Sequence.end(); I != E; ++I) { E = Sequence.end(); I != E; ++I) {
const SCEV *Reg = *I; const SCEV *Reg = *I;
@ -2727,10 +2684,10 @@ void LSRInstance::GenerateCrossUseConstantOffsets() {
// Compute the difference between the two. // Compute the difference between the two.
int64_t Imm = (uint64_t)JImm - M->first; int64_t Imm = (uint64_t)JImm - M->first;
for (int LUIdx = UsedByIndices.find_first(); LUIdx != -1; for (int LUIdx = UsedByIndices.find_first(); LUIdx != -1;
LUIdx = UsedByIndices.find_next(LUIdx)) { LUIdx = UsedByIndices.find_next(LUIdx))
// Make a memo of this use, offset, and register tuple. // Make a memo of this use, offset, and register tuple.
WorkItems.insert(WorkItem(LUIdx, Imm, OrigReg)); if (UniqueItems.insert(std::make_pair(LUIdx, Imm)))
} WorkItems.push_back(WorkItem(LUIdx, Imm, OrigReg));
} }
} }
} }
@ -2738,6 +2695,7 @@ void LSRInstance::GenerateCrossUseConstantOffsets() {
Map.clear(); Map.clear();
Sequence.clear(); Sequence.clear();
UsedByIndicesMap.clear(); UsedByIndicesMap.clear();
UniqueItems.clear();
// Now iterate through the worklist and add new formulae. // Now iterate through the worklist and add new formulae.
for (SmallVectorImpl<WorkItem>::const_iterator I = WorkItems.begin(), for (SmallVectorImpl<WorkItem>::const_iterator I = WorkItems.begin(),
@ -3034,12 +2992,8 @@ void LSRInstance::NarrowSearchSpaceByCollapsingUnrolledCode() {
E = LU.Formulae.end(); I != E; ++I) { E = LU.Formulae.end(); I != E; ++I) {
const Formula &F = *I; const Formula &F = *I;
if (F.AM.BaseOffs != 0 && F.AM.Scale == 0) { if (F.AM.BaseOffs != 0 && F.AM.Scale == 0) {
int64_t NewBaseOffs; if (LSRUse *LUThatHas = FindUseWithSimilarFormula(F, LU)) {
if (LSRUse *LUThatHas = FindUseWithSimilarFormula(F, LU, if (reconcileNewOffset(*LUThatHas, F.AM.BaseOffs,
NewBaseOffs)) {
if (reconcileNewOffset(*LUThatHas,
F.AM.BaseOffs + LU.MinOffset - NewBaseOffs,
F.AM.BaseOffs + LU.MaxOffset - NewBaseOffs,
/*HasBaseReg=*/false, /*HasBaseReg=*/false,
LU.Kind, LU.AccessTy)) { LU.Kind, LU.AccessTy)) {
DEBUG(dbgs() << " Deleting use "; LU.print(dbgs()); DEBUG(dbgs() << " Deleting use "; LU.print(dbgs());
@ -3047,30 +3001,6 @@ void LSRInstance::NarrowSearchSpaceByCollapsingUnrolledCode() {
LUThatHas->AllFixupsOutsideLoop &= LU.AllFixupsOutsideLoop; LUThatHas->AllFixupsOutsideLoop &= LU.AllFixupsOutsideLoop;
// Update the relocs to reference the new use.
// Do this first so that MinOffset and MaxOffset are updated
// before we begin to determine which formulae to delete.
for (SmallVectorImpl<LSRFixup>::iterator I = Fixups.begin(),
E = Fixups.end(); I != E; ++I) {
LSRFixup &Fixup = *I;
if (Fixup.LUIdx == LUIdx) {
Fixup.LUIdx = LUThatHas - &Uses.front();
Fixup.Offset += F.AM.BaseOffs - NewBaseOffs;
DEBUG(dbgs() << "New fixup has offset "
<< Fixup.Offset << '\n');
LUThatHas->Offsets.push_back(Fixup.Offset);
if (Fixup.Offset > LUThatHas->MaxOffset)
LUThatHas->MaxOffset = Fixup.Offset;
if (Fixup.Offset < LUThatHas->MinOffset)
LUThatHas->MinOffset = Fixup.Offset;
}
// DeleteUse will do a swap+pop_back, so if this fixup is
// now pointing to the last LSRUse, update it to point to the
// position it'll be swapped to.
if (Fixup.LUIdx == NumUses-1)
Fixup.LUIdx = LUIdx;
}
// Delete formulae from the new use which are no longer legal. // Delete formulae from the new use which are no longer legal.
bool Any = false; bool Any = false;
for (size_t i = 0, e = LUThatHas->Formulae.size(); i != e; ++i) { for (size_t i = 0, e = LUThatHas->Formulae.size(); i != e; ++i) {
@ -3089,6 +3019,20 @@ void LSRInstance::NarrowSearchSpaceByCollapsingUnrolledCode() {
if (Any) if (Any)
LUThatHas->RecomputeRegs(LUThatHas - &Uses.front(), RegUses); LUThatHas->RecomputeRegs(LUThatHas - &Uses.front(), RegUses);
// Update the relocs to reference the new use.
for (SmallVectorImpl<LSRFixup>::iterator I = Fixups.begin(),
E = Fixups.end(); I != E; ++I) {
LSRFixup &Fixup = *I;
if (Fixup.LUIdx == LUIdx) {
Fixup.LUIdx = LUThatHas - &Uses.front();
Fixup.Offset += F.AM.BaseOffs;
DEBUG(dbgs() << "New fixup has offset "
<< Fixup.Offset << '\n');
}
if (Fixup.LUIdx == NumUses-1)
Fixup.LUIdx = LUIdx;
}
// Delete the old use. // Delete the old use.
DeleteUse(LU); DeleteUse(LU);
--LUIdx; --LUIdx;

View File

@ -452,8 +452,8 @@ bb5: ; preds = %bb3, %entry
; CHECK-NEXT: addss %xmm{{.*}}, %xmm{{.*}} ; CHECK-NEXT: addss %xmm{{.*}}, %xmm{{.*}}
; CHECK-NEXT: mulss (%r{{[^,]*}}), %xmm{{.*}} ; CHECK-NEXT: mulss (%r{{[^,]*}}), %xmm{{.*}}
; CHECK-NEXT: movss %xmm{{.*}}, (%r{{[^,]*}}) ; CHECK-NEXT: movss %xmm{{.*}}, (%r{{[^,]*}})
; CHECK-NEXT: decq %r{{.*}}
; CHECK-NEXT: addq $4, %r{{.*}} ; CHECK-NEXT: addq $4, %r{{.*}}
; CHECK-NEXT: decq %r{{.*}}
; CHECK-NEXT: addq $4, %r{{.*}} ; CHECK-NEXT: addq $4, %r{{.*}}
; CHECK-NEXT: movaps %xmm{{.*}}, %xmm{{.*}} ; CHECK-NEXT: movaps %xmm{{.*}}, %xmm{{.*}}
; CHECK-NEXT: BB10_2: ; CHECK-NEXT: BB10_2: