mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2025-06-14 14:24:05 +00:00
Clean whitespaces.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@160668 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
@ -83,7 +83,7 @@ namespace {
|
||||
const TargetLibraryInfo *TLInfo;
|
||||
DominatorTree *DT;
|
||||
ProfileInfo *PFI;
|
||||
|
||||
|
||||
/// CurInstIterator - As we scan instructions optimizing them, this is the
|
||||
/// next instruction to optimize. Xforms that can invalidate this should
|
||||
/// update it.
|
||||
@ -157,7 +157,7 @@ bool CodeGenPrepare::runOnFunction(Function &F) {
|
||||
EverMadeChange |= EliminateMostlyEmptyBlocks(F);
|
||||
|
||||
// llvm.dbg.value is far away from the value then iSel may not be able
|
||||
// handle it properly. iSel will drop llvm.dbg.value if it can not
|
||||
// handle it properly. iSel will drop llvm.dbg.value if it can not
|
||||
// find a node corresponding to the value.
|
||||
EverMadeChange |= PlaceDbgValues(F);
|
||||
|
||||
@ -336,7 +336,7 @@ void CodeGenPrepare::EliminateMostlyEmptyBlock(BasicBlock *BB) {
|
||||
|
||||
if (isEntry && BB != &BB->getParent()->getEntryBlock())
|
||||
BB->moveBefore(&BB->getParent()->getEntryBlock());
|
||||
|
||||
|
||||
DEBUG(dbgs() << "AFTER:\n" << *DestBB << "\n\n\n");
|
||||
return;
|
||||
}
|
||||
@ -547,7 +547,7 @@ protected:
|
||||
|
||||
bool CodeGenPrepare::OptimizeCallInst(CallInst *CI) {
|
||||
BasicBlock *BB = CI->getParent();
|
||||
|
||||
|
||||
// Lower inline assembly if we can.
|
||||
// If we found an inline asm expession, and if the target knows how to
|
||||
// lower it to normal LLVM code, do so now.
|
||||
@ -564,19 +564,19 @@ bool CodeGenPrepare::OptimizeCallInst(CallInst *CI) {
|
||||
if (OptimizeInlineAsmInst(CI))
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
// Lower all uses of llvm.objectsize.*
|
||||
IntrinsicInst *II = dyn_cast<IntrinsicInst>(CI);
|
||||
if (II && II->getIntrinsicID() == Intrinsic::objectsize) {
|
||||
bool Min = (cast<ConstantInt>(II->getArgOperand(1))->getZExtValue() == 1);
|
||||
Type *ReturnTy = CI->getType();
|
||||
Constant *RetVal = ConstantInt::get(ReturnTy, Min ? 0 : -1ULL);
|
||||
|
||||
Constant *RetVal = ConstantInt::get(ReturnTy, Min ? 0 : -1ULL);
|
||||
|
||||
// Substituting this can cause recursive simplifications, which can
|
||||
// invalidate our iterator. Use a WeakVH to hold onto it in case this
|
||||
// happens.
|
||||
WeakVH IterHandle(CurInstIterator);
|
||||
|
||||
|
||||
replaceAndRecursivelySimplify(CI, RetVal, TLI ? TLI->getTargetData() : 0,
|
||||
TLInfo, ModifiedDT ? 0 : DT);
|
||||
|
||||
@ -604,7 +604,7 @@ bool CodeGenPrepare::OptimizeCallInst(CallInst *CI) {
|
||||
// We'll need TargetData from here on out.
|
||||
const TargetData *TD = TLI ? TLI->getTargetData() : 0;
|
||||
if (!TD) return false;
|
||||
|
||||
|
||||
// Lower all default uses of _chk calls. This is very similar
|
||||
// to what InstCombineCalls does, but here we are only lowering calls
|
||||
// that have the default "don't know" as the objectsize. Anything else
|
||||
@ -760,13 +760,13 @@ static bool IsNonLocalValue(Value *V, BasicBlock *BB) {
|
||||
bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
Type *AccessTy) {
|
||||
Value *Repl = Addr;
|
||||
|
||||
// Try to collapse single-value PHI nodes. This is necessary to undo
|
||||
|
||||
// Try to collapse single-value PHI nodes. This is necessary to undo
|
||||
// unprofitable PRE transformations.
|
||||
SmallVector<Value*, 8> worklist;
|
||||
SmallPtrSet<Value*, 16> Visited;
|
||||
worklist.push_back(Addr);
|
||||
|
||||
|
||||
// Use a worklist to iteratively look through PHI nodes, and ensure that
|
||||
// the addressing mode obtained from the non-PHI roots of the graph
|
||||
// are equivalent.
|
||||
@ -778,20 +778,20 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
while (!worklist.empty()) {
|
||||
Value *V = worklist.back();
|
||||
worklist.pop_back();
|
||||
|
||||
|
||||
// Break use-def graph loops.
|
||||
if (!Visited.insert(V)) {
|
||||
Consensus = 0;
|
||||
break;
|
||||
}
|
||||
|
||||
|
||||
// For a PHI node, push all of its incoming values.
|
||||
if (PHINode *P = dyn_cast<PHINode>(V)) {
|
||||
for (unsigned i = 0, e = P->getNumIncomingValues(); i != e; ++i)
|
||||
worklist.push_back(P->getIncomingValue(i));
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
// For non-PHIs, determine the addressing mode being computed.
|
||||
SmallVector<Instruction*, 16> NewAddrModeInsts;
|
||||
ExtAddrMode NewAddrMode =
|
||||
@ -826,15 +826,15 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
Consensus = 0;
|
||||
break;
|
||||
}
|
||||
|
||||
|
||||
// If the addressing mode couldn't be determined, or if multiple different
|
||||
// ones were determined, bail out now.
|
||||
if (!Consensus) return false;
|
||||
|
||||
|
||||
// Check to see if any of the instructions supersumed by this addr mode are
|
||||
// non-local to I's BB.
|
||||
bool AnyNonLocal = false;
|
||||
@ -943,7 +943,7 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
// Use a WeakVH to hold onto it in case this happens.
|
||||
WeakVH IterHandle(CurInstIterator);
|
||||
BasicBlock *BB = CurInstIterator->getParent();
|
||||
|
||||
|
||||
RecursivelyDeleteTriviallyDeadInstructions(Repl);
|
||||
|
||||
if (IterHandle != CurInstIterator) {
|
||||
@ -955,7 +955,7 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
// This address is now available for reassignment, so erase the table
|
||||
// entry; we don't want to match some completely different instruction.
|
||||
SunkAddrs[Addr] = 0;
|
||||
}
|
||||
}
|
||||
}
|
||||
++NumMemoryInsts;
|
||||
return true;
|
||||
@ -967,12 +967,12 @@ bool CodeGenPrepare::OptimizeMemoryInst(Instruction *MemoryInst, Value *Addr,
|
||||
bool CodeGenPrepare::OptimizeInlineAsmInst(CallInst *CS) {
|
||||
bool MadeChange = false;
|
||||
|
||||
TargetLowering::AsmOperandInfoVector
|
||||
TargetLowering::AsmOperandInfoVector
|
||||
TargetConstraints = TLI->ParseConstraints(CS);
|
||||
unsigned ArgNo = 0;
|
||||
for (unsigned i = 0, e = TargetConstraints.size(); i != e; ++i) {
|
||||
TargetLowering::AsmOperandInfo &OpInfo = TargetConstraints[i];
|
||||
|
||||
|
||||
// Compute the constraint code and ConstraintType to use.
|
||||
TLI->ComputeConstraintToUse(OpInfo, SDValue());
|
||||
|
||||
@ -1187,7 +1187,7 @@ bool CodeGenPrepare::OptimizeInst(Instruction *I) {
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
if (CastInst *CI = dyn_cast<CastInst>(I)) {
|
||||
// If the source of the cast is a constant, then this should have
|
||||
// already been constant folded. The only reason NOT to constant fold
|
||||
@ -1207,23 +1207,23 @@ bool CodeGenPrepare::OptimizeInst(Instruction *I) {
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
if (CmpInst *CI = dyn_cast<CmpInst>(I))
|
||||
return OptimizeCmpExpression(CI);
|
||||
|
||||
|
||||
if (LoadInst *LI = dyn_cast<LoadInst>(I)) {
|
||||
if (TLI)
|
||||
return OptimizeMemoryInst(I, I->getOperand(0), LI->getType());
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
if (StoreInst *SI = dyn_cast<StoreInst>(I)) {
|
||||
if (TLI)
|
||||
return OptimizeMemoryInst(I, SI->getOperand(1),
|
||||
SI->getOperand(0)->getType());
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
if (GetElementPtrInst *GEPI = dyn_cast<GetElementPtrInst>(I)) {
|
||||
if (GEPI->hasAllZeroIndices()) {
|
||||
/// The GEP operand must be a pointer, so must its result -> BitCast
|
||||
@ -1237,7 +1237,7 @@ bool CodeGenPrepare::OptimizeInst(Instruction *I) {
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
if (CallInst *CI = dyn_cast<CallInst>(I))
|
||||
return OptimizeCallInst(CI);
|
||||
|
||||
@ -1265,7 +1265,7 @@ bool CodeGenPrepare::OptimizeBlock(BasicBlock &BB) {
|
||||
}
|
||||
|
||||
// llvm.dbg.value is far away from the value then iSel may not be able
|
||||
// handle it properly. iSel will drop llvm.dbg.value if it can not
|
||||
// handle it properly. iSel will drop llvm.dbg.value if it can not
|
||||
// find a node corresponding to the value.
|
||||
bool CodeGenPrepare::PlaceDbgValues(Function &F) {
|
||||
bool MadeChange = false;
|
||||
|
Reference in New Issue
Block a user