mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-11-11 23:05:31 +00:00
0b8c9a80f2
into their new header subdirectory: include/llvm/IR. This matches the directory structure of lib, and begins to correct a long standing point of file layout clutter in LLVM. There are still more header files to move here, but I wanted to handle them in separate commits to make tracking what files make sense at each layer easier. The only really questionable files here are the target intrinsic tablegen files. But that's a battle I'd rather not fight today. I've updated both CMake and Makefile build systems (I think, and my tests think, but I may have missed something). I've also re-sorted the includes throughout the project. I'll be committing updates to Clang, DragonEgg, and Polly momentarily. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@171366 91177308-0d34-0410-b5e6-96231b3b80d8
213 lines
7.9 KiB
C++
213 lines
7.9 KiB
C++
//===- CodeMetrics.cpp - Code cost measurements ---------------------------===//
|
|
//
|
|
// The LLVM Compiler Infrastructure
|
|
//
|
|
// This file is distributed under the University of Illinois Open Source
|
|
// License. See LICENSE.TXT for details.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file implements code cost measurement utilities.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/Analysis/CodeMetrics.h"
|
|
#include "llvm/IR/DataLayout.h"
|
|
#include "llvm/IR/Function.h"
|
|
#include "llvm/IR/IntrinsicInst.h"
|
|
#include "llvm/Support/CallSite.h"
|
|
|
|
using namespace llvm;
|
|
|
|
/// callIsSmall - If a call is likely to lower to a single target instruction,
|
|
/// or is otherwise deemed small return true.
|
|
/// TODO: Perhaps calls like memcpy, strcpy, etc?
|
|
bool llvm::callIsSmall(ImmutableCallSite CS) {
|
|
if (isa<IntrinsicInst>(CS.getInstruction()))
|
|
return true;
|
|
|
|
const Function *F = CS.getCalledFunction();
|
|
if (!F) return false;
|
|
|
|
if (F->hasLocalLinkage()) return false;
|
|
|
|
if (!F->hasName()) return false;
|
|
|
|
StringRef Name = F->getName();
|
|
|
|
// These will all likely lower to a single selection DAG node.
|
|
if (Name == "copysign" || Name == "copysignf" || Name == "copysignl" ||
|
|
Name == "fabs" || Name == "fabsf" || Name == "fabsl" ||
|
|
Name == "sin" || Name == "sinf" || Name == "sinl" ||
|
|
Name == "cos" || Name == "cosf" || Name == "cosl" ||
|
|
Name == "sqrt" || Name == "sqrtf" || Name == "sqrtl" )
|
|
return true;
|
|
|
|
// These are all likely to be optimized into something smaller.
|
|
if (Name == "pow" || Name == "powf" || Name == "powl" ||
|
|
Name == "exp2" || Name == "exp2l" || Name == "exp2f" ||
|
|
Name == "floor" || Name == "floorf" || Name == "ceil" ||
|
|
Name == "round" || Name == "ffs" || Name == "ffsl" ||
|
|
Name == "abs" || Name == "labs" || Name == "llabs")
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
bool llvm::isInstructionFree(const Instruction *I, const DataLayout *TD) {
|
|
if (isa<PHINode>(I))
|
|
return true;
|
|
|
|
// If a GEP has all constant indices, it will probably be folded with
|
|
// a load/store.
|
|
if (const GetElementPtrInst *GEP = dyn_cast<GetElementPtrInst>(I))
|
|
return GEP->hasAllConstantIndices();
|
|
|
|
if (const IntrinsicInst *II = dyn_cast<IntrinsicInst>(I)) {
|
|
switch (II->getIntrinsicID()) {
|
|
default:
|
|
return false;
|
|
case Intrinsic::dbg_declare:
|
|
case Intrinsic::dbg_value:
|
|
case Intrinsic::invariant_start:
|
|
case Intrinsic::invariant_end:
|
|
case Intrinsic::lifetime_start:
|
|
case Intrinsic::lifetime_end:
|
|
case Intrinsic::objectsize:
|
|
case Intrinsic::ptr_annotation:
|
|
case Intrinsic::var_annotation:
|
|
// These intrinsics don't count as size.
|
|
return true;
|
|
}
|
|
}
|
|
|
|
if (const CastInst *CI = dyn_cast<CastInst>(I)) {
|
|
// Noop casts, including ptr <-> int, don't count.
|
|
if (CI->isLosslessCast())
|
|
return true;
|
|
|
|
Value *Op = CI->getOperand(0);
|
|
// An inttoptr cast is free so long as the input is a legal integer type
|
|
// which doesn't contain values outside the range of a pointer.
|
|
if (isa<IntToPtrInst>(CI) && TD &&
|
|
TD->isLegalInteger(Op->getType()->getScalarSizeInBits()) &&
|
|
Op->getType()->getScalarSizeInBits() <= TD->getPointerSizeInBits())
|
|
return true;
|
|
|
|
// A ptrtoint cast is free so long as the result is large enough to store
|
|
// the pointer, and a legal integer type.
|
|
if (isa<PtrToIntInst>(CI) && TD &&
|
|
TD->isLegalInteger(Op->getType()->getScalarSizeInBits()) &&
|
|
Op->getType()->getScalarSizeInBits() >= TD->getPointerSizeInBits())
|
|
return true;
|
|
|
|
// trunc to a native type is free (assuming the target has compare and
|
|
// shift-right of the same width).
|
|
if (TD && isa<TruncInst>(CI) &&
|
|
TD->isLegalInteger(TD->getTypeSizeInBits(CI->getType())))
|
|
return true;
|
|
// Result of a cmp instruction is often extended (to be used by other
|
|
// cmp instructions, logical or return instructions). These are usually
|
|
// nop on most sane targets.
|
|
if (isa<CmpInst>(CI->getOperand(0)))
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/// analyzeBasicBlock - Fill in the current structure with information gleaned
|
|
/// from the specified block.
|
|
void CodeMetrics::analyzeBasicBlock(const BasicBlock *BB,
|
|
const DataLayout *TD) {
|
|
++NumBlocks;
|
|
unsigned NumInstsBeforeThisBB = NumInsts;
|
|
for (BasicBlock::const_iterator II = BB->begin(), E = BB->end();
|
|
II != E; ++II) {
|
|
if (isInstructionFree(II, TD))
|
|
continue;
|
|
|
|
// Special handling for calls.
|
|
if (isa<CallInst>(II) || isa<InvokeInst>(II)) {
|
|
ImmutableCallSite CS(cast<Instruction>(II));
|
|
|
|
if (const Function *F = CS.getCalledFunction()) {
|
|
// If a function is both internal and has a single use, then it is
|
|
// extremely likely to get inlined in the future (it was probably
|
|
// exposed by an interleaved devirtualization pass).
|
|
if (!CS.isNoInline() && F->hasInternalLinkage() && F->hasOneUse())
|
|
++NumInlineCandidates;
|
|
|
|
// If this call is to function itself, then the function is recursive.
|
|
// Inlining it into other functions is a bad idea, because this is
|
|
// basically just a form of loop peeling, and our metrics aren't useful
|
|
// for that case.
|
|
if (F == BB->getParent())
|
|
isRecursive = true;
|
|
}
|
|
|
|
if (!callIsSmall(CS)) {
|
|
// Each argument to a call takes on average one instruction to set up.
|
|
NumInsts += CS.arg_size();
|
|
|
|
// We don't want inline asm to count as a call - that would prevent loop
|
|
// unrolling. The argument setup cost is still real, though.
|
|
if (!isa<InlineAsm>(CS.getCalledValue()))
|
|
++NumCalls;
|
|
}
|
|
}
|
|
|
|
if (const AllocaInst *AI = dyn_cast<AllocaInst>(II)) {
|
|
if (!AI->isStaticAlloca())
|
|
this->usesDynamicAlloca = true;
|
|
}
|
|
|
|
if (isa<ExtractElementInst>(II) || II->getType()->isVectorTy())
|
|
++NumVectorInsts;
|
|
|
|
if (const CallInst *CI = dyn_cast<CallInst>(II))
|
|
if (CI->hasFnAttr(Attribute::NoDuplicate))
|
|
notDuplicatable = true;
|
|
|
|
if (const InvokeInst *InvI = dyn_cast<InvokeInst>(II))
|
|
if (InvI->hasFnAttr(Attribute::NoDuplicate))
|
|
notDuplicatable = true;
|
|
|
|
++NumInsts;
|
|
}
|
|
|
|
if (isa<ReturnInst>(BB->getTerminator()))
|
|
++NumRets;
|
|
|
|
// We never want to inline functions that contain an indirectbr. This is
|
|
// incorrect because all the blockaddress's (in static global initializers
|
|
// for example) would be referring to the original function, and this indirect
|
|
// jump would jump from the inlined copy of the function into the original
|
|
// function which is extremely undefined behavior.
|
|
// FIXME: This logic isn't really right; we can safely inline functions
|
|
// with indirectbr's as long as no other function or global references the
|
|
// blockaddress of a block within the current function. And as a QOI issue,
|
|
// if someone is using a blockaddress without an indirectbr, and that
|
|
// reference somehow ends up in another function or global, we probably
|
|
// don't want to inline this function.
|
|
notDuplicatable |= isa<IndirectBrInst>(BB->getTerminator());
|
|
|
|
// Remember NumInsts for this BB.
|
|
NumBBInsts[BB] = NumInsts - NumInstsBeforeThisBB;
|
|
}
|
|
|
|
void CodeMetrics::analyzeFunction(Function *F, const DataLayout *TD) {
|
|
// If this function contains a call that "returns twice" (e.g., setjmp or
|
|
// _setjmp) and it isn't marked with "returns twice" itself, never inline it.
|
|
// This is a hack because we depend on the user marking their local variables
|
|
// as volatile if they are live across a setjmp call, and they probably
|
|
// won't do this in callers.
|
|
exposesReturnsTwice = F->callsFunctionThatReturnsTwice() &&
|
|
!F->getAttributes().hasAttribute(AttributeSet::FunctionIndex,
|
|
Attribute::ReturnsTwice);
|
|
|
|
// Look at the size of the callee.
|
|
for (Function::const_iterator BB = F->begin(), E = F->end(); BB != E; ++BB)
|
|
analyzeBasicBlock(&*BB, TD);
|
|
}
|