mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2025-08-02 08:26:39 +00:00
Make some DataLayout pointers const.
No functionality change. Just reduces the noise of an upcoming patch. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@202087 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
@@ -84,7 +84,7 @@ namespace {
|
||||
const GlobalStatus &GS);
|
||||
bool OptimizeEmptyGlobalCXXDtors(Function *CXAAtExitFn);
|
||||
|
||||
DataLayout *DL;
|
||||
const DataLayout *DL;
|
||||
TargetLibraryInfo *TLI;
|
||||
};
|
||||
}
|
||||
@@ -266,7 +266,8 @@ static bool CleanupPointerRootUsers(GlobalVariable *GV,
|
||||
/// quick scan over the use list to clean up the easy and obvious cruft. This
|
||||
/// returns true if it made a change.
|
||||
static bool CleanupConstantGlobalUsers(Value *V, Constant *Init,
|
||||
DataLayout *DL, TargetLibraryInfo *TLI) {
|
||||
const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
bool Changed = false;
|
||||
// Note that we need to use a weak value handle for the worklist items. When
|
||||
// we delete a constant array, we may also be holding pointer to one of its
|
||||
@@ -743,7 +744,7 @@ static bool OptimizeAwayTrappingUsesOfValue(Value *V, Constant *NewV) {
|
||||
/// if the loaded value is dynamically null, then we know that they cannot be
|
||||
/// reachable with a null optimize away the load.
|
||||
static bool OptimizeAwayTrappingUsesOfLoads(GlobalVariable *GV, Constant *LV,
|
||||
DataLayout *DL,
|
||||
const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
bool Changed = false;
|
||||
|
||||
@@ -806,8 +807,8 @@ static bool OptimizeAwayTrappingUsesOfLoads(GlobalVariable *GV, Constant *LV,
|
||||
|
||||
/// ConstantPropUsersOf - Walk the use list of V, constant folding all of the
|
||||
/// instructions that are foldable.
|
||||
static void ConstantPropUsersOf(Value *V,
|
||||
DataLayout *DL, TargetLibraryInfo *TLI) {
|
||||
static void ConstantPropUsersOf(Value *V, const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
for (Value::use_iterator UI = V->use_begin(), E = V->use_end(); UI != E; )
|
||||
if (Instruction *I = dyn_cast<Instruction>(*UI++))
|
||||
if (Constant *NewC = ConstantFoldInstruction(I, DL, TLI)) {
|
||||
@@ -830,7 +831,7 @@ static GlobalVariable *OptimizeGlobalAddressOfMalloc(GlobalVariable *GV,
|
||||
CallInst *CI,
|
||||
Type *AllocTy,
|
||||
ConstantInt *NElements,
|
||||
DataLayout *DL,
|
||||
const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
DEBUG(errs() << "PROMOTING GLOBAL: " << *GV << " CALL = " << *CI << '\n');
|
||||
|
||||
@@ -1278,7 +1279,7 @@ static void RewriteUsesOfLoadForHeapSRoA(LoadInst *Load,
|
||||
/// PerformHeapAllocSRoA - CI is an allocation of an array of structures. Break
|
||||
/// it up into multiple allocations of arrays of the fields.
|
||||
static GlobalVariable *PerformHeapAllocSRoA(GlobalVariable *GV, CallInst *CI,
|
||||
Value *NElems, DataLayout *DL,
|
||||
Value *NElems, const DataLayout *DL,
|
||||
const TargetLibraryInfo *TLI) {
|
||||
DEBUG(dbgs() << "SROA HEAP ALLOC: " << *GV << " MALLOC = " << *CI << '\n');
|
||||
Type *MAT = getMallocAllocatedType(CI, TLI);
|
||||
@@ -1470,7 +1471,7 @@ static bool TryToOptimizeStoreOfMallocToGlobal(GlobalVariable *GV,
|
||||
Type *AllocTy,
|
||||
AtomicOrdering Ordering,
|
||||
Module::global_iterator &GVI,
|
||||
DataLayout *DL,
|
||||
const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
if (!DL)
|
||||
return false;
|
||||
@@ -1569,7 +1570,8 @@ static bool TryToOptimizeStoreOfMallocToGlobal(GlobalVariable *GV,
|
||||
static bool OptimizeOnceStoredGlobal(GlobalVariable *GV, Value *StoredOnceVal,
|
||||
AtomicOrdering Ordering,
|
||||
Module::global_iterator &GVI,
|
||||
DataLayout *DL, TargetLibraryInfo *TLI) {
|
||||
const DataLayout *DL,
|
||||
TargetLibraryInfo *TLI) {
|
||||
// Ignore no-op GEPs and bitcasts.
|
||||
StoredOnceVal = StoredOnceVal->stripPointerCasts();
|
||||
|
||||
|
@@ -108,12 +108,12 @@ public:
|
||||
static const ComparableFunction TombstoneKey;
|
||||
static DataLayout * const LookupOnly;
|
||||
|
||||
ComparableFunction(Function *Func, DataLayout *DL)
|
||||
ComparableFunction(Function *Func, const DataLayout *DL)
|
||||
: Func(Func), Hash(profileFunction(Func)), DL(DL) {}
|
||||
|
||||
Function *getFunc() const { return Func; }
|
||||
unsigned getHash() const { return Hash; }
|
||||
DataLayout *getDataLayout() const { return DL; }
|
||||
const DataLayout *getDataLayout() const { return DL; }
|
||||
|
||||
// Drops AssertingVH reference to the function. Outside of debug mode, this
|
||||
// does nothing.
|
||||
@@ -129,7 +129,7 @@ private:
|
||||
|
||||
AssertingVH<Function> Func;
|
||||
unsigned Hash;
|
||||
DataLayout *DL;
|
||||
const DataLayout *DL;
|
||||
};
|
||||
|
||||
const ComparableFunction ComparableFunction::EmptyKey = ComparableFunction(0);
|
||||
|
Reference in New Issue
Block a user