mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2026-04-26 12:20:42 +00:00
Rename many DataLayout variables from TD to DL.
I am really sorry for the noise, but the current state where some parts of the code use TD (from the old name: TargetData) and other parts use DL makes it hard to write a patch that changes where those variables come from and how they are passed along. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@201827 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
@@ -218,7 +218,7 @@ Instruction *InstCombiner::
|
||||
FoldCmpLoadFromIndexedGlobal(GetElementPtrInst *GEP, GlobalVariable *GV,
|
||||
CmpInst &ICI, ConstantInt *AndCst) {
|
||||
// We need TD information to know the pointer size unless this is inbounds.
|
||||
if (!GEP->isInBounds() && TD == 0)
|
||||
if (!GEP->isInBounds() && DL == 0)
|
||||
return 0;
|
||||
|
||||
Constant *Init = GV->getInitializer();
|
||||
@@ -307,7 +307,7 @@ FoldCmpLoadFromIndexedGlobal(GetElementPtrInst *GEP, GlobalVariable *GV,
|
||||
|
||||
// Find out if the comparison would be true or false for the i'th element.
|
||||
Constant *C = ConstantFoldCompareInstOperands(ICI.getPredicate(), Elt,
|
||||
CompareRHS, TD, TLI);
|
||||
CompareRHS, DL, TLI);
|
||||
// If the result is undef for this element, ignore it.
|
||||
if (isa<UndefValue>(C)) {
|
||||
// Extend range state machines to cover this element in case there is an
|
||||
@@ -386,7 +386,7 @@ FoldCmpLoadFromIndexedGlobal(GetElementPtrInst *GEP, GlobalVariable *GV,
|
||||
// index down like the GEP would do implicitly. We don't have to do this for
|
||||
// an inbounds GEP because the index can't be out of range.
|
||||
if (!GEP->isInBounds()) {
|
||||
Type *IntPtrTy = TD->getIntPtrType(GEP->getType());
|
||||
Type *IntPtrTy = DL->getIntPtrType(GEP->getType());
|
||||
unsigned PtrSize = IntPtrTy->getIntegerBitWidth();
|
||||
if (Idx->getType()->getPrimitiveSizeInBits() > PtrSize)
|
||||
Idx = Builder->CreateTrunc(Idx, IntPtrTy);
|
||||
@@ -475,8 +475,8 @@ FoldCmpLoadFromIndexedGlobal(GetElementPtrInst *GEP, GlobalVariable *GV,
|
||||
// - Default to i32
|
||||
if (ArrayElementCount <= Idx->getType()->getIntegerBitWidth())
|
||||
Ty = Idx->getType();
|
||||
else if (TD)
|
||||
Ty = TD->getSmallestLegalIntType(Init->getContext(), ArrayElementCount);
|
||||
else if (DL)
|
||||
Ty = DL->getSmallestLegalIntType(Init->getContext(), ArrayElementCount);
|
||||
else if (ArrayElementCount <= 32)
|
||||
Ty = Type::getInt32Ty(Init->getContext());
|
||||
|
||||
@@ -503,7 +503,7 @@ FoldCmpLoadFromIndexedGlobal(GetElementPtrInst *GEP, GlobalVariable *GV,
|
||||
/// If we can't emit an optimized form for this expression, this returns null.
|
||||
///
|
||||
static Value *EvaluateGEPOffsetExpression(User *GEP, InstCombiner &IC) {
|
||||
DataLayout &TD = *IC.getDataLayout();
|
||||
DataLayout &DL = *IC.getDataLayout();
|
||||
gep_type_iterator GTI = gep_type_begin(GEP);
|
||||
|
||||
// Check to see if this gep only has a single variable index. If so, and if
|
||||
@@ -520,9 +520,9 @@ static Value *EvaluateGEPOffsetExpression(User *GEP, InstCombiner &IC) {
|
||||
|
||||
// Handle a struct index, which adds its field offset to the pointer.
|
||||
if (StructType *STy = dyn_cast<StructType>(*GTI)) {
|
||||
Offset += TD.getStructLayout(STy)->getElementOffset(CI->getZExtValue());
|
||||
Offset += DL.getStructLayout(STy)->getElementOffset(CI->getZExtValue());
|
||||
} else {
|
||||
uint64_t Size = TD.getTypeAllocSize(GTI.getIndexedType());
|
||||
uint64_t Size = DL.getTypeAllocSize(GTI.getIndexedType());
|
||||
Offset += Size*CI->getSExtValue();
|
||||
}
|
||||
} else {
|
||||
@@ -538,7 +538,7 @@ static Value *EvaluateGEPOffsetExpression(User *GEP, InstCombiner &IC) {
|
||||
Value *VariableIdx = GEP->getOperand(i);
|
||||
// Determine the scale factor of the variable element. For example, this is
|
||||
// 4 if the variable index is into an array of i32.
|
||||
uint64_t VariableScale = TD.getTypeAllocSize(GTI.getIndexedType());
|
||||
uint64_t VariableScale = DL.getTypeAllocSize(GTI.getIndexedType());
|
||||
|
||||
// Verify that there are no other variable indices. If so, emit the hard way.
|
||||
for (++i, ++GTI; i != e; ++i, ++GTI) {
|
||||
@@ -550,9 +550,9 @@ static Value *EvaluateGEPOffsetExpression(User *GEP, InstCombiner &IC) {
|
||||
|
||||
// Handle a struct index, which adds its field offset to the pointer.
|
||||
if (StructType *STy = dyn_cast<StructType>(*GTI)) {
|
||||
Offset += TD.getStructLayout(STy)->getElementOffset(CI->getZExtValue());
|
||||
Offset += DL.getStructLayout(STy)->getElementOffset(CI->getZExtValue());
|
||||
} else {
|
||||
uint64_t Size = TD.getTypeAllocSize(GTI.getIndexedType());
|
||||
uint64_t Size = DL.getTypeAllocSize(GTI.getIndexedType());
|
||||
Offset += Size*CI->getSExtValue();
|
||||
}
|
||||
}
|
||||
@@ -562,7 +562,7 @@ static Value *EvaluateGEPOffsetExpression(User *GEP, InstCombiner &IC) {
|
||||
// Okay, we know we have a single variable index, which must be a
|
||||
// pointer/array/vector index. If there is no offset, life is simple, return
|
||||
// the index.
|
||||
Type *IntPtrTy = TD.getIntPtrType(GEP->getOperand(0)->getType());
|
||||
Type *IntPtrTy = DL.getIntPtrType(GEP->getOperand(0)->getType());
|
||||
unsigned IntPtrWidth = IntPtrTy->getIntegerBitWidth();
|
||||
if (Offset == 0) {
|
||||
// Cast to intptrty in case a truncation occurs. If an extension is needed,
|
||||
@@ -615,7 +615,7 @@ Instruction *InstCombiner::FoldGEPICmp(GEPOperator *GEPLHS, Value *RHS,
|
||||
RHS = BCI->getOperand(0);
|
||||
|
||||
Value *PtrBase = GEPLHS->getOperand(0);
|
||||
if (TD && PtrBase == RHS && GEPLHS->isInBounds()) {
|
||||
if (DL && PtrBase == RHS && GEPLHS->isInBounds()) {
|
||||
// ((gep Ptr, OFFSET) cmp Ptr) ---> (OFFSET cmp 0).
|
||||
// This transformation (ignoring the base and scales) is valid because we
|
||||
// know pointers can't overflow since the gep is inbounds. See if we can
|
||||
@@ -648,7 +648,7 @@ Instruction *InstCombiner::FoldGEPICmp(GEPOperator *GEPLHS, Value *RHS,
|
||||
// If we're comparing GEPs with two base pointers that only differ in type
|
||||
// and both GEPs have only constant indices or just one use, then fold
|
||||
// the compare with the adjusted indices.
|
||||
if (TD && GEPLHS->isInBounds() && GEPRHS->isInBounds() &&
|
||||
if (DL && GEPLHS->isInBounds() && GEPRHS->isInBounds() &&
|
||||
(GEPLHS->hasAllConstantIndices() || GEPLHS->hasOneUse()) &&
|
||||
(GEPRHS->hasAllConstantIndices() || GEPRHS->hasOneUse()) &&
|
||||
PtrBase->stripPointerCasts() ==
|
||||
@@ -719,7 +719,7 @@ Instruction *InstCombiner::FoldGEPICmp(GEPOperator *GEPLHS, Value *RHS,
|
||||
|
||||
// Only lower this if the icmp is the only user of the GEP or if we expect
|
||||
// the result to fold to a constant!
|
||||
if (TD &&
|
||||
if (DL &&
|
||||
GEPsInBounds &&
|
||||
(isa<ConstantExpr>(GEPLHS) || GEPLHS->hasOneUse()) &&
|
||||
(isa<ConstantExpr>(GEPRHS) || GEPRHS->hasOneUse())) {
|
||||
@@ -1792,8 +1792,8 @@ Instruction *InstCombiner::visitICmpInstWithCastAndCast(ICmpInst &ICI) {
|
||||
|
||||
// Turn icmp (ptrtoint x), (ptrtoint/c) into a compare of the input if the
|
||||
// integer type is the same size as the pointer type.
|
||||
if (TD && LHSCI->getOpcode() == Instruction::PtrToInt &&
|
||||
TD->getPointerTypeSizeInBits(SrcTy) == DestTy->getIntegerBitWidth()) {
|
||||
if (DL && LHSCI->getOpcode() == Instruction::PtrToInt &&
|
||||
DL->getPointerTypeSizeInBits(SrcTy) == DestTy->getIntegerBitWidth()) {
|
||||
Value *RHSOp = 0;
|
||||
if (Constant *RHSC = dyn_cast<Constant>(ICI.getOperand(1))) {
|
||||
RHSOp = ConstantExpr::getIntToPtr(RHSC, SrcTy);
|
||||
@@ -2104,7 +2104,7 @@ Instruction *InstCombiner::visitICmpInst(ICmpInst &I) {
|
||||
Changed = true;
|
||||
}
|
||||
|
||||
if (Value *V = SimplifyICmpInst(I.getPredicate(), Op0, Op1, TD))
|
||||
if (Value *V = SimplifyICmpInst(I.getPredicate(), Op0, Op1, DL))
|
||||
return ReplaceInstUsesWith(I, V);
|
||||
|
||||
// comparing -val or val with non-zero is the same as just comparing val
|
||||
@@ -2172,8 +2172,8 @@ Instruction *InstCombiner::visitICmpInst(ICmpInst &I) {
|
||||
unsigned BitWidth = 0;
|
||||
if (Ty->isIntOrIntVectorTy())
|
||||
BitWidth = Ty->getScalarSizeInBits();
|
||||
else if (TD) // Pointers require TD info to get their size.
|
||||
BitWidth = TD->getTypeSizeInBits(Ty->getScalarType());
|
||||
else if (DL) // Pointers require DL info to get their size.
|
||||
BitWidth = DL->getTypeSizeInBits(Ty->getScalarType());
|
||||
|
||||
bool isSignBit = false;
|
||||
|
||||
@@ -2532,8 +2532,8 @@ Instruction *InstCombiner::visitICmpInst(ICmpInst &I) {
|
||||
}
|
||||
case Instruction::IntToPtr:
|
||||
// icmp pred inttoptr(X), null -> icmp pred X, 0
|
||||
if (RHSC->isNullValue() && TD &&
|
||||
TD->getIntPtrType(RHSC->getType()) ==
|
||||
if (RHSC->isNullValue() && DL &&
|
||||
DL->getIntPtrType(RHSC->getType()) ==
|
||||
LHSI->getOperand(0)->getType())
|
||||
return new ICmpInst(I.getPredicate(), LHSI->getOperand(0),
|
||||
Constant::getNullValue(LHSI->getOperand(0)->getType()));
|
||||
@@ -3229,7 +3229,7 @@ Instruction *InstCombiner::visitFCmpInst(FCmpInst &I) {
|
||||
|
||||
Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
|
||||
|
||||
if (Value *V = SimplifyFCmpInst(I.getPredicate(), Op0, Op1, TD))
|
||||
if (Value *V = SimplifyFCmpInst(I.getPredicate(), Op0, Op1, DL))
|
||||
return ReplaceInstUsesWith(I, V);
|
||||
|
||||
// Simplify 'fcmp pred X, X'
|
||||
|
||||
Reference in New Issue
Block a user