mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-07-25 14:30:32 +00:00
Slightly simplify code with helper functions
e.g. Use Ty->getPointerElementType() instead of cast<PointerType>(Ty)->getElementType() git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@188223 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
d9a84efe44
commit
89fa38e15c
@ -338,7 +338,7 @@ static bool ReadDataFromGlobal(Constant *C, uint64_t ByteOffset,
|
|||||||
|
|
||||||
if (isa<ConstantArray>(C) || isa<ConstantVector>(C) ||
|
if (isa<ConstantArray>(C) || isa<ConstantVector>(C) ||
|
||||||
isa<ConstantDataSequential>(C)) {
|
isa<ConstantDataSequential>(C)) {
|
||||||
Type *EltTy = cast<SequentialType>(C->getType())->getElementType();
|
Type *EltTy = C->getType()->getSequentialElementType();
|
||||||
uint64_t EltSize = TD.getTypeAllocSize(EltTy);
|
uint64_t EltSize = TD.getTypeAllocSize(EltTy);
|
||||||
uint64_t Index = ByteOffset / EltSize;
|
uint64_t Index = ByteOffset / EltSize;
|
||||||
uint64_t Offset = ByteOffset - Index * EltSize;
|
uint64_t Offset = ByteOffset - Index * EltSize;
|
||||||
@ -346,7 +346,7 @@ static bool ReadDataFromGlobal(Constant *C, uint64_t ByteOffset,
|
|||||||
if (ArrayType *AT = dyn_cast<ArrayType>(C->getType()))
|
if (ArrayType *AT = dyn_cast<ArrayType>(C->getType()))
|
||||||
NumElts = AT->getNumElements();
|
NumElts = AT->getNumElements();
|
||||||
else
|
else
|
||||||
NumElts = cast<VectorType>(C->getType())->getNumElements();
|
NumElts = C->getType()->getVectorNumElements();
|
||||||
|
|
||||||
for (; Index != NumElts; ++Index) {
|
for (; Index != NumElts; ++Index) {
|
||||||
if (!ReadDataFromGlobal(C->getAggregateElement(Index), Offset, CurPtr,
|
if (!ReadDataFromGlobal(C->getAggregateElement(Index), Offset, CurPtr,
|
||||||
@ -409,7 +409,8 @@ static Constant *FoldReinterpretLoadFromConstPtr(Constant *C,
|
|||||||
}
|
}
|
||||||
|
|
||||||
unsigned BytesLoaded = (IntType->getBitWidth() + 7) / 8;
|
unsigned BytesLoaded = (IntType->getBitWidth() + 7) / 8;
|
||||||
if (BytesLoaded > 32 || BytesLoaded == 0) return 0;
|
if (BytesLoaded > 32 || BytesLoaded == 0)
|
||||||
|
return 0;
|
||||||
|
|
||||||
GlobalValue *GVal;
|
GlobalValue *GVal;
|
||||||
APInt Offset(TD.getPointerSizeInBits(), 0);
|
APInt Offset(TD.getPointerSizeInBits(), 0);
|
||||||
@ -423,7 +424,8 @@ static Constant *FoldReinterpretLoadFromConstPtr(Constant *C,
|
|||||||
|
|
||||||
// If we're loading off the beginning of the global, some bytes may be valid,
|
// If we're loading off the beginning of the global, some bytes may be valid,
|
||||||
// but we don't try to handle this.
|
// but we don't try to handle this.
|
||||||
if (Offset.isNegative()) return 0;
|
if (Offset.isNegative())
|
||||||
|
return 0;
|
||||||
|
|
||||||
// If we're not accessing anything in this constant, the result is undefined.
|
// If we're not accessing anything in this constant, the result is undefined.
|
||||||
if (Offset.getZExtValue() >=
|
if (Offset.getZExtValue() >=
|
||||||
@ -659,11 +661,12 @@ static Constant *SymbolicallyEvaluateGEP(ArrayRef<Constant *> Ops,
|
|||||||
Type *ResultTy, const DataLayout *TD,
|
Type *ResultTy, const DataLayout *TD,
|
||||||
const TargetLibraryInfo *TLI) {
|
const TargetLibraryInfo *TLI) {
|
||||||
Constant *Ptr = Ops[0];
|
Constant *Ptr = Ops[0];
|
||||||
if (!TD || !cast<PointerType>(Ptr->getType())->getElementType()->isSized() ||
|
if (!TD || !Ptr->getType()->getPointerElementType()->isSized() ||
|
||||||
!Ptr->getType()->isPointerTy())
|
!Ptr->getType()->isPointerTy())
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
Type *IntPtrTy = TD->getIntPtrType(Ptr->getContext());
|
Type *IntPtrTy = TD->getIntPtrType(Ptr->getContext());
|
||||||
|
Type *ResultElementTy = ResultTy->getPointerElementType();
|
||||||
|
|
||||||
// If this is a constant expr gep that is effectively computing an
|
// If this is a constant expr gep that is effectively computing an
|
||||||
// "offsetof", fold it into 'cast int Size to T*' instead of 'gep 0, 0, 12'
|
// "offsetof", fold it into 'cast int Size to T*' instead of 'gep 0, 0, 12'
|
||||||
@ -672,8 +675,7 @@ static Constant *SymbolicallyEvaluateGEP(ArrayRef<Constant *> Ops,
|
|||||||
|
|
||||||
// If this is "gep i8* Ptr, (sub 0, V)", fold this as:
|
// If this is "gep i8* Ptr, (sub 0, V)", fold this as:
|
||||||
// "inttoptr (sub (ptrtoint Ptr), V)"
|
// "inttoptr (sub (ptrtoint Ptr), V)"
|
||||||
if (Ops.size() == 2 &&
|
if (Ops.size() == 2 && ResultElementTy->isIntegerTy(8)) {
|
||||||
cast<PointerType>(ResultTy)->getElementType()->isIntegerTy(8)) {
|
|
||||||
ConstantExpr *CE = dyn_cast<ConstantExpr>(Ops[1]);
|
ConstantExpr *CE = dyn_cast<ConstantExpr>(Ops[1]);
|
||||||
assert((CE == 0 || CE->getType() == IntPtrTy) &&
|
assert((CE == 0 || CE->getType() == IntPtrTy) &&
|
||||||
"CastGEPIndices didn't canonicalize index types!");
|
"CastGEPIndices didn't canonicalize index types!");
|
||||||
@ -789,7 +791,7 @@ static Constant *SymbolicallyEvaluateGEP(ArrayRef<Constant *> Ops,
|
|||||||
// We've reached some non-indexable type.
|
// We've reached some non-indexable type.
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
} while (Ty != cast<PointerType>(ResultTy)->getElementType());
|
} while (Ty != ResultElementTy);
|
||||||
|
|
||||||
// If we haven't used up the entire offset by descending the static
|
// If we haven't used up the entire offset by descending the static
|
||||||
// type, then the offset is pointing into the middle of an indivisible
|
// type, then the offset is pointing into the middle of an indivisible
|
||||||
@ -799,12 +801,12 @@ static Constant *SymbolicallyEvaluateGEP(ArrayRef<Constant *> Ops,
|
|||||||
|
|
||||||
// Create a GEP.
|
// Create a GEP.
|
||||||
Constant *C = ConstantExpr::getGetElementPtr(Ptr, NewIdxs);
|
Constant *C = ConstantExpr::getGetElementPtr(Ptr, NewIdxs);
|
||||||
assert(cast<PointerType>(C->getType())->getElementType() == Ty &&
|
assert(C->getType()->getPointerElementType() == Ty &&
|
||||||
"Computed GetElementPtr has unexpected type!");
|
"Computed GetElementPtr has unexpected type!");
|
||||||
|
|
||||||
// If we ended up indexing a member with a type that doesn't match
|
// If we ended up indexing a member with a type that doesn't match
|
||||||
// the type of what the original indices indexed, add a cast.
|
// the type of what the original indices indexed, add a cast.
|
||||||
if (Ty != cast<PointerType>(ResultTy)->getElementType())
|
if (Ty != ResultElementTy)
|
||||||
C = FoldBitCast(C, ResultTy, *TD);
|
C = FoldBitCast(C, ResultTy, *TD);
|
||||||
|
|
||||||
return C;
|
return C;
|
||||||
@ -1265,7 +1267,7 @@ static Constant *ConstantFoldBinaryFP(double (*NativeFP)(double, double),
|
|||||||
static Constant *ConstantFoldConvertToInt(const APFloat &Val,
|
static Constant *ConstantFoldConvertToInt(const APFloat &Val,
|
||||||
bool roundTowardZero, Type *Ty) {
|
bool roundTowardZero, Type *Ty) {
|
||||||
// All of these conversion intrinsics form an integer of at most 64bits.
|
// All of these conversion intrinsics form an integer of at most 64bits.
|
||||||
unsigned ResultWidth = cast<IntegerType>(Ty)->getBitWidth();
|
unsigned ResultWidth = Ty->getIntegerBitWidth();
|
||||||
assert(ResultWidth <= 64 &&
|
assert(ResultWidth <= 64 &&
|
||||||
"Can only constant fold conversions to 64 and 32 bit ints");
|
"Can only constant fold conversions to 64 and 32 bit ints");
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user