Add a -no-implicit-float flag. This acts like -soft-float, but may generate

floating point instructions that are explicitly specified by the user.


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@66719 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
Bill Wendling
2009-03-11 22:30:01 +00:00
parent a891518c4d
commit f9abd7e33e
3 changed files with 78 additions and 75 deletions

View File

@ -65,6 +65,12 @@ namespace llvm {
/// target FP instructions. /// target FP instructions.
extern bool UseSoftFloat; extern bool UseSoftFloat;
/// NoImplicitFloat - This flag is enabled when the -no-implicit-float flag is
/// specified on the command line. When this flag is on, the code generator
/// won't generate any implicit floating point instructions. I.e., no XMM or
/// x87 or vectorized memcpy/memmove instructions. This is for X86 only.
extern bool NoImplicitFloat;
/// NoZerosInBSS - By default some codegens place zero-initialized data to /// NoZerosInBSS - By default some codegens place zero-initialized data to
/// .bss section. This flag disables such behaviour (necessary, e.g. for /// .bss section. This flag disables such behaviour (necessary, e.g. for
/// crt*.o compiling). /// crt*.o compiling).

View File

@ -29,6 +29,7 @@ namespace llvm {
bool FiniteOnlyFPMathOption; bool FiniteOnlyFPMathOption;
bool HonorSignDependentRoundingFPMathOption; bool HonorSignDependentRoundingFPMathOption;
bool UseSoftFloat; bool UseSoftFloat;
bool NoImplicitFloat;
bool NoZerosInBSS; bool NoZerosInBSS;
bool ExceptionHandling; bool ExceptionHandling;
bool UnwindTablesMandatory; bool UnwindTablesMandatory;
@ -43,10 +44,10 @@ namespace llvm {
bool DisableRedZone; bool DisableRedZone;
} }
static cl::opt<bool, true> PrintCode("print-machineinstrs", static cl::opt<bool, true>
PrintCode("print-machineinstrs",
cl::desc("Print generated machine code"), cl::desc("Print generated machine code"),
cl::location(PrintMachineCode), cl::init(false)); cl::location(PrintMachineCode), cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
DisableFPElim("disable-fp-elim", DisableFPElim("disable-fp-elim",
cl::desc("Disable frame pointer elimination optimization"), cl::desc("Disable frame pointer elimination optimization"),
@ -68,18 +69,22 @@ EnableFiniteOnlyFPMath("enable-finite-only-fp-math",
cl::location(FiniteOnlyFPMathOption), cl::location(FiniteOnlyFPMathOption),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
EnableHonorSignDependentRoundingFPMath(cl::Hidden, EnableHonorSignDependentRoundingFPMath("enable-sign-dependent-rounding-fp-math",
"enable-sign-dependent-rounding-fp-math", cl::Hidden,
cl::desc("Force codegen to assume rounding mode can change dynamically"), cl::desc("Force codegen to assume rounding mode can change dynamically"),
cl::location(HonorSignDependentRoundingFPMathOption), cl::location(HonorSignDependentRoundingFPMathOption),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
GenerateSoftFloatCalls("soft-float", GenerateSoftFloatCalls("soft-float",
cl::desc("Generate software floating point library calls"), cl::desc("Generate software floating point library calls"),
cl::location(UseSoftFloat), cl::location(UseSoftFloat),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
GenerateNoImplicitFloats("no-implicit-float",
cl::desc("Don't generate implicit floating point instructions (x86-only)"),
cl::location(NoImplicitFloat),
cl::init(false));
static cl::opt<bool, true>
DontPlaceZerosInBSS("nozero-initialized-in-bss", DontPlaceZerosInBSS("nozero-initialized-in-bss",
cl::desc("Don't place zero-initialized symbols into bss section"), cl::desc("Don't place zero-initialized symbols into bss section"),
cl::location(NoZerosInBSS), cl::location(NoZerosInBSS),
@ -96,8 +101,7 @@ EnableUnwindTables("unwind-tables",
cl::init(false)); cl::init(false));
static cl::opt<llvm::Reloc::Model, true> static cl::opt<llvm::Reloc::Model, true>
DefRelocationModel( DefRelocationModel("relocation-model",
"relocation-model",
cl::desc("Choose relocation model"), cl::desc("Choose relocation model"),
cl::location(RelocationModel), cl::location(RelocationModel),
cl::init(Reloc::Default), cl::init(Reloc::Default),
@ -112,8 +116,7 @@ DefRelocationModel(
"Relocatable external references, non-relocatable code"), "Relocatable external references, non-relocatable code"),
clEnumValEnd)); clEnumValEnd));
static cl::opt<llvm::CodeModel::Model, true> static cl::opt<llvm::CodeModel::Model, true>
DefCodeModel( DefCodeModel("code-model",
"code-model",
cl::desc("Choose code model"), cl::desc("Choose code model"),
cl::location(CMModel), cl::location(CMModel),
cl::init(CodeModel::Default), cl::init(CodeModel::Default),
@ -129,42 +132,35 @@ DefCodeModel(
clEnumValN(CodeModel::Large, "large", clEnumValN(CodeModel::Large, "large",
"Large code model"), "Large code model"),
clEnumValEnd)); clEnumValEnd));
static cl::opt<bool, true> static cl::opt<bool, true>
EnablePerformTailCallOpt("tailcallopt", EnablePerformTailCallOpt("tailcallopt",
cl::desc("Turn on tail call optimization."), cl::desc("Turn on tail call optimization."),
cl::location(PerformTailCallOpt), cl::location(PerformTailCallOpt),
cl::init(false)); cl::init(false));
static cl::opt<unsigned, true> static cl::opt<unsigned, true>
OverrideStackAlignment("stack-alignment", OverrideStackAlignment("stack-alignment",
cl::desc("Override default stack alignment"), cl::desc("Override default stack alignment"),
cl::location(StackAlignment), cl::location(StackAlignment),
cl::init(0)); cl::init(0));
static cl::opt<bool, true> static cl::opt<bool, true>
EnableRealignStack("realign-stack", EnableRealignStack("realign-stack",
cl::desc("Realign stack if needed"), cl::desc("Realign stack if needed"),
cl::location(RealignStack), cl::location(RealignStack),
cl::init(true)); cl::init(true));
static cl::opt<bool, true> static cl::opt<bool, true>
AsmVerbose("asm-verbose", cl::desc("Add comments to directives."), AsmVerbose("asm-verbose", cl::desc("Add comments to directives."),
cl::location(VerboseAsm), cl::location(VerboseAsm),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
DisableSwitchTables(cl::Hidden, "disable-jump-tables", DisableSwitchTables(cl::Hidden, "disable-jump-tables",
cl::desc("Do not generate jump tables."), cl::desc("Do not generate jump tables."),
cl::location(DisableJumpTables), cl::location(DisableJumpTables),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
EnableStrongPHIElim(cl::Hidden, "strong-phi-elim", EnableStrongPHIElim(cl::Hidden, "strong-phi-elim",
cl::desc("Use strong PHI elimination."), cl::desc("Use strong PHI elimination."),
cl::location(StrongPHIElim), cl::location(StrongPHIElim),
cl::init(false)); cl::init(false));
static cl::opt<bool, true> static cl::opt<bool, true>
DisableRedZoneOption("disable-red-zone", DisableRedZoneOption("disable-red-zone",
cl::desc("Do not emit code that uses the red zone."), cl::desc("Do not emit code that uses the red zone."),

View File

@ -493,7 +493,7 @@ X86TargetLowering::X86TargetLowering(X86TargetMachine &TM)
} }
// Long double always uses X87. // Long double always uses X87.
if (!UseSoftFloat) { if (!UseSoftFloat && !NoImplicitFloat) {
addRegisterClass(MVT::f80, X86::RFP80RegisterClass); addRegisterClass(MVT::f80, X86::RFP80RegisterClass);
setOperationAction(ISD::UNDEF, MVT::f80, Expand); setOperationAction(ISD::UNDEF, MVT::f80, Expand);
setOperationAction(ISD::FCOPYSIGN, MVT::f80, Expand); setOperationAction(ISD::FCOPYSIGN, MVT::f80, Expand);
@ -582,7 +582,7 @@ X86TargetLowering::X86TargetLowering(X86TargetMachine &TM)
// FIXME: In order to prevent SSE instructions being expanded to MMX ones // FIXME: In order to prevent SSE instructions being expanded to MMX ones
// with -msoft-float, disable use of MMX as well. // with -msoft-float, disable use of MMX as well.
if (!UseSoftFloat && !DisableMMX && Subtarget->hasMMX()) { if (!UseSoftFloat && !NoImplicitFloat && !DisableMMX && Subtarget->hasMMX()) {
addRegisterClass(MVT::v8i8, X86::VR64RegisterClass); addRegisterClass(MVT::v8i8, X86::VR64RegisterClass);
addRegisterClass(MVT::v4i16, X86::VR64RegisterClass); addRegisterClass(MVT::v4i16, X86::VR64RegisterClass);
addRegisterClass(MVT::v2i32, X86::VR64RegisterClass); addRegisterClass(MVT::v2i32, X86::VR64RegisterClass);
@ -662,7 +662,7 @@ X86TargetLowering::X86TargetLowering(X86TargetMachine &TM)
setOperationAction(ISD::SELECT, MVT::v1i64, Custom); setOperationAction(ISD::SELECT, MVT::v1i64, Custom);
} }
if (!UseSoftFloat && Subtarget->hasSSE1()) { if (!UseSoftFloat && !NoImplicitFloat && Subtarget->hasSSE1()) {
addRegisterClass(MVT::v4f32, X86::VR128RegisterClass); addRegisterClass(MVT::v4f32, X86::VR128RegisterClass);
setOperationAction(ISD::FADD, MVT::v4f32, Legal); setOperationAction(ISD::FADD, MVT::v4f32, Legal);
@ -679,11 +679,11 @@ X86TargetLowering::X86TargetLowering(X86TargetMachine &TM)
setOperationAction(ISD::VSETCC, MVT::v4f32, Custom); setOperationAction(ISD::VSETCC, MVT::v4f32, Custom);
} }
if (!UseSoftFloat && Subtarget->hasSSE2()) { if (!UseSoftFloat && !NoImplicitFloat && Subtarget->hasSSE2()) {
addRegisterClass(MVT::v2f64, X86::VR128RegisterClass); addRegisterClass(MVT::v2f64, X86::VR128RegisterClass);
// FIXME: Unfortunately -soft-float means XMM registers cannot be used even // FIXME: Unfortunately -soft-float and -no-implicit-float means XMM
// for integer operations. // registers cannot be used even for integer operations.
addRegisterClass(MVT::v16i8, X86::VR128RegisterClass); addRegisterClass(MVT::v16i8, X86::VR128RegisterClass);
addRegisterClass(MVT::v8i16, X86::VR128RegisterClass); addRegisterClass(MVT::v8i16, X86::VR128RegisterClass);
addRegisterClass(MVT::v4i32, X86::VR128RegisterClass); addRegisterClass(MVT::v4i32, X86::VR128RegisterClass);
@ -727,12 +727,14 @@ X86TargetLowering::X86TargetLowering(X86TargetMachine &TM)
setOperationAction(ISD::VECTOR_SHUFFLE, VT, Custom); setOperationAction(ISD::VECTOR_SHUFFLE, VT, Custom);
setOperationAction(ISD::EXTRACT_VECTOR_ELT, VT, Custom); setOperationAction(ISD::EXTRACT_VECTOR_ELT, VT, Custom);
} }
setOperationAction(ISD::BUILD_VECTOR, MVT::v2f64, Custom); setOperationAction(ISD::BUILD_VECTOR, MVT::v2f64, Custom);
setOperationAction(ISD::BUILD_VECTOR, MVT::v2i64, Custom); setOperationAction(ISD::BUILD_VECTOR, MVT::v2i64, Custom);
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2f64, Custom); setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2f64, Custom);
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2i64, Custom); setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v2i64, Custom);
setOperationAction(ISD::INSERT_VECTOR_ELT, MVT::v2f64, Custom); setOperationAction(ISD::INSERT_VECTOR_ELT, MVT::v2f64, Custom);
setOperationAction(ISD::EXTRACT_VECTOR_ELT, MVT::v2f64, Custom); setOperationAction(ISD::EXTRACT_VECTOR_ELT, MVT::v2f64, Custom);
if (Subtarget->is64Bit()) { if (Subtarget->is64Bit()) {
setOperationAction(ISD::INSERT_VECTOR_ELT, MVT::v2i64, Custom); setOperationAction(ISD::INSERT_VECTOR_ELT, MVT::v2i64, Custom);
setOperationAction(ISD::EXTRACT_VECTOR_ELT, MVT::v2i64, Custom); setOperationAction(ISD::EXTRACT_VECTOR_ELT, MVT::v2i64, Custom);
@ -888,7 +890,7 @@ X86TargetLowering::getOptimalMemOpType(uint64_t Size, unsigned Align,
// FIXME: This turns off use of xmm stores for memset/memcpy on targets like // FIXME: This turns off use of xmm stores for memset/memcpy on targets like
// linux. This is because the stack realignment code can't handle certain // linux. This is because the stack realignment code can't handle certain
// cases like PR2962. This should be removed when PR2962 is fixed. // cases like PR2962. This should be removed when PR2962 is fixed.
if (Subtarget->getStackAlignment() >= 16) { if (!NoImplicitFloat && Subtarget->getStackAlignment() >= 16) {
if ((isSrcConst || isSrcStr) && Subtarget->hasSSE2() && Size >= 16) if ((isSrcConst || isSrcStr) && Subtarget->hasSSE2() && Size >= 16)
return MVT::v4i32; return MVT::v4i32;
if ((isSrcConst || isSrcStr) && Subtarget->hasSSE1() && Size >= 16) if ((isSrcConst || isSrcStr) && Subtarget->hasSSE1() && Size >= 16)
@ -899,7 +901,6 @@ X86TargetLowering::getOptimalMemOpType(uint64_t Size, unsigned Align,
return MVT::i32; return MVT::i32;
} }
/// getPICJumpTableRelocaBase - Returns relocation base for the given PIC /// getPICJumpTableRelocaBase - Returns relocation base for the given PIC
/// jumptable. /// jumptable.
SDValue X86TargetLowering::getPICJumpTableRelocBase(SDValue Table, SDValue X86TargetLowering::getPICJumpTableRelocBase(SDValue Table,
@ -1434,13 +1435,13 @@ X86TargetLowering::LowerFORMAL_ARGUMENTS(SDValue Op, SelectionDAG &DAG) {
assert(!(NumXMMRegs && !Subtarget->hasSSE1()) && assert(!(NumXMMRegs && !Subtarget->hasSSE1()) &&
"SSE register cannot be used when SSE is disabled!"); "SSE register cannot be used when SSE is disabled!");
assert(!(NumXMMRegs && UseSoftFloat) && assert(!(NumXMMRegs && UseSoftFloat && NoImplicitFloat) &&
"SSE register cannot be used when SSE is disabled!"); "SSE register cannot be used when SSE is disabled!");
if (UseSoftFloat || !Subtarget->hasSSE1()) { if (UseSoftFloat || NoImplicitFloat || !Subtarget->hasSSE1())
// Kernel mode asks for SSE to be disabled, so don't push them // Kernel mode asks for SSE to be disabled, so don't push them
// on the stack. // on the stack.
TotalNumXMMRegs = 0; TotalNumXMMRegs = 0;
}
// For X86-64, if there are vararg parameters that are passed via // For X86-64, if there are vararg parameters that are passed via
// registers, then we must store them to their spots on the stack so they // registers, then we must store them to their spots on the stack so they
// may be loaded by deferencing the result of va_next. // may be loaded by deferencing the result of va_next.