mirror of
https://github.com/ldc-developers/ldc.git
synced 2025-05-02 16:11:08 +03:00
156 lines
4.6 KiB
C++
156 lines
4.6 KiB
C++
//===-- targetmachine.cpp -------------------------------------------------===//
|
||
//
|
||
// LDC – the LLVM D compiler
|
||
//
|
||
// This file is distributed under the BSD-style LDC license. See the LICENSE
|
||
// file for details.
|
||
//
|
||
//===----------------------------------------------------------------------===//
|
||
//
|
||
// Note: The target CPU detection logic has been adapted from Clang
|
||
// (lib/Driver/Tools.cpp).
|
||
//
|
||
//===----------------------------------------------------------------------===//
|
||
|
||
#include "driver/targetmachine.h"
|
||
#include "llvm/ADT/Triple.h"
|
||
#include "llvm/MC/SubtargetFeature.h"
|
||
#include "llvm/Support/Host.h"
|
||
#include "llvm/Support/TargetRegistry.h"
|
||
#include "llvm/Support/TargetSelect.h"
|
||
#include "llvm/Target/TargetMachine.h"
|
||
#include "llvm/Target/TargetOptions.h"
|
||
#include "mars.h"
|
||
|
||
|
||
static std::string getX86TargetCPU(const std::string &cpu,
|
||
const llvm::Triple &triple)
|
||
{
|
||
if (!cpu.empty()) {
|
||
if (cpu != "native")
|
||
return cpu;
|
||
|
||
// FIXME: Reject attempts to use -mcpu=native unless the target matches
|
||
// the host.
|
||
//
|
||
// FIXME: We should also incorporate the detected target features for use
|
||
// with -native.
|
||
std::string cpu = llvm::sys::getHostCPUName();
|
||
if (!cpu.empty() && cpu != "generic")
|
||
return cpu;
|
||
}
|
||
|
||
// Select the default CPU if none was given (or detection failed).
|
||
|
||
bool is64Bit = triple.getArch() == llvm::Triple::x86_64;
|
||
|
||
if (triple.isOSDarwin())
|
||
return is64Bit ? "core2" : "yonah";
|
||
|
||
// Everything else goes to x86-64 in 64-bit mode.
|
||
if (is64Bit)
|
||
return "x86-64";
|
||
|
||
if (triple.getOSName().startswith("haiku"))
|
||
return "i586";
|
||
if (triple.getOSName().startswith("openbsd"))
|
||
return "i486";
|
||
if (triple.getOSName().startswith("bitrig"))
|
||
return "i686";
|
||
if (triple.getOSName().startswith("freebsd"))
|
||
return "i486";
|
||
if (triple.getOSName().startswith("netbsd"))
|
||
return "i486";
|
||
#if LDC_LLVM_VER >= 302
|
||
// All x86 devices running Android have core2 as their common
|
||
// denominator. This makes a better choice than pentium4.
|
||
if (triple.getEnvironment() == llvm::Triple::Android)
|
||
return "core2";
|
||
#endif
|
||
|
||
// Fallback to p4.
|
||
return "pentium4";
|
||
}
|
||
|
||
llvm::TargetMachine* createTargetMachine(
|
||
std::string targetTriple,
|
||
std::string arch,
|
||
std::string cpu,
|
||
std::vector<std::string> attrs,
|
||
ExplicitBitness::Type bitness,
|
||
llvm::Reloc::Model relocModel,
|
||
llvm::CodeModel::Model codeModel,
|
||
llvm::CodeGenOpt::Level codeGenOptLevel,
|
||
bool genDebugInfo)
|
||
{
|
||
// Determine target triple. If the user didn't explicitly specify one, use
|
||
// the one set at LLVM configure time.
|
||
llvm::Triple triple;
|
||
if (targetTriple.empty())
|
||
{
|
||
triple = llvm::Triple(llvm::sys::getDefaultTargetTriple());
|
||
|
||
// Handle -m32/-m64.
|
||
if (sizeof(void*) == 4 && bitness == ExplicitBitness::M64)
|
||
{
|
||
triple = triple.get64BitArchVariant();
|
||
}
|
||
else if (sizeof(void*) == 8 && bitness == ExplicitBitness::M32)
|
||
{
|
||
triple = triple.get32BitArchVariant();
|
||
}
|
||
}
|
||
else
|
||
{
|
||
triple = llvm::Triple(llvm::Triple::normalize(targetTriple));
|
||
}
|
||
|
||
// Look up the LLVM backend to use.
|
||
std::string errMsg;
|
||
const llvm::Target *target =
|
||
llvm::TargetRegistry::lookupTarget(arch, triple, errMsg);
|
||
if (target == 0)
|
||
{
|
||
error("%s", errMsg.c_str());
|
||
fatal();
|
||
}
|
||
|
||
// With an empty CPU string, LLVM will default to the host CPU, which is
|
||
// usually not what we want (expected behavior from other compilers is
|
||
// to default to "generic").
|
||
if (triple.getArch() == llvm::Triple::x86_64 ||
|
||
triple.getArch() == llvm::Triple::x86)
|
||
{
|
||
cpu = getX86TargetCPU(cpu, triple);
|
||
}
|
||
|
||
// Package up features to be passed to target/subtarget
|
||
std::string FeaturesStr;
|
||
if (cpu.size() || attrs.size())
|
||
{
|
||
llvm::SubtargetFeatures Features;
|
||
for (unsigned i = 0; i != attrs.size(); ++i)
|
||
Features.AddFeature(attrs[i]);
|
||
FeaturesStr = Features.getString();
|
||
}
|
||
|
||
if (triple.isMacOSX() && relocModel == llvm::Reloc::Default)
|
||
{
|
||
// OS X defaults to PIC (and as of 10.7.5/LLVM 3.1-3.3, TLS use leads
|
||
// to crashes for non-PIC code). LLVM doesn't handle this.
|
||
relocModel = llvm::Reloc::PIC_;
|
||
}
|
||
|
||
llvm::TargetOptions targetOptions;
|
||
targetOptions.NoFramePointerElim = genDebugInfo;
|
||
|
||
return target->createTargetMachine(
|
||
triple.str(),
|
||
cpu,
|
||
FeaturesStr,
|
||
targetOptions,
|
||
relocModel,
|
||
codeModel,
|
||
codeGenOptLevel
|
||
);
|
||
}
|