Reverts llvm/llvm-project#118734 There are currently some specific versions of MSVC that are miscompiling this code (we think). We don't know why as all the other build bots and at least some folks' local Windows builds work fine. This is a candidate revert to help the relevant folks catch their builders up and have time to debug the issue. However, the expectation is to roll forward at some point with a workaround if at all possible.
302 lines
11 KiB
C++
302 lines
11 KiB
C++
//===--- NVPTX.cpp - Implement NVPTX target feature support ---------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file implements NVPTX TargetInfo objects.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "NVPTX.h"
|
|
#include "Targets.h"
|
|
#include "clang/Basic/Builtins.h"
|
|
#include "clang/Basic/MacroBuilder.h"
|
|
#include "clang/Basic/TargetBuiltins.h"
|
|
#include "llvm/ADT/StringSwitch.h"
|
|
|
|
using namespace clang;
|
|
using namespace clang::targets;
|
|
|
|
static constexpr Builtin::Info BuiltinInfo[] = {
|
|
#define BUILTIN(ID, TYPE, ATTRS) \
|
|
{#ID, TYPE, ATTRS, nullptr, HeaderDesc::NO_HEADER, ALL_LANGUAGES},
|
|
#define LIBBUILTIN(ID, TYPE, ATTRS, HEADER) \
|
|
{#ID, TYPE, ATTRS, nullptr, HeaderDesc::HEADER, ALL_LANGUAGES},
|
|
#define TARGET_BUILTIN(ID, TYPE, ATTRS, FEATURE) \
|
|
{#ID, TYPE, ATTRS, FEATURE, HeaderDesc::NO_HEADER, ALL_LANGUAGES},
|
|
#include "clang/Basic/BuiltinsNVPTX.def"
|
|
};
|
|
|
|
const char *const NVPTXTargetInfo::GCCRegNames[] = {"r0"};
|
|
|
|
NVPTXTargetInfo::NVPTXTargetInfo(const llvm::Triple &Triple,
|
|
const TargetOptions &Opts,
|
|
unsigned TargetPointerWidth)
|
|
: TargetInfo(Triple) {
|
|
assert((TargetPointerWidth == 32 || TargetPointerWidth == 64) &&
|
|
"NVPTX only supports 32- and 64-bit modes.");
|
|
|
|
PTXVersion = 32;
|
|
for (const StringRef Feature : Opts.FeaturesAsWritten) {
|
|
int PTXV;
|
|
if (!Feature.starts_with("+ptx") ||
|
|
Feature.drop_front(4).getAsInteger(10, PTXV))
|
|
continue;
|
|
PTXVersion = PTXV; // TODO: should it be max(PTXVersion, PTXV)?
|
|
}
|
|
|
|
TLSSupported = false;
|
|
VLASupported = false;
|
|
AddrSpaceMap = &NVPTXAddrSpaceMap;
|
|
UseAddrSpaceMapMangling = true;
|
|
// __bf16 is always available as a load/store only type.
|
|
BFloat16Width = BFloat16Align = 16;
|
|
BFloat16Format = &llvm::APFloat::BFloat();
|
|
|
|
// Define available target features
|
|
// These must be defined in sorted order!
|
|
NoAsmVariants = true;
|
|
GPU = OffloadArch::UNUSED;
|
|
|
|
// PTX supports f16 as a fundamental type.
|
|
HasLegalHalfType = true;
|
|
HasFloat16 = true;
|
|
|
|
if (TargetPointerWidth == 32)
|
|
resetDataLayout("e-p:32:32-i64:64-i128:128-v16:16-v32:32-n16:32:64");
|
|
else if (Opts.NVPTXUseShortPointers)
|
|
resetDataLayout(
|
|
"e-p3:32:32-p4:32:32-p5:32:32-i64:64-i128:128-v16:16-v32:32-n16:32:64");
|
|
else
|
|
resetDataLayout("e-i64:64-i128:128-v16:16-v32:32-n16:32:64");
|
|
|
|
// If possible, get a TargetInfo for our host triple, so we can match its
|
|
// types.
|
|
llvm::Triple HostTriple(Opts.HostTriple);
|
|
if (!HostTriple.isNVPTX())
|
|
HostTarget = AllocateTarget(llvm::Triple(Opts.HostTriple), Opts);
|
|
|
|
// If no host target, make some guesses about the data layout and return.
|
|
if (!HostTarget) {
|
|
LongWidth = LongAlign = TargetPointerWidth;
|
|
PointerWidth = PointerAlign = TargetPointerWidth;
|
|
switch (TargetPointerWidth) {
|
|
case 32:
|
|
SizeType = TargetInfo::UnsignedInt;
|
|
PtrDiffType = TargetInfo::SignedInt;
|
|
IntPtrType = TargetInfo::SignedInt;
|
|
break;
|
|
case 64:
|
|
SizeType = TargetInfo::UnsignedLong;
|
|
PtrDiffType = TargetInfo::SignedLong;
|
|
IntPtrType = TargetInfo::SignedLong;
|
|
break;
|
|
default:
|
|
llvm_unreachable("TargetPointerWidth must be 32 or 64");
|
|
}
|
|
|
|
MaxAtomicInlineWidth = TargetPointerWidth;
|
|
return;
|
|
}
|
|
|
|
// Copy properties from host target.
|
|
PointerWidth = HostTarget->getPointerWidth(LangAS::Default);
|
|
PointerAlign = HostTarget->getPointerAlign(LangAS::Default);
|
|
BoolWidth = HostTarget->getBoolWidth();
|
|
BoolAlign = HostTarget->getBoolAlign();
|
|
IntWidth = HostTarget->getIntWidth();
|
|
IntAlign = HostTarget->getIntAlign();
|
|
HalfWidth = HostTarget->getHalfWidth();
|
|
HalfAlign = HostTarget->getHalfAlign();
|
|
FloatWidth = HostTarget->getFloatWidth();
|
|
FloatAlign = HostTarget->getFloatAlign();
|
|
DoubleWidth = HostTarget->getDoubleWidth();
|
|
DoubleAlign = HostTarget->getDoubleAlign();
|
|
LongWidth = HostTarget->getLongWidth();
|
|
LongAlign = HostTarget->getLongAlign();
|
|
LongLongWidth = HostTarget->getLongLongWidth();
|
|
LongLongAlign = HostTarget->getLongLongAlign();
|
|
MinGlobalAlign = HostTarget->getMinGlobalAlign(/* TypeSize = */ 0,
|
|
/* HasNonWeakDef = */ true);
|
|
NewAlign = HostTarget->getNewAlign();
|
|
DefaultAlignForAttributeAligned =
|
|
HostTarget->getDefaultAlignForAttributeAligned();
|
|
SizeType = HostTarget->getSizeType();
|
|
IntMaxType = HostTarget->getIntMaxType();
|
|
PtrDiffType = HostTarget->getPtrDiffType(LangAS::Default);
|
|
IntPtrType = HostTarget->getIntPtrType();
|
|
WCharType = HostTarget->getWCharType();
|
|
WIntType = HostTarget->getWIntType();
|
|
Char16Type = HostTarget->getChar16Type();
|
|
Char32Type = HostTarget->getChar32Type();
|
|
Int64Type = HostTarget->getInt64Type();
|
|
SigAtomicType = HostTarget->getSigAtomicType();
|
|
ProcessIDType = HostTarget->getProcessIDType();
|
|
|
|
UseBitFieldTypeAlignment = HostTarget->useBitFieldTypeAlignment();
|
|
UseZeroLengthBitfieldAlignment = HostTarget->useZeroLengthBitfieldAlignment();
|
|
UseExplicitBitFieldAlignment = HostTarget->useExplicitBitFieldAlignment();
|
|
ZeroLengthBitfieldBoundary = HostTarget->getZeroLengthBitfieldBoundary();
|
|
|
|
// This is a bit of a lie, but it controls __GCC_ATOMIC_XXX_LOCK_FREE, and
|
|
// we need those macros to be identical on host and device, because (among
|
|
// other things) they affect which standard library classes are defined, and
|
|
// we need all classes to be defined on both the host and device.
|
|
MaxAtomicInlineWidth = HostTarget->getMaxAtomicInlineWidth();
|
|
|
|
// Properties intentionally not copied from host:
|
|
// - LargeArrayMinWidth, LargeArrayAlign: Not visible across the
|
|
// host/device boundary.
|
|
// - SuitableAlign: Not visible across the host/device boundary, and may
|
|
// correctly be different on host/device, e.g. if host has wider vector
|
|
// types than device.
|
|
// - LongDoubleWidth, LongDoubleAlign: nvptx's long double type is the same
|
|
// as its double type, but that's not necessarily true on the host.
|
|
// TODO: nvcc emits a warning when using long double on device; we should
|
|
// do the same.
|
|
}
|
|
|
|
ArrayRef<const char *> NVPTXTargetInfo::getGCCRegNames() const {
|
|
return llvm::ArrayRef(GCCRegNames);
|
|
}
|
|
|
|
bool NVPTXTargetInfo::hasFeature(StringRef Feature) const {
|
|
return llvm::StringSwitch<bool>(Feature)
|
|
.Cases("ptx", "nvptx", true)
|
|
.Default(false);
|
|
}
|
|
|
|
void NVPTXTargetInfo::getTargetDefines(const LangOptions &Opts,
|
|
MacroBuilder &Builder) const {
|
|
Builder.defineMacro("__PTX__");
|
|
Builder.defineMacro("__NVPTX__");
|
|
|
|
// Skip setting architecture dependent macros if undefined.
|
|
if (GPU == OffloadArch::UNUSED && !HostTarget)
|
|
return;
|
|
|
|
if (Opts.CUDAIsDevice || Opts.OpenMPIsTargetDevice || !HostTarget) {
|
|
// Set __CUDA_ARCH__ for the GPU specified.
|
|
std::string CUDAArchCode = [this] {
|
|
switch (GPU) {
|
|
case OffloadArch::GFX600:
|
|
case OffloadArch::GFX601:
|
|
case OffloadArch::GFX602:
|
|
case OffloadArch::GFX700:
|
|
case OffloadArch::GFX701:
|
|
case OffloadArch::GFX702:
|
|
case OffloadArch::GFX703:
|
|
case OffloadArch::GFX704:
|
|
case OffloadArch::GFX705:
|
|
case OffloadArch::GFX801:
|
|
case OffloadArch::GFX802:
|
|
case OffloadArch::GFX803:
|
|
case OffloadArch::GFX805:
|
|
case OffloadArch::GFX810:
|
|
case OffloadArch::GFX9_GENERIC:
|
|
case OffloadArch::GFX900:
|
|
case OffloadArch::GFX902:
|
|
case OffloadArch::GFX904:
|
|
case OffloadArch::GFX906:
|
|
case OffloadArch::GFX908:
|
|
case OffloadArch::GFX909:
|
|
case OffloadArch::GFX90a:
|
|
case OffloadArch::GFX90c:
|
|
case OffloadArch::GFX9_4_GENERIC:
|
|
case OffloadArch::GFX940:
|
|
case OffloadArch::GFX941:
|
|
case OffloadArch::GFX942:
|
|
case OffloadArch::GFX950:
|
|
case OffloadArch::GFX10_1_GENERIC:
|
|
case OffloadArch::GFX1010:
|
|
case OffloadArch::GFX1011:
|
|
case OffloadArch::GFX1012:
|
|
case OffloadArch::GFX1013:
|
|
case OffloadArch::GFX10_3_GENERIC:
|
|
case OffloadArch::GFX1030:
|
|
case OffloadArch::GFX1031:
|
|
case OffloadArch::GFX1032:
|
|
case OffloadArch::GFX1033:
|
|
case OffloadArch::GFX1034:
|
|
case OffloadArch::GFX1035:
|
|
case OffloadArch::GFX1036:
|
|
case OffloadArch::GFX11_GENERIC:
|
|
case OffloadArch::GFX1100:
|
|
case OffloadArch::GFX1101:
|
|
case OffloadArch::GFX1102:
|
|
case OffloadArch::GFX1103:
|
|
case OffloadArch::GFX1150:
|
|
case OffloadArch::GFX1151:
|
|
case OffloadArch::GFX1152:
|
|
case OffloadArch::GFX1153:
|
|
case OffloadArch::GFX12_GENERIC:
|
|
case OffloadArch::GFX1200:
|
|
case OffloadArch::GFX1201:
|
|
case OffloadArch::AMDGCNSPIRV:
|
|
case OffloadArch::Generic:
|
|
case OffloadArch::LAST:
|
|
break;
|
|
case OffloadArch::UNKNOWN:
|
|
assert(false && "No GPU arch when compiling CUDA device code.");
|
|
return "";
|
|
case OffloadArch::UNUSED:
|
|
case OffloadArch::SM_20:
|
|
return "200";
|
|
case OffloadArch::SM_21:
|
|
return "210";
|
|
case OffloadArch::SM_30:
|
|
return "300";
|
|
case OffloadArch::SM_32_:
|
|
return "320";
|
|
case OffloadArch::SM_35:
|
|
return "350";
|
|
case OffloadArch::SM_37:
|
|
return "370";
|
|
case OffloadArch::SM_50:
|
|
return "500";
|
|
case OffloadArch::SM_52:
|
|
return "520";
|
|
case OffloadArch::SM_53:
|
|
return "530";
|
|
case OffloadArch::SM_60:
|
|
return "600";
|
|
case OffloadArch::SM_61:
|
|
return "610";
|
|
case OffloadArch::SM_62:
|
|
return "620";
|
|
case OffloadArch::SM_70:
|
|
return "700";
|
|
case OffloadArch::SM_72:
|
|
return "720";
|
|
case OffloadArch::SM_75:
|
|
return "750";
|
|
case OffloadArch::SM_80:
|
|
return "800";
|
|
case OffloadArch::SM_86:
|
|
return "860";
|
|
case OffloadArch::SM_87:
|
|
return "870";
|
|
case OffloadArch::SM_89:
|
|
return "890";
|
|
case OffloadArch::SM_90:
|
|
case OffloadArch::SM_90a:
|
|
return "900";
|
|
case OffloadArch::SM_100:
|
|
return "1000";
|
|
}
|
|
llvm_unreachable("unhandled OffloadArch");
|
|
}();
|
|
Builder.defineMacro("__CUDA_ARCH__", CUDAArchCode);
|
|
if (GPU == OffloadArch::SM_90a)
|
|
Builder.defineMacro("__CUDA_ARCH_FEAT_SM90_ALL", "1");
|
|
}
|
|
}
|
|
|
|
ArrayRef<Builtin::Info> NVPTXTargetInfo::getTargetBuiltins() const {
|
|
return llvm::ArrayRef(BuiltinInfo,
|
|
clang::NVPTX::LastTSBuiltin - Builtin::FirstTSBuiltin);
|
|
}
|