1 | //===----- CGCUDANV.cpp - Interface to NVIDIA CUDA Runtime ----------------===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | // |
9 | // This provides a class for CUDA code generation targeting the NVIDIA CUDA |
10 | // runtime library. |
11 | // |
12 | //===----------------------------------------------------------------------===// |
13 | |
14 | #include "CGCUDARuntime.h" |
15 | #include "CGCXXABI.h" |
16 | #include "CodeGenFunction.h" |
17 | #include "CodeGenModule.h" |
18 | #include "clang/AST/CharUnits.h" |
19 | #include "clang/AST/Decl.h" |
20 | #include "clang/Basic/Cuda.h" |
21 | #include "clang/CodeGen/CodeGenABITypes.h" |
22 | #include "clang/CodeGen/ConstantInitBuilder.h" |
23 | #include "llvm/ADT/StringRef.h" |
24 | #include "llvm/Frontend/Offloading/Utility.h" |
25 | #include "llvm/IR/BasicBlock.h" |
26 | #include "llvm/IR/Constants.h" |
27 | #include "llvm/IR/DerivedTypes.h" |
28 | #include "llvm/IR/ReplaceConstant.h" |
29 | #include "llvm/Support/Format.h" |
30 | #include "llvm/Support/VirtualFileSystem.h" |
31 | |
32 | using namespace clang; |
33 | using namespace CodeGen; |
34 | |
35 | namespace { |
36 | constexpr unsigned CudaFatMagic = 0x466243b1; |
37 | constexpr unsigned HIPFatMagic = 0x48495046; // "HIPF" |
38 | |
39 | class CGNVCUDARuntime : public CGCUDARuntime { |
40 | |
41 | /// The prefix used for function calls and section names (CUDA, HIP, LLVM) |
42 | StringRef Prefix; |
43 | |
44 | private: |
45 | llvm::IntegerType *IntTy, *SizeTy; |
46 | llvm::Type *VoidTy; |
47 | llvm::PointerType *PtrTy; |
48 | |
49 | /// Convenience reference to LLVM Context |
50 | llvm::LLVMContext &Context; |
51 | /// Convenience reference to the current module |
52 | llvm::Module &TheModule; |
53 | /// Keeps track of kernel launch stubs and handles emitted in this module |
54 | struct KernelInfo { |
55 | llvm::Function *Kernel; // stub function to help launch kernel |
56 | const Decl *D; |
57 | }; |
58 | llvm::SmallVector<KernelInfo, 16> EmittedKernels; |
59 | // Map a kernel mangled name to a symbol for identifying kernel in host code |
60 | // For CUDA, the symbol for identifying the kernel is the same as the device |
61 | // stub function. For HIP, they are different. |
62 | llvm::DenseMap<StringRef, llvm::GlobalValue *> KernelHandles; |
63 | // Map a kernel handle to the kernel stub. |
64 | llvm::DenseMap<llvm::GlobalValue *, llvm::Function *> KernelStubs; |
65 | struct VarInfo { |
66 | llvm::GlobalVariable *Var; |
67 | const VarDecl *D; |
68 | DeviceVarFlags Flags; |
69 | }; |
70 | llvm::SmallVector<VarInfo, 16> DeviceVars; |
71 | /// Keeps track of variable containing handle of GPU binary. Populated by |
72 | /// ModuleCtorFunction() and used to create corresponding cleanup calls in |
73 | /// ModuleDtorFunction() |
74 | llvm::GlobalVariable *GpuBinaryHandle = nullptr; |
75 | /// Whether we generate relocatable device code. |
76 | bool RelocatableDeviceCode; |
77 | /// Mangle context for device. |
78 | std::unique_ptr<MangleContext> DeviceMC; |
79 | |
80 | llvm::FunctionCallee getSetupArgumentFn() const; |
81 | llvm::FunctionCallee getLaunchFn() const; |
82 | |
83 | llvm::FunctionType *getRegisterGlobalsFnTy() const; |
84 | llvm::FunctionType *getCallbackFnTy() const; |
85 | llvm::FunctionType *getRegisterLinkedBinaryFnTy() const; |
86 | std::string addPrefixToName(StringRef FuncName) const; |
87 | std::string addUnderscoredPrefixToName(StringRef FuncName) const; |
88 | |
89 | /// Creates a function to register all kernel stubs generated in this module. |
90 | llvm::Function *makeRegisterGlobalsFn(); |
91 | |
92 | /// Helper function that generates a constant string and returns a pointer to |
93 | /// the start of the string. The result of this function can be used anywhere |
94 | /// where the C code specifies const char*. |
95 | llvm::Constant *makeConstantString(const std::string &Str, |
96 | const std::string &Name = "" ) { |
97 | return CGM.GetAddrOfConstantCString(Str, GlobalName: Name.c_str()).getPointer(); |
98 | } |
99 | |
100 | /// Helper function which generates an initialized constant array from Str, |
101 | /// and optionally sets section name and alignment. AddNull specifies whether |
102 | /// the array should nave NUL termination. |
103 | llvm::Constant *makeConstantArray(StringRef Str, |
104 | StringRef Name = "" , |
105 | StringRef SectionName = "" , |
106 | unsigned Alignment = 0, |
107 | bool AddNull = false) { |
108 | llvm::Constant *Value = |
109 | llvm::ConstantDataArray::getString(Context, Initializer: Str, AddNull); |
110 | auto *GV = new llvm::GlobalVariable( |
111 | TheModule, Value->getType(), /*isConstant=*/true, |
112 | llvm::GlobalValue::PrivateLinkage, Value, Name); |
113 | if (!SectionName.empty()) { |
114 | GV->setSection(SectionName); |
115 | // Mark the address as used which make sure that this section isn't |
116 | // merged and we will really have it in the object file. |
117 | GV->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::None); |
118 | } |
119 | if (Alignment) |
120 | GV->setAlignment(llvm::Align(Alignment)); |
121 | return GV; |
122 | } |
123 | |
124 | /// Helper function that generates an empty dummy function returning void. |
125 | llvm::Function *makeDummyFunction(llvm::FunctionType *FnTy) { |
126 | assert(FnTy->getReturnType()->isVoidTy() && |
127 | "Can only generate dummy functions returning void!" ); |
128 | llvm::Function *DummyFunc = llvm::Function::Create( |
129 | Ty: FnTy, Linkage: llvm::GlobalValue::InternalLinkage, N: "dummy" , M: &TheModule); |
130 | |
131 | llvm::BasicBlock *DummyBlock = |
132 | llvm::BasicBlock::Create(Context, Name: "" , Parent: DummyFunc); |
133 | CGBuilderTy FuncBuilder(CGM, Context); |
134 | FuncBuilder.SetInsertPoint(DummyBlock); |
135 | FuncBuilder.CreateRetVoid(); |
136 | |
137 | return DummyFunc; |
138 | } |
139 | |
140 | Address prepareKernelArgs(CodeGenFunction &CGF, FunctionArgList &Args); |
141 | Address prepareKernelArgsLLVMOffload(CodeGenFunction &CGF, |
142 | FunctionArgList &Args); |
143 | void emitDeviceStubBodyLegacy(CodeGenFunction &CGF, FunctionArgList &Args); |
144 | void emitDeviceStubBodyNew(CodeGenFunction &CGF, FunctionArgList &Args); |
145 | std::string getDeviceSideName(const NamedDecl *ND) override; |
146 | |
147 | void registerDeviceVar(const VarDecl *VD, llvm::GlobalVariable &Var, |
148 | bool Extern, bool Constant) { |
149 | DeviceVars.push_back(Elt: {.Var: &Var, |
150 | .D: VD, |
151 | .Flags: {DeviceVarFlags::Variable, Extern, Constant, |
152 | VD->hasAttr<HIPManagedAttr>(), |
153 | /*Normalized*/ false, 0}}); |
154 | } |
155 | void registerDeviceSurf(const VarDecl *VD, llvm::GlobalVariable &Var, |
156 | bool Extern, int Type) { |
157 | DeviceVars.push_back(Elt: {.Var: &Var, |
158 | .D: VD, |
159 | .Flags: {DeviceVarFlags::Surface, Extern, /*Constant*/ false, |
160 | /*Managed*/ false, |
161 | /*Normalized*/ false, Type}}); |
162 | } |
163 | void registerDeviceTex(const VarDecl *VD, llvm::GlobalVariable &Var, |
164 | bool Extern, int Type, bool Normalized) { |
165 | DeviceVars.push_back(Elt: {.Var: &Var, |
166 | .D: VD, |
167 | .Flags: {DeviceVarFlags::Texture, Extern, /*Constant*/ false, |
168 | /*Managed*/ false, Normalized, Type}}); |
169 | } |
170 | |
171 | /// Creates module constructor function |
172 | llvm::Function *makeModuleCtorFunction(); |
173 | /// Creates module destructor function |
174 | llvm::Function *makeModuleDtorFunction(); |
175 | /// Transform managed variables for device compilation. |
176 | void transformManagedVars(); |
177 | /// Create offloading entries to register globals in RDC mode. |
178 | void createOffloadingEntries(); |
179 | |
180 | public: |
181 | CGNVCUDARuntime(CodeGenModule &CGM); |
182 | |
183 | llvm::GlobalValue *getKernelHandle(llvm::Function *F, GlobalDecl GD) override; |
184 | llvm::Function *getKernelStub(llvm::GlobalValue *Handle) override { |
185 | auto Loc = KernelStubs.find(Val: Handle); |
186 | assert(Loc != KernelStubs.end()); |
187 | return Loc->second; |
188 | } |
189 | void emitDeviceStub(CodeGenFunction &CGF, FunctionArgList &Args) override; |
190 | void handleVarRegistration(const VarDecl *VD, |
191 | llvm::GlobalVariable &Var) override; |
192 | void |
193 | internalizeDeviceSideVar(const VarDecl *D, |
194 | llvm::GlobalValue::LinkageTypes &Linkage) override; |
195 | |
196 | llvm::Function *finalizeModule() override; |
197 | }; |
198 | |
199 | } // end anonymous namespace |
200 | |
201 | std::string CGNVCUDARuntime::addPrefixToName(StringRef FuncName) const { |
202 | return (Prefix + FuncName).str(); |
203 | } |
204 | std::string |
205 | CGNVCUDARuntime::addUnderscoredPrefixToName(StringRef FuncName) const { |
206 | return ("__" + Prefix + FuncName).str(); |
207 | } |
208 | |
209 | static std::unique_ptr<MangleContext> InitDeviceMC(CodeGenModule &CGM) { |
210 | // If the host and device have different C++ ABIs, mark it as the device |
211 | // mangle context so that the mangling needs to retrieve the additional |
212 | // device lambda mangling number instead of the regular host one. |
213 | if (CGM.getContext().getAuxTargetInfo() && |
214 | CGM.getContext().getTargetInfo().getCXXABI().isMicrosoft() && |
215 | CGM.getContext().getAuxTargetInfo()->getCXXABI().isItaniumFamily()) { |
216 | return std::unique_ptr<MangleContext>( |
217 | CGM.getContext().createDeviceMangleContext( |
218 | T: *CGM.getContext().getAuxTargetInfo())); |
219 | } |
220 | |
221 | return std::unique_ptr<MangleContext>(CGM.getContext().createMangleContext( |
222 | T: CGM.getContext().getAuxTargetInfo())); |
223 | } |
224 | |
225 | CGNVCUDARuntime::CGNVCUDARuntime(CodeGenModule &CGM) |
226 | : CGCUDARuntime(CGM), Context(CGM.getLLVMContext()), |
227 | TheModule(CGM.getModule()), |
228 | RelocatableDeviceCode(CGM.getLangOpts().GPURelocatableDeviceCode), |
229 | DeviceMC(InitDeviceMC(CGM)) { |
230 | IntTy = CGM.IntTy; |
231 | SizeTy = CGM.SizeTy; |
232 | VoidTy = CGM.VoidTy; |
233 | PtrTy = CGM.UnqualPtrTy; |
234 | |
235 | if (CGM.getLangOpts().OffloadViaLLVM) |
236 | Prefix = "llvm" ; |
237 | else if (CGM.getLangOpts().HIP) |
238 | Prefix = "hip" ; |
239 | else |
240 | Prefix = "cuda" ; |
241 | } |
242 | |
243 | llvm::FunctionCallee CGNVCUDARuntime::getSetupArgumentFn() const { |
244 | // cudaError_t cudaSetupArgument(void *, size_t, size_t) |
245 | llvm::Type *Params[] = {PtrTy, SizeTy, SizeTy}; |
246 | return CGM.CreateRuntimeFunction( |
247 | Ty: llvm::FunctionType::get(Result: IntTy, Params, isVarArg: false), |
248 | Name: addPrefixToName(FuncName: "SetupArgument" )); |
249 | } |
250 | |
251 | llvm::FunctionCallee CGNVCUDARuntime::getLaunchFn() const { |
252 | if (CGM.getLangOpts().HIP) { |
253 | // hipError_t hipLaunchByPtr(char *); |
254 | return CGM.CreateRuntimeFunction( |
255 | Ty: llvm::FunctionType::get(Result: IntTy, Params: PtrTy, isVarArg: false), Name: "hipLaunchByPtr" ); |
256 | } |
257 | // cudaError_t cudaLaunch(char *); |
258 | return CGM.CreateRuntimeFunction(Ty: llvm::FunctionType::get(Result: IntTy, Params: PtrTy, isVarArg: false), |
259 | Name: "cudaLaunch" ); |
260 | } |
261 | |
262 | llvm::FunctionType *CGNVCUDARuntime::getRegisterGlobalsFnTy() const { |
263 | return llvm::FunctionType::get(Result: VoidTy, Params: PtrTy, isVarArg: false); |
264 | } |
265 | |
266 | llvm::FunctionType *CGNVCUDARuntime::getCallbackFnTy() const { |
267 | return llvm::FunctionType::get(Result: VoidTy, Params: PtrTy, isVarArg: false); |
268 | } |
269 | |
270 | llvm::FunctionType *CGNVCUDARuntime::getRegisterLinkedBinaryFnTy() const { |
271 | llvm::Type *Params[] = {llvm::PointerType::getUnqual(C&: Context), PtrTy, PtrTy, |
272 | llvm::PointerType::getUnqual(C&: Context)}; |
273 | return llvm::FunctionType::get(Result: VoidTy, Params, isVarArg: false); |
274 | } |
275 | |
276 | std::string CGNVCUDARuntime::getDeviceSideName(const NamedDecl *ND) { |
277 | GlobalDecl GD; |
278 | // D could be either a kernel or a variable. |
279 | if (auto *FD = dyn_cast<FunctionDecl>(Val: ND)) |
280 | GD = GlobalDecl(FD, KernelReferenceKind::Kernel); |
281 | else |
282 | GD = GlobalDecl(ND); |
283 | std::string DeviceSideName; |
284 | MangleContext *MC; |
285 | if (CGM.getLangOpts().CUDAIsDevice) |
286 | MC = &CGM.getCXXABI().getMangleContext(); |
287 | else |
288 | MC = DeviceMC.get(); |
289 | if (MC->shouldMangleDeclName(D: ND)) { |
290 | SmallString<256> Buffer; |
291 | llvm::raw_svector_ostream Out(Buffer); |
292 | MC->mangleName(GD, Out); |
293 | DeviceSideName = std::string(Out.str()); |
294 | } else |
295 | DeviceSideName = std::string(ND->getIdentifier()->getName()); |
296 | |
297 | // Make unique name for device side static file-scope variable for HIP. |
298 | if (CGM.getContext().shouldExternalize(D: ND) && |
299 | CGM.getLangOpts().GPURelocatableDeviceCode) { |
300 | SmallString<256> Buffer; |
301 | llvm::raw_svector_ostream Out(Buffer); |
302 | Out << DeviceSideName; |
303 | CGM.printPostfixForExternalizedDecl(OS&: Out, D: ND); |
304 | DeviceSideName = std::string(Out.str()); |
305 | } |
306 | return DeviceSideName; |
307 | } |
308 | |
309 | void CGNVCUDARuntime::emitDeviceStub(CodeGenFunction &CGF, |
310 | FunctionArgList &Args) { |
311 | EmittedKernels.push_back(Elt: {.Kernel: CGF.CurFn, .D: CGF.CurFuncDecl}); |
312 | if (auto *GV = |
313 | dyn_cast<llvm::GlobalVariable>(Val: KernelHandles[CGF.CurFn->getName()])) { |
314 | GV->setLinkage(CGF.CurFn->getLinkage()); |
315 | GV->setInitializer(CGF.CurFn); |
316 | } |
317 | if (CudaFeatureEnabled(CGM.getTarget().getSDKVersion(), |
318 | CudaFeature::CUDA_USES_NEW_LAUNCH) || |
319 | (CGF.getLangOpts().HIP && CGF.getLangOpts().HIPUseNewLaunchAPI) || |
320 | (CGF.getLangOpts().OffloadViaLLVM)) |
321 | emitDeviceStubBodyNew(CGF, Args); |
322 | else |
323 | emitDeviceStubBodyLegacy(CGF, Args); |
324 | } |
325 | |
326 | /// CUDA passes the arguments with a level of indirection. For example, a |
327 | /// (void*, short, void*) is passed as {void **, short *, void **} to the launch |
328 | /// function. For the LLVM/offload launch we flatten the arguments into the |
329 | /// struct directly. In addition, we include the size of the arguments, thus |
330 | /// pass {sizeof({void *, short, void *}), ptr to {void *, short, void *}, |
331 | /// nullptr}. The last nullptr needs to be initialized to an array of pointers |
332 | /// pointing to the arguments if we want to offload to the host. |
333 | Address CGNVCUDARuntime::prepareKernelArgsLLVMOffload(CodeGenFunction &CGF, |
334 | FunctionArgList &Args) { |
335 | SmallVector<llvm::Type *> ArgTypes, KernelLaunchParamsTypes; |
336 | for (auto &Arg : Args) |
337 | ArgTypes.push_back(Elt: CGF.ConvertTypeForMem(T: Arg->getType())); |
338 | llvm::StructType *KernelArgsTy = llvm::StructType::create(Elements: ArgTypes); |
339 | |
340 | auto *Int64Ty = CGF.Builder.getInt64Ty(); |
341 | KernelLaunchParamsTypes.push_back(Elt: Int64Ty); |
342 | KernelLaunchParamsTypes.push_back(Elt: PtrTy); |
343 | KernelLaunchParamsTypes.push_back(Elt: PtrTy); |
344 | |
345 | llvm::StructType *KernelLaunchParamsTy = |
346 | llvm::StructType::create(Elements: KernelLaunchParamsTypes); |
347 | Address KernelArgs = CGF.CreateTempAllocaWithoutCast( |
348 | Ty: KernelArgsTy, align: CharUnits::fromQuantity(Quantity: 16), Name: "kernel_args" ); |
349 | Address KernelLaunchParams = CGF.CreateTempAllocaWithoutCast( |
350 | Ty: KernelLaunchParamsTy, align: CharUnits::fromQuantity(Quantity: 16), |
351 | Name: "kernel_launch_params" ); |
352 | |
353 | auto KernelArgsSize = CGM.getDataLayout().getTypeAllocSize(Ty: KernelArgsTy); |
354 | CGF.Builder.CreateStore(Val: llvm::ConstantInt::get(Ty: Int64Ty, V: KernelArgsSize), |
355 | Addr: CGF.Builder.CreateStructGEP(Addr: KernelLaunchParams, Index: 0)); |
356 | CGF.Builder.CreateStore(Val: KernelArgs.emitRawPointer(CGF), |
357 | Addr: CGF.Builder.CreateStructGEP(Addr: KernelLaunchParams, Index: 1)); |
358 | CGF.Builder.CreateStore(Val: llvm::Constant::getNullValue(Ty: PtrTy), |
359 | Addr: CGF.Builder.CreateStructGEP(Addr: KernelLaunchParams, Index: 2)); |
360 | |
361 | for (unsigned i = 0; i < Args.size(); ++i) { |
362 | auto *ArgVal = CGF.Builder.CreateLoad(Addr: CGF.GetAddrOfLocalVar(VD: Args[i])); |
363 | CGF.Builder.CreateStore(Val: ArgVal, Addr: CGF.Builder.CreateStructGEP(Addr: KernelArgs, Index: i)); |
364 | } |
365 | |
366 | return KernelLaunchParams; |
367 | } |
368 | |
369 | Address CGNVCUDARuntime::prepareKernelArgs(CodeGenFunction &CGF, |
370 | FunctionArgList &Args) { |
371 | // Calculate amount of space we will need for all arguments. If we have no |
372 | // args, allocate a single pointer so we still have a valid pointer to the |
373 | // argument array that we can pass to runtime, even if it will be unused. |
374 | Address KernelArgs = CGF.CreateTempAlloca( |
375 | Ty: PtrTy, align: CharUnits::fromQuantity(Quantity: 16), Name: "kernel_args" , |
376 | ArraySize: llvm::ConstantInt::get(Ty: SizeTy, V: std::max<size_t>(a: 1, b: Args.size()))); |
377 | // Store pointers to the arguments in a locally allocated launch_args. |
378 | for (unsigned i = 0; i < Args.size(); ++i) { |
379 | llvm::Value *VarPtr = CGF.GetAddrOfLocalVar(VD: Args[i]).emitRawPointer(CGF); |
380 | llvm::Value *VoidVarPtr = CGF.Builder.CreatePointerCast(V: VarPtr, DestTy: PtrTy); |
381 | CGF.Builder.CreateDefaultAlignedStore( |
382 | Val: VoidVarPtr, Addr: CGF.Builder.CreateConstGEP1_32( |
383 | Ty: PtrTy, Ptr: KernelArgs.emitRawPointer(CGF), Idx0: i)); |
384 | } |
385 | return KernelArgs; |
386 | } |
387 | |
388 | // CUDA 9.0+ uses new way to launch kernels. Parameters are packed in a local |
389 | // array and kernels are launched using cudaLaunchKernel(). |
390 | void CGNVCUDARuntime::emitDeviceStubBodyNew(CodeGenFunction &CGF, |
391 | FunctionArgList &Args) { |
392 | // Build the shadow stack entry at the very start of the function. |
393 | Address KernelArgs = CGF.getLangOpts().OffloadViaLLVM |
394 | ? prepareKernelArgsLLVMOffload(CGF, Args) |
395 | : prepareKernelArgs(CGF, Args); |
396 | |
397 | llvm::BasicBlock *EndBlock = CGF.createBasicBlock(name: "setup.end" ); |
398 | |
399 | // Lookup cudaLaunchKernel/hipLaunchKernel function. |
400 | // HIP kernel launching API name depends on -fgpu-default-stream option. For |
401 | // the default value 'legacy', it is hipLaunchKernel. For 'per-thread', |
402 | // it is hipLaunchKernel_spt. |
403 | // cudaError_t cudaLaunchKernel(const void *func, dim3 gridDim, dim3 blockDim, |
404 | // void **args, size_t sharedMem, |
405 | // cudaStream_t stream); |
406 | // hipError_t hipLaunchKernel[_spt](const void *func, dim3 gridDim, |
407 | // dim3 blockDim, void **args, |
408 | // size_t sharedMem, hipStream_t stream); |
409 | TranslationUnitDecl *TUDecl = CGM.getContext().getTranslationUnitDecl(); |
410 | DeclContext *DC = TranslationUnitDecl::castToDeclContext(D: TUDecl); |
411 | std::string KernelLaunchAPI = "LaunchKernel" ; |
412 | if (CGF.getLangOpts().GPUDefaultStream == |
413 | LangOptions::GPUDefaultStreamKind::PerThread) { |
414 | if (CGF.getLangOpts().HIP) |
415 | KernelLaunchAPI = KernelLaunchAPI + "_spt" ; |
416 | else if (CGF.getLangOpts().CUDA) |
417 | KernelLaunchAPI = KernelLaunchAPI + "_ptsz" ; |
418 | } |
419 | auto LaunchKernelName = addPrefixToName(FuncName: KernelLaunchAPI); |
420 | const IdentifierInfo &cudaLaunchKernelII = |
421 | CGM.getContext().Idents.get(Name: LaunchKernelName); |
422 | FunctionDecl *cudaLaunchKernelFD = nullptr; |
423 | for (auto *Result : DC->lookup(Name: &cudaLaunchKernelII)) { |
424 | if (FunctionDecl *FD = dyn_cast<FunctionDecl>(Val: Result)) |
425 | cudaLaunchKernelFD = FD; |
426 | } |
427 | |
428 | if (cudaLaunchKernelFD == nullptr) { |
429 | CGM.Error(loc: CGF.CurFuncDecl->getLocation(), |
430 | error: "Can't find declaration for " + LaunchKernelName); |
431 | return; |
432 | } |
433 | // Create temporary dim3 grid_dim, block_dim. |
434 | ParmVarDecl *GridDimParam = cudaLaunchKernelFD->getParamDecl(i: 1); |
435 | QualType Dim3Ty = GridDimParam->getType(); |
436 | Address GridDim = |
437 | CGF.CreateMemTemp(T: Dim3Ty, Align: CharUnits::fromQuantity(Quantity: 8), Name: "grid_dim" ); |
438 | Address BlockDim = |
439 | CGF.CreateMemTemp(T: Dim3Ty, Align: CharUnits::fromQuantity(Quantity: 8), Name: "block_dim" ); |
440 | Address ShmemSize = |
441 | CGF.CreateTempAlloca(Ty: SizeTy, align: CGM.getSizeAlign(), Name: "shmem_size" ); |
442 | Address Stream = CGF.CreateTempAlloca(Ty: PtrTy, align: CGM.getPointerAlign(), Name: "stream" ); |
443 | llvm::FunctionCallee cudaPopConfigFn = CGM.CreateRuntimeFunction( |
444 | Ty: llvm::FunctionType::get(Result: IntTy, |
445 | Params: {/*gridDim=*/GridDim.getType(), |
446 | /*blockDim=*/BlockDim.getType(), |
447 | /*ShmemSize=*/ShmemSize.getType(), |
448 | /*Stream=*/Stream.getType()}, |
449 | /*isVarArg=*/false), |
450 | Name: addUnderscoredPrefixToName(FuncName: "PopCallConfiguration" )); |
451 | |
452 | CGF.EmitRuntimeCallOrInvoke(callee: cudaPopConfigFn, args: {GridDim.emitRawPointer(CGF), |
453 | BlockDim.emitRawPointer(CGF), |
454 | ShmemSize.emitRawPointer(CGF), |
455 | Stream.emitRawPointer(CGF)}); |
456 | |
457 | // Emit the call to cudaLaunch |
458 | llvm::Value *Kernel = |
459 | CGF.Builder.CreatePointerCast(V: KernelHandles[CGF.CurFn->getName()], DestTy: PtrTy); |
460 | CallArgList LaunchKernelArgs; |
461 | LaunchKernelArgs.add(rvalue: RValue::get(V: Kernel), |
462 | type: cudaLaunchKernelFD->getParamDecl(i: 0)->getType()); |
463 | LaunchKernelArgs.add(rvalue: RValue::getAggregate(addr: GridDim), type: Dim3Ty); |
464 | LaunchKernelArgs.add(rvalue: RValue::getAggregate(addr: BlockDim), type: Dim3Ty); |
465 | LaunchKernelArgs.add(rvalue: RValue::get(Addr: KernelArgs, CGF), |
466 | type: cudaLaunchKernelFD->getParamDecl(i: 3)->getType()); |
467 | LaunchKernelArgs.add(rvalue: RValue::get(V: CGF.Builder.CreateLoad(Addr: ShmemSize)), |
468 | type: cudaLaunchKernelFD->getParamDecl(i: 4)->getType()); |
469 | LaunchKernelArgs.add(rvalue: RValue::get(V: CGF.Builder.CreateLoad(Addr: Stream)), |
470 | type: cudaLaunchKernelFD->getParamDecl(i: 5)->getType()); |
471 | |
472 | QualType QT = cudaLaunchKernelFD->getType(); |
473 | QualType CQT = QT.getCanonicalType(); |
474 | llvm::Type *Ty = CGM.getTypes().ConvertType(T: CQT); |
475 | llvm::FunctionType *FTy = cast<llvm::FunctionType>(Val: Ty); |
476 | |
477 | const CGFunctionInfo &FI = |
478 | CGM.getTypes().arrangeFunctionDeclaration(GD: cudaLaunchKernelFD); |
479 | llvm::FunctionCallee cudaLaunchKernelFn = |
480 | CGM.CreateRuntimeFunction(Ty: FTy, Name: LaunchKernelName); |
481 | CGF.EmitCall(CallInfo: FI, Callee: CGCallee::forDirect(functionPtr: cudaLaunchKernelFn), ReturnValue: ReturnValueSlot(), |
482 | Args: LaunchKernelArgs); |
483 | |
484 | // To prevent CUDA device stub functions from being merged by ICF in MSVC |
485 | // environment, create an unique global variable for each kernel and write to |
486 | // the variable in the device stub. |
487 | if (CGM.getContext().getTargetInfo().getCXXABI().isMicrosoft() && |
488 | !CGF.getLangOpts().HIP) { |
489 | llvm::Function *KernelFunction = llvm::cast<llvm::Function>(Val: Kernel); |
490 | std::string GlobalVarName = (KernelFunction->getName() + ".id" ).str(); |
491 | |
492 | llvm::GlobalVariable *HandleVar = |
493 | CGM.getModule().getNamedGlobal(Name: GlobalVarName); |
494 | if (!HandleVar) { |
495 | HandleVar = new llvm::GlobalVariable( |
496 | CGM.getModule(), CGM.Int8Ty, |
497 | /*Constant=*/false, KernelFunction->getLinkage(), |
498 | llvm::ConstantInt::get(Ty: CGM.Int8Ty, V: 0), GlobalVarName); |
499 | HandleVar->setDSOLocal(KernelFunction->isDSOLocal()); |
500 | HandleVar->setVisibility(KernelFunction->getVisibility()); |
501 | if (KernelFunction->hasComdat()) |
502 | HandleVar->setComdat(CGM.getModule().getOrInsertComdat(Name: GlobalVarName)); |
503 | } |
504 | |
505 | CGF.Builder.CreateAlignedStore(Val: llvm::ConstantInt::get(Ty: CGM.Int8Ty, V: 1), |
506 | Addr: HandleVar, Align: CharUnits::One(), |
507 | /*IsVolatile=*/true); |
508 | } |
509 | |
510 | CGF.EmitBranch(Block: EndBlock); |
511 | |
512 | CGF.EmitBlock(BB: EndBlock); |
513 | } |
514 | |
515 | void CGNVCUDARuntime::emitDeviceStubBodyLegacy(CodeGenFunction &CGF, |
516 | FunctionArgList &Args) { |
517 | // Emit a call to cudaSetupArgument for each arg in Args. |
518 | llvm::FunctionCallee cudaSetupArgFn = getSetupArgumentFn(); |
519 | llvm::BasicBlock *EndBlock = CGF.createBasicBlock(name: "setup.end" ); |
520 | CharUnits Offset = CharUnits::Zero(); |
521 | for (const VarDecl *A : Args) { |
522 | auto TInfo = CGM.getContext().getTypeInfoInChars(T: A->getType()); |
523 | Offset = Offset.alignTo(Align: TInfo.Align); |
524 | llvm::Value *Args[] = { |
525 | CGF.Builder.CreatePointerCast( |
526 | V: CGF.GetAddrOfLocalVar(VD: A).emitRawPointer(CGF), DestTy: PtrTy), |
527 | llvm::ConstantInt::get(Ty: SizeTy, V: TInfo.Width.getQuantity()), |
528 | llvm::ConstantInt::get(Ty: SizeTy, V: Offset.getQuantity()), |
529 | }; |
530 | llvm::CallBase *CB = CGF.EmitRuntimeCallOrInvoke(callee: cudaSetupArgFn, args: Args); |
531 | llvm::Constant *Zero = llvm::ConstantInt::get(Ty: IntTy, V: 0); |
532 | llvm::Value *CBZero = CGF.Builder.CreateICmpEQ(LHS: CB, RHS: Zero); |
533 | llvm::BasicBlock *NextBlock = CGF.createBasicBlock(name: "setup.next" ); |
534 | CGF.Builder.CreateCondBr(Cond: CBZero, True: NextBlock, False: EndBlock); |
535 | CGF.EmitBlock(BB: NextBlock); |
536 | Offset += TInfo.Width; |
537 | } |
538 | |
539 | // Emit the call to cudaLaunch |
540 | llvm::FunctionCallee cudaLaunchFn = getLaunchFn(); |
541 | llvm::Value *Arg = |
542 | CGF.Builder.CreatePointerCast(V: KernelHandles[CGF.CurFn->getName()], DestTy: PtrTy); |
543 | CGF.EmitRuntimeCallOrInvoke(callee: cudaLaunchFn, args: Arg); |
544 | CGF.EmitBranch(Block: EndBlock); |
545 | |
546 | CGF.EmitBlock(BB: EndBlock); |
547 | } |
548 | |
549 | // Replace the original variable Var with the address loaded from variable |
550 | // ManagedVar populated by HIP runtime. |
551 | static void replaceManagedVar(llvm::GlobalVariable *Var, |
552 | llvm::GlobalVariable *ManagedVar) { |
553 | SmallVector<SmallVector<llvm::User *, 8>, 8> WorkList; |
554 | for (auto &&VarUse : Var->uses()) { |
555 | WorkList.push_back(Elt: {VarUse.getUser()}); |
556 | } |
557 | while (!WorkList.empty()) { |
558 | auto &&WorkItem = WorkList.pop_back_val(); |
559 | auto *U = WorkItem.back(); |
560 | if (isa<llvm::ConstantExpr>(Val: U)) { |
561 | for (auto &&UU : U->uses()) { |
562 | WorkItem.push_back(Elt: UU.getUser()); |
563 | WorkList.push_back(Elt: WorkItem); |
564 | WorkItem.pop_back(); |
565 | } |
566 | continue; |
567 | } |
568 | if (auto *I = dyn_cast<llvm::Instruction>(Val: U)) { |
569 | llvm::Value *OldV = Var; |
570 | llvm::Instruction *NewV = new llvm::LoadInst( |
571 | Var->getType(), ManagedVar, "ld.managed" , false, |
572 | llvm::Align(Var->getAlignment()), I->getIterator()); |
573 | WorkItem.pop_back(); |
574 | // Replace constant expressions directly or indirectly using the managed |
575 | // variable with instructions. |
576 | for (auto &&Op : WorkItem) { |
577 | auto *CE = cast<llvm::ConstantExpr>(Val: Op); |
578 | auto *NewInst = CE->getAsInstruction(); |
579 | NewInst->insertBefore(BB&: *I->getParent(), InsertPos: I->getIterator()); |
580 | NewInst->replaceUsesOfWith(From: OldV, To: NewV); |
581 | OldV = CE; |
582 | NewV = NewInst; |
583 | } |
584 | I->replaceUsesOfWith(From: OldV, To: NewV); |
585 | } else { |
586 | llvm_unreachable("Invalid use of managed variable" ); |
587 | } |
588 | } |
589 | } |
590 | |
591 | /// Creates a function that sets up state on the host side for CUDA objects that |
592 | /// have a presence on both the host and device sides. Specifically, registers |
593 | /// the host side of kernel functions and device global variables with the CUDA |
594 | /// runtime. |
595 | /// \code |
596 | /// void __cuda_register_globals(void** GpuBinaryHandle) { |
597 | /// __cudaRegisterFunction(GpuBinaryHandle,Kernel0,...); |
598 | /// ... |
599 | /// __cudaRegisterFunction(GpuBinaryHandle,KernelM,...); |
600 | /// __cudaRegisterVar(GpuBinaryHandle, GlobalVar0, ...); |
601 | /// ... |
602 | /// __cudaRegisterVar(GpuBinaryHandle, GlobalVarN, ...); |
603 | /// } |
604 | /// \endcode |
605 | llvm::Function *CGNVCUDARuntime::makeRegisterGlobalsFn() { |
606 | // No need to register anything |
607 | if (EmittedKernels.empty() && DeviceVars.empty()) |
608 | return nullptr; |
609 | |
610 | llvm::Function *RegisterKernelsFunc = llvm::Function::Create( |
611 | Ty: getRegisterGlobalsFnTy(), Linkage: llvm::GlobalValue::InternalLinkage, |
612 | N: addUnderscoredPrefixToName(FuncName: "_register_globals" ), M: &TheModule); |
613 | llvm::BasicBlock *EntryBB = |
614 | llvm::BasicBlock::Create(Context, Name: "entry" , Parent: RegisterKernelsFunc); |
615 | CGBuilderTy Builder(CGM, Context); |
616 | Builder.SetInsertPoint(EntryBB); |
617 | |
618 | // void __cudaRegisterFunction(void **, const char *, char *, const char *, |
619 | // int, uint3*, uint3*, dim3*, dim3*, int*) |
620 | llvm::Type *RegisterFuncParams[] = { |
621 | PtrTy, PtrTy, PtrTy, PtrTy, IntTy, |
622 | PtrTy, PtrTy, PtrTy, PtrTy, llvm::PointerType::getUnqual(C&: Context)}; |
623 | llvm::FunctionCallee RegisterFunc = CGM.CreateRuntimeFunction( |
624 | Ty: llvm::FunctionType::get(Result: IntTy, Params: RegisterFuncParams, isVarArg: false), |
625 | Name: addUnderscoredPrefixToName(FuncName: "RegisterFunction" )); |
626 | |
627 | // Extract GpuBinaryHandle passed as the first argument passed to |
628 | // __cuda_register_globals() and generate __cudaRegisterFunction() call for |
629 | // each emitted kernel. |
630 | llvm::Argument &GpuBinaryHandlePtr = *RegisterKernelsFunc->arg_begin(); |
631 | for (auto &&I : EmittedKernels) { |
632 | llvm::Constant *KernelName = |
633 | makeConstantString(Str: getDeviceSideName(ND: cast<NamedDecl>(Val: I.D))); |
634 | llvm::Constant *NullPtr = llvm::ConstantPointerNull::get(T: PtrTy); |
635 | llvm::Value *Args[] = { |
636 | &GpuBinaryHandlePtr, |
637 | KernelHandles[I.Kernel->getName()], |
638 | KernelName, |
639 | KernelName, |
640 | llvm::ConstantInt::get(Ty: IntTy, V: -1), |
641 | NullPtr, |
642 | NullPtr, |
643 | NullPtr, |
644 | NullPtr, |
645 | llvm::ConstantPointerNull::get(T: llvm::PointerType::getUnqual(C&: Context))}; |
646 | Builder.CreateCall(Callee: RegisterFunc, Args); |
647 | } |
648 | |
649 | llvm::Type *VarSizeTy = IntTy; |
650 | // For HIP or CUDA 9.0+, device variable size is type of `size_t`. |
651 | if (CGM.getLangOpts().HIP || |
652 | ToCudaVersion(CGM.getTarget().getSDKVersion()) >= CudaVersion::CUDA_90) |
653 | VarSizeTy = SizeTy; |
654 | |
655 | // void __cudaRegisterVar(void **, char *, char *, const char *, |
656 | // int, int, int, int) |
657 | llvm::Type *RegisterVarParams[] = {PtrTy, PtrTy, PtrTy, PtrTy, |
658 | IntTy, VarSizeTy, IntTy, IntTy}; |
659 | llvm::FunctionCallee RegisterVar = CGM.CreateRuntimeFunction( |
660 | Ty: llvm::FunctionType::get(Result: VoidTy, Params: RegisterVarParams, isVarArg: false), |
661 | Name: addUnderscoredPrefixToName(FuncName: "RegisterVar" )); |
662 | // void __hipRegisterManagedVar(void **, char *, char *, const char *, |
663 | // size_t, unsigned) |
664 | llvm::Type *RegisterManagedVarParams[] = {PtrTy, PtrTy, PtrTy, |
665 | PtrTy, VarSizeTy, IntTy}; |
666 | llvm::FunctionCallee RegisterManagedVar = CGM.CreateRuntimeFunction( |
667 | Ty: llvm::FunctionType::get(Result: VoidTy, Params: RegisterManagedVarParams, isVarArg: false), |
668 | Name: addUnderscoredPrefixToName(FuncName: "RegisterManagedVar" )); |
669 | // void __cudaRegisterSurface(void **, const struct surfaceReference *, |
670 | // const void **, const char *, int, int); |
671 | llvm::FunctionCallee RegisterSurf = CGM.CreateRuntimeFunction( |
672 | Ty: llvm::FunctionType::get( |
673 | Result: VoidTy, Params: {PtrTy, PtrTy, PtrTy, PtrTy, IntTy, IntTy}, isVarArg: false), |
674 | Name: addUnderscoredPrefixToName(FuncName: "RegisterSurface" )); |
675 | // void __cudaRegisterTexture(void **, const struct textureReference *, |
676 | // const void **, const char *, int, int, int) |
677 | llvm::FunctionCallee RegisterTex = CGM.CreateRuntimeFunction( |
678 | Ty: llvm::FunctionType::get( |
679 | Result: VoidTy, Params: {PtrTy, PtrTy, PtrTy, PtrTy, IntTy, IntTy, IntTy}, isVarArg: false), |
680 | Name: addUnderscoredPrefixToName(FuncName: "RegisterTexture" )); |
681 | for (auto &&Info : DeviceVars) { |
682 | llvm::GlobalVariable *Var = Info.Var; |
683 | assert((!Var->isDeclaration() || Info.Flags.isManaged()) && |
684 | "External variables should not show up here, except HIP managed " |
685 | "variables" ); |
686 | llvm::Constant *VarName = makeConstantString(Str: getDeviceSideName(ND: Info.D)); |
687 | switch (Info.Flags.getKind()) { |
688 | case DeviceVarFlags::Variable: { |
689 | uint64_t VarSize = |
690 | CGM.getDataLayout().getTypeAllocSize(Ty: Var->getValueType()); |
691 | if (Info.Flags.isManaged()) { |
692 | assert(Var->getName().ends_with(".managed" ) && |
693 | "HIP managed variables not transformed" ); |
694 | auto *ManagedVar = CGM.getModule().getNamedGlobal( |
695 | Name: Var->getName().drop_back(N: StringRef(".managed" ).size())); |
696 | llvm::Value *Args[] = { |
697 | &GpuBinaryHandlePtr, |
698 | ManagedVar, |
699 | Var, |
700 | VarName, |
701 | llvm::ConstantInt::get(Ty: VarSizeTy, V: VarSize), |
702 | llvm::ConstantInt::get(Ty: IntTy, V: Var->getAlignment())}; |
703 | if (!Var->isDeclaration()) |
704 | Builder.CreateCall(Callee: RegisterManagedVar, Args); |
705 | } else { |
706 | llvm::Value *Args[] = { |
707 | &GpuBinaryHandlePtr, |
708 | Var, |
709 | VarName, |
710 | VarName, |
711 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.isExtern()), |
712 | llvm::ConstantInt::get(Ty: VarSizeTy, V: VarSize), |
713 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.isConstant()), |
714 | llvm::ConstantInt::get(Ty: IntTy, V: 0)}; |
715 | Builder.CreateCall(Callee: RegisterVar, Args); |
716 | } |
717 | break; |
718 | } |
719 | case DeviceVarFlags::Surface: |
720 | Builder.CreateCall( |
721 | Callee: RegisterSurf, |
722 | Args: {&GpuBinaryHandlePtr, Var, VarName, VarName, |
723 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.getSurfTexType()), |
724 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.isExtern())}); |
725 | break; |
726 | case DeviceVarFlags::Texture: |
727 | Builder.CreateCall( |
728 | Callee: RegisterTex, |
729 | Args: {&GpuBinaryHandlePtr, Var, VarName, VarName, |
730 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.getSurfTexType()), |
731 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.isNormalized()), |
732 | llvm::ConstantInt::get(Ty: IntTy, V: Info.Flags.isExtern())}); |
733 | break; |
734 | } |
735 | } |
736 | |
737 | Builder.CreateRetVoid(); |
738 | return RegisterKernelsFunc; |
739 | } |
740 | |
741 | /// Creates a global constructor function for the module: |
742 | /// |
743 | /// For CUDA: |
744 | /// \code |
745 | /// void __cuda_module_ctor() { |
746 | /// Handle = __cudaRegisterFatBinary(GpuBinaryBlob); |
747 | /// __cuda_register_globals(Handle); |
748 | /// } |
749 | /// \endcode |
750 | /// |
751 | /// For HIP: |
752 | /// \code |
753 | /// void __hip_module_ctor() { |
754 | /// if (__hip_gpubin_handle == 0) { |
755 | /// __hip_gpubin_handle = __hipRegisterFatBinary(GpuBinaryBlob); |
756 | /// __hip_register_globals(__hip_gpubin_handle); |
757 | /// } |
758 | /// } |
759 | /// \endcode |
760 | llvm::Function *CGNVCUDARuntime::makeModuleCtorFunction() { |
761 | bool IsHIP = CGM.getLangOpts().HIP; |
762 | bool IsCUDA = CGM.getLangOpts().CUDA; |
763 | // No need to generate ctors/dtors if there is no GPU binary. |
764 | StringRef CudaGpuBinaryFileName = CGM.getCodeGenOpts().CudaGpuBinaryFileName; |
765 | if (CudaGpuBinaryFileName.empty() && !IsHIP) |
766 | return nullptr; |
767 | if ((IsHIP || (IsCUDA && !RelocatableDeviceCode)) && EmittedKernels.empty() && |
768 | DeviceVars.empty()) |
769 | return nullptr; |
770 | |
771 | // void __{cuda|hip}_register_globals(void* handle); |
772 | llvm::Function *RegisterGlobalsFunc = makeRegisterGlobalsFn(); |
773 | // We always need a function to pass in as callback. Create a dummy |
774 | // implementation if we don't need to register anything. |
775 | if (RelocatableDeviceCode && !RegisterGlobalsFunc) |
776 | RegisterGlobalsFunc = makeDummyFunction(FnTy: getRegisterGlobalsFnTy()); |
777 | |
778 | // void ** __{cuda|hip}RegisterFatBinary(void *); |
779 | llvm::FunctionCallee RegisterFatbinFunc = CGM.CreateRuntimeFunction( |
780 | Ty: llvm::FunctionType::get(Result: PtrTy, Params: PtrTy, isVarArg: false), |
781 | Name: addUnderscoredPrefixToName(FuncName: "RegisterFatBinary" )); |
782 | // struct { int magic, int version, void * gpu_binary, void * dont_care }; |
783 | llvm::StructType *FatbinWrapperTy = |
784 | llvm::StructType::get(elt1: IntTy, elts: IntTy, elts: PtrTy, elts: PtrTy); |
785 | |
786 | // Register GPU binary with the CUDA runtime, store returned handle in a |
787 | // global variable and save a reference in GpuBinaryHandle to be cleaned up |
788 | // in destructor on exit. Then associate all known kernels with the GPU binary |
789 | // handle so CUDA runtime can figure out what to call on the GPU side. |
790 | std::unique_ptr<llvm::MemoryBuffer> CudaGpuBinary = nullptr; |
791 | if (!CudaGpuBinaryFileName.empty()) { |
792 | auto VFS = CGM.getFileSystem(); |
793 | auto CudaGpuBinaryOrErr = |
794 | VFS->getBufferForFile(Name: CudaGpuBinaryFileName, FileSize: -1, RequiresNullTerminator: false); |
795 | if (std::error_code EC = CudaGpuBinaryOrErr.getError()) { |
796 | CGM.getDiags().Report(DiagID: diag::err_cannot_open_file) |
797 | << CudaGpuBinaryFileName << EC.message(); |
798 | return nullptr; |
799 | } |
800 | CudaGpuBinary = std::move(CudaGpuBinaryOrErr.get()); |
801 | } |
802 | |
803 | llvm::Function *ModuleCtorFunc = llvm::Function::Create( |
804 | Ty: llvm::FunctionType::get(Result: VoidTy, isVarArg: false), |
805 | Linkage: llvm::GlobalValue::InternalLinkage, |
806 | N: addUnderscoredPrefixToName(FuncName: "_module_ctor" ), M: &TheModule); |
807 | llvm::BasicBlock *CtorEntryBB = |
808 | llvm::BasicBlock::Create(Context, Name: "entry" , Parent: ModuleCtorFunc); |
809 | CGBuilderTy CtorBuilder(CGM, Context); |
810 | |
811 | CtorBuilder.SetInsertPoint(CtorEntryBB); |
812 | |
813 | const char *FatbinConstantName; |
814 | const char *FatbinSectionName; |
815 | const char *ModuleIDSectionName; |
816 | StringRef ModuleIDPrefix; |
817 | llvm::Constant *FatBinStr; |
818 | unsigned FatMagic; |
819 | if (IsHIP) { |
820 | FatbinConstantName = ".hip_fatbin" ; |
821 | FatbinSectionName = ".hipFatBinSegment" ; |
822 | |
823 | ModuleIDSectionName = "__hip_module_id" ; |
824 | ModuleIDPrefix = "__hip_" ; |
825 | |
826 | if (CudaGpuBinary) { |
827 | // If fatbin is available from early finalization, create a string |
828 | // literal containing the fat binary loaded from the given file. |
829 | const unsigned HIPCodeObjectAlign = 4096; |
830 | FatBinStr = makeConstantArray(Str: std::string(CudaGpuBinary->getBuffer()), Name: "" , |
831 | SectionName: FatbinConstantName, Alignment: HIPCodeObjectAlign); |
832 | } else { |
833 | // If fatbin is not available, create an external symbol |
834 | // __hip_fatbin in section .hip_fatbin. The external symbol is supposed |
835 | // to contain the fat binary but will be populated somewhere else, |
836 | // e.g. by lld through link script. |
837 | FatBinStr = new llvm::GlobalVariable( |
838 | CGM.getModule(), CGM.Int8Ty, |
839 | /*isConstant=*/true, llvm::GlobalValue::ExternalLinkage, nullptr, |
840 | "__hip_fatbin" + (CGM.getLangOpts().CUID.empty() |
841 | ? "" |
842 | : "_" + CGM.getContext().getCUIDHash()), |
843 | nullptr, llvm::GlobalVariable::NotThreadLocal); |
844 | cast<llvm::GlobalVariable>(Val: FatBinStr)->setSection(FatbinConstantName); |
845 | } |
846 | |
847 | FatMagic = HIPFatMagic; |
848 | } else { |
849 | if (RelocatableDeviceCode) |
850 | FatbinConstantName = CGM.getTriple().isMacOSX() |
851 | ? "__NV_CUDA,__nv_relfatbin" |
852 | : "__nv_relfatbin" ; |
853 | else |
854 | FatbinConstantName = |
855 | CGM.getTriple().isMacOSX() ? "__NV_CUDA,__nv_fatbin" : ".nv_fatbin" ; |
856 | // NVIDIA's cuobjdump looks for fatbins in this section. |
857 | FatbinSectionName = |
858 | CGM.getTriple().isMacOSX() ? "__NV_CUDA,__fatbin" : ".nvFatBinSegment" ; |
859 | |
860 | ModuleIDSectionName = CGM.getTriple().isMacOSX() |
861 | ? "__NV_CUDA,__nv_module_id" |
862 | : "__nv_module_id" ; |
863 | ModuleIDPrefix = "__nv_" ; |
864 | |
865 | // For CUDA, create a string literal containing the fat binary loaded from |
866 | // the given file. |
867 | FatBinStr = makeConstantArray(Str: std::string(CudaGpuBinary->getBuffer()), Name: "" , |
868 | SectionName: FatbinConstantName, Alignment: 8); |
869 | FatMagic = CudaFatMagic; |
870 | } |
871 | |
872 | // Create initialized wrapper structure that points to the loaded GPU binary |
873 | ConstantInitBuilder Builder(CGM); |
874 | auto Values = Builder.beginStruct(structTy: FatbinWrapperTy); |
875 | // Fatbin wrapper magic. |
876 | Values.addInt(intTy: IntTy, value: FatMagic); |
877 | // Fatbin version. |
878 | Values.addInt(intTy: IntTy, value: 1); |
879 | // Data. |
880 | Values.add(value: FatBinStr); |
881 | // Unused in fatbin v1. |
882 | Values.add(value: llvm::ConstantPointerNull::get(T: PtrTy)); |
883 | llvm::GlobalVariable *FatbinWrapper = Values.finishAndCreateGlobal( |
884 | args: addUnderscoredPrefixToName(FuncName: "_fatbin_wrapper" ), args: CGM.getPointerAlign(), |
885 | /*constant*/ args: true); |
886 | FatbinWrapper->setSection(FatbinSectionName); |
887 | |
888 | // There is only one HIP fat binary per linked module, however there are |
889 | // multiple constructor functions. Make sure the fat binary is registered |
890 | // only once. The constructor functions are executed by the dynamic loader |
891 | // before the program gains control. The dynamic loader cannot execute the |
892 | // constructor functions concurrently since doing that would not guarantee |
893 | // thread safety of the loaded program. Therefore we can assume sequential |
894 | // execution of constructor functions here. |
895 | if (IsHIP) { |
896 | auto Linkage = RelocatableDeviceCode ? llvm::GlobalValue::ExternalLinkage |
897 | : llvm::GlobalValue::InternalLinkage; |
898 | llvm::BasicBlock *IfBlock = |
899 | llvm::BasicBlock::Create(Context, Name: "if" , Parent: ModuleCtorFunc); |
900 | llvm::BasicBlock *ExitBlock = |
901 | llvm::BasicBlock::Create(Context, Name: "exit" , Parent: ModuleCtorFunc); |
902 | // The name, size, and initialization pattern of this variable is part |
903 | // of HIP ABI. |
904 | GpuBinaryHandle = new llvm::GlobalVariable( |
905 | TheModule, PtrTy, /*isConstant=*/false, Linkage, |
906 | /*Initializer=*/ |
907 | !RelocatableDeviceCode ? llvm::ConstantPointerNull::get(T: PtrTy) |
908 | : nullptr, |
909 | "__hip_gpubin_handle" + (CGM.getLangOpts().CUID.empty() |
910 | ? "" |
911 | : "_" + CGM.getContext().getCUIDHash())); |
912 | GpuBinaryHandle->setAlignment(CGM.getPointerAlign().getAsAlign()); |
913 | // Prevent the weak symbol in different shared libraries being merged. |
914 | if (Linkage != llvm::GlobalValue::InternalLinkage) |
915 | GpuBinaryHandle->setVisibility(llvm::GlobalValue::HiddenVisibility); |
916 | Address GpuBinaryAddr( |
917 | GpuBinaryHandle, PtrTy, |
918 | CharUnits::fromQuantity(Quantity: GpuBinaryHandle->getAlignment())); |
919 | { |
920 | auto *HandleValue = CtorBuilder.CreateLoad(Addr: GpuBinaryAddr); |
921 | llvm::Constant *Zero = |
922 | llvm::Constant::getNullValue(Ty: HandleValue->getType()); |
923 | llvm::Value *EQZero = CtorBuilder.CreateICmpEQ(LHS: HandleValue, RHS: Zero); |
924 | CtorBuilder.CreateCondBr(Cond: EQZero, True: IfBlock, False: ExitBlock); |
925 | } |
926 | { |
927 | CtorBuilder.SetInsertPoint(IfBlock); |
928 | // GpuBinaryHandle = __hipRegisterFatBinary(&FatbinWrapper); |
929 | llvm::CallInst *RegisterFatbinCall = |
930 | CtorBuilder.CreateCall(Callee: RegisterFatbinFunc, Args: FatbinWrapper); |
931 | CtorBuilder.CreateStore(Val: RegisterFatbinCall, Addr: GpuBinaryAddr); |
932 | CtorBuilder.CreateBr(Dest: ExitBlock); |
933 | } |
934 | { |
935 | CtorBuilder.SetInsertPoint(ExitBlock); |
936 | // Call __hip_register_globals(GpuBinaryHandle); |
937 | if (RegisterGlobalsFunc) { |
938 | auto *HandleValue = CtorBuilder.CreateLoad(Addr: GpuBinaryAddr); |
939 | CtorBuilder.CreateCall(Callee: RegisterGlobalsFunc, Args: HandleValue); |
940 | } |
941 | } |
942 | } else if (!RelocatableDeviceCode) { |
943 | // Register binary with CUDA runtime. This is substantially different in |
944 | // default mode vs. separate compilation! |
945 | // GpuBinaryHandle = __cudaRegisterFatBinary(&FatbinWrapper); |
946 | llvm::CallInst *RegisterFatbinCall = |
947 | CtorBuilder.CreateCall(Callee: RegisterFatbinFunc, Args: FatbinWrapper); |
948 | GpuBinaryHandle = new llvm::GlobalVariable( |
949 | TheModule, PtrTy, false, llvm::GlobalValue::InternalLinkage, |
950 | llvm::ConstantPointerNull::get(T: PtrTy), "__cuda_gpubin_handle" ); |
951 | GpuBinaryHandle->setAlignment(CGM.getPointerAlign().getAsAlign()); |
952 | CtorBuilder.CreateAlignedStore(Val: RegisterFatbinCall, Addr: GpuBinaryHandle, |
953 | Align: CGM.getPointerAlign()); |
954 | |
955 | // Call __cuda_register_globals(GpuBinaryHandle); |
956 | if (RegisterGlobalsFunc) |
957 | CtorBuilder.CreateCall(Callee: RegisterGlobalsFunc, Args: RegisterFatbinCall); |
958 | |
959 | // Call __cudaRegisterFatBinaryEnd(Handle) if this CUDA version needs it. |
960 | if (CudaFeatureEnabled(CGM.getTarget().getSDKVersion(), |
961 | CudaFeature::CUDA_USES_FATBIN_REGISTER_END)) { |
962 | // void __cudaRegisterFatBinaryEnd(void **); |
963 | llvm::FunctionCallee RegisterFatbinEndFunc = CGM.CreateRuntimeFunction( |
964 | Ty: llvm::FunctionType::get(Result: VoidTy, Params: PtrTy, isVarArg: false), |
965 | Name: "__cudaRegisterFatBinaryEnd" ); |
966 | CtorBuilder.CreateCall(Callee: RegisterFatbinEndFunc, Args: RegisterFatbinCall); |
967 | } |
968 | } else { |
969 | // Generate a unique module ID. |
970 | SmallString<64> ModuleID; |
971 | llvm::raw_svector_ostream OS(ModuleID); |
972 | OS << ModuleIDPrefix << llvm::format(Fmt: "%" PRIx64, Vals: FatbinWrapper->getGUID()); |
973 | llvm::Constant *ModuleIDConstant = makeConstantArray( |
974 | Str: std::string(ModuleID), Name: "" , SectionName: ModuleIDSectionName, Alignment: 32, /*AddNull=*/true); |
975 | |
976 | // Create an alias for the FatbinWrapper that nvcc will look for. |
977 | llvm::GlobalAlias::create(Linkage: llvm::GlobalValue::ExternalLinkage, |
978 | Name: Twine("__fatbinwrap" ) + ModuleID, Aliasee: FatbinWrapper); |
979 | |
980 | // void __cudaRegisterLinkedBinary%ModuleID%(void (*)(void *), void *, |
981 | // void *, void (*)(void **)) |
982 | SmallString<128> RegisterLinkedBinaryName("__cudaRegisterLinkedBinary" ); |
983 | RegisterLinkedBinaryName += ModuleID; |
984 | llvm::FunctionCallee RegisterLinkedBinaryFunc = CGM.CreateRuntimeFunction( |
985 | Ty: getRegisterLinkedBinaryFnTy(), Name: RegisterLinkedBinaryName); |
986 | |
987 | assert(RegisterGlobalsFunc && "Expecting at least dummy function!" ); |
988 | llvm::Value *Args[] = {RegisterGlobalsFunc, FatbinWrapper, ModuleIDConstant, |
989 | makeDummyFunction(FnTy: getCallbackFnTy())}; |
990 | CtorBuilder.CreateCall(Callee: RegisterLinkedBinaryFunc, Args); |
991 | } |
992 | |
993 | // Create destructor and register it with atexit() the way NVCC does it. Doing |
994 | // it during regular destructor phase worked in CUDA before 9.2 but results in |
995 | // double-free in 9.2. |
996 | if (llvm::Function *CleanupFn = makeModuleDtorFunction()) { |
997 | // extern "C" int atexit(void (*f)(void)); |
998 | llvm::FunctionType *AtExitTy = |
999 | llvm::FunctionType::get(Result: IntTy, Params: CleanupFn->getType(), isVarArg: false); |
1000 | llvm::FunctionCallee AtExitFunc = |
1001 | CGM.CreateRuntimeFunction(Ty: AtExitTy, Name: "atexit" , ExtraAttrs: llvm::AttributeList(), |
1002 | /*Local=*/true); |
1003 | CtorBuilder.CreateCall(Callee: AtExitFunc, Args: CleanupFn); |
1004 | } |
1005 | |
1006 | CtorBuilder.CreateRetVoid(); |
1007 | return ModuleCtorFunc; |
1008 | } |
1009 | |
1010 | /// Creates a global destructor function that unregisters the GPU code blob |
1011 | /// registered by constructor. |
1012 | /// |
1013 | /// For CUDA: |
1014 | /// \code |
1015 | /// void __cuda_module_dtor() { |
1016 | /// __cudaUnregisterFatBinary(Handle); |
1017 | /// } |
1018 | /// \endcode |
1019 | /// |
1020 | /// For HIP: |
1021 | /// \code |
1022 | /// void __hip_module_dtor() { |
1023 | /// if (__hip_gpubin_handle) { |
1024 | /// __hipUnregisterFatBinary(__hip_gpubin_handle); |
1025 | /// __hip_gpubin_handle = 0; |
1026 | /// } |
1027 | /// } |
1028 | /// \endcode |
1029 | llvm::Function *CGNVCUDARuntime::makeModuleDtorFunction() { |
1030 | // No need for destructor if we don't have a handle to unregister. |
1031 | if (!GpuBinaryHandle) |
1032 | return nullptr; |
1033 | |
1034 | // void __cudaUnregisterFatBinary(void ** handle); |
1035 | llvm::FunctionCallee UnregisterFatbinFunc = CGM.CreateRuntimeFunction( |
1036 | Ty: llvm::FunctionType::get(Result: VoidTy, Params: PtrTy, isVarArg: false), |
1037 | Name: addUnderscoredPrefixToName(FuncName: "UnregisterFatBinary" )); |
1038 | |
1039 | llvm::Function *ModuleDtorFunc = llvm::Function::Create( |
1040 | Ty: llvm::FunctionType::get(Result: VoidTy, isVarArg: false), |
1041 | Linkage: llvm::GlobalValue::InternalLinkage, |
1042 | N: addUnderscoredPrefixToName(FuncName: "_module_dtor" ), M: &TheModule); |
1043 | |
1044 | llvm::BasicBlock *DtorEntryBB = |
1045 | llvm::BasicBlock::Create(Context, Name: "entry" , Parent: ModuleDtorFunc); |
1046 | CGBuilderTy DtorBuilder(CGM, Context); |
1047 | DtorBuilder.SetInsertPoint(DtorEntryBB); |
1048 | |
1049 | Address GpuBinaryAddr( |
1050 | GpuBinaryHandle, GpuBinaryHandle->getValueType(), |
1051 | CharUnits::fromQuantity(Quantity: GpuBinaryHandle->getAlignment())); |
1052 | auto *HandleValue = DtorBuilder.CreateLoad(Addr: GpuBinaryAddr); |
1053 | // There is only one HIP fat binary per linked module, however there are |
1054 | // multiple destructor functions. Make sure the fat binary is unregistered |
1055 | // only once. |
1056 | if (CGM.getLangOpts().HIP) { |
1057 | llvm::BasicBlock *IfBlock = |
1058 | llvm::BasicBlock::Create(Context, Name: "if" , Parent: ModuleDtorFunc); |
1059 | llvm::BasicBlock *ExitBlock = |
1060 | llvm::BasicBlock::Create(Context, Name: "exit" , Parent: ModuleDtorFunc); |
1061 | llvm::Constant *Zero = llvm::Constant::getNullValue(Ty: HandleValue->getType()); |
1062 | llvm::Value *NEZero = DtorBuilder.CreateICmpNE(LHS: HandleValue, RHS: Zero); |
1063 | DtorBuilder.CreateCondBr(Cond: NEZero, True: IfBlock, False: ExitBlock); |
1064 | |
1065 | DtorBuilder.SetInsertPoint(IfBlock); |
1066 | DtorBuilder.CreateCall(Callee: UnregisterFatbinFunc, Args: HandleValue); |
1067 | DtorBuilder.CreateStore(Val: Zero, Addr: GpuBinaryAddr); |
1068 | DtorBuilder.CreateBr(Dest: ExitBlock); |
1069 | |
1070 | DtorBuilder.SetInsertPoint(ExitBlock); |
1071 | } else { |
1072 | DtorBuilder.CreateCall(Callee: UnregisterFatbinFunc, Args: HandleValue); |
1073 | } |
1074 | DtorBuilder.CreateRetVoid(); |
1075 | return ModuleDtorFunc; |
1076 | } |
1077 | |
1078 | CGCUDARuntime *CodeGen::CreateNVCUDARuntime(CodeGenModule &CGM) { |
1079 | return new CGNVCUDARuntime(CGM); |
1080 | } |
1081 | |
1082 | void CGNVCUDARuntime::internalizeDeviceSideVar( |
1083 | const VarDecl *D, llvm::GlobalValue::LinkageTypes &Linkage) { |
1084 | // For -fno-gpu-rdc, host-side shadows of external declarations of device-side |
1085 | // global variables become internal definitions. These have to be internal in |
1086 | // order to prevent name conflicts with global host variables with the same |
1087 | // name in a different TUs. |
1088 | // |
1089 | // For -fgpu-rdc, the shadow variables should not be internalized because |
1090 | // they may be accessed by different TU. |
1091 | if (CGM.getLangOpts().GPURelocatableDeviceCode) |
1092 | return; |
1093 | |
1094 | // __shared__ variables are odd. Shadows do get created, but |
1095 | // they are not registered with the CUDA runtime, so they |
1096 | // can't really be used to access their device-side |
1097 | // counterparts. It's not clear yet whether it's nvcc's bug or |
1098 | // a feature, but we've got to do the same for compatibility. |
1099 | if (D->hasAttr<CUDADeviceAttr>() || D->hasAttr<CUDAConstantAttr>() || |
1100 | D->hasAttr<CUDASharedAttr>() || |
1101 | D->getType()->isCUDADeviceBuiltinSurfaceType() || |
1102 | D->getType()->isCUDADeviceBuiltinTextureType()) { |
1103 | Linkage = llvm::GlobalValue::InternalLinkage; |
1104 | } |
1105 | } |
1106 | |
1107 | void CGNVCUDARuntime::handleVarRegistration(const VarDecl *D, |
1108 | llvm::GlobalVariable &GV) { |
1109 | if (D->hasAttr<CUDADeviceAttr>() || D->hasAttr<CUDAConstantAttr>()) { |
1110 | // Shadow variables and their properties must be registered with CUDA |
1111 | // runtime. Skip Extern global variables, which will be registered in |
1112 | // the TU where they are defined. |
1113 | // |
1114 | // Don't register a C++17 inline variable. The local symbol can be |
1115 | // discarded and referencing a discarded local symbol from outside the |
1116 | // comdat (__cuda_register_globals) is disallowed by the ELF spec. |
1117 | // |
1118 | // HIP managed variables need to be always recorded in device and host |
1119 | // compilations for transformation. |
1120 | // |
1121 | // HIP managed variables and variables in CUDADeviceVarODRUsedByHost are |
1122 | // added to llvm.compiler-used, therefore they are safe to be registered. |
1123 | if ((!D->hasExternalStorage() && !D->isInline()) || |
1124 | CGM.getContext().CUDADeviceVarODRUsedByHost.contains(V: D) || |
1125 | D->hasAttr<HIPManagedAttr>()) { |
1126 | registerDeviceVar(VD: D, Var&: GV, Extern: !D->hasDefinition(), |
1127 | Constant: D->hasAttr<CUDAConstantAttr>()); |
1128 | } |
1129 | } else if (D->getType()->isCUDADeviceBuiltinSurfaceType() || |
1130 | D->getType()->isCUDADeviceBuiltinTextureType()) { |
1131 | // Builtin surfaces and textures and their template arguments are |
1132 | // also registered with CUDA runtime. |
1133 | const auto *TD = cast<ClassTemplateSpecializationDecl>( |
1134 | Val: D->getType()->castAs<RecordType>()->getDecl()); |
1135 | const TemplateArgumentList &Args = TD->getTemplateArgs(); |
1136 | if (TD->hasAttr<CUDADeviceBuiltinSurfaceTypeAttr>()) { |
1137 | assert(Args.size() == 2 && |
1138 | "Unexpected number of template arguments of CUDA device " |
1139 | "builtin surface type." ); |
1140 | auto SurfType = Args[1].getAsIntegral(); |
1141 | if (!D->hasExternalStorage()) |
1142 | registerDeviceSurf(VD: D, Var&: GV, Extern: !D->hasDefinition(), Type: SurfType.getSExtValue()); |
1143 | } else { |
1144 | assert(Args.size() == 3 && |
1145 | "Unexpected number of template arguments of CUDA device " |
1146 | "builtin texture type." ); |
1147 | auto TexType = Args[1].getAsIntegral(); |
1148 | auto Normalized = Args[2].getAsIntegral(); |
1149 | if (!D->hasExternalStorage()) |
1150 | registerDeviceTex(VD: D, Var&: GV, Extern: !D->hasDefinition(), Type: TexType.getSExtValue(), |
1151 | Normalized: Normalized.getZExtValue()); |
1152 | } |
1153 | } |
1154 | } |
1155 | |
1156 | // Transform managed variables to pointers to managed variables in device code. |
1157 | // Each use of the original managed variable is replaced by a load from the |
1158 | // transformed managed variable. The transformed managed variable contains |
1159 | // the address of managed memory which will be allocated by the runtime. |
1160 | void CGNVCUDARuntime::transformManagedVars() { |
1161 | for (auto &&Info : DeviceVars) { |
1162 | llvm::GlobalVariable *Var = Info.Var; |
1163 | if (Info.Flags.getKind() == DeviceVarFlags::Variable && |
1164 | Info.Flags.isManaged()) { |
1165 | auto *ManagedVar = new llvm::GlobalVariable( |
1166 | CGM.getModule(), Var->getType(), |
1167 | /*isConstant=*/false, Var->getLinkage(), |
1168 | /*Init=*/Var->isDeclaration() |
1169 | ? nullptr |
1170 | : llvm::ConstantPointerNull::get(T: Var->getType()), |
1171 | /*Name=*/"" , /*InsertBefore=*/nullptr, |
1172 | llvm::GlobalVariable::NotThreadLocal, |
1173 | CGM.getContext().getTargetAddressSpace(AS: CGM.getLangOpts().CUDAIsDevice |
1174 | ? LangAS::cuda_device |
1175 | : LangAS::Default)); |
1176 | ManagedVar->setDSOLocal(Var->isDSOLocal()); |
1177 | ManagedVar->setVisibility(Var->getVisibility()); |
1178 | ManagedVar->setExternallyInitialized(true); |
1179 | replaceManagedVar(Var, ManagedVar); |
1180 | ManagedVar->takeName(V: Var); |
1181 | Var->setName(Twine(ManagedVar->getName()) + ".managed" ); |
1182 | // Keep managed variables even if they are not used in device code since |
1183 | // they need to be allocated by the runtime. |
1184 | if (CGM.getLangOpts().CUDAIsDevice && !Var->isDeclaration()) { |
1185 | assert(!ManagedVar->isDeclaration()); |
1186 | CGM.addCompilerUsedGlobal(GV: Var); |
1187 | CGM.addCompilerUsedGlobal(GV: ManagedVar); |
1188 | } |
1189 | } |
1190 | } |
1191 | } |
1192 | |
1193 | // Creates offloading entries for all the kernels and globals that must be |
1194 | // registered. The linker will provide a pointer to this section so we can |
1195 | // register the symbols with the linked device image. |
1196 | void CGNVCUDARuntime::createOffloadingEntries() { |
1197 | llvm::object::OffloadKind Kind = CGM.getLangOpts().HIP |
1198 | ? llvm::object::OffloadKind::OFK_HIP |
1199 | : llvm::object::OffloadKind::OFK_Cuda; |
1200 | // For now, just spoof this as OpenMP because that's the runtime it uses. |
1201 | if (CGM.getLangOpts().OffloadViaLLVM) |
1202 | Kind = llvm::object::OffloadKind::OFK_OpenMP; |
1203 | |
1204 | llvm::Module &M = CGM.getModule(); |
1205 | for (KernelInfo &I : EmittedKernels) |
1206 | llvm::offloading::emitOffloadingEntry( |
1207 | M, Kind, Addr: KernelHandles[I.Kernel->getName()], |
1208 | Name: getDeviceSideName(ND: cast<NamedDecl>(Val: I.D)), /*Flags=*/Size: 0, /*Data=*/Flags: 0, |
1209 | Data: llvm::offloading::OffloadGlobalEntry); |
1210 | |
1211 | for (VarInfo &I : DeviceVars) { |
1212 | uint64_t VarSize = |
1213 | CGM.getDataLayout().getTypeAllocSize(Ty: I.Var->getValueType()); |
1214 | int32_t Flags = |
1215 | (I.Flags.isExtern() |
1216 | ? static_cast<int32_t>(llvm::offloading::OffloadGlobalExtern) |
1217 | : 0) | |
1218 | (I.Flags.isConstant() |
1219 | ? static_cast<int32_t>(llvm::offloading::OffloadGlobalConstant) |
1220 | : 0) | |
1221 | (I.Flags.isNormalized() |
1222 | ? static_cast<int32_t>(llvm::offloading::OffloadGlobalNormalized) |
1223 | : 0); |
1224 | if (I.Flags.getKind() == DeviceVarFlags::Variable) { |
1225 | if (I.Flags.isManaged()) { |
1226 | assert(I.Var->getName().ends_with(".managed" ) && |
1227 | "HIP managed variables not transformed" ); |
1228 | |
1229 | auto *ManagedVar = M.getNamedGlobal( |
1230 | Name: I.Var->getName().drop_back(N: StringRef(".managed" ).size())); |
1231 | llvm::offloading::emitOffloadingEntry( |
1232 | M, Kind, Addr: I.Var, Name: getDeviceSideName(ND: I.D), Size: VarSize, |
1233 | Flags: llvm::offloading::OffloadGlobalManagedEntry | Flags, |
1234 | /*Data=*/I.Var->getAlignment(), AuxAddr: ManagedVar); |
1235 | } else { |
1236 | llvm::offloading::emitOffloadingEntry( |
1237 | M, Kind, Addr: I.Var, Name: getDeviceSideName(ND: I.D), Size: VarSize, |
1238 | Flags: llvm::offloading::OffloadGlobalEntry | Flags, |
1239 | /*Data=*/0); |
1240 | } |
1241 | } else if (I.Flags.getKind() == DeviceVarFlags::Surface) { |
1242 | llvm::offloading::emitOffloadingEntry( |
1243 | M, Kind, Addr: I.Var, Name: getDeviceSideName(ND: I.D), Size: VarSize, |
1244 | Flags: llvm::offloading::OffloadGlobalSurfaceEntry | Flags, |
1245 | Data: I.Flags.getSurfTexType()); |
1246 | } else if (I.Flags.getKind() == DeviceVarFlags::Texture) { |
1247 | llvm::offloading::emitOffloadingEntry( |
1248 | M, Kind, Addr: I.Var, Name: getDeviceSideName(ND: I.D), Size: VarSize, |
1249 | Flags: llvm::offloading::OffloadGlobalTextureEntry | Flags, |
1250 | Data: I.Flags.getSurfTexType()); |
1251 | } |
1252 | } |
1253 | } |
1254 | |
1255 | // Returns module constructor to be added. |
1256 | llvm::Function *CGNVCUDARuntime::finalizeModule() { |
1257 | transformManagedVars(); |
1258 | if (CGM.getLangOpts().CUDAIsDevice) { |
1259 | // Mark ODR-used device variables as compiler used to prevent it from being |
1260 | // eliminated by optimization. This is necessary for device variables |
1261 | // ODR-used by host functions. Sema correctly marks them as ODR-used no |
1262 | // matter whether they are ODR-used by device or host functions. |
1263 | // |
1264 | // We do not need to do this if the variable has used attribute since it |
1265 | // has already been added. |
1266 | // |
1267 | // Static device variables have been externalized at this point, therefore |
1268 | // variables with LLVM private or internal linkage need not be added. |
1269 | for (auto &&Info : DeviceVars) { |
1270 | auto Kind = Info.Flags.getKind(); |
1271 | if (!Info.Var->isDeclaration() && |
1272 | !llvm::GlobalValue::isLocalLinkage(Linkage: Info.Var->getLinkage()) && |
1273 | (Kind == DeviceVarFlags::Variable || |
1274 | Kind == DeviceVarFlags::Surface || |
1275 | Kind == DeviceVarFlags::Texture) && |
1276 | Info.D->isUsed() && !Info.D->hasAttr<UsedAttr>()) { |
1277 | CGM.addCompilerUsedGlobal(GV: Info.Var); |
1278 | } |
1279 | } |
1280 | return nullptr; |
1281 | } |
1282 | if (CGM.getLangOpts().OffloadViaLLVM || |
1283 | (CGM.getLangOpts().OffloadingNewDriver && |
1284 | (CGM.getLangOpts().HIP || RelocatableDeviceCode))) |
1285 | createOffloadingEntries(); |
1286 | else |
1287 | return makeModuleCtorFunction(); |
1288 | |
1289 | return nullptr; |
1290 | } |
1291 | |
1292 | llvm::GlobalValue *CGNVCUDARuntime::getKernelHandle(llvm::Function *F, |
1293 | GlobalDecl GD) { |
1294 | auto Loc = KernelHandles.find(Val: F->getName()); |
1295 | if (Loc != KernelHandles.end()) { |
1296 | auto OldHandle = Loc->second; |
1297 | if (KernelStubs[OldHandle] == F) |
1298 | return OldHandle; |
1299 | |
1300 | // We've found the function name, but F itself has changed, so we need to |
1301 | // update the references. |
1302 | if (CGM.getLangOpts().HIP) { |
1303 | // For HIP compilation the handle itself does not change, so we only need |
1304 | // to update the Stub value. |
1305 | KernelStubs[OldHandle] = F; |
1306 | return OldHandle; |
1307 | } |
1308 | // For non-HIP compilation, erase the old Stub and fall-through to creating |
1309 | // new entries. |
1310 | KernelStubs.erase(Val: OldHandle); |
1311 | } |
1312 | |
1313 | if (!CGM.getLangOpts().HIP) { |
1314 | KernelHandles[F->getName()] = F; |
1315 | KernelStubs[F] = F; |
1316 | return F; |
1317 | } |
1318 | |
1319 | auto *Var = new llvm::GlobalVariable( |
1320 | TheModule, F->getType(), /*isConstant=*/true, F->getLinkage(), |
1321 | /*Initializer=*/nullptr, |
1322 | CGM.getMangledName( |
1323 | GD: GD.getWithKernelReferenceKind(Kind: KernelReferenceKind::Kernel))); |
1324 | Var->setAlignment(CGM.getPointerAlign().getAsAlign()); |
1325 | Var->setDSOLocal(F->isDSOLocal()); |
1326 | Var->setVisibility(F->getVisibility()); |
1327 | auto *FD = cast<FunctionDecl>(Val: GD.getDecl()); |
1328 | auto *FT = FD->getPrimaryTemplate(); |
1329 | if (!FT || FT->isThisDeclarationADefinition()) |
1330 | CGM.maybeSetTrivialComdat(D: *FD, GO&: *Var); |
1331 | KernelHandles[F->getName()] = Var; |
1332 | KernelStubs[Var] = F; |
1333 | return Var; |
1334 | } |
1335 | |