1 //===--- CGDeclCXX.cpp - Emit LLVM Code for C++ declarations --------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This contains code dealing with code generation of C++ declarations 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenFunction.h" 15 #include "CGCXXABI.h" 16 #include "CGObjCRuntime.h" 17 #include "CGOpenMPRuntime.h" 18 #include "clang/Frontend/CodeGenOptions.h" 19 #include "llvm/ADT/StringExtras.h" 20 #include "llvm/IR/Intrinsics.h" 21 #include "llvm/IR/MDBuilder.h" 22 #include "llvm/Support/Path.h" 23 24 using namespace clang; 25 using namespace CodeGen; 26 27 static void EmitDeclInit(CodeGenFunction &CGF, const VarDecl &D, 28 ConstantAddress DeclPtr) { 29 assert(D.hasGlobalStorage() && "VarDecl must have global storage!"); 30 assert(!D.getType()->isReferenceType() && 31 "Should not call EmitDeclInit on a reference!"); 32 33 QualType type = D.getType(); 34 LValue lv = CGF.MakeAddrLValue(DeclPtr, type); 35 36 const Expr *Init = D.getInit(); 37 switch (CGF.getEvaluationKind(type)) { 38 case TEK_Scalar: { 39 CodeGenModule &CGM = CGF.CGM; 40 if (lv.isObjCStrong()) 41 CGM.getObjCRuntime().EmitObjCGlobalAssign(CGF, CGF.EmitScalarExpr(Init), 42 DeclPtr, D.getTLSKind()); 43 else if (lv.isObjCWeak()) 44 CGM.getObjCRuntime().EmitObjCWeakAssign(CGF, CGF.EmitScalarExpr(Init), 45 DeclPtr); 46 else 47 CGF.EmitScalarInit(Init, &D, lv, false); 48 return; 49 } 50 case TEK_Complex: 51 CGF.EmitComplexExprIntoLValue(Init, lv, /*isInit*/ true); 52 return; 53 case TEK_Aggregate: 54 CGF.EmitAggExpr(Init, AggValueSlot::forLValue(lv,AggValueSlot::IsDestructed, 55 AggValueSlot::DoesNotNeedGCBarriers, 56 AggValueSlot::IsNotAliased)); 57 return; 58 } 59 llvm_unreachable("bad evaluation kind"); 60 } 61 62 /// Emit code to cause the destruction of the given variable with 63 /// static storage duration. 64 static void EmitDeclDestroy(CodeGenFunction &CGF, const VarDecl &D, 65 ConstantAddress addr) { 66 CodeGenModule &CGM = CGF.CGM; 67 68 // FIXME: __attribute__((cleanup)) ? 69 70 QualType type = D.getType(); 71 QualType::DestructionKind dtorKind = type.isDestructedType(); 72 73 switch (dtorKind) { 74 case QualType::DK_none: 75 return; 76 77 case QualType::DK_cxx_destructor: 78 break; 79 80 case QualType::DK_objc_strong_lifetime: 81 case QualType::DK_objc_weak_lifetime: 82 // We don't care about releasing objects during process teardown. 83 assert(!D.getTLSKind() && "should have rejected this"); 84 return; 85 } 86 87 llvm::Constant *function; 88 llvm::Constant *argument; 89 90 // Special-case non-array C++ destructors, if they have the right signature. 91 // Under some ABIs, destructors return this instead of void, and cannot be 92 // passed directly to __cxa_atexit if the target does not allow this mismatch. 93 const CXXRecordDecl *Record = type->getAsCXXRecordDecl(); 94 bool CanRegisterDestructor = 95 Record && (!CGM.getCXXABI().HasThisReturn( 96 GlobalDecl(Record->getDestructor(), Dtor_Complete)) || 97 CGM.getCXXABI().canCallMismatchedFunctionType()); 98 // If __cxa_atexit is disabled via a flag, a different helper function is 99 // generated elsewhere which uses atexit instead, and it takes the destructor 100 // directly. 101 bool UsingExternalHelper = !CGM.getCodeGenOpts().CXAAtExit; 102 if (Record && (CanRegisterDestructor || UsingExternalHelper)) { 103 assert(!Record->hasTrivialDestructor()); 104 CXXDestructorDecl *dtor = Record->getDestructor(); 105 106 function = CGM.getAddrOfCXXStructor(dtor, StructorType::Complete); 107 argument = llvm::ConstantExpr::getBitCast( 108 addr.getPointer(), CGF.getTypes().ConvertType(type)->getPointerTo()); 109 110 // Otherwise, the standard logic requires a helper function. 111 } else { 112 function = CodeGenFunction(CGM) 113 .generateDestroyHelper(addr, type, CGF.getDestroyer(dtorKind), 114 CGF.needsEHCleanup(dtorKind), &D); 115 argument = llvm::Constant::getNullValue(CGF.Int8PtrTy); 116 } 117 118 CGM.getCXXABI().registerGlobalDtor(CGF, D, function, argument); 119 } 120 121 /// Emit code to cause the variable at the given address to be considered as 122 /// constant from this point onwards. 123 static void EmitDeclInvariant(CodeGenFunction &CGF, const VarDecl &D, 124 llvm::Constant *Addr) { 125 // Do not emit the intrinsic if we're not optimizing. 126 if (!CGF.CGM.getCodeGenOpts().OptimizationLevel) 127 return; 128 129 // Grab the llvm.invariant.start intrinsic. 130 llvm::Intrinsic::ID InvStartID = llvm::Intrinsic::invariant_start; 131 // Overloaded address space type. 132 llvm::Type *ObjectPtr[1] = {CGF.Int8PtrTy}; 133 llvm::Constant *InvariantStart = CGF.CGM.getIntrinsic(InvStartID, ObjectPtr); 134 135 // Emit a call with the size in bytes of the object. 136 CharUnits WidthChars = CGF.getContext().getTypeSizeInChars(D.getType()); 137 uint64_t Width = WidthChars.getQuantity(); 138 llvm::Value *Args[2] = { llvm::ConstantInt::getSigned(CGF.Int64Ty, Width), 139 llvm::ConstantExpr::getBitCast(Addr, CGF.Int8PtrTy)}; 140 CGF.Builder.CreateCall(InvariantStart, Args); 141 } 142 143 void CodeGenFunction::EmitCXXGlobalVarDeclInit(const VarDecl &D, 144 llvm::Constant *DeclPtr, 145 bool PerformInit) { 146 147 const Expr *Init = D.getInit(); 148 QualType T = D.getType(); 149 150 // The address space of a static local variable (DeclPtr) may be different 151 // from the address space of the "this" argument of the constructor. In that 152 // case, we need an addrspacecast before calling the constructor. 153 // 154 // struct StructWithCtor { 155 // __device__ StructWithCtor() {...} 156 // }; 157 // __device__ void foo() { 158 // __shared__ StructWithCtor s; 159 // ... 160 // } 161 // 162 // For example, in the above CUDA code, the static local variable s has a 163 // "shared" address space qualifier, but the constructor of StructWithCtor 164 // expects "this" in the "generic" address space. 165 unsigned ExpectedAddrSpace = getContext().getTargetAddressSpace(T); 166 unsigned ActualAddrSpace = DeclPtr->getType()->getPointerAddressSpace(); 167 if (ActualAddrSpace != ExpectedAddrSpace) { 168 llvm::Type *LTy = CGM.getTypes().ConvertTypeForMem(T); 169 llvm::PointerType *PTy = llvm::PointerType::get(LTy, ExpectedAddrSpace); 170 DeclPtr = llvm::ConstantExpr::getAddrSpaceCast(DeclPtr, PTy); 171 } 172 173 ConstantAddress DeclAddr(DeclPtr, getContext().getDeclAlign(&D)); 174 175 if (!T->isReferenceType()) { 176 if (getLangOpts().OpenMP && !getLangOpts().OpenMPSimd && 177 D.hasAttr<OMPThreadPrivateDeclAttr>()) { 178 (void)CGM.getOpenMPRuntime().emitThreadPrivateVarDefinition( 179 &D, DeclAddr, D.getAttr<OMPThreadPrivateDeclAttr>()->getLocation(), 180 PerformInit, this); 181 } 182 if (PerformInit) 183 EmitDeclInit(*this, D, DeclAddr); 184 if (CGM.isTypeConstant(D.getType(), true)) 185 EmitDeclInvariant(*this, D, DeclPtr); 186 else 187 EmitDeclDestroy(*this, D, DeclAddr); 188 return; 189 } 190 191 assert(PerformInit && "cannot have constant initializer which needs " 192 "destruction for reference"); 193 RValue RV = EmitReferenceBindingToExpr(Init); 194 EmitStoreOfScalar(RV.getScalarVal(), DeclAddr, false, T); 195 } 196 197 /// Create a stub function, suitable for being passed to atexit, 198 /// which passes the given address to the given destructor function. 199 llvm::Constant *CodeGenFunction::createAtExitStub(const VarDecl &VD, 200 llvm::Constant *dtor, 201 llvm::Constant *addr) { 202 // Get the destructor function type, void(*)(void). 203 llvm::FunctionType *ty = llvm::FunctionType::get(CGM.VoidTy, false); 204 SmallString<256> FnName; 205 { 206 llvm::raw_svector_ostream Out(FnName); 207 CGM.getCXXABI().getMangleContext().mangleDynamicAtExitDestructor(&VD, Out); 208 } 209 210 const CGFunctionInfo &FI = CGM.getTypes().arrangeNullaryFunction(); 211 llvm::Function *fn = CGM.CreateGlobalInitOrDestructFunction(ty, FnName.str(), 212 FI, 213 VD.getLocation()); 214 215 CodeGenFunction CGF(CGM); 216 217 CGF.StartFunction(&VD, CGM.getContext().VoidTy, fn, FI, FunctionArgList()); 218 219 llvm::CallInst *call = CGF.Builder.CreateCall(dtor, addr); 220 221 // Make sure the call and the callee agree on calling convention. 222 if (llvm::Function *dtorFn = 223 dyn_cast<llvm::Function>(dtor->stripPointerCasts())) 224 call->setCallingConv(dtorFn->getCallingConv()); 225 226 CGF.FinishFunction(); 227 228 return fn; 229 } 230 231 /// Register a global destructor using the C atexit runtime function. 232 void CodeGenFunction::registerGlobalDtorWithAtExit(const VarDecl &VD, 233 llvm::Constant *dtor, 234 llvm::Constant *addr) { 235 // Create a function which calls the destructor. 236 llvm::Constant *dtorStub = createAtExitStub(VD, dtor, addr); 237 238 // extern "C" int atexit(void (*f)(void)); 239 llvm::FunctionType *atexitTy = 240 llvm::FunctionType::get(IntTy, dtorStub->getType(), false); 241 242 llvm::Constant *atexit = 243 CGM.CreateRuntimeFunction(atexitTy, "atexit", llvm::AttributeList(), 244 /*Local=*/true); 245 if (llvm::Function *atexitFn = dyn_cast<llvm::Function>(atexit)) 246 atexitFn->setDoesNotThrow(); 247 248 EmitNounwindRuntimeCall(atexit, dtorStub); 249 } 250 251 void CodeGenFunction::EmitCXXGuardedInit(const VarDecl &D, 252 llvm::GlobalVariable *DeclPtr, 253 bool PerformInit) { 254 // If we've been asked to forbid guard variables, emit an error now. 255 // This diagnostic is hard-coded for Darwin's use case; we can find 256 // better phrasing if someone else needs it. 257 if (CGM.getCodeGenOpts().ForbidGuardVariables) 258 CGM.Error(D.getLocation(), 259 "this initialization requires a guard variable, which " 260 "the kernel does not support"); 261 262 CGM.getCXXABI().EmitGuardedInit(*this, D, DeclPtr, PerformInit); 263 } 264 265 void CodeGenFunction::EmitCXXGuardedInitBranch(llvm::Value *NeedsInit, 266 llvm::BasicBlock *InitBlock, 267 llvm::BasicBlock *NoInitBlock, 268 GuardKind Kind, 269 const VarDecl *D) { 270 assert((Kind == GuardKind::TlsGuard || D) && "no guarded variable"); 271 272 // A guess at how many times we will enter the initialization of a 273 // variable, depending on the kind of variable. 274 static const uint64_t InitsPerTLSVar = 1024; 275 static const uint64_t InitsPerLocalVar = 1024 * 1024; 276 277 llvm::MDNode *Weights; 278 if (Kind == GuardKind::VariableGuard && !D->isLocalVarDecl()) { 279 // For non-local variables, don't apply any weighting for now. Due to our 280 // use of COMDATs, we expect there to be at most one initialization of the 281 // variable per DSO, but we have no way to know how many DSOs will try to 282 // initialize the variable. 283 Weights = nullptr; 284 } else { 285 uint64_t NumInits; 286 // FIXME: For the TLS case, collect and use profiling information to 287 // determine a more accurate brach weight. 288 if (Kind == GuardKind::TlsGuard || D->getTLSKind()) 289 NumInits = InitsPerTLSVar; 290 else 291 NumInits = InitsPerLocalVar; 292 293 // The probability of us entering the initializer is 294 // 1 / (total number of times we attempt to initialize the variable). 295 llvm::MDBuilder MDHelper(CGM.getLLVMContext()); 296 Weights = MDHelper.createBranchWeights(1, NumInits - 1); 297 } 298 299 Builder.CreateCondBr(NeedsInit, InitBlock, NoInitBlock, Weights); 300 } 301 302 llvm::Function *CodeGenModule::CreateGlobalInitOrDestructFunction( 303 llvm::FunctionType *FTy, const Twine &Name, const CGFunctionInfo &FI, 304 SourceLocation Loc, bool TLS) { 305 llvm::Function *Fn = 306 llvm::Function::Create(FTy, llvm::GlobalValue::InternalLinkage, 307 Name, &getModule()); 308 if (!getLangOpts().AppleKext && !TLS) { 309 // Set the section if needed. 310 if (const char *Section = getTarget().getStaticInitSectionSpecifier()) 311 Fn->setSection(Section); 312 } 313 314 SetInternalFunctionAttributes(nullptr, Fn, FI); 315 316 Fn->setCallingConv(getRuntimeCC()); 317 318 if (!getLangOpts().Exceptions) 319 Fn->setDoesNotThrow(); 320 321 if (getLangOpts().Sanitize.has(SanitizerKind::Address) && 322 !isInSanitizerBlacklist(SanitizerKind::Address, Fn, Loc)) 323 Fn->addFnAttr(llvm::Attribute::SanitizeAddress); 324 325 if (getLangOpts().Sanitize.has(SanitizerKind::KernelAddress) && 326 !isInSanitizerBlacklist(SanitizerKind::KernelAddress, Fn, Loc)) 327 Fn->addFnAttr(llvm::Attribute::SanitizeAddress); 328 329 if (getLangOpts().Sanitize.has(SanitizerKind::HWAddress) && 330 !isInSanitizerBlacklist(SanitizerKind::HWAddress, Fn, Loc)) 331 Fn->addFnAttr(llvm::Attribute::SanitizeHWAddress); 332 333 if (getLangOpts().Sanitize.has(SanitizerKind::Thread) && 334 !isInSanitizerBlacklist(SanitizerKind::Thread, Fn, Loc)) 335 Fn->addFnAttr(llvm::Attribute::SanitizeThread); 336 337 if (getLangOpts().Sanitize.has(SanitizerKind::Memory) && 338 !isInSanitizerBlacklist(SanitizerKind::Memory, Fn, Loc)) 339 Fn->addFnAttr(llvm::Attribute::SanitizeMemory); 340 341 if (getLangOpts().Sanitize.has(SanitizerKind::SafeStack) && 342 !isInSanitizerBlacklist(SanitizerKind::SafeStack, Fn, Loc)) 343 Fn->addFnAttr(llvm::Attribute::SafeStack); 344 345 return Fn; 346 } 347 348 /// Create a global pointer to a function that will initialize a global 349 /// variable. The user has requested that this pointer be emitted in a specific 350 /// section. 351 void CodeGenModule::EmitPointerToInitFunc(const VarDecl *D, 352 llvm::GlobalVariable *GV, 353 llvm::Function *InitFunc, 354 InitSegAttr *ISA) { 355 llvm::GlobalVariable *PtrArray = new llvm::GlobalVariable( 356 TheModule, InitFunc->getType(), /*isConstant=*/true, 357 llvm::GlobalValue::PrivateLinkage, InitFunc, "__cxx_init_fn_ptr"); 358 PtrArray->setSection(ISA->getSection()); 359 addUsedGlobal(PtrArray); 360 361 // If the GV is already in a comdat group, then we have to join it. 362 if (llvm::Comdat *C = GV->getComdat()) 363 PtrArray->setComdat(C); 364 } 365 366 void 367 CodeGenModule::EmitCXXGlobalVarDeclInitFunc(const VarDecl *D, 368 llvm::GlobalVariable *Addr, 369 bool PerformInit) { 370 371 // According to E.2.3.1 in CUDA-7.5 Programming guide: __device__, 372 // __constant__ and __shared__ variables defined in namespace scope, 373 // that are of class type, cannot have a non-empty constructor. All 374 // the checks have been done in Sema by now. Whatever initializers 375 // are allowed are empty and we just need to ignore them here. 376 if (getLangOpts().CUDA && getLangOpts().CUDAIsDevice && 377 (D->hasAttr<CUDADeviceAttr>() || D->hasAttr<CUDAConstantAttr>() || 378 D->hasAttr<CUDASharedAttr>())) 379 return; 380 381 // Check if we've already initialized this decl. 382 auto I = DelayedCXXInitPosition.find(D); 383 if (I != DelayedCXXInitPosition.end() && I->second == ~0U) 384 return; 385 386 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 387 SmallString<256> FnName; 388 { 389 llvm::raw_svector_ostream Out(FnName); 390 getCXXABI().getMangleContext().mangleDynamicInitializer(D, Out); 391 } 392 393 // Create a variable initialization function. 394 llvm::Function *Fn = 395 CreateGlobalInitOrDestructFunction(FTy, FnName.str(), 396 getTypes().arrangeNullaryFunction(), 397 D->getLocation()); 398 399 auto *ISA = D->getAttr<InitSegAttr>(); 400 CodeGenFunction(*this).GenerateCXXGlobalVarDeclInitFunc(Fn, D, Addr, 401 PerformInit); 402 403 llvm::GlobalVariable *COMDATKey = 404 supportsCOMDAT() && D->isExternallyVisible() ? Addr : nullptr; 405 406 if (D->getTLSKind()) { 407 // FIXME: Should we support init_priority for thread_local? 408 // FIXME: We only need to register one __cxa_thread_atexit function for the 409 // entire TU. 410 CXXThreadLocalInits.push_back(Fn); 411 CXXThreadLocalInitVars.push_back(D); 412 } else if (PerformInit && ISA) { 413 EmitPointerToInitFunc(D, Addr, Fn, ISA); 414 } else if (auto *IPA = D->getAttr<InitPriorityAttr>()) { 415 OrderGlobalInits Key(IPA->getPriority(), PrioritizedCXXGlobalInits.size()); 416 PrioritizedCXXGlobalInits.push_back(std::make_pair(Key, Fn)); 417 } else if (isTemplateInstantiation(D->getTemplateSpecializationKind())) { 418 // C++ [basic.start.init]p2: 419 // Definitions of explicitly specialized class template static data 420 // members have ordered initialization. Other class template static data 421 // members (i.e., implicitly or explicitly instantiated specializations) 422 // have unordered initialization. 423 // 424 // As a consequence, we can put them into their own llvm.global_ctors entry. 425 // 426 // If the global is externally visible, put the initializer into a COMDAT 427 // group with the global being initialized. On most platforms, this is a 428 // minor startup time optimization. In the MS C++ ABI, there are no guard 429 // variables, so this COMDAT key is required for correctness. 430 AddGlobalCtor(Fn, 65535, COMDATKey); 431 } else if (D->hasAttr<SelectAnyAttr>()) { 432 // SelectAny globals will be comdat-folded. Put the initializer into a 433 // COMDAT group associated with the global, so the initializers get folded 434 // too. 435 AddGlobalCtor(Fn, 65535, COMDATKey); 436 } else { 437 I = DelayedCXXInitPosition.find(D); // Re-do lookup in case of re-hash. 438 if (I == DelayedCXXInitPosition.end()) { 439 CXXGlobalInits.push_back(Fn); 440 } else if (I->second != ~0U) { 441 assert(I->second < CXXGlobalInits.size() && 442 CXXGlobalInits[I->second] == nullptr); 443 CXXGlobalInits[I->second] = Fn; 444 } 445 } 446 447 // Remember that we already emitted the initializer for this global. 448 DelayedCXXInitPosition[D] = ~0U; 449 } 450 451 void CodeGenModule::EmitCXXThreadLocalInitFunc() { 452 getCXXABI().EmitThreadLocalInitFuncs( 453 *this, CXXThreadLocals, CXXThreadLocalInits, CXXThreadLocalInitVars); 454 455 CXXThreadLocalInits.clear(); 456 CXXThreadLocalInitVars.clear(); 457 CXXThreadLocals.clear(); 458 } 459 460 void 461 CodeGenModule::EmitCXXGlobalInitFunc() { 462 while (!CXXGlobalInits.empty() && !CXXGlobalInits.back()) 463 CXXGlobalInits.pop_back(); 464 465 if (CXXGlobalInits.empty() && PrioritizedCXXGlobalInits.empty()) 466 return; 467 468 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 469 const CGFunctionInfo &FI = getTypes().arrangeNullaryFunction(); 470 471 // Create our global initialization function. 472 if (!PrioritizedCXXGlobalInits.empty()) { 473 SmallVector<llvm::Function *, 8> LocalCXXGlobalInits; 474 llvm::array_pod_sort(PrioritizedCXXGlobalInits.begin(), 475 PrioritizedCXXGlobalInits.end()); 476 // Iterate over "chunks" of ctors with same priority and emit each chunk 477 // into separate function. Note - everything is sorted first by priority, 478 // second - by lex order, so we emit ctor functions in proper order. 479 for (SmallVectorImpl<GlobalInitData >::iterator 480 I = PrioritizedCXXGlobalInits.begin(), 481 E = PrioritizedCXXGlobalInits.end(); I != E; ) { 482 SmallVectorImpl<GlobalInitData >::iterator 483 PrioE = std::upper_bound(I + 1, E, *I, GlobalInitPriorityCmp()); 484 485 LocalCXXGlobalInits.clear(); 486 unsigned Priority = I->first.priority; 487 // Compute the function suffix from priority. Prepend with zeroes to make 488 // sure the function names are also ordered as priorities. 489 std::string PrioritySuffix = llvm::utostr(Priority); 490 // Priority is always <= 65535 (enforced by sema). 491 PrioritySuffix = std::string(6-PrioritySuffix.size(), '0')+PrioritySuffix; 492 llvm::Function *Fn = CreateGlobalInitOrDestructFunction( 493 FTy, "_GLOBAL__I_" + PrioritySuffix, FI); 494 495 for (; I < PrioE; ++I) 496 LocalCXXGlobalInits.push_back(I->second); 497 498 CodeGenFunction(*this).GenerateCXXGlobalInitFunc(Fn, LocalCXXGlobalInits); 499 AddGlobalCtor(Fn, Priority); 500 } 501 PrioritizedCXXGlobalInits.clear(); 502 } 503 504 // Include the filename in the symbol name. Including "sub_" matches gcc and 505 // makes sure these symbols appear lexicographically behind the symbols with 506 // priority emitted above. 507 SmallString<128> FileName = llvm::sys::path::filename(getModule().getName()); 508 if (FileName.empty()) 509 FileName = "<null>"; 510 511 for (size_t i = 0; i < FileName.size(); ++i) { 512 // Replace everything that's not [a-zA-Z0-9._] with a _. This set happens 513 // to be the set of C preprocessing numbers. 514 if (!isPreprocessingNumberBody(FileName[i])) 515 FileName[i] = '_'; 516 } 517 518 llvm::Function *Fn = CreateGlobalInitOrDestructFunction( 519 FTy, llvm::Twine("_GLOBAL__sub_I_", FileName), FI); 520 521 CodeGenFunction(*this).GenerateCXXGlobalInitFunc(Fn, CXXGlobalInits); 522 AddGlobalCtor(Fn); 523 524 CXXGlobalInits.clear(); 525 } 526 527 void CodeGenModule::EmitCXXGlobalDtorFunc() { 528 if (CXXGlobalDtors.empty()) 529 return; 530 531 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 532 533 // Create our global destructor function. 534 const CGFunctionInfo &FI = getTypes().arrangeNullaryFunction(); 535 llvm::Function *Fn = 536 CreateGlobalInitOrDestructFunction(FTy, "_GLOBAL__D_a", FI); 537 538 CodeGenFunction(*this).GenerateCXXGlobalDtorsFunc(Fn, CXXGlobalDtors); 539 AddGlobalDtor(Fn); 540 } 541 542 /// Emit the code necessary to initialize the given global variable. 543 void CodeGenFunction::GenerateCXXGlobalVarDeclInitFunc(llvm::Function *Fn, 544 const VarDecl *D, 545 llvm::GlobalVariable *Addr, 546 bool PerformInit) { 547 // Check if we need to emit debug info for variable initializer. 548 if (D->hasAttr<NoDebugAttr>()) 549 DebugInfo = nullptr; // disable debug info indefinitely for this function 550 551 CurEHLocation = D->getLocStart(); 552 553 StartFunction(GlobalDecl(D), getContext().VoidTy, Fn, 554 getTypes().arrangeNullaryFunction(), 555 FunctionArgList(), D->getLocation(), 556 D->getInit()->getExprLoc()); 557 558 // Use guarded initialization if the global variable is weak. This 559 // occurs for, e.g., instantiated static data members and 560 // definitions explicitly marked weak. 561 if (Addr->hasWeakLinkage() || Addr->hasLinkOnceLinkage()) { 562 EmitCXXGuardedInit(*D, Addr, PerformInit); 563 } else { 564 EmitCXXGlobalVarDeclInit(*D, Addr, PerformInit); 565 } 566 567 FinishFunction(); 568 } 569 570 void 571 CodeGenFunction::GenerateCXXGlobalInitFunc(llvm::Function *Fn, 572 ArrayRef<llvm::Function *> Decls, 573 Address Guard) { 574 { 575 auto NL = ApplyDebugLocation::CreateEmpty(*this); 576 StartFunction(GlobalDecl(), getContext().VoidTy, Fn, 577 getTypes().arrangeNullaryFunction(), FunctionArgList()); 578 // Emit an artificial location for this function. 579 auto AL = ApplyDebugLocation::CreateArtificial(*this); 580 581 llvm::BasicBlock *ExitBlock = nullptr; 582 if (Guard.isValid()) { 583 // If we have a guard variable, check whether we've already performed 584 // these initializations. This happens for TLS initialization functions. 585 llvm::Value *GuardVal = Builder.CreateLoad(Guard); 586 llvm::Value *Uninit = Builder.CreateIsNull(GuardVal, 587 "guard.uninitialized"); 588 llvm::BasicBlock *InitBlock = createBasicBlock("init"); 589 ExitBlock = createBasicBlock("exit"); 590 EmitCXXGuardedInitBranch(Uninit, InitBlock, ExitBlock, 591 GuardKind::TlsGuard, nullptr); 592 EmitBlock(InitBlock); 593 // Mark as initialized before initializing anything else. If the 594 // initializers use previously-initialized thread_local vars, that's 595 // probably supposed to be OK, but the standard doesn't say. 596 Builder.CreateStore(llvm::ConstantInt::get(GuardVal->getType(),1), Guard); 597 } 598 599 RunCleanupsScope Scope(*this); 600 601 // When building in Objective-C++ ARC mode, create an autorelease pool 602 // around the global initializers. 603 if (getLangOpts().ObjCAutoRefCount && getLangOpts().CPlusPlus) { 604 llvm::Value *token = EmitObjCAutoreleasePoolPush(); 605 EmitObjCAutoreleasePoolCleanup(token); 606 } 607 608 for (unsigned i = 0, e = Decls.size(); i != e; ++i) 609 if (Decls[i]) 610 EmitRuntimeCall(Decls[i]); 611 612 Scope.ForceCleanup(); 613 614 if (ExitBlock) { 615 Builder.CreateBr(ExitBlock); 616 EmitBlock(ExitBlock); 617 } 618 } 619 620 FinishFunction(); 621 } 622 623 void CodeGenFunction::GenerateCXXGlobalDtorsFunc( 624 llvm::Function *Fn, 625 const std::vector<std::pair<llvm::WeakTrackingVH, llvm::Constant *>> 626 &DtorsAndObjects) { 627 { 628 auto NL = ApplyDebugLocation::CreateEmpty(*this); 629 StartFunction(GlobalDecl(), getContext().VoidTy, Fn, 630 getTypes().arrangeNullaryFunction(), FunctionArgList()); 631 // Emit an artificial location for this function. 632 auto AL = ApplyDebugLocation::CreateArtificial(*this); 633 634 // Emit the dtors, in reverse order from construction. 635 for (unsigned i = 0, e = DtorsAndObjects.size(); i != e; ++i) { 636 llvm::Value *Callee = DtorsAndObjects[e - i - 1].first; 637 llvm::CallInst *CI = Builder.CreateCall(Callee, 638 DtorsAndObjects[e - i - 1].second); 639 // Make sure the call and the callee agree on calling convention. 640 if (llvm::Function *F = dyn_cast<llvm::Function>(Callee)) 641 CI->setCallingConv(F->getCallingConv()); 642 } 643 } 644 645 FinishFunction(); 646 } 647 648 /// generateDestroyHelper - Generates a helper function which, when 649 /// invoked, destroys the given object. The address of the object 650 /// should be in global memory. 651 llvm::Function *CodeGenFunction::generateDestroyHelper( 652 Address addr, QualType type, Destroyer *destroyer, 653 bool useEHCleanupForArray, const VarDecl *VD) { 654 FunctionArgList args; 655 ImplicitParamDecl Dst(getContext(), getContext().VoidPtrTy, 656 ImplicitParamDecl::Other); 657 args.push_back(&Dst); 658 659 const CGFunctionInfo &FI = 660 CGM.getTypes().arrangeBuiltinFunctionDeclaration(getContext().VoidTy, args); 661 llvm::FunctionType *FTy = CGM.getTypes().GetFunctionType(FI); 662 llvm::Function *fn = CGM.CreateGlobalInitOrDestructFunction( 663 FTy, "__cxx_global_array_dtor", FI, VD->getLocation()); 664 665 CurEHLocation = VD->getLocStart(); 666 667 StartFunction(VD, getContext().VoidTy, fn, FI, args); 668 669 emitDestroy(addr, type, destroyer, useEHCleanupForArray); 670 671 FinishFunction(); 672 673 return fn; 674 } 675