1 //===--- CGDeclCXX.cpp - Emit LLVM Code for C++ declarations --------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This contains code dealing with code generation of C++ declarations 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenFunction.h" 15 #include "CGCXXABI.h" 16 #include "CGObjCRuntime.h" 17 #include "CGOpenMPRuntime.h" 18 #include "clang/Frontend/CodeGenOptions.h" 19 #include "llvm/ADT/StringExtras.h" 20 #include "llvm/IR/Intrinsics.h" 21 #include "llvm/IR/MDBuilder.h" 22 #include "llvm/Support/Path.h" 23 24 using namespace clang; 25 using namespace CodeGen; 26 27 static void EmitDeclInit(CodeGenFunction &CGF, const VarDecl &D, 28 ConstantAddress DeclPtr) { 29 assert(D.hasGlobalStorage() && "VarDecl must have global storage!"); 30 assert(!D.getType()->isReferenceType() && 31 "Should not call EmitDeclInit on a reference!"); 32 33 QualType type = D.getType(); 34 LValue lv = CGF.MakeAddrLValue(DeclPtr, type); 35 36 const Expr *Init = D.getInit(); 37 switch (CGF.getEvaluationKind(type)) { 38 case TEK_Scalar: { 39 CodeGenModule &CGM = CGF.CGM; 40 if (lv.isObjCStrong()) 41 CGM.getObjCRuntime().EmitObjCGlobalAssign(CGF, CGF.EmitScalarExpr(Init), 42 DeclPtr, D.getTLSKind()); 43 else if (lv.isObjCWeak()) 44 CGM.getObjCRuntime().EmitObjCWeakAssign(CGF, CGF.EmitScalarExpr(Init), 45 DeclPtr); 46 else 47 CGF.EmitScalarInit(Init, &D, lv, false); 48 return; 49 } 50 case TEK_Complex: 51 CGF.EmitComplexExprIntoLValue(Init, lv, /*isInit*/ true); 52 return; 53 case TEK_Aggregate: 54 CGF.EmitAggExpr(Init, AggValueSlot::forLValue(lv,AggValueSlot::IsDestructed, 55 AggValueSlot::DoesNotNeedGCBarriers, 56 AggValueSlot::IsNotAliased)); 57 return; 58 } 59 llvm_unreachable("bad evaluation kind"); 60 } 61 62 /// Emit code to cause the destruction of the given variable with 63 /// static storage duration. 64 static void EmitDeclDestroy(CodeGenFunction &CGF, const VarDecl &D, 65 ConstantAddress addr) { 66 CodeGenModule &CGM = CGF.CGM; 67 68 // FIXME: __attribute__((cleanup)) ? 69 70 QualType type = D.getType(); 71 QualType::DestructionKind dtorKind = type.isDestructedType(); 72 73 switch (dtorKind) { 74 case QualType::DK_none: 75 return; 76 77 case QualType::DK_cxx_destructor: 78 break; 79 80 case QualType::DK_objc_strong_lifetime: 81 case QualType::DK_objc_weak_lifetime: 82 case QualType::DK_nontrivial_c_struct: 83 // We don't care about releasing objects during process teardown. 84 assert(!D.getTLSKind() && "should have rejected this"); 85 return; 86 } 87 88 llvm::Constant *function; 89 llvm::Constant *argument; 90 91 // Special-case non-array C++ destructors, if they have the right signature. 92 // Under some ABIs, destructors return this instead of void, and cannot be 93 // passed directly to __cxa_atexit if the target does not allow this mismatch. 94 const CXXRecordDecl *Record = type->getAsCXXRecordDecl(); 95 bool CanRegisterDestructor = 96 Record && (!CGM.getCXXABI().HasThisReturn( 97 GlobalDecl(Record->getDestructor(), Dtor_Complete)) || 98 CGM.getCXXABI().canCallMismatchedFunctionType()); 99 // If __cxa_atexit is disabled via a flag, a different helper function is 100 // generated elsewhere which uses atexit instead, and it takes the destructor 101 // directly. 102 bool UsingExternalHelper = !CGM.getCodeGenOpts().CXAAtExit; 103 if (Record && (CanRegisterDestructor || UsingExternalHelper)) { 104 assert(!Record->hasTrivialDestructor()); 105 CXXDestructorDecl *dtor = Record->getDestructor(); 106 107 function = CGM.getAddrOfCXXStructor(dtor, StructorType::Complete); 108 argument = llvm::ConstantExpr::getBitCast( 109 addr.getPointer(), CGF.getTypes().ConvertType(type)->getPointerTo()); 110 111 // Otherwise, the standard logic requires a helper function. 112 } else { 113 function = CodeGenFunction(CGM) 114 .generateDestroyHelper(addr, type, CGF.getDestroyer(dtorKind), 115 CGF.needsEHCleanup(dtorKind), &D); 116 argument = llvm::Constant::getNullValue(CGF.Int8PtrTy); 117 } 118 119 CGM.getCXXABI().registerGlobalDtor(CGF, D, function, argument); 120 } 121 122 /// Emit code to cause the variable at the given address to be considered as 123 /// constant from this point onwards. 124 static void EmitDeclInvariant(CodeGenFunction &CGF, const VarDecl &D, 125 llvm::Constant *Addr) { 126 // Do not emit the intrinsic if we're not optimizing. 127 if (!CGF.CGM.getCodeGenOpts().OptimizationLevel) 128 return; 129 130 // Grab the llvm.invariant.start intrinsic. 131 llvm::Intrinsic::ID InvStartID = llvm::Intrinsic::invariant_start; 132 // Overloaded address space type. 133 llvm::Type *ObjectPtr[1] = {CGF.Int8PtrTy}; 134 llvm::Constant *InvariantStart = CGF.CGM.getIntrinsic(InvStartID, ObjectPtr); 135 136 // Emit a call with the size in bytes of the object. 137 CharUnits WidthChars = CGF.getContext().getTypeSizeInChars(D.getType()); 138 uint64_t Width = WidthChars.getQuantity(); 139 llvm::Value *Args[2] = { llvm::ConstantInt::getSigned(CGF.Int64Ty, Width), 140 llvm::ConstantExpr::getBitCast(Addr, CGF.Int8PtrTy)}; 141 CGF.Builder.CreateCall(InvariantStart, Args); 142 } 143 144 void CodeGenFunction::EmitCXXGlobalVarDeclInit(const VarDecl &D, 145 llvm::Constant *DeclPtr, 146 bool PerformInit) { 147 148 const Expr *Init = D.getInit(); 149 QualType T = D.getType(); 150 151 // The address space of a static local variable (DeclPtr) may be different 152 // from the address space of the "this" argument of the constructor. In that 153 // case, we need an addrspacecast before calling the constructor. 154 // 155 // struct StructWithCtor { 156 // __device__ StructWithCtor() {...} 157 // }; 158 // __device__ void foo() { 159 // __shared__ StructWithCtor s; 160 // ... 161 // } 162 // 163 // For example, in the above CUDA code, the static local variable s has a 164 // "shared" address space qualifier, but the constructor of StructWithCtor 165 // expects "this" in the "generic" address space. 166 unsigned ExpectedAddrSpace = getContext().getTargetAddressSpace(T); 167 unsigned ActualAddrSpace = DeclPtr->getType()->getPointerAddressSpace(); 168 if (ActualAddrSpace != ExpectedAddrSpace) { 169 llvm::Type *LTy = CGM.getTypes().ConvertTypeForMem(T); 170 llvm::PointerType *PTy = llvm::PointerType::get(LTy, ExpectedAddrSpace); 171 DeclPtr = llvm::ConstantExpr::getAddrSpaceCast(DeclPtr, PTy); 172 } 173 174 ConstantAddress DeclAddr(DeclPtr, getContext().getDeclAlign(&D)); 175 176 if (!T->isReferenceType()) { 177 if (getLangOpts().OpenMP && !getLangOpts().OpenMPSimd && 178 D.hasAttr<OMPThreadPrivateDeclAttr>()) { 179 (void)CGM.getOpenMPRuntime().emitThreadPrivateVarDefinition( 180 &D, DeclAddr, D.getAttr<OMPThreadPrivateDeclAttr>()->getLocation(), 181 PerformInit, this); 182 } 183 if (PerformInit) 184 EmitDeclInit(*this, D, DeclAddr); 185 if (CGM.isTypeConstant(D.getType(), true)) 186 EmitDeclInvariant(*this, D, DeclPtr); 187 else 188 EmitDeclDestroy(*this, D, DeclAddr); 189 return; 190 } 191 192 assert(PerformInit && "cannot have constant initializer which needs " 193 "destruction for reference"); 194 RValue RV = EmitReferenceBindingToExpr(Init); 195 EmitStoreOfScalar(RV.getScalarVal(), DeclAddr, false, T); 196 } 197 198 /// Create a stub function, suitable for being passed to atexit, 199 /// which passes the given address to the given destructor function. 200 llvm::Constant *CodeGenFunction::createAtExitStub(const VarDecl &VD, 201 llvm::Constant *dtor, 202 llvm::Constant *addr) { 203 // Get the destructor function type, void(*)(void). 204 llvm::FunctionType *ty = llvm::FunctionType::get(CGM.VoidTy, false); 205 SmallString<256> FnName; 206 { 207 llvm::raw_svector_ostream Out(FnName); 208 CGM.getCXXABI().getMangleContext().mangleDynamicAtExitDestructor(&VD, Out); 209 } 210 211 const CGFunctionInfo &FI = CGM.getTypes().arrangeNullaryFunction(); 212 llvm::Function *fn = CGM.CreateGlobalInitOrDestructFunction(ty, FnName.str(), 213 FI, 214 VD.getLocation()); 215 216 CodeGenFunction CGF(CGM); 217 218 CGF.StartFunction(&VD, CGM.getContext().VoidTy, fn, FI, FunctionArgList()); 219 220 llvm::CallInst *call = CGF.Builder.CreateCall(dtor, addr); 221 222 // Make sure the call and the callee agree on calling convention. 223 if (llvm::Function *dtorFn = 224 dyn_cast<llvm::Function>(dtor->stripPointerCasts())) 225 call->setCallingConv(dtorFn->getCallingConv()); 226 227 CGF.FinishFunction(); 228 229 return fn; 230 } 231 232 /// Register a global destructor using the C atexit runtime function. 233 void CodeGenFunction::registerGlobalDtorWithAtExit(const VarDecl &VD, 234 llvm::Constant *dtor, 235 llvm::Constant *addr) { 236 // Create a function which calls the destructor. 237 llvm::Constant *dtorStub = createAtExitStub(VD, dtor, addr); 238 239 // extern "C" int atexit(void (*f)(void)); 240 llvm::FunctionType *atexitTy = 241 llvm::FunctionType::get(IntTy, dtorStub->getType(), false); 242 243 llvm::Constant *atexit = 244 CGM.CreateRuntimeFunction(atexitTy, "atexit", llvm::AttributeList(), 245 /*Local=*/true); 246 if (llvm::Function *atexitFn = dyn_cast<llvm::Function>(atexit)) 247 atexitFn->setDoesNotThrow(); 248 249 EmitNounwindRuntimeCall(atexit, dtorStub); 250 } 251 252 void CodeGenFunction::EmitCXXGuardedInit(const VarDecl &D, 253 llvm::GlobalVariable *DeclPtr, 254 bool PerformInit) { 255 // If we've been asked to forbid guard variables, emit an error now. 256 // This diagnostic is hard-coded for Darwin's use case; we can find 257 // better phrasing if someone else needs it. 258 if (CGM.getCodeGenOpts().ForbidGuardVariables) 259 CGM.Error(D.getLocation(), 260 "this initialization requires a guard variable, which " 261 "the kernel does not support"); 262 263 CGM.getCXXABI().EmitGuardedInit(*this, D, DeclPtr, PerformInit); 264 } 265 266 void CodeGenFunction::EmitCXXGuardedInitBranch(llvm::Value *NeedsInit, 267 llvm::BasicBlock *InitBlock, 268 llvm::BasicBlock *NoInitBlock, 269 GuardKind Kind, 270 const VarDecl *D) { 271 assert((Kind == GuardKind::TlsGuard || D) && "no guarded variable"); 272 273 // A guess at how many times we will enter the initialization of a 274 // variable, depending on the kind of variable. 275 static const uint64_t InitsPerTLSVar = 1024; 276 static const uint64_t InitsPerLocalVar = 1024 * 1024; 277 278 llvm::MDNode *Weights; 279 if (Kind == GuardKind::VariableGuard && !D->isLocalVarDecl()) { 280 // For non-local variables, don't apply any weighting for now. Due to our 281 // use of COMDATs, we expect there to be at most one initialization of the 282 // variable per DSO, but we have no way to know how many DSOs will try to 283 // initialize the variable. 284 Weights = nullptr; 285 } else { 286 uint64_t NumInits; 287 // FIXME: For the TLS case, collect and use profiling information to 288 // determine a more accurate brach weight. 289 if (Kind == GuardKind::TlsGuard || D->getTLSKind()) 290 NumInits = InitsPerTLSVar; 291 else 292 NumInits = InitsPerLocalVar; 293 294 // The probability of us entering the initializer is 295 // 1 / (total number of times we attempt to initialize the variable). 296 llvm::MDBuilder MDHelper(CGM.getLLVMContext()); 297 Weights = MDHelper.createBranchWeights(1, NumInits - 1); 298 } 299 300 Builder.CreateCondBr(NeedsInit, InitBlock, NoInitBlock, Weights); 301 } 302 303 llvm::Function *CodeGenModule::CreateGlobalInitOrDestructFunction( 304 llvm::FunctionType *FTy, const Twine &Name, const CGFunctionInfo &FI, 305 SourceLocation Loc, bool TLS) { 306 llvm::Function *Fn = 307 llvm::Function::Create(FTy, llvm::GlobalValue::InternalLinkage, 308 Name, &getModule()); 309 if (!getLangOpts().AppleKext && !TLS) { 310 // Set the section if needed. 311 if (const char *Section = getTarget().getStaticInitSectionSpecifier()) 312 Fn->setSection(Section); 313 } 314 315 SetInternalFunctionAttributes(GlobalDecl(), Fn, FI); 316 317 Fn->setCallingConv(getRuntimeCC()); 318 319 if (!getLangOpts().Exceptions) 320 Fn->setDoesNotThrow(); 321 322 if (getLangOpts().Sanitize.has(SanitizerKind::Address) && 323 !isInSanitizerBlacklist(SanitizerKind::Address, Fn, Loc)) 324 Fn->addFnAttr(llvm::Attribute::SanitizeAddress); 325 326 if (getLangOpts().Sanitize.has(SanitizerKind::KernelAddress) && 327 !isInSanitizerBlacklist(SanitizerKind::KernelAddress, Fn, Loc)) 328 Fn->addFnAttr(llvm::Attribute::SanitizeAddress); 329 330 if (getLangOpts().Sanitize.has(SanitizerKind::HWAddress) && 331 !isInSanitizerBlacklist(SanitizerKind::HWAddress, Fn, Loc)) 332 Fn->addFnAttr(llvm::Attribute::SanitizeHWAddress); 333 334 if (getLangOpts().Sanitize.has(SanitizerKind::Thread) && 335 !isInSanitizerBlacklist(SanitizerKind::Thread, Fn, Loc)) 336 Fn->addFnAttr(llvm::Attribute::SanitizeThread); 337 338 if (getLangOpts().Sanitize.has(SanitizerKind::Memory) && 339 !isInSanitizerBlacklist(SanitizerKind::Memory, Fn, Loc)) 340 Fn->addFnAttr(llvm::Attribute::SanitizeMemory); 341 342 if (getLangOpts().Sanitize.has(SanitizerKind::SafeStack) && 343 !isInSanitizerBlacklist(SanitizerKind::SafeStack, Fn, Loc)) 344 Fn->addFnAttr(llvm::Attribute::SafeStack); 345 346 return Fn; 347 } 348 349 /// Create a global pointer to a function that will initialize a global 350 /// variable. The user has requested that this pointer be emitted in a specific 351 /// section. 352 void CodeGenModule::EmitPointerToInitFunc(const VarDecl *D, 353 llvm::GlobalVariable *GV, 354 llvm::Function *InitFunc, 355 InitSegAttr *ISA) { 356 llvm::GlobalVariable *PtrArray = new llvm::GlobalVariable( 357 TheModule, InitFunc->getType(), /*isConstant=*/true, 358 llvm::GlobalValue::PrivateLinkage, InitFunc, "__cxx_init_fn_ptr"); 359 PtrArray->setSection(ISA->getSection()); 360 addUsedGlobal(PtrArray); 361 362 // If the GV is already in a comdat group, then we have to join it. 363 if (llvm::Comdat *C = GV->getComdat()) 364 PtrArray->setComdat(C); 365 } 366 367 void 368 CodeGenModule::EmitCXXGlobalVarDeclInitFunc(const VarDecl *D, 369 llvm::GlobalVariable *Addr, 370 bool PerformInit) { 371 372 // According to E.2.3.1 in CUDA-7.5 Programming guide: __device__, 373 // __constant__ and __shared__ variables defined in namespace scope, 374 // that are of class type, cannot have a non-empty constructor. All 375 // the checks have been done in Sema by now. Whatever initializers 376 // are allowed are empty and we just need to ignore them here. 377 if (getLangOpts().CUDA && getLangOpts().CUDAIsDevice && 378 (D->hasAttr<CUDADeviceAttr>() || D->hasAttr<CUDAConstantAttr>() || 379 D->hasAttr<CUDASharedAttr>())) 380 return; 381 382 // Check if we've already initialized this decl. 383 auto I = DelayedCXXInitPosition.find(D); 384 if (I != DelayedCXXInitPosition.end() && I->second == ~0U) 385 return; 386 387 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 388 SmallString<256> FnName; 389 { 390 llvm::raw_svector_ostream Out(FnName); 391 getCXXABI().getMangleContext().mangleDynamicInitializer(D, Out); 392 } 393 394 // Create a variable initialization function. 395 llvm::Function *Fn = 396 CreateGlobalInitOrDestructFunction(FTy, FnName.str(), 397 getTypes().arrangeNullaryFunction(), 398 D->getLocation()); 399 400 auto *ISA = D->getAttr<InitSegAttr>(); 401 CodeGenFunction(*this).GenerateCXXGlobalVarDeclInitFunc(Fn, D, Addr, 402 PerformInit); 403 404 llvm::GlobalVariable *COMDATKey = 405 supportsCOMDAT() && D->isExternallyVisible() ? Addr : nullptr; 406 407 if (D->getTLSKind()) { 408 // FIXME: Should we support init_priority for thread_local? 409 // FIXME: We only need to register one __cxa_thread_atexit function for the 410 // entire TU. 411 CXXThreadLocalInits.push_back(Fn); 412 CXXThreadLocalInitVars.push_back(D); 413 } else if (PerformInit && ISA) { 414 EmitPointerToInitFunc(D, Addr, Fn, ISA); 415 } else if (auto *IPA = D->getAttr<InitPriorityAttr>()) { 416 OrderGlobalInits Key(IPA->getPriority(), PrioritizedCXXGlobalInits.size()); 417 PrioritizedCXXGlobalInits.push_back(std::make_pair(Key, Fn)); 418 } else if (isTemplateInstantiation(D->getTemplateSpecializationKind())) { 419 // C++ [basic.start.init]p2: 420 // Definitions of explicitly specialized class template static data 421 // members have ordered initialization. Other class template static data 422 // members (i.e., implicitly or explicitly instantiated specializations) 423 // have unordered initialization. 424 // 425 // As a consequence, we can put them into their own llvm.global_ctors entry. 426 // 427 // If the global is externally visible, put the initializer into a COMDAT 428 // group with the global being initialized. On most platforms, this is a 429 // minor startup time optimization. In the MS C++ ABI, there are no guard 430 // variables, so this COMDAT key is required for correctness. 431 AddGlobalCtor(Fn, 65535, COMDATKey); 432 } else if (D->hasAttr<SelectAnyAttr>()) { 433 // SelectAny globals will be comdat-folded. Put the initializer into a 434 // COMDAT group associated with the global, so the initializers get folded 435 // too. 436 AddGlobalCtor(Fn, 65535, COMDATKey); 437 } else { 438 I = DelayedCXXInitPosition.find(D); // Re-do lookup in case of re-hash. 439 if (I == DelayedCXXInitPosition.end()) { 440 CXXGlobalInits.push_back(Fn); 441 } else if (I->second != ~0U) { 442 assert(I->second < CXXGlobalInits.size() && 443 CXXGlobalInits[I->second] == nullptr); 444 CXXGlobalInits[I->second] = Fn; 445 } 446 } 447 448 // Remember that we already emitted the initializer for this global. 449 DelayedCXXInitPosition[D] = ~0U; 450 } 451 452 void CodeGenModule::EmitCXXThreadLocalInitFunc() { 453 getCXXABI().EmitThreadLocalInitFuncs( 454 *this, CXXThreadLocals, CXXThreadLocalInits, CXXThreadLocalInitVars); 455 456 CXXThreadLocalInits.clear(); 457 CXXThreadLocalInitVars.clear(); 458 CXXThreadLocals.clear(); 459 } 460 461 void 462 CodeGenModule::EmitCXXGlobalInitFunc() { 463 while (!CXXGlobalInits.empty() && !CXXGlobalInits.back()) 464 CXXGlobalInits.pop_back(); 465 466 if (CXXGlobalInits.empty() && PrioritizedCXXGlobalInits.empty()) 467 return; 468 469 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 470 const CGFunctionInfo &FI = getTypes().arrangeNullaryFunction(); 471 472 // Create our global initialization function. 473 if (!PrioritizedCXXGlobalInits.empty()) { 474 SmallVector<llvm::Function *, 8> LocalCXXGlobalInits; 475 llvm::array_pod_sort(PrioritizedCXXGlobalInits.begin(), 476 PrioritizedCXXGlobalInits.end()); 477 // Iterate over "chunks" of ctors with same priority and emit each chunk 478 // into separate function. Note - everything is sorted first by priority, 479 // second - by lex order, so we emit ctor functions in proper order. 480 for (SmallVectorImpl<GlobalInitData >::iterator 481 I = PrioritizedCXXGlobalInits.begin(), 482 E = PrioritizedCXXGlobalInits.end(); I != E; ) { 483 SmallVectorImpl<GlobalInitData >::iterator 484 PrioE = std::upper_bound(I + 1, E, *I, GlobalInitPriorityCmp()); 485 486 LocalCXXGlobalInits.clear(); 487 unsigned Priority = I->first.priority; 488 // Compute the function suffix from priority. Prepend with zeroes to make 489 // sure the function names are also ordered as priorities. 490 std::string PrioritySuffix = llvm::utostr(Priority); 491 // Priority is always <= 65535 (enforced by sema). 492 PrioritySuffix = std::string(6-PrioritySuffix.size(), '0')+PrioritySuffix; 493 llvm::Function *Fn = CreateGlobalInitOrDestructFunction( 494 FTy, "_GLOBAL__I_" + PrioritySuffix, FI); 495 496 for (; I < PrioE; ++I) 497 LocalCXXGlobalInits.push_back(I->second); 498 499 CodeGenFunction(*this).GenerateCXXGlobalInitFunc(Fn, LocalCXXGlobalInits); 500 AddGlobalCtor(Fn, Priority); 501 } 502 PrioritizedCXXGlobalInits.clear(); 503 } 504 505 // Include the filename in the symbol name. Including "sub_" matches gcc and 506 // makes sure these symbols appear lexicographically behind the symbols with 507 // priority emitted above. 508 SmallString<128> FileName = llvm::sys::path::filename(getModule().getName()); 509 if (FileName.empty()) 510 FileName = "<null>"; 511 512 for (size_t i = 0; i < FileName.size(); ++i) { 513 // Replace everything that's not [a-zA-Z0-9._] with a _. This set happens 514 // to be the set of C preprocessing numbers. 515 if (!isPreprocessingNumberBody(FileName[i])) 516 FileName[i] = '_'; 517 } 518 519 llvm::Function *Fn = CreateGlobalInitOrDestructFunction( 520 FTy, llvm::Twine("_GLOBAL__sub_I_", FileName), FI); 521 522 CodeGenFunction(*this).GenerateCXXGlobalInitFunc(Fn, CXXGlobalInits); 523 AddGlobalCtor(Fn); 524 525 CXXGlobalInits.clear(); 526 } 527 528 void CodeGenModule::EmitCXXGlobalDtorFunc() { 529 if (CXXGlobalDtors.empty()) 530 return; 531 532 llvm::FunctionType *FTy = llvm::FunctionType::get(VoidTy, false); 533 534 // Create our global destructor function. 535 const CGFunctionInfo &FI = getTypes().arrangeNullaryFunction(); 536 llvm::Function *Fn = 537 CreateGlobalInitOrDestructFunction(FTy, "_GLOBAL__D_a", FI); 538 539 CodeGenFunction(*this).GenerateCXXGlobalDtorsFunc(Fn, CXXGlobalDtors); 540 AddGlobalDtor(Fn); 541 } 542 543 /// Emit the code necessary to initialize the given global variable. 544 void CodeGenFunction::GenerateCXXGlobalVarDeclInitFunc(llvm::Function *Fn, 545 const VarDecl *D, 546 llvm::GlobalVariable *Addr, 547 bool PerformInit) { 548 // Check if we need to emit debug info for variable initializer. 549 if (D->hasAttr<NoDebugAttr>()) 550 DebugInfo = nullptr; // disable debug info indefinitely for this function 551 552 CurEHLocation = D->getLocStart(); 553 554 StartFunction(GlobalDecl(D), getContext().VoidTy, Fn, 555 getTypes().arrangeNullaryFunction(), 556 FunctionArgList(), D->getLocation(), 557 D->getInit()->getExprLoc()); 558 559 // Use guarded initialization if the global variable is weak. This 560 // occurs for, e.g., instantiated static data members and 561 // definitions explicitly marked weak. 562 if (Addr->hasWeakLinkage() || Addr->hasLinkOnceLinkage()) { 563 EmitCXXGuardedInit(*D, Addr, PerformInit); 564 } else { 565 EmitCXXGlobalVarDeclInit(*D, Addr, PerformInit); 566 } 567 568 FinishFunction(); 569 } 570 571 void 572 CodeGenFunction::GenerateCXXGlobalInitFunc(llvm::Function *Fn, 573 ArrayRef<llvm::Function *> Decls, 574 Address Guard) { 575 { 576 auto NL = ApplyDebugLocation::CreateEmpty(*this); 577 StartFunction(GlobalDecl(), getContext().VoidTy, Fn, 578 getTypes().arrangeNullaryFunction(), FunctionArgList()); 579 // Emit an artificial location for this function. 580 auto AL = ApplyDebugLocation::CreateArtificial(*this); 581 582 llvm::BasicBlock *ExitBlock = nullptr; 583 if (Guard.isValid()) { 584 // If we have a guard variable, check whether we've already performed 585 // these initializations. This happens for TLS initialization functions. 586 llvm::Value *GuardVal = Builder.CreateLoad(Guard); 587 llvm::Value *Uninit = Builder.CreateIsNull(GuardVal, 588 "guard.uninitialized"); 589 llvm::BasicBlock *InitBlock = createBasicBlock("init"); 590 ExitBlock = createBasicBlock("exit"); 591 EmitCXXGuardedInitBranch(Uninit, InitBlock, ExitBlock, 592 GuardKind::TlsGuard, nullptr); 593 EmitBlock(InitBlock); 594 // Mark as initialized before initializing anything else. If the 595 // initializers use previously-initialized thread_local vars, that's 596 // probably supposed to be OK, but the standard doesn't say. 597 Builder.CreateStore(llvm::ConstantInt::get(GuardVal->getType(),1), Guard); 598 } 599 600 RunCleanupsScope Scope(*this); 601 602 // When building in Objective-C++ ARC mode, create an autorelease pool 603 // around the global initializers. 604 if (getLangOpts().ObjCAutoRefCount && getLangOpts().CPlusPlus) { 605 llvm::Value *token = EmitObjCAutoreleasePoolPush(); 606 EmitObjCAutoreleasePoolCleanup(token); 607 } 608 609 for (unsigned i = 0, e = Decls.size(); i != e; ++i) 610 if (Decls[i]) 611 EmitRuntimeCall(Decls[i]); 612 613 Scope.ForceCleanup(); 614 615 if (ExitBlock) { 616 Builder.CreateBr(ExitBlock); 617 EmitBlock(ExitBlock); 618 } 619 } 620 621 FinishFunction(); 622 } 623 624 void CodeGenFunction::GenerateCXXGlobalDtorsFunc( 625 llvm::Function *Fn, 626 const std::vector<std::pair<llvm::WeakTrackingVH, llvm::Constant *>> 627 &DtorsAndObjects) { 628 { 629 auto NL = ApplyDebugLocation::CreateEmpty(*this); 630 StartFunction(GlobalDecl(), getContext().VoidTy, Fn, 631 getTypes().arrangeNullaryFunction(), FunctionArgList()); 632 // Emit an artificial location for this function. 633 auto AL = ApplyDebugLocation::CreateArtificial(*this); 634 635 // Emit the dtors, in reverse order from construction. 636 for (unsigned i = 0, e = DtorsAndObjects.size(); i != e; ++i) { 637 llvm::Value *Callee = DtorsAndObjects[e - i - 1].first; 638 llvm::CallInst *CI = Builder.CreateCall(Callee, 639 DtorsAndObjects[e - i - 1].second); 640 // Make sure the call and the callee agree on calling convention. 641 if (llvm::Function *F = dyn_cast<llvm::Function>(Callee)) 642 CI->setCallingConv(F->getCallingConv()); 643 } 644 } 645 646 FinishFunction(); 647 } 648 649 /// generateDestroyHelper - Generates a helper function which, when 650 /// invoked, destroys the given object. The address of the object 651 /// should be in global memory. 652 llvm::Function *CodeGenFunction::generateDestroyHelper( 653 Address addr, QualType type, Destroyer *destroyer, 654 bool useEHCleanupForArray, const VarDecl *VD) { 655 FunctionArgList args; 656 ImplicitParamDecl Dst(getContext(), getContext().VoidPtrTy, 657 ImplicitParamDecl::Other); 658 args.push_back(&Dst); 659 660 const CGFunctionInfo &FI = 661 CGM.getTypes().arrangeBuiltinFunctionDeclaration(getContext().VoidTy, args); 662 llvm::FunctionType *FTy = CGM.getTypes().GetFunctionType(FI); 663 llvm::Function *fn = CGM.CreateGlobalInitOrDestructFunction( 664 FTy, "__cxx_global_array_dtor", FI, VD->getLocation()); 665 666 CurEHLocation = VD->getLocStart(); 667 668 StartFunction(VD, getContext().VoidTy, fn, FI, args); 669 670 emitDestroy(addr, type, destroyer, useEHCleanupForArray); 671 672 FinishFunction(); 673 674 return fn; 675 } 676