1 //===- ASTContext.cpp - Context to hold long-lived AST nodes --------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file implements the ASTContext interface. 10 // 11 //===----------------------------------------------------------------------===// 12 13 #include "clang/AST/ASTContext.h" 14 #include "CXXABI.h" 15 #include "Interp/Context.h" 16 #include "clang/AST/APValue.h" 17 #include "clang/AST/ASTConcept.h" 18 #include "clang/AST/ASTMutationListener.h" 19 #include "clang/AST/ASTTypeTraits.h" 20 #include "clang/AST/Attr.h" 21 #include "clang/AST/AttrIterator.h" 22 #include "clang/AST/CharUnits.h" 23 #include "clang/AST/Comment.h" 24 #include "clang/AST/Decl.h" 25 #include "clang/AST/DeclBase.h" 26 #include "clang/AST/DeclCXX.h" 27 #include "clang/AST/DeclContextInternals.h" 28 #include "clang/AST/DeclObjC.h" 29 #include "clang/AST/DeclOpenMP.h" 30 #include "clang/AST/DeclTemplate.h" 31 #include "clang/AST/DeclarationName.h" 32 #include "clang/AST/DependenceFlags.h" 33 #include "clang/AST/Expr.h" 34 #include "clang/AST/ExprCXX.h" 35 #include "clang/AST/ExprConcepts.h" 36 #include "clang/AST/ExternalASTSource.h" 37 #include "clang/AST/Mangle.h" 38 #include "clang/AST/MangleNumberingContext.h" 39 #include "clang/AST/NestedNameSpecifier.h" 40 #include "clang/AST/ParentMapContext.h" 41 #include "clang/AST/RawCommentList.h" 42 #include "clang/AST/RecordLayout.h" 43 #include "clang/AST/Stmt.h" 44 #include "clang/AST/TemplateBase.h" 45 #include "clang/AST/TemplateName.h" 46 #include "clang/AST/Type.h" 47 #include "clang/AST/TypeLoc.h" 48 #include "clang/AST/UnresolvedSet.h" 49 #include "clang/AST/VTableBuilder.h" 50 #include "clang/Basic/AddressSpaces.h" 51 #include "clang/Basic/Builtins.h" 52 #include "clang/Basic/CommentOptions.h" 53 #include "clang/Basic/ExceptionSpecificationType.h" 54 #include "clang/Basic/IdentifierTable.h" 55 #include "clang/Basic/LLVM.h" 56 #include "clang/Basic/LangOptions.h" 57 #include "clang/Basic/Linkage.h" 58 #include "clang/Basic/Module.h" 59 #include "clang/Basic/NoSanitizeList.h" 60 #include "clang/Basic/ObjCRuntime.h" 61 #include "clang/Basic/SourceLocation.h" 62 #include "clang/Basic/SourceManager.h" 63 #include "clang/Basic/Specifiers.h" 64 #include "clang/Basic/TargetCXXABI.h" 65 #include "clang/Basic/TargetInfo.h" 66 #include "clang/Basic/XRayLists.h" 67 #include "llvm/ADT/APFixedPoint.h" 68 #include "llvm/ADT/APInt.h" 69 #include "llvm/ADT/APSInt.h" 70 #include "llvm/ADT/ArrayRef.h" 71 #include "llvm/ADT/DenseMap.h" 72 #include "llvm/ADT/DenseSet.h" 73 #include "llvm/ADT/FoldingSet.h" 74 #include "llvm/ADT/None.h" 75 #include "llvm/ADT/Optional.h" 76 #include "llvm/ADT/PointerUnion.h" 77 #include "llvm/ADT/STLExtras.h" 78 #include "llvm/ADT/SmallPtrSet.h" 79 #include "llvm/ADT/SmallVector.h" 80 #include "llvm/ADT/StringExtras.h" 81 #include "llvm/ADT/StringRef.h" 82 #include "llvm/ADT/Triple.h" 83 #include "llvm/Support/Capacity.h" 84 #include "llvm/Support/Casting.h" 85 #include "llvm/Support/Compiler.h" 86 #include "llvm/Support/ErrorHandling.h" 87 #include "llvm/Support/MD5.h" 88 #include "llvm/Support/MathExtras.h" 89 #include "llvm/Support/raw_ostream.h" 90 #include <algorithm> 91 #include <cassert> 92 #include <cstddef> 93 #include <cstdint> 94 #include <cstdlib> 95 #include <map> 96 #include <memory> 97 #include <string> 98 #include <tuple> 99 #include <utility> 100 101 using namespace clang; 102 103 enum FloatingRank { 104 BFloat16Rank, 105 Float16Rank, 106 HalfRank, 107 FloatRank, 108 DoubleRank, 109 LongDoubleRank, 110 Float128Rank, 111 Ibm128Rank 112 }; 113 114 /// \returns location that is relevant when searching for Doc comments related 115 /// to \p D. 116 static SourceLocation getDeclLocForCommentSearch(const Decl *D, 117 SourceManager &SourceMgr) { 118 assert(D); 119 120 // User can not attach documentation to implicit declarations. 121 if (D->isImplicit()) 122 return {}; 123 124 // User can not attach documentation to implicit instantiations. 125 if (const auto *FD = dyn_cast<FunctionDecl>(D)) { 126 if (FD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation) 127 return {}; 128 } 129 130 if (const auto *VD = dyn_cast<VarDecl>(D)) { 131 if (VD->isStaticDataMember() && 132 VD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation) 133 return {}; 134 } 135 136 if (const auto *CRD = dyn_cast<CXXRecordDecl>(D)) { 137 if (CRD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation) 138 return {}; 139 } 140 141 if (const auto *CTSD = dyn_cast<ClassTemplateSpecializationDecl>(D)) { 142 TemplateSpecializationKind TSK = CTSD->getSpecializationKind(); 143 if (TSK == TSK_ImplicitInstantiation || 144 TSK == TSK_Undeclared) 145 return {}; 146 } 147 148 if (const auto *ED = dyn_cast<EnumDecl>(D)) { 149 if (ED->getTemplateSpecializationKind() == TSK_ImplicitInstantiation) 150 return {}; 151 } 152 if (const auto *TD = dyn_cast<TagDecl>(D)) { 153 // When tag declaration (but not definition!) is part of the 154 // decl-specifier-seq of some other declaration, it doesn't get comment 155 if (TD->isEmbeddedInDeclarator() && !TD->isCompleteDefinition()) 156 return {}; 157 } 158 // TODO: handle comments for function parameters properly. 159 if (isa<ParmVarDecl>(D)) 160 return {}; 161 162 // TODO: we could look up template parameter documentation in the template 163 // documentation. 164 if (isa<TemplateTypeParmDecl>(D) || 165 isa<NonTypeTemplateParmDecl>(D) || 166 isa<TemplateTemplateParmDecl>(D)) 167 return {}; 168 169 // Find declaration location. 170 // For Objective-C declarations we generally don't expect to have multiple 171 // declarators, thus use declaration starting location as the "declaration 172 // location". 173 // For all other declarations multiple declarators are used quite frequently, 174 // so we use the location of the identifier as the "declaration location". 175 if (isa<ObjCMethodDecl>(D) || isa<ObjCContainerDecl>(D) || 176 isa<ObjCPropertyDecl>(D) || 177 isa<RedeclarableTemplateDecl>(D) || 178 isa<ClassTemplateSpecializationDecl>(D) || 179 // Allow association with Y across {} in `typedef struct X {} Y`. 180 isa<TypedefDecl>(D)) 181 return D->getBeginLoc(); 182 183 const SourceLocation DeclLoc = D->getLocation(); 184 if (DeclLoc.isMacroID()) { 185 if (isa<TypedefDecl>(D)) { 186 // If location of the typedef name is in a macro, it is because being 187 // declared via a macro. Try using declaration's starting location as 188 // the "declaration location". 189 return D->getBeginLoc(); 190 } 191 192 if (const auto *TD = dyn_cast<TagDecl>(D)) { 193 // If location of the tag decl is inside a macro, but the spelling of 194 // the tag name comes from a macro argument, it looks like a special 195 // macro like NS_ENUM is being used to define the tag decl. In that 196 // case, adjust the source location to the expansion loc so that we can 197 // attach the comment to the tag decl. 198 if (SourceMgr.isMacroArgExpansion(DeclLoc) && TD->isCompleteDefinition()) 199 return SourceMgr.getExpansionLoc(DeclLoc); 200 } 201 } 202 203 return DeclLoc; 204 } 205 206 RawComment *ASTContext::getRawCommentForDeclNoCacheImpl( 207 const Decl *D, const SourceLocation RepresentativeLocForDecl, 208 const std::map<unsigned, RawComment *> &CommentsInTheFile) const { 209 // If the declaration doesn't map directly to a location in a file, we 210 // can't find the comment. 211 if (RepresentativeLocForDecl.isInvalid() || 212 !RepresentativeLocForDecl.isFileID()) 213 return nullptr; 214 215 // If there are no comments anywhere, we won't find anything. 216 if (CommentsInTheFile.empty()) 217 return nullptr; 218 219 // Decompose the location for the declaration and find the beginning of the 220 // file buffer. 221 const std::pair<FileID, unsigned> DeclLocDecomp = 222 SourceMgr.getDecomposedLoc(RepresentativeLocForDecl); 223 224 // Slow path. 225 auto OffsetCommentBehindDecl = 226 CommentsInTheFile.lower_bound(DeclLocDecomp.second); 227 228 // First check whether we have a trailing comment. 229 if (OffsetCommentBehindDecl != CommentsInTheFile.end()) { 230 RawComment *CommentBehindDecl = OffsetCommentBehindDecl->second; 231 if ((CommentBehindDecl->isDocumentation() || 232 LangOpts.CommentOpts.ParseAllComments) && 233 CommentBehindDecl->isTrailingComment() && 234 (isa<FieldDecl>(D) || isa<EnumConstantDecl>(D) || isa<VarDecl>(D) || 235 isa<ObjCMethodDecl>(D) || isa<ObjCPropertyDecl>(D))) { 236 237 // Check that Doxygen trailing comment comes after the declaration, starts 238 // on the same line and in the same file as the declaration. 239 if (SourceMgr.getLineNumber(DeclLocDecomp.first, DeclLocDecomp.second) == 240 Comments.getCommentBeginLine(CommentBehindDecl, DeclLocDecomp.first, 241 OffsetCommentBehindDecl->first)) { 242 return CommentBehindDecl; 243 } 244 } 245 } 246 247 // The comment just after the declaration was not a trailing comment. 248 // Let's look at the previous comment. 249 if (OffsetCommentBehindDecl == CommentsInTheFile.begin()) 250 return nullptr; 251 252 auto OffsetCommentBeforeDecl = --OffsetCommentBehindDecl; 253 RawComment *CommentBeforeDecl = OffsetCommentBeforeDecl->second; 254 255 // Check that we actually have a non-member Doxygen comment. 256 if (!(CommentBeforeDecl->isDocumentation() || 257 LangOpts.CommentOpts.ParseAllComments) || 258 CommentBeforeDecl->isTrailingComment()) 259 return nullptr; 260 261 // Decompose the end of the comment. 262 const unsigned CommentEndOffset = 263 Comments.getCommentEndOffset(CommentBeforeDecl); 264 265 // Get the corresponding buffer. 266 bool Invalid = false; 267 const char *Buffer = SourceMgr.getBufferData(DeclLocDecomp.first, 268 &Invalid).data(); 269 if (Invalid) 270 return nullptr; 271 272 // Extract text between the comment and declaration. 273 StringRef Text(Buffer + CommentEndOffset, 274 DeclLocDecomp.second - CommentEndOffset); 275 276 // There should be no other declarations or preprocessor directives between 277 // comment and declaration. 278 if (Text.find_first_of(";{}#@") != StringRef::npos) 279 return nullptr; 280 281 return CommentBeforeDecl; 282 } 283 284 RawComment *ASTContext::getRawCommentForDeclNoCache(const Decl *D) const { 285 const SourceLocation DeclLoc = getDeclLocForCommentSearch(D, SourceMgr); 286 287 // If the declaration doesn't map directly to a location in a file, we 288 // can't find the comment. 289 if (DeclLoc.isInvalid() || !DeclLoc.isFileID()) 290 return nullptr; 291 292 if (ExternalSource && !CommentsLoaded) { 293 ExternalSource->ReadComments(); 294 CommentsLoaded = true; 295 } 296 297 if (Comments.empty()) 298 return nullptr; 299 300 const FileID File = SourceMgr.getDecomposedLoc(DeclLoc).first; 301 const auto CommentsInThisFile = Comments.getCommentsInFile(File); 302 if (!CommentsInThisFile || CommentsInThisFile->empty()) 303 return nullptr; 304 305 return getRawCommentForDeclNoCacheImpl(D, DeclLoc, *CommentsInThisFile); 306 } 307 308 void ASTContext::addComment(const RawComment &RC) { 309 assert(LangOpts.RetainCommentsFromSystemHeaders || 310 !SourceMgr.isInSystemHeader(RC.getSourceRange().getBegin())); 311 Comments.addComment(RC, LangOpts.CommentOpts, BumpAlloc); 312 } 313 314 /// If we have a 'templated' declaration for a template, adjust 'D' to 315 /// refer to the actual template. 316 /// If we have an implicit instantiation, adjust 'D' to refer to template. 317 static const Decl &adjustDeclToTemplate(const Decl &D) { 318 if (const auto *FD = dyn_cast<FunctionDecl>(&D)) { 319 // Is this function declaration part of a function template? 320 if (const FunctionTemplateDecl *FTD = FD->getDescribedFunctionTemplate()) 321 return *FTD; 322 323 // Nothing to do if function is not an implicit instantiation. 324 if (FD->getTemplateSpecializationKind() != TSK_ImplicitInstantiation) 325 return D; 326 327 // Function is an implicit instantiation of a function template? 328 if (const FunctionTemplateDecl *FTD = FD->getPrimaryTemplate()) 329 return *FTD; 330 331 // Function is instantiated from a member definition of a class template? 332 if (const FunctionDecl *MemberDecl = 333 FD->getInstantiatedFromMemberFunction()) 334 return *MemberDecl; 335 336 return D; 337 } 338 if (const auto *VD = dyn_cast<VarDecl>(&D)) { 339 // Static data member is instantiated from a member definition of a class 340 // template? 341 if (VD->isStaticDataMember()) 342 if (const VarDecl *MemberDecl = VD->getInstantiatedFromStaticDataMember()) 343 return *MemberDecl; 344 345 return D; 346 } 347 if (const auto *CRD = dyn_cast<CXXRecordDecl>(&D)) { 348 // Is this class declaration part of a class template? 349 if (const ClassTemplateDecl *CTD = CRD->getDescribedClassTemplate()) 350 return *CTD; 351 352 // Class is an implicit instantiation of a class template or partial 353 // specialization? 354 if (const auto *CTSD = dyn_cast<ClassTemplateSpecializationDecl>(CRD)) { 355 if (CTSD->getSpecializationKind() != TSK_ImplicitInstantiation) 356 return D; 357 llvm::PointerUnion<ClassTemplateDecl *, 358 ClassTemplatePartialSpecializationDecl *> 359 PU = CTSD->getSpecializedTemplateOrPartial(); 360 return PU.is<ClassTemplateDecl *>() 361 ? *static_cast<const Decl *>(PU.get<ClassTemplateDecl *>()) 362 : *static_cast<const Decl *>( 363 PU.get<ClassTemplatePartialSpecializationDecl *>()); 364 } 365 366 // Class is instantiated from a member definition of a class template? 367 if (const MemberSpecializationInfo *Info = 368 CRD->getMemberSpecializationInfo()) 369 return *Info->getInstantiatedFrom(); 370 371 return D; 372 } 373 if (const auto *ED = dyn_cast<EnumDecl>(&D)) { 374 // Enum is instantiated from a member definition of a class template? 375 if (const EnumDecl *MemberDecl = ED->getInstantiatedFromMemberEnum()) 376 return *MemberDecl; 377 378 return D; 379 } 380 // FIXME: Adjust alias templates? 381 return D; 382 } 383 384 const RawComment *ASTContext::getRawCommentForAnyRedecl( 385 const Decl *D, 386 const Decl **OriginalDecl) const { 387 if (!D) { 388 if (OriginalDecl) 389 OriginalDecl = nullptr; 390 return nullptr; 391 } 392 393 D = &adjustDeclToTemplate(*D); 394 395 // Any comment directly attached to D? 396 { 397 auto DeclComment = DeclRawComments.find(D); 398 if (DeclComment != DeclRawComments.end()) { 399 if (OriginalDecl) 400 *OriginalDecl = D; 401 return DeclComment->second; 402 } 403 } 404 405 // Any comment attached to any redeclaration of D? 406 const Decl *CanonicalD = D->getCanonicalDecl(); 407 if (!CanonicalD) 408 return nullptr; 409 410 { 411 auto RedeclComment = RedeclChainComments.find(CanonicalD); 412 if (RedeclComment != RedeclChainComments.end()) { 413 if (OriginalDecl) 414 *OriginalDecl = RedeclComment->second; 415 auto CommentAtRedecl = DeclRawComments.find(RedeclComment->second); 416 assert(CommentAtRedecl != DeclRawComments.end() && 417 "This decl is supposed to have comment attached."); 418 return CommentAtRedecl->second; 419 } 420 } 421 422 // Any redeclarations of D that we haven't checked for comments yet? 423 // We can't use DenseMap::iterator directly since it'd get invalid. 424 auto LastCheckedRedecl = [this, CanonicalD]() -> const Decl * { 425 auto LookupRes = CommentlessRedeclChains.find(CanonicalD); 426 if (LookupRes != CommentlessRedeclChains.end()) 427 return LookupRes->second; 428 return nullptr; 429 }(); 430 431 for (const auto Redecl : D->redecls()) { 432 assert(Redecl); 433 // Skip all redeclarations that have been checked previously. 434 if (LastCheckedRedecl) { 435 if (LastCheckedRedecl == Redecl) { 436 LastCheckedRedecl = nullptr; 437 } 438 continue; 439 } 440 const RawComment *RedeclComment = getRawCommentForDeclNoCache(Redecl); 441 if (RedeclComment) { 442 cacheRawCommentForDecl(*Redecl, *RedeclComment); 443 if (OriginalDecl) 444 *OriginalDecl = Redecl; 445 return RedeclComment; 446 } 447 CommentlessRedeclChains[CanonicalD] = Redecl; 448 } 449 450 if (OriginalDecl) 451 *OriginalDecl = nullptr; 452 return nullptr; 453 } 454 455 void ASTContext::cacheRawCommentForDecl(const Decl &OriginalD, 456 const RawComment &Comment) const { 457 assert(Comment.isDocumentation() || LangOpts.CommentOpts.ParseAllComments); 458 DeclRawComments.try_emplace(&OriginalD, &Comment); 459 const Decl *const CanonicalDecl = OriginalD.getCanonicalDecl(); 460 RedeclChainComments.try_emplace(CanonicalDecl, &OriginalD); 461 CommentlessRedeclChains.erase(CanonicalDecl); 462 } 463 464 static void addRedeclaredMethods(const ObjCMethodDecl *ObjCMethod, 465 SmallVectorImpl<const NamedDecl *> &Redeclared) { 466 const DeclContext *DC = ObjCMethod->getDeclContext(); 467 if (const auto *IMD = dyn_cast<ObjCImplDecl>(DC)) { 468 const ObjCInterfaceDecl *ID = IMD->getClassInterface(); 469 if (!ID) 470 return; 471 // Add redeclared method here. 472 for (const auto *Ext : ID->known_extensions()) { 473 if (ObjCMethodDecl *RedeclaredMethod = 474 Ext->getMethod(ObjCMethod->getSelector(), 475 ObjCMethod->isInstanceMethod())) 476 Redeclared.push_back(RedeclaredMethod); 477 } 478 } 479 } 480 481 void ASTContext::attachCommentsToJustParsedDecls(ArrayRef<Decl *> Decls, 482 const Preprocessor *PP) { 483 if (Comments.empty() || Decls.empty()) 484 return; 485 486 FileID File; 487 for (Decl *D : Decls) { 488 SourceLocation Loc = D->getLocation(); 489 if (Loc.isValid()) { 490 // See if there are any new comments that are not attached to a decl. 491 // The location doesn't have to be precise - we care only about the file. 492 File = SourceMgr.getDecomposedLoc(Loc).first; 493 break; 494 } 495 } 496 497 if (File.isInvalid()) 498 return; 499 500 auto CommentsInThisFile = Comments.getCommentsInFile(File); 501 if (!CommentsInThisFile || CommentsInThisFile->empty() || 502 CommentsInThisFile->rbegin()->second->isAttached()) 503 return; 504 505 // There is at least one comment not attached to a decl. 506 // Maybe it should be attached to one of Decls? 507 // 508 // Note that this way we pick up not only comments that precede the 509 // declaration, but also comments that *follow* the declaration -- thanks to 510 // the lookahead in the lexer: we've consumed the semicolon and looked 511 // ahead through comments. 512 513 for (const Decl *D : Decls) { 514 assert(D); 515 if (D->isInvalidDecl()) 516 continue; 517 518 D = &adjustDeclToTemplate(*D); 519 520 const SourceLocation DeclLoc = getDeclLocForCommentSearch(D, SourceMgr); 521 522 if (DeclLoc.isInvalid() || !DeclLoc.isFileID()) 523 continue; 524 525 if (DeclRawComments.count(D) > 0) 526 continue; 527 528 if (RawComment *const DocComment = 529 getRawCommentForDeclNoCacheImpl(D, DeclLoc, *CommentsInThisFile)) { 530 cacheRawCommentForDecl(*D, *DocComment); 531 comments::FullComment *FC = DocComment->parse(*this, PP, D); 532 ParsedComments[D->getCanonicalDecl()] = FC; 533 } 534 } 535 } 536 537 comments::FullComment *ASTContext::cloneFullComment(comments::FullComment *FC, 538 const Decl *D) const { 539 auto *ThisDeclInfo = new (*this) comments::DeclInfo; 540 ThisDeclInfo->CommentDecl = D; 541 ThisDeclInfo->IsFilled = false; 542 ThisDeclInfo->fill(); 543 ThisDeclInfo->CommentDecl = FC->getDecl(); 544 if (!ThisDeclInfo->TemplateParameters) 545 ThisDeclInfo->TemplateParameters = FC->getDeclInfo()->TemplateParameters; 546 comments::FullComment *CFC = 547 new (*this) comments::FullComment(FC->getBlocks(), 548 ThisDeclInfo); 549 return CFC; 550 } 551 552 comments::FullComment *ASTContext::getLocalCommentForDeclUncached(const Decl *D) const { 553 const RawComment *RC = getRawCommentForDeclNoCache(D); 554 return RC ? RC->parse(*this, nullptr, D) : nullptr; 555 } 556 557 comments::FullComment *ASTContext::getCommentForDecl( 558 const Decl *D, 559 const Preprocessor *PP) const { 560 if (!D || D->isInvalidDecl()) 561 return nullptr; 562 D = &adjustDeclToTemplate(*D); 563 564 const Decl *Canonical = D->getCanonicalDecl(); 565 llvm::DenseMap<const Decl *, comments::FullComment *>::iterator Pos = 566 ParsedComments.find(Canonical); 567 568 if (Pos != ParsedComments.end()) { 569 if (Canonical != D) { 570 comments::FullComment *FC = Pos->second; 571 comments::FullComment *CFC = cloneFullComment(FC, D); 572 return CFC; 573 } 574 return Pos->second; 575 } 576 577 const Decl *OriginalDecl = nullptr; 578 579 const RawComment *RC = getRawCommentForAnyRedecl(D, &OriginalDecl); 580 if (!RC) { 581 if (isa<ObjCMethodDecl>(D) || isa<FunctionDecl>(D)) { 582 SmallVector<const NamedDecl*, 8> Overridden; 583 const auto *OMD = dyn_cast<ObjCMethodDecl>(D); 584 if (OMD && OMD->isPropertyAccessor()) 585 if (const ObjCPropertyDecl *PDecl = OMD->findPropertyDecl()) 586 if (comments::FullComment *FC = getCommentForDecl(PDecl, PP)) 587 return cloneFullComment(FC, D); 588 if (OMD) 589 addRedeclaredMethods(OMD, Overridden); 590 getOverriddenMethods(dyn_cast<NamedDecl>(D), Overridden); 591 for (unsigned i = 0, e = Overridden.size(); i < e; i++) 592 if (comments::FullComment *FC = getCommentForDecl(Overridden[i], PP)) 593 return cloneFullComment(FC, D); 594 } 595 else if (const auto *TD = dyn_cast<TypedefNameDecl>(D)) { 596 // Attach any tag type's documentation to its typedef if latter 597 // does not have one of its own. 598 QualType QT = TD->getUnderlyingType(); 599 if (const auto *TT = QT->getAs<TagType>()) 600 if (const Decl *TD = TT->getDecl()) 601 if (comments::FullComment *FC = getCommentForDecl(TD, PP)) 602 return cloneFullComment(FC, D); 603 } 604 else if (const auto *IC = dyn_cast<ObjCInterfaceDecl>(D)) { 605 while (IC->getSuperClass()) { 606 IC = IC->getSuperClass(); 607 if (comments::FullComment *FC = getCommentForDecl(IC, PP)) 608 return cloneFullComment(FC, D); 609 } 610 } 611 else if (const auto *CD = dyn_cast<ObjCCategoryDecl>(D)) { 612 if (const ObjCInterfaceDecl *IC = CD->getClassInterface()) 613 if (comments::FullComment *FC = getCommentForDecl(IC, PP)) 614 return cloneFullComment(FC, D); 615 } 616 else if (const auto *RD = dyn_cast<CXXRecordDecl>(D)) { 617 if (!(RD = RD->getDefinition())) 618 return nullptr; 619 // Check non-virtual bases. 620 for (const auto &I : RD->bases()) { 621 if (I.isVirtual() || (I.getAccessSpecifier() != AS_public)) 622 continue; 623 QualType Ty = I.getType(); 624 if (Ty.isNull()) 625 continue; 626 if (const CXXRecordDecl *NonVirtualBase = Ty->getAsCXXRecordDecl()) { 627 if (!(NonVirtualBase= NonVirtualBase->getDefinition())) 628 continue; 629 630 if (comments::FullComment *FC = getCommentForDecl((NonVirtualBase), PP)) 631 return cloneFullComment(FC, D); 632 } 633 } 634 // Check virtual bases. 635 for (const auto &I : RD->vbases()) { 636 if (I.getAccessSpecifier() != AS_public) 637 continue; 638 QualType Ty = I.getType(); 639 if (Ty.isNull()) 640 continue; 641 if (const CXXRecordDecl *VirtualBase = Ty->getAsCXXRecordDecl()) { 642 if (!(VirtualBase= VirtualBase->getDefinition())) 643 continue; 644 if (comments::FullComment *FC = getCommentForDecl((VirtualBase), PP)) 645 return cloneFullComment(FC, D); 646 } 647 } 648 } 649 return nullptr; 650 } 651 652 // If the RawComment was attached to other redeclaration of this Decl, we 653 // should parse the comment in context of that other Decl. This is important 654 // because comments can contain references to parameter names which can be 655 // different across redeclarations. 656 if (D != OriginalDecl && OriginalDecl) 657 return getCommentForDecl(OriginalDecl, PP); 658 659 comments::FullComment *FC = RC->parse(*this, PP, D); 660 ParsedComments[Canonical] = FC; 661 return FC; 662 } 663 664 void 665 ASTContext::CanonicalTemplateTemplateParm::Profile(llvm::FoldingSetNodeID &ID, 666 const ASTContext &C, 667 TemplateTemplateParmDecl *Parm) { 668 ID.AddInteger(Parm->getDepth()); 669 ID.AddInteger(Parm->getPosition()); 670 ID.AddBoolean(Parm->isParameterPack()); 671 672 TemplateParameterList *Params = Parm->getTemplateParameters(); 673 ID.AddInteger(Params->size()); 674 for (TemplateParameterList::const_iterator P = Params->begin(), 675 PEnd = Params->end(); 676 P != PEnd; ++P) { 677 if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(*P)) { 678 ID.AddInteger(0); 679 ID.AddBoolean(TTP->isParameterPack()); 680 const TypeConstraint *TC = TTP->getTypeConstraint(); 681 ID.AddBoolean(TC != nullptr); 682 if (TC) 683 TC->getImmediatelyDeclaredConstraint()->Profile(ID, C, 684 /*Canonical=*/true); 685 if (TTP->isExpandedParameterPack()) { 686 ID.AddBoolean(true); 687 ID.AddInteger(TTP->getNumExpansionParameters()); 688 } else 689 ID.AddBoolean(false); 690 continue; 691 } 692 693 if (const auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(*P)) { 694 ID.AddInteger(1); 695 ID.AddBoolean(NTTP->isParameterPack()); 696 ID.AddPointer(NTTP->getType().getCanonicalType().getAsOpaquePtr()); 697 if (NTTP->isExpandedParameterPack()) { 698 ID.AddBoolean(true); 699 ID.AddInteger(NTTP->getNumExpansionTypes()); 700 for (unsigned I = 0, N = NTTP->getNumExpansionTypes(); I != N; ++I) { 701 QualType T = NTTP->getExpansionType(I); 702 ID.AddPointer(T.getCanonicalType().getAsOpaquePtr()); 703 } 704 } else 705 ID.AddBoolean(false); 706 continue; 707 } 708 709 auto *TTP = cast<TemplateTemplateParmDecl>(*P); 710 ID.AddInteger(2); 711 Profile(ID, C, TTP); 712 } 713 Expr *RequiresClause = Parm->getTemplateParameters()->getRequiresClause(); 714 ID.AddBoolean(RequiresClause != nullptr); 715 if (RequiresClause) 716 RequiresClause->Profile(ID, C, /*Canonical=*/true); 717 } 718 719 static Expr * 720 canonicalizeImmediatelyDeclaredConstraint(const ASTContext &C, Expr *IDC, 721 QualType ConstrainedType) { 722 // This is a bit ugly - we need to form a new immediately-declared 723 // constraint that references the new parameter; this would ideally 724 // require semantic analysis (e.g. template<C T> struct S {}; - the 725 // converted arguments of C<T> could be an argument pack if C is 726 // declared as template<typename... T> concept C = ...). 727 // We don't have semantic analysis here so we dig deep into the 728 // ready-made constraint expr and change the thing manually. 729 ConceptSpecializationExpr *CSE; 730 if (const auto *Fold = dyn_cast<CXXFoldExpr>(IDC)) 731 CSE = cast<ConceptSpecializationExpr>(Fold->getLHS()); 732 else 733 CSE = cast<ConceptSpecializationExpr>(IDC); 734 ArrayRef<TemplateArgument> OldConverted = CSE->getTemplateArguments(); 735 SmallVector<TemplateArgument, 3> NewConverted; 736 NewConverted.reserve(OldConverted.size()); 737 if (OldConverted.front().getKind() == TemplateArgument::Pack) { 738 // The case: 739 // template<typename... T> concept C = true; 740 // template<C<int> T> struct S; -> constraint is C<{T, int}> 741 NewConverted.push_back(ConstrainedType); 742 for (auto &Arg : OldConverted.front().pack_elements().drop_front(1)) 743 NewConverted.push_back(Arg); 744 TemplateArgument NewPack(NewConverted); 745 746 NewConverted.clear(); 747 NewConverted.push_back(NewPack); 748 assert(OldConverted.size() == 1 && 749 "Template parameter pack should be the last parameter"); 750 } else { 751 assert(OldConverted.front().getKind() == TemplateArgument::Type && 752 "Unexpected first argument kind for immediately-declared " 753 "constraint"); 754 NewConverted.push_back(ConstrainedType); 755 for (auto &Arg : OldConverted.drop_front(1)) 756 NewConverted.push_back(Arg); 757 } 758 Expr *NewIDC = ConceptSpecializationExpr::Create( 759 C, CSE->getNamedConcept(), NewConverted, nullptr, 760 CSE->isInstantiationDependent(), CSE->containsUnexpandedParameterPack()); 761 762 if (auto *OrigFold = dyn_cast<CXXFoldExpr>(IDC)) 763 NewIDC = new (C) CXXFoldExpr( 764 OrigFold->getType(), /*Callee*/nullptr, SourceLocation(), NewIDC, 765 BinaryOperatorKind::BO_LAnd, SourceLocation(), /*RHS=*/nullptr, 766 SourceLocation(), /*NumExpansions=*/None); 767 return NewIDC; 768 } 769 770 TemplateTemplateParmDecl * 771 ASTContext::getCanonicalTemplateTemplateParmDecl( 772 TemplateTemplateParmDecl *TTP) const { 773 // Check if we already have a canonical template template parameter. 774 llvm::FoldingSetNodeID ID; 775 CanonicalTemplateTemplateParm::Profile(ID, *this, TTP); 776 void *InsertPos = nullptr; 777 CanonicalTemplateTemplateParm *Canonical 778 = CanonTemplateTemplateParms.FindNodeOrInsertPos(ID, InsertPos); 779 if (Canonical) 780 return Canonical->getParam(); 781 782 // Build a canonical template parameter list. 783 TemplateParameterList *Params = TTP->getTemplateParameters(); 784 SmallVector<NamedDecl *, 4> CanonParams; 785 CanonParams.reserve(Params->size()); 786 for (TemplateParameterList::const_iterator P = Params->begin(), 787 PEnd = Params->end(); 788 P != PEnd; ++P) { 789 if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(*P)) { 790 TemplateTypeParmDecl *NewTTP = TemplateTypeParmDecl::Create(*this, 791 getTranslationUnitDecl(), SourceLocation(), SourceLocation(), 792 TTP->getDepth(), TTP->getIndex(), nullptr, false, 793 TTP->isParameterPack(), TTP->hasTypeConstraint(), 794 TTP->isExpandedParameterPack() ? 795 llvm::Optional<unsigned>(TTP->getNumExpansionParameters()) : None); 796 if (const auto *TC = TTP->getTypeConstraint()) { 797 QualType ParamAsArgument(NewTTP->getTypeForDecl(), 0); 798 Expr *NewIDC = canonicalizeImmediatelyDeclaredConstraint( 799 *this, TC->getImmediatelyDeclaredConstraint(), 800 ParamAsArgument); 801 TemplateArgumentListInfo CanonArgsAsWritten; 802 if (auto *Args = TC->getTemplateArgsAsWritten()) 803 for (const auto &ArgLoc : Args->arguments()) 804 CanonArgsAsWritten.addArgument( 805 TemplateArgumentLoc(ArgLoc.getArgument(), 806 TemplateArgumentLocInfo())); 807 NewTTP->setTypeConstraint( 808 NestedNameSpecifierLoc(), 809 DeclarationNameInfo(TC->getNamedConcept()->getDeclName(), 810 SourceLocation()), /*FoundDecl=*/nullptr, 811 // Actually canonicalizing a TemplateArgumentLoc is difficult so we 812 // simply omit the ArgsAsWritten 813 TC->getNamedConcept(), /*ArgsAsWritten=*/nullptr, NewIDC); 814 } 815 CanonParams.push_back(NewTTP); 816 } else if (const auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(*P)) { 817 QualType T = getCanonicalType(NTTP->getType()); 818 TypeSourceInfo *TInfo = getTrivialTypeSourceInfo(T); 819 NonTypeTemplateParmDecl *Param; 820 if (NTTP->isExpandedParameterPack()) { 821 SmallVector<QualType, 2> ExpandedTypes; 822 SmallVector<TypeSourceInfo *, 2> ExpandedTInfos; 823 for (unsigned I = 0, N = NTTP->getNumExpansionTypes(); I != N; ++I) { 824 ExpandedTypes.push_back(getCanonicalType(NTTP->getExpansionType(I))); 825 ExpandedTInfos.push_back( 826 getTrivialTypeSourceInfo(ExpandedTypes.back())); 827 } 828 829 Param = NonTypeTemplateParmDecl::Create(*this, getTranslationUnitDecl(), 830 SourceLocation(), 831 SourceLocation(), 832 NTTP->getDepth(), 833 NTTP->getPosition(), nullptr, 834 T, 835 TInfo, 836 ExpandedTypes, 837 ExpandedTInfos); 838 } else { 839 Param = NonTypeTemplateParmDecl::Create(*this, getTranslationUnitDecl(), 840 SourceLocation(), 841 SourceLocation(), 842 NTTP->getDepth(), 843 NTTP->getPosition(), nullptr, 844 T, 845 NTTP->isParameterPack(), 846 TInfo); 847 } 848 if (AutoType *AT = T->getContainedAutoType()) { 849 if (AT->isConstrained()) { 850 Param->setPlaceholderTypeConstraint( 851 canonicalizeImmediatelyDeclaredConstraint( 852 *this, NTTP->getPlaceholderTypeConstraint(), T)); 853 } 854 } 855 CanonParams.push_back(Param); 856 857 } else 858 CanonParams.push_back(getCanonicalTemplateTemplateParmDecl( 859 cast<TemplateTemplateParmDecl>(*P))); 860 } 861 862 Expr *CanonRequiresClause = nullptr; 863 if (Expr *RequiresClause = TTP->getTemplateParameters()->getRequiresClause()) 864 CanonRequiresClause = RequiresClause; 865 866 TemplateTemplateParmDecl *CanonTTP 867 = TemplateTemplateParmDecl::Create(*this, getTranslationUnitDecl(), 868 SourceLocation(), TTP->getDepth(), 869 TTP->getPosition(), 870 TTP->isParameterPack(), 871 nullptr, 872 TemplateParameterList::Create(*this, SourceLocation(), 873 SourceLocation(), 874 CanonParams, 875 SourceLocation(), 876 CanonRequiresClause)); 877 878 // Get the new insert position for the node we care about. 879 Canonical = CanonTemplateTemplateParms.FindNodeOrInsertPos(ID, InsertPos); 880 assert(!Canonical && "Shouldn't be in the map!"); 881 (void)Canonical; 882 883 // Create the canonical template template parameter entry. 884 Canonical = new (*this) CanonicalTemplateTemplateParm(CanonTTP); 885 CanonTemplateTemplateParms.InsertNode(Canonical, InsertPos); 886 return CanonTTP; 887 } 888 889 TargetCXXABI::Kind ASTContext::getCXXABIKind() const { 890 auto Kind = getTargetInfo().getCXXABI().getKind(); 891 return getLangOpts().CXXABI.getValueOr(Kind); 892 } 893 894 CXXABI *ASTContext::createCXXABI(const TargetInfo &T) { 895 if (!LangOpts.CPlusPlus) return nullptr; 896 897 switch (getCXXABIKind()) { 898 case TargetCXXABI::AppleARM64: 899 case TargetCXXABI::Fuchsia: 900 case TargetCXXABI::GenericARM: // Same as Itanium at this level 901 case TargetCXXABI::iOS: 902 case TargetCXXABI::WatchOS: 903 case TargetCXXABI::GenericAArch64: 904 case TargetCXXABI::GenericMIPS: 905 case TargetCXXABI::GenericItanium: 906 case TargetCXXABI::WebAssembly: 907 case TargetCXXABI::XL: 908 return CreateItaniumCXXABI(*this); 909 case TargetCXXABI::Microsoft: 910 return CreateMicrosoftCXXABI(*this); 911 } 912 llvm_unreachable("Invalid CXXABI type!"); 913 } 914 915 interp::Context &ASTContext::getInterpContext() { 916 if (!InterpContext) { 917 InterpContext.reset(new interp::Context(*this)); 918 } 919 return *InterpContext.get(); 920 } 921 922 ParentMapContext &ASTContext::getParentMapContext() { 923 if (!ParentMapCtx) 924 ParentMapCtx.reset(new ParentMapContext(*this)); 925 return *ParentMapCtx.get(); 926 } 927 928 static const LangASMap *getAddressSpaceMap(const TargetInfo &T, 929 const LangOptions &LOpts) { 930 if (LOpts.FakeAddressSpaceMap) { 931 // The fake address space map must have a distinct entry for each 932 // language-specific address space. 933 static const unsigned FakeAddrSpaceMap[] = { 934 0, // Default 935 1, // opencl_global 936 3, // opencl_local 937 2, // opencl_constant 938 0, // opencl_private 939 4, // opencl_generic 940 5, // opencl_global_device 941 6, // opencl_global_host 942 7, // cuda_device 943 8, // cuda_constant 944 9, // cuda_shared 945 1, // sycl_global 946 5, // sycl_global_device 947 6, // sycl_global_host 948 3, // sycl_local 949 0, // sycl_private 950 10, // ptr32_sptr 951 11, // ptr32_uptr 952 12 // ptr64 953 }; 954 return &FakeAddrSpaceMap; 955 } else { 956 return &T.getAddressSpaceMap(); 957 } 958 } 959 960 static bool isAddrSpaceMapManglingEnabled(const TargetInfo &TI, 961 const LangOptions &LangOpts) { 962 switch (LangOpts.getAddressSpaceMapMangling()) { 963 case LangOptions::ASMM_Target: 964 return TI.useAddressSpaceMapMangling(); 965 case LangOptions::ASMM_On: 966 return true; 967 case LangOptions::ASMM_Off: 968 return false; 969 } 970 llvm_unreachable("getAddressSpaceMapMangling() doesn't cover anything."); 971 } 972 973 ASTContext::ASTContext(LangOptions &LOpts, SourceManager &SM, 974 IdentifierTable &idents, SelectorTable &sels, 975 Builtin::Context &builtins, TranslationUnitKind TUKind) 976 : ConstantArrayTypes(this_()), FunctionProtoTypes(this_()), 977 TemplateSpecializationTypes(this_()), 978 DependentTemplateSpecializationTypes(this_()), AutoTypes(this_()), 979 SubstTemplateTemplateParmPacks(this_()), 980 CanonTemplateTemplateParms(this_()), SourceMgr(SM), LangOpts(LOpts), 981 NoSanitizeL(new NoSanitizeList(LangOpts.NoSanitizeFiles, SM)), 982 XRayFilter(new XRayFunctionFilter(LangOpts.XRayAlwaysInstrumentFiles, 983 LangOpts.XRayNeverInstrumentFiles, 984 LangOpts.XRayAttrListFiles, SM)), 985 ProfList(new ProfileList(LangOpts.ProfileListFiles, SM)), 986 PrintingPolicy(LOpts), Idents(idents), Selectors(sels), 987 BuiltinInfo(builtins), TUKind(TUKind), DeclarationNames(*this), 988 Comments(SM), CommentCommandTraits(BumpAlloc, LOpts.CommentOpts), 989 CompCategories(this_()), LastSDM(nullptr, 0) { 990 addTranslationUnitDecl(); 991 } 992 993 void ASTContext::cleanup() { 994 // Release the DenseMaps associated with DeclContext objects. 995 // FIXME: Is this the ideal solution? 996 ReleaseDeclContextMaps(); 997 998 // Call all of the deallocation functions on all of their targets. 999 for (auto &Pair : Deallocations) 1000 (Pair.first)(Pair.second); 1001 Deallocations.clear(); 1002 1003 // ASTRecordLayout objects in ASTRecordLayouts must always be destroyed 1004 // because they can contain DenseMaps. 1005 for (llvm::DenseMap<const ObjCContainerDecl*, 1006 const ASTRecordLayout*>::iterator 1007 I = ObjCLayouts.begin(), E = ObjCLayouts.end(); I != E; ) 1008 // Increment in loop to prevent using deallocated memory. 1009 if (auto *R = const_cast<ASTRecordLayout *>((I++)->second)) 1010 R->Destroy(*this); 1011 ObjCLayouts.clear(); 1012 1013 for (llvm::DenseMap<const RecordDecl*, const ASTRecordLayout*>::iterator 1014 I = ASTRecordLayouts.begin(), E = ASTRecordLayouts.end(); I != E; ) { 1015 // Increment in loop to prevent using deallocated memory. 1016 if (auto *R = const_cast<ASTRecordLayout *>((I++)->second)) 1017 R->Destroy(*this); 1018 } 1019 ASTRecordLayouts.clear(); 1020 1021 for (llvm::DenseMap<const Decl*, AttrVec*>::iterator A = DeclAttrs.begin(), 1022 AEnd = DeclAttrs.end(); 1023 A != AEnd; ++A) 1024 A->second->~AttrVec(); 1025 DeclAttrs.clear(); 1026 1027 for (const auto &Value : ModuleInitializers) 1028 Value.second->~PerModuleInitializers(); 1029 ModuleInitializers.clear(); 1030 } 1031 1032 ASTContext::~ASTContext() { cleanup(); } 1033 1034 void ASTContext::setTraversalScope(const std::vector<Decl *> &TopLevelDecls) { 1035 TraversalScope = TopLevelDecls; 1036 getParentMapContext().clear(); 1037 } 1038 1039 void ASTContext::AddDeallocation(void (*Callback)(void *), void *Data) const { 1040 Deallocations.push_back({Callback, Data}); 1041 } 1042 1043 void 1044 ASTContext::setExternalSource(IntrusiveRefCntPtr<ExternalASTSource> Source) { 1045 ExternalSource = std::move(Source); 1046 } 1047 1048 void ASTContext::PrintStats() const { 1049 llvm::errs() << "\n*** AST Context Stats:\n"; 1050 llvm::errs() << " " << Types.size() << " types total.\n"; 1051 1052 unsigned counts[] = { 1053 #define TYPE(Name, Parent) 0, 1054 #define ABSTRACT_TYPE(Name, Parent) 1055 #include "clang/AST/TypeNodes.inc" 1056 0 // Extra 1057 }; 1058 1059 for (unsigned i = 0, e = Types.size(); i != e; ++i) { 1060 Type *T = Types[i]; 1061 counts[(unsigned)T->getTypeClass()]++; 1062 } 1063 1064 unsigned Idx = 0; 1065 unsigned TotalBytes = 0; 1066 #define TYPE(Name, Parent) \ 1067 if (counts[Idx]) \ 1068 llvm::errs() << " " << counts[Idx] << " " << #Name \ 1069 << " types, " << sizeof(Name##Type) << " each " \ 1070 << "(" << counts[Idx] * sizeof(Name##Type) \ 1071 << " bytes)\n"; \ 1072 TotalBytes += counts[Idx] * sizeof(Name##Type); \ 1073 ++Idx; 1074 #define ABSTRACT_TYPE(Name, Parent) 1075 #include "clang/AST/TypeNodes.inc" 1076 1077 llvm::errs() << "Total bytes = " << TotalBytes << "\n"; 1078 1079 // Implicit special member functions. 1080 llvm::errs() << NumImplicitDefaultConstructorsDeclared << "/" 1081 << NumImplicitDefaultConstructors 1082 << " implicit default constructors created\n"; 1083 llvm::errs() << NumImplicitCopyConstructorsDeclared << "/" 1084 << NumImplicitCopyConstructors 1085 << " implicit copy constructors created\n"; 1086 if (getLangOpts().CPlusPlus) 1087 llvm::errs() << NumImplicitMoveConstructorsDeclared << "/" 1088 << NumImplicitMoveConstructors 1089 << " implicit move constructors created\n"; 1090 llvm::errs() << NumImplicitCopyAssignmentOperatorsDeclared << "/" 1091 << NumImplicitCopyAssignmentOperators 1092 << " implicit copy assignment operators created\n"; 1093 if (getLangOpts().CPlusPlus) 1094 llvm::errs() << NumImplicitMoveAssignmentOperatorsDeclared << "/" 1095 << NumImplicitMoveAssignmentOperators 1096 << " implicit move assignment operators created\n"; 1097 llvm::errs() << NumImplicitDestructorsDeclared << "/" 1098 << NumImplicitDestructors 1099 << " implicit destructors created\n"; 1100 1101 if (ExternalSource) { 1102 llvm::errs() << "\n"; 1103 ExternalSource->PrintStats(); 1104 } 1105 1106 BumpAlloc.PrintStats(); 1107 } 1108 1109 void ASTContext::mergeDefinitionIntoModule(NamedDecl *ND, Module *M, 1110 bool NotifyListeners) { 1111 if (NotifyListeners) 1112 if (auto *Listener = getASTMutationListener()) 1113 Listener->RedefinedHiddenDefinition(ND, M); 1114 1115 MergedDefModules[cast<NamedDecl>(ND->getCanonicalDecl())].push_back(M); 1116 } 1117 1118 void ASTContext::deduplicateMergedDefinitonsFor(NamedDecl *ND) { 1119 auto It = MergedDefModules.find(cast<NamedDecl>(ND->getCanonicalDecl())); 1120 if (It == MergedDefModules.end()) 1121 return; 1122 1123 auto &Merged = It->second; 1124 llvm::DenseSet<Module*> Found; 1125 for (Module *&M : Merged) 1126 if (!Found.insert(M).second) 1127 M = nullptr; 1128 llvm::erase_value(Merged, nullptr); 1129 } 1130 1131 ArrayRef<Module *> 1132 ASTContext::getModulesWithMergedDefinition(const NamedDecl *Def) { 1133 auto MergedIt = 1134 MergedDefModules.find(cast<NamedDecl>(Def->getCanonicalDecl())); 1135 if (MergedIt == MergedDefModules.end()) 1136 return None; 1137 return MergedIt->second; 1138 } 1139 1140 void ASTContext::PerModuleInitializers::resolve(ASTContext &Ctx) { 1141 if (LazyInitializers.empty()) 1142 return; 1143 1144 auto *Source = Ctx.getExternalSource(); 1145 assert(Source && "lazy initializers but no external source"); 1146 1147 auto LazyInits = std::move(LazyInitializers); 1148 LazyInitializers.clear(); 1149 1150 for (auto ID : LazyInits) 1151 Initializers.push_back(Source->GetExternalDecl(ID)); 1152 1153 assert(LazyInitializers.empty() && 1154 "GetExternalDecl for lazy module initializer added more inits"); 1155 } 1156 1157 void ASTContext::addModuleInitializer(Module *M, Decl *D) { 1158 // One special case: if we add a module initializer that imports another 1159 // module, and that module's only initializer is an ImportDecl, simplify. 1160 if (const auto *ID = dyn_cast<ImportDecl>(D)) { 1161 auto It = ModuleInitializers.find(ID->getImportedModule()); 1162 1163 // Maybe the ImportDecl does nothing at all. (Common case.) 1164 if (It == ModuleInitializers.end()) 1165 return; 1166 1167 // Maybe the ImportDecl only imports another ImportDecl. 1168 auto &Imported = *It->second; 1169 if (Imported.Initializers.size() + Imported.LazyInitializers.size() == 1) { 1170 Imported.resolve(*this); 1171 auto *OnlyDecl = Imported.Initializers.front(); 1172 if (isa<ImportDecl>(OnlyDecl)) 1173 D = OnlyDecl; 1174 } 1175 } 1176 1177 auto *&Inits = ModuleInitializers[M]; 1178 if (!Inits) 1179 Inits = new (*this) PerModuleInitializers; 1180 Inits->Initializers.push_back(D); 1181 } 1182 1183 void ASTContext::addLazyModuleInitializers(Module *M, ArrayRef<uint32_t> IDs) { 1184 auto *&Inits = ModuleInitializers[M]; 1185 if (!Inits) 1186 Inits = new (*this) PerModuleInitializers; 1187 Inits->LazyInitializers.insert(Inits->LazyInitializers.end(), 1188 IDs.begin(), IDs.end()); 1189 } 1190 1191 ArrayRef<Decl *> ASTContext::getModuleInitializers(Module *M) { 1192 auto It = ModuleInitializers.find(M); 1193 if (It == ModuleInitializers.end()) 1194 return None; 1195 1196 auto *Inits = It->second; 1197 Inits->resolve(*this); 1198 return Inits->Initializers; 1199 } 1200 1201 ExternCContextDecl *ASTContext::getExternCContextDecl() const { 1202 if (!ExternCContext) 1203 ExternCContext = ExternCContextDecl::Create(*this, getTranslationUnitDecl()); 1204 1205 return ExternCContext; 1206 } 1207 1208 BuiltinTemplateDecl * 1209 ASTContext::buildBuiltinTemplateDecl(BuiltinTemplateKind BTK, 1210 const IdentifierInfo *II) const { 1211 auto *BuiltinTemplate = 1212 BuiltinTemplateDecl::Create(*this, getTranslationUnitDecl(), II, BTK); 1213 BuiltinTemplate->setImplicit(); 1214 getTranslationUnitDecl()->addDecl(BuiltinTemplate); 1215 1216 return BuiltinTemplate; 1217 } 1218 1219 BuiltinTemplateDecl * 1220 ASTContext::getMakeIntegerSeqDecl() const { 1221 if (!MakeIntegerSeqDecl) 1222 MakeIntegerSeqDecl = buildBuiltinTemplateDecl(BTK__make_integer_seq, 1223 getMakeIntegerSeqName()); 1224 return MakeIntegerSeqDecl; 1225 } 1226 1227 BuiltinTemplateDecl * 1228 ASTContext::getTypePackElementDecl() const { 1229 if (!TypePackElementDecl) 1230 TypePackElementDecl = buildBuiltinTemplateDecl(BTK__type_pack_element, 1231 getTypePackElementName()); 1232 return TypePackElementDecl; 1233 } 1234 1235 RecordDecl *ASTContext::buildImplicitRecord(StringRef Name, 1236 RecordDecl::TagKind TK) const { 1237 SourceLocation Loc; 1238 RecordDecl *NewDecl; 1239 if (getLangOpts().CPlusPlus) 1240 NewDecl = CXXRecordDecl::Create(*this, TK, getTranslationUnitDecl(), Loc, 1241 Loc, &Idents.get(Name)); 1242 else 1243 NewDecl = RecordDecl::Create(*this, TK, getTranslationUnitDecl(), Loc, Loc, 1244 &Idents.get(Name)); 1245 NewDecl->setImplicit(); 1246 NewDecl->addAttr(TypeVisibilityAttr::CreateImplicit( 1247 const_cast<ASTContext &>(*this), TypeVisibilityAttr::Default)); 1248 return NewDecl; 1249 } 1250 1251 TypedefDecl *ASTContext::buildImplicitTypedef(QualType T, 1252 StringRef Name) const { 1253 TypeSourceInfo *TInfo = getTrivialTypeSourceInfo(T); 1254 TypedefDecl *NewDecl = TypedefDecl::Create( 1255 const_cast<ASTContext &>(*this), getTranslationUnitDecl(), 1256 SourceLocation(), SourceLocation(), &Idents.get(Name), TInfo); 1257 NewDecl->setImplicit(); 1258 return NewDecl; 1259 } 1260 1261 TypedefDecl *ASTContext::getInt128Decl() const { 1262 if (!Int128Decl) 1263 Int128Decl = buildImplicitTypedef(Int128Ty, "__int128_t"); 1264 return Int128Decl; 1265 } 1266 1267 TypedefDecl *ASTContext::getUInt128Decl() const { 1268 if (!UInt128Decl) 1269 UInt128Decl = buildImplicitTypedef(UnsignedInt128Ty, "__uint128_t"); 1270 return UInt128Decl; 1271 } 1272 1273 void ASTContext::InitBuiltinType(CanQualType &R, BuiltinType::Kind K) { 1274 auto *Ty = new (*this, TypeAlignment) BuiltinType(K); 1275 R = CanQualType::CreateUnsafe(QualType(Ty, 0)); 1276 Types.push_back(Ty); 1277 } 1278 1279 void ASTContext::InitBuiltinTypes(const TargetInfo &Target, 1280 const TargetInfo *AuxTarget) { 1281 assert((!this->Target || this->Target == &Target) && 1282 "Incorrect target reinitialization"); 1283 assert(VoidTy.isNull() && "Context reinitialized?"); 1284 1285 this->Target = &Target; 1286 this->AuxTarget = AuxTarget; 1287 1288 ABI.reset(createCXXABI(Target)); 1289 AddrSpaceMap = getAddressSpaceMap(Target, LangOpts); 1290 AddrSpaceMapMangling = isAddrSpaceMapManglingEnabled(Target, LangOpts); 1291 1292 // C99 6.2.5p19. 1293 InitBuiltinType(VoidTy, BuiltinType::Void); 1294 1295 // C99 6.2.5p2. 1296 InitBuiltinType(BoolTy, BuiltinType::Bool); 1297 // C99 6.2.5p3. 1298 if (LangOpts.CharIsSigned) 1299 InitBuiltinType(CharTy, BuiltinType::Char_S); 1300 else 1301 InitBuiltinType(CharTy, BuiltinType::Char_U); 1302 // C99 6.2.5p4. 1303 InitBuiltinType(SignedCharTy, BuiltinType::SChar); 1304 InitBuiltinType(ShortTy, BuiltinType::Short); 1305 InitBuiltinType(IntTy, BuiltinType::Int); 1306 InitBuiltinType(LongTy, BuiltinType::Long); 1307 InitBuiltinType(LongLongTy, BuiltinType::LongLong); 1308 1309 // C99 6.2.5p6. 1310 InitBuiltinType(UnsignedCharTy, BuiltinType::UChar); 1311 InitBuiltinType(UnsignedShortTy, BuiltinType::UShort); 1312 InitBuiltinType(UnsignedIntTy, BuiltinType::UInt); 1313 InitBuiltinType(UnsignedLongTy, BuiltinType::ULong); 1314 InitBuiltinType(UnsignedLongLongTy, BuiltinType::ULongLong); 1315 1316 // C99 6.2.5p10. 1317 InitBuiltinType(FloatTy, BuiltinType::Float); 1318 InitBuiltinType(DoubleTy, BuiltinType::Double); 1319 InitBuiltinType(LongDoubleTy, BuiltinType::LongDouble); 1320 1321 // GNU extension, __float128 for IEEE quadruple precision 1322 InitBuiltinType(Float128Ty, BuiltinType::Float128); 1323 1324 // __ibm128 for IBM extended precision 1325 InitBuiltinType(Ibm128Ty, BuiltinType::Ibm128); 1326 1327 // C11 extension ISO/IEC TS 18661-3 1328 InitBuiltinType(Float16Ty, BuiltinType::Float16); 1329 1330 // ISO/IEC JTC1 SC22 WG14 N1169 Extension 1331 InitBuiltinType(ShortAccumTy, BuiltinType::ShortAccum); 1332 InitBuiltinType(AccumTy, BuiltinType::Accum); 1333 InitBuiltinType(LongAccumTy, BuiltinType::LongAccum); 1334 InitBuiltinType(UnsignedShortAccumTy, BuiltinType::UShortAccum); 1335 InitBuiltinType(UnsignedAccumTy, BuiltinType::UAccum); 1336 InitBuiltinType(UnsignedLongAccumTy, BuiltinType::ULongAccum); 1337 InitBuiltinType(ShortFractTy, BuiltinType::ShortFract); 1338 InitBuiltinType(FractTy, BuiltinType::Fract); 1339 InitBuiltinType(LongFractTy, BuiltinType::LongFract); 1340 InitBuiltinType(UnsignedShortFractTy, BuiltinType::UShortFract); 1341 InitBuiltinType(UnsignedFractTy, BuiltinType::UFract); 1342 InitBuiltinType(UnsignedLongFractTy, BuiltinType::ULongFract); 1343 InitBuiltinType(SatShortAccumTy, BuiltinType::SatShortAccum); 1344 InitBuiltinType(SatAccumTy, BuiltinType::SatAccum); 1345 InitBuiltinType(SatLongAccumTy, BuiltinType::SatLongAccum); 1346 InitBuiltinType(SatUnsignedShortAccumTy, BuiltinType::SatUShortAccum); 1347 InitBuiltinType(SatUnsignedAccumTy, BuiltinType::SatUAccum); 1348 InitBuiltinType(SatUnsignedLongAccumTy, BuiltinType::SatULongAccum); 1349 InitBuiltinType(SatShortFractTy, BuiltinType::SatShortFract); 1350 InitBuiltinType(SatFractTy, BuiltinType::SatFract); 1351 InitBuiltinType(SatLongFractTy, BuiltinType::SatLongFract); 1352 InitBuiltinType(SatUnsignedShortFractTy, BuiltinType::SatUShortFract); 1353 InitBuiltinType(SatUnsignedFractTy, BuiltinType::SatUFract); 1354 InitBuiltinType(SatUnsignedLongFractTy, BuiltinType::SatULongFract); 1355 1356 // GNU extension, 128-bit integers. 1357 InitBuiltinType(Int128Ty, BuiltinType::Int128); 1358 InitBuiltinType(UnsignedInt128Ty, BuiltinType::UInt128); 1359 1360 // C++ 3.9.1p5 1361 if (TargetInfo::isTypeSigned(Target.getWCharType())) 1362 InitBuiltinType(WCharTy, BuiltinType::WChar_S); 1363 else // -fshort-wchar makes wchar_t be unsigned. 1364 InitBuiltinType(WCharTy, BuiltinType::WChar_U); 1365 if (LangOpts.CPlusPlus && LangOpts.WChar) 1366 WideCharTy = WCharTy; 1367 else { 1368 // C99 (or C++ using -fno-wchar). 1369 WideCharTy = getFromTargetType(Target.getWCharType()); 1370 } 1371 1372 WIntTy = getFromTargetType(Target.getWIntType()); 1373 1374 // C++20 (proposed) 1375 InitBuiltinType(Char8Ty, BuiltinType::Char8); 1376 1377 if (LangOpts.CPlusPlus) // C++0x 3.9.1p5, extension for C++ 1378 InitBuiltinType(Char16Ty, BuiltinType::Char16); 1379 else // C99 1380 Char16Ty = getFromTargetType(Target.getChar16Type()); 1381 1382 if (LangOpts.CPlusPlus) // C++0x 3.9.1p5, extension for C++ 1383 InitBuiltinType(Char32Ty, BuiltinType::Char32); 1384 else // C99 1385 Char32Ty = getFromTargetType(Target.getChar32Type()); 1386 1387 // Placeholder type for type-dependent expressions whose type is 1388 // completely unknown. No code should ever check a type against 1389 // DependentTy and users should never see it; however, it is here to 1390 // help diagnose failures to properly check for type-dependent 1391 // expressions. 1392 InitBuiltinType(DependentTy, BuiltinType::Dependent); 1393 1394 // Placeholder type for functions. 1395 InitBuiltinType(OverloadTy, BuiltinType::Overload); 1396 1397 // Placeholder type for bound members. 1398 InitBuiltinType(BoundMemberTy, BuiltinType::BoundMember); 1399 1400 // Placeholder type for pseudo-objects. 1401 InitBuiltinType(PseudoObjectTy, BuiltinType::PseudoObject); 1402 1403 // "any" type; useful for debugger-like clients. 1404 InitBuiltinType(UnknownAnyTy, BuiltinType::UnknownAny); 1405 1406 // Placeholder type for unbridged ARC casts. 1407 InitBuiltinType(ARCUnbridgedCastTy, BuiltinType::ARCUnbridgedCast); 1408 1409 // Placeholder type for builtin functions. 1410 InitBuiltinType(BuiltinFnTy, BuiltinType::BuiltinFn); 1411 1412 // Placeholder type for OMP array sections. 1413 if (LangOpts.OpenMP) { 1414 InitBuiltinType(OMPArraySectionTy, BuiltinType::OMPArraySection); 1415 InitBuiltinType(OMPArrayShapingTy, BuiltinType::OMPArrayShaping); 1416 InitBuiltinType(OMPIteratorTy, BuiltinType::OMPIterator); 1417 } 1418 if (LangOpts.MatrixTypes) 1419 InitBuiltinType(IncompleteMatrixIdxTy, BuiltinType::IncompleteMatrixIdx); 1420 1421 // Builtin types for 'id', 'Class', and 'SEL'. 1422 InitBuiltinType(ObjCBuiltinIdTy, BuiltinType::ObjCId); 1423 InitBuiltinType(ObjCBuiltinClassTy, BuiltinType::ObjCClass); 1424 InitBuiltinType(ObjCBuiltinSelTy, BuiltinType::ObjCSel); 1425 1426 if (LangOpts.OpenCL) { 1427 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \ 1428 InitBuiltinType(SingletonId, BuiltinType::Id); 1429 #include "clang/Basic/OpenCLImageTypes.def" 1430 1431 InitBuiltinType(OCLSamplerTy, BuiltinType::OCLSampler); 1432 InitBuiltinType(OCLEventTy, BuiltinType::OCLEvent); 1433 InitBuiltinType(OCLClkEventTy, BuiltinType::OCLClkEvent); 1434 InitBuiltinType(OCLQueueTy, BuiltinType::OCLQueue); 1435 InitBuiltinType(OCLReserveIDTy, BuiltinType::OCLReserveID); 1436 1437 #define EXT_OPAQUE_TYPE(ExtType, Id, Ext) \ 1438 InitBuiltinType(Id##Ty, BuiltinType::Id); 1439 #include "clang/Basic/OpenCLExtensionTypes.def" 1440 } 1441 1442 if (Target.hasAArch64SVETypes()) { 1443 #define SVE_TYPE(Name, Id, SingletonId) \ 1444 InitBuiltinType(SingletonId, BuiltinType::Id); 1445 #include "clang/Basic/AArch64SVEACLETypes.def" 1446 } 1447 1448 if (Target.getTriple().isPPC64()) { 1449 #define PPC_VECTOR_MMA_TYPE(Name, Id, Size) \ 1450 InitBuiltinType(Id##Ty, BuiltinType::Id); 1451 #include "clang/Basic/PPCTypes.def" 1452 #define PPC_VECTOR_VSX_TYPE(Name, Id, Size) \ 1453 InitBuiltinType(Id##Ty, BuiltinType::Id); 1454 #include "clang/Basic/PPCTypes.def" 1455 } 1456 1457 if (Target.hasRISCVVTypes()) { 1458 #define RVV_TYPE(Name, Id, SingletonId) \ 1459 InitBuiltinType(SingletonId, BuiltinType::Id); 1460 #include "clang/Basic/RISCVVTypes.def" 1461 } 1462 1463 // Builtin type for __objc_yes and __objc_no 1464 ObjCBuiltinBoolTy = (Target.useSignedCharForObjCBool() ? 1465 SignedCharTy : BoolTy); 1466 1467 ObjCConstantStringType = QualType(); 1468 1469 ObjCSuperType = QualType(); 1470 1471 // void * type 1472 if (LangOpts.OpenCLGenericAddressSpace) { 1473 auto Q = VoidTy.getQualifiers(); 1474 Q.setAddressSpace(LangAS::opencl_generic); 1475 VoidPtrTy = getPointerType(getCanonicalType( 1476 getQualifiedType(VoidTy.getUnqualifiedType(), Q))); 1477 } else { 1478 VoidPtrTy = getPointerType(VoidTy); 1479 } 1480 1481 // nullptr type (C++0x 2.14.7) 1482 InitBuiltinType(NullPtrTy, BuiltinType::NullPtr); 1483 1484 // half type (OpenCL 6.1.1.1) / ARM NEON __fp16 1485 InitBuiltinType(HalfTy, BuiltinType::Half); 1486 1487 InitBuiltinType(BFloat16Ty, BuiltinType::BFloat16); 1488 1489 // Builtin type used to help define __builtin_va_list. 1490 VaListTagDecl = nullptr; 1491 1492 // MSVC predeclares struct _GUID, and we need it to create MSGuidDecls. 1493 if (LangOpts.MicrosoftExt || LangOpts.Borland) { 1494 MSGuidTagDecl = buildImplicitRecord("_GUID"); 1495 getTranslationUnitDecl()->addDecl(MSGuidTagDecl); 1496 } 1497 } 1498 1499 DiagnosticsEngine &ASTContext::getDiagnostics() const { 1500 return SourceMgr.getDiagnostics(); 1501 } 1502 1503 AttrVec& ASTContext::getDeclAttrs(const Decl *D) { 1504 AttrVec *&Result = DeclAttrs[D]; 1505 if (!Result) { 1506 void *Mem = Allocate(sizeof(AttrVec)); 1507 Result = new (Mem) AttrVec; 1508 } 1509 1510 return *Result; 1511 } 1512 1513 /// Erase the attributes corresponding to the given declaration. 1514 void ASTContext::eraseDeclAttrs(const Decl *D) { 1515 llvm::DenseMap<const Decl*, AttrVec*>::iterator Pos = DeclAttrs.find(D); 1516 if (Pos != DeclAttrs.end()) { 1517 Pos->second->~AttrVec(); 1518 DeclAttrs.erase(Pos); 1519 } 1520 } 1521 1522 // FIXME: Remove ? 1523 MemberSpecializationInfo * 1524 ASTContext::getInstantiatedFromStaticDataMember(const VarDecl *Var) { 1525 assert(Var->isStaticDataMember() && "Not a static data member"); 1526 return getTemplateOrSpecializationInfo(Var) 1527 .dyn_cast<MemberSpecializationInfo *>(); 1528 } 1529 1530 ASTContext::TemplateOrSpecializationInfo 1531 ASTContext::getTemplateOrSpecializationInfo(const VarDecl *Var) { 1532 llvm::DenseMap<const VarDecl *, TemplateOrSpecializationInfo>::iterator Pos = 1533 TemplateOrInstantiation.find(Var); 1534 if (Pos == TemplateOrInstantiation.end()) 1535 return {}; 1536 1537 return Pos->second; 1538 } 1539 1540 void 1541 ASTContext::setInstantiatedFromStaticDataMember(VarDecl *Inst, VarDecl *Tmpl, 1542 TemplateSpecializationKind TSK, 1543 SourceLocation PointOfInstantiation) { 1544 assert(Inst->isStaticDataMember() && "Not a static data member"); 1545 assert(Tmpl->isStaticDataMember() && "Not a static data member"); 1546 setTemplateOrSpecializationInfo(Inst, new (*this) MemberSpecializationInfo( 1547 Tmpl, TSK, PointOfInstantiation)); 1548 } 1549 1550 void 1551 ASTContext::setTemplateOrSpecializationInfo(VarDecl *Inst, 1552 TemplateOrSpecializationInfo TSI) { 1553 assert(!TemplateOrInstantiation[Inst] && 1554 "Already noted what the variable was instantiated from"); 1555 TemplateOrInstantiation[Inst] = TSI; 1556 } 1557 1558 NamedDecl * 1559 ASTContext::getInstantiatedFromUsingDecl(NamedDecl *UUD) { 1560 auto Pos = InstantiatedFromUsingDecl.find(UUD); 1561 if (Pos == InstantiatedFromUsingDecl.end()) 1562 return nullptr; 1563 1564 return Pos->second; 1565 } 1566 1567 void 1568 ASTContext::setInstantiatedFromUsingDecl(NamedDecl *Inst, NamedDecl *Pattern) { 1569 assert((isa<UsingDecl>(Pattern) || 1570 isa<UnresolvedUsingValueDecl>(Pattern) || 1571 isa<UnresolvedUsingTypenameDecl>(Pattern)) && 1572 "pattern decl is not a using decl"); 1573 assert((isa<UsingDecl>(Inst) || 1574 isa<UnresolvedUsingValueDecl>(Inst) || 1575 isa<UnresolvedUsingTypenameDecl>(Inst)) && 1576 "instantiation did not produce a using decl"); 1577 assert(!InstantiatedFromUsingDecl[Inst] && "pattern already exists"); 1578 InstantiatedFromUsingDecl[Inst] = Pattern; 1579 } 1580 1581 UsingEnumDecl * 1582 ASTContext::getInstantiatedFromUsingEnumDecl(UsingEnumDecl *UUD) { 1583 auto Pos = InstantiatedFromUsingEnumDecl.find(UUD); 1584 if (Pos == InstantiatedFromUsingEnumDecl.end()) 1585 return nullptr; 1586 1587 return Pos->second; 1588 } 1589 1590 void ASTContext::setInstantiatedFromUsingEnumDecl(UsingEnumDecl *Inst, 1591 UsingEnumDecl *Pattern) { 1592 assert(!InstantiatedFromUsingEnumDecl[Inst] && "pattern already exists"); 1593 InstantiatedFromUsingEnumDecl[Inst] = Pattern; 1594 } 1595 1596 UsingShadowDecl * 1597 ASTContext::getInstantiatedFromUsingShadowDecl(UsingShadowDecl *Inst) { 1598 llvm::DenseMap<UsingShadowDecl*, UsingShadowDecl*>::const_iterator Pos 1599 = InstantiatedFromUsingShadowDecl.find(Inst); 1600 if (Pos == InstantiatedFromUsingShadowDecl.end()) 1601 return nullptr; 1602 1603 return Pos->second; 1604 } 1605 1606 void 1607 ASTContext::setInstantiatedFromUsingShadowDecl(UsingShadowDecl *Inst, 1608 UsingShadowDecl *Pattern) { 1609 assert(!InstantiatedFromUsingShadowDecl[Inst] && "pattern already exists"); 1610 InstantiatedFromUsingShadowDecl[Inst] = Pattern; 1611 } 1612 1613 FieldDecl *ASTContext::getInstantiatedFromUnnamedFieldDecl(FieldDecl *Field) { 1614 llvm::DenseMap<FieldDecl *, FieldDecl *>::iterator Pos 1615 = InstantiatedFromUnnamedFieldDecl.find(Field); 1616 if (Pos == InstantiatedFromUnnamedFieldDecl.end()) 1617 return nullptr; 1618 1619 return Pos->second; 1620 } 1621 1622 void ASTContext::setInstantiatedFromUnnamedFieldDecl(FieldDecl *Inst, 1623 FieldDecl *Tmpl) { 1624 assert(!Inst->getDeclName() && "Instantiated field decl is not unnamed"); 1625 assert(!Tmpl->getDeclName() && "Template field decl is not unnamed"); 1626 assert(!InstantiatedFromUnnamedFieldDecl[Inst] && 1627 "Already noted what unnamed field was instantiated from"); 1628 1629 InstantiatedFromUnnamedFieldDecl[Inst] = Tmpl; 1630 } 1631 1632 ASTContext::overridden_cxx_method_iterator 1633 ASTContext::overridden_methods_begin(const CXXMethodDecl *Method) const { 1634 return overridden_methods(Method).begin(); 1635 } 1636 1637 ASTContext::overridden_cxx_method_iterator 1638 ASTContext::overridden_methods_end(const CXXMethodDecl *Method) const { 1639 return overridden_methods(Method).end(); 1640 } 1641 1642 unsigned 1643 ASTContext::overridden_methods_size(const CXXMethodDecl *Method) const { 1644 auto Range = overridden_methods(Method); 1645 return Range.end() - Range.begin(); 1646 } 1647 1648 ASTContext::overridden_method_range 1649 ASTContext::overridden_methods(const CXXMethodDecl *Method) const { 1650 llvm::DenseMap<const CXXMethodDecl *, CXXMethodVector>::const_iterator Pos = 1651 OverriddenMethods.find(Method->getCanonicalDecl()); 1652 if (Pos == OverriddenMethods.end()) 1653 return overridden_method_range(nullptr, nullptr); 1654 return overridden_method_range(Pos->second.begin(), Pos->second.end()); 1655 } 1656 1657 void ASTContext::addOverriddenMethod(const CXXMethodDecl *Method, 1658 const CXXMethodDecl *Overridden) { 1659 assert(Method->isCanonicalDecl() && Overridden->isCanonicalDecl()); 1660 OverriddenMethods[Method].push_back(Overridden); 1661 } 1662 1663 void ASTContext::getOverriddenMethods( 1664 const NamedDecl *D, 1665 SmallVectorImpl<const NamedDecl *> &Overridden) const { 1666 assert(D); 1667 1668 if (const auto *CXXMethod = dyn_cast<CXXMethodDecl>(D)) { 1669 Overridden.append(overridden_methods_begin(CXXMethod), 1670 overridden_methods_end(CXXMethod)); 1671 return; 1672 } 1673 1674 const auto *Method = dyn_cast<ObjCMethodDecl>(D); 1675 if (!Method) 1676 return; 1677 1678 SmallVector<const ObjCMethodDecl *, 8> OverDecls; 1679 Method->getOverriddenMethods(OverDecls); 1680 Overridden.append(OverDecls.begin(), OverDecls.end()); 1681 } 1682 1683 void ASTContext::addedLocalImportDecl(ImportDecl *Import) { 1684 assert(!Import->getNextLocalImport() && 1685 "Import declaration already in the chain"); 1686 assert(!Import->isFromASTFile() && "Non-local import declaration"); 1687 if (!FirstLocalImport) { 1688 FirstLocalImport = Import; 1689 LastLocalImport = Import; 1690 return; 1691 } 1692 1693 LastLocalImport->setNextLocalImport(Import); 1694 LastLocalImport = Import; 1695 } 1696 1697 //===----------------------------------------------------------------------===// 1698 // Type Sizing and Analysis 1699 //===----------------------------------------------------------------------===// 1700 1701 /// getFloatTypeSemantics - Return the APFloat 'semantics' for the specified 1702 /// scalar floating point type. 1703 const llvm::fltSemantics &ASTContext::getFloatTypeSemantics(QualType T) const { 1704 switch (T->castAs<BuiltinType>()->getKind()) { 1705 default: 1706 llvm_unreachable("Not a floating point type!"); 1707 case BuiltinType::BFloat16: 1708 return Target->getBFloat16Format(); 1709 case BuiltinType::Float16: 1710 case BuiltinType::Half: 1711 return Target->getHalfFormat(); 1712 case BuiltinType::Float: return Target->getFloatFormat(); 1713 case BuiltinType::Double: return Target->getDoubleFormat(); 1714 case BuiltinType::Ibm128: 1715 return Target->getIbm128Format(); 1716 case BuiltinType::LongDouble: 1717 if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice) 1718 return AuxTarget->getLongDoubleFormat(); 1719 return Target->getLongDoubleFormat(); 1720 case BuiltinType::Float128: 1721 if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice) 1722 return AuxTarget->getFloat128Format(); 1723 return Target->getFloat128Format(); 1724 } 1725 } 1726 1727 CharUnits ASTContext::getDeclAlign(const Decl *D, bool ForAlignof) const { 1728 unsigned Align = Target->getCharWidth(); 1729 1730 bool UseAlignAttrOnly = false; 1731 if (unsigned AlignFromAttr = D->getMaxAlignment()) { 1732 Align = AlignFromAttr; 1733 1734 // __attribute__((aligned)) can increase or decrease alignment 1735 // *except* on a struct or struct member, where it only increases 1736 // alignment unless 'packed' is also specified. 1737 // 1738 // It is an error for alignas to decrease alignment, so we can 1739 // ignore that possibility; Sema should diagnose it. 1740 if (isa<FieldDecl>(D)) { 1741 UseAlignAttrOnly = D->hasAttr<PackedAttr>() || 1742 cast<FieldDecl>(D)->getParent()->hasAttr<PackedAttr>(); 1743 } else { 1744 UseAlignAttrOnly = true; 1745 } 1746 } 1747 else if (isa<FieldDecl>(D)) 1748 UseAlignAttrOnly = 1749 D->hasAttr<PackedAttr>() || 1750 cast<FieldDecl>(D)->getParent()->hasAttr<PackedAttr>(); 1751 1752 // If we're using the align attribute only, just ignore everything 1753 // else about the declaration and its type. 1754 if (UseAlignAttrOnly) { 1755 // do nothing 1756 } else if (const auto *VD = dyn_cast<ValueDecl>(D)) { 1757 QualType T = VD->getType(); 1758 if (const auto *RT = T->getAs<ReferenceType>()) { 1759 if (ForAlignof) 1760 T = RT->getPointeeType(); 1761 else 1762 T = getPointerType(RT->getPointeeType()); 1763 } 1764 QualType BaseT = getBaseElementType(T); 1765 if (T->isFunctionType()) 1766 Align = getTypeInfoImpl(T.getTypePtr()).Align; 1767 else if (!BaseT->isIncompleteType()) { 1768 // Adjust alignments of declarations with array type by the 1769 // large-array alignment on the target. 1770 if (const ArrayType *arrayType = getAsArrayType(T)) { 1771 unsigned MinWidth = Target->getLargeArrayMinWidth(); 1772 if (!ForAlignof && MinWidth) { 1773 if (isa<VariableArrayType>(arrayType)) 1774 Align = std::max(Align, Target->getLargeArrayAlign()); 1775 else if (isa<ConstantArrayType>(arrayType) && 1776 MinWidth <= getTypeSize(cast<ConstantArrayType>(arrayType))) 1777 Align = std::max(Align, Target->getLargeArrayAlign()); 1778 } 1779 } 1780 Align = std::max(Align, getPreferredTypeAlign(T.getTypePtr())); 1781 if (BaseT.getQualifiers().hasUnaligned()) 1782 Align = Target->getCharWidth(); 1783 if (const auto *VD = dyn_cast<VarDecl>(D)) { 1784 if (VD->hasGlobalStorage() && !ForAlignof) { 1785 uint64_t TypeSize = getTypeSize(T.getTypePtr()); 1786 Align = std::max(Align, getTargetInfo().getMinGlobalAlign(TypeSize)); 1787 } 1788 } 1789 } 1790 1791 // Fields can be subject to extra alignment constraints, like if 1792 // the field is packed, the struct is packed, or the struct has a 1793 // a max-field-alignment constraint (#pragma pack). So calculate 1794 // the actual alignment of the field within the struct, and then 1795 // (as we're expected to) constrain that by the alignment of the type. 1796 if (const auto *Field = dyn_cast<FieldDecl>(VD)) { 1797 const RecordDecl *Parent = Field->getParent(); 1798 // We can only produce a sensible answer if the record is valid. 1799 if (!Parent->isInvalidDecl()) { 1800 const ASTRecordLayout &Layout = getASTRecordLayout(Parent); 1801 1802 // Start with the record's overall alignment. 1803 unsigned FieldAlign = toBits(Layout.getAlignment()); 1804 1805 // Use the GCD of that and the offset within the record. 1806 uint64_t Offset = Layout.getFieldOffset(Field->getFieldIndex()); 1807 if (Offset > 0) { 1808 // Alignment is always a power of 2, so the GCD will be a power of 2, 1809 // which means we get to do this crazy thing instead of Euclid's. 1810 uint64_t LowBitOfOffset = Offset & (~Offset + 1); 1811 if (LowBitOfOffset < FieldAlign) 1812 FieldAlign = static_cast<unsigned>(LowBitOfOffset); 1813 } 1814 1815 Align = std::min(Align, FieldAlign); 1816 } 1817 } 1818 } 1819 1820 // Some targets have hard limitation on the maximum requestable alignment in 1821 // aligned attribute for static variables. 1822 const unsigned MaxAlignedAttr = getTargetInfo().getMaxAlignedAttribute(); 1823 const auto *VD = dyn_cast<VarDecl>(D); 1824 if (MaxAlignedAttr && VD && VD->getStorageClass() == SC_Static) 1825 Align = std::min(Align, MaxAlignedAttr); 1826 1827 return toCharUnitsFromBits(Align); 1828 } 1829 1830 CharUnits ASTContext::getExnObjectAlignment() const { 1831 return toCharUnitsFromBits(Target->getExnObjectAlignment()); 1832 } 1833 1834 // getTypeInfoDataSizeInChars - Return the size of a type, in 1835 // chars. If the type is a record, its data size is returned. This is 1836 // the size of the memcpy that's performed when assigning this type 1837 // using a trivial copy/move assignment operator. 1838 TypeInfoChars ASTContext::getTypeInfoDataSizeInChars(QualType T) const { 1839 TypeInfoChars Info = getTypeInfoInChars(T); 1840 1841 // In C++, objects can sometimes be allocated into the tail padding 1842 // of a base-class subobject. We decide whether that's possible 1843 // during class layout, so here we can just trust the layout results. 1844 if (getLangOpts().CPlusPlus) { 1845 if (const auto *RT = T->getAs<RecordType>()) { 1846 const ASTRecordLayout &layout = getASTRecordLayout(RT->getDecl()); 1847 Info.Width = layout.getDataSize(); 1848 } 1849 } 1850 1851 return Info; 1852 } 1853 1854 /// getConstantArrayInfoInChars - Performing the computation in CharUnits 1855 /// instead of in bits prevents overflowing the uint64_t for some large arrays. 1856 TypeInfoChars 1857 static getConstantArrayInfoInChars(const ASTContext &Context, 1858 const ConstantArrayType *CAT) { 1859 TypeInfoChars EltInfo = Context.getTypeInfoInChars(CAT->getElementType()); 1860 uint64_t Size = CAT->getSize().getZExtValue(); 1861 assert((Size == 0 || static_cast<uint64_t>(EltInfo.Width.getQuantity()) <= 1862 (uint64_t)(-1)/Size) && 1863 "Overflow in array type char size evaluation"); 1864 uint64_t Width = EltInfo.Width.getQuantity() * Size; 1865 unsigned Align = EltInfo.Align.getQuantity(); 1866 if (!Context.getTargetInfo().getCXXABI().isMicrosoft() || 1867 Context.getTargetInfo().getPointerWidth(0) == 64) 1868 Width = llvm::alignTo(Width, Align); 1869 return TypeInfoChars(CharUnits::fromQuantity(Width), 1870 CharUnits::fromQuantity(Align), 1871 EltInfo.AlignRequirement); 1872 } 1873 1874 TypeInfoChars ASTContext::getTypeInfoInChars(const Type *T) const { 1875 if (const auto *CAT = dyn_cast<ConstantArrayType>(T)) 1876 return getConstantArrayInfoInChars(*this, CAT); 1877 TypeInfo Info = getTypeInfo(T); 1878 return TypeInfoChars(toCharUnitsFromBits(Info.Width), 1879 toCharUnitsFromBits(Info.Align), Info.AlignRequirement); 1880 } 1881 1882 TypeInfoChars ASTContext::getTypeInfoInChars(QualType T) const { 1883 return getTypeInfoInChars(T.getTypePtr()); 1884 } 1885 1886 bool ASTContext::isAlignmentRequired(const Type *T) const { 1887 return getTypeInfo(T).AlignRequirement != AlignRequirementKind::None; 1888 } 1889 1890 bool ASTContext::isAlignmentRequired(QualType T) const { 1891 return isAlignmentRequired(T.getTypePtr()); 1892 } 1893 1894 unsigned ASTContext::getTypeAlignIfKnown(QualType T, 1895 bool NeedsPreferredAlignment) const { 1896 // An alignment on a typedef overrides anything else. 1897 if (const auto *TT = T->getAs<TypedefType>()) 1898 if (unsigned Align = TT->getDecl()->getMaxAlignment()) 1899 return Align; 1900 1901 // If we have an (array of) complete type, we're done. 1902 T = getBaseElementType(T); 1903 if (!T->isIncompleteType()) 1904 return NeedsPreferredAlignment ? getPreferredTypeAlign(T) : getTypeAlign(T); 1905 1906 // If we had an array type, its element type might be a typedef 1907 // type with an alignment attribute. 1908 if (const auto *TT = T->getAs<TypedefType>()) 1909 if (unsigned Align = TT->getDecl()->getMaxAlignment()) 1910 return Align; 1911 1912 // Otherwise, see if the declaration of the type had an attribute. 1913 if (const auto *TT = T->getAs<TagType>()) 1914 return TT->getDecl()->getMaxAlignment(); 1915 1916 return 0; 1917 } 1918 1919 TypeInfo ASTContext::getTypeInfo(const Type *T) const { 1920 TypeInfoMap::iterator I = MemoizedTypeInfo.find(T); 1921 if (I != MemoizedTypeInfo.end()) 1922 return I->second; 1923 1924 // This call can invalidate MemoizedTypeInfo[T], so we need a second lookup. 1925 TypeInfo TI = getTypeInfoImpl(T); 1926 MemoizedTypeInfo[T] = TI; 1927 return TI; 1928 } 1929 1930 /// getTypeInfoImpl - Return the size of the specified type, in bits. This 1931 /// method does not work on incomplete types. 1932 /// 1933 /// FIXME: Pointers into different addr spaces could have different sizes and 1934 /// alignment requirements: getPointerInfo should take an AddrSpace, this 1935 /// should take a QualType, &c. 1936 TypeInfo ASTContext::getTypeInfoImpl(const Type *T) const { 1937 uint64_t Width = 0; 1938 unsigned Align = 8; 1939 AlignRequirementKind AlignRequirement = AlignRequirementKind::None; 1940 unsigned AS = 0; 1941 switch (T->getTypeClass()) { 1942 #define TYPE(Class, Base) 1943 #define ABSTRACT_TYPE(Class, Base) 1944 #define NON_CANONICAL_TYPE(Class, Base) 1945 #define DEPENDENT_TYPE(Class, Base) case Type::Class: 1946 #define NON_CANONICAL_UNLESS_DEPENDENT_TYPE(Class, Base) \ 1947 case Type::Class: \ 1948 assert(!T->isDependentType() && "should not see dependent types here"); \ 1949 return getTypeInfo(cast<Class##Type>(T)->desugar().getTypePtr()); 1950 #include "clang/AST/TypeNodes.inc" 1951 llvm_unreachable("Should not see dependent types"); 1952 1953 case Type::FunctionNoProto: 1954 case Type::FunctionProto: 1955 // GCC extension: alignof(function) = 32 bits 1956 Width = 0; 1957 Align = 32; 1958 break; 1959 1960 case Type::IncompleteArray: 1961 case Type::VariableArray: 1962 case Type::ConstantArray: { 1963 // Model non-constant sized arrays as size zero, but track the alignment. 1964 uint64_t Size = 0; 1965 if (const auto *CAT = dyn_cast<ConstantArrayType>(T)) 1966 Size = CAT->getSize().getZExtValue(); 1967 1968 TypeInfo EltInfo = getTypeInfo(cast<ArrayType>(T)->getElementType()); 1969 assert((Size == 0 || EltInfo.Width <= (uint64_t)(-1) / Size) && 1970 "Overflow in array type bit size evaluation"); 1971 Width = EltInfo.Width * Size; 1972 Align = EltInfo.Align; 1973 AlignRequirement = EltInfo.AlignRequirement; 1974 if (!getTargetInfo().getCXXABI().isMicrosoft() || 1975 getTargetInfo().getPointerWidth(0) == 64) 1976 Width = llvm::alignTo(Width, Align); 1977 break; 1978 } 1979 1980 case Type::ExtVector: 1981 case Type::Vector: { 1982 const auto *VT = cast<VectorType>(T); 1983 TypeInfo EltInfo = getTypeInfo(VT->getElementType()); 1984 Width = EltInfo.Width * VT->getNumElements(); 1985 Align = Width; 1986 // If the alignment is not a power of 2, round up to the next power of 2. 1987 // This happens for non-power-of-2 length vectors. 1988 if (Align & (Align-1)) { 1989 Align = llvm::NextPowerOf2(Align); 1990 Width = llvm::alignTo(Width, Align); 1991 } 1992 // Adjust the alignment based on the target max. 1993 uint64_t TargetVectorAlign = Target->getMaxVectorAlign(); 1994 if (TargetVectorAlign && TargetVectorAlign < Align) 1995 Align = TargetVectorAlign; 1996 if (VT->getVectorKind() == VectorType::SveFixedLengthDataVector) 1997 // Adjust the alignment for fixed-length SVE vectors. This is important 1998 // for non-power-of-2 vector lengths. 1999 Align = 128; 2000 else if (VT->getVectorKind() == VectorType::SveFixedLengthPredicateVector) 2001 // Adjust the alignment for fixed-length SVE predicates. 2002 Align = 16; 2003 break; 2004 } 2005 2006 case Type::ConstantMatrix: { 2007 const auto *MT = cast<ConstantMatrixType>(T); 2008 TypeInfo ElementInfo = getTypeInfo(MT->getElementType()); 2009 // The internal layout of a matrix value is implementation defined. 2010 // Initially be ABI compatible with arrays with respect to alignment and 2011 // size. 2012 Width = ElementInfo.Width * MT->getNumRows() * MT->getNumColumns(); 2013 Align = ElementInfo.Align; 2014 break; 2015 } 2016 2017 case Type::Builtin: 2018 switch (cast<BuiltinType>(T)->getKind()) { 2019 default: llvm_unreachable("Unknown builtin type!"); 2020 case BuiltinType::Void: 2021 // GCC extension: alignof(void) = 8 bits. 2022 Width = 0; 2023 Align = 8; 2024 break; 2025 case BuiltinType::Bool: 2026 Width = Target->getBoolWidth(); 2027 Align = Target->getBoolAlign(); 2028 break; 2029 case BuiltinType::Char_S: 2030 case BuiltinType::Char_U: 2031 case BuiltinType::UChar: 2032 case BuiltinType::SChar: 2033 case BuiltinType::Char8: 2034 Width = Target->getCharWidth(); 2035 Align = Target->getCharAlign(); 2036 break; 2037 case BuiltinType::WChar_S: 2038 case BuiltinType::WChar_U: 2039 Width = Target->getWCharWidth(); 2040 Align = Target->getWCharAlign(); 2041 break; 2042 case BuiltinType::Char16: 2043 Width = Target->getChar16Width(); 2044 Align = Target->getChar16Align(); 2045 break; 2046 case BuiltinType::Char32: 2047 Width = Target->getChar32Width(); 2048 Align = Target->getChar32Align(); 2049 break; 2050 case BuiltinType::UShort: 2051 case BuiltinType::Short: 2052 Width = Target->getShortWidth(); 2053 Align = Target->getShortAlign(); 2054 break; 2055 case BuiltinType::UInt: 2056 case BuiltinType::Int: 2057 Width = Target->getIntWidth(); 2058 Align = Target->getIntAlign(); 2059 break; 2060 case BuiltinType::ULong: 2061 case BuiltinType::Long: 2062 Width = Target->getLongWidth(); 2063 Align = Target->getLongAlign(); 2064 break; 2065 case BuiltinType::ULongLong: 2066 case BuiltinType::LongLong: 2067 Width = Target->getLongLongWidth(); 2068 Align = Target->getLongLongAlign(); 2069 break; 2070 case BuiltinType::Int128: 2071 case BuiltinType::UInt128: 2072 Width = 128; 2073 Align = 128; // int128_t is 128-bit aligned on all targets. 2074 break; 2075 case BuiltinType::ShortAccum: 2076 case BuiltinType::UShortAccum: 2077 case BuiltinType::SatShortAccum: 2078 case BuiltinType::SatUShortAccum: 2079 Width = Target->getShortAccumWidth(); 2080 Align = Target->getShortAccumAlign(); 2081 break; 2082 case BuiltinType::Accum: 2083 case BuiltinType::UAccum: 2084 case BuiltinType::SatAccum: 2085 case BuiltinType::SatUAccum: 2086 Width = Target->getAccumWidth(); 2087 Align = Target->getAccumAlign(); 2088 break; 2089 case BuiltinType::LongAccum: 2090 case BuiltinType::ULongAccum: 2091 case BuiltinType::SatLongAccum: 2092 case BuiltinType::SatULongAccum: 2093 Width = Target->getLongAccumWidth(); 2094 Align = Target->getLongAccumAlign(); 2095 break; 2096 case BuiltinType::ShortFract: 2097 case BuiltinType::UShortFract: 2098 case BuiltinType::SatShortFract: 2099 case BuiltinType::SatUShortFract: 2100 Width = Target->getShortFractWidth(); 2101 Align = Target->getShortFractAlign(); 2102 break; 2103 case BuiltinType::Fract: 2104 case BuiltinType::UFract: 2105 case BuiltinType::SatFract: 2106 case BuiltinType::SatUFract: 2107 Width = Target->getFractWidth(); 2108 Align = Target->getFractAlign(); 2109 break; 2110 case BuiltinType::LongFract: 2111 case BuiltinType::ULongFract: 2112 case BuiltinType::SatLongFract: 2113 case BuiltinType::SatULongFract: 2114 Width = Target->getLongFractWidth(); 2115 Align = Target->getLongFractAlign(); 2116 break; 2117 case BuiltinType::BFloat16: 2118 Width = Target->getBFloat16Width(); 2119 Align = Target->getBFloat16Align(); 2120 break; 2121 case BuiltinType::Float16: 2122 case BuiltinType::Half: 2123 if (Target->hasFloat16Type() || !getLangOpts().OpenMP || 2124 !getLangOpts().OpenMPIsDevice) { 2125 Width = Target->getHalfWidth(); 2126 Align = Target->getHalfAlign(); 2127 } else { 2128 assert(getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice && 2129 "Expected OpenMP device compilation."); 2130 Width = AuxTarget->getHalfWidth(); 2131 Align = AuxTarget->getHalfAlign(); 2132 } 2133 break; 2134 case BuiltinType::Float: 2135 Width = Target->getFloatWidth(); 2136 Align = Target->getFloatAlign(); 2137 break; 2138 case BuiltinType::Double: 2139 Width = Target->getDoubleWidth(); 2140 Align = Target->getDoubleAlign(); 2141 break; 2142 case BuiltinType::Ibm128: 2143 Width = Target->getIbm128Width(); 2144 Align = Target->getIbm128Align(); 2145 break; 2146 case BuiltinType::LongDouble: 2147 if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice && 2148 (Target->getLongDoubleWidth() != AuxTarget->getLongDoubleWidth() || 2149 Target->getLongDoubleAlign() != AuxTarget->getLongDoubleAlign())) { 2150 Width = AuxTarget->getLongDoubleWidth(); 2151 Align = AuxTarget->getLongDoubleAlign(); 2152 } else { 2153 Width = Target->getLongDoubleWidth(); 2154 Align = Target->getLongDoubleAlign(); 2155 } 2156 break; 2157 case BuiltinType::Float128: 2158 if (Target->hasFloat128Type() || !getLangOpts().OpenMP || 2159 !getLangOpts().OpenMPIsDevice) { 2160 Width = Target->getFloat128Width(); 2161 Align = Target->getFloat128Align(); 2162 } else { 2163 assert(getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice && 2164 "Expected OpenMP device compilation."); 2165 Width = AuxTarget->getFloat128Width(); 2166 Align = AuxTarget->getFloat128Align(); 2167 } 2168 break; 2169 case BuiltinType::NullPtr: 2170 Width = Target->getPointerWidth(0); // C++ 3.9.1p11: sizeof(nullptr_t) 2171 Align = Target->getPointerAlign(0); // == sizeof(void*) 2172 break; 2173 case BuiltinType::ObjCId: 2174 case BuiltinType::ObjCClass: 2175 case BuiltinType::ObjCSel: 2176 Width = Target->getPointerWidth(0); 2177 Align = Target->getPointerAlign(0); 2178 break; 2179 case BuiltinType::OCLSampler: 2180 case BuiltinType::OCLEvent: 2181 case BuiltinType::OCLClkEvent: 2182 case BuiltinType::OCLQueue: 2183 case BuiltinType::OCLReserveID: 2184 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \ 2185 case BuiltinType::Id: 2186 #include "clang/Basic/OpenCLImageTypes.def" 2187 #define EXT_OPAQUE_TYPE(ExtType, Id, Ext) \ 2188 case BuiltinType::Id: 2189 #include "clang/Basic/OpenCLExtensionTypes.def" 2190 AS = getTargetAddressSpace( 2191 Target->getOpenCLTypeAddrSpace(getOpenCLTypeKind(T))); 2192 Width = Target->getPointerWidth(AS); 2193 Align = Target->getPointerAlign(AS); 2194 break; 2195 // The SVE types are effectively target-specific. The length of an 2196 // SVE_VECTOR_TYPE is only known at runtime, but it is always a multiple 2197 // of 128 bits. There is one predicate bit for each vector byte, so the 2198 // length of an SVE_PREDICATE_TYPE is always a multiple of 16 bits. 2199 // 2200 // Because the length is only known at runtime, we use a dummy value 2201 // of 0 for the static length. The alignment values are those defined 2202 // by the Procedure Call Standard for the Arm Architecture. 2203 #define SVE_VECTOR_TYPE(Name, MangledName, Id, SingletonId, NumEls, ElBits, \ 2204 IsSigned, IsFP, IsBF) \ 2205 case BuiltinType::Id: \ 2206 Width = 0; \ 2207 Align = 128; \ 2208 break; 2209 #define SVE_PREDICATE_TYPE(Name, MangledName, Id, SingletonId, NumEls) \ 2210 case BuiltinType::Id: \ 2211 Width = 0; \ 2212 Align = 16; \ 2213 break; 2214 #include "clang/Basic/AArch64SVEACLETypes.def" 2215 #define PPC_VECTOR_TYPE(Name, Id, Size) \ 2216 case BuiltinType::Id: \ 2217 Width = Size; \ 2218 Align = Size; \ 2219 break; 2220 #include "clang/Basic/PPCTypes.def" 2221 #define RVV_VECTOR_TYPE(Name, Id, SingletonId, ElKind, ElBits, NF, IsSigned, \ 2222 IsFP) \ 2223 case BuiltinType::Id: \ 2224 Width = 0; \ 2225 Align = ElBits; \ 2226 break; 2227 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, ElKind) \ 2228 case BuiltinType::Id: \ 2229 Width = 0; \ 2230 Align = 8; \ 2231 break; 2232 #include "clang/Basic/RISCVVTypes.def" 2233 } 2234 break; 2235 case Type::ObjCObjectPointer: 2236 Width = Target->getPointerWidth(0); 2237 Align = Target->getPointerAlign(0); 2238 break; 2239 case Type::BlockPointer: 2240 AS = getTargetAddressSpace(cast<BlockPointerType>(T)->getPointeeType()); 2241 Width = Target->getPointerWidth(AS); 2242 Align = Target->getPointerAlign(AS); 2243 break; 2244 case Type::LValueReference: 2245 case Type::RValueReference: 2246 // alignof and sizeof should never enter this code path here, so we go 2247 // the pointer route. 2248 AS = getTargetAddressSpace(cast<ReferenceType>(T)->getPointeeType()); 2249 Width = Target->getPointerWidth(AS); 2250 Align = Target->getPointerAlign(AS); 2251 break; 2252 case Type::Pointer: 2253 AS = getTargetAddressSpace(cast<PointerType>(T)->getPointeeType()); 2254 Width = Target->getPointerWidth(AS); 2255 Align = Target->getPointerAlign(AS); 2256 break; 2257 case Type::MemberPointer: { 2258 const auto *MPT = cast<MemberPointerType>(T); 2259 CXXABI::MemberPointerInfo MPI = ABI->getMemberPointerInfo(MPT); 2260 Width = MPI.Width; 2261 Align = MPI.Align; 2262 break; 2263 } 2264 case Type::Complex: { 2265 // Complex types have the same alignment as their elements, but twice the 2266 // size. 2267 TypeInfo EltInfo = getTypeInfo(cast<ComplexType>(T)->getElementType()); 2268 Width = EltInfo.Width * 2; 2269 Align = EltInfo.Align; 2270 break; 2271 } 2272 case Type::ObjCObject: 2273 return getTypeInfo(cast<ObjCObjectType>(T)->getBaseType().getTypePtr()); 2274 case Type::Adjusted: 2275 case Type::Decayed: 2276 return getTypeInfo(cast<AdjustedType>(T)->getAdjustedType().getTypePtr()); 2277 case Type::ObjCInterface: { 2278 const auto *ObjCI = cast<ObjCInterfaceType>(T); 2279 if (ObjCI->getDecl()->isInvalidDecl()) { 2280 Width = 8; 2281 Align = 8; 2282 break; 2283 } 2284 const ASTRecordLayout &Layout = getASTObjCInterfaceLayout(ObjCI->getDecl()); 2285 Width = toBits(Layout.getSize()); 2286 Align = toBits(Layout.getAlignment()); 2287 break; 2288 } 2289 case Type::ExtInt: { 2290 const auto *EIT = cast<ExtIntType>(T); 2291 Align = 2292 std::min(static_cast<unsigned>(std::max( 2293 getCharWidth(), llvm::PowerOf2Ceil(EIT->getNumBits()))), 2294 Target->getLongLongAlign()); 2295 Width = llvm::alignTo(EIT->getNumBits(), Align); 2296 break; 2297 } 2298 case Type::Record: 2299 case Type::Enum: { 2300 const auto *TT = cast<TagType>(T); 2301 2302 if (TT->getDecl()->isInvalidDecl()) { 2303 Width = 8; 2304 Align = 8; 2305 break; 2306 } 2307 2308 if (const auto *ET = dyn_cast<EnumType>(TT)) { 2309 const EnumDecl *ED = ET->getDecl(); 2310 TypeInfo Info = 2311 getTypeInfo(ED->getIntegerType()->getUnqualifiedDesugaredType()); 2312 if (unsigned AttrAlign = ED->getMaxAlignment()) { 2313 Info.Align = AttrAlign; 2314 Info.AlignRequirement = AlignRequirementKind::RequiredByEnum; 2315 } 2316 return Info; 2317 } 2318 2319 const auto *RT = cast<RecordType>(TT); 2320 const RecordDecl *RD = RT->getDecl(); 2321 const ASTRecordLayout &Layout = getASTRecordLayout(RD); 2322 Width = toBits(Layout.getSize()); 2323 Align = toBits(Layout.getAlignment()); 2324 AlignRequirement = RD->hasAttr<AlignedAttr>() 2325 ? AlignRequirementKind::RequiredByRecord 2326 : AlignRequirementKind::None; 2327 break; 2328 } 2329 2330 case Type::SubstTemplateTypeParm: 2331 return getTypeInfo(cast<SubstTemplateTypeParmType>(T)-> 2332 getReplacementType().getTypePtr()); 2333 2334 case Type::Auto: 2335 case Type::DeducedTemplateSpecialization: { 2336 const auto *A = cast<DeducedType>(T); 2337 assert(!A->getDeducedType().isNull() && 2338 "cannot request the size of an undeduced or dependent auto type"); 2339 return getTypeInfo(A->getDeducedType().getTypePtr()); 2340 } 2341 2342 case Type::Paren: 2343 return getTypeInfo(cast<ParenType>(T)->getInnerType().getTypePtr()); 2344 2345 case Type::MacroQualified: 2346 return getTypeInfo( 2347 cast<MacroQualifiedType>(T)->getUnderlyingType().getTypePtr()); 2348 2349 case Type::ObjCTypeParam: 2350 return getTypeInfo(cast<ObjCTypeParamType>(T)->desugar().getTypePtr()); 2351 2352 case Type::Typedef: { 2353 const TypedefNameDecl *Typedef = cast<TypedefType>(T)->getDecl(); 2354 TypeInfo Info = getTypeInfo(Typedef->getUnderlyingType().getTypePtr()); 2355 // If the typedef has an aligned attribute on it, it overrides any computed 2356 // alignment we have. This violates the GCC documentation (which says that 2357 // attribute(aligned) can only round up) but matches its implementation. 2358 if (unsigned AttrAlign = Typedef->getMaxAlignment()) { 2359 Align = AttrAlign; 2360 AlignRequirement = AlignRequirementKind::RequiredByTypedef; 2361 } else { 2362 Align = Info.Align; 2363 AlignRequirement = Info.AlignRequirement; 2364 } 2365 Width = Info.Width; 2366 break; 2367 } 2368 2369 case Type::Elaborated: 2370 return getTypeInfo(cast<ElaboratedType>(T)->getNamedType().getTypePtr()); 2371 2372 case Type::Attributed: 2373 return getTypeInfo( 2374 cast<AttributedType>(T)->getEquivalentType().getTypePtr()); 2375 2376 case Type::Atomic: { 2377 // Start with the base type information. 2378 TypeInfo Info = getTypeInfo(cast<AtomicType>(T)->getValueType()); 2379 Width = Info.Width; 2380 Align = Info.Align; 2381 2382 if (!Width) { 2383 // An otherwise zero-sized type should still generate an 2384 // atomic operation. 2385 Width = Target->getCharWidth(); 2386 assert(Align); 2387 } else if (Width <= Target->getMaxAtomicPromoteWidth()) { 2388 // If the size of the type doesn't exceed the platform's max 2389 // atomic promotion width, make the size and alignment more 2390 // favorable to atomic operations: 2391 2392 // Round the size up to a power of 2. 2393 if (!llvm::isPowerOf2_64(Width)) 2394 Width = llvm::NextPowerOf2(Width); 2395 2396 // Set the alignment equal to the size. 2397 Align = static_cast<unsigned>(Width); 2398 } 2399 } 2400 break; 2401 2402 case Type::Pipe: 2403 Width = Target->getPointerWidth(getTargetAddressSpace(LangAS::opencl_global)); 2404 Align = Target->getPointerAlign(getTargetAddressSpace(LangAS::opencl_global)); 2405 break; 2406 } 2407 2408 assert(llvm::isPowerOf2_32(Align) && "Alignment must be power of 2"); 2409 return TypeInfo(Width, Align, AlignRequirement); 2410 } 2411 2412 unsigned ASTContext::getTypeUnadjustedAlign(const Type *T) const { 2413 UnadjustedAlignMap::iterator I = MemoizedUnadjustedAlign.find(T); 2414 if (I != MemoizedUnadjustedAlign.end()) 2415 return I->second; 2416 2417 unsigned UnadjustedAlign; 2418 if (const auto *RT = T->getAs<RecordType>()) { 2419 const RecordDecl *RD = RT->getDecl(); 2420 const ASTRecordLayout &Layout = getASTRecordLayout(RD); 2421 UnadjustedAlign = toBits(Layout.getUnadjustedAlignment()); 2422 } else if (const auto *ObjCI = T->getAs<ObjCInterfaceType>()) { 2423 const ASTRecordLayout &Layout = getASTObjCInterfaceLayout(ObjCI->getDecl()); 2424 UnadjustedAlign = toBits(Layout.getUnadjustedAlignment()); 2425 } else { 2426 UnadjustedAlign = getTypeAlign(T->getUnqualifiedDesugaredType()); 2427 } 2428 2429 MemoizedUnadjustedAlign[T] = UnadjustedAlign; 2430 return UnadjustedAlign; 2431 } 2432 2433 unsigned ASTContext::getOpenMPDefaultSimdAlign(QualType T) const { 2434 unsigned SimdAlign = getTargetInfo().getSimdDefaultAlign(); 2435 return SimdAlign; 2436 } 2437 2438 /// toCharUnitsFromBits - Convert a size in bits to a size in characters. 2439 CharUnits ASTContext::toCharUnitsFromBits(int64_t BitSize) const { 2440 return CharUnits::fromQuantity(BitSize / getCharWidth()); 2441 } 2442 2443 /// toBits - Convert a size in characters to a size in characters. 2444 int64_t ASTContext::toBits(CharUnits CharSize) const { 2445 return CharSize.getQuantity() * getCharWidth(); 2446 } 2447 2448 /// getTypeSizeInChars - Return the size of the specified type, in characters. 2449 /// This method does not work on incomplete types. 2450 CharUnits ASTContext::getTypeSizeInChars(QualType T) const { 2451 return getTypeInfoInChars(T).Width; 2452 } 2453 CharUnits ASTContext::getTypeSizeInChars(const Type *T) const { 2454 return getTypeInfoInChars(T).Width; 2455 } 2456 2457 /// getTypeAlignInChars - Return the ABI-specified alignment of a type, in 2458 /// characters. This method does not work on incomplete types. 2459 CharUnits ASTContext::getTypeAlignInChars(QualType T) const { 2460 return toCharUnitsFromBits(getTypeAlign(T)); 2461 } 2462 CharUnits ASTContext::getTypeAlignInChars(const Type *T) const { 2463 return toCharUnitsFromBits(getTypeAlign(T)); 2464 } 2465 2466 /// getTypeUnadjustedAlignInChars - Return the ABI-specified alignment of a 2467 /// type, in characters, before alignment adustments. This method does 2468 /// not work on incomplete types. 2469 CharUnits ASTContext::getTypeUnadjustedAlignInChars(QualType T) const { 2470 return toCharUnitsFromBits(getTypeUnadjustedAlign(T)); 2471 } 2472 CharUnits ASTContext::getTypeUnadjustedAlignInChars(const Type *T) const { 2473 return toCharUnitsFromBits(getTypeUnadjustedAlign(T)); 2474 } 2475 2476 /// getPreferredTypeAlign - Return the "preferred" alignment of the specified 2477 /// type for the current target in bits. This can be different than the ABI 2478 /// alignment in cases where it is beneficial for performance or backwards 2479 /// compatibility preserving to overalign a data type. (Note: despite the name, 2480 /// the preferred alignment is ABI-impacting, and not an optimization.) 2481 unsigned ASTContext::getPreferredTypeAlign(const Type *T) const { 2482 TypeInfo TI = getTypeInfo(T); 2483 unsigned ABIAlign = TI.Align; 2484 2485 T = T->getBaseElementTypeUnsafe(); 2486 2487 // The preferred alignment of member pointers is that of a pointer. 2488 if (T->isMemberPointerType()) 2489 return getPreferredTypeAlign(getPointerDiffType().getTypePtr()); 2490 2491 if (!Target->allowsLargerPreferedTypeAlignment()) 2492 return ABIAlign; 2493 2494 if (const auto *RT = T->getAs<RecordType>()) { 2495 const RecordDecl *RD = RT->getDecl(); 2496 2497 // When used as part of a typedef, or together with a 'packed' attribute, 2498 // the 'aligned' attribute can be used to decrease alignment. Note that the 2499 // 'packed' case is already taken into consideration when computing the 2500 // alignment, we only need to handle the typedef case here. 2501 if (TI.AlignRequirement == AlignRequirementKind::RequiredByTypedef || 2502 RD->isInvalidDecl()) 2503 return ABIAlign; 2504 2505 unsigned PreferredAlign = static_cast<unsigned>( 2506 toBits(getASTRecordLayout(RD).PreferredAlignment)); 2507 assert(PreferredAlign >= ABIAlign && 2508 "PreferredAlign should be at least as large as ABIAlign."); 2509 return PreferredAlign; 2510 } 2511 2512 // Double (and, for targets supporting AIX `power` alignment, long double) and 2513 // long long should be naturally aligned (despite requiring less alignment) if 2514 // possible. 2515 if (const auto *CT = T->getAs<ComplexType>()) 2516 T = CT->getElementType().getTypePtr(); 2517 if (const auto *ET = T->getAs<EnumType>()) 2518 T = ET->getDecl()->getIntegerType().getTypePtr(); 2519 if (T->isSpecificBuiltinType(BuiltinType::Double) || 2520 T->isSpecificBuiltinType(BuiltinType::LongLong) || 2521 T->isSpecificBuiltinType(BuiltinType::ULongLong) || 2522 (T->isSpecificBuiltinType(BuiltinType::LongDouble) && 2523 Target->defaultsToAIXPowerAlignment())) 2524 // Don't increase the alignment if an alignment attribute was specified on a 2525 // typedef declaration. 2526 if (!TI.isAlignRequired()) 2527 return std::max(ABIAlign, (unsigned)getTypeSize(T)); 2528 2529 return ABIAlign; 2530 } 2531 2532 /// getTargetDefaultAlignForAttributeAligned - Return the default alignment 2533 /// for __attribute__((aligned)) on this target, to be used if no alignment 2534 /// value is specified. 2535 unsigned ASTContext::getTargetDefaultAlignForAttributeAligned() const { 2536 return getTargetInfo().getDefaultAlignForAttributeAligned(); 2537 } 2538 2539 /// getAlignOfGlobalVar - Return the alignment in bits that should be given 2540 /// to a global variable of the specified type. 2541 unsigned ASTContext::getAlignOfGlobalVar(QualType T) const { 2542 uint64_t TypeSize = getTypeSize(T.getTypePtr()); 2543 return std::max(getPreferredTypeAlign(T), 2544 getTargetInfo().getMinGlobalAlign(TypeSize)); 2545 } 2546 2547 /// getAlignOfGlobalVarInChars - Return the alignment in characters that 2548 /// should be given to a global variable of the specified type. 2549 CharUnits ASTContext::getAlignOfGlobalVarInChars(QualType T) const { 2550 return toCharUnitsFromBits(getAlignOfGlobalVar(T)); 2551 } 2552 2553 CharUnits ASTContext::getOffsetOfBaseWithVBPtr(const CXXRecordDecl *RD) const { 2554 CharUnits Offset = CharUnits::Zero(); 2555 const ASTRecordLayout *Layout = &getASTRecordLayout(RD); 2556 while (const CXXRecordDecl *Base = Layout->getBaseSharingVBPtr()) { 2557 Offset += Layout->getBaseClassOffset(Base); 2558 Layout = &getASTRecordLayout(Base); 2559 } 2560 return Offset; 2561 } 2562 2563 CharUnits ASTContext::getMemberPointerPathAdjustment(const APValue &MP) const { 2564 const ValueDecl *MPD = MP.getMemberPointerDecl(); 2565 CharUnits ThisAdjustment = CharUnits::Zero(); 2566 ArrayRef<const CXXRecordDecl*> Path = MP.getMemberPointerPath(); 2567 bool DerivedMember = MP.isMemberPointerToDerivedMember(); 2568 const CXXRecordDecl *RD = cast<CXXRecordDecl>(MPD->getDeclContext()); 2569 for (unsigned I = 0, N = Path.size(); I != N; ++I) { 2570 const CXXRecordDecl *Base = RD; 2571 const CXXRecordDecl *Derived = Path[I]; 2572 if (DerivedMember) 2573 std::swap(Base, Derived); 2574 ThisAdjustment += getASTRecordLayout(Derived).getBaseClassOffset(Base); 2575 RD = Path[I]; 2576 } 2577 if (DerivedMember) 2578 ThisAdjustment = -ThisAdjustment; 2579 return ThisAdjustment; 2580 } 2581 2582 /// DeepCollectObjCIvars - 2583 /// This routine first collects all declared, but not synthesized, ivars in 2584 /// super class and then collects all ivars, including those synthesized for 2585 /// current class. This routine is used for implementation of current class 2586 /// when all ivars, declared and synthesized are known. 2587 void ASTContext::DeepCollectObjCIvars(const ObjCInterfaceDecl *OI, 2588 bool leafClass, 2589 SmallVectorImpl<const ObjCIvarDecl*> &Ivars) const { 2590 if (const ObjCInterfaceDecl *SuperClass = OI->getSuperClass()) 2591 DeepCollectObjCIvars(SuperClass, false, Ivars); 2592 if (!leafClass) { 2593 for (const auto *I : OI->ivars()) 2594 Ivars.push_back(I); 2595 } else { 2596 auto *IDecl = const_cast<ObjCInterfaceDecl *>(OI); 2597 for (const ObjCIvarDecl *Iv = IDecl->all_declared_ivar_begin(); Iv; 2598 Iv= Iv->getNextIvar()) 2599 Ivars.push_back(Iv); 2600 } 2601 } 2602 2603 /// CollectInheritedProtocols - Collect all protocols in current class and 2604 /// those inherited by it. 2605 void ASTContext::CollectInheritedProtocols(const Decl *CDecl, 2606 llvm::SmallPtrSet<ObjCProtocolDecl*, 8> &Protocols) { 2607 if (const auto *OI = dyn_cast<ObjCInterfaceDecl>(CDecl)) { 2608 // We can use protocol_iterator here instead of 2609 // all_referenced_protocol_iterator since we are walking all categories. 2610 for (auto *Proto : OI->all_referenced_protocols()) { 2611 CollectInheritedProtocols(Proto, Protocols); 2612 } 2613 2614 // Categories of this Interface. 2615 for (const auto *Cat : OI->visible_categories()) 2616 CollectInheritedProtocols(Cat, Protocols); 2617 2618 if (ObjCInterfaceDecl *SD = OI->getSuperClass()) 2619 while (SD) { 2620 CollectInheritedProtocols(SD, Protocols); 2621 SD = SD->getSuperClass(); 2622 } 2623 } else if (const auto *OC = dyn_cast<ObjCCategoryDecl>(CDecl)) { 2624 for (auto *Proto : OC->protocols()) { 2625 CollectInheritedProtocols(Proto, Protocols); 2626 } 2627 } else if (const auto *OP = dyn_cast<ObjCProtocolDecl>(CDecl)) { 2628 // Insert the protocol. 2629 if (!Protocols.insert( 2630 const_cast<ObjCProtocolDecl *>(OP->getCanonicalDecl())).second) 2631 return; 2632 2633 for (auto *Proto : OP->protocols()) 2634 CollectInheritedProtocols(Proto, Protocols); 2635 } 2636 } 2637 2638 static bool unionHasUniqueObjectRepresentations(const ASTContext &Context, 2639 const RecordDecl *RD) { 2640 assert(RD->isUnion() && "Must be union type"); 2641 CharUnits UnionSize = Context.getTypeSizeInChars(RD->getTypeForDecl()); 2642 2643 for (const auto *Field : RD->fields()) { 2644 if (!Context.hasUniqueObjectRepresentations(Field->getType())) 2645 return false; 2646 CharUnits FieldSize = Context.getTypeSizeInChars(Field->getType()); 2647 if (FieldSize != UnionSize) 2648 return false; 2649 } 2650 return !RD->field_empty(); 2651 } 2652 2653 static int64_t getSubobjectOffset(const FieldDecl *Field, 2654 const ASTContext &Context, 2655 const clang::ASTRecordLayout & /*Layout*/) { 2656 return Context.getFieldOffset(Field); 2657 } 2658 2659 static int64_t getSubobjectOffset(const CXXRecordDecl *RD, 2660 const ASTContext &Context, 2661 const clang::ASTRecordLayout &Layout) { 2662 return Context.toBits(Layout.getBaseClassOffset(RD)); 2663 } 2664 2665 static llvm::Optional<int64_t> 2666 structHasUniqueObjectRepresentations(const ASTContext &Context, 2667 const RecordDecl *RD); 2668 2669 static llvm::Optional<int64_t> 2670 getSubobjectSizeInBits(const FieldDecl *Field, const ASTContext &Context) { 2671 if (Field->getType()->isRecordType()) { 2672 const RecordDecl *RD = Field->getType()->getAsRecordDecl(); 2673 if (!RD->isUnion()) 2674 return structHasUniqueObjectRepresentations(Context, RD); 2675 } 2676 if (!Field->getType()->isReferenceType() && 2677 !Context.hasUniqueObjectRepresentations(Field->getType())) 2678 return llvm::None; 2679 2680 int64_t FieldSizeInBits = 2681 Context.toBits(Context.getTypeSizeInChars(Field->getType())); 2682 if (Field->isBitField()) { 2683 int64_t BitfieldSize = Field->getBitWidthValue(Context); 2684 if (BitfieldSize > FieldSizeInBits) 2685 return llvm::None; 2686 FieldSizeInBits = BitfieldSize; 2687 } 2688 return FieldSizeInBits; 2689 } 2690 2691 static llvm::Optional<int64_t> 2692 getSubobjectSizeInBits(const CXXRecordDecl *RD, const ASTContext &Context) { 2693 return structHasUniqueObjectRepresentations(Context, RD); 2694 } 2695 2696 template <typename RangeT> 2697 static llvm::Optional<int64_t> structSubobjectsHaveUniqueObjectRepresentations( 2698 const RangeT &Subobjects, int64_t CurOffsetInBits, 2699 const ASTContext &Context, const clang::ASTRecordLayout &Layout) { 2700 for (const auto *Subobject : Subobjects) { 2701 llvm::Optional<int64_t> SizeInBits = 2702 getSubobjectSizeInBits(Subobject, Context); 2703 if (!SizeInBits) 2704 return llvm::None; 2705 if (*SizeInBits != 0) { 2706 int64_t Offset = getSubobjectOffset(Subobject, Context, Layout); 2707 if (Offset != CurOffsetInBits) 2708 return llvm::None; 2709 CurOffsetInBits += *SizeInBits; 2710 } 2711 } 2712 return CurOffsetInBits; 2713 } 2714 2715 static llvm::Optional<int64_t> 2716 structHasUniqueObjectRepresentations(const ASTContext &Context, 2717 const RecordDecl *RD) { 2718 assert(!RD->isUnion() && "Must be struct/class type"); 2719 const auto &Layout = Context.getASTRecordLayout(RD); 2720 2721 int64_t CurOffsetInBits = 0; 2722 if (const auto *ClassDecl = dyn_cast<CXXRecordDecl>(RD)) { 2723 if (ClassDecl->isDynamicClass()) 2724 return llvm::None; 2725 2726 SmallVector<CXXRecordDecl *, 4> Bases; 2727 for (const auto &Base : ClassDecl->bases()) { 2728 // Empty types can be inherited from, and non-empty types can potentially 2729 // have tail padding, so just make sure there isn't an error. 2730 Bases.emplace_back(Base.getType()->getAsCXXRecordDecl()); 2731 } 2732 2733 llvm::sort(Bases, [&](const CXXRecordDecl *L, const CXXRecordDecl *R) { 2734 return Layout.getBaseClassOffset(L) < Layout.getBaseClassOffset(R); 2735 }); 2736 2737 llvm::Optional<int64_t> OffsetAfterBases = 2738 structSubobjectsHaveUniqueObjectRepresentations(Bases, CurOffsetInBits, 2739 Context, Layout); 2740 if (!OffsetAfterBases) 2741 return llvm::None; 2742 CurOffsetInBits = *OffsetAfterBases; 2743 } 2744 2745 llvm::Optional<int64_t> OffsetAfterFields = 2746 structSubobjectsHaveUniqueObjectRepresentations( 2747 RD->fields(), CurOffsetInBits, Context, Layout); 2748 if (!OffsetAfterFields) 2749 return llvm::None; 2750 CurOffsetInBits = *OffsetAfterFields; 2751 2752 return CurOffsetInBits; 2753 } 2754 2755 bool ASTContext::hasUniqueObjectRepresentations(QualType Ty) const { 2756 // C++17 [meta.unary.prop]: 2757 // The predicate condition for a template specialization 2758 // has_unique_object_representations<T> shall be 2759 // satisfied if and only if: 2760 // (9.1) - T is trivially copyable, and 2761 // (9.2) - any two objects of type T with the same value have the same 2762 // object representation, where two objects 2763 // of array or non-union class type are considered to have the same value 2764 // if their respective sequences of 2765 // direct subobjects have the same values, and two objects of union type 2766 // are considered to have the same 2767 // value if they have the same active member and the corresponding members 2768 // have the same value. 2769 // The set of scalar types for which this condition holds is 2770 // implementation-defined. [ Note: If a type has padding 2771 // bits, the condition does not hold; otherwise, the condition holds true 2772 // for unsigned integral types. -- end note ] 2773 assert(!Ty.isNull() && "Null QualType sent to unique object rep check"); 2774 2775 // Arrays are unique only if their element type is unique. 2776 if (Ty->isArrayType()) 2777 return hasUniqueObjectRepresentations(getBaseElementType(Ty)); 2778 2779 // (9.1) - T is trivially copyable... 2780 if (!Ty.isTriviallyCopyableType(*this)) 2781 return false; 2782 2783 // All integrals and enums are unique. 2784 if (Ty->isIntegralOrEnumerationType()) 2785 return true; 2786 2787 // All other pointers are unique. 2788 if (Ty->isPointerType()) 2789 return true; 2790 2791 if (Ty->isMemberPointerType()) { 2792 const auto *MPT = Ty->getAs<MemberPointerType>(); 2793 return !ABI->getMemberPointerInfo(MPT).HasPadding; 2794 } 2795 2796 if (Ty->isRecordType()) { 2797 const RecordDecl *Record = Ty->castAs<RecordType>()->getDecl(); 2798 2799 if (Record->isInvalidDecl()) 2800 return false; 2801 2802 if (Record->isUnion()) 2803 return unionHasUniqueObjectRepresentations(*this, Record); 2804 2805 Optional<int64_t> StructSize = 2806 structHasUniqueObjectRepresentations(*this, Record); 2807 2808 return StructSize && 2809 StructSize.getValue() == static_cast<int64_t>(getTypeSize(Ty)); 2810 } 2811 2812 // FIXME: More cases to handle here (list by rsmith): 2813 // vectors (careful about, eg, vector of 3 foo) 2814 // _Complex int and friends 2815 // _Atomic T 2816 // Obj-C block pointers 2817 // Obj-C object pointers 2818 // and perhaps OpenCL's various builtin types (pipe, sampler_t, event_t, 2819 // clk_event_t, queue_t, reserve_id_t) 2820 // There're also Obj-C class types and the Obj-C selector type, but I think it 2821 // makes sense for those to return false here. 2822 2823 return false; 2824 } 2825 2826 unsigned ASTContext::CountNonClassIvars(const ObjCInterfaceDecl *OI) const { 2827 unsigned count = 0; 2828 // Count ivars declared in class extension. 2829 for (const auto *Ext : OI->known_extensions()) 2830 count += Ext->ivar_size(); 2831 2832 // Count ivar defined in this class's implementation. This 2833 // includes synthesized ivars. 2834 if (ObjCImplementationDecl *ImplDecl = OI->getImplementation()) 2835 count += ImplDecl->ivar_size(); 2836 2837 return count; 2838 } 2839 2840 bool ASTContext::isSentinelNullExpr(const Expr *E) { 2841 if (!E) 2842 return false; 2843 2844 // nullptr_t is always treated as null. 2845 if (E->getType()->isNullPtrType()) return true; 2846 2847 if (E->getType()->isAnyPointerType() && 2848 E->IgnoreParenCasts()->isNullPointerConstant(*this, 2849 Expr::NPC_ValueDependentIsNull)) 2850 return true; 2851 2852 // Unfortunately, __null has type 'int'. 2853 if (isa<GNUNullExpr>(E)) return true; 2854 2855 return false; 2856 } 2857 2858 /// Get the implementation of ObjCInterfaceDecl, or nullptr if none 2859 /// exists. 2860 ObjCImplementationDecl *ASTContext::getObjCImplementation(ObjCInterfaceDecl *D) { 2861 llvm::DenseMap<ObjCContainerDecl*, ObjCImplDecl*>::iterator 2862 I = ObjCImpls.find(D); 2863 if (I != ObjCImpls.end()) 2864 return cast<ObjCImplementationDecl>(I->second); 2865 return nullptr; 2866 } 2867 2868 /// Get the implementation of ObjCCategoryDecl, or nullptr if none 2869 /// exists. 2870 ObjCCategoryImplDecl *ASTContext::getObjCImplementation(ObjCCategoryDecl *D) { 2871 llvm::DenseMap<ObjCContainerDecl*, ObjCImplDecl*>::iterator 2872 I = ObjCImpls.find(D); 2873 if (I != ObjCImpls.end()) 2874 return cast<ObjCCategoryImplDecl>(I->second); 2875 return nullptr; 2876 } 2877 2878 /// Set the implementation of ObjCInterfaceDecl. 2879 void ASTContext::setObjCImplementation(ObjCInterfaceDecl *IFaceD, 2880 ObjCImplementationDecl *ImplD) { 2881 assert(IFaceD && ImplD && "Passed null params"); 2882 ObjCImpls[IFaceD] = ImplD; 2883 } 2884 2885 /// Set the implementation of ObjCCategoryDecl. 2886 void ASTContext::setObjCImplementation(ObjCCategoryDecl *CatD, 2887 ObjCCategoryImplDecl *ImplD) { 2888 assert(CatD && ImplD && "Passed null params"); 2889 ObjCImpls[CatD] = ImplD; 2890 } 2891 2892 const ObjCMethodDecl * 2893 ASTContext::getObjCMethodRedeclaration(const ObjCMethodDecl *MD) const { 2894 return ObjCMethodRedecls.lookup(MD); 2895 } 2896 2897 void ASTContext::setObjCMethodRedeclaration(const ObjCMethodDecl *MD, 2898 const ObjCMethodDecl *Redecl) { 2899 assert(!getObjCMethodRedeclaration(MD) && "MD already has a redeclaration"); 2900 ObjCMethodRedecls[MD] = Redecl; 2901 } 2902 2903 const ObjCInterfaceDecl *ASTContext::getObjContainingInterface( 2904 const NamedDecl *ND) const { 2905 if (const auto *ID = dyn_cast<ObjCInterfaceDecl>(ND->getDeclContext())) 2906 return ID; 2907 if (const auto *CD = dyn_cast<ObjCCategoryDecl>(ND->getDeclContext())) 2908 return CD->getClassInterface(); 2909 if (const auto *IMD = dyn_cast<ObjCImplDecl>(ND->getDeclContext())) 2910 return IMD->getClassInterface(); 2911 2912 return nullptr; 2913 } 2914 2915 /// Get the copy initialization expression of VarDecl, or nullptr if 2916 /// none exists. 2917 BlockVarCopyInit ASTContext::getBlockVarCopyInit(const VarDecl *VD) const { 2918 assert(VD && "Passed null params"); 2919 assert(VD->hasAttr<BlocksAttr>() && 2920 "getBlockVarCopyInits - not __block var"); 2921 auto I = BlockVarCopyInits.find(VD); 2922 if (I != BlockVarCopyInits.end()) 2923 return I->second; 2924 return {nullptr, false}; 2925 } 2926 2927 /// Set the copy initialization expression of a block var decl. 2928 void ASTContext::setBlockVarCopyInit(const VarDecl*VD, Expr *CopyExpr, 2929 bool CanThrow) { 2930 assert(VD && CopyExpr && "Passed null params"); 2931 assert(VD->hasAttr<BlocksAttr>() && 2932 "setBlockVarCopyInits - not __block var"); 2933 BlockVarCopyInits[VD].setExprAndFlag(CopyExpr, CanThrow); 2934 } 2935 2936 TypeSourceInfo *ASTContext::CreateTypeSourceInfo(QualType T, 2937 unsigned DataSize) const { 2938 if (!DataSize) 2939 DataSize = TypeLoc::getFullDataSizeForType(T); 2940 else 2941 assert(DataSize == TypeLoc::getFullDataSizeForType(T) && 2942 "incorrect data size provided to CreateTypeSourceInfo!"); 2943 2944 auto *TInfo = 2945 (TypeSourceInfo*)BumpAlloc.Allocate(sizeof(TypeSourceInfo) + DataSize, 8); 2946 new (TInfo) TypeSourceInfo(T); 2947 return TInfo; 2948 } 2949 2950 TypeSourceInfo *ASTContext::getTrivialTypeSourceInfo(QualType T, 2951 SourceLocation L) const { 2952 TypeSourceInfo *DI = CreateTypeSourceInfo(T); 2953 DI->getTypeLoc().initialize(const_cast<ASTContext &>(*this), L); 2954 return DI; 2955 } 2956 2957 const ASTRecordLayout & 2958 ASTContext::getASTObjCInterfaceLayout(const ObjCInterfaceDecl *D) const { 2959 return getObjCLayout(D, nullptr); 2960 } 2961 2962 const ASTRecordLayout & 2963 ASTContext::getASTObjCImplementationLayout( 2964 const ObjCImplementationDecl *D) const { 2965 return getObjCLayout(D->getClassInterface(), D); 2966 } 2967 2968 //===----------------------------------------------------------------------===// 2969 // Type creation/memoization methods 2970 //===----------------------------------------------------------------------===// 2971 2972 QualType 2973 ASTContext::getExtQualType(const Type *baseType, Qualifiers quals) const { 2974 unsigned fastQuals = quals.getFastQualifiers(); 2975 quals.removeFastQualifiers(); 2976 2977 // Check if we've already instantiated this type. 2978 llvm::FoldingSetNodeID ID; 2979 ExtQuals::Profile(ID, baseType, quals); 2980 void *insertPos = nullptr; 2981 if (ExtQuals *eq = ExtQualNodes.FindNodeOrInsertPos(ID, insertPos)) { 2982 assert(eq->getQualifiers() == quals); 2983 return QualType(eq, fastQuals); 2984 } 2985 2986 // If the base type is not canonical, make the appropriate canonical type. 2987 QualType canon; 2988 if (!baseType->isCanonicalUnqualified()) { 2989 SplitQualType canonSplit = baseType->getCanonicalTypeInternal().split(); 2990 canonSplit.Quals.addConsistentQualifiers(quals); 2991 canon = getExtQualType(canonSplit.Ty, canonSplit.Quals); 2992 2993 // Re-find the insert position. 2994 (void) ExtQualNodes.FindNodeOrInsertPos(ID, insertPos); 2995 } 2996 2997 auto *eq = new (*this, TypeAlignment) ExtQuals(baseType, canon, quals); 2998 ExtQualNodes.InsertNode(eq, insertPos); 2999 return QualType(eq, fastQuals); 3000 } 3001 3002 QualType ASTContext::getAddrSpaceQualType(QualType T, 3003 LangAS AddressSpace) const { 3004 QualType CanT = getCanonicalType(T); 3005 if (CanT.getAddressSpace() == AddressSpace) 3006 return T; 3007 3008 // If we are composing extended qualifiers together, merge together 3009 // into one ExtQuals node. 3010 QualifierCollector Quals; 3011 const Type *TypeNode = Quals.strip(T); 3012 3013 // If this type already has an address space specified, it cannot get 3014 // another one. 3015 assert(!Quals.hasAddressSpace() && 3016 "Type cannot be in multiple addr spaces!"); 3017 Quals.addAddressSpace(AddressSpace); 3018 3019 return getExtQualType(TypeNode, Quals); 3020 } 3021 3022 QualType ASTContext::removeAddrSpaceQualType(QualType T) const { 3023 // If the type is not qualified with an address space, just return it 3024 // immediately. 3025 if (!T.hasAddressSpace()) 3026 return T; 3027 3028 // If we are composing extended qualifiers together, merge together 3029 // into one ExtQuals node. 3030 QualifierCollector Quals; 3031 const Type *TypeNode; 3032 3033 while (T.hasAddressSpace()) { 3034 TypeNode = Quals.strip(T); 3035 3036 // If the type no longer has an address space after stripping qualifiers, 3037 // jump out. 3038 if (!QualType(TypeNode, 0).hasAddressSpace()) 3039 break; 3040 3041 // There might be sugar in the way. Strip it and try again. 3042 T = T.getSingleStepDesugaredType(*this); 3043 } 3044 3045 Quals.removeAddressSpace(); 3046 3047 // Removal of the address space can mean there are no longer any 3048 // non-fast qualifiers, so creating an ExtQualType isn't possible (asserts) 3049 // or required. 3050 if (Quals.hasNonFastQualifiers()) 3051 return getExtQualType(TypeNode, Quals); 3052 else 3053 return QualType(TypeNode, Quals.getFastQualifiers()); 3054 } 3055 3056 QualType ASTContext::getObjCGCQualType(QualType T, 3057 Qualifiers::GC GCAttr) const { 3058 QualType CanT = getCanonicalType(T); 3059 if (CanT.getObjCGCAttr() == GCAttr) 3060 return T; 3061 3062 if (const auto *ptr = T->getAs<PointerType>()) { 3063 QualType Pointee = ptr->getPointeeType(); 3064 if (Pointee->isAnyPointerType()) { 3065 QualType ResultType = getObjCGCQualType(Pointee, GCAttr); 3066 return getPointerType(ResultType); 3067 } 3068 } 3069 3070 // If we are composing extended qualifiers together, merge together 3071 // into one ExtQuals node. 3072 QualifierCollector Quals; 3073 const Type *TypeNode = Quals.strip(T); 3074 3075 // If this type already has an ObjCGC specified, it cannot get 3076 // another one. 3077 assert(!Quals.hasObjCGCAttr() && 3078 "Type cannot have multiple ObjCGCs!"); 3079 Quals.addObjCGCAttr(GCAttr); 3080 3081 return getExtQualType(TypeNode, Quals); 3082 } 3083 3084 QualType ASTContext::removePtrSizeAddrSpace(QualType T) const { 3085 if (const PointerType *Ptr = T->getAs<PointerType>()) { 3086 QualType Pointee = Ptr->getPointeeType(); 3087 if (isPtrSizeAddressSpace(Pointee.getAddressSpace())) { 3088 return getPointerType(removeAddrSpaceQualType(Pointee)); 3089 } 3090 } 3091 return T; 3092 } 3093 3094 const FunctionType *ASTContext::adjustFunctionType(const FunctionType *T, 3095 FunctionType::ExtInfo Info) { 3096 if (T->getExtInfo() == Info) 3097 return T; 3098 3099 QualType Result; 3100 if (const auto *FNPT = dyn_cast<FunctionNoProtoType>(T)) { 3101 Result = getFunctionNoProtoType(FNPT->getReturnType(), Info); 3102 } else { 3103 const auto *FPT = cast<FunctionProtoType>(T); 3104 FunctionProtoType::ExtProtoInfo EPI = FPT->getExtProtoInfo(); 3105 EPI.ExtInfo = Info; 3106 Result = getFunctionType(FPT->getReturnType(), FPT->getParamTypes(), EPI); 3107 } 3108 3109 return cast<FunctionType>(Result.getTypePtr()); 3110 } 3111 3112 void ASTContext::adjustDeducedFunctionResultType(FunctionDecl *FD, 3113 QualType ResultType) { 3114 FD = FD->getMostRecentDecl(); 3115 while (true) { 3116 const auto *FPT = FD->getType()->castAs<FunctionProtoType>(); 3117 FunctionProtoType::ExtProtoInfo EPI = FPT->getExtProtoInfo(); 3118 FD->setType(getFunctionType(ResultType, FPT->getParamTypes(), EPI)); 3119 if (FunctionDecl *Next = FD->getPreviousDecl()) 3120 FD = Next; 3121 else 3122 break; 3123 } 3124 if (ASTMutationListener *L = getASTMutationListener()) 3125 L->DeducedReturnType(FD, ResultType); 3126 } 3127 3128 /// Get a function type and produce the equivalent function type with the 3129 /// specified exception specification. Type sugar that can be present on a 3130 /// declaration of a function with an exception specification is permitted 3131 /// and preserved. Other type sugar (for instance, typedefs) is not. 3132 QualType ASTContext::getFunctionTypeWithExceptionSpec( 3133 QualType Orig, const FunctionProtoType::ExceptionSpecInfo &ESI) { 3134 // Might have some parens. 3135 if (const auto *PT = dyn_cast<ParenType>(Orig)) 3136 return getParenType( 3137 getFunctionTypeWithExceptionSpec(PT->getInnerType(), ESI)); 3138 3139 // Might be wrapped in a macro qualified type. 3140 if (const auto *MQT = dyn_cast<MacroQualifiedType>(Orig)) 3141 return getMacroQualifiedType( 3142 getFunctionTypeWithExceptionSpec(MQT->getUnderlyingType(), ESI), 3143 MQT->getMacroIdentifier()); 3144 3145 // Might have a calling-convention attribute. 3146 if (const auto *AT = dyn_cast<AttributedType>(Orig)) 3147 return getAttributedType( 3148 AT->getAttrKind(), 3149 getFunctionTypeWithExceptionSpec(AT->getModifiedType(), ESI), 3150 getFunctionTypeWithExceptionSpec(AT->getEquivalentType(), ESI)); 3151 3152 // Anything else must be a function type. Rebuild it with the new exception 3153 // specification. 3154 const auto *Proto = Orig->castAs<FunctionProtoType>(); 3155 return getFunctionType( 3156 Proto->getReturnType(), Proto->getParamTypes(), 3157 Proto->getExtProtoInfo().withExceptionSpec(ESI)); 3158 } 3159 3160 bool ASTContext::hasSameFunctionTypeIgnoringExceptionSpec(QualType T, 3161 QualType U) { 3162 return hasSameType(T, U) || 3163 (getLangOpts().CPlusPlus17 && 3164 hasSameType(getFunctionTypeWithExceptionSpec(T, EST_None), 3165 getFunctionTypeWithExceptionSpec(U, EST_None))); 3166 } 3167 3168 QualType ASTContext::getFunctionTypeWithoutPtrSizes(QualType T) { 3169 if (const auto *Proto = T->getAs<FunctionProtoType>()) { 3170 QualType RetTy = removePtrSizeAddrSpace(Proto->getReturnType()); 3171 SmallVector<QualType, 16> Args(Proto->param_types()); 3172 for (unsigned i = 0, n = Args.size(); i != n; ++i) 3173 Args[i] = removePtrSizeAddrSpace(Args[i]); 3174 return getFunctionType(RetTy, Args, Proto->getExtProtoInfo()); 3175 } 3176 3177 if (const FunctionNoProtoType *Proto = T->getAs<FunctionNoProtoType>()) { 3178 QualType RetTy = removePtrSizeAddrSpace(Proto->getReturnType()); 3179 return getFunctionNoProtoType(RetTy, Proto->getExtInfo()); 3180 } 3181 3182 return T; 3183 } 3184 3185 bool ASTContext::hasSameFunctionTypeIgnoringPtrSizes(QualType T, QualType U) { 3186 return hasSameType(T, U) || 3187 hasSameType(getFunctionTypeWithoutPtrSizes(T), 3188 getFunctionTypeWithoutPtrSizes(U)); 3189 } 3190 3191 void ASTContext::adjustExceptionSpec( 3192 FunctionDecl *FD, const FunctionProtoType::ExceptionSpecInfo &ESI, 3193 bool AsWritten) { 3194 // Update the type. 3195 QualType Updated = 3196 getFunctionTypeWithExceptionSpec(FD->getType(), ESI); 3197 FD->setType(Updated); 3198 3199 if (!AsWritten) 3200 return; 3201 3202 // Update the type in the type source information too. 3203 if (TypeSourceInfo *TSInfo = FD->getTypeSourceInfo()) { 3204 // If the type and the type-as-written differ, we may need to update 3205 // the type-as-written too. 3206 if (TSInfo->getType() != FD->getType()) 3207 Updated = getFunctionTypeWithExceptionSpec(TSInfo->getType(), ESI); 3208 3209 // FIXME: When we get proper type location information for exceptions, 3210 // we'll also have to rebuild the TypeSourceInfo. For now, we just patch 3211 // up the TypeSourceInfo; 3212 assert(TypeLoc::getFullDataSizeForType(Updated) == 3213 TypeLoc::getFullDataSizeForType(TSInfo->getType()) && 3214 "TypeLoc size mismatch from updating exception specification"); 3215 TSInfo->overrideType(Updated); 3216 } 3217 } 3218 3219 /// getComplexType - Return the uniqued reference to the type for a complex 3220 /// number with the specified element type. 3221 QualType ASTContext::getComplexType(QualType T) const { 3222 // Unique pointers, to guarantee there is only one pointer of a particular 3223 // structure. 3224 llvm::FoldingSetNodeID ID; 3225 ComplexType::Profile(ID, T); 3226 3227 void *InsertPos = nullptr; 3228 if (ComplexType *CT = ComplexTypes.FindNodeOrInsertPos(ID, InsertPos)) 3229 return QualType(CT, 0); 3230 3231 // If the pointee type isn't canonical, this won't be a canonical type either, 3232 // so fill in the canonical type field. 3233 QualType Canonical; 3234 if (!T.isCanonical()) { 3235 Canonical = getComplexType(getCanonicalType(T)); 3236 3237 // Get the new insert position for the node we care about. 3238 ComplexType *NewIP = ComplexTypes.FindNodeOrInsertPos(ID, InsertPos); 3239 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3240 } 3241 auto *New = new (*this, TypeAlignment) ComplexType(T, Canonical); 3242 Types.push_back(New); 3243 ComplexTypes.InsertNode(New, InsertPos); 3244 return QualType(New, 0); 3245 } 3246 3247 /// getPointerType - Return the uniqued reference to the type for a pointer to 3248 /// the specified type. 3249 QualType ASTContext::getPointerType(QualType T) const { 3250 // Unique pointers, to guarantee there is only one pointer of a particular 3251 // structure. 3252 llvm::FoldingSetNodeID ID; 3253 PointerType::Profile(ID, T); 3254 3255 void *InsertPos = nullptr; 3256 if (PointerType *PT = PointerTypes.FindNodeOrInsertPos(ID, InsertPos)) 3257 return QualType(PT, 0); 3258 3259 // If the pointee type isn't canonical, this won't be a canonical type either, 3260 // so fill in the canonical type field. 3261 QualType Canonical; 3262 if (!T.isCanonical()) { 3263 Canonical = getPointerType(getCanonicalType(T)); 3264 3265 // Get the new insert position for the node we care about. 3266 PointerType *NewIP = PointerTypes.FindNodeOrInsertPos(ID, InsertPos); 3267 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3268 } 3269 auto *New = new (*this, TypeAlignment) PointerType(T, Canonical); 3270 Types.push_back(New); 3271 PointerTypes.InsertNode(New, InsertPos); 3272 return QualType(New, 0); 3273 } 3274 3275 QualType ASTContext::getAdjustedType(QualType Orig, QualType New) const { 3276 llvm::FoldingSetNodeID ID; 3277 AdjustedType::Profile(ID, Orig, New); 3278 void *InsertPos = nullptr; 3279 AdjustedType *AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos); 3280 if (AT) 3281 return QualType(AT, 0); 3282 3283 QualType Canonical = getCanonicalType(New); 3284 3285 // Get the new insert position for the node we care about. 3286 AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos); 3287 assert(!AT && "Shouldn't be in the map!"); 3288 3289 AT = new (*this, TypeAlignment) 3290 AdjustedType(Type::Adjusted, Orig, New, Canonical); 3291 Types.push_back(AT); 3292 AdjustedTypes.InsertNode(AT, InsertPos); 3293 return QualType(AT, 0); 3294 } 3295 3296 QualType ASTContext::getDecayedType(QualType T) const { 3297 assert((T->isArrayType() || T->isFunctionType()) && "T does not decay"); 3298 3299 QualType Decayed; 3300 3301 // C99 6.7.5.3p7: 3302 // A declaration of a parameter as "array of type" shall be 3303 // adjusted to "qualified pointer to type", where the type 3304 // qualifiers (if any) are those specified within the [ and ] of 3305 // the array type derivation. 3306 if (T->isArrayType()) 3307 Decayed = getArrayDecayedType(T); 3308 3309 // C99 6.7.5.3p8: 3310 // A declaration of a parameter as "function returning type" 3311 // shall be adjusted to "pointer to function returning type", as 3312 // in 6.3.2.1. 3313 if (T->isFunctionType()) 3314 Decayed = getPointerType(T); 3315 3316 llvm::FoldingSetNodeID ID; 3317 AdjustedType::Profile(ID, T, Decayed); 3318 void *InsertPos = nullptr; 3319 AdjustedType *AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos); 3320 if (AT) 3321 return QualType(AT, 0); 3322 3323 QualType Canonical = getCanonicalType(Decayed); 3324 3325 // Get the new insert position for the node we care about. 3326 AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos); 3327 assert(!AT && "Shouldn't be in the map!"); 3328 3329 AT = new (*this, TypeAlignment) DecayedType(T, Decayed, Canonical); 3330 Types.push_back(AT); 3331 AdjustedTypes.InsertNode(AT, InsertPos); 3332 return QualType(AT, 0); 3333 } 3334 3335 /// getBlockPointerType - Return the uniqued reference to the type for 3336 /// a pointer to the specified block. 3337 QualType ASTContext::getBlockPointerType(QualType T) const { 3338 assert(T->isFunctionType() && "block of function types only"); 3339 // Unique pointers, to guarantee there is only one block of a particular 3340 // structure. 3341 llvm::FoldingSetNodeID ID; 3342 BlockPointerType::Profile(ID, T); 3343 3344 void *InsertPos = nullptr; 3345 if (BlockPointerType *PT = 3346 BlockPointerTypes.FindNodeOrInsertPos(ID, InsertPos)) 3347 return QualType(PT, 0); 3348 3349 // If the block pointee type isn't canonical, this won't be a canonical 3350 // type either so fill in the canonical type field. 3351 QualType Canonical; 3352 if (!T.isCanonical()) { 3353 Canonical = getBlockPointerType(getCanonicalType(T)); 3354 3355 // Get the new insert position for the node we care about. 3356 BlockPointerType *NewIP = 3357 BlockPointerTypes.FindNodeOrInsertPos(ID, InsertPos); 3358 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3359 } 3360 auto *New = new (*this, TypeAlignment) BlockPointerType(T, Canonical); 3361 Types.push_back(New); 3362 BlockPointerTypes.InsertNode(New, InsertPos); 3363 return QualType(New, 0); 3364 } 3365 3366 /// getLValueReferenceType - Return the uniqued reference to the type for an 3367 /// lvalue reference to the specified type. 3368 QualType 3369 ASTContext::getLValueReferenceType(QualType T, bool SpelledAsLValue) const { 3370 assert(getCanonicalType(T) != OverloadTy && 3371 "Unresolved overloaded function type"); 3372 3373 // Unique pointers, to guarantee there is only one pointer of a particular 3374 // structure. 3375 llvm::FoldingSetNodeID ID; 3376 ReferenceType::Profile(ID, T, SpelledAsLValue); 3377 3378 void *InsertPos = nullptr; 3379 if (LValueReferenceType *RT = 3380 LValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos)) 3381 return QualType(RT, 0); 3382 3383 const auto *InnerRef = T->getAs<ReferenceType>(); 3384 3385 // If the referencee type isn't canonical, this won't be a canonical type 3386 // either, so fill in the canonical type field. 3387 QualType Canonical; 3388 if (!SpelledAsLValue || InnerRef || !T.isCanonical()) { 3389 QualType PointeeType = (InnerRef ? InnerRef->getPointeeType() : T); 3390 Canonical = getLValueReferenceType(getCanonicalType(PointeeType)); 3391 3392 // Get the new insert position for the node we care about. 3393 LValueReferenceType *NewIP = 3394 LValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos); 3395 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3396 } 3397 3398 auto *New = new (*this, TypeAlignment) LValueReferenceType(T, Canonical, 3399 SpelledAsLValue); 3400 Types.push_back(New); 3401 LValueReferenceTypes.InsertNode(New, InsertPos); 3402 3403 return QualType(New, 0); 3404 } 3405 3406 /// getRValueReferenceType - Return the uniqued reference to the type for an 3407 /// rvalue reference to the specified type. 3408 QualType ASTContext::getRValueReferenceType(QualType T) const { 3409 // Unique pointers, to guarantee there is only one pointer of a particular 3410 // structure. 3411 llvm::FoldingSetNodeID ID; 3412 ReferenceType::Profile(ID, T, false); 3413 3414 void *InsertPos = nullptr; 3415 if (RValueReferenceType *RT = 3416 RValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos)) 3417 return QualType(RT, 0); 3418 3419 const auto *InnerRef = T->getAs<ReferenceType>(); 3420 3421 // If the referencee type isn't canonical, this won't be a canonical type 3422 // either, so fill in the canonical type field. 3423 QualType Canonical; 3424 if (InnerRef || !T.isCanonical()) { 3425 QualType PointeeType = (InnerRef ? InnerRef->getPointeeType() : T); 3426 Canonical = getRValueReferenceType(getCanonicalType(PointeeType)); 3427 3428 // Get the new insert position for the node we care about. 3429 RValueReferenceType *NewIP = 3430 RValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos); 3431 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3432 } 3433 3434 auto *New = new (*this, TypeAlignment) RValueReferenceType(T, Canonical); 3435 Types.push_back(New); 3436 RValueReferenceTypes.InsertNode(New, InsertPos); 3437 return QualType(New, 0); 3438 } 3439 3440 /// getMemberPointerType - Return the uniqued reference to the type for a 3441 /// member pointer to the specified type, in the specified class. 3442 QualType ASTContext::getMemberPointerType(QualType T, const Type *Cls) const { 3443 // Unique pointers, to guarantee there is only one pointer of a particular 3444 // structure. 3445 llvm::FoldingSetNodeID ID; 3446 MemberPointerType::Profile(ID, T, Cls); 3447 3448 void *InsertPos = nullptr; 3449 if (MemberPointerType *PT = 3450 MemberPointerTypes.FindNodeOrInsertPos(ID, InsertPos)) 3451 return QualType(PT, 0); 3452 3453 // If the pointee or class type isn't canonical, this won't be a canonical 3454 // type either, so fill in the canonical type field. 3455 QualType Canonical; 3456 if (!T.isCanonical() || !Cls->isCanonicalUnqualified()) { 3457 Canonical = getMemberPointerType(getCanonicalType(T),getCanonicalType(Cls)); 3458 3459 // Get the new insert position for the node we care about. 3460 MemberPointerType *NewIP = 3461 MemberPointerTypes.FindNodeOrInsertPos(ID, InsertPos); 3462 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3463 } 3464 auto *New = new (*this, TypeAlignment) MemberPointerType(T, Cls, Canonical); 3465 Types.push_back(New); 3466 MemberPointerTypes.InsertNode(New, InsertPos); 3467 return QualType(New, 0); 3468 } 3469 3470 /// getConstantArrayType - Return the unique reference to the type for an 3471 /// array of the specified element type. 3472 QualType ASTContext::getConstantArrayType(QualType EltTy, 3473 const llvm::APInt &ArySizeIn, 3474 const Expr *SizeExpr, 3475 ArrayType::ArraySizeModifier ASM, 3476 unsigned IndexTypeQuals) const { 3477 assert((EltTy->isDependentType() || 3478 EltTy->isIncompleteType() || EltTy->isConstantSizeType()) && 3479 "Constant array of VLAs is illegal!"); 3480 3481 // We only need the size as part of the type if it's instantiation-dependent. 3482 if (SizeExpr && !SizeExpr->isInstantiationDependent()) 3483 SizeExpr = nullptr; 3484 3485 // Convert the array size into a canonical width matching the pointer size for 3486 // the target. 3487 llvm::APInt ArySize(ArySizeIn); 3488 ArySize = ArySize.zextOrTrunc(Target->getMaxPointerWidth()); 3489 3490 llvm::FoldingSetNodeID ID; 3491 ConstantArrayType::Profile(ID, *this, EltTy, ArySize, SizeExpr, ASM, 3492 IndexTypeQuals); 3493 3494 void *InsertPos = nullptr; 3495 if (ConstantArrayType *ATP = 3496 ConstantArrayTypes.FindNodeOrInsertPos(ID, InsertPos)) 3497 return QualType(ATP, 0); 3498 3499 // If the element type isn't canonical or has qualifiers, or the array bound 3500 // is instantiation-dependent, this won't be a canonical type either, so fill 3501 // in the canonical type field. 3502 QualType Canon; 3503 if (!EltTy.isCanonical() || EltTy.hasLocalQualifiers() || SizeExpr) { 3504 SplitQualType canonSplit = getCanonicalType(EltTy).split(); 3505 Canon = getConstantArrayType(QualType(canonSplit.Ty, 0), ArySize, nullptr, 3506 ASM, IndexTypeQuals); 3507 Canon = getQualifiedType(Canon, canonSplit.Quals); 3508 3509 // Get the new insert position for the node we care about. 3510 ConstantArrayType *NewIP = 3511 ConstantArrayTypes.FindNodeOrInsertPos(ID, InsertPos); 3512 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3513 } 3514 3515 void *Mem = Allocate( 3516 ConstantArrayType::totalSizeToAlloc<const Expr *>(SizeExpr ? 1 : 0), 3517 TypeAlignment); 3518 auto *New = new (Mem) 3519 ConstantArrayType(EltTy, Canon, ArySize, SizeExpr, ASM, IndexTypeQuals); 3520 ConstantArrayTypes.InsertNode(New, InsertPos); 3521 Types.push_back(New); 3522 return QualType(New, 0); 3523 } 3524 3525 /// getVariableArrayDecayedType - Turns the given type, which may be 3526 /// variably-modified, into the corresponding type with all the known 3527 /// sizes replaced with [*]. 3528 QualType ASTContext::getVariableArrayDecayedType(QualType type) const { 3529 // Vastly most common case. 3530 if (!type->isVariablyModifiedType()) return type; 3531 3532 QualType result; 3533 3534 SplitQualType split = type.getSplitDesugaredType(); 3535 const Type *ty = split.Ty; 3536 switch (ty->getTypeClass()) { 3537 #define TYPE(Class, Base) 3538 #define ABSTRACT_TYPE(Class, Base) 3539 #define NON_CANONICAL_TYPE(Class, Base) case Type::Class: 3540 #include "clang/AST/TypeNodes.inc" 3541 llvm_unreachable("didn't desugar past all non-canonical types?"); 3542 3543 // These types should never be variably-modified. 3544 case Type::Builtin: 3545 case Type::Complex: 3546 case Type::Vector: 3547 case Type::DependentVector: 3548 case Type::ExtVector: 3549 case Type::DependentSizedExtVector: 3550 case Type::ConstantMatrix: 3551 case Type::DependentSizedMatrix: 3552 case Type::DependentAddressSpace: 3553 case Type::ObjCObject: 3554 case Type::ObjCInterface: 3555 case Type::ObjCObjectPointer: 3556 case Type::Record: 3557 case Type::Enum: 3558 case Type::UnresolvedUsing: 3559 case Type::TypeOfExpr: 3560 case Type::TypeOf: 3561 case Type::Decltype: 3562 case Type::UnaryTransform: 3563 case Type::DependentName: 3564 case Type::InjectedClassName: 3565 case Type::TemplateSpecialization: 3566 case Type::DependentTemplateSpecialization: 3567 case Type::TemplateTypeParm: 3568 case Type::SubstTemplateTypeParmPack: 3569 case Type::Auto: 3570 case Type::DeducedTemplateSpecialization: 3571 case Type::PackExpansion: 3572 case Type::ExtInt: 3573 case Type::DependentExtInt: 3574 llvm_unreachable("type should never be variably-modified"); 3575 3576 // These types can be variably-modified but should never need to 3577 // further decay. 3578 case Type::FunctionNoProto: 3579 case Type::FunctionProto: 3580 case Type::BlockPointer: 3581 case Type::MemberPointer: 3582 case Type::Pipe: 3583 return type; 3584 3585 // These types can be variably-modified. All these modifications 3586 // preserve structure except as noted by comments. 3587 // TODO: if we ever care about optimizing VLAs, there are no-op 3588 // optimizations available here. 3589 case Type::Pointer: 3590 result = getPointerType(getVariableArrayDecayedType( 3591 cast<PointerType>(ty)->getPointeeType())); 3592 break; 3593 3594 case Type::LValueReference: { 3595 const auto *lv = cast<LValueReferenceType>(ty); 3596 result = getLValueReferenceType( 3597 getVariableArrayDecayedType(lv->getPointeeType()), 3598 lv->isSpelledAsLValue()); 3599 break; 3600 } 3601 3602 case Type::RValueReference: { 3603 const auto *lv = cast<RValueReferenceType>(ty); 3604 result = getRValueReferenceType( 3605 getVariableArrayDecayedType(lv->getPointeeType())); 3606 break; 3607 } 3608 3609 case Type::Atomic: { 3610 const auto *at = cast<AtomicType>(ty); 3611 result = getAtomicType(getVariableArrayDecayedType(at->getValueType())); 3612 break; 3613 } 3614 3615 case Type::ConstantArray: { 3616 const auto *cat = cast<ConstantArrayType>(ty); 3617 result = getConstantArrayType( 3618 getVariableArrayDecayedType(cat->getElementType()), 3619 cat->getSize(), 3620 cat->getSizeExpr(), 3621 cat->getSizeModifier(), 3622 cat->getIndexTypeCVRQualifiers()); 3623 break; 3624 } 3625 3626 case Type::DependentSizedArray: { 3627 const auto *dat = cast<DependentSizedArrayType>(ty); 3628 result = getDependentSizedArrayType( 3629 getVariableArrayDecayedType(dat->getElementType()), 3630 dat->getSizeExpr(), 3631 dat->getSizeModifier(), 3632 dat->getIndexTypeCVRQualifiers(), 3633 dat->getBracketsRange()); 3634 break; 3635 } 3636 3637 // Turn incomplete types into [*] types. 3638 case Type::IncompleteArray: { 3639 const auto *iat = cast<IncompleteArrayType>(ty); 3640 result = getVariableArrayType( 3641 getVariableArrayDecayedType(iat->getElementType()), 3642 /*size*/ nullptr, 3643 ArrayType::Normal, 3644 iat->getIndexTypeCVRQualifiers(), 3645 SourceRange()); 3646 break; 3647 } 3648 3649 // Turn VLA types into [*] types. 3650 case Type::VariableArray: { 3651 const auto *vat = cast<VariableArrayType>(ty); 3652 result = getVariableArrayType( 3653 getVariableArrayDecayedType(vat->getElementType()), 3654 /*size*/ nullptr, 3655 ArrayType::Star, 3656 vat->getIndexTypeCVRQualifiers(), 3657 vat->getBracketsRange()); 3658 break; 3659 } 3660 } 3661 3662 // Apply the top-level qualifiers from the original. 3663 return getQualifiedType(result, split.Quals); 3664 } 3665 3666 /// getVariableArrayType - Returns a non-unique reference to the type for a 3667 /// variable array of the specified element type. 3668 QualType ASTContext::getVariableArrayType(QualType EltTy, 3669 Expr *NumElts, 3670 ArrayType::ArraySizeModifier ASM, 3671 unsigned IndexTypeQuals, 3672 SourceRange Brackets) const { 3673 // Since we don't unique expressions, it isn't possible to unique VLA's 3674 // that have an expression provided for their size. 3675 QualType Canon; 3676 3677 // Be sure to pull qualifiers off the element type. 3678 if (!EltTy.isCanonical() || EltTy.hasLocalQualifiers()) { 3679 SplitQualType canonSplit = getCanonicalType(EltTy).split(); 3680 Canon = getVariableArrayType(QualType(canonSplit.Ty, 0), NumElts, ASM, 3681 IndexTypeQuals, Brackets); 3682 Canon = getQualifiedType(Canon, canonSplit.Quals); 3683 } 3684 3685 auto *New = new (*this, TypeAlignment) 3686 VariableArrayType(EltTy, Canon, NumElts, ASM, IndexTypeQuals, Brackets); 3687 3688 VariableArrayTypes.push_back(New); 3689 Types.push_back(New); 3690 return QualType(New, 0); 3691 } 3692 3693 /// getDependentSizedArrayType - Returns a non-unique reference to 3694 /// the type for a dependently-sized array of the specified element 3695 /// type. 3696 QualType ASTContext::getDependentSizedArrayType(QualType elementType, 3697 Expr *numElements, 3698 ArrayType::ArraySizeModifier ASM, 3699 unsigned elementTypeQuals, 3700 SourceRange brackets) const { 3701 assert((!numElements || numElements->isTypeDependent() || 3702 numElements->isValueDependent()) && 3703 "Size must be type- or value-dependent!"); 3704 3705 // Dependently-sized array types that do not have a specified number 3706 // of elements will have their sizes deduced from a dependent 3707 // initializer. We do no canonicalization here at all, which is okay 3708 // because they can't be used in most locations. 3709 if (!numElements) { 3710 auto *newType 3711 = new (*this, TypeAlignment) 3712 DependentSizedArrayType(*this, elementType, QualType(), 3713 numElements, ASM, elementTypeQuals, 3714 brackets); 3715 Types.push_back(newType); 3716 return QualType(newType, 0); 3717 } 3718 3719 // Otherwise, we actually build a new type every time, but we 3720 // also build a canonical type. 3721 3722 SplitQualType canonElementType = getCanonicalType(elementType).split(); 3723 3724 void *insertPos = nullptr; 3725 llvm::FoldingSetNodeID ID; 3726 DependentSizedArrayType::Profile(ID, *this, 3727 QualType(canonElementType.Ty, 0), 3728 ASM, elementTypeQuals, numElements); 3729 3730 // Look for an existing type with these properties. 3731 DependentSizedArrayType *canonTy = 3732 DependentSizedArrayTypes.FindNodeOrInsertPos(ID, insertPos); 3733 3734 // If we don't have one, build one. 3735 if (!canonTy) { 3736 canonTy = new (*this, TypeAlignment) 3737 DependentSizedArrayType(*this, QualType(canonElementType.Ty, 0), 3738 QualType(), numElements, ASM, elementTypeQuals, 3739 brackets); 3740 DependentSizedArrayTypes.InsertNode(canonTy, insertPos); 3741 Types.push_back(canonTy); 3742 } 3743 3744 // Apply qualifiers from the element type to the array. 3745 QualType canon = getQualifiedType(QualType(canonTy,0), 3746 canonElementType.Quals); 3747 3748 // If we didn't need extra canonicalization for the element type or the size 3749 // expression, then just use that as our result. 3750 if (QualType(canonElementType.Ty, 0) == elementType && 3751 canonTy->getSizeExpr() == numElements) 3752 return canon; 3753 3754 // Otherwise, we need to build a type which follows the spelling 3755 // of the element type. 3756 auto *sugaredType 3757 = new (*this, TypeAlignment) 3758 DependentSizedArrayType(*this, elementType, canon, numElements, 3759 ASM, elementTypeQuals, brackets); 3760 Types.push_back(sugaredType); 3761 return QualType(sugaredType, 0); 3762 } 3763 3764 QualType ASTContext::getIncompleteArrayType(QualType elementType, 3765 ArrayType::ArraySizeModifier ASM, 3766 unsigned elementTypeQuals) const { 3767 llvm::FoldingSetNodeID ID; 3768 IncompleteArrayType::Profile(ID, elementType, ASM, elementTypeQuals); 3769 3770 void *insertPos = nullptr; 3771 if (IncompleteArrayType *iat = 3772 IncompleteArrayTypes.FindNodeOrInsertPos(ID, insertPos)) 3773 return QualType(iat, 0); 3774 3775 // If the element type isn't canonical, this won't be a canonical type 3776 // either, so fill in the canonical type field. We also have to pull 3777 // qualifiers off the element type. 3778 QualType canon; 3779 3780 if (!elementType.isCanonical() || elementType.hasLocalQualifiers()) { 3781 SplitQualType canonSplit = getCanonicalType(elementType).split(); 3782 canon = getIncompleteArrayType(QualType(canonSplit.Ty, 0), 3783 ASM, elementTypeQuals); 3784 canon = getQualifiedType(canon, canonSplit.Quals); 3785 3786 // Get the new insert position for the node we care about. 3787 IncompleteArrayType *existing = 3788 IncompleteArrayTypes.FindNodeOrInsertPos(ID, insertPos); 3789 assert(!existing && "Shouldn't be in the map!"); (void) existing; 3790 } 3791 3792 auto *newType = new (*this, TypeAlignment) 3793 IncompleteArrayType(elementType, canon, ASM, elementTypeQuals); 3794 3795 IncompleteArrayTypes.InsertNode(newType, insertPos); 3796 Types.push_back(newType); 3797 return QualType(newType, 0); 3798 } 3799 3800 ASTContext::BuiltinVectorTypeInfo 3801 ASTContext::getBuiltinVectorTypeInfo(const BuiltinType *Ty) const { 3802 #define SVE_INT_ELTTY(BITS, ELTS, SIGNED, NUMVECTORS) \ 3803 {getIntTypeForBitwidth(BITS, SIGNED), llvm::ElementCount::getScalable(ELTS), \ 3804 NUMVECTORS}; 3805 3806 #define SVE_ELTTY(ELTTY, ELTS, NUMVECTORS) \ 3807 {ELTTY, llvm::ElementCount::getScalable(ELTS), NUMVECTORS}; 3808 3809 switch (Ty->getKind()) { 3810 default: 3811 llvm_unreachable("Unsupported builtin vector type"); 3812 case BuiltinType::SveInt8: 3813 return SVE_INT_ELTTY(8, 16, true, 1); 3814 case BuiltinType::SveUint8: 3815 return SVE_INT_ELTTY(8, 16, false, 1); 3816 case BuiltinType::SveInt8x2: 3817 return SVE_INT_ELTTY(8, 16, true, 2); 3818 case BuiltinType::SveUint8x2: 3819 return SVE_INT_ELTTY(8, 16, false, 2); 3820 case BuiltinType::SveInt8x3: 3821 return SVE_INT_ELTTY(8, 16, true, 3); 3822 case BuiltinType::SveUint8x3: 3823 return SVE_INT_ELTTY(8, 16, false, 3); 3824 case BuiltinType::SveInt8x4: 3825 return SVE_INT_ELTTY(8, 16, true, 4); 3826 case BuiltinType::SveUint8x4: 3827 return SVE_INT_ELTTY(8, 16, false, 4); 3828 case BuiltinType::SveInt16: 3829 return SVE_INT_ELTTY(16, 8, true, 1); 3830 case BuiltinType::SveUint16: 3831 return SVE_INT_ELTTY(16, 8, false, 1); 3832 case BuiltinType::SveInt16x2: 3833 return SVE_INT_ELTTY(16, 8, true, 2); 3834 case BuiltinType::SveUint16x2: 3835 return SVE_INT_ELTTY(16, 8, false, 2); 3836 case BuiltinType::SveInt16x3: 3837 return SVE_INT_ELTTY(16, 8, true, 3); 3838 case BuiltinType::SveUint16x3: 3839 return SVE_INT_ELTTY(16, 8, false, 3); 3840 case BuiltinType::SveInt16x4: 3841 return SVE_INT_ELTTY(16, 8, true, 4); 3842 case BuiltinType::SveUint16x4: 3843 return SVE_INT_ELTTY(16, 8, false, 4); 3844 case BuiltinType::SveInt32: 3845 return SVE_INT_ELTTY(32, 4, true, 1); 3846 case BuiltinType::SveUint32: 3847 return SVE_INT_ELTTY(32, 4, false, 1); 3848 case BuiltinType::SveInt32x2: 3849 return SVE_INT_ELTTY(32, 4, true, 2); 3850 case BuiltinType::SveUint32x2: 3851 return SVE_INT_ELTTY(32, 4, false, 2); 3852 case BuiltinType::SveInt32x3: 3853 return SVE_INT_ELTTY(32, 4, true, 3); 3854 case BuiltinType::SveUint32x3: 3855 return SVE_INT_ELTTY(32, 4, false, 3); 3856 case BuiltinType::SveInt32x4: 3857 return SVE_INT_ELTTY(32, 4, true, 4); 3858 case BuiltinType::SveUint32x4: 3859 return SVE_INT_ELTTY(32, 4, false, 4); 3860 case BuiltinType::SveInt64: 3861 return SVE_INT_ELTTY(64, 2, true, 1); 3862 case BuiltinType::SveUint64: 3863 return SVE_INT_ELTTY(64, 2, false, 1); 3864 case BuiltinType::SveInt64x2: 3865 return SVE_INT_ELTTY(64, 2, true, 2); 3866 case BuiltinType::SveUint64x2: 3867 return SVE_INT_ELTTY(64, 2, false, 2); 3868 case BuiltinType::SveInt64x3: 3869 return SVE_INT_ELTTY(64, 2, true, 3); 3870 case BuiltinType::SveUint64x3: 3871 return SVE_INT_ELTTY(64, 2, false, 3); 3872 case BuiltinType::SveInt64x4: 3873 return SVE_INT_ELTTY(64, 2, true, 4); 3874 case BuiltinType::SveUint64x4: 3875 return SVE_INT_ELTTY(64, 2, false, 4); 3876 case BuiltinType::SveBool: 3877 return SVE_ELTTY(BoolTy, 16, 1); 3878 case BuiltinType::SveFloat16: 3879 return SVE_ELTTY(HalfTy, 8, 1); 3880 case BuiltinType::SveFloat16x2: 3881 return SVE_ELTTY(HalfTy, 8, 2); 3882 case BuiltinType::SveFloat16x3: 3883 return SVE_ELTTY(HalfTy, 8, 3); 3884 case BuiltinType::SveFloat16x4: 3885 return SVE_ELTTY(HalfTy, 8, 4); 3886 case BuiltinType::SveFloat32: 3887 return SVE_ELTTY(FloatTy, 4, 1); 3888 case BuiltinType::SveFloat32x2: 3889 return SVE_ELTTY(FloatTy, 4, 2); 3890 case BuiltinType::SveFloat32x3: 3891 return SVE_ELTTY(FloatTy, 4, 3); 3892 case BuiltinType::SveFloat32x4: 3893 return SVE_ELTTY(FloatTy, 4, 4); 3894 case BuiltinType::SveFloat64: 3895 return SVE_ELTTY(DoubleTy, 2, 1); 3896 case BuiltinType::SveFloat64x2: 3897 return SVE_ELTTY(DoubleTy, 2, 2); 3898 case BuiltinType::SveFloat64x3: 3899 return SVE_ELTTY(DoubleTy, 2, 3); 3900 case BuiltinType::SveFloat64x4: 3901 return SVE_ELTTY(DoubleTy, 2, 4); 3902 case BuiltinType::SveBFloat16: 3903 return SVE_ELTTY(BFloat16Ty, 8, 1); 3904 case BuiltinType::SveBFloat16x2: 3905 return SVE_ELTTY(BFloat16Ty, 8, 2); 3906 case BuiltinType::SveBFloat16x3: 3907 return SVE_ELTTY(BFloat16Ty, 8, 3); 3908 case BuiltinType::SveBFloat16x4: 3909 return SVE_ELTTY(BFloat16Ty, 8, 4); 3910 #define RVV_VECTOR_TYPE_INT(Name, Id, SingletonId, NumEls, ElBits, NF, \ 3911 IsSigned) \ 3912 case BuiltinType::Id: \ 3913 return {getIntTypeForBitwidth(ElBits, IsSigned), \ 3914 llvm::ElementCount::getScalable(NumEls), NF}; 3915 #define RVV_VECTOR_TYPE_FLOAT(Name, Id, SingletonId, NumEls, ElBits, NF) \ 3916 case BuiltinType::Id: \ 3917 return {ElBits == 16 ? Float16Ty : (ElBits == 32 ? FloatTy : DoubleTy), \ 3918 llvm::ElementCount::getScalable(NumEls), NF}; 3919 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, NumEls) \ 3920 case BuiltinType::Id: \ 3921 return {BoolTy, llvm::ElementCount::getScalable(NumEls), 1}; 3922 #include "clang/Basic/RISCVVTypes.def" 3923 } 3924 } 3925 3926 /// getScalableVectorType - Return the unique reference to a scalable vector 3927 /// type of the specified element type and size. VectorType must be a built-in 3928 /// type. 3929 QualType ASTContext::getScalableVectorType(QualType EltTy, 3930 unsigned NumElts) const { 3931 if (Target->hasAArch64SVETypes()) { 3932 uint64_t EltTySize = getTypeSize(EltTy); 3933 #define SVE_VECTOR_TYPE(Name, MangledName, Id, SingletonId, NumEls, ElBits, \ 3934 IsSigned, IsFP, IsBF) \ 3935 if (!EltTy->isBooleanType() && \ 3936 ((EltTy->hasIntegerRepresentation() && \ 3937 EltTy->hasSignedIntegerRepresentation() == IsSigned) || \ 3938 (EltTy->hasFloatingRepresentation() && !EltTy->isBFloat16Type() && \ 3939 IsFP && !IsBF) || \ 3940 (EltTy->hasFloatingRepresentation() && EltTy->isBFloat16Type() && \ 3941 IsBF && !IsFP)) && \ 3942 EltTySize == ElBits && NumElts == NumEls) { \ 3943 return SingletonId; \ 3944 } 3945 #define SVE_PREDICATE_TYPE(Name, MangledName, Id, SingletonId, NumEls) \ 3946 if (EltTy->isBooleanType() && NumElts == NumEls) \ 3947 return SingletonId; 3948 #include "clang/Basic/AArch64SVEACLETypes.def" 3949 } else if (Target->hasRISCVVTypes()) { 3950 uint64_t EltTySize = getTypeSize(EltTy); 3951 #define RVV_VECTOR_TYPE(Name, Id, SingletonId, NumEls, ElBits, NF, IsSigned, \ 3952 IsFP) \ 3953 if (!EltTy->isBooleanType() && \ 3954 ((EltTy->hasIntegerRepresentation() && \ 3955 EltTy->hasSignedIntegerRepresentation() == IsSigned) || \ 3956 (EltTy->hasFloatingRepresentation() && IsFP)) && \ 3957 EltTySize == ElBits && NumElts == NumEls) \ 3958 return SingletonId; 3959 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, NumEls) \ 3960 if (EltTy->isBooleanType() && NumElts == NumEls) \ 3961 return SingletonId; 3962 #include "clang/Basic/RISCVVTypes.def" 3963 } 3964 return QualType(); 3965 } 3966 3967 /// getVectorType - Return the unique reference to a vector type of 3968 /// the specified element type and size. VectorType must be a built-in type. 3969 QualType ASTContext::getVectorType(QualType vecType, unsigned NumElts, 3970 VectorType::VectorKind VecKind) const { 3971 assert(vecType->isBuiltinType()); 3972 3973 // Check if we've already instantiated a vector of this type. 3974 llvm::FoldingSetNodeID ID; 3975 VectorType::Profile(ID, vecType, NumElts, Type::Vector, VecKind); 3976 3977 void *InsertPos = nullptr; 3978 if (VectorType *VTP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos)) 3979 return QualType(VTP, 0); 3980 3981 // If the element type isn't canonical, this won't be a canonical type either, 3982 // so fill in the canonical type field. 3983 QualType Canonical; 3984 if (!vecType.isCanonical()) { 3985 Canonical = getVectorType(getCanonicalType(vecType), NumElts, VecKind); 3986 3987 // Get the new insert position for the node we care about. 3988 VectorType *NewIP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos); 3989 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 3990 } 3991 auto *New = new (*this, TypeAlignment) 3992 VectorType(vecType, NumElts, Canonical, VecKind); 3993 VectorTypes.InsertNode(New, InsertPos); 3994 Types.push_back(New); 3995 return QualType(New, 0); 3996 } 3997 3998 QualType 3999 ASTContext::getDependentVectorType(QualType VecType, Expr *SizeExpr, 4000 SourceLocation AttrLoc, 4001 VectorType::VectorKind VecKind) const { 4002 llvm::FoldingSetNodeID ID; 4003 DependentVectorType::Profile(ID, *this, getCanonicalType(VecType), SizeExpr, 4004 VecKind); 4005 void *InsertPos = nullptr; 4006 DependentVectorType *Canon = 4007 DependentVectorTypes.FindNodeOrInsertPos(ID, InsertPos); 4008 DependentVectorType *New; 4009 4010 if (Canon) { 4011 New = new (*this, TypeAlignment) DependentVectorType( 4012 *this, VecType, QualType(Canon, 0), SizeExpr, AttrLoc, VecKind); 4013 } else { 4014 QualType CanonVecTy = getCanonicalType(VecType); 4015 if (CanonVecTy == VecType) { 4016 New = new (*this, TypeAlignment) DependentVectorType( 4017 *this, VecType, QualType(), SizeExpr, AttrLoc, VecKind); 4018 4019 DependentVectorType *CanonCheck = 4020 DependentVectorTypes.FindNodeOrInsertPos(ID, InsertPos); 4021 assert(!CanonCheck && 4022 "Dependent-sized vector_size canonical type broken"); 4023 (void)CanonCheck; 4024 DependentVectorTypes.InsertNode(New, InsertPos); 4025 } else { 4026 QualType CanonTy = getDependentVectorType(CanonVecTy, SizeExpr, 4027 SourceLocation(), VecKind); 4028 New = new (*this, TypeAlignment) DependentVectorType( 4029 *this, VecType, CanonTy, SizeExpr, AttrLoc, VecKind); 4030 } 4031 } 4032 4033 Types.push_back(New); 4034 return QualType(New, 0); 4035 } 4036 4037 /// getExtVectorType - Return the unique reference to an extended vector type of 4038 /// the specified element type and size. VectorType must be a built-in type. 4039 QualType 4040 ASTContext::getExtVectorType(QualType vecType, unsigned NumElts) const { 4041 assert(vecType->isBuiltinType() || vecType->isDependentType()); 4042 4043 // Check if we've already instantiated a vector of this type. 4044 llvm::FoldingSetNodeID ID; 4045 VectorType::Profile(ID, vecType, NumElts, Type::ExtVector, 4046 VectorType::GenericVector); 4047 void *InsertPos = nullptr; 4048 if (VectorType *VTP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos)) 4049 return QualType(VTP, 0); 4050 4051 // If the element type isn't canonical, this won't be a canonical type either, 4052 // so fill in the canonical type field. 4053 QualType Canonical; 4054 if (!vecType.isCanonical()) { 4055 Canonical = getExtVectorType(getCanonicalType(vecType), NumElts); 4056 4057 // Get the new insert position for the node we care about. 4058 VectorType *NewIP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos); 4059 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 4060 } 4061 auto *New = new (*this, TypeAlignment) 4062 ExtVectorType(vecType, NumElts, Canonical); 4063 VectorTypes.InsertNode(New, InsertPos); 4064 Types.push_back(New); 4065 return QualType(New, 0); 4066 } 4067 4068 QualType 4069 ASTContext::getDependentSizedExtVectorType(QualType vecType, 4070 Expr *SizeExpr, 4071 SourceLocation AttrLoc) const { 4072 llvm::FoldingSetNodeID ID; 4073 DependentSizedExtVectorType::Profile(ID, *this, getCanonicalType(vecType), 4074 SizeExpr); 4075 4076 void *InsertPos = nullptr; 4077 DependentSizedExtVectorType *Canon 4078 = DependentSizedExtVectorTypes.FindNodeOrInsertPos(ID, InsertPos); 4079 DependentSizedExtVectorType *New; 4080 if (Canon) { 4081 // We already have a canonical version of this array type; use it as 4082 // the canonical type for a newly-built type. 4083 New = new (*this, TypeAlignment) 4084 DependentSizedExtVectorType(*this, vecType, QualType(Canon, 0), 4085 SizeExpr, AttrLoc); 4086 } else { 4087 QualType CanonVecTy = getCanonicalType(vecType); 4088 if (CanonVecTy == vecType) { 4089 New = new (*this, TypeAlignment) 4090 DependentSizedExtVectorType(*this, vecType, QualType(), SizeExpr, 4091 AttrLoc); 4092 4093 DependentSizedExtVectorType *CanonCheck 4094 = DependentSizedExtVectorTypes.FindNodeOrInsertPos(ID, InsertPos); 4095 assert(!CanonCheck && "Dependent-sized ext_vector canonical type broken"); 4096 (void)CanonCheck; 4097 DependentSizedExtVectorTypes.InsertNode(New, InsertPos); 4098 } else { 4099 QualType CanonExtTy = getDependentSizedExtVectorType(CanonVecTy, SizeExpr, 4100 SourceLocation()); 4101 New = new (*this, TypeAlignment) DependentSizedExtVectorType( 4102 *this, vecType, CanonExtTy, SizeExpr, AttrLoc); 4103 } 4104 } 4105 4106 Types.push_back(New); 4107 return QualType(New, 0); 4108 } 4109 4110 QualType ASTContext::getConstantMatrixType(QualType ElementTy, unsigned NumRows, 4111 unsigned NumColumns) const { 4112 llvm::FoldingSetNodeID ID; 4113 ConstantMatrixType::Profile(ID, ElementTy, NumRows, NumColumns, 4114 Type::ConstantMatrix); 4115 4116 assert(MatrixType::isValidElementType(ElementTy) && 4117 "need a valid element type"); 4118 assert(ConstantMatrixType::isDimensionValid(NumRows) && 4119 ConstantMatrixType::isDimensionValid(NumColumns) && 4120 "need valid matrix dimensions"); 4121 void *InsertPos = nullptr; 4122 if (ConstantMatrixType *MTP = MatrixTypes.FindNodeOrInsertPos(ID, InsertPos)) 4123 return QualType(MTP, 0); 4124 4125 QualType Canonical; 4126 if (!ElementTy.isCanonical()) { 4127 Canonical = 4128 getConstantMatrixType(getCanonicalType(ElementTy), NumRows, NumColumns); 4129 4130 ConstantMatrixType *NewIP = MatrixTypes.FindNodeOrInsertPos(ID, InsertPos); 4131 assert(!NewIP && "Matrix type shouldn't already exist in the map"); 4132 (void)NewIP; 4133 } 4134 4135 auto *New = new (*this, TypeAlignment) 4136 ConstantMatrixType(ElementTy, NumRows, NumColumns, Canonical); 4137 MatrixTypes.InsertNode(New, InsertPos); 4138 Types.push_back(New); 4139 return QualType(New, 0); 4140 } 4141 4142 QualType ASTContext::getDependentSizedMatrixType(QualType ElementTy, 4143 Expr *RowExpr, 4144 Expr *ColumnExpr, 4145 SourceLocation AttrLoc) const { 4146 QualType CanonElementTy = getCanonicalType(ElementTy); 4147 llvm::FoldingSetNodeID ID; 4148 DependentSizedMatrixType::Profile(ID, *this, CanonElementTy, RowExpr, 4149 ColumnExpr); 4150 4151 void *InsertPos = nullptr; 4152 DependentSizedMatrixType *Canon = 4153 DependentSizedMatrixTypes.FindNodeOrInsertPos(ID, InsertPos); 4154 4155 if (!Canon) { 4156 Canon = new (*this, TypeAlignment) DependentSizedMatrixType( 4157 *this, CanonElementTy, QualType(), RowExpr, ColumnExpr, AttrLoc); 4158 #ifndef NDEBUG 4159 DependentSizedMatrixType *CanonCheck = 4160 DependentSizedMatrixTypes.FindNodeOrInsertPos(ID, InsertPos); 4161 assert(!CanonCheck && "Dependent-sized matrix canonical type broken"); 4162 #endif 4163 DependentSizedMatrixTypes.InsertNode(Canon, InsertPos); 4164 Types.push_back(Canon); 4165 } 4166 4167 // Already have a canonical version of the matrix type 4168 // 4169 // If it exactly matches the requested type, use it directly. 4170 if (Canon->getElementType() == ElementTy && Canon->getRowExpr() == RowExpr && 4171 Canon->getRowExpr() == ColumnExpr) 4172 return QualType(Canon, 0); 4173 4174 // Use Canon as the canonical type for newly-built type. 4175 DependentSizedMatrixType *New = new (*this, TypeAlignment) 4176 DependentSizedMatrixType(*this, ElementTy, QualType(Canon, 0), RowExpr, 4177 ColumnExpr, AttrLoc); 4178 Types.push_back(New); 4179 return QualType(New, 0); 4180 } 4181 4182 QualType ASTContext::getDependentAddressSpaceType(QualType PointeeType, 4183 Expr *AddrSpaceExpr, 4184 SourceLocation AttrLoc) const { 4185 assert(AddrSpaceExpr->isInstantiationDependent()); 4186 4187 QualType canonPointeeType = getCanonicalType(PointeeType); 4188 4189 void *insertPos = nullptr; 4190 llvm::FoldingSetNodeID ID; 4191 DependentAddressSpaceType::Profile(ID, *this, canonPointeeType, 4192 AddrSpaceExpr); 4193 4194 DependentAddressSpaceType *canonTy = 4195 DependentAddressSpaceTypes.FindNodeOrInsertPos(ID, insertPos); 4196 4197 if (!canonTy) { 4198 canonTy = new (*this, TypeAlignment) 4199 DependentAddressSpaceType(*this, canonPointeeType, 4200 QualType(), AddrSpaceExpr, AttrLoc); 4201 DependentAddressSpaceTypes.InsertNode(canonTy, insertPos); 4202 Types.push_back(canonTy); 4203 } 4204 4205 if (canonPointeeType == PointeeType && 4206 canonTy->getAddrSpaceExpr() == AddrSpaceExpr) 4207 return QualType(canonTy, 0); 4208 4209 auto *sugaredType 4210 = new (*this, TypeAlignment) 4211 DependentAddressSpaceType(*this, PointeeType, QualType(canonTy, 0), 4212 AddrSpaceExpr, AttrLoc); 4213 Types.push_back(sugaredType); 4214 return QualType(sugaredType, 0); 4215 } 4216 4217 /// Determine whether \p T is canonical as the result type of a function. 4218 static bool isCanonicalResultType(QualType T) { 4219 return T.isCanonical() && 4220 (T.getObjCLifetime() == Qualifiers::OCL_None || 4221 T.getObjCLifetime() == Qualifiers::OCL_ExplicitNone); 4222 } 4223 4224 /// getFunctionNoProtoType - Return a K&R style C function type like 'int()'. 4225 QualType 4226 ASTContext::getFunctionNoProtoType(QualType ResultTy, 4227 const FunctionType::ExtInfo &Info) const { 4228 // Unique functions, to guarantee there is only one function of a particular 4229 // structure. 4230 llvm::FoldingSetNodeID ID; 4231 FunctionNoProtoType::Profile(ID, ResultTy, Info); 4232 4233 void *InsertPos = nullptr; 4234 if (FunctionNoProtoType *FT = 4235 FunctionNoProtoTypes.FindNodeOrInsertPos(ID, InsertPos)) 4236 return QualType(FT, 0); 4237 4238 QualType Canonical; 4239 if (!isCanonicalResultType(ResultTy)) { 4240 Canonical = 4241 getFunctionNoProtoType(getCanonicalFunctionResultType(ResultTy), Info); 4242 4243 // Get the new insert position for the node we care about. 4244 FunctionNoProtoType *NewIP = 4245 FunctionNoProtoTypes.FindNodeOrInsertPos(ID, InsertPos); 4246 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 4247 } 4248 4249 auto *New = new (*this, TypeAlignment) 4250 FunctionNoProtoType(ResultTy, Canonical, Info); 4251 Types.push_back(New); 4252 FunctionNoProtoTypes.InsertNode(New, InsertPos); 4253 return QualType(New, 0); 4254 } 4255 4256 CanQualType 4257 ASTContext::getCanonicalFunctionResultType(QualType ResultType) const { 4258 CanQualType CanResultType = getCanonicalType(ResultType); 4259 4260 // Canonical result types do not have ARC lifetime qualifiers. 4261 if (CanResultType.getQualifiers().hasObjCLifetime()) { 4262 Qualifiers Qs = CanResultType.getQualifiers(); 4263 Qs.removeObjCLifetime(); 4264 return CanQualType::CreateUnsafe( 4265 getQualifiedType(CanResultType.getUnqualifiedType(), Qs)); 4266 } 4267 4268 return CanResultType; 4269 } 4270 4271 static bool isCanonicalExceptionSpecification( 4272 const FunctionProtoType::ExceptionSpecInfo &ESI, bool NoexceptInType) { 4273 if (ESI.Type == EST_None) 4274 return true; 4275 if (!NoexceptInType) 4276 return false; 4277 4278 // C++17 onwards: exception specification is part of the type, as a simple 4279 // boolean "can this function type throw". 4280 if (ESI.Type == EST_BasicNoexcept) 4281 return true; 4282 4283 // A noexcept(expr) specification is (possibly) canonical if expr is 4284 // value-dependent. 4285 if (ESI.Type == EST_DependentNoexcept) 4286 return true; 4287 4288 // A dynamic exception specification is canonical if it only contains pack 4289 // expansions (so we can't tell whether it's non-throwing) and all its 4290 // contained types are canonical. 4291 if (ESI.Type == EST_Dynamic) { 4292 bool AnyPackExpansions = false; 4293 for (QualType ET : ESI.Exceptions) { 4294 if (!ET.isCanonical()) 4295 return false; 4296 if (ET->getAs<PackExpansionType>()) 4297 AnyPackExpansions = true; 4298 } 4299 return AnyPackExpansions; 4300 } 4301 4302 return false; 4303 } 4304 4305 QualType ASTContext::getFunctionTypeInternal( 4306 QualType ResultTy, ArrayRef<QualType> ArgArray, 4307 const FunctionProtoType::ExtProtoInfo &EPI, bool OnlyWantCanonical) const { 4308 size_t NumArgs = ArgArray.size(); 4309 4310 // Unique functions, to guarantee there is only one function of a particular 4311 // structure. 4312 llvm::FoldingSetNodeID ID; 4313 FunctionProtoType::Profile(ID, ResultTy, ArgArray.begin(), NumArgs, EPI, 4314 *this, true); 4315 4316 QualType Canonical; 4317 bool Unique = false; 4318 4319 void *InsertPos = nullptr; 4320 if (FunctionProtoType *FPT = 4321 FunctionProtoTypes.FindNodeOrInsertPos(ID, InsertPos)) { 4322 QualType Existing = QualType(FPT, 0); 4323 4324 // If we find a pre-existing equivalent FunctionProtoType, we can just reuse 4325 // it so long as our exception specification doesn't contain a dependent 4326 // noexcept expression, or we're just looking for a canonical type. 4327 // Otherwise, we're going to need to create a type 4328 // sugar node to hold the concrete expression. 4329 if (OnlyWantCanonical || !isComputedNoexcept(EPI.ExceptionSpec.Type) || 4330 EPI.ExceptionSpec.NoexceptExpr == FPT->getNoexceptExpr()) 4331 return Existing; 4332 4333 // We need a new type sugar node for this one, to hold the new noexcept 4334 // expression. We do no canonicalization here, but that's OK since we don't 4335 // expect to see the same noexcept expression much more than once. 4336 Canonical = getCanonicalType(Existing); 4337 Unique = true; 4338 } 4339 4340 bool NoexceptInType = getLangOpts().CPlusPlus17; 4341 bool IsCanonicalExceptionSpec = 4342 isCanonicalExceptionSpecification(EPI.ExceptionSpec, NoexceptInType); 4343 4344 // Determine whether the type being created is already canonical or not. 4345 bool isCanonical = !Unique && IsCanonicalExceptionSpec && 4346 isCanonicalResultType(ResultTy) && !EPI.HasTrailingReturn; 4347 for (unsigned i = 0; i != NumArgs && isCanonical; ++i) 4348 if (!ArgArray[i].isCanonicalAsParam()) 4349 isCanonical = false; 4350 4351 if (OnlyWantCanonical) 4352 assert(isCanonical && 4353 "given non-canonical parameters constructing canonical type"); 4354 4355 // If this type isn't canonical, get the canonical version of it if we don't 4356 // already have it. The exception spec is only partially part of the 4357 // canonical type, and only in C++17 onwards. 4358 if (!isCanonical && Canonical.isNull()) { 4359 SmallVector<QualType, 16> CanonicalArgs; 4360 CanonicalArgs.reserve(NumArgs); 4361 for (unsigned i = 0; i != NumArgs; ++i) 4362 CanonicalArgs.push_back(getCanonicalParamType(ArgArray[i])); 4363 4364 llvm::SmallVector<QualType, 8> ExceptionTypeStorage; 4365 FunctionProtoType::ExtProtoInfo CanonicalEPI = EPI; 4366 CanonicalEPI.HasTrailingReturn = false; 4367 4368 if (IsCanonicalExceptionSpec) { 4369 // Exception spec is already OK. 4370 } else if (NoexceptInType) { 4371 switch (EPI.ExceptionSpec.Type) { 4372 case EST_Unparsed: case EST_Unevaluated: case EST_Uninstantiated: 4373 // We don't know yet. It shouldn't matter what we pick here; no-one 4374 // should ever look at this. 4375 LLVM_FALLTHROUGH; 4376 case EST_None: case EST_MSAny: case EST_NoexceptFalse: 4377 CanonicalEPI.ExceptionSpec.Type = EST_None; 4378 break; 4379 4380 // A dynamic exception specification is almost always "not noexcept", 4381 // with the exception that a pack expansion might expand to no types. 4382 case EST_Dynamic: { 4383 bool AnyPacks = false; 4384 for (QualType ET : EPI.ExceptionSpec.Exceptions) { 4385 if (ET->getAs<PackExpansionType>()) 4386 AnyPacks = true; 4387 ExceptionTypeStorage.push_back(getCanonicalType(ET)); 4388 } 4389 if (!AnyPacks) 4390 CanonicalEPI.ExceptionSpec.Type = EST_None; 4391 else { 4392 CanonicalEPI.ExceptionSpec.Type = EST_Dynamic; 4393 CanonicalEPI.ExceptionSpec.Exceptions = ExceptionTypeStorage; 4394 } 4395 break; 4396 } 4397 4398 case EST_DynamicNone: 4399 case EST_BasicNoexcept: 4400 case EST_NoexceptTrue: 4401 case EST_NoThrow: 4402 CanonicalEPI.ExceptionSpec.Type = EST_BasicNoexcept; 4403 break; 4404 4405 case EST_DependentNoexcept: 4406 llvm_unreachable("dependent noexcept is already canonical"); 4407 } 4408 } else { 4409 CanonicalEPI.ExceptionSpec = FunctionProtoType::ExceptionSpecInfo(); 4410 } 4411 4412 // Adjust the canonical function result type. 4413 CanQualType CanResultTy = getCanonicalFunctionResultType(ResultTy); 4414 Canonical = 4415 getFunctionTypeInternal(CanResultTy, CanonicalArgs, CanonicalEPI, true); 4416 4417 // Get the new insert position for the node we care about. 4418 FunctionProtoType *NewIP = 4419 FunctionProtoTypes.FindNodeOrInsertPos(ID, InsertPos); 4420 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 4421 } 4422 4423 // Compute the needed size to hold this FunctionProtoType and the 4424 // various trailing objects. 4425 auto ESH = FunctionProtoType::getExceptionSpecSize( 4426 EPI.ExceptionSpec.Type, EPI.ExceptionSpec.Exceptions.size()); 4427 size_t Size = FunctionProtoType::totalSizeToAlloc< 4428 QualType, SourceLocation, FunctionType::FunctionTypeExtraBitfields, 4429 FunctionType::ExceptionType, Expr *, FunctionDecl *, 4430 FunctionProtoType::ExtParameterInfo, Qualifiers>( 4431 NumArgs, EPI.Variadic, 4432 FunctionProtoType::hasExtraBitfields(EPI.ExceptionSpec.Type), 4433 ESH.NumExceptionType, ESH.NumExprPtr, ESH.NumFunctionDeclPtr, 4434 EPI.ExtParameterInfos ? NumArgs : 0, 4435 EPI.TypeQuals.hasNonFastQualifiers() ? 1 : 0); 4436 4437 auto *FTP = (FunctionProtoType *)Allocate(Size, TypeAlignment); 4438 FunctionProtoType::ExtProtoInfo newEPI = EPI; 4439 new (FTP) FunctionProtoType(ResultTy, ArgArray, Canonical, newEPI); 4440 Types.push_back(FTP); 4441 if (!Unique) 4442 FunctionProtoTypes.InsertNode(FTP, InsertPos); 4443 return QualType(FTP, 0); 4444 } 4445 4446 QualType ASTContext::getPipeType(QualType T, bool ReadOnly) const { 4447 llvm::FoldingSetNodeID ID; 4448 PipeType::Profile(ID, T, ReadOnly); 4449 4450 void *InsertPos = nullptr; 4451 if (PipeType *PT = PipeTypes.FindNodeOrInsertPos(ID, InsertPos)) 4452 return QualType(PT, 0); 4453 4454 // If the pipe element type isn't canonical, this won't be a canonical type 4455 // either, so fill in the canonical type field. 4456 QualType Canonical; 4457 if (!T.isCanonical()) { 4458 Canonical = getPipeType(getCanonicalType(T), ReadOnly); 4459 4460 // Get the new insert position for the node we care about. 4461 PipeType *NewIP = PipeTypes.FindNodeOrInsertPos(ID, InsertPos); 4462 assert(!NewIP && "Shouldn't be in the map!"); 4463 (void)NewIP; 4464 } 4465 auto *New = new (*this, TypeAlignment) PipeType(T, Canonical, ReadOnly); 4466 Types.push_back(New); 4467 PipeTypes.InsertNode(New, InsertPos); 4468 return QualType(New, 0); 4469 } 4470 4471 QualType ASTContext::adjustStringLiteralBaseType(QualType Ty) const { 4472 // OpenCL v1.1 s6.5.3: a string literal is in the constant address space. 4473 return LangOpts.OpenCL ? getAddrSpaceQualType(Ty, LangAS::opencl_constant) 4474 : Ty; 4475 } 4476 4477 QualType ASTContext::getReadPipeType(QualType T) const { 4478 return getPipeType(T, true); 4479 } 4480 4481 QualType ASTContext::getWritePipeType(QualType T) const { 4482 return getPipeType(T, false); 4483 } 4484 4485 QualType ASTContext::getExtIntType(bool IsUnsigned, unsigned NumBits) const { 4486 llvm::FoldingSetNodeID ID; 4487 ExtIntType::Profile(ID, IsUnsigned, NumBits); 4488 4489 void *InsertPos = nullptr; 4490 if (ExtIntType *EIT = ExtIntTypes.FindNodeOrInsertPos(ID, InsertPos)) 4491 return QualType(EIT, 0); 4492 4493 auto *New = new (*this, TypeAlignment) ExtIntType(IsUnsigned, NumBits); 4494 ExtIntTypes.InsertNode(New, InsertPos); 4495 Types.push_back(New); 4496 return QualType(New, 0); 4497 } 4498 4499 QualType ASTContext::getDependentExtIntType(bool IsUnsigned, 4500 Expr *NumBitsExpr) const { 4501 assert(NumBitsExpr->isInstantiationDependent() && "Only good for dependent"); 4502 llvm::FoldingSetNodeID ID; 4503 DependentExtIntType::Profile(ID, *this, IsUnsigned, NumBitsExpr); 4504 4505 void *InsertPos = nullptr; 4506 if (DependentExtIntType *Existing = 4507 DependentExtIntTypes.FindNodeOrInsertPos(ID, InsertPos)) 4508 return QualType(Existing, 0); 4509 4510 auto *New = new (*this, TypeAlignment) 4511 DependentExtIntType(*this, IsUnsigned, NumBitsExpr); 4512 DependentExtIntTypes.InsertNode(New, InsertPos); 4513 4514 Types.push_back(New); 4515 return QualType(New, 0); 4516 } 4517 4518 #ifndef NDEBUG 4519 static bool NeedsInjectedClassNameType(const RecordDecl *D) { 4520 if (!isa<CXXRecordDecl>(D)) return false; 4521 const auto *RD = cast<CXXRecordDecl>(D); 4522 if (isa<ClassTemplatePartialSpecializationDecl>(RD)) 4523 return true; 4524 if (RD->getDescribedClassTemplate() && 4525 !isa<ClassTemplateSpecializationDecl>(RD)) 4526 return true; 4527 return false; 4528 } 4529 #endif 4530 4531 /// getInjectedClassNameType - Return the unique reference to the 4532 /// injected class name type for the specified templated declaration. 4533 QualType ASTContext::getInjectedClassNameType(CXXRecordDecl *Decl, 4534 QualType TST) const { 4535 assert(NeedsInjectedClassNameType(Decl)); 4536 if (Decl->TypeForDecl) { 4537 assert(isa<InjectedClassNameType>(Decl->TypeForDecl)); 4538 } else if (CXXRecordDecl *PrevDecl = Decl->getPreviousDecl()) { 4539 assert(PrevDecl->TypeForDecl && "previous declaration has no type"); 4540 Decl->TypeForDecl = PrevDecl->TypeForDecl; 4541 assert(isa<InjectedClassNameType>(Decl->TypeForDecl)); 4542 } else { 4543 Type *newType = 4544 new (*this, TypeAlignment) InjectedClassNameType(Decl, TST); 4545 Decl->TypeForDecl = newType; 4546 Types.push_back(newType); 4547 } 4548 return QualType(Decl->TypeForDecl, 0); 4549 } 4550 4551 /// getTypeDeclType - Return the unique reference to the type for the 4552 /// specified type declaration. 4553 QualType ASTContext::getTypeDeclTypeSlow(const TypeDecl *Decl) const { 4554 assert(Decl && "Passed null for Decl param"); 4555 assert(!Decl->TypeForDecl && "TypeForDecl present in slow case"); 4556 4557 if (const auto *Typedef = dyn_cast<TypedefNameDecl>(Decl)) 4558 return getTypedefType(Typedef); 4559 4560 assert(!isa<TemplateTypeParmDecl>(Decl) && 4561 "Template type parameter types are always available."); 4562 4563 if (const auto *Record = dyn_cast<RecordDecl>(Decl)) { 4564 assert(Record->isFirstDecl() && "struct/union has previous declaration"); 4565 assert(!NeedsInjectedClassNameType(Record)); 4566 return getRecordType(Record); 4567 } else if (const auto *Enum = dyn_cast<EnumDecl>(Decl)) { 4568 assert(Enum->isFirstDecl() && "enum has previous declaration"); 4569 return getEnumType(Enum); 4570 } else if (const auto *Using = dyn_cast<UnresolvedUsingTypenameDecl>(Decl)) { 4571 Type *newType = new (*this, TypeAlignment) UnresolvedUsingType(Using); 4572 Decl->TypeForDecl = newType; 4573 Types.push_back(newType); 4574 } else 4575 llvm_unreachable("TypeDecl without a type?"); 4576 4577 return QualType(Decl->TypeForDecl, 0); 4578 } 4579 4580 /// getTypedefType - Return the unique reference to the type for the 4581 /// specified typedef name decl. 4582 QualType ASTContext::getTypedefType(const TypedefNameDecl *Decl, 4583 QualType Underlying) const { 4584 if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0); 4585 4586 if (Underlying.isNull()) 4587 Underlying = Decl->getUnderlyingType(); 4588 QualType Canonical = getCanonicalType(Underlying); 4589 auto *newType = new (*this, TypeAlignment) 4590 TypedefType(Type::Typedef, Decl, Underlying, Canonical); 4591 Decl->TypeForDecl = newType; 4592 Types.push_back(newType); 4593 return QualType(newType, 0); 4594 } 4595 4596 QualType ASTContext::getRecordType(const RecordDecl *Decl) const { 4597 if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0); 4598 4599 if (const RecordDecl *PrevDecl = Decl->getPreviousDecl()) 4600 if (PrevDecl->TypeForDecl) 4601 return QualType(Decl->TypeForDecl = PrevDecl->TypeForDecl, 0); 4602 4603 auto *newType = new (*this, TypeAlignment) RecordType(Decl); 4604 Decl->TypeForDecl = newType; 4605 Types.push_back(newType); 4606 return QualType(newType, 0); 4607 } 4608 4609 QualType ASTContext::getEnumType(const EnumDecl *Decl) const { 4610 if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0); 4611 4612 if (const EnumDecl *PrevDecl = Decl->getPreviousDecl()) 4613 if (PrevDecl->TypeForDecl) 4614 return QualType(Decl->TypeForDecl = PrevDecl->TypeForDecl, 0); 4615 4616 auto *newType = new (*this, TypeAlignment) EnumType(Decl); 4617 Decl->TypeForDecl = newType; 4618 Types.push_back(newType); 4619 return QualType(newType, 0); 4620 } 4621 4622 QualType ASTContext::getAttributedType(attr::Kind attrKind, 4623 QualType modifiedType, 4624 QualType equivalentType) { 4625 llvm::FoldingSetNodeID id; 4626 AttributedType::Profile(id, attrKind, modifiedType, equivalentType); 4627 4628 void *insertPos = nullptr; 4629 AttributedType *type = AttributedTypes.FindNodeOrInsertPos(id, insertPos); 4630 if (type) return QualType(type, 0); 4631 4632 QualType canon = getCanonicalType(equivalentType); 4633 type = new (*this, TypeAlignment) 4634 AttributedType(canon, attrKind, modifiedType, equivalentType); 4635 4636 Types.push_back(type); 4637 AttributedTypes.InsertNode(type, insertPos); 4638 4639 return QualType(type, 0); 4640 } 4641 4642 /// Retrieve a substitution-result type. 4643 QualType 4644 ASTContext::getSubstTemplateTypeParmType(const TemplateTypeParmType *Parm, 4645 QualType Replacement) const { 4646 assert(Replacement.isCanonical() 4647 && "replacement types must always be canonical"); 4648 4649 llvm::FoldingSetNodeID ID; 4650 SubstTemplateTypeParmType::Profile(ID, Parm, Replacement); 4651 void *InsertPos = nullptr; 4652 SubstTemplateTypeParmType *SubstParm 4653 = SubstTemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos); 4654 4655 if (!SubstParm) { 4656 SubstParm = new (*this, TypeAlignment) 4657 SubstTemplateTypeParmType(Parm, Replacement); 4658 Types.push_back(SubstParm); 4659 SubstTemplateTypeParmTypes.InsertNode(SubstParm, InsertPos); 4660 } 4661 4662 return QualType(SubstParm, 0); 4663 } 4664 4665 /// Retrieve a 4666 QualType ASTContext::getSubstTemplateTypeParmPackType( 4667 const TemplateTypeParmType *Parm, 4668 const TemplateArgument &ArgPack) { 4669 #ifndef NDEBUG 4670 for (const auto &P : ArgPack.pack_elements()) { 4671 assert(P.getKind() == TemplateArgument::Type &&"Pack contains a non-type"); 4672 assert(P.getAsType().isCanonical() && "Pack contains non-canonical type"); 4673 } 4674 #endif 4675 4676 llvm::FoldingSetNodeID ID; 4677 SubstTemplateTypeParmPackType::Profile(ID, Parm, ArgPack); 4678 void *InsertPos = nullptr; 4679 if (SubstTemplateTypeParmPackType *SubstParm 4680 = SubstTemplateTypeParmPackTypes.FindNodeOrInsertPos(ID, InsertPos)) 4681 return QualType(SubstParm, 0); 4682 4683 QualType Canon; 4684 if (!Parm->isCanonicalUnqualified()) { 4685 Canon = getCanonicalType(QualType(Parm, 0)); 4686 Canon = getSubstTemplateTypeParmPackType(cast<TemplateTypeParmType>(Canon), 4687 ArgPack); 4688 SubstTemplateTypeParmPackTypes.FindNodeOrInsertPos(ID, InsertPos); 4689 } 4690 4691 auto *SubstParm 4692 = new (*this, TypeAlignment) SubstTemplateTypeParmPackType(Parm, Canon, 4693 ArgPack); 4694 Types.push_back(SubstParm); 4695 SubstTemplateTypeParmPackTypes.InsertNode(SubstParm, InsertPos); 4696 return QualType(SubstParm, 0); 4697 } 4698 4699 /// Retrieve the template type parameter type for a template 4700 /// parameter or parameter pack with the given depth, index, and (optionally) 4701 /// name. 4702 QualType ASTContext::getTemplateTypeParmType(unsigned Depth, unsigned Index, 4703 bool ParameterPack, 4704 TemplateTypeParmDecl *TTPDecl) const { 4705 llvm::FoldingSetNodeID ID; 4706 TemplateTypeParmType::Profile(ID, Depth, Index, ParameterPack, TTPDecl); 4707 void *InsertPos = nullptr; 4708 TemplateTypeParmType *TypeParm 4709 = TemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos); 4710 4711 if (TypeParm) 4712 return QualType(TypeParm, 0); 4713 4714 if (TTPDecl) { 4715 QualType Canon = getTemplateTypeParmType(Depth, Index, ParameterPack); 4716 TypeParm = new (*this, TypeAlignment) TemplateTypeParmType(TTPDecl, Canon); 4717 4718 TemplateTypeParmType *TypeCheck 4719 = TemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos); 4720 assert(!TypeCheck && "Template type parameter canonical type broken"); 4721 (void)TypeCheck; 4722 } else 4723 TypeParm = new (*this, TypeAlignment) 4724 TemplateTypeParmType(Depth, Index, ParameterPack); 4725 4726 Types.push_back(TypeParm); 4727 TemplateTypeParmTypes.InsertNode(TypeParm, InsertPos); 4728 4729 return QualType(TypeParm, 0); 4730 } 4731 4732 TypeSourceInfo * 4733 ASTContext::getTemplateSpecializationTypeInfo(TemplateName Name, 4734 SourceLocation NameLoc, 4735 const TemplateArgumentListInfo &Args, 4736 QualType Underlying) const { 4737 assert(!Name.getAsDependentTemplateName() && 4738 "No dependent template names here!"); 4739 QualType TST = getTemplateSpecializationType(Name, Args, Underlying); 4740 4741 TypeSourceInfo *DI = CreateTypeSourceInfo(TST); 4742 TemplateSpecializationTypeLoc TL = 4743 DI->getTypeLoc().castAs<TemplateSpecializationTypeLoc>(); 4744 TL.setTemplateKeywordLoc(SourceLocation()); 4745 TL.setTemplateNameLoc(NameLoc); 4746 TL.setLAngleLoc(Args.getLAngleLoc()); 4747 TL.setRAngleLoc(Args.getRAngleLoc()); 4748 for (unsigned i = 0, e = TL.getNumArgs(); i != e; ++i) 4749 TL.setArgLocInfo(i, Args[i].getLocInfo()); 4750 return DI; 4751 } 4752 4753 QualType 4754 ASTContext::getTemplateSpecializationType(TemplateName Template, 4755 const TemplateArgumentListInfo &Args, 4756 QualType Underlying) const { 4757 assert(!Template.getAsDependentTemplateName() && 4758 "No dependent template names here!"); 4759 4760 SmallVector<TemplateArgument, 4> ArgVec; 4761 ArgVec.reserve(Args.size()); 4762 for (const TemplateArgumentLoc &Arg : Args.arguments()) 4763 ArgVec.push_back(Arg.getArgument()); 4764 4765 return getTemplateSpecializationType(Template, ArgVec, Underlying); 4766 } 4767 4768 #ifndef NDEBUG 4769 static bool hasAnyPackExpansions(ArrayRef<TemplateArgument> Args) { 4770 for (const TemplateArgument &Arg : Args) 4771 if (Arg.isPackExpansion()) 4772 return true; 4773 4774 return true; 4775 } 4776 #endif 4777 4778 QualType 4779 ASTContext::getTemplateSpecializationType(TemplateName Template, 4780 ArrayRef<TemplateArgument> Args, 4781 QualType Underlying) const { 4782 assert(!Template.getAsDependentTemplateName() && 4783 "No dependent template names here!"); 4784 // Look through qualified template names. 4785 if (QualifiedTemplateName *QTN = Template.getAsQualifiedTemplateName()) 4786 Template = TemplateName(QTN->getTemplateDecl()); 4787 4788 bool IsTypeAlias = 4789 Template.getAsTemplateDecl() && 4790 isa<TypeAliasTemplateDecl>(Template.getAsTemplateDecl()); 4791 QualType CanonType; 4792 if (!Underlying.isNull()) 4793 CanonType = getCanonicalType(Underlying); 4794 else { 4795 // We can get here with an alias template when the specialization contains 4796 // a pack expansion that does not match up with a parameter pack. 4797 assert((!IsTypeAlias || hasAnyPackExpansions(Args)) && 4798 "Caller must compute aliased type"); 4799 IsTypeAlias = false; 4800 CanonType = getCanonicalTemplateSpecializationType(Template, Args); 4801 } 4802 4803 // Allocate the (non-canonical) template specialization type, but don't 4804 // try to unique it: these types typically have location information that 4805 // we don't unique and don't want to lose. 4806 void *Mem = Allocate(sizeof(TemplateSpecializationType) + 4807 sizeof(TemplateArgument) * Args.size() + 4808 (IsTypeAlias? sizeof(QualType) : 0), 4809 TypeAlignment); 4810 auto *Spec 4811 = new (Mem) TemplateSpecializationType(Template, Args, CanonType, 4812 IsTypeAlias ? Underlying : QualType()); 4813 4814 Types.push_back(Spec); 4815 return QualType(Spec, 0); 4816 } 4817 4818 QualType ASTContext::getCanonicalTemplateSpecializationType( 4819 TemplateName Template, ArrayRef<TemplateArgument> Args) const { 4820 assert(!Template.getAsDependentTemplateName() && 4821 "No dependent template names here!"); 4822 4823 // Look through qualified template names. 4824 if (QualifiedTemplateName *QTN = Template.getAsQualifiedTemplateName()) 4825 Template = TemplateName(QTN->getTemplateDecl()); 4826 4827 // Build the canonical template specialization type. 4828 TemplateName CanonTemplate = getCanonicalTemplateName(Template); 4829 SmallVector<TemplateArgument, 4> CanonArgs; 4830 unsigned NumArgs = Args.size(); 4831 CanonArgs.reserve(NumArgs); 4832 for (const TemplateArgument &Arg : Args) 4833 CanonArgs.push_back(getCanonicalTemplateArgument(Arg)); 4834 4835 // Determine whether this canonical template specialization type already 4836 // exists. 4837 llvm::FoldingSetNodeID ID; 4838 TemplateSpecializationType::Profile(ID, CanonTemplate, 4839 CanonArgs, *this); 4840 4841 void *InsertPos = nullptr; 4842 TemplateSpecializationType *Spec 4843 = TemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos); 4844 4845 if (!Spec) { 4846 // Allocate a new canonical template specialization type. 4847 void *Mem = Allocate((sizeof(TemplateSpecializationType) + 4848 sizeof(TemplateArgument) * NumArgs), 4849 TypeAlignment); 4850 Spec = new (Mem) TemplateSpecializationType(CanonTemplate, 4851 CanonArgs, 4852 QualType(), QualType()); 4853 Types.push_back(Spec); 4854 TemplateSpecializationTypes.InsertNode(Spec, InsertPos); 4855 } 4856 4857 assert(Spec->isDependentType() && 4858 "Non-dependent template-id type must have a canonical type"); 4859 return QualType(Spec, 0); 4860 } 4861 4862 QualType ASTContext::getElaboratedType(ElaboratedTypeKeyword Keyword, 4863 NestedNameSpecifier *NNS, 4864 QualType NamedType, 4865 TagDecl *OwnedTagDecl) const { 4866 llvm::FoldingSetNodeID ID; 4867 ElaboratedType::Profile(ID, Keyword, NNS, NamedType, OwnedTagDecl); 4868 4869 void *InsertPos = nullptr; 4870 ElaboratedType *T = ElaboratedTypes.FindNodeOrInsertPos(ID, InsertPos); 4871 if (T) 4872 return QualType(T, 0); 4873 4874 QualType Canon = NamedType; 4875 if (!Canon.isCanonical()) { 4876 Canon = getCanonicalType(NamedType); 4877 ElaboratedType *CheckT = ElaboratedTypes.FindNodeOrInsertPos(ID, InsertPos); 4878 assert(!CheckT && "Elaborated canonical type broken"); 4879 (void)CheckT; 4880 } 4881 4882 void *Mem = Allocate(ElaboratedType::totalSizeToAlloc<TagDecl *>(!!OwnedTagDecl), 4883 TypeAlignment); 4884 T = new (Mem) ElaboratedType(Keyword, NNS, NamedType, Canon, OwnedTagDecl); 4885 4886 Types.push_back(T); 4887 ElaboratedTypes.InsertNode(T, InsertPos); 4888 return QualType(T, 0); 4889 } 4890 4891 QualType 4892 ASTContext::getParenType(QualType InnerType) const { 4893 llvm::FoldingSetNodeID ID; 4894 ParenType::Profile(ID, InnerType); 4895 4896 void *InsertPos = nullptr; 4897 ParenType *T = ParenTypes.FindNodeOrInsertPos(ID, InsertPos); 4898 if (T) 4899 return QualType(T, 0); 4900 4901 QualType Canon = InnerType; 4902 if (!Canon.isCanonical()) { 4903 Canon = getCanonicalType(InnerType); 4904 ParenType *CheckT = ParenTypes.FindNodeOrInsertPos(ID, InsertPos); 4905 assert(!CheckT && "Paren canonical type broken"); 4906 (void)CheckT; 4907 } 4908 4909 T = new (*this, TypeAlignment) ParenType(InnerType, Canon); 4910 Types.push_back(T); 4911 ParenTypes.InsertNode(T, InsertPos); 4912 return QualType(T, 0); 4913 } 4914 4915 QualType 4916 ASTContext::getMacroQualifiedType(QualType UnderlyingTy, 4917 const IdentifierInfo *MacroII) const { 4918 QualType Canon = UnderlyingTy; 4919 if (!Canon.isCanonical()) 4920 Canon = getCanonicalType(UnderlyingTy); 4921 4922 auto *newType = new (*this, TypeAlignment) 4923 MacroQualifiedType(UnderlyingTy, Canon, MacroII); 4924 Types.push_back(newType); 4925 return QualType(newType, 0); 4926 } 4927 4928 QualType ASTContext::getDependentNameType(ElaboratedTypeKeyword Keyword, 4929 NestedNameSpecifier *NNS, 4930 const IdentifierInfo *Name, 4931 QualType Canon) const { 4932 if (Canon.isNull()) { 4933 NestedNameSpecifier *CanonNNS = getCanonicalNestedNameSpecifier(NNS); 4934 if (CanonNNS != NNS) 4935 Canon = getDependentNameType(Keyword, CanonNNS, Name); 4936 } 4937 4938 llvm::FoldingSetNodeID ID; 4939 DependentNameType::Profile(ID, Keyword, NNS, Name); 4940 4941 void *InsertPos = nullptr; 4942 DependentNameType *T 4943 = DependentNameTypes.FindNodeOrInsertPos(ID, InsertPos); 4944 if (T) 4945 return QualType(T, 0); 4946 4947 T = new (*this, TypeAlignment) DependentNameType(Keyword, NNS, Name, Canon); 4948 Types.push_back(T); 4949 DependentNameTypes.InsertNode(T, InsertPos); 4950 return QualType(T, 0); 4951 } 4952 4953 QualType 4954 ASTContext::getDependentTemplateSpecializationType( 4955 ElaboratedTypeKeyword Keyword, 4956 NestedNameSpecifier *NNS, 4957 const IdentifierInfo *Name, 4958 const TemplateArgumentListInfo &Args) const { 4959 // TODO: avoid this copy 4960 SmallVector<TemplateArgument, 16> ArgCopy; 4961 for (unsigned I = 0, E = Args.size(); I != E; ++I) 4962 ArgCopy.push_back(Args[I].getArgument()); 4963 return getDependentTemplateSpecializationType(Keyword, NNS, Name, ArgCopy); 4964 } 4965 4966 QualType 4967 ASTContext::getDependentTemplateSpecializationType( 4968 ElaboratedTypeKeyword Keyword, 4969 NestedNameSpecifier *NNS, 4970 const IdentifierInfo *Name, 4971 ArrayRef<TemplateArgument> Args) const { 4972 assert((!NNS || NNS->isDependent()) && 4973 "nested-name-specifier must be dependent"); 4974 4975 llvm::FoldingSetNodeID ID; 4976 DependentTemplateSpecializationType::Profile(ID, *this, Keyword, NNS, 4977 Name, Args); 4978 4979 void *InsertPos = nullptr; 4980 DependentTemplateSpecializationType *T 4981 = DependentTemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos); 4982 if (T) 4983 return QualType(T, 0); 4984 4985 NestedNameSpecifier *CanonNNS = getCanonicalNestedNameSpecifier(NNS); 4986 4987 ElaboratedTypeKeyword CanonKeyword = Keyword; 4988 if (Keyword == ETK_None) CanonKeyword = ETK_Typename; 4989 4990 bool AnyNonCanonArgs = false; 4991 unsigned NumArgs = Args.size(); 4992 SmallVector<TemplateArgument, 16> CanonArgs(NumArgs); 4993 for (unsigned I = 0; I != NumArgs; ++I) { 4994 CanonArgs[I] = getCanonicalTemplateArgument(Args[I]); 4995 if (!CanonArgs[I].structurallyEquals(Args[I])) 4996 AnyNonCanonArgs = true; 4997 } 4998 4999 QualType Canon; 5000 if (AnyNonCanonArgs || CanonNNS != NNS || CanonKeyword != Keyword) { 5001 Canon = getDependentTemplateSpecializationType(CanonKeyword, CanonNNS, 5002 Name, 5003 CanonArgs); 5004 5005 // Find the insert position again. 5006 DependentTemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos); 5007 } 5008 5009 void *Mem = Allocate((sizeof(DependentTemplateSpecializationType) + 5010 sizeof(TemplateArgument) * NumArgs), 5011 TypeAlignment); 5012 T = new (Mem) DependentTemplateSpecializationType(Keyword, NNS, 5013 Name, Args, Canon); 5014 Types.push_back(T); 5015 DependentTemplateSpecializationTypes.InsertNode(T, InsertPos); 5016 return QualType(T, 0); 5017 } 5018 5019 TemplateArgument ASTContext::getInjectedTemplateArg(NamedDecl *Param) { 5020 TemplateArgument Arg; 5021 if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(Param)) { 5022 QualType ArgType = getTypeDeclType(TTP); 5023 if (TTP->isParameterPack()) 5024 ArgType = getPackExpansionType(ArgType, None); 5025 5026 Arg = TemplateArgument(ArgType); 5027 } else if (auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(Param)) { 5028 QualType T = 5029 NTTP->getType().getNonPackExpansionType().getNonLValueExprType(*this); 5030 // For class NTTPs, ensure we include the 'const' so the type matches that 5031 // of a real template argument. 5032 // FIXME: It would be more faithful to model this as something like an 5033 // lvalue-to-rvalue conversion applied to a const-qualified lvalue. 5034 if (T->isRecordType()) 5035 T.addConst(); 5036 Expr *E = new (*this) DeclRefExpr( 5037 *this, NTTP, /*enclosing*/ false, T, 5038 Expr::getValueKindForType(NTTP->getType()), NTTP->getLocation()); 5039 5040 if (NTTP->isParameterPack()) 5041 E = new (*this) PackExpansionExpr(DependentTy, E, NTTP->getLocation(), 5042 None); 5043 Arg = TemplateArgument(E); 5044 } else { 5045 auto *TTP = cast<TemplateTemplateParmDecl>(Param); 5046 if (TTP->isParameterPack()) 5047 Arg = TemplateArgument(TemplateName(TTP), Optional<unsigned>()); 5048 else 5049 Arg = TemplateArgument(TemplateName(TTP)); 5050 } 5051 5052 if (Param->isTemplateParameterPack()) 5053 Arg = TemplateArgument::CreatePackCopy(*this, Arg); 5054 5055 return Arg; 5056 } 5057 5058 void 5059 ASTContext::getInjectedTemplateArgs(const TemplateParameterList *Params, 5060 SmallVectorImpl<TemplateArgument> &Args) { 5061 Args.reserve(Args.size() + Params->size()); 5062 5063 for (NamedDecl *Param : *Params) 5064 Args.push_back(getInjectedTemplateArg(Param)); 5065 } 5066 5067 QualType ASTContext::getPackExpansionType(QualType Pattern, 5068 Optional<unsigned> NumExpansions, 5069 bool ExpectPackInType) { 5070 assert((!ExpectPackInType || Pattern->containsUnexpandedParameterPack()) && 5071 "Pack expansions must expand one or more parameter packs"); 5072 5073 llvm::FoldingSetNodeID ID; 5074 PackExpansionType::Profile(ID, Pattern, NumExpansions); 5075 5076 void *InsertPos = nullptr; 5077 PackExpansionType *T = PackExpansionTypes.FindNodeOrInsertPos(ID, InsertPos); 5078 if (T) 5079 return QualType(T, 0); 5080 5081 QualType Canon; 5082 if (!Pattern.isCanonical()) { 5083 Canon = getPackExpansionType(getCanonicalType(Pattern), NumExpansions, 5084 /*ExpectPackInType=*/false); 5085 5086 // Find the insert position again, in case we inserted an element into 5087 // PackExpansionTypes and invalidated our insert position. 5088 PackExpansionTypes.FindNodeOrInsertPos(ID, InsertPos); 5089 } 5090 5091 T = new (*this, TypeAlignment) 5092 PackExpansionType(Pattern, Canon, NumExpansions); 5093 Types.push_back(T); 5094 PackExpansionTypes.InsertNode(T, InsertPos); 5095 return QualType(T, 0); 5096 } 5097 5098 /// CmpProtocolNames - Comparison predicate for sorting protocols 5099 /// alphabetically. 5100 static int CmpProtocolNames(ObjCProtocolDecl *const *LHS, 5101 ObjCProtocolDecl *const *RHS) { 5102 return DeclarationName::compare((*LHS)->getDeclName(), (*RHS)->getDeclName()); 5103 } 5104 5105 static bool areSortedAndUniqued(ArrayRef<ObjCProtocolDecl *> Protocols) { 5106 if (Protocols.empty()) return true; 5107 5108 if (Protocols[0]->getCanonicalDecl() != Protocols[0]) 5109 return false; 5110 5111 for (unsigned i = 1; i != Protocols.size(); ++i) 5112 if (CmpProtocolNames(&Protocols[i - 1], &Protocols[i]) >= 0 || 5113 Protocols[i]->getCanonicalDecl() != Protocols[i]) 5114 return false; 5115 return true; 5116 } 5117 5118 static void 5119 SortAndUniqueProtocols(SmallVectorImpl<ObjCProtocolDecl *> &Protocols) { 5120 // Sort protocols, keyed by name. 5121 llvm::array_pod_sort(Protocols.begin(), Protocols.end(), CmpProtocolNames); 5122 5123 // Canonicalize. 5124 for (ObjCProtocolDecl *&P : Protocols) 5125 P = P->getCanonicalDecl(); 5126 5127 // Remove duplicates. 5128 auto ProtocolsEnd = std::unique(Protocols.begin(), Protocols.end()); 5129 Protocols.erase(ProtocolsEnd, Protocols.end()); 5130 } 5131 5132 QualType ASTContext::getObjCObjectType(QualType BaseType, 5133 ObjCProtocolDecl * const *Protocols, 5134 unsigned NumProtocols) const { 5135 return getObjCObjectType(BaseType, {}, 5136 llvm::makeArrayRef(Protocols, NumProtocols), 5137 /*isKindOf=*/false); 5138 } 5139 5140 QualType ASTContext::getObjCObjectType( 5141 QualType baseType, 5142 ArrayRef<QualType> typeArgs, 5143 ArrayRef<ObjCProtocolDecl *> protocols, 5144 bool isKindOf) const { 5145 // If the base type is an interface and there aren't any protocols or 5146 // type arguments to add, then the interface type will do just fine. 5147 if (typeArgs.empty() && protocols.empty() && !isKindOf && 5148 isa<ObjCInterfaceType>(baseType)) 5149 return baseType; 5150 5151 // Look in the folding set for an existing type. 5152 llvm::FoldingSetNodeID ID; 5153 ObjCObjectTypeImpl::Profile(ID, baseType, typeArgs, protocols, isKindOf); 5154 void *InsertPos = nullptr; 5155 if (ObjCObjectType *QT = ObjCObjectTypes.FindNodeOrInsertPos(ID, InsertPos)) 5156 return QualType(QT, 0); 5157 5158 // Determine the type arguments to be used for canonicalization, 5159 // which may be explicitly specified here or written on the base 5160 // type. 5161 ArrayRef<QualType> effectiveTypeArgs = typeArgs; 5162 if (effectiveTypeArgs.empty()) { 5163 if (const auto *baseObject = baseType->getAs<ObjCObjectType>()) 5164 effectiveTypeArgs = baseObject->getTypeArgs(); 5165 } 5166 5167 // Build the canonical type, which has the canonical base type and a 5168 // sorted-and-uniqued list of protocols and the type arguments 5169 // canonicalized. 5170 QualType canonical; 5171 bool typeArgsAreCanonical = llvm::all_of( 5172 effectiveTypeArgs, [&](QualType type) { return type.isCanonical(); }); 5173 bool protocolsSorted = areSortedAndUniqued(protocols); 5174 if (!typeArgsAreCanonical || !protocolsSorted || !baseType.isCanonical()) { 5175 // Determine the canonical type arguments. 5176 ArrayRef<QualType> canonTypeArgs; 5177 SmallVector<QualType, 4> canonTypeArgsVec; 5178 if (!typeArgsAreCanonical) { 5179 canonTypeArgsVec.reserve(effectiveTypeArgs.size()); 5180 for (auto typeArg : effectiveTypeArgs) 5181 canonTypeArgsVec.push_back(getCanonicalType(typeArg)); 5182 canonTypeArgs = canonTypeArgsVec; 5183 } else { 5184 canonTypeArgs = effectiveTypeArgs; 5185 } 5186 5187 ArrayRef<ObjCProtocolDecl *> canonProtocols; 5188 SmallVector<ObjCProtocolDecl*, 8> canonProtocolsVec; 5189 if (!protocolsSorted) { 5190 canonProtocolsVec.append(protocols.begin(), protocols.end()); 5191 SortAndUniqueProtocols(canonProtocolsVec); 5192 canonProtocols = canonProtocolsVec; 5193 } else { 5194 canonProtocols = protocols; 5195 } 5196 5197 canonical = getObjCObjectType(getCanonicalType(baseType), canonTypeArgs, 5198 canonProtocols, isKindOf); 5199 5200 // Regenerate InsertPos. 5201 ObjCObjectTypes.FindNodeOrInsertPos(ID, InsertPos); 5202 } 5203 5204 unsigned size = sizeof(ObjCObjectTypeImpl); 5205 size += typeArgs.size() * sizeof(QualType); 5206 size += protocols.size() * sizeof(ObjCProtocolDecl *); 5207 void *mem = Allocate(size, TypeAlignment); 5208 auto *T = 5209 new (mem) ObjCObjectTypeImpl(canonical, baseType, typeArgs, protocols, 5210 isKindOf); 5211 5212 Types.push_back(T); 5213 ObjCObjectTypes.InsertNode(T, InsertPos); 5214 return QualType(T, 0); 5215 } 5216 5217 /// Apply Objective-C protocol qualifiers to the given type. 5218 /// If this is for the canonical type of a type parameter, we can apply 5219 /// protocol qualifiers on the ObjCObjectPointerType. 5220 QualType 5221 ASTContext::applyObjCProtocolQualifiers(QualType type, 5222 ArrayRef<ObjCProtocolDecl *> protocols, bool &hasError, 5223 bool allowOnPointerType) const { 5224 hasError = false; 5225 5226 if (const auto *objT = dyn_cast<ObjCTypeParamType>(type.getTypePtr())) { 5227 return getObjCTypeParamType(objT->getDecl(), protocols); 5228 } 5229 5230 // Apply protocol qualifiers to ObjCObjectPointerType. 5231 if (allowOnPointerType) { 5232 if (const auto *objPtr = 5233 dyn_cast<ObjCObjectPointerType>(type.getTypePtr())) { 5234 const ObjCObjectType *objT = objPtr->getObjectType(); 5235 // Merge protocol lists and construct ObjCObjectType. 5236 SmallVector<ObjCProtocolDecl*, 8> protocolsVec; 5237 protocolsVec.append(objT->qual_begin(), 5238 objT->qual_end()); 5239 protocolsVec.append(protocols.begin(), protocols.end()); 5240 ArrayRef<ObjCProtocolDecl *> protocols = protocolsVec; 5241 type = getObjCObjectType( 5242 objT->getBaseType(), 5243 objT->getTypeArgsAsWritten(), 5244 protocols, 5245 objT->isKindOfTypeAsWritten()); 5246 return getObjCObjectPointerType(type); 5247 } 5248 } 5249 5250 // Apply protocol qualifiers to ObjCObjectType. 5251 if (const auto *objT = dyn_cast<ObjCObjectType>(type.getTypePtr())){ 5252 // FIXME: Check for protocols to which the class type is already 5253 // known to conform. 5254 5255 return getObjCObjectType(objT->getBaseType(), 5256 objT->getTypeArgsAsWritten(), 5257 protocols, 5258 objT->isKindOfTypeAsWritten()); 5259 } 5260 5261 // If the canonical type is ObjCObjectType, ... 5262 if (type->isObjCObjectType()) { 5263 // Silently overwrite any existing protocol qualifiers. 5264 // TODO: determine whether that's the right thing to do. 5265 5266 // FIXME: Check for protocols to which the class type is already 5267 // known to conform. 5268 return getObjCObjectType(type, {}, protocols, false); 5269 } 5270 5271 // id<protocol-list> 5272 if (type->isObjCIdType()) { 5273 const auto *objPtr = type->castAs<ObjCObjectPointerType>(); 5274 type = getObjCObjectType(ObjCBuiltinIdTy, {}, protocols, 5275 objPtr->isKindOfType()); 5276 return getObjCObjectPointerType(type); 5277 } 5278 5279 // Class<protocol-list> 5280 if (type->isObjCClassType()) { 5281 const auto *objPtr = type->castAs<ObjCObjectPointerType>(); 5282 type = getObjCObjectType(ObjCBuiltinClassTy, {}, protocols, 5283 objPtr->isKindOfType()); 5284 return getObjCObjectPointerType(type); 5285 } 5286 5287 hasError = true; 5288 return type; 5289 } 5290 5291 QualType 5292 ASTContext::getObjCTypeParamType(const ObjCTypeParamDecl *Decl, 5293 ArrayRef<ObjCProtocolDecl *> protocols) const { 5294 // Look in the folding set for an existing type. 5295 llvm::FoldingSetNodeID ID; 5296 ObjCTypeParamType::Profile(ID, Decl, Decl->getUnderlyingType(), protocols); 5297 void *InsertPos = nullptr; 5298 if (ObjCTypeParamType *TypeParam = 5299 ObjCTypeParamTypes.FindNodeOrInsertPos(ID, InsertPos)) 5300 return QualType(TypeParam, 0); 5301 5302 // We canonicalize to the underlying type. 5303 QualType Canonical = getCanonicalType(Decl->getUnderlyingType()); 5304 if (!protocols.empty()) { 5305 // Apply the protocol qualifers. 5306 bool hasError; 5307 Canonical = getCanonicalType(applyObjCProtocolQualifiers( 5308 Canonical, protocols, hasError, true /*allowOnPointerType*/)); 5309 assert(!hasError && "Error when apply protocol qualifier to bound type"); 5310 } 5311 5312 unsigned size = sizeof(ObjCTypeParamType); 5313 size += protocols.size() * sizeof(ObjCProtocolDecl *); 5314 void *mem = Allocate(size, TypeAlignment); 5315 auto *newType = new (mem) ObjCTypeParamType(Decl, Canonical, protocols); 5316 5317 Types.push_back(newType); 5318 ObjCTypeParamTypes.InsertNode(newType, InsertPos); 5319 return QualType(newType, 0); 5320 } 5321 5322 void ASTContext::adjustObjCTypeParamBoundType(const ObjCTypeParamDecl *Orig, 5323 ObjCTypeParamDecl *New) const { 5324 New->setTypeSourceInfo(getTrivialTypeSourceInfo(Orig->getUnderlyingType())); 5325 // Update TypeForDecl after updating TypeSourceInfo. 5326 auto NewTypeParamTy = cast<ObjCTypeParamType>(New->getTypeForDecl()); 5327 SmallVector<ObjCProtocolDecl *, 8> protocols; 5328 protocols.append(NewTypeParamTy->qual_begin(), NewTypeParamTy->qual_end()); 5329 QualType UpdatedTy = getObjCTypeParamType(New, protocols); 5330 New->setTypeForDecl(UpdatedTy.getTypePtr()); 5331 } 5332 5333 /// ObjCObjectAdoptsQTypeProtocols - Checks that protocols in IC's 5334 /// protocol list adopt all protocols in QT's qualified-id protocol 5335 /// list. 5336 bool ASTContext::ObjCObjectAdoptsQTypeProtocols(QualType QT, 5337 ObjCInterfaceDecl *IC) { 5338 if (!QT->isObjCQualifiedIdType()) 5339 return false; 5340 5341 if (const auto *OPT = QT->getAs<ObjCObjectPointerType>()) { 5342 // If both the right and left sides have qualifiers. 5343 for (auto *Proto : OPT->quals()) { 5344 if (!IC->ClassImplementsProtocol(Proto, false)) 5345 return false; 5346 } 5347 return true; 5348 } 5349 return false; 5350 } 5351 5352 /// QIdProtocolsAdoptObjCObjectProtocols - Checks that protocols in 5353 /// QT's qualified-id protocol list adopt all protocols in IDecl's list 5354 /// of protocols. 5355 bool ASTContext::QIdProtocolsAdoptObjCObjectProtocols(QualType QT, 5356 ObjCInterfaceDecl *IDecl) { 5357 if (!QT->isObjCQualifiedIdType()) 5358 return false; 5359 const auto *OPT = QT->getAs<ObjCObjectPointerType>(); 5360 if (!OPT) 5361 return false; 5362 if (!IDecl->hasDefinition()) 5363 return false; 5364 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> InheritedProtocols; 5365 CollectInheritedProtocols(IDecl, InheritedProtocols); 5366 if (InheritedProtocols.empty()) 5367 return false; 5368 // Check that if every protocol in list of id<plist> conforms to a protocol 5369 // of IDecl's, then bridge casting is ok. 5370 bool Conforms = false; 5371 for (auto *Proto : OPT->quals()) { 5372 Conforms = false; 5373 for (auto *PI : InheritedProtocols) { 5374 if (ProtocolCompatibleWithProtocol(Proto, PI)) { 5375 Conforms = true; 5376 break; 5377 } 5378 } 5379 if (!Conforms) 5380 break; 5381 } 5382 if (Conforms) 5383 return true; 5384 5385 for (auto *PI : InheritedProtocols) { 5386 // If both the right and left sides have qualifiers. 5387 bool Adopts = false; 5388 for (auto *Proto : OPT->quals()) { 5389 // return 'true' if 'PI' is in the inheritance hierarchy of Proto 5390 if ((Adopts = ProtocolCompatibleWithProtocol(PI, Proto))) 5391 break; 5392 } 5393 if (!Adopts) 5394 return false; 5395 } 5396 return true; 5397 } 5398 5399 /// getObjCObjectPointerType - Return a ObjCObjectPointerType type for 5400 /// the given object type. 5401 QualType ASTContext::getObjCObjectPointerType(QualType ObjectT) const { 5402 llvm::FoldingSetNodeID ID; 5403 ObjCObjectPointerType::Profile(ID, ObjectT); 5404 5405 void *InsertPos = nullptr; 5406 if (ObjCObjectPointerType *QT = 5407 ObjCObjectPointerTypes.FindNodeOrInsertPos(ID, InsertPos)) 5408 return QualType(QT, 0); 5409 5410 // Find the canonical object type. 5411 QualType Canonical; 5412 if (!ObjectT.isCanonical()) { 5413 Canonical = getObjCObjectPointerType(getCanonicalType(ObjectT)); 5414 5415 // Regenerate InsertPos. 5416 ObjCObjectPointerTypes.FindNodeOrInsertPos(ID, InsertPos); 5417 } 5418 5419 // No match. 5420 void *Mem = Allocate(sizeof(ObjCObjectPointerType), TypeAlignment); 5421 auto *QType = 5422 new (Mem) ObjCObjectPointerType(Canonical, ObjectT); 5423 5424 Types.push_back(QType); 5425 ObjCObjectPointerTypes.InsertNode(QType, InsertPos); 5426 return QualType(QType, 0); 5427 } 5428 5429 /// getObjCInterfaceType - Return the unique reference to the type for the 5430 /// specified ObjC interface decl. The list of protocols is optional. 5431 QualType ASTContext::getObjCInterfaceType(const ObjCInterfaceDecl *Decl, 5432 ObjCInterfaceDecl *PrevDecl) const { 5433 if (Decl->TypeForDecl) 5434 return QualType(Decl->TypeForDecl, 0); 5435 5436 if (PrevDecl) { 5437 assert(PrevDecl->TypeForDecl && "previous decl has no TypeForDecl"); 5438 Decl->TypeForDecl = PrevDecl->TypeForDecl; 5439 return QualType(PrevDecl->TypeForDecl, 0); 5440 } 5441 5442 // Prefer the definition, if there is one. 5443 if (const ObjCInterfaceDecl *Def = Decl->getDefinition()) 5444 Decl = Def; 5445 5446 void *Mem = Allocate(sizeof(ObjCInterfaceType), TypeAlignment); 5447 auto *T = new (Mem) ObjCInterfaceType(Decl); 5448 Decl->TypeForDecl = T; 5449 Types.push_back(T); 5450 return QualType(T, 0); 5451 } 5452 5453 /// getTypeOfExprType - Unlike many "get<Type>" functions, we can't unique 5454 /// TypeOfExprType AST's (since expression's are never shared). For example, 5455 /// multiple declarations that refer to "typeof(x)" all contain different 5456 /// DeclRefExpr's. This doesn't effect the type checker, since it operates 5457 /// on canonical type's (which are always unique). 5458 QualType ASTContext::getTypeOfExprType(Expr *tofExpr) const { 5459 TypeOfExprType *toe; 5460 if (tofExpr->isTypeDependent()) { 5461 llvm::FoldingSetNodeID ID; 5462 DependentTypeOfExprType::Profile(ID, *this, tofExpr); 5463 5464 void *InsertPos = nullptr; 5465 DependentTypeOfExprType *Canon 5466 = DependentTypeOfExprTypes.FindNodeOrInsertPos(ID, InsertPos); 5467 if (Canon) { 5468 // We already have a "canonical" version of an identical, dependent 5469 // typeof(expr) type. Use that as our canonical type. 5470 toe = new (*this, TypeAlignment) TypeOfExprType(tofExpr, 5471 QualType((TypeOfExprType*)Canon, 0)); 5472 } else { 5473 // Build a new, canonical typeof(expr) type. 5474 Canon 5475 = new (*this, TypeAlignment) DependentTypeOfExprType(*this, tofExpr); 5476 DependentTypeOfExprTypes.InsertNode(Canon, InsertPos); 5477 toe = Canon; 5478 } 5479 } else { 5480 QualType Canonical = getCanonicalType(tofExpr->getType()); 5481 toe = new (*this, TypeAlignment) TypeOfExprType(tofExpr, Canonical); 5482 } 5483 Types.push_back(toe); 5484 return QualType(toe, 0); 5485 } 5486 5487 /// getTypeOfType - Unlike many "get<Type>" functions, we don't unique 5488 /// TypeOfType nodes. The only motivation to unique these nodes would be 5489 /// memory savings. Since typeof(t) is fairly uncommon, space shouldn't be 5490 /// an issue. This doesn't affect the type checker, since it operates 5491 /// on canonical types (which are always unique). 5492 QualType ASTContext::getTypeOfType(QualType tofType) const { 5493 QualType Canonical = getCanonicalType(tofType); 5494 auto *tot = new (*this, TypeAlignment) TypeOfType(tofType, Canonical); 5495 Types.push_back(tot); 5496 return QualType(tot, 0); 5497 } 5498 5499 /// getReferenceQualifiedType - Given an expr, will return the type for 5500 /// that expression, as in [dcl.type.simple]p4 but without taking id-expressions 5501 /// and class member access into account. 5502 QualType ASTContext::getReferenceQualifiedType(const Expr *E) const { 5503 // C++11 [dcl.type.simple]p4: 5504 // [...] 5505 QualType T = E->getType(); 5506 switch (E->getValueKind()) { 5507 // - otherwise, if e is an xvalue, decltype(e) is T&&, where T is the 5508 // type of e; 5509 case VK_XValue: 5510 return getRValueReferenceType(T); 5511 // - otherwise, if e is an lvalue, decltype(e) is T&, where T is the 5512 // type of e; 5513 case VK_LValue: 5514 return getLValueReferenceType(T); 5515 // - otherwise, decltype(e) is the type of e. 5516 case VK_PRValue: 5517 return T; 5518 } 5519 llvm_unreachable("Unknown value kind"); 5520 } 5521 5522 /// Unlike many "get<Type>" functions, we don't unique DecltypeType 5523 /// nodes. This would never be helpful, since each such type has its own 5524 /// expression, and would not give a significant memory saving, since there 5525 /// is an Expr tree under each such type. 5526 QualType ASTContext::getDecltypeType(Expr *e, QualType UnderlyingType) const { 5527 DecltypeType *dt; 5528 5529 // C++11 [temp.type]p2: 5530 // If an expression e involves a template parameter, decltype(e) denotes a 5531 // unique dependent type. Two such decltype-specifiers refer to the same 5532 // type only if their expressions are equivalent (14.5.6.1). 5533 if (e->isInstantiationDependent()) { 5534 llvm::FoldingSetNodeID ID; 5535 DependentDecltypeType::Profile(ID, *this, e); 5536 5537 void *InsertPos = nullptr; 5538 DependentDecltypeType *Canon 5539 = DependentDecltypeTypes.FindNodeOrInsertPos(ID, InsertPos); 5540 if (!Canon) { 5541 // Build a new, canonical decltype(expr) type. 5542 Canon = new (*this, TypeAlignment) DependentDecltypeType(*this, e); 5543 DependentDecltypeTypes.InsertNode(Canon, InsertPos); 5544 } 5545 dt = new (*this, TypeAlignment) 5546 DecltypeType(e, UnderlyingType, QualType((DecltypeType *)Canon, 0)); 5547 } else { 5548 dt = new (*this, TypeAlignment) 5549 DecltypeType(e, UnderlyingType, getCanonicalType(UnderlyingType)); 5550 } 5551 Types.push_back(dt); 5552 return QualType(dt, 0); 5553 } 5554 5555 /// getUnaryTransformationType - We don't unique these, since the memory 5556 /// savings are minimal and these are rare. 5557 QualType ASTContext::getUnaryTransformType(QualType BaseType, 5558 QualType UnderlyingType, 5559 UnaryTransformType::UTTKind Kind) 5560 const { 5561 UnaryTransformType *ut = nullptr; 5562 5563 if (BaseType->isDependentType()) { 5564 // Look in the folding set for an existing type. 5565 llvm::FoldingSetNodeID ID; 5566 DependentUnaryTransformType::Profile(ID, getCanonicalType(BaseType), Kind); 5567 5568 void *InsertPos = nullptr; 5569 DependentUnaryTransformType *Canon 5570 = DependentUnaryTransformTypes.FindNodeOrInsertPos(ID, InsertPos); 5571 5572 if (!Canon) { 5573 // Build a new, canonical __underlying_type(type) type. 5574 Canon = new (*this, TypeAlignment) 5575 DependentUnaryTransformType(*this, getCanonicalType(BaseType), 5576 Kind); 5577 DependentUnaryTransformTypes.InsertNode(Canon, InsertPos); 5578 } 5579 ut = new (*this, TypeAlignment) UnaryTransformType (BaseType, 5580 QualType(), Kind, 5581 QualType(Canon, 0)); 5582 } else { 5583 QualType CanonType = getCanonicalType(UnderlyingType); 5584 ut = new (*this, TypeAlignment) UnaryTransformType (BaseType, 5585 UnderlyingType, Kind, 5586 CanonType); 5587 } 5588 Types.push_back(ut); 5589 return QualType(ut, 0); 5590 } 5591 5592 /// getAutoType - Return the uniqued reference to the 'auto' type which has been 5593 /// deduced to the given type, or to the canonical undeduced 'auto' type, or the 5594 /// canonical deduced-but-dependent 'auto' type. 5595 QualType 5596 ASTContext::getAutoType(QualType DeducedType, AutoTypeKeyword Keyword, 5597 bool IsDependent, bool IsPack, 5598 ConceptDecl *TypeConstraintConcept, 5599 ArrayRef<TemplateArgument> TypeConstraintArgs) const { 5600 assert((!IsPack || IsDependent) && "only use IsPack for a dependent pack"); 5601 if (DeducedType.isNull() && Keyword == AutoTypeKeyword::Auto && 5602 !TypeConstraintConcept && !IsDependent) 5603 return getAutoDeductType(); 5604 5605 // Look in the folding set for an existing type. 5606 void *InsertPos = nullptr; 5607 llvm::FoldingSetNodeID ID; 5608 AutoType::Profile(ID, *this, DeducedType, Keyword, IsDependent, 5609 TypeConstraintConcept, TypeConstraintArgs); 5610 if (AutoType *AT = AutoTypes.FindNodeOrInsertPos(ID, InsertPos)) 5611 return QualType(AT, 0); 5612 5613 void *Mem = Allocate(sizeof(AutoType) + 5614 sizeof(TemplateArgument) * TypeConstraintArgs.size(), 5615 TypeAlignment); 5616 auto *AT = new (Mem) AutoType( 5617 DeducedType, Keyword, 5618 (IsDependent ? TypeDependence::DependentInstantiation 5619 : TypeDependence::None) | 5620 (IsPack ? TypeDependence::UnexpandedPack : TypeDependence::None), 5621 TypeConstraintConcept, TypeConstraintArgs); 5622 Types.push_back(AT); 5623 if (InsertPos) 5624 AutoTypes.InsertNode(AT, InsertPos); 5625 return QualType(AT, 0); 5626 } 5627 5628 /// Return the uniqued reference to the deduced template specialization type 5629 /// which has been deduced to the given type, or to the canonical undeduced 5630 /// such type, or the canonical deduced-but-dependent such type. 5631 QualType ASTContext::getDeducedTemplateSpecializationType( 5632 TemplateName Template, QualType DeducedType, bool IsDependent) const { 5633 // Look in the folding set for an existing type. 5634 void *InsertPos = nullptr; 5635 llvm::FoldingSetNodeID ID; 5636 DeducedTemplateSpecializationType::Profile(ID, Template, DeducedType, 5637 IsDependent); 5638 if (DeducedTemplateSpecializationType *DTST = 5639 DeducedTemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos)) 5640 return QualType(DTST, 0); 5641 5642 auto *DTST = new (*this, TypeAlignment) 5643 DeducedTemplateSpecializationType(Template, DeducedType, IsDependent); 5644 Types.push_back(DTST); 5645 if (InsertPos) 5646 DeducedTemplateSpecializationTypes.InsertNode(DTST, InsertPos); 5647 return QualType(DTST, 0); 5648 } 5649 5650 /// getAtomicType - Return the uniqued reference to the atomic type for 5651 /// the given value type. 5652 QualType ASTContext::getAtomicType(QualType T) const { 5653 // Unique pointers, to guarantee there is only one pointer of a particular 5654 // structure. 5655 llvm::FoldingSetNodeID ID; 5656 AtomicType::Profile(ID, T); 5657 5658 void *InsertPos = nullptr; 5659 if (AtomicType *AT = AtomicTypes.FindNodeOrInsertPos(ID, InsertPos)) 5660 return QualType(AT, 0); 5661 5662 // If the atomic value type isn't canonical, this won't be a canonical type 5663 // either, so fill in the canonical type field. 5664 QualType Canonical; 5665 if (!T.isCanonical()) { 5666 Canonical = getAtomicType(getCanonicalType(T)); 5667 5668 // Get the new insert position for the node we care about. 5669 AtomicType *NewIP = AtomicTypes.FindNodeOrInsertPos(ID, InsertPos); 5670 assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP; 5671 } 5672 auto *New = new (*this, TypeAlignment) AtomicType(T, Canonical); 5673 Types.push_back(New); 5674 AtomicTypes.InsertNode(New, InsertPos); 5675 return QualType(New, 0); 5676 } 5677 5678 /// getAutoDeductType - Get type pattern for deducing against 'auto'. 5679 QualType ASTContext::getAutoDeductType() const { 5680 if (AutoDeductTy.isNull()) 5681 AutoDeductTy = QualType(new (*this, TypeAlignment) 5682 AutoType(QualType(), AutoTypeKeyword::Auto, 5683 TypeDependence::None, 5684 /*concept*/ nullptr, /*args*/ {}), 5685 0); 5686 return AutoDeductTy; 5687 } 5688 5689 /// getAutoRRefDeductType - Get type pattern for deducing against 'auto &&'. 5690 QualType ASTContext::getAutoRRefDeductType() const { 5691 if (AutoRRefDeductTy.isNull()) 5692 AutoRRefDeductTy = getRValueReferenceType(getAutoDeductType()); 5693 assert(!AutoRRefDeductTy.isNull() && "can't build 'auto &&' pattern"); 5694 return AutoRRefDeductTy; 5695 } 5696 5697 /// getTagDeclType - Return the unique reference to the type for the 5698 /// specified TagDecl (struct/union/class/enum) decl. 5699 QualType ASTContext::getTagDeclType(const TagDecl *Decl) const { 5700 assert(Decl); 5701 // FIXME: What is the design on getTagDeclType when it requires casting 5702 // away const? mutable? 5703 return getTypeDeclType(const_cast<TagDecl*>(Decl)); 5704 } 5705 5706 /// getSizeType - Return the unique type for "size_t" (C99 7.17), the result 5707 /// of the sizeof operator (C99 6.5.3.4p4). The value is target dependent and 5708 /// needs to agree with the definition in <stddef.h>. 5709 CanQualType ASTContext::getSizeType() const { 5710 return getFromTargetType(Target->getSizeType()); 5711 } 5712 5713 /// Return the unique signed counterpart of the integer type 5714 /// corresponding to size_t. 5715 CanQualType ASTContext::getSignedSizeType() const { 5716 return getFromTargetType(Target->getSignedSizeType()); 5717 } 5718 5719 /// getIntMaxType - Return the unique type for "intmax_t" (C99 7.18.1.5). 5720 CanQualType ASTContext::getIntMaxType() const { 5721 return getFromTargetType(Target->getIntMaxType()); 5722 } 5723 5724 /// getUIntMaxType - Return the unique type for "uintmax_t" (C99 7.18.1.5). 5725 CanQualType ASTContext::getUIntMaxType() const { 5726 return getFromTargetType(Target->getUIntMaxType()); 5727 } 5728 5729 /// getSignedWCharType - Return the type of "signed wchar_t". 5730 /// Used when in C++, as a GCC extension. 5731 QualType ASTContext::getSignedWCharType() const { 5732 // FIXME: derive from "Target" ? 5733 return WCharTy; 5734 } 5735 5736 /// getUnsignedWCharType - Return the type of "unsigned wchar_t". 5737 /// Used when in C++, as a GCC extension. 5738 QualType ASTContext::getUnsignedWCharType() const { 5739 // FIXME: derive from "Target" ? 5740 return UnsignedIntTy; 5741 } 5742 5743 QualType ASTContext::getIntPtrType() const { 5744 return getFromTargetType(Target->getIntPtrType()); 5745 } 5746 5747 QualType ASTContext::getUIntPtrType() const { 5748 return getCorrespondingUnsignedType(getIntPtrType()); 5749 } 5750 5751 /// getPointerDiffType - Return the unique type for "ptrdiff_t" (C99 7.17) 5752 /// defined in <stddef.h>. Pointer - pointer requires this (C99 6.5.6p9). 5753 QualType ASTContext::getPointerDiffType() const { 5754 return getFromTargetType(Target->getPtrDiffType(0)); 5755 } 5756 5757 /// Return the unique unsigned counterpart of "ptrdiff_t" 5758 /// integer type. The standard (C11 7.21.6.1p7) refers to this type 5759 /// in the definition of %tu format specifier. 5760 QualType ASTContext::getUnsignedPointerDiffType() const { 5761 return getFromTargetType(Target->getUnsignedPtrDiffType(0)); 5762 } 5763 5764 /// Return the unique type for "pid_t" defined in 5765 /// <sys/types.h>. We need this to compute the correct type for vfork(). 5766 QualType ASTContext::getProcessIDType() const { 5767 return getFromTargetType(Target->getProcessIDType()); 5768 } 5769 5770 //===----------------------------------------------------------------------===// 5771 // Type Operators 5772 //===----------------------------------------------------------------------===// 5773 5774 CanQualType ASTContext::getCanonicalParamType(QualType T) const { 5775 // Push qualifiers into arrays, and then discard any remaining 5776 // qualifiers. 5777 T = getCanonicalType(T); 5778 T = getVariableArrayDecayedType(T); 5779 const Type *Ty = T.getTypePtr(); 5780 QualType Result; 5781 if (isa<ArrayType>(Ty)) { 5782 Result = getArrayDecayedType(QualType(Ty,0)); 5783 } else if (isa<FunctionType>(Ty)) { 5784 Result = getPointerType(QualType(Ty, 0)); 5785 } else { 5786 Result = QualType(Ty, 0); 5787 } 5788 5789 return CanQualType::CreateUnsafe(Result); 5790 } 5791 5792 QualType ASTContext::getUnqualifiedArrayType(QualType type, 5793 Qualifiers &quals) { 5794 SplitQualType splitType = type.getSplitUnqualifiedType(); 5795 5796 // FIXME: getSplitUnqualifiedType() actually walks all the way to 5797 // the unqualified desugared type and then drops it on the floor. 5798 // We then have to strip that sugar back off with 5799 // getUnqualifiedDesugaredType(), which is silly. 5800 const auto *AT = 5801 dyn_cast<ArrayType>(splitType.Ty->getUnqualifiedDesugaredType()); 5802 5803 // If we don't have an array, just use the results in splitType. 5804 if (!AT) { 5805 quals = splitType.Quals; 5806 return QualType(splitType.Ty, 0); 5807 } 5808 5809 // Otherwise, recurse on the array's element type. 5810 QualType elementType = AT->getElementType(); 5811 QualType unqualElementType = getUnqualifiedArrayType(elementType, quals); 5812 5813 // If that didn't change the element type, AT has no qualifiers, so we 5814 // can just use the results in splitType. 5815 if (elementType == unqualElementType) { 5816 assert(quals.empty()); // from the recursive call 5817 quals = splitType.Quals; 5818 return QualType(splitType.Ty, 0); 5819 } 5820 5821 // Otherwise, add in the qualifiers from the outermost type, then 5822 // build the type back up. 5823 quals.addConsistentQualifiers(splitType.Quals); 5824 5825 if (const auto *CAT = dyn_cast<ConstantArrayType>(AT)) { 5826 return getConstantArrayType(unqualElementType, CAT->getSize(), 5827 CAT->getSizeExpr(), CAT->getSizeModifier(), 0); 5828 } 5829 5830 if (const auto *IAT = dyn_cast<IncompleteArrayType>(AT)) { 5831 return getIncompleteArrayType(unqualElementType, IAT->getSizeModifier(), 0); 5832 } 5833 5834 if (const auto *VAT = dyn_cast<VariableArrayType>(AT)) { 5835 return getVariableArrayType(unqualElementType, 5836 VAT->getSizeExpr(), 5837 VAT->getSizeModifier(), 5838 VAT->getIndexTypeCVRQualifiers(), 5839 VAT->getBracketsRange()); 5840 } 5841 5842 const auto *DSAT = cast<DependentSizedArrayType>(AT); 5843 return getDependentSizedArrayType(unqualElementType, DSAT->getSizeExpr(), 5844 DSAT->getSizeModifier(), 0, 5845 SourceRange()); 5846 } 5847 5848 /// Attempt to unwrap two types that may both be array types with the same bound 5849 /// (or both be array types of unknown bound) for the purpose of comparing the 5850 /// cv-decomposition of two types per C++ [conv.qual]. 5851 /// 5852 /// \param AllowPiMismatch Allow the Pi1 and Pi2 to differ as described in 5853 /// C++20 [conv.qual], if permitted by the current language mode. 5854 void ASTContext::UnwrapSimilarArrayTypes(QualType &T1, QualType &T2, 5855 bool AllowPiMismatch) { 5856 while (true) { 5857 auto *AT1 = getAsArrayType(T1); 5858 if (!AT1) 5859 return; 5860 5861 auto *AT2 = getAsArrayType(T2); 5862 if (!AT2) 5863 return; 5864 5865 // If we don't have two array types with the same constant bound nor two 5866 // incomplete array types, we've unwrapped everything we can. 5867 // C++20 also permits one type to be a constant array type and the other 5868 // to be an incomplete array type. 5869 // FIXME: Consider also unwrapping array of unknown bound and VLA. 5870 if (auto *CAT1 = dyn_cast<ConstantArrayType>(AT1)) { 5871 auto *CAT2 = dyn_cast<ConstantArrayType>(AT2); 5872 if (!((CAT2 && CAT1->getSize() == CAT2->getSize()) || 5873 (AllowPiMismatch && getLangOpts().CPlusPlus20 && 5874 isa<IncompleteArrayType>(AT2)))) 5875 return; 5876 } else if (isa<IncompleteArrayType>(AT1)) { 5877 if (!(isa<IncompleteArrayType>(AT2) || 5878 (AllowPiMismatch && getLangOpts().CPlusPlus20 && 5879 isa<ConstantArrayType>(AT2)))) 5880 return; 5881 } else { 5882 return; 5883 } 5884 5885 T1 = AT1->getElementType(); 5886 T2 = AT2->getElementType(); 5887 } 5888 } 5889 5890 /// Attempt to unwrap two types that may be similar (C++ [conv.qual]). 5891 /// 5892 /// If T1 and T2 are both pointer types of the same kind, or both array types 5893 /// with the same bound, unwraps layers from T1 and T2 until a pointer type is 5894 /// unwrapped. Top-level qualifiers on T1 and T2 are ignored. 5895 /// 5896 /// This function will typically be called in a loop that successively 5897 /// "unwraps" pointer and pointer-to-member types to compare them at each 5898 /// level. 5899 /// 5900 /// \param AllowPiMismatch Allow the Pi1 and Pi2 to differ as described in 5901 /// C++20 [conv.qual], if permitted by the current language mode. 5902 /// 5903 /// \return \c true if a pointer type was unwrapped, \c false if we reached a 5904 /// pair of types that can't be unwrapped further. 5905 bool ASTContext::UnwrapSimilarTypes(QualType &T1, QualType &T2, 5906 bool AllowPiMismatch) { 5907 UnwrapSimilarArrayTypes(T1, T2, AllowPiMismatch); 5908 5909 const auto *T1PtrType = T1->getAs<PointerType>(); 5910 const auto *T2PtrType = T2->getAs<PointerType>(); 5911 if (T1PtrType && T2PtrType) { 5912 T1 = T1PtrType->getPointeeType(); 5913 T2 = T2PtrType->getPointeeType(); 5914 return true; 5915 } 5916 5917 const auto *T1MPType = T1->getAs<MemberPointerType>(); 5918 const auto *T2MPType = T2->getAs<MemberPointerType>(); 5919 if (T1MPType && T2MPType && 5920 hasSameUnqualifiedType(QualType(T1MPType->getClass(), 0), 5921 QualType(T2MPType->getClass(), 0))) { 5922 T1 = T1MPType->getPointeeType(); 5923 T2 = T2MPType->getPointeeType(); 5924 return true; 5925 } 5926 5927 if (getLangOpts().ObjC) { 5928 const auto *T1OPType = T1->getAs<ObjCObjectPointerType>(); 5929 const auto *T2OPType = T2->getAs<ObjCObjectPointerType>(); 5930 if (T1OPType && T2OPType) { 5931 T1 = T1OPType->getPointeeType(); 5932 T2 = T2OPType->getPointeeType(); 5933 return true; 5934 } 5935 } 5936 5937 // FIXME: Block pointers, too? 5938 5939 return false; 5940 } 5941 5942 bool ASTContext::hasSimilarType(QualType T1, QualType T2) { 5943 while (true) { 5944 Qualifiers Quals; 5945 T1 = getUnqualifiedArrayType(T1, Quals); 5946 T2 = getUnqualifiedArrayType(T2, Quals); 5947 if (hasSameType(T1, T2)) 5948 return true; 5949 if (!UnwrapSimilarTypes(T1, T2)) 5950 return false; 5951 } 5952 } 5953 5954 bool ASTContext::hasCvrSimilarType(QualType T1, QualType T2) { 5955 while (true) { 5956 Qualifiers Quals1, Quals2; 5957 T1 = getUnqualifiedArrayType(T1, Quals1); 5958 T2 = getUnqualifiedArrayType(T2, Quals2); 5959 5960 Quals1.removeCVRQualifiers(); 5961 Quals2.removeCVRQualifiers(); 5962 if (Quals1 != Quals2) 5963 return false; 5964 5965 if (hasSameType(T1, T2)) 5966 return true; 5967 5968 if (!UnwrapSimilarTypes(T1, T2, /*AllowPiMismatch*/ false)) 5969 return false; 5970 } 5971 } 5972 5973 DeclarationNameInfo 5974 ASTContext::getNameForTemplate(TemplateName Name, 5975 SourceLocation NameLoc) const { 5976 switch (Name.getKind()) { 5977 case TemplateName::QualifiedTemplate: 5978 case TemplateName::Template: 5979 // DNInfo work in progress: CHECKME: what about DNLoc? 5980 return DeclarationNameInfo(Name.getAsTemplateDecl()->getDeclName(), 5981 NameLoc); 5982 5983 case TemplateName::OverloadedTemplate: { 5984 OverloadedTemplateStorage *Storage = Name.getAsOverloadedTemplate(); 5985 // DNInfo work in progress: CHECKME: what about DNLoc? 5986 return DeclarationNameInfo((*Storage->begin())->getDeclName(), NameLoc); 5987 } 5988 5989 case TemplateName::AssumedTemplate: { 5990 AssumedTemplateStorage *Storage = Name.getAsAssumedTemplateName(); 5991 return DeclarationNameInfo(Storage->getDeclName(), NameLoc); 5992 } 5993 5994 case TemplateName::DependentTemplate: { 5995 DependentTemplateName *DTN = Name.getAsDependentTemplateName(); 5996 DeclarationName DName; 5997 if (DTN->isIdentifier()) { 5998 DName = DeclarationNames.getIdentifier(DTN->getIdentifier()); 5999 return DeclarationNameInfo(DName, NameLoc); 6000 } else { 6001 DName = DeclarationNames.getCXXOperatorName(DTN->getOperator()); 6002 // DNInfo work in progress: FIXME: source locations? 6003 DeclarationNameLoc DNLoc = 6004 DeclarationNameLoc::makeCXXOperatorNameLoc(SourceRange()); 6005 return DeclarationNameInfo(DName, NameLoc, DNLoc); 6006 } 6007 } 6008 6009 case TemplateName::SubstTemplateTemplateParm: { 6010 SubstTemplateTemplateParmStorage *subst 6011 = Name.getAsSubstTemplateTemplateParm(); 6012 return DeclarationNameInfo(subst->getParameter()->getDeclName(), 6013 NameLoc); 6014 } 6015 6016 case TemplateName::SubstTemplateTemplateParmPack: { 6017 SubstTemplateTemplateParmPackStorage *subst 6018 = Name.getAsSubstTemplateTemplateParmPack(); 6019 return DeclarationNameInfo(subst->getParameterPack()->getDeclName(), 6020 NameLoc); 6021 } 6022 } 6023 6024 llvm_unreachable("bad template name kind!"); 6025 } 6026 6027 TemplateName ASTContext::getCanonicalTemplateName(TemplateName Name) const { 6028 switch (Name.getKind()) { 6029 case TemplateName::QualifiedTemplate: 6030 case TemplateName::Template: { 6031 TemplateDecl *Template = Name.getAsTemplateDecl(); 6032 if (auto *TTP = dyn_cast<TemplateTemplateParmDecl>(Template)) 6033 Template = getCanonicalTemplateTemplateParmDecl(TTP); 6034 6035 // The canonical template name is the canonical template declaration. 6036 return TemplateName(cast<TemplateDecl>(Template->getCanonicalDecl())); 6037 } 6038 6039 case TemplateName::OverloadedTemplate: 6040 case TemplateName::AssumedTemplate: 6041 llvm_unreachable("cannot canonicalize unresolved template"); 6042 6043 case TemplateName::DependentTemplate: { 6044 DependentTemplateName *DTN = Name.getAsDependentTemplateName(); 6045 assert(DTN && "Non-dependent template names must refer to template decls."); 6046 return DTN->CanonicalTemplateName; 6047 } 6048 6049 case TemplateName::SubstTemplateTemplateParm: { 6050 SubstTemplateTemplateParmStorage *subst 6051 = Name.getAsSubstTemplateTemplateParm(); 6052 return getCanonicalTemplateName(subst->getReplacement()); 6053 } 6054 6055 case TemplateName::SubstTemplateTemplateParmPack: { 6056 SubstTemplateTemplateParmPackStorage *subst 6057 = Name.getAsSubstTemplateTemplateParmPack(); 6058 TemplateTemplateParmDecl *canonParameter 6059 = getCanonicalTemplateTemplateParmDecl(subst->getParameterPack()); 6060 TemplateArgument canonArgPack 6061 = getCanonicalTemplateArgument(subst->getArgumentPack()); 6062 return getSubstTemplateTemplateParmPack(canonParameter, canonArgPack); 6063 } 6064 } 6065 6066 llvm_unreachable("bad template name!"); 6067 } 6068 6069 bool ASTContext::hasSameTemplateName(TemplateName X, TemplateName Y) { 6070 X = getCanonicalTemplateName(X); 6071 Y = getCanonicalTemplateName(Y); 6072 return X.getAsVoidPointer() == Y.getAsVoidPointer(); 6073 } 6074 6075 TemplateArgument 6076 ASTContext::getCanonicalTemplateArgument(const TemplateArgument &Arg) const { 6077 switch (Arg.getKind()) { 6078 case TemplateArgument::Null: 6079 return Arg; 6080 6081 case TemplateArgument::Expression: 6082 return Arg; 6083 6084 case TemplateArgument::Declaration: { 6085 auto *D = cast<ValueDecl>(Arg.getAsDecl()->getCanonicalDecl()); 6086 return TemplateArgument(D, Arg.getParamTypeForDecl()); 6087 } 6088 6089 case TemplateArgument::NullPtr: 6090 return TemplateArgument(getCanonicalType(Arg.getNullPtrType()), 6091 /*isNullPtr*/true); 6092 6093 case TemplateArgument::Template: 6094 return TemplateArgument(getCanonicalTemplateName(Arg.getAsTemplate())); 6095 6096 case TemplateArgument::TemplateExpansion: 6097 return TemplateArgument(getCanonicalTemplateName( 6098 Arg.getAsTemplateOrTemplatePattern()), 6099 Arg.getNumTemplateExpansions()); 6100 6101 case TemplateArgument::Integral: 6102 return TemplateArgument(Arg, getCanonicalType(Arg.getIntegralType())); 6103 6104 case TemplateArgument::Type: 6105 return TemplateArgument(getCanonicalType(Arg.getAsType())); 6106 6107 case TemplateArgument::Pack: { 6108 if (Arg.pack_size() == 0) 6109 return Arg; 6110 6111 auto *CanonArgs = new (*this) TemplateArgument[Arg.pack_size()]; 6112 unsigned Idx = 0; 6113 for (TemplateArgument::pack_iterator A = Arg.pack_begin(), 6114 AEnd = Arg.pack_end(); 6115 A != AEnd; (void)++A, ++Idx) 6116 CanonArgs[Idx] = getCanonicalTemplateArgument(*A); 6117 6118 return TemplateArgument(llvm::makeArrayRef(CanonArgs, Arg.pack_size())); 6119 } 6120 } 6121 6122 // Silence GCC warning 6123 llvm_unreachable("Unhandled template argument kind"); 6124 } 6125 6126 NestedNameSpecifier * 6127 ASTContext::getCanonicalNestedNameSpecifier(NestedNameSpecifier *NNS) const { 6128 if (!NNS) 6129 return nullptr; 6130 6131 switch (NNS->getKind()) { 6132 case NestedNameSpecifier::Identifier: 6133 // Canonicalize the prefix but keep the identifier the same. 6134 return NestedNameSpecifier::Create(*this, 6135 getCanonicalNestedNameSpecifier(NNS->getPrefix()), 6136 NNS->getAsIdentifier()); 6137 6138 case NestedNameSpecifier::Namespace: 6139 // A namespace is canonical; build a nested-name-specifier with 6140 // this namespace and no prefix. 6141 return NestedNameSpecifier::Create(*this, nullptr, 6142 NNS->getAsNamespace()->getOriginalNamespace()); 6143 6144 case NestedNameSpecifier::NamespaceAlias: 6145 // A namespace is canonical; build a nested-name-specifier with 6146 // this namespace and no prefix. 6147 return NestedNameSpecifier::Create(*this, nullptr, 6148 NNS->getAsNamespaceAlias()->getNamespace() 6149 ->getOriginalNamespace()); 6150 6151 // The difference between TypeSpec and TypeSpecWithTemplate is that the 6152 // latter will have the 'template' keyword when printed. 6153 case NestedNameSpecifier::TypeSpec: 6154 case NestedNameSpecifier::TypeSpecWithTemplate: { 6155 const Type *T = getCanonicalType(NNS->getAsType()); 6156 6157 // If we have some kind of dependent-named type (e.g., "typename T::type"), 6158 // break it apart into its prefix and identifier, then reconsititute those 6159 // as the canonical nested-name-specifier. This is required to canonicalize 6160 // a dependent nested-name-specifier involving typedefs of dependent-name 6161 // types, e.g., 6162 // typedef typename T::type T1; 6163 // typedef typename T1::type T2; 6164 if (const auto *DNT = T->getAs<DependentNameType>()) 6165 return NestedNameSpecifier::Create( 6166 *this, DNT->getQualifier(), 6167 const_cast<IdentifierInfo *>(DNT->getIdentifier())); 6168 if (const auto *DTST = T->getAs<DependentTemplateSpecializationType>()) 6169 return NestedNameSpecifier::Create(*this, DTST->getQualifier(), true, 6170 const_cast<Type *>(T)); 6171 6172 // TODO: Set 'Template' parameter to true for other template types. 6173 return NestedNameSpecifier::Create(*this, nullptr, false, 6174 const_cast<Type *>(T)); 6175 } 6176 6177 case NestedNameSpecifier::Global: 6178 case NestedNameSpecifier::Super: 6179 // The global specifier and __super specifer are canonical and unique. 6180 return NNS; 6181 } 6182 6183 llvm_unreachable("Invalid NestedNameSpecifier::Kind!"); 6184 } 6185 6186 const ArrayType *ASTContext::getAsArrayType(QualType T) const { 6187 // Handle the non-qualified case efficiently. 6188 if (!T.hasLocalQualifiers()) { 6189 // Handle the common positive case fast. 6190 if (const auto *AT = dyn_cast<ArrayType>(T)) 6191 return AT; 6192 } 6193 6194 // Handle the common negative case fast. 6195 if (!isa<ArrayType>(T.getCanonicalType())) 6196 return nullptr; 6197 6198 // Apply any qualifiers from the array type to the element type. This 6199 // implements C99 6.7.3p8: "If the specification of an array type includes 6200 // any type qualifiers, the element type is so qualified, not the array type." 6201 6202 // If we get here, we either have type qualifiers on the type, or we have 6203 // sugar such as a typedef in the way. If we have type qualifiers on the type 6204 // we must propagate them down into the element type. 6205 6206 SplitQualType split = T.getSplitDesugaredType(); 6207 Qualifiers qs = split.Quals; 6208 6209 // If we have a simple case, just return now. 6210 const auto *ATy = dyn_cast<ArrayType>(split.Ty); 6211 if (!ATy || qs.empty()) 6212 return ATy; 6213 6214 // Otherwise, we have an array and we have qualifiers on it. Push the 6215 // qualifiers into the array element type and return a new array type. 6216 QualType NewEltTy = getQualifiedType(ATy->getElementType(), qs); 6217 6218 if (const auto *CAT = dyn_cast<ConstantArrayType>(ATy)) 6219 return cast<ArrayType>(getConstantArrayType(NewEltTy, CAT->getSize(), 6220 CAT->getSizeExpr(), 6221 CAT->getSizeModifier(), 6222 CAT->getIndexTypeCVRQualifiers())); 6223 if (const auto *IAT = dyn_cast<IncompleteArrayType>(ATy)) 6224 return cast<ArrayType>(getIncompleteArrayType(NewEltTy, 6225 IAT->getSizeModifier(), 6226 IAT->getIndexTypeCVRQualifiers())); 6227 6228 if (const auto *DSAT = dyn_cast<DependentSizedArrayType>(ATy)) 6229 return cast<ArrayType>( 6230 getDependentSizedArrayType(NewEltTy, 6231 DSAT->getSizeExpr(), 6232 DSAT->getSizeModifier(), 6233 DSAT->getIndexTypeCVRQualifiers(), 6234 DSAT->getBracketsRange())); 6235 6236 const auto *VAT = cast<VariableArrayType>(ATy); 6237 return cast<ArrayType>(getVariableArrayType(NewEltTy, 6238 VAT->getSizeExpr(), 6239 VAT->getSizeModifier(), 6240 VAT->getIndexTypeCVRQualifiers(), 6241 VAT->getBracketsRange())); 6242 } 6243 6244 QualType ASTContext::getAdjustedParameterType(QualType T) const { 6245 if (T->isArrayType() || T->isFunctionType()) 6246 return getDecayedType(T); 6247 return T; 6248 } 6249 6250 QualType ASTContext::getSignatureParameterType(QualType T) const { 6251 T = getVariableArrayDecayedType(T); 6252 T = getAdjustedParameterType(T); 6253 return T.getUnqualifiedType(); 6254 } 6255 6256 QualType ASTContext::getExceptionObjectType(QualType T) const { 6257 // C++ [except.throw]p3: 6258 // A throw-expression initializes a temporary object, called the exception 6259 // object, the type of which is determined by removing any top-level 6260 // cv-qualifiers from the static type of the operand of throw and adjusting 6261 // the type from "array of T" or "function returning T" to "pointer to T" 6262 // or "pointer to function returning T", [...] 6263 T = getVariableArrayDecayedType(T); 6264 if (T->isArrayType() || T->isFunctionType()) 6265 T = getDecayedType(T); 6266 return T.getUnqualifiedType(); 6267 } 6268 6269 /// getArrayDecayedType - Return the properly qualified result of decaying the 6270 /// specified array type to a pointer. This operation is non-trivial when 6271 /// handling typedefs etc. The canonical type of "T" must be an array type, 6272 /// this returns a pointer to a properly qualified element of the array. 6273 /// 6274 /// See C99 6.7.5.3p7 and C99 6.3.2.1p3. 6275 QualType ASTContext::getArrayDecayedType(QualType Ty) const { 6276 // Get the element type with 'getAsArrayType' so that we don't lose any 6277 // typedefs in the element type of the array. This also handles propagation 6278 // of type qualifiers from the array type into the element type if present 6279 // (C99 6.7.3p8). 6280 const ArrayType *PrettyArrayType = getAsArrayType(Ty); 6281 assert(PrettyArrayType && "Not an array type!"); 6282 6283 QualType PtrTy = getPointerType(PrettyArrayType->getElementType()); 6284 6285 // int x[restrict 4] -> int *restrict 6286 QualType Result = getQualifiedType(PtrTy, 6287 PrettyArrayType->getIndexTypeQualifiers()); 6288 6289 // int x[_Nullable] -> int * _Nullable 6290 if (auto Nullability = Ty->getNullability(*this)) { 6291 Result = const_cast<ASTContext *>(this)->getAttributedType( 6292 AttributedType::getNullabilityAttrKind(*Nullability), Result, Result); 6293 } 6294 return Result; 6295 } 6296 6297 QualType ASTContext::getBaseElementType(const ArrayType *array) const { 6298 return getBaseElementType(array->getElementType()); 6299 } 6300 6301 QualType ASTContext::getBaseElementType(QualType type) const { 6302 Qualifiers qs; 6303 while (true) { 6304 SplitQualType split = type.getSplitDesugaredType(); 6305 const ArrayType *array = split.Ty->getAsArrayTypeUnsafe(); 6306 if (!array) break; 6307 6308 type = array->getElementType(); 6309 qs.addConsistentQualifiers(split.Quals); 6310 } 6311 6312 return getQualifiedType(type, qs); 6313 } 6314 6315 /// getConstantArrayElementCount - Returns number of constant array elements. 6316 uint64_t 6317 ASTContext::getConstantArrayElementCount(const ConstantArrayType *CA) const { 6318 uint64_t ElementCount = 1; 6319 do { 6320 ElementCount *= CA->getSize().getZExtValue(); 6321 CA = dyn_cast_or_null<ConstantArrayType>( 6322 CA->getElementType()->getAsArrayTypeUnsafe()); 6323 } while (CA); 6324 return ElementCount; 6325 } 6326 6327 /// getFloatingRank - Return a relative rank for floating point types. 6328 /// This routine will assert if passed a built-in type that isn't a float. 6329 static FloatingRank getFloatingRank(QualType T) { 6330 if (const auto *CT = T->getAs<ComplexType>()) 6331 return getFloatingRank(CT->getElementType()); 6332 6333 switch (T->castAs<BuiltinType>()->getKind()) { 6334 default: llvm_unreachable("getFloatingRank(): not a floating type"); 6335 case BuiltinType::Float16: return Float16Rank; 6336 case BuiltinType::Half: return HalfRank; 6337 case BuiltinType::Float: return FloatRank; 6338 case BuiltinType::Double: return DoubleRank; 6339 case BuiltinType::LongDouble: return LongDoubleRank; 6340 case BuiltinType::Float128: return Float128Rank; 6341 case BuiltinType::BFloat16: return BFloat16Rank; 6342 case BuiltinType::Ibm128: return Ibm128Rank; 6343 } 6344 } 6345 6346 /// getFloatingTypeOfSizeWithinDomain - Returns a real floating 6347 /// point or a complex type (based on typeDomain/typeSize). 6348 /// 'typeDomain' is a real floating point or complex type. 6349 /// 'typeSize' is a real floating point or complex type. 6350 QualType ASTContext::getFloatingTypeOfSizeWithinDomain(QualType Size, 6351 QualType Domain) const { 6352 FloatingRank EltRank = getFloatingRank(Size); 6353 if (Domain->isComplexType()) { 6354 switch (EltRank) { 6355 case BFloat16Rank: llvm_unreachable("Complex bfloat16 is not supported"); 6356 case Float16Rank: 6357 case HalfRank: llvm_unreachable("Complex half is not supported"); 6358 case Ibm128Rank: return getComplexType(Ibm128Ty); 6359 case FloatRank: return getComplexType(FloatTy); 6360 case DoubleRank: return getComplexType(DoubleTy); 6361 case LongDoubleRank: return getComplexType(LongDoubleTy); 6362 case Float128Rank: return getComplexType(Float128Ty); 6363 } 6364 } 6365 6366 assert(Domain->isRealFloatingType() && "Unknown domain!"); 6367 switch (EltRank) { 6368 case Float16Rank: return HalfTy; 6369 case BFloat16Rank: return BFloat16Ty; 6370 case HalfRank: return HalfTy; 6371 case FloatRank: return FloatTy; 6372 case DoubleRank: return DoubleTy; 6373 case LongDoubleRank: return LongDoubleTy; 6374 case Float128Rank: return Float128Ty; 6375 case Ibm128Rank: 6376 return Ibm128Ty; 6377 } 6378 llvm_unreachable("getFloatingRank(): illegal value for rank"); 6379 } 6380 6381 /// getFloatingTypeOrder - Compare the rank of the two specified floating 6382 /// point types, ignoring the domain of the type (i.e. 'double' == 6383 /// '_Complex double'). If LHS > RHS, return 1. If LHS == RHS, return 0. If 6384 /// LHS < RHS, return -1. 6385 int ASTContext::getFloatingTypeOrder(QualType LHS, QualType RHS) const { 6386 FloatingRank LHSR = getFloatingRank(LHS); 6387 FloatingRank RHSR = getFloatingRank(RHS); 6388 6389 if (LHSR == RHSR) 6390 return 0; 6391 if (LHSR > RHSR) 6392 return 1; 6393 return -1; 6394 } 6395 6396 int ASTContext::getFloatingTypeSemanticOrder(QualType LHS, QualType RHS) const { 6397 if (&getFloatTypeSemantics(LHS) == &getFloatTypeSemantics(RHS)) 6398 return 0; 6399 return getFloatingTypeOrder(LHS, RHS); 6400 } 6401 6402 /// getIntegerRank - Return an integer conversion rank (C99 6.3.1.1p1). This 6403 /// routine will assert if passed a built-in type that isn't an integer or enum, 6404 /// or if it is not canonicalized. 6405 unsigned ASTContext::getIntegerRank(const Type *T) const { 6406 assert(T->isCanonicalUnqualified() && "T should be canonicalized"); 6407 6408 // Results in this 'losing' to any type of the same size, but winning if 6409 // larger. 6410 if (const auto *EIT = dyn_cast<ExtIntType>(T)) 6411 return 0 + (EIT->getNumBits() << 3); 6412 6413 switch (cast<BuiltinType>(T)->getKind()) { 6414 default: llvm_unreachable("getIntegerRank(): not a built-in integer"); 6415 case BuiltinType::Bool: 6416 return 1 + (getIntWidth(BoolTy) << 3); 6417 case BuiltinType::Char_S: 6418 case BuiltinType::Char_U: 6419 case BuiltinType::SChar: 6420 case BuiltinType::UChar: 6421 return 2 + (getIntWidth(CharTy) << 3); 6422 case BuiltinType::Short: 6423 case BuiltinType::UShort: 6424 return 3 + (getIntWidth(ShortTy) << 3); 6425 case BuiltinType::Int: 6426 case BuiltinType::UInt: 6427 return 4 + (getIntWidth(IntTy) << 3); 6428 case BuiltinType::Long: 6429 case BuiltinType::ULong: 6430 return 5 + (getIntWidth(LongTy) << 3); 6431 case BuiltinType::LongLong: 6432 case BuiltinType::ULongLong: 6433 return 6 + (getIntWidth(LongLongTy) << 3); 6434 case BuiltinType::Int128: 6435 case BuiltinType::UInt128: 6436 return 7 + (getIntWidth(Int128Ty) << 3); 6437 } 6438 } 6439 6440 /// Whether this is a promotable bitfield reference according 6441 /// to C99 6.3.1.1p2, bullet 2 (and GCC extensions). 6442 /// 6443 /// \returns the type this bit-field will promote to, or NULL if no 6444 /// promotion occurs. 6445 QualType ASTContext::isPromotableBitField(Expr *E) const { 6446 if (E->isTypeDependent() || E->isValueDependent()) 6447 return {}; 6448 6449 // C++ [conv.prom]p5: 6450 // If the bit-field has an enumerated type, it is treated as any other 6451 // value of that type for promotion purposes. 6452 if (getLangOpts().CPlusPlus && E->getType()->isEnumeralType()) 6453 return {}; 6454 6455 // FIXME: We should not do this unless E->refersToBitField() is true. This 6456 // matters in C where getSourceBitField() will find bit-fields for various 6457 // cases where the source expression is not a bit-field designator. 6458 6459 FieldDecl *Field = E->getSourceBitField(); // FIXME: conditional bit-fields? 6460 if (!Field) 6461 return {}; 6462 6463 QualType FT = Field->getType(); 6464 6465 uint64_t BitWidth = Field->getBitWidthValue(*this); 6466 uint64_t IntSize = getTypeSize(IntTy); 6467 // C++ [conv.prom]p5: 6468 // A prvalue for an integral bit-field can be converted to a prvalue of type 6469 // int if int can represent all the values of the bit-field; otherwise, it 6470 // can be converted to unsigned int if unsigned int can represent all the 6471 // values of the bit-field. If the bit-field is larger yet, no integral 6472 // promotion applies to it. 6473 // C11 6.3.1.1/2: 6474 // [For a bit-field of type _Bool, int, signed int, or unsigned int:] 6475 // If an int can represent all values of the original type (as restricted by 6476 // the width, for a bit-field), the value is converted to an int; otherwise, 6477 // it is converted to an unsigned int. 6478 // 6479 // FIXME: C does not permit promotion of a 'long : 3' bitfield to int. 6480 // We perform that promotion here to match GCC and C++. 6481 // FIXME: C does not permit promotion of an enum bit-field whose rank is 6482 // greater than that of 'int'. We perform that promotion to match GCC. 6483 if (BitWidth < IntSize) 6484 return IntTy; 6485 6486 if (BitWidth == IntSize) 6487 return FT->isSignedIntegerType() ? IntTy : UnsignedIntTy; 6488 6489 // Bit-fields wider than int are not subject to promotions, and therefore act 6490 // like the base type. GCC has some weird bugs in this area that we 6491 // deliberately do not follow (GCC follows a pre-standard resolution to 6492 // C's DR315 which treats bit-width as being part of the type, and this leaks 6493 // into their semantics in some cases). 6494 return {}; 6495 } 6496 6497 /// getPromotedIntegerType - Returns the type that Promotable will 6498 /// promote to: C99 6.3.1.1p2, assuming that Promotable is a promotable 6499 /// integer type. 6500 QualType ASTContext::getPromotedIntegerType(QualType Promotable) const { 6501 assert(!Promotable.isNull()); 6502 assert(Promotable->isPromotableIntegerType()); 6503 if (const auto *ET = Promotable->getAs<EnumType>()) 6504 return ET->getDecl()->getPromotionType(); 6505 6506 if (const auto *BT = Promotable->getAs<BuiltinType>()) { 6507 // C++ [conv.prom]: A prvalue of type char16_t, char32_t, or wchar_t 6508 // (3.9.1) can be converted to a prvalue of the first of the following 6509 // types that can represent all the values of its underlying type: 6510 // int, unsigned int, long int, unsigned long int, long long int, or 6511 // unsigned long long int [...] 6512 // FIXME: Is there some better way to compute this? 6513 if (BT->getKind() == BuiltinType::WChar_S || 6514 BT->getKind() == BuiltinType::WChar_U || 6515 BT->getKind() == BuiltinType::Char8 || 6516 BT->getKind() == BuiltinType::Char16 || 6517 BT->getKind() == BuiltinType::Char32) { 6518 bool FromIsSigned = BT->getKind() == BuiltinType::WChar_S; 6519 uint64_t FromSize = getTypeSize(BT); 6520 QualType PromoteTypes[] = { IntTy, UnsignedIntTy, LongTy, UnsignedLongTy, 6521 LongLongTy, UnsignedLongLongTy }; 6522 for (size_t Idx = 0; Idx < llvm::array_lengthof(PromoteTypes); ++Idx) { 6523 uint64_t ToSize = getTypeSize(PromoteTypes[Idx]); 6524 if (FromSize < ToSize || 6525 (FromSize == ToSize && 6526 FromIsSigned == PromoteTypes[Idx]->isSignedIntegerType())) 6527 return PromoteTypes[Idx]; 6528 } 6529 llvm_unreachable("char type should fit into long long"); 6530 } 6531 } 6532 6533 // At this point, we should have a signed or unsigned integer type. 6534 if (Promotable->isSignedIntegerType()) 6535 return IntTy; 6536 uint64_t PromotableSize = getIntWidth(Promotable); 6537 uint64_t IntSize = getIntWidth(IntTy); 6538 assert(Promotable->isUnsignedIntegerType() && PromotableSize <= IntSize); 6539 return (PromotableSize != IntSize) ? IntTy : UnsignedIntTy; 6540 } 6541 6542 /// Recurses in pointer/array types until it finds an objc retainable 6543 /// type and returns its ownership. 6544 Qualifiers::ObjCLifetime ASTContext::getInnerObjCOwnership(QualType T) const { 6545 while (!T.isNull()) { 6546 if (T.getObjCLifetime() != Qualifiers::OCL_None) 6547 return T.getObjCLifetime(); 6548 if (T->isArrayType()) 6549 T = getBaseElementType(T); 6550 else if (const auto *PT = T->getAs<PointerType>()) 6551 T = PT->getPointeeType(); 6552 else if (const auto *RT = T->getAs<ReferenceType>()) 6553 T = RT->getPointeeType(); 6554 else 6555 break; 6556 } 6557 6558 return Qualifiers::OCL_None; 6559 } 6560 6561 static const Type *getIntegerTypeForEnum(const EnumType *ET) { 6562 // Incomplete enum types are not treated as integer types. 6563 // FIXME: In C++, enum types are never integer types. 6564 if (ET->getDecl()->isComplete() && !ET->getDecl()->isScoped()) 6565 return ET->getDecl()->getIntegerType().getTypePtr(); 6566 return nullptr; 6567 } 6568 6569 /// getIntegerTypeOrder - Returns the highest ranked integer type: 6570 /// C99 6.3.1.8p1. If LHS > RHS, return 1. If LHS == RHS, return 0. If 6571 /// LHS < RHS, return -1. 6572 int ASTContext::getIntegerTypeOrder(QualType LHS, QualType RHS) const { 6573 const Type *LHSC = getCanonicalType(LHS).getTypePtr(); 6574 const Type *RHSC = getCanonicalType(RHS).getTypePtr(); 6575 6576 // Unwrap enums to their underlying type. 6577 if (const auto *ET = dyn_cast<EnumType>(LHSC)) 6578 LHSC = getIntegerTypeForEnum(ET); 6579 if (const auto *ET = dyn_cast<EnumType>(RHSC)) 6580 RHSC = getIntegerTypeForEnum(ET); 6581 6582 if (LHSC == RHSC) return 0; 6583 6584 bool LHSUnsigned = LHSC->isUnsignedIntegerType(); 6585 bool RHSUnsigned = RHSC->isUnsignedIntegerType(); 6586 6587 unsigned LHSRank = getIntegerRank(LHSC); 6588 unsigned RHSRank = getIntegerRank(RHSC); 6589 6590 if (LHSUnsigned == RHSUnsigned) { // Both signed or both unsigned. 6591 if (LHSRank == RHSRank) return 0; 6592 return LHSRank > RHSRank ? 1 : -1; 6593 } 6594 6595 // Otherwise, the LHS is signed and the RHS is unsigned or visa versa. 6596 if (LHSUnsigned) { 6597 // If the unsigned [LHS] type is larger, return it. 6598 if (LHSRank >= RHSRank) 6599 return 1; 6600 6601 // If the signed type can represent all values of the unsigned type, it 6602 // wins. Because we are dealing with 2's complement and types that are 6603 // powers of two larger than each other, this is always safe. 6604 return -1; 6605 } 6606 6607 // If the unsigned [RHS] type is larger, return it. 6608 if (RHSRank >= LHSRank) 6609 return -1; 6610 6611 // If the signed type can represent all values of the unsigned type, it 6612 // wins. Because we are dealing with 2's complement and types that are 6613 // powers of two larger than each other, this is always safe. 6614 return 1; 6615 } 6616 6617 TypedefDecl *ASTContext::getCFConstantStringDecl() const { 6618 if (CFConstantStringTypeDecl) 6619 return CFConstantStringTypeDecl; 6620 6621 assert(!CFConstantStringTagDecl && 6622 "tag and typedef should be initialized together"); 6623 CFConstantStringTagDecl = buildImplicitRecord("__NSConstantString_tag"); 6624 CFConstantStringTagDecl->startDefinition(); 6625 6626 struct { 6627 QualType Type; 6628 const char *Name; 6629 } Fields[5]; 6630 unsigned Count = 0; 6631 6632 /// Objective-C ABI 6633 /// 6634 /// typedef struct __NSConstantString_tag { 6635 /// const int *isa; 6636 /// int flags; 6637 /// const char *str; 6638 /// long length; 6639 /// } __NSConstantString; 6640 /// 6641 /// Swift ABI (4.1, 4.2) 6642 /// 6643 /// typedef struct __NSConstantString_tag { 6644 /// uintptr_t _cfisa; 6645 /// uintptr_t _swift_rc; 6646 /// _Atomic(uint64_t) _cfinfoa; 6647 /// const char *_ptr; 6648 /// uint32_t _length; 6649 /// } __NSConstantString; 6650 /// 6651 /// Swift ABI (5.0) 6652 /// 6653 /// typedef struct __NSConstantString_tag { 6654 /// uintptr_t _cfisa; 6655 /// uintptr_t _swift_rc; 6656 /// _Atomic(uint64_t) _cfinfoa; 6657 /// const char *_ptr; 6658 /// uintptr_t _length; 6659 /// } __NSConstantString; 6660 6661 const auto CFRuntime = getLangOpts().CFRuntime; 6662 if (static_cast<unsigned>(CFRuntime) < 6663 static_cast<unsigned>(LangOptions::CoreFoundationABI::Swift)) { 6664 Fields[Count++] = { getPointerType(IntTy.withConst()), "isa" }; 6665 Fields[Count++] = { IntTy, "flags" }; 6666 Fields[Count++] = { getPointerType(CharTy.withConst()), "str" }; 6667 Fields[Count++] = { LongTy, "length" }; 6668 } else { 6669 Fields[Count++] = { getUIntPtrType(), "_cfisa" }; 6670 Fields[Count++] = { getUIntPtrType(), "_swift_rc" }; 6671 Fields[Count++] = { getFromTargetType(Target->getUInt64Type()), "_swift_rc" }; 6672 Fields[Count++] = { getPointerType(CharTy.withConst()), "_ptr" }; 6673 if (CFRuntime == LangOptions::CoreFoundationABI::Swift4_1 || 6674 CFRuntime == LangOptions::CoreFoundationABI::Swift4_2) 6675 Fields[Count++] = { IntTy, "_ptr" }; 6676 else 6677 Fields[Count++] = { getUIntPtrType(), "_ptr" }; 6678 } 6679 6680 // Create fields 6681 for (unsigned i = 0; i < Count; ++i) { 6682 FieldDecl *Field = 6683 FieldDecl::Create(*this, CFConstantStringTagDecl, SourceLocation(), 6684 SourceLocation(), &Idents.get(Fields[i].Name), 6685 Fields[i].Type, /*TInfo=*/nullptr, 6686 /*BitWidth=*/nullptr, /*Mutable=*/false, ICIS_NoInit); 6687 Field->setAccess(AS_public); 6688 CFConstantStringTagDecl->addDecl(Field); 6689 } 6690 6691 CFConstantStringTagDecl->completeDefinition(); 6692 // This type is designed to be compatible with NSConstantString, but cannot 6693 // use the same name, since NSConstantString is an interface. 6694 auto tagType = getTagDeclType(CFConstantStringTagDecl); 6695 CFConstantStringTypeDecl = 6696 buildImplicitTypedef(tagType, "__NSConstantString"); 6697 6698 return CFConstantStringTypeDecl; 6699 } 6700 6701 RecordDecl *ASTContext::getCFConstantStringTagDecl() const { 6702 if (!CFConstantStringTagDecl) 6703 getCFConstantStringDecl(); // Build the tag and the typedef. 6704 return CFConstantStringTagDecl; 6705 } 6706 6707 // getCFConstantStringType - Return the type used for constant CFStrings. 6708 QualType ASTContext::getCFConstantStringType() const { 6709 return getTypedefType(getCFConstantStringDecl()); 6710 } 6711 6712 QualType ASTContext::getObjCSuperType() const { 6713 if (ObjCSuperType.isNull()) { 6714 RecordDecl *ObjCSuperTypeDecl = buildImplicitRecord("objc_super"); 6715 getTranslationUnitDecl()->addDecl(ObjCSuperTypeDecl); 6716 ObjCSuperType = getTagDeclType(ObjCSuperTypeDecl); 6717 } 6718 return ObjCSuperType; 6719 } 6720 6721 void ASTContext::setCFConstantStringType(QualType T) { 6722 const auto *TD = T->castAs<TypedefType>(); 6723 CFConstantStringTypeDecl = cast<TypedefDecl>(TD->getDecl()); 6724 const auto *TagType = 6725 CFConstantStringTypeDecl->getUnderlyingType()->castAs<RecordType>(); 6726 CFConstantStringTagDecl = TagType->getDecl(); 6727 } 6728 6729 QualType ASTContext::getBlockDescriptorType() const { 6730 if (BlockDescriptorType) 6731 return getTagDeclType(BlockDescriptorType); 6732 6733 RecordDecl *RD; 6734 // FIXME: Needs the FlagAppleBlock bit. 6735 RD = buildImplicitRecord("__block_descriptor"); 6736 RD->startDefinition(); 6737 6738 QualType FieldTypes[] = { 6739 UnsignedLongTy, 6740 UnsignedLongTy, 6741 }; 6742 6743 static const char *const FieldNames[] = { 6744 "reserved", 6745 "Size" 6746 }; 6747 6748 for (size_t i = 0; i < 2; ++i) { 6749 FieldDecl *Field = FieldDecl::Create( 6750 *this, RD, SourceLocation(), SourceLocation(), 6751 &Idents.get(FieldNames[i]), FieldTypes[i], /*TInfo=*/nullptr, 6752 /*BitWidth=*/nullptr, /*Mutable=*/false, ICIS_NoInit); 6753 Field->setAccess(AS_public); 6754 RD->addDecl(Field); 6755 } 6756 6757 RD->completeDefinition(); 6758 6759 BlockDescriptorType = RD; 6760 6761 return getTagDeclType(BlockDescriptorType); 6762 } 6763 6764 QualType ASTContext::getBlockDescriptorExtendedType() const { 6765 if (BlockDescriptorExtendedType) 6766 return getTagDeclType(BlockDescriptorExtendedType); 6767 6768 RecordDecl *RD; 6769 // FIXME: Needs the FlagAppleBlock bit. 6770 RD = buildImplicitRecord("__block_descriptor_withcopydispose"); 6771 RD->startDefinition(); 6772 6773 QualType FieldTypes[] = { 6774 UnsignedLongTy, 6775 UnsignedLongTy, 6776 getPointerType(VoidPtrTy), 6777 getPointerType(VoidPtrTy) 6778 }; 6779 6780 static const char *const FieldNames[] = { 6781 "reserved", 6782 "Size", 6783 "CopyFuncPtr", 6784 "DestroyFuncPtr" 6785 }; 6786 6787 for (size_t i = 0; i < 4; ++i) { 6788 FieldDecl *Field = FieldDecl::Create( 6789 *this, RD, SourceLocation(), SourceLocation(), 6790 &Idents.get(FieldNames[i]), FieldTypes[i], /*TInfo=*/nullptr, 6791 /*BitWidth=*/nullptr, 6792 /*Mutable=*/false, ICIS_NoInit); 6793 Field->setAccess(AS_public); 6794 RD->addDecl(Field); 6795 } 6796 6797 RD->completeDefinition(); 6798 6799 BlockDescriptorExtendedType = RD; 6800 return getTagDeclType(BlockDescriptorExtendedType); 6801 } 6802 6803 OpenCLTypeKind ASTContext::getOpenCLTypeKind(const Type *T) const { 6804 const auto *BT = dyn_cast<BuiltinType>(T); 6805 6806 if (!BT) { 6807 if (isa<PipeType>(T)) 6808 return OCLTK_Pipe; 6809 6810 return OCLTK_Default; 6811 } 6812 6813 switch (BT->getKind()) { 6814 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \ 6815 case BuiltinType::Id: \ 6816 return OCLTK_Image; 6817 #include "clang/Basic/OpenCLImageTypes.def" 6818 6819 case BuiltinType::OCLClkEvent: 6820 return OCLTK_ClkEvent; 6821 6822 case BuiltinType::OCLEvent: 6823 return OCLTK_Event; 6824 6825 case BuiltinType::OCLQueue: 6826 return OCLTK_Queue; 6827 6828 case BuiltinType::OCLReserveID: 6829 return OCLTK_ReserveID; 6830 6831 case BuiltinType::OCLSampler: 6832 return OCLTK_Sampler; 6833 6834 default: 6835 return OCLTK_Default; 6836 } 6837 } 6838 6839 LangAS ASTContext::getOpenCLTypeAddrSpace(const Type *T) const { 6840 return Target->getOpenCLTypeAddrSpace(getOpenCLTypeKind(T)); 6841 } 6842 6843 /// BlockRequiresCopying - Returns true if byref variable "D" of type "Ty" 6844 /// requires copy/dispose. Note that this must match the logic 6845 /// in buildByrefHelpers. 6846 bool ASTContext::BlockRequiresCopying(QualType Ty, 6847 const VarDecl *D) { 6848 if (const CXXRecordDecl *record = Ty->getAsCXXRecordDecl()) { 6849 const Expr *copyExpr = getBlockVarCopyInit(D).getCopyExpr(); 6850 if (!copyExpr && record->hasTrivialDestructor()) return false; 6851 6852 return true; 6853 } 6854 6855 // The block needs copy/destroy helpers if Ty is non-trivial to destructively 6856 // move or destroy. 6857 if (Ty.isNonTrivialToPrimitiveDestructiveMove() || Ty.isDestructedType()) 6858 return true; 6859 6860 if (!Ty->isObjCRetainableType()) return false; 6861 6862 Qualifiers qs = Ty.getQualifiers(); 6863 6864 // If we have lifetime, that dominates. 6865 if (Qualifiers::ObjCLifetime lifetime = qs.getObjCLifetime()) { 6866 switch (lifetime) { 6867 case Qualifiers::OCL_None: llvm_unreachable("impossible"); 6868 6869 // These are just bits as far as the runtime is concerned. 6870 case Qualifiers::OCL_ExplicitNone: 6871 case Qualifiers::OCL_Autoreleasing: 6872 return false; 6873 6874 // These cases should have been taken care of when checking the type's 6875 // non-triviality. 6876 case Qualifiers::OCL_Weak: 6877 case Qualifiers::OCL_Strong: 6878 llvm_unreachable("impossible"); 6879 } 6880 llvm_unreachable("fell out of lifetime switch!"); 6881 } 6882 return (Ty->isBlockPointerType() || isObjCNSObjectType(Ty) || 6883 Ty->isObjCObjectPointerType()); 6884 } 6885 6886 bool ASTContext::getByrefLifetime(QualType Ty, 6887 Qualifiers::ObjCLifetime &LifeTime, 6888 bool &HasByrefExtendedLayout) const { 6889 if (!getLangOpts().ObjC || 6890 getLangOpts().getGC() != LangOptions::NonGC) 6891 return false; 6892 6893 HasByrefExtendedLayout = false; 6894 if (Ty->isRecordType()) { 6895 HasByrefExtendedLayout = true; 6896 LifeTime = Qualifiers::OCL_None; 6897 } else if ((LifeTime = Ty.getObjCLifetime())) { 6898 // Honor the ARC qualifiers. 6899 } else if (Ty->isObjCObjectPointerType() || Ty->isBlockPointerType()) { 6900 // The MRR rule. 6901 LifeTime = Qualifiers::OCL_ExplicitNone; 6902 } else { 6903 LifeTime = Qualifiers::OCL_None; 6904 } 6905 return true; 6906 } 6907 6908 CanQualType ASTContext::getNSUIntegerType() const { 6909 assert(Target && "Expected target to be initialized"); 6910 const llvm::Triple &T = Target->getTriple(); 6911 // Windows is LLP64 rather than LP64 6912 if (T.isOSWindows() && T.isArch64Bit()) 6913 return UnsignedLongLongTy; 6914 return UnsignedLongTy; 6915 } 6916 6917 CanQualType ASTContext::getNSIntegerType() const { 6918 assert(Target && "Expected target to be initialized"); 6919 const llvm::Triple &T = Target->getTriple(); 6920 // Windows is LLP64 rather than LP64 6921 if (T.isOSWindows() && T.isArch64Bit()) 6922 return LongLongTy; 6923 return LongTy; 6924 } 6925 6926 TypedefDecl *ASTContext::getObjCInstanceTypeDecl() { 6927 if (!ObjCInstanceTypeDecl) 6928 ObjCInstanceTypeDecl = 6929 buildImplicitTypedef(getObjCIdType(), "instancetype"); 6930 return ObjCInstanceTypeDecl; 6931 } 6932 6933 // This returns true if a type has been typedefed to BOOL: 6934 // typedef <type> BOOL; 6935 static bool isTypeTypedefedAsBOOL(QualType T) { 6936 if (const auto *TT = dyn_cast<TypedefType>(T)) 6937 if (IdentifierInfo *II = TT->getDecl()->getIdentifier()) 6938 return II->isStr("BOOL"); 6939 6940 return false; 6941 } 6942 6943 /// getObjCEncodingTypeSize returns size of type for objective-c encoding 6944 /// purpose. 6945 CharUnits ASTContext::getObjCEncodingTypeSize(QualType type) const { 6946 if (!type->isIncompleteArrayType() && type->isIncompleteType()) 6947 return CharUnits::Zero(); 6948 6949 CharUnits sz = getTypeSizeInChars(type); 6950 6951 // Make all integer and enum types at least as large as an int 6952 if (sz.isPositive() && type->isIntegralOrEnumerationType()) 6953 sz = std::max(sz, getTypeSizeInChars(IntTy)); 6954 // Treat arrays as pointers, since that's how they're passed in. 6955 else if (type->isArrayType()) 6956 sz = getTypeSizeInChars(VoidPtrTy); 6957 return sz; 6958 } 6959 6960 bool ASTContext::isMSStaticDataMemberInlineDefinition(const VarDecl *VD) const { 6961 return getTargetInfo().getCXXABI().isMicrosoft() && 6962 VD->isStaticDataMember() && 6963 VD->getType()->isIntegralOrEnumerationType() && 6964 !VD->getFirstDecl()->isOutOfLine() && VD->getFirstDecl()->hasInit(); 6965 } 6966 6967 ASTContext::InlineVariableDefinitionKind 6968 ASTContext::getInlineVariableDefinitionKind(const VarDecl *VD) const { 6969 if (!VD->isInline()) 6970 return InlineVariableDefinitionKind::None; 6971 6972 // In almost all cases, it's a weak definition. 6973 auto *First = VD->getFirstDecl(); 6974 if (First->isInlineSpecified() || !First->isStaticDataMember()) 6975 return InlineVariableDefinitionKind::Weak; 6976 6977 // If there's a file-context declaration in this translation unit, it's a 6978 // non-discardable definition. 6979 for (auto *D : VD->redecls()) 6980 if (D->getLexicalDeclContext()->isFileContext() && 6981 !D->isInlineSpecified() && (D->isConstexpr() || First->isConstexpr())) 6982 return InlineVariableDefinitionKind::Strong; 6983 6984 // If we've not seen one yet, we don't know. 6985 return InlineVariableDefinitionKind::WeakUnknown; 6986 } 6987 6988 static std::string charUnitsToString(const CharUnits &CU) { 6989 return llvm::itostr(CU.getQuantity()); 6990 } 6991 6992 /// getObjCEncodingForBlock - Return the encoded type for this block 6993 /// declaration. 6994 std::string ASTContext::getObjCEncodingForBlock(const BlockExpr *Expr) const { 6995 std::string S; 6996 6997 const BlockDecl *Decl = Expr->getBlockDecl(); 6998 QualType BlockTy = 6999 Expr->getType()->castAs<BlockPointerType>()->getPointeeType(); 7000 QualType BlockReturnTy = BlockTy->castAs<FunctionType>()->getReturnType(); 7001 // Encode result type. 7002 if (getLangOpts().EncodeExtendedBlockSig) 7003 getObjCEncodingForMethodParameter(Decl::OBJC_TQ_None, BlockReturnTy, S, 7004 true /*Extended*/); 7005 else 7006 getObjCEncodingForType(BlockReturnTy, S); 7007 // Compute size of all parameters. 7008 // Start with computing size of a pointer in number of bytes. 7009 // FIXME: There might(should) be a better way of doing this computation! 7010 CharUnits PtrSize = getTypeSizeInChars(VoidPtrTy); 7011 CharUnits ParmOffset = PtrSize; 7012 for (auto PI : Decl->parameters()) { 7013 QualType PType = PI->getType(); 7014 CharUnits sz = getObjCEncodingTypeSize(PType); 7015 if (sz.isZero()) 7016 continue; 7017 assert(sz.isPositive() && "BlockExpr - Incomplete param type"); 7018 ParmOffset += sz; 7019 } 7020 // Size of the argument frame 7021 S += charUnitsToString(ParmOffset); 7022 // Block pointer and offset. 7023 S += "@?0"; 7024 7025 // Argument types. 7026 ParmOffset = PtrSize; 7027 for (auto PVDecl : Decl->parameters()) { 7028 QualType PType = PVDecl->getOriginalType(); 7029 if (const auto *AT = 7030 dyn_cast<ArrayType>(PType->getCanonicalTypeInternal())) { 7031 // Use array's original type only if it has known number of 7032 // elements. 7033 if (!isa<ConstantArrayType>(AT)) 7034 PType = PVDecl->getType(); 7035 } else if (PType->isFunctionType()) 7036 PType = PVDecl->getType(); 7037 if (getLangOpts().EncodeExtendedBlockSig) 7038 getObjCEncodingForMethodParameter(Decl::OBJC_TQ_None, PType, 7039 S, true /*Extended*/); 7040 else 7041 getObjCEncodingForType(PType, S); 7042 S += charUnitsToString(ParmOffset); 7043 ParmOffset += getObjCEncodingTypeSize(PType); 7044 } 7045 7046 return S; 7047 } 7048 7049 std::string 7050 ASTContext::getObjCEncodingForFunctionDecl(const FunctionDecl *Decl) const { 7051 std::string S; 7052 // Encode result type. 7053 getObjCEncodingForType(Decl->getReturnType(), S); 7054 CharUnits ParmOffset; 7055 // Compute size of all parameters. 7056 for (auto PI : Decl->parameters()) { 7057 QualType PType = PI->getType(); 7058 CharUnits sz = getObjCEncodingTypeSize(PType); 7059 if (sz.isZero()) 7060 continue; 7061 7062 assert(sz.isPositive() && 7063 "getObjCEncodingForFunctionDecl - Incomplete param type"); 7064 ParmOffset += sz; 7065 } 7066 S += charUnitsToString(ParmOffset); 7067 ParmOffset = CharUnits::Zero(); 7068 7069 // Argument types. 7070 for (auto PVDecl : Decl->parameters()) { 7071 QualType PType = PVDecl->getOriginalType(); 7072 if (const auto *AT = 7073 dyn_cast<ArrayType>(PType->getCanonicalTypeInternal())) { 7074 // Use array's original type only if it has known number of 7075 // elements. 7076 if (!isa<ConstantArrayType>(AT)) 7077 PType = PVDecl->getType(); 7078 } else if (PType->isFunctionType()) 7079 PType = PVDecl->getType(); 7080 getObjCEncodingForType(PType, S); 7081 S += charUnitsToString(ParmOffset); 7082 ParmOffset += getObjCEncodingTypeSize(PType); 7083 } 7084 7085 return S; 7086 } 7087 7088 /// getObjCEncodingForMethodParameter - Return the encoded type for a single 7089 /// method parameter or return type. If Extended, include class names and 7090 /// block object types. 7091 void ASTContext::getObjCEncodingForMethodParameter(Decl::ObjCDeclQualifier QT, 7092 QualType T, std::string& S, 7093 bool Extended) const { 7094 // Encode type qualifier, 'in', 'inout', etc. for the parameter. 7095 getObjCEncodingForTypeQualifier(QT, S); 7096 // Encode parameter type. 7097 ObjCEncOptions Options = ObjCEncOptions() 7098 .setExpandPointedToStructures() 7099 .setExpandStructures() 7100 .setIsOutermostType(); 7101 if (Extended) 7102 Options.setEncodeBlockParameters().setEncodeClassNames(); 7103 getObjCEncodingForTypeImpl(T, S, Options, /*Field=*/nullptr); 7104 } 7105 7106 /// getObjCEncodingForMethodDecl - Return the encoded type for this method 7107 /// declaration. 7108 std::string ASTContext::getObjCEncodingForMethodDecl(const ObjCMethodDecl *Decl, 7109 bool Extended) const { 7110 // FIXME: This is not very efficient. 7111 // Encode return type. 7112 std::string S; 7113 getObjCEncodingForMethodParameter(Decl->getObjCDeclQualifier(), 7114 Decl->getReturnType(), S, Extended); 7115 // Compute size of all parameters. 7116 // Start with computing size of a pointer in number of bytes. 7117 // FIXME: There might(should) be a better way of doing this computation! 7118 CharUnits PtrSize = getTypeSizeInChars(VoidPtrTy); 7119 // The first two arguments (self and _cmd) are pointers; account for 7120 // their size. 7121 CharUnits ParmOffset = 2 * PtrSize; 7122 for (ObjCMethodDecl::param_const_iterator PI = Decl->param_begin(), 7123 E = Decl->sel_param_end(); PI != E; ++PI) { 7124 QualType PType = (*PI)->getType(); 7125 CharUnits sz = getObjCEncodingTypeSize(PType); 7126 if (sz.isZero()) 7127 continue; 7128 7129 assert(sz.isPositive() && 7130 "getObjCEncodingForMethodDecl - Incomplete param type"); 7131 ParmOffset += sz; 7132 } 7133 S += charUnitsToString(ParmOffset); 7134 S += "@0:"; 7135 S += charUnitsToString(PtrSize); 7136 7137 // Argument types. 7138 ParmOffset = 2 * PtrSize; 7139 for (ObjCMethodDecl::param_const_iterator PI = Decl->param_begin(), 7140 E = Decl->sel_param_end(); PI != E; ++PI) { 7141 const ParmVarDecl *PVDecl = *PI; 7142 QualType PType = PVDecl->getOriginalType(); 7143 if (const auto *AT = 7144 dyn_cast<ArrayType>(PType->getCanonicalTypeInternal())) { 7145 // Use array's original type only if it has known number of 7146 // elements. 7147 if (!isa<ConstantArrayType>(AT)) 7148 PType = PVDecl->getType(); 7149 } else if (PType->isFunctionType()) 7150 PType = PVDecl->getType(); 7151 getObjCEncodingForMethodParameter(PVDecl->getObjCDeclQualifier(), 7152 PType, S, Extended); 7153 S += charUnitsToString(ParmOffset); 7154 ParmOffset += getObjCEncodingTypeSize(PType); 7155 } 7156 7157 return S; 7158 } 7159 7160 ObjCPropertyImplDecl * 7161 ASTContext::getObjCPropertyImplDeclForPropertyDecl( 7162 const ObjCPropertyDecl *PD, 7163 const Decl *Container) const { 7164 if (!Container) 7165 return nullptr; 7166 if (const auto *CID = dyn_cast<ObjCCategoryImplDecl>(Container)) { 7167 for (auto *PID : CID->property_impls()) 7168 if (PID->getPropertyDecl() == PD) 7169 return PID; 7170 } else { 7171 const auto *OID = cast<ObjCImplementationDecl>(Container); 7172 for (auto *PID : OID->property_impls()) 7173 if (PID->getPropertyDecl() == PD) 7174 return PID; 7175 } 7176 return nullptr; 7177 } 7178 7179 /// getObjCEncodingForPropertyDecl - Return the encoded type for this 7180 /// property declaration. If non-NULL, Container must be either an 7181 /// ObjCCategoryImplDecl or ObjCImplementationDecl; it should only be 7182 /// NULL when getting encodings for protocol properties. 7183 /// Property attributes are stored as a comma-delimited C string. The simple 7184 /// attributes readonly and bycopy are encoded as single characters. The 7185 /// parametrized attributes, getter=name, setter=name, and ivar=name, are 7186 /// encoded as single characters, followed by an identifier. Property types 7187 /// are also encoded as a parametrized attribute. The characters used to encode 7188 /// these attributes are defined by the following enumeration: 7189 /// @code 7190 /// enum PropertyAttributes { 7191 /// kPropertyReadOnly = 'R', // property is read-only. 7192 /// kPropertyBycopy = 'C', // property is a copy of the value last assigned 7193 /// kPropertyByref = '&', // property is a reference to the value last assigned 7194 /// kPropertyDynamic = 'D', // property is dynamic 7195 /// kPropertyGetter = 'G', // followed by getter selector name 7196 /// kPropertySetter = 'S', // followed by setter selector name 7197 /// kPropertyInstanceVariable = 'V' // followed by instance variable name 7198 /// kPropertyType = 'T' // followed by old-style type encoding. 7199 /// kPropertyWeak = 'W' // 'weak' property 7200 /// kPropertyStrong = 'P' // property GC'able 7201 /// kPropertyNonAtomic = 'N' // property non-atomic 7202 /// }; 7203 /// @endcode 7204 std::string 7205 ASTContext::getObjCEncodingForPropertyDecl(const ObjCPropertyDecl *PD, 7206 const Decl *Container) const { 7207 // Collect information from the property implementation decl(s). 7208 bool Dynamic = false; 7209 ObjCPropertyImplDecl *SynthesizePID = nullptr; 7210 7211 if (ObjCPropertyImplDecl *PropertyImpDecl = 7212 getObjCPropertyImplDeclForPropertyDecl(PD, Container)) { 7213 if (PropertyImpDecl->getPropertyImplementation() == ObjCPropertyImplDecl::Dynamic) 7214 Dynamic = true; 7215 else 7216 SynthesizePID = PropertyImpDecl; 7217 } 7218 7219 // FIXME: This is not very efficient. 7220 std::string S = "T"; 7221 7222 // Encode result type. 7223 // GCC has some special rules regarding encoding of properties which 7224 // closely resembles encoding of ivars. 7225 getObjCEncodingForPropertyType(PD->getType(), S); 7226 7227 if (PD->isReadOnly()) { 7228 S += ",R"; 7229 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_copy) 7230 S += ",C"; 7231 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_retain) 7232 S += ",&"; 7233 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_weak) 7234 S += ",W"; 7235 } else { 7236 switch (PD->getSetterKind()) { 7237 case ObjCPropertyDecl::Assign: break; 7238 case ObjCPropertyDecl::Copy: S += ",C"; break; 7239 case ObjCPropertyDecl::Retain: S += ",&"; break; 7240 case ObjCPropertyDecl::Weak: S += ",W"; break; 7241 } 7242 } 7243 7244 // It really isn't clear at all what this means, since properties 7245 // are "dynamic by default". 7246 if (Dynamic) 7247 S += ",D"; 7248 7249 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_nonatomic) 7250 S += ",N"; 7251 7252 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_getter) { 7253 S += ",G"; 7254 S += PD->getGetterName().getAsString(); 7255 } 7256 7257 if (PD->getPropertyAttributes() & ObjCPropertyAttribute::kind_setter) { 7258 S += ",S"; 7259 S += PD->getSetterName().getAsString(); 7260 } 7261 7262 if (SynthesizePID) { 7263 const ObjCIvarDecl *OID = SynthesizePID->getPropertyIvarDecl(); 7264 S += ",V"; 7265 S += OID->getNameAsString(); 7266 } 7267 7268 // FIXME: OBJCGC: weak & strong 7269 return S; 7270 } 7271 7272 /// getLegacyIntegralTypeEncoding - 7273 /// Another legacy compatibility encoding: 32-bit longs are encoded as 7274 /// 'l' or 'L' , but not always. For typedefs, we need to use 7275 /// 'i' or 'I' instead if encoding a struct field, or a pointer! 7276 void ASTContext::getLegacyIntegralTypeEncoding (QualType &PointeeTy) const { 7277 if (isa<TypedefType>(PointeeTy.getTypePtr())) { 7278 if (const auto *BT = PointeeTy->getAs<BuiltinType>()) { 7279 if (BT->getKind() == BuiltinType::ULong && getIntWidth(PointeeTy) == 32) 7280 PointeeTy = UnsignedIntTy; 7281 else 7282 if (BT->getKind() == BuiltinType::Long && getIntWidth(PointeeTy) == 32) 7283 PointeeTy = IntTy; 7284 } 7285 } 7286 } 7287 7288 void ASTContext::getObjCEncodingForType(QualType T, std::string& S, 7289 const FieldDecl *Field, 7290 QualType *NotEncodedT) const { 7291 // We follow the behavior of gcc, expanding structures which are 7292 // directly pointed to, and expanding embedded structures. Note that 7293 // these rules are sufficient to prevent recursive encoding of the 7294 // same type. 7295 getObjCEncodingForTypeImpl(T, S, 7296 ObjCEncOptions() 7297 .setExpandPointedToStructures() 7298 .setExpandStructures() 7299 .setIsOutermostType(), 7300 Field, NotEncodedT); 7301 } 7302 7303 void ASTContext::getObjCEncodingForPropertyType(QualType T, 7304 std::string& S) const { 7305 // Encode result type. 7306 // GCC has some special rules regarding encoding of properties which 7307 // closely resembles encoding of ivars. 7308 getObjCEncodingForTypeImpl(T, S, 7309 ObjCEncOptions() 7310 .setExpandPointedToStructures() 7311 .setExpandStructures() 7312 .setIsOutermostType() 7313 .setEncodingProperty(), 7314 /*Field=*/nullptr); 7315 } 7316 7317 static char getObjCEncodingForPrimitiveType(const ASTContext *C, 7318 const BuiltinType *BT) { 7319 BuiltinType::Kind kind = BT->getKind(); 7320 switch (kind) { 7321 case BuiltinType::Void: return 'v'; 7322 case BuiltinType::Bool: return 'B'; 7323 case BuiltinType::Char8: 7324 case BuiltinType::Char_U: 7325 case BuiltinType::UChar: return 'C'; 7326 case BuiltinType::Char16: 7327 case BuiltinType::UShort: return 'S'; 7328 case BuiltinType::Char32: 7329 case BuiltinType::UInt: return 'I'; 7330 case BuiltinType::ULong: 7331 return C->getTargetInfo().getLongWidth() == 32 ? 'L' : 'Q'; 7332 case BuiltinType::UInt128: return 'T'; 7333 case BuiltinType::ULongLong: return 'Q'; 7334 case BuiltinType::Char_S: 7335 case BuiltinType::SChar: return 'c'; 7336 case BuiltinType::Short: return 's'; 7337 case BuiltinType::WChar_S: 7338 case BuiltinType::WChar_U: 7339 case BuiltinType::Int: return 'i'; 7340 case BuiltinType::Long: 7341 return C->getTargetInfo().getLongWidth() == 32 ? 'l' : 'q'; 7342 case BuiltinType::LongLong: return 'q'; 7343 case BuiltinType::Int128: return 't'; 7344 case BuiltinType::Float: return 'f'; 7345 case BuiltinType::Double: return 'd'; 7346 case BuiltinType::LongDouble: return 'D'; 7347 case BuiltinType::NullPtr: return '*'; // like char* 7348 7349 case BuiltinType::BFloat16: 7350 case BuiltinType::Float16: 7351 case BuiltinType::Float128: 7352 case BuiltinType::Ibm128: 7353 case BuiltinType::Half: 7354 case BuiltinType::ShortAccum: 7355 case BuiltinType::Accum: 7356 case BuiltinType::LongAccum: 7357 case BuiltinType::UShortAccum: 7358 case BuiltinType::UAccum: 7359 case BuiltinType::ULongAccum: 7360 case BuiltinType::ShortFract: 7361 case BuiltinType::Fract: 7362 case BuiltinType::LongFract: 7363 case BuiltinType::UShortFract: 7364 case BuiltinType::UFract: 7365 case BuiltinType::ULongFract: 7366 case BuiltinType::SatShortAccum: 7367 case BuiltinType::SatAccum: 7368 case BuiltinType::SatLongAccum: 7369 case BuiltinType::SatUShortAccum: 7370 case BuiltinType::SatUAccum: 7371 case BuiltinType::SatULongAccum: 7372 case BuiltinType::SatShortFract: 7373 case BuiltinType::SatFract: 7374 case BuiltinType::SatLongFract: 7375 case BuiltinType::SatUShortFract: 7376 case BuiltinType::SatUFract: 7377 case BuiltinType::SatULongFract: 7378 // FIXME: potentially need @encodes for these! 7379 return ' '; 7380 7381 #define SVE_TYPE(Name, Id, SingletonId) \ 7382 case BuiltinType::Id: 7383 #include "clang/Basic/AArch64SVEACLETypes.def" 7384 #define RVV_TYPE(Name, Id, SingletonId) case BuiltinType::Id: 7385 #include "clang/Basic/RISCVVTypes.def" 7386 { 7387 DiagnosticsEngine &Diags = C->getDiagnostics(); 7388 unsigned DiagID = Diags.getCustomDiagID(DiagnosticsEngine::Error, 7389 "cannot yet @encode type %0"); 7390 Diags.Report(DiagID) << BT->getName(C->getPrintingPolicy()); 7391 return ' '; 7392 } 7393 7394 case BuiltinType::ObjCId: 7395 case BuiltinType::ObjCClass: 7396 case BuiltinType::ObjCSel: 7397 llvm_unreachable("@encoding ObjC primitive type"); 7398 7399 // OpenCL and placeholder types don't need @encodings. 7400 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \ 7401 case BuiltinType::Id: 7402 #include "clang/Basic/OpenCLImageTypes.def" 7403 #define EXT_OPAQUE_TYPE(ExtType, Id, Ext) \ 7404 case BuiltinType::Id: 7405 #include "clang/Basic/OpenCLExtensionTypes.def" 7406 case BuiltinType::OCLEvent: 7407 case BuiltinType::OCLClkEvent: 7408 case BuiltinType::OCLQueue: 7409 case BuiltinType::OCLReserveID: 7410 case BuiltinType::OCLSampler: 7411 case BuiltinType::Dependent: 7412 #define PPC_VECTOR_TYPE(Name, Id, Size) \ 7413 case BuiltinType::Id: 7414 #include "clang/Basic/PPCTypes.def" 7415 #define BUILTIN_TYPE(KIND, ID) 7416 #define PLACEHOLDER_TYPE(KIND, ID) \ 7417 case BuiltinType::KIND: 7418 #include "clang/AST/BuiltinTypes.def" 7419 llvm_unreachable("invalid builtin type for @encode"); 7420 } 7421 llvm_unreachable("invalid BuiltinType::Kind value"); 7422 } 7423 7424 static char ObjCEncodingForEnumType(const ASTContext *C, const EnumType *ET) { 7425 EnumDecl *Enum = ET->getDecl(); 7426 7427 // The encoding of an non-fixed enum type is always 'i', regardless of size. 7428 if (!Enum->isFixed()) 7429 return 'i'; 7430 7431 // The encoding of a fixed enum type matches its fixed underlying type. 7432 const auto *BT = Enum->getIntegerType()->castAs<BuiltinType>(); 7433 return getObjCEncodingForPrimitiveType(C, BT); 7434 } 7435 7436 static void EncodeBitField(const ASTContext *Ctx, std::string& S, 7437 QualType T, const FieldDecl *FD) { 7438 assert(FD->isBitField() && "not a bitfield - getObjCEncodingForTypeImpl"); 7439 S += 'b'; 7440 // The NeXT runtime encodes bit fields as b followed by the number of bits. 7441 // The GNU runtime requires more information; bitfields are encoded as b, 7442 // then the offset (in bits) of the first element, then the type of the 7443 // bitfield, then the size in bits. For example, in this structure: 7444 // 7445 // struct 7446 // { 7447 // int integer; 7448 // int flags:2; 7449 // }; 7450 // On a 32-bit system, the encoding for flags would be b2 for the NeXT 7451 // runtime, but b32i2 for the GNU runtime. The reason for this extra 7452 // information is not especially sensible, but we're stuck with it for 7453 // compatibility with GCC, although providing it breaks anything that 7454 // actually uses runtime introspection and wants to work on both runtimes... 7455 if (Ctx->getLangOpts().ObjCRuntime.isGNUFamily()) { 7456 uint64_t Offset; 7457 7458 if (const auto *IVD = dyn_cast<ObjCIvarDecl>(FD)) { 7459 Offset = Ctx->lookupFieldBitOffset(IVD->getContainingInterface(), nullptr, 7460 IVD); 7461 } else { 7462 const RecordDecl *RD = FD->getParent(); 7463 const ASTRecordLayout &RL = Ctx->getASTRecordLayout(RD); 7464 Offset = RL.getFieldOffset(FD->getFieldIndex()); 7465 } 7466 7467 S += llvm::utostr(Offset); 7468 7469 if (const auto *ET = T->getAs<EnumType>()) 7470 S += ObjCEncodingForEnumType(Ctx, ET); 7471 else { 7472 const auto *BT = T->castAs<BuiltinType>(); 7473 S += getObjCEncodingForPrimitiveType(Ctx, BT); 7474 } 7475 } 7476 S += llvm::utostr(FD->getBitWidthValue(*Ctx)); 7477 } 7478 7479 // Helper function for determining whether the encoded type string would include 7480 // a template specialization type. 7481 static bool hasTemplateSpecializationInEncodedString(const Type *T, 7482 bool VisitBasesAndFields) { 7483 T = T->getBaseElementTypeUnsafe(); 7484 7485 if (auto *PT = T->getAs<PointerType>()) 7486 return hasTemplateSpecializationInEncodedString( 7487 PT->getPointeeType().getTypePtr(), false); 7488 7489 auto *CXXRD = T->getAsCXXRecordDecl(); 7490 7491 if (!CXXRD) 7492 return false; 7493 7494 if (isa<ClassTemplateSpecializationDecl>(CXXRD)) 7495 return true; 7496 7497 if (!CXXRD->hasDefinition() || !VisitBasesAndFields) 7498 return false; 7499 7500 for (auto B : CXXRD->bases()) 7501 if (hasTemplateSpecializationInEncodedString(B.getType().getTypePtr(), 7502 true)) 7503 return true; 7504 7505 for (auto *FD : CXXRD->fields()) 7506 if (hasTemplateSpecializationInEncodedString(FD->getType().getTypePtr(), 7507 true)) 7508 return true; 7509 7510 return false; 7511 } 7512 7513 // FIXME: Use SmallString for accumulating string. 7514 void ASTContext::getObjCEncodingForTypeImpl(QualType T, std::string &S, 7515 const ObjCEncOptions Options, 7516 const FieldDecl *FD, 7517 QualType *NotEncodedT) const { 7518 CanQualType CT = getCanonicalType(T); 7519 switch (CT->getTypeClass()) { 7520 case Type::Builtin: 7521 case Type::Enum: 7522 if (FD && FD->isBitField()) 7523 return EncodeBitField(this, S, T, FD); 7524 if (const auto *BT = dyn_cast<BuiltinType>(CT)) 7525 S += getObjCEncodingForPrimitiveType(this, BT); 7526 else 7527 S += ObjCEncodingForEnumType(this, cast<EnumType>(CT)); 7528 return; 7529 7530 case Type::Complex: 7531 S += 'j'; 7532 getObjCEncodingForTypeImpl(T->castAs<ComplexType>()->getElementType(), S, 7533 ObjCEncOptions(), 7534 /*Field=*/nullptr); 7535 return; 7536 7537 case Type::Atomic: 7538 S += 'A'; 7539 getObjCEncodingForTypeImpl(T->castAs<AtomicType>()->getValueType(), S, 7540 ObjCEncOptions(), 7541 /*Field=*/nullptr); 7542 return; 7543 7544 // encoding for pointer or reference types. 7545 case Type::Pointer: 7546 case Type::LValueReference: 7547 case Type::RValueReference: { 7548 QualType PointeeTy; 7549 if (isa<PointerType>(CT)) { 7550 const auto *PT = T->castAs<PointerType>(); 7551 if (PT->isObjCSelType()) { 7552 S += ':'; 7553 return; 7554 } 7555 PointeeTy = PT->getPointeeType(); 7556 } else { 7557 PointeeTy = T->castAs<ReferenceType>()->getPointeeType(); 7558 } 7559 7560 bool isReadOnly = false; 7561 // For historical/compatibility reasons, the read-only qualifier of the 7562 // pointee gets emitted _before_ the '^'. The read-only qualifier of 7563 // the pointer itself gets ignored, _unless_ we are looking at a typedef! 7564 // Also, do not emit the 'r' for anything but the outermost type! 7565 if (isa<TypedefType>(T.getTypePtr())) { 7566 if (Options.IsOutermostType() && T.isConstQualified()) { 7567 isReadOnly = true; 7568 S += 'r'; 7569 } 7570 } else if (Options.IsOutermostType()) { 7571 QualType P = PointeeTy; 7572 while (auto PT = P->getAs<PointerType>()) 7573 P = PT->getPointeeType(); 7574 if (P.isConstQualified()) { 7575 isReadOnly = true; 7576 S += 'r'; 7577 } 7578 } 7579 if (isReadOnly) { 7580 // Another legacy compatibility encoding. Some ObjC qualifier and type 7581 // combinations need to be rearranged. 7582 // Rewrite "in const" from "nr" to "rn" 7583 if (StringRef(S).endswith("nr")) 7584 S.replace(S.end()-2, S.end(), "rn"); 7585 } 7586 7587 if (PointeeTy->isCharType()) { 7588 // char pointer types should be encoded as '*' unless it is a 7589 // type that has been typedef'd to 'BOOL'. 7590 if (!isTypeTypedefedAsBOOL(PointeeTy)) { 7591 S += '*'; 7592 return; 7593 } 7594 } else if (const auto *RTy = PointeeTy->getAs<RecordType>()) { 7595 // GCC binary compat: Need to convert "struct objc_class *" to "#". 7596 if (RTy->getDecl()->getIdentifier() == &Idents.get("objc_class")) { 7597 S += '#'; 7598 return; 7599 } 7600 // GCC binary compat: Need to convert "struct objc_object *" to "@". 7601 if (RTy->getDecl()->getIdentifier() == &Idents.get("objc_object")) { 7602 S += '@'; 7603 return; 7604 } 7605 // If the encoded string for the class includes template names, just emit 7606 // "^v" for pointers to the class. 7607 if (getLangOpts().CPlusPlus && 7608 (!getLangOpts().EncodeCXXClassTemplateSpec && 7609 hasTemplateSpecializationInEncodedString( 7610 RTy, Options.ExpandPointedToStructures()))) { 7611 S += "^v"; 7612 return; 7613 } 7614 // fall through... 7615 } 7616 S += '^'; 7617 getLegacyIntegralTypeEncoding(PointeeTy); 7618 7619 ObjCEncOptions NewOptions; 7620 if (Options.ExpandPointedToStructures()) 7621 NewOptions.setExpandStructures(); 7622 getObjCEncodingForTypeImpl(PointeeTy, S, NewOptions, 7623 /*Field=*/nullptr, NotEncodedT); 7624 return; 7625 } 7626 7627 case Type::ConstantArray: 7628 case Type::IncompleteArray: 7629 case Type::VariableArray: { 7630 const auto *AT = cast<ArrayType>(CT); 7631 7632 if (isa<IncompleteArrayType>(AT) && !Options.IsStructField()) { 7633 // Incomplete arrays are encoded as a pointer to the array element. 7634 S += '^'; 7635 7636 getObjCEncodingForTypeImpl( 7637 AT->getElementType(), S, 7638 Options.keepingOnly(ObjCEncOptions().setExpandStructures()), FD); 7639 } else { 7640 S += '['; 7641 7642 if (const auto *CAT = dyn_cast<ConstantArrayType>(AT)) 7643 S += llvm::utostr(CAT->getSize().getZExtValue()); 7644 else { 7645 //Variable length arrays are encoded as a regular array with 0 elements. 7646 assert((isa<VariableArrayType>(AT) || isa<IncompleteArrayType>(AT)) && 7647 "Unknown array type!"); 7648 S += '0'; 7649 } 7650 7651 getObjCEncodingForTypeImpl( 7652 AT->getElementType(), S, 7653 Options.keepingOnly(ObjCEncOptions().setExpandStructures()), FD, 7654 NotEncodedT); 7655 S += ']'; 7656 } 7657 return; 7658 } 7659 7660 case Type::FunctionNoProto: 7661 case Type::FunctionProto: 7662 S += '?'; 7663 return; 7664 7665 case Type::Record: { 7666 RecordDecl *RDecl = cast<RecordType>(CT)->getDecl(); 7667 S += RDecl->isUnion() ? '(' : '{'; 7668 // Anonymous structures print as '?' 7669 if (const IdentifierInfo *II = RDecl->getIdentifier()) { 7670 S += II->getName(); 7671 if (const auto *Spec = dyn_cast<ClassTemplateSpecializationDecl>(RDecl)) { 7672 const TemplateArgumentList &TemplateArgs = Spec->getTemplateArgs(); 7673 llvm::raw_string_ostream OS(S); 7674 printTemplateArgumentList(OS, TemplateArgs.asArray(), 7675 getPrintingPolicy()); 7676 } 7677 } else { 7678 S += '?'; 7679 } 7680 if (Options.ExpandStructures()) { 7681 S += '='; 7682 if (!RDecl->isUnion()) { 7683 getObjCEncodingForStructureImpl(RDecl, S, FD, true, NotEncodedT); 7684 } else { 7685 for (const auto *Field : RDecl->fields()) { 7686 if (FD) { 7687 S += '"'; 7688 S += Field->getNameAsString(); 7689 S += '"'; 7690 } 7691 7692 // Special case bit-fields. 7693 if (Field->isBitField()) { 7694 getObjCEncodingForTypeImpl(Field->getType(), S, 7695 ObjCEncOptions().setExpandStructures(), 7696 Field); 7697 } else { 7698 QualType qt = Field->getType(); 7699 getLegacyIntegralTypeEncoding(qt); 7700 getObjCEncodingForTypeImpl( 7701 qt, S, 7702 ObjCEncOptions().setExpandStructures().setIsStructField(), FD, 7703 NotEncodedT); 7704 } 7705 } 7706 } 7707 } 7708 S += RDecl->isUnion() ? ')' : '}'; 7709 return; 7710 } 7711 7712 case Type::BlockPointer: { 7713 const auto *BT = T->castAs<BlockPointerType>(); 7714 S += "@?"; // Unlike a pointer-to-function, which is "^?". 7715 if (Options.EncodeBlockParameters()) { 7716 const auto *FT = BT->getPointeeType()->castAs<FunctionType>(); 7717 7718 S += '<'; 7719 // Block return type 7720 getObjCEncodingForTypeImpl(FT->getReturnType(), S, 7721 Options.forComponentType(), FD, NotEncodedT); 7722 // Block self 7723 S += "@?"; 7724 // Block parameters 7725 if (const auto *FPT = dyn_cast<FunctionProtoType>(FT)) { 7726 for (const auto &I : FPT->param_types()) 7727 getObjCEncodingForTypeImpl(I, S, Options.forComponentType(), FD, 7728 NotEncodedT); 7729 } 7730 S += '>'; 7731 } 7732 return; 7733 } 7734 7735 case Type::ObjCObject: { 7736 // hack to match legacy encoding of *id and *Class 7737 QualType Ty = getObjCObjectPointerType(CT); 7738 if (Ty->isObjCIdType()) { 7739 S += "{objc_object=}"; 7740 return; 7741 } 7742 else if (Ty->isObjCClassType()) { 7743 S += "{objc_class=}"; 7744 return; 7745 } 7746 // TODO: Double check to make sure this intentionally falls through. 7747 LLVM_FALLTHROUGH; 7748 } 7749 7750 case Type::ObjCInterface: { 7751 // Ignore protocol qualifiers when mangling at this level. 7752 // @encode(class_name) 7753 ObjCInterfaceDecl *OI = T->castAs<ObjCObjectType>()->getInterface(); 7754 S += '{'; 7755 S += OI->getObjCRuntimeNameAsString(); 7756 if (Options.ExpandStructures()) { 7757 S += '='; 7758 SmallVector<const ObjCIvarDecl*, 32> Ivars; 7759 DeepCollectObjCIvars(OI, true, Ivars); 7760 for (unsigned i = 0, e = Ivars.size(); i != e; ++i) { 7761 const FieldDecl *Field = Ivars[i]; 7762 if (Field->isBitField()) 7763 getObjCEncodingForTypeImpl(Field->getType(), S, 7764 ObjCEncOptions().setExpandStructures(), 7765 Field); 7766 else 7767 getObjCEncodingForTypeImpl(Field->getType(), S, 7768 ObjCEncOptions().setExpandStructures(), FD, 7769 NotEncodedT); 7770 } 7771 } 7772 S += '}'; 7773 return; 7774 } 7775 7776 case Type::ObjCObjectPointer: { 7777 const auto *OPT = T->castAs<ObjCObjectPointerType>(); 7778 if (OPT->isObjCIdType()) { 7779 S += '@'; 7780 return; 7781 } 7782 7783 if (OPT->isObjCClassType() || OPT->isObjCQualifiedClassType()) { 7784 // FIXME: Consider if we need to output qualifiers for 'Class<p>'. 7785 // Since this is a binary compatibility issue, need to consult with 7786 // runtime folks. Fortunately, this is a *very* obscure construct. 7787 S += '#'; 7788 return; 7789 } 7790 7791 if (OPT->isObjCQualifiedIdType()) { 7792 getObjCEncodingForTypeImpl( 7793 getObjCIdType(), S, 7794 Options.keepingOnly(ObjCEncOptions() 7795 .setExpandPointedToStructures() 7796 .setExpandStructures()), 7797 FD); 7798 if (FD || Options.EncodingProperty() || Options.EncodeClassNames()) { 7799 // Note that we do extended encoding of protocol qualifier list 7800 // Only when doing ivar or property encoding. 7801 S += '"'; 7802 for (const auto *I : OPT->quals()) { 7803 S += '<'; 7804 S += I->getObjCRuntimeNameAsString(); 7805 S += '>'; 7806 } 7807 S += '"'; 7808 } 7809 return; 7810 } 7811 7812 S += '@'; 7813 if (OPT->getInterfaceDecl() && 7814 (FD || Options.EncodingProperty() || Options.EncodeClassNames())) { 7815 S += '"'; 7816 S += OPT->getInterfaceDecl()->getObjCRuntimeNameAsString(); 7817 for (const auto *I : OPT->quals()) { 7818 S += '<'; 7819 S += I->getObjCRuntimeNameAsString(); 7820 S += '>'; 7821 } 7822 S += '"'; 7823 } 7824 return; 7825 } 7826 7827 // gcc just blithely ignores member pointers. 7828 // FIXME: we should do better than that. 'M' is available. 7829 case Type::MemberPointer: 7830 // This matches gcc's encoding, even though technically it is insufficient. 7831 //FIXME. We should do a better job than gcc. 7832 case Type::Vector: 7833 case Type::ExtVector: 7834 // Until we have a coherent encoding of these three types, issue warning. 7835 if (NotEncodedT) 7836 *NotEncodedT = T; 7837 return; 7838 7839 case Type::ConstantMatrix: 7840 if (NotEncodedT) 7841 *NotEncodedT = T; 7842 return; 7843 7844 // We could see an undeduced auto type here during error recovery. 7845 // Just ignore it. 7846 case Type::Auto: 7847 case Type::DeducedTemplateSpecialization: 7848 return; 7849 7850 case Type::Pipe: 7851 case Type::ExtInt: 7852 #define ABSTRACT_TYPE(KIND, BASE) 7853 #define TYPE(KIND, BASE) 7854 #define DEPENDENT_TYPE(KIND, BASE) \ 7855 case Type::KIND: 7856 #define NON_CANONICAL_TYPE(KIND, BASE) \ 7857 case Type::KIND: 7858 #define NON_CANONICAL_UNLESS_DEPENDENT_TYPE(KIND, BASE) \ 7859 case Type::KIND: 7860 #include "clang/AST/TypeNodes.inc" 7861 llvm_unreachable("@encode for dependent type!"); 7862 } 7863 llvm_unreachable("bad type kind!"); 7864 } 7865 7866 void ASTContext::getObjCEncodingForStructureImpl(RecordDecl *RDecl, 7867 std::string &S, 7868 const FieldDecl *FD, 7869 bool includeVBases, 7870 QualType *NotEncodedT) const { 7871 assert(RDecl && "Expected non-null RecordDecl"); 7872 assert(!RDecl->isUnion() && "Should not be called for unions"); 7873 if (!RDecl->getDefinition() || RDecl->getDefinition()->isInvalidDecl()) 7874 return; 7875 7876 const auto *CXXRec = dyn_cast<CXXRecordDecl>(RDecl); 7877 std::multimap<uint64_t, NamedDecl *> FieldOrBaseOffsets; 7878 const ASTRecordLayout &layout = getASTRecordLayout(RDecl); 7879 7880 if (CXXRec) { 7881 for (const auto &BI : CXXRec->bases()) { 7882 if (!BI.isVirtual()) { 7883 CXXRecordDecl *base = BI.getType()->getAsCXXRecordDecl(); 7884 if (base->isEmpty()) 7885 continue; 7886 uint64_t offs = toBits(layout.getBaseClassOffset(base)); 7887 FieldOrBaseOffsets.insert(FieldOrBaseOffsets.upper_bound(offs), 7888 std::make_pair(offs, base)); 7889 } 7890 } 7891 } 7892 7893 unsigned i = 0; 7894 for (FieldDecl *Field : RDecl->fields()) { 7895 if (!Field->isZeroLengthBitField(*this) && Field->isZeroSize(*this)) 7896 continue; 7897 uint64_t offs = layout.getFieldOffset(i); 7898 FieldOrBaseOffsets.insert(FieldOrBaseOffsets.upper_bound(offs), 7899 std::make_pair(offs, Field)); 7900 ++i; 7901 } 7902 7903 if (CXXRec && includeVBases) { 7904 for (const auto &BI : CXXRec->vbases()) { 7905 CXXRecordDecl *base = BI.getType()->getAsCXXRecordDecl(); 7906 if (base->isEmpty()) 7907 continue; 7908 uint64_t offs = toBits(layout.getVBaseClassOffset(base)); 7909 if (offs >= uint64_t(toBits(layout.getNonVirtualSize())) && 7910 FieldOrBaseOffsets.find(offs) == FieldOrBaseOffsets.end()) 7911 FieldOrBaseOffsets.insert(FieldOrBaseOffsets.end(), 7912 std::make_pair(offs, base)); 7913 } 7914 } 7915 7916 CharUnits size; 7917 if (CXXRec) { 7918 size = includeVBases ? layout.getSize() : layout.getNonVirtualSize(); 7919 } else { 7920 size = layout.getSize(); 7921 } 7922 7923 #ifndef NDEBUG 7924 uint64_t CurOffs = 0; 7925 #endif 7926 std::multimap<uint64_t, NamedDecl *>::iterator 7927 CurLayObj = FieldOrBaseOffsets.begin(); 7928 7929 if (CXXRec && CXXRec->isDynamicClass() && 7930 (CurLayObj == FieldOrBaseOffsets.end() || CurLayObj->first != 0)) { 7931 if (FD) { 7932 S += "\"_vptr$"; 7933 std::string recname = CXXRec->getNameAsString(); 7934 if (recname.empty()) recname = "?"; 7935 S += recname; 7936 S += '"'; 7937 } 7938 S += "^^?"; 7939 #ifndef NDEBUG 7940 CurOffs += getTypeSize(VoidPtrTy); 7941 #endif 7942 } 7943 7944 if (!RDecl->hasFlexibleArrayMember()) { 7945 // Mark the end of the structure. 7946 uint64_t offs = toBits(size); 7947 FieldOrBaseOffsets.insert(FieldOrBaseOffsets.upper_bound(offs), 7948 std::make_pair(offs, nullptr)); 7949 } 7950 7951 for (; CurLayObj != FieldOrBaseOffsets.end(); ++CurLayObj) { 7952 #ifndef NDEBUG 7953 assert(CurOffs <= CurLayObj->first); 7954 if (CurOffs < CurLayObj->first) { 7955 uint64_t padding = CurLayObj->first - CurOffs; 7956 // FIXME: There doesn't seem to be a way to indicate in the encoding that 7957 // packing/alignment of members is different that normal, in which case 7958 // the encoding will be out-of-sync with the real layout. 7959 // If the runtime switches to just consider the size of types without 7960 // taking into account alignment, we could make padding explicit in the 7961 // encoding (e.g. using arrays of chars). The encoding strings would be 7962 // longer then though. 7963 CurOffs += padding; 7964 } 7965 #endif 7966 7967 NamedDecl *dcl = CurLayObj->second; 7968 if (!dcl) 7969 break; // reached end of structure. 7970 7971 if (auto *base = dyn_cast<CXXRecordDecl>(dcl)) { 7972 // We expand the bases without their virtual bases since those are going 7973 // in the initial structure. Note that this differs from gcc which 7974 // expands virtual bases each time one is encountered in the hierarchy, 7975 // making the encoding type bigger than it really is. 7976 getObjCEncodingForStructureImpl(base, S, FD, /*includeVBases*/false, 7977 NotEncodedT); 7978 assert(!base->isEmpty()); 7979 #ifndef NDEBUG 7980 CurOffs += toBits(getASTRecordLayout(base).getNonVirtualSize()); 7981 #endif 7982 } else { 7983 const auto *field = cast<FieldDecl>(dcl); 7984 if (FD) { 7985 S += '"'; 7986 S += field->getNameAsString(); 7987 S += '"'; 7988 } 7989 7990 if (field->isBitField()) { 7991 EncodeBitField(this, S, field->getType(), field); 7992 #ifndef NDEBUG 7993 CurOffs += field->getBitWidthValue(*this); 7994 #endif 7995 } else { 7996 QualType qt = field->getType(); 7997 getLegacyIntegralTypeEncoding(qt); 7998 getObjCEncodingForTypeImpl( 7999 qt, S, ObjCEncOptions().setExpandStructures().setIsStructField(), 8000 FD, NotEncodedT); 8001 #ifndef NDEBUG 8002 CurOffs += getTypeSize(field->getType()); 8003 #endif 8004 } 8005 } 8006 } 8007 } 8008 8009 void ASTContext::getObjCEncodingForTypeQualifier(Decl::ObjCDeclQualifier QT, 8010 std::string& S) const { 8011 if (QT & Decl::OBJC_TQ_In) 8012 S += 'n'; 8013 if (QT & Decl::OBJC_TQ_Inout) 8014 S += 'N'; 8015 if (QT & Decl::OBJC_TQ_Out) 8016 S += 'o'; 8017 if (QT & Decl::OBJC_TQ_Bycopy) 8018 S += 'O'; 8019 if (QT & Decl::OBJC_TQ_Byref) 8020 S += 'R'; 8021 if (QT & Decl::OBJC_TQ_Oneway) 8022 S += 'V'; 8023 } 8024 8025 TypedefDecl *ASTContext::getObjCIdDecl() const { 8026 if (!ObjCIdDecl) { 8027 QualType T = getObjCObjectType(ObjCBuiltinIdTy, {}, {}); 8028 T = getObjCObjectPointerType(T); 8029 ObjCIdDecl = buildImplicitTypedef(T, "id"); 8030 } 8031 return ObjCIdDecl; 8032 } 8033 8034 TypedefDecl *ASTContext::getObjCSelDecl() const { 8035 if (!ObjCSelDecl) { 8036 QualType T = getPointerType(ObjCBuiltinSelTy); 8037 ObjCSelDecl = buildImplicitTypedef(T, "SEL"); 8038 } 8039 return ObjCSelDecl; 8040 } 8041 8042 TypedefDecl *ASTContext::getObjCClassDecl() const { 8043 if (!ObjCClassDecl) { 8044 QualType T = getObjCObjectType(ObjCBuiltinClassTy, {}, {}); 8045 T = getObjCObjectPointerType(T); 8046 ObjCClassDecl = buildImplicitTypedef(T, "Class"); 8047 } 8048 return ObjCClassDecl; 8049 } 8050 8051 ObjCInterfaceDecl *ASTContext::getObjCProtocolDecl() const { 8052 if (!ObjCProtocolClassDecl) { 8053 ObjCProtocolClassDecl 8054 = ObjCInterfaceDecl::Create(*this, getTranslationUnitDecl(), 8055 SourceLocation(), 8056 &Idents.get("Protocol"), 8057 /*typeParamList=*/nullptr, 8058 /*PrevDecl=*/nullptr, 8059 SourceLocation(), true); 8060 } 8061 8062 return ObjCProtocolClassDecl; 8063 } 8064 8065 //===----------------------------------------------------------------------===// 8066 // __builtin_va_list Construction Functions 8067 //===----------------------------------------------------------------------===// 8068 8069 static TypedefDecl *CreateCharPtrNamedVaListDecl(const ASTContext *Context, 8070 StringRef Name) { 8071 // typedef char* __builtin[_ms]_va_list; 8072 QualType T = Context->getPointerType(Context->CharTy); 8073 return Context->buildImplicitTypedef(T, Name); 8074 } 8075 8076 static TypedefDecl *CreateMSVaListDecl(const ASTContext *Context) { 8077 return CreateCharPtrNamedVaListDecl(Context, "__builtin_ms_va_list"); 8078 } 8079 8080 static TypedefDecl *CreateCharPtrBuiltinVaListDecl(const ASTContext *Context) { 8081 return CreateCharPtrNamedVaListDecl(Context, "__builtin_va_list"); 8082 } 8083 8084 static TypedefDecl *CreateVoidPtrBuiltinVaListDecl(const ASTContext *Context) { 8085 // typedef void* __builtin_va_list; 8086 QualType T = Context->getPointerType(Context->VoidTy); 8087 return Context->buildImplicitTypedef(T, "__builtin_va_list"); 8088 } 8089 8090 static TypedefDecl * 8091 CreateAArch64ABIBuiltinVaListDecl(const ASTContext *Context) { 8092 RecordDecl *VaListTagDecl = Context->buildImplicitRecord("__va_list"); 8093 // namespace std { struct __va_list { 8094 // Note that we create the namespace even in C. This is intentional so that 8095 // the type is consistent between C and C++, which is important in cases where 8096 // the types need to match between translation units (e.g. with 8097 // -fsanitize=cfi-icall). Ideally we wouldn't have created this namespace at 8098 // all, but it's now part of the ABI (e.g. in mangled names), so we can't 8099 // change it. 8100 auto *NS = NamespaceDecl::Create( 8101 const_cast<ASTContext &>(*Context), Context->getTranslationUnitDecl(), 8102 /*Inline*/ false, SourceLocation(), SourceLocation(), 8103 &Context->Idents.get("std"), 8104 /*PrevDecl*/ nullptr); 8105 NS->setImplicit(); 8106 VaListTagDecl->setDeclContext(NS); 8107 8108 VaListTagDecl->startDefinition(); 8109 8110 const size_t NumFields = 5; 8111 QualType FieldTypes[NumFields]; 8112 const char *FieldNames[NumFields]; 8113 8114 // void *__stack; 8115 FieldTypes[0] = Context->getPointerType(Context->VoidTy); 8116 FieldNames[0] = "__stack"; 8117 8118 // void *__gr_top; 8119 FieldTypes[1] = Context->getPointerType(Context->VoidTy); 8120 FieldNames[1] = "__gr_top"; 8121 8122 // void *__vr_top; 8123 FieldTypes[2] = Context->getPointerType(Context->VoidTy); 8124 FieldNames[2] = "__vr_top"; 8125 8126 // int __gr_offs; 8127 FieldTypes[3] = Context->IntTy; 8128 FieldNames[3] = "__gr_offs"; 8129 8130 // int __vr_offs; 8131 FieldTypes[4] = Context->IntTy; 8132 FieldNames[4] = "__vr_offs"; 8133 8134 // Create fields 8135 for (unsigned i = 0; i < NumFields; ++i) { 8136 FieldDecl *Field = FieldDecl::Create(const_cast<ASTContext &>(*Context), 8137 VaListTagDecl, 8138 SourceLocation(), 8139 SourceLocation(), 8140 &Context->Idents.get(FieldNames[i]), 8141 FieldTypes[i], /*TInfo=*/nullptr, 8142 /*BitWidth=*/nullptr, 8143 /*Mutable=*/false, 8144 ICIS_NoInit); 8145 Field->setAccess(AS_public); 8146 VaListTagDecl->addDecl(Field); 8147 } 8148 VaListTagDecl->completeDefinition(); 8149 Context->VaListTagDecl = VaListTagDecl; 8150 QualType VaListTagType = Context->getRecordType(VaListTagDecl); 8151 8152 // } __builtin_va_list; 8153 return Context->buildImplicitTypedef(VaListTagType, "__builtin_va_list"); 8154 } 8155 8156 static TypedefDecl *CreatePowerABIBuiltinVaListDecl(const ASTContext *Context) { 8157 // typedef struct __va_list_tag { 8158 RecordDecl *VaListTagDecl; 8159 8160 VaListTagDecl = Context->buildImplicitRecord("__va_list_tag"); 8161 VaListTagDecl->startDefinition(); 8162 8163 const size_t NumFields = 5; 8164 QualType FieldTypes[NumFields]; 8165 const char *FieldNames[NumFields]; 8166 8167 // unsigned char gpr; 8168 FieldTypes[0] = Context->UnsignedCharTy; 8169 FieldNames[0] = "gpr"; 8170 8171 // unsigned char fpr; 8172 FieldTypes[1] = Context->UnsignedCharTy; 8173 FieldNames[1] = "fpr"; 8174 8175 // unsigned short reserved; 8176 FieldTypes[2] = Context->UnsignedShortTy; 8177 FieldNames[2] = "reserved"; 8178 8179 // void* overflow_arg_area; 8180 FieldTypes[3] = Context->getPointerType(Context->VoidTy); 8181 FieldNames[3] = "overflow_arg_area"; 8182 8183 // void* reg_save_area; 8184 FieldTypes[4] = Context->getPointerType(Context->VoidTy); 8185 FieldNames[4] = "reg_save_area"; 8186 8187 // Create fields 8188 for (unsigned i = 0; i < NumFields; ++i) { 8189 FieldDecl *Field = FieldDecl::Create(*Context, VaListTagDecl, 8190 SourceLocation(), 8191 SourceLocation(), 8192 &Context->Idents.get(FieldNames[i]), 8193 FieldTypes[i], /*TInfo=*/nullptr, 8194 /*BitWidth=*/nullptr, 8195 /*Mutable=*/false, 8196 ICIS_NoInit); 8197 Field->setAccess(AS_public); 8198 VaListTagDecl->addDecl(Field); 8199 } 8200 VaListTagDecl->completeDefinition(); 8201 Context->VaListTagDecl = VaListTagDecl; 8202 QualType VaListTagType = Context->getRecordType(VaListTagDecl); 8203 8204 // } __va_list_tag; 8205 TypedefDecl *VaListTagTypedefDecl = 8206 Context->buildImplicitTypedef(VaListTagType, "__va_list_tag"); 8207 8208 QualType VaListTagTypedefType = 8209 Context->getTypedefType(VaListTagTypedefDecl); 8210 8211 // typedef __va_list_tag __builtin_va_list[1]; 8212 llvm::APInt Size(Context->getTypeSize(Context->getSizeType()), 1); 8213 QualType VaListTagArrayType 8214 = Context->getConstantArrayType(VaListTagTypedefType, 8215 Size, nullptr, ArrayType::Normal, 0); 8216 return Context->buildImplicitTypedef(VaListTagArrayType, "__builtin_va_list"); 8217 } 8218 8219 static TypedefDecl * 8220 CreateX86_64ABIBuiltinVaListDecl(const ASTContext *Context) { 8221 // struct __va_list_tag { 8222 RecordDecl *VaListTagDecl; 8223 VaListTagDecl = Context->buildImplicitRecord("__va_list_tag"); 8224 VaListTagDecl->startDefinition(); 8225 8226 const size_t NumFields = 4; 8227 QualType FieldTypes[NumFields]; 8228 const char *FieldNames[NumFields]; 8229 8230 // unsigned gp_offset; 8231 FieldTypes[0] = Context->UnsignedIntTy; 8232 FieldNames[0] = "gp_offset"; 8233 8234 // unsigned fp_offset; 8235 FieldTypes[1] = Context->UnsignedIntTy; 8236 FieldNames[1] = "fp_offset"; 8237 8238 // void* overflow_arg_area; 8239 FieldTypes[2] = Context->getPointerType(Context->VoidTy); 8240 FieldNames[2] = "overflow_arg_area"; 8241 8242 // void* reg_save_area; 8243 FieldTypes[3] = Context->getPointerType(Context->VoidTy); 8244 FieldNames[3] = "reg_save_area"; 8245 8246 // Create fields 8247 for (unsigned i = 0; i < NumFields; ++i) { 8248 FieldDecl *Field = FieldDecl::Create(const_cast<ASTContext &>(*Context), 8249 VaListTagDecl, 8250 SourceLocation(), 8251 SourceLocation(), 8252 &Context->Idents.get(FieldNames[i]), 8253 FieldTypes[i], /*TInfo=*/nullptr, 8254 /*BitWidth=*/nullptr, 8255 /*Mutable=*/false, 8256 ICIS_NoInit); 8257 Field->setAccess(AS_public); 8258 VaListTagDecl->addDecl(Field); 8259 } 8260 VaListTagDecl->completeDefinition(); 8261 Context->VaListTagDecl = VaListTagDecl; 8262 QualType VaListTagType = Context->getRecordType(VaListTagDecl); 8263 8264 // }; 8265 8266 // typedef struct __va_list_tag __builtin_va_list[1]; 8267 llvm::APInt Size(Context->getTypeSize(Context->getSizeType()), 1); 8268 QualType VaListTagArrayType = Context->getConstantArrayType( 8269 VaListTagType, Size, nullptr, ArrayType::Normal, 0); 8270 return Context->buildImplicitTypedef(VaListTagArrayType, "__builtin_va_list"); 8271 } 8272 8273 static TypedefDecl *CreatePNaClABIBuiltinVaListDecl(const ASTContext *Context) { 8274 // typedef int __builtin_va_list[4]; 8275 llvm::APInt Size(Context->getTypeSize(Context->getSizeType()), 4); 8276 QualType IntArrayType = Context->getConstantArrayType( 8277 Context->IntTy, Size, nullptr, ArrayType::Normal, 0); 8278 return Context->buildImplicitTypedef(IntArrayType, "__builtin_va_list"); 8279 } 8280 8281 static TypedefDecl * 8282 CreateAAPCSABIBuiltinVaListDecl(const ASTContext *Context) { 8283 // struct __va_list 8284 RecordDecl *VaListDecl = Context->buildImplicitRecord("__va_list"); 8285 if (Context->getLangOpts().CPlusPlus) { 8286 // namespace std { struct __va_list { 8287 NamespaceDecl *NS; 8288 NS = NamespaceDecl::Create(const_cast<ASTContext &>(*Context), 8289 Context->getTranslationUnitDecl(), 8290 /*Inline*/false, SourceLocation(), 8291 SourceLocation(), &Context->Idents.get("std"), 8292 /*PrevDecl*/ nullptr); 8293 NS->setImplicit(); 8294 VaListDecl->setDeclContext(NS); 8295 } 8296 8297 VaListDecl->startDefinition(); 8298 8299 // void * __ap; 8300 FieldDecl *Field = FieldDecl::Create(const_cast<ASTContext &>(*Context), 8301 VaListDecl, 8302 SourceLocation(), 8303 SourceLocation(), 8304 &Context->Idents.get("__ap"), 8305 Context->getPointerType(Context->VoidTy), 8306 /*TInfo=*/nullptr, 8307 /*BitWidth=*/nullptr, 8308 /*Mutable=*/false, 8309 ICIS_NoInit); 8310 Field->setAccess(AS_public); 8311 VaListDecl->addDecl(Field); 8312 8313 // }; 8314 VaListDecl->completeDefinition(); 8315 Context->VaListTagDecl = VaListDecl; 8316 8317 // typedef struct __va_list __builtin_va_list; 8318 QualType T = Context->getRecordType(VaListDecl); 8319 return Context->buildImplicitTypedef(T, "__builtin_va_list"); 8320 } 8321 8322 static TypedefDecl * 8323 CreateSystemZBuiltinVaListDecl(const ASTContext *Context) { 8324 // struct __va_list_tag { 8325 RecordDecl *VaListTagDecl; 8326 VaListTagDecl = Context->buildImplicitRecord("__va_list_tag"); 8327 VaListTagDecl->startDefinition(); 8328 8329 const size_t NumFields = 4; 8330 QualType FieldTypes[NumFields]; 8331 const char *FieldNames[NumFields]; 8332 8333 // long __gpr; 8334 FieldTypes[0] = Context->LongTy; 8335 FieldNames[0] = "__gpr"; 8336 8337 // long __fpr; 8338 FieldTypes[1] = Context->LongTy; 8339 FieldNames[1] = "__fpr"; 8340 8341 // void *__overflow_arg_area; 8342 FieldTypes[2] = Context->getPointerType(Context->VoidTy); 8343 FieldNames[2] = "__overflow_arg_area"; 8344 8345 // void *__reg_save_area; 8346 FieldTypes[3] = Context->getPointerType(Context->VoidTy); 8347 FieldNames[3] = "__reg_save_area"; 8348 8349 // Create fields 8350 for (unsigned i = 0; i < NumFields; ++i) { 8351 FieldDecl *Field = FieldDecl::Create(const_cast<ASTContext &>(*Context), 8352 VaListTagDecl, 8353 SourceLocation(), 8354 SourceLocation(), 8355 &Context->Idents.get(FieldNames[i]), 8356 FieldTypes[i], /*TInfo=*/nullptr, 8357 /*BitWidth=*/nullptr, 8358 /*Mutable=*/false, 8359 ICIS_NoInit); 8360 Field->setAccess(AS_public); 8361 VaListTagDecl->addDecl(Field); 8362 } 8363 VaListTagDecl->completeDefinition(); 8364 Context->VaListTagDecl = VaListTagDecl; 8365 QualType VaListTagType = Context->getRecordType(VaListTagDecl); 8366 8367 // }; 8368 8369 // typedef __va_list_tag __builtin_va_list[1]; 8370 llvm::APInt Size(Context->getTypeSize(Context->getSizeType()), 1); 8371 QualType VaListTagArrayType = Context->getConstantArrayType( 8372 VaListTagType, Size, nullptr, ArrayType::Normal, 0); 8373 8374 return Context->buildImplicitTypedef(VaListTagArrayType, "__builtin_va_list"); 8375 } 8376 8377 static TypedefDecl *CreateHexagonBuiltinVaListDecl(const ASTContext *Context) { 8378 // typedef struct __va_list_tag { 8379 RecordDecl *VaListTagDecl; 8380 VaListTagDecl = Context->buildImplicitRecord("__va_list_tag"); 8381 VaListTagDecl->startDefinition(); 8382 8383 const size_t NumFields = 3; 8384 QualType FieldTypes[NumFields]; 8385 const char *FieldNames[NumFields]; 8386 8387 // void *CurrentSavedRegisterArea; 8388 FieldTypes[0] = Context->getPointerType(Context->VoidTy); 8389 FieldNames[0] = "__current_saved_reg_area_pointer"; 8390 8391 // void *SavedRegAreaEnd; 8392 FieldTypes[1] = Context->getPointerType(Context->VoidTy); 8393 FieldNames[1] = "__saved_reg_area_end_pointer"; 8394 8395 // void *OverflowArea; 8396 FieldTypes[2] = Context->getPointerType(Context->VoidTy); 8397 FieldNames[2] = "__overflow_area_pointer"; 8398 8399 // Create fields 8400 for (unsigned i = 0; i < NumFields; ++i) { 8401 FieldDecl *Field = FieldDecl::Create( 8402 const_cast<ASTContext &>(*Context), VaListTagDecl, SourceLocation(), 8403 SourceLocation(), &Context->Idents.get(FieldNames[i]), FieldTypes[i], 8404 /*TInfo=*/0, 8405 /*BitWidth=*/0, 8406 /*Mutable=*/false, ICIS_NoInit); 8407 Field->setAccess(AS_public); 8408 VaListTagDecl->addDecl(Field); 8409 } 8410 VaListTagDecl->completeDefinition(); 8411 Context->VaListTagDecl = VaListTagDecl; 8412 QualType VaListTagType = Context->getRecordType(VaListTagDecl); 8413 8414 // } __va_list_tag; 8415 TypedefDecl *VaListTagTypedefDecl = 8416 Context->buildImplicitTypedef(VaListTagType, "__va_list_tag"); 8417 8418 QualType VaListTagTypedefType = Context->getTypedefType(VaListTagTypedefDecl); 8419 8420 // typedef __va_list_tag __builtin_va_list[1]; 8421 llvm::APInt Size(Context->getTypeSize(Context->getSizeType()), 1); 8422 QualType VaListTagArrayType = Context->getConstantArrayType( 8423 VaListTagTypedefType, Size, nullptr, ArrayType::Normal, 0); 8424 8425 return Context->buildImplicitTypedef(VaListTagArrayType, "__builtin_va_list"); 8426 } 8427 8428 static TypedefDecl *CreateVaListDecl(const ASTContext *Context, 8429 TargetInfo::BuiltinVaListKind Kind) { 8430 switch (Kind) { 8431 case TargetInfo::CharPtrBuiltinVaList: 8432 return CreateCharPtrBuiltinVaListDecl(Context); 8433 case TargetInfo::VoidPtrBuiltinVaList: 8434 return CreateVoidPtrBuiltinVaListDecl(Context); 8435 case TargetInfo::AArch64ABIBuiltinVaList: 8436 return CreateAArch64ABIBuiltinVaListDecl(Context); 8437 case TargetInfo::PowerABIBuiltinVaList: 8438 return CreatePowerABIBuiltinVaListDecl(Context); 8439 case TargetInfo::X86_64ABIBuiltinVaList: 8440 return CreateX86_64ABIBuiltinVaListDecl(Context); 8441 case TargetInfo::PNaClABIBuiltinVaList: 8442 return CreatePNaClABIBuiltinVaListDecl(Context); 8443 case TargetInfo::AAPCSABIBuiltinVaList: 8444 return CreateAAPCSABIBuiltinVaListDecl(Context); 8445 case TargetInfo::SystemZBuiltinVaList: 8446 return CreateSystemZBuiltinVaListDecl(Context); 8447 case TargetInfo::HexagonBuiltinVaList: 8448 return CreateHexagonBuiltinVaListDecl(Context); 8449 } 8450 8451 llvm_unreachable("Unhandled __builtin_va_list type kind"); 8452 } 8453 8454 TypedefDecl *ASTContext::getBuiltinVaListDecl() const { 8455 if (!BuiltinVaListDecl) { 8456 BuiltinVaListDecl = CreateVaListDecl(this, Target->getBuiltinVaListKind()); 8457 assert(BuiltinVaListDecl->isImplicit()); 8458 } 8459 8460 return BuiltinVaListDecl; 8461 } 8462 8463 Decl *ASTContext::getVaListTagDecl() const { 8464 // Force the creation of VaListTagDecl by building the __builtin_va_list 8465 // declaration. 8466 if (!VaListTagDecl) 8467 (void)getBuiltinVaListDecl(); 8468 8469 return VaListTagDecl; 8470 } 8471 8472 TypedefDecl *ASTContext::getBuiltinMSVaListDecl() const { 8473 if (!BuiltinMSVaListDecl) 8474 BuiltinMSVaListDecl = CreateMSVaListDecl(this); 8475 8476 return BuiltinMSVaListDecl; 8477 } 8478 8479 bool ASTContext::canBuiltinBeRedeclared(const FunctionDecl *FD) const { 8480 return BuiltinInfo.canBeRedeclared(FD->getBuiltinID()); 8481 } 8482 8483 void ASTContext::setObjCConstantStringInterface(ObjCInterfaceDecl *Decl) { 8484 assert(ObjCConstantStringType.isNull() && 8485 "'NSConstantString' type already set!"); 8486 8487 ObjCConstantStringType = getObjCInterfaceType(Decl); 8488 } 8489 8490 /// Retrieve the template name that corresponds to a non-empty 8491 /// lookup. 8492 TemplateName 8493 ASTContext::getOverloadedTemplateName(UnresolvedSetIterator Begin, 8494 UnresolvedSetIterator End) const { 8495 unsigned size = End - Begin; 8496 assert(size > 1 && "set is not overloaded!"); 8497 8498 void *memory = Allocate(sizeof(OverloadedTemplateStorage) + 8499 size * sizeof(FunctionTemplateDecl*)); 8500 auto *OT = new (memory) OverloadedTemplateStorage(size); 8501 8502 NamedDecl **Storage = OT->getStorage(); 8503 for (UnresolvedSetIterator I = Begin; I != End; ++I) { 8504 NamedDecl *D = *I; 8505 assert(isa<FunctionTemplateDecl>(D) || 8506 isa<UnresolvedUsingValueDecl>(D) || 8507 (isa<UsingShadowDecl>(D) && 8508 isa<FunctionTemplateDecl>(D->getUnderlyingDecl()))); 8509 *Storage++ = D; 8510 } 8511 8512 return TemplateName(OT); 8513 } 8514 8515 /// Retrieve a template name representing an unqualified-id that has been 8516 /// assumed to name a template for ADL purposes. 8517 TemplateName ASTContext::getAssumedTemplateName(DeclarationName Name) const { 8518 auto *OT = new (*this) AssumedTemplateStorage(Name); 8519 return TemplateName(OT); 8520 } 8521 8522 /// Retrieve the template name that represents a qualified 8523 /// template name such as \c std::vector. 8524 TemplateName 8525 ASTContext::getQualifiedTemplateName(NestedNameSpecifier *NNS, 8526 bool TemplateKeyword, 8527 TemplateDecl *Template) const { 8528 assert(NNS && "Missing nested-name-specifier in qualified template name"); 8529 8530 // FIXME: Canonicalization? 8531 llvm::FoldingSetNodeID ID; 8532 QualifiedTemplateName::Profile(ID, NNS, TemplateKeyword, Template); 8533 8534 void *InsertPos = nullptr; 8535 QualifiedTemplateName *QTN = 8536 QualifiedTemplateNames.FindNodeOrInsertPos(ID, InsertPos); 8537 if (!QTN) { 8538 QTN = new (*this, alignof(QualifiedTemplateName)) 8539 QualifiedTemplateName(NNS, TemplateKeyword, Template); 8540 QualifiedTemplateNames.InsertNode(QTN, InsertPos); 8541 } 8542 8543 return TemplateName(QTN); 8544 } 8545 8546 /// Retrieve the template name that represents a dependent 8547 /// template name such as \c MetaFun::template apply. 8548 TemplateName 8549 ASTContext::getDependentTemplateName(NestedNameSpecifier *NNS, 8550 const IdentifierInfo *Name) const { 8551 assert((!NNS || NNS->isDependent()) && 8552 "Nested name specifier must be dependent"); 8553 8554 llvm::FoldingSetNodeID ID; 8555 DependentTemplateName::Profile(ID, NNS, Name); 8556 8557 void *InsertPos = nullptr; 8558 DependentTemplateName *QTN = 8559 DependentTemplateNames.FindNodeOrInsertPos(ID, InsertPos); 8560 8561 if (QTN) 8562 return TemplateName(QTN); 8563 8564 NestedNameSpecifier *CanonNNS = getCanonicalNestedNameSpecifier(NNS); 8565 if (CanonNNS == NNS) { 8566 QTN = new (*this, alignof(DependentTemplateName)) 8567 DependentTemplateName(NNS, Name); 8568 } else { 8569 TemplateName Canon = getDependentTemplateName(CanonNNS, Name); 8570 QTN = new (*this, alignof(DependentTemplateName)) 8571 DependentTemplateName(NNS, Name, Canon); 8572 DependentTemplateName *CheckQTN = 8573 DependentTemplateNames.FindNodeOrInsertPos(ID, InsertPos); 8574 assert(!CheckQTN && "Dependent type name canonicalization broken"); 8575 (void)CheckQTN; 8576 } 8577 8578 DependentTemplateNames.InsertNode(QTN, InsertPos); 8579 return TemplateName(QTN); 8580 } 8581 8582 /// Retrieve the template name that represents a dependent 8583 /// template name such as \c MetaFun::template operator+. 8584 TemplateName 8585 ASTContext::getDependentTemplateName(NestedNameSpecifier *NNS, 8586 OverloadedOperatorKind Operator) const { 8587 assert((!NNS || NNS->isDependent()) && 8588 "Nested name specifier must be dependent"); 8589 8590 llvm::FoldingSetNodeID ID; 8591 DependentTemplateName::Profile(ID, NNS, Operator); 8592 8593 void *InsertPos = nullptr; 8594 DependentTemplateName *QTN 8595 = DependentTemplateNames.FindNodeOrInsertPos(ID, InsertPos); 8596 8597 if (QTN) 8598 return TemplateName(QTN); 8599 8600 NestedNameSpecifier *CanonNNS = getCanonicalNestedNameSpecifier(NNS); 8601 if (CanonNNS == NNS) { 8602 QTN = new (*this, alignof(DependentTemplateName)) 8603 DependentTemplateName(NNS, Operator); 8604 } else { 8605 TemplateName Canon = getDependentTemplateName(CanonNNS, Operator); 8606 QTN = new (*this, alignof(DependentTemplateName)) 8607 DependentTemplateName(NNS, Operator, Canon); 8608 8609 DependentTemplateName *CheckQTN 8610 = DependentTemplateNames.FindNodeOrInsertPos(ID, InsertPos); 8611 assert(!CheckQTN && "Dependent template name canonicalization broken"); 8612 (void)CheckQTN; 8613 } 8614 8615 DependentTemplateNames.InsertNode(QTN, InsertPos); 8616 return TemplateName(QTN); 8617 } 8618 8619 TemplateName 8620 ASTContext::getSubstTemplateTemplateParm(TemplateTemplateParmDecl *param, 8621 TemplateName replacement) const { 8622 llvm::FoldingSetNodeID ID; 8623 SubstTemplateTemplateParmStorage::Profile(ID, param, replacement); 8624 8625 void *insertPos = nullptr; 8626 SubstTemplateTemplateParmStorage *subst 8627 = SubstTemplateTemplateParms.FindNodeOrInsertPos(ID, insertPos); 8628 8629 if (!subst) { 8630 subst = new (*this) SubstTemplateTemplateParmStorage(param, replacement); 8631 SubstTemplateTemplateParms.InsertNode(subst, insertPos); 8632 } 8633 8634 return TemplateName(subst); 8635 } 8636 8637 TemplateName 8638 ASTContext::getSubstTemplateTemplateParmPack(TemplateTemplateParmDecl *Param, 8639 const TemplateArgument &ArgPack) const { 8640 auto &Self = const_cast<ASTContext &>(*this); 8641 llvm::FoldingSetNodeID ID; 8642 SubstTemplateTemplateParmPackStorage::Profile(ID, Self, Param, ArgPack); 8643 8644 void *InsertPos = nullptr; 8645 SubstTemplateTemplateParmPackStorage *Subst 8646 = SubstTemplateTemplateParmPacks.FindNodeOrInsertPos(ID, InsertPos); 8647 8648 if (!Subst) { 8649 Subst = new (*this) SubstTemplateTemplateParmPackStorage(Param, 8650 ArgPack.pack_size(), 8651 ArgPack.pack_begin()); 8652 SubstTemplateTemplateParmPacks.InsertNode(Subst, InsertPos); 8653 } 8654 8655 return TemplateName(Subst); 8656 } 8657 8658 /// getFromTargetType - Given one of the integer types provided by 8659 /// TargetInfo, produce the corresponding type. The unsigned @p Type 8660 /// is actually a value of type @c TargetInfo::IntType. 8661 CanQualType ASTContext::getFromTargetType(unsigned Type) const { 8662 switch (Type) { 8663 case TargetInfo::NoInt: return {}; 8664 case TargetInfo::SignedChar: return SignedCharTy; 8665 case TargetInfo::UnsignedChar: return UnsignedCharTy; 8666 case TargetInfo::SignedShort: return ShortTy; 8667 case TargetInfo::UnsignedShort: return UnsignedShortTy; 8668 case TargetInfo::SignedInt: return IntTy; 8669 case TargetInfo::UnsignedInt: return UnsignedIntTy; 8670 case TargetInfo::SignedLong: return LongTy; 8671 case TargetInfo::UnsignedLong: return UnsignedLongTy; 8672 case TargetInfo::SignedLongLong: return LongLongTy; 8673 case TargetInfo::UnsignedLongLong: return UnsignedLongLongTy; 8674 } 8675 8676 llvm_unreachable("Unhandled TargetInfo::IntType value"); 8677 } 8678 8679 //===----------------------------------------------------------------------===// 8680 // Type Predicates. 8681 //===----------------------------------------------------------------------===// 8682 8683 /// getObjCGCAttr - Returns one of GCNone, Weak or Strong objc's 8684 /// garbage collection attribute. 8685 /// 8686 Qualifiers::GC ASTContext::getObjCGCAttrKind(QualType Ty) const { 8687 if (getLangOpts().getGC() == LangOptions::NonGC) 8688 return Qualifiers::GCNone; 8689 8690 assert(getLangOpts().ObjC); 8691 Qualifiers::GC GCAttrs = Ty.getObjCGCAttr(); 8692 8693 // Default behaviour under objective-C's gc is for ObjC pointers 8694 // (or pointers to them) be treated as though they were declared 8695 // as __strong. 8696 if (GCAttrs == Qualifiers::GCNone) { 8697 if (Ty->isObjCObjectPointerType() || Ty->isBlockPointerType()) 8698 return Qualifiers::Strong; 8699 else if (Ty->isPointerType()) 8700 return getObjCGCAttrKind(Ty->castAs<PointerType>()->getPointeeType()); 8701 } else { 8702 // It's not valid to set GC attributes on anything that isn't a 8703 // pointer. 8704 #ifndef NDEBUG 8705 QualType CT = Ty->getCanonicalTypeInternal(); 8706 while (const auto *AT = dyn_cast<ArrayType>(CT)) 8707 CT = AT->getElementType(); 8708 assert(CT->isAnyPointerType() || CT->isBlockPointerType()); 8709 #endif 8710 } 8711 return GCAttrs; 8712 } 8713 8714 //===----------------------------------------------------------------------===// 8715 // Type Compatibility Testing 8716 //===----------------------------------------------------------------------===// 8717 8718 /// areCompatVectorTypes - Return true if the two specified vector types are 8719 /// compatible. 8720 static bool areCompatVectorTypes(const VectorType *LHS, 8721 const VectorType *RHS) { 8722 assert(LHS->isCanonicalUnqualified() && RHS->isCanonicalUnqualified()); 8723 return LHS->getElementType() == RHS->getElementType() && 8724 LHS->getNumElements() == RHS->getNumElements(); 8725 } 8726 8727 /// areCompatMatrixTypes - Return true if the two specified matrix types are 8728 /// compatible. 8729 static bool areCompatMatrixTypes(const ConstantMatrixType *LHS, 8730 const ConstantMatrixType *RHS) { 8731 assert(LHS->isCanonicalUnqualified() && RHS->isCanonicalUnqualified()); 8732 return LHS->getElementType() == RHS->getElementType() && 8733 LHS->getNumRows() == RHS->getNumRows() && 8734 LHS->getNumColumns() == RHS->getNumColumns(); 8735 } 8736 8737 bool ASTContext::areCompatibleVectorTypes(QualType FirstVec, 8738 QualType SecondVec) { 8739 assert(FirstVec->isVectorType() && "FirstVec should be a vector type"); 8740 assert(SecondVec->isVectorType() && "SecondVec should be a vector type"); 8741 8742 if (hasSameUnqualifiedType(FirstVec, SecondVec)) 8743 return true; 8744 8745 // Treat Neon vector types and most AltiVec vector types as if they are the 8746 // equivalent GCC vector types. 8747 const auto *First = FirstVec->castAs<VectorType>(); 8748 const auto *Second = SecondVec->castAs<VectorType>(); 8749 if (First->getNumElements() == Second->getNumElements() && 8750 hasSameType(First->getElementType(), Second->getElementType()) && 8751 First->getVectorKind() != VectorType::AltiVecPixel && 8752 First->getVectorKind() != VectorType::AltiVecBool && 8753 Second->getVectorKind() != VectorType::AltiVecPixel && 8754 Second->getVectorKind() != VectorType::AltiVecBool && 8755 First->getVectorKind() != VectorType::SveFixedLengthDataVector && 8756 First->getVectorKind() != VectorType::SveFixedLengthPredicateVector && 8757 Second->getVectorKind() != VectorType::SveFixedLengthDataVector && 8758 Second->getVectorKind() != VectorType::SveFixedLengthPredicateVector) 8759 return true; 8760 8761 return false; 8762 } 8763 8764 /// getSVETypeSize - Return SVE vector or predicate register size. 8765 static uint64_t getSVETypeSize(ASTContext &Context, const BuiltinType *Ty) { 8766 assert(Ty->isVLSTBuiltinType() && "Invalid SVE Type"); 8767 return Ty->getKind() == BuiltinType::SveBool 8768 ? (Context.getLangOpts().VScaleMin * 128) / Context.getCharWidth() 8769 : Context.getLangOpts().VScaleMin * 128; 8770 } 8771 8772 bool ASTContext::areCompatibleSveTypes(QualType FirstType, 8773 QualType SecondType) { 8774 assert(((FirstType->isSizelessBuiltinType() && SecondType->isVectorType()) || 8775 (FirstType->isVectorType() && SecondType->isSizelessBuiltinType())) && 8776 "Expected SVE builtin type and vector type!"); 8777 8778 auto IsValidCast = [this](QualType FirstType, QualType SecondType) { 8779 if (const auto *BT = FirstType->getAs<BuiltinType>()) { 8780 if (const auto *VT = SecondType->getAs<VectorType>()) { 8781 // Predicates have the same representation as uint8 so we also have to 8782 // check the kind to make these types incompatible. 8783 if (VT->getVectorKind() == VectorType::SveFixedLengthPredicateVector) 8784 return BT->getKind() == BuiltinType::SveBool; 8785 else if (VT->getVectorKind() == VectorType::SveFixedLengthDataVector) 8786 return VT->getElementType().getCanonicalType() == 8787 FirstType->getSveEltType(*this); 8788 else if (VT->getVectorKind() == VectorType::GenericVector) 8789 return getTypeSize(SecondType) == getSVETypeSize(*this, BT) && 8790 hasSameType(VT->getElementType(), 8791 getBuiltinVectorTypeInfo(BT).ElementType); 8792 } 8793 } 8794 return false; 8795 }; 8796 8797 return IsValidCast(FirstType, SecondType) || 8798 IsValidCast(SecondType, FirstType); 8799 } 8800 8801 bool ASTContext::areLaxCompatibleSveTypes(QualType FirstType, 8802 QualType SecondType) { 8803 assert(((FirstType->isSizelessBuiltinType() && SecondType->isVectorType()) || 8804 (FirstType->isVectorType() && SecondType->isSizelessBuiltinType())) && 8805 "Expected SVE builtin type and vector type!"); 8806 8807 auto IsLaxCompatible = [this](QualType FirstType, QualType SecondType) { 8808 const auto *BT = FirstType->getAs<BuiltinType>(); 8809 if (!BT) 8810 return false; 8811 8812 const auto *VecTy = SecondType->getAs<VectorType>(); 8813 if (VecTy && 8814 (VecTy->getVectorKind() == VectorType::SveFixedLengthDataVector || 8815 VecTy->getVectorKind() == VectorType::GenericVector)) { 8816 const LangOptions::LaxVectorConversionKind LVCKind = 8817 getLangOpts().getLaxVectorConversions(); 8818 8819 // Can not convert between sve predicates and sve vectors because of 8820 // different size. 8821 if (BT->getKind() == BuiltinType::SveBool && 8822 VecTy->getVectorKind() == VectorType::SveFixedLengthDataVector) 8823 return false; 8824 8825 // If __ARM_FEATURE_SVE_BITS != N do not allow GNU vector lax conversion. 8826 // "Whenever __ARM_FEATURE_SVE_BITS==N, GNUT implicitly 8827 // converts to VLAT and VLAT implicitly converts to GNUT." 8828 // ACLE Spec Version 00bet6, 3.7.3.2. Behavior common to vectors and 8829 // predicates. 8830 if (VecTy->getVectorKind() == VectorType::GenericVector && 8831 getTypeSize(SecondType) != getSVETypeSize(*this, BT)) 8832 return false; 8833 8834 // If -flax-vector-conversions=all is specified, the types are 8835 // certainly compatible. 8836 if (LVCKind == LangOptions::LaxVectorConversionKind::All) 8837 return true; 8838 8839 // If -flax-vector-conversions=integer is specified, the types are 8840 // compatible if the elements are integer types. 8841 if (LVCKind == LangOptions::LaxVectorConversionKind::Integer) 8842 return VecTy->getElementType().getCanonicalType()->isIntegerType() && 8843 FirstType->getSveEltType(*this)->isIntegerType(); 8844 } 8845 8846 return false; 8847 }; 8848 8849 return IsLaxCompatible(FirstType, SecondType) || 8850 IsLaxCompatible(SecondType, FirstType); 8851 } 8852 8853 bool ASTContext::hasDirectOwnershipQualifier(QualType Ty) const { 8854 while (true) { 8855 // __strong id 8856 if (const AttributedType *Attr = dyn_cast<AttributedType>(Ty)) { 8857 if (Attr->getAttrKind() == attr::ObjCOwnership) 8858 return true; 8859 8860 Ty = Attr->getModifiedType(); 8861 8862 // X *__strong (...) 8863 } else if (const ParenType *Paren = dyn_cast<ParenType>(Ty)) { 8864 Ty = Paren->getInnerType(); 8865 8866 // We do not want to look through typedefs, typeof(expr), 8867 // typeof(type), or any other way that the type is somehow 8868 // abstracted. 8869 } else { 8870 return false; 8871 } 8872 } 8873 } 8874 8875 //===----------------------------------------------------------------------===// 8876 // ObjCQualifiedIdTypesAreCompatible - Compatibility testing for qualified id's. 8877 //===----------------------------------------------------------------------===// 8878 8879 /// ProtocolCompatibleWithProtocol - return 'true' if 'lProto' is in the 8880 /// inheritance hierarchy of 'rProto'. 8881 bool 8882 ASTContext::ProtocolCompatibleWithProtocol(ObjCProtocolDecl *lProto, 8883 ObjCProtocolDecl *rProto) const { 8884 if (declaresSameEntity(lProto, rProto)) 8885 return true; 8886 for (auto *PI : rProto->protocols()) 8887 if (ProtocolCompatibleWithProtocol(lProto, PI)) 8888 return true; 8889 return false; 8890 } 8891 8892 /// ObjCQualifiedClassTypesAreCompatible - compare Class<pr,...> and 8893 /// Class<pr1, ...>. 8894 bool ASTContext::ObjCQualifiedClassTypesAreCompatible( 8895 const ObjCObjectPointerType *lhs, const ObjCObjectPointerType *rhs) { 8896 for (auto *lhsProto : lhs->quals()) { 8897 bool match = false; 8898 for (auto *rhsProto : rhs->quals()) { 8899 if (ProtocolCompatibleWithProtocol(lhsProto, rhsProto)) { 8900 match = true; 8901 break; 8902 } 8903 } 8904 if (!match) 8905 return false; 8906 } 8907 return true; 8908 } 8909 8910 /// ObjCQualifiedIdTypesAreCompatible - We know that one of lhs/rhs is an 8911 /// ObjCQualifiedIDType. 8912 bool ASTContext::ObjCQualifiedIdTypesAreCompatible( 8913 const ObjCObjectPointerType *lhs, const ObjCObjectPointerType *rhs, 8914 bool compare) { 8915 // Allow id<P..> and an 'id' in all cases. 8916 if (lhs->isObjCIdType() || rhs->isObjCIdType()) 8917 return true; 8918 8919 // Don't allow id<P..> to convert to Class or Class<P..> in either direction. 8920 if (lhs->isObjCClassType() || lhs->isObjCQualifiedClassType() || 8921 rhs->isObjCClassType() || rhs->isObjCQualifiedClassType()) 8922 return false; 8923 8924 if (lhs->isObjCQualifiedIdType()) { 8925 if (rhs->qual_empty()) { 8926 // If the RHS is a unqualified interface pointer "NSString*", 8927 // make sure we check the class hierarchy. 8928 if (ObjCInterfaceDecl *rhsID = rhs->getInterfaceDecl()) { 8929 for (auto *I : lhs->quals()) { 8930 // when comparing an id<P> on lhs with a static type on rhs, 8931 // see if static class implements all of id's protocols, directly or 8932 // through its super class and categories. 8933 if (!rhsID->ClassImplementsProtocol(I, true)) 8934 return false; 8935 } 8936 } 8937 // If there are no qualifiers and no interface, we have an 'id'. 8938 return true; 8939 } 8940 // Both the right and left sides have qualifiers. 8941 for (auto *lhsProto : lhs->quals()) { 8942 bool match = false; 8943 8944 // when comparing an id<P> on lhs with a static type on rhs, 8945 // see if static class implements all of id's protocols, directly or 8946 // through its super class and categories. 8947 for (auto *rhsProto : rhs->quals()) { 8948 if (ProtocolCompatibleWithProtocol(lhsProto, rhsProto) || 8949 (compare && ProtocolCompatibleWithProtocol(rhsProto, lhsProto))) { 8950 match = true; 8951 break; 8952 } 8953 } 8954 // If the RHS is a qualified interface pointer "NSString<P>*", 8955 // make sure we check the class hierarchy. 8956 if (ObjCInterfaceDecl *rhsID = rhs->getInterfaceDecl()) { 8957 for (auto *I : lhs->quals()) { 8958 // when comparing an id<P> on lhs with a static type on rhs, 8959 // see if static class implements all of id's protocols, directly or 8960 // through its super class and categories. 8961 if (rhsID->ClassImplementsProtocol(I, true)) { 8962 match = true; 8963 break; 8964 } 8965 } 8966 } 8967 if (!match) 8968 return false; 8969 } 8970 8971 return true; 8972 } 8973 8974 assert(rhs->isObjCQualifiedIdType() && "One of the LHS/RHS should be id<x>"); 8975 8976 if (lhs->getInterfaceType()) { 8977 // If both the right and left sides have qualifiers. 8978 for (auto *lhsProto : lhs->quals()) { 8979 bool match = false; 8980 8981 // when comparing an id<P> on rhs with a static type on lhs, 8982 // see if static class implements all of id's protocols, directly or 8983 // through its super class and categories. 8984 // First, lhs protocols in the qualifier list must be found, direct 8985 // or indirect in rhs's qualifier list or it is a mismatch. 8986 for (auto *rhsProto : rhs->quals()) { 8987 if (ProtocolCompatibleWithProtocol(lhsProto, rhsProto) || 8988 (compare && ProtocolCompatibleWithProtocol(rhsProto, lhsProto))) { 8989 match = true; 8990 break; 8991 } 8992 } 8993 if (!match) 8994 return false; 8995 } 8996 8997 // Static class's protocols, or its super class or category protocols 8998 // must be found, direct or indirect in rhs's qualifier list or it is a mismatch. 8999 if (ObjCInterfaceDecl *lhsID = lhs->getInterfaceDecl()) { 9000 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> LHSInheritedProtocols; 9001 CollectInheritedProtocols(lhsID, LHSInheritedProtocols); 9002 // This is rather dubious but matches gcc's behavior. If lhs has 9003 // no type qualifier and its class has no static protocol(s) 9004 // assume that it is mismatch. 9005 if (LHSInheritedProtocols.empty() && lhs->qual_empty()) 9006 return false; 9007 for (auto *lhsProto : LHSInheritedProtocols) { 9008 bool match = false; 9009 for (auto *rhsProto : rhs->quals()) { 9010 if (ProtocolCompatibleWithProtocol(lhsProto, rhsProto) || 9011 (compare && ProtocolCompatibleWithProtocol(rhsProto, lhsProto))) { 9012 match = true; 9013 break; 9014 } 9015 } 9016 if (!match) 9017 return false; 9018 } 9019 } 9020 return true; 9021 } 9022 return false; 9023 } 9024 9025 /// canAssignObjCInterfaces - Return true if the two interface types are 9026 /// compatible for assignment from RHS to LHS. This handles validation of any 9027 /// protocol qualifiers on the LHS or RHS. 9028 bool ASTContext::canAssignObjCInterfaces(const ObjCObjectPointerType *LHSOPT, 9029 const ObjCObjectPointerType *RHSOPT) { 9030 const ObjCObjectType* LHS = LHSOPT->getObjectType(); 9031 const ObjCObjectType* RHS = RHSOPT->getObjectType(); 9032 9033 // If either type represents the built-in 'id' type, return true. 9034 if (LHS->isObjCUnqualifiedId() || RHS->isObjCUnqualifiedId()) 9035 return true; 9036 9037 // Function object that propagates a successful result or handles 9038 // __kindof types. 9039 auto finish = [&](bool succeeded) -> bool { 9040 if (succeeded) 9041 return true; 9042 9043 if (!RHS->isKindOfType()) 9044 return false; 9045 9046 // Strip off __kindof and protocol qualifiers, then check whether 9047 // we can assign the other way. 9048 return canAssignObjCInterfaces(RHSOPT->stripObjCKindOfTypeAndQuals(*this), 9049 LHSOPT->stripObjCKindOfTypeAndQuals(*this)); 9050 }; 9051 9052 // Casts from or to id<P> are allowed when the other side has compatible 9053 // protocols. 9054 if (LHS->isObjCQualifiedId() || RHS->isObjCQualifiedId()) { 9055 return finish(ObjCQualifiedIdTypesAreCompatible(LHSOPT, RHSOPT, false)); 9056 } 9057 9058 // Verify protocol compatibility for casts from Class<P1> to Class<P2>. 9059 if (LHS->isObjCQualifiedClass() && RHS->isObjCQualifiedClass()) { 9060 return finish(ObjCQualifiedClassTypesAreCompatible(LHSOPT, RHSOPT)); 9061 } 9062 9063 // Casts from Class to Class<Foo>, or vice-versa, are allowed. 9064 if (LHS->isObjCClass() && RHS->isObjCClass()) { 9065 return true; 9066 } 9067 9068 // If we have 2 user-defined types, fall into that path. 9069 if (LHS->getInterface() && RHS->getInterface()) { 9070 return finish(canAssignObjCInterfaces(LHS, RHS)); 9071 } 9072 9073 return false; 9074 } 9075 9076 /// canAssignObjCInterfacesInBlockPointer - This routine is specifically written 9077 /// for providing type-safety for objective-c pointers used to pass/return 9078 /// arguments in block literals. When passed as arguments, passing 'A*' where 9079 /// 'id' is expected is not OK. Passing 'Sub *" where 'Super *" is expected is 9080 /// not OK. For the return type, the opposite is not OK. 9081 bool ASTContext::canAssignObjCInterfacesInBlockPointer( 9082 const ObjCObjectPointerType *LHSOPT, 9083 const ObjCObjectPointerType *RHSOPT, 9084 bool BlockReturnType) { 9085 9086 // Function object that propagates a successful result or handles 9087 // __kindof types. 9088 auto finish = [&](bool succeeded) -> bool { 9089 if (succeeded) 9090 return true; 9091 9092 const ObjCObjectPointerType *Expected = BlockReturnType ? RHSOPT : LHSOPT; 9093 if (!Expected->isKindOfType()) 9094 return false; 9095 9096 // Strip off __kindof and protocol qualifiers, then check whether 9097 // we can assign the other way. 9098 return canAssignObjCInterfacesInBlockPointer( 9099 RHSOPT->stripObjCKindOfTypeAndQuals(*this), 9100 LHSOPT->stripObjCKindOfTypeAndQuals(*this), 9101 BlockReturnType); 9102 }; 9103 9104 if (RHSOPT->isObjCBuiltinType() || LHSOPT->isObjCIdType()) 9105 return true; 9106 9107 if (LHSOPT->isObjCBuiltinType()) { 9108 return finish(RHSOPT->isObjCBuiltinType() || 9109 RHSOPT->isObjCQualifiedIdType()); 9110 } 9111 9112 if (LHSOPT->isObjCQualifiedIdType() || RHSOPT->isObjCQualifiedIdType()) { 9113 if (getLangOpts().CompatibilityQualifiedIdBlockParamTypeChecking) 9114 // Use for block parameters previous type checking for compatibility. 9115 return finish(ObjCQualifiedIdTypesAreCompatible(LHSOPT, RHSOPT, false) || 9116 // Or corrected type checking as in non-compat mode. 9117 (!BlockReturnType && 9118 ObjCQualifiedIdTypesAreCompatible(RHSOPT, LHSOPT, false))); 9119 else 9120 return finish(ObjCQualifiedIdTypesAreCompatible( 9121 (BlockReturnType ? LHSOPT : RHSOPT), 9122 (BlockReturnType ? RHSOPT : LHSOPT), false)); 9123 } 9124 9125 const ObjCInterfaceType* LHS = LHSOPT->getInterfaceType(); 9126 const ObjCInterfaceType* RHS = RHSOPT->getInterfaceType(); 9127 if (LHS && RHS) { // We have 2 user-defined types. 9128 if (LHS != RHS) { 9129 if (LHS->getDecl()->isSuperClassOf(RHS->getDecl())) 9130 return finish(BlockReturnType); 9131 if (RHS->getDecl()->isSuperClassOf(LHS->getDecl())) 9132 return finish(!BlockReturnType); 9133 } 9134 else 9135 return true; 9136 } 9137 return false; 9138 } 9139 9140 /// Comparison routine for Objective-C protocols to be used with 9141 /// llvm::array_pod_sort. 9142 static int compareObjCProtocolsByName(ObjCProtocolDecl * const *lhs, 9143 ObjCProtocolDecl * const *rhs) { 9144 return (*lhs)->getName().compare((*rhs)->getName()); 9145 } 9146 9147 /// getIntersectionOfProtocols - This routine finds the intersection of set 9148 /// of protocols inherited from two distinct objective-c pointer objects with 9149 /// the given common base. 9150 /// It is used to build composite qualifier list of the composite type of 9151 /// the conditional expression involving two objective-c pointer objects. 9152 static 9153 void getIntersectionOfProtocols(ASTContext &Context, 9154 const ObjCInterfaceDecl *CommonBase, 9155 const ObjCObjectPointerType *LHSOPT, 9156 const ObjCObjectPointerType *RHSOPT, 9157 SmallVectorImpl<ObjCProtocolDecl *> &IntersectionSet) { 9158 9159 const ObjCObjectType* LHS = LHSOPT->getObjectType(); 9160 const ObjCObjectType* RHS = RHSOPT->getObjectType(); 9161 assert(LHS->getInterface() && "LHS must have an interface base"); 9162 assert(RHS->getInterface() && "RHS must have an interface base"); 9163 9164 // Add all of the protocols for the LHS. 9165 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> LHSProtocolSet; 9166 9167 // Start with the protocol qualifiers. 9168 for (auto proto : LHS->quals()) { 9169 Context.CollectInheritedProtocols(proto, LHSProtocolSet); 9170 } 9171 9172 // Also add the protocols associated with the LHS interface. 9173 Context.CollectInheritedProtocols(LHS->getInterface(), LHSProtocolSet); 9174 9175 // Add all of the protocols for the RHS. 9176 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> RHSProtocolSet; 9177 9178 // Start with the protocol qualifiers. 9179 for (auto proto : RHS->quals()) { 9180 Context.CollectInheritedProtocols(proto, RHSProtocolSet); 9181 } 9182 9183 // Also add the protocols associated with the RHS interface. 9184 Context.CollectInheritedProtocols(RHS->getInterface(), RHSProtocolSet); 9185 9186 // Compute the intersection of the collected protocol sets. 9187 for (auto proto : LHSProtocolSet) { 9188 if (RHSProtocolSet.count(proto)) 9189 IntersectionSet.push_back(proto); 9190 } 9191 9192 // Compute the set of protocols that is implied by either the common type or 9193 // the protocols within the intersection. 9194 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> ImpliedProtocols; 9195 Context.CollectInheritedProtocols(CommonBase, ImpliedProtocols); 9196 9197 // Remove any implied protocols from the list of inherited protocols. 9198 if (!ImpliedProtocols.empty()) { 9199 llvm::erase_if(IntersectionSet, [&](ObjCProtocolDecl *proto) -> bool { 9200 return ImpliedProtocols.count(proto) > 0; 9201 }); 9202 } 9203 9204 // Sort the remaining protocols by name. 9205 llvm::array_pod_sort(IntersectionSet.begin(), IntersectionSet.end(), 9206 compareObjCProtocolsByName); 9207 } 9208 9209 /// Determine whether the first type is a subtype of the second. 9210 static bool canAssignObjCObjectTypes(ASTContext &ctx, QualType lhs, 9211 QualType rhs) { 9212 // Common case: two object pointers. 9213 const auto *lhsOPT = lhs->getAs<ObjCObjectPointerType>(); 9214 const auto *rhsOPT = rhs->getAs<ObjCObjectPointerType>(); 9215 if (lhsOPT && rhsOPT) 9216 return ctx.canAssignObjCInterfaces(lhsOPT, rhsOPT); 9217 9218 // Two block pointers. 9219 const auto *lhsBlock = lhs->getAs<BlockPointerType>(); 9220 const auto *rhsBlock = rhs->getAs<BlockPointerType>(); 9221 if (lhsBlock && rhsBlock) 9222 return ctx.typesAreBlockPointerCompatible(lhs, rhs); 9223 9224 // If either is an unqualified 'id' and the other is a block, it's 9225 // acceptable. 9226 if ((lhsOPT && lhsOPT->isObjCIdType() && rhsBlock) || 9227 (rhsOPT && rhsOPT->isObjCIdType() && lhsBlock)) 9228 return true; 9229 9230 return false; 9231 } 9232 9233 // Check that the given Objective-C type argument lists are equivalent. 9234 static bool sameObjCTypeArgs(ASTContext &ctx, 9235 const ObjCInterfaceDecl *iface, 9236 ArrayRef<QualType> lhsArgs, 9237 ArrayRef<QualType> rhsArgs, 9238 bool stripKindOf) { 9239 if (lhsArgs.size() != rhsArgs.size()) 9240 return false; 9241 9242 ObjCTypeParamList *typeParams = iface->getTypeParamList(); 9243 for (unsigned i = 0, n = lhsArgs.size(); i != n; ++i) { 9244 if (ctx.hasSameType(lhsArgs[i], rhsArgs[i])) 9245 continue; 9246 9247 switch (typeParams->begin()[i]->getVariance()) { 9248 case ObjCTypeParamVariance::Invariant: 9249 if (!stripKindOf || 9250 !ctx.hasSameType(lhsArgs[i].stripObjCKindOfType(ctx), 9251 rhsArgs[i].stripObjCKindOfType(ctx))) { 9252 return false; 9253 } 9254 break; 9255 9256 case ObjCTypeParamVariance::Covariant: 9257 if (!canAssignObjCObjectTypes(ctx, lhsArgs[i], rhsArgs[i])) 9258 return false; 9259 break; 9260 9261 case ObjCTypeParamVariance::Contravariant: 9262 if (!canAssignObjCObjectTypes(ctx, rhsArgs[i], lhsArgs[i])) 9263 return false; 9264 break; 9265 } 9266 } 9267 9268 return true; 9269 } 9270 9271 QualType ASTContext::areCommonBaseCompatible( 9272 const ObjCObjectPointerType *Lptr, 9273 const ObjCObjectPointerType *Rptr) { 9274 const ObjCObjectType *LHS = Lptr->getObjectType(); 9275 const ObjCObjectType *RHS = Rptr->getObjectType(); 9276 const ObjCInterfaceDecl* LDecl = LHS->getInterface(); 9277 const ObjCInterfaceDecl* RDecl = RHS->getInterface(); 9278 9279 if (!LDecl || !RDecl) 9280 return {}; 9281 9282 // When either LHS or RHS is a kindof type, we should return a kindof type. 9283 // For example, for common base of kindof(ASub1) and kindof(ASub2), we return 9284 // kindof(A). 9285 bool anyKindOf = LHS->isKindOfType() || RHS->isKindOfType(); 9286 9287 // Follow the left-hand side up the class hierarchy until we either hit a 9288 // root or find the RHS. Record the ancestors in case we don't find it. 9289 llvm::SmallDenseMap<const ObjCInterfaceDecl *, const ObjCObjectType *, 4> 9290 LHSAncestors; 9291 while (true) { 9292 // Record this ancestor. We'll need this if the common type isn't in the 9293 // path from the LHS to the root. 9294 LHSAncestors[LHS->getInterface()->getCanonicalDecl()] = LHS; 9295 9296 if (declaresSameEntity(LHS->getInterface(), RDecl)) { 9297 // Get the type arguments. 9298 ArrayRef<QualType> LHSTypeArgs = LHS->getTypeArgsAsWritten(); 9299 bool anyChanges = false; 9300 if (LHS->isSpecialized() && RHS->isSpecialized()) { 9301 // Both have type arguments, compare them. 9302 if (!sameObjCTypeArgs(*this, LHS->getInterface(), 9303 LHS->getTypeArgs(), RHS->getTypeArgs(), 9304 /*stripKindOf=*/true)) 9305 return {}; 9306 } else if (LHS->isSpecialized() != RHS->isSpecialized()) { 9307 // If only one has type arguments, the result will not have type 9308 // arguments. 9309 LHSTypeArgs = {}; 9310 anyChanges = true; 9311 } 9312 9313 // Compute the intersection of protocols. 9314 SmallVector<ObjCProtocolDecl *, 8> Protocols; 9315 getIntersectionOfProtocols(*this, LHS->getInterface(), Lptr, Rptr, 9316 Protocols); 9317 if (!Protocols.empty()) 9318 anyChanges = true; 9319 9320 // If anything in the LHS will have changed, build a new result type. 9321 // If we need to return a kindof type but LHS is not a kindof type, we 9322 // build a new result type. 9323 if (anyChanges || LHS->isKindOfType() != anyKindOf) { 9324 QualType Result = getObjCInterfaceType(LHS->getInterface()); 9325 Result = getObjCObjectType(Result, LHSTypeArgs, Protocols, 9326 anyKindOf || LHS->isKindOfType()); 9327 return getObjCObjectPointerType(Result); 9328 } 9329 9330 return getObjCObjectPointerType(QualType(LHS, 0)); 9331 } 9332 9333 // Find the superclass. 9334 QualType LHSSuperType = LHS->getSuperClassType(); 9335 if (LHSSuperType.isNull()) 9336 break; 9337 9338 LHS = LHSSuperType->castAs<ObjCObjectType>(); 9339 } 9340 9341 // We didn't find anything by following the LHS to its root; now check 9342 // the RHS against the cached set of ancestors. 9343 while (true) { 9344 auto KnownLHS = LHSAncestors.find(RHS->getInterface()->getCanonicalDecl()); 9345 if (KnownLHS != LHSAncestors.end()) { 9346 LHS = KnownLHS->second; 9347 9348 // Get the type arguments. 9349 ArrayRef<QualType> RHSTypeArgs = RHS->getTypeArgsAsWritten(); 9350 bool anyChanges = false; 9351 if (LHS->isSpecialized() && RHS->isSpecialized()) { 9352 // Both have type arguments, compare them. 9353 if (!sameObjCTypeArgs(*this, LHS->getInterface(), 9354 LHS->getTypeArgs(), RHS->getTypeArgs(), 9355 /*stripKindOf=*/true)) 9356 return {}; 9357 } else if (LHS->isSpecialized() != RHS->isSpecialized()) { 9358 // If only one has type arguments, the result will not have type 9359 // arguments. 9360 RHSTypeArgs = {}; 9361 anyChanges = true; 9362 } 9363 9364 // Compute the intersection of protocols. 9365 SmallVector<ObjCProtocolDecl *, 8> Protocols; 9366 getIntersectionOfProtocols(*this, RHS->getInterface(), Lptr, Rptr, 9367 Protocols); 9368 if (!Protocols.empty()) 9369 anyChanges = true; 9370 9371 // If we need to return a kindof type but RHS is not a kindof type, we 9372 // build a new result type. 9373 if (anyChanges || RHS->isKindOfType() != anyKindOf) { 9374 QualType Result = getObjCInterfaceType(RHS->getInterface()); 9375 Result = getObjCObjectType(Result, RHSTypeArgs, Protocols, 9376 anyKindOf || RHS->isKindOfType()); 9377 return getObjCObjectPointerType(Result); 9378 } 9379 9380 return getObjCObjectPointerType(QualType(RHS, 0)); 9381 } 9382 9383 // Find the superclass of the RHS. 9384 QualType RHSSuperType = RHS->getSuperClassType(); 9385 if (RHSSuperType.isNull()) 9386 break; 9387 9388 RHS = RHSSuperType->castAs<ObjCObjectType>(); 9389 } 9390 9391 return {}; 9392 } 9393 9394 bool ASTContext::canAssignObjCInterfaces(const ObjCObjectType *LHS, 9395 const ObjCObjectType *RHS) { 9396 assert(LHS->getInterface() && "LHS is not an interface type"); 9397 assert(RHS->getInterface() && "RHS is not an interface type"); 9398 9399 // Verify that the base decls are compatible: the RHS must be a subclass of 9400 // the LHS. 9401 ObjCInterfaceDecl *LHSInterface = LHS->getInterface(); 9402 bool IsSuperClass = LHSInterface->isSuperClassOf(RHS->getInterface()); 9403 if (!IsSuperClass) 9404 return false; 9405 9406 // If the LHS has protocol qualifiers, determine whether all of them are 9407 // satisfied by the RHS (i.e., the RHS has a superset of the protocols in the 9408 // LHS). 9409 if (LHS->getNumProtocols() > 0) { 9410 // OK if conversion of LHS to SuperClass results in narrowing of types 9411 // ; i.e., SuperClass may implement at least one of the protocols 9412 // in LHS's protocol list. Example, SuperObj<P1> = lhs<P1,P2> is ok. 9413 // But not SuperObj<P1,P2,P3> = lhs<P1,P2>. 9414 llvm::SmallPtrSet<ObjCProtocolDecl *, 8> SuperClassInheritedProtocols; 9415 CollectInheritedProtocols(RHS->getInterface(), SuperClassInheritedProtocols); 9416 // Also, if RHS has explicit quelifiers, include them for comparing with LHS's 9417 // qualifiers. 9418 for (auto *RHSPI : RHS->quals()) 9419 CollectInheritedProtocols(RHSPI, SuperClassInheritedProtocols); 9420 // If there is no protocols associated with RHS, it is not a match. 9421 if (SuperClassInheritedProtocols.empty()) 9422 return false; 9423 9424 for (const auto *LHSProto : LHS->quals()) { 9425 bool SuperImplementsProtocol = false; 9426 for (auto *SuperClassProto : SuperClassInheritedProtocols) 9427 if (SuperClassProto->lookupProtocolNamed(LHSProto->getIdentifier())) { 9428 SuperImplementsProtocol = true; 9429 break; 9430 } 9431 if (!SuperImplementsProtocol) 9432 return false; 9433 } 9434 } 9435 9436 // If the LHS is specialized, we may need to check type arguments. 9437 if (LHS->isSpecialized()) { 9438 // Follow the superclass chain until we've matched the LHS class in the 9439 // hierarchy. This substitutes type arguments through. 9440 const ObjCObjectType *RHSSuper = RHS; 9441 while (!declaresSameEntity(RHSSuper->getInterface(), LHSInterface)) 9442 RHSSuper = RHSSuper->getSuperClassType()->castAs<ObjCObjectType>(); 9443 9444 // If the RHS is specializd, compare type arguments. 9445 if (RHSSuper->isSpecialized() && 9446 !sameObjCTypeArgs(*this, LHS->getInterface(), 9447 LHS->getTypeArgs(), RHSSuper->getTypeArgs(), 9448 /*stripKindOf=*/true)) { 9449 return false; 9450 } 9451 } 9452 9453 return true; 9454 } 9455 9456 bool ASTContext::areComparableObjCPointerTypes(QualType LHS, QualType RHS) { 9457 // get the "pointed to" types 9458 const auto *LHSOPT = LHS->getAs<ObjCObjectPointerType>(); 9459 const auto *RHSOPT = RHS->getAs<ObjCObjectPointerType>(); 9460 9461 if (!LHSOPT || !RHSOPT) 9462 return false; 9463 9464 return canAssignObjCInterfaces(LHSOPT, RHSOPT) || 9465 canAssignObjCInterfaces(RHSOPT, LHSOPT); 9466 } 9467 9468 bool ASTContext::canBindObjCObjectType(QualType To, QualType From) { 9469 return canAssignObjCInterfaces( 9470 getObjCObjectPointerType(To)->castAs<ObjCObjectPointerType>(), 9471 getObjCObjectPointerType(From)->castAs<ObjCObjectPointerType>()); 9472 } 9473 9474 /// typesAreCompatible - C99 6.7.3p9: For two qualified types to be compatible, 9475 /// both shall have the identically qualified version of a compatible type. 9476 /// C99 6.2.7p1: Two types have compatible types if their types are the 9477 /// same. See 6.7.[2,3,5] for additional rules. 9478 bool ASTContext::typesAreCompatible(QualType LHS, QualType RHS, 9479 bool CompareUnqualified) { 9480 if (getLangOpts().CPlusPlus) 9481 return hasSameType(LHS, RHS); 9482 9483 return !mergeTypes(LHS, RHS, false, CompareUnqualified).isNull(); 9484 } 9485 9486 bool ASTContext::propertyTypesAreCompatible(QualType LHS, QualType RHS) { 9487 return typesAreCompatible(LHS, RHS); 9488 } 9489 9490 bool ASTContext::typesAreBlockPointerCompatible(QualType LHS, QualType RHS) { 9491 return !mergeTypes(LHS, RHS, true).isNull(); 9492 } 9493 9494 /// mergeTransparentUnionType - if T is a transparent union type and a member 9495 /// of T is compatible with SubType, return the merged type, else return 9496 /// QualType() 9497 QualType ASTContext::mergeTransparentUnionType(QualType T, QualType SubType, 9498 bool OfBlockPointer, 9499 bool Unqualified) { 9500 if (const RecordType *UT = T->getAsUnionType()) { 9501 RecordDecl *UD = UT->getDecl(); 9502 if (UD->hasAttr<TransparentUnionAttr>()) { 9503 for (const auto *I : UD->fields()) { 9504 QualType ET = I->getType().getUnqualifiedType(); 9505 QualType MT = mergeTypes(ET, SubType, OfBlockPointer, Unqualified); 9506 if (!MT.isNull()) 9507 return MT; 9508 } 9509 } 9510 } 9511 9512 return {}; 9513 } 9514 9515 /// mergeFunctionParameterTypes - merge two types which appear as function 9516 /// parameter types 9517 QualType ASTContext::mergeFunctionParameterTypes(QualType lhs, QualType rhs, 9518 bool OfBlockPointer, 9519 bool Unqualified) { 9520 // GNU extension: two types are compatible if they appear as a function 9521 // argument, one of the types is a transparent union type and the other 9522 // type is compatible with a union member 9523 QualType lmerge = mergeTransparentUnionType(lhs, rhs, OfBlockPointer, 9524 Unqualified); 9525 if (!lmerge.isNull()) 9526 return lmerge; 9527 9528 QualType rmerge = mergeTransparentUnionType(rhs, lhs, OfBlockPointer, 9529 Unqualified); 9530 if (!rmerge.isNull()) 9531 return rmerge; 9532 9533 return mergeTypes(lhs, rhs, OfBlockPointer, Unqualified); 9534 } 9535 9536 QualType ASTContext::mergeFunctionTypes(QualType lhs, QualType rhs, 9537 bool OfBlockPointer, bool Unqualified, 9538 bool AllowCXX) { 9539 const auto *lbase = lhs->castAs<FunctionType>(); 9540 const auto *rbase = rhs->castAs<FunctionType>(); 9541 const auto *lproto = dyn_cast<FunctionProtoType>(lbase); 9542 const auto *rproto = dyn_cast<FunctionProtoType>(rbase); 9543 bool allLTypes = true; 9544 bool allRTypes = true; 9545 9546 // Check return type 9547 QualType retType; 9548 if (OfBlockPointer) { 9549 QualType RHS = rbase->getReturnType(); 9550 QualType LHS = lbase->getReturnType(); 9551 bool UnqualifiedResult = Unqualified; 9552 if (!UnqualifiedResult) 9553 UnqualifiedResult = (!RHS.hasQualifiers() && LHS.hasQualifiers()); 9554 retType = mergeTypes(LHS, RHS, true, UnqualifiedResult, true); 9555 } 9556 else 9557 retType = mergeTypes(lbase->getReturnType(), rbase->getReturnType(), false, 9558 Unqualified); 9559 if (retType.isNull()) 9560 return {}; 9561 9562 if (Unqualified) 9563 retType = retType.getUnqualifiedType(); 9564 9565 CanQualType LRetType = getCanonicalType(lbase->getReturnType()); 9566 CanQualType RRetType = getCanonicalType(rbase->getReturnType()); 9567 if (Unqualified) { 9568 LRetType = LRetType.getUnqualifiedType(); 9569 RRetType = RRetType.getUnqualifiedType(); 9570 } 9571 9572 if (getCanonicalType(retType) != LRetType) 9573 allLTypes = false; 9574 if (getCanonicalType(retType) != RRetType) 9575 allRTypes = false; 9576 9577 // FIXME: double check this 9578 // FIXME: should we error if lbase->getRegParmAttr() != 0 && 9579 // rbase->getRegParmAttr() != 0 && 9580 // lbase->getRegParmAttr() != rbase->getRegParmAttr()? 9581 FunctionType::ExtInfo lbaseInfo = lbase->getExtInfo(); 9582 FunctionType::ExtInfo rbaseInfo = rbase->getExtInfo(); 9583 9584 // Compatible functions must have compatible calling conventions 9585 if (lbaseInfo.getCC() != rbaseInfo.getCC()) 9586 return {}; 9587 9588 // Regparm is part of the calling convention. 9589 if (lbaseInfo.getHasRegParm() != rbaseInfo.getHasRegParm()) 9590 return {}; 9591 if (lbaseInfo.getRegParm() != rbaseInfo.getRegParm()) 9592 return {}; 9593 9594 if (lbaseInfo.getProducesResult() != rbaseInfo.getProducesResult()) 9595 return {}; 9596 if (lbaseInfo.getNoCallerSavedRegs() != rbaseInfo.getNoCallerSavedRegs()) 9597 return {}; 9598 if (lbaseInfo.getNoCfCheck() != rbaseInfo.getNoCfCheck()) 9599 return {}; 9600 9601 // FIXME: some uses, e.g. conditional exprs, really want this to be 'both'. 9602 bool NoReturn = lbaseInfo.getNoReturn() || rbaseInfo.getNoReturn(); 9603 9604 if (lbaseInfo.getNoReturn() != NoReturn) 9605 allLTypes = false; 9606 if (rbaseInfo.getNoReturn() != NoReturn) 9607 allRTypes = false; 9608 9609 FunctionType::ExtInfo einfo = lbaseInfo.withNoReturn(NoReturn); 9610 9611 if (lproto && rproto) { // two C99 style function prototypes 9612 assert((AllowCXX || 9613 (!lproto->hasExceptionSpec() && !rproto->hasExceptionSpec())) && 9614 "C++ shouldn't be here"); 9615 // Compatible functions must have the same number of parameters 9616 if (lproto->getNumParams() != rproto->getNumParams()) 9617 return {}; 9618 9619 // Variadic and non-variadic functions aren't compatible 9620 if (lproto->isVariadic() != rproto->isVariadic()) 9621 return {}; 9622 9623 if (lproto->getMethodQuals() != rproto->getMethodQuals()) 9624 return {}; 9625 9626 SmallVector<FunctionProtoType::ExtParameterInfo, 4> newParamInfos; 9627 bool canUseLeft, canUseRight; 9628 if (!mergeExtParameterInfo(lproto, rproto, canUseLeft, canUseRight, 9629 newParamInfos)) 9630 return {}; 9631 9632 if (!canUseLeft) 9633 allLTypes = false; 9634 if (!canUseRight) 9635 allRTypes = false; 9636 9637 // Check parameter type compatibility 9638 SmallVector<QualType, 10> types; 9639 for (unsigned i = 0, n = lproto->getNumParams(); i < n; i++) { 9640 QualType lParamType = lproto->getParamType(i).getUnqualifiedType(); 9641 QualType rParamType = rproto->getParamType(i).getUnqualifiedType(); 9642 QualType paramType = mergeFunctionParameterTypes( 9643 lParamType, rParamType, OfBlockPointer, Unqualified); 9644 if (paramType.isNull()) 9645 return {}; 9646 9647 if (Unqualified) 9648 paramType = paramType.getUnqualifiedType(); 9649 9650 types.push_back(paramType); 9651 if (Unqualified) { 9652 lParamType = lParamType.getUnqualifiedType(); 9653 rParamType = rParamType.getUnqualifiedType(); 9654 } 9655 9656 if (getCanonicalType(paramType) != getCanonicalType(lParamType)) 9657 allLTypes = false; 9658 if (getCanonicalType(paramType) != getCanonicalType(rParamType)) 9659 allRTypes = false; 9660 } 9661 9662 if (allLTypes) return lhs; 9663 if (allRTypes) return rhs; 9664 9665 FunctionProtoType::ExtProtoInfo EPI = lproto->getExtProtoInfo(); 9666 EPI.ExtInfo = einfo; 9667 EPI.ExtParameterInfos = 9668 newParamInfos.empty() ? nullptr : newParamInfos.data(); 9669 return getFunctionType(retType, types, EPI); 9670 } 9671 9672 if (lproto) allRTypes = false; 9673 if (rproto) allLTypes = false; 9674 9675 const FunctionProtoType *proto = lproto ? lproto : rproto; 9676 if (proto) { 9677 assert((AllowCXX || !proto->hasExceptionSpec()) && "C++ shouldn't be here"); 9678 if (proto->isVariadic()) 9679 return {}; 9680 // Check that the types are compatible with the types that 9681 // would result from default argument promotions (C99 6.7.5.3p15). 9682 // The only types actually affected are promotable integer 9683 // types and floats, which would be passed as a different 9684 // type depending on whether the prototype is visible. 9685 for (unsigned i = 0, n = proto->getNumParams(); i < n; ++i) { 9686 QualType paramTy = proto->getParamType(i); 9687 9688 // Look at the converted type of enum types, since that is the type used 9689 // to pass enum values. 9690 if (const auto *Enum = paramTy->getAs<EnumType>()) { 9691 paramTy = Enum->getDecl()->getIntegerType(); 9692 if (paramTy.isNull()) 9693 return {}; 9694 } 9695 9696 if (paramTy->isPromotableIntegerType() || 9697 getCanonicalType(paramTy).getUnqualifiedType() == FloatTy) 9698 return {}; 9699 } 9700 9701 if (allLTypes) return lhs; 9702 if (allRTypes) return rhs; 9703 9704 FunctionProtoType::ExtProtoInfo EPI = proto->getExtProtoInfo(); 9705 EPI.ExtInfo = einfo; 9706 return getFunctionType(retType, proto->getParamTypes(), EPI); 9707 } 9708 9709 if (allLTypes) return lhs; 9710 if (allRTypes) return rhs; 9711 return getFunctionNoProtoType(retType, einfo); 9712 } 9713 9714 /// Given that we have an enum type and a non-enum type, try to merge them. 9715 static QualType mergeEnumWithInteger(ASTContext &Context, const EnumType *ET, 9716 QualType other, bool isBlockReturnType) { 9717 // C99 6.7.2.2p4: Each enumerated type shall be compatible with char, 9718 // a signed integer type, or an unsigned integer type. 9719 // Compatibility is based on the underlying type, not the promotion 9720 // type. 9721 QualType underlyingType = ET->getDecl()->getIntegerType(); 9722 if (underlyingType.isNull()) 9723 return {}; 9724 if (Context.hasSameType(underlyingType, other)) 9725 return other; 9726 9727 // In block return types, we're more permissive and accept any 9728 // integral type of the same size. 9729 if (isBlockReturnType && other->isIntegerType() && 9730 Context.getTypeSize(underlyingType) == Context.getTypeSize(other)) 9731 return other; 9732 9733 return {}; 9734 } 9735 9736 QualType ASTContext::mergeTypes(QualType LHS, QualType RHS, 9737 bool OfBlockPointer, 9738 bool Unqualified, bool BlockReturnType) { 9739 // For C++ we will not reach this code with reference types (see below), 9740 // for OpenMP variant call overloading we might. 9741 // 9742 // C++ [expr]: If an expression initially has the type "reference to T", the 9743 // type is adjusted to "T" prior to any further analysis, the expression 9744 // designates the object or function denoted by the reference, and the 9745 // expression is an lvalue unless the reference is an rvalue reference and 9746 // the expression is a function call (possibly inside parentheses). 9747 if (LangOpts.OpenMP && LHS->getAs<ReferenceType>() && 9748 RHS->getAs<ReferenceType>() && LHS->getTypeClass() == RHS->getTypeClass()) 9749 return mergeTypes(LHS->getAs<ReferenceType>()->getPointeeType(), 9750 RHS->getAs<ReferenceType>()->getPointeeType(), 9751 OfBlockPointer, Unqualified, BlockReturnType); 9752 if (LHS->getAs<ReferenceType>() || RHS->getAs<ReferenceType>()) 9753 return {}; 9754 9755 if (Unqualified) { 9756 LHS = LHS.getUnqualifiedType(); 9757 RHS = RHS.getUnqualifiedType(); 9758 } 9759 9760 QualType LHSCan = getCanonicalType(LHS), 9761 RHSCan = getCanonicalType(RHS); 9762 9763 // If two types are identical, they are compatible. 9764 if (LHSCan == RHSCan) 9765 return LHS; 9766 9767 // If the qualifiers are different, the types aren't compatible... mostly. 9768 Qualifiers LQuals = LHSCan.getLocalQualifiers(); 9769 Qualifiers RQuals = RHSCan.getLocalQualifiers(); 9770 if (LQuals != RQuals) { 9771 // If any of these qualifiers are different, we have a type 9772 // mismatch. 9773 if (LQuals.getCVRQualifiers() != RQuals.getCVRQualifiers() || 9774 LQuals.getAddressSpace() != RQuals.getAddressSpace() || 9775 LQuals.getObjCLifetime() != RQuals.getObjCLifetime() || 9776 LQuals.hasUnaligned() != RQuals.hasUnaligned()) 9777 return {}; 9778 9779 // Exactly one GC qualifier difference is allowed: __strong is 9780 // okay if the other type has no GC qualifier but is an Objective 9781 // C object pointer (i.e. implicitly strong by default). We fix 9782 // this by pretending that the unqualified type was actually 9783 // qualified __strong. 9784 Qualifiers::GC GC_L = LQuals.getObjCGCAttr(); 9785 Qualifiers::GC GC_R = RQuals.getObjCGCAttr(); 9786 assert((GC_L != GC_R) && "unequal qualifier sets had only equal elements"); 9787 9788 if (GC_L == Qualifiers::Weak || GC_R == Qualifiers::Weak) 9789 return {}; 9790 9791 if (GC_L == Qualifiers::Strong && RHSCan->isObjCObjectPointerType()) { 9792 return mergeTypes(LHS, getObjCGCQualType(RHS, Qualifiers::Strong)); 9793 } 9794 if (GC_R == Qualifiers::Strong && LHSCan->isObjCObjectPointerType()) { 9795 return mergeTypes(getObjCGCQualType(LHS, Qualifiers::Strong), RHS); 9796 } 9797 return {}; 9798 } 9799 9800 // Okay, qualifiers are equal. 9801 9802 Type::TypeClass LHSClass = LHSCan->getTypeClass(); 9803 Type::TypeClass RHSClass = RHSCan->getTypeClass(); 9804 9805 // We want to consider the two function types to be the same for these 9806 // comparisons, just force one to the other. 9807 if (LHSClass == Type::FunctionProto) LHSClass = Type::FunctionNoProto; 9808 if (RHSClass == Type::FunctionProto) RHSClass = Type::FunctionNoProto; 9809 9810 // Same as above for arrays 9811 if (LHSClass == Type::VariableArray || LHSClass == Type::IncompleteArray) 9812 LHSClass = Type::ConstantArray; 9813 if (RHSClass == Type::VariableArray || RHSClass == Type::IncompleteArray) 9814 RHSClass = Type::ConstantArray; 9815 9816 // ObjCInterfaces are just specialized ObjCObjects. 9817 if (LHSClass == Type::ObjCInterface) LHSClass = Type::ObjCObject; 9818 if (RHSClass == Type::ObjCInterface) RHSClass = Type::ObjCObject; 9819 9820 // Canonicalize ExtVector -> Vector. 9821 if (LHSClass == Type::ExtVector) LHSClass = Type::Vector; 9822 if (RHSClass == Type::ExtVector) RHSClass = Type::Vector; 9823 9824 // If the canonical type classes don't match. 9825 if (LHSClass != RHSClass) { 9826 // Note that we only have special rules for turning block enum 9827 // returns into block int returns, not vice-versa. 9828 if (const auto *ETy = LHS->getAs<EnumType>()) { 9829 return mergeEnumWithInteger(*this, ETy, RHS, false); 9830 } 9831 if (const EnumType* ETy = RHS->getAs<EnumType>()) { 9832 return mergeEnumWithInteger(*this, ETy, LHS, BlockReturnType); 9833 } 9834 // allow block pointer type to match an 'id' type. 9835 if (OfBlockPointer && !BlockReturnType) { 9836 if (LHS->isObjCIdType() && RHS->isBlockPointerType()) 9837 return LHS; 9838 if (RHS->isObjCIdType() && LHS->isBlockPointerType()) 9839 return RHS; 9840 } 9841 9842 return {}; 9843 } 9844 9845 // The canonical type classes match. 9846 switch (LHSClass) { 9847 #define TYPE(Class, Base) 9848 #define ABSTRACT_TYPE(Class, Base) 9849 #define NON_CANONICAL_UNLESS_DEPENDENT_TYPE(Class, Base) case Type::Class: 9850 #define NON_CANONICAL_TYPE(Class, Base) case Type::Class: 9851 #define DEPENDENT_TYPE(Class, Base) case Type::Class: 9852 #include "clang/AST/TypeNodes.inc" 9853 llvm_unreachable("Non-canonical and dependent types shouldn't get here"); 9854 9855 case Type::Auto: 9856 case Type::DeducedTemplateSpecialization: 9857 case Type::LValueReference: 9858 case Type::RValueReference: 9859 case Type::MemberPointer: 9860 llvm_unreachable("C++ should never be in mergeTypes"); 9861 9862 case Type::ObjCInterface: 9863 case Type::IncompleteArray: 9864 case Type::VariableArray: 9865 case Type::FunctionProto: 9866 case Type::ExtVector: 9867 llvm_unreachable("Types are eliminated above"); 9868 9869 case Type::Pointer: 9870 { 9871 // Merge two pointer types, while trying to preserve typedef info 9872 QualType LHSPointee = LHS->castAs<PointerType>()->getPointeeType(); 9873 QualType RHSPointee = RHS->castAs<PointerType>()->getPointeeType(); 9874 if (Unqualified) { 9875 LHSPointee = LHSPointee.getUnqualifiedType(); 9876 RHSPointee = RHSPointee.getUnqualifiedType(); 9877 } 9878 QualType ResultType = mergeTypes(LHSPointee, RHSPointee, false, 9879 Unqualified); 9880 if (ResultType.isNull()) 9881 return {}; 9882 if (getCanonicalType(LHSPointee) == getCanonicalType(ResultType)) 9883 return LHS; 9884 if (getCanonicalType(RHSPointee) == getCanonicalType(ResultType)) 9885 return RHS; 9886 return getPointerType(ResultType); 9887 } 9888 case Type::BlockPointer: 9889 { 9890 // Merge two block pointer types, while trying to preserve typedef info 9891 QualType LHSPointee = LHS->castAs<BlockPointerType>()->getPointeeType(); 9892 QualType RHSPointee = RHS->castAs<BlockPointerType>()->getPointeeType(); 9893 if (Unqualified) { 9894 LHSPointee = LHSPointee.getUnqualifiedType(); 9895 RHSPointee = RHSPointee.getUnqualifiedType(); 9896 } 9897 if (getLangOpts().OpenCL) { 9898 Qualifiers LHSPteeQual = LHSPointee.getQualifiers(); 9899 Qualifiers RHSPteeQual = RHSPointee.getQualifiers(); 9900 // Blocks can't be an expression in a ternary operator (OpenCL v2.0 9901 // 6.12.5) thus the following check is asymmetric. 9902 if (!LHSPteeQual.isAddressSpaceSupersetOf(RHSPteeQual)) 9903 return {}; 9904 LHSPteeQual.removeAddressSpace(); 9905 RHSPteeQual.removeAddressSpace(); 9906 LHSPointee = 9907 QualType(LHSPointee.getTypePtr(), LHSPteeQual.getAsOpaqueValue()); 9908 RHSPointee = 9909 QualType(RHSPointee.getTypePtr(), RHSPteeQual.getAsOpaqueValue()); 9910 } 9911 QualType ResultType = mergeTypes(LHSPointee, RHSPointee, OfBlockPointer, 9912 Unqualified); 9913 if (ResultType.isNull()) 9914 return {}; 9915 if (getCanonicalType(LHSPointee) == getCanonicalType(ResultType)) 9916 return LHS; 9917 if (getCanonicalType(RHSPointee) == getCanonicalType(ResultType)) 9918 return RHS; 9919 return getBlockPointerType(ResultType); 9920 } 9921 case Type::Atomic: 9922 { 9923 // Merge two pointer types, while trying to preserve typedef info 9924 QualType LHSValue = LHS->castAs<AtomicType>()->getValueType(); 9925 QualType RHSValue = RHS->castAs<AtomicType>()->getValueType(); 9926 if (Unqualified) { 9927 LHSValue = LHSValue.getUnqualifiedType(); 9928 RHSValue = RHSValue.getUnqualifiedType(); 9929 } 9930 QualType ResultType = mergeTypes(LHSValue, RHSValue, false, 9931 Unqualified); 9932 if (ResultType.isNull()) 9933 return {}; 9934 if (getCanonicalType(LHSValue) == getCanonicalType(ResultType)) 9935 return LHS; 9936 if (getCanonicalType(RHSValue) == getCanonicalType(ResultType)) 9937 return RHS; 9938 return getAtomicType(ResultType); 9939 } 9940 case Type::ConstantArray: 9941 { 9942 const ConstantArrayType* LCAT = getAsConstantArrayType(LHS); 9943 const ConstantArrayType* RCAT = getAsConstantArrayType(RHS); 9944 if (LCAT && RCAT && RCAT->getSize() != LCAT->getSize()) 9945 return {}; 9946 9947 QualType LHSElem = getAsArrayType(LHS)->getElementType(); 9948 QualType RHSElem = getAsArrayType(RHS)->getElementType(); 9949 if (Unqualified) { 9950 LHSElem = LHSElem.getUnqualifiedType(); 9951 RHSElem = RHSElem.getUnqualifiedType(); 9952 } 9953 9954 QualType ResultType = mergeTypes(LHSElem, RHSElem, false, Unqualified); 9955 if (ResultType.isNull()) 9956 return {}; 9957 9958 const VariableArrayType* LVAT = getAsVariableArrayType(LHS); 9959 const VariableArrayType* RVAT = getAsVariableArrayType(RHS); 9960 9961 // If either side is a variable array, and both are complete, check whether 9962 // the current dimension is definite. 9963 if (LVAT || RVAT) { 9964 auto SizeFetch = [this](const VariableArrayType* VAT, 9965 const ConstantArrayType* CAT) 9966 -> std::pair<bool,llvm::APInt> { 9967 if (VAT) { 9968 Optional<llvm::APSInt> TheInt; 9969 Expr *E = VAT->getSizeExpr(); 9970 if (E && (TheInt = E->getIntegerConstantExpr(*this))) 9971 return std::make_pair(true, *TheInt); 9972 return std::make_pair(false, llvm::APSInt()); 9973 } 9974 if (CAT) 9975 return std::make_pair(true, CAT->getSize()); 9976 return std::make_pair(false, llvm::APInt()); 9977 }; 9978 9979 bool HaveLSize, HaveRSize; 9980 llvm::APInt LSize, RSize; 9981 std::tie(HaveLSize, LSize) = SizeFetch(LVAT, LCAT); 9982 std::tie(HaveRSize, RSize) = SizeFetch(RVAT, RCAT); 9983 if (HaveLSize && HaveRSize && !llvm::APInt::isSameValue(LSize, RSize)) 9984 return {}; // Definite, but unequal, array dimension 9985 } 9986 9987 if (LCAT && getCanonicalType(LHSElem) == getCanonicalType(ResultType)) 9988 return LHS; 9989 if (RCAT && getCanonicalType(RHSElem) == getCanonicalType(ResultType)) 9990 return RHS; 9991 if (LCAT) 9992 return getConstantArrayType(ResultType, LCAT->getSize(), 9993 LCAT->getSizeExpr(), 9994 ArrayType::ArraySizeModifier(), 0); 9995 if (RCAT) 9996 return getConstantArrayType(ResultType, RCAT->getSize(), 9997 RCAT->getSizeExpr(), 9998 ArrayType::ArraySizeModifier(), 0); 9999 if (LVAT && getCanonicalType(LHSElem) == getCanonicalType(ResultType)) 10000 return LHS; 10001 if (RVAT && getCanonicalType(RHSElem) == getCanonicalType(ResultType)) 10002 return RHS; 10003 if (LVAT) { 10004 // FIXME: This isn't correct! But tricky to implement because 10005 // the array's size has to be the size of LHS, but the type 10006 // has to be different. 10007 return LHS; 10008 } 10009 if (RVAT) { 10010 // FIXME: This isn't correct! But tricky to implement because 10011 // the array's size has to be the size of RHS, but the type 10012 // has to be different. 10013 return RHS; 10014 } 10015 if (getCanonicalType(LHSElem) == getCanonicalType(ResultType)) return LHS; 10016 if (getCanonicalType(RHSElem) == getCanonicalType(ResultType)) return RHS; 10017 return getIncompleteArrayType(ResultType, 10018 ArrayType::ArraySizeModifier(), 0); 10019 } 10020 case Type::FunctionNoProto: 10021 return mergeFunctionTypes(LHS, RHS, OfBlockPointer, Unqualified); 10022 case Type::Record: 10023 case Type::Enum: 10024 return {}; 10025 case Type::Builtin: 10026 // Only exactly equal builtin types are compatible, which is tested above. 10027 return {}; 10028 case Type::Complex: 10029 // Distinct complex types are incompatible. 10030 return {}; 10031 case Type::Vector: 10032 // FIXME: The merged type should be an ExtVector! 10033 if (areCompatVectorTypes(LHSCan->castAs<VectorType>(), 10034 RHSCan->castAs<VectorType>())) 10035 return LHS; 10036 return {}; 10037 case Type::ConstantMatrix: 10038 if (areCompatMatrixTypes(LHSCan->castAs<ConstantMatrixType>(), 10039 RHSCan->castAs<ConstantMatrixType>())) 10040 return LHS; 10041 return {}; 10042 case Type::ObjCObject: { 10043 // Check if the types are assignment compatible. 10044 // FIXME: This should be type compatibility, e.g. whether 10045 // "LHS x; RHS x;" at global scope is legal. 10046 if (canAssignObjCInterfaces(LHS->castAs<ObjCObjectType>(), 10047 RHS->castAs<ObjCObjectType>())) 10048 return LHS; 10049 return {}; 10050 } 10051 case Type::ObjCObjectPointer: 10052 if (OfBlockPointer) { 10053 if (canAssignObjCInterfacesInBlockPointer( 10054 LHS->castAs<ObjCObjectPointerType>(), 10055 RHS->castAs<ObjCObjectPointerType>(), BlockReturnType)) 10056 return LHS; 10057 return {}; 10058 } 10059 if (canAssignObjCInterfaces(LHS->castAs<ObjCObjectPointerType>(), 10060 RHS->castAs<ObjCObjectPointerType>())) 10061 return LHS; 10062 return {}; 10063 case Type::Pipe: 10064 assert(LHS != RHS && 10065 "Equivalent pipe types should have already been handled!"); 10066 return {}; 10067 case Type::ExtInt: { 10068 // Merge two ext-int types, while trying to preserve typedef info. 10069 bool LHSUnsigned = LHS->castAs<ExtIntType>()->isUnsigned(); 10070 bool RHSUnsigned = RHS->castAs<ExtIntType>()->isUnsigned(); 10071 unsigned LHSBits = LHS->castAs<ExtIntType>()->getNumBits(); 10072 unsigned RHSBits = RHS->castAs<ExtIntType>()->getNumBits(); 10073 10074 // Like unsigned/int, shouldn't have a type if they don't match. 10075 if (LHSUnsigned != RHSUnsigned) 10076 return {}; 10077 10078 if (LHSBits != RHSBits) 10079 return {}; 10080 return LHS; 10081 } 10082 } 10083 10084 llvm_unreachable("Invalid Type::Class!"); 10085 } 10086 10087 bool ASTContext::mergeExtParameterInfo( 10088 const FunctionProtoType *FirstFnType, const FunctionProtoType *SecondFnType, 10089 bool &CanUseFirst, bool &CanUseSecond, 10090 SmallVectorImpl<FunctionProtoType::ExtParameterInfo> &NewParamInfos) { 10091 assert(NewParamInfos.empty() && "param info list not empty"); 10092 CanUseFirst = CanUseSecond = true; 10093 bool FirstHasInfo = FirstFnType->hasExtParameterInfos(); 10094 bool SecondHasInfo = SecondFnType->hasExtParameterInfos(); 10095 10096 // Fast path: if the first type doesn't have ext parameter infos, 10097 // we match if and only if the second type also doesn't have them. 10098 if (!FirstHasInfo && !SecondHasInfo) 10099 return true; 10100 10101 bool NeedParamInfo = false; 10102 size_t E = FirstHasInfo ? FirstFnType->getExtParameterInfos().size() 10103 : SecondFnType->getExtParameterInfos().size(); 10104 10105 for (size_t I = 0; I < E; ++I) { 10106 FunctionProtoType::ExtParameterInfo FirstParam, SecondParam; 10107 if (FirstHasInfo) 10108 FirstParam = FirstFnType->getExtParameterInfo(I); 10109 if (SecondHasInfo) 10110 SecondParam = SecondFnType->getExtParameterInfo(I); 10111 10112 // Cannot merge unless everything except the noescape flag matches. 10113 if (FirstParam.withIsNoEscape(false) != SecondParam.withIsNoEscape(false)) 10114 return false; 10115 10116 bool FirstNoEscape = FirstParam.isNoEscape(); 10117 bool SecondNoEscape = SecondParam.isNoEscape(); 10118 bool IsNoEscape = FirstNoEscape && SecondNoEscape; 10119 NewParamInfos.push_back(FirstParam.withIsNoEscape(IsNoEscape)); 10120 if (NewParamInfos.back().getOpaqueValue()) 10121 NeedParamInfo = true; 10122 if (FirstNoEscape != IsNoEscape) 10123 CanUseFirst = false; 10124 if (SecondNoEscape != IsNoEscape) 10125 CanUseSecond = false; 10126 } 10127 10128 if (!NeedParamInfo) 10129 NewParamInfos.clear(); 10130 10131 return true; 10132 } 10133 10134 void ASTContext::ResetObjCLayout(const ObjCContainerDecl *CD) { 10135 ObjCLayouts[CD] = nullptr; 10136 } 10137 10138 /// mergeObjCGCQualifiers - This routine merges ObjC's GC attribute of 'LHS' and 10139 /// 'RHS' attributes and returns the merged version; including for function 10140 /// return types. 10141 QualType ASTContext::mergeObjCGCQualifiers(QualType LHS, QualType RHS) { 10142 QualType LHSCan = getCanonicalType(LHS), 10143 RHSCan = getCanonicalType(RHS); 10144 // If two types are identical, they are compatible. 10145 if (LHSCan == RHSCan) 10146 return LHS; 10147 if (RHSCan->isFunctionType()) { 10148 if (!LHSCan->isFunctionType()) 10149 return {}; 10150 QualType OldReturnType = 10151 cast<FunctionType>(RHSCan.getTypePtr())->getReturnType(); 10152 QualType NewReturnType = 10153 cast<FunctionType>(LHSCan.getTypePtr())->getReturnType(); 10154 QualType ResReturnType = 10155 mergeObjCGCQualifiers(NewReturnType, OldReturnType); 10156 if (ResReturnType.isNull()) 10157 return {}; 10158 if (ResReturnType == NewReturnType || ResReturnType == OldReturnType) { 10159 // id foo(); ... __strong id foo(); or: __strong id foo(); ... id foo(); 10160 // In either case, use OldReturnType to build the new function type. 10161 const auto *F = LHS->castAs<FunctionType>(); 10162 if (const auto *FPT = cast<FunctionProtoType>(F)) { 10163 FunctionProtoType::ExtProtoInfo EPI = FPT->getExtProtoInfo(); 10164 EPI.ExtInfo = getFunctionExtInfo(LHS); 10165 QualType ResultType = 10166 getFunctionType(OldReturnType, FPT->getParamTypes(), EPI); 10167 return ResultType; 10168 } 10169 } 10170 return {}; 10171 } 10172 10173 // If the qualifiers are different, the types can still be merged. 10174 Qualifiers LQuals = LHSCan.getLocalQualifiers(); 10175 Qualifiers RQuals = RHSCan.getLocalQualifiers(); 10176 if (LQuals != RQuals) { 10177 // If any of these qualifiers are different, we have a type mismatch. 10178 if (LQuals.getCVRQualifiers() != RQuals.getCVRQualifiers() || 10179 LQuals.getAddressSpace() != RQuals.getAddressSpace()) 10180 return {}; 10181 10182 // Exactly one GC qualifier difference is allowed: __strong is 10183 // okay if the other type has no GC qualifier but is an Objective 10184 // C object pointer (i.e. implicitly strong by default). We fix 10185 // this by pretending that the unqualified type was actually 10186 // qualified __strong. 10187 Qualifiers::GC GC_L = LQuals.getObjCGCAttr(); 10188 Qualifiers::GC GC_R = RQuals.getObjCGCAttr(); 10189 assert((GC_L != GC_R) && "unequal qualifier sets had only equal elements"); 10190 10191 if (GC_L == Qualifiers::Weak || GC_R == Qualifiers::Weak) 10192 return {}; 10193 10194 if (GC_L == Qualifiers::Strong) 10195 return LHS; 10196 if (GC_R == Qualifiers::Strong) 10197 return RHS; 10198 return {}; 10199 } 10200 10201 if (LHSCan->isObjCObjectPointerType() && RHSCan->isObjCObjectPointerType()) { 10202 QualType LHSBaseQT = LHS->castAs<ObjCObjectPointerType>()->getPointeeType(); 10203 QualType RHSBaseQT = RHS->castAs<ObjCObjectPointerType>()->getPointeeType(); 10204 QualType ResQT = mergeObjCGCQualifiers(LHSBaseQT, RHSBaseQT); 10205 if (ResQT == LHSBaseQT) 10206 return LHS; 10207 if (ResQT == RHSBaseQT) 10208 return RHS; 10209 } 10210 return {}; 10211 } 10212 10213 //===----------------------------------------------------------------------===// 10214 // Integer Predicates 10215 //===----------------------------------------------------------------------===// 10216 10217 unsigned ASTContext::getIntWidth(QualType T) const { 10218 if (const auto *ET = T->getAs<EnumType>()) 10219 T = ET->getDecl()->getIntegerType(); 10220 if (T->isBooleanType()) 10221 return 1; 10222 if(const auto *EIT = T->getAs<ExtIntType>()) 10223 return EIT->getNumBits(); 10224 // For builtin types, just use the standard type sizing method 10225 return (unsigned)getTypeSize(T); 10226 } 10227 10228 QualType ASTContext::getCorrespondingUnsignedType(QualType T) const { 10229 assert((T->hasSignedIntegerRepresentation() || T->isSignedFixedPointType()) && 10230 "Unexpected type"); 10231 10232 // Turn <4 x signed int> -> <4 x unsigned int> 10233 if (const auto *VTy = T->getAs<VectorType>()) 10234 return getVectorType(getCorrespondingUnsignedType(VTy->getElementType()), 10235 VTy->getNumElements(), VTy->getVectorKind()); 10236 10237 // For _ExtInt, return an unsigned _ExtInt with same width. 10238 if (const auto *EITy = T->getAs<ExtIntType>()) 10239 return getExtIntType(/*IsUnsigned=*/true, EITy->getNumBits()); 10240 10241 // For enums, get the underlying integer type of the enum, and let the general 10242 // integer type signchanging code handle it. 10243 if (const auto *ETy = T->getAs<EnumType>()) 10244 T = ETy->getDecl()->getIntegerType(); 10245 10246 switch (T->castAs<BuiltinType>()->getKind()) { 10247 case BuiltinType::Char_S: 10248 case BuiltinType::SChar: 10249 return UnsignedCharTy; 10250 case BuiltinType::Short: 10251 return UnsignedShortTy; 10252 case BuiltinType::Int: 10253 return UnsignedIntTy; 10254 case BuiltinType::Long: 10255 return UnsignedLongTy; 10256 case BuiltinType::LongLong: 10257 return UnsignedLongLongTy; 10258 case BuiltinType::Int128: 10259 return UnsignedInt128Ty; 10260 // wchar_t is special. It is either signed or not, but when it's signed, 10261 // there's no matching "unsigned wchar_t". Therefore we return the unsigned 10262 // version of it's underlying type instead. 10263 case BuiltinType::WChar_S: 10264 return getUnsignedWCharType(); 10265 10266 case BuiltinType::ShortAccum: 10267 return UnsignedShortAccumTy; 10268 case BuiltinType::Accum: 10269 return UnsignedAccumTy; 10270 case BuiltinType::LongAccum: 10271 return UnsignedLongAccumTy; 10272 case BuiltinType::SatShortAccum: 10273 return SatUnsignedShortAccumTy; 10274 case BuiltinType::SatAccum: 10275 return SatUnsignedAccumTy; 10276 case BuiltinType::SatLongAccum: 10277 return SatUnsignedLongAccumTy; 10278 case BuiltinType::ShortFract: 10279 return UnsignedShortFractTy; 10280 case BuiltinType::Fract: 10281 return UnsignedFractTy; 10282 case BuiltinType::LongFract: 10283 return UnsignedLongFractTy; 10284 case BuiltinType::SatShortFract: 10285 return SatUnsignedShortFractTy; 10286 case BuiltinType::SatFract: 10287 return SatUnsignedFractTy; 10288 case BuiltinType::SatLongFract: 10289 return SatUnsignedLongFractTy; 10290 default: 10291 llvm_unreachable("Unexpected signed integer or fixed point type"); 10292 } 10293 } 10294 10295 QualType ASTContext::getCorrespondingSignedType(QualType T) const { 10296 assert((T->hasUnsignedIntegerRepresentation() || 10297 T->isUnsignedFixedPointType()) && 10298 "Unexpected type"); 10299 10300 // Turn <4 x unsigned int> -> <4 x signed int> 10301 if (const auto *VTy = T->getAs<VectorType>()) 10302 return getVectorType(getCorrespondingSignedType(VTy->getElementType()), 10303 VTy->getNumElements(), VTy->getVectorKind()); 10304 10305 // For _ExtInt, return a signed _ExtInt with same width. 10306 if (const auto *EITy = T->getAs<ExtIntType>()) 10307 return getExtIntType(/*IsUnsigned=*/false, EITy->getNumBits()); 10308 10309 // For enums, get the underlying integer type of the enum, and let the general 10310 // integer type signchanging code handle it. 10311 if (const auto *ETy = T->getAs<EnumType>()) 10312 T = ETy->getDecl()->getIntegerType(); 10313 10314 switch (T->castAs<BuiltinType>()->getKind()) { 10315 case BuiltinType::Char_U: 10316 case BuiltinType::UChar: 10317 return SignedCharTy; 10318 case BuiltinType::UShort: 10319 return ShortTy; 10320 case BuiltinType::UInt: 10321 return IntTy; 10322 case BuiltinType::ULong: 10323 return LongTy; 10324 case BuiltinType::ULongLong: 10325 return LongLongTy; 10326 case BuiltinType::UInt128: 10327 return Int128Ty; 10328 // wchar_t is special. It is either unsigned or not, but when it's unsigned, 10329 // there's no matching "signed wchar_t". Therefore we return the signed 10330 // version of it's underlying type instead. 10331 case BuiltinType::WChar_U: 10332 return getSignedWCharType(); 10333 10334 case BuiltinType::UShortAccum: 10335 return ShortAccumTy; 10336 case BuiltinType::UAccum: 10337 return AccumTy; 10338 case BuiltinType::ULongAccum: 10339 return LongAccumTy; 10340 case BuiltinType::SatUShortAccum: 10341 return SatShortAccumTy; 10342 case BuiltinType::SatUAccum: 10343 return SatAccumTy; 10344 case BuiltinType::SatULongAccum: 10345 return SatLongAccumTy; 10346 case BuiltinType::UShortFract: 10347 return ShortFractTy; 10348 case BuiltinType::UFract: 10349 return FractTy; 10350 case BuiltinType::ULongFract: 10351 return LongFractTy; 10352 case BuiltinType::SatUShortFract: 10353 return SatShortFractTy; 10354 case BuiltinType::SatUFract: 10355 return SatFractTy; 10356 case BuiltinType::SatULongFract: 10357 return SatLongFractTy; 10358 default: 10359 llvm_unreachable("Unexpected unsigned integer or fixed point type"); 10360 } 10361 } 10362 10363 ASTMutationListener::~ASTMutationListener() = default; 10364 10365 void ASTMutationListener::DeducedReturnType(const FunctionDecl *FD, 10366 QualType ReturnType) {} 10367 10368 //===----------------------------------------------------------------------===// 10369 // Builtin Type Computation 10370 //===----------------------------------------------------------------------===// 10371 10372 /// DecodeTypeFromStr - This decodes one type descriptor from Str, advancing the 10373 /// pointer over the consumed characters. This returns the resultant type. If 10374 /// AllowTypeModifiers is false then modifier like * are not parsed, just basic 10375 /// types. This allows "v2i*" to be parsed as a pointer to a v2i instead of 10376 /// a vector of "i*". 10377 /// 10378 /// RequiresICE is filled in on return to indicate whether the value is required 10379 /// to be an Integer Constant Expression. 10380 static QualType DecodeTypeFromStr(const char *&Str, const ASTContext &Context, 10381 ASTContext::GetBuiltinTypeError &Error, 10382 bool &RequiresICE, 10383 bool AllowTypeModifiers) { 10384 // Modifiers. 10385 int HowLong = 0; 10386 bool Signed = false, Unsigned = false; 10387 RequiresICE = false; 10388 10389 // Read the prefixed modifiers first. 10390 bool Done = false; 10391 #ifndef NDEBUG 10392 bool IsSpecial = false; 10393 #endif 10394 while (!Done) { 10395 switch (*Str++) { 10396 default: Done = true; --Str; break; 10397 case 'I': 10398 RequiresICE = true; 10399 break; 10400 case 'S': 10401 assert(!Unsigned && "Can't use both 'S' and 'U' modifiers!"); 10402 assert(!Signed && "Can't use 'S' modifier multiple times!"); 10403 Signed = true; 10404 break; 10405 case 'U': 10406 assert(!Signed && "Can't use both 'S' and 'U' modifiers!"); 10407 assert(!Unsigned && "Can't use 'U' modifier multiple times!"); 10408 Unsigned = true; 10409 break; 10410 case 'L': 10411 assert(!IsSpecial && "Can't use 'L' with 'W', 'N', 'Z' or 'O' modifiers"); 10412 assert(HowLong <= 2 && "Can't have LLLL modifier"); 10413 ++HowLong; 10414 break; 10415 case 'N': 10416 // 'N' behaves like 'L' for all non LP64 targets and 'int' otherwise. 10417 assert(!IsSpecial && "Can't use two 'N', 'W', 'Z' or 'O' modifiers!"); 10418 assert(HowLong == 0 && "Can't use both 'L' and 'N' modifiers!"); 10419 #ifndef NDEBUG 10420 IsSpecial = true; 10421 #endif 10422 if (Context.getTargetInfo().getLongWidth() == 32) 10423 ++HowLong; 10424 break; 10425 case 'W': 10426 // This modifier represents int64 type. 10427 assert(!IsSpecial && "Can't use two 'N', 'W', 'Z' or 'O' modifiers!"); 10428 assert(HowLong == 0 && "Can't use both 'L' and 'W' modifiers!"); 10429 #ifndef NDEBUG 10430 IsSpecial = true; 10431 #endif 10432 switch (Context.getTargetInfo().getInt64Type()) { 10433 default: 10434 llvm_unreachable("Unexpected integer type"); 10435 case TargetInfo::SignedLong: 10436 HowLong = 1; 10437 break; 10438 case TargetInfo::SignedLongLong: 10439 HowLong = 2; 10440 break; 10441 } 10442 break; 10443 case 'Z': 10444 // This modifier represents int32 type. 10445 assert(!IsSpecial && "Can't use two 'N', 'W', 'Z' or 'O' modifiers!"); 10446 assert(HowLong == 0 && "Can't use both 'L' and 'Z' modifiers!"); 10447 #ifndef NDEBUG 10448 IsSpecial = true; 10449 #endif 10450 switch (Context.getTargetInfo().getIntTypeByWidth(32, true)) { 10451 default: 10452 llvm_unreachable("Unexpected integer type"); 10453 case TargetInfo::SignedInt: 10454 HowLong = 0; 10455 break; 10456 case TargetInfo::SignedLong: 10457 HowLong = 1; 10458 break; 10459 case TargetInfo::SignedLongLong: 10460 HowLong = 2; 10461 break; 10462 } 10463 break; 10464 case 'O': 10465 assert(!IsSpecial && "Can't use two 'N', 'W', 'Z' or 'O' modifiers!"); 10466 assert(HowLong == 0 && "Can't use both 'L' and 'O' modifiers!"); 10467 #ifndef NDEBUG 10468 IsSpecial = true; 10469 #endif 10470 if (Context.getLangOpts().OpenCL) 10471 HowLong = 1; 10472 else 10473 HowLong = 2; 10474 break; 10475 } 10476 } 10477 10478 QualType Type; 10479 10480 // Read the base type. 10481 switch (*Str++) { 10482 default: llvm_unreachable("Unknown builtin type letter!"); 10483 case 'x': 10484 assert(HowLong == 0 && !Signed && !Unsigned && 10485 "Bad modifiers used with 'x'!"); 10486 Type = Context.Float16Ty; 10487 break; 10488 case 'y': 10489 assert(HowLong == 0 && !Signed && !Unsigned && 10490 "Bad modifiers used with 'y'!"); 10491 Type = Context.BFloat16Ty; 10492 break; 10493 case 'v': 10494 assert(HowLong == 0 && !Signed && !Unsigned && 10495 "Bad modifiers used with 'v'!"); 10496 Type = Context.VoidTy; 10497 break; 10498 case 'h': 10499 assert(HowLong == 0 && !Signed && !Unsigned && 10500 "Bad modifiers used with 'h'!"); 10501 Type = Context.HalfTy; 10502 break; 10503 case 'f': 10504 assert(HowLong == 0 && !Signed && !Unsigned && 10505 "Bad modifiers used with 'f'!"); 10506 Type = Context.FloatTy; 10507 break; 10508 case 'd': 10509 assert(HowLong < 3 && !Signed && !Unsigned && 10510 "Bad modifiers used with 'd'!"); 10511 if (HowLong == 1) 10512 Type = Context.LongDoubleTy; 10513 else if (HowLong == 2) 10514 Type = Context.Float128Ty; 10515 else 10516 Type = Context.DoubleTy; 10517 break; 10518 case 's': 10519 assert(HowLong == 0 && "Bad modifiers used with 's'!"); 10520 if (Unsigned) 10521 Type = Context.UnsignedShortTy; 10522 else 10523 Type = Context.ShortTy; 10524 break; 10525 case 'i': 10526 if (HowLong == 3) 10527 Type = Unsigned ? Context.UnsignedInt128Ty : Context.Int128Ty; 10528 else if (HowLong == 2) 10529 Type = Unsigned ? Context.UnsignedLongLongTy : Context.LongLongTy; 10530 else if (HowLong == 1) 10531 Type = Unsigned ? Context.UnsignedLongTy : Context.LongTy; 10532 else 10533 Type = Unsigned ? Context.UnsignedIntTy : Context.IntTy; 10534 break; 10535 case 'c': 10536 assert(HowLong == 0 && "Bad modifiers used with 'c'!"); 10537 if (Signed) 10538 Type = Context.SignedCharTy; 10539 else if (Unsigned) 10540 Type = Context.UnsignedCharTy; 10541 else 10542 Type = Context.CharTy; 10543 break; 10544 case 'b': // boolean 10545 assert(HowLong == 0 && !Signed && !Unsigned && "Bad modifiers for 'b'!"); 10546 Type = Context.BoolTy; 10547 break; 10548 case 'z': // size_t. 10549 assert(HowLong == 0 && !Signed && !Unsigned && "Bad modifiers for 'z'!"); 10550 Type = Context.getSizeType(); 10551 break; 10552 case 'w': // wchar_t. 10553 assert(HowLong == 0 && !Signed && !Unsigned && "Bad modifiers for 'w'!"); 10554 Type = Context.getWideCharType(); 10555 break; 10556 case 'F': 10557 Type = Context.getCFConstantStringType(); 10558 break; 10559 case 'G': 10560 Type = Context.getObjCIdType(); 10561 break; 10562 case 'H': 10563 Type = Context.getObjCSelType(); 10564 break; 10565 case 'M': 10566 Type = Context.getObjCSuperType(); 10567 break; 10568 case 'a': 10569 Type = Context.getBuiltinVaListType(); 10570 assert(!Type.isNull() && "builtin va list type not initialized!"); 10571 break; 10572 case 'A': 10573 // This is a "reference" to a va_list; however, what exactly 10574 // this means depends on how va_list is defined. There are two 10575 // different kinds of va_list: ones passed by value, and ones 10576 // passed by reference. An example of a by-value va_list is 10577 // x86, where va_list is a char*. An example of by-ref va_list 10578 // is x86-64, where va_list is a __va_list_tag[1]. For x86, 10579 // we want this argument to be a char*&; for x86-64, we want 10580 // it to be a __va_list_tag*. 10581 Type = Context.getBuiltinVaListType(); 10582 assert(!Type.isNull() && "builtin va list type not initialized!"); 10583 if (Type->isArrayType()) 10584 Type = Context.getArrayDecayedType(Type); 10585 else 10586 Type = Context.getLValueReferenceType(Type); 10587 break; 10588 case 'q': { 10589 char *End; 10590 unsigned NumElements = strtoul(Str, &End, 10); 10591 assert(End != Str && "Missing vector size"); 10592 Str = End; 10593 10594 QualType ElementType = DecodeTypeFromStr(Str, Context, Error, 10595 RequiresICE, false); 10596 assert(!RequiresICE && "Can't require vector ICE"); 10597 10598 Type = Context.getScalableVectorType(ElementType, NumElements); 10599 break; 10600 } 10601 case 'V': { 10602 char *End; 10603 unsigned NumElements = strtoul(Str, &End, 10); 10604 assert(End != Str && "Missing vector size"); 10605 Str = End; 10606 10607 QualType ElementType = DecodeTypeFromStr(Str, Context, Error, 10608 RequiresICE, false); 10609 assert(!RequiresICE && "Can't require vector ICE"); 10610 10611 // TODO: No way to make AltiVec vectors in builtins yet. 10612 Type = Context.getVectorType(ElementType, NumElements, 10613 VectorType::GenericVector); 10614 break; 10615 } 10616 case 'E': { 10617 char *End; 10618 10619 unsigned NumElements = strtoul(Str, &End, 10); 10620 assert(End != Str && "Missing vector size"); 10621 10622 Str = End; 10623 10624 QualType ElementType = DecodeTypeFromStr(Str, Context, Error, RequiresICE, 10625 false); 10626 Type = Context.getExtVectorType(ElementType, NumElements); 10627 break; 10628 } 10629 case 'X': { 10630 QualType ElementType = DecodeTypeFromStr(Str, Context, Error, RequiresICE, 10631 false); 10632 assert(!RequiresICE && "Can't require complex ICE"); 10633 Type = Context.getComplexType(ElementType); 10634 break; 10635 } 10636 case 'Y': 10637 Type = Context.getPointerDiffType(); 10638 break; 10639 case 'P': 10640 Type = Context.getFILEType(); 10641 if (Type.isNull()) { 10642 Error = ASTContext::GE_Missing_stdio; 10643 return {}; 10644 } 10645 break; 10646 case 'J': 10647 if (Signed) 10648 Type = Context.getsigjmp_bufType(); 10649 else 10650 Type = Context.getjmp_bufType(); 10651 10652 if (Type.isNull()) { 10653 Error = ASTContext::GE_Missing_setjmp; 10654 return {}; 10655 } 10656 break; 10657 case 'K': 10658 assert(HowLong == 0 && !Signed && !Unsigned && "Bad modifiers for 'K'!"); 10659 Type = Context.getucontext_tType(); 10660 10661 if (Type.isNull()) { 10662 Error = ASTContext::GE_Missing_ucontext; 10663 return {}; 10664 } 10665 break; 10666 case 'p': 10667 Type = Context.getProcessIDType(); 10668 break; 10669 } 10670 10671 // If there are modifiers and if we're allowed to parse them, go for it. 10672 Done = !AllowTypeModifiers; 10673 while (!Done) { 10674 switch (char c = *Str++) { 10675 default: Done = true; --Str; break; 10676 case '*': 10677 case '&': { 10678 // Both pointers and references can have their pointee types 10679 // qualified with an address space. 10680 char *End; 10681 unsigned AddrSpace = strtoul(Str, &End, 10); 10682 if (End != Str) { 10683 // Note AddrSpace == 0 is not the same as an unspecified address space. 10684 Type = Context.getAddrSpaceQualType( 10685 Type, 10686 Context.getLangASForBuiltinAddressSpace(AddrSpace)); 10687 Str = End; 10688 } 10689 if (c == '*') 10690 Type = Context.getPointerType(Type); 10691 else 10692 Type = Context.getLValueReferenceType(Type); 10693 break; 10694 } 10695 // FIXME: There's no way to have a built-in with an rvalue ref arg. 10696 case 'C': 10697 Type = Type.withConst(); 10698 break; 10699 case 'D': 10700 Type = Context.getVolatileType(Type); 10701 break; 10702 case 'R': 10703 Type = Type.withRestrict(); 10704 break; 10705 } 10706 } 10707 10708 assert((!RequiresICE || Type->isIntegralOrEnumerationType()) && 10709 "Integer constant 'I' type must be an integer"); 10710 10711 return Type; 10712 } 10713 10714 // On some targets such as PowerPC, some of the builtins are defined with custom 10715 // type descriptors for target-dependent types. These descriptors are decoded in 10716 // other functions, but it may be useful to be able to fall back to default 10717 // descriptor decoding to define builtins mixing target-dependent and target- 10718 // independent types. This function allows decoding one type descriptor with 10719 // default decoding. 10720 QualType ASTContext::DecodeTypeStr(const char *&Str, const ASTContext &Context, 10721 GetBuiltinTypeError &Error, bool &RequireICE, 10722 bool AllowTypeModifiers) const { 10723 return DecodeTypeFromStr(Str, Context, Error, RequireICE, AllowTypeModifiers); 10724 } 10725 10726 /// GetBuiltinType - Return the type for the specified builtin. 10727 QualType ASTContext::GetBuiltinType(unsigned Id, 10728 GetBuiltinTypeError &Error, 10729 unsigned *IntegerConstantArgs) const { 10730 const char *TypeStr = BuiltinInfo.getTypeString(Id); 10731 if (TypeStr[0] == '\0') { 10732 Error = GE_Missing_type; 10733 return {}; 10734 } 10735 10736 SmallVector<QualType, 8> ArgTypes; 10737 10738 bool RequiresICE = false; 10739 Error = GE_None; 10740 QualType ResType = DecodeTypeFromStr(TypeStr, *this, Error, 10741 RequiresICE, true); 10742 if (Error != GE_None) 10743 return {}; 10744 10745 assert(!RequiresICE && "Result of intrinsic cannot be required to be an ICE"); 10746 10747 while (TypeStr[0] && TypeStr[0] != '.') { 10748 QualType Ty = DecodeTypeFromStr(TypeStr, *this, Error, RequiresICE, true); 10749 if (Error != GE_None) 10750 return {}; 10751 10752 // If this argument is required to be an IntegerConstantExpression and the 10753 // caller cares, fill in the bitmask we return. 10754 if (RequiresICE && IntegerConstantArgs) 10755 *IntegerConstantArgs |= 1 << ArgTypes.size(); 10756 10757 // Do array -> pointer decay. The builtin should use the decayed type. 10758 if (Ty->isArrayType()) 10759 Ty = getArrayDecayedType(Ty); 10760 10761 ArgTypes.push_back(Ty); 10762 } 10763 10764 if (Id == Builtin::BI__GetExceptionInfo) 10765 return {}; 10766 10767 assert((TypeStr[0] != '.' || TypeStr[1] == 0) && 10768 "'.' should only occur at end of builtin type list!"); 10769 10770 bool Variadic = (TypeStr[0] == '.'); 10771 10772 FunctionType::ExtInfo EI(getDefaultCallingConvention( 10773 Variadic, /*IsCXXMethod=*/false, /*IsBuiltin=*/true)); 10774 if (BuiltinInfo.isNoReturn(Id)) EI = EI.withNoReturn(true); 10775 10776 10777 // We really shouldn't be making a no-proto type here. 10778 if (ArgTypes.empty() && Variadic && !getLangOpts().CPlusPlus) 10779 return getFunctionNoProtoType(ResType, EI); 10780 10781 FunctionProtoType::ExtProtoInfo EPI; 10782 EPI.ExtInfo = EI; 10783 EPI.Variadic = Variadic; 10784 if (getLangOpts().CPlusPlus && BuiltinInfo.isNoThrow(Id)) 10785 EPI.ExceptionSpec.Type = 10786 getLangOpts().CPlusPlus11 ? EST_BasicNoexcept : EST_DynamicNone; 10787 10788 return getFunctionType(ResType, ArgTypes, EPI); 10789 } 10790 10791 static GVALinkage basicGVALinkageForFunction(const ASTContext &Context, 10792 const FunctionDecl *FD) { 10793 if (!FD->isExternallyVisible()) 10794 return GVA_Internal; 10795 10796 // Non-user-provided functions get emitted as weak definitions with every 10797 // use, no matter whether they've been explicitly instantiated etc. 10798 if (const auto *MD = dyn_cast<CXXMethodDecl>(FD)) 10799 if (!MD->isUserProvided()) 10800 return GVA_DiscardableODR; 10801 10802 GVALinkage External; 10803 switch (FD->getTemplateSpecializationKind()) { 10804 case TSK_Undeclared: 10805 case TSK_ExplicitSpecialization: 10806 External = GVA_StrongExternal; 10807 break; 10808 10809 case TSK_ExplicitInstantiationDefinition: 10810 return GVA_StrongODR; 10811 10812 // C++11 [temp.explicit]p10: 10813 // [ Note: The intent is that an inline function that is the subject of 10814 // an explicit instantiation declaration will still be implicitly 10815 // instantiated when used so that the body can be considered for 10816 // inlining, but that no out-of-line copy of the inline function would be 10817 // generated in the translation unit. -- end note ] 10818 case TSK_ExplicitInstantiationDeclaration: 10819 return GVA_AvailableExternally; 10820 10821 case TSK_ImplicitInstantiation: 10822 External = GVA_DiscardableODR; 10823 break; 10824 } 10825 10826 if (!FD->isInlined()) 10827 return External; 10828 10829 if ((!Context.getLangOpts().CPlusPlus && 10830 !Context.getTargetInfo().getCXXABI().isMicrosoft() && 10831 !FD->hasAttr<DLLExportAttr>()) || 10832 FD->hasAttr<GNUInlineAttr>()) { 10833 // FIXME: This doesn't match gcc's behavior for dllexport inline functions. 10834 10835 // GNU or C99 inline semantics. Determine whether this symbol should be 10836 // externally visible. 10837 if (FD->isInlineDefinitionExternallyVisible()) 10838 return External; 10839 10840 // C99 inline semantics, where the symbol is not externally visible. 10841 return GVA_AvailableExternally; 10842 } 10843 10844 // Functions specified with extern and inline in -fms-compatibility mode 10845 // forcibly get emitted. While the body of the function cannot be later 10846 // replaced, the function definition cannot be discarded. 10847 if (FD->isMSExternInline()) 10848 return GVA_StrongODR; 10849 10850 return GVA_DiscardableODR; 10851 } 10852 10853 static GVALinkage adjustGVALinkageForAttributes(const ASTContext &Context, 10854 const Decl *D, GVALinkage L) { 10855 // See http://msdn.microsoft.com/en-us/library/xa0d9ste.aspx 10856 // dllexport/dllimport on inline functions. 10857 if (D->hasAttr<DLLImportAttr>()) { 10858 if (L == GVA_DiscardableODR || L == GVA_StrongODR) 10859 return GVA_AvailableExternally; 10860 } else if (D->hasAttr<DLLExportAttr>()) { 10861 if (L == GVA_DiscardableODR) 10862 return GVA_StrongODR; 10863 } else if (Context.getLangOpts().CUDA && Context.getLangOpts().CUDAIsDevice) { 10864 // Device-side functions with __global__ attribute must always be 10865 // visible externally so they can be launched from host. 10866 if (D->hasAttr<CUDAGlobalAttr>() && 10867 (L == GVA_DiscardableODR || L == GVA_Internal)) 10868 return GVA_StrongODR; 10869 // Single source offloading languages like CUDA/HIP need to be able to 10870 // access static device variables from host code of the same compilation 10871 // unit. This is done by externalizing the static variable with a shared 10872 // name between the host and device compilation which is the same for the 10873 // same compilation unit whereas different among different compilation 10874 // units. 10875 if (Context.shouldExternalizeStaticVar(D)) 10876 return GVA_StrongExternal; 10877 } 10878 return L; 10879 } 10880 10881 /// Adjust the GVALinkage for a declaration based on what an external AST source 10882 /// knows about whether there can be other definitions of this declaration. 10883 static GVALinkage 10884 adjustGVALinkageForExternalDefinitionKind(const ASTContext &Ctx, const Decl *D, 10885 GVALinkage L) { 10886 ExternalASTSource *Source = Ctx.getExternalSource(); 10887 if (!Source) 10888 return L; 10889 10890 switch (Source->hasExternalDefinitions(D)) { 10891 case ExternalASTSource::EK_Never: 10892 // Other translation units rely on us to provide the definition. 10893 if (L == GVA_DiscardableODR) 10894 return GVA_StrongODR; 10895 break; 10896 10897 case ExternalASTSource::EK_Always: 10898 return GVA_AvailableExternally; 10899 10900 case ExternalASTSource::EK_ReplyHazy: 10901 break; 10902 } 10903 return L; 10904 } 10905 10906 GVALinkage ASTContext::GetGVALinkageForFunction(const FunctionDecl *FD) const { 10907 return adjustGVALinkageForExternalDefinitionKind(*this, FD, 10908 adjustGVALinkageForAttributes(*this, FD, 10909 basicGVALinkageForFunction(*this, FD))); 10910 } 10911 10912 static GVALinkage basicGVALinkageForVariable(const ASTContext &Context, 10913 const VarDecl *VD) { 10914 if (!VD->isExternallyVisible()) 10915 return GVA_Internal; 10916 10917 if (VD->isStaticLocal()) { 10918 const DeclContext *LexicalContext = VD->getParentFunctionOrMethod(); 10919 while (LexicalContext && !isa<FunctionDecl>(LexicalContext)) 10920 LexicalContext = LexicalContext->getLexicalParent(); 10921 10922 // ObjC Blocks can create local variables that don't have a FunctionDecl 10923 // LexicalContext. 10924 if (!LexicalContext) 10925 return GVA_DiscardableODR; 10926 10927 // Otherwise, let the static local variable inherit its linkage from the 10928 // nearest enclosing function. 10929 auto StaticLocalLinkage = 10930 Context.GetGVALinkageForFunction(cast<FunctionDecl>(LexicalContext)); 10931 10932 // Itanium ABI 5.2.2: "Each COMDAT group [for a static local variable] must 10933 // be emitted in any object with references to the symbol for the object it 10934 // contains, whether inline or out-of-line." 10935 // Similar behavior is observed with MSVC. An alternative ABI could use 10936 // StrongODR/AvailableExternally to match the function, but none are 10937 // known/supported currently. 10938 if (StaticLocalLinkage == GVA_StrongODR || 10939 StaticLocalLinkage == GVA_AvailableExternally) 10940 return GVA_DiscardableODR; 10941 return StaticLocalLinkage; 10942 } 10943 10944 // MSVC treats in-class initialized static data members as definitions. 10945 // By giving them non-strong linkage, out-of-line definitions won't 10946 // cause link errors. 10947 if (Context.isMSStaticDataMemberInlineDefinition(VD)) 10948 return GVA_DiscardableODR; 10949 10950 // Most non-template variables have strong linkage; inline variables are 10951 // linkonce_odr or (occasionally, for compatibility) weak_odr. 10952 GVALinkage StrongLinkage; 10953 switch (Context.getInlineVariableDefinitionKind(VD)) { 10954 case ASTContext::InlineVariableDefinitionKind::None: 10955 StrongLinkage = GVA_StrongExternal; 10956 break; 10957 case ASTContext::InlineVariableDefinitionKind::Weak: 10958 case ASTContext::InlineVariableDefinitionKind::WeakUnknown: 10959 StrongLinkage = GVA_DiscardableODR; 10960 break; 10961 case ASTContext::InlineVariableDefinitionKind::Strong: 10962 StrongLinkage = GVA_StrongODR; 10963 break; 10964 } 10965 10966 switch (VD->getTemplateSpecializationKind()) { 10967 case TSK_Undeclared: 10968 return StrongLinkage; 10969 10970 case TSK_ExplicitSpecialization: 10971 return Context.getTargetInfo().getCXXABI().isMicrosoft() && 10972 VD->isStaticDataMember() 10973 ? GVA_StrongODR 10974 : StrongLinkage; 10975 10976 case TSK_ExplicitInstantiationDefinition: 10977 return GVA_StrongODR; 10978 10979 case TSK_ExplicitInstantiationDeclaration: 10980 return GVA_AvailableExternally; 10981 10982 case TSK_ImplicitInstantiation: 10983 return GVA_DiscardableODR; 10984 } 10985 10986 llvm_unreachable("Invalid Linkage!"); 10987 } 10988 10989 GVALinkage ASTContext::GetGVALinkageForVariable(const VarDecl *VD) { 10990 return adjustGVALinkageForExternalDefinitionKind(*this, VD, 10991 adjustGVALinkageForAttributes(*this, VD, 10992 basicGVALinkageForVariable(*this, VD))); 10993 } 10994 10995 bool ASTContext::DeclMustBeEmitted(const Decl *D) { 10996 if (const auto *VD = dyn_cast<VarDecl>(D)) { 10997 if (!VD->isFileVarDecl()) 10998 return false; 10999 // Global named register variables (GNU extension) are never emitted. 11000 if (VD->getStorageClass() == SC_Register) 11001 return false; 11002 if (VD->getDescribedVarTemplate() || 11003 isa<VarTemplatePartialSpecializationDecl>(VD)) 11004 return false; 11005 } else if (const auto *FD = dyn_cast<FunctionDecl>(D)) { 11006 // We never need to emit an uninstantiated function template. 11007 if (FD->getTemplatedKind() == FunctionDecl::TK_FunctionTemplate) 11008 return false; 11009 } else if (isa<PragmaCommentDecl>(D)) 11010 return true; 11011 else if (isa<PragmaDetectMismatchDecl>(D)) 11012 return true; 11013 else if (isa<OMPRequiresDecl>(D)) 11014 return true; 11015 else if (isa<OMPThreadPrivateDecl>(D)) 11016 return !D->getDeclContext()->isDependentContext(); 11017 else if (isa<OMPAllocateDecl>(D)) 11018 return !D->getDeclContext()->isDependentContext(); 11019 else if (isa<OMPDeclareReductionDecl>(D) || isa<OMPDeclareMapperDecl>(D)) 11020 return !D->getDeclContext()->isDependentContext(); 11021 else if (isa<ImportDecl>(D)) 11022 return true; 11023 else 11024 return false; 11025 11026 // If this is a member of a class template, we do not need to emit it. 11027 if (D->getDeclContext()->isDependentContext()) 11028 return false; 11029 11030 // Weak references don't produce any output by themselves. 11031 if (D->hasAttr<WeakRefAttr>()) 11032 return false; 11033 11034 // Aliases and used decls are required. 11035 if (D->hasAttr<AliasAttr>() || D->hasAttr<UsedAttr>()) 11036 return true; 11037 11038 if (const auto *FD = dyn_cast<FunctionDecl>(D)) { 11039 // Forward declarations aren't required. 11040 if (!FD->doesThisDeclarationHaveABody()) 11041 return FD->doesDeclarationForceExternallyVisibleDefinition(); 11042 11043 // Constructors and destructors are required. 11044 if (FD->hasAttr<ConstructorAttr>() || FD->hasAttr<DestructorAttr>()) 11045 return true; 11046 11047 // The key function for a class is required. This rule only comes 11048 // into play when inline functions can be key functions, though. 11049 if (getTargetInfo().getCXXABI().canKeyFunctionBeInline()) { 11050 if (const auto *MD = dyn_cast<CXXMethodDecl>(FD)) { 11051 const CXXRecordDecl *RD = MD->getParent(); 11052 if (MD->isOutOfLine() && RD->isDynamicClass()) { 11053 const CXXMethodDecl *KeyFunc = getCurrentKeyFunction(RD); 11054 if (KeyFunc && KeyFunc->getCanonicalDecl() == MD->getCanonicalDecl()) 11055 return true; 11056 } 11057 } 11058 } 11059 11060 GVALinkage Linkage = GetGVALinkageForFunction(FD); 11061 11062 // static, static inline, always_inline, and extern inline functions can 11063 // always be deferred. Normal inline functions can be deferred in C99/C++. 11064 // Implicit template instantiations can also be deferred in C++. 11065 return !isDiscardableGVALinkage(Linkage); 11066 } 11067 11068 const auto *VD = cast<VarDecl>(D); 11069 assert(VD->isFileVarDecl() && "Expected file scoped var"); 11070 11071 // If the decl is marked as `declare target to`, it should be emitted for the 11072 // host and for the device. 11073 if (LangOpts.OpenMP && 11074 OMPDeclareTargetDeclAttr::isDeclareTargetDeclaration(VD)) 11075 return true; 11076 11077 if (VD->isThisDeclarationADefinition() == VarDecl::DeclarationOnly && 11078 !isMSStaticDataMemberInlineDefinition(VD)) 11079 return false; 11080 11081 // Variables that can be needed in other TUs are required. 11082 auto Linkage = GetGVALinkageForVariable(VD); 11083 if (!isDiscardableGVALinkage(Linkage)) 11084 return true; 11085 11086 // We never need to emit a variable that is available in another TU. 11087 if (Linkage == GVA_AvailableExternally) 11088 return false; 11089 11090 // Variables that have destruction with side-effects are required. 11091 if (VD->needsDestruction(*this)) 11092 return true; 11093 11094 // Variables that have initialization with side-effects are required. 11095 if (VD->getInit() && VD->getInit()->HasSideEffects(*this) && 11096 // We can get a value-dependent initializer during error recovery. 11097 (VD->getInit()->isValueDependent() || !VD->evaluateValue())) 11098 return true; 11099 11100 // Likewise, variables with tuple-like bindings are required if their 11101 // bindings have side-effects. 11102 if (const auto *DD = dyn_cast<DecompositionDecl>(VD)) 11103 for (const auto *BD : DD->bindings()) 11104 if (const auto *BindingVD = BD->getHoldingVar()) 11105 if (DeclMustBeEmitted(BindingVD)) 11106 return true; 11107 11108 return false; 11109 } 11110 11111 void ASTContext::forEachMultiversionedFunctionVersion( 11112 const FunctionDecl *FD, 11113 llvm::function_ref<void(FunctionDecl *)> Pred) const { 11114 assert(FD->isMultiVersion() && "Only valid for multiversioned functions"); 11115 llvm::SmallDenseSet<const FunctionDecl*, 4> SeenDecls; 11116 FD = FD->getMostRecentDecl(); 11117 // FIXME: The order of traversal here matters and depends on the order of 11118 // lookup results, which happens to be (mostly) oldest-to-newest, but we 11119 // shouldn't rely on that. 11120 for (auto *CurDecl : 11121 FD->getDeclContext()->getRedeclContext()->lookup(FD->getDeclName())) { 11122 FunctionDecl *CurFD = CurDecl->getAsFunction()->getMostRecentDecl(); 11123 if (CurFD && hasSameType(CurFD->getType(), FD->getType()) && 11124 std::end(SeenDecls) == llvm::find(SeenDecls, CurFD)) { 11125 SeenDecls.insert(CurFD); 11126 Pred(CurFD); 11127 } 11128 } 11129 } 11130 11131 CallingConv ASTContext::getDefaultCallingConvention(bool IsVariadic, 11132 bool IsCXXMethod, 11133 bool IsBuiltin) const { 11134 // Pass through to the C++ ABI object 11135 if (IsCXXMethod) 11136 return ABI->getDefaultMethodCallConv(IsVariadic); 11137 11138 // Builtins ignore user-specified default calling convention and remain the 11139 // Target's default calling convention. 11140 if (!IsBuiltin) { 11141 switch (LangOpts.getDefaultCallingConv()) { 11142 case LangOptions::DCC_None: 11143 break; 11144 case LangOptions::DCC_CDecl: 11145 return CC_C; 11146 case LangOptions::DCC_FastCall: 11147 if (getTargetInfo().hasFeature("sse2") && !IsVariadic) 11148 return CC_X86FastCall; 11149 break; 11150 case LangOptions::DCC_StdCall: 11151 if (!IsVariadic) 11152 return CC_X86StdCall; 11153 break; 11154 case LangOptions::DCC_VectorCall: 11155 // __vectorcall cannot be applied to variadic functions. 11156 if (!IsVariadic) 11157 return CC_X86VectorCall; 11158 break; 11159 case LangOptions::DCC_RegCall: 11160 // __regcall cannot be applied to variadic functions. 11161 if (!IsVariadic) 11162 return CC_X86RegCall; 11163 break; 11164 } 11165 } 11166 return Target->getDefaultCallingConv(); 11167 } 11168 11169 bool ASTContext::isNearlyEmpty(const CXXRecordDecl *RD) const { 11170 // Pass through to the C++ ABI object 11171 return ABI->isNearlyEmpty(RD); 11172 } 11173 11174 VTableContextBase *ASTContext::getVTableContext() { 11175 if (!VTContext.get()) { 11176 auto ABI = Target->getCXXABI(); 11177 if (ABI.isMicrosoft()) 11178 VTContext.reset(new MicrosoftVTableContext(*this)); 11179 else { 11180 auto ComponentLayout = getLangOpts().RelativeCXXABIVTables 11181 ? ItaniumVTableContext::Relative 11182 : ItaniumVTableContext::Pointer; 11183 VTContext.reset(new ItaniumVTableContext(*this, ComponentLayout)); 11184 } 11185 } 11186 return VTContext.get(); 11187 } 11188 11189 MangleContext *ASTContext::createMangleContext(const TargetInfo *T) { 11190 if (!T) 11191 T = Target; 11192 switch (T->getCXXABI().getKind()) { 11193 case TargetCXXABI::AppleARM64: 11194 case TargetCXXABI::Fuchsia: 11195 case TargetCXXABI::GenericAArch64: 11196 case TargetCXXABI::GenericItanium: 11197 case TargetCXXABI::GenericARM: 11198 case TargetCXXABI::GenericMIPS: 11199 case TargetCXXABI::iOS: 11200 case TargetCXXABI::WebAssembly: 11201 case TargetCXXABI::WatchOS: 11202 case TargetCXXABI::XL: 11203 return ItaniumMangleContext::create(*this, getDiagnostics()); 11204 case TargetCXXABI::Microsoft: 11205 return MicrosoftMangleContext::create(*this, getDiagnostics()); 11206 } 11207 llvm_unreachable("Unsupported ABI"); 11208 } 11209 11210 MangleContext *ASTContext::createDeviceMangleContext(const TargetInfo &T) { 11211 assert(T.getCXXABI().getKind() != TargetCXXABI::Microsoft && 11212 "Device mangle context does not support Microsoft mangling."); 11213 switch (T.getCXXABI().getKind()) { 11214 case TargetCXXABI::AppleARM64: 11215 case TargetCXXABI::Fuchsia: 11216 case TargetCXXABI::GenericAArch64: 11217 case TargetCXXABI::GenericItanium: 11218 case TargetCXXABI::GenericARM: 11219 case TargetCXXABI::GenericMIPS: 11220 case TargetCXXABI::iOS: 11221 case TargetCXXABI::WebAssembly: 11222 case TargetCXXABI::WatchOS: 11223 case TargetCXXABI::XL: 11224 return ItaniumMangleContext::create( 11225 *this, getDiagnostics(), 11226 [](ASTContext &, const NamedDecl *ND) -> llvm::Optional<unsigned> { 11227 if (const auto *RD = dyn_cast<CXXRecordDecl>(ND)) 11228 return RD->getDeviceLambdaManglingNumber(); 11229 return llvm::None; 11230 }); 11231 case TargetCXXABI::Microsoft: 11232 return MicrosoftMangleContext::create(*this, getDiagnostics()); 11233 } 11234 llvm_unreachable("Unsupported ABI"); 11235 } 11236 11237 CXXABI::~CXXABI() = default; 11238 11239 size_t ASTContext::getSideTableAllocatedMemory() const { 11240 return ASTRecordLayouts.getMemorySize() + 11241 llvm::capacity_in_bytes(ObjCLayouts) + 11242 llvm::capacity_in_bytes(KeyFunctions) + 11243 llvm::capacity_in_bytes(ObjCImpls) + 11244 llvm::capacity_in_bytes(BlockVarCopyInits) + 11245 llvm::capacity_in_bytes(DeclAttrs) + 11246 llvm::capacity_in_bytes(TemplateOrInstantiation) + 11247 llvm::capacity_in_bytes(InstantiatedFromUsingDecl) + 11248 llvm::capacity_in_bytes(InstantiatedFromUsingShadowDecl) + 11249 llvm::capacity_in_bytes(InstantiatedFromUnnamedFieldDecl) + 11250 llvm::capacity_in_bytes(OverriddenMethods) + 11251 llvm::capacity_in_bytes(Types) + 11252 llvm::capacity_in_bytes(VariableArrayTypes); 11253 } 11254 11255 /// getIntTypeForBitwidth - 11256 /// sets integer QualTy according to specified details: 11257 /// bitwidth, signed/unsigned. 11258 /// Returns empty type if there is no appropriate target types. 11259 QualType ASTContext::getIntTypeForBitwidth(unsigned DestWidth, 11260 unsigned Signed) const { 11261 TargetInfo::IntType Ty = getTargetInfo().getIntTypeByWidth(DestWidth, Signed); 11262 CanQualType QualTy = getFromTargetType(Ty); 11263 if (!QualTy && DestWidth == 128) 11264 return Signed ? Int128Ty : UnsignedInt128Ty; 11265 return QualTy; 11266 } 11267 11268 /// getRealTypeForBitwidth - 11269 /// sets floating point QualTy according to specified bitwidth. 11270 /// Returns empty type if there is no appropriate target types. 11271 QualType ASTContext::getRealTypeForBitwidth(unsigned DestWidth, 11272 FloatModeKind ExplicitType) const { 11273 FloatModeKind Ty = 11274 getTargetInfo().getRealTypeByWidth(DestWidth, ExplicitType); 11275 switch (Ty) { 11276 case FloatModeKind::Float: 11277 return FloatTy; 11278 case FloatModeKind::Double: 11279 return DoubleTy; 11280 case FloatModeKind::LongDouble: 11281 return LongDoubleTy; 11282 case FloatModeKind::Float128: 11283 return Float128Ty; 11284 case FloatModeKind::Ibm128: 11285 return Ibm128Ty; 11286 case FloatModeKind::NoFloat: 11287 return {}; 11288 } 11289 11290 llvm_unreachable("Unhandled TargetInfo::RealType value"); 11291 } 11292 11293 void ASTContext::setManglingNumber(const NamedDecl *ND, unsigned Number) { 11294 if (Number > 1) 11295 MangleNumbers[ND] = Number; 11296 } 11297 11298 unsigned ASTContext::getManglingNumber(const NamedDecl *ND) const { 11299 auto I = MangleNumbers.find(ND); 11300 return I != MangleNumbers.end() ? I->second : 1; 11301 } 11302 11303 void ASTContext::setStaticLocalNumber(const VarDecl *VD, unsigned Number) { 11304 if (Number > 1) 11305 StaticLocalNumbers[VD] = Number; 11306 } 11307 11308 unsigned ASTContext::getStaticLocalNumber(const VarDecl *VD) const { 11309 auto I = StaticLocalNumbers.find(VD); 11310 return I != StaticLocalNumbers.end() ? I->second : 1; 11311 } 11312 11313 MangleNumberingContext & 11314 ASTContext::getManglingNumberContext(const DeclContext *DC) { 11315 assert(LangOpts.CPlusPlus); // We don't need mangling numbers for plain C. 11316 std::unique_ptr<MangleNumberingContext> &MCtx = MangleNumberingContexts[DC]; 11317 if (!MCtx) 11318 MCtx = createMangleNumberingContext(); 11319 return *MCtx; 11320 } 11321 11322 MangleNumberingContext & 11323 ASTContext::getManglingNumberContext(NeedExtraManglingDecl_t, const Decl *D) { 11324 assert(LangOpts.CPlusPlus); // We don't need mangling numbers for plain C. 11325 std::unique_ptr<MangleNumberingContext> &MCtx = 11326 ExtraMangleNumberingContexts[D]; 11327 if (!MCtx) 11328 MCtx = createMangleNumberingContext(); 11329 return *MCtx; 11330 } 11331 11332 std::unique_ptr<MangleNumberingContext> 11333 ASTContext::createMangleNumberingContext() const { 11334 return ABI->createMangleNumberingContext(); 11335 } 11336 11337 const CXXConstructorDecl * 11338 ASTContext::getCopyConstructorForExceptionObject(CXXRecordDecl *RD) { 11339 return ABI->getCopyConstructorForExceptionObject( 11340 cast<CXXRecordDecl>(RD->getFirstDecl())); 11341 } 11342 11343 void ASTContext::addCopyConstructorForExceptionObject(CXXRecordDecl *RD, 11344 CXXConstructorDecl *CD) { 11345 return ABI->addCopyConstructorForExceptionObject( 11346 cast<CXXRecordDecl>(RD->getFirstDecl()), 11347 cast<CXXConstructorDecl>(CD->getFirstDecl())); 11348 } 11349 11350 void ASTContext::addTypedefNameForUnnamedTagDecl(TagDecl *TD, 11351 TypedefNameDecl *DD) { 11352 return ABI->addTypedefNameForUnnamedTagDecl(TD, DD); 11353 } 11354 11355 TypedefNameDecl * 11356 ASTContext::getTypedefNameForUnnamedTagDecl(const TagDecl *TD) { 11357 return ABI->getTypedefNameForUnnamedTagDecl(TD); 11358 } 11359 11360 void ASTContext::addDeclaratorForUnnamedTagDecl(TagDecl *TD, 11361 DeclaratorDecl *DD) { 11362 return ABI->addDeclaratorForUnnamedTagDecl(TD, DD); 11363 } 11364 11365 DeclaratorDecl *ASTContext::getDeclaratorForUnnamedTagDecl(const TagDecl *TD) { 11366 return ABI->getDeclaratorForUnnamedTagDecl(TD); 11367 } 11368 11369 void ASTContext::setParameterIndex(const ParmVarDecl *D, unsigned int index) { 11370 ParamIndices[D] = index; 11371 } 11372 11373 unsigned ASTContext::getParameterIndex(const ParmVarDecl *D) const { 11374 ParameterIndexTable::const_iterator I = ParamIndices.find(D); 11375 assert(I != ParamIndices.end() && 11376 "ParmIndices lacks entry set by ParmVarDecl"); 11377 return I->second; 11378 } 11379 11380 QualType ASTContext::getStringLiteralArrayType(QualType EltTy, 11381 unsigned Length) const { 11382 // A C++ string literal has a const-qualified element type (C++ 2.13.4p1). 11383 if (getLangOpts().CPlusPlus || getLangOpts().ConstStrings) 11384 EltTy = EltTy.withConst(); 11385 11386 EltTy = adjustStringLiteralBaseType(EltTy); 11387 11388 // Get an array type for the string, according to C99 6.4.5. This includes 11389 // the null terminator character. 11390 return getConstantArrayType(EltTy, llvm::APInt(32, Length + 1), nullptr, 11391 ArrayType::Normal, /*IndexTypeQuals*/ 0); 11392 } 11393 11394 StringLiteral * 11395 ASTContext::getPredefinedStringLiteralFromCache(StringRef Key) const { 11396 StringLiteral *&Result = StringLiteralCache[Key]; 11397 if (!Result) 11398 Result = StringLiteral::Create( 11399 *this, Key, StringLiteral::Ascii, 11400 /*Pascal*/ false, getStringLiteralArrayType(CharTy, Key.size()), 11401 SourceLocation()); 11402 return Result; 11403 } 11404 11405 MSGuidDecl * 11406 ASTContext::getMSGuidDecl(MSGuidDecl::Parts Parts) const { 11407 assert(MSGuidTagDecl && "building MS GUID without MS extensions?"); 11408 11409 llvm::FoldingSetNodeID ID; 11410 MSGuidDecl::Profile(ID, Parts); 11411 11412 void *InsertPos; 11413 if (MSGuidDecl *Existing = MSGuidDecls.FindNodeOrInsertPos(ID, InsertPos)) 11414 return Existing; 11415 11416 QualType GUIDType = getMSGuidType().withConst(); 11417 MSGuidDecl *New = MSGuidDecl::Create(*this, GUIDType, Parts); 11418 MSGuidDecls.InsertNode(New, InsertPos); 11419 return New; 11420 } 11421 11422 TemplateParamObjectDecl * 11423 ASTContext::getTemplateParamObjectDecl(QualType T, const APValue &V) const { 11424 assert(T->isRecordType() && "template param object of unexpected type"); 11425 11426 // C++ [temp.param]p8: 11427 // [...] a static storage duration object of type 'const T' [...] 11428 T.addConst(); 11429 11430 llvm::FoldingSetNodeID ID; 11431 TemplateParamObjectDecl::Profile(ID, T, V); 11432 11433 void *InsertPos; 11434 if (TemplateParamObjectDecl *Existing = 11435 TemplateParamObjectDecls.FindNodeOrInsertPos(ID, InsertPos)) 11436 return Existing; 11437 11438 TemplateParamObjectDecl *New = TemplateParamObjectDecl::Create(*this, T, V); 11439 TemplateParamObjectDecls.InsertNode(New, InsertPos); 11440 return New; 11441 } 11442 11443 bool ASTContext::AtomicUsesUnsupportedLibcall(const AtomicExpr *E) const { 11444 const llvm::Triple &T = getTargetInfo().getTriple(); 11445 if (!T.isOSDarwin()) 11446 return false; 11447 11448 if (!(T.isiOS() && T.isOSVersionLT(7)) && 11449 !(T.isMacOSX() && T.isOSVersionLT(10, 9))) 11450 return false; 11451 11452 QualType AtomicTy = E->getPtr()->getType()->getPointeeType(); 11453 CharUnits sizeChars = getTypeSizeInChars(AtomicTy); 11454 uint64_t Size = sizeChars.getQuantity(); 11455 CharUnits alignChars = getTypeAlignInChars(AtomicTy); 11456 unsigned Align = alignChars.getQuantity(); 11457 unsigned MaxInlineWidthInBits = getTargetInfo().getMaxAtomicInlineWidth(); 11458 return (Size != Align || toBits(sizeChars) > MaxInlineWidthInBits); 11459 } 11460 11461 bool 11462 ASTContext::ObjCMethodsAreEqual(const ObjCMethodDecl *MethodDecl, 11463 const ObjCMethodDecl *MethodImpl) { 11464 // No point trying to match an unavailable/deprecated mothod. 11465 if (MethodDecl->hasAttr<UnavailableAttr>() 11466 || MethodDecl->hasAttr<DeprecatedAttr>()) 11467 return false; 11468 if (MethodDecl->getObjCDeclQualifier() != 11469 MethodImpl->getObjCDeclQualifier()) 11470 return false; 11471 if (!hasSameType(MethodDecl->getReturnType(), MethodImpl->getReturnType())) 11472 return false; 11473 11474 if (MethodDecl->param_size() != MethodImpl->param_size()) 11475 return false; 11476 11477 for (ObjCMethodDecl::param_const_iterator IM = MethodImpl->param_begin(), 11478 IF = MethodDecl->param_begin(), EM = MethodImpl->param_end(), 11479 EF = MethodDecl->param_end(); 11480 IM != EM && IF != EF; ++IM, ++IF) { 11481 const ParmVarDecl *DeclVar = (*IF); 11482 const ParmVarDecl *ImplVar = (*IM); 11483 if (ImplVar->getObjCDeclQualifier() != DeclVar->getObjCDeclQualifier()) 11484 return false; 11485 if (!hasSameType(DeclVar->getType(), ImplVar->getType())) 11486 return false; 11487 } 11488 11489 return (MethodDecl->isVariadic() == MethodImpl->isVariadic()); 11490 } 11491 11492 uint64_t ASTContext::getTargetNullPointerValue(QualType QT) const { 11493 LangAS AS; 11494 if (QT->getUnqualifiedDesugaredType()->isNullPtrType()) 11495 AS = LangAS::Default; 11496 else 11497 AS = QT->getPointeeType().getAddressSpace(); 11498 11499 return getTargetInfo().getNullPointerValue(AS); 11500 } 11501 11502 unsigned ASTContext::getTargetAddressSpace(LangAS AS) const { 11503 if (isTargetAddressSpace(AS)) 11504 return toTargetAddressSpace(AS); 11505 else 11506 return (*AddrSpaceMap)[(unsigned)AS]; 11507 } 11508 11509 QualType ASTContext::getCorrespondingSaturatedType(QualType Ty) const { 11510 assert(Ty->isFixedPointType()); 11511 11512 if (Ty->isSaturatedFixedPointType()) return Ty; 11513 11514 switch (Ty->castAs<BuiltinType>()->getKind()) { 11515 default: 11516 llvm_unreachable("Not a fixed point type!"); 11517 case BuiltinType::ShortAccum: 11518 return SatShortAccumTy; 11519 case BuiltinType::Accum: 11520 return SatAccumTy; 11521 case BuiltinType::LongAccum: 11522 return SatLongAccumTy; 11523 case BuiltinType::UShortAccum: 11524 return SatUnsignedShortAccumTy; 11525 case BuiltinType::UAccum: 11526 return SatUnsignedAccumTy; 11527 case BuiltinType::ULongAccum: 11528 return SatUnsignedLongAccumTy; 11529 case BuiltinType::ShortFract: 11530 return SatShortFractTy; 11531 case BuiltinType::Fract: 11532 return SatFractTy; 11533 case BuiltinType::LongFract: 11534 return SatLongFractTy; 11535 case BuiltinType::UShortFract: 11536 return SatUnsignedShortFractTy; 11537 case BuiltinType::UFract: 11538 return SatUnsignedFractTy; 11539 case BuiltinType::ULongFract: 11540 return SatUnsignedLongFractTy; 11541 } 11542 } 11543 11544 LangAS ASTContext::getLangASForBuiltinAddressSpace(unsigned AS) const { 11545 if (LangOpts.OpenCL) 11546 return getTargetInfo().getOpenCLBuiltinAddressSpace(AS); 11547 11548 if (LangOpts.CUDA) 11549 return getTargetInfo().getCUDABuiltinAddressSpace(AS); 11550 11551 return getLangASFromTargetAS(AS); 11552 } 11553 11554 // Explicitly instantiate this in case a Redeclarable<T> is used from a TU that 11555 // doesn't include ASTContext.h 11556 template 11557 clang::LazyGenerationalUpdatePtr< 11558 const Decl *, Decl *, &ExternalASTSource::CompleteRedeclChain>::ValueType 11559 clang::LazyGenerationalUpdatePtr< 11560 const Decl *, Decl *, &ExternalASTSource::CompleteRedeclChain>::makeValue( 11561 const clang::ASTContext &Ctx, Decl *Value); 11562 11563 unsigned char ASTContext::getFixedPointScale(QualType Ty) const { 11564 assert(Ty->isFixedPointType()); 11565 11566 const TargetInfo &Target = getTargetInfo(); 11567 switch (Ty->castAs<BuiltinType>()->getKind()) { 11568 default: 11569 llvm_unreachable("Not a fixed point type!"); 11570 case BuiltinType::ShortAccum: 11571 case BuiltinType::SatShortAccum: 11572 return Target.getShortAccumScale(); 11573 case BuiltinType::Accum: 11574 case BuiltinType::SatAccum: 11575 return Target.getAccumScale(); 11576 case BuiltinType::LongAccum: 11577 case BuiltinType::SatLongAccum: 11578 return Target.getLongAccumScale(); 11579 case BuiltinType::UShortAccum: 11580 case BuiltinType::SatUShortAccum: 11581 return Target.getUnsignedShortAccumScale(); 11582 case BuiltinType::UAccum: 11583 case BuiltinType::SatUAccum: 11584 return Target.getUnsignedAccumScale(); 11585 case BuiltinType::ULongAccum: 11586 case BuiltinType::SatULongAccum: 11587 return Target.getUnsignedLongAccumScale(); 11588 case BuiltinType::ShortFract: 11589 case BuiltinType::SatShortFract: 11590 return Target.getShortFractScale(); 11591 case BuiltinType::Fract: 11592 case BuiltinType::SatFract: 11593 return Target.getFractScale(); 11594 case BuiltinType::LongFract: 11595 case BuiltinType::SatLongFract: 11596 return Target.getLongFractScale(); 11597 case BuiltinType::UShortFract: 11598 case BuiltinType::SatUShortFract: 11599 return Target.getUnsignedShortFractScale(); 11600 case BuiltinType::UFract: 11601 case BuiltinType::SatUFract: 11602 return Target.getUnsignedFractScale(); 11603 case BuiltinType::ULongFract: 11604 case BuiltinType::SatULongFract: 11605 return Target.getUnsignedLongFractScale(); 11606 } 11607 } 11608 11609 unsigned char ASTContext::getFixedPointIBits(QualType Ty) const { 11610 assert(Ty->isFixedPointType()); 11611 11612 const TargetInfo &Target = getTargetInfo(); 11613 switch (Ty->castAs<BuiltinType>()->getKind()) { 11614 default: 11615 llvm_unreachable("Not a fixed point type!"); 11616 case BuiltinType::ShortAccum: 11617 case BuiltinType::SatShortAccum: 11618 return Target.getShortAccumIBits(); 11619 case BuiltinType::Accum: 11620 case BuiltinType::SatAccum: 11621 return Target.getAccumIBits(); 11622 case BuiltinType::LongAccum: 11623 case BuiltinType::SatLongAccum: 11624 return Target.getLongAccumIBits(); 11625 case BuiltinType::UShortAccum: 11626 case BuiltinType::SatUShortAccum: 11627 return Target.getUnsignedShortAccumIBits(); 11628 case BuiltinType::UAccum: 11629 case BuiltinType::SatUAccum: 11630 return Target.getUnsignedAccumIBits(); 11631 case BuiltinType::ULongAccum: 11632 case BuiltinType::SatULongAccum: 11633 return Target.getUnsignedLongAccumIBits(); 11634 case BuiltinType::ShortFract: 11635 case BuiltinType::SatShortFract: 11636 case BuiltinType::Fract: 11637 case BuiltinType::SatFract: 11638 case BuiltinType::LongFract: 11639 case BuiltinType::SatLongFract: 11640 case BuiltinType::UShortFract: 11641 case BuiltinType::SatUShortFract: 11642 case BuiltinType::UFract: 11643 case BuiltinType::SatUFract: 11644 case BuiltinType::ULongFract: 11645 case BuiltinType::SatULongFract: 11646 return 0; 11647 } 11648 } 11649 11650 llvm::FixedPointSemantics 11651 ASTContext::getFixedPointSemantics(QualType Ty) const { 11652 assert((Ty->isFixedPointType() || Ty->isIntegerType()) && 11653 "Can only get the fixed point semantics for a " 11654 "fixed point or integer type."); 11655 if (Ty->isIntegerType()) 11656 return llvm::FixedPointSemantics::GetIntegerSemantics( 11657 getIntWidth(Ty), Ty->isSignedIntegerType()); 11658 11659 bool isSigned = Ty->isSignedFixedPointType(); 11660 return llvm::FixedPointSemantics( 11661 static_cast<unsigned>(getTypeSize(Ty)), getFixedPointScale(Ty), isSigned, 11662 Ty->isSaturatedFixedPointType(), 11663 !isSigned && getTargetInfo().doUnsignedFixedPointTypesHavePadding()); 11664 } 11665 11666 llvm::APFixedPoint ASTContext::getFixedPointMax(QualType Ty) const { 11667 assert(Ty->isFixedPointType()); 11668 return llvm::APFixedPoint::getMax(getFixedPointSemantics(Ty)); 11669 } 11670 11671 llvm::APFixedPoint ASTContext::getFixedPointMin(QualType Ty) const { 11672 assert(Ty->isFixedPointType()); 11673 return llvm::APFixedPoint::getMin(getFixedPointSemantics(Ty)); 11674 } 11675 11676 QualType ASTContext::getCorrespondingSignedFixedPointType(QualType Ty) const { 11677 assert(Ty->isUnsignedFixedPointType() && 11678 "Expected unsigned fixed point type"); 11679 11680 switch (Ty->castAs<BuiltinType>()->getKind()) { 11681 case BuiltinType::UShortAccum: 11682 return ShortAccumTy; 11683 case BuiltinType::UAccum: 11684 return AccumTy; 11685 case BuiltinType::ULongAccum: 11686 return LongAccumTy; 11687 case BuiltinType::SatUShortAccum: 11688 return SatShortAccumTy; 11689 case BuiltinType::SatUAccum: 11690 return SatAccumTy; 11691 case BuiltinType::SatULongAccum: 11692 return SatLongAccumTy; 11693 case BuiltinType::UShortFract: 11694 return ShortFractTy; 11695 case BuiltinType::UFract: 11696 return FractTy; 11697 case BuiltinType::ULongFract: 11698 return LongFractTy; 11699 case BuiltinType::SatUShortFract: 11700 return SatShortFractTy; 11701 case BuiltinType::SatUFract: 11702 return SatFractTy; 11703 case BuiltinType::SatULongFract: 11704 return SatLongFractTy; 11705 default: 11706 llvm_unreachable("Unexpected unsigned fixed point type"); 11707 } 11708 } 11709 11710 ParsedTargetAttr 11711 ASTContext::filterFunctionTargetAttrs(const TargetAttr *TD) const { 11712 assert(TD != nullptr); 11713 ParsedTargetAttr ParsedAttr = TD->parse(); 11714 11715 llvm::erase_if(ParsedAttr.Features, [&](const std::string &Feat) { 11716 return !Target->isValidFeatureName(StringRef{Feat}.substr(1)); 11717 }); 11718 return ParsedAttr; 11719 } 11720 11721 void ASTContext::getFunctionFeatureMap(llvm::StringMap<bool> &FeatureMap, 11722 const FunctionDecl *FD) const { 11723 if (FD) 11724 getFunctionFeatureMap(FeatureMap, GlobalDecl().getWithDecl(FD)); 11725 else 11726 Target->initFeatureMap(FeatureMap, getDiagnostics(), 11727 Target->getTargetOpts().CPU, 11728 Target->getTargetOpts().Features); 11729 } 11730 11731 // Fills in the supplied string map with the set of target features for the 11732 // passed in function. 11733 void ASTContext::getFunctionFeatureMap(llvm::StringMap<bool> &FeatureMap, 11734 GlobalDecl GD) const { 11735 StringRef TargetCPU = Target->getTargetOpts().CPU; 11736 const FunctionDecl *FD = GD.getDecl()->getAsFunction(); 11737 if (const auto *TD = FD->getAttr<TargetAttr>()) { 11738 ParsedTargetAttr ParsedAttr = filterFunctionTargetAttrs(TD); 11739 11740 // Make a copy of the features as passed on the command line into the 11741 // beginning of the additional features from the function to override. 11742 ParsedAttr.Features.insert( 11743 ParsedAttr.Features.begin(), 11744 Target->getTargetOpts().FeaturesAsWritten.begin(), 11745 Target->getTargetOpts().FeaturesAsWritten.end()); 11746 11747 if (ParsedAttr.Architecture != "" && 11748 Target->isValidCPUName(ParsedAttr.Architecture)) 11749 TargetCPU = ParsedAttr.Architecture; 11750 11751 // Now populate the feature map, first with the TargetCPU which is either 11752 // the default or a new one from the target attribute string. Then we'll use 11753 // the passed in features (FeaturesAsWritten) along with the new ones from 11754 // the attribute. 11755 Target->initFeatureMap(FeatureMap, getDiagnostics(), TargetCPU, 11756 ParsedAttr.Features); 11757 } else if (const auto *SD = FD->getAttr<CPUSpecificAttr>()) { 11758 llvm::SmallVector<StringRef, 32> FeaturesTmp; 11759 Target->getCPUSpecificCPUDispatchFeatures( 11760 SD->getCPUName(GD.getMultiVersionIndex())->getName(), FeaturesTmp); 11761 std::vector<std::string> Features(FeaturesTmp.begin(), FeaturesTmp.end()); 11762 Features.insert(Features.begin(), 11763 Target->getTargetOpts().FeaturesAsWritten.begin(), 11764 Target->getTargetOpts().FeaturesAsWritten.end()); 11765 Target->initFeatureMap(FeatureMap, getDiagnostics(), TargetCPU, Features); 11766 } else { 11767 FeatureMap = Target->getTargetOpts().FeatureMap; 11768 } 11769 } 11770 11771 OMPTraitInfo &ASTContext::getNewOMPTraitInfo() { 11772 OMPTraitInfoVector.emplace_back(new OMPTraitInfo()); 11773 return *OMPTraitInfoVector.back(); 11774 } 11775 11776 const StreamingDiagnostic &clang:: 11777 operator<<(const StreamingDiagnostic &DB, 11778 const ASTContext::SectionInfo &Section) { 11779 if (Section.Decl) 11780 return DB << Section.Decl; 11781 return DB << "a prior #pragma section"; 11782 } 11783 11784 bool ASTContext::mayExternalizeStaticVar(const Decl *D) const { 11785 bool IsStaticVar = 11786 isa<VarDecl>(D) && cast<VarDecl>(D)->getStorageClass() == SC_Static; 11787 bool IsExplicitDeviceVar = (D->hasAttr<CUDADeviceAttr>() && 11788 !D->getAttr<CUDADeviceAttr>()->isImplicit()) || 11789 (D->hasAttr<CUDAConstantAttr>() && 11790 !D->getAttr<CUDAConstantAttr>()->isImplicit()); 11791 // CUDA/HIP: static managed variables need to be externalized since it is 11792 // a declaration in IR, therefore cannot have internal linkage. 11793 return IsStaticVar && 11794 (D->hasAttr<HIPManagedAttr>() || IsExplicitDeviceVar); 11795 } 11796 11797 bool ASTContext::shouldExternalizeStaticVar(const Decl *D) const { 11798 return mayExternalizeStaticVar(D) && 11799 (D->hasAttr<HIPManagedAttr>() || 11800 CUDADeviceVarODRUsedByHost.count(cast<VarDecl>(D))); 11801 } 11802 11803 StringRef ASTContext::getCUIDHash() const { 11804 if (!CUIDHash.empty()) 11805 return CUIDHash; 11806 if (LangOpts.CUID.empty()) 11807 return StringRef(); 11808 CUIDHash = llvm::utohexstr(llvm::MD5Hash(LangOpts.CUID), /*LowerCase=*/true); 11809 return CUIDHash; 11810 } 11811