clang  14.0.0git
ASTContext.cpp
Go to the documentation of this file.
1 //===- ASTContext.cpp - Context to hold long-lived AST nodes --------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file implements the ASTContext interface.
10 //
11 //===----------------------------------------------------------------------===//
12 
13 #include "clang/AST/ASTContext.h"
14 #include "CXXABI.h"
15 #include "Interp/Context.h"
16 #include "clang/AST/APValue.h"
17 #include "clang/AST/ASTConcept.h"
20 #include "clang/AST/Attr.h"
21 #include "clang/AST/AttrIterator.h"
22 #include "clang/AST/CharUnits.h"
23 #include "clang/AST/Comment.h"
24 #include "clang/AST/Decl.h"
25 #include "clang/AST/DeclBase.h"
26 #include "clang/AST/DeclCXX.h"
28 #include "clang/AST/DeclObjC.h"
29 #include "clang/AST/DeclOpenMP.h"
30 #include "clang/AST/DeclTemplate.h"
33 #include "clang/AST/Expr.h"
34 #include "clang/AST/ExprCXX.h"
35 #include "clang/AST/ExprConcepts.h"
37 #include "clang/AST/Mangle.h"
42 #include "clang/AST/RecordLayout.h"
43 #include "clang/AST/Stmt.h"
44 #include "clang/AST/TemplateBase.h"
45 #include "clang/AST/TemplateName.h"
46 #include "clang/AST/Type.h"
47 #include "clang/AST/TypeLoc.h"
51 #include "clang/Basic/Builtins.h"
55 #include "clang/Basic/LLVM.h"
57 #include "clang/Basic/Linkage.h"
58 #include "clang/Basic/Module.h"
63 #include "clang/Basic/Specifiers.h"
65 #include "clang/Basic/TargetInfo.h"
66 #include "clang/Basic/XRayLists.h"
67 #include "llvm/ADT/APFixedPoint.h"
68 #include "llvm/ADT/APInt.h"
69 #include "llvm/ADT/APSInt.h"
70 #include "llvm/ADT/ArrayRef.h"
71 #include "llvm/ADT/DenseMap.h"
72 #include "llvm/ADT/DenseSet.h"
73 #include "llvm/ADT/FoldingSet.h"
74 #include "llvm/ADT/None.h"
75 #include "llvm/ADT/Optional.h"
76 #include "llvm/ADT/PointerUnion.h"
77 #include "llvm/ADT/STLExtras.h"
78 #include "llvm/ADT/SmallPtrSet.h"
79 #include "llvm/ADT/SmallVector.h"
80 #include "llvm/ADT/StringExtras.h"
81 #include "llvm/ADT/StringRef.h"
82 #include "llvm/ADT/Triple.h"
83 #include "llvm/Support/Capacity.h"
84 #include "llvm/Support/Casting.h"
85 #include "llvm/Support/Compiler.h"
86 #include "llvm/Support/ErrorHandling.h"
87 #include "llvm/Support/MD5.h"
88 #include "llvm/Support/MathExtras.h"
89 #include "llvm/Support/raw_ostream.h"
90 #include <algorithm>
91 #include <cassert>
92 #include <cstddef>
93 #include <cstdint>
94 #include <cstdlib>
95 #include <map>
96 #include <memory>
97 #include <string>
98 #include <tuple>
99 #include <utility>
100 
101 using namespace clang;
102 
112 };
113 
114 /// \returns location that is relevant when searching for Doc comments related
115 /// to \p D.
117  SourceManager &SourceMgr) {
118  assert(D);
119 
120  // User can not attach documentation to implicit declarations.
121  if (D->isImplicit())
122  return {};
123 
124  // User can not attach documentation to implicit instantiations.
125  if (const auto *FD = dyn_cast<FunctionDecl>(D)) {
126  if (FD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation)
127  return {};
128  }
129 
130  if (const auto *VD = dyn_cast<VarDecl>(D)) {
131  if (VD->isStaticDataMember() &&
132  VD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation)
133  return {};
134  }
135 
136  if (const auto *CRD = dyn_cast<CXXRecordDecl>(D)) {
137  if (CRD->getTemplateSpecializationKind() == TSK_ImplicitInstantiation)
138  return {};
139  }
140 
141  if (const auto *CTSD = dyn_cast<ClassTemplateSpecializationDecl>(D)) {
142  TemplateSpecializationKind TSK = CTSD->getSpecializationKind();
143  if (TSK == TSK_ImplicitInstantiation ||
144  TSK == TSK_Undeclared)
145  return {};
146  }
147 
148  if (const auto *ED = dyn_cast<EnumDecl>(D)) {
149  if (ED->getTemplateSpecializationKind() == TSK_ImplicitInstantiation)
150  return {};
151  }
152  if (const auto *TD = dyn_cast<TagDecl>(D)) {
153  // When tag declaration (but not definition!) is part of the
154  // decl-specifier-seq of some other declaration, it doesn't get comment
155  if (TD->isEmbeddedInDeclarator() && !TD->isCompleteDefinition())
156  return {};
157  }
158  // TODO: handle comments for function parameters properly.
159  if (isa<ParmVarDecl>(D))
160  return {};
161 
162  // TODO: we could look up template parameter documentation in the template
163  // documentation.
164  if (isa<TemplateTypeParmDecl>(D) ||
165  isa<NonTypeTemplateParmDecl>(D) ||
166  isa<TemplateTemplateParmDecl>(D))
167  return {};
168 
169  // Find declaration location.
170  // For Objective-C declarations we generally don't expect to have multiple
171  // declarators, thus use declaration starting location as the "declaration
172  // location".
173  // For all other declarations multiple declarators are used quite frequently,
174  // so we use the location of the identifier as the "declaration location".
175  if (isa<ObjCMethodDecl>(D) || isa<ObjCContainerDecl>(D) ||
176  isa<ObjCPropertyDecl>(D) ||
177  isa<RedeclarableTemplateDecl>(D) ||
178  isa<ClassTemplateSpecializationDecl>(D) ||
179  // Allow association with Y across {} in `typedef struct X {} Y`.
180  isa<TypedefDecl>(D))
181  return D->getBeginLoc();
182 
183  const SourceLocation DeclLoc = D->getLocation();
184  if (DeclLoc.isMacroID()) {
185  if (isa<TypedefDecl>(D)) {
186  // If location of the typedef name is in a macro, it is because being
187  // declared via a macro. Try using declaration's starting location as
188  // the "declaration location".
189  return D->getBeginLoc();
190  }
191 
192  if (const auto *TD = dyn_cast<TagDecl>(D)) {
193  // If location of the tag decl is inside a macro, but the spelling of
194  // the tag name comes from a macro argument, it looks like a special
195  // macro like NS_ENUM is being used to define the tag decl. In that
196  // case, adjust the source location to the expansion loc so that we can
197  // attach the comment to the tag decl.
198  if (SourceMgr.isMacroArgExpansion(DeclLoc) && TD->isCompleteDefinition())
199  return SourceMgr.getExpansionLoc(DeclLoc);
200  }
201  }
202 
203  return DeclLoc;
204 }
205 
207  const Decl *D, const SourceLocation RepresentativeLocForDecl,
208  const std::map<unsigned, RawComment *> &CommentsInTheFile) const {
209  // If the declaration doesn't map directly to a location in a file, we
210  // can't find the comment.
211  if (RepresentativeLocForDecl.isInvalid() ||
212  !RepresentativeLocForDecl.isFileID())
213  return nullptr;
214 
215  // If there are no comments anywhere, we won't find anything.
216  if (CommentsInTheFile.empty())
217  return nullptr;
218 
219  // Decompose the location for the declaration and find the beginning of the
220  // file buffer.
221  const std::pair<FileID, unsigned> DeclLocDecomp =
222  SourceMgr.getDecomposedLoc(RepresentativeLocForDecl);
223 
224  // Slow path.
225  auto OffsetCommentBehindDecl =
226  CommentsInTheFile.lower_bound(DeclLocDecomp.second);
227 
228  // First check whether we have a trailing comment.
229  if (OffsetCommentBehindDecl != CommentsInTheFile.end()) {
230  RawComment *CommentBehindDecl = OffsetCommentBehindDecl->second;
231  if ((CommentBehindDecl->isDocumentation() ||
232  LangOpts.CommentOpts.ParseAllComments) &&
233  CommentBehindDecl->isTrailingComment() &&
234  (isa<FieldDecl>(D) || isa<EnumConstantDecl>(D) || isa<VarDecl>(D) ||
235  isa<ObjCMethodDecl>(D) || isa<ObjCPropertyDecl>(D))) {
236 
237  // Check that Doxygen trailing comment comes after the declaration, starts
238  // on the same line and in the same file as the declaration.
239  if (SourceMgr.getLineNumber(DeclLocDecomp.first, DeclLocDecomp.second) ==
240  Comments.getCommentBeginLine(CommentBehindDecl, DeclLocDecomp.first,
241  OffsetCommentBehindDecl->first)) {
242  return CommentBehindDecl;
243  }
244  }
245  }
246 
247  // The comment just after the declaration was not a trailing comment.
248  // Let's look at the previous comment.
249  if (OffsetCommentBehindDecl == CommentsInTheFile.begin())
250  return nullptr;
251 
252  auto OffsetCommentBeforeDecl = --OffsetCommentBehindDecl;
253  RawComment *CommentBeforeDecl = OffsetCommentBeforeDecl->second;
254 
255  // Check that we actually have a non-member Doxygen comment.
256  if (!(CommentBeforeDecl->isDocumentation() ||
257  LangOpts.CommentOpts.ParseAllComments) ||
258  CommentBeforeDecl->isTrailingComment())
259  return nullptr;
260 
261  // Decompose the end of the comment.
262  const unsigned CommentEndOffset =
263  Comments.getCommentEndOffset(CommentBeforeDecl);
264 
265  // Get the corresponding buffer.
266  bool Invalid = false;
267  const char *Buffer = SourceMgr.getBufferData(DeclLocDecomp.first,
268  &Invalid).data();
269  if (Invalid)
270  return nullptr;
271 
272  // Extract text between the comment and declaration.
273  StringRef Text(Buffer + CommentEndOffset,
274  DeclLocDecomp.second - CommentEndOffset);
275 
276  // There should be no other declarations or preprocessor directives between
277  // comment and declaration.
278  if (Text.find_first_of(";{}#@") != StringRef::npos)
279  return nullptr;
280 
281  return CommentBeforeDecl;
282 }
283 
285  const SourceLocation DeclLoc = getDeclLocForCommentSearch(D, SourceMgr);
286 
287  // If the declaration doesn't map directly to a location in a file, we
288  // can't find the comment.
289  if (DeclLoc.isInvalid() || !DeclLoc.isFileID())
290  return nullptr;
291 
292  if (ExternalSource && !CommentsLoaded) {
293  ExternalSource->ReadComments();
294  CommentsLoaded = true;
295  }
296 
297  if (Comments.empty())
298  return nullptr;
299 
300  const FileID File = SourceMgr.getDecomposedLoc(DeclLoc).first;
301  const auto CommentsInThisFile = Comments.getCommentsInFile(File);
302  if (!CommentsInThisFile || CommentsInThisFile->empty())
303  return nullptr;
304 
305  return getRawCommentForDeclNoCacheImpl(D, DeclLoc, *CommentsInThisFile);
306 }
307 
309  assert(LangOpts.RetainCommentsFromSystemHeaders ||
310  !SourceMgr.isInSystemHeader(RC.getSourceRange().getBegin()));
311  Comments.addComment(RC, LangOpts.CommentOpts, BumpAlloc);
312 }
313 
314 /// If we have a 'templated' declaration for a template, adjust 'D' to
315 /// refer to the actual template.
316 /// If we have an implicit instantiation, adjust 'D' to refer to template.
317 static const Decl &adjustDeclToTemplate(const Decl &D) {
318  if (const auto *FD = dyn_cast<FunctionDecl>(&D)) {
319  // Is this function declaration part of a function template?
320  if (const FunctionTemplateDecl *FTD = FD->getDescribedFunctionTemplate())
321  return *FTD;
322 
323  // Nothing to do if function is not an implicit instantiation.
324  if (FD->getTemplateSpecializationKind() != TSK_ImplicitInstantiation)
325  return D;
326 
327  // Function is an implicit instantiation of a function template?
328  if (const FunctionTemplateDecl *FTD = FD->getPrimaryTemplate())
329  return *FTD;
330 
331  // Function is instantiated from a member definition of a class template?
332  if (const FunctionDecl *MemberDecl =
334  return *MemberDecl;
335 
336  return D;
337  }
338  if (const auto *VD = dyn_cast<VarDecl>(&D)) {
339  // Static data member is instantiated from a member definition of a class
340  // template?
341  if (VD->isStaticDataMember())
342  if (const VarDecl *MemberDecl = VD->getInstantiatedFromStaticDataMember())
343  return *MemberDecl;
344 
345  return D;
346  }
347  if (const auto *CRD = dyn_cast<CXXRecordDecl>(&D)) {
348  // Is this class declaration part of a class template?
349  if (const ClassTemplateDecl *CTD = CRD->getDescribedClassTemplate())
350  return *CTD;
351 
352  // Class is an implicit instantiation of a class template or partial
353  // specialization?
354  if (const auto *CTSD = dyn_cast<ClassTemplateSpecializationDecl>(CRD)) {
355  if (CTSD->getSpecializationKind() != TSK_ImplicitInstantiation)
356  return D;
357  llvm::PointerUnion<ClassTemplateDecl *,
359  PU = CTSD->getSpecializedTemplateOrPartial();
360  return PU.is<ClassTemplateDecl *>()
361  ? *static_cast<const Decl *>(PU.get<ClassTemplateDecl *>())
362  : *static_cast<const Decl *>(
364  }
365 
366  // Class is instantiated from a member definition of a class template?
367  if (const MemberSpecializationInfo *Info =
368  CRD->getMemberSpecializationInfo())
369  return *Info->getInstantiatedFrom();
370 
371  return D;
372  }
373  if (const auto *ED = dyn_cast<EnumDecl>(&D)) {
374  // Enum is instantiated from a member definition of a class template?
375  if (const EnumDecl *MemberDecl = ED->getInstantiatedFromMemberEnum())
376  return *MemberDecl;
377 
378  return D;
379  }
380  // FIXME: Adjust alias templates?
381  return D;
382 }
383 
385  const Decl *D,
386  const Decl **OriginalDecl) const {
387  if (!D) {
388  if (OriginalDecl)
389  OriginalDecl = nullptr;
390  return nullptr;
391  }
392 
393  D = &adjustDeclToTemplate(*D);
394 
395  // Any comment directly attached to D?
396  {
397  auto DeclComment = DeclRawComments.find(D);
398  if (DeclComment != DeclRawComments.end()) {
399  if (OriginalDecl)
400  *OriginalDecl = D;
401  return DeclComment->second;
402  }
403  }
404 
405  // Any comment attached to any redeclaration of D?
406  const Decl *CanonicalD = D->getCanonicalDecl();
407  if (!CanonicalD)
408  return nullptr;
409 
410  {
411  auto RedeclComment = RedeclChainComments.find(CanonicalD);
412  if (RedeclComment != RedeclChainComments.end()) {
413  if (OriginalDecl)
414  *OriginalDecl = RedeclComment->second;
415  auto CommentAtRedecl = DeclRawComments.find(RedeclComment->second);
416  assert(CommentAtRedecl != DeclRawComments.end() &&
417  "This decl is supposed to have comment attached.");
418  return CommentAtRedecl->second;
419  }
420  }
421 
422  // Any redeclarations of D that we haven't checked for comments yet?
423  // We can't use DenseMap::iterator directly since it'd get invalid.
424  auto LastCheckedRedecl = [this, CanonicalD]() -> const Decl * {
425  auto LookupRes = CommentlessRedeclChains.find(CanonicalD);
426  if (LookupRes != CommentlessRedeclChains.end())
427  return LookupRes->second;
428  return nullptr;
429  }();
430 
431  for (const auto Redecl : D->redecls()) {
432  assert(Redecl);
433  // Skip all redeclarations that have been checked previously.
434  if (LastCheckedRedecl) {
435  if (LastCheckedRedecl == Redecl) {
436  LastCheckedRedecl = nullptr;
437  }
438  continue;
439  }
440  const RawComment *RedeclComment = getRawCommentForDeclNoCache(Redecl);
441  if (RedeclComment) {
442  cacheRawCommentForDecl(*Redecl, *RedeclComment);
443  if (OriginalDecl)
444  *OriginalDecl = Redecl;
445  return RedeclComment;
446  }
447  CommentlessRedeclChains[CanonicalD] = Redecl;
448  }
449 
450  if (OriginalDecl)
451  *OriginalDecl = nullptr;
452  return nullptr;
453 }
454 
456  const RawComment &Comment) const {
457  assert(Comment.isDocumentation() || LangOpts.CommentOpts.ParseAllComments);
458  DeclRawComments.try_emplace(&OriginalD, &Comment);
459  const Decl *const CanonicalDecl = OriginalD.getCanonicalDecl();
460  RedeclChainComments.try_emplace(CanonicalDecl, &OriginalD);
461  CommentlessRedeclChains.erase(CanonicalDecl);
462 }
463 
464 static void addRedeclaredMethods(const ObjCMethodDecl *ObjCMethod,
466  const DeclContext *DC = ObjCMethod->getDeclContext();
467  if (const auto *IMD = dyn_cast<ObjCImplDecl>(DC)) {
468  const ObjCInterfaceDecl *ID = IMD->getClassInterface();
469  if (!ID)
470  return;
471  // Add redeclared method here.
472  for (const auto *Ext : ID->known_extensions()) {
473  if (ObjCMethodDecl *RedeclaredMethod =
474  Ext->getMethod(ObjCMethod->getSelector(),
475  ObjCMethod->isInstanceMethod()))
476  Redeclared.push_back(RedeclaredMethod);
477  }
478  }
479 }
480 
482  const Preprocessor *PP) {
483  if (Comments.empty() || Decls.empty())
484  return;
485 
486  FileID File;
487  for (Decl *D : Decls) {
488  SourceLocation Loc = D->getLocation();
489  if (Loc.isValid()) {
490  // See if there are any new comments that are not attached to a decl.
491  // The location doesn't have to be precise - we care only about the file.
492  File = SourceMgr.getDecomposedLoc(Loc).first;
493  break;
494  }
495  }
496 
497  if (File.isInvalid())
498  return;
499 
500  auto CommentsInThisFile = Comments.getCommentsInFile(File);
501  if (!CommentsInThisFile || CommentsInThisFile->empty() ||
502  CommentsInThisFile->rbegin()->second->isAttached())
503  return;
504 
505  // There is at least one comment not attached to a decl.
506  // Maybe it should be attached to one of Decls?
507  //
508  // Note that this way we pick up not only comments that precede the
509  // declaration, but also comments that *follow* the declaration -- thanks to
510  // the lookahead in the lexer: we've consumed the semicolon and looked
511  // ahead through comments.
512 
513  for (const Decl *D : Decls) {
514  assert(D);
515  if (D->isInvalidDecl())
516  continue;
517 
518  D = &adjustDeclToTemplate(*D);
519 
520  const SourceLocation DeclLoc = getDeclLocForCommentSearch(D, SourceMgr);
521 
522  if (DeclLoc.isInvalid() || !DeclLoc.isFileID())
523  continue;
524 
525  if (DeclRawComments.count(D) > 0)
526  continue;
527 
528  if (RawComment *const DocComment =
529  getRawCommentForDeclNoCacheImpl(D, DeclLoc, *CommentsInThisFile)) {
530  cacheRawCommentForDecl(*D, *DocComment);
531  comments::FullComment *FC = DocComment->parse(*this, PP, D);
532  ParsedComments[D->getCanonicalDecl()] = FC;
533  }
534  }
535 }
536 
538  const Decl *D) const {
539  auto *ThisDeclInfo = new (*this) comments::DeclInfo;
540  ThisDeclInfo->CommentDecl = D;
541  ThisDeclInfo->IsFilled = false;
542  ThisDeclInfo->fill();
543  ThisDeclInfo->CommentDecl = FC->getDecl();
544  if (!ThisDeclInfo->TemplateParameters)
545  ThisDeclInfo->TemplateParameters = FC->getDeclInfo()->TemplateParameters;
546  comments::FullComment *CFC =
547  new (*this) comments::FullComment(FC->getBlocks(),
548  ThisDeclInfo);
549  return CFC;
550 }
551 
554  return RC ? RC->parse(*this, nullptr, D) : nullptr;
555 }
556 
558  const Decl *D,
559  const Preprocessor *PP) const {
560  if (!D || D->isInvalidDecl())
561  return nullptr;
562  D = &adjustDeclToTemplate(*D);
563 
564  const Decl *Canonical = D->getCanonicalDecl();
565  llvm::DenseMap<const Decl *, comments::FullComment *>::iterator Pos =
566  ParsedComments.find(Canonical);
567 
568  if (Pos != ParsedComments.end()) {
569  if (Canonical != D) {
570  comments::FullComment *FC = Pos->second;
572  return CFC;
573  }
574  return Pos->second;
575  }
576 
577  const Decl *OriginalDecl = nullptr;
578 
579  const RawComment *RC = getRawCommentForAnyRedecl(D, &OriginalDecl);
580  if (!RC) {
581  if (isa<ObjCMethodDecl>(D) || isa<FunctionDecl>(D)) {
583  const auto *OMD = dyn_cast<ObjCMethodDecl>(D);
584  if (OMD && OMD->isPropertyAccessor())
585  if (const ObjCPropertyDecl *PDecl = OMD->findPropertyDecl())
586  if (comments::FullComment *FC = getCommentForDecl(PDecl, PP))
587  return cloneFullComment(FC, D);
588  if (OMD)
589  addRedeclaredMethods(OMD, Overridden);
590  getOverriddenMethods(dyn_cast<NamedDecl>(D), Overridden);
591  for (unsigned i = 0, e = Overridden.size(); i < e; i++)
592  if (comments::FullComment *FC = getCommentForDecl(Overridden[i], PP))
593  return cloneFullComment(FC, D);
594  }
595  else if (const auto *TD = dyn_cast<TypedefNameDecl>(D)) {
596  // Attach any tag type's documentation to its typedef if latter
597  // does not have one of its own.
598  QualType QT = TD->getUnderlyingType();
599  if (const auto *TT = QT->getAs<TagType>())
600  if (const Decl *TD = TT->getDecl())
601  if (comments::FullComment *FC = getCommentForDecl(TD, PP))
602  return cloneFullComment(FC, D);
603  }
604  else if (const auto *IC = dyn_cast<ObjCInterfaceDecl>(D)) {
605  while (IC->getSuperClass()) {
606  IC = IC->getSuperClass();
607  if (comments::FullComment *FC = getCommentForDecl(IC, PP))
608  return cloneFullComment(FC, D);
609  }
610  }
611  else if (const auto *CD = dyn_cast<ObjCCategoryDecl>(D)) {
612  if (const ObjCInterfaceDecl *IC = CD->getClassInterface())
613  if (comments::FullComment *FC = getCommentForDecl(IC, PP))
614  return cloneFullComment(FC, D);
615  }
616  else if (const auto *RD = dyn_cast<CXXRecordDecl>(D)) {
617  if (!(RD = RD->getDefinition()))
618  return nullptr;
619  // Check non-virtual bases.
620  for (const auto &I : RD->bases()) {
621  if (I.isVirtual() || (I.getAccessSpecifier() != AS_public))
622  continue;
623  QualType Ty = I.getType();
624  if (Ty.isNull())
625  continue;
626  if (const CXXRecordDecl *NonVirtualBase = Ty->getAsCXXRecordDecl()) {
627  if (!(NonVirtualBase= NonVirtualBase->getDefinition()))
628  continue;
629 
630  if (comments::FullComment *FC = getCommentForDecl((NonVirtualBase), PP))
631  return cloneFullComment(FC, D);
632  }
633  }
634  // Check virtual bases.
635  for (const auto &I : RD->vbases()) {
636  if (I.getAccessSpecifier() != AS_public)
637  continue;
638  QualType Ty = I.getType();
639  if (Ty.isNull())
640  continue;
641  if (const CXXRecordDecl *VirtualBase = Ty->getAsCXXRecordDecl()) {
642  if (!(VirtualBase= VirtualBase->getDefinition()))
643  continue;
644  if (comments::FullComment *FC = getCommentForDecl((VirtualBase), PP))
645  return cloneFullComment(FC, D);
646  }
647  }
648  }
649  return nullptr;
650  }
651 
652  // If the RawComment was attached to other redeclaration of this Decl, we
653  // should parse the comment in context of that other Decl. This is important
654  // because comments can contain references to parameter names which can be
655  // different across redeclarations.
656  if (D != OriginalDecl && OriginalDecl)
657  return getCommentForDecl(OriginalDecl, PP);
658 
659  comments::FullComment *FC = RC->parse(*this, PP, D);
660  ParsedComments[Canonical] = FC;
661  return FC;
662 }
663 
664 void
665 ASTContext::CanonicalTemplateTemplateParm::Profile(llvm::FoldingSetNodeID &ID,
666  const ASTContext &C,
667  TemplateTemplateParmDecl *Parm) {
668  ID.AddInteger(Parm->getDepth());
669  ID.AddInteger(Parm->getPosition());
670  ID.AddBoolean(Parm->isParameterPack());
671 
673  ID.AddInteger(Params->size());
675  PEnd = Params->end();
676  P != PEnd; ++P) {
677  if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(*P)) {
678  ID.AddInteger(0);
679  ID.AddBoolean(TTP->isParameterPack());
680  const TypeConstraint *TC = TTP->getTypeConstraint();
681  ID.AddBoolean(TC != nullptr);
682  if (TC)
684  /*Canonical=*/true);
685  if (TTP->isExpandedParameterPack()) {
686  ID.AddBoolean(true);
687  ID.AddInteger(TTP->getNumExpansionParameters());
688  } else
689  ID.AddBoolean(false);
690  continue;
691  }
692 
693  if (const auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(*P)) {
694  ID.AddInteger(1);
695  ID.AddBoolean(NTTP->isParameterPack());
696  ID.AddPointer(NTTP->getType().getCanonicalType().getAsOpaquePtr());
697  if (NTTP->isExpandedParameterPack()) {
698  ID.AddBoolean(true);
699  ID.AddInteger(NTTP->getNumExpansionTypes());
700  for (unsigned I = 0, N = NTTP->getNumExpansionTypes(); I != N; ++I) {
701  QualType T = NTTP->getExpansionType(I);
702  ID.AddPointer(T.getCanonicalType().getAsOpaquePtr());
703  }
704  } else
705  ID.AddBoolean(false);
706  continue;
707  }
708 
709  auto *TTP = cast<TemplateTemplateParmDecl>(*P);
710  ID.AddInteger(2);
711  Profile(ID, C, TTP);
712  }
713  Expr *RequiresClause = Parm->getTemplateParameters()->getRequiresClause();
714  ID.AddBoolean(RequiresClause != nullptr);
715  if (RequiresClause)
716  RequiresClause->Profile(ID, C, /*Canonical=*/true);
717 }
718 
719 static Expr *
721  QualType ConstrainedType) {
722  // This is a bit ugly - we need to form a new immediately-declared
723  // constraint that references the new parameter; this would ideally
724  // require semantic analysis (e.g. template<C T> struct S {}; - the
725  // converted arguments of C<T> could be an argument pack if C is
726  // declared as template<typename... T> concept C = ...).
727  // We don't have semantic analysis here so we dig deep into the
728  // ready-made constraint expr and change the thing manually.
730  if (const auto *Fold = dyn_cast<CXXFoldExpr>(IDC))
731  CSE = cast<ConceptSpecializationExpr>(Fold->getLHS());
732  else
733  CSE = cast<ConceptSpecializationExpr>(IDC);
734  ArrayRef<TemplateArgument> OldConverted = CSE->getTemplateArguments();
736  NewConverted.reserve(OldConverted.size());
737  if (OldConverted.front().getKind() == TemplateArgument::Pack) {
738  // The case:
739  // template<typename... T> concept C = true;
740  // template<C<int> T> struct S; -> constraint is C<{T, int}>
741  NewConverted.push_back(ConstrainedType);
742  for (auto &Arg : OldConverted.front().pack_elements().drop_front(1))
743  NewConverted.push_back(Arg);
744  TemplateArgument NewPack(NewConverted);
745 
746  NewConverted.clear();
747  NewConverted.push_back(NewPack);
748  assert(OldConverted.size() == 1 &&
749  "Template parameter pack should be the last parameter");
750  } else {
751  assert(OldConverted.front().getKind() == TemplateArgument::Type &&
752  "Unexpected first argument kind for immediately-declared "
753  "constraint");
754  NewConverted.push_back(ConstrainedType);
755  for (auto &Arg : OldConverted.drop_front(1))
756  NewConverted.push_back(Arg);
757  }
759  C, CSE->getNamedConcept(), NewConverted, nullptr,
761 
762  if (auto *OrigFold = dyn_cast<CXXFoldExpr>(IDC))
763  NewIDC = new (C) CXXFoldExpr(
764  OrigFold->getType(), /*Callee*/nullptr, SourceLocation(), NewIDC,
765  BinaryOperatorKind::BO_LAnd, SourceLocation(), /*RHS=*/nullptr,
766  SourceLocation(), /*NumExpansions=*/None);
767  return NewIDC;
768 }
769 
771 ASTContext::getCanonicalTemplateTemplateParmDecl(
772  TemplateTemplateParmDecl *TTP) const {
773  // Check if we already have a canonical template template parameter.
774  llvm::FoldingSetNodeID ID;
775  CanonicalTemplateTemplateParm::Profile(ID, *this, TTP);
776  void *InsertPos = nullptr;
777  CanonicalTemplateTemplateParm *Canonical
778  = CanonTemplateTemplateParms.FindNodeOrInsertPos(ID, InsertPos);
779  if (Canonical)
780  return Canonical->getParam();
781 
782  // Build a canonical template parameter list.
784  SmallVector<NamedDecl *, 4> CanonParams;
785  CanonParams.reserve(Params->size());
787  PEnd = Params->end();
788  P != PEnd; ++P) {
789  if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(*P)) {
792  TTP->getDepth(), TTP->getIndex(), nullptr, false,
793  TTP->isParameterPack(), TTP->hasTypeConstraint(),
794  TTP->isExpandedParameterPack() ?
795  llvm::Optional<unsigned>(TTP->getNumExpansionParameters()) : None);
796  if (const auto *TC = TTP->getTypeConstraint()) {
797  QualType ParamAsArgument(NewTTP->getTypeForDecl(), 0);
800  ParamAsArgument);
801  TemplateArgumentListInfo CanonArgsAsWritten;
802  if (auto *Args = TC->getTemplateArgsAsWritten())
803  for (const auto &ArgLoc : Args->arguments())
804  CanonArgsAsWritten.addArgument(
805  TemplateArgumentLoc(ArgLoc.getArgument(),
807  NewTTP->setTypeConstraint(
810  SourceLocation()), /*FoundDecl=*/nullptr,
811  // Actually canonicalizing a TemplateArgumentLoc is difficult so we
812  // simply omit the ArgsAsWritten
813  TC->getNamedConcept(), /*ArgsAsWritten=*/nullptr, NewIDC);
814  }
815  CanonParams.push_back(NewTTP);
816  } else if (const auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(*P)) {
817  QualType T = getCanonicalType(NTTP->getType());
820  if (NTTP->isExpandedParameterPack()) {
821  SmallVector<QualType, 2> ExpandedTypes;
822  SmallVector<TypeSourceInfo *, 2> ExpandedTInfos;
823  for (unsigned I = 0, N = NTTP->getNumExpansionTypes(); I != N; ++I) {
824  ExpandedTypes.push_back(getCanonicalType(NTTP->getExpansionType(I)));
825  ExpandedTInfos.push_back(
826  getTrivialTypeSourceInfo(ExpandedTypes.back()));
827  }
828 
830  SourceLocation(),
831  SourceLocation(),
832  NTTP->getDepth(),
833  NTTP->getPosition(), nullptr,
834  T,
835  TInfo,
836  ExpandedTypes,
837  ExpandedTInfos);
838  } else {
840  SourceLocation(),
841  SourceLocation(),
842  NTTP->getDepth(),
843  NTTP->getPosition(), nullptr,
844  T,
845  NTTP->isParameterPack(),
846  TInfo);
847  }
848  if (AutoType *AT = T->getContainedAutoType()) {
849  if (AT->isConstrained()) {
852  *this, NTTP->getPlaceholderTypeConstraint(), T));
853  }
854  }
855  CanonParams.push_back(Param);
856 
857  } else
858  CanonParams.push_back(getCanonicalTemplateTemplateParmDecl(
859  cast<TemplateTemplateParmDecl>(*P)));
860  }
861 
862  Expr *CanonRequiresClause = nullptr;
863  if (Expr *RequiresClause = TTP->getTemplateParameters()->getRequiresClause())
864  CanonRequiresClause = RequiresClause;
865 
866  TemplateTemplateParmDecl *CanonTTP
868  SourceLocation(), TTP->getDepth(),
869  TTP->getPosition(),
870  TTP->isParameterPack(),
871  nullptr,
873  SourceLocation(),
874  CanonParams,
875  SourceLocation(),
876  CanonRequiresClause));
877 
878  // Get the new insert position for the node we care about.
879  Canonical = CanonTemplateTemplateParms.FindNodeOrInsertPos(ID, InsertPos);
880  assert(!Canonical && "Shouldn't be in the map!");
881  (void)Canonical;
882 
883  // Create the canonical template template parameter entry.
884  Canonical = new (*this) CanonicalTemplateTemplateParm(CanonTTP);
885  CanonTemplateTemplateParms.InsertNode(Canonical, InsertPos);
886  return CanonTTP;
887 }
888 
890  auto Kind = getTargetInfo().getCXXABI().getKind();
891  return getLangOpts().CXXABI.getValueOr(Kind);
892 }
893 
894 CXXABI *ASTContext::createCXXABI(const TargetInfo &T) {
895  if (!LangOpts.CPlusPlus) return nullptr;
896 
897  switch (getCXXABIKind()) {
898  case TargetCXXABI::AppleARM64:
899  case TargetCXXABI::Fuchsia:
900  case TargetCXXABI::GenericARM: // Same as Itanium at this level
901  case TargetCXXABI::iOS:
902  case TargetCXXABI::WatchOS:
903  case TargetCXXABI::GenericAArch64:
904  case TargetCXXABI::GenericMIPS:
905  case TargetCXXABI::GenericItanium:
906  case TargetCXXABI::WebAssembly:
907  case TargetCXXABI::XL:
908  return CreateItaniumCXXABI(*this);
909  case TargetCXXABI::Microsoft:
910  return CreateMicrosoftCXXABI(*this);
911  }
912  llvm_unreachable("Invalid CXXABI type!");
913 }
914 
916  if (!InterpContext) {
917  InterpContext.reset(new interp::Context(*this));
918  }
919  return *InterpContext.get();
920 }
921 
923  if (!ParentMapCtx)
924  ParentMapCtx.reset(new ParentMapContext(*this));
925  return *ParentMapCtx.get();
926 }
927 
928 static const LangASMap *getAddressSpaceMap(const TargetInfo &T,
929  const LangOptions &LOpts) {
930  if (LOpts.FakeAddressSpaceMap) {
931  // The fake address space map must have a distinct entry for each
932  // language-specific address space.
933  static const unsigned FakeAddrSpaceMap[] = {
934  0, // Default
935  1, // opencl_global
936  3, // opencl_local
937  2, // opencl_constant
938  0, // opencl_private
939  4, // opencl_generic
940  5, // opencl_global_device
941  6, // opencl_global_host
942  7, // cuda_device
943  8, // cuda_constant
944  9, // cuda_shared
945  1, // sycl_global
946  5, // sycl_global_device
947  6, // sycl_global_host
948  3, // sycl_local
949  0, // sycl_private
950  10, // ptr32_sptr
951  11, // ptr32_uptr
952  12 // ptr64
953  };
954  return &FakeAddrSpaceMap;
955  } else {
956  return &T.getAddressSpaceMap();
957  }
958 }
959 
961  const LangOptions &LangOpts) {
962  switch (LangOpts.getAddressSpaceMapMangling()) {
964  return TI.useAddressSpaceMapMangling();
966  return true;
968  return false;
969  }
970  llvm_unreachable("getAddressSpaceMapMangling() doesn't cover anything.");
971 }
972 
974  IdentifierTable &idents, SelectorTable &sels,
975  Builtin::Context &builtins, TranslationUnitKind TUKind)
976  : ConstantArrayTypes(this_()), FunctionProtoTypes(this_()),
977  TemplateSpecializationTypes(this_()),
978  DependentTemplateSpecializationTypes(this_()), AutoTypes(this_()),
979  SubstTemplateTemplateParmPacks(this_()),
980  CanonTemplateTemplateParms(this_()), SourceMgr(SM), LangOpts(LOpts),
981  NoSanitizeL(new NoSanitizeList(LangOpts.NoSanitizeFiles, SM)),
982  XRayFilter(new XRayFunctionFilter(LangOpts.XRayAlwaysInstrumentFiles,
983  LangOpts.XRayNeverInstrumentFiles,
984  LangOpts.XRayAttrListFiles, SM)),
985  ProfList(new ProfileList(LangOpts.ProfileListFiles, SM)),
986  PrintingPolicy(LOpts), Idents(idents), Selectors(sels),
987  BuiltinInfo(builtins), TUKind(TUKind), DeclarationNames(*this),
988  Comments(SM), CommentCommandTraits(BumpAlloc, LOpts.CommentOpts),
989  CompCategories(this_()), LastSDM(nullptr, 0) {
991 }
992 
994  // Release the DenseMaps associated with DeclContext objects.
995  // FIXME: Is this the ideal solution?
996  ReleaseDeclContextMaps();
997 
998  // Call all of the deallocation functions on all of their targets.
999  for (auto &Pair : Deallocations)
1000  (Pair.first)(Pair.second);
1001  Deallocations.clear();
1002 
1003  // ASTRecordLayout objects in ASTRecordLayouts must always be destroyed
1004  // because they can contain DenseMaps.
1005  for (llvm::DenseMap<const ObjCContainerDecl*,
1006  const ASTRecordLayout*>::iterator
1007  I = ObjCLayouts.begin(), E = ObjCLayouts.end(); I != E; )
1008  // Increment in loop to prevent using deallocated memory.
1009  if (auto *R = const_cast<ASTRecordLayout *>((I++)->second))
1010  R->Destroy(*this);
1011  ObjCLayouts.clear();
1012 
1013  for (llvm::DenseMap<const RecordDecl*, const ASTRecordLayout*>::iterator
1014  I = ASTRecordLayouts.begin(), E = ASTRecordLayouts.end(); I != E; ) {
1015  // Increment in loop to prevent using deallocated memory.
1016  if (auto *R = const_cast<ASTRecordLayout *>((I++)->second))
1017  R->Destroy(*this);
1018  }
1019  ASTRecordLayouts.clear();
1020 
1021  for (llvm::DenseMap<const Decl*, AttrVec*>::iterator A = DeclAttrs.begin(),
1022  AEnd = DeclAttrs.end();
1023  A != AEnd; ++A)
1024  A->second->~AttrVec();
1025  DeclAttrs.clear();
1026 
1027  for (const auto &Value : ModuleInitializers)
1028  Value.second->~PerModuleInitializers();
1029  ModuleInitializers.clear();
1030 }
1031 
1033 
1034 void ASTContext::setTraversalScope(const std::vector<Decl *> &TopLevelDecls) {
1035  TraversalScope = TopLevelDecls;
1037 }
1038 
1039 void ASTContext::AddDeallocation(void (*Callback)(void *), void *Data) const {
1040  Deallocations.push_back({Callback, Data});
1041 }
1042 
1043 void
1045  ExternalSource = std::move(Source);
1046 }
1047 
1049  llvm::errs() << "\n*** AST Context Stats:\n";
1050  llvm::errs() << " " << Types.size() << " types total.\n";
1051 
1052  unsigned counts[] = {
1053 #define TYPE(Name, Parent) 0,
1054 #define ABSTRACT_TYPE(Name, Parent)
1055 #include "clang/AST/TypeNodes.inc"
1056  0 // Extra
1057  };
1058 
1059  for (unsigned i = 0, e = Types.size(); i != e; ++i) {
1060  Type *T = Types[i];
1061  counts[(unsigned)T->getTypeClass()]++;
1062  }
1063 
1064  unsigned Idx = 0;
1065  unsigned TotalBytes = 0;
1066 #define TYPE(Name, Parent) \
1067  if (counts[Idx]) \
1068  llvm::errs() << " " << counts[Idx] << " " << #Name \
1069  << " types, " << sizeof(Name##Type) << " each " \
1070  << "(" << counts[Idx] * sizeof(Name##Type) \
1071  << " bytes)\n"; \
1072  TotalBytes += counts[Idx] * sizeof(Name##Type); \
1073  ++Idx;
1074 #define ABSTRACT_TYPE(Name, Parent)
1075 #include "clang/AST/TypeNodes.inc"
1076 
1077  llvm::errs() << "Total bytes = " << TotalBytes << "\n";
1078 
1079  // Implicit special member functions.
1080  llvm::errs() << NumImplicitDefaultConstructorsDeclared << "/"
1082  << " implicit default constructors created\n";
1083  llvm::errs() << NumImplicitCopyConstructorsDeclared << "/"
1085  << " implicit copy constructors created\n";
1086  if (getLangOpts().CPlusPlus)
1087  llvm::errs() << NumImplicitMoveConstructorsDeclared << "/"
1089  << " implicit move constructors created\n";
1090  llvm::errs() << NumImplicitCopyAssignmentOperatorsDeclared << "/"
1092  << " implicit copy assignment operators created\n";
1093  if (getLangOpts().CPlusPlus)
1094  llvm::errs() << NumImplicitMoveAssignmentOperatorsDeclared << "/"
1096  << " implicit move assignment operators created\n";
1097  llvm::errs() << NumImplicitDestructorsDeclared << "/"
1099  << " implicit destructors created\n";
1100 
1101  if (ExternalSource) {
1102  llvm::errs() << "\n";
1103  ExternalSource->PrintStats();
1104  }
1105 
1106  BumpAlloc.PrintStats();
1107 }
1108 
1110  bool NotifyListeners) {
1111  if (NotifyListeners)
1112  if (auto *Listener = getASTMutationListener())
1114 
1115  MergedDefModules[cast<NamedDecl>(ND->getCanonicalDecl())].push_back(M);
1116 }
1117 
1119  auto It = MergedDefModules.find(cast<NamedDecl>(ND->getCanonicalDecl()));
1120  if (It == MergedDefModules.end())
1121  return;
1122 
1123  auto &Merged = It->second;
1125  for (Module *&M : Merged)
1126  if (!Found.insert(M).second)
1127  M = nullptr;
1128  llvm::erase_value(Merged, nullptr);
1129 }
1130 
1133  auto MergedIt =
1134  MergedDefModules.find(cast<NamedDecl>(Def->getCanonicalDecl()));
1135  if (MergedIt == MergedDefModules.end())
1136  return None;
1137  return MergedIt->second;
1138 }
1139 
1140 void ASTContext::PerModuleInitializers::resolve(ASTContext &Ctx) {
1141  if (LazyInitializers.empty())
1142  return;
1143 
1144  auto *Source = Ctx.getExternalSource();
1145  assert(Source && "lazy initializers but no external source");
1146 
1147  auto LazyInits = std::move(LazyInitializers);
1148  LazyInitializers.clear();
1149 
1150  for (auto ID : LazyInits)
1151  Initializers.push_back(Source->GetExternalDecl(ID));
1152 
1153  assert(LazyInitializers.empty() &&
1154  "GetExternalDecl for lazy module initializer added more inits");
1155 }
1156 
1158  // One special case: if we add a module initializer that imports another
1159  // module, and that module's only initializer is an ImportDecl, simplify.
1160  if (const auto *ID = dyn_cast<ImportDecl>(D)) {
1161  auto It = ModuleInitializers.find(ID->getImportedModule());
1162 
1163  // Maybe the ImportDecl does nothing at all. (Common case.)
1164  if (It == ModuleInitializers.end())
1165  return;
1166 
1167  // Maybe the ImportDecl only imports another ImportDecl.
1168  auto &Imported = *It->second;
1169  if (Imported.Initializers.size() + Imported.LazyInitializers.size() == 1) {
1170  Imported.resolve(*this);
1171  auto *OnlyDecl = Imported.Initializers.front();
1172  if (isa<ImportDecl>(OnlyDecl))
1173  D = OnlyDecl;
1174  }
1175  }
1176 
1177  auto *&Inits = ModuleInitializers[M];
1178  if (!Inits)
1179  Inits = new (*this) PerModuleInitializers;
1180  Inits->Initializers.push_back(D);
1181 }
1182 
1184  auto *&Inits = ModuleInitializers[M];
1185  if (!Inits)
1186  Inits = new (*this) PerModuleInitializers;
1187  Inits->LazyInitializers.insert(Inits->LazyInitializers.end(),
1188  IDs.begin(), IDs.end());
1189 }
1190 
1192  auto It = ModuleInitializers.find(M);
1193  if (It == ModuleInitializers.end())
1194  return None;
1195 
1196  auto *Inits = It->second;
1197  Inits->resolve(*this);
1198  return Inits->Initializers;
1199 }
1200 
1202  if (!ExternCContext)
1203  ExternCContext = ExternCContextDecl::Create(*this, getTranslationUnitDecl());
1204 
1205  return ExternCContext;
1206 }
1207 
1210  const IdentifierInfo *II) const {
1211  auto *BuiltinTemplate =
1213  BuiltinTemplate->setImplicit();
1214  getTranslationUnitDecl()->addDecl(BuiltinTemplate);
1215 
1216  return BuiltinTemplate;
1217 }
1218 
1221  if (!MakeIntegerSeqDecl)
1222  MakeIntegerSeqDecl = buildBuiltinTemplateDecl(BTK__make_integer_seq,
1224  return MakeIntegerSeqDecl;
1225 }
1226 
1229  if (!TypePackElementDecl)
1230  TypePackElementDecl = buildBuiltinTemplateDecl(BTK__type_pack_element,
1232  return TypePackElementDecl;
1233 }
1234 
1236  RecordDecl::TagKind TK) const {
1237  SourceLocation Loc;
1238  RecordDecl *NewDecl;
1239  if (getLangOpts().CPlusPlus)
1240  NewDecl = CXXRecordDecl::Create(*this, TK, getTranslationUnitDecl(), Loc,
1241  Loc, &Idents.get(Name));
1242  else
1243  NewDecl = RecordDecl::Create(*this, TK, getTranslationUnitDecl(), Loc, Loc,
1244  &Idents.get(Name));
1245  NewDecl->setImplicit();
1246  NewDecl->addAttr(TypeVisibilityAttr::CreateImplicit(
1247  const_cast<ASTContext &>(*this), TypeVisibilityAttr::Default));
1248  return NewDecl;
1249 }
1250 
1252  StringRef Name) const {
1254  TypedefDecl *NewDecl = TypedefDecl::Create(
1255  const_cast<ASTContext &>(*this), getTranslationUnitDecl(),
1256  SourceLocation(), SourceLocation(), &Idents.get(Name), TInfo);
1257  NewDecl->setImplicit();
1258  return NewDecl;
1259 }
1260 
1262  if (!Int128Decl)
1263  Int128Decl = buildImplicitTypedef(Int128Ty, "__int128_t");
1264  return Int128Decl;
1265 }
1266 
1268  if (!UInt128Decl)
1269  UInt128Decl = buildImplicitTypedef(UnsignedInt128Ty, "__uint128_t");
1270  return UInt128Decl;
1271 }
1272 
1273 void ASTContext::InitBuiltinType(CanQualType &R, BuiltinType::Kind K) {
1274  auto *Ty = new (*this, TypeAlignment) BuiltinType(K);
1275  R = CanQualType::CreateUnsafe(QualType(Ty, 0));
1276  Types.push_back(Ty);
1277 }
1278 
1280  const TargetInfo *AuxTarget) {
1281  assert((!this->Target || this->Target == &Target) &&
1282  "Incorrect target reinitialization");
1283  assert(VoidTy.isNull() && "Context reinitialized?");
1284 
1285  this->Target = &Target;
1286  this->AuxTarget = AuxTarget;
1287 
1288  ABI.reset(createCXXABI(Target));
1289  AddrSpaceMap = getAddressSpaceMap(Target, LangOpts);
1290  AddrSpaceMapMangling = isAddrSpaceMapManglingEnabled(Target, LangOpts);
1291 
1292  // C99 6.2.5p19.
1293  InitBuiltinType(VoidTy, BuiltinType::Void);
1294 
1295  // C99 6.2.5p2.
1296  InitBuiltinType(BoolTy, BuiltinType::Bool);
1297  // C99 6.2.5p3.
1298  if (LangOpts.CharIsSigned)
1299  InitBuiltinType(CharTy, BuiltinType::Char_S);
1300  else
1301  InitBuiltinType(CharTy, BuiltinType::Char_U);
1302  // C99 6.2.5p4.
1303  InitBuiltinType(SignedCharTy, BuiltinType::SChar);
1304  InitBuiltinType(ShortTy, BuiltinType::Short);
1305  InitBuiltinType(IntTy, BuiltinType::Int);
1306  InitBuiltinType(LongTy, BuiltinType::Long);
1307  InitBuiltinType(LongLongTy, BuiltinType::LongLong);
1308 
1309  // C99 6.2.5p6.
1310  InitBuiltinType(UnsignedCharTy, BuiltinType::UChar);
1311  InitBuiltinType(UnsignedShortTy, BuiltinType::UShort);
1312  InitBuiltinType(UnsignedIntTy, BuiltinType::UInt);
1313  InitBuiltinType(UnsignedLongTy, BuiltinType::ULong);
1314  InitBuiltinType(UnsignedLongLongTy, BuiltinType::ULongLong);
1315 
1316  // C99 6.2.5p10.
1317  InitBuiltinType(FloatTy, BuiltinType::Float);
1318  InitBuiltinType(DoubleTy, BuiltinType::Double);
1319  InitBuiltinType(LongDoubleTy, BuiltinType::LongDouble);
1320 
1321  // GNU extension, __float128 for IEEE quadruple precision
1322  InitBuiltinType(Float128Ty, BuiltinType::Float128);
1323 
1324  // __ibm128 for IBM extended precision
1325  InitBuiltinType(Ibm128Ty, BuiltinType::Ibm128);
1326 
1327  // C11 extension ISO/IEC TS 18661-3
1328  InitBuiltinType(Float16Ty, BuiltinType::Float16);
1329 
1330  // ISO/IEC JTC1 SC22 WG14 N1169 Extension
1331  InitBuiltinType(ShortAccumTy, BuiltinType::ShortAccum);
1332  InitBuiltinType(AccumTy, BuiltinType::Accum);
1333  InitBuiltinType(LongAccumTy, BuiltinType::LongAccum);
1334  InitBuiltinType(UnsignedShortAccumTy, BuiltinType::UShortAccum);
1335  InitBuiltinType(UnsignedAccumTy, BuiltinType::UAccum);
1336  InitBuiltinType(UnsignedLongAccumTy, BuiltinType::ULongAccum);
1337  InitBuiltinType(ShortFractTy, BuiltinType::ShortFract);
1338  InitBuiltinType(FractTy, BuiltinType::Fract);
1339  InitBuiltinType(LongFractTy, BuiltinType::LongFract);
1340  InitBuiltinType(UnsignedShortFractTy, BuiltinType::UShortFract);
1341  InitBuiltinType(UnsignedFractTy, BuiltinType::UFract);
1342  InitBuiltinType(UnsignedLongFractTy, BuiltinType::ULongFract);
1343  InitBuiltinType(SatShortAccumTy, BuiltinType::SatShortAccum);
1344  InitBuiltinType(SatAccumTy, BuiltinType::SatAccum);
1345  InitBuiltinType(SatLongAccumTy, BuiltinType::SatLongAccum);
1346  InitBuiltinType(SatUnsignedShortAccumTy, BuiltinType::SatUShortAccum);
1347  InitBuiltinType(SatUnsignedAccumTy, BuiltinType::SatUAccum);
1348  InitBuiltinType(SatUnsignedLongAccumTy, BuiltinType::SatULongAccum);
1349  InitBuiltinType(SatShortFractTy, BuiltinType::SatShortFract);
1350  InitBuiltinType(SatFractTy, BuiltinType::SatFract);
1351  InitBuiltinType(SatLongFractTy, BuiltinType::SatLongFract);
1352  InitBuiltinType(SatUnsignedShortFractTy, BuiltinType::SatUShortFract);
1353  InitBuiltinType(SatUnsignedFractTy, BuiltinType::SatUFract);
1354  InitBuiltinType(SatUnsignedLongFractTy, BuiltinType::SatULongFract);
1355 
1356  // GNU extension, 128-bit integers.
1357  InitBuiltinType(Int128Ty, BuiltinType::Int128);
1358  InitBuiltinType(UnsignedInt128Ty, BuiltinType::UInt128);
1359 
1360  // C++ 3.9.1p5
1361  if (TargetInfo::isTypeSigned(Target.getWCharType()))
1362  InitBuiltinType(WCharTy, BuiltinType::WChar_S);
1363  else // -fshort-wchar makes wchar_t be unsigned.
1364  InitBuiltinType(WCharTy, BuiltinType::WChar_U);
1365  if (LangOpts.CPlusPlus && LangOpts.WChar)
1366  WideCharTy = WCharTy;
1367  else {
1368  // C99 (or C++ using -fno-wchar).
1369  WideCharTy = getFromTargetType(Target.getWCharType());
1370  }
1371 
1372  WIntTy = getFromTargetType(Target.getWIntType());
1373 
1374  // C++20 (proposed)
1375  InitBuiltinType(Char8Ty, BuiltinType::Char8);
1376 
1377  if (LangOpts.CPlusPlus) // C++0x 3.9.1p5, extension for C++
1378  InitBuiltinType(Char16Ty, BuiltinType::Char16);
1379  else // C99
1380  Char16Ty = getFromTargetType(Target.getChar16Type());
1381 
1382  if (LangOpts.CPlusPlus) // C++0x 3.9.1p5, extension for C++
1383  InitBuiltinType(Char32Ty, BuiltinType::Char32);
1384  else // C99
1385  Char32Ty = getFromTargetType(Target.getChar32Type());
1386 
1387  // Placeholder type for type-dependent expressions whose type is
1388  // completely unknown. No code should ever check a type against
1389  // DependentTy and users should never see it; however, it is here to
1390  // help diagnose failures to properly check for type-dependent
1391  // expressions.
1392  InitBuiltinType(DependentTy, BuiltinType::Dependent);
1393 
1394  // Placeholder type for functions.
1395  InitBuiltinType(OverloadTy, BuiltinType::Overload);
1396 
1397  // Placeholder type for bound members.
1398  InitBuiltinType(BoundMemberTy, BuiltinType::BoundMember);
1399 
1400  // Placeholder type for pseudo-objects.
1401  InitBuiltinType(PseudoObjectTy, BuiltinType::PseudoObject);
1402 
1403  // "any" type; useful for debugger-like clients.
1404  InitBuiltinType(UnknownAnyTy, BuiltinType::UnknownAny);
1405 
1406  // Placeholder type for unbridged ARC casts.
1407  InitBuiltinType(ARCUnbridgedCastTy, BuiltinType::ARCUnbridgedCast);
1408 
1409  // Placeholder type for builtin functions.
1410  InitBuiltinType(BuiltinFnTy, BuiltinType::BuiltinFn);
1411 
1412  // Placeholder type for OMP array sections.
1413  if (LangOpts.OpenMP) {
1414  InitBuiltinType(OMPArraySectionTy, BuiltinType::OMPArraySection);
1415  InitBuiltinType(OMPArrayShapingTy, BuiltinType::OMPArrayShaping);
1416  InitBuiltinType(OMPIteratorTy, BuiltinType::OMPIterator);
1417  }
1418  if (LangOpts.MatrixTypes)
1419  InitBuiltinType(IncompleteMatrixIdxTy, BuiltinType::IncompleteMatrixIdx);
1420 
1421  // Builtin types for 'id', 'Class', and 'SEL'.
1422  InitBuiltinType(ObjCBuiltinIdTy, BuiltinType::ObjCId);
1423  InitBuiltinType(ObjCBuiltinClassTy, BuiltinType::ObjCClass);
1424  InitBuiltinType(ObjCBuiltinSelTy, BuiltinType::ObjCSel);
1425 
1426  if (LangOpts.OpenCL) {
1427 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \
1428  InitBuiltinType(SingletonId, BuiltinType::Id);
1429 #include "clang/Basic/OpenCLImageTypes.def"
1430 
1431  InitBuiltinType(OCLSamplerTy, BuiltinType::OCLSampler);
1432  InitBuiltinType(OCLEventTy, BuiltinType::OCLEvent);
1433  InitBuiltinType(OCLClkEventTy, BuiltinType::OCLClkEvent);
1434  InitBuiltinType(OCLQueueTy, BuiltinType::OCLQueue);
1435  InitBuiltinType(OCLReserveIDTy, BuiltinType::OCLReserveID);
1436 
1437 #define EXT_OPAQUE_TYPE(ExtType, Id, Ext) \
1438  InitBuiltinType(Id##Ty, BuiltinType::Id);
1439 #include "clang/Basic/OpenCLExtensionTypes.def"
1440  }
1441 
1442  if (Target.hasAArch64SVETypes()) {
1443 #define SVE_TYPE(Name, Id, SingletonId) \
1444  InitBuiltinType(SingletonId, BuiltinType::Id);
1445 #include "clang/Basic/AArch64SVEACLETypes.def"
1446  }
1447 
1448  if (Target.getTriple().isPPC64()) {
1449 #define PPC_VECTOR_MMA_TYPE(Name, Id, Size) \
1450  InitBuiltinType(Id##Ty, BuiltinType::Id);
1451 #include "clang/Basic/PPCTypes.def"
1452 #define PPC_VECTOR_VSX_TYPE(Name, Id, Size) \
1453  InitBuiltinType(Id##Ty, BuiltinType::Id);
1454 #include "clang/Basic/PPCTypes.def"
1455  }
1456 
1457  if (Target.hasRISCVVTypes()) {
1458 #define RVV_TYPE(Name, Id, SingletonId) \
1459  InitBuiltinType(SingletonId, BuiltinType::Id);
1460 #include "clang/Basic/RISCVVTypes.def"
1461  }
1462 
1463  // Builtin type for __objc_yes and __objc_no
1464  ObjCBuiltinBoolTy = (Target.useSignedCharForObjCBool() ?
1465  SignedCharTy : BoolTy);
1466 
1467  ObjCConstantStringType = QualType();
1468 
1469  ObjCSuperType = QualType();
1470 
1471  // void * type
1472  if (LangOpts.OpenCLGenericAddressSpace) {
1473  auto Q = VoidTy.getQualifiers();
1477  } else {
1479  }
1480 
1481  // nullptr type (C++0x 2.14.7)
1482  InitBuiltinType(NullPtrTy, BuiltinType::NullPtr);
1483 
1484  // half type (OpenCL 6.1.1.1) / ARM NEON __fp16
1485  InitBuiltinType(HalfTy, BuiltinType::Half);
1486 
1487  InitBuiltinType(BFloat16Ty, BuiltinType::BFloat16);
1488 
1489  // Builtin type used to help define __builtin_va_list.
1490  VaListTagDecl = nullptr;
1491 
1492  // MSVC predeclares struct _GUID, and we need it to create MSGuidDecls.
1493  if (LangOpts.MicrosoftExt || LangOpts.Borland) {
1496  }
1497 }
1498 
1500  return SourceMgr.getDiagnostics();
1501 }
1502 
1504  AttrVec *&Result = DeclAttrs[D];
1505  if (!Result) {
1506  void *Mem = Allocate(sizeof(AttrVec));
1507  Result = new (Mem) AttrVec;
1508  }
1509 
1510  return *Result;
1511 }
1512 
1513 /// Erase the attributes corresponding to the given declaration.
1515  llvm::DenseMap<const Decl*, AttrVec*>::iterator Pos = DeclAttrs.find(D);
1516  if (Pos != DeclAttrs.end()) {
1517  Pos->second->~AttrVec();
1518  DeclAttrs.erase(Pos);
1519  }
1520 }
1521 
1522 // FIXME: Remove ?
1525  assert(Var->isStaticDataMember() && "Not a static data member");
1527  .dyn_cast<MemberSpecializationInfo *>();
1528 }
1529 
1532  llvm::DenseMap<const VarDecl *, TemplateOrSpecializationInfo>::iterator Pos =
1533  TemplateOrInstantiation.find(Var);
1534  if (Pos == TemplateOrInstantiation.end())
1535  return {};
1536 
1537  return Pos->second;
1538 }
1539 
1540 void
1543  SourceLocation PointOfInstantiation) {
1544  assert(Inst->isStaticDataMember() && "Not a static data member");
1545  assert(Tmpl->isStaticDataMember() && "Not a static data member");
1547  Tmpl, TSK, PointOfInstantiation));
1548 }
1549 
1550 void
1553  assert(!TemplateOrInstantiation[Inst] &&
1554  "Already noted what the variable was instantiated from");
1555  TemplateOrInstantiation[Inst] = TSI;
1556 }
1557 
1558 NamedDecl *
1560  auto Pos = InstantiatedFromUsingDecl.find(UUD);
1561  if (Pos == InstantiatedFromUsingDecl.end())
1562  return nullptr;
1563 
1564  return Pos->second;
1565 }
1566 
1567 void
1569  assert((isa<UsingDecl>(Pattern) ||
1570  isa<UnresolvedUsingValueDecl>(Pattern) ||
1571  isa<UnresolvedUsingTypenameDecl>(Pattern)) &&
1572  "pattern decl is not a using decl");
1573  assert((isa<UsingDecl>(Inst) ||
1574  isa<UnresolvedUsingValueDecl>(Inst) ||
1575  isa<UnresolvedUsingTypenameDecl>(Inst)) &&
1576  "instantiation did not produce a using decl");
1577  assert(!InstantiatedFromUsingDecl[Inst] && "pattern already exists");
1578  InstantiatedFromUsingDecl[Inst] = Pattern;
1579 }
1580 
1581 UsingEnumDecl *
1583  auto Pos = InstantiatedFromUsingEnumDecl.find(UUD);
1584  if (Pos == InstantiatedFromUsingEnumDecl.end())
1585  return nullptr;
1586 
1587  return Pos->second;
1588 }
1589 
1591  UsingEnumDecl *Pattern) {
1592  assert(!InstantiatedFromUsingEnumDecl[Inst] && "pattern already exists");
1593  InstantiatedFromUsingEnumDecl[Inst] = Pattern;
1594 }
1595 
1598  llvm::DenseMap<UsingShadowDecl*, UsingShadowDecl*>::const_iterator Pos
1599  = InstantiatedFromUsingShadowDecl.find(Inst);
1600  if (Pos == InstantiatedFromUsingShadowDecl.end())
1601  return nullptr;
1602 
1603  return Pos->second;
1604 }
1605 
1606 void
1608  UsingShadowDecl *Pattern) {
1609  assert(!InstantiatedFromUsingShadowDecl[Inst] && "pattern already exists");
1610  InstantiatedFromUsingShadowDecl[Inst] = Pattern;
1611 }
1612 
1614  llvm::DenseMap<FieldDecl *, FieldDecl *>::iterator Pos
1615  = InstantiatedFromUnnamedFieldDecl.find(Field);
1616  if (Pos == InstantiatedFromUnnamedFieldDecl.end())
1617  return nullptr;
1618 
1619  return Pos->second;
1620 }
1621 
1623  FieldDecl *Tmpl) {
1624  assert(!Inst->getDeclName() && "Instantiated field decl is not unnamed");
1625  assert(!Tmpl->getDeclName() && "Template field decl is not unnamed");
1626  assert(!InstantiatedFromUnnamedFieldDecl[Inst] &&
1627  "Already noted what unnamed field was instantiated from");
1628 
1629  InstantiatedFromUnnamedFieldDecl[Inst] = Tmpl;
1630 }
1631 
1634  return overridden_methods(Method).begin();
1635 }
1636 
1639  return overridden_methods(Method).end();
1640 }
1641 
1642 unsigned
1644  auto Range = overridden_methods(Method);
1645  return Range.end() - Range.begin();
1646 }
1647 
1650  llvm::DenseMap<const CXXMethodDecl *, CXXMethodVector>::const_iterator Pos =
1651  OverriddenMethods.find(Method->getCanonicalDecl());
1652  if (Pos == OverriddenMethods.end())
1653  return overridden_method_range(nullptr, nullptr);
1654  return overridden_method_range(Pos->second.begin(), Pos->second.end());
1655 }
1656 
1658  const CXXMethodDecl *Overridden) {
1659  assert(Method->isCanonicalDecl() && Overridden->isCanonicalDecl());
1660  OverriddenMethods[Method].push_back(Overridden);
1661 }
1662 
1664  const NamedDecl *D,
1665  SmallVectorImpl<const NamedDecl *> &Overridden) const {
1666  assert(D);
1667 
1668  if (const auto *CXXMethod = dyn_cast<CXXMethodDecl>(D)) {
1669  Overridden.append(overridden_methods_begin(CXXMethod),
1670  overridden_methods_end(CXXMethod));
1671  return;
1672  }
1673 
1674  const auto *Method = dyn_cast<ObjCMethodDecl>(D);
1675  if (!Method)
1676  return;
1677 
1679  Method->getOverriddenMethods(OverDecls);
1680  Overridden.append(OverDecls.begin(), OverDecls.end());
1681 }
1682 
1684  assert(!Import->getNextLocalImport() &&
1685  "Import declaration already in the chain");
1686  assert(!Import->isFromASTFile() && "Non-local import declaration");
1687  if (!FirstLocalImport) {
1688  FirstLocalImport = Import;
1689  LastLocalImport = Import;
1690  return;
1691  }
1692 
1693  LastLocalImport->setNextLocalImport(Import);
1694  LastLocalImport = Import;
1695 }
1696 
1697 //===----------------------------------------------------------------------===//
1698 // Type Sizing and Analysis
1699 //===----------------------------------------------------------------------===//
1700 
1701 /// getFloatTypeSemantics - Return the APFloat 'semantics' for the specified
1702 /// scalar floating point type.
1703 const llvm::fltSemantics &ASTContext::getFloatTypeSemantics(QualType T) const {
1704  switch (T->castAs<BuiltinType>()->getKind()) {
1705  default:
1706  llvm_unreachable("Not a floating point type!");
1707  case BuiltinType::BFloat16:
1708  return Target->getBFloat16Format();
1709  case BuiltinType::Float16:
1710  case BuiltinType::Half:
1711  return Target->getHalfFormat();
1712  case BuiltinType::Float: return Target->getFloatFormat();
1713  case BuiltinType::Double: return Target->getDoubleFormat();
1714  case BuiltinType::Ibm128:
1715  return Target->getIbm128Format();
1716  case BuiltinType::LongDouble:
1717  if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice)
1718  return AuxTarget->getLongDoubleFormat();
1719  return Target->getLongDoubleFormat();
1720  case BuiltinType::Float128:
1721  if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice)
1722  return AuxTarget->getFloat128Format();
1723  return Target->getFloat128Format();
1724  }
1725 }
1726 
1727 CharUnits ASTContext::getDeclAlign(const Decl *D, bool ForAlignof) const {
1728  unsigned Align = Target->getCharWidth();
1729 
1730  bool UseAlignAttrOnly = false;
1731  if (unsigned AlignFromAttr = D->getMaxAlignment()) {
1732  Align = AlignFromAttr;
1733 
1734  // __attribute__((aligned)) can increase or decrease alignment
1735  // *except* on a struct or struct member, where it only increases
1736  // alignment unless 'packed' is also specified.
1737  //
1738  // It is an error for alignas to decrease alignment, so we can
1739  // ignore that possibility; Sema should diagnose it.
1740  if (isa<FieldDecl>(D)) {
1741  UseAlignAttrOnly = D->hasAttr<PackedAttr>() ||
1742  cast<FieldDecl>(D)->getParent()->hasAttr<PackedAttr>();
1743  } else {
1744  UseAlignAttrOnly = true;
1745  }
1746  }
1747  else if (isa<FieldDecl>(D))
1748  UseAlignAttrOnly =
1749  D->hasAttr<PackedAttr>() ||
1750  cast<FieldDecl>(D)->getParent()->hasAttr<PackedAttr>();
1751 
1752  // If we're using the align attribute only, just ignore everything
1753  // else about the declaration and its type.
1754  if (UseAlignAttrOnly) {
1755  // do nothing
1756  } else if (const auto *VD = dyn_cast<ValueDecl>(D)) {
1757  QualType T = VD->getType();
1758  if (const auto *RT = T->getAs<ReferenceType>()) {
1759  if (ForAlignof)
1760  T = RT->getPointeeType();
1761  else
1762  T = getPointerType(RT->getPointeeType());
1763  }
1764  QualType BaseT = getBaseElementType(T);
1765  if (T->isFunctionType())
1766  Align = getTypeInfoImpl(T.getTypePtr()).Align;
1767  else if (!BaseT->isIncompleteType()) {
1768  // Adjust alignments of declarations with array type by the
1769  // large-array alignment on the target.
1770  if (const ArrayType *arrayType = getAsArrayType(T)) {
1771  unsigned MinWidth = Target->getLargeArrayMinWidth();
1772  if (!ForAlignof && MinWidth) {
1773  if (isa<VariableArrayType>(arrayType))
1774  Align = std::max(Align, Target->getLargeArrayAlign());
1775  else if (isa<ConstantArrayType>(arrayType) &&
1776  MinWidth <= getTypeSize(cast<ConstantArrayType>(arrayType)))
1777  Align = std::max(Align, Target->getLargeArrayAlign());
1778  }
1779  }
1780  Align = std::max(Align, getPreferredTypeAlign(T.getTypePtr()));
1781  if (BaseT.getQualifiers().hasUnaligned())
1782  Align = Target->getCharWidth();
1783  if (const auto *VD = dyn_cast<VarDecl>(D)) {
1784  if (VD->hasGlobalStorage() && !ForAlignof) {
1785  uint64_t TypeSize = getTypeSize(T.getTypePtr());
1786  Align = std::max(Align, getTargetInfo().getMinGlobalAlign(TypeSize));
1787  }
1788  }
1789  }
1790 
1791  // Fields can be subject to extra alignment constraints, like if
1792  // the field is packed, the struct is packed, or the struct has a
1793  // a max-field-alignment constraint (#pragma pack). So calculate
1794  // the actual alignment of the field within the struct, and then
1795  // (as we're expected to) constrain that by the alignment of the type.
1796  if (const auto *Field = dyn_cast<FieldDecl>(VD)) {
1797  const RecordDecl *Parent = Field->getParent();
1798  // We can only produce a sensible answer if the record is valid.
1799  if (!Parent->isInvalidDecl()) {
1800  const ASTRecordLayout &Layout = getASTRecordLayout(Parent);
1801 
1802  // Start with the record's overall alignment.
1803  unsigned FieldAlign = toBits(Layout.getAlignment());
1804 
1805  // Use the GCD of that and the offset within the record.
1806  uint64_t Offset = Layout.getFieldOffset(Field->getFieldIndex());
1807  if (Offset > 0) {
1808  // Alignment is always a power of 2, so the GCD will be a power of 2,
1809  // which means we get to do this crazy thing instead of Euclid's.
1810  uint64_t LowBitOfOffset = Offset & (~Offset + 1);
1811  if (LowBitOfOffset < FieldAlign)
1812  FieldAlign = static_cast<unsigned>(LowBitOfOffset);
1813  }
1814 
1815  Align = std::min(Align, FieldAlign);
1816  }
1817  }
1818  }
1819 
1820  // Some targets have hard limitation on the maximum requestable alignment in
1821  // aligned attribute for static variables.
1822  const unsigned MaxAlignedAttr = getTargetInfo().getMaxAlignedAttribute();
1823  const auto *VD = dyn_cast<VarDecl>(D);
1824  if (MaxAlignedAttr && VD && VD->getStorageClass() == SC_Static)
1825  Align = std::min(Align, MaxAlignedAttr);
1826 
1827  return toCharUnitsFromBits(Align);
1828 }
1829 
1831  return toCharUnitsFromBits(Target->getExnObjectAlignment());
1832 }
1833 
1834 // getTypeInfoDataSizeInChars - Return the size of a type, in
1835 // chars. If the type is a record, its data size is returned. This is
1836 // the size of the memcpy that's performed when assigning this type
1837 // using a trivial copy/move assignment operator.
1840 
1841  // In C++, objects can sometimes be allocated into the tail padding
1842  // of a base-class subobject. We decide whether that's possible
1843  // during class layout, so here we can just trust the layout results.
1844  if (getLangOpts().CPlusPlus) {
1845  if (const auto *RT = T->getAs<RecordType>()) {
1846  const ASTRecordLayout &layout = getASTRecordLayout(RT->getDecl());
1847  Info.Width = layout.getDataSize();
1848  }
1849  }
1850 
1851  return Info;
1852 }
1853 
1854 /// getConstantArrayInfoInChars - Performing the computation in CharUnits
1855 /// instead of in bits prevents overflowing the uint64_t for some large arrays.
1858  const ConstantArrayType *CAT) {
1859  TypeInfoChars EltInfo = Context.getTypeInfoInChars(CAT->getElementType());
1860  uint64_t Size = CAT->getSize().getZExtValue();
1861  assert((Size == 0 || static_cast<uint64_t>(EltInfo.Width.getQuantity()) <=
1862  (uint64_t)(-1)/Size) &&
1863  "Overflow in array type char size evaluation");
1864  uint64_t Width = EltInfo.Width.getQuantity() * Size;
1865  unsigned Align = EltInfo.Align.getQuantity();
1866  if (!Context.getTargetInfo().getCXXABI().isMicrosoft() ||
1867  Context.getTargetInfo().getPointerWidth(0) == 64)
1868  Width = llvm::alignTo(Width, Align);
1869  return TypeInfoChars(CharUnits::fromQuantity(Width),
1870  CharUnits::fromQuantity(Align),
1871  EltInfo.AlignRequirement);
1872 }
1873 
1875  if (const auto *CAT = dyn_cast<ConstantArrayType>(T))
1876  return getConstantArrayInfoInChars(*this, CAT);
1877  TypeInfo Info = getTypeInfo(T);
1880 }
1881 
1883  return getTypeInfoInChars(T.getTypePtr());
1884 }
1885 
1888 }
1889 
1891  return isAlignmentRequired(T.getTypePtr());
1892 }
1893 
1895  bool NeedsPreferredAlignment) const {
1896  // An alignment on a typedef overrides anything else.
1897  if (const auto *TT = T->getAs<TypedefType>())
1898  if (unsigned Align = TT->getDecl()->getMaxAlignment())
1899  return Align;
1900 
1901  // If we have an (array of) complete type, we're done.
1902  T = getBaseElementType(T);
1903  if (!T->isIncompleteType())
1904  return NeedsPreferredAlignment ? getPreferredTypeAlign(T) : getTypeAlign(T);
1905 
1906  // If we had an array type, its element type might be a typedef
1907  // type with an alignment attribute.
1908  if (const auto *TT = T->getAs<TypedefType>())
1909  if (unsigned Align = TT->getDecl()->getMaxAlignment())
1910  return Align;
1911 
1912  // Otherwise, see if the declaration of the type had an attribute.
1913  if (const auto *TT = T->getAs<TagType>())
1914  return TT->getDecl()->getMaxAlignment();
1915 
1916  return 0;
1917 }
1918 
1920  TypeInfoMap::iterator I = MemoizedTypeInfo.find(T);
1921  if (I != MemoizedTypeInfo.end())
1922  return I->second;
1923 
1924  // This call can invalidate MemoizedTypeInfo[T], so we need a second lookup.
1925  TypeInfo TI = getTypeInfoImpl(T);
1926  MemoizedTypeInfo[T] = TI;
1927  return TI;
1928 }
1929 
1930 /// getTypeInfoImpl - Return the size of the specified type, in bits. This
1931 /// method does not work on incomplete types.
1932 ///
1933 /// FIXME: Pointers into different addr spaces could have different sizes and
1934 /// alignment requirements: getPointerInfo should take an AddrSpace, this
1935 /// should take a QualType, &c.
1936 TypeInfo ASTContext::getTypeInfoImpl(const Type *T) const {
1937  uint64_t Width = 0;
1938  unsigned Align = 8;
1940  unsigned AS = 0;
1941  switch (T->getTypeClass()) {
1942 #define TYPE(Class, Base)
1943 #define ABSTRACT_TYPE(Class, Base)
1944 #define NON_CANONICAL_TYPE(Class, Base)
1945 #define DEPENDENT_TYPE(Class, Base) case Type::Class:
1946 #define NON_CANONICAL_UNLESS_DEPENDENT_TYPE(Class, Base) \
1947  case Type::Class: \
1948  assert(!T->isDependentType() && "should not see dependent types here"); \
1949  return getTypeInfo(cast<Class##Type>(T)->desugar().getTypePtr());
1950 #include "clang/AST/TypeNodes.inc"
1951  llvm_unreachable("Should not see dependent types");
1952 
1953  case Type::FunctionNoProto:
1954  case Type::FunctionProto:
1955  // GCC extension: alignof(function) = 32 bits
1956  Width = 0;
1957  Align = 32;
1958  break;
1959 
1960  case Type::IncompleteArray:
1961  case Type::VariableArray:
1962  case Type::ConstantArray: {
1963  // Model non-constant sized arrays as size zero, but track the alignment.
1964  uint64_t Size = 0;
1965  if (const auto *CAT = dyn_cast<ConstantArrayType>(T))
1966  Size = CAT->getSize().getZExtValue();
1967 
1968  TypeInfo EltInfo = getTypeInfo(cast<ArrayType>(T)->getElementType());
1969  assert((Size == 0 || EltInfo.Width <= (uint64_t)(-1) / Size) &&
1970  "Overflow in array type bit size evaluation");
1971  Width = EltInfo.Width * Size;
1972  Align = EltInfo.Align;
1973  AlignRequirement = EltInfo.AlignRequirement;
1974  if (!getTargetInfo().getCXXABI().isMicrosoft() ||
1975  getTargetInfo().getPointerWidth(0) == 64)
1976  Width = llvm::alignTo(Width, Align);
1977  break;
1978  }
1979 
1980  case Type::ExtVector:
1981  case Type::Vector: {
1982  const auto *VT = cast<VectorType>(T);
1983  TypeInfo EltInfo = getTypeInfo(VT->getElementType());
1984  Width = EltInfo.Width * VT->getNumElements();
1985  Align = Width;
1986  // If the alignment is not a power of 2, round up to the next power of 2.
1987  // This happens for non-power-of-2 length vectors.
1988  if (Align & (Align-1)) {
1989  Align = llvm::NextPowerOf2(Align);
1990  Width = llvm::alignTo(Width, Align);
1991  }
1992  // Adjust the alignment based on the target max.
1993  uint64_t TargetVectorAlign = Target->getMaxVectorAlign();
1994  if (TargetVectorAlign && TargetVectorAlign < Align)
1995  Align = TargetVectorAlign;
1996  if (VT->getVectorKind() == VectorType::SveFixedLengthDataVector)
1997  // Adjust the alignment for fixed-length SVE vectors. This is important
1998  // for non-power-of-2 vector lengths.
1999  Align = 128;
2000  else if (VT->getVectorKind() == VectorType::SveFixedLengthPredicateVector)
2001  // Adjust the alignment for fixed-length SVE predicates.
2002  Align = 16;
2003  break;
2004  }
2005 
2006  case Type::ConstantMatrix: {
2007  const auto *MT = cast<ConstantMatrixType>(T);
2008  TypeInfo ElementInfo = getTypeInfo(MT->getElementType());
2009  // The internal layout of a matrix value is implementation defined.
2010  // Initially be ABI compatible with arrays with respect to alignment and
2011  // size.
2012  Width = ElementInfo.Width * MT->getNumRows() * MT->getNumColumns();
2013  Align = ElementInfo.Align;
2014  break;
2015  }
2016 
2017  case Type::Builtin:
2018  switch (cast<BuiltinType>(T)->getKind()) {
2019  default: llvm_unreachable("Unknown builtin type!");
2020  case BuiltinType::Void:
2021  // GCC extension: alignof(void) = 8 bits.
2022  Width = 0;
2023  Align = 8;
2024  break;
2025  case BuiltinType::Bool:
2026  Width = Target->getBoolWidth();
2027  Align = Target->getBoolAlign();
2028  break;
2029  case BuiltinType::Char_S:
2030  case BuiltinType::Char_U:
2031  case BuiltinType::UChar:
2032  case BuiltinType::SChar:
2033  case BuiltinType::Char8:
2034  Width = Target->getCharWidth();
2035  Align = Target->getCharAlign();
2036  break;
2037  case BuiltinType::WChar_S:
2038  case BuiltinType::WChar_U:
2039  Width = Target->getWCharWidth();
2040  Align = Target->getWCharAlign();
2041  break;
2042  case BuiltinType::Char16:
2043  Width = Target->getChar16Width();
2044  Align = Target->getChar16Align();
2045  break;
2046  case BuiltinType::Char32:
2047  Width = Target->getChar32Width();
2048  Align = Target->getChar32Align();
2049  break;
2050  case BuiltinType::UShort:
2051  case BuiltinType::Short:
2052  Width = Target->getShortWidth();
2053  Align = Target->getShortAlign();
2054  break;
2055  case BuiltinType::UInt:
2056  case BuiltinType::Int:
2057  Width = Target->getIntWidth();
2058  Align = Target->getIntAlign();
2059  break;
2060  case BuiltinType::ULong:
2061  case BuiltinType::Long:
2062  Width = Target->getLongWidth();
2063  Align = Target->getLongAlign();
2064  break;
2065  case BuiltinType::ULongLong:
2066  case BuiltinType::LongLong:
2067  Width = Target->getLongLongWidth();
2068  Align = Target->getLongLongAlign();
2069  break;
2070  case BuiltinType::Int128:
2071  case BuiltinType::UInt128:
2072  Width = 128;
2073  Align = 128; // int128_t is 128-bit aligned on all targets.
2074  break;
2075  case BuiltinType::ShortAccum:
2076  case BuiltinType::UShortAccum:
2077  case BuiltinType::SatShortAccum:
2078  case BuiltinType::SatUShortAccum:
2079  Width = Target->getShortAccumWidth();
2080  Align = Target->getShortAccumAlign();
2081  break;
2082  case BuiltinType::Accum:
2083  case BuiltinType::UAccum:
2084  case BuiltinType::SatAccum:
2085  case BuiltinType::SatUAccum:
2086  Width = Target->getAccumWidth();
2087  Align = Target->getAccumAlign();
2088  break;
2089  case BuiltinType::LongAccum:
2090  case BuiltinType::ULongAccum:
2091  case BuiltinType::SatLongAccum:
2092  case BuiltinType::SatULongAccum:
2093  Width = Target->getLongAccumWidth();
2094  Align = Target->getLongAccumAlign();
2095  break;
2096  case BuiltinType::ShortFract:
2097  case BuiltinType::UShortFract:
2098  case BuiltinType::SatShortFract:
2099  case BuiltinType::SatUShortFract:
2100  Width = Target->getShortFractWidth();
2101  Align = Target->getShortFractAlign();
2102  break;
2103  case BuiltinType::Fract:
2104  case BuiltinType::UFract:
2105  case BuiltinType::SatFract:
2106  case BuiltinType::SatUFract:
2107  Width = Target->getFractWidth();
2108  Align = Target->getFractAlign();
2109  break;
2110  case BuiltinType::LongFract:
2111  case BuiltinType::ULongFract:
2112  case BuiltinType::SatLongFract:
2113  case BuiltinType::SatULongFract:
2114  Width = Target->getLongFractWidth();
2115  Align = Target->getLongFractAlign();
2116  break;
2117  case BuiltinType::BFloat16:
2118  Width = Target->getBFloat16Width();
2119  Align = Target->getBFloat16Align();
2120  break;
2121  case BuiltinType::Float16:
2122  case BuiltinType::Half:
2123  if (Target->hasFloat16Type() || !getLangOpts().OpenMP ||
2124  !getLangOpts().OpenMPIsDevice) {
2125  Width = Target->getHalfWidth();
2126  Align = Target->getHalfAlign();
2127  } else {
2128  assert(getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice &&
2129  "Expected OpenMP device compilation.");
2130  Width = AuxTarget->getHalfWidth();
2131  Align = AuxTarget->getHalfAlign();
2132  }
2133  break;
2134  case BuiltinType::Float:
2135  Width = Target->getFloatWidth();
2136  Align = Target->getFloatAlign();
2137  break;
2138  case BuiltinType::Double:
2139  Width = Target->getDoubleWidth();
2140  Align = Target->getDoubleAlign();
2141  break;
2142  case BuiltinType::Ibm128:
2143  Width = Target->getIbm128Width();
2144  Align = Target->getIbm128Align();
2145  break;
2146  case BuiltinType::LongDouble:
2147  if (getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice &&
2148  (Target->getLongDoubleWidth() != AuxTarget->getLongDoubleWidth() ||
2149  Target->getLongDoubleAlign() != AuxTarget->getLongDoubleAlign())) {
2150  Width = AuxTarget->getLongDoubleWidth();
2151  Align = AuxTarget->getLongDoubleAlign();
2152  } else {
2153  Width = Target->getLongDoubleWidth();
2154  Align = Target->getLongDoubleAlign();
2155  }
2156  break;
2157  case BuiltinType::Float128:
2158  if (Target->hasFloat128Type() || !getLangOpts().OpenMP ||
2159  !getLangOpts().OpenMPIsDevice) {
2160  Width = Target->getFloat128Width();
2161  Align = Target->getFloat128Align();
2162  } else {
2163  assert(getLangOpts().OpenMP && getLangOpts().OpenMPIsDevice &&
2164  "Expected OpenMP device compilation.");
2165  Width = AuxTarget->getFloat128Width();
2166  Align = AuxTarget->getFloat128Align();
2167  }
2168  break;
2169  case BuiltinType::NullPtr:
2170  Width = Target->getPointerWidth(0); // C++ 3.9.1p11: sizeof(nullptr_t)
2171  Align = Target->getPointerAlign(0); // == sizeof(void*)
2172  break;
2173  case BuiltinType::ObjCId:
2174  case BuiltinType::ObjCClass:
2175  case BuiltinType::ObjCSel:
2176  Width = Target->getPointerWidth(0);
2177  Align = Target->getPointerAlign(0);
2178  break;
2179  case BuiltinType::OCLSampler:
2180  case BuiltinType::OCLEvent:
2181  case BuiltinType::OCLClkEvent:
2182  case BuiltinType::OCLQueue:
2183  case BuiltinType::OCLReserveID:
2184 #define IMAGE_TYPE(ImgType, Id, SingletonId, Access, Suffix) \
2185  case BuiltinType::Id:
2186 #include "clang/Basic/OpenCLImageTypes.def"
2187 #define EXT_OPAQUE_TYPE(ExtType, Id, Ext) \
2188  case BuiltinType::Id:
2189 #include "clang/Basic/OpenCLExtensionTypes.def"
2190  AS = getTargetAddressSpace(
2192  Width = Target->getPointerWidth(AS);
2193  Align = Target->getPointerAlign(AS);
2194  break;
2195  // The SVE types are effectively target-specific. The length of an
2196  // SVE_VECTOR_TYPE is only known at runtime, but it is always a multiple
2197  // of 128 bits. There is one predicate bit for each vector byte, so the
2198  // length of an SVE_PREDICATE_TYPE is always a multiple of 16 bits.
2199  //
2200  // Because the length is only known at runtime, we use a dummy value
2201  // of 0 for the static length. The alignment values are those defined
2202  // by the Procedure Call Standard for the Arm Architecture.
2203 #define SVE_VECTOR_TYPE(Name, MangledName, Id, SingletonId, NumEls, ElBits, \
2204  IsSigned, IsFP, IsBF) \
2205  case BuiltinType::Id: \
2206  Width = 0; \
2207  Align = 128; \
2208  break;
2209 #define SVE_PREDICATE_TYPE(Name, MangledName, Id, SingletonId, NumEls) \
2210  case BuiltinType::Id: \
2211  Width = 0; \
2212  Align = 16; \
2213  break;
2214 #include "clang/Basic/AArch64SVEACLETypes.def"
2215 #define PPC_VECTOR_TYPE(Name, Id, Size) \
2216  case BuiltinType::Id: \
2217  Width = Size; \
2218  Align = Size; \
2219  break;
2220 #include "clang/Basic/PPCTypes.def"
2221 #define RVV_VECTOR_TYPE(Name, Id, SingletonId, ElKind, ElBits, NF, IsSigned, \
2222  IsFP) \
2223  case BuiltinType::Id: \
2224  Width = 0; \
2225  Align = ElBits; \
2226  break;
2227 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, ElKind) \
2228  case BuiltinType::Id: \
2229  Width = 0; \
2230  Align = 8; \
2231  break;
2232 #include "clang/Basic/RISCVVTypes.def"
2233  }
2234  break;
2235  case Type::ObjCObjectPointer:
2236  Width = Target->getPointerWidth(0);
2237  Align = Target->getPointerAlign(0);
2238  break;
2239  case Type::BlockPointer:
2240  AS = getTargetAddressSpace(cast<BlockPointerType>(T)->getPointeeType());
2241  Width = Target->getPointerWidth(AS);
2242  Align = Target->getPointerAlign(AS);
2243  break;
2244  case Type::LValueReference:
2245  case Type::RValueReference:
2246  // alignof and sizeof should never enter this code path here, so we go
2247  // the pointer route.
2248  AS = getTargetAddressSpace(cast<ReferenceType>(T)->getPointeeType());
2249  Width = Target->getPointerWidth(AS);
2250  Align = Target->getPointerAlign(AS);
2251  break;
2252  case Type::Pointer:
2253  AS = getTargetAddressSpace(cast<PointerType>(T)->getPointeeType());
2254  Width = Target->getPointerWidth(AS);
2255  Align = Target->getPointerAlign(AS);
2256  break;
2257  case Type::MemberPointer: {
2258  const auto *MPT = cast<MemberPointerType>(T);
2259  CXXABI::MemberPointerInfo MPI = ABI->getMemberPointerInfo(MPT);
2260  Width = MPI.Width;
2261  Align = MPI.Align;
2262  break;
2263  }
2264  case Type::Complex: {
2265  // Complex types have the same alignment as their elements, but twice the
2266  // size.
2267  TypeInfo EltInfo = getTypeInfo(cast<ComplexType>(T)->getElementType());
2268  Width = EltInfo.Width * 2;
2269  Align = EltInfo.Align;
2270  break;
2271  }
2272  case Type::ObjCObject:
2273  return getTypeInfo(cast<ObjCObjectType>(T)->getBaseType().getTypePtr());
2274  case Type::Adjusted:
2275  case Type::Decayed:
2276  return getTypeInfo(cast<AdjustedType>(T)->getAdjustedType().getTypePtr());
2277  case Type::ObjCInterface: {
2278  const auto *ObjCI = cast<ObjCInterfaceType>(T);
2279  if (ObjCI->getDecl()->isInvalidDecl()) {
2280  Width = 8;
2281  Align = 8;
2282  break;
2283  }
2284  const ASTRecordLayout &Layout = getASTObjCInterfaceLayout(ObjCI->getDecl());
2285  Width = toBits(Layout.getSize());
2286  Align = toBits(Layout.getAlignment());
2287  break;
2288  }
2289  case Type::BitInt: {
2290  const auto *EIT = cast<BitIntType>(T);
2291  Align =
2292  std::min(static_cast<unsigned>(std::max(
2293  getCharWidth(), llvm::PowerOf2Ceil(EIT->getNumBits()))),
2294  Target->getLongLongAlign());
2295  Width = llvm::alignTo(EIT->getNumBits(), Align);
2296  break;
2297  }
2298  case Type::Record:
2299  case Type::Enum: {
2300  const auto *TT = cast<TagType>(T);
2301 
2302  if (TT->getDecl()->isInvalidDecl()) {
2303  Width = 8;
2304  Align = 8;
2305  break;
2306  }
2307 
2308  if (const auto *ET = dyn_cast<EnumType>(TT)) {
2309  const EnumDecl *ED = ET->getDecl();
2310  TypeInfo Info =
2312  if (unsigned AttrAlign = ED->getMaxAlignment()) {
2313  Info.Align = AttrAlign;
2315  }
2316  return Info;
2317  }
2318 
2319  const auto *RT = cast<RecordType>(TT);
2320  const RecordDecl *RD = RT->getDecl();
2321  const ASTRecordLayout &Layout = getASTRecordLayout(RD);
2322  Width = toBits(Layout.getSize());
2323  Align = toBits(Layout.getAlignment());
2324  AlignRequirement = RD->hasAttr<AlignedAttr>()
2327  break;
2328  }
2329 
2330  case Type::SubstTemplateTypeParm:
2331  return getTypeInfo(cast<SubstTemplateTypeParmType>(T)->
2332  getReplacementType().getTypePtr());
2333 
2334  case Type::Auto:
2335  case Type::DeducedTemplateSpecialization: {
2336  const auto *A = cast<DeducedType>(T);
2337  assert(!A->getDeducedType().isNull() &&
2338  "cannot request the size of an undeduced or dependent auto type");
2339  return getTypeInfo(A->getDeducedType().getTypePtr());
2340  }
2341 
2342  case Type::Paren:
2343  return getTypeInfo(cast<ParenType>(T)->getInnerType().getTypePtr());
2344 
2345  case Type::MacroQualified:
2346  return getTypeInfo(
2347  cast<MacroQualifiedType>(T)->getUnderlyingType().getTypePtr());
2348 
2349  case Type::ObjCTypeParam:
2350  return getTypeInfo(cast<ObjCTypeParamType>(T)->desugar().getTypePtr());
2351 
2352  case Type::Using:
2353  return getTypeInfo(cast<UsingType>(T)->desugar().getTypePtr());
2354 
2355  case Type::Typedef: {
2356  const TypedefNameDecl *Typedef = cast<TypedefType>(T)->getDecl();
2357  TypeInfo Info = getTypeInfo(Typedef->getUnderlyingType().getTypePtr());
2358  // If the typedef has an aligned attribute on it, it overrides any computed
2359  // alignment we have. This violates the GCC documentation (which says that
2360  // attribute(aligned) can only round up) but matches its implementation.
2361  if (unsigned AttrAlign = Typedef->getMaxAlignment()) {
2362  Align = AttrAlign;
2363  AlignRequirement = AlignRequirementKind::RequiredByTypedef;
2364  } else {
2365  Align = Info.Align;
2366  AlignRequirement = Info.AlignRequirement;
2367  }
2368  Width = Info.Width;
2369  break;
2370  }
2371 
2372  case Type::Elaborated:
2373  return getTypeInfo(cast<ElaboratedType>(T)->getNamedType().getTypePtr());
2374 
2375  case Type::Attributed:
2376  return getTypeInfo(
2377  cast<AttributedType>(T)->getEquivalentType().getTypePtr());
2378 
2379  case Type::Atomic: {
2380  // Start with the base type information.
2381  TypeInfo Info = getTypeInfo(cast<AtomicType>(T)->getValueType());
2382  Width = Info.Width;
2383  Align = Info.Align;
2384 
2385  if (!Width) {
2386  // An otherwise zero-sized type should still generate an
2387  // atomic operation.
2388  Width = Target->getCharWidth();
2389  assert(Align);
2390  } else if (Width <= Target->getMaxAtomicPromoteWidth()) {
2391  // If the size of the type doesn't exceed the platform's max
2392  // atomic promotion width, make the size and alignment more
2393  // favorable to atomic operations:
2394 
2395  // Round the size up to a power of 2.
2396  if (!llvm::isPowerOf2_64(Width))
2397  Width = llvm::NextPowerOf2(Width);
2398 
2399  // Set the alignment equal to the size.
2400  Align = static_cast<unsigned>(Width);
2401  }
2402  }
2403  break;
2404 
2405  case Type::Pipe:
2408  break;
2409  }
2410 
2411  assert(llvm::isPowerOf2_32(Align) && "Alignment must be power of 2");
2412  return TypeInfo(Width, Align, AlignRequirement);
2413 }
2414 
2415 unsigned ASTContext::getTypeUnadjustedAlign(const Type *T) const {
2416  UnadjustedAlignMap::iterator I = MemoizedUnadjustedAlign.find(T);
2417  if (I != MemoizedUnadjustedAlign.end())
2418  return I->second;
2419 
2420  unsigned UnadjustedAlign;
2421  if (const auto *RT = T->getAs<RecordType>()) {
2422  const RecordDecl *RD = RT->getDecl();
2423  const ASTRecordLayout &Layout = getASTRecordLayout(RD);
2424  UnadjustedAlign = toBits(Layout.getUnadjustedAlignment());
2425  } else if (const auto *ObjCI = T->getAs<ObjCInterfaceType>()) {
2426  const ASTRecordLayout &Layout = getASTObjCInterfaceLayout(ObjCI->getDecl());
2427  UnadjustedAlign = toBits(Layout.getUnadjustedAlignment());
2428  } else {
2429  UnadjustedAlign = getTypeAlign(T->getUnqualifiedDesugaredType());
2430  }
2431 
2432  MemoizedUnadjustedAlign[T] = UnadjustedAlign;
2433  return UnadjustedAlign;
2434 }
2435 
2437  unsigned SimdAlign = getTargetInfo().getSimdDefaultAlign();
2438  return SimdAlign;
2439 }
2440 
2441 /// toCharUnitsFromBits - Convert a size in bits to a size in characters.
2443  return CharUnits::fromQuantity(BitSize / getCharWidth());
2444 }
2445 
2446 /// toBits - Convert a size in characters to a size in characters.
2447 int64_t ASTContext::toBits(CharUnits CharSize) const {
2448  return CharSize.getQuantity() * getCharWidth();
2449 }
2450 
2451 /// getTypeSizeInChars - Return the size of the specified type, in characters.
2452 /// This method does not work on incomplete types.
2454  return getTypeInfoInChars(T).Width;
2455 }
2457  return getTypeInfoInChars(T).Width;
2458 }
2459 
2460 /// getTypeAlignInChars - Return the ABI-specified alignment of a type, in
2461 /// characters. This method does not work on incomplete types.
2463  return toCharUnitsFromBits(getTypeAlign(T));
2464 }
2466  return toCharUnitsFromBits(getTypeAlign(T));
2467 }
2468 
2469 /// getTypeUnadjustedAlignInChars - Return the ABI-specified alignment of a
2470 /// type, in characters, before alignment adustments. This method does
2471 /// not work on incomplete types.
2474 }
2477 }
2478 
2479 /// getPreferredTypeAlign - Return the "preferred" alignment of the specified
2480 /// type for the current target in bits. This can be different than the ABI
2481 /// alignment in cases where it is beneficial for performance or backwards
2482 /// compatibility preserving to overalign a data type. (Note: despite the name,
2483 /// the preferred alignment is ABI-impacting, and not an optimization.)
2484 unsigned ASTContext::getPreferredTypeAlign(const Type *T) const {
2485  TypeInfo TI = getTypeInfo(T);
2486  unsigned ABIAlign = TI.Align;
2487 
2488  T = T->getBaseElementTypeUnsafe();
2489 
2490  // The preferred alignment of member pointers is that of a pointer.
2491  if (T->isMemberPointerType())
2492  return getPreferredTypeAlign(getPointerDiffType().getTypePtr());
2493 
2494  if (!Target->allowsLargerPreferedTypeAlignment())
2495  return ABIAlign;
2496 
2497  if (const auto *RT = T->getAs<RecordType>()) {
2498  const RecordDecl *RD = RT->getDecl();
2499 
2500  // When used as part of a typedef, or together with a 'packed' attribute,
2501  // the 'aligned' attribute can be used to decrease alignment. Note that the
2502  // 'packed' case is already taken into consideration when computing the
2503  // alignment, we only need to handle the typedef case here.
2505  RD->isInvalidDecl())
2506  return ABIAlign;
2507 
2508  unsigned PreferredAlign = static_cast<unsigned>(
2509  toBits(getASTRecordLayout(RD).PreferredAlignment));
2510  assert(PreferredAlign >= ABIAlign &&
2511  "PreferredAlign should be at least as large as ABIAlign.");
2512  return PreferredAlign;
2513  }
2514 
2515  // Double (and, for targets supporting AIX `power` alignment, long double) and
2516  // long long should be naturally aligned (despite requiring less alignment) if
2517  // possible.
2518  if (const auto *CT = T->getAs<ComplexType>())
2519  T = CT->getElementType().getTypePtr();
2520  if (const auto *ET = T->getAs<EnumType>())
2521  T = ET->getDecl()->getIntegerType().getTypePtr();
2522  if (T->isSpecificBuiltinType(BuiltinType::Double) ||
2523  T->isSpecificBuiltinType(BuiltinType::LongLong) ||
2524  T->isSpecificBuiltinType(BuiltinType::ULongLong) ||
2525  (T->isSpecificBuiltinType(BuiltinType::LongDouble) &&
2526  Target->defaultsToAIXPowerAlignment()))
2527  // Don't increase the alignment if an alignment attribute was specified on a
2528  // typedef declaration.
2529  if (!TI.isAlignRequired())
2530  return std::max(ABIAlign, (unsigned)getTypeSize(T));
2531 
2532  return ABIAlign;
2533 }
2534 
2535 /// getTargetDefaultAlignForAttributeAligned - Return the default alignment
2536 /// for __attribute__((aligned)) on this target, to be used if no alignment
2537 /// value is specified.
2540 }
2541 
2542 /// getAlignOfGlobalVar - Return the alignment in bits that should be given
2543 /// to a global variable of the specified type.
2545  uint64_t TypeSize = getTypeSize(T.getTypePtr());
2546  return std::max(getPreferredTypeAlign(T),
2547  getTargetInfo().getMinGlobalAlign(TypeSize));
2548 }
2549 
2550 /// getAlignOfGlobalVarInChars - Return the alignment in characters that
2551 /// should be given to a global variable of the specified type.
2554 }
2555 
2558  const ASTRecordLayout *Layout = &getASTRecordLayout(RD);
2559  while (const CXXRecordDecl *Base = Layout->getBaseSharingVBPtr()) {
2560  Offset += Layout->getBaseClassOffset(Base);
2561  Layout = &getASTRecordLayout(Base);
2562  }
2563  return Offset;
2564 }
2565 
2567  const ValueDecl *MPD = MP.getMemberPointerDecl();
2570  bool DerivedMember = MP.isMemberPointerToDerivedMember();
2571  const CXXRecordDecl *RD = cast<CXXRecordDecl>(MPD->getDeclContext());
2572  for (unsigned I = 0, N = Path.size(); I != N; ++I) {
2573  const CXXRecordDecl *Base = RD;
2574  const CXXRecordDecl *Derived = Path[I];
2575  if (DerivedMember)
2576  std::swap(Base, Derived);
2578  RD = Path[I];
2579  }
2580  if (DerivedMember)
2582  return ThisAdjustment;
2583 }
2584 
2585 /// DeepCollectObjCIvars -
2586 /// This routine first collects all declared, but not synthesized, ivars in
2587 /// super class and then collects all ivars, including those synthesized for
2588 /// current class. This routine is used for implementation of current class
2589 /// when all ivars, declared and synthesized are known.
2591  bool leafClass,
2592  SmallVectorImpl<const ObjCIvarDecl*> &Ivars) const {
2593  if (const ObjCInterfaceDecl *SuperClass = OI->getSuperClass())
2594  DeepCollectObjCIvars(SuperClass, false, Ivars);
2595  if (!leafClass) {
2596  for (const auto *I : OI->ivars())
2597  Ivars.push_back(I);
2598  } else {
2599  auto *IDecl = const_cast<ObjCInterfaceDecl *>(OI);
2600  for (const ObjCIvarDecl *Iv = IDecl->all_declared_ivar_begin(); Iv;
2601  Iv= Iv->getNextIvar())
2602  Ivars.push_back(Iv);
2603  }
2604 }
2605 
2606 /// CollectInheritedProtocols - Collect all protocols in current class and
2607 /// those inherited by it.
2610  if (const auto *OI = dyn_cast<ObjCInterfaceDecl>(CDecl)) {
2611  // We can use protocol_iterator here instead of
2612  // all_referenced_protocol_iterator since we are walking all categories.
2613  for (auto *Proto : OI->all_referenced_protocols()) {
2614  CollectInheritedProtocols(Proto, Protocols);
2615  }
2616 
2617  // Categories of this Interface.
2618  for (const auto *Cat : OI->visible_categories())
2619  CollectInheritedProtocols(Cat, Protocols);
2620 
2621  if (ObjCInterfaceDecl *SD = OI->getSuperClass())
2622  while (SD) {
2623  CollectInheritedProtocols(SD, Protocols);
2624  SD = SD->getSuperClass();
2625  }
2626  } else if (const auto *OC = dyn_cast<ObjCCategoryDecl>(CDecl)) {
2627  for (auto *Proto : OC->protocols()) {
2628  CollectInheritedProtocols(Proto, Protocols);
2629  }
2630  } else if (const auto *OP = dyn_cast<ObjCProtocolDecl>(CDecl)) {
2631  // Insert the protocol.
2632  if (!Protocols.insert(
2633  const_cast<ObjCProtocolDecl *>(OP->getCanonicalDecl())).second)
2634  return;
2635 
2636  for (auto *Proto : OP->protocols())
2637  CollectInheritedProtocols(Proto, Protocols);
2638  }
2639 }
2640 
2642  const RecordDecl *RD) {
2643  assert(RD->isUnion() && "Must be union type");
2644  CharUnits UnionSize = Context.getTypeSizeInChars(RD->getTypeForDecl());
2645 
2646  for (const auto *Field : RD->fields()) {
2647  if (!Context.hasUniqueObjectRepresentations(Field->getType()))
2648  return false;
2649  CharUnits FieldSize = Context.getTypeSizeInChars(Field->getType());
2650  if (FieldSize != UnionSize)
2651  return false;
2652  }
2653  return !RD->field_empty();
2654 }
2655 
2656 static int64_t getSubobjectOffset(const FieldDecl *Field,
2657  const ASTContext &Context,
2658  const clang::ASTRecordLayout & /*Layout*/) {
2659  return Context.getFieldOffset(Field);
2660 }
2661 
2662 static int64_t getSubobjectOffset(const CXXRecordDecl *RD,
2663  const ASTContext &Context,
2664  const clang::ASTRecordLayout &Layout) {
2665  return Context.toBits(Layout.getBaseClassOffset(RD));
2666 }
2667 
2670  const RecordDecl *RD);
2671 
2673 getSubobjectSizeInBits(const FieldDecl *Field, const ASTContext &Context) {
2674  if (Field->getType()->isRecordType()) {
2675  const RecordDecl *RD = Field->getType()->getAsRecordDecl();
2676  if (!RD->isUnion())
2677  return structHasUniqueObjectRepresentations(Context, RD);
2678  }
2679  if (!Field->getType()->isReferenceType() &&
2680  !Context.hasUniqueObjectRepresentations(Field->getType()))
2681  return llvm::None;
2682 
2683  int64_t FieldSizeInBits =
2684  Context.toBits(Context.getTypeSizeInChars(Field->getType()));
2685  if (Field->isBitField()) {
2686  int64_t BitfieldSize = Field->getBitWidthValue(Context);
2687  if (BitfieldSize > FieldSizeInBits)
2688  return llvm::None;
2689  FieldSizeInBits = BitfieldSize;
2690  }
2691  return FieldSizeInBits;
2692 }
2693 
2696  return structHasUniqueObjectRepresentations(Context, RD);
2697 }
2698 
2699 template <typename RangeT>
2701  const RangeT &Subobjects, int64_t CurOffsetInBits,
2702  const ASTContext &Context, const clang::ASTRecordLayout &Layout) {
2703  for (const auto *Subobject : Subobjects) {
2704  llvm::Optional<int64_t> SizeInBits =
2705  getSubobjectSizeInBits(Subobject, Context);
2706  if (!SizeInBits)
2707  return llvm::None;
2708  if (*SizeInBits != 0) {
2709  int64_t Offset = getSubobjectOffset(Subobject, Context, Layout);
2710  if (Offset != CurOffsetInBits)
2711  return llvm::None;
2712  CurOffsetInBits += *SizeInBits;
2713  }
2714  }
2715  return CurOffsetInBits;
2716 }
2717 
2720  const RecordDecl *RD) {
2721  assert(!RD->isUnion() && "Must be struct/class type");
2722  const auto &Layout = Context.getASTRecordLayout(RD);
2723 
2724  int64_t CurOffsetInBits = 0;
2725  if (const auto *ClassDecl = dyn_cast<CXXRecordDecl>(RD)) {
2726  if (ClassDecl->isDynamicClass())
2727  return llvm::None;
2728 
2730  for (const auto &Base : ClassDecl->bases()) {
2731  // Empty types can be inherited from, and non-empty types can potentially
2732  // have tail padding, so just make sure there isn't an error.
2733  Bases.emplace_back(Base.getType()->getAsCXXRecordDecl());
2734  }
2735 
2736  llvm::sort(Bases, [&](const CXXRecordDecl *L, const CXXRecordDecl *R) {
2737  return Layout.getBaseClassOffset(L) < Layout.getBaseClassOffset(R);
2738  });
2739 
2740  llvm::Optional<int64_t> OffsetAfterBases =
2741  structSubobjectsHaveUniqueObjectRepresentations(Bases, CurOffsetInBits,
2742  Context, Layout);
2743  if (!OffsetAfterBases)
2744  return llvm::None;
2745  CurOffsetInBits = *OffsetAfterBases;
2746  }
2747 
2748  llvm::Optional<int64_t> OffsetAfterFields =
2750  RD->fields(), CurOffsetInBits, Context, Layout);
2751  if (!OffsetAfterFields)
2752  return llvm::None;
2753  CurOffsetInBits = *OffsetAfterFields;
2754 
2755  return CurOffsetInBits;
2756 }
2757 
2759  // C++17 [meta.unary.prop]:
2760  // The predicate condition for a template specialization
2761  // has_unique_object_representations<T> shall be
2762  // satisfied if and only if:
2763  // (9.1) - T is trivially copyable, and
2764  // (9.2) - any two objects of type T with the same value have the same
2765  // object representation, where two objects
2766  // of array or non-union class type are considered to have the same value
2767  // if their respective sequences of
2768  // direct subobjects have the same values, and two objects of union type
2769  // are considered to have the same
2770  // value if they have the same active member and the corresponding members
2771  // have the same value.
2772  // The set of scalar types for which this condition holds is
2773  // implementation-defined. [ Note: If a type has padding
2774  // bits, the condition does not hold; otherwise, the condition holds true
2775  // for unsigned integral types. -- end note ]
2776  assert(!Ty.isNull() && "Null QualType sent to unique object rep check");
2777 
2778  // Arrays are unique only if their element type is unique.
2779  if (Ty->isArrayType())
2781 
2782  // (9.1) - T is trivially copyable...
2783  if (!Ty.isTriviallyCopyableType(*this))
2784  return false;
2785 
2786  // All integrals and enums are unique.
2787  if (Ty->isIntegralOrEnumerationType())
2788  return true;
2789 
2790  // All other pointers are unique.
2791  if (Ty->isPointerType())
2792  return true;
2793 
2794  if (Ty->isMemberPointerType()) {
2795  const auto *MPT = Ty->getAs<MemberPointerType>();
2796  return !ABI->getMemberPointerInfo(MPT).HasPadding;
2797  }
2798 
2799  if (Ty->isRecordType()) {
2800  const RecordDecl *Record = Ty->castAs<RecordType>()->getDecl();
2801 
2802  if (Record->isInvalidDecl())
2803  return false;
2804 
2805  if (Record->isUnion())
2806  return unionHasUniqueObjectRepresentations(*this, Record);
2807 
2808  Optional<int64_t> StructSize =
2809  structHasUniqueObjectRepresentations(*this, Record);
2810 
2811  return StructSize &&
2812  StructSize.getValue() == static_cast<int64_t>(getTypeSize(Ty));
2813  }
2814 
2815  // FIXME: More cases to handle here (list by rsmith):
2816  // vectors (careful about, eg, vector of 3 foo)
2817  // _Complex int and friends
2818  // _Atomic T
2819  // Obj-C block pointers
2820  // Obj-C object pointers
2821  // and perhaps OpenCL's various builtin types (pipe, sampler_t, event_t,
2822  // clk_event_t, queue_t, reserve_id_t)
2823  // There're also Obj-C class types and the Obj-C selector type, but I think it
2824  // makes sense for those to return false here.
2825 
2826  return false;
2827 }
2828 
2830  unsigned count = 0;
2831  // Count ivars declared in class extension.
2832  for (const auto *Ext : OI->known_extensions())
2833  count += Ext->ivar_size();
2834 
2835  // Count ivar defined in this class's implementation. This
2836  // includes synthesized ivars.
2837  if (ObjCImplementationDecl *ImplDecl = OI->getImplementation())
2838  count += ImplDecl->ivar_size();
2839 
2840  return count;
2841 }
2842 
2844  if (!E)
2845  return false;
2846 
2847  // nullptr_t is always treated as null.
2848  if (E->getType()->isNullPtrType()) return true;
2849 
2850  if (E->getType()->isAnyPointerType() &&
2853  return true;
2854 
2855  // Unfortunately, __null has type 'int'.
2856  if (isa<GNUNullExpr>(E)) return true;
2857 
2858  return false;
2859 }
2860 
2861 /// Get the implementation of ObjCInterfaceDecl, or nullptr if none
2862 /// exists.
2864  llvm::DenseMap<ObjCContainerDecl*, ObjCImplDecl*>::iterator
2865  I = ObjCImpls.find(D);
2866  if (I != ObjCImpls.end())
2867  return cast<ObjCImplementationDecl>(I->second);
2868  return nullptr;
2869 }
2870 
2871 /// Get the implementation of ObjCCategoryDecl, or nullptr if none
2872 /// exists.
2874  llvm::DenseMap<ObjCContainerDecl*, ObjCImplDecl*>::iterator
2875  I = ObjCImpls.find(D);
2876  if (I != ObjCImpls.end())
2877  return cast<ObjCCategoryImplDecl>(I->second);
2878  return nullptr;
2879 }
2880 
2881 /// Set the implementation of ObjCInterfaceDecl.
2883  ObjCImplementationDecl *ImplD) {
2884  assert(IFaceD && ImplD && "Passed null params");
2885  ObjCImpls[IFaceD] = ImplD;
2886 }
2887 
2888 /// Set the implementation of ObjCCategoryDecl.
2890  ObjCCategoryImplDecl *ImplD) {
2891  assert(CatD && ImplD && "Passed null params");
2892  ObjCImpls[CatD] = ImplD;
2893 }
2894 
2895 const ObjCMethodDecl *
2897  return ObjCMethodRedecls.lookup(MD);
2898 }
2899 
2901  const ObjCMethodDecl *Redecl) {
2902  assert(!getObjCMethodRedeclaration(MD) && "MD already has a redeclaration");
2903  ObjCMethodRedecls[MD] = Redecl;
2904 }
2905 
2907  const NamedDecl *ND) const {
2908  if (const auto *ID = dyn_cast<ObjCInterfaceDecl>(ND->getDeclContext()))
2909  return ID;
2910  if (const auto *CD = dyn_cast<ObjCCategoryDecl>(ND->getDeclContext()))
2911  return CD->getClassInterface();
2912  if (const auto *IMD = dyn_cast<ObjCImplDecl>(ND->getDeclContext()))
2913  return IMD->getClassInterface();
2914 
2915  return nullptr;
2916 }
2917 
2918 /// Get the copy initialization expression of VarDecl, or nullptr if
2919 /// none exists.
2921  assert(VD && "Passed null params");
2922  assert(VD->hasAttr<BlocksAttr>() &&
2923  "getBlockVarCopyInits - not __block var");
2924  auto I = BlockVarCopyInits.find(VD);
2925  if (I != BlockVarCopyInits.end())
2926  return I->second;
2927  return {nullptr, false};
2928 }
2929 
2930 /// Set the copy initialization expression of a block var decl.
2932  bool CanThrow) {
2933  assert(VD && CopyExpr && "Passed null params");
2934  assert(VD->hasAttr<BlocksAttr>() &&
2935  "setBlockVarCopyInits - not __block var");
2936  BlockVarCopyInits[VD].setExprAndFlag(CopyExpr, CanThrow);
2937 }
2938 
2940  unsigned DataSize) const {
2941  if (!DataSize)
2942  DataSize = TypeLoc::getFullDataSizeForType(T);
2943  else
2944  assert(DataSize == TypeLoc::getFullDataSizeForType(T) &&
2945  "incorrect data size provided to CreateTypeSourceInfo!");
2946 
2947  auto *TInfo =
2948  (TypeSourceInfo*)BumpAlloc.Allocate(sizeof(TypeSourceInfo) + DataSize, 8);
2949  new (TInfo) TypeSourceInfo(T);
2950  return TInfo;
2951 }
2952 
2954  SourceLocation L) const {
2956  DI->getTypeLoc().initialize(const_cast<ASTContext &>(*this), L);
2957  return DI;
2958 }
2959 
2960 const ASTRecordLayout &
2962  return getObjCLayout(D, nullptr);
2963 }
2964 
2965 const ASTRecordLayout &
2967  const ObjCImplementationDecl *D) const {
2968  return getObjCLayout(D->getClassInterface(), D);
2969 }
2970 
2971 //===----------------------------------------------------------------------===//
2972 // Type creation/memoization methods
2973 //===----------------------------------------------------------------------===//
2974 
2975 QualType
2976 ASTContext::getExtQualType(const Type *baseType, Qualifiers quals) const {
2977  unsigned fastQuals = quals.getFastQualifiers();
2978  quals.removeFastQualifiers();
2979 
2980  // Check if we've already instantiated this type.
2981  llvm::FoldingSetNodeID ID;
2982  ExtQuals::Profile(ID, baseType, quals);
2983  void *insertPos = nullptr;
2984  if (ExtQuals *eq = ExtQualNodes.FindNodeOrInsertPos(ID, insertPos)) {
2985  assert(eq->getQualifiers() == quals);
2986  return QualType(eq, fastQuals);
2987  }
2988 
2989  // If the base type is not canonical, make the appropriate canonical type.
2990  QualType canon;
2991  if (!baseType->isCanonicalUnqualified()) {
2992  SplitQualType canonSplit = baseType->getCanonicalTypeInternal().split();
2993  canonSplit.Quals.addConsistentQualifiers(quals);
2994  canon = getExtQualType(canonSplit.Ty, canonSplit.Quals);
2995 
2996  // Re-find the insert position.
2997  (void) ExtQualNodes.FindNodeOrInsertPos(ID, insertPos);
2998  }
2999 
3000  auto *eq = new (*this, TypeAlignment) ExtQuals(baseType, canon, quals);
3001  ExtQualNodes.InsertNode(eq, insertPos);
3002  return QualType(eq, fastQuals);
3003 }
3004 
3006  LangAS AddressSpace) const {
3007  QualType CanT = getCanonicalType(T);
3008  if (CanT.getAddressSpace() == AddressSpace)
3009  return T;
3010 
3011  // If we are composing extended qualifiers together, merge together
3012  // into one ExtQuals node.
3013  QualifierCollector Quals;
3014  const Type *TypeNode = Quals.strip(T);
3015 
3016  // If this type already has an address space specified, it cannot get
3017  // another one.
3018  assert(!Quals.hasAddressSpace() &&
3019  "Type cannot be in multiple addr spaces!");
3020  Quals.addAddressSpace(AddressSpace);
3021 
3022  return getExtQualType(TypeNode, Quals);
3023 }
3024 
3026  // If the type is not qualified with an address space, just return it
3027  // immediately.
3028  if (!T.hasAddressSpace())
3029  return T;
3030 
3031  // If we are composing extended qualifiers together, merge together
3032  // into one ExtQuals node.
3033  QualifierCollector Quals;
3034  const Type *TypeNode;
3035 
3036  while (T.hasAddressSpace()) {
3037  TypeNode = Quals.strip(T);
3038 
3039  // If the type no longer has an address space after stripping qualifiers,
3040  // jump out.
3041  if (!QualType(TypeNode, 0).hasAddressSpace())
3042  break;
3043 
3044  // There might be sugar in the way. Strip it and try again.
3045  T = T.getSingleStepDesugaredType(*this);
3046  }
3047 
3048  Quals.removeAddressSpace();
3049 
3050  // Removal of the address space can mean there are no longer any
3051  // non-fast qualifiers, so creating an ExtQualType isn't possible (asserts)
3052  // or required.
3053  if (Quals.hasNonFastQualifiers())
3054  return getExtQualType(TypeNode, Quals);
3055  else
3056  return QualType(TypeNode, Quals.getFastQualifiers());
3057 }
3058 
3060  Qualifiers::GC GCAttr) const {
3061  QualType CanT = getCanonicalType(T);
3062  if (CanT.getObjCGCAttr() == GCAttr)
3063  return T;
3064 
3065  if (const auto *ptr = T->getAs<PointerType>()) {
3066  QualType Pointee = ptr->getPointeeType();
3067  if (Pointee->isAnyPointerType()) {
3068  QualType ResultType = getObjCGCQualType(Pointee, GCAttr);
3069  return getPointerType(ResultType);
3070  }
3071  }
3072 
3073  // If we are composing extended qualifiers together, merge together
3074  // into one ExtQuals node.
3075  QualifierCollector Quals;
3076  const Type *TypeNode = Quals.strip(T);
3077 
3078  // If this type already has an ObjCGC specified, it cannot get
3079  // another one.
3080  assert(!Quals.hasObjCGCAttr() &&
3081  "Type cannot have multiple ObjCGCs!");
3082  Quals.addObjCGCAttr(GCAttr);
3083 
3084  return getExtQualType(TypeNode, Quals);
3085 }
3086 
3088  if (const PointerType *Ptr = T->getAs<PointerType>()) {
3089  QualType Pointee = Ptr->getPointeeType();
3090  if (isPtrSizeAddressSpace(Pointee.getAddressSpace())) {
3091  return getPointerType(removeAddrSpaceQualType(Pointee));
3092  }
3093  }
3094  return T;
3095 }
3096 
3098  FunctionType::ExtInfo Info) {
3099  if (T->getExtInfo() == Info)
3100  return T;
3101 
3102  QualType Result;
3103  if (const auto *FNPT = dyn_cast<FunctionNoProtoType>(T)) {
3104  Result = getFunctionNoProtoType(FNPT->getReturnType(), Info);
3105  } else {
3106  const auto *FPT = cast<FunctionProtoType>(T);
3107  FunctionProtoType::ExtProtoInfo EPI = FPT->getExtProtoInfo();
3108  EPI.ExtInfo = Info;
3109  Result = getFunctionType(FPT->getReturnType(), FPT->getParamTypes(), EPI);
3110  }
3111 
3112  return cast<FunctionType>(Result.getTypePtr());
3113 }
3114 
3116  QualType ResultType) {
3117  FD = FD->getMostRecentDecl();
3118  while (true) {
3119  const auto *FPT = FD->getType()->castAs<FunctionProtoType>();
3120  FunctionProtoType::ExtProtoInfo EPI = FPT->getExtProtoInfo();
3121  FD->setType(getFunctionType(ResultType, FPT->getParamTypes(), EPI));
3122  if (FunctionDecl *Next = FD->getPreviousDecl())
3123  FD = Next;
3124  else
3125  break;
3126  }
3128  L->DeducedReturnType(FD, ResultType);
3129 }
3130 
3131 /// Get a function type and produce the equivalent function type with the
3132 /// specified exception specification. Type sugar that can be present on a
3133 /// declaration of a function with an exception specification is permitted
3134 /// and preserved. Other type sugar (for instance, typedefs) is not.
3137  // Might have some parens.
3138  if (const auto *PT = dyn_cast<ParenType>(Orig))
3139  return getParenType(
3140  getFunctionTypeWithExceptionSpec(PT->getInnerType(), ESI));
3141 
3142  // Might be wrapped in a macro qualified type.
3143  if (const auto *MQT = dyn_cast<MacroQualifiedType>(Orig))
3144  return getMacroQualifiedType(
3145  getFunctionTypeWithExceptionSpec(MQT->getUnderlyingType(), ESI),
3146  MQT->getMacroIdentifier());
3147 
3148  // Might have a calling-convention attribute.
3149  if (const auto *AT = dyn_cast<AttributedType>(Orig))
3150  return getAttributedType(
3151  AT->getAttrKind(),
3152  getFunctionTypeWithExceptionSpec(AT->getModifiedType(), ESI),
3153  getFunctionTypeWithExceptionSpec(AT->getEquivalentType(), ESI));
3154 
3155  // Anything else must be a function type. Rebuild it with the new exception
3156  // specification.
3157  const auto *Proto = Orig->castAs<FunctionProtoType>();
3158  return getFunctionType(
3159  Proto->getReturnType(), Proto->getParamTypes(),
3160  Proto->getExtProtoInfo().withExceptionSpec(ESI));
3161 }
3162 
3164  QualType U) {
3165  return hasSameType(T, U) ||
3166  (getLangOpts().CPlusPlus17 &&
3169 }
3170 
3172  if (const auto *Proto = T->getAs<FunctionProtoType>()) {
3173  QualType RetTy = removePtrSizeAddrSpace(Proto->getReturnType());
3174  SmallVector<QualType, 16> Args(Proto->param_types());
3175  for (unsigned i = 0, n = Args.size(); i != n; ++i)
3176  Args[i] = removePtrSizeAddrSpace(Args[i]);
3177  return getFunctionType(RetTy, Args, Proto->getExtProtoInfo());
3178  }
3179 
3180  if (const FunctionNoProtoType *Proto = T->getAs<FunctionNoProtoType>()) {
3181  QualType RetTy = removePtrSizeAddrSpace(Proto->getReturnType());
3182  return getFunctionNoProtoType(RetTy, Proto->getExtInfo());
3183  }
3184 
3185  return T;
3186 }
3187 
3189  return hasSameType(T, U) ||
3192 }
3193 
3196  bool AsWritten) {
3197  // Update the type.
3198  QualType Updated =
3200  FD->setType(Updated);
3201 
3202  if (!AsWritten)
3203  return;
3204 
3205  // Update the type in the type source information too.
3206  if (TypeSourceInfo *TSInfo = FD->getTypeSourceInfo()) {
3207  // If the type and the type-as-written differ, we may need to update
3208  // the type-as-written too.
3209  if (TSInfo->getType() != FD->getType())
3210  Updated = getFunctionTypeWithExceptionSpec(TSInfo->getType(), ESI);
3211 
3212  // FIXME: When we get proper type location information for exceptions,
3213  // we'll also have to rebuild the TypeSourceInfo. For now, we just patch
3214  // up the TypeSourceInfo;
3215  assert(TypeLoc::getFullDataSizeForType(Updated) ==
3216  TypeLoc::getFullDataSizeForType(TSInfo->getType()) &&
3217  "TypeLoc size mismatch from updating exception specification");
3218  TSInfo->overrideType(Updated);
3219  }
3220 }
3221 
3222 /// getComplexType - Return the uniqued reference to the type for a complex
3223 /// number with the specified element type.
3225  // Unique pointers, to guarantee there is only one pointer of a particular
3226  // structure.
3227  llvm::FoldingSetNodeID ID;
3229 
3230  void *InsertPos = nullptr;
3231  if (ComplexType *CT = ComplexTypes.FindNodeOrInsertPos(ID, InsertPos))
3232  return QualType(CT, 0);
3233 
3234  // If the pointee type isn't canonical, this won't be a canonical type either,
3235  // so fill in the canonical type field.
3236  QualType Canonical;
3237  if (!T.isCanonical()) {
3238  Canonical = getComplexType(getCanonicalType(T));
3239 
3240  // Get the new insert position for the node we care about.
3241  ComplexType *NewIP = ComplexTypes.FindNodeOrInsertPos(ID, InsertPos);
3242  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3243  }
3244  auto *New = new (*this, TypeAlignment) ComplexType(T, Canonical);
3245  Types.push_back(New);
3246  ComplexTypes.InsertNode(New, InsertPos);
3247  return QualType(New, 0);
3248 }
3249 
3250 /// getPointerType - Return the uniqued reference to the type for a pointer to
3251 /// the specified type.
3253  // Unique pointers, to guarantee there is only one pointer of a particular
3254  // structure.
3255  llvm::FoldingSetNodeID ID;
3257 
3258  void *InsertPos = nullptr;
3259  if (PointerType *PT = PointerTypes.FindNodeOrInsertPos(ID, InsertPos))
3260  return QualType(PT, 0);
3261 
3262  // If the pointee type isn't canonical, this won't be a canonical type either,
3263  // so fill in the canonical type field.
3264  QualType Canonical;
3265  if (!T.isCanonical()) {
3266  Canonical = getPointerType(getCanonicalType(T));
3267 
3268  // Get the new insert position for the node we care about.
3269  PointerType *NewIP = PointerTypes.FindNodeOrInsertPos(ID, InsertPos);
3270  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3271  }
3272  auto *New = new (*this, TypeAlignment) PointerType(T, Canonical);
3273  Types.push_back(New);
3274  PointerTypes.InsertNode(New, InsertPos);
3275  return QualType(New, 0);
3276 }
3277 
3279  llvm::FoldingSetNodeID ID;
3280  AdjustedType::Profile(ID, Orig, New);
3281  void *InsertPos = nullptr;
3282  AdjustedType *AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos);
3283  if (AT)
3284  return QualType(AT, 0);
3285 
3286  QualType Canonical = getCanonicalType(New);
3287 
3288  // Get the new insert position for the node we care about.
3289  AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos);
3290  assert(!AT && "Shouldn't be in the map!");
3291 
3292  AT = new (*this, TypeAlignment)
3293  AdjustedType(Type::Adjusted, Orig, New, Canonical);
3294  Types.push_back(AT);
3295  AdjustedTypes.InsertNode(AT, InsertPos);
3296  return QualType(AT, 0);
3297 }
3298 
3300  assert((T->isArrayType() || T->isFunctionType()) && "T does not decay");
3301 
3302  QualType Decayed;
3303 
3304  // C99 6.7.5.3p7:
3305  // A declaration of a parameter as "array of type" shall be
3306  // adjusted to "qualified pointer to type", where the type
3307  // qualifiers (if any) are those specified within the [ and ] of
3308  // the array type derivation.
3309  if (T->isArrayType())
3310  Decayed = getArrayDecayedType(T);
3311 
3312  // C99 6.7.5.3p8:
3313  // A declaration of a parameter as "function returning type"
3314  // shall be adjusted to "pointer to function returning type", as
3315  // in 6.3.2.1.
3316  if (T->isFunctionType())
3317  Decayed = getPointerType(T);
3318 
3319  llvm::FoldingSetNodeID ID;
3320  AdjustedType::Profile(ID, T, Decayed);
3321  void *InsertPos = nullptr;
3322  AdjustedType *AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos);
3323  if (AT)
3324  return QualType(AT, 0);
3325 
3326  QualType Canonical = getCanonicalType(Decayed);
3327 
3328  // Get the new insert position for the node we care about.
3329  AT = AdjustedTypes.FindNodeOrInsertPos(ID, InsertPos);
3330  assert(!AT && "Shouldn't be in the map!");
3331 
3332  AT = new (*this, TypeAlignment) DecayedType(T, Decayed, Canonical);
3333  Types.push_back(AT);
3334  AdjustedTypes.InsertNode(AT, InsertPos);
3335  return QualType(AT, 0);
3336 }
3337 
3338 /// getBlockPointerType - Return the uniqued reference to the type for
3339 /// a pointer to the specified block.
3341  assert(T->isFunctionType() && "block of function types only");
3342  // Unique pointers, to guarantee there is only one block of a particular
3343  // structure.
3344  llvm::FoldingSetNodeID ID;
3346 
3347  void *InsertPos = nullptr;
3348  if (BlockPointerType *PT =
3349  BlockPointerTypes.FindNodeOrInsertPos(ID, InsertPos))
3350  return QualType(PT, 0);
3351 
3352  // If the block pointee type isn't canonical, this won't be a canonical
3353  // type either so fill in the canonical type field.
3354  QualType Canonical;
3355  if (!T.isCanonical()) {
3356  Canonical = getBlockPointerType(getCanonicalType(T));
3357 
3358  // Get the new insert position for the node we care about.
3359  BlockPointerType *NewIP =
3360  BlockPointerTypes.FindNodeOrInsertPos(ID, InsertPos);
3361  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3362  }
3363  auto *New = new (*this, TypeAlignment) BlockPointerType(T, Canonical);
3364  Types.push_back(New);
3365  BlockPointerTypes.InsertNode(New, InsertPos);
3366  return QualType(New, 0);
3367 }
3368 
3369 /// getLValueReferenceType - Return the uniqued reference to the type for an
3370 /// lvalue reference to the specified type.
3371 QualType
3372 ASTContext::getLValueReferenceType(QualType T, bool SpelledAsLValue) const {
3373  assert(getCanonicalType(T) != OverloadTy &&
3374  "Unresolved overloaded function type");
3375 
3376  // Unique pointers, to guarantee there is only one pointer of a particular
3377  // structure.
3378  llvm::FoldingSetNodeID ID;
3379  ReferenceType::Profile(ID, T, SpelledAsLValue);
3380 
3381  void *InsertPos = nullptr;
3382  if (LValueReferenceType *RT =
3383  LValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos))
3384  return QualType(RT, 0);
3385 
3386  const auto *InnerRef = T->getAs<ReferenceType>();
3387 
3388  // If the referencee type isn't canonical, this won't be a canonical type
3389  // either, so fill in the canonical type field.
3390  QualType Canonical;
3391  if (!SpelledAsLValue || InnerRef || !T.isCanonical()) {
3392  QualType PointeeType = (InnerRef ? InnerRef->getPointeeType() : T);
3393  Canonical = getLValueReferenceType(getCanonicalType(PointeeType));
3394 
3395  // Get the new insert position for the node we care about.
3396  LValueReferenceType *NewIP =
3397  LValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos);
3398  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3399  }
3400 
3401  auto *New = new (*this, TypeAlignment) LValueReferenceType(T, Canonical,
3402  SpelledAsLValue);
3403  Types.push_back(New);
3404  LValueReferenceTypes.InsertNode(New, InsertPos);
3405 
3406  return QualType(New, 0);
3407 }
3408 
3409 /// getRValueReferenceType - Return the uniqued reference to the type for an
3410 /// rvalue reference to the specified type.
3412  // Unique pointers, to guarantee there is only one pointer of a particular
3413  // structure.
3414  llvm::FoldingSetNodeID ID;
3415  ReferenceType::Profile(ID, T, false);
3416 
3417  void *InsertPos = nullptr;
3418  if (RValueReferenceType *RT =
3419  RValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos))
3420  return QualType(RT, 0);
3421 
3422  const auto *InnerRef = T->getAs<ReferenceType>();
3423 
3424  // If the referencee type isn't canonical, this won't be a canonical type
3425  // either, so fill in the canonical type field.
3426  QualType Canonical;
3427  if (InnerRef || !T.isCanonical()) {
3428  QualType PointeeType = (InnerRef ? InnerRef->getPointeeType() : T);
3429  Canonical = getRValueReferenceType(getCanonicalType(PointeeType));
3430 
3431  // Get the new insert position for the node we care about.
3432  RValueReferenceType *NewIP =
3433  RValueReferenceTypes.FindNodeOrInsertPos(ID, InsertPos);
3434  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3435  }
3436 
3437  auto *New = new (*this, TypeAlignment) RValueReferenceType(T, Canonical);
3438  Types.push_back(New);
3439  RValueReferenceTypes.InsertNode(New, InsertPos);
3440  return QualType(New, 0);
3441 }
3442 
3443 /// getMemberPointerType - Return the uniqued reference to the type for a
3444 /// member pointer to the specified type, in the specified class.
3446  // Unique pointers, to guarantee there is only one pointer of a particular
3447  // structure.
3448  llvm::FoldingSetNodeID ID;
3449  MemberPointerType::Profile(ID, T, Cls);
3450 
3451  void *InsertPos = nullptr;
3452  if (MemberPointerType *PT =
3453  MemberPointerTypes.FindNodeOrInsertPos(ID, InsertPos))
3454  return QualType(PT, 0);
3455 
3456  // If the pointee or class type isn't canonical, this won't be a canonical
3457  // type either, so fill in the canonical type field.
3458  QualType Canonical;
3459  if (!T.isCanonical() || !Cls->isCanonicalUnqualified()) {
3461 
3462  // Get the new insert position for the node we care about.
3463  MemberPointerType *NewIP =
3464  MemberPointerTypes.FindNodeOrInsertPos(ID, InsertPos);
3465  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3466  }
3467  auto *New = new (*this, TypeAlignment) MemberPointerType(T, Cls, Canonical);
3468  Types.push_back(New);
3469  MemberPointerTypes.InsertNode(New, InsertPos);
3470  return QualType(New, 0);
3471 }
3472 
3473 /// getConstantArrayType - Return the unique reference to the type for an
3474 /// array of the specified element type.
3476  const llvm::APInt &ArySizeIn,
3477  const Expr *SizeExpr,
3479  unsigned IndexTypeQuals) const {
3480  assert((EltTy->isDependentType() ||
3481  EltTy->isIncompleteType() || EltTy->isConstantSizeType()) &&
3482  "Constant array of VLAs is illegal!");
3483 
3484  // We only need the size as part of the type if it's instantiation-dependent.
3485  if (SizeExpr && !SizeExpr->isInstantiationDependent())
3486  SizeExpr = nullptr;
3487 
3488  // Convert the array size into a canonical width matching the pointer size for
3489  // the target.
3490  llvm::APInt ArySize(ArySizeIn);
3491  ArySize = ArySize.zextOrTrunc(Target->getMaxPointerWidth());
3492 
3493  llvm::FoldingSetNodeID ID;
3494  ConstantArrayType::Profile(ID, *this, EltTy, ArySize, SizeExpr, ASM,
3495  IndexTypeQuals);
3496 
3497  void *InsertPos = nullptr;
3498  if (ConstantArrayType *ATP =
3499  ConstantArrayTypes.FindNodeOrInsertPos(ID, InsertPos))
3500  return QualType(ATP, 0);
3501 
3502  // If the element type isn't canonical or has qualifiers, or the array bound
3503  // is instantiation-dependent, this won't be a canonical type either, so fill
3504  // in the canonical type field.
3505  QualType Canon;
3506  if (!EltTy.isCanonical() || EltTy.hasLocalQualifiers() || SizeExpr) {
3507  SplitQualType canonSplit = getCanonicalType(EltTy).split();
3508  Canon = getConstantArrayType(QualType(canonSplit.Ty, 0), ArySize, nullptr,
3509  ASM, IndexTypeQuals);
3510  Canon = getQualifiedType(Canon, canonSplit.Quals);
3511 
3512  // Get the new insert position for the node we care about.
3513  ConstantArrayType *NewIP =
3514  ConstantArrayTypes.FindNodeOrInsertPos(ID, InsertPos);
3515  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3516  }
3517 
3518  void *Mem = Allocate(
3519  ConstantArrayType::totalSizeToAlloc<const Expr *>(SizeExpr ? 1 : 0),
3520  TypeAlignment);
3521  auto *New = new (Mem)
3522  ConstantArrayType(EltTy, Canon, ArySize, SizeExpr, ASM, IndexTypeQuals);
3523  ConstantArrayTypes.InsertNode(New, InsertPos);
3524  Types.push_back(New);
3525  return QualType(New, 0);
3526 }
3527 
3528 /// getVariableArrayDecayedType - Turns the given type, which may be
3529 /// variably-modified, into the corresponding type with all the known
3530 /// sizes replaced with [*].
3532  // Vastly most common case.
3533  if (!type->isVariablyModifiedType()) return type;
3534 
3535  QualType result;
3536 
3537  SplitQualType split = type.getSplitDesugaredType();
3538  const Type *ty = split.Ty;
3539  switch (ty->getTypeClass()) {
3540 #define TYPE(Class, Base)
3541 #define ABSTRACT_TYPE(Class, Base)
3542 #define NON_CANONICAL_TYPE(Class, Base) case Type::Class:
3543 #include "clang/AST/TypeNodes.inc"
3544  llvm_unreachable("didn't desugar past all non-canonical types?");
3545 
3546  // These types should never be variably-modified.
3547  case Type::Builtin:
3548  case Type::Complex:
3549  case Type::Vector:
3550  case Type::DependentVector:
3551  case Type::ExtVector:
3552  case Type::DependentSizedExtVector:
3553  case Type::ConstantMatrix:
3554  case Type::DependentSizedMatrix:
3555  case Type::DependentAddressSpace:
3556  case Type::ObjCObject:
3557  case Type::ObjCInterface:
3558  case Type::ObjCObjectPointer:
3559  case Type::Record:
3560  case Type::Enum:
3561  case Type::UnresolvedUsing:
3562  case Type::TypeOfExpr:
3563  case Type::TypeOf:
3564  case Type::Decltype:
3565  case Type::UnaryTransform:
3566  case Type::DependentName:
3567  case Type::InjectedClassName:
3568  case Type::TemplateSpecialization:
3569  case Type::DependentTemplateSpecialization:
3570  case Type::TemplateTypeParm:
3571  case Type::SubstTemplateTypeParmPack:
3572  case Type::Auto:
3573  case Type::DeducedTemplateSpecialization:
3574  case Type::PackExpansion:
3575  case Type::BitInt:
3576  case Type::DependentBitInt:
3577  llvm_unreachable("type should never be variably-modified");
3578 
3579  // These types can be variably-modified but should never need to
3580  // further decay.
3581  case Type::FunctionNoProto:
3582  case Type::FunctionProto:
3583  case Type::BlockPointer:
3584  case Type::MemberPointer:
3585  case Type::Pipe:
3586  return type;
3587 
3588  // These types can be variably-modified. All these modifications
3589  // preserve structure except as noted by comments.
3590  // TODO: if we ever care about optimizing VLAs, there are no-op
3591  // optimizations available here.
3592  case Type::Pointer:
3594  cast<PointerType>(ty)->getPointeeType()));
3595  break;
3596 
3597  case Type::LValueReference: {
3598  const auto *lv = cast<LValueReferenceType>(ty);
3599  result = getLValueReferenceType(
3600  getVariableArrayDecayedType(lv->getPointeeType()),
3601  lv->isSpelledAsLValue());
3602  break;
3603  }
3604 
3605  case Type::RValueReference: {
3606  const auto *lv = cast<RValueReferenceType>(ty);
3607  result = getRValueReferenceType(
3608  getVariableArrayDecayedType(lv->getPointeeType()));
3609  break;
3610  }
3611 
3612  case Type::Atomic: {
3613  const auto *at = cast<AtomicType>(ty);
3614  result = getAtomicType(getVariableArrayDecayedType(at->getValueType()));
3615  break;
3616  }
3617 
3618  case Type::ConstantArray: {
3619  const auto *cat = cast<ConstantArrayType>(ty);
3620  result = getConstantArrayType(
3621  getVariableArrayDecayedType(cat->getElementType()),
3622  cat->getSize(),
3623  cat->getSizeExpr(),
3624  cat->getSizeModifier(),
3625  cat->getIndexTypeCVRQualifiers());
3626  break;
3627  }
3628 
3629  case Type::DependentSizedArray: {
3630  const auto *dat = cast<DependentSizedArrayType>(ty);
3631  result = getDependentSizedArrayType(
3632  getVariableArrayDecayedType(dat->getElementType()),
3633  dat->getSizeExpr(),
3634  dat->getSizeModifier(),
3635  dat->getIndexTypeCVRQualifiers(),
3636  dat->getBracketsRange());
3637  break;
3638  }
3639 
3640  // Turn incomplete types into [*] types.
3641  case Type::IncompleteArray: {
3642  const auto *iat = cast<IncompleteArrayType>(ty);
3643  result = getVariableArrayType(
3644  getVariableArrayDecayedType(iat->getElementType()),
3645  /*size*/ nullptr,
3647  iat->getIndexTypeCVRQualifiers(),
3648  SourceRange());
3649  break;
3650  }
3651 
3652  // Turn VLA types into [*] types.
3653  case Type::VariableArray: {
3654  const auto *vat = cast<VariableArrayType>(ty);
3655  result = getVariableArrayType(
3656  getVariableArrayDecayedType(vat->getElementType()),
3657  /*size*/ nullptr,
3659  vat->getIndexTypeCVRQualifiers(),
3660  vat->getBracketsRange());
3661  break;
3662  }
3663  }
3664 
3665  // Apply the top-level qualifiers from the original.
3666  return getQualifiedType(result, split.Quals);
3667 }
3668 
3669 /// getVariableArrayType - Returns a non-unique reference to the type for a
3670 /// variable array of the specified element type.
3672  Expr *NumElts,
3674  unsigned IndexTypeQuals,
3675  SourceRange Brackets) const {
3676  // Since we don't unique expressions, it isn't possible to unique VLA's
3677  // that have an expression provided for their size.
3678  QualType Canon;
3679 
3680  // Be sure to pull qualifiers off the element type.
3681  if (!EltTy.isCanonical() || EltTy.hasLocalQualifiers()) {
3682  SplitQualType canonSplit = getCanonicalType(EltTy).split();
3683  Canon = getVariableArrayType(QualType(canonSplit.Ty, 0), NumElts, ASM,
3684  IndexTypeQuals, Brackets);
3685  Canon = getQualifiedType(Canon, canonSplit.Quals);
3686  }
3687 
3688  auto *New = new (*this, TypeAlignment)
3689  VariableArrayType(EltTy, Canon, NumElts, ASM, IndexTypeQuals, Brackets);
3690 
3691  VariableArrayTypes.push_back(New);
3692  Types.push_back(New);
3693  return QualType(New, 0);
3694 }
3695 
3696 /// getDependentSizedArrayType - Returns a non-unique reference to
3697 /// the type for a dependently-sized array of the specified element
3698 /// type.
3700  Expr *numElements,
3702  unsigned elementTypeQuals,
3703  SourceRange brackets) const {
3704  assert((!numElements || numElements->isTypeDependent() ||
3705  numElements->isValueDependent()) &&
3706  "Size must be type- or value-dependent!");
3707 
3708  // Dependently-sized array types that do not have a specified number
3709  // of elements will have their sizes deduced from a dependent
3710  // initializer. We do no canonicalization here at all, which is okay
3711  // because they can't be used in most locations.
3712  if (!numElements) {
3713  auto *newType
3714  = new (*this, TypeAlignment)
3715  DependentSizedArrayType(*this, elementType, QualType(),
3716  numElements, ASM, elementTypeQuals,
3717  brackets);
3718  Types.push_back(newType);
3719  return QualType(newType, 0);
3720  }
3721 
3722  // Otherwise, we actually build a new type every time, but we
3723  // also build a canonical type.
3724 
3725  SplitQualType canonElementType = getCanonicalType(elementType).split();
3726 
3727  void *insertPos = nullptr;
3728  llvm::FoldingSetNodeID ID;
3730  QualType(canonElementType.Ty, 0),
3731  ASM, elementTypeQuals, numElements);
3732 
3733  // Look for an existing type with these properties.
3734  DependentSizedArrayType *canonTy =
3735  DependentSizedArrayTypes.FindNodeOrInsertPos(ID, insertPos);
3736 
3737  // If we don't have one, build one.
3738  if (!canonTy) {
3739  canonTy = new (*this, TypeAlignment)
3740  DependentSizedArrayType(*this, QualType(canonElementType.Ty, 0),
3741  QualType(), numElements, ASM, elementTypeQuals,
3742  brackets);
3743  DependentSizedArrayTypes.InsertNode(canonTy, insertPos);
3744  Types.push_back(canonTy);
3745  }
3746 
3747  // Apply qualifiers from the element type to the array.
3748  QualType canon = getQualifiedType(QualType(canonTy,0),
3749  canonElementType.Quals);
3750 
3751  // If we didn't need extra canonicalization for the element type or the size
3752  // expression, then just use that as our result.
3753  if (QualType(canonElementType.Ty, 0) == elementType &&
3754  canonTy->getSizeExpr() == numElements)
3755  return canon;
3756 
3757  // Otherwise, we need to build a type which follows the spelling
3758  // of the element type.
3759  auto *sugaredType
3760  = new (*this, TypeAlignment)
3761  DependentSizedArrayType(*this, elementType, canon, numElements,
3762  ASM, elementTypeQuals, brackets);
3763  Types.push_back(sugaredType);
3764  return QualType(sugaredType, 0);
3765 }
3766 
3769  unsigned elementTypeQuals) const {
3770  llvm::FoldingSetNodeID ID;
3771  IncompleteArrayType::Profile(ID, elementType, ASM, elementTypeQuals);
3772 
3773  void *insertPos = nullptr;
3774  if (IncompleteArrayType *iat =
3775  IncompleteArrayTypes.FindNodeOrInsertPos(ID, insertPos))
3776  return QualType(iat, 0);
3777 
3778  // If the element type isn't canonical, this won't be a canonical type
3779  // either, so fill in the canonical type field. We also have to pull
3780  // qualifiers off the element type.
3781  QualType canon;
3782 
3783  if (!elementType.isCanonical() || elementType.hasLocalQualifiers()) {
3784  SplitQualType canonSplit = getCanonicalType(elementType).split();
3785  canon = getIncompleteArrayType(QualType(canonSplit.Ty, 0),
3786  ASM, elementTypeQuals);
3787  canon = getQualifiedType(canon, canonSplit.Quals);
3788 
3789  // Get the new insert position for the node we care about.
3790  IncompleteArrayType *existing =
3791  IncompleteArrayTypes.FindNodeOrInsertPos(ID, insertPos);
3792  assert(!existing && "Shouldn't be in the map!"); (void) existing;
3793  }
3794 
3795  auto *newType = new (*this, TypeAlignment)
3796  IncompleteArrayType(elementType, canon, ASM, elementTypeQuals);
3797 
3798  IncompleteArrayTypes.InsertNode(newType, insertPos);
3799  Types.push_back(newType);
3800  return QualType(newType, 0);
3801 }
3802 
3805 #define SVE_INT_ELTTY(BITS, ELTS, SIGNED, NUMVECTORS) \
3806  {getIntTypeForBitwidth(BITS, SIGNED), llvm::ElementCount::getScalable(ELTS), \
3807  NUMVECTORS};
3808 
3809 #define SVE_ELTTY(ELTTY, ELTS, NUMVECTORS) \
3810  {ELTTY, llvm::ElementCount::getScalable(ELTS), NUMVECTORS};
3811 
3812  switch (Ty->getKind()) {
3813  default:
3814  llvm_unreachable("Unsupported builtin vector type");
3815  case BuiltinType::SveInt8:
3816  return SVE_INT_ELTTY(8, 16, true, 1);
3817  case BuiltinType::SveUint8:
3818  return SVE_INT_ELTTY(8, 16, false, 1);
3819  case BuiltinType::SveInt8x2:
3820  return SVE_INT_ELTTY(8, 16, true, 2);
3821  case BuiltinType::SveUint8x2:
3822  return SVE_INT_ELTTY(8, 16, false, 2);
3823  case BuiltinType::SveInt8x3:
3824  return SVE_INT_ELTTY(8, 16, true, 3);
3825  case BuiltinType::SveUint8x3:
3826  return SVE_INT_ELTTY(8, 16, false, 3);
3827  case BuiltinType::SveInt8x4:
3828  return SVE_INT_ELTTY(8, 16, true, 4);
3829  case BuiltinType::SveUint8x4:
3830  return SVE_INT_ELTTY(8, 16, false, 4);
3831  case BuiltinType::SveInt16:
3832  return SVE_INT_ELTTY(16, 8, true, 1);
3833  case BuiltinType::SveUint16:
3834  return SVE_INT_ELTTY(16, 8, false, 1);
3835  case BuiltinType::SveInt16x2:
3836  return SVE_INT_ELTTY(16, 8, true, 2);
3837  case BuiltinType::SveUint16x2:
3838  return SVE_INT_ELTTY(16, 8, false, 2);
3839  case BuiltinType::SveInt16x3:
3840  return SVE_INT_ELTTY(16, 8, true, 3);
3841  case BuiltinType::SveUint16x3:
3842  return SVE_INT_ELTTY(16, 8, false, 3);
3843  case BuiltinType::SveInt16x4:
3844  return SVE_INT_ELTTY(16, 8, true, 4);
3845  case BuiltinType::SveUint16x4:
3846  return SVE_INT_ELTTY(16, 8, false, 4);
3847  case BuiltinType::SveInt32:
3848  return SVE_INT_ELTTY(32, 4, true, 1);
3849  case BuiltinType::SveUint32:
3850  return SVE_INT_ELTTY(32, 4, false, 1);
3851  case BuiltinType::SveInt32x2:
3852  return SVE_INT_ELTTY(32, 4, true, 2);
3853  case BuiltinType::SveUint32x2:
3854  return SVE_INT_ELTTY(32, 4, false, 2);
3855  case BuiltinType::SveInt32x3:
3856  return SVE_INT_ELTTY(32, 4, true, 3);
3857  case BuiltinType::SveUint32x3:
3858  return SVE_INT_ELTTY(32, 4, false, 3);
3859  case BuiltinType::SveInt32x4:
3860  return SVE_INT_ELTTY(32, 4, true, 4);
3861  case BuiltinType::SveUint32x4:
3862  return SVE_INT_ELTTY(32, 4, false, 4);
3863  case BuiltinType::SveInt64:
3864  return SVE_INT_ELTTY(64, 2, true, 1);
3865  case BuiltinType::SveUint64:
3866  return SVE_INT_ELTTY(64, 2, false, 1);
3867  case BuiltinType::SveInt64x2:
3868  return SVE_INT_ELTTY(64, 2, true, 2);
3869  case BuiltinType::SveUint64x2:
3870  return SVE_INT_ELTTY(64, 2, false, 2);
3871  case BuiltinType::SveInt64x3:
3872  return SVE_INT_ELTTY(64, 2, true, 3);
3873  case BuiltinType::SveUint64x3:
3874  return SVE_INT_ELTTY(64, 2, false, 3);
3875  case BuiltinType::SveInt64x4:
3876  return SVE_INT_ELTTY(64, 2, true, 4);
3877  case BuiltinType::SveUint64x4:
3878  return SVE_INT_ELTTY(64, 2, false, 4);
3879  case BuiltinType::SveBool:
3880  return SVE_ELTTY(BoolTy, 16, 1);
3881  case BuiltinType::SveFloat16:
3882  return SVE_ELTTY(HalfTy, 8, 1);
3883  case BuiltinType::SveFloat16x2:
3884  return SVE_ELTTY(HalfTy, 8, 2);
3885  case BuiltinType::SveFloat16x3:
3886  return SVE_ELTTY(HalfTy, 8, 3);
3887  case BuiltinType::SveFloat16x4:
3888  return SVE_ELTTY(HalfTy, 8, 4);
3889  case BuiltinType::SveFloat32:
3890  return SVE_ELTTY(FloatTy, 4, 1);
3891  case BuiltinType::SveFloat32x2:
3892  return SVE_ELTTY(FloatTy, 4, 2);
3893  case BuiltinType::SveFloat32x3:
3894  return SVE_ELTTY(FloatTy, 4, 3);
3895  case BuiltinType::SveFloat32x4:
3896  return SVE_ELTTY(FloatTy, 4, 4);
3897  case BuiltinType::SveFloat64:
3898  return SVE_ELTTY(DoubleTy, 2, 1);
3899  case BuiltinType::SveFloat64x2:
3900  return SVE_ELTTY(DoubleTy, 2, 2);
3901  case BuiltinType::SveFloat64x3:
3902  return SVE_ELTTY(DoubleTy, 2, 3);
3903  case BuiltinType::SveFloat64x4:
3904  return SVE_ELTTY(DoubleTy, 2, 4);
3905  case BuiltinType::SveBFloat16:
3906  return SVE_ELTTY(BFloat16Ty, 8, 1);
3907  case BuiltinType::SveBFloat16x2:
3908  return SVE_ELTTY(BFloat16Ty, 8, 2);
3909  case BuiltinType::SveBFloat16x3:
3910  return SVE_ELTTY(BFloat16Ty, 8, 3);
3911  case BuiltinType::SveBFloat16x4:
3912  return SVE_ELTTY(BFloat16Ty, 8, 4);
3913 #define RVV_VECTOR_TYPE_INT(Name, Id, SingletonId, NumEls, ElBits, NF, \
3914  IsSigned) \
3915  case BuiltinType::Id: \
3916  return {getIntTypeForBitwidth(ElBits, IsSigned), \
3917  llvm::ElementCount::getScalable(NumEls), NF};
3918 #define RVV_VECTOR_TYPE_FLOAT(Name, Id, SingletonId, NumEls, ElBits, NF) \
3919  case BuiltinType::Id: \
3920  return {ElBits == 16 ? Float16Ty : (ElBits == 32 ? FloatTy : DoubleTy), \
3921  llvm::ElementCount::getScalable(NumEls), NF};
3922 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, NumEls) \
3923  case BuiltinType::Id: \
3924  return {BoolTy, llvm::ElementCount::getScalable(NumEls), 1};
3925 #include "clang/Basic/RISCVVTypes.def"
3926  }
3927 }
3928 
3929 /// getScalableVectorType - Return the unique reference to a scalable vector
3930 /// type of the specified element type and size. VectorType must be a built-in
3931 /// type.
3933  unsigned NumElts) const {
3934  if (Target->hasAArch64SVETypes()) {
3935  uint64_t EltTySize = getTypeSize(EltTy);
3936 #define SVE_VECTOR_TYPE(Name, MangledName, Id, SingletonId, NumEls, ElBits, \
3937  IsSigned, IsFP, IsBF) \
3938  if (!EltTy->isBooleanType() && \
3939  ((EltTy->hasIntegerRepresentation() && \
3940  EltTy->hasSignedIntegerRepresentation() == IsSigned) || \
3941  (EltTy->hasFloatingRepresentation() && !EltTy->isBFloat16Type() && \
3942  IsFP && !IsBF) || \
3943  (EltTy->hasFloatingRepresentation() && EltTy->isBFloat16Type() && \
3944  IsBF && !IsFP)) && \
3945  EltTySize == ElBits && NumElts == NumEls) { \
3946  return SingletonId; \
3947  }
3948 #define SVE_PREDICATE_TYPE(Name, MangledName, Id, SingletonId, NumEls) \
3949  if (EltTy->isBooleanType() && NumElts == NumEls) \
3950  return SingletonId;
3951 #include "clang/Basic/AArch64SVEACLETypes.def"
3952  } else if (Target->hasRISCVVTypes()) {
3953  uint64_t EltTySize = getTypeSize(EltTy);
3954 #define RVV_VECTOR_TYPE(Name, Id, SingletonId, NumEls, ElBits, NF, IsSigned, \
3955  IsFP) \
3956  if (!EltTy->isBooleanType() && \
3957  ((EltTy->hasIntegerRepresentation() && \
3958  EltTy->hasSignedIntegerRepresentation() == IsSigned) || \
3959  (EltTy->hasFloatingRepresentation() && IsFP)) && \
3960  EltTySize == ElBits && NumElts == NumEls) \
3961  return SingletonId;
3962 #define RVV_PREDICATE_TYPE(Name, Id, SingletonId, NumEls) \
3963  if (EltTy->isBooleanType() && NumElts == NumEls) \
3964  return SingletonId;
3965 #include "clang/Basic/RISCVVTypes.def"
3966  }
3967  return QualType();
3968 }
3969 
3970 /// getVectorType - Return the unique reference to a vector type of
3971 /// the specified element type and size. VectorType must be a built-in type.
3972 QualType ASTContext::getVectorType(QualType vecType, unsigned NumElts,
3973  VectorType::VectorKind VecKind) const {
3974  assert(vecType->isBuiltinType());
3975 
3976  // Check if we've already instantiated a vector of this type.
3977  llvm::FoldingSetNodeID ID;
3978  VectorType::Profile(ID, vecType, NumElts, Type::Vector, VecKind);
3979 
3980  void *InsertPos = nullptr;
3981  if (VectorType *VTP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos))
3982  return QualType(VTP, 0);
3983 
3984  // If the element type isn't canonical, this won't be a canonical type either,
3985  // so fill in the canonical type field.
3986  QualType Canonical;
3987  if (!vecType.isCanonical()) {
3988  Canonical = getVectorType(getCanonicalType(vecType), NumElts, VecKind);
3989 
3990  // Get the new insert position for the node we care about.
3991  VectorType *NewIP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos);
3992  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
3993  }
3994  auto *New = new (*this, TypeAlignment)
3995  VectorType(vecType, NumElts, Canonical, VecKind);
3996  VectorTypes.InsertNode(New, InsertPos);
3997  Types.push_back(New);
3998  return QualType(New, 0);
3999 }
4000 
4001 QualType
4003  SourceLocation AttrLoc,
4004  VectorType::VectorKind VecKind) const {
4005  llvm::FoldingSetNodeID ID;
4006  DependentVectorType::Profile(ID, *this, getCanonicalType(VecType), SizeExpr,
4007  VecKind);
4008  void *InsertPos = nullptr;
4009  DependentVectorType *Canon =
4010  DependentVectorTypes.FindNodeOrInsertPos(ID, InsertPos);
4011  DependentVectorType *New;
4012 
4013  if (Canon) {
4014  New = new (*this, TypeAlignment) DependentVectorType(
4015  *this, VecType, QualType(Canon, 0), SizeExpr, AttrLoc, VecKind);
4016  } else {
4017  QualType CanonVecTy = getCanonicalType(VecType);
4018  if (CanonVecTy == VecType) {
4019  New = new (*this, TypeAlignment) DependentVectorType(
4020  *this, VecType, QualType(), SizeExpr, AttrLoc, VecKind);
4021 
4022  DependentVectorType *CanonCheck =
4023  DependentVectorTypes.FindNodeOrInsertPos(ID, InsertPos);
4024  assert(!CanonCheck &&
4025  "Dependent-sized vector_size canonical type broken");
4026  (void)CanonCheck;
4027  DependentVectorTypes.InsertNode(New, InsertPos);
4028  } else {
4029  QualType CanonTy = getDependentVectorType(CanonVecTy, SizeExpr,
4030  SourceLocation(), VecKind);
4031  New = new (*this, TypeAlignment) DependentVectorType(
4032  *this, VecType, CanonTy, SizeExpr, AttrLoc, VecKind);
4033  }
4034  }
4035 
4036  Types.push_back(New);
4037  return QualType(New, 0);
4038 }
4039 
4040 /// getExtVectorType - Return the unique reference to an extended vector type of
4041 /// the specified element type and size. VectorType must be a built-in type.
4042 QualType
4043 ASTContext::getExtVectorType(QualType vecType, unsigned NumElts) const {
4044  assert(vecType->isBuiltinType() || vecType->isDependentType());
4045 
4046  // Check if we've already instantiated a vector of this type.
4047  llvm::FoldingSetNodeID ID;
4048  VectorType::Profile(ID, vecType, NumElts, Type::ExtVector,
4050  void *InsertPos = nullptr;
4051  if (VectorType *VTP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos))
4052  return QualType(VTP, 0);
4053 
4054  // If the element type isn't canonical, this won't be a canonical type either,
4055  // so fill in the canonical type field.
4056  QualType Canonical;
4057  if (!vecType.isCanonical()) {
4058  Canonical = getExtVectorType(getCanonicalType(vecType), NumElts);
4059 
4060  // Get the new insert position for the node we care about.
4061  VectorType *NewIP = VectorTypes.FindNodeOrInsertPos(ID, InsertPos);
4062  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
4063  }
4064  auto *New = new (*this, TypeAlignment)
4065  ExtVectorType(vecType, NumElts, Canonical);
4066  VectorTypes.InsertNode(New, InsertPos);
4067  Types.push_back(New);
4068  return QualType(New, 0);
4069 }
4070 
4071 QualType
4073  Expr *SizeExpr,
4074  SourceLocation AttrLoc) const {
4075  llvm::FoldingSetNodeID ID;
4077  SizeExpr);
4078 
4079  void *InsertPos = nullptr;
4081  = DependentSizedExtVectorTypes.FindNodeOrInsertPos(ID, InsertPos);
4083  if (Canon) {
4084  // We already have a canonical version of this array type; use it as
4085  // the canonical type for a newly-built type.
4086  New = new (*this, TypeAlignment)
4087  DependentSizedExtVectorType(*this, vecType, QualType(Canon, 0),
4088  SizeExpr, AttrLoc);
4089  } else {
4090  QualType CanonVecTy = getCanonicalType(vecType);
4091  if (CanonVecTy == vecType) {
4092  New = new (*this, TypeAlignment)
4093  DependentSizedExtVectorType(*this, vecType, QualType(), SizeExpr,
4094  AttrLoc);
4095 
4096  DependentSizedExtVectorType *CanonCheck
4097  = DependentSizedExtVectorTypes.FindNodeOrInsertPos(ID, InsertPos);
4098  assert(!CanonCheck && "Dependent-sized ext_vector canonical type broken");
4099  (void)CanonCheck;
4100  DependentSizedExtVectorTypes.InsertNode(New, InsertPos);
4101  } else {
4102  QualType CanonExtTy = getDependentSizedExtVectorType(CanonVecTy, SizeExpr,
4103  SourceLocation());
4104  New = new (*this, TypeAlignment) DependentSizedExtVectorType(
4105  *this, vecType, CanonExtTy, SizeExpr, AttrLoc);
4106  }
4107  }
4108 
4109  Types.push_back(New);
4110  return QualType(New, 0);
4111 }
4112 
4114  unsigned NumColumns) const {
4115  llvm::FoldingSetNodeID ID;
4116  ConstantMatrixType::Profile(ID, ElementTy, NumRows, NumColumns,
4117  Type::ConstantMatrix);
4118 
4119  assert(MatrixType::isValidElementType(ElementTy) &&
4120  "need a valid element type");
4121  assert(ConstantMatrixType::isDimensionValid(NumRows) &&
4123  "need valid matrix dimensions");
4124  void *InsertPos = nullptr;
4125  if (ConstantMatrixType *MTP = MatrixTypes.FindNodeOrInsertPos(ID, InsertPos))
4126  return QualType(MTP, 0);
4127 
4128  QualType Canonical;
4129  if (!ElementTy.isCanonical()) {
4130  Canonical =
4131  getConstantMatrixType(getCanonicalType(ElementTy), NumRows, NumColumns);
4132 
4133  ConstantMatrixType *NewIP = MatrixTypes.FindNodeOrInsertPos(ID, InsertPos);
4134  assert(!NewIP && "Matrix type shouldn't already exist in the map");
4135  (void)NewIP;
4136  }
4137 
4138  auto *New = new (*this, TypeAlignment)
4139  ConstantMatrixType(ElementTy, NumRows, NumColumns, Canonical);
4140  MatrixTypes.InsertNode(New, InsertPos);
4141  Types.push_back(New);
4142  return QualType(New, 0);
4143 }
4144 
4146  Expr *RowExpr,
4147  Expr *ColumnExpr,
4148  SourceLocation AttrLoc) const {
4149  QualType CanonElementTy = getCanonicalType(ElementTy);
4150  llvm::FoldingSetNodeID ID;
4151  DependentSizedMatrixType::Profile(ID, *this, CanonElementTy, RowExpr,
4152  ColumnExpr);
4153 
4154  void *InsertPos = nullptr;
4155  DependentSizedMatrixType *Canon =
4156  DependentSizedMatrixTypes.FindNodeOrInsertPos(ID, InsertPos);
4157 
4158  if (!Canon) {
4159  Canon = new (*this, TypeAlignment) DependentSizedMatrixType(
4160  *this, CanonElementTy, QualType(), RowExpr, ColumnExpr, AttrLoc);
4161 #ifndef NDEBUG
4162  DependentSizedMatrixType *CanonCheck =
4163  DependentSizedMatrixTypes.FindNodeOrInsertPos(ID, InsertPos);
4164  assert(!CanonCheck && "Dependent-sized matrix canonical type broken");
4165 #endif
4166  DependentSizedMatrixTypes.InsertNode(Canon, InsertPos);
4167  Types.push_back(Canon);
4168  }
4169 
4170  // Already have a canonical version of the matrix type
4171  //
4172  // If it exactly matches the requested type, use it directly.
4173  if (Canon->getElementType() == ElementTy && Canon->getRowExpr() == RowExpr &&
4174  Canon->getRowExpr() == ColumnExpr)
4175  return QualType(Canon, 0);
4176 
4177  // Use Canon as the canonical type for newly-built type.
4178  DependentSizedMatrixType *New = new (*this, TypeAlignment)
4179  DependentSizedMatrixType(*this, ElementTy, QualType(Canon, 0), RowExpr,
4180  ColumnExpr, AttrLoc);
4181  Types.push_back(New);
4182  return QualType(New, 0);
4183 }
4184 
4186  Expr *AddrSpaceExpr,
4187  SourceLocation AttrLoc) const {
4188  assert(AddrSpaceExpr->isInstantiationDependent());
4189 
4190  QualType canonPointeeType = getCanonicalType(PointeeType);
4191 
4192  void *insertPos = nullptr;
4193  llvm::FoldingSetNodeID ID;
4194  DependentAddressSpaceType::Profile(ID, *this, canonPointeeType,
4195  AddrSpaceExpr);
4196 
4197  DependentAddressSpaceType *canonTy =
4198  DependentAddressSpaceTypes.FindNodeOrInsertPos(ID, insertPos);
4199 
4200  if (!canonTy) {
4201  canonTy = new (*this, TypeAlignment)
4202  DependentAddressSpaceType(*this, canonPointeeType,
4203  QualType(), AddrSpaceExpr, AttrLoc);
4204  DependentAddressSpaceTypes.InsertNode(canonTy, insertPos);
4205  Types.push_back(canonTy);
4206  }
4207 
4208  if (canonPointeeType == PointeeType &&
4209  canonTy->getAddrSpaceExpr() == AddrSpaceExpr)
4210  return QualType(canonTy, 0);
4211 
4212  auto *sugaredType
4213  = new (*this, TypeAlignment)
4214  DependentAddressSpaceType(*this, PointeeType, QualType(canonTy, 0),
4215  AddrSpaceExpr, AttrLoc);
4216  Types.push_back(sugaredType);
4217  return QualType(sugaredType, 0);
4218 }
4219 
4220 /// Determine whether \p T is canonical as the result type of a function.
4222  return T.isCanonical() &&
4225 }
4226 
4227 /// getFunctionNoProtoType - Return a K&R style C function type like 'int()'.
4228 QualType
4230  const FunctionType::ExtInfo &Info) const {
4231  // Unique functions, to guarantee there is only one function of a particular
4232  // structure.
4233  llvm::FoldingSetNodeID ID;
4234  FunctionNoProtoType::Profile(ID, ResultTy, Info);
4235 
4236  void *InsertPos = nullptr;
4237  if (FunctionNoProtoType *FT =
4238  FunctionNoProtoTypes.FindNodeOrInsertPos(ID, InsertPos))
4239  return QualType(FT, 0);
4240 
4241  QualType Canonical;
4242  if (!isCanonicalResultType(ResultTy)) {
4243  Canonical =
4245 
4246  // Get the new insert position for the node we care about.
4247  FunctionNoProtoType *NewIP =
4248  FunctionNoProtoTypes.FindNodeOrInsertPos(ID, InsertPos);
4249  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
4250  }
4251 
4252  auto *New = new (*this, TypeAlignment)
4253  FunctionNoProtoType(ResultTy, Canonical, Info);
4254  Types.push_back(New);
4255  FunctionNoProtoTypes.InsertNode(New, InsertPos);
4256  return QualType(New, 0);
4257 }
4258 
4261  CanQualType CanResultType = getCanonicalType(ResultType);
4262 
4263  // Canonical result types do not have ARC lifetime qualifiers.
4264  if (CanResultType.getQualifiers().hasObjCLifetime()) {
4265  Qualifiers Qs = CanResultType.getQualifiers();
4266  Qs.removeObjCLifetime();
4268  getQualifiedType(CanResultType.getUnqualifiedType(), Qs));
4269  }
4270 
4271  return CanResultType;
4272 }
4273 
4275  const FunctionProtoType::ExceptionSpecInfo &ESI, bool NoexceptInType) {
4276  if (ESI.Type == EST_None)
4277  return true;
4278  if (!NoexceptInType)
4279  return false;
4280 
4281  // C++17 onwards: exception specification is part of the type, as a simple
4282  // boolean "can this function type throw".
4283  if (ESI.Type == EST_BasicNoexcept)
4284  return true;
4285 
4286  // A noexcept(expr) specification is (possibly) canonical if expr is
4287  // value-dependent.
4288  if (ESI.Type == EST_DependentNoexcept)
4289  return true;
4290 
4291  // A dynamic exception specification is canonical if it only contains pack
4292  // expansions (so we can't tell whether it's non-throwing) and all its
4293  // contained types are canonical.
4294  if (ESI.Type == EST_Dynamic) {
4295  bool AnyPackExpansions = false;
4296  for (QualType ET : ESI.Exceptions) {
4297  if (!ET.isCanonical())
4298  return false;
4299  if (ET->getAs<PackExpansionType>())
4300  AnyPackExpansions = true;
4301  }
4302  return AnyPackExpansions;
4303  }
4304 
4305  return false;
4306 }
4307 
4308 QualType ASTContext::getFunctionTypeInternal(
4309  QualType ResultTy, ArrayRef<QualType> ArgArray,
4310  const FunctionProtoType::ExtProtoInfo &EPI, bool OnlyWantCanonical) const {
4311  size_t NumArgs = ArgArray.size();
4312 
4313  // Unique functions, to guarantee there is only one function of a particular
4314  // structure.
4315  llvm::FoldingSetNodeID ID;
4316  FunctionProtoType::Profile(ID, ResultTy, ArgArray.begin(), NumArgs, EPI,
4317  *this, true);
4318 
4319  QualType Canonical;
4320  bool Unique = false;
4321 
4322  void *InsertPos = nullptr;
4323  if (FunctionProtoType *FPT =
4324  FunctionProtoTypes.FindNodeOrInsertPos(ID, InsertPos)) {
4325  QualType Existing = QualType(FPT, 0);
4326 
4327  // If we find a pre-existing equivalent FunctionProtoType, we can just reuse
4328  // it so long as our exception specification doesn't contain a dependent
4329  // noexcept expression, or we're just looking for a canonical type.
4330  // Otherwise, we're going to need to create a type
4331  // sugar node to hold the concrete expression.
4332  if (OnlyWantCanonical || !isComputedNoexcept(EPI.ExceptionSpec.Type) ||
4333  EPI.ExceptionSpec.NoexceptExpr == FPT->getNoexceptExpr())
4334  return Existing;
4335 
4336  // We need a new type sugar node for this one, to hold the new noexcept
4337  // expression. We do no canonicalization here, but that's OK since we don't
4338  // expect to see the same noexcept expression much more than once.
4339  Canonical = getCanonicalType(Existing);
4340  Unique = true;
4341  }
4342 
4343  bool NoexceptInType = getLangOpts().CPlusPlus17;
4344  bool IsCanonicalExceptionSpec =
4345  isCanonicalExceptionSpecification(EPI.ExceptionSpec, NoexceptInType);
4346 
4347  // Determine whether the type being created is already canonical or not.
4348  bool isCanonical = !Unique && IsCanonicalExceptionSpec &&
4349  isCanonicalResultType(ResultTy) && !EPI.HasTrailingReturn;
4350  for (unsigned i = 0; i != NumArgs && isCanonical; ++i)
4351  if (!ArgArray[i].isCanonicalAsParam())
4352  isCanonical = false;
4353 
4354  if (OnlyWantCanonical)
4355  assert(isCanonical &&
4356  "given non-canonical parameters constructing canonical type");
4357 
4358  // If this type isn't canonical, get the canonical version of it if we don't
4359  // already have it. The exception spec is only partially part of the
4360  // canonical type, and only in C++17 onwards.
4361  if (!isCanonical && Canonical.isNull()) {
4362  SmallVector<QualType, 16> CanonicalArgs;
4363  CanonicalArgs.reserve(NumArgs);
4364  for (unsigned i = 0; i != NumArgs; ++i)
4365  CanonicalArgs.push_back(getCanonicalParamType(ArgArray[i]));
4366 
4367  llvm::SmallVector<QualType, 8> ExceptionTypeStorage;
4368  FunctionProtoType::ExtProtoInfo CanonicalEPI = EPI;
4369  CanonicalEPI.HasTrailingReturn = false;
4370 
4371  if (IsCanonicalExceptionSpec) {
4372  // Exception spec is already OK.
4373  } else if (NoexceptInType) {
4374  switch (EPI.ExceptionSpec.Type) {
4376  // We don't know yet. It shouldn't matter what we pick here; no-one
4377  // should ever look at this.
4378  LLVM_FALLTHROUGH;
4379  case EST_None: case EST_MSAny: case EST_NoexceptFalse:
4380  CanonicalEPI.ExceptionSpec.Type = EST_None;
4381  break;
4382 
4383  // A dynamic exception specification is almost always "not noexcept",
4384  // with the exception that a pack expansion might expand to no types.
4385  case EST_Dynamic: {
4386  bool AnyPacks = false;
4387  for (QualType ET : EPI.ExceptionSpec.Exceptions) {
4388  if (ET->getAs<PackExpansionType>())
4389  AnyPacks = true;
4390  ExceptionTypeStorage.push_back(getCanonicalType(ET));
4391  }
4392  if (!AnyPacks)
4393  CanonicalEPI.ExceptionSpec.Type = EST_None;
4394  else {
4395  CanonicalEPI.ExceptionSpec.Type = EST_Dynamic;
4396  CanonicalEPI.ExceptionSpec.Exceptions = ExceptionTypeStorage;
4397  }
4398  break;
4399  }
4400 
4401  case EST_DynamicNone:
4402  case EST_BasicNoexcept:
4403  case EST_NoexceptTrue:
4404  case EST_NoThrow:
4405  CanonicalEPI.ExceptionSpec.Type = EST_BasicNoexcept;
4406  break;
4407 
4408  case EST_DependentNoexcept:
4409  llvm_unreachable("dependent noexcept is already canonical");
4410  }
4411  } else {
4413  }
4414 
4415  // Adjust the canonical function result type.
4416  CanQualType CanResultTy = getCanonicalFunctionResultType(ResultTy);
4417  Canonical =
4418  getFunctionTypeInternal(CanResultTy, CanonicalArgs, CanonicalEPI, true);
4419 
4420  // Get the new insert position for the node we care about.
4421  FunctionProtoType *NewIP =
4422  FunctionProtoTypes.FindNodeOrInsertPos(ID, InsertPos);
4423  assert(!NewIP && "Shouldn't be in the map!"); (void)NewIP;
4424  }
4425 
4426  // Compute the needed size to hold this FunctionProtoType and the
4427  // various trailing objects.
4428  auto ESH = FunctionProtoType::getExceptionSpecSize(
4429  EPI.ExceptionSpec.Type, EPI.ExceptionSpec.Exceptions.size());
4430  size_t Size = FunctionProtoType::totalSizeToAlloc<
4433  FunctionProtoType::ExtParameterInfo, Qualifiers>(
4434  NumArgs, EPI.Variadic,
4435  FunctionProtoType::hasExtraBitfields(EPI.ExceptionSpec.Type),
4436  ESH.NumExceptionType, ESH.NumExprPtr, ESH.NumFunctionDeclPtr,
4437  EPI.ExtParameterInfos ? NumArgs : 0,
4438  EPI.TypeQuals.hasNonFastQualifiers() ? 1 : 0);
4439 
4440  auto *FTP = (FunctionProtoType *)Allocate(Size, TypeAlignment);
4441  FunctionProtoType::ExtProtoInfo newEPI = EPI;
4442  new (FTP) FunctionProtoType(ResultTy, ArgArray, Canonical, newEPI);
4443  Types.push_back(FTP);
4444  if (!Unique)
4445  FunctionProtoTypes.InsertNode(FTP, InsertPos);
4446  return QualType(FTP, 0);
4447 }
4448 
4449 QualType ASTContext::getPipeType(QualType T, bool ReadOnly) const {
4450  llvm::FoldingSetNodeID ID;
4451  PipeType::Profile(ID, T, ReadOnly);
4452 
4453  void *InsertPos = nullptr;
4454  if (PipeType *PT = PipeTypes.FindNodeOrInsertPos(ID, InsertPos))
4455  return QualType(PT, 0);
4456 
4457  // If the pipe element type isn't canonical, this won't be a canonical type
4458  // either, so fill in the canonical type field.
4459  QualType Canonical;
4460  if (!T.isCanonical()) {
4461  Canonical = getPipeType(getCanonicalType(T), ReadOnly);
4462 
4463  // Get the new insert position for the node we care about.
4464  PipeType *NewIP = PipeTypes.FindNodeOrInsertPos(ID, InsertPos);
4465  assert(!NewIP && "Shouldn't be in the map!");
4466  (void)NewIP;
4467  }
4468  auto *New = new (*this, TypeAlignment) PipeType(T, Canonical, ReadOnly);
4469  Types.push_back(New);
4470  PipeTypes.InsertNode(New, InsertPos);
4471  return QualType(New, 0);
4472 }
4473 
4475  // OpenCL v1.1 s6.5.3: a string literal is in the constant address space.
4476  return LangOpts.OpenCL ? getAddrSpaceQualType(Ty, LangAS::opencl_constant)
4477  : Ty;
4478 }
4479 
4481  return getPipeType(T, true);
4482 }
4483 
4485  return getPipeType(T, false);
4486 }
4487 
4488 QualType ASTContext::getBitIntType(bool IsUnsigned, unsigned NumBits) const {
4489  llvm::FoldingSetNodeID ID;
4490  BitIntType::Profile(ID, IsUnsigned, NumBits);
4491 
4492  void *InsertPos = nullptr;
4493  if (BitIntType *EIT = BitIntTypes.FindNodeOrInsertPos(ID, InsertPos))
4494  return QualType(EIT, 0);
4495 
4496  auto *New = new (*this, TypeAlignment) BitIntType(IsUnsigned, NumBits);
4497  BitIntTypes.InsertNode(New, InsertPos);
4498  Types.push_back(New);
4499  return QualType(New, 0);
4500 }
4501 
4503  Expr *NumBitsExpr) const {
4504  assert(NumBitsExpr->isInstantiationDependent() && "Only good for dependent");
4505  llvm::FoldingSetNodeID ID;
4506  DependentBitIntType::Profile(ID, *this, IsUnsigned, NumBitsExpr);
4507 
4508  void *InsertPos = nullptr;
4509  if (DependentBitIntType *Existing =
4510  DependentBitIntTypes.FindNodeOrInsertPos(ID, InsertPos))
4511  return QualType(Existing, 0);
4512 
4513  auto *New = new (*this, TypeAlignment)
4514  DependentBitIntType(*this, IsUnsigned, NumBitsExpr);
4515  DependentBitIntTypes.InsertNode(New, InsertPos);
4516 
4517  Types.push_back(New);
4518  return QualType(New, 0);
4519 }
4520 
4521 #ifndef NDEBUG
4523  if (!isa<CXXRecordDecl>(D)) return false;
4524  const auto *RD = cast<CXXRecordDecl>(D);
4525  if (isa<ClassTemplatePartialSpecializationDecl>(RD))
4526  return true;
4527  if (RD->getDescribedClassTemplate() &&
4528  !isa<ClassTemplateSpecializationDecl>(RD))
4529  return true;
4530  return false;
4531 }
4532 #endif
4533 
4534 /// getInjectedClassNameType - Return the unique reference to the
4535 /// injected class name type for the specified templated declaration.
4537  QualType TST) const {
4539  if (Decl->TypeForDecl) {
4540  assert(isa<InjectedClassNameType>(Decl->TypeForDecl));
4541  } else if (CXXRecordDecl *PrevDecl = Decl->getPreviousDecl()) {
4542  assert(PrevDecl->TypeForDecl && "previous declaration has no type");
4543  Decl->TypeForDecl = PrevDecl->TypeForDecl;
4544  assert(isa<InjectedClassNameType>(Decl->TypeForDecl));
4545  } else {
4546  Type *newType =
4547  new (*this, TypeAlignment) InjectedClassNameType(Decl, TST);
4548  Decl->TypeForDecl = newType;
4549  Types.push_back(newType);
4550  }
4551  return QualType(Decl->TypeForDecl, 0);
4552 }
4553 
4554 /// getTypeDeclType - Return the unique reference to the type for the
4555 /// specified type declaration.
4556 QualType ASTContext::getTypeDeclTypeSlow(const TypeDecl *Decl) const {
4557  assert(Decl && "Passed null for Decl param");
4558  assert(!Decl->TypeForDecl && "TypeForDecl present in slow case");
4559 
4560  if (const auto *Typedef = dyn_cast<TypedefNameDecl>(Decl))
4561  return getTypedefType(Typedef);
4562 
4563  assert(!isa<TemplateTypeParmDecl>(Decl) &&
4564  "Template type parameter types are always available.");
4565 
4566  if (const auto *Record = dyn_cast<RecordDecl>(Decl)) {
4567  assert(Record->isFirstDecl() && "struct/union has previous declaration");
4568  assert(!NeedsInjectedClassNameType(Record));
4569  return getRecordType(Record);
4570  } else if (const auto *Enum = dyn_cast<EnumDecl>(Decl)) {
4571  assert(Enum->isFirstDecl() && "enum has previous declaration");
4572  return getEnumType(Enum);
4573  } else if (const auto *Using = dyn_cast<UnresolvedUsingTypenameDecl>(Decl)) {
4574  return getUnresolvedUsingType(Using);
4575  } else
4576  llvm_unreachable("TypeDecl without a type?");
4577 
4578  return QualType(Decl->TypeForDecl, 0);
4579 }
4580 
4581 /// getTypedefType - Return the unique reference to the type for the
4582 /// specified typedef name decl.
4584  QualType Underlying) const {
4585  if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0);
4586 
4587  if (Underlying.isNull())
4588  Underlying = Decl->getUnderlyingType();
4589  QualType Canonical = getCanonicalType(Underlying);
4590  auto *newType = new (*this, TypeAlignment)
4591  TypedefType(Type::Typedef, Decl, Underlying, Canonical);
4592  Decl->TypeForDecl = newType;
4593  Types.push_back(newType);
4594  return QualType(newType, 0);
4595 }
4596 
4598  QualType Underlying) const {
4599  llvm::FoldingSetNodeID ID;
4600  UsingType::Profile(ID, Found);
4601 
4602  void *InsertPos = nullptr;
4603  UsingType *T = UsingTypes.FindNodeOrInsertPos(ID, InsertPos);
4604  if (T)
4605  return QualType(T, 0);
4606 
4607  assert(!Underlying.hasLocalQualifiers());
4608  assert(Underlying == getTypeDeclType(cast<TypeDecl>(Found->getTargetDecl())));
4609  QualType Canon = Underlying.getCanonicalType();
4610 
4611  UsingType *NewType =
4612  new (*this, TypeAlignment) UsingType(Found, Underlying, Canon);
4613  Types.push_back(NewType);
4614  UsingTypes.InsertNode(NewType, InsertPos);
4615  return QualType(NewType, 0);
4616 }
4617 
4619  if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0);
4620 
4621  if (const RecordDecl *PrevDecl = Decl->getPreviousDecl())
4622  if (PrevDecl->TypeForDecl)
4623  return QualType(Decl->TypeForDecl = PrevDecl->TypeForDecl, 0);
4624 
4625  auto *newType = new (*this, TypeAlignment) RecordType(Decl);
4626  Decl->TypeForDecl = newType;
4627  Types.push_back(newType);
4628  return QualType(newType, 0);
4629 }
4630 
4632  if (Decl->TypeForDecl) return QualType(Decl->TypeForDecl, 0);
4633 
4634  if (const EnumDecl *PrevDecl = Decl->getPreviousDecl())
4635  if (PrevDecl->TypeForDecl)
4636  return QualType(Decl->TypeForDecl = PrevDecl->TypeForDecl, 0);
4637 
4638  auto *newType = new (*this, TypeAlignment) EnumType(Decl);
4639  Decl->TypeForDecl = newType;
4640  Types.push_back(newType);
4641  return QualType(newType, 0);
4642 }
4643 
4645  const UnresolvedUsingTypenameDecl *Decl) const {
4646  if (Decl->TypeForDecl)
4647  return QualType(Decl->TypeForDecl, 0);
4648 
4649  if (const UnresolvedUsingTypenameDecl *CanonicalDecl =
4651  if (CanonicalDecl->TypeForDecl)
4652  return QualType(Decl->TypeForDecl = CanonicalDecl->TypeForDecl, 0);
4653 
4654  Type *newType = new (*this, TypeAlignment) UnresolvedUsingType(Decl);
4655  Decl->TypeForDecl = newType;
4656  Types.push_back(newType);
4657  return QualType(newType, 0);
4658 }
4659 
4661  QualType modifiedType,
4662  QualType equivalentType) {
4663  llvm::FoldingSetNodeID id;
4664  AttributedType::Profile(id, attrKind, modifiedType, equivalentType);
4665 
4666  void *insertPos = nullptr;
4667  AttributedType *type = AttributedTypes.FindNodeOrInsertPos(id, insertPos);
4668  if (type) return QualType(type, 0);
4669 
4670  QualType canon = getCanonicalType(equivalentType);
4671  type = new (*this, TypeAlignment)
4672  AttributedType(canon, attrKind, modifiedType, equivalentType);
4673 
4674  Types.push_back(type);
4675  AttributedTypes.InsertNode(type, insertPos);
4676 
4677  return QualType(type, 0);
4678 }
4679 
4680 /// Retrieve a substitution-result type.
4681 QualType
4683  QualType Replacement) const {
4684  assert(Replacement.isCanonical()
4685  && "replacement types must always be canonical");
4686 
4687  llvm::FoldingSetNodeID ID;
4688  SubstTemplateTypeParmType::Profile(ID, Parm, Replacement);
4689  void *InsertPos = nullptr;
4690  SubstTemplateTypeParmType *SubstParm
4691  = SubstTemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos);
4692 
4693  if (!SubstParm) {
4694  SubstParm = new (*this, TypeAlignment)
4695  SubstTemplateTypeParmType(Parm, Replacement);
4696  Types.push_back(SubstParm);
4697  SubstTemplateTypeParmTypes.InsertNode(SubstParm, InsertPos);
4698  }
4699 
4700  return QualType(SubstParm, 0);
4701 }
4702 
4703 /// Retrieve a
4705  const TemplateTypeParmType *Parm,
4706  const TemplateArgument &ArgPack) {
4707 #ifndef NDEBUG
4708  for (const auto &P : ArgPack.pack_elements()) {
4709  assert(P.getKind() == TemplateArgument::Type &&"Pack contains a non-type");
4710  assert(P.getAsType().isCanonical() && "Pack contains non-canonical type");
4711  }
4712 #endif
4713 
4714  llvm::FoldingSetNodeID ID;
4716  void *InsertPos = nullptr;
4717  if (SubstTemplateTypeParmPackType *SubstParm
4718  = SubstTemplateTypeParmPackTypes.FindNodeOrInsertPos(ID, InsertPos))
4719  return QualType(SubstParm, 0);
4720 
4721  QualType Canon;
4722  if (!Parm->isCanonicalUnqualified()) {
4723  Canon = getCanonicalType(QualType(Parm, 0));
4724  Canon = getSubstTemplateTypeParmPackType(cast<TemplateTypeParmType>(Canon),
4725  ArgPack);
4726  SubstTemplateTypeParmPackTypes.FindNodeOrInsertPos(ID, InsertPos);
4727  }
4728 
4729  auto *SubstParm
4730  = new (*this, TypeAlignment) SubstTemplateTypeParmPackType(Parm, Canon,
4731  ArgPack);
4732  Types.push_back(SubstParm);
4733  SubstTemplateTypeParmPackTypes.InsertNode(SubstParm, InsertPos);
4734  return QualType(SubstParm, 0);
4735 }
4736 
4737 /// Retrieve the template type parameter type for a template
4738 /// parameter or parameter pack with the given depth, index, and (optionally)
4739 /// name.
4741  bool ParameterPack,
4742  TemplateTypeParmDecl *TTPDecl) const {
4743  llvm::FoldingSetNodeID ID;
4744  TemplateTypeParmType::Profile(ID, Depth, Index, ParameterPack, TTPDecl);
4745  void *InsertPos = nullptr;
4746  TemplateTypeParmType *TypeParm
4747  = TemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos);
4748 
4749  if (TypeParm)
4750  return QualType(TypeParm, 0);
4751 
4752  if (TTPDecl) {
4753  QualType Canon = getTemplateTypeParmType(Depth, Index, ParameterPack);
4754  TypeParm = new (*this, TypeAlignment) TemplateTypeParmType(TTPDecl, Canon);
4755 
4756  TemplateTypeParmType *TypeCheck
4757  = TemplateTypeParmTypes.FindNodeOrInsertPos(ID, InsertPos);
4758  assert(!TypeCheck && "Template type parameter canonical type broken");
4759  (void)TypeCheck;
4760  } else
4761  TypeParm = new (*this, TypeAlignment)
4762  TemplateTypeParmType(Depth, Index, ParameterPack);
4763 
4764  Types.push_back(TypeParm);
4765  TemplateTypeParmTypes.InsertNode(TypeParm, InsertPos);
4766 
4767  return QualType(TypeParm, 0);
4768 }
4769 
4772  SourceLocation NameLoc,
4773  const TemplateArgumentListInfo &Args,
4774  QualType Underlying) const {
4775  assert(!Name.getAsDependentTemplateName() &&
4776  "No dependent template names here!");
4777  QualType TST = getTemplateSpecializationType(Name, Args, Underlying);
4778 
4783  TL.setTemplateNameLoc(NameLoc);
4784  TL.setLAngleLoc(Args.getLAngleLoc());
4785  TL.setRAngleLoc(Args.getRAngleLoc());
4786  for (unsigned i = 0, e = TL.getNumArgs(); i != e; ++i)
4787  TL.setArgLocInfo(i, Args[i].getLocInfo());
4788  return DI;
4789 }
4790 
4791 QualType
4793  const TemplateArgumentListInfo &Args,
4794  QualType Underlying) const {
4795  assert(!Template.getAsDependentTemplateName() &&
4796  "No dependent template names here!");
4797 
4799  ArgVec.reserve(Args.size());
4800  for (const TemplateArgumentLoc &Arg : Args.arguments())
4801  ArgVec.push_back(Arg.getArgument());
4802 
4803  return getTemplateSpecializationType(Template, ArgVec, Underlying);
4804 }
4805 
4806 #ifndef NDEBUG
4808  for (const TemplateArgument &Arg : Args)
4809  if (Arg.isPackExpansion())
4810  return true;
4811 
4812  return true;
4813 }
4814 #endif
4815 
4816 QualType
4819  QualType Underlying) const {
4820  assert(!Template.getAsDependentTemplateName() &&
4821  "No dependent template names here!");
4822  // Look through qualified template names.
4823  if (QualifiedTemplateName *QTN = Template.getAsQualifiedTemplateName())
4824  Template = TemplateName(QTN->getTemplateDecl());
4825 
4826  bool IsTypeAlias =
4827  Template.getAsTemplateDecl() &&
4828  isa<TypeAliasTemplateDecl>(Template.getAsTemplateDecl());
4829  QualType CanonType;
4830  if (!Underlying.isNull())
4831  CanonType = getCanonicalType(Underlying);
4832  else {
4833  // We can get here with an alias template when the specialization contains
4834  // a pack expansion that does not match up with a parameter pack.
4835  assert((!IsTypeAlias || hasAnyPackExpansions(Args)) &&
4836  "Caller must compute aliased type");
4837  IsTypeAlias = false;
4838  CanonType = getCanonicalTemplateSpecializationType(Template, Args);
4839  }
4840 
4841  // Allocate the (non-canonical) template specialization type, but don't
4842  // try to unique it: these types typically have location information that
4843  // we don't unique and don't want to lose.
4844  void *Mem = Allocate(sizeof(TemplateSpecializationType) +
4845  sizeof(TemplateArgument) * Args.size() +
4846  (IsTypeAlias? sizeof(QualType) : 0),
4847  TypeAlignment);
4848  auto *Spec
4849  = new (Mem) TemplateSpecializationType(Template, Args, CanonType,
4850  IsTypeAlias ? Underlying : QualType());
4851 
4852  Types.push_back(Spec);
4853  return QualType(Spec, 0);
4854 }
4855 
4856 static bool
4858  ArrayRef<TemplateArgument> OrigArgs,
4859  SmallVectorImpl<TemplateArgument> &CanonArgs) {
4860  bool AnyNonCanonArgs = false;
4861  unsigned NumArgs = OrigArgs.size();
4862  CanonArgs.resize(NumArgs);
4863  for (unsigned I = 0; I != NumArgs; ++I) {
4864  const TemplateArgument &OrigArg = OrigArgs[I];
4865  TemplateArgument &CanonArg = CanonArgs[I];
4866  CanonArg = C.getCanonicalTemplateArgument(OrigArg);
4867  if (!CanonArg.structurallyEquals(OrigArg))
4868  AnyNonCanonArgs = true;
4869  }
4870  return AnyNonCanonArgs;
4871 }
4872 
4874  TemplateName Template, ArrayRef<TemplateArgument> Args) const {
4875  assert(!Template.getAsDependentTemplateName() &&
4876  "No dependent template names here!");
4877 
4878  // Look through qualified template names.
4879  if (QualifiedTemplateName *QTN = Template.getAsQualifiedTemplateName())
4880  Template = TemplateName(QTN->getTemplateDecl());
4881 
4882  // Build the canonical template specialization type.
4883  TemplateName CanonTemplate = getCanonicalTemplateName(Template);
4885  ::getCanonicalTemplateArguments(*this, Args, CanonArgs);
4886 
4887  // Determine whether this canonical template specialization type already
4888  // exists.
4889  llvm::FoldingSetNodeID ID;
4890  TemplateSpecializationType::Profile(ID, CanonTemplate,
4891  CanonArgs, *this);
4892 
4893  void *InsertPos = nullptr;
4895  = TemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos);
4896 
4897  if (!Spec) {
4898  // Allocate a new canonical template specialization type.
4899  void *Mem = Allocate((sizeof(TemplateSpecializationType) +
4900  sizeof(TemplateArgument) * CanonArgs.size()),
4901  TypeAlignment);
4902  Spec = new (Mem) TemplateSpecializationType(CanonTemplate,
4903  CanonArgs,
4904  QualType(), QualType());
4905  Types.push_back(Spec);
4906  TemplateSpecializationTypes.InsertNode(Spec, InsertPos);
4907  }
4908 
4909  assert(Spec->isDependentType() &&
4910  "Non-dependent template-id type must have a canonical type");
4911  return QualType(Spec, 0);
4912 }
4913 
4915  NestedNameSpecifier *NNS,
4916  QualType NamedType,
4917  TagDecl *OwnedTagDecl) const {
4918  llvm::FoldingSetNodeID ID;
4919  ElaboratedType::Profile(ID, Keyword, NNS, NamedType, OwnedTagDecl);
4920 
4921  void *InsertPos = nullptr;
4922  ElaboratedType *T = ElaboratedTypes.FindNodeOrInsertPos(ID, InsertPos);
4923  if (T)
4924  return QualType(T, 0);
4925 
4926  QualType Canon = NamedType;
4927  if (!Canon.isCanonical()) {
4928  Canon = getCanonicalType(NamedType);
4929  ElaboratedType *CheckT = ElaboratedTypes.FindNodeOrInsertPos(ID, InsertPos);
4930  assert(!CheckT && "Elaborated canonical type broken");
4931  (void)CheckT;
4932  }
4933 
4934  void *Mem = Allocate(ElaboratedType::totalSizeToAlloc<TagDecl *>(!!OwnedTagDecl),
4935  TypeAlignment);
4936  T = new (Mem) ElaboratedType(Keyword, NNS, NamedType, Canon, OwnedTagDecl);
4937 
4938  Types.push_back(T);
4939  ElaboratedTypes.InsertNode(T, InsertPos);
4940  return QualType(T, 0);
4941 }
4942 
4943 QualType
4945  llvm::FoldingSetNodeID ID;
4946  ParenType::Profile(ID, InnerType);
4947 
4948  void *InsertPos = nullptr;
4949  ParenType *T = ParenTypes.FindNodeOrInsertPos(ID, InsertPos);
4950  if (T)
4951  return QualType(T, 0);
4952 
4953  QualType Canon = InnerType;
4954  if (!Canon.isCanonical()) {
4955  Canon = getCanonicalType(InnerType);
4956  ParenType *CheckT = ParenTypes.FindNodeOrInsertPos(ID, InsertPos);
4957  assert(!CheckT && "Paren canonical type broken");
4958  (void)CheckT;
4959  }
4960 
4961  T = new (*this, TypeAlignment) ParenType(InnerType, Canon);
4962  Types.push_back(T);
4963  ParenTypes.InsertNode(T, InsertPos);
4964  return QualType(T, 0);
4965 }
4966 
4967 QualType
4969  const IdentifierInfo *MacroII) const {
4970  QualType Canon = UnderlyingTy;
4971  if (!Canon.isCanonical())
4972  Canon = getCanonicalType(UnderlyingTy);
4973 
4974  auto *newType = new (*this, TypeAlignment)
4975  MacroQualifiedType(UnderlyingTy, Canon, MacroII);
4976  Types.push_back(newType);
4977  return QualType(newType, 0);
4978 }
4979 
4981  NestedNameSpecifier *NNS,
4982  const IdentifierInfo *Name,
4983  QualType Canon) const {
4984  if (Canon.isNull()) {
4986  if (CanonNNS != NNS)
4987  Canon = getDependentNameType(Keyword, CanonNNS, Name);
4988  }
4989 
4990  llvm::FoldingSetNodeID ID;
4991  DependentNameType::Profile(ID, Keyword, NNS, Name);
4992 
4993  void *InsertPos = nullptr;
4995  = DependentNameTypes.FindNodeOrInsertPos(ID, InsertPos);
4996  if (T)
4997  return QualType(T, 0);
4998 
4999  T = new (*this, TypeAlignment) DependentNameType(Keyword, NNS, Name, Canon);
5000  Types.push_back(T);
5001  DependentNameTypes.InsertNode(T, InsertPos);
5002  return QualType(T, 0);
5003 }
5004 
5005 QualType
5007  ElaboratedTypeKeyword Keyword,
5008  NestedNameSpecifier *NNS,
5009  const IdentifierInfo *Name,
5010  const TemplateArgumentListInfo &Args) const {
5011  // TODO: avoid this copy
5013  for (unsigned I = 0, E = Args.size(); I != E; ++I)
5014  ArgCopy.push_back(Args[I].getArgument());
5015  return getDependentTemplateSpecializationType(Keyword, NNS, Name, ArgCopy);
5016 }
5017 
5018 QualType
5020  ElaboratedTypeKeyword Keyword,
5021  NestedNameSpecifier *NNS,
5022  const IdentifierInfo *Name,
5023  ArrayRef<TemplateArgument> Args) const {
5024  assert((!NNS || NNS->isDependent()) &&
5025  "nested-name-specifier must be dependent");
5026 
5027  llvm::FoldingSetNodeID ID;
5028  DependentTemplateSpecializationType::Profile(ID, *this, Keyword, NNS,
5029  Name, Args);
5030 
5031  void *InsertPos = nullptr;
5033  = DependentTemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos);
5034  if (T)
5035  return QualType(T, 0);
5036 
5038 
5039  ElaboratedTypeKeyword CanonKeyword = Keyword;
5040  if (Keyword == ETK_None) CanonKeyword = ETK_Typename;
5041 
5043  bool AnyNonCanonArgs =
5044  ::getCanonicalTemplateArguments(*this, Args, CanonArgs);
5045 
5046  QualType Canon;
5047  if (AnyNonCanonArgs || CanonNNS != NNS || CanonKeyword != Keyword) {
5048  Canon = getDependentTemplateSpecializationType(CanonKeyword, CanonNNS,
5049  Name,
5050  CanonArgs);
5051 
5052  // Find the insert position again.
5053  DependentTemplateSpecializationTypes.FindNodeOrInsertPos(ID, InsertPos);
5054  }
5055 
5056  void *Mem = Allocate((sizeof(DependentTemplateSpecializationType) +
5057  sizeof(TemplateArgument) * Args.size()),
5058  TypeAlignment);
5059  T = new (Mem) DependentTemplateSpecializationType(Keyword, NNS,
5060  Name, Args, Canon);
5061  Types.push_back(T);
5062  DependentTemplateSpecializationTypes.InsertNode(T, InsertPos);
5063  return QualType(T, 0);
5064 }
5065 
5067  TemplateArgument Arg;
5068  if (const auto *TTP = dyn_cast<TemplateTypeParmDecl>(Param)) {
5069  QualType ArgType = getTypeDeclType(TTP);
5070  if (TTP->isParameterPack())
5071  ArgType = getPackExpansionType(ArgType, None);
5072 
5073  Arg = TemplateArgument(ArgType);
5074  } else if (auto *NTTP = dyn_cast<NonTypeTemplateParmDecl>(Param)) {
5075  QualType T =
5076  NTTP->getType().getNonPackExpansionType().getNonLValueExprType(*this);
5077  // For class NTTPs, ensure we include the 'const' so the type matches that
5078  // of a real template argument.
5079  // FIXME: It would be more faithful to model this as something like an
5080  // lvalue-to-rvalue conversion applied to a const-qualified lvalue.
5081  if (T->isRecordType())
5082  T.addConst();
5083  Expr *E = new (*this) DeclRefExpr(
5084  *this, NTTP, /*enclosing*/ false, T,
5085  Expr::getValueKindForType(NTTP->getType()), NTTP->getLocation());
5086 
5087  if (NTTP->isParameterPack())
5088  E = new (*this) PackExpansionExpr(DependentTy, E, NTTP->getLocation(),
5089  None);
5090  Arg = TemplateArgument(E);
5091  } else {
5092  auto *TTP = cast<TemplateTemplateParmDecl>(Param);
5093  if (TTP->isParameterPack())
5095  else
5096  Arg = TemplateArgument(TemplateName(TTP));
5097  }
5098 
5099  if (Param->isTemplateParameterPack())
5100  Arg = TemplateArgument::CreatePackCopy(*this, Arg);
5101 
5102  return Arg;
5103 }
5104 
5105 void
5108  Args.reserve(Args.size() + Params->size());
5109 
5110  for (NamedDecl *Param : *Params)
5111  Args.push_back(getInjectedTemplateArg(Param));
5112 }
5113 
5115  Optional<unsigned> NumExpansions,
5116  bool ExpectPackInType) {
5117  assert((!ExpectPackInType || Pattern->containsUnexpandedParameterPack()) &&
5118  "Pack expansions must expand one or more parameter packs");
5119 
5120  llvm::FoldingSetNodeID ID;
5121  PackExpansionType::Profile(ID, Pattern, NumExpansions);
5122 
5123  void *InsertPos = nullptr;
5124  PackExpansionType *T = PackExpansionTypes.FindNodeOrInsertPos(ID, InsertPos);
5125  if (T)
5126  return QualType(T, 0);
5127 
5128  QualType Canon;
5129  if (!Pattern.isCanonical()) {
5130  Canon = getPackExpansionType(getCanonicalType(Pattern), NumExpansions,
5131  /*ExpectPackInType=*/false);
5132 
5133  // Find the insert position again, in case we inserted an element into
5134  // PackExpansionTypes and invalidated our insert position.
5135  PackExpansionTypes.FindNodeOrInsertPos(ID, InsertPos);
5136  }
5137 
5138  T = new (*this, TypeAlignment)
5139  PackExpansionType(Pattern, Canon, NumExpansions);
5140  Types.push_back(T);
5141  PackExpansionTypes.InsertNode(T, InsertPos);
5142  return QualType(T, 0);
5143 }
5144 
5145 /// CmpProtocolNames - Comparison predicate for sorting protocols
5146 /// alphabetically.
5147 static int CmpProtocolNames(ObjCProtocolDecl *const *LHS,
5148  ObjCProtocolDecl *const *RHS) {
5149  return DeclarationName::compare((*LHS)->getDeclName(), (*RHS)->getDeclName());
5150 }
5151 
5153  if (Protocols.empty()) return true;
5154 
5155  if (Protocols[0]->getCanonicalDecl() != Protocols[0])
5156  return false;
5157 
5158  for (unsigned i = 1; i != Protocols.size(); ++i)
5159  if (CmpProtocolNames(&Protocols[i - 1], &Protocols[i]) >= 0 ||
5160  Protocols[i]->getCanonicalDecl() != Protocols[i])
5161  return false;
5162  return true;
5163 }
5164 
5165 static void
5167  // Sort protocols, keyed by name.
5168  llvm::array_pod_sort(Protocols.begin(), Protocols.end(), CmpProtocolNames);
5169 
5170  // Canonicalize.
5171  for (ObjCProtocolDecl *&P : Protocols)
5172  P = P->getCanonicalDecl();
5173 
5174  // Remove duplicates.
5175  auto ProtocolsEnd = std::unique(Protocols.begin(), Protocols.end());
5176  Protocols.erase(ProtocolsEnd, Protocols.end());
5177 }
5178 
5180  ObjCProtocolDecl * const *Protocols,
5181  unsigned NumProtocols) const {
5182  return getObjCObjectType(BaseType, {},
5183  llvm::makeArrayRef(Protocols, NumProtocols),
5184  /*isKindOf=*/false);
5185 }
5186 
5188  QualType baseType,
5189  ArrayRef<QualType> typeArgs,
5190  ArrayRef<ObjCProtocolDecl *> protocols,
5191  bool isKindOf) const {
5192  // If the base type is an interface and there aren't any protocols or
5193  // type arguments to add, then the interface type will do just fine.
5194  if (typeArgs.empty() && protocols.empty() && !isKindOf &&
5195  isa<ObjCInterfaceType>(baseType))
5196  return baseType;
5197 
5198  // Look in the folding set for an existing type.
5199  llvm::FoldingSetNodeID ID;
5200  ObjCObjectTypeImpl::Profile(ID, baseType, typeArgs, protocols, isKindOf);
5201  void *InsertPos = nullptr;
5202  if (ObjCObjectType *QT = ObjCObjectTypes.FindNodeOrInsertPos(ID, InsertPos))
5203  return QualType(QT, 0);
5204 
5205  // Determine the type arguments to be used for canonicalization,
5206  // which may be explicitly specified here or written on the base
5207  // type.
5208  ArrayRef<QualType> effectiveTypeArgs = typeArgs;
5209  if (effectiveTypeArgs.empty()) {
5210  if (const auto *baseObject = baseType->getAs<ObjCObjectType>())
5211  effectiveTypeArgs = baseObject->getTypeArgs();
5212  }
5213 
5214  // Build the canonical type, which has the canonical base type and a
5215  // sorted-and-uniqued list of protocols and the type arguments
5216  // canonicalized.
5217  QualType canonical;
5218  bool typeArgsAreCanonical = llvm::all_of(
5219  effectiveTypeArgs, [&](QualType type) { return type.isCanonical(); });
5220  bool protocolsSorted = areSortedAndUniqued(protocols);
5221  if (!typeArgsAreCanonical || !protocolsSorted || !baseType.isCanonical()) {
5222  // Determine the canonical type arguments.
5223  ArrayRef<QualType> canonTypeArgs;
5224  SmallVector<QualType, 4> canonTypeArgsVec;
5225  if (!typeArgsAreCanonical) {
5226  canonTypeArgsVec.reserve(effectiveTypeArgs.size());
5227  for (auto typeArg : effectiveTypeArgs)
5228  canonTypeArgsVec.push_back(getCanonicalType(typeArg));
5229  canonTypeArgs = canonTypeArgsVec;
5230  } else {
5231  canonTypeArgs = effectiveTypeArgs;
5232  }
5233 
5234  ArrayRef<ObjCProtocolDecl *> canonProtocols;
5235  SmallVector<ObjCProtocolDecl*, 8> canonProtocolsVec;
5236  if (!protocolsSorted) {
5237  canonProtocolsVec.append(protocols.begin(), protocols.end());
5238  SortAndUniqueProtocols(canonProtocolsVec);
5239  canonProtocols = canonProtocolsVec;
5240  } else {
5241  canonProtocols = protocols;
5242  }
5243 
5244  canonical = getObjCObjectType(getCanonicalType(baseType), canonTypeArgs,
5245  canonProtocols, isKindOf);
5246 
5247  // Regenerate InsertPos.
5248  ObjCObjectTypes.FindNodeOrInsertPos(ID, InsertPos);
5249  }
5250 
5251  unsigned size = sizeof(ObjCObjectTypeImpl);
5252  size += typeArgs.size() * sizeof(QualType);
5253  size += protocols.size() * sizeof(ObjCProtocolDecl *);
5254  void *mem = Allocate(size, TypeAlignment);
5255  auto *T =
5256  new (mem) ObjCObjectTypeImpl(canonical, baseType, typeArgs, protocols,
5257  isKindOf);
5258 
5259  Types.push_back(T);
5260  ObjCObjectTypes.InsertNode(T, InsertPos);
5261  return QualType(T, 0);
5262 }
5263 
5264 /// Apply Objective-C protocol qualifiers to the given type.
5265 /// If this is for the canonical type of a type parameter, we can apply
5266 /// protocol qualifiers on the ObjCObjectPointerType.
5267 QualType
5269  ArrayRef<ObjCProtocolDecl *> protocols, bool &hasError,
5270  bool allowOnPointerType) const {
5271  hasError = false;
5272 
5273  if (const auto *objT = dyn_cast<ObjCTypeParamType>(type.getTypePtr())) {
5274  return getObjCTypeParamType(objT->getDecl(), protocols);
5275  }
5276 
5277  // Apply protocol qualifiers to ObjCObjectPointerType.
5278  if (allowOnPointerType) {
5279  if (const auto *objPtr =
5280  dyn_cast<ObjCObjectPointerType>(type.getTypePtr())) {
5281  const ObjCObjectType *objT = objPtr->getObjectType();
5282  // Merge protocol lists and construct ObjCObjectType.
5283  SmallVector<ObjCProtocolDecl*, 8> protocolsVec;
5284  protocolsVec.append(objT->qual_begin(),
5285  objT->qual_end());
5286  protocolsVec.append(protocols.begin(), protocols.end());
5287  ArrayRef<ObjCProtocolDecl *> protocols = protocolsVec;
5289  objT->getBaseType(),
5290  objT->getTypeArgsAsWritten(),
5291  protocols,
5292  objT->isKindOfTypeAsWritten());
5294  }
5295  }
5296 
5297  // Apply protocol qualifiers to ObjCObjectType.
5298  if (const auto *objT = dyn_cast<ObjCObjectType>(type.getTypePtr())){
5299  // FIXME: Check for protocols to which the class type is already
5300  // known to conform.
5301 
5302  return getObjCObjectType(objT->getBaseType(),
5303  objT->getTypeArgsAsWritten(),
5304  protocols,
5305  objT->isKindOfTypeAsWritten());
5306  }
5307 
5308  // If the canonical type is ObjCObjectType, ...
5309  if (type->isObjCObjectType()) {
5310  // Silently overwrite any existing protocol qualifiers.
5311  // TODO: determine whether that's the right thing to do.
5312 
5313  // FIXME: Check for protocols to which the class type is already
5314  // known to conform.
5315  return getObjCObjectType(type, {}, protocols, false);
5316  }
5317 
5318  // id<protocol-list>
5319  if (type->isObjCIdType()) {
5320  const auto *objPtr = type->castAs<ObjCObjectPointerType>();
5321  type = getObjCObjectType(ObjCBuiltinIdTy, {}, protocols,
5322  objPtr->isKindOfType());
5324  }
5325 
5326  // Class<protocol-list>
5327  if (type->isObjCClassType()) {
5328  const auto *objPtr = type->castAs<ObjCObjectPointerType>();
5329  type = getObjCObjectType(ObjCBuiltinClassTy, {}, protocols,
5330  objPtr->isKindOfType());
5332  }
5333 
5334  hasError = true;
5335  return type;
5336 }
5337 
5338 QualType
5340  ArrayRef<ObjCProtocolDecl *> protocols) const {
5341  // Look in the folding set for an existing type.
5342  llvm::FoldingSetNodeID ID;
5343  ObjCTypeParamType::Profile(ID, Decl, Decl->getUnderlyingType(), protocols);
5344  void *InsertPos = nullptr;
5345  if (ObjCTypeParamType *TypeParam =
5346  ObjCTypeParamTypes.FindNodeOrInsertPos(ID, InsertPos))
5347  return QualType(TypeParam, 0);
5348 
5349  // We canonicalize to the underlying type.
5350  QualType Canonical = getCanonicalType(Decl->getUnderlyingType());
5351  if (!protocols.empty()) {
5352  // Apply the protocol qualifers.
5353  bool hasError;
5355  Canonical, protocols, hasError, true /*allowOnPointerType*/));
5356  assert(!hasError && "Error when apply protocol qualifier to bound type");
5357  }
5358 
5359  unsigned size = sizeof(ObjCTypeParamType);
5360  size += protocols.size() * sizeof(ObjCProtocolDecl *);
5361  void *mem = Allocate(size, TypeAlignment);
5362  auto *newType = new (mem) ObjCTypeParamType(Decl, Canonical, protocols);
5363 
5364  Types.push_back(newType);
5365  ObjCTypeParamTypes.InsertNode(newType, InsertPos);
5366  return QualType(newType, 0);
5367 }
5368 
5370  ObjCTypeParamDecl *New) const {
5372  // Update TypeForDecl after updating TypeSourceInfo.
5373  auto NewTypeParamTy = cast<ObjCTypeParamType>(New->getTypeForDecl());
5375  protocols.append(NewTypeParamTy->qual_begin(), NewTypeParamTy->qual_end());
5376  QualType UpdatedTy = getObjCTypeParamType(New, protocols);
5377  New->setTypeForDecl(UpdatedTy.getTypePtr());
5378 }
5379 
5380 /// ObjCObjectAdoptsQTypeProtocols - Checks that protocols in IC's
5381 /// protocol list adopt all protocols in QT's qualified-id protocol
5382 /// list.
5384  ObjCInterfaceDecl *IC) {
5385  if (!QT->isObjCQualifiedIdType())
5386  return false;
5387 
5388  if (const auto *OPT = QT->getAs<ObjCObjectPointerType>()) {
5389  // If both the right and left sides have qualifiers.
5390  for (auto *Proto : OPT->quals()) {
5391  if (!IC->ClassImplementsProtocol(Proto, false))
5392  return false;
5393  }
5394  return true;
5395  }
5396  return false;
5397 }
5398 
5399 /// QIdProtocolsAdoptObjCObjectProtocols - Checks that protocols in
5400 /// QT's qualified-id protocol list adopt all protocols in IDecl's list
5401 /// of protocols.
5403  ObjCInterfaceDecl *IDecl) {
5404  if (!QT->isObjCQualifiedIdType())
5405  return false;
5406  const auto *OPT = QT->getAs<ObjCObjectPointerType>();
5407  if (!OPT)
<