1 //===--- CGExpr.cpp - Emit LLVM Code from Expressions ---------------------===//
2 //
3 //                     The LLVM Compiler Infrastructure
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 //===----------------------------------------------------------------------===//
9 //
10 // This contains code to emit Expr nodes as LLVM code.
11 //
12 //===----------------------------------------------------------------------===//
13 
14 #include "CodeGenFunction.h"
15 #include "CGCXXABI.h"
16 #include "CGCall.h"
17 #include "CGDebugInfo.h"
18 #include "CGObjCRuntime.h"
19 #include "CGRecordLayout.h"
20 #include "CodeGenModule.h"
21 #include "TargetInfo.h"
22 #include "clang/AST/ASTContext.h"
23 #include "clang/AST/DeclObjC.h"
24 #include "clang/AST/Attr.h"
25 #include "clang/Frontend/CodeGenOptions.h"
26 #include "llvm/ADT/Hashing.h"
27 #include "llvm/ADT/StringExtras.h"
28 #include "llvm/IR/DataLayout.h"
29 #include "llvm/IR/Intrinsics.h"
30 #include "llvm/IR/LLVMContext.h"
31 #include "llvm/IR/MDBuilder.h"
32 #include "llvm/Support/ConvertUTF.h"
33 
34 using namespace clang;
35 using namespace CodeGen;
36 
37 //===--------------------------------------------------------------------===//
38 //                        Miscellaneous Helper Methods
39 //===--------------------------------------------------------------------===//
40 
41 llvm::Value *CodeGenFunction::EmitCastToVoidPtr(llvm::Value *value) {
42   unsigned addressSpace =
43     cast<llvm::PointerType>(value->getType())->getAddressSpace();
44 
45   llvm::PointerType *destType = Int8PtrTy;
46   if (addressSpace)
47     destType = llvm::Type::getInt8PtrTy(getLLVMContext(), addressSpace);
48 
49   if (value->getType() == destType) return value;
50   return Builder.CreateBitCast(value, destType);
51 }
52 
53 /// CreateTempAlloca - This creates a alloca and inserts it into the entry
54 /// block.
55 llvm::AllocaInst *CodeGenFunction::CreateTempAlloca(llvm::Type *Ty,
56                                                     const Twine &Name) {
57   if (!Builder.isNamePreserving())
58     return new llvm::AllocaInst(Ty, nullptr, "", AllocaInsertPt);
59   return new llvm::AllocaInst(Ty, nullptr, Name, AllocaInsertPt);
60 }
61 
62 void CodeGenFunction::InitTempAlloca(llvm::AllocaInst *Var,
63                                      llvm::Value *Init) {
64   auto *Store = new llvm::StoreInst(Init, Var);
65   llvm::BasicBlock *Block = AllocaInsertPt->getParent();
66   Block->getInstList().insertAfter(&*AllocaInsertPt, Store);
67 }
68 
69 llvm::AllocaInst *CodeGenFunction::CreateIRTemp(QualType Ty,
70                                                 const Twine &Name) {
71   llvm::AllocaInst *Alloc = CreateTempAlloca(ConvertType(Ty), Name);
72   // FIXME: Should we prefer the preferred type alignment here?
73   CharUnits Align = getContext().getTypeAlignInChars(Ty);
74   Alloc->setAlignment(Align.getQuantity());
75   return Alloc;
76 }
77 
78 llvm::AllocaInst *CodeGenFunction::CreateMemTemp(QualType Ty,
79                                                  const Twine &Name) {
80   llvm::AllocaInst *Alloc = CreateTempAlloca(ConvertTypeForMem(Ty), Name);
81   // FIXME: Should we prefer the preferred type alignment here?
82   CharUnits Align = getContext().getTypeAlignInChars(Ty);
83   Alloc->setAlignment(Align.getQuantity());
84   return Alloc;
85 }
86 
87 /// EvaluateExprAsBool - Perform the usual unary conversions on the specified
88 /// expression and compare the result against zero, returning an Int1Ty value.
89 llvm::Value *CodeGenFunction::EvaluateExprAsBool(const Expr *E) {
90   PGO.setCurrentStmt(E);
91   if (const MemberPointerType *MPT = E->getType()->getAs<MemberPointerType>()) {
92     llvm::Value *MemPtr = EmitScalarExpr(E);
93     return CGM.getCXXABI().EmitMemberPointerIsNotNull(*this, MemPtr, MPT);
94   }
95 
96   QualType BoolTy = getContext().BoolTy;
97   if (!E->getType()->isAnyComplexType())
98     return EmitScalarConversion(EmitScalarExpr(E), E->getType(), BoolTy);
99 
100   return EmitComplexToScalarConversion(EmitComplexExpr(E), E->getType(),BoolTy);
101 }
102 
103 /// EmitIgnoredExpr - Emit code to compute the specified expression,
104 /// ignoring the result.
105 void CodeGenFunction::EmitIgnoredExpr(const Expr *E) {
106   if (E->isRValue())
107     return (void) EmitAnyExpr(E, AggValueSlot::ignored(), true);
108 
109   // Just emit it as an l-value and drop the result.
110   EmitLValue(E);
111 }
112 
113 /// EmitAnyExpr - Emit code to compute the specified expression which
114 /// can have any type.  The result is returned as an RValue struct.
115 /// If this is an aggregate expression, AggSlot indicates where the
116 /// result should be returned.
117 RValue CodeGenFunction::EmitAnyExpr(const Expr *E,
118                                     AggValueSlot aggSlot,
119                                     bool ignoreResult) {
120   switch (getEvaluationKind(E->getType())) {
121   case TEK_Scalar:
122     return RValue::get(EmitScalarExpr(E, ignoreResult));
123   case TEK_Complex:
124     return RValue::getComplex(EmitComplexExpr(E, ignoreResult, ignoreResult));
125   case TEK_Aggregate:
126     if (!ignoreResult && aggSlot.isIgnored())
127       aggSlot = CreateAggTemp(E->getType(), "agg-temp");
128     EmitAggExpr(E, aggSlot);
129     return aggSlot.asRValue();
130   }
131   llvm_unreachable("bad evaluation kind");
132 }
133 
134 /// EmitAnyExprToTemp - Similary to EmitAnyExpr(), however, the result will
135 /// always be accessible even if no aggregate location is provided.
136 RValue CodeGenFunction::EmitAnyExprToTemp(const Expr *E) {
137   AggValueSlot AggSlot = AggValueSlot::ignored();
138 
139   if (hasAggregateEvaluationKind(E->getType()))
140     AggSlot = CreateAggTemp(E->getType(), "agg.tmp");
141   return EmitAnyExpr(E, AggSlot);
142 }
143 
144 /// EmitAnyExprToMem - Evaluate an expression into a given memory
145 /// location.
146 void CodeGenFunction::EmitAnyExprToMem(const Expr *E,
147                                        llvm::Value *Location,
148                                        Qualifiers Quals,
149                                        bool IsInit) {
150   // FIXME: This function should take an LValue as an argument.
151   switch (getEvaluationKind(E->getType())) {
152   case TEK_Complex:
153     EmitComplexExprIntoLValue(E,
154                          MakeNaturalAlignAddrLValue(Location, E->getType()),
155                               /*isInit*/ false);
156     return;
157 
158   case TEK_Aggregate: {
159     CharUnits Alignment = getContext().getTypeAlignInChars(E->getType());
160     EmitAggExpr(E, AggValueSlot::forAddr(Location, Alignment, Quals,
161                                          AggValueSlot::IsDestructed_t(IsInit),
162                                          AggValueSlot::DoesNotNeedGCBarriers,
163                                          AggValueSlot::IsAliased_t(!IsInit)));
164     return;
165   }
166 
167   case TEK_Scalar: {
168     RValue RV = RValue::get(EmitScalarExpr(E, /*Ignore*/ false));
169     LValue LV = MakeAddrLValue(Location, E->getType());
170     EmitStoreThroughLValue(RV, LV);
171     return;
172   }
173   }
174   llvm_unreachable("bad evaluation kind");
175 }
176 
177 static void
178 pushTemporaryCleanup(CodeGenFunction &CGF, const MaterializeTemporaryExpr *M,
179                      const Expr *E, llvm::Value *ReferenceTemporary) {
180   // Objective-C++ ARC:
181   //   If we are binding a reference to a temporary that has ownership, we
182   //   need to perform retain/release operations on the temporary.
183   //
184   // FIXME: This should be looking at E, not M.
185   if (CGF.getLangOpts().ObjCAutoRefCount &&
186       M->getType()->isObjCLifetimeType()) {
187     QualType ObjCARCReferenceLifetimeType = M->getType();
188     switch (Qualifiers::ObjCLifetime Lifetime =
189                 ObjCARCReferenceLifetimeType.getObjCLifetime()) {
190     case Qualifiers::OCL_None:
191     case Qualifiers::OCL_ExplicitNone:
192       // Carry on to normal cleanup handling.
193       break;
194 
195     case Qualifiers::OCL_Autoreleasing:
196       // Nothing to do; cleaned up by an autorelease pool.
197       return;
198 
199     case Qualifiers::OCL_Strong:
200     case Qualifiers::OCL_Weak:
201       switch (StorageDuration Duration = M->getStorageDuration()) {
202       case SD_Static:
203         // Note: we intentionally do not register a cleanup to release
204         // the object on program termination.
205         return;
206 
207       case SD_Thread:
208         // FIXME: We should probably register a cleanup in this case.
209         return;
210 
211       case SD_Automatic:
212       case SD_FullExpression:
213         CodeGenFunction::Destroyer *Destroy;
214         CleanupKind CleanupKind;
215         if (Lifetime == Qualifiers::OCL_Strong) {
216           const ValueDecl *VD = M->getExtendingDecl();
217           bool Precise =
218               VD && isa<VarDecl>(VD) && VD->hasAttr<ObjCPreciseLifetimeAttr>();
219           CleanupKind = CGF.getARCCleanupKind();
220           Destroy = Precise ? &CodeGenFunction::destroyARCStrongPrecise
221                             : &CodeGenFunction::destroyARCStrongImprecise;
222         } else {
223           // __weak objects always get EH cleanups; otherwise, exceptions
224           // could cause really nasty crashes instead of mere leaks.
225           CleanupKind = NormalAndEHCleanup;
226           Destroy = &CodeGenFunction::destroyARCWeak;
227         }
228         if (Duration == SD_FullExpression)
229           CGF.pushDestroy(CleanupKind, ReferenceTemporary,
230                           ObjCARCReferenceLifetimeType, *Destroy,
231                           CleanupKind & EHCleanup);
232         else
233           CGF.pushLifetimeExtendedDestroy(CleanupKind, ReferenceTemporary,
234                                           ObjCARCReferenceLifetimeType,
235                                           *Destroy, CleanupKind & EHCleanup);
236         return;
237 
238       case SD_Dynamic:
239         llvm_unreachable("temporary cannot have dynamic storage duration");
240       }
241       llvm_unreachable("unknown storage duration");
242     }
243   }
244 
245   CXXDestructorDecl *ReferenceTemporaryDtor = nullptr;
246   if (const RecordType *RT =
247           E->getType()->getBaseElementTypeUnsafe()->getAs<RecordType>()) {
248     // Get the destructor for the reference temporary.
249     auto *ClassDecl = cast<CXXRecordDecl>(RT->getDecl());
250     if (!ClassDecl->hasTrivialDestructor())
251       ReferenceTemporaryDtor = ClassDecl->getDestructor();
252   }
253 
254   if (!ReferenceTemporaryDtor)
255     return;
256 
257   // Call the destructor for the temporary.
258   switch (M->getStorageDuration()) {
259   case SD_Static:
260   case SD_Thread: {
261     llvm::Constant *CleanupFn;
262     llvm::Constant *CleanupArg;
263     if (E->getType()->isArrayType()) {
264       CleanupFn = CodeGenFunction(CGF.CGM).generateDestroyHelper(
265           cast<llvm::Constant>(ReferenceTemporary), E->getType(),
266           CodeGenFunction::destroyCXXObject, CGF.getLangOpts().Exceptions,
267           dyn_cast_or_null<VarDecl>(M->getExtendingDecl()));
268       CleanupArg = llvm::Constant::getNullValue(CGF.Int8PtrTy);
269     } else {
270       CleanupFn = CGF.CGM.getAddrOfCXXStructor(ReferenceTemporaryDtor,
271                                                StructorType::Complete);
272       CleanupArg = cast<llvm::Constant>(ReferenceTemporary);
273     }
274     CGF.CGM.getCXXABI().registerGlobalDtor(
275         CGF, *cast<VarDecl>(M->getExtendingDecl()), CleanupFn, CleanupArg);
276     break;
277   }
278 
279   case SD_FullExpression:
280     CGF.pushDestroy(NormalAndEHCleanup, ReferenceTemporary, E->getType(),
281                     CodeGenFunction::destroyCXXObject,
282                     CGF.getLangOpts().Exceptions);
283     break;
284 
285   case SD_Automatic:
286     CGF.pushLifetimeExtendedDestroy(NormalAndEHCleanup,
287                                     ReferenceTemporary, E->getType(),
288                                     CodeGenFunction::destroyCXXObject,
289                                     CGF.getLangOpts().Exceptions);
290     break;
291 
292   case SD_Dynamic:
293     llvm_unreachable("temporary cannot have dynamic storage duration");
294   }
295 }
296 
297 static llvm::Value *
298 createReferenceTemporary(CodeGenFunction &CGF,
299                          const MaterializeTemporaryExpr *M, const Expr *Inner) {
300   switch (M->getStorageDuration()) {
301   case SD_FullExpression:
302   case SD_Automatic:
303     return CGF.CreateMemTemp(Inner->getType(), "ref.tmp");
304 
305   case SD_Thread:
306   case SD_Static:
307     return CGF.CGM.GetAddrOfGlobalTemporary(M, Inner);
308 
309   case SD_Dynamic:
310     llvm_unreachable("temporary can't have dynamic storage duration");
311   }
312   llvm_unreachable("unknown storage duration");
313 }
314 
315 LValue CodeGenFunction::
316 EmitMaterializeTemporaryExpr(const MaterializeTemporaryExpr *M) {
317   const Expr *E = M->GetTemporaryExpr();
318 
319     // FIXME: ideally this would use EmitAnyExprToMem, however, we cannot do so
320     // as that will cause the lifetime adjustment to be lost for ARC
321   if (getLangOpts().ObjCAutoRefCount &&
322       M->getType()->isObjCLifetimeType() &&
323       M->getType().getObjCLifetime() != Qualifiers::OCL_None &&
324       M->getType().getObjCLifetime() != Qualifiers::OCL_ExplicitNone) {
325     llvm::Value *Object = createReferenceTemporary(*this, M, E);
326     LValue RefTempDst = MakeAddrLValue(Object, M->getType());
327 
328     if (auto *Var = dyn_cast<llvm::GlobalVariable>(Object)) {
329       // We should not have emitted the initializer for this temporary as a
330       // constant.
331       assert(!Var->hasInitializer());
332       Var->setInitializer(CGM.EmitNullConstant(E->getType()));
333     }
334 
335     switch (getEvaluationKind(E->getType())) {
336     default: llvm_unreachable("expected scalar or aggregate expression");
337     case TEK_Scalar:
338       EmitScalarInit(E, M->getExtendingDecl(), RefTempDst, false);
339       break;
340     case TEK_Aggregate: {
341       CharUnits Alignment = getContext().getTypeAlignInChars(E->getType());
342       EmitAggExpr(E, AggValueSlot::forAddr(Object, Alignment,
343                                            E->getType().getQualifiers(),
344                                            AggValueSlot::IsDestructed,
345                                            AggValueSlot::DoesNotNeedGCBarriers,
346                                            AggValueSlot::IsNotAliased));
347       break;
348     }
349     }
350 
351     pushTemporaryCleanup(*this, M, E, Object);
352     return RefTempDst;
353   }
354 
355   SmallVector<const Expr *, 2> CommaLHSs;
356   SmallVector<SubobjectAdjustment, 2> Adjustments;
357   E = E->skipRValueSubobjectAdjustments(CommaLHSs, Adjustments);
358 
359   for (const auto &Ignored : CommaLHSs)
360     EmitIgnoredExpr(Ignored);
361 
362   if (const auto *opaque = dyn_cast<OpaqueValueExpr>(E)) {
363     if (opaque->getType()->isRecordType()) {
364       assert(Adjustments.empty());
365       return EmitOpaqueValueLValue(opaque);
366     }
367   }
368 
369   // Create and initialize the reference temporary.
370   llvm::Value *Object = createReferenceTemporary(*this, M, E);
371   if (auto *Var = dyn_cast<llvm::GlobalVariable>(Object)) {
372     // If the temporary is a global and has a constant initializer, we may
373     // have already initialized it.
374     if (!Var->hasInitializer()) {
375       Var->setInitializer(CGM.EmitNullConstant(E->getType()));
376       EmitAnyExprToMem(E, Object, Qualifiers(), /*IsInit*/true);
377     }
378   } else {
379     EmitAnyExprToMem(E, Object, Qualifiers(), /*IsInit*/true);
380   }
381   pushTemporaryCleanup(*this, M, E, Object);
382 
383   // Perform derived-to-base casts and/or field accesses, to get from the
384   // temporary object we created (and, potentially, for which we extended
385   // the lifetime) to the subobject we're binding the reference to.
386   for (unsigned I = Adjustments.size(); I != 0; --I) {
387     SubobjectAdjustment &Adjustment = Adjustments[I-1];
388     switch (Adjustment.Kind) {
389     case SubobjectAdjustment::DerivedToBaseAdjustment:
390       Object =
391           GetAddressOfBaseClass(Object, Adjustment.DerivedToBase.DerivedClass,
392                                 Adjustment.DerivedToBase.BasePath->path_begin(),
393                                 Adjustment.DerivedToBase.BasePath->path_end(),
394                                 /*NullCheckValue=*/ false, E->getExprLoc());
395       break;
396 
397     case SubobjectAdjustment::FieldAdjustment: {
398       LValue LV = MakeAddrLValue(Object, E->getType());
399       LV = EmitLValueForField(LV, Adjustment.Field);
400       assert(LV.isSimple() &&
401              "materialized temporary field is not a simple lvalue");
402       Object = LV.getAddress();
403       break;
404     }
405 
406     case SubobjectAdjustment::MemberPointerAdjustment: {
407       llvm::Value *Ptr = EmitScalarExpr(Adjustment.Ptr.RHS);
408       Object = CGM.getCXXABI().EmitMemberDataPointerAddress(
409           *this, E, Object, Ptr, Adjustment.Ptr.MPT);
410       break;
411     }
412     }
413   }
414 
415   return MakeAddrLValue(Object, M->getType());
416 }
417 
418 RValue
419 CodeGenFunction::EmitReferenceBindingToExpr(const Expr *E) {
420   // Emit the expression as an lvalue.
421   LValue LV = EmitLValue(E);
422   assert(LV.isSimple());
423   llvm::Value *Value = LV.getAddress();
424 
425   if (sanitizePerformTypeCheck() && !E->getType()->isFunctionType()) {
426     // C++11 [dcl.ref]p5 (as amended by core issue 453):
427     //   If a glvalue to which a reference is directly bound designates neither
428     //   an existing object or function of an appropriate type nor a region of
429     //   storage of suitable size and alignment to contain an object of the
430     //   reference's type, the behavior is undefined.
431     QualType Ty = E->getType();
432     EmitTypeCheck(TCK_ReferenceBinding, E->getExprLoc(), Value, Ty);
433   }
434 
435   return RValue::get(Value);
436 }
437 
438 
439 /// getAccessedFieldNo - Given an encoded value and a result number, return the
440 /// input field number being accessed.
441 unsigned CodeGenFunction::getAccessedFieldNo(unsigned Idx,
442                                              const llvm::Constant *Elts) {
443   return cast<llvm::ConstantInt>(Elts->getAggregateElement(Idx))
444       ->getZExtValue();
445 }
446 
447 /// Emit the hash_16_bytes function from include/llvm/ADT/Hashing.h.
448 static llvm::Value *emitHash16Bytes(CGBuilderTy &Builder, llvm::Value *Low,
449                                     llvm::Value *High) {
450   llvm::Value *KMul = Builder.getInt64(0x9ddfea08eb382d69ULL);
451   llvm::Value *K47 = Builder.getInt64(47);
452   llvm::Value *A0 = Builder.CreateMul(Builder.CreateXor(Low, High), KMul);
453   llvm::Value *A1 = Builder.CreateXor(Builder.CreateLShr(A0, K47), A0);
454   llvm::Value *B0 = Builder.CreateMul(Builder.CreateXor(High, A1), KMul);
455   llvm::Value *B1 = Builder.CreateXor(Builder.CreateLShr(B0, K47), B0);
456   return Builder.CreateMul(B1, KMul);
457 }
458 
459 bool CodeGenFunction::sanitizePerformTypeCheck() const {
460   return SanOpts.Null | SanOpts.Alignment | SanOpts.ObjectSize |
461          SanOpts.Vptr;
462 }
463 
464 void CodeGenFunction::EmitTypeCheck(TypeCheckKind TCK, SourceLocation Loc,
465                                     llvm::Value *Address, QualType Ty,
466                                     CharUnits Alignment, bool SkipNullCheck) {
467   if (!sanitizePerformTypeCheck())
468     return;
469 
470   // Don't check pointers outside the default address space. The null check
471   // isn't correct, the object-size check isn't supported by LLVM, and we can't
472   // communicate the addresses to the runtime handler for the vptr check.
473   if (Address->getType()->getPointerAddressSpace())
474     return;
475 
476   SanitizerScope SanScope(this);
477 
478   llvm::Value *Cond = nullptr;
479   llvm::BasicBlock *Done = nullptr;
480 
481   bool AllowNullPointers = TCK == TCK_DowncastPointer || TCK == TCK_Upcast ||
482                            TCK == TCK_UpcastToVirtualBase;
483   if ((SanOpts.Null || AllowNullPointers) && !SkipNullCheck) {
484     // The glvalue must not be an empty glvalue.
485     Cond = Builder.CreateICmpNE(
486         Address, llvm::Constant::getNullValue(Address->getType()));
487 
488     if (AllowNullPointers) {
489       // When performing pointer casts, it's OK if the value is null.
490       // Skip the remaining checks in that case.
491       Done = createBasicBlock("null");
492       llvm::BasicBlock *Rest = createBasicBlock("not.null");
493       Builder.CreateCondBr(Cond, Rest, Done);
494       EmitBlock(Rest);
495       Cond = nullptr;
496     }
497   }
498 
499   if (SanOpts.ObjectSize && !Ty->isIncompleteType()) {
500     uint64_t Size = getContext().getTypeSizeInChars(Ty).getQuantity();
501 
502     // The glvalue must refer to a large enough storage region.
503     // FIXME: If Address Sanitizer is enabled, insert dynamic instrumentation
504     //        to check this.
505     // FIXME: Get object address space
506     llvm::Type *Tys[2] = { IntPtrTy, Int8PtrTy };
507     llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::objectsize, Tys);
508     llvm::Value *Min = Builder.getFalse();
509     llvm::Value *CastAddr = Builder.CreateBitCast(Address, Int8PtrTy);
510     llvm::Value *LargeEnough =
511         Builder.CreateICmpUGE(Builder.CreateCall2(F, CastAddr, Min),
512                               llvm::ConstantInt::get(IntPtrTy, Size));
513     Cond = Cond ? Builder.CreateAnd(Cond, LargeEnough) : LargeEnough;
514   }
515 
516   uint64_t AlignVal = 0;
517 
518   if (SanOpts.Alignment) {
519     AlignVal = Alignment.getQuantity();
520     if (!Ty->isIncompleteType() && !AlignVal)
521       AlignVal = getContext().getTypeAlignInChars(Ty).getQuantity();
522 
523     // The glvalue must be suitably aligned.
524     if (AlignVal) {
525       llvm::Value *Align =
526           Builder.CreateAnd(Builder.CreatePtrToInt(Address, IntPtrTy),
527                             llvm::ConstantInt::get(IntPtrTy, AlignVal - 1));
528       llvm::Value *Aligned =
529         Builder.CreateICmpEQ(Align, llvm::ConstantInt::get(IntPtrTy, 0));
530       Cond = Cond ? Builder.CreateAnd(Cond, Aligned) : Aligned;
531     }
532   }
533 
534   if (Cond) {
535     llvm::Constant *StaticData[] = {
536       EmitCheckSourceLocation(Loc),
537       EmitCheckTypeDescriptor(Ty),
538       llvm::ConstantInt::get(SizeTy, AlignVal),
539       llvm::ConstantInt::get(Int8Ty, TCK)
540     };
541     EmitCheck(Cond, "type_mismatch", StaticData, Address, CRK_Recoverable);
542   }
543 
544   // If possible, check that the vptr indicates that there is a subobject of
545   // type Ty at offset zero within this object.
546   //
547   // C++11 [basic.life]p5,6:
548   //   [For storage which does not refer to an object within its lifetime]
549   //   The program has undefined behavior if:
550   //    -- the [pointer or glvalue] is used to access a non-static data member
551   //       or call a non-static member function
552   CXXRecordDecl *RD = Ty->getAsCXXRecordDecl();
553   if (SanOpts.Vptr &&
554       (TCK == TCK_MemberAccess || TCK == TCK_MemberCall ||
555        TCK == TCK_DowncastPointer || TCK == TCK_DowncastReference ||
556        TCK == TCK_UpcastToVirtualBase) &&
557       RD && RD->hasDefinition() && RD->isDynamicClass()) {
558     // Compute a hash of the mangled name of the type.
559     //
560     // FIXME: This is not guaranteed to be deterministic! Move to a
561     //        fingerprinting mechanism once LLVM provides one. For the time
562     //        being the implementation happens to be deterministic.
563     SmallString<64> MangledName;
564     llvm::raw_svector_ostream Out(MangledName);
565     CGM.getCXXABI().getMangleContext().mangleCXXRTTI(Ty.getUnqualifiedType(),
566                                                      Out);
567 
568     // Blacklist based on the mangled type.
569     if (!CGM.getContext().getSanitizerBlacklist().isBlacklistedType(
570             Out.str())) {
571       llvm::hash_code TypeHash = hash_value(Out.str());
572 
573       // Load the vptr, and compute hash_16_bytes(TypeHash, vptr).
574       llvm::Value *Low = llvm::ConstantInt::get(Int64Ty, TypeHash);
575       llvm::Type *VPtrTy = llvm::PointerType::get(IntPtrTy, 0);
576       llvm::Value *VPtrAddr = Builder.CreateBitCast(Address, VPtrTy);
577       llvm::Value *VPtrVal = Builder.CreateLoad(VPtrAddr);
578       llvm::Value *High = Builder.CreateZExt(VPtrVal, Int64Ty);
579 
580       llvm::Value *Hash = emitHash16Bytes(Builder, Low, High);
581       Hash = Builder.CreateTrunc(Hash, IntPtrTy);
582 
583       // Look the hash up in our cache.
584       const int CacheSize = 128;
585       llvm::Type *HashTable = llvm::ArrayType::get(IntPtrTy, CacheSize);
586       llvm::Value *Cache = CGM.CreateRuntimeVariable(HashTable,
587                                                      "__ubsan_vptr_type_cache");
588       llvm::Value *Slot = Builder.CreateAnd(Hash,
589                                             llvm::ConstantInt::get(IntPtrTy,
590                                                                    CacheSize-1));
591       llvm::Value *Indices[] = { Builder.getInt32(0), Slot };
592       llvm::Value *CacheVal =
593         Builder.CreateLoad(Builder.CreateInBoundsGEP(Cache, Indices));
594 
595       // If the hash isn't in the cache, call a runtime handler to perform the
596       // hard work of checking whether the vptr is for an object of the right
597       // type. This will either fill in the cache and return, or produce a
598       // diagnostic.
599       llvm::Constant *StaticData[] = {
600         EmitCheckSourceLocation(Loc),
601         EmitCheckTypeDescriptor(Ty),
602         CGM.GetAddrOfRTTIDescriptor(Ty.getUnqualifiedType()),
603         llvm::ConstantInt::get(Int8Ty, TCK)
604       };
605       llvm::Value *DynamicData[] = { Address, Hash };
606       EmitCheck(Builder.CreateICmpEQ(CacheVal, Hash),
607                 "dynamic_type_cache_miss", StaticData, DynamicData,
608                 CRK_AlwaysRecoverable);
609     }
610   }
611 
612   if (Done) {
613     Builder.CreateBr(Done);
614     EmitBlock(Done);
615   }
616 }
617 
618 /// Determine whether this expression refers to a flexible array member in a
619 /// struct. We disable array bounds checks for such members.
620 static bool isFlexibleArrayMemberExpr(const Expr *E) {
621   // For compatibility with existing code, we treat arrays of length 0 or
622   // 1 as flexible array members.
623   const ArrayType *AT = E->getType()->castAsArrayTypeUnsafe();
624   if (const auto *CAT = dyn_cast<ConstantArrayType>(AT)) {
625     if (CAT->getSize().ugt(1))
626       return false;
627   } else if (!isa<IncompleteArrayType>(AT))
628     return false;
629 
630   E = E->IgnoreParens();
631 
632   // A flexible array member must be the last member in the class.
633   if (const auto *ME = dyn_cast<MemberExpr>(E)) {
634     // FIXME: If the base type of the member expr is not FD->getParent(),
635     // this should not be treated as a flexible array member access.
636     if (const auto *FD = dyn_cast<FieldDecl>(ME->getMemberDecl())) {
637       RecordDecl::field_iterator FI(
638           DeclContext::decl_iterator(const_cast<FieldDecl *>(FD)));
639       return ++FI == FD->getParent()->field_end();
640     }
641   }
642 
643   return false;
644 }
645 
646 /// If Base is known to point to the start of an array, return the length of
647 /// that array. Return 0 if the length cannot be determined.
648 static llvm::Value *getArrayIndexingBound(
649     CodeGenFunction &CGF, const Expr *Base, QualType &IndexedType) {
650   // For the vector indexing extension, the bound is the number of elements.
651   if (const VectorType *VT = Base->getType()->getAs<VectorType>()) {
652     IndexedType = Base->getType();
653     return CGF.Builder.getInt32(VT->getNumElements());
654   }
655 
656   Base = Base->IgnoreParens();
657 
658   if (const auto *CE = dyn_cast<CastExpr>(Base)) {
659     if (CE->getCastKind() == CK_ArrayToPointerDecay &&
660         !isFlexibleArrayMemberExpr(CE->getSubExpr())) {
661       IndexedType = CE->getSubExpr()->getType();
662       const ArrayType *AT = IndexedType->castAsArrayTypeUnsafe();
663       if (const auto *CAT = dyn_cast<ConstantArrayType>(AT))
664         return CGF.Builder.getInt(CAT->getSize());
665       else if (const auto *VAT = dyn_cast<VariableArrayType>(AT))
666         return CGF.getVLASize(VAT).first;
667     }
668   }
669 
670   return nullptr;
671 }
672 
673 void CodeGenFunction::EmitBoundsCheck(const Expr *E, const Expr *Base,
674                                       llvm::Value *Index, QualType IndexType,
675                                       bool Accessed) {
676   assert(SanOpts.ArrayBounds &&
677          "should not be called unless adding bounds checks");
678   SanitizerScope SanScope(this);
679 
680   QualType IndexedType;
681   llvm::Value *Bound = getArrayIndexingBound(*this, Base, IndexedType);
682   if (!Bound)
683     return;
684 
685   bool IndexSigned = IndexType->isSignedIntegerOrEnumerationType();
686   llvm::Value *IndexVal = Builder.CreateIntCast(Index, SizeTy, IndexSigned);
687   llvm::Value *BoundVal = Builder.CreateIntCast(Bound, SizeTy, false);
688 
689   llvm::Constant *StaticData[] = {
690     EmitCheckSourceLocation(E->getExprLoc()),
691     EmitCheckTypeDescriptor(IndexedType),
692     EmitCheckTypeDescriptor(IndexType)
693   };
694   llvm::Value *Check = Accessed ? Builder.CreateICmpULT(IndexVal, BoundVal)
695                                 : Builder.CreateICmpULE(IndexVal, BoundVal);
696   EmitCheck(Check, "out_of_bounds", StaticData, Index, CRK_Recoverable);
697 }
698 
699 
700 CodeGenFunction::ComplexPairTy CodeGenFunction::
701 EmitComplexPrePostIncDec(const UnaryOperator *E, LValue LV,
702                          bool isInc, bool isPre) {
703   ComplexPairTy InVal = EmitLoadOfComplex(LV, E->getExprLoc());
704 
705   llvm::Value *NextVal;
706   if (isa<llvm::IntegerType>(InVal.first->getType())) {
707     uint64_t AmountVal = isInc ? 1 : -1;
708     NextVal = llvm::ConstantInt::get(InVal.first->getType(), AmountVal, true);
709 
710     // Add the inc/dec to the real part.
711     NextVal = Builder.CreateAdd(InVal.first, NextVal, isInc ? "inc" : "dec");
712   } else {
713     QualType ElemTy = E->getType()->getAs<ComplexType>()->getElementType();
714     llvm::APFloat FVal(getContext().getFloatTypeSemantics(ElemTy), 1);
715     if (!isInc)
716       FVal.changeSign();
717     NextVal = llvm::ConstantFP::get(getLLVMContext(), FVal);
718 
719     // Add the inc/dec to the real part.
720     NextVal = Builder.CreateFAdd(InVal.first, NextVal, isInc ? "inc" : "dec");
721   }
722 
723   ComplexPairTy IncVal(NextVal, InVal.second);
724 
725   // Store the updated result through the lvalue.
726   EmitStoreOfComplex(IncVal, LV, /*init*/ false);
727 
728   // If this is a postinc, return the value read from memory, otherwise use the
729   // updated value.
730   return isPre ? IncVal : InVal;
731 }
732 
733 //===----------------------------------------------------------------------===//
734 //                         LValue Expression Emission
735 //===----------------------------------------------------------------------===//
736 
737 RValue CodeGenFunction::GetUndefRValue(QualType Ty) {
738   if (Ty->isVoidType())
739     return RValue::get(nullptr);
740 
741   switch (getEvaluationKind(Ty)) {
742   case TEK_Complex: {
743     llvm::Type *EltTy =
744       ConvertType(Ty->castAs<ComplexType>()->getElementType());
745     llvm::Value *U = llvm::UndefValue::get(EltTy);
746     return RValue::getComplex(std::make_pair(U, U));
747   }
748 
749   // If this is a use of an undefined aggregate type, the aggregate must have an
750   // identifiable address.  Just because the contents of the value are undefined
751   // doesn't mean that the address can't be taken and compared.
752   case TEK_Aggregate: {
753     llvm::Value *DestPtr = CreateMemTemp(Ty, "undef.agg.tmp");
754     return RValue::getAggregate(DestPtr);
755   }
756 
757   case TEK_Scalar:
758     return RValue::get(llvm::UndefValue::get(ConvertType(Ty)));
759   }
760   llvm_unreachable("bad evaluation kind");
761 }
762 
763 RValue CodeGenFunction::EmitUnsupportedRValue(const Expr *E,
764                                               const char *Name) {
765   ErrorUnsupported(E, Name);
766   return GetUndefRValue(E->getType());
767 }
768 
769 LValue CodeGenFunction::EmitUnsupportedLValue(const Expr *E,
770                                               const char *Name) {
771   ErrorUnsupported(E, Name);
772   llvm::Type *Ty = llvm::PointerType::getUnqual(ConvertType(E->getType()));
773   return MakeAddrLValue(llvm::UndefValue::get(Ty), E->getType());
774 }
775 
776 LValue CodeGenFunction::EmitCheckedLValue(const Expr *E, TypeCheckKind TCK) {
777   LValue LV;
778   if (SanOpts.ArrayBounds && isa<ArraySubscriptExpr>(E))
779     LV = EmitArraySubscriptExpr(cast<ArraySubscriptExpr>(E), /*Accessed*/true);
780   else
781     LV = EmitLValue(E);
782   if (!isa<DeclRefExpr>(E) && !LV.isBitField() && LV.isSimple())
783     EmitTypeCheck(TCK, E->getExprLoc(), LV.getAddress(),
784                   E->getType(), LV.getAlignment());
785   return LV;
786 }
787 
788 /// EmitLValue - Emit code to compute a designator that specifies the location
789 /// of the expression.
790 ///
791 /// This can return one of two things: a simple address or a bitfield reference.
792 /// In either case, the LLVM Value* in the LValue structure is guaranteed to be
793 /// an LLVM pointer type.
794 ///
795 /// If this returns a bitfield reference, nothing about the pointee type of the
796 /// LLVM value is known: For example, it may not be a pointer to an integer.
797 ///
798 /// If this returns a normal address, and if the lvalue's C type is fixed size,
799 /// this method guarantees that the returned pointer type will point to an LLVM
800 /// type of the same size of the lvalue's type.  If the lvalue has a variable
801 /// length type, this is not possible.
802 ///
803 LValue CodeGenFunction::EmitLValue(const Expr *E) {
804   switch (E->getStmtClass()) {
805   default: return EmitUnsupportedLValue(E, "l-value expression");
806 
807   case Expr::ObjCPropertyRefExprClass:
808     llvm_unreachable("cannot emit a property reference directly");
809 
810   case Expr::ObjCSelectorExprClass:
811     return EmitObjCSelectorLValue(cast<ObjCSelectorExpr>(E));
812   case Expr::ObjCIsaExprClass:
813     return EmitObjCIsaExpr(cast<ObjCIsaExpr>(E));
814   case Expr::BinaryOperatorClass:
815     return EmitBinaryOperatorLValue(cast<BinaryOperator>(E));
816   case Expr::CompoundAssignOperatorClass:
817     if (!E->getType()->isAnyComplexType())
818       return EmitCompoundAssignmentLValue(cast<CompoundAssignOperator>(E));
819     return EmitComplexCompoundAssignmentLValue(cast<CompoundAssignOperator>(E));
820   case Expr::CallExprClass:
821   case Expr::CXXMemberCallExprClass:
822   case Expr::CXXOperatorCallExprClass:
823   case Expr::UserDefinedLiteralClass:
824     return EmitCallExprLValue(cast<CallExpr>(E));
825   case Expr::VAArgExprClass:
826     return EmitVAArgExprLValue(cast<VAArgExpr>(E));
827   case Expr::DeclRefExprClass:
828     return EmitDeclRefLValue(cast<DeclRefExpr>(E));
829   case Expr::ParenExprClass:
830     return EmitLValue(cast<ParenExpr>(E)->getSubExpr());
831   case Expr::GenericSelectionExprClass:
832     return EmitLValue(cast<GenericSelectionExpr>(E)->getResultExpr());
833   case Expr::PredefinedExprClass:
834     return EmitPredefinedLValue(cast<PredefinedExpr>(E));
835   case Expr::StringLiteralClass:
836     return EmitStringLiteralLValue(cast<StringLiteral>(E));
837   case Expr::ObjCEncodeExprClass:
838     return EmitObjCEncodeExprLValue(cast<ObjCEncodeExpr>(E));
839   case Expr::PseudoObjectExprClass:
840     return EmitPseudoObjectLValue(cast<PseudoObjectExpr>(E));
841   case Expr::InitListExprClass:
842     return EmitInitListLValue(cast<InitListExpr>(E));
843   case Expr::CXXTemporaryObjectExprClass:
844   case Expr::CXXConstructExprClass:
845     return EmitCXXConstructLValue(cast<CXXConstructExpr>(E));
846   case Expr::CXXBindTemporaryExprClass:
847     return EmitCXXBindTemporaryLValue(cast<CXXBindTemporaryExpr>(E));
848   case Expr::CXXUuidofExprClass:
849     return EmitCXXUuidofLValue(cast<CXXUuidofExpr>(E));
850   case Expr::LambdaExprClass:
851     return EmitLambdaLValue(cast<LambdaExpr>(E));
852 
853   case Expr::ExprWithCleanupsClass: {
854     const auto *cleanups = cast<ExprWithCleanups>(E);
855     enterFullExpression(cleanups);
856     RunCleanupsScope Scope(*this);
857     return EmitLValue(cleanups->getSubExpr());
858   }
859 
860   case Expr::CXXDefaultArgExprClass:
861     return EmitLValue(cast<CXXDefaultArgExpr>(E)->getExpr());
862   case Expr::CXXDefaultInitExprClass: {
863     CXXDefaultInitExprScope Scope(*this);
864     return EmitLValue(cast<CXXDefaultInitExpr>(E)->getExpr());
865   }
866   case Expr::CXXTypeidExprClass:
867     return EmitCXXTypeidLValue(cast<CXXTypeidExpr>(E));
868 
869   case Expr::ObjCMessageExprClass:
870     return EmitObjCMessageExprLValue(cast<ObjCMessageExpr>(E));
871   case Expr::ObjCIvarRefExprClass:
872     return EmitObjCIvarRefLValue(cast<ObjCIvarRefExpr>(E));
873   case Expr::StmtExprClass:
874     return EmitStmtExprLValue(cast<StmtExpr>(E));
875   case Expr::UnaryOperatorClass:
876     return EmitUnaryOpLValue(cast<UnaryOperator>(E));
877   case Expr::ArraySubscriptExprClass:
878     return EmitArraySubscriptExpr(cast<ArraySubscriptExpr>(E));
879   case Expr::ExtVectorElementExprClass:
880     return EmitExtVectorElementExpr(cast<ExtVectorElementExpr>(E));
881   case Expr::MemberExprClass:
882     return EmitMemberExpr(cast<MemberExpr>(E));
883   case Expr::CompoundLiteralExprClass:
884     return EmitCompoundLiteralLValue(cast<CompoundLiteralExpr>(E));
885   case Expr::ConditionalOperatorClass:
886     return EmitConditionalOperatorLValue(cast<ConditionalOperator>(E));
887   case Expr::BinaryConditionalOperatorClass:
888     return EmitConditionalOperatorLValue(cast<BinaryConditionalOperator>(E));
889   case Expr::ChooseExprClass:
890     return EmitLValue(cast<ChooseExpr>(E)->getChosenSubExpr());
891   case Expr::OpaqueValueExprClass:
892     return EmitOpaqueValueLValue(cast<OpaqueValueExpr>(E));
893   case Expr::SubstNonTypeTemplateParmExprClass:
894     return EmitLValue(cast<SubstNonTypeTemplateParmExpr>(E)->getReplacement());
895   case Expr::ImplicitCastExprClass:
896   case Expr::CStyleCastExprClass:
897   case Expr::CXXFunctionalCastExprClass:
898   case Expr::CXXStaticCastExprClass:
899   case Expr::CXXDynamicCastExprClass:
900   case Expr::CXXReinterpretCastExprClass:
901   case Expr::CXXConstCastExprClass:
902   case Expr::ObjCBridgedCastExprClass:
903     return EmitCastLValue(cast<CastExpr>(E));
904 
905   case Expr::MaterializeTemporaryExprClass:
906     return EmitMaterializeTemporaryExpr(cast<MaterializeTemporaryExpr>(E));
907   }
908 }
909 
910 /// Given an object of the given canonical type, can we safely copy a
911 /// value out of it based on its initializer?
912 static bool isConstantEmittableObjectType(QualType type) {
913   assert(type.isCanonical());
914   assert(!type->isReferenceType());
915 
916   // Must be const-qualified but non-volatile.
917   Qualifiers qs = type.getLocalQualifiers();
918   if (!qs.hasConst() || qs.hasVolatile()) return false;
919 
920   // Otherwise, all object types satisfy this except C++ classes with
921   // mutable subobjects or non-trivial copy/destroy behavior.
922   if (const auto *RT = dyn_cast<RecordType>(type))
923     if (const auto *RD = dyn_cast<CXXRecordDecl>(RT->getDecl()))
924       if (RD->hasMutableFields() || !RD->isTrivial())
925         return false;
926 
927   return true;
928 }
929 
930 /// Can we constant-emit a load of a reference to a variable of the
931 /// given type?  This is different from predicates like
932 /// Decl::isUsableInConstantExpressions because we do want it to apply
933 /// in situations that don't necessarily satisfy the language's rules
934 /// for this (e.g. C++'s ODR-use rules).  For example, we want to able
935 /// to do this with const float variables even if those variables
936 /// aren't marked 'constexpr'.
937 enum ConstantEmissionKind {
938   CEK_None,
939   CEK_AsReferenceOnly,
940   CEK_AsValueOrReference,
941   CEK_AsValueOnly
942 };
943 static ConstantEmissionKind checkVarTypeForConstantEmission(QualType type) {
944   type = type.getCanonicalType();
945   if (const auto *ref = dyn_cast<ReferenceType>(type)) {
946     if (isConstantEmittableObjectType(ref->getPointeeType()))
947       return CEK_AsValueOrReference;
948     return CEK_AsReferenceOnly;
949   }
950   if (isConstantEmittableObjectType(type))
951     return CEK_AsValueOnly;
952   return CEK_None;
953 }
954 
955 /// Try to emit a reference to the given value without producing it as
956 /// an l-value.  This is actually more than an optimization: we can't
957 /// produce an l-value for variables that we never actually captured
958 /// in a block or lambda, which means const int variables or constexpr
959 /// literals or similar.
960 CodeGenFunction::ConstantEmission
961 CodeGenFunction::tryEmitAsConstant(DeclRefExpr *refExpr) {
962   ValueDecl *value = refExpr->getDecl();
963 
964   // The value needs to be an enum constant or a constant variable.
965   ConstantEmissionKind CEK;
966   if (isa<ParmVarDecl>(value)) {
967     CEK = CEK_None;
968   } else if (auto *var = dyn_cast<VarDecl>(value)) {
969     CEK = checkVarTypeForConstantEmission(var->getType());
970   } else if (isa<EnumConstantDecl>(value)) {
971     CEK = CEK_AsValueOnly;
972   } else {
973     CEK = CEK_None;
974   }
975   if (CEK == CEK_None) return ConstantEmission();
976 
977   Expr::EvalResult result;
978   bool resultIsReference;
979   QualType resultType;
980 
981   // It's best to evaluate all the way as an r-value if that's permitted.
982   if (CEK != CEK_AsReferenceOnly &&
983       refExpr->EvaluateAsRValue(result, getContext())) {
984     resultIsReference = false;
985     resultType = refExpr->getType();
986 
987   // Otherwise, try to evaluate as an l-value.
988   } else if (CEK != CEK_AsValueOnly &&
989              refExpr->EvaluateAsLValue(result, getContext())) {
990     resultIsReference = true;
991     resultType = value->getType();
992 
993   // Failure.
994   } else {
995     return ConstantEmission();
996   }
997 
998   // In any case, if the initializer has side-effects, abandon ship.
999   if (result.HasSideEffects)
1000     return ConstantEmission();
1001 
1002   // Emit as a constant.
1003   llvm::Constant *C = CGM.EmitConstantValue(result.Val, resultType, this);
1004 
1005   // Make sure we emit a debug reference to the global variable.
1006   // This should probably fire even for
1007   if (isa<VarDecl>(value)) {
1008     if (!getContext().DeclMustBeEmitted(cast<VarDecl>(value)))
1009       EmitDeclRefExprDbgValue(refExpr, C);
1010   } else {
1011     assert(isa<EnumConstantDecl>(value));
1012     EmitDeclRefExprDbgValue(refExpr, C);
1013   }
1014 
1015   // If we emitted a reference constant, we need to dereference that.
1016   if (resultIsReference)
1017     return ConstantEmission::forReference(C);
1018 
1019   return ConstantEmission::forValue(C);
1020 }
1021 
1022 llvm::Value *CodeGenFunction::EmitLoadOfScalar(LValue lvalue,
1023                                                SourceLocation Loc) {
1024   return EmitLoadOfScalar(lvalue.getAddress(), lvalue.isVolatile(),
1025                           lvalue.getAlignment().getQuantity(),
1026                           lvalue.getType(), Loc, lvalue.getTBAAInfo(),
1027                           lvalue.getTBAABaseType(), lvalue.getTBAAOffset());
1028 }
1029 
1030 static bool hasBooleanRepresentation(QualType Ty) {
1031   if (Ty->isBooleanType())
1032     return true;
1033 
1034   if (const EnumType *ET = Ty->getAs<EnumType>())
1035     return ET->getDecl()->getIntegerType()->isBooleanType();
1036 
1037   if (const AtomicType *AT = Ty->getAs<AtomicType>())
1038     return hasBooleanRepresentation(AT->getValueType());
1039 
1040   return false;
1041 }
1042 
1043 static bool getRangeForType(CodeGenFunction &CGF, QualType Ty,
1044                             llvm::APInt &Min, llvm::APInt &End,
1045                             bool StrictEnums) {
1046   const EnumType *ET = Ty->getAs<EnumType>();
1047   bool IsRegularCPlusPlusEnum = CGF.getLangOpts().CPlusPlus && StrictEnums &&
1048                                 ET && !ET->getDecl()->isFixed();
1049   bool IsBool = hasBooleanRepresentation(Ty);
1050   if (!IsBool && !IsRegularCPlusPlusEnum)
1051     return false;
1052 
1053   if (IsBool) {
1054     Min = llvm::APInt(CGF.getContext().getTypeSize(Ty), 0);
1055     End = llvm::APInt(CGF.getContext().getTypeSize(Ty), 2);
1056   } else {
1057     const EnumDecl *ED = ET->getDecl();
1058     llvm::Type *LTy = CGF.ConvertTypeForMem(ED->getIntegerType());
1059     unsigned Bitwidth = LTy->getScalarSizeInBits();
1060     unsigned NumNegativeBits = ED->getNumNegativeBits();
1061     unsigned NumPositiveBits = ED->getNumPositiveBits();
1062 
1063     if (NumNegativeBits) {
1064       unsigned NumBits = std::max(NumNegativeBits, NumPositiveBits + 1);
1065       assert(NumBits <= Bitwidth);
1066       End = llvm::APInt(Bitwidth, 1) << (NumBits - 1);
1067       Min = -End;
1068     } else {
1069       assert(NumPositiveBits <= Bitwidth);
1070       End = llvm::APInt(Bitwidth, 1) << NumPositiveBits;
1071       Min = llvm::APInt(Bitwidth, 0);
1072     }
1073   }
1074   return true;
1075 }
1076 
1077 llvm::MDNode *CodeGenFunction::getRangeForLoadFromType(QualType Ty) {
1078   llvm::APInt Min, End;
1079   if (!getRangeForType(*this, Ty, Min, End,
1080                        CGM.getCodeGenOpts().StrictEnums))
1081     return nullptr;
1082 
1083   llvm::MDBuilder MDHelper(getLLVMContext());
1084   return MDHelper.createRange(Min, End);
1085 }
1086 
1087 llvm::Value *CodeGenFunction::EmitLoadOfScalar(llvm::Value *Addr, bool Volatile,
1088                                                unsigned Alignment, QualType Ty,
1089                                                SourceLocation Loc,
1090                                                llvm::MDNode *TBAAInfo,
1091                                                QualType TBAABaseType,
1092                                                uint64_t TBAAOffset) {
1093   // For better performance, handle vector loads differently.
1094   if (Ty->isVectorType()) {
1095     llvm::Value *V;
1096     const llvm::Type *EltTy =
1097     cast<llvm::PointerType>(Addr->getType())->getElementType();
1098 
1099     const auto *VTy = cast<llvm::VectorType>(EltTy);
1100 
1101     // Handle vectors of size 3, like size 4 for better performance.
1102     if (VTy->getNumElements() == 3) {
1103 
1104       // Bitcast to vec4 type.
1105       llvm::VectorType *vec4Ty = llvm::VectorType::get(VTy->getElementType(),
1106                                                          4);
1107       llvm::PointerType *ptVec4Ty =
1108       llvm::PointerType::get(vec4Ty,
1109                              (cast<llvm::PointerType>(
1110                                       Addr->getType()))->getAddressSpace());
1111       llvm::Value *Cast = Builder.CreateBitCast(Addr, ptVec4Ty,
1112                                                 "castToVec4");
1113       // Now load value.
1114       llvm::Value *LoadVal = Builder.CreateLoad(Cast, Volatile, "loadVec4");
1115 
1116       // Shuffle vector to get vec3.
1117       llvm::Constant *Mask[] = {
1118         llvm::ConstantInt::get(llvm::Type::getInt32Ty(getLLVMContext()), 0),
1119         llvm::ConstantInt::get(llvm::Type::getInt32Ty(getLLVMContext()), 1),
1120         llvm::ConstantInt::get(llvm::Type::getInt32Ty(getLLVMContext()), 2)
1121       };
1122 
1123       llvm::Value *MaskV = llvm::ConstantVector::get(Mask);
1124       V = Builder.CreateShuffleVector(LoadVal,
1125                                       llvm::UndefValue::get(vec4Ty),
1126                                       MaskV, "extractVec");
1127       return EmitFromMemory(V, Ty);
1128     }
1129   }
1130 
1131   // Atomic operations have to be done on integral types.
1132   if (Ty->isAtomicType()) {
1133     LValue lvalue = LValue::MakeAddr(Addr, Ty,
1134                                      CharUnits::fromQuantity(Alignment),
1135                                      getContext(), TBAAInfo);
1136     return EmitAtomicLoad(lvalue, Loc).getScalarVal();
1137   }
1138 
1139   llvm::LoadInst *Load = Builder.CreateLoad(Addr);
1140   if (Volatile)
1141     Load->setVolatile(true);
1142   if (Alignment)
1143     Load->setAlignment(Alignment);
1144   if (TBAAInfo) {
1145     llvm::MDNode *TBAAPath = CGM.getTBAAStructTagInfo(TBAABaseType, TBAAInfo,
1146                                                       TBAAOffset);
1147     if (TBAAPath)
1148       CGM.DecorateInstruction(Load, TBAAPath, false/*ConvertTypeToTag*/);
1149   }
1150 
1151   if ((SanOpts.Bool && hasBooleanRepresentation(Ty)) ||
1152       (SanOpts.Enum && Ty->getAs<EnumType>())) {
1153     SanitizerScope SanScope(this);
1154     llvm::APInt Min, End;
1155     if (getRangeForType(*this, Ty, Min, End, true)) {
1156       --End;
1157       llvm::Value *Check;
1158       if (!Min)
1159         Check = Builder.CreateICmpULE(
1160           Load, llvm::ConstantInt::get(getLLVMContext(), End));
1161       else {
1162         llvm::Value *Upper = Builder.CreateICmpSLE(
1163           Load, llvm::ConstantInt::get(getLLVMContext(), End));
1164         llvm::Value *Lower = Builder.CreateICmpSGE(
1165           Load, llvm::ConstantInt::get(getLLVMContext(), Min));
1166         Check = Builder.CreateAnd(Upper, Lower);
1167       }
1168       llvm::Constant *StaticArgs[] = {
1169         EmitCheckSourceLocation(Loc),
1170         EmitCheckTypeDescriptor(Ty)
1171       };
1172       EmitCheck(Check, "load_invalid_value", StaticArgs, EmitCheckValue(Load),
1173                 CRK_Recoverable);
1174     }
1175   } else if (CGM.getCodeGenOpts().OptimizationLevel > 0)
1176     if (llvm::MDNode *RangeInfo = getRangeForLoadFromType(Ty))
1177       Load->setMetadata(llvm::LLVMContext::MD_range, RangeInfo);
1178 
1179   return EmitFromMemory(Load, Ty);
1180 }
1181 
1182 llvm::Value *CodeGenFunction::EmitToMemory(llvm::Value *Value, QualType Ty) {
1183   // Bool has a different representation in memory than in registers.
1184   if (hasBooleanRepresentation(Ty)) {
1185     // This should really always be an i1, but sometimes it's already
1186     // an i8, and it's awkward to track those cases down.
1187     if (Value->getType()->isIntegerTy(1))
1188       return Builder.CreateZExt(Value, ConvertTypeForMem(Ty), "frombool");
1189     assert(Value->getType()->isIntegerTy(getContext().getTypeSize(Ty)) &&
1190            "wrong value rep of bool");
1191   }
1192 
1193   return Value;
1194 }
1195 
1196 llvm::Value *CodeGenFunction::EmitFromMemory(llvm::Value *Value, QualType Ty) {
1197   // Bool has a different representation in memory than in registers.
1198   if (hasBooleanRepresentation(Ty)) {
1199     assert(Value->getType()->isIntegerTy(getContext().getTypeSize(Ty)) &&
1200            "wrong value rep of bool");
1201     return Builder.CreateTrunc(Value, Builder.getInt1Ty(), "tobool");
1202   }
1203 
1204   return Value;
1205 }
1206 
1207 void CodeGenFunction::EmitStoreOfScalar(llvm::Value *Value, llvm::Value *Addr,
1208                                         bool Volatile, unsigned Alignment,
1209                                         QualType Ty, llvm::MDNode *TBAAInfo,
1210                                         bool isInit, QualType TBAABaseType,
1211                                         uint64_t TBAAOffset) {
1212 
1213   // Handle vectors differently to get better performance.
1214   if (Ty->isVectorType()) {
1215     llvm::Type *SrcTy = Value->getType();
1216     auto *VecTy = cast<llvm::VectorType>(SrcTy);
1217     // Handle vec3 special.
1218     if (VecTy->getNumElements() == 3) {
1219       llvm::LLVMContext &VMContext = getLLVMContext();
1220 
1221       // Our source is a vec3, do a shuffle vector to make it a vec4.
1222       SmallVector<llvm::Constant*, 4> Mask;
1223       Mask.push_back(llvm::ConstantInt::get(llvm::Type::getInt32Ty(VMContext),
1224                                             0));
1225       Mask.push_back(llvm::ConstantInt::get(llvm::Type::getInt32Ty(VMContext),
1226                                             1));
1227       Mask.push_back(llvm::ConstantInt::get(llvm::Type::getInt32Ty(VMContext),
1228                                             2));
1229       Mask.push_back(llvm::UndefValue::get(llvm::Type::getInt32Ty(VMContext)));
1230 
1231       llvm::Value *MaskV = llvm::ConstantVector::get(Mask);
1232       Value = Builder.CreateShuffleVector(Value,
1233                                           llvm::UndefValue::get(VecTy),
1234                                           MaskV, "extractVec");
1235       SrcTy = llvm::VectorType::get(VecTy->getElementType(), 4);
1236     }
1237     auto *DstPtr = cast<llvm::PointerType>(Addr->getType());
1238     if (DstPtr->getElementType() != SrcTy) {
1239       llvm::Type *MemTy =
1240       llvm::PointerType::get(SrcTy, DstPtr->getAddressSpace());
1241       Addr = Builder.CreateBitCast(Addr, MemTy, "storetmp");
1242     }
1243   }
1244 
1245   Value = EmitToMemory(Value, Ty);
1246 
1247   if (Ty->isAtomicType()) {
1248     EmitAtomicStore(RValue::get(Value),
1249                     LValue::MakeAddr(Addr, Ty,
1250                                      CharUnits::fromQuantity(Alignment),
1251                                      getContext(), TBAAInfo),
1252                     isInit);
1253     return;
1254   }
1255 
1256   llvm::StoreInst *Store = Builder.CreateStore(Value, Addr, Volatile);
1257   if (Alignment)
1258     Store->setAlignment(Alignment);
1259   if (TBAAInfo) {
1260     llvm::MDNode *TBAAPath = CGM.getTBAAStructTagInfo(TBAABaseType, TBAAInfo,
1261                                                       TBAAOffset);
1262     if (TBAAPath)
1263       CGM.DecorateInstruction(Store, TBAAPath, false/*ConvertTypeToTag*/);
1264   }
1265 }
1266 
1267 void CodeGenFunction::EmitStoreOfScalar(llvm::Value *value, LValue lvalue,
1268                                         bool isInit) {
1269   EmitStoreOfScalar(value, lvalue.getAddress(), lvalue.isVolatile(),
1270                     lvalue.getAlignment().getQuantity(), lvalue.getType(),
1271                     lvalue.getTBAAInfo(), isInit, lvalue.getTBAABaseType(),
1272                     lvalue.getTBAAOffset());
1273 }
1274 
1275 /// EmitLoadOfLValue - Given an expression that represents a value lvalue, this
1276 /// method emits the address of the lvalue, then loads the result as an rvalue,
1277 /// returning the rvalue.
1278 RValue CodeGenFunction::EmitLoadOfLValue(LValue LV, SourceLocation Loc) {
1279   if (LV.isObjCWeak()) {
1280     // load of a __weak object.
1281     llvm::Value *AddrWeakObj = LV.getAddress();
1282     return RValue::get(CGM.getObjCRuntime().EmitObjCWeakRead(*this,
1283                                                              AddrWeakObj));
1284   }
1285   if (LV.getQuals().getObjCLifetime() == Qualifiers::OCL_Weak) {
1286     llvm::Value *Object = EmitARCLoadWeakRetained(LV.getAddress());
1287     Object = EmitObjCConsumeObject(LV.getType(), Object);
1288     return RValue::get(Object);
1289   }
1290 
1291   if (LV.isSimple()) {
1292     assert(!LV.getType()->isFunctionType());
1293 
1294     // Everything needs a load.
1295     return RValue::get(EmitLoadOfScalar(LV, Loc));
1296   }
1297 
1298   if (LV.isVectorElt()) {
1299     llvm::LoadInst *Load = Builder.CreateLoad(LV.getVectorAddr(),
1300                                               LV.isVolatileQualified());
1301     Load->setAlignment(LV.getAlignment().getQuantity());
1302     return RValue::get(Builder.CreateExtractElement(Load, LV.getVectorIdx(),
1303                                                     "vecext"));
1304   }
1305 
1306   // If this is a reference to a subset of the elements of a vector, either
1307   // shuffle the input or extract/insert them as appropriate.
1308   if (LV.isExtVectorElt())
1309     return EmitLoadOfExtVectorElementLValue(LV);
1310 
1311   // Global Register variables always invoke intrinsics
1312   if (LV.isGlobalReg())
1313     return EmitLoadOfGlobalRegLValue(LV);
1314 
1315   assert(LV.isBitField() && "Unknown LValue type!");
1316   return EmitLoadOfBitfieldLValue(LV);
1317 }
1318 
1319 RValue CodeGenFunction::EmitLoadOfBitfieldLValue(LValue LV) {
1320   const CGBitFieldInfo &Info = LV.getBitFieldInfo();
1321 
1322   // Get the output type.
1323   llvm::Type *ResLTy = ConvertType(LV.getType());
1324 
1325   llvm::Value *Ptr = LV.getBitFieldAddr();
1326   llvm::Value *Val = Builder.CreateLoad(Ptr, LV.isVolatileQualified(),
1327                                         "bf.load");
1328   cast<llvm::LoadInst>(Val)->setAlignment(Info.StorageAlignment);
1329 
1330   if (Info.IsSigned) {
1331     assert(static_cast<unsigned>(Info.Offset + Info.Size) <= Info.StorageSize);
1332     unsigned HighBits = Info.StorageSize - Info.Offset - Info.Size;
1333     if (HighBits)
1334       Val = Builder.CreateShl(Val, HighBits, "bf.shl");
1335     if (Info.Offset + HighBits)
1336       Val = Builder.CreateAShr(Val, Info.Offset + HighBits, "bf.ashr");
1337   } else {
1338     if (Info.Offset)
1339       Val = Builder.CreateLShr(Val, Info.Offset, "bf.lshr");
1340     if (static_cast<unsigned>(Info.Offset) + Info.Size < Info.StorageSize)
1341       Val = Builder.CreateAnd(Val, llvm::APInt::getLowBitsSet(Info.StorageSize,
1342                                                               Info.Size),
1343                               "bf.clear");
1344   }
1345   Val = Builder.CreateIntCast(Val, ResLTy, Info.IsSigned, "bf.cast");
1346 
1347   return RValue::get(Val);
1348 }
1349 
1350 // If this is a reference to a subset of the elements of a vector, create an
1351 // appropriate shufflevector.
1352 RValue CodeGenFunction::EmitLoadOfExtVectorElementLValue(LValue LV) {
1353   llvm::LoadInst *Load = Builder.CreateLoad(LV.getExtVectorAddr(),
1354                                             LV.isVolatileQualified());
1355   Load->setAlignment(LV.getAlignment().getQuantity());
1356   llvm::Value *Vec = Load;
1357 
1358   const llvm::Constant *Elts = LV.getExtVectorElts();
1359 
1360   // If the result of the expression is a non-vector type, we must be extracting
1361   // a single element.  Just codegen as an extractelement.
1362   const VectorType *ExprVT = LV.getType()->getAs<VectorType>();
1363   if (!ExprVT) {
1364     unsigned InIdx = getAccessedFieldNo(0, Elts);
1365     llvm::Value *Elt = llvm::ConstantInt::get(SizeTy, InIdx);
1366     return RValue::get(Builder.CreateExtractElement(Vec, Elt));
1367   }
1368 
1369   // Always use shuffle vector to try to retain the original program structure
1370   unsigned NumResultElts = ExprVT->getNumElements();
1371 
1372   SmallVector<llvm::Constant*, 4> Mask;
1373   for (unsigned i = 0; i != NumResultElts; ++i)
1374     Mask.push_back(Builder.getInt32(getAccessedFieldNo(i, Elts)));
1375 
1376   llvm::Value *MaskV = llvm::ConstantVector::get(Mask);
1377   Vec = Builder.CreateShuffleVector(Vec, llvm::UndefValue::get(Vec->getType()),
1378                                     MaskV);
1379   return RValue::get(Vec);
1380 }
1381 
1382 /// @brief Generates lvalue for partial ext_vector access.
1383 llvm::Value *CodeGenFunction::EmitExtVectorElementLValue(LValue LV) {
1384   llvm::Value *VectorAddress = LV.getExtVectorAddr();
1385   const VectorType *ExprVT = LV.getType()->getAs<VectorType>();
1386   QualType EQT = ExprVT->getElementType();
1387   llvm::Type *VectorElementTy = CGM.getTypes().ConvertType(EQT);
1388   llvm::Type *VectorElementPtrToTy = VectorElementTy->getPointerTo();
1389 
1390   llvm::Value *CastToPointerElement =
1391     Builder.CreateBitCast(VectorAddress,
1392                           VectorElementPtrToTy, "conv.ptr.element");
1393 
1394   const llvm::Constant *Elts = LV.getExtVectorElts();
1395   unsigned ix = getAccessedFieldNo(0, Elts);
1396 
1397   llvm::Value *VectorBasePtrPlusIx =
1398     Builder.CreateInBoundsGEP(CastToPointerElement,
1399                               llvm::ConstantInt::get(SizeTy, ix), "add.ptr");
1400 
1401   return VectorBasePtrPlusIx;
1402 }
1403 
1404 /// @brief Load of global gamed gegisters are always calls to intrinsics.
1405 RValue CodeGenFunction::EmitLoadOfGlobalRegLValue(LValue LV) {
1406   assert((LV.getType()->isIntegerType() || LV.getType()->isPointerType()) &&
1407          "Bad type for register variable");
1408   llvm::MDNode *RegName = dyn_cast<llvm::MDNode>(LV.getGlobalReg());
1409   assert(RegName && "Register LValue is not metadata");
1410 
1411   // We accept integer and pointer types only
1412   llvm::Type *OrigTy = CGM.getTypes().ConvertType(LV.getType());
1413   llvm::Type *Ty = OrigTy;
1414   if (OrigTy->isPointerTy())
1415     Ty = CGM.getTypes().getDataLayout().getIntPtrType(OrigTy);
1416   llvm::Type *Types[] = { Ty };
1417 
1418   llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::read_register, Types);
1419   llvm::Value *Call = Builder.CreateCall(F, RegName);
1420   if (OrigTy->isPointerTy())
1421     Call = Builder.CreateIntToPtr(Call, OrigTy);
1422   return RValue::get(Call);
1423 }
1424 
1425 
1426 /// EmitStoreThroughLValue - Store the specified rvalue into the specified
1427 /// lvalue, where both are guaranteed to the have the same type, and that type
1428 /// is 'Ty'.
1429 void CodeGenFunction::EmitStoreThroughLValue(RValue Src, LValue Dst,
1430                                              bool isInit) {
1431   if (!Dst.isSimple()) {
1432     if (Dst.isVectorElt()) {
1433       // Read/modify/write the vector, inserting the new element.
1434       llvm::LoadInst *Load = Builder.CreateLoad(Dst.getVectorAddr(),
1435                                                 Dst.isVolatileQualified());
1436       Load->setAlignment(Dst.getAlignment().getQuantity());
1437       llvm::Value *Vec = Load;
1438       Vec = Builder.CreateInsertElement(Vec, Src.getScalarVal(),
1439                                         Dst.getVectorIdx(), "vecins");
1440       llvm::StoreInst *Store = Builder.CreateStore(Vec, Dst.getVectorAddr(),
1441                                                    Dst.isVolatileQualified());
1442       Store->setAlignment(Dst.getAlignment().getQuantity());
1443       return;
1444     }
1445 
1446     // If this is an update of extended vector elements, insert them as
1447     // appropriate.
1448     if (Dst.isExtVectorElt())
1449       return EmitStoreThroughExtVectorComponentLValue(Src, Dst);
1450 
1451     if (Dst.isGlobalReg())
1452       return EmitStoreThroughGlobalRegLValue(Src, Dst);
1453 
1454     assert(Dst.isBitField() && "Unknown LValue type");
1455     return EmitStoreThroughBitfieldLValue(Src, Dst);
1456   }
1457 
1458   // There's special magic for assigning into an ARC-qualified l-value.
1459   if (Qualifiers::ObjCLifetime Lifetime = Dst.getQuals().getObjCLifetime()) {
1460     switch (Lifetime) {
1461     case Qualifiers::OCL_None:
1462       llvm_unreachable("present but none");
1463 
1464     case Qualifiers::OCL_ExplicitNone:
1465       // nothing special
1466       break;
1467 
1468     case Qualifiers::OCL_Strong:
1469       EmitARCStoreStrong(Dst, Src.getScalarVal(), /*ignore*/ true);
1470       return;
1471 
1472     case Qualifiers::OCL_Weak:
1473       EmitARCStoreWeak(Dst.getAddress(), Src.getScalarVal(), /*ignore*/ true);
1474       return;
1475 
1476     case Qualifiers::OCL_Autoreleasing:
1477       Src = RValue::get(EmitObjCExtendObjectLifetime(Dst.getType(),
1478                                                      Src.getScalarVal()));
1479       // fall into the normal path
1480       break;
1481     }
1482   }
1483 
1484   if (Dst.isObjCWeak() && !Dst.isNonGC()) {
1485     // load of a __weak object.
1486     llvm::Value *LvalueDst = Dst.getAddress();
1487     llvm::Value *src = Src.getScalarVal();
1488      CGM.getObjCRuntime().EmitObjCWeakAssign(*this, src, LvalueDst);
1489     return;
1490   }
1491 
1492   if (Dst.isObjCStrong() && !Dst.isNonGC()) {
1493     // load of a __strong object.
1494     llvm::Value *LvalueDst = Dst.getAddress();
1495     llvm::Value *src = Src.getScalarVal();
1496     if (Dst.isObjCIvar()) {
1497       assert(Dst.getBaseIvarExp() && "BaseIvarExp is NULL");
1498       llvm::Type *ResultType = ConvertType(getContext().LongTy);
1499       llvm::Value *RHS = EmitScalarExpr(Dst.getBaseIvarExp());
1500       llvm::Value *dst = RHS;
1501       RHS = Builder.CreatePtrToInt(RHS, ResultType, "sub.ptr.rhs.cast");
1502       llvm::Value *LHS =
1503         Builder.CreatePtrToInt(LvalueDst, ResultType, "sub.ptr.lhs.cast");
1504       llvm::Value *BytesBetween = Builder.CreateSub(LHS, RHS, "ivar.offset");
1505       CGM.getObjCRuntime().EmitObjCIvarAssign(*this, src, dst,
1506                                               BytesBetween);
1507     } else if (Dst.isGlobalObjCRef()) {
1508       CGM.getObjCRuntime().EmitObjCGlobalAssign(*this, src, LvalueDst,
1509                                                 Dst.isThreadLocalRef());
1510     }
1511     else
1512       CGM.getObjCRuntime().EmitObjCStrongCastAssign(*this, src, LvalueDst);
1513     return;
1514   }
1515 
1516   assert(Src.isScalar() && "Can't emit an agg store with this method");
1517   EmitStoreOfScalar(Src.getScalarVal(), Dst, isInit);
1518 }
1519 
1520 void CodeGenFunction::EmitStoreThroughBitfieldLValue(RValue Src, LValue Dst,
1521                                                      llvm::Value **Result) {
1522   const CGBitFieldInfo &Info = Dst.getBitFieldInfo();
1523   llvm::Type *ResLTy = ConvertTypeForMem(Dst.getType());
1524   llvm::Value *Ptr = Dst.getBitFieldAddr();
1525 
1526   // Get the source value, truncated to the width of the bit-field.
1527   llvm::Value *SrcVal = Src.getScalarVal();
1528 
1529   // Cast the source to the storage type and shift it into place.
1530   SrcVal = Builder.CreateIntCast(SrcVal,
1531                                  Ptr->getType()->getPointerElementType(),
1532                                  /*IsSigned=*/false);
1533   llvm::Value *MaskedVal = SrcVal;
1534 
1535   // See if there are other bits in the bitfield's storage we'll need to load
1536   // and mask together with source before storing.
1537   if (Info.StorageSize != Info.Size) {
1538     assert(Info.StorageSize > Info.Size && "Invalid bitfield size.");
1539     llvm::Value *Val = Builder.CreateLoad(Ptr, Dst.isVolatileQualified(),
1540                                           "bf.load");
1541     cast<llvm::LoadInst>(Val)->setAlignment(Info.StorageAlignment);
1542 
1543     // Mask the source value as needed.
1544     if (!hasBooleanRepresentation(Dst.getType()))
1545       SrcVal = Builder.CreateAnd(SrcVal,
1546                                  llvm::APInt::getLowBitsSet(Info.StorageSize,
1547                                                             Info.Size),
1548                                  "bf.value");
1549     MaskedVal = SrcVal;
1550     if (Info.Offset)
1551       SrcVal = Builder.CreateShl(SrcVal, Info.Offset, "bf.shl");
1552 
1553     // Mask out the original value.
1554     Val = Builder.CreateAnd(Val,
1555                             ~llvm::APInt::getBitsSet(Info.StorageSize,
1556                                                      Info.Offset,
1557                                                      Info.Offset + Info.Size),
1558                             "bf.clear");
1559 
1560     // Or together the unchanged values and the source value.
1561     SrcVal = Builder.CreateOr(Val, SrcVal, "bf.set");
1562   } else {
1563     assert(Info.Offset == 0);
1564   }
1565 
1566   // Write the new value back out.
1567   llvm::StoreInst *Store = Builder.CreateStore(SrcVal, Ptr,
1568                                                Dst.isVolatileQualified());
1569   Store->setAlignment(Info.StorageAlignment);
1570 
1571   // Return the new value of the bit-field, if requested.
1572   if (Result) {
1573     llvm::Value *ResultVal = MaskedVal;
1574 
1575     // Sign extend the value if needed.
1576     if (Info.IsSigned) {
1577       assert(Info.Size <= Info.StorageSize);
1578       unsigned HighBits = Info.StorageSize - Info.Size;
1579       if (HighBits) {
1580         ResultVal = Builder.CreateShl(ResultVal, HighBits, "bf.result.shl");
1581         ResultVal = Builder.CreateAShr(ResultVal, HighBits, "bf.result.ashr");
1582       }
1583     }
1584 
1585     ResultVal = Builder.CreateIntCast(ResultVal, ResLTy, Info.IsSigned,
1586                                       "bf.result.cast");
1587     *Result = EmitFromMemory(ResultVal, Dst.getType());
1588   }
1589 }
1590 
1591 void CodeGenFunction::EmitStoreThroughExtVectorComponentLValue(RValue Src,
1592                                                                LValue Dst) {
1593   // This access turns into a read/modify/write of the vector.  Load the input
1594   // value now.
1595   llvm::LoadInst *Load = Builder.CreateLoad(Dst.getExtVectorAddr(),
1596                                             Dst.isVolatileQualified());
1597   Load->setAlignment(Dst.getAlignment().getQuantity());
1598   llvm::Value *Vec = Load;
1599   const llvm::Constant *Elts = Dst.getExtVectorElts();
1600 
1601   llvm::Value *SrcVal = Src.getScalarVal();
1602 
1603   if (const VectorType *VTy = Dst.getType()->getAs<VectorType>()) {
1604     unsigned NumSrcElts = VTy->getNumElements();
1605     unsigned NumDstElts =
1606        cast<llvm::VectorType>(Vec->getType())->getNumElements();
1607     if (NumDstElts == NumSrcElts) {
1608       // Use shuffle vector is the src and destination are the same number of
1609       // elements and restore the vector mask since it is on the side it will be
1610       // stored.
1611       SmallVector<llvm::Constant*, 4> Mask(NumDstElts);
1612       for (unsigned i = 0; i != NumSrcElts; ++i)
1613         Mask[getAccessedFieldNo(i, Elts)] = Builder.getInt32(i);
1614 
1615       llvm::Value *MaskV = llvm::ConstantVector::get(Mask);
1616       Vec = Builder.CreateShuffleVector(SrcVal,
1617                                         llvm::UndefValue::get(Vec->getType()),
1618                                         MaskV);
1619     } else if (NumDstElts > NumSrcElts) {
1620       // Extended the source vector to the same length and then shuffle it
1621       // into the destination.
1622       // FIXME: since we're shuffling with undef, can we just use the indices
1623       //        into that?  This could be simpler.
1624       SmallVector<llvm::Constant*, 4> ExtMask;
1625       for (unsigned i = 0; i != NumSrcElts; ++i)
1626         ExtMask.push_back(Builder.getInt32(i));
1627       ExtMask.resize(NumDstElts, llvm::UndefValue::get(Int32Ty));
1628       llvm::Value *ExtMaskV = llvm::ConstantVector::get(ExtMask);
1629       llvm::Value *ExtSrcVal =
1630         Builder.CreateShuffleVector(SrcVal,
1631                                     llvm::UndefValue::get(SrcVal->getType()),
1632                                     ExtMaskV);
1633       // build identity
1634       SmallVector<llvm::Constant*, 4> Mask;
1635       for (unsigned i = 0; i != NumDstElts; ++i)
1636         Mask.push_back(Builder.getInt32(i));
1637 
1638       // When the vector size is odd and .odd or .hi is used, the last element
1639       // of the Elts constant array will be one past the size of the vector.
1640       // Ignore the last element here, if it is greater than the mask size.
1641       if (getAccessedFieldNo(NumSrcElts - 1, Elts) == Mask.size())
1642         NumSrcElts--;
1643 
1644       // modify when what gets shuffled in
1645       for (unsigned i = 0; i != NumSrcElts; ++i)
1646         Mask[getAccessedFieldNo(i, Elts)] = Builder.getInt32(i+NumDstElts);
1647       llvm::Value *MaskV = llvm::ConstantVector::get(Mask);
1648       Vec = Builder.CreateShuffleVector(Vec, ExtSrcVal, MaskV);
1649     } else {
1650       // We should never shorten the vector
1651       llvm_unreachable("unexpected shorten vector length");
1652     }
1653   } else {
1654     // If the Src is a scalar (not a vector) it must be updating one element.
1655     unsigned InIdx = getAccessedFieldNo(0, Elts);
1656     llvm::Value *Elt = llvm::ConstantInt::get(SizeTy, InIdx);
1657     Vec = Builder.CreateInsertElement(Vec, SrcVal, Elt);
1658   }
1659 
1660   llvm::StoreInst *Store = Builder.CreateStore(Vec, Dst.getExtVectorAddr(),
1661                                                Dst.isVolatileQualified());
1662   Store->setAlignment(Dst.getAlignment().getQuantity());
1663 }
1664 
1665 /// @brief Store of global named registers are always calls to intrinsics.
1666 void CodeGenFunction::EmitStoreThroughGlobalRegLValue(RValue Src, LValue Dst) {
1667   assert((Dst.getType()->isIntegerType() || Dst.getType()->isPointerType()) &&
1668          "Bad type for register variable");
1669   llvm::MDNode *RegName = dyn_cast<llvm::MDNode>(Dst.getGlobalReg());
1670   assert(RegName && "Register LValue is not metadata");
1671 
1672   // We accept integer and pointer types only
1673   llvm::Type *OrigTy = CGM.getTypes().ConvertType(Dst.getType());
1674   llvm::Type *Ty = OrigTy;
1675   if (OrigTy->isPointerTy())
1676     Ty = CGM.getTypes().getDataLayout().getIntPtrType(OrigTy);
1677   llvm::Type *Types[] = { Ty };
1678 
1679   llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::write_register, Types);
1680   llvm::Value *Value = Src.getScalarVal();
1681   if (OrigTy->isPointerTy())
1682     Value = Builder.CreatePtrToInt(Value, Ty);
1683   Builder.CreateCall2(F, RegName, Value);
1684 }
1685 
1686 // setObjCGCLValueClass - sets class of the lvalue for the purpose of
1687 // generating write-barries API. It is currently a global, ivar,
1688 // or neither.
1689 static void setObjCGCLValueClass(const ASTContext &Ctx, const Expr *E,
1690                                  LValue &LV,
1691                                  bool IsMemberAccess=false) {
1692   if (Ctx.getLangOpts().getGC() == LangOptions::NonGC)
1693     return;
1694 
1695   if (isa<ObjCIvarRefExpr>(E)) {
1696     QualType ExpTy = E->getType();
1697     if (IsMemberAccess && ExpTy->isPointerType()) {
1698       // If ivar is a structure pointer, assigning to field of
1699       // this struct follows gcc's behavior and makes it a non-ivar
1700       // writer-barrier conservatively.
1701       ExpTy = ExpTy->getAs<PointerType>()->getPointeeType();
1702       if (ExpTy->isRecordType()) {
1703         LV.setObjCIvar(false);
1704         return;
1705       }
1706     }
1707     LV.setObjCIvar(true);
1708     auto *Exp = cast<ObjCIvarRefExpr>(const_cast<Expr *>(E));
1709     LV.setBaseIvarExp(Exp->getBase());
1710     LV.setObjCArray(E->getType()->isArrayType());
1711     return;
1712   }
1713 
1714   if (const auto *Exp = dyn_cast<DeclRefExpr>(E)) {
1715     if (const auto *VD = dyn_cast<VarDecl>(Exp->getDecl())) {
1716       if (VD->hasGlobalStorage()) {
1717         LV.setGlobalObjCRef(true);
1718         LV.setThreadLocalRef(VD->getTLSKind() != VarDecl::TLS_None);
1719       }
1720     }
1721     LV.setObjCArray(E->getType()->isArrayType());
1722     return;
1723   }
1724 
1725   if (const auto *Exp = dyn_cast<UnaryOperator>(E)) {
1726     setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess);
1727     return;
1728   }
1729 
1730   if (const auto *Exp = dyn_cast<ParenExpr>(E)) {
1731     setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess);
1732     if (LV.isObjCIvar()) {
1733       // If cast is to a structure pointer, follow gcc's behavior and make it
1734       // a non-ivar write-barrier.
1735       QualType ExpTy = E->getType();
1736       if (ExpTy->isPointerType())
1737         ExpTy = ExpTy->getAs<PointerType>()->getPointeeType();
1738       if (ExpTy->isRecordType())
1739         LV.setObjCIvar(false);
1740     }
1741     return;
1742   }
1743 
1744   if (const auto *Exp = dyn_cast<GenericSelectionExpr>(E)) {
1745     setObjCGCLValueClass(Ctx, Exp->getResultExpr(), LV);
1746     return;
1747   }
1748 
1749   if (const auto *Exp = dyn_cast<ImplicitCastExpr>(E)) {
1750     setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess);
1751     return;
1752   }
1753 
1754   if (const auto *Exp = dyn_cast<CStyleCastExpr>(E)) {
1755     setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess);
1756     return;
1757   }
1758 
1759   if (const auto *Exp = dyn_cast<ObjCBridgedCastExpr>(E)) {
1760     setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess);
1761     return;
1762   }
1763 
1764   if (const auto *Exp = dyn_cast<ArraySubscriptExpr>(E)) {
1765     setObjCGCLValueClass(Ctx, Exp->getBase(), LV);
1766     if (LV.isObjCIvar() && !LV.isObjCArray())
1767       // Using array syntax to assigning to what an ivar points to is not
1768       // same as assigning to the ivar itself. {id *Names;} Names[i] = 0;
1769       LV.setObjCIvar(false);
1770     else if (LV.isGlobalObjCRef() && !LV.isObjCArray())
1771       // Using array syntax to assigning to what global points to is not
1772       // same as assigning to the global itself. {id *G;} G[i] = 0;
1773       LV.setGlobalObjCRef(false);
1774     return;
1775   }
1776 
1777   if (const auto *Exp = dyn_cast<MemberExpr>(E)) {
1778     setObjCGCLValueClass(Ctx, Exp->getBase(), LV, true);
1779     // We don't know if member is an 'ivar', but this flag is looked at
1780     // only in the context of LV.isObjCIvar().
1781     LV.setObjCArray(E->getType()->isArrayType());
1782     return;
1783   }
1784 }
1785 
1786 static llvm::Value *
1787 EmitBitCastOfLValueToProperType(CodeGenFunction &CGF,
1788                                 llvm::Value *V, llvm::Type *IRType,
1789                                 StringRef Name = StringRef()) {
1790   unsigned AS = cast<llvm::PointerType>(V->getType())->getAddressSpace();
1791   return CGF.Builder.CreateBitCast(V, IRType->getPointerTo(AS), Name);
1792 }
1793 
1794 static LValue EmitGlobalVarDeclLValue(CodeGenFunction &CGF,
1795                                       const Expr *E, const VarDecl *VD) {
1796   QualType T = E->getType();
1797 
1798   // If it's thread_local, emit a call to its wrapper function instead.
1799   if (VD->getTLSKind() == VarDecl::TLS_Dynamic &&
1800       CGF.CGM.getCXXABI().usesThreadWrapperFunction())
1801     return CGF.CGM.getCXXABI().EmitThreadLocalVarDeclLValue(CGF, VD, T);
1802 
1803   llvm::Value *V = CGF.CGM.GetAddrOfGlobalVar(VD);
1804   llvm::Type *RealVarTy = CGF.getTypes().ConvertTypeForMem(VD->getType());
1805   V = EmitBitCastOfLValueToProperType(CGF, V, RealVarTy);
1806   CharUnits Alignment = CGF.getContext().getDeclAlign(VD);
1807   LValue LV;
1808   if (VD->getType()->isReferenceType()) {
1809     llvm::LoadInst *LI = CGF.Builder.CreateLoad(V);
1810     LI->setAlignment(Alignment.getQuantity());
1811     V = LI;
1812     LV = CGF.MakeNaturalAlignAddrLValue(V, T);
1813   } else {
1814     LV = CGF.MakeAddrLValue(V, T, Alignment);
1815   }
1816   setObjCGCLValueClass(CGF.getContext(), E, LV);
1817   return LV;
1818 }
1819 
1820 static LValue EmitFunctionDeclLValue(CodeGenFunction &CGF,
1821                                      const Expr *E, const FunctionDecl *FD) {
1822   llvm::Value *V = CGF.CGM.GetAddrOfFunction(FD);
1823   if (!FD->hasPrototype()) {
1824     if (const FunctionProtoType *Proto =
1825             FD->getType()->getAs<FunctionProtoType>()) {
1826       // Ugly case: for a K&R-style definition, the type of the definition
1827       // isn't the same as the type of a use.  Correct for this with a
1828       // bitcast.
1829       QualType NoProtoType =
1830           CGF.getContext().getFunctionNoProtoType(Proto->getReturnType());
1831       NoProtoType = CGF.getContext().getPointerType(NoProtoType);
1832       V = CGF.Builder.CreateBitCast(V, CGF.ConvertType(NoProtoType));
1833     }
1834   }
1835   CharUnits Alignment = CGF.getContext().getDeclAlign(FD);
1836   return CGF.MakeAddrLValue(V, E->getType(), Alignment);
1837 }
1838 
1839 static LValue EmitCapturedFieldLValue(CodeGenFunction &CGF, const FieldDecl *FD,
1840                                       llvm::Value *ThisValue) {
1841   QualType TagType = CGF.getContext().getTagDeclType(FD->getParent());
1842   LValue LV = CGF.MakeNaturalAlignAddrLValue(ThisValue, TagType);
1843   return CGF.EmitLValueForField(LV, FD);
1844 }
1845 
1846 /// Named Registers are named metadata pointing to the register name
1847 /// which will be read from/written to as an argument to the intrinsic
1848 /// @llvm.read/write_register.
1849 /// So far, only the name is being passed down, but other options such as
1850 /// register type, allocation type or even optimization options could be
1851 /// passed down via the metadata node.
1852 static LValue EmitGlobalNamedRegister(const VarDecl *VD,
1853                                       CodeGenModule &CGM,
1854                                       CharUnits Alignment) {
1855   SmallString<64> Name("llvm.named.register.");
1856   AsmLabelAttr *Asm = VD->getAttr<AsmLabelAttr>();
1857   assert(Asm->getLabel().size() < 64-Name.size() &&
1858       "Register name too big");
1859   Name.append(Asm->getLabel());
1860   llvm::NamedMDNode *M =
1861     CGM.getModule().getOrInsertNamedMetadata(Name);
1862   if (M->getNumOperands() == 0) {
1863     llvm::MDString *Str = llvm::MDString::get(CGM.getLLVMContext(),
1864                                               Asm->getLabel());
1865     llvm::Value *Ops[] = { Str };
1866     M->addOperand(llvm::MDNode::get(CGM.getLLVMContext(), Ops));
1867   }
1868   return LValue::MakeGlobalReg(M->getOperand(0), VD->getType(), Alignment);
1869 }
1870 
1871 LValue CodeGenFunction::EmitDeclRefLValue(const DeclRefExpr *E) {
1872   const NamedDecl *ND = E->getDecl();
1873   CharUnits Alignment = getContext().getDeclAlign(ND);
1874   QualType T = E->getType();
1875 
1876   if (const auto *VD = dyn_cast<VarDecl>(ND)) {
1877     // Global Named registers access via intrinsics only
1878     if (VD->getStorageClass() == SC_Register &&
1879         VD->hasAttr<AsmLabelAttr>() && !VD->isLocalVarDecl())
1880       return EmitGlobalNamedRegister(VD, CGM, Alignment);
1881 
1882     // A DeclRefExpr for a reference initialized by a constant expression can
1883     // appear without being odr-used. Directly emit the constant initializer.
1884     const Expr *Init = VD->getAnyInitializer(VD);
1885     if (Init && !isa<ParmVarDecl>(VD) && VD->getType()->isReferenceType() &&
1886         VD->isUsableInConstantExpressions(getContext()) &&
1887         VD->checkInitIsICE()) {
1888       llvm::Constant *Val =
1889         CGM.EmitConstantValue(*VD->evaluateValue(), VD->getType(), this);
1890       assert(Val && "failed to emit reference constant expression");
1891       // FIXME: Eventually we will want to emit vector element references.
1892       return MakeAddrLValue(Val, T, Alignment);
1893     }
1894   }
1895 
1896   // FIXME: We should be able to assert this for FunctionDecls as well!
1897   // FIXME: We should be able to assert this for all DeclRefExprs, not just
1898   // those with a valid source location.
1899   assert((ND->isUsed(false) || !isa<VarDecl>(ND) ||
1900           !E->getLocation().isValid()) &&
1901          "Should not use decl without marking it used!");
1902 
1903   if (ND->hasAttr<WeakRefAttr>()) {
1904     const auto *VD = cast<ValueDecl>(ND);
1905     llvm::Constant *Aliasee = CGM.GetWeakRefReference(VD);
1906     return MakeAddrLValue(Aliasee, T, Alignment);
1907   }
1908 
1909   if (const auto *VD = dyn_cast<VarDecl>(ND)) {
1910     // Check if this is a global variable.
1911     if (VD->hasLinkage() || VD->isStaticDataMember())
1912       return EmitGlobalVarDeclLValue(*this, E, VD);
1913 
1914     bool isBlockVariable = VD->hasAttr<BlocksAttr>();
1915 
1916     llvm::Value *V = LocalDeclMap.lookup(VD);
1917     if (!V && VD->isStaticLocal())
1918       V = CGM.getOrCreateStaticVarDecl(
1919           *VD, CGM.getLLVMLinkageVarDefinition(VD, /*isConstant=*/false));
1920 
1921     // Use special handling for lambdas.
1922     if (!V) {
1923       if (FieldDecl *FD = LambdaCaptureFields.lookup(VD)) {
1924         return EmitCapturedFieldLValue(*this, FD, CXXABIThisValue);
1925       } else if (CapturedStmtInfo) {
1926         if (const FieldDecl *FD = CapturedStmtInfo->lookup(VD))
1927           return EmitCapturedFieldLValue(*this, FD,
1928                                          CapturedStmtInfo->getContextValue());
1929       }
1930 
1931       assert(isa<BlockDecl>(CurCodeDecl) && E->refersToEnclosingLocal());
1932       return MakeAddrLValue(GetAddrOfBlockDecl(VD, isBlockVariable),
1933                             T, Alignment);
1934     }
1935 
1936     assert(V && "DeclRefExpr not entered in LocalDeclMap?");
1937 
1938     if (isBlockVariable)
1939       V = BuildBlockByrefAddress(V, VD);
1940 
1941     LValue LV;
1942     if (VD->getType()->isReferenceType()) {
1943       llvm::LoadInst *LI = Builder.CreateLoad(V);
1944       LI->setAlignment(Alignment.getQuantity());
1945       V = LI;
1946       LV = MakeNaturalAlignAddrLValue(V, T);
1947     } else {
1948       LV = MakeAddrLValue(V, T, Alignment);
1949     }
1950 
1951     bool isLocalStorage = VD->hasLocalStorage();
1952 
1953     bool NonGCable = isLocalStorage &&
1954                      !VD->getType()->isReferenceType() &&
1955                      !isBlockVariable;
1956     if (NonGCable) {
1957       LV.getQuals().removeObjCGCAttr();
1958       LV.setNonGC(true);
1959     }
1960 
1961     bool isImpreciseLifetime =
1962       (isLocalStorage && !VD->hasAttr<ObjCPreciseLifetimeAttr>());
1963     if (isImpreciseLifetime)
1964       LV.setARCPreciseLifetime(ARCImpreciseLifetime);
1965     setObjCGCLValueClass(getContext(), E, LV);
1966     return LV;
1967   }
1968 
1969   if (const auto *FD = dyn_cast<FunctionDecl>(ND))
1970     return EmitFunctionDeclLValue(*this, E, FD);
1971 
1972   llvm_unreachable("Unhandled DeclRefExpr");
1973 }
1974 
1975 LValue CodeGenFunction::EmitUnaryOpLValue(const UnaryOperator *E) {
1976   // __extension__ doesn't affect lvalue-ness.
1977   if (E->getOpcode() == UO_Extension)
1978     return EmitLValue(E->getSubExpr());
1979 
1980   QualType ExprTy = getContext().getCanonicalType(E->getSubExpr()->getType());
1981   switch (E->getOpcode()) {
1982   default: llvm_unreachable("Unknown unary operator lvalue!");
1983   case UO_Deref: {
1984     QualType T = E->getSubExpr()->getType()->getPointeeType();
1985     assert(!T.isNull() && "CodeGenFunction::EmitUnaryOpLValue: Illegal type");
1986 
1987     LValue LV = MakeNaturalAlignAddrLValue(EmitScalarExpr(E->getSubExpr()), T);
1988     LV.getQuals().setAddressSpace(ExprTy.getAddressSpace());
1989 
1990     // We should not generate __weak write barrier on indirect reference
1991     // of a pointer to object; as in void foo (__weak id *param); *param = 0;
1992     // But, we continue to generate __strong write barrier on indirect write
1993     // into a pointer to object.
1994     if (getLangOpts().ObjC1 &&
1995         getLangOpts().getGC() != LangOptions::NonGC &&
1996         LV.isObjCWeak())
1997       LV.setNonGC(!E->isOBJCGCCandidate(getContext()));
1998     return LV;
1999   }
2000   case UO_Real:
2001   case UO_Imag: {
2002     LValue LV = EmitLValue(E->getSubExpr());
2003     assert(LV.isSimple() && "real/imag on non-ordinary l-value");
2004     llvm::Value *Addr = LV.getAddress();
2005 
2006     // __real is valid on scalars.  This is a faster way of testing that.
2007     // __imag can only produce an rvalue on scalars.
2008     if (E->getOpcode() == UO_Real &&
2009         !cast<llvm::PointerType>(Addr->getType())
2010            ->getElementType()->isStructTy()) {
2011       assert(E->getSubExpr()->getType()->isArithmeticType());
2012       return LV;
2013     }
2014 
2015     assert(E->getSubExpr()->getType()->isAnyComplexType());
2016 
2017     unsigned Idx = E->getOpcode() == UO_Imag;
2018     return MakeAddrLValue(Builder.CreateStructGEP(LV.getAddress(),
2019                                                   Idx, "idx"),
2020                           ExprTy);
2021   }
2022   case UO_PreInc:
2023   case UO_PreDec: {
2024     LValue LV = EmitLValue(E->getSubExpr());
2025     bool isInc = E->getOpcode() == UO_PreInc;
2026 
2027     if (E->getType()->isAnyComplexType())
2028       EmitComplexPrePostIncDec(E, LV, isInc, true/*isPre*/);
2029     else
2030       EmitScalarPrePostIncDec(E, LV, isInc, true/*isPre*/);
2031     return LV;
2032   }
2033   }
2034 }
2035 
2036 LValue CodeGenFunction::EmitStringLiteralLValue(const StringLiteral *E) {
2037   return MakeAddrLValue(CGM.GetAddrOfConstantStringFromLiteral(E),
2038                         E->getType());
2039 }
2040 
2041 LValue CodeGenFunction::EmitObjCEncodeExprLValue(const ObjCEncodeExpr *E) {
2042   return MakeAddrLValue(CGM.GetAddrOfConstantStringFromObjCEncode(E),
2043                         E->getType());
2044 }
2045 
2046 LValue CodeGenFunction::EmitPredefinedLValue(const PredefinedExpr *E) {
2047   auto SL = E->getFunctionName();
2048   assert(SL != nullptr && "No StringLiteral name in PredefinedExpr");
2049   StringRef FnName = CurFn->getName();
2050   if (FnName.startswith("\01"))
2051     FnName = FnName.substr(1);
2052   StringRef NameItems[] = {
2053       PredefinedExpr::getIdentTypeName(E->getIdentType()), FnName};
2054   std::string GVName = llvm::join(NameItems, NameItems + 2, ".");
2055 
2056   auto C = CGM.GetAddrOfConstantStringFromLiteral(SL, GVName);
2057   return MakeAddrLValue(C, E->getType());
2058 }
2059 
2060 /// Emit a type description suitable for use by a runtime sanitizer library. The
2061 /// format of a type descriptor is
2062 ///
2063 /// \code
2064 ///   { i16 TypeKind, i16 TypeInfo }
2065 /// \endcode
2066 ///
2067 /// followed by an array of i8 containing the type name. TypeKind is 0 for an
2068 /// integer, 1 for a floating point value, and -1 for anything else.
2069 llvm::Constant *CodeGenFunction::EmitCheckTypeDescriptor(QualType T) {
2070   // Only emit each type's descriptor once.
2071   if (llvm::Constant *C = CGM.getTypeDescriptorFromMap(T))
2072     return C;
2073 
2074   uint16_t TypeKind = -1;
2075   uint16_t TypeInfo = 0;
2076 
2077   if (T->isIntegerType()) {
2078     TypeKind = 0;
2079     TypeInfo = (llvm::Log2_32(getContext().getTypeSize(T)) << 1) |
2080                (T->isSignedIntegerType() ? 1 : 0);
2081   } else if (T->isFloatingType()) {
2082     TypeKind = 1;
2083     TypeInfo = getContext().getTypeSize(T);
2084   }
2085 
2086   // Format the type name as if for a diagnostic, including quotes and
2087   // optionally an 'aka'.
2088   SmallString<32> Buffer;
2089   CGM.getDiags().ConvertArgToString(DiagnosticsEngine::ak_qualtype,
2090                                     (intptr_t)T.getAsOpaquePtr(),
2091                                     StringRef(), StringRef(), None, Buffer,
2092                                     None);
2093 
2094   llvm::Constant *Components[] = {
2095     Builder.getInt16(TypeKind), Builder.getInt16(TypeInfo),
2096     llvm::ConstantDataArray::getString(getLLVMContext(), Buffer)
2097   };
2098   llvm::Constant *Descriptor = llvm::ConstantStruct::getAnon(Components);
2099 
2100   auto *GV = new llvm::GlobalVariable(
2101       CGM.getModule(), Descriptor->getType(),
2102       /*isConstant=*/true, llvm::GlobalVariable::PrivateLinkage, Descriptor);
2103   GV->setUnnamedAddr(true);
2104   CGM.getSanitizerMetadata()->disableSanitizerForGlobal(GV);
2105 
2106   // Remember the descriptor for this type.
2107   CGM.setTypeDescriptorInMap(T, GV);
2108 
2109   return GV;
2110 }
2111 
2112 llvm::Value *CodeGenFunction::EmitCheckValue(llvm::Value *V) {
2113   llvm::Type *TargetTy = IntPtrTy;
2114 
2115   // Floating-point types which fit into intptr_t are bitcast to integers
2116   // and then passed directly (after zero-extension, if necessary).
2117   if (V->getType()->isFloatingPointTy()) {
2118     unsigned Bits = V->getType()->getPrimitiveSizeInBits();
2119     if (Bits <= TargetTy->getIntegerBitWidth())
2120       V = Builder.CreateBitCast(V, llvm::Type::getIntNTy(getLLVMContext(),
2121                                                          Bits));
2122   }
2123 
2124   // Integers which fit in intptr_t are zero-extended and passed directly.
2125   if (V->getType()->isIntegerTy() &&
2126       V->getType()->getIntegerBitWidth() <= TargetTy->getIntegerBitWidth())
2127     return Builder.CreateZExt(V, TargetTy);
2128 
2129   // Pointers are passed directly, everything else is passed by address.
2130   if (!V->getType()->isPointerTy()) {
2131     llvm::Value *Ptr = CreateTempAlloca(V->getType());
2132     Builder.CreateStore(V, Ptr);
2133     V = Ptr;
2134   }
2135   return Builder.CreatePtrToInt(V, TargetTy);
2136 }
2137 
2138 /// \brief Emit a representation of a SourceLocation for passing to a handler
2139 /// in a sanitizer runtime library. The format for this data is:
2140 /// \code
2141 ///   struct SourceLocation {
2142 ///     const char *Filename;
2143 ///     int32_t Line, Column;
2144 ///   };
2145 /// \endcode
2146 /// For an invalid SourceLocation, the Filename pointer is null.
2147 llvm::Constant *CodeGenFunction::EmitCheckSourceLocation(SourceLocation Loc) {
2148   llvm::Constant *Filename;
2149   int Line, Column;
2150 
2151   PresumedLoc PLoc = getContext().getSourceManager().getPresumedLoc(Loc);
2152   if (PLoc.isValid()) {
2153     auto FilenameGV = CGM.GetAddrOfConstantCString(PLoc.getFilename(), ".src");
2154     CGM.getSanitizerMetadata()->disableSanitizerForGlobal(FilenameGV);
2155     Filename = FilenameGV;
2156     Line = PLoc.getLine();
2157     Column = PLoc.getColumn();
2158   } else {
2159     Filename = llvm::Constant::getNullValue(Int8PtrTy);
2160     Line = Column = 0;
2161   }
2162 
2163   llvm::Constant *Data[] = {Filename, Builder.getInt32(Line),
2164                             Builder.getInt32(Column)};
2165 
2166   return llvm::ConstantStruct::getAnon(Data);
2167 }
2168 
2169 void CodeGenFunction::EmitCheck(llvm::Value *Checked, StringRef CheckName,
2170                                 ArrayRef<llvm::Constant *> StaticArgs,
2171                                 ArrayRef<llvm::Value *> DynamicArgs,
2172                                 CheckRecoverableKind RecoverKind) {
2173   assert(IsSanitizerScope);
2174 
2175   if (CGM.getCodeGenOpts().SanitizeUndefinedTrapOnError) {
2176     assert (RecoverKind != CRK_AlwaysRecoverable &&
2177             "Runtime call required for AlwaysRecoverable kind!");
2178     return EmitTrapCheck(Checked);
2179   }
2180 
2181   llvm::BasicBlock *Cont = createBasicBlock("cont");
2182 
2183   llvm::BasicBlock *Handler = createBasicBlock("handler." + CheckName);
2184 
2185   llvm::Instruction *Branch = Builder.CreateCondBr(Checked, Cont, Handler);
2186 
2187   // Give hint that we very much don't expect to execute the handler
2188   // Value chosen to match UR_NONTAKEN_WEIGHT, see BranchProbabilityInfo.cpp
2189   llvm::MDBuilder MDHelper(getLLVMContext());
2190   llvm::MDNode *Node = MDHelper.createBranchWeights((1U << 20) - 1, 1);
2191   Branch->setMetadata(llvm::LLVMContext::MD_prof, Node);
2192 
2193   EmitBlock(Handler);
2194 
2195   llvm::Constant *Info = llvm::ConstantStruct::getAnon(StaticArgs);
2196   auto *InfoPtr =
2197       new llvm::GlobalVariable(CGM.getModule(), Info->getType(), false,
2198                                llvm::GlobalVariable::PrivateLinkage, Info);
2199   InfoPtr->setUnnamedAddr(true);
2200   CGM.getSanitizerMetadata()->disableSanitizerForGlobal(InfoPtr);
2201 
2202   SmallVector<llvm::Value *, 4> Args;
2203   SmallVector<llvm::Type *, 4> ArgTypes;
2204   Args.reserve(DynamicArgs.size() + 1);
2205   ArgTypes.reserve(DynamicArgs.size() + 1);
2206 
2207   // Handler functions take an i8* pointing to the (handler-specific) static
2208   // information block, followed by a sequence of intptr_t arguments
2209   // representing operand values.
2210   Args.push_back(Builder.CreateBitCast(InfoPtr, Int8PtrTy));
2211   ArgTypes.push_back(Int8PtrTy);
2212   for (size_t i = 0, n = DynamicArgs.size(); i != n; ++i) {
2213     Args.push_back(EmitCheckValue(DynamicArgs[i]));
2214     ArgTypes.push_back(IntPtrTy);
2215   }
2216 
2217   bool Recover = RecoverKind == CRK_AlwaysRecoverable ||
2218                  (RecoverKind == CRK_Recoverable &&
2219                   CGM.getCodeGenOpts().SanitizeRecover);
2220 
2221   llvm::FunctionType *FnType =
2222     llvm::FunctionType::get(CGM.VoidTy, ArgTypes, false);
2223   llvm::AttrBuilder B;
2224   if (!Recover) {
2225     B.addAttribute(llvm::Attribute::NoReturn)
2226      .addAttribute(llvm::Attribute::NoUnwind);
2227   }
2228   B.addAttribute(llvm::Attribute::UWTable);
2229 
2230   // Checks that have two variants use a suffix to differentiate them
2231   bool NeedsAbortSuffix = RecoverKind != CRK_Unrecoverable &&
2232                           !CGM.getCodeGenOpts().SanitizeRecover;
2233   std::string FunctionName = ("__ubsan_handle_" + CheckName +
2234                               (NeedsAbortSuffix? "_abort" : "")).str();
2235   llvm::Value *Fn = CGM.CreateRuntimeFunction(
2236       FnType, FunctionName,
2237       llvm::AttributeSet::get(getLLVMContext(),
2238                               llvm::AttributeSet::FunctionIndex, B));
2239   llvm::CallInst *HandlerCall = EmitNounwindRuntimeCall(Fn, Args);
2240   if (Recover) {
2241     Builder.CreateBr(Cont);
2242   } else {
2243     HandlerCall->setDoesNotReturn();
2244     Builder.CreateUnreachable();
2245   }
2246 
2247   EmitBlock(Cont);
2248 }
2249 
2250 void CodeGenFunction::EmitTrapCheck(llvm::Value *Checked) {
2251   llvm::BasicBlock *Cont = createBasicBlock("cont");
2252 
2253   // If we're optimizing, collapse all calls to trap down to just one per
2254   // function to save on code size.
2255   if (!CGM.getCodeGenOpts().OptimizationLevel || !TrapBB) {
2256     TrapBB = createBasicBlock("trap");
2257     Builder.CreateCondBr(Checked, Cont, TrapBB);
2258     EmitBlock(TrapBB);
2259     llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::trap);
2260     llvm::CallInst *TrapCall = Builder.CreateCall(F);
2261     TrapCall->setDoesNotReturn();
2262     TrapCall->setDoesNotThrow();
2263     Builder.CreateUnreachable();
2264   } else {
2265     Builder.CreateCondBr(Checked, Cont, TrapBB);
2266   }
2267 
2268   EmitBlock(Cont);
2269 }
2270 
2271 /// isSimpleArrayDecayOperand - If the specified expr is a simple decay from an
2272 /// array to pointer, return the array subexpression.
2273 static const Expr *isSimpleArrayDecayOperand(const Expr *E) {
2274   // If this isn't just an array->pointer decay, bail out.
2275   const auto *CE = dyn_cast<CastExpr>(E);
2276   if (!CE || CE->getCastKind() != CK_ArrayToPointerDecay)
2277     return nullptr;
2278 
2279   // If this is a decay from variable width array, bail out.
2280   const Expr *SubExpr = CE->getSubExpr();
2281   if (SubExpr->getType()->isVariableArrayType())
2282     return nullptr;
2283 
2284   return SubExpr;
2285 }
2286 
2287 LValue CodeGenFunction::EmitArraySubscriptExpr(const ArraySubscriptExpr *E,
2288                                                bool Accessed) {
2289   // The index must always be an integer, which is not an aggregate.  Emit it.
2290   llvm::Value *Idx = EmitScalarExpr(E->getIdx());
2291   QualType IdxTy  = E->getIdx()->getType();
2292   bool IdxSigned = IdxTy->isSignedIntegerOrEnumerationType();
2293 
2294   if (SanOpts.ArrayBounds)
2295     EmitBoundsCheck(E, E->getBase(), Idx, IdxTy, Accessed);
2296 
2297   // If the base is a vector type, then we are forming a vector element lvalue
2298   // with this subscript.
2299   if (E->getBase()->getType()->isVectorType() &&
2300       !isa<ExtVectorElementExpr>(E->getBase())) {
2301     // Emit the vector as an lvalue to get its address.
2302     LValue LHS = EmitLValue(E->getBase());
2303     assert(LHS.isSimple() && "Can only subscript lvalue vectors here!");
2304     return LValue::MakeVectorElt(LHS.getAddress(), Idx,
2305                                  E->getBase()->getType(), LHS.getAlignment());
2306   }
2307 
2308   // Extend or truncate the index type to 32 or 64-bits.
2309   if (Idx->getType() != IntPtrTy)
2310     Idx = Builder.CreateIntCast(Idx, IntPtrTy, IdxSigned, "idxprom");
2311 
2312   // We know that the pointer points to a type of the correct size, unless the
2313   // size is a VLA or Objective-C interface.
2314   llvm::Value *Address = nullptr;
2315   CharUnits ArrayAlignment;
2316   if (isa<ExtVectorElementExpr>(E->getBase())) {
2317     LValue LV = EmitLValue(E->getBase());
2318     Address = EmitExtVectorElementLValue(LV);
2319     Address = Builder.CreateInBoundsGEP(Address, Idx, "arrayidx");
2320     const VectorType *ExprVT = LV.getType()->getAs<VectorType>();
2321     QualType EQT = ExprVT->getElementType();
2322     return MakeAddrLValue(Address, EQT,
2323                           getContext().getTypeAlignInChars(EQT));
2324   }
2325   else if (const VariableArrayType *vla =
2326            getContext().getAsVariableArrayType(E->getType())) {
2327     // The base must be a pointer, which is not an aggregate.  Emit
2328     // it.  It needs to be emitted first in case it's what captures
2329     // the VLA bounds.
2330     Address = EmitScalarExpr(E->getBase());
2331 
2332     // The element count here is the total number of non-VLA elements.
2333     llvm::Value *numElements = getVLASize(vla).first;
2334 
2335     // Effectively, the multiply by the VLA size is part of the GEP.
2336     // GEP indexes are signed, and scaling an index isn't permitted to
2337     // signed-overflow, so we use the same semantics for our explicit
2338     // multiply.  We suppress this if overflow is not undefined behavior.
2339     if (getLangOpts().isSignedOverflowDefined()) {
2340       Idx = Builder.CreateMul(Idx, numElements);
2341       Address = Builder.CreateGEP(Address, Idx, "arrayidx");
2342     } else {
2343       Idx = Builder.CreateNSWMul(Idx, numElements);
2344       Address = Builder.CreateInBoundsGEP(Address, Idx, "arrayidx");
2345     }
2346   } else if (const ObjCObjectType *OIT = E->getType()->getAs<ObjCObjectType>()){
2347     // Indexing over an interface, as in "NSString *P; P[4];"
2348     llvm::Value *InterfaceSize =
2349       llvm::ConstantInt::get(Idx->getType(),
2350           getContext().getTypeSizeInChars(OIT).getQuantity());
2351 
2352     Idx = Builder.CreateMul(Idx, InterfaceSize);
2353 
2354     // The base must be a pointer, which is not an aggregate.  Emit it.
2355     llvm::Value *Base = EmitScalarExpr(E->getBase());
2356     Address = EmitCastToVoidPtr(Base);
2357     Address = Builder.CreateGEP(Address, Idx, "arrayidx");
2358     Address = Builder.CreateBitCast(Address, Base->getType());
2359   } else if (const Expr *Array = isSimpleArrayDecayOperand(E->getBase())) {
2360     // If this is A[i] where A is an array, the frontend will have decayed the
2361     // base to be a ArrayToPointerDecay implicit cast.  While correct, it is
2362     // inefficient at -O0 to emit a "gep A, 0, 0" when codegen'ing it, then a
2363     // "gep x, i" here.  Emit one "gep A, 0, i".
2364     assert(Array->getType()->isArrayType() &&
2365            "Array to pointer decay must have array source type!");
2366     LValue ArrayLV;
2367     // For simple multidimensional array indexing, set the 'accessed' flag for
2368     // better bounds-checking of the base expression.
2369     if (const auto *ASE = dyn_cast<ArraySubscriptExpr>(Array))
2370       ArrayLV = EmitArraySubscriptExpr(ASE, /*Accessed*/ true);
2371     else
2372       ArrayLV = EmitLValue(Array);
2373     llvm::Value *ArrayPtr = ArrayLV.getAddress();
2374     llvm::Value *Zero = llvm::ConstantInt::get(Int32Ty, 0);
2375     llvm::Value *Args[] = { Zero, Idx };
2376 
2377     // Propagate the alignment from the array itself to the result.
2378     ArrayAlignment = ArrayLV.getAlignment();
2379 
2380     if (getLangOpts().isSignedOverflowDefined())
2381       Address = Builder.CreateGEP(ArrayPtr, Args, "arrayidx");
2382     else
2383       Address = Builder.CreateInBoundsGEP(ArrayPtr, Args, "arrayidx");
2384   } else {
2385     // The base must be a pointer, which is not an aggregate.  Emit it.
2386     llvm::Value *Base = EmitScalarExpr(E->getBase());
2387     if (getLangOpts().isSignedOverflowDefined())
2388       Address = Builder.CreateGEP(Base, Idx, "arrayidx");
2389     else
2390       Address = Builder.CreateInBoundsGEP(Base, Idx, "arrayidx");
2391   }
2392 
2393   QualType T = E->getBase()->getType()->getPointeeType();
2394   assert(!T.isNull() &&
2395          "CodeGenFunction::EmitArraySubscriptExpr(): Illegal base type");
2396 
2397 
2398   // Limit the alignment to that of the result type.
2399   LValue LV;
2400   if (!ArrayAlignment.isZero()) {
2401     CharUnits Align = getContext().getTypeAlignInChars(T);
2402     ArrayAlignment = std::min(Align, ArrayAlignment);
2403     LV = MakeAddrLValue(Address, T, ArrayAlignment);
2404   } else {
2405     LV = MakeNaturalAlignAddrLValue(Address, T);
2406   }
2407 
2408   LV.getQuals().setAddressSpace(E->getBase()->getType().getAddressSpace());
2409 
2410   if (getLangOpts().ObjC1 &&
2411       getLangOpts().getGC() != LangOptions::NonGC) {
2412     LV.setNonGC(!E->isOBJCGCCandidate(getContext()));
2413     setObjCGCLValueClass(getContext(), E, LV);
2414   }
2415   return LV;
2416 }
2417 
2418 static
2419 llvm::Constant *GenerateConstantVector(CGBuilderTy &Builder,
2420                                        SmallVectorImpl<unsigned> &Elts) {
2421   SmallVector<llvm::Constant*, 4> CElts;
2422   for (unsigned i = 0, e = Elts.size(); i != e; ++i)
2423     CElts.push_back(Builder.getInt32(Elts[i]));
2424 
2425   return llvm::ConstantVector::get(CElts);
2426 }
2427 
2428 LValue CodeGenFunction::
2429 EmitExtVectorElementExpr(const ExtVectorElementExpr *E) {
2430   // Emit the base vector as an l-value.
2431   LValue Base;
2432 
2433   // ExtVectorElementExpr's base can either be a vector or pointer to vector.
2434   if (E->isArrow()) {
2435     // If it is a pointer to a vector, emit the address and form an lvalue with
2436     // it.
2437     llvm::Value *Ptr = EmitScalarExpr(E->getBase());
2438     const PointerType *PT = E->getBase()->getType()->getAs<PointerType>();
2439     Base = MakeAddrLValue(Ptr, PT->getPointeeType());
2440     Base.getQuals().removeObjCGCAttr();
2441   } else if (E->getBase()->isGLValue()) {
2442     // Otherwise, if the base is an lvalue ( as in the case of foo.x.x),
2443     // emit the base as an lvalue.
2444     assert(E->getBase()->getType()->isVectorType());
2445     Base = EmitLValue(E->getBase());
2446   } else {
2447     // Otherwise, the base is a normal rvalue (as in (V+V).x), emit it as such.
2448     assert(E->getBase()->getType()->isVectorType() &&
2449            "Result must be a vector");
2450     llvm::Value *Vec = EmitScalarExpr(E->getBase());
2451 
2452     // Store the vector to memory (because LValue wants an address).
2453     llvm::Value *VecMem = CreateMemTemp(E->getBase()->getType());
2454     Builder.CreateStore(Vec, VecMem);
2455     Base = MakeAddrLValue(VecMem, E->getBase()->getType());
2456   }
2457 
2458   QualType type =
2459     E->getType().withCVRQualifiers(Base.getQuals().getCVRQualifiers());
2460 
2461   // Encode the element access list into a vector of unsigned indices.
2462   SmallVector<unsigned, 4> Indices;
2463   E->getEncodedElementAccess(Indices);
2464 
2465   if (Base.isSimple()) {
2466     llvm::Constant *CV = GenerateConstantVector(Builder, Indices);
2467     return LValue::MakeExtVectorElt(Base.getAddress(), CV, type,
2468                                     Base.getAlignment());
2469   }
2470   assert(Base.isExtVectorElt() && "Can only subscript lvalue vec elts here!");
2471 
2472   llvm::Constant *BaseElts = Base.getExtVectorElts();
2473   SmallVector<llvm::Constant *, 4> CElts;
2474 
2475   for (unsigned i = 0, e = Indices.size(); i != e; ++i)
2476     CElts.push_back(BaseElts->getAggregateElement(Indices[i]));
2477   llvm::Constant *CV = llvm::ConstantVector::get(CElts);
2478   return LValue::MakeExtVectorElt(Base.getExtVectorAddr(), CV, type,
2479                                   Base.getAlignment());
2480 }
2481 
2482 LValue CodeGenFunction::EmitMemberExpr(const MemberExpr *E) {
2483   Expr *BaseExpr = E->getBase();
2484 
2485   // If this is s.x, emit s as an lvalue.  If it is s->x, emit s as a scalar.
2486   LValue BaseLV;
2487   if (E->isArrow()) {
2488     llvm::Value *Ptr = EmitScalarExpr(BaseExpr);
2489     QualType PtrTy = BaseExpr->getType()->getPointeeType();
2490     EmitTypeCheck(TCK_MemberAccess, E->getExprLoc(), Ptr, PtrTy);
2491     BaseLV = MakeNaturalAlignAddrLValue(Ptr, PtrTy);
2492   } else
2493     BaseLV = EmitCheckedLValue(BaseExpr, TCK_MemberAccess);
2494 
2495   NamedDecl *ND = E->getMemberDecl();
2496   if (auto *Field = dyn_cast<FieldDecl>(ND)) {
2497     LValue LV = EmitLValueForField(BaseLV, Field);
2498     setObjCGCLValueClass(getContext(), E, LV);
2499     return LV;
2500   }
2501 
2502   if (auto *VD = dyn_cast<VarDecl>(ND))
2503     return EmitGlobalVarDeclLValue(*this, E, VD);
2504 
2505   if (const auto *FD = dyn_cast<FunctionDecl>(ND))
2506     return EmitFunctionDeclLValue(*this, E, FD);
2507 
2508   llvm_unreachable("Unhandled member declaration!");
2509 }
2510 
2511 /// Given that we are currently emitting a lambda, emit an l-value for
2512 /// one of its members.
2513 LValue CodeGenFunction::EmitLValueForLambdaField(const FieldDecl *Field) {
2514   assert(cast<CXXMethodDecl>(CurCodeDecl)->getParent()->isLambda());
2515   assert(cast<CXXMethodDecl>(CurCodeDecl)->getParent() == Field->getParent());
2516   QualType LambdaTagType =
2517     getContext().getTagDeclType(Field->getParent());
2518   LValue LambdaLV = MakeNaturalAlignAddrLValue(CXXABIThisValue, LambdaTagType);
2519   return EmitLValueForField(LambdaLV, Field);
2520 }
2521 
2522 LValue CodeGenFunction::EmitLValueForField(LValue base,
2523                                            const FieldDecl *field) {
2524   if (field->isBitField()) {
2525     const CGRecordLayout &RL =
2526       CGM.getTypes().getCGRecordLayout(field->getParent());
2527     const CGBitFieldInfo &Info = RL.getBitFieldInfo(field);
2528     llvm::Value *Addr = base.getAddress();
2529     unsigned Idx = RL.getLLVMFieldNo(field);
2530     if (Idx != 0)
2531       // For structs, we GEP to the field that the record layout suggests.
2532       Addr = Builder.CreateStructGEP(Addr, Idx, field->getName());
2533     // Get the access type.
2534     llvm::Type *PtrTy = llvm::Type::getIntNPtrTy(
2535       getLLVMContext(), Info.StorageSize,
2536       CGM.getContext().getTargetAddressSpace(base.getType()));
2537     if (Addr->getType() != PtrTy)
2538       Addr = Builder.CreateBitCast(Addr, PtrTy);
2539 
2540     QualType fieldType =
2541       field->getType().withCVRQualifiers(base.getVRQualifiers());
2542     return LValue::MakeBitfield(Addr, Info, fieldType, base.getAlignment());
2543   }
2544 
2545   const RecordDecl *rec = field->getParent();
2546   QualType type = field->getType();
2547   CharUnits alignment = getContext().getDeclAlign(field);
2548 
2549   // FIXME: It should be impossible to have an LValue without alignment for a
2550   // complete type.
2551   if (!base.getAlignment().isZero())
2552     alignment = std::min(alignment, base.getAlignment());
2553 
2554   bool mayAlias = rec->hasAttr<MayAliasAttr>();
2555 
2556   llvm::Value *addr = base.getAddress();
2557   unsigned cvr = base.getVRQualifiers();
2558   bool TBAAPath = CGM.getCodeGenOpts().StructPathTBAA;
2559   if (rec->isUnion()) {
2560     // For unions, there is no pointer adjustment.
2561     assert(!type->isReferenceType() && "union has reference member");
2562     // TODO: handle path-aware TBAA for union.
2563     TBAAPath = false;
2564   } else {
2565     // For structs, we GEP to the field that the record layout suggests.
2566     unsigned idx = CGM.getTypes().getCGRecordLayout(rec).getLLVMFieldNo(field);
2567     addr = Builder.CreateStructGEP(addr, idx, field->getName());
2568 
2569     // If this is a reference field, load the reference right now.
2570     if (const ReferenceType *refType = type->getAs<ReferenceType>()) {
2571       llvm::LoadInst *load = Builder.CreateLoad(addr, "ref");
2572       if (cvr & Qualifiers::Volatile) load->setVolatile(true);
2573       load->setAlignment(alignment.getQuantity());
2574 
2575       // Loading the reference will disable path-aware TBAA.
2576       TBAAPath = false;
2577       if (CGM.shouldUseTBAA()) {
2578         llvm::MDNode *tbaa;
2579         if (mayAlias)
2580           tbaa = CGM.getTBAAInfo(getContext().CharTy);
2581         else
2582           tbaa = CGM.getTBAAInfo(type);
2583         if (tbaa)
2584           CGM.DecorateInstruction(load, tbaa);
2585       }
2586 
2587       addr = load;
2588       mayAlias = false;
2589       type = refType->getPointeeType();
2590       if (type->isIncompleteType())
2591         alignment = CharUnits();
2592       else
2593         alignment = getContext().getTypeAlignInChars(type);
2594       cvr = 0; // qualifiers don't recursively apply to referencee
2595     }
2596   }
2597 
2598   // Make sure that the address is pointing to the right type.  This is critical
2599   // for both unions and structs.  A union needs a bitcast, a struct element
2600   // will need a bitcast if the LLVM type laid out doesn't match the desired
2601   // type.
2602   addr = EmitBitCastOfLValueToProperType(*this, addr,
2603                                          CGM.getTypes().ConvertTypeForMem(type),
2604                                          field->getName());
2605 
2606   if (field->hasAttr<AnnotateAttr>())
2607     addr = EmitFieldAnnotations(field, addr);
2608 
2609   LValue LV = MakeAddrLValue(addr, type, alignment);
2610   LV.getQuals().addCVRQualifiers(cvr);
2611   if (TBAAPath) {
2612     const ASTRecordLayout &Layout =
2613         getContext().getASTRecordLayout(field->getParent());
2614     // Set the base type to be the base type of the base LValue and
2615     // update offset to be relative to the base type.
2616     LV.setTBAABaseType(mayAlias ? getContext().CharTy : base.getTBAABaseType());
2617     LV.setTBAAOffset(mayAlias ? 0 : base.getTBAAOffset() +
2618                      Layout.getFieldOffset(field->getFieldIndex()) /
2619                                            getContext().getCharWidth());
2620   }
2621 
2622   // __weak attribute on a field is ignored.
2623   if (LV.getQuals().getObjCGCAttr() == Qualifiers::Weak)
2624     LV.getQuals().removeObjCGCAttr();
2625 
2626   // Fields of may_alias structs act like 'char' for TBAA purposes.
2627   // FIXME: this should get propagated down through anonymous structs
2628   // and unions.
2629   if (mayAlias && LV.getTBAAInfo())
2630     LV.setTBAAInfo(CGM.getTBAAInfo(getContext().CharTy));
2631 
2632   return LV;
2633 }
2634 
2635 LValue
2636 CodeGenFunction::EmitLValueForFieldInitialization(LValue Base,
2637                                                   const FieldDecl *Field) {
2638   QualType FieldType = Field->getType();
2639 
2640   if (!FieldType->isReferenceType())
2641     return EmitLValueForField(Base, Field);
2642 
2643   const CGRecordLayout &RL =
2644     CGM.getTypes().getCGRecordLayout(Field->getParent());
2645   unsigned idx = RL.getLLVMFieldNo(Field);
2646   llvm::Value *V = Builder.CreateStructGEP(Base.getAddress(), idx);
2647   assert(!FieldType.getObjCGCAttr() && "fields cannot have GC attrs");
2648 
2649   // Make sure that the address is pointing to the right type.  This is critical
2650   // for both unions and structs.  A union needs a bitcast, a struct element
2651   // will need a bitcast if the LLVM type laid out doesn't match the desired
2652   // type.
2653   llvm::Type *llvmType = ConvertTypeForMem(FieldType);
2654   V = EmitBitCastOfLValueToProperType(*this, V, llvmType, Field->getName());
2655 
2656   CharUnits Alignment = getContext().getDeclAlign(Field);
2657 
2658   // FIXME: It should be impossible to have an LValue without alignment for a
2659   // complete type.
2660   if (!Base.getAlignment().isZero())
2661     Alignment = std::min(Alignment, Base.getAlignment());
2662 
2663   return MakeAddrLValue(V, FieldType, Alignment);
2664 }
2665 
2666 LValue CodeGenFunction::EmitCompoundLiteralLValue(const CompoundLiteralExpr *E){
2667   if (E->isFileScope()) {
2668     llvm::Value *GlobalPtr = CGM.GetAddrOfConstantCompoundLiteral(E);
2669     return MakeAddrLValue(GlobalPtr, E->getType());
2670   }
2671   if (E->getType()->isVariablyModifiedType())
2672     // make sure to emit the VLA size.
2673     EmitVariablyModifiedType(E->getType());
2674 
2675   llvm::Value *DeclPtr = CreateMemTemp(E->getType(), ".compoundliteral");
2676   const Expr *InitExpr = E->getInitializer();
2677   LValue Result = MakeAddrLValue(DeclPtr, E->getType());
2678 
2679   EmitAnyExprToMem(InitExpr, DeclPtr, E->getType().getQualifiers(),
2680                    /*Init*/ true);
2681 
2682   return Result;
2683 }
2684 
2685 LValue CodeGenFunction::EmitInitListLValue(const InitListExpr *E) {
2686   if (!E->isGLValue())
2687     // Initializing an aggregate temporary in C++11: T{...}.
2688     return EmitAggExprToLValue(E);
2689 
2690   // An lvalue initializer list must be initializing a reference.
2691   assert(E->getNumInits() == 1 && "reference init with multiple values");
2692   return EmitLValue(E->getInit(0));
2693 }
2694 
2695 /// Emit the operand of a glvalue conditional operator. This is either a glvalue
2696 /// or a (possibly-parenthesized) throw-expression. If this is a throw, no
2697 /// LValue is returned and the current block has been terminated.
2698 static Optional<LValue> EmitLValueOrThrowExpression(CodeGenFunction &CGF,
2699                                                     const Expr *Operand) {
2700   if (auto *ThrowExpr = dyn_cast<CXXThrowExpr>(Operand->IgnoreParens())) {
2701     CGF.EmitCXXThrowExpr(ThrowExpr, /*KeepInsertionPoint*/false);
2702     return None;
2703   }
2704 
2705   return CGF.EmitLValue(Operand);
2706 }
2707 
2708 LValue CodeGenFunction::
2709 EmitConditionalOperatorLValue(const AbstractConditionalOperator *expr) {
2710   if (!expr->isGLValue()) {
2711     // ?: here should be an aggregate.
2712     assert(hasAggregateEvaluationKind(expr->getType()) &&
2713            "Unexpected conditional operator!");
2714     return EmitAggExprToLValue(expr);
2715   }
2716 
2717   OpaqueValueMapping binding(*this, expr);
2718   RegionCounter Cnt = getPGORegionCounter(expr);
2719 
2720   const Expr *condExpr = expr->getCond();
2721   bool CondExprBool;
2722   if (ConstantFoldsToSimpleInteger(condExpr, CondExprBool)) {
2723     const Expr *live = expr->getTrueExpr(), *dead = expr->getFalseExpr();
2724     if (!CondExprBool) std::swap(live, dead);
2725 
2726     if (!ContainsLabel(dead)) {
2727       // If the true case is live, we need to track its region.
2728       if (CondExprBool)
2729         Cnt.beginRegion(Builder);
2730       return EmitLValue(live);
2731     }
2732   }
2733 
2734   llvm::BasicBlock *lhsBlock = createBasicBlock("cond.true");
2735   llvm::BasicBlock *rhsBlock = createBasicBlock("cond.false");
2736   llvm::BasicBlock *contBlock = createBasicBlock("cond.end");
2737 
2738   ConditionalEvaluation eval(*this);
2739   EmitBranchOnBoolExpr(condExpr, lhsBlock, rhsBlock, Cnt.getCount());
2740 
2741   // Any temporaries created here are conditional.
2742   EmitBlock(lhsBlock);
2743   Cnt.beginRegion(Builder);
2744   eval.begin(*this);
2745   Optional<LValue> lhs =
2746       EmitLValueOrThrowExpression(*this, expr->getTrueExpr());
2747   eval.end(*this);
2748 
2749   if (lhs && !lhs->isSimple())
2750     return EmitUnsupportedLValue(expr, "conditional operator");
2751 
2752   lhsBlock = Builder.GetInsertBlock();
2753   if (lhs)
2754     Builder.CreateBr(contBlock);
2755 
2756   // Any temporaries created here are conditional.
2757   EmitBlock(rhsBlock);
2758   eval.begin(*this);
2759   Optional<LValue> rhs =
2760       EmitLValueOrThrowExpression(*this, expr->getFalseExpr());
2761   eval.end(*this);
2762   if (rhs && !rhs->isSimple())
2763     return EmitUnsupportedLValue(expr, "conditional operator");
2764   rhsBlock = Builder.GetInsertBlock();
2765 
2766   EmitBlock(contBlock);
2767 
2768   if (lhs && rhs) {
2769     llvm::PHINode *phi = Builder.CreatePHI(lhs->getAddress()->getType(),
2770                                            2, "cond-lvalue");
2771     phi->addIncoming(lhs->getAddress(), lhsBlock);
2772     phi->addIncoming(rhs->getAddress(), rhsBlock);
2773     return MakeAddrLValue(phi, expr->getType());
2774   } else {
2775     assert((lhs || rhs) &&
2776            "both operands of glvalue conditional are throw-expressions?");
2777     return lhs ? *lhs : *rhs;
2778   }
2779 }
2780 
2781 /// EmitCastLValue - Casts are never lvalues unless that cast is to a reference
2782 /// type. If the cast is to a reference, we can have the usual lvalue result,
2783 /// otherwise if a cast is needed by the code generator in an lvalue context,
2784 /// then it must mean that we need the address of an aggregate in order to
2785 /// access one of its members.  This can happen for all the reasons that casts
2786 /// are permitted with aggregate result, including noop aggregate casts, and
2787 /// cast from scalar to union.
2788 LValue CodeGenFunction::EmitCastLValue(const CastExpr *E) {
2789   switch (E->getCastKind()) {
2790   case CK_ToVoid:
2791   case CK_BitCast:
2792   case CK_ArrayToPointerDecay:
2793   case CK_FunctionToPointerDecay:
2794   case CK_NullToMemberPointer:
2795   case CK_NullToPointer:
2796   case CK_IntegralToPointer:
2797   case CK_PointerToIntegral:
2798   case CK_PointerToBoolean:
2799   case CK_VectorSplat:
2800   case CK_IntegralCast:
2801   case CK_IntegralToBoolean:
2802   case CK_IntegralToFloating:
2803   case CK_FloatingToIntegral:
2804   case CK_FloatingToBoolean:
2805   case CK_FloatingCast:
2806   case CK_FloatingRealToComplex:
2807   case CK_FloatingComplexToReal:
2808   case CK_FloatingComplexToBoolean:
2809   case CK_FloatingComplexCast:
2810   case CK_FloatingComplexToIntegralComplex:
2811   case CK_IntegralRealToComplex:
2812   case CK_IntegralComplexToReal:
2813   case CK_IntegralComplexToBoolean:
2814   case CK_IntegralComplexCast:
2815   case CK_IntegralComplexToFloatingComplex:
2816   case CK_DerivedToBaseMemberPointer:
2817   case CK_BaseToDerivedMemberPointer:
2818   case CK_MemberPointerToBoolean:
2819   case CK_ReinterpretMemberPointer:
2820   case CK_AnyPointerToBlockPointerCast:
2821   case CK_ARCProduceObject:
2822   case CK_ARCConsumeObject:
2823   case CK_ARCReclaimReturnedObject:
2824   case CK_ARCExtendBlockObject:
2825   case CK_CopyAndAutoreleaseBlockObject:
2826   case CK_AddressSpaceConversion:
2827     return EmitUnsupportedLValue(E, "unexpected cast lvalue");
2828 
2829   case CK_Dependent:
2830     llvm_unreachable("dependent cast kind in IR gen!");
2831 
2832   case CK_BuiltinFnToFnPtr:
2833     llvm_unreachable("builtin functions are handled elsewhere");
2834 
2835   // These are never l-values; just use the aggregate emission code.
2836   case CK_NonAtomicToAtomic:
2837   case CK_AtomicToNonAtomic:
2838     return EmitAggExprToLValue(E);
2839 
2840   case CK_Dynamic: {
2841     LValue LV = EmitLValue(E->getSubExpr());
2842     llvm::Value *V = LV.getAddress();
2843     const auto *DCE = cast<CXXDynamicCastExpr>(E);
2844     return MakeAddrLValue(EmitDynamicCast(V, DCE), E->getType());
2845   }
2846 
2847   case CK_ConstructorConversion:
2848   case CK_UserDefinedConversion:
2849   case CK_CPointerToObjCPointerCast:
2850   case CK_BlockPointerToObjCPointerCast:
2851   case CK_NoOp:
2852   case CK_LValueToRValue:
2853     return EmitLValue(E->getSubExpr());
2854 
2855   case CK_UncheckedDerivedToBase:
2856   case CK_DerivedToBase: {
2857     const RecordType *DerivedClassTy =
2858       E->getSubExpr()->getType()->getAs<RecordType>();
2859     auto *DerivedClassDecl = cast<CXXRecordDecl>(DerivedClassTy->getDecl());
2860 
2861     LValue LV = EmitLValue(E->getSubExpr());
2862     llvm::Value *This = LV.getAddress();
2863 
2864     // Perform the derived-to-base conversion
2865     llvm::Value *Base = GetAddressOfBaseClass(
2866         This, DerivedClassDecl, E->path_begin(), E->path_end(),
2867         /*NullCheckValue=*/false, E->getExprLoc());
2868 
2869     return MakeAddrLValue(Base, E->getType());
2870   }
2871   case CK_ToUnion:
2872     return EmitAggExprToLValue(E);
2873   case CK_BaseToDerived: {
2874     const RecordType *DerivedClassTy = E->getType()->getAs<RecordType>();
2875     auto *DerivedClassDecl = cast<CXXRecordDecl>(DerivedClassTy->getDecl());
2876 
2877     LValue LV = EmitLValue(E->getSubExpr());
2878 
2879     // Perform the base-to-derived conversion
2880     llvm::Value *Derived =
2881       GetAddressOfDerivedClass(LV.getAddress(), DerivedClassDecl,
2882                                E->path_begin(), E->path_end(),
2883                                /*NullCheckValue=*/false);
2884 
2885     // C++11 [expr.static.cast]p2: Behavior is undefined if a downcast is
2886     // performed and the object is not of the derived type.
2887     if (sanitizePerformTypeCheck())
2888       EmitTypeCheck(TCK_DowncastReference, E->getExprLoc(),
2889                     Derived, E->getType());
2890 
2891     return MakeAddrLValue(Derived, E->getType());
2892   }
2893   case CK_LValueBitCast: {
2894     // This must be a reinterpret_cast (or c-style equivalent).
2895     const auto *CE = cast<ExplicitCastExpr>(E);
2896 
2897     LValue LV = EmitLValue(E->getSubExpr());
2898     llvm::Value *V = Builder.CreateBitCast(LV.getAddress(),
2899                                            ConvertType(CE->getTypeAsWritten()));
2900     return MakeAddrLValue(V, E->getType());
2901   }
2902   case CK_ObjCObjectLValueCast: {
2903     LValue LV = EmitLValue(E->getSubExpr());
2904     QualType ToType = getContext().getLValueReferenceType(E->getType());
2905     llvm::Value *V = Builder.CreateBitCast(LV.getAddress(),
2906                                            ConvertType(ToType));
2907     return MakeAddrLValue(V, E->getType());
2908   }
2909   case CK_ZeroToOCLEvent:
2910     llvm_unreachable("NULL to OpenCL event lvalue cast is not valid");
2911   }
2912 
2913   llvm_unreachable("Unhandled lvalue cast kind?");
2914 }
2915 
2916 LValue CodeGenFunction::EmitOpaqueValueLValue(const OpaqueValueExpr *e) {
2917   assert(OpaqueValueMappingData::shouldBindAsLValue(e));
2918   return getOpaqueLValueMapping(e);
2919 }
2920 
2921 RValue CodeGenFunction::EmitRValueForField(LValue LV,
2922                                            const FieldDecl *FD,
2923                                            SourceLocation Loc) {
2924   QualType FT = FD->getType();
2925   LValue FieldLV = EmitLValueForField(LV, FD);
2926   switch (getEvaluationKind(FT)) {
2927   case TEK_Complex:
2928     return RValue::getComplex(EmitLoadOfComplex(FieldLV, Loc));
2929   case TEK_Aggregate:
2930     return FieldLV.asAggregateRValue();
2931   case TEK_Scalar:
2932     return EmitLoadOfLValue(FieldLV, Loc);
2933   }
2934   llvm_unreachable("bad evaluation kind");
2935 }
2936 
2937 //===--------------------------------------------------------------------===//
2938 //                             Expression Emission
2939 //===--------------------------------------------------------------------===//
2940 
2941 RValue CodeGenFunction::EmitCallExpr(const CallExpr *E,
2942                                      ReturnValueSlot ReturnValue) {
2943   if (CGDebugInfo *DI = getDebugInfo()) {
2944     SourceLocation Loc = E->getLocStart();
2945     // Force column info to be generated so we can differentiate
2946     // multiple call sites on the same line in the debug info.
2947     // FIXME: This is insufficient. Two calls coming from the same macro
2948     // expansion will still get the same line/column and break debug info. It's
2949     // possible that LLVM can be fixed to not rely on this uniqueness, at which
2950     // point this workaround can be removed.
2951     const FunctionDecl* Callee = E->getDirectCallee();
2952     bool ForceColumnInfo = Callee && Callee->isInlineSpecified();
2953     DI->EmitLocation(Builder, Loc, ForceColumnInfo);
2954   }
2955 
2956   // Builtins never have block type.
2957   if (E->getCallee()->getType()->isBlockPointerType())
2958     return EmitBlockCallExpr(E, ReturnValue);
2959 
2960   if (const auto *CE = dyn_cast<CXXMemberCallExpr>(E))
2961     return EmitCXXMemberCallExpr(CE, ReturnValue);
2962 
2963   if (const auto *CE = dyn_cast<CUDAKernelCallExpr>(E))
2964     return EmitCUDAKernelCallExpr(CE, ReturnValue);
2965 
2966   const Decl *TargetDecl = E->getCalleeDecl();
2967   if (const FunctionDecl *FD = dyn_cast_or_null<FunctionDecl>(TargetDecl)) {
2968     if (unsigned builtinID = FD->getBuiltinID())
2969       return EmitBuiltinExpr(FD, builtinID, E);
2970   }
2971 
2972   if (const auto *CE = dyn_cast<CXXOperatorCallExpr>(E))
2973     if (const CXXMethodDecl *MD = dyn_cast_or_null<CXXMethodDecl>(TargetDecl))
2974       return EmitCXXOperatorMemberCallExpr(CE, MD, ReturnValue);
2975 
2976   if (const auto *PseudoDtor =
2977           dyn_cast<CXXPseudoDestructorExpr>(E->getCallee()->IgnoreParens())) {
2978     QualType DestroyedType = PseudoDtor->getDestroyedType();
2979     if (getLangOpts().ObjCAutoRefCount &&
2980         DestroyedType->isObjCLifetimeType() &&
2981         (DestroyedType.getObjCLifetime() == Qualifiers::OCL_Strong ||
2982          DestroyedType.getObjCLifetime() == Qualifiers::OCL_Weak)) {
2983       // Automatic Reference Counting:
2984       //   If the pseudo-expression names a retainable object with weak or
2985       //   strong lifetime, the object shall be released.
2986       Expr *BaseExpr = PseudoDtor->getBase();
2987       llvm::Value *BaseValue = nullptr;
2988       Qualifiers BaseQuals;
2989 
2990       // If this is s.x, emit s as an lvalue. If it is s->x, emit s as a scalar.
2991       if (PseudoDtor->isArrow()) {
2992         BaseValue = EmitScalarExpr(BaseExpr);
2993         const PointerType *PTy = BaseExpr->getType()->getAs<PointerType>();
2994         BaseQuals = PTy->getPointeeType().getQualifiers();
2995       } else {
2996         LValue BaseLV = EmitLValue(BaseExpr);
2997         BaseValue = BaseLV.getAddress();
2998         QualType BaseTy = BaseExpr->getType();
2999         BaseQuals = BaseTy.getQualifiers();
3000       }
3001 
3002       switch (PseudoDtor->getDestroyedType().getObjCLifetime()) {
3003       case Qualifiers::OCL_None:
3004       case Qualifiers::OCL_ExplicitNone:
3005       case Qualifiers::OCL_Autoreleasing:
3006         break;
3007 
3008       case Qualifiers::OCL_Strong:
3009         EmitARCRelease(Builder.CreateLoad(BaseValue,
3010                           PseudoDtor->getDestroyedType().isVolatileQualified()),
3011                        ARCPreciseLifetime);
3012         break;
3013 
3014       case Qualifiers::OCL_Weak:
3015         EmitARCDestroyWeak(BaseValue);
3016         break;
3017       }
3018     } else {
3019       // C++ [expr.pseudo]p1:
3020       //   The result shall only be used as the operand for the function call
3021       //   operator (), and the result of such a call has type void. The only
3022       //   effect is the evaluation of the postfix-expression before the dot or
3023       //   arrow.
3024       EmitScalarExpr(E->getCallee());
3025     }
3026 
3027     return RValue::get(nullptr);
3028   }
3029 
3030   llvm::Value *Callee = EmitScalarExpr(E->getCallee());
3031   return EmitCall(E->getCallee()->getType(), Callee, E, ReturnValue,
3032                   TargetDecl);
3033 }
3034 
3035 LValue CodeGenFunction::EmitBinaryOperatorLValue(const BinaryOperator *E) {
3036   // Comma expressions just emit their LHS then their RHS as an l-value.
3037   if (E->getOpcode() == BO_Comma) {
3038     EmitIgnoredExpr(E->getLHS());
3039     EnsureInsertPoint();
3040     return EmitLValue(E->getRHS());
3041   }
3042 
3043   if (E->getOpcode() == BO_PtrMemD ||
3044       E->getOpcode() == BO_PtrMemI)
3045     return EmitPointerToDataMemberBinaryExpr(E);
3046 
3047   assert(E->getOpcode() == BO_Assign && "unexpected binary l-value");
3048 
3049   // Note that in all of these cases, __block variables need the RHS
3050   // evaluated first just in case the variable gets moved by the RHS.
3051 
3052   switch (getEvaluationKind(E->getType())) {
3053   case TEK_Scalar: {
3054     switch (E->getLHS()->getType().getObjCLifetime()) {
3055     case Qualifiers::OCL_Strong:
3056       return EmitARCStoreStrong(E, /*ignored*/ false).first;
3057 
3058     case Qualifiers::OCL_Autoreleasing:
3059       return EmitARCStoreAutoreleasing(E).first;
3060 
3061     // No reason to do any of these differently.
3062     case Qualifiers::OCL_None:
3063     case Qualifiers::OCL_ExplicitNone:
3064     case Qualifiers::OCL_Weak:
3065       break;
3066     }
3067 
3068     RValue RV = EmitAnyExpr(E->getRHS());
3069     LValue LV = EmitCheckedLValue(E->getLHS(), TCK_Store);
3070     EmitStoreThroughLValue(RV, LV);
3071     return LV;
3072   }
3073 
3074   case TEK_Complex:
3075     return EmitComplexAssignmentLValue(E);
3076 
3077   case TEK_Aggregate:
3078     return EmitAggExprToLValue(E);
3079   }
3080   llvm_unreachable("bad evaluation kind");
3081 }
3082 
3083 LValue CodeGenFunction::EmitCallExprLValue(const CallExpr *E) {
3084   RValue RV = EmitCallExpr(E);
3085 
3086   if (!RV.isScalar())
3087     return MakeAddrLValue(RV.getAggregateAddr(), E->getType());
3088 
3089   assert(E->getCallReturnType()->isReferenceType() &&
3090          "Can't have a scalar return unless the return type is a "
3091          "reference type!");
3092 
3093   return MakeAddrLValue(RV.getScalarVal(), E->getType());
3094 }
3095 
3096 LValue CodeGenFunction::EmitVAArgExprLValue(const VAArgExpr *E) {
3097   // FIXME: This shouldn't require another copy.
3098   return EmitAggExprToLValue(E);
3099 }
3100 
3101 LValue CodeGenFunction::EmitCXXConstructLValue(const CXXConstructExpr *E) {
3102   assert(E->getType()->getAsCXXRecordDecl()->hasTrivialDestructor()
3103          && "binding l-value to type which needs a temporary");
3104   AggValueSlot Slot = CreateAggTemp(E->getType());
3105   EmitCXXConstructExpr(E, Slot);
3106   return MakeAddrLValue(Slot.getAddr(), E->getType());
3107 }
3108 
3109 LValue
3110 CodeGenFunction::EmitCXXTypeidLValue(const CXXTypeidExpr *E) {
3111   return MakeAddrLValue(EmitCXXTypeidExpr(E), E->getType());
3112 }
3113 
3114 llvm::Value *CodeGenFunction::EmitCXXUuidofExpr(const CXXUuidofExpr *E) {
3115   return Builder.CreateBitCast(CGM.GetAddrOfUuidDescriptor(E),
3116                                ConvertType(E->getType())->getPointerTo());
3117 }
3118 
3119 LValue CodeGenFunction::EmitCXXUuidofLValue(const CXXUuidofExpr *E) {
3120   return MakeAddrLValue(EmitCXXUuidofExpr(E), E->getType());
3121 }
3122 
3123 LValue
3124 CodeGenFunction::EmitCXXBindTemporaryLValue(const CXXBindTemporaryExpr *E) {
3125   AggValueSlot Slot = CreateAggTemp(E->getType(), "temp.lvalue");
3126   Slot.setExternallyDestructed();
3127   EmitAggExpr(E->getSubExpr(), Slot);
3128   EmitCXXTemporary(E->getTemporary(), E->getType(), Slot.getAddr());
3129   return MakeAddrLValue(Slot.getAddr(), E->getType());
3130 }
3131 
3132 LValue
3133 CodeGenFunction::EmitLambdaLValue(const LambdaExpr *E) {
3134   AggValueSlot Slot = CreateAggTemp(E->getType(), "temp.lvalue");
3135   EmitLambdaExpr(E, Slot);
3136   return MakeAddrLValue(Slot.getAddr(), E->getType());
3137 }
3138 
3139 LValue CodeGenFunction::EmitObjCMessageExprLValue(const ObjCMessageExpr *E) {
3140   RValue RV = EmitObjCMessageExpr(E);
3141 
3142   if (!RV.isScalar())
3143     return MakeAddrLValue(RV.getAggregateAddr(), E->getType());
3144 
3145   assert(E->getMethodDecl()->getReturnType()->isReferenceType() &&
3146          "Can't have a scalar return unless the return type is a "
3147          "reference type!");
3148 
3149   return MakeAddrLValue(RV.getScalarVal(), E->getType());
3150 }
3151 
3152 LValue CodeGenFunction::EmitObjCSelectorLValue(const ObjCSelectorExpr *E) {
3153   llvm::Value *V =
3154     CGM.getObjCRuntime().GetSelector(*this, E->getSelector(), true);
3155   return MakeAddrLValue(V, E->getType());
3156 }
3157 
3158 llvm::Value *CodeGenFunction::EmitIvarOffset(const ObjCInterfaceDecl *Interface,
3159                                              const ObjCIvarDecl *Ivar) {
3160   return CGM.getObjCRuntime().EmitIvarOffset(*this, Interface, Ivar);
3161 }
3162 
3163 LValue CodeGenFunction::EmitLValueForIvar(QualType ObjectTy,
3164                                           llvm::Value *BaseValue,
3165                                           const ObjCIvarDecl *Ivar,
3166                                           unsigned CVRQualifiers) {
3167   return CGM.getObjCRuntime().EmitObjCValueForIvar(*this, ObjectTy, BaseValue,
3168                                                    Ivar, CVRQualifiers);
3169 }
3170 
3171 LValue CodeGenFunction::EmitObjCIvarRefLValue(const ObjCIvarRefExpr *E) {
3172   // FIXME: A lot of the code below could be shared with EmitMemberExpr.
3173   llvm::Value *BaseValue = nullptr;
3174   const Expr *BaseExpr = E->getBase();
3175   Qualifiers BaseQuals;
3176   QualType ObjectTy;
3177   if (E->isArrow()) {
3178     BaseValue = EmitScalarExpr(BaseExpr);
3179     ObjectTy = BaseExpr->getType()->getPointeeType();
3180     BaseQuals = ObjectTy.getQualifiers();
3181   } else {
3182     LValue BaseLV = EmitLValue(BaseExpr);
3183     // FIXME: this isn't right for bitfields.
3184     BaseValue = BaseLV.getAddress();
3185     ObjectTy = BaseExpr->getType();
3186     BaseQuals = ObjectTy.getQualifiers();
3187   }
3188 
3189   LValue LV =
3190     EmitLValueForIvar(ObjectTy, BaseValue, E->getDecl(),
3191                       BaseQuals.getCVRQualifiers());
3192   setObjCGCLValueClass(getContext(), E, LV);
3193   return LV;
3194 }
3195 
3196 LValue CodeGenFunction::EmitStmtExprLValue(const StmtExpr *E) {
3197   // Can only get l-value for message expression returning aggregate type
3198   RValue RV = EmitAnyExprToTemp(E);
3199   return MakeAddrLValue(RV.getAggregateAddr(), E->getType());
3200 }
3201 
3202 RValue CodeGenFunction::EmitCall(QualType CalleeType, llvm::Value *Callee,
3203                                  const CallExpr *E, ReturnValueSlot ReturnValue,
3204                                  const Decl *TargetDecl) {
3205   // Get the actual function type. The callee type will always be a pointer to
3206   // function type or a block pointer type.
3207   assert(CalleeType->isFunctionPointerType() &&
3208          "Call must have function pointer type!");
3209 
3210   CalleeType = getContext().getCanonicalType(CalleeType);
3211 
3212   const auto *FnType =
3213       cast<FunctionType>(cast<PointerType>(CalleeType)->getPointeeType());
3214 
3215   // Force column info to differentiate multiple inlined call sites on
3216   // the same line, analoguous to EmitCallExpr.
3217   // FIXME: This is insufficient. Two calls coming from the same macro expansion
3218   // will still get the same line/column and break debug info. It's possible
3219   // that LLVM can be fixed to not rely on this uniqueness, at which point this
3220   // workaround can be removed.
3221   bool ForceColumnInfo = false;
3222   if (const FunctionDecl* FD = dyn_cast_or_null<const FunctionDecl>(TargetDecl))
3223     ForceColumnInfo = FD->isInlineSpecified();
3224 
3225   if (getLangOpts().CPlusPlus && SanOpts.Function &&
3226       (!TargetDecl || !isa<FunctionDecl>(TargetDecl))) {
3227     if (llvm::Constant *PrefixSig =
3228             CGM.getTargetCodeGenInfo().getUBSanFunctionSignature(CGM)) {
3229       SanitizerScope SanScope(this);
3230       llvm::Constant *FTRTTIConst =
3231           CGM.GetAddrOfRTTIDescriptor(QualType(FnType, 0), /*ForEH=*/true);
3232       llvm::Type *PrefixStructTyElems[] = {
3233         PrefixSig->getType(),
3234         FTRTTIConst->getType()
3235       };
3236       llvm::StructType *PrefixStructTy = llvm::StructType::get(
3237           CGM.getLLVMContext(), PrefixStructTyElems, /*isPacked=*/true);
3238 
3239       llvm::Value *CalleePrefixStruct = Builder.CreateBitCast(
3240           Callee, llvm::PointerType::getUnqual(PrefixStructTy));
3241       llvm::Value *CalleeSigPtr =
3242           Builder.CreateConstGEP2_32(CalleePrefixStruct, 0, 0);
3243       llvm::Value *CalleeSig = Builder.CreateLoad(CalleeSigPtr);
3244       llvm::Value *CalleeSigMatch = Builder.CreateICmpEQ(CalleeSig, PrefixSig);
3245 
3246       llvm::BasicBlock *Cont = createBasicBlock("cont");
3247       llvm::BasicBlock *TypeCheck = createBasicBlock("typecheck");
3248       Builder.CreateCondBr(CalleeSigMatch, TypeCheck, Cont);
3249 
3250       EmitBlock(TypeCheck);
3251       llvm::Value *CalleeRTTIPtr =
3252           Builder.CreateConstGEP2_32(CalleePrefixStruct, 0, 1);
3253       llvm::Value *CalleeRTTI = Builder.CreateLoad(CalleeRTTIPtr);
3254       llvm::Value *CalleeRTTIMatch =
3255           Builder.CreateICmpEQ(CalleeRTTI, FTRTTIConst);
3256       llvm::Constant *StaticData[] = {
3257         EmitCheckSourceLocation(E->getLocStart()),
3258         EmitCheckTypeDescriptor(CalleeType)
3259       };
3260       EmitCheck(CalleeRTTIMatch,
3261                 "function_type_mismatch",
3262                 StaticData,
3263                 Callee,
3264                 CRK_Recoverable);
3265 
3266       Builder.CreateBr(Cont);
3267       EmitBlock(Cont);
3268     }
3269   }
3270 
3271   CallArgList Args;
3272   EmitCallArgs(Args, dyn_cast<FunctionProtoType>(FnType), E->arg_begin(),
3273                E->arg_end(), E->getDirectCallee(), /*ParamsToSkip*/ 0,
3274                ForceColumnInfo);
3275 
3276   const CGFunctionInfo &FnInfo =
3277     CGM.getTypes().arrangeFreeFunctionCall(Args, FnType);
3278 
3279   // C99 6.5.2.2p6:
3280   //   If the expression that denotes the called function has a type
3281   //   that does not include a prototype, [the default argument
3282   //   promotions are performed]. If the number of arguments does not
3283   //   equal the number of parameters, the behavior is undefined. If
3284   //   the function is defined with a type that includes a prototype,
3285   //   and either the prototype ends with an ellipsis (, ...) or the
3286   //   types of the arguments after promotion are not compatible with
3287   //   the types of the parameters, the behavior is undefined. If the
3288   //   function is defined with a type that does not include a
3289   //   prototype, and the types of the arguments after promotion are
3290   //   not compatible with those of the parameters after promotion,
3291   //   the behavior is undefined [except in some trivial cases].
3292   // That is, in the general case, we should assume that a call
3293   // through an unprototyped function type works like a *non-variadic*
3294   // call.  The way we make this work is to cast to the exact type
3295   // of the promoted arguments.
3296   if (isa<FunctionNoProtoType>(FnType)) {
3297     llvm::Type *CalleeTy = getTypes().GetFunctionType(FnInfo);
3298     CalleeTy = CalleeTy->getPointerTo();
3299     Callee = Builder.CreateBitCast(Callee, CalleeTy, "callee.knr.cast");
3300   }
3301 
3302   return EmitCall(FnInfo, Callee, ReturnValue, Args, TargetDecl);
3303 }
3304 
3305 LValue CodeGenFunction::
3306 EmitPointerToDataMemberBinaryExpr(const BinaryOperator *E) {
3307   llvm::Value *BaseV;
3308   if (E->getOpcode() == BO_PtrMemI)
3309     BaseV = EmitScalarExpr(E->getLHS());
3310   else
3311     BaseV = EmitLValue(E->getLHS()).getAddress();
3312 
3313   llvm::Value *OffsetV = EmitScalarExpr(E->getRHS());
3314 
3315   const MemberPointerType *MPT
3316     = E->getRHS()->getType()->getAs<MemberPointerType>();
3317 
3318   llvm::Value *AddV = CGM.getCXXABI().EmitMemberDataPointerAddress(
3319       *this, E, BaseV, OffsetV, MPT);
3320 
3321   return MakeAddrLValue(AddV, MPT->getPointeeType());
3322 }
3323 
3324 /// Given the address of a temporary variable, produce an r-value of
3325 /// its type.
3326 RValue CodeGenFunction::convertTempToRValue(llvm::Value *addr,
3327                                             QualType type,
3328                                             SourceLocation loc) {
3329   LValue lvalue = MakeNaturalAlignAddrLValue(addr, type);
3330   switch (getEvaluationKind(type)) {
3331   case TEK_Complex:
3332     return RValue::getComplex(EmitLoadOfComplex(lvalue, loc));
3333   case TEK_Aggregate:
3334     return lvalue.asAggregateRValue();
3335   case TEK_Scalar:
3336     return RValue::get(EmitLoadOfScalar(lvalue, loc));
3337   }
3338   llvm_unreachable("bad evaluation kind");
3339 }
3340 
3341 void CodeGenFunction::SetFPAccuracy(llvm::Value *Val, float Accuracy) {
3342   assert(Val->getType()->isFPOrFPVectorTy());
3343   if (Accuracy == 0.0 || !isa<llvm::Instruction>(Val))
3344     return;
3345 
3346   llvm::MDBuilder MDHelper(getLLVMContext());
3347   llvm::MDNode *Node = MDHelper.createFPMath(Accuracy);
3348 
3349   cast<llvm::Instruction>(Val)->setMetadata(llvm::LLVMContext::MD_fpmath, Node);
3350 }
3351 
3352 namespace {
3353   struct LValueOrRValue {
3354     LValue LV;
3355     RValue RV;
3356   };
3357 }
3358 
3359 static LValueOrRValue emitPseudoObjectExpr(CodeGenFunction &CGF,
3360                                            const PseudoObjectExpr *E,
3361                                            bool forLValue,
3362                                            AggValueSlot slot) {
3363   SmallVector<CodeGenFunction::OpaqueValueMappingData, 4> opaques;
3364 
3365   // Find the result expression, if any.
3366   const Expr *resultExpr = E->getResultExpr();
3367   LValueOrRValue result;
3368 
3369   for (PseudoObjectExpr::const_semantics_iterator
3370          i = E->semantics_begin(), e = E->semantics_end(); i != e; ++i) {
3371     const Expr *semantic = *i;
3372 
3373     // If this semantic expression is an opaque value, bind it
3374     // to the result of its source expression.
3375     if (const auto *ov = dyn_cast<OpaqueValueExpr>(semantic)) {
3376 
3377       // If this is the result expression, we may need to evaluate
3378       // directly into the slot.
3379       typedef CodeGenFunction::OpaqueValueMappingData OVMA;
3380       OVMA opaqueData;
3381       if (ov == resultExpr && ov->isRValue() && !forLValue &&
3382           CodeGenFunction::hasAggregateEvaluationKind(ov->getType())) {
3383         CGF.EmitAggExpr(ov->getSourceExpr(), slot);
3384 
3385         LValue LV = CGF.MakeAddrLValue(slot.getAddr(), ov->getType());
3386         opaqueData = OVMA::bind(CGF, ov, LV);
3387         result.RV = slot.asRValue();
3388 
3389       // Otherwise, emit as normal.
3390       } else {
3391         opaqueData = OVMA::bind(CGF, ov, ov->getSourceExpr());
3392 
3393         // If this is the result, also evaluate the result now.
3394         if (ov == resultExpr) {
3395           if (forLValue)
3396             result.LV = CGF.EmitLValue(ov);
3397           else
3398             result.RV = CGF.EmitAnyExpr(ov, slot);
3399         }
3400       }
3401 
3402       opaques.push_back(opaqueData);
3403 
3404     // Otherwise, if the expression is the result, evaluate it
3405     // and remember the result.
3406     } else if (semantic == resultExpr) {
3407       if (forLValue)
3408         result.LV = CGF.EmitLValue(semantic);
3409       else
3410         result.RV = CGF.EmitAnyExpr(semantic, slot);
3411 
3412     // Otherwise, evaluate the expression in an ignored context.
3413     } else {
3414       CGF.EmitIgnoredExpr(semantic);
3415     }
3416   }
3417 
3418   // Unbind all the opaques now.
3419   for (unsigned i = 0, e = opaques.size(); i != e; ++i)
3420     opaques[i].unbind(CGF);
3421 
3422   return result;
3423 }
3424 
3425 RValue CodeGenFunction::EmitPseudoObjectRValue(const PseudoObjectExpr *E,
3426                                                AggValueSlot slot) {
3427   return emitPseudoObjectExpr(*this, E, false, slot).RV;
3428 }
3429 
3430 LValue CodeGenFunction::EmitPseudoObjectLValue(const PseudoObjectExpr *E) {
3431   return emitPseudoObjectExpr(*this, E, true, AggValueSlot::ignored()).LV;
3432 }
3433