1 //===--- CGExpr.cpp - Emit LLVM Code from Expressions ---------------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This contains code to emit Expr nodes as LLVM code. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenFunction.h" 15 #include "CodeGenModule.h" 16 #include "CGCall.h" 17 #include "CGCXXABI.h" 18 #include "CGDebugInfo.h" 19 #include "CGRecordLayout.h" 20 #include "CGObjCRuntime.h" 21 #include "TargetInfo.h" 22 #include "clang/AST/ASTContext.h" 23 #include "clang/AST/DeclObjC.h" 24 #include "clang/Frontend/CodeGenOptions.h" 25 #include "llvm/Intrinsics.h" 26 #include "llvm/LLVMContext.h" 27 #include "llvm/Target/TargetData.h" 28 using namespace clang; 29 using namespace CodeGen; 30 31 //===--------------------------------------------------------------------===// 32 // Miscellaneous Helper Methods 33 //===--------------------------------------------------------------------===// 34 35 llvm::Value *CodeGenFunction::EmitCastToVoidPtr(llvm::Value *value) { 36 unsigned addressSpace = 37 cast<llvm::PointerType>(value->getType())->getAddressSpace(); 38 39 llvm::PointerType *destType = Int8PtrTy; 40 if (addressSpace) 41 destType = llvm::Type::getInt8PtrTy(getLLVMContext(), addressSpace); 42 43 if (value->getType() == destType) return value; 44 return Builder.CreateBitCast(value, destType); 45 } 46 47 /// CreateTempAlloca - This creates a alloca and inserts it into the entry 48 /// block. 49 llvm::AllocaInst *CodeGenFunction::CreateTempAlloca(llvm::Type *Ty, 50 const Twine &Name) { 51 if (!Builder.isNamePreserving()) 52 return new llvm::AllocaInst(Ty, 0, "", AllocaInsertPt); 53 return new llvm::AllocaInst(Ty, 0, Name, AllocaInsertPt); 54 } 55 56 void CodeGenFunction::InitTempAlloca(llvm::AllocaInst *Var, 57 llvm::Value *Init) { 58 llvm::StoreInst *Store = new llvm::StoreInst(Init, Var); 59 llvm::BasicBlock *Block = AllocaInsertPt->getParent(); 60 Block->getInstList().insertAfter(&*AllocaInsertPt, Store); 61 } 62 63 llvm::AllocaInst *CodeGenFunction::CreateIRTemp(QualType Ty, 64 const Twine &Name) { 65 llvm::AllocaInst *Alloc = CreateTempAlloca(ConvertType(Ty), Name); 66 // FIXME: Should we prefer the preferred type alignment here? 67 CharUnits Align = getContext().getTypeAlignInChars(Ty); 68 Alloc->setAlignment(Align.getQuantity()); 69 return Alloc; 70 } 71 72 llvm::AllocaInst *CodeGenFunction::CreateMemTemp(QualType Ty, 73 const Twine &Name) { 74 llvm::AllocaInst *Alloc = CreateTempAlloca(ConvertTypeForMem(Ty), Name); 75 // FIXME: Should we prefer the preferred type alignment here? 76 CharUnits Align = getContext().getTypeAlignInChars(Ty); 77 Alloc->setAlignment(Align.getQuantity()); 78 return Alloc; 79 } 80 81 /// EvaluateExprAsBool - Perform the usual unary conversions on the specified 82 /// expression and compare the result against zero, returning an Int1Ty value. 83 llvm::Value *CodeGenFunction::EvaluateExprAsBool(const Expr *E) { 84 if (const MemberPointerType *MPT = E->getType()->getAs<MemberPointerType>()) { 85 llvm::Value *MemPtr = EmitScalarExpr(E); 86 return CGM.getCXXABI().EmitMemberPointerIsNotNull(*this, MemPtr, MPT); 87 } 88 89 QualType BoolTy = getContext().BoolTy; 90 if (!E->getType()->isAnyComplexType()) 91 return EmitScalarConversion(EmitScalarExpr(E), E->getType(), BoolTy); 92 93 return EmitComplexToScalarConversion(EmitComplexExpr(E), E->getType(),BoolTy); 94 } 95 96 /// EmitIgnoredExpr - Emit code to compute the specified expression, 97 /// ignoring the result. 98 void CodeGenFunction::EmitIgnoredExpr(const Expr *E) { 99 if (E->isRValue()) 100 return (void) EmitAnyExpr(E, AggValueSlot::ignored(), true); 101 102 // Just emit it as an l-value and drop the result. 103 EmitLValue(E); 104 } 105 106 /// EmitAnyExpr - Emit code to compute the specified expression which 107 /// can have any type. The result is returned as an RValue struct. 108 /// If this is an aggregate expression, AggSlot indicates where the 109 /// result should be returned. 110 RValue CodeGenFunction::EmitAnyExpr(const Expr *E, AggValueSlot AggSlot, 111 bool IgnoreResult) { 112 if (!hasAggregateLLVMType(E->getType())) 113 return RValue::get(EmitScalarExpr(E, IgnoreResult)); 114 else if (E->getType()->isAnyComplexType()) 115 return RValue::getComplex(EmitComplexExpr(E, IgnoreResult, IgnoreResult)); 116 117 EmitAggExpr(E, AggSlot, IgnoreResult); 118 return AggSlot.asRValue(); 119 } 120 121 /// EmitAnyExprToTemp - Similary to EmitAnyExpr(), however, the result will 122 /// always be accessible even if no aggregate location is provided. 123 RValue CodeGenFunction::EmitAnyExprToTemp(const Expr *E) { 124 AggValueSlot AggSlot = AggValueSlot::ignored(); 125 126 if (hasAggregateLLVMType(E->getType()) && 127 !E->getType()->isAnyComplexType()) 128 AggSlot = CreateAggTemp(E->getType(), "agg.tmp"); 129 return EmitAnyExpr(E, AggSlot); 130 } 131 132 /// EmitAnyExprToMem - Evaluate an expression into a given memory 133 /// location. 134 void CodeGenFunction::EmitAnyExprToMem(const Expr *E, 135 llvm::Value *Location, 136 Qualifiers Quals, 137 bool IsInit) { 138 if (E->getType()->isAnyComplexType()) 139 EmitComplexExprIntoAddr(E, Location, Quals.hasVolatile()); 140 else if (hasAggregateLLVMType(E->getType())) 141 EmitAggExpr(E, AggValueSlot::forAddr(Location, Quals, 142 AggValueSlot::IsDestructed_t(IsInit), 143 AggValueSlot::DoesNotNeedGCBarriers, 144 AggValueSlot::IsAliased_t(!IsInit))); 145 else { 146 RValue RV = RValue::get(EmitScalarExpr(E, /*Ignore*/ false)); 147 LValue LV = MakeAddrLValue(Location, E->getType()); 148 EmitStoreThroughLValue(RV, LV); 149 } 150 } 151 152 namespace { 153 /// \brief An adjustment to be made to the temporary created when emitting a 154 /// reference binding, which accesses a particular subobject of that temporary. 155 struct SubobjectAdjustment { 156 enum { DerivedToBaseAdjustment, FieldAdjustment } Kind; 157 158 union { 159 struct { 160 const CastExpr *BasePath; 161 const CXXRecordDecl *DerivedClass; 162 } DerivedToBase; 163 164 FieldDecl *Field; 165 }; 166 167 SubobjectAdjustment(const CastExpr *BasePath, 168 const CXXRecordDecl *DerivedClass) 169 : Kind(DerivedToBaseAdjustment) { 170 DerivedToBase.BasePath = BasePath; 171 DerivedToBase.DerivedClass = DerivedClass; 172 } 173 174 SubobjectAdjustment(FieldDecl *Field) 175 : Kind(FieldAdjustment) { 176 this->Field = Field; 177 } 178 }; 179 } 180 181 static llvm::Value * 182 CreateReferenceTemporary(CodeGenFunction &CGF, QualType Type, 183 const NamedDecl *InitializedDecl) { 184 if (const VarDecl *VD = dyn_cast_or_null<VarDecl>(InitializedDecl)) { 185 if (VD->hasGlobalStorage()) { 186 llvm::SmallString<256> Name; 187 llvm::raw_svector_ostream Out(Name); 188 CGF.CGM.getCXXABI().getMangleContext().mangleReferenceTemporary(VD, Out); 189 Out.flush(); 190 191 llvm::Type *RefTempTy = CGF.ConvertTypeForMem(Type); 192 193 // Create the reference temporary. 194 llvm::GlobalValue *RefTemp = 195 new llvm::GlobalVariable(CGF.CGM.getModule(), 196 RefTempTy, /*isConstant=*/false, 197 llvm::GlobalValue::InternalLinkage, 198 llvm::Constant::getNullValue(RefTempTy), 199 Name.str()); 200 return RefTemp; 201 } 202 } 203 204 return CGF.CreateMemTemp(Type, "ref.tmp"); 205 } 206 207 static llvm::Value * 208 EmitExprForReferenceBinding(CodeGenFunction &CGF, const Expr *E, 209 llvm::Value *&ReferenceTemporary, 210 const CXXDestructorDecl *&ReferenceTemporaryDtor, 211 QualType &ObjCARCReferenceLifetimeType, 212 const NamedDecl *InitializedDecl) { 213 // Look through single-element init lists that claim to be lvalues. They're 214 // just syntactic wrappers in this case. 215 if (const InitListExpr *ILE = dyn_cast<InitListExpr>(E)) { 216 if (ILE->getNumInits() == 1 && ILE->isGLValue()) 217 E = ILE->getInit(0); 218 } 219 220 // Look through expressions for materialized temporaries (for now). 221 if (const MaterializeTemporaryExpr *M 222 = dyn_cast<MaterializeTemporaryExpr>(E)) { 223 // Objective-C++ ARC: 224 // If we are binding a reference to a temporary that has ownership, we 225 // need to perform retain/release operations on the temporary. 226 if (CGF.getContext().getLangOptions().ObjCAutoRefCount && 227 E->getType()->isObjCLifetimeType() && 228 (E->getType().getObjCLifetime() == Qualifiers::OCL_Strong || 229 E->getType().getObjCLifetime() == Qualifiers::OCL_Weak || 230 E->getType().getObjCLifetime() == Qualifiers::OCL_Autoreleasing)) 231 ObjCARCReferenceLifetimeType = E->getType(); 232 233 E = M->GetTemporaryExpr(); 234 } 235 236 if (const CXXDefaultArgExpr *DAE = dyn_cast<CXXDefaultArgExpr>(E)) 237 E = DAE->getExpr(); 238 239 if (const ExprWithCleanups *EWC = dyn_cast<ExprWithCleanups>(E)) { 240 CGF.enterFullExpression(EWC); 241 CodeGenFunction::RunCleanupsScope Scope(CGF); 242 243 return EmitExprForReferenceBinding(CGF, EWC->getSubExpr(), 244 ReferenceTemporary, 245 ReferenceTemporaryDtor, 246 ObjCARCReferenceLifetimeType, 247 InitializedDecl); 248 } 249 250 RValue RV; 251 if (E->isGLValue()) { 252 // Emit the expression as an lvalue. 253 LValue LV = CGF.EmitLValue(E); 254 255 if (LV.isSimple()) 256 return LV.getAddress(); 257 258 // We have to load the lvalue. 259 RV = CGF.EmitLoadOfLValue(LV); 260 } else { 261 if (!ObjCARCReferenceLifetimeType.isNull()) { 262 ReferenceTemporary = CreateReferenceTemporary(CGF, 263 ObjCARCReferenceLifetimeType, 264 InitializedDecl); 265 266 267 LValue RefTempDst = CGF.MakeAddrLValue(ReferenceTemporary, 268 ObjCARCReferenceLifetimeType); 269 270 CGF.EmitScalarInit(E, dyn_cast_or_null<ValueDecl>(InitializedDecl), 271 RefTempDst, false); 272 273 bool ExtendsLifeOfTemporary = false; 274 if (const VarDecl *Var = dyn_cast_or_null<VarDecl>(InitializedDecl)) { 275 if (Var->extendsLifetimeOfTemporary()) 276 ExtendsLifeOfTemporary = true; 277 } else if (InitializedDecl && isa<FieldDecl>(InitializedDecl)) { 278 ExtendsLifeOfTemporary = true; 279 } 280 281 if (!ExtendsLifeOfTemporary) { 282 // Since the lifetime of this temporary isn't going to be extended, 283 // we need to clean it up ourselves at the end of the full expression. 284 switch (ObjCARCReferenceLifetimeType.getObjCLifetime()) { 285 case Qualifiers::OCL_None: 286 case Qualifiers::OCL_ExplicitNone: 287 case Qualifiers::OCL_Autoreleasing: 288 break; 289 290 case Qualifiers::OCL_Strong: { 291 assert(!ObjCARCReferenceLifetimeType->isArrayType()); 292 CleanupKind cleanupKind = CGF.getARCCleanupKind(); 293 CGF.pushDestroy(cleanupKind, 294 ReferenceTemporary, 295 ObjCARCReferenceLifetimeType, 296 CodeGenFunction::destroyARCStrongImprecise, 297 cleanupKind & EHCleanup); 298 break; 299 } 300 301 case Qualifiers::OCL_Weak: 302 assert(!ObjCARCReferenceLifetimeType->isArrayType()); 303 CGF.pushDestroy(NormalAndEHCleanup, 304 ReferenceTemporary, 305 ObjCARCReferenceLifetimeType, 306 CodeGenFunction::destroyARCWeak, 307 /*useEHCleanupForArray*/ true); 308 break; 309 } 310 311 ObjCARCReferenceLifetimeType = QualType(); 312 } 313 314 return ReferenceTemporary; 315 } 316 317 SmallVector<SubobjectAdjustment, 2> Adjustments; 318 while (true) { 319 E = E->IgnoreParens(); 320 321 if (const CastExpr *CE = dyn_cast<CastExpr>(E)) { 322 if ((CE->getCastKind() == CK_DerivedToBase || 323 CE->getCastKind() == CK_UncheckedDerivedToBase) && 324 E->getType()->isRecordType()) { 325 E = CE->getSubExpr(); 326 CXXRecordDecl *Derived 327 = cast<CXXRecordDecl>(E->getType()->getAs<RecordType>()->getDecl()); 328 Adjustments.push_back(SubobjectAdjustment(CE, Derived)); 329 continue; 330 } 331 332 if (CE->getCastKind() == CK_NoOp) { 333 E = CE->getSubExpr(); 334 continue; 335 } 336 } else if (const MemberExpr *ME = dyn_cast<MemberExpr>(E)) { 337 if (!ME->isArrow() && ME->getBase()->isRValue()) { 338 assert(ME->getBase()->getType()->isRecordType()); 339 if (FieldDecl *Field = dyn_cast<FieldDecl>(ME->getMemberDecl())) { 340 E = ME->getBase(); 341 Adjustments.push_back(SubobjectAdjustment(Field)); 342 continue; 343 } 344 } 345 } 346 347 if (const OpaqueValueExpr *opaque = dyn_cast<OpaqueValueExpr>(E)) 348 if (opaque->getType()->isRecordType()) 349 return CGF.EmitOpaqueValueLValue(opaque).getAddress(); 350 351 // Nothing changed. 352 break; 353 } 354 355 // Create a reference temporary if necessary. 356 AggValueSlot AggSlot = AggValueSlot::ignored(); 357 if (CGF.hasAggregateLLVMType(E->getType()) && 358 !E->getType()->isAnyComplexType()) { 359 ReferenceTemporary = CreateReferenceTemporary(CGF, E->getType(), 360 InitializedDecl); 361 AggValueSlot::IsDestructed_t isDestructed 362 = AggValueSlot::IsDestructed_t(InitializedDecl != 0); 363 AggSlot = AggValueSlot::forAddr(ReferenceTemporary, Qualifiers(), 364 isDestructed, 365 AggValueSlot::DoesNotNeedGCBarriers, 366 AggValueSlot::IsNotAliased); 367 } 368 369 if (InitializedDecl) { 370 // Get the destructor for the reference temporary. 371 if (const RecordType *RT = E->getType()->getAs<RecordType>()) { 372 CXXRecordDecl *ClassDecl = cast<CXXRecordDecl>(RT->getDecl()); 373 if (!ClassDecl->hasTrivialDestructor()) 374 ReferenceTemporaryDtor = ClassDecl->getDestructor(); 375 } 376 } 377 378 RV = CGF.EmitAnyExpr(E, AggSlot); 379 380 // Check if need to perform derived-to-base casts and/or field accesses, to 381 // get from the temporary object we created (and, potentially, for which we 382 // extended the lifetime) to the subobject we're binding the reference to. 383 if (!Adjustments.empty()) { 384 llvm::Value *Object = RV.getAggregateAddr(); 385 for (unsigned I = Adjustments.size(); I != 0; --I) { 386 SubobjectAdjustment &Adjustment = Adjustments[I-1]; 387 switch (Adjustment.Kind) { 388 case SubobjectAdjustment::DerivedToBaseAdjustment: 389 Object = 390 CGF.GetAddressOfBaseClass(Object, 391 Adjustment.DerivedToBase.DerivedClass, 392 Adjustment.DerivedToBase.BasePath->path_begin(), 393 Adjustment.DerivedToBase.BasePath->path_end(), 394 /*NullCheckValue=*/false); 395 break; 396 397 case SubobjectAdjustment::FieldAdjustment: { 398 LValue LV = 399 CGF.EmitLValueForField(Object, Adjustment.Field, 0); 400 if (LV.isSimple()) { 401 Object = LV.getAddress(); 402 break; 403 } 404 405 // For non-simple lvalues, we actually have to create a copy of 406 // the object we're binding to. 407 QualType T = Adjustment.Field->getType().getNonReferenceType() 408 .getUnqualifiedType(); 409 Object = CreateReferenceTemporary(CGF, T, InitializedDecl); 410 LValue TempLV = CGF.MakeAddrLValue(Object, 411 Adjustment.Field->getType()); 412 CGF.EmitStoreThroughLValue(CGF.EmitLoadOfLValue(LV), TempLV); 413 break; 414 } 415 416 } 417 } 418 419 return Object; 420 } 421 } 422 423 if (RV.isAggregate()) 424 return RV.getAggregateAddr(); 425 426 // Create a temporary variable that we can bind the reference to. 427 ReferenceTemporary = CreateReferenceTemporary(CGF, E->getType(), 428 InitializedDecl); 429 430 431 unsigned Alignment = 432 CGF.getContext().getTypeAlignInChars(E->getType()).getQuantity(); 433 if (RV.isScalar()) 434 CGF.EmitStoreOfScalar(RV.getScalarVal(), ReferenceTemporary, 435 /*Volatile=*/false, Alignment, E->getType()); 436 else 437 CGF.StoreComplexToAddr(RV.getComplexVal(), ReferenceTemporary, 438 /*Volatile=*/false); 439 return ReferenceTemporary; 440 } 441 442 RValue 443 CodeGenFunction::EmitReferenceBindingToExpr(const Expr *E, 444 const NamedDecl *InitializedDecl) { 445 llvm::Value *ReferenceTemporary = 0; 446 const CXXDestructorDecl *ReferenceTemporaryDtor = 0; 447 QualType ObjCARCReferenceLifetimeType; 448 llvm::Value *Value = EmitExprForReferenceBinding(*this, E, ReferenceTemporary, 449 ReferenceTemporaryDtor, 450 ObjCARCReferenceLifetimeType, 451 InitializedDecl); 452 if (!ReferenceTemporaryDtor && ObjCARCReferenceLifetimeType.isNull()) 453 return RValue::get(Value); 454 455 // Make sure to call the destructor for the reference temporary. 456 const VarDecl *VD = dyn_cast_or_null<VarDecl>(InitializedDecl); 457 if (VD && VD->hasGlobalStorage()) { 458 if (ReferenceTemporaryDtor) { 459 llvm::Constant *DtorFn = 460 CGM.GetAddrOfCXXDestructor(ReferenceTemporaryDtor, Dtor_Complete); 461 EmitCXXGlobalDtorRegistration(DtorFn, 462 cast<llvm::Constant>(ReferenceTemporary)); 463 } else { 464 assert(!ObjCARCReferenceLifetimeType.isNull()); 465 // Note: We intentionally do not register a global "destructor" to 466 // release the object. 467 } 468 469 return RValue::get(Value); 470 } 471 472 if (ReferenceTemporaryDtor) 473 PushDestructorCleanup(ReferenceTemporaryDtor, ReferenceTemporary); 474 else { 475 switch (ObjCARCReferenceLifetimeType.getObjCLifetime()) { 476 case Qualifiers::OCL_None: 477 llvm_unreachable( 478 "Not a reference temporary that needs to be deallocated"); 479 case Qualifiers::OCL_ExplicitNone: 480 case Qualifiers::OCL_Autoreleasing: 481 // Nothing to do. 482 break; 483 484 case Qualifiers::OCL_Strong: { 485 bool precise = VD && VD->hasAttr<ObjCPreciseLifetimeAttr>(); 486 CleanupKind cleanupKind = getARCCleanupKind(); 487 // This local is a GCC and MSVC compiler workaround. 488 Destroyer *destroyer = precise ? &destroyARCStrongPrecise : 489 &destroyARCStrongImprecise; 490 pushDestroy(cleanupKind, ReferenceTemporary, ObjCARCReferenceLifetimeType, 491 *destroyer, cleanupKind & EHCleanup); 492 break; 493 } 494 495 case Qualifiers::OCL_Weak: { 496 // This local is a GCC and MSVC compiler workaround. 497 Destroyer *destroyer = &destroyARCWeak; 498 // __weak objects always get EH cleanups; otherwise, exceptions 499 // could cause really nasty crashes instead of mere leaks. 500 pushDestroy(NormalAndEHCleanup, ReferenceTemporary, 501 ObjCARCReferenceLifetimeType, *destroyer, true); 502 break; 503 } 504 } 505 } 506 507 return RValue::get(Value); 508 } 509 510 511 /// getAccessedFieldNo - Given an encoded value and a result number, return the 512 /// input field number being accessed. 513 unsigned CodeGenFunction::getAccessedFieldNo(unsigned Idx, 514 const llvm::Constant *Elts) { 515 if (isa<llvm::ConstantAggregateZero>(Elts)) 516 return 0; 517 518 return cast<llvm::ConstantInt>(Elts->getOperand(Idx))->getZExtValue(); 519 } 520 521 void CodeGenFunction::EmitCheck(llvm::Value *Address, unsigned Size) { 522 if (!CatchUndefined) 523 return; 524 525 // This needs to be to the standard address space. 526 Address = Builder.CreateBitCast(Address, Int8PtrTy); 527 528 llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::objectsize, IntPtrTy); 529 530 // In time, people may want to control this and use a 1 here. 531 llvm::Value *Arg = Builder.getFalse(); 532 llvm::Value *C = Builder.CreateCall2(F, Address, Arg); 533 llvm::BasicBlock *Cont = createBasicBlock(); 534 llvm::BasicBlock *Check = createBasicBlock(); 535 llvm::Value *NegativeOne = llvm::ConstantInt::get(IntPtrTy, -1ULL); 536 Builder.CreateCondBr(Builder.CreateICmpEQ(C, NegativeOne), Cont, Check); 537 538 EmitBlock(Check); 539 Builder.CreateCondBr(Builder.CreateICmpUGE(C, 540 llvm::ConstantInt::get(IntPtrTy, Size)), 541 Cont, getTrapBB()); 542 EmitBlock(Cont); 543 } 544 545 546 CodeGenFunction::ComplexPairTy CodeGenFunction:: 547 EmitComplexPrePostIncDec(const UnaryOperator *E, LValue LV, 548 bool isInc, bool isPre) { 549 ComplexPairTy InVal = LoadComplexFromAddr(LV.getAddress(), 550 LV.isVolatileQualified()); 551 552 llvm::Value *NextVal; 553 if (isa<llvm::IntegerType>(InVal.first->getType())) { 554 uint64_t AmountVal = isInc ? 1 : -1; 555 NextVal = llvm::ConstantInt::get(InVal.first->getType(), AmountVal, true); 556 557 // Add the inc/dec to the real part. 558 NextVal = Builder.CreateAdd(InVal.first, NextVal, isInc ? "inc" : "dec"); 559 } else { 560 QualType ElemTy = E->getType()->getAs<ComplexType>()->getElementType(); 561 llvm::APFloat FVal(getContext().getFloatTypeSemantics(ElemTy), 1); 562 if (!isInc) 563 FVal.changeSign(); 564 NextVal = llvm::ConstantFP::get(getLLVMContext(), FVal); 565 566 // Add the inc/dec to the real part. 567 NextVal = Builder.CreateFAdd(InVal.first, NextVal, isInc ? "inc" : "dec"); 568 } 569 570 ComplexPairTy IncVal(NextVal, InVal.second); 571 572 // Store the updated result through the lvalue. 573 StoreComplexToAddr(IncVal, LV.getAddress(), LV.isVolatileQualified()); 574 575 // If this is a postinc, return the value read from memory, otherwise use the 576 // updated value. 577 return isPre ? IncVal : InVal; 578 } 579 580 581 //===----------------------------------------------------------------------===// 582 // LValue Expression Emission 583 //===----------------------------------------------------------------------===// 584 585 RValue CodeGenFunction::GetUndefRValue(QualType Ty) { 586 if (Ty->isVoidType()) 587 return RValue::get(0); 588 589 if (const ComplexType *CTy = Ty->getAs<ComplexType>()) { 590 llvm::Type *EltTy = ConvertType(CTy->getElementType()); 591 llvm::Value *U = llvm::UndefValue::get(EltTy); 592 return RValue::getComplex(std::make_pair(U, U)); 593 } 594 595 // If this is a use of an undefined aggregate type, the aggregate must have an 596 // identifiable address. Just because the contents of the value are undefined 597 // doesn't mean that the address can't be taken and compared. 598 if (hasAggregateLLVMType(Ty)) { 599 llvm::Value *DestPtr = CreateMemTemp(Ty, "undef.agg.tmp"); 600 return RValue::getAggregate(DestPtr); 601 } 602 603 return RValue::get(llvm::UndefValue::get(ConvertType(Ty))); 604 } 605 606 RValue CodeGenFunction::EmitUnsupportedRValue(const Expr *E, 607 const char *Name) { 608 ErrorUnsupported(E, Name); 609 return GetUndefRValue(E->getType()); 610 } 611 612 LValue CodeGenFunction::EmitUnsupportedLValue(const Expr *E, 613 const char *Name) { 614 ErrorUnsupported(E, Name); 615 llvm::Type *Ty = llvm::PointerType::getUnqual(ConvertType(E->getType())); 616 return MakeAddrLValue(llvm::UndefValue::get(Ty), E->getType()); 617 } 618 619 LValue CodeGenFunction::EmitCheckedLValue(const Expr *E) { 620 LValue LV = EmitLValue(E); 621 if (!isa<DeclRefExpr>(E) && !LV.isBitField() && LV.isSimple()) 622 EmitCheck(LV.getAddress(), 623 getContext().getTypeSizeInChars(E->getType()).getQuantity()); 624 return LV; 625 } 626 627 /// EmitLValue - Emit code to compute a designator that specifies the location 628 /// of the expression. 629 /// 630 /// This can return one of two things: a simple address or a bitfield reference. 631 /// In either case, the LLVM Value* in the LValue structure is guaranteed to be 632 /// an LLVM pointer type. 633 /// 634 /// If this returns a bitfield reference, nothing about the pointee type of the 635 /// LLVM value is known: For example, it may not be a pointer to an integer. 636 /// 637 /// If this returns a normal address, and if the lvalue's C type is fixed size, 638 /// this method guarantees that the returned pointer type will point to an LLVM 639 /// type of the same size of the lvalue's type. If the lvalue has a variable 640 /// length type, this is not possible. 641 /// 642 LValue CodeGenFunction::EmitLValue(const Expr *E) { 643 switch (E->getStmtClass()) { 644 default: return EmitUnsupportedLValue(E, "l-value expression"); 645 646 case Expr::ObjCPropertyRefExprClass: 647 llvm_unreachable("cannot emit a property reference directly"); 648 649 case Expr::ObjCSelectorExprClass: 650 return EmitObjCSelectorLValue(cast<ObjCSelectorExpr>(E)); 651 case Expr::ObjCIsaExprClass: 652 return EmitObjCIsaExpr(cast<ObjCIsaExpr>(E)); 653 case Expr::BinaryOperatorClass: 654 return EmitBinaryOperatorLValue(cast<BinaryOperator>(E)); 655 case Expr::CompoundAssignOperatorClass: 656 if (!E->getType()->isAnyComplexType()) 657 return EmitCompoundAssignmentLValue(cast<CompoundAssignOperator>(E)); 658 return EmitComplexCompoundAssignmentLValue(cast<CompoundAssignOperator>(E)); 659 case Expr::CallExprClass: 660 case Expr::CXXMemberCallExprClass: 661 case Expr::CXXOperatorCallExprClass: 662 return EmitCallExprLValue(cast<CallExpr>(E)); 663 case Expr::VAArgExprClass: 664 return EmitVAArgExprLValue(cast<VAArgExpr>(E)); 665 case Expr::DeclRefExprClass: 666 return EmitDeclRefLValue(cast<DeclRefExpr>(E)); 667 case Expr::ParenExprClass: 668 return EmitLValue(cast<ParenExpr>(E)->getSubExpr()); 669 case Expr::GenericSelectionExprClass: 670 return EmitLValue(cast<GenericSelectionExpr>(E)->getResultExpr()); 671 case Expr::PredefinedExprClass: 672 return EmitPredefinedLValue(cast<PredefinedExpr>(E)); 673 case Expr::StringLiteralClass: 674 return EmitStringLiteralLValue(cast<StringLiteral>(E)); 675 case Expr::ObjCEncodeExprClass: 676 return EmitObjCEncodeExprLValue(cast<ObjCEncodeExpr>(E)); 677 case Expr::PseudoObjectExprClass: 678 return EmitPseudoObjectLValue(cast<PseudoObjectExpr>(E)); 679 case Expr::InitListExprClass: 680 assert(cast<InitListExpr>(E)->getNumInits() == 1 && 681 "Only single-element init list can be lvalue."); 682 return EmitLValue(cast<InitListExpr>(E)->getInit(0)); 683 684 case Expr::BlockDeclRefExprClass: 685 return EmitBlockDeclRefLValue(cast<BlockDeclRefExpr>(E)); 686 687 case Expr::CXXTemporaryObjectExprClass: 688 case Expr::CXXConstructExprClass: 689 return EmitCXXConstructLValue(cast<CXXConstructExpr>(E)); 690 case Expr::CXXBindTemporaryExprClass: 691 return EmitCXXBindTemporaryLValue(cast<CXXBindTemporaryExpr>(E)); 692 693 case Expr::ExprWithCleanupsClass: { 694 const ExprWithCleanups *cleanups = cast<ExprWithCleanups>(E); 695 enterFullExpression(cleanups); 696 RunCleanupsScope Scope(*this); 697 return EmitLValue(cleanups->getSubExpr()); 698 } 699 700 case Expr::CXXScalarValueInitExprClass: 701 return EmitNullInitializationLValue(cast<CXXScalarValueInitExpr>(E)); 702 case Expr::CXXDefaultArgExprClass: 703 return EmitLValue(cast<CXXDefaultArgExpr>(E)->getExpr()); 704 case Expr::CXXTypeidExprClass: 705 return EmitCXXTypeidLValue(cast<CXXTypeidExpr>(E)); 706 707 case Expr::ObjCMessageExprClass: 708 return EmitObjCMessageExprLValue(cast<ObjCMessageExpr>(E)); 709 case Expr::ObjCIvarRefExprClass: 710 return EmitObjCIvarRefLValue(cast<ObjCIvarRefExpr>(E)); 711 case Expr::StmtExprClass: 712 return EmitStmtExprLValue(cast<StmtExpr>(E)); 713 case Expr::UnaryOperatorClass: 714 return EmitUnaryOpLValue(cast<UnaryOperator>(E)); 715 case Expr::ArraySubscriptExprClass: 716 return EmitArraySubscriptExpr(cast<ArraySubscriptExpr>(E)); 717 case Expr::ExtVectorElementExprClass: 718 return EmitExtVectorElementExpr(cast<ExtVectorElementExpr>(E)); 719 case Expr::MemberExprClass: 720 return EmitMemberExpr(cast<MemberExpr>(E)); 721 case Expr::CompoundLiteralExprClass: 722 return EmitCompoundLiteralLValue(cast<CompoundLiteralExpr>(E)); 723 case Expr::ConditionalOperatorClass: 724 return EmitConditionalOperatorLValue(cast<ConditionalOperator>(E)); 725 case Expr::BinaryConditionalOperatorClass: 726 return EmitConditionalOperatorLValue(cast<BinaryConditionalOperator>(E)); 727 case Expr::ChooseExprClass: 728 return EmitLValue(cast<ChooseExpr>(E)->getChosenSubExpr(getContext())); 729 case Expr::OpaqueValueExprClass: 730 return EmitOpaqueValueLValue(cast<OpaqueValueExpr>(E)); 731 case Expr::SubstNonTypeTemplateParmExprClass: 732 return EmitLValue(cast<SubstNonTypeTemplateParmExpr>(E)->getReplacement()); 733 case Expr::ImplicitCastExprClass: 734 case Expr::CStyleCastExprClass: 735 case Expr::CXXFunctionalCastExprClass: 736 case Expr::CXXStaticCastExprClass: 737 case Expr::CXXDynamicCastExprClass: 738 case Expr::CXXReinterpretCastExprClass: 739 case Expr::CXXConstCastExprClass: 740 case Expr::ObjCBridgedCastExprClass: 741 return EmitCastLValue(cast<CastExpr>(E)); 742 743 case Expr::MaterializeTemporaryExprClass: 744 return EmitMaterializeTemporaryExpr(cast<MaterializeTemporaryExpr>(E)); 745 } 746 } 747 748 llvm::Value *CodeGenFunction::EmitLoadOfScalar(LValue lvalue) { 749 return EmitLoadOfScalar(lvalue.getAddress(), lvalue.isVolatile(), 750 lvalue.getAlignment(), lvalue.getType(), 751 lvalue.getTBAAInfo()); 752 } 753 754 llvm::Value *CodeGenFunction::EmitLoadOfScalar(llvm::Value *Addr, bool Volatile, 755 unsigned Alignment, QualType Ty, 756 llvm::MDNode *TBAAInfo) { 757 llvm::LoadInst *Load = Builder.CreateLoad(Addr); 758 if (Volatile) 759 Load->setVolatile(true); 760 if (Alignment) 761 Load->setAlignment(Alignment); 762 if (TBAAInfo) 763 CGM.DecorateInstruction(Load, TBAAInfo); 764 765 return EmitFromMemory(Load, Ty); 766 } 767 768 static bool isBooleanUnderlyingType(QualType Ty) { 769 if (const EnumType *ET = dyn_cast<EnumType>(Ty)) 770 return ET->getDecl()->getIntegerType()->isBooleanType(); 771 return false; 772 } 773 774 llvm::Value *CodeGenFunction::EmitToMemory(llvm::Value *Value, QualType Ty) { 775 // Bool has a different representation in memory than in registers. 776 if (Ty->isBooleanType() || isBooleanUnderlyingType(Ty)) { 777 // This should really always be an i1, but sometimes it's already 778 // an i8, and it's awkward to track those cases down. 779 if (Value->getType()->isIntegerTy(1)) 780 return Builder.CreateZExt(Value, Builder.getInt8Ty(), "frombool"); 781 assert(Value->getType()->isIntegerTy(8) && "value rep of bool not i1/i8"); 782 } 783 784 return Value; 785 } 786 787 llvm::Value *CodeGenFunction::EmitFromMemory(llvm::Value *Value, QualType Ty) { 788 // Bool has a different representation in memory than in registers. 789 if (Ty->isBooleanType() || isBooleanUnderlyingType(Ty)) { 790 assert(Value->getType()->isIntegerTy(8) && "memory rep of bool not i8"); 791 return Builder.CreateTrunc(Value, Builder.getInt1Ty(), "tobool"); 792 } 793 794 return Value; 795 } 796 797 void CodeGenFunction::EmitStoreOfScalar(llvm::Value *Value, llvm::Value *Addr, 798 bool Volatile, unsigned Alignment, 799 QualType Ty, 800 llvm::MDNode *TBAAInfo) { 801 Value = EmitToMemory(Value, Ty); 802 803 llvm::StoreInst *Store = Builder.CreateStore(Value, Addr, Volatile); 804 if (Alignment) 805 Store->setAlignment(Alignment); 806 if (TBAAInfo) 807 CGM.DecorateInstruction(Store, TBAAInfo); 808 } 809 810 void CodeGenFunction::EmitStoreOfScalar(llvm::Value *value, LValue lvalue) { 811 EmitStoreOfScalar(value, lvalue.getAddress(), lvalue.isVolatile(), 812 lvalue.getAlignment(), lvalue.getType(), 813 lvalue.getTBAAInfo()); 814 } 815 816 /// EmitLoadOfLValue - Given an expression that represents a value lvalue, this 817 /// method emits the address of the lvalue, then loads the result as an rvalue, 818 /// returning the rvalue. 819 RValue CodeGenFunction::EmitLoadOfLValue(LValue LV) { 820 if (LV.isObjCWeak()) { 821 // load of a __weak object. 822 llvm::Value *AddrWeakObj = LV.getAddress(); 823 return RValue::get(CGM.getObjCRuntime().EmitObjCWeakRead(*this, 824 AddrWeakObj)); 825 } 826 if (LV.getQuals().getObjCLifetime() == Qualifiers::OCL_Weak) 827 return RValue::get(EmitARCLoadWeak(LV.getAddress())); 828 829 if (LV.isSimple()) { 830 assert(!LV.getType()->isFunctionType()); 831 832 // Everything needs a load. 833 return RValue::get(EmitLoadOfScalar(LV)); 834 } 835 836 if (LV.isVectorElt()) { 837 llvm::Value *Vec = Builder.CreateLoad(LV.getVectorAddr(), 838 LV.isVolatileQualified()); 839 return RValue::get(Builder.CreateExtractElement(Vec, LV.getVectorIdx(), 840 "vecext")); 841 } 842 843 // If this is a reference to a subset of the elements of a vector, either 844 // shuffle the input or extract/insert them as appropriate. 845 if (LV.isExtVectorElt()) 846 return EmitLoadOfExtVectorElementLValue(LV); 847 848 assert(LV.isBitField() && "Unknown LValue type!"); 849 return EmitLoadOfBitfieldLValue(LV); 850 } 851 852 RValue CodeGenFunction::EmitLoadOfBitfieldLValue(LValue LV) { 853 const CGBitFieldInfo &Info = LV.getBitFieldInfo(); 854 855 // Get the output type. 856 llvm::Type *ResLTy = ConvertType(LV.getType()); 857 unsigned ResSizeInBits = CGM.getTargetData().getTypeSizeInBits(ResLTy); 858 859 // Compute the result as an OR of all of the individual component accesses. 860 llvm::Value *Res = 0; 861 for (unsigned i = 0, e = Info.getNumComponents(); i != e; ++i) { 862 const CGBitFieldInfo::AccessInfo &AI = Info.getComponent(i); 863 864 // Get the field pointer. 865 llvm::Value *Ptr = LV.getBitFieldBaseAddr(); 866 867 // Only offset by the field index if used, so that incoming values are not 868 // required to be structures. 869 if (AI.FieldIndex) 870 Ptr = Builder.CreateStructGEP(Ptr, AI.FieldIndex, "bf.field"); 871 872 // Offset by the byte offset, if used. 873 if (!AI.FieldByteOffset.isZero()) { 874 Ptr = EmitCastToVoidPtr(Ptr); 875 Ptr = Builder.CreateConstGEP1_32(Ptr, AI.FieldByteOffset.getQuantity(), 876 "bf.field.offs"); 877 } 878 879 // Cast to the access type. 880 llvm::Type *PTy = llvm::Type::getIntNPtrTy(getLLVMContext(), 881 AI.AccessWidth, 882 CGM.getContext().getTargetAddressSpace(LV.getType())); 883 Ptr = Builder.CreateBitCast(Ptr, PTy); 884 885 // Perform the load. 886 llvm::LoadInst *Load = Builder.CreateLoad(Ptr, LV.isVolatileQualified()); 887 if (!AI.AccessAlignment.isZero()) 888 Load->setAlignment(AI.AccessAlignment.getQuantity()); 889 890 // Shift out unused low bits and mask out unused high bits. 891 llvm::Value *Val = Load; 892 if (AI.FieldBitStart) 893 Val = Builder.CreateLShr(Load, AI.FieldBitStart); 894 Val = Builder.CreateAnd(Val, llvm::APInt::getLowBitsSet(AI.AccessWidth, 895 AI.TargetBitWidth), 896 "bf.clear"); 897 898 // Extend or truncate to the target size. 899 if (AI.AccessWidth < ResSizeInBits) 900 Val = Builder.CreateZExt(Val, ResLTy); 901 else if (AI.AccessWidth > ResSizeInBits) 902 Val = Builder.CreateTrunc(Val, ResLTy); 903 904 // Shift into place, and OR into the result. 905 if (AI.TargetBitOffset) 906 Val = Builder.CreateShl(Val, AI.TargetBitOffset); 907 Res = Res ? Builder.CreateOr(Res, Val) : Val; 908 } 909 910 // If the bit-field is signed, perform the sign-extension. 911 // 912 // FIXME: This can easily be folded into the load of the high bits, which 913 // could also eliminate the mask of high bits in some situations. 914 if (Info.isSigned()) { 915 unsigned ExtraBits = ResSizeInBits - Info.getSize(); 916 if (ExtraBits) 917 Res = Builder.CreateAShr(Builder.CreateShl(Res, ExtraBits), 918 ExtraBits, "bf.val.sext"); 919 } 920 921 return RValue::get(Res); 922 } 923 924 // If this is a reference to a subset of the elements of a vector, create an 925 // appropriate shufflevector. 926 RValue CodeGenFunction::EmitLoadOfExtVectorElementLValue(LValue LV) { 927 llvm::Value *Vec = Builder.CreateLoad(LV.getExtVectorAddr(), 928 LV.isVolatileQualified()); 929 930 const llvm::Constant *Elts = LV.getExtVectorElts(); 931 932 // If the result of the expression is a non-vector type, we must be extracting 933 // a single element. Just codegen as an extractelement. 934 const VectorType *ExprVT = LV.getType()->getAs<VectorType>(); 935 if (!ExprVT) { 936 unsigned InIdx = getAccessedFieldNo(0, Elts); 937 llvm::Value *Elt = llvm::ConstantInt::get(Int32Ty, InIdx); 938 return RValue::get(Builder.CreateExtractElement(Vec, Elt)); 939 } 940 941 // Always use shuffle vector to try to retain the original program structure 942 unsigned NumResultElts = ExprVT->getNumElements(); 943 944 SmallVector<llvm::Constant*, 4> Mask; 945 for (unsigned i = 0; i != NumResultElts; ++i) { 946 unsigned InIdx = getAccessedFieldNo(i, Elts); 947 Mask.push_back(llvm::ConstantInt::get(Int32Ty, InIdx)); 948 } 949 950 llvm::Value *MaskV = llvm::ConstantVector::get(Mask); 951 Vec = Builder.CreateShuffleVector(Vec, llvm::UndefValue::get(Vec->getType()), 952 MaskV); 953 return RValue::get(Vec); 954 } 955 956 957 958 /// EmitStoreThroughLValue - Store the specified rvalue into the specified 959 /// lvalue, where both are guaranteed to the have the same type, and that type 960 /// is 'Ty'. 961 void CodeGenFunction::EmitStoreThroughLValue(RValue Src, LValue Dst) { 962 if (!Dst.isSimple()) { 963 if (Dst.isVectorElt()) { 964 // Read/modify/write the vector, inserting the new element. 965 llvm::Value *Vec = Builder.CreateLoad(Dst.getVectorAddr(), 966 Dst.isVolatileQualified()); 967 Vec = Builder.CreateInsertElement(Vec, Src.getScalarVal(), 968 Dst.getVectorIdx(), "vecins"); 969 Builder.CreateStore(Vec, Dst.getVectorAddr(),Dst.isVolatileQualified()); 970 return; 971 } 972 973 // If this is an update of extended vector elements, insert them as 974 // appropriate. 975 if (Dst.isExtVectorElt()) 976 return EmitStoreThroughExtVectorComponentLValue(Src, Dst); 977 978 assert(Dst.isBitField() && "Unknown LValue type"); 979 return EmitStoreThroughBitfieldLValue(Src, Dst); 980 } 981 982 // There's special magic for assigning into an ARC-qualified l-value. 983 if (Qualifiers::ObjCLifetime Lifetime = Dst.getQuals().getObjCLifetime()) { 984 switch (Lifetime) { 985 case Qualifiers::OCL_None: 986 llvm_unreachable("present but none"); 987 988 case Qualifiers::OCL_ExplicitNone: 989 // nothing special 990 break; 991 992 case Qualifiers::OCL_Strong: 993 EmitARCStoreStrong(Dst, Src.getScalarVal(), /*ignore*/ true); 994 return; 995 996 case Qualifiers::OCL_Weak: 997 EmitARCStoreWeak(Dst.getAddress(), Src.getScalarVal(), /*ignore*/ true); 998 return; 999 1000 case Qualifiers::OCL_Autoreleasing: 1001 Src = RValue::get(EmitObjCExtendObjectLifetime(Dst.getType(), 1002 Src.getScalarVal())); 1003 // fall into the normal path 1004 break; 1005 } 1006 } 1007 1008 if (Dst.isObjCWeak() && !Dst.isNonGC()) { 1009 // load of a __weak object. 1010 llvm::Value *LvalueDst = Dst.getAddress(); 1011 llvm::Value *src = Src.getScalarVal(); 1012 CGM.getObjCRuntime().EmitObjCWeakAssign(*this, src, LvalueDst); 1013 return; 1014 } 1015 1016 if (Dst.isObjCStrong() && !Dst.isNonGC()) { 1017 // load of a __strong object. 1018 llvm::Value *LvalueDst = Dst.getAddress(); 1019 llvm::Value *src = Src.getScalarVal(); 1020 if (Dst.isObjCIvar()) { 1021 assert(Dst.getBaseIvarExp() && "BaseIvarExp is NULL"); 1022 llvm::Type *ResultType = ConvertType(getContext().LongTy); 1023 llvm::Value *RHS = EmitScalarExpr(Dst.getBaseIvarExp()); 1024 llvm::Value *dst = RHS; 1025 RHS = Builder.CreatePtrToInt(RHS, ResultType, "sub.ptr.rhs.cast"); 1026 llvm::Value *LHS = 1027 Builder.CreatePtrToInt(LvalueDst, ResultType, "sub.ptr.lhs.cast"); 1028 llvm::Value *BytesBetween = Builder.CreateSub(LHS, RHS, "ivar.offset"); 1029 CGM.getObjCRuntime().EmitObjCIvarAssign(*this, src, dst, 1030 BytesBetween); 1031 } else if (Dst.isGlobalObjCRef()) { 1032 CGM.getObjCRuntime().EmitObjCGlobalAssign(*this, src, LvalueDst, 1033 Dst.isThreadLocalRef()); 1034 } 1035 else 1036 CGM.getObjCRuntime().EmitObjCStrongCastAssign(*this, src, LvalueDst); 1037 return; 1038 } 1039 1040 assert(Src.isScalar() && "Can't emit an agg store with this method"); 1041 EmitStoreOfScalar(Src.getScalarVal(), Dst); 1042 } 1043 1044 void CodeGenFunction::EmitStoreThroughBitfieldLValue(RValue Src, LValue Dst, 1045 llvm::Value **Result) { 1046 const CGBitFieldInfo &Info = Dst.getBitFieldInfo(); 1047 1048 // Get the output type. 1049 llvm::Type *ResLTy = ConvertTypeForMem(Dst.getType()); 1050 unsigned ResSizeInBits = CGM.getTargetData().getTypeSizeInBits(ResLTy); 1051 1052 // Get the source value, truncated to the width of the bit-field. 1053 llvm::Value *SrcVal = Src.getScalarVal(); 1054 1055 if (Dst.getType()->isBooleanType()) 1056 SrcVal = Builder.CreateIntCast(SrcVal, ResLTy, /*IsSigned=*/false); 1057 1058 SrcVal = Builder.CreateAnd(SrcVal, llvm::APInt::getLowBitsSet(ResSizeInBits, 1059 Info.getSize()), 1060 "bf.value"); 1061 1062 // Return the new value of the bit-field, if requested. 1063 if (Result) { 1064 // Cast back to the proper type for result. 1065 llvm::Type *SrcTy = Src.getScalarVal()->getType(); 1066 llvm::Value *ReloadVal = Builder.CreateIntCast(SrcVal, SrcTy, false, 1067 "bf.reload.val"); 1068 1069 // Sign extend if necessary. 1070 if (Info.isSigned()) { 1071 unsigned ExtraBits = ResSizeInBits - Info.getSize(); 1072 if (ExtraBits) 1073 ReloadVal = Builder.CreateAShr(Builder.CreateShl(ReloadVal, ExtraBits), 1074 ExtraBits, "bf.reload.sext"); 1075 } 1076 1077 *Result = ReloadVal; 1078 } 1079 1080 // Iterate over the components, writing each piece to memory. 1081 for (unsigned i = 0, e = Info.getNumComponents(); i != e; ++i) { 1082 const CGBitFieldInfo::AccessInfo &AI = Info.getComponent(i); 1083 1084 // Get the field pointer. 1085 llvm::Value *Ptr = Dst.getBitFieldBaseAddr(); 1086 unsigned addressSpace = 1087 cast<llvm::PointerType>(Ptr->getType())->getAddressSpace(); 1088 1089 // Only offset by the field index if used, so that incoming values are not 1090 // required to be structures. 1091 if (AI.FieldIndex) 1092 Ptr = Builder.CreateStructGEP(Ptr, AI.FieldIndex, "bf.field"); 1093 1094 // Offset by the byte offset, if used. 1095 if (!AI.FieldByteOffset.isZero()) { 1096 Ptr = EmitCastToVoidPtr(Ptr); 1097 Ptr = Builder.CreateConstGEP1_32(Ptr, AI.FieldByteOffset.getQuantity(), 1098 "bf.field.offs"); 1099 } 1100 1101 // Cast to the access type. 1102 llvm::Type *AccessLTy = 1103 llvm::Type::getIntNTy(getLLVMContext(), AI.AccessWidth); 1104 1105 llvm::Type *PTy = AccessLTy->getPointerTo(addressSpace); 1106 Ptr = Builder.CreateBitCast(Ptr, PTy); 1107 1108 // Extract the piece of the bit-field value to write in this access, limited 1109 // to the values that are part of this access. 1110 llvm::Value *Val = SrcVal; 1111 if (AI.TargetBitOffset) 1112 Val = Builder.CreateLShr(Val, AI.TargetBitOffset); 1113 Val = Builder.CreateAnd(Val, llvm::APInt::getLowBitsSet(ResSizeInBits, 1114 AI.TargetBitWidth)); 1115 1116 // Extend or truncate to the access size. 1117 if (ResSizeInBits < AI.AccessWidth) 1118 Val = Builder.CreateZExt(Val, AccessLTy); 1119 else if (ResSizeInBits > AI.AccessWidth) 1120 Val = Builder.CreateTrunc(Val, AccessLTy); 1121 1122 // Shift into the position in memory. 1123 if (AI.FieldBitStart) 1124 Val = Builder.CreateShl(Val, AI.FieldBitStart); 1125 1126 // If necessary, load and OR in bits that are outside of the bit-field. 1127 if (AI.TargetBitWidth != AI.AccessWidth) { 1128 llvm::LoadInst *Load = Builder.CreateLoad(Ptr, Dst.isVolatileQualified()); 1129 if (!AI.AccessAlignment.isZero()) 1130 Load->setAlignment(AI.AccessAlignment.getQuantity()); 1131 1132 // Compute the mask for zeroing the bits that are part of the bit-field. 1133 llvm::APInt InvMask = 1134 ~llvm::APInt::getBitsSet(AI.AccessWidth, AI.FieldBitStart, 1135 AI.FieldBitStart + AI.TargetBitWidth); 1136 1137 // Apply the mask and OR in to the value to write. 1138 Val = Builder.CreateOr(Builder.CreateAnd(Load, InvMask), Val); 1139 } 1140 1141 // Write the value. 1142 llvm::StoreInst *Store = Builder.CreateStore(Val, Ptr, 1143 Dst.isVolatileQualified()); 1144 if (!AI.AccessAlignment.isZero()) 1145 Store->setAlignment(AI.AccessAlignment.getQuantity()); 1146 } 1147 } 1148 1149 void CodeGenFunction::EmitStoreThroughExtVectorComponentLValue(RValue Src, 1150 LValue Dst) { 1151 // This access turns into a read/modify/write of the vector. Load the input 1152 // value now. 1153 llvm::Value *Vec = Builder.CreateLoad(Dst.getExtVectorAddr(), 1154 Dst.isVolatileQualified()); 1155 const llvm::Constant *Elts = Dst.getExtVectorElts(); 1156 1157 llvm::Value *SrcVal = Src.getScalarVal(); 1158 1159 if (const VectorType *VTy = Dst.getType()->getAs<VectorType>()) { 1160 unsigned NumSrcElts = VTy->getNumElements(); 1161 unsigned NumDstElts = 1162 cast<llvm::VectorType>(Vec->getType())->getNumElements(); 1163 if (NumDstElts == NumSrcElts) { 1164 // Use shuffle vector is the src and destination are the same number of 1165 // elements and restore the vector mask since it is on the side it will be 1166 // stored. 1167 SmallVector<llvm::Constant*, 4> Mask(NumDstElts); 1168 for (unsigned i = 0; i != NumSrcElts; ++i) { 1169 unsigned InIdx = getAccessedFieldNo(i, Elts); 1170 Mask[InIdx] = llvm::ConstantInt::get(Int32Ty, i); 1171 } 1172 1173 llvm::Value *MaskV = llvm::ConstantVector::get(Mask); 1174 Vec = Builder.CreateShuffleVector(SrcVal, 1175 llvm::UndefValue::get(Vec->getType()), 1176 MaskV); 1177 } else if (NumDstElts > NumSrcElts) { 1178 // Extended the source vector to the same length and then shuffle it 1179 // into the destination. 1180 // FIXME: since we're shuffling with undef, can we just use the indices 1181 // into that? This could be simpler. 1182 SmallVector<llvm::Constant*, 4> ExtMask; 1183 unsigned i; 1184 for (i = 0; i != NumSrcElts; ++i) 1185 ExtMask.push_back(llvm::ConstantInt::get(Int32Ty, i)); 1186 for (; i != NumDstElts; ++i) 1187 ExtMask.push_back(llvm::UndefValue::get(Int32Ty)); 1188 llvm::Value *ExtMaskV = llvm::ConstantVector::get(ExtMask); 1189 llvm::Value *ExtSrcVal = 1190 Builder.CreateShuffleVector(SrcVal, 1191 llvm::UndefValue::get(SrcVal->getType()), 1192 ExtMaskV); 1193 // build identity 1194 SmallVector<llvm::Constant*, 4> Mask; 1195 for (unsigned i = 0; i != NumDstElts; ++i) 1196 Mask.push_back(llvm::ConstantInt::get(Int32Ty, i)); 1197 1198 // modify when what gets shuffled in 1199 for (unsigned i = 0; i != NumSrcElts; ++i) { 1200 unsigned Idx = getAccessedFieldNo(i, Elts); 1201 Mask[Idx] = llvm::ConstantInt::get(Int32Ty, i+NumDstElts); 1202 } 1203 llvm::Value *MaskV = llvm::ConstantVector::get(Mask); 1204 Vec = Builder.CreateShuffleVector(Vec, ExtSrcVal, MaskV); 1205 } else { 1206 // We should never shorten the vector 1207 llvm_unreachable("unexpected shorten vector length"); 1208 } 1209 } else { 1210 // If the Src is a scalar (not a vector) it must be updating one element. 1211 unsigned InIdx = getAccessedFieldNo(0, Elts); 1212 llvm::Value *Elt = llvm::ConstantInt::get(Int32Ty, InIdx); 1213 Vec = Builder.CreateInsertElement(Vec, SrcVal, Elt); 1214 } 1215 1216 Builder.CreateStore(Vec, Dst.getExtVectorAddr(), Dst.isVolatileQualified()); 1217 } 1218 1219 // setObjCGCLValueClass - sets class of he lvalue for the purpose of 1220 // generating write-barries API. It is currently a global, ivar, 1221 // or neither. 1222 static void setObjCGCLValueClass(const ASTContext &Ctx, const Expr *E, 1223 LValue &LV, 1224 bool IsMemberAccess=false) { 1225 if (Ctx.getLangOptions().getGC() == LangOptions::NonGC) 1226 return; 1227 1228 if (isa<ObjCIvarRefExpr>(E)) { 1229 QualType ExpTy = E->getType(); 1230 if (IsMemberAccess && ExpTy->isPointerType()) { 1231 // If ivar is a structure pointer, assigning to field of 1232 // this struct follows gcc's behavior and makes it a non-ivar 1233 // writer-barrier conservatively. 1234 ExpTy = ExpTy->getAs<PointerType>()->getPointeeType(); 1235 if (ExpTy->isRecordType()) { 1236 LV.setObjCIvar(false); 1237 return; 1238 } 1239 } 1240 LV.setObjCIvar(true); 1241 ObjCIvarRefExpr *Exp = cast<ObjCIvarRefExpr>(const_cast<Expr*>(E)); 1242 LV.setBaseIvarExp(Exp->getBase()); 1243 LV.setObjCArray(E->getType()->isArrayType()); 1244 return; 1245 } 1246 1247 if (const DeclRefExpr *Exp = dyn_cast<DeclRefExpr>(E)) { 1248 if (const VarDecl *VD = dyn_cast<VarDecl>(Exp->getDecl())) { 1249 if (VD->hasGlobalStorage()) { 1250 LV.setGlobalObjCRef(true); 1251 LV.setThreadLocalRef(VD->isThreadSpecified()); 1252 } 1253 } 1254 LV.setObjCArray(E->getType()->isArrayType()); 1255 return; 1256 } 1257 1258 if (const UnaryOperator *Exp = dyn_cast<UnaryOperator>(E)) { 1259 setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess); 1260 return; 1261 } 1262 1263 if (const ParenExpr *Exp = dyn_cast<ParenExpr>(E)) { 1264 setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess); 1265 if (LV.isObjCIvar()) { 1266 // If cast is to a structure pointer, follow gcc's behavior and make it 1267 // a non-ivar write-barrier. 1268 QualType ExpTy = E->getType(); 1269 if (ExpTy->isPointerType()) 1270 ExpTy = ExpTy->getAs<PointerType>()->getPointeeType(); 1271 if (ExpTy->isRecordType()) 1272 LV.setObjCIvar(false); 1273 } 1274 return; 1275 } 1276 1277 if (const GenericSelectionExpr *Exp = dyn_cast<GenericSelectionExpr>(E)) { 1278 setObjCGCLValueClass(Ctx, Exp->getResultExpr(), LV); 1279 return; 1280 } 1281 1282 if (const ImplicitCastExpr *Exp = dyn_cast<ImplicitCastExpr>(E)) { 1283 setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess); 1284 return; 1285 } 1286 1287 if (const CStyleCastExpr *Exp = dyn_cast<CStyleCastExpr>(E)) { 1288 setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess); 1289 return; 1290 } 1291 1292 if (const ObjCBridgedCastExpr *Exp = dyn_cast<ObjCBridgedCastExpr>(E)) { 1293 setObjCGCLValueClass(Ctx, Exp->getSubExpr(), LV, IsMemberAccess); 1294 return; 1295 } 1296 1297 if (const ArraySubscriptExpr *Exp = dyn_cast<ArraySubscriptExpr>(E)) { 1298 setObjCGCLValueClass(Ctx, Exp->getBase(), LV); 1299 if (LV.isObjCIvar() && !LV.isObjCArray()) 1300 // Using array syntax to assigning to what an ivar points to is not 1301 // same as assigning to the ivar itself. {id *Names;} Names[i] = 0; 1302 LV.setObjCIvar(false); 1303 else if (LV.isGlobalObjCRef() && !LV.isObjCArray()) 1304 // Using array syntax to assigning to what global points to is not 1305 // same as assigning to the global itself. {id *G;} G[i] = 0; 1306 LV.setGlobalObjCRef(false); 1307 return; 1308 } 1309 1310 if (const MemberExpr *Exp = dyn_cast<MemberExpr>(E)) { 1311 setObjCGCLValueClass(Ctx, Exp->getBase(), LV, true); 1312 // We don't know if member is an 'ivar', but this flag is looked at 1313 // only in the context of LV.isObjCIvar(). 1314 LV.setObjCArray(E->getType()->isArrayType()); 1315 return; 1316 } 1317 } 1318 1319 static llvm::Value * 1320 EmitBitCastOfLValueToProperType(CodeGenFunction &CGF, 1321 llvm::Value *V, llvm::Type *IRType, 1322 StringRef Name = StringRef()) { 1323 unsigned AS = cast<llvm::PointerType>(V->getType())->getAddressSpace(); 1324 return CGF.Builder.CreateBitCast(V, IRType->getPointerTo(AS), Name); 1325 } 1326 1327 static LValue EmitGlobalVarDeclLValue(CodeGenFunction &CGF, 1328 const Expr *E, const VarDecl *VD) { 1329 assert((VD->hasExternalStorage() || VD->isFileVarDecl()) && 1330 "Var decl must have external storage or be a file var decl!"); 1331 1332 llvm::Value *V = CGF.CGM.GetAddrOfGlobalVar(VD); 1333 llvm::Type *RealVarTy = CGF.getTypes().ConvertTypeForMem(VD->getType()); 1334 V = EmitBitCastOfLValueToProperType(CGF, V, RealVarTy); 1335 unsigned Alignment = CGF.getContext().getDeclAlign(VD).getQuantity(); 1336 QualType T = E->getType(); 1337 LValue LV; 1338 if (VD->getType()->isReferenceType()) { 1339 llvm::LoadInst *LI = CGF.Builder.CreateLoad(V); 1340 LI->setAlignment(Alignment); 1341 V = LI; 1342 LV = CGF.MakeNaturalAlignAddrLValue(V, T); 1343 } else { 1344 LV = CGF.MakeAddrLValue(V, E->getType(), Alignment); 1345 } 1346 setObjCGCLValueClass(CGF.getContext(), E, LV); 1347 return LV; 1348 } 1349 1350 static LValue EmitFunctionDeclLValue(CodeGenFunction &CGF, 1351 const Expr *E, const FunctionDecl *FD) { 1352 llvm::Value *V = CGF.CGM.GetAddrOfFunction(FD); 1353 if (!FD->hasPrototype()) { 1354 if (const FunctionProtoType *Proto = 1355 FD->getType()->getAs<FunctionProtoType>()) { 1356 // Ugly case: for a K&R-style definition, the type of the definition 1357 // isn't the same as the type of a use. Correct for this with a 1358 // bitcast. 1359 QualType NoProtoType = 1360 CGF.getContext().getFunctionNoProtoType(Proto->getResultType()); 1361 NoProtoType = CGF.getContext().getPointerType(NoProtoType); 1362 V = CGF.Builder.CreateBitCast(V, CGF.ConvertType(NoProtoType)); 1363 } 1364 } 1365 unsigned Alignment = CGF.getContext().getDeclAlign(FD).getQuantity(); 1366 return CGF.MakeAddrLValue(V, E->getType(), Alignment); 1367 } 1368 1369 LValue CodeGenFunction::EmitDeclRefLValue(const DeclRefExpr *E) { 1370 const NamedDecl *ND = E->getDecl(); 1371 unsigned Alignment = getContext().getDeclAlign(ND).getQuantity(); 1372 QualType T = E->getType(); 1373 1374 if (ND->hasAttr<WeakRefAttr>()) { 1375 const ValueDecl *VD = cast<ValueDecl>(ND); 1376 llvm::Constant *Aliasee = CGM.GetWeakRefReference(VD); 1377 return MakeAddrLValue(Aliasee, E->getType(), Alignment); 1378 } 1379 1380 if (const VarDecl *VD = dyn_cast<VarDecl>(ND)) { 1381 1382 // Check if this is a global variable. 1383 if (VD->hasExternalStorage() || VD->isFileVarDecl()) 1384 return EmitGlobalVarDeclLValue(*this, E, VD); 1385 1386 bool NonGCable = VD->hasLocalStorage() && 1387 !VD->getType()->isReferenceType() && 1388 !VD->hasAttr<BlocksAttr>(); 1389 1390 llvm::Value *V = LocalDeclMap[VD]; 1391 if (!V && VD->isStaticLocal()) 1392 V = CGM.getStaticLocalDeclAddress(VD); 1393 assert(V && "DeclRefExpr not entered in LocalDeclMap?"); 1394 1395 if (VD->hasAttr<BlocksAttr>()) 1396 V = BuildBlockByrefAddress(V, VD); 1397 1398 LValue LV; 1399 if (VD->getType()->isReferenceType()) { 1400 llvm::LoadInst *LI = Builder.CreateLoad(V); 1401 LI->setAlignment(Alignment); 1402 V = LI; 1403 LV = MakeNaturalAlignAddrLValue(V, T); 1404 } else { 1405 LV = MakeAddrLValue(V, T, Alignment); 1406 } 1407 1408 if (NonGCable) { 1409 LV.getQuals().removeObjCGCAttr(); 1410 LV.setNonGC(true); 1411 } 1412 setObjCGCLValueClass(getContext(), E, LV); 1413 return LV; 1414 } 1415 1416 if (const FunctionDecl *fn = dyn_cast<FunctionDecl>(ND)) 1417 return EmitFunctionDeclLValue(*this, E, fn); 1418 1419 llvm_unreachable("Unhandled DeclRefExpr"); 1420 1421 // an invalid LValue, but the assert will 1422 // ensure that this point is never reached. 1423 return LValue(); 1424 } 1425 1426 LValue CodeGenFunction::EmitBlockDeclRefLValue(const BlockDeclRefExpr *E) { 1427 unsigned Alignment = 1428 getContext().getDeclAlign(E->getDecl()).getQuantity(); 1429 return MakeAddrLValue(GetAddrOfBlockDecl(E), E->getType(), Alignment); 1430 } 1431 1432 LValue CodeGenFunction::EmitUnaryOpLValue(const UnaryOperator *E) { 1433 // __extension__ doesn't affect lvalue-ness. 1434 if (E->getOpcode() == UO_Extension) 1435 return EmitLValue(E->getSubExpr()); 1436 1437 QualType ExprTy = getContext().getCanonicalType(E->getSubExpr()->getType()); 1438 switch (E->getOpcode()) { 1439 default: llvm_unreachable("Unknown unary operator lvalue!"); 1440 case UO_Deref: { 1441 QualType T = E->getSubExpr()->getType()->getPointeeType(); 1442 assert(!T.isNull() && "CodeGenFunction::EmitUnaryOpLValue: Illegal type"); 1443 1444 LValue LV = MakeAddrLValue(EmitScalarExpr(E->getSubExpr()), T); 1445 LV.getQuals().setAddressSpace(ExprTy.getAddressSpace()); 1446 1447 // We should not generate __weak write barrier on indirect reference 1448 // of a pointer to object; as in void foo (__weak id *param); *param = 0; 1449 // But, we continue to generate __strong write barrier on indirect write 1450 // into a pointer to object. 1451 if (getContext().getLangOptions().ObjC1 && 1452 getContext().getLangOptions().getGC() != LangOptions::NonGC && 1453 LV.isObjCWeak()) 1454 LV.setNonGC(!E->isOBJCGCCandidate(getContext())); 1455 return LV; 1456 } 1457 case UO_Real: 1458 case UO_Imag: { 1459 LValue LV = EmitLValue(E->getSubExpr()); 1460 assert(LV.isSimple() && "real/imag on non-ordinary l-value"); 1461 llvm::Value *Addr = LV.getAddress(); 1462 1463 // real and imag are valid on scalars. This is a faster way of 1464 // testing that. 1465 if (!cast<llvm::PointerType>(Addr->getType()) 1466 ->getElementType()->isStructTy()) { 1467 assert(E->getSubExpr()->getType()->isArithmeticType()); 1468 return LV; 1469 } 1470 1471 assert(E->getSubExpr()->getType()->isAnyComplexType()); 1472 1473 unsigned Idx = E->getOpcode() == UO_Imag; 1474 return MakeAddrLValue(Builder.CreateStructGEP(LV.getAddress(), 1475 Idx, "idx"), 1476 ExprTy); 1477 } 1478 case UO_PreInc: 1479 case UO_PreDec: { 1480 LValue LV = EmitLValue(E->getSubExpr()); 1481 bool isInc = E->getOpcode() == UO_PreInc; 1482 1483 if (E->getType()->isAnyComplexType()) 1484 EmitComplexPrePostIncDec(E, LV, isInc, true/*isPre*/); 1485 else 1486 EmitScalarPrePostIncDec(E, LV, isInc, true/*isPre*/); 1487 return LV; 1488 } 1489 } 1490 } 1491 1492 LValue CodeGenFunction::EmitStringLiteralLValue(const StringLiteral *E) { 1493 return MakeAddrLValue(CGM.GetAddrOfConstantStringFromLiteral(E), 1494 E->getType()); 1495 } 1496 1497 LValue CodeGenFunction::EmitObjCEncodeExprLValue(const ObjCEncodeExpr *E) { 1498 return MakeAddrLValue(CGM.GetAddrOfConstantStringFromObjCEncode(E), 1499 E->getType()); 1500 } 1501 1502 1503 LValue CodeGenFunction::EmitPredefinedLValue(const PredefinedExpr *E) { 1504 switch (E->getIdentType()) { 1505 default: 1506 return EmitUnsupportedLValue(E, "predefined expression"); 1507 1508 case PredefinedExpr::Func: 1509 case PredefinedExpr::Function: 1510 case PredefinedExpr::PrettyFunction: { 1511 unsigned Type = E->getIdentType(); 1512 std::string GlobalVarName; 1513 1514 switch (Type) { 1515 default: llvm_unreachable("Invalid type"); 1516 case PredefinedExpr::Func: 1517 GlobalVarName = "__func__."; 1518 break; 1519 case PredefinedExpr::Function: 1520 GlobalVarName = "__FUNCTION__."; 1521 break; 1522 case PredefinedExpr::PrettyFunction: 1523 GlobalVarName = "__PRETTY_FUNCTION__."; 1524 break; 1525 } 1526 1527 StringRef FnName = CurFn->getName(); 1528 if (FnName.startswith("\01")) 1529 FnName = FnName.substr(1); 1530 GlobalVarName += FnName; 1531 1532 const Decl *CurDecl = CurCodeDecl; 1533 if (CurDecl == 0) 1534 CurDecl = getContext().getTranslationUnitDecl(); 1535 1536 std::string FunctionName = 1537 (isa<BlockDecl>(CurDecl) 1538 ? FnName.str() 1539 : PredefinedExpr::ComputeName((PredefinedExpr::IdentType)Type, CurDecl)); 1540 1541 llvm::Constant *C = 1542 CGM.GetAddrOfConstantCString(FunctionName, GlobalVarName.c_str()); 1543 return MakeAddrLValue(C, E->getType()); 1544 } 1545 } 1546 } 1547 1548 llvm::BasicBlock *CodeGenFunction::getTrapBB() { 1549 const CodeGenOptions &GCO = CGM.getCodeGenOpts(); 1550 1551 // If we are not optimzing, don't collapse all calls to trap in the function 1552 // to the same call, that way, in the debugger they can see which operation 1553 // did in fact fail. If we are optimizing, we collapse all calls to trap down 1554 // to just one per function to save on codesize. 1555 if (GCO.OptimizationLevel && TrapBB) 1556 return TrapBB; 1557 1558 llvm::BasicBlock *Cont = 0; 1559 if (HaveInsertPoint()) { 1560 Cont = createBasicBlock("cont"); 1561 EmitBranch(Cont); 1562 } 1563 TrapBB = createBasicBlock("trap"); 1564 EmitBlock(TrapBB); 1565 1566 llvm::Value *F = CGM.getIntrinsic(llvm::Intrinsic::trap); 1567 llvm::CallInst *TrapCall = Builder.CreateCall(F); 1568 TrapCall->setDoesNotReturn(); 1569 TrapCall->setDoesNotThrow(); 1570 Builder.CreateUnreachable(); 1571 1572 if (Cont) 1573 EmitBlock(Cont); 1574 return TrapBB; 1575 } 1576 1577 /// isSimpleArrayDecayOperand - If the specified expr is a simple decay from an 1578 /// array to pointer, return the array subexpression. 1579 static const Expr *isSimpleArrayDecayOperand(const Expr *E) { 1580 // If this isn't just an array->pointer decay, bail out. 1581 const CastExpr *CE = dyn_cast<CastExpr>(E); 1582 if (CE == 0 || CE->getCastKind() != CK_ArrayToPointerDecay) 1583 return 0; 1584 1585 // If this is a decay from variable width array, bail out. 1586 const Expr *SubExpr = CE->getSubExpr(); 1587 if (SubExpr->getType()->isVariableArrayType()) 1588 return 0; 1589 1590 return SubExpr; 1591 } 1592 1593 LValue CodeGenFunction::EmitArraySubscriptExpr(const ArraySubscriptExpr *E) { 1594 // The index must always be an integer, which is not an aggregate. Emit it. 1595 llvm::Value *Idx = EmitScalarExpr(E->getIdx()); 1596 QualType IdxTy = E->getIdx()->getType(); 1597 bool IdxSigned = IdxTy->isSignedIntegerOrEnumerationType(); 1598 1599 // If the base is a vector type, then we are forming a vector element lvalue 1600 // with this subscript. 1601 if (E->getBase()->getType()->isVectorType()) { 1602 // Emit the vector as an lvalue to get its address. 1603 LValue LHS = EmitLValue(E->getBase()); 1604 assert(LHS.isSimple() && "Can only subscript lvalue vectors here!"); 1605 Idx = Builder.CreateIntCast(Idx, Int32Ty, IdxSigned, "vidx"); 1606 return LValue::MakeVectorElt(LHS.getAddress(), Idx, 1607 E->getBase()->getType()); 1608 } 1609 1610 // Extend or truncate the index type to 32 or 64-bits. 1611 if (Idx->getType() != IntPtrTy) 1612 Idx = Builder.CreateIntCast(Idx, IntPtrTy, IdxSigned, "idxprom"); 1613 1614 // FIXME: As llvm implements the object size checking, this can come out. 1615 if (CatchUndefined) { 1616 if (const ImplicitCastExpr *ICE = dyn_cast<ImplicitCastExpr>(E->getBase())){ 1617 if (const DeclRefExpr *DRE = dyn_cast<DeclRefExpr>(ICE->getSubExpr())) { 1618 if (ICE->getCastKind() == CK_ArrayToPointerDecay) { 1619 if (const ConstantArrayType *CAT 1620 = getContext().getAsConstantArrayType(DRE->getType())) { 1621 llvm::APInt Size = CAT->getSize(); 1622 llvm::BasicBlock *Cont = createBasicBlock("cont"); 1623 Builder.CreateCondBr(Builder.CreateICmpULE(Idx, 1624 llvm::ConstantInt::get(Idx->getType(), Size)), 1625 Cont, getTrapBB()); 1626 EmitBlock(Cont); 1627 } 1628 } 1629 } 1630 } 1631 } 1632 1633 // We know that the pointer points to a type of the correct size, unless the 1634 // size is a VLA or Objective-C interface. 1635 llvm::Value *Address = 0; 1636 unsigned ArrayAlignment = 0; 1637 if (const VariableArrayType *vla = 1638 getContext().getAsVariableArrayType(E->getType())) { 1639 // The base must be a pointer, which is not an aggregate. Emit 1640 // it. It needs to be emitted first in case it's what captures 1641 // the VLA bounds. 1642 Address = EmitScalarExpr(E->getBase()); 1643 1644 // The element count here is the total number of non-VLA elements. 1645 llvm::Value *numElements = getVLASize(vla).first; 1646 1647 // Effectively, the multiply by the VLA size is part of the GEP. 1648 // GEP indexes are signed, and scaling an index isn't permitted to 1649 // signed-overflow, so we use the same semantics for our explicit 1650 // multiply. We suppress this if overflow is not undefined behavior. 1651 if (getLangOptions().isSignedOverflowDefined()) { 1652 Idx = Builder.CreateMul(Idx, numElements); 1653 Address = Builder.CreateGEP(Address, Idx, "arrayidx"); 1654 } else { 1655 Idx = Builder.CreateNSWMul(Idx, numElements); 1656 Address = Builder.CreateInBoundsGEP(Address, Idx, "arrayidx"); 1657 } 1658 } else if (const ObjCObjectType *OIT = E->getType()->getAs<ObjCObjectType>()){ 1659 // Indexing over an interface, as in "NSString *P; P[4];" 1660 llvm::Value *InterfaceSize = 1661 llvm::ConstantInt::get(Idx->getType(), 1662 getContext().getTypeSizeInChars(OIT).getQuantity()); 1663 1664 Idx = Builder.CreateMul(Idx, InterfaceSize); 1665 1666 // The base must be a pointer, which is not an aggregate. Emit it. 1667 llvm::Value *Base = EmitScalarExpr(E->getBase()); 1668 Address = EmitCastToVoidPtr(Base); 1669 Address = Builder.CreateGEP(Address, Idx, "arrayidx"); 1670 Address = Builder.CreateBitCast(Address, Base->getType()); 1671 } else if (const Expr *Array = isSimpleArrayDecayOperand(E->getBase())) { 1672 // If this is A[i] where A is an array, the frontend will have decayed the 1673 // base to be a ArrayToPointerDecay implicit cast. While correct, it is 1674 // inefficient at -O0 to emit a "gep A, 0, 0" when codegen'ing it, then a 1675 // "gep x, i" here. Emit one "gep A, 0, i". 1676 assert(Array->getType()->isArrayType() && 1677 "Array to pointer decay must have array source type!"); 1678 LValue ArrayLV = EmitLValue(Array); 1679 llvm::Value *ArrayPtr = ArrayLV.getAddress(); 1680 llvm::Value *Zero = llvm::ConstantInt::get(Int32Ty, 0); 1681 llvm::Value *Args[] = { Zero, Idx }; 1682 1683 // Propagate the alignment from the array itself to the result. 1684 ArrayAlignment = ArrayLV.getAlignment(); 1685 1686 if (getContext().getLangOptions().isSignedOverflowDefined()) 1687 Address = Builder.CreateGEP(ArrayPtr, Args, "arrayidx"); 1688 else 1689 Address = Builder.CreateInBoundsGEP(ArrayPtr, Args, "arrayidx"); 1690 } else { 1691 // The base must be a pointer, which is not an aggregate. Emit it. 1692 llvm::Value *Base = EmitScalarExpr(E->getBase()); 1693 if (getContext().getLangOptions().isSignedOverflowDefined()) 1694 Address = Builder.CreateGEP(Base, Idx, "arrayidx"); 1695 else 1696 Address = Builder.CreateInBoundsGEP(Base, Idx, "arrayidx"); 1697 } 1698 1699 QualType T = E->getBase()->getType()->getPointeeType(); 1700 assert(!T.isNull() && 1701 "CodeGenFunction::EmitArraySubscriptExpr(): Illegal base type"); 1702 1703 // Limit the alignment to that of the result type. 1704 if (ArrayAlignment) { 1705 unsigned Align = getContext().getTypeAlignInChars(T).getQuantity(); 1706 ArrayAlignment = std::min(Align, ArrayAlignment); 1707 } 1708 1709 LValue LV = MakeAddrLValue(Address, T, ArrayAlignment); 1710 LV.getQuals().setAddressSpace(E->getBase()->getType().getAddressSpace()); 1711 1712 if (getContext().getLangOptions().ObjC1 && 1713 getContext().getLangOptions().getGC() != LangOptions::NonGC) { 1714 LV.setNonGC(!E->isOBJCGCCandidate(getContext())); 1715 setObjCGCLValueClass(getContext(), E, LV); 1716 } 1717 return LV; 1718 } 1719 1720 static 1721 llvm::Constant *GenerateConstantVector(llvm::LLVMContext &VMContext, 1722 SmallVector<unsigned, 4> &Elts) { 1723 SmallVector<llvm::Constant*, 4> CElts; 1724 1725 llvm::Type *Int32Ty = llvm::Type::getInt32Ty(VMContext); 1726 for (unsigned i = 0, e = Elts.size(); i != e; ++i) 1727 CElts.push_back(llvm::ConstantInt::get(Int32Ty, Elts[i])); 1728 1729 return llvm::ConstantVector::get(CElts); 1730 } 1731 1732 LValue CodeGenFunction:: 1733 EmitExtVectorElementExpr(const ExtVectorElementExpr *E) { 1734 // Emit the base vector as an l-value. 1735 LValue Base; 1736 1737 // ExtVectorElementExpr's base can either be a vector or pointer to vector. 1738 if (E->isArrow()) { 1739 // If it is a pointer to a vector, emit the address and form an lvalue with 1740 // it. 1741 llvm::Value *Ptr = EmitScalarExpr(E->getBase()); 1742 const PointerType *PT = E->getBase()->getType()->getAs<PointerType>(); 1743 Base = MakeAddrLValue(Ptr, PT->getPointeeType()); 1744 Base.getQuals().removeObjCGCAttr(); 1745 } else if (E->getBase()->isGLValue()) { 1746 // Otherwise, if the base is an lvalue ( as in the case of foo.x.x), 1747 // emit the base as an lvalue. 1748 assert(E->getBase()->getType()->isVectorType()); 1749 Base = EmitLValue(E->getBase()); 1750 } else { 1751 // Otherwise, the base is a normal rvalue (as in (V+V).x), emit it as such. 1752 assert(E->getBase()->getType()->isVectorType() && 1753 "Result must be a vector"); 1754 llvm::Value *Vec = EmitScalarExpr(E->getBase()); 1755 1756 // Store the vector to memory (because LValue wants an address). 1757 llvm::Value *VecMem = CreateMemTemp(E->getBase()->getType()); 1758 Builder.CreateStore(Vec, VecMem); 1759 Base = MakeAddrLValue(VecMem, E->getBase()->getType()); 1760 } 1761 1762 QualType type = 1763 E->getType().withCVRQualifiers(Base.getQuals().getCVRQualifiers()); 1764 1765 // Encode the element access list into a vector of unsigned indices. 1766 SmallVector<unsigned, 4> Indices; 1767 E->getEncodedElementAccess(Indices); 1768 1769 if (Base.isSimple()) { 1770 llvm::Constant *CV = GenerateConstantVector(getLLVMContext(), Indices); 1771 return LValue::MakeExtVectorElt(Base.getAddress(), CV, type); 1772 } 1773 assert(Base.isExtVectorElt() && "Can only subscript lvalue vec elts here!"); 1774 1775 llvm::Constant *BaseElts = Base.getExtVectorElts(); 1776 SmallVector<llvm::Constant *, 4> CElts; 1777 1778 for (unsigned i = 0, e = Indices.size(); i != e; ++i) { 1779 if (isa<llvm::ConstantAggregateZero>(BaseElts)) 1780 CElts.push_back(llvm::ConstantInt::get(Int32Ty, 0)); 1781 else 1782 CElts.push_back(cast<llvm::Constant>(BaseElts->getOperand(Indices[i]))); 1783 } 1784 llvm::Constant *CV = llvm::ConstantVector::get(CElts); 1785 return LValue::MakeExtVectorElt(Base.getExtVectorAddr(), CV, type); 1786 } 1787 1788 LValue CodeGenFunction::EmitMemberExpr(const MemberExpr *E) { 1789 bool isNonGC = false; 1790 Expr *BaseExpr = E->getBase(); 1791 llvm::Value *BaseValue = NULL; 1792 Qualifiers BaseQuals; 1793 1794 // If this is s.x, emit s as an lvalue. If it is s->x, emit s as a scalar. 1795 if (E->isArrow()) { 1796 BaseValue = EmitScalarExpr(BaseExpr); 1797 const PointerType *PTy = 1798 BaseExpr->getType()->getAs<PointerType>(); 1799 BaseQuals = PTy->getPointeeType().getQualifiers(); 1800 } else { 1801 LValue BaseLV = EmitLValue(BaseExpr); 1802 if (BaseLV.isNonGC()) 1803 isNonGC = true; 1804 // FIXME: this isn't right for bitfields. 1805 BaseValue = BaseLV.getAddress(); 1806 QualType BaseTy = BaseExpr->getType(); 1807 BaseQuals = BaseTy.getQualifiers(); 1808 } 1809 1810 NamedDecl *ND = E->getMemberDecl(); 1811 if (FieldDecl *Field = dyn_cast<FieldDecl>(ND)) { 1812 LValue LV = EmitLValueForField(BaseValue, Field, 1813 BaseQuals.getCVRQualifiers()); 1814 LV.setNonGC(isNonGC); 1815 setObjCGCLValueClass(getContext(), E, LV); 1816 return LV; 1817 } 1818 1819 if (VarDecl *VD = dyn_cast<VarDecl>(ND)) 1820 return EmitGlobalVarDeclLValue(*this, E, VD); 1821 1822 if (const FunctionDecl *FD = dyn_cast<FunctionDecl>(ND)) 1823 return EmitFunctionDeclLValue(*this, E, FD); 1824 1825 llvm_unreachable("Unhandled member declaration!"); 1826 } 1827 1828 LValue CodeGenFunction::EmitLValueForBitfield(llvm::Value *BaseValue, 1829 const FieldDecl *Field, 1830 unsigned CVRQualifiers) { 1831 const CGRecordLayout &RL = 1832 CGM.getTypes().getCGRecordLayout(Field->getParent()); 1833 const CGBitFieldInfo &Info = RL.getBitFieldInfo(Field); 1834 return LValue::MakeBitfield(BaseValue, Info, 1835 Field->getType().withCVRQualifiers(CVRQualifiers)); 1836 } 1837 1838 /// EmitLValueForAnonRecordField - Given that the field is a member of 1839 /// an anonymous struct or union buried inside a record, and given 1840 /// that the base value is a pointer to the enclosing record, derive 1841 /// an lvalue for the ultimate field. 1842 LValue CodeGenFunction::EmitLValueForAnonRecordField(llvm::Value *BaseValue, 1843 const IndirectFieldDecl *Field, 1844 unsigned CVRQualifiers) { 1845 IndirectFieldDecl::chain_iterator I = Field->chain_begin(), 1846 IEnd = Field->chain_end(); 1847 while (true) { 1848 LValue LV = EmitLValueForField(BaseValue, cast<FieldDecl>(*I), 1849 CVRQualifiers); 1850 if (++I == IEnd) return LV; 1851 1852 assert(LV.isSimple()); 1853 BaseValue = LV.getAddress(); 1854 CVRQualifiers |= LV.getVRQualifiers(); 1855 } 1856 } 1857 1858 LValue CodeGenFunction::EmitLValueForField(llvm::Value *baseAddr, 1859 const FieldDecl *field, 1860 unsigned cvr) { 1861 if (field->isBitField()) 1862 return EmitLValueForBitfield(baseAddr, field, cvr); 1863 1864 const RecordDecl *rec = field->getParent(); 1865 QualType type = field->getType(); 1866 unsigned alignment = getContext().getDeclAlign(field).getQuantity(); 1867 1868 bool mayAlias = rec->hasAttr<MayAliasAttr>(); 1869 1870 llvm::Value *addr = baseAddr; 1871 if (rec->isUnion()) { 1872 // For unions, there is no pointer adjustment. 1873 assert(!type->isReferenceType() && "union has reference member"); 1874 } else { 1875 // For structs, we GEP to the field that the record layout suggests. 1876 unsigned idx = CGM.getTypes().getCGRecordLayout(rec).getLLVMFieldNo(field); 1877 addr = Builder.CreateStructGEP(addr, idx, field->getName()); 1878 1879 // If this is a reference field, load the reference right now. 1880 if (const ReferenceType *refType = type->getAs<ReferenceType>()) { 1881 llvm::LoadInst *load = Builder.CreateLoad(addr, "ref"); 1882 if (cvr & Qualifiers::Volatile) load->setVolatile(true); 1883 load->setAlignment(alignment); 1884 1885 if (CGM.shouldUseTBAA()) { 1886 llvm::MDNode *tbaa; 1887 if (mayAlias) 1888 tbaa = CGM.getTBAAInfo(getContext().CharTy); 1889 else 1890 tbaa = CGM.getTBAAInfo(type); 1891 CGM.DecorateInstruction(load, tbaa); 1892 } 1893 1894 addr = load; 1895 mayAlias = false; 1896 type = refType->getPointeeType(); 1897 if (type->isIncompleteType()) 1898 alignment = 0; 1899 else 1900 alignment = getContext().getTypeAlignInChars(type).getQuantity(); 1901 cvr = 0; // qualifiers don't recursively apply to referencee 1902 } 1903 } 1904 1905 // Make sure that the address is pointing to the right type. This is critical 1906 // for both unions and structs. A union needs a bitcast, a struct element 1907 // will need a bitcast if the LLVM type laid out doesn't match the desired 1908 // type. 1909 addr = EmitBitCastOfLValueToProperType(*this, addr, 1910 CGM.getTypes().ConvertTypeForMem(type), 1911 field->getName()); 1912 1913 if (field->hasAttr<AnnotateAttr>()) 1914 addr = EmitFieldAnnotations(field, addr); 1915 1916 LValue LV = MakeAddrLValue(addr, type, alignment); 1917 LV.getQuals().addCVRQualifiers(cvr); 1918 1919 // __weak attribute on a field is ignored. 1920 if (LV.getQuals().getObjCGCAttr() == Qualifiers::Weak) 1921 LV.getQuals().removeObjCGCAttr(); 1922 1923 // Fields of may_alias structs act like 'char' for TBAA purposes. 1924 // FIXME: this should get propagated down through anonymous structs 1925 // and unions. 1926 if (mayAlias && LV.getTBAAInfo()) 1927 LV.setTBAAInfo(CGM.getTBAAInfo(getContext().CharTy)); 1928 1929 return LV; 1930 } 1931 1932 LValue 1933 CodeGenFunction::EmitLValueForFieldInitialization(llvm::Value *BaseValue, 1934 const FieldDecl *Field, 1935 unsigned CVRQualifiers) { 1936 QualType FieldType = Field->getType(); 1937 1938 if (!FieldType->isReferenceType()) 1939 return EmitLValueForField(BaseValue, Field, CVRQualifiers); 1940 1941 const CGRecordLayout &RL = 1942 CGM.getTypes().getCGRecordLayout(Field->getParent()); 1943 unsigned idx = RL.getLLVMFieldNo(Field); 1944 llvm::Value *V = Builder.CreateStructGEP(BaseValue, idx); 1945 assert(!FieldType.getObjCGCAttr() && "fields cannot have GC attrs"); 1946 1947 1948 // Make sure that the address is pointing to the right type. This is critical 1949 // for both unions and structs. A union needs a bitcast, a struct element 1950 // will need a bitcast if the LLVM type laid out doesn't match the desired 1951 // type. 1952 llvm::Type *llvmType = ConvertTypeForMem(FieldType); 1953 unsigned AS = cast<llvm::PointerType>(V->getType())->getAddressSpace(); 1954 V = Builder.CreateBitCast(V, llvmType->getPointerTo(AS)); 1955 1956 unsigned Alignment = getContext().getDeclAlign(Field).getQuantity(); 1957 return MakeAddrLValue(V, FieldType, Alignment); 1958 } 1959 1960 LValue CodeGenFunction::EmitCompoundLiteralLValue(const CompoundLiteralExpr *E){ 1961 if (E->isFileScope()) { 1962 llvm::Value *GlobalPtr = CGM.GetAddrOfConstantCompoundLiteral(E); 1963 return MakeAddrLValue(GlobalPtr, E->getType()); 1964 } 1965 1966 llvm::Value *DeclPtr = CreateMemTemp(E->getType(), ".compoundliteral"); 1967 const Expr *InitExpr = E->getInitializer(); 1968 LValue Result = MakeAddrLValue(DeclPtr, E->getType()); 1969 1970 EmitAnyExprToMem(InitExpr, DeclPtr, E->getType().getQualifiers(), 1971 /*Init*/ true); 1972 1973 return Result; 1974 } 1975 1976 LValue CodeGenFunction:: 1977 EmitConditionalOperatorLValue(const AbstractConditionalOperator *expr) { 1978 if (!expr->isGLValue()) { 1979 // ?: here should be an aggregate. 1980 assert((hasAggregateLLVMType(expr->getType()) && 1981 !expr->getType()->isAnyComplexType()) && 1982 "Unexpected conditional operator!"); 1983 return EmitAggExprToLValue(expr); 1984 } 1985 1986 const Expr *condExpr = expr->getCond(); 1987 bool CondExprBool; 1988 if (ConstantFoldsToSimpleInteger(condExpr, CondExprBool)) { 1989 const Expr *live = expr->getTrueExpr(), *dead = expr->getFalseExpr(); 1990 if (!CondExprBool) std::swap(live, dead); 1991 1992 if (!ContainsLabel(dead)) 1993 return EmitLValue(live); 1994 } 1995 1996 OpaqueValueMapping binding(*this, expr); 1997 1998 llvm::BasicBlock *lhsBlock = createBasicBlock("cond.true"); 1999 llvm::BasicBlock *rhsBlock = createBasicBlock("cond.false"); 2000 llvm::BasicBlock *contBlock = createBasicBlock("cond.end"); 2001 2002 ConditionalEvaluation eval(*this); 2003 EmitBranchOnBoolExpr(condExpr, lhsBlock, rhsBlock); 2004 2005 // Any temporaries created here are conditional. 2006 EmitBlock(lhsBlock); 2007 eval.begin(*this); 2008 LValue lhs = EmitLValue(expr->getTrueExpr()); 2009 eval.end(*this); 2010 2011 if (!lhs.isSimple()) 2012 return EmitUnsupportedLValue(expr, "conditional operator"); 2013 2014 lhsBlock = Builder.GetInsertBlock(); 2015 Builder.CreateBr(contBlock); 2016 2017 // Any temporaries created here are conditional. 2018 EmitBlock(rhsBlock); 2019 eval.begin(*this); 2020 LValue rhs = EmitLValue(expr->getFalseExpr()); 2021 eval.end(*this); 2022 if (!rhs.isSimple()) 2023 return EmitUnsupportedLValue(expr, "conditional operator"); 2024 rhsBlock = Builder.GetInsertBlock(); 2025 2026 EmitBlock(contBlock); 2027 2028 llvm::PHINode *phi = Builder.CreatePHI(lhs.getAddress()->getType(), 2, 2029 "cond-lvalue"); 2030 phi->addIncoming(lhs.getAddress(), lhsBlock); 2031 phi->addIncoming(rhs.getAddress(), rhsBlock); 2032 return MakeAddrLValue(phi, expr->getType()); 2033 } 2034 2035 /// EmitCastLValue - Casts are never lvalues unless that cast is a dynamic_cast. 2036 /// If the cast is a dynamic_cast, we can have the usual lvalue result, 2037 /// otherwise if a cast is needed by the code generator in an lvalue context, 2038 /// then it must mean that we need the address of an aggregate in order to 2039 /// access one of its fields. This can happen for all the reasons that casts 2040 /// are permitted with aggregate result, including noop aggregate casts, and 2041 /// cast from scalar to union. 2042 LValue CodeGenFunction::EmitCastLValue(const CastExpr *E) { 2043 switch (E->getCastKind()) { 2044 case CK_ToVoid: 2045 return EmitUnsupportedLValue(E, "unexpected cast lvalue"); 2046 2047 case CK_Dependent: 2048 llvm_unreachable("dependent cast kind in IR gen!"); 2049 2050 case CK_NoOp: 2051 case CK_LValueToRValue: 2052 if (!E->getSubExpr()->Classify(getContext()).isPRValue() 2053 || E->getType()->isRecordType()) 2054 return EmitLValue(E->getSubExpr()); 2055 // Fall through to synthesize a temporary. 2056 2057 case CK_BitCast: 2058 case CK_ArrayToPointerDecay: 2059 case CK_FunctionToPointerDecay: 2060 case CK_NullToMemberPointer: 2061 case CK_NullToPointer: 2062 case CK_IntegralToPointer: 2063 case CK_PointerToIntegral: 2064 case CK_PointerToBoolean: 2065 case CK_VectorSplat: 2066 case CK_IntegralCast: 2067 case CK_IntegralToBoolean: 2068 case CK_IntegralToFloating: 2069 case CK_FloatingToIntegral: 2070 case CK_FloatingToBoolean: 2071 case CK_FloatingCast: 2072 case CK_FloatingRealToComplex: 2073 case CK_FloatingComplexToReal: 2074 case CK_FloatingComplexToBoolean: 2075 case CK_FloatingComplexCast: 2076 case CK_FloatingComplexToIntegralComplex: 2077 case CK_IntegralRealToComplex: 2078 case CK_IntegralComplexToReal: 2079 case CK_IntegralComplexToBoolean: 2080 case CK_IntegralComplexCast: 2081 case CK_IntegralComplexToFloatingComplex: 2082 case CK_DerivedToBaseMemberPointer: 2083 case CK_BaseToDerivedMemberPointer: 2084 case CK_MemberPointerToBoolean: 2085 case CK_AnyPointerToBlockPointerCast: 2086 case CK_ARCProduceObject: 2087 case CK_ARCConsumeObject: 2088 case CK_ARCReclaimReturnedObject: 2089 case CK_ARCExtendBlockObject: { 2090 // These casts only produce lvalues when we're binding a reference to a 2091 // temporary realized from a (converted) pure rvalue. Emit the expression 2092 // as a value, copy it into a temporary, and return an lvalue referring to 2093 // that temporary. 2094 llvm::Value *V = CreateMemTemp(E->getType(), "ref.temp"); 2095 EmitAnyExprToMem(E, V, E->getType().getQualifiers(), false); 2096 return MakeAddrLValue(V, E->getType()); 2097 } 2098 2099 case CK_Dynamic: { 2100 LValue LV = EmitLValue(E->getSubExpr()); 2101 llvm::Value *V = LV.getAddress(); 2102 const CXXDynamicCastExpr *DCE = cast<CXXDynamicCastExpr>(E); 2103 return MakeAddrLValue(EmitDynamicCast(V, DCE), E->getType()); 2104 } 2105 2106 case CK_ConstructorConversion: 2107 case CK_UserDefinedConversion: 2108 case CK_CPointerToObjCPointerCast: 2109 case CK_BlockPointerToObjCPointerCast: 2110 return EmitLValue(E->getSubExpr()); 2111 2112 case CK_UncheckedDerivedToBase: 2113 case CK_DerivedToBase: { 2114 const RecordType *DerivedClassTy = 2115 E->getSubExpr()->getType()->getAs<RecordType>(); 2116 CXXRecordDecl *DerivedClassDecl = 2117 cast<CXXRecordDecl>(DerivedClassTy->getDecl()); 2118 2119 LValue LV = EmitLValue(E->getSubExpr()); 2120 llvm::Value *This = LV.getAddress(); 2121 2122 // Perform the derived-to-base conversion 2123 llvm::Value *Base = 2124 GetAddressOfBaseClass(This, DerivedClassDecl, 2125 E->path_begin(), E->path_end(), 2126 /*NullCheckValue=*/false); 2127 2128 return MakeAddrLValue(Base, E->getType()); 2129 } 2130 case CK_ToUnion: 2131 return EmitAggExprToLValue(E); 2132 case CK_BaseToDerived: { 2133 const RecordType *DerivedClassTy = E->getType()->getAs<RecordType>(); 2134 CXXRecordDecl *DerivedClassDecl = 2135 cast<CXXRecordDecl>(DerivedClassTy->getDecl()); 2136 2137 LValue LV = EmitLValue(E->getSubExpr()); 2138 2139 // Perform the base-to-derived conversion 2140 llvm::Value *Derived = 2141 GetAddressOfDerivedClass(LV.getAddress(), DerivedClassDecl, 2142 E->path_begin(), E->path_end(), 2143 /*NullCheckValue=*/false); 2144 2145 return MakeAddrLValue(Derived, E->getType()); 2146 } 2147 case CK_LValueBitCast: { 2148 // This must be a reinterpret_cast (or c-style equivalent). 2149 const ExplicitCastExpr *CE = cast<ExplicitCastExpr>(E); 2150 2151 LValue LV = EmitLValue(E->getSubExpr()); 2152 llvm::Value *V = Builder.CreateBitCast(LV.getAddress(), 2153 ConvertType(CE->getTypeAsWritten())); 2154 return MakeAddrLValue(V, E->getType()); 2155 } 2156 case CK_ObjCObjectLValueCast: { 2157 LValue LV = EmitLValue(E->getSubExpr()); 2158 QualType ToType = getContext().getLValueReferenceType(E->getType()); 2159 llvm::Value *V = Builder.CreateBitCast(LV.getAddress(), 2160 ConvertType(ToType)); 2161 return MakeAddrLValue(V, E->getType()); 2162 } 2163 } 2164 2165 llvm_unreachable("Unhandled lvalue cast kind?"); 2166 } 2167 2168 LValue CodeGenFunction::EmitNullInitializationLValue( 2169 const CXXScalarValueInitExpr *E) { 2170 QualType Ty = E->getType(); 2171 LValue LV = MakeAddrLValue(CreateMemTemp(Ty), Ty); 2172 EmitNullInitialization(LV.getAddress(), Ty); 2173 return LV; 2174 } 2175 2176 LValue CodeGenFunction::EmitOpaqueValueLValue(const OpaqueValueExpr *e) { 2177 assert(OpaqueValueMappingData::shouldBindAsLValue(e)); 2178 return getOpaqueLValueMapping(e); 2179 } 2180 2181 LValue CodeGenFunction::EmitMaterializeTemporaryExpr( 2182 const MaterializeTemporaryExpr *E) { 2183 RValue RV = EmitReferenceBindingToExpr(E, /*InitializedDecl=*/0); 2184 return MakeAddrLValue(RV.getScalarVal(), E->getType()); 2185 } 2186 2187 2188 //===--------------------------------------------------------------------===// 2189 // Expression Emission 2190 //===--------------------------------------------------------------------===// 2191 2192 RValue CodeGenFunction::EmitCallExpr(const CallExpr *E, 2193 ReturnValueSlot ReturnValue) { 2194 if (CGDebugInfo *DI = getDebugInfo()) 2195 DI->EmitLocation(Builder, E->getLocStart()); 2196 2197 // Builtins never have block type. 2198 if (E->getCallee()->getType()->isBlockPointerType()) 2199 return EmitBlockCallExpr(E, ReturnValue); 2200 2201 if (const CXXMemberCallExpr *CE = dyn_cast<CXXMemberCallExpr>(E)) 2202 return EmitCXXMemberCallExpr(CE, ReturnValue); 2203 2204 if (const CUDAKernelCallExpr *CE = dyn_cast<CUDAKernelCallExpr>(E)) 2205 return EmitCUDAKernelCallExpr(CE, ReturnValue); 2206 2207 const Decl *TargetDecl = E->getCalleeDecl(); 2208 if (const FunctionDecl *FD = dyn_cast_or_null<FunctionDecl>(TargetDecl)) { 2209 if (unsigned builtinID = FD->getBuiltinID()) 2210 return EmitBuiltinExpr(FD, builtinID, E); 2211 } 2212 2213 if (const CXXOperatorCallExpr *CE = dyn_cast<CXXOperatorCallExpr>(E)) 2214 if (const CXXMethodDecl *MD = dyn_cast_or_null<CXXMethodDecl>(TargetDecl)) 2215 return EmitCXXOperatorMemberCallExpr(CE, MD, ReturnValue); 2216 2217 if (const CXXPseudoDestructorExpr *PseudoDtor 2218 = dyn_cast<CXXPseudoDestructorExpr>(E->getCallee()->IgnoreParens())) { 2219 QualType DestroyedType = PseudoDtor->getDestroyedType(); 2220 if (getContext().getLangOptions().ObjCAutoRefCount && 2221 DestroyedType->isObjCLifetimeType() && 2222 (DestroyedType.getObjCLifetime() == Qualifiers::OCL_Strong || 2223 DestroyedType.getObjCLifetime() == Qualifiers::OCL_Weak)) { 2224 // Automatic Reference Counting: 2225 // If the pseudo-expression names a retainable object with weak or 2226 // strong lifetime, the object shall be released. 2227 Expr *BaseExpr = PseudoDtor->getBase(); 2228 llvm::Value *BaseValue = NULL; 2229 Qualifiers BaseQuals; 2230 2231 // If this is s.x, emit s as an lvalue. If it is s->x, emit s as a scalar. 2232 if (PseudoDtor->isArrow()) { 2233 BaseValue = EmitScalarExpr(BaseExpr); 2234 const PointerType *PTy = BaseExpr->getType()->getAs<PointerType>(); 2235 BaseQuals = PTy->getPointeeType().getQualifiers(); 2236 } else { 2237 LValue BaseLV = EmitLValue(BaseExpr); 2238 BaseValue = BaseLV.getAddress(); 2239 QualType BaseTy = BaseExpr->getType(); 2240 BaseQuals = BaseTy.getQualifiers(); 2241 } 2242 2243 switch (PseudoDtor->getDestroyedType().getObjCLifetime()) { 2244 case Qualifiers::OCL_None: 2245 case Qualifiers::OCL_ExplicitNone: 2246 case Qualifiers::OCL_Autoreleasing: 2247 break; 2248 2249 case Qualifiers::OCL_Strong: 2250 EmitARCRelease(Builder.CreateLoad(BaseValue, 2251 PseudoDtor->getDestroyedType().isVolatileQualified()), 2252 /*precise*/ true); 2253 break; 2254 2255 case Qualifiers::OCL_Weak: 2256 EmitARCDestroyWeak(BaseValue); 2257 break; 2258 } 2259 } else { 2260 // C++ [expr.pseudo]p1: 2261 // The result shall only be used as the operand for the function call 2262 // operator (), and the result of such a call has type void. The only 2263 // effect is the evaluation of the postfix-expression before the dot or 2264 // arrow. 2265 EmitScalarExpr(E->getCallee()); 2266 } 2267 2268 return RValue::get(0); 2269 } 2270 2271 llvm::Value *Callee = EmitScalarExpr(E->getCallee()); 2272 return EmitCall(E->getCallee()->getType(), Callee, ReturnValue, 2273 E->arg_begin(), E->arg_end(), TargetDecl); 2274 } 2275 2276 LValue CodeGenFunction::EmitBinaryOperatorLValue(const BinaryOperator *E) { 2277 // Comma expressions just emit their LHS then their RHS as an l-value. 2278 if (E->getOpcode() == BO_Comma) { 2279 EmitIgnoredExpr(E->getLHS()); 2280 EnsureInsertPoint(); 2281 return EmitLValue(E->getRHS()); 2282 } 2283 2284 if (E->getOpcode() == BO_PtrMemD || 2285 E->getOpcode() == BO_PtrMemI) 2286 return EmitPointerToDataMemberBinaryExpr(E); 2287 2288 assert(E->getOpcode() == BO_Assign && "unexpected binary l-value"); 2289 2290 // Note that in all of these cases, __block variables need the RHS 2291 // evaluated first just in case the variable gets moved by the RHS. 2292 2293 if (!hasAggregateLLVMType(E->getType())) { 2294 switch (E->getLHS()->getType().getObjCLifetime()) { 2295 case Qualifiers::OCL_Strong: 2296 return EmitARCStoreStrong(E, /*ignored*/ false).first; 2297 2298 case Qualifiers::OCL_Autoreleasing: 2299 return EmitARCStoreAutoreleasing(E).first; 2300 2301 // No reason to do any of these differently. 2302 case Qualifiers::OCL_None: 2303 case Qualifiers::OCL_ExplicitNone: 2304 case Qualifiers::OCL_Weak: 2305 break; 2306 } 2307 2308 RValue RV = EmitAnyExpr(E->getRHS()); 2309 LValue LV = EmitLValue(E->getLHS()); 2310 EmitStoreThroughLValue(RV, LV); 2311 return LV; 2312 } 2313 2314 if (E->getType()->isAnyComplexType()) 2315 return EmitComplexAssignmentLValue(E); 2316 2317 return EmitAggExprToLValue(E); 2318 } 2319 2320 LValue CodeGenFunction::EmitCallExprLValue(const CallExpr *E) { 2321 RValue RV = EmitCallExpr(E); 2322 2323 if (!RV.isScalar()) 2324 return MakeAddrLValue(RV.getAggregateAddr(), E->getType()); 2325 2326 assert(E->getCallReturnType()->isReferenceType() && 2327 "Can't have a scalar return unless the return type is a " 2328 "reference type!"); 2329 2330 return MakeAddrLValue(RV.getScalarVal(), E->getType()); 2331 } 2332 2333 LValue CodeGenFunction::EmitVAArgExprLValue(const VAArgExpr *E) { 2334 // FIXME: This shouldn't require another copy. 2335 return EmitAggExprToLValue(E); 2336 } 2337 2338 LValue CodeGenFunction::EmitCXXConstructLValue(const CXXConstructExpr *E) { 2339 assert(E->getType()->getAsCXXRecordDecl()->hasTrivialDestructor() 2340 && "binding l-value to type which needs a temporary"); 2341 AggValueSlot Slot = CreateAggTemp(E->getType()); 2342 EmitCXXConstructExpr(E, Slot); 2343 return MakeAddrLValue(Slot.getAddr(), E->getType()); 2344 } 2345 2346 LValue 2347 CodeGenFunction::EmitCXXTypeidLValue(const CXXTypeidExpr *E) { 2348 return MakeAddrLValue(EmitCXXTypeidExpr(E), E->getType()); 2349 } 2350 2351 LValue 2352 CodeGenFunction::EmitCXXBindTemporaryLValue(const CXXBindTemporaryExpr *E) { 2353 AggValueSlot Slot = CreateAggTemp(E->getType(), "temp.lvalue"); 2354 Slot.setExternallyDestructed(); 2355 EmitAggExpr(E->getSubExpr(), Slot); 2356 EmitCXXTemporary(E->getTemporary(), E->getType(), Slot.getAddr()); 2357 return MakeAddrLValue(Slot.getAddr(), E->getType()); 2358 } 2359 2360 LValue CodeGenFunction::EmitObjCMessageExprLValue(const ObjCMessageExpr *E) { 2361 RValue RV = EmitObjCMessageExpr(E); 2362 2363 if (!RV.isScalar()) 2364 return MakeAddrLValue(RV.getAggregateAddr(), E->getType()); 2365 2366 assert(E->getMethodDecl()->getResultType()->isReferenceType() && 2367 "Can't have a scalar return unless the return type is a " 2368 "reference type!"); 2369 2370 return MakeAddrLValue(RV.getScalarVal(), E->getType()); 2371 } 2372 2373 LValue CodeGenFunction::EmitObjCSelectorLValue(const ObjCSelectorExpr *E) { 2374 llvm::Value *V = 2375 CGM.getObjCRuntime().GetSelector(Builder, E->getSelector(), true); 2376 return MakeAddrLValue(V, E->getType()); 2377 } 2378 2379 llvm::Value *CodeGenFunction::EmitIvarOffset(const ObjCInterfaceDecl *Interface, 2380 const ObjCIvarDecl *Ivar) { 2381 return CGM.getObjCRuntime().EmitIvarOffset(*this, Interface, Ivar); 2382 } 2383 2384 LValue CodeGenFunction::EmitLValueForIvar(QualType ObjectTy, 2385 llvm::Value *BaseValue, 2386 const ObjCIvarDecl *Ivar, 2387 unsigned CVRQualifiers) { 2388 return CGM.getObjCRuntime().EmitObjCValueForIvar(*this, ObjectTy, BaseValue, 2389 Ivar, CVRQualifiers); 2390 } 2391 2392 LValue CodeGenFunction::EmitObjCIvarRefLValue(const ObjCIvarRefExpr *E) { 2393 // FIXME: A lot of the code below could be shared with EmitMemberExpr. 2394 llvm::Value *BaseValue = 0; 2395 const Expr *BaseExpr = E->getBase(); 2396 Qualifiers BaseQuals; 2397 QualType ObjectTy; 2398 if (E->isArrow()) { 2399 BaseValue = EmitScalarExpr(BaseExpr); 2400 ObjectTy = BaseExpr->getType()->getPointeeType(); 2401 BaseQuals = ObjectTy.getQualifiers(); 2402 } else { 2403 LValue BaseLV = EmitLValue(BaseExpr); 2404 // FIXME: this isn't right for bitfields. 2405 BaseValue = BaseLV.getAddress(); 2406 ObjectTy = BaseExpr->getType(); 2407 BaseQuals = ObjectTy.getQualifiers(); 2408 } 2409 2410 LValue LV = 2411 EmitLValueForIvar(ObjectTy, BaseValue, E->getDecl(), 2412 BaseQuals.getCVRQualifiers()); 2413 setObjCGCLValueClass(getContext(), E, LV); 2414 return LV; 2415 } 2416 2417 LValue CodeGenFunction::EmitStmtExprLValue(const StmtExpr *E) { 2418 // Can only get l-value for message expression returning aggregate type 2419 RValue RV = EmitAnyExprToTemp(E); 2420 return MakeAddrLValue(RV.getAggregateAddr(), E->getType()); 2421 } 2422 2423 RValue CodeGenFunction::EmitCall(QualType CalleeType, llvm::Value *Callee, 2424 ReturnValueSlot ReturnValue, 2425 CallExpr::const_arg_iterator ArgBeg, 2426 CallExpr::const_arg_iterator ArgEnd, 2427 const Decl *TargetDecl) { 2428 // Get the actual function type. The callee type will always be a pointer to 2429 // function type or a block pointer type. 2430 assert(CalleeType->isFunctionPointerType() && 2431 "Call must have function pointer type!"); 2432 2433 CalleeType = getContext().getCanonicalType(CalleeType); 2434 2435 const FunctionType *FnType 2436 = cast<FunctionType>(cast<PointerType>(CalleeType)->getPointeeType()); 2437 2438 CallArgList Args; 2439 EmitCallArgs(Args, dyn_cast<FunctionProtoType>(FnType), ArgBeg, ArgEnd); 2440 2441 const CGFunctionInfo &FnInfo = CGM.getTypes().getFunctionInfo(Args, FnType); 2442 2443 // C99 6.5.2.2p6: 2444 // If the expression that denotes the called function has a type 2445 // that does not include a prototype, [the default argument 2446 // promotions are performed]. If the number of arguments does not 2447 // equal the number of parameters, the behavior is undefined. If 2448 // the function is defined with a type that includes a prototype, 2449 // and either the prototype ends with an ellipsis (, ...) or the 2450 // types of the arguments after promotion are not compatible with 2451 // the types of the parameters, the behavior is undefined. If the 2452 // function is defined with a type that does not include a 2453 // prototype, and the types of the arguments after promotion are 2454 // not compatible with those of the parameters after promotion, 2455 // the behavior is undefined [except in some trivial cases]. 2456 // That is, in the general case, we should assume that a call 2457 // through an unprototyped function type works like a *non-variadic* 2458 // call. The way we make this work is to cast to the exact type 2459 // of the promoted arguments. 2460 if (isa<FunctionNoProtoType>(FnType) && 2461 !getTargetHooks().isNoProtoCallVariadic(FnType->getCallConv())) { 2462 assert(cast<llvm::FunctionType>(Callee->getType()->getContainedType(0)) 2463 ->isVarArg()); 2464 llvm::Type *CalleeTy = getTypes().GetFunctionType(FnInfo, false); 2465 CalleeTy = CalleeTy->getPointerTo(); 2466 Callee = Builder.CreateBitCast(Callee, CalleeTy, "callee.knr.cast"); 2467 } 2468 2469 return EmitCall(FnInfo, Callee, ReturnValue, Args, TargetDecl); 2470 } 2471 2472 LValue CodeGenFunction:: 2473 EmitPointerToDataMemberBinaryExpr(const BinaryOperator *E) { 2474 llvm::Value *BaseV; 2475 if (E->getOpcode() == BO_PtrMemI) 2476 BaseV = EmitScalarExpr(E->getLHS()); 2477 else 2478 BaseV = EmitLValue(E->getLHS()).getAddress(); 2479 2480 llvm::Value *OffsetV = EmitScalarExpr(E->getRHS()); 2481 2482 const MemberPointerType *MPT 2483 = E->getRHS()->getType()->getAs<MemberPointerType>(); 2484 2485 llvm::Value *AddV = 2486 CGM.getCXXABI().EmitMemberDataPointerAddress(*this, BaseV, OffsetV, MPT); 2487 2488 return MakeAddrLValue(AddV, MPT->getPointeeType()); 2489 } 2490 2491 static void 2492 EmitAtomicOp(CodeGenFunction &CGF, AtomicExpr *E, llvm::Value *Dest, 2493 llvm::Value *Ptr, llvm::Value *Val1, llvm::Value *Val2, 2494 uint64_t Size, unsigned Align, llvm::AtomicOrdering Order) { 2495 if (E->isCmpXChg()) { 2496 // Note that cmpxchg only supports specifying one ordering and 2497 // doesn't support weak cmpxchg, at least at the moment. 2498 llvm::LoadInst *LoadVal1 = CGF.Builder.CreateLoad(Val1); 2499 LoadVal1->setAlignment(Align); 2500 llvm::LoadInst *LoadVal2 = CGF.Builder.CreateLoad(Val2); 2501 LoadVal2->setAlignment(Align); 2502 llvm::AtomicCmpXchgInst *CXI = 2503 CGF.Builder.CreateAtomicCmpXchg(Ptr, LoadVal1, LoadVal2, Order); 2504 CXI->setVolatile(E->isVolatile()); 2505 llvm::StoreInst *StoreVal1 = CGF.Builder.CreateStore(CXI, Val1); 2506 StoreVal1->setAlignment(Align); 2507 llvm::Value *Cmp = CGF.Builder.CreateICmpEQ(CXI, LoadVal1); 2508 CGF.EmitStoreOfScalar(Cmp, CGF.MakeAddrLValue(Dest, E->getType())); 2509 return; 2510 } 2511 2512 if (E->getOp() == AtomicExpr::Load) { 2513 llvm::LoadInst *Load = CGF.Builder.CreateLoad(Ptr); 2514 Load->setAtomic(Order); 2515 Load->setAlignment(Size); 2516 Load->setVolatile(E->isVolatile()); 2517 llvm::StoreInst *StoreDest = CGF.Builder.CreateStore(Load, Dest); 2518 StoreDest->setAlignment(Align); 2519 return; 2520 } 2521 2522 if (E->getOp() == AtomicExpr::Store) { 2523 assert(!Dest && "Store does not return a value"); 2524 llvm::LoadInst *LoadVal1 = CGF.Builder.CreateLoad(Val1); 2525 LoadVal1->setAlignment(Align); 2526 llvm::StoreInst *Store = CGF.Builder.CreateStore(LoadVal1, Ptr); 2527 Store->setAtomic(Order); 2528 Store->setAlignment(Size); 2529 Store->setVolatile(E->isVolatile()); 2530 return; 2531 } 2532 2533 llvm::AtomicRMWInst::BinOp Op = llvm::AtomicRMWInst::Add; 2534 switch (E->getOp()) { 2535 case AtomicExpr::CmpXchgWeak: 2536 case AtomicExpr::CmpXchgStrong: 2537 case AtomicExpr::Store: 2538 case AtomicExpr::Load: assert(0 && "Already handled!"); 2539 case AtomicExpr::Add: Op = llvm::AtomicRMWInst::Add; break; 2540 case AtomicExpr::Sub: Op = llvm::AtomicRMWInst::Sub; break; 2541 case AtomicExpr::And: Op = llvm::AtomicRMWInst::And; break; 2542 case AtomicExpr::Or: Op = llvm::AtomicRMWInst::Or; break; 2543 case AtomicExpr::Xor: Op = llvm::AtomicRMWInst::Xor; break; 2544 case AtomicExpr::Xchg: Op = llvm::AtomicRMWInst::Xchg; break; 2545 } 2546 llvm::LoadInst *LoadVal1 = CGF.Builder.CreateLoad(Val1); 2547 LoadVal1->setAlignment(Align); 2548 llvm::AtomicRMWInst *RMWI = 2549 CGF.Builder.CreateAtomicRMW(Op, Ptr, LoadVal1, Order); 2550 RMWI->setVolatile(E->isVolatile()); 2551 llvm::StoreInst *StoreDest = CGF.Builder.CreateStore(RMWI, Dest); 2552 StoreDest->setAlignment(Align); 2553 } 2554 2555 // This function emits any expression (scalar, complex, or aggregate) 2556 // into a temporary alloca. 2557 static llvm::Value * 2558 EmitValToTemp(CodeGenFunction &CGF, Expr *E) { 2559 llvm::Value *DeclPtr = CGF.CreateMemTemp(E->getType(), ".atomictmp"); 2560 CGF.EmitAnyExprToMem(E, DeclPtr, E->getType().getQualifiers(), 2561 /*Init*/ true); 2562 return DeclPtr; 2563 } 2564 2565 static RValue ConvertTempToRValue(CodeGenFunction &CGF, QualType Ty, 2566 llvm::Value *Dest) { 2567 if (Ty->isAnyComplexType()) 2568 return RValue::getComplex(CGF.LoadComplexFromAddr(Dest, false)); 2569 if (CGF.hasAggregateLLVMType(Ty)) 2570 return RValue::getAggregate(Dest); 2571 return RValue::get(CGF.EmitLoadOfScalar(CGF.MakeAddrLValue(Dest, Ty))); 2572 } 2573 2574 RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr *E, llvm::Value *Dest) { 2575 QualType AtomicTy = E->getPtr()->getType()->getPointeeType(); 2576 QualType MemTy = AtomicTy->getAs<AtomicType>()->getValueType(); 2577 CharUnits sizeChars = getContext().getTypeSizeInChars(AtomicTy); 2578 uint64_t Size = sizeChars.getQuantity(); 2579 CharUnits alignChars = getContext().getTypeAlignInChars(AtomicTy); 2580 unsigned Align = alignChars.getQuantity(); 2581 unsigned MaxInlineWidth = 2582 getContext().getTargetInfo().getMaxAtomicInlineWidth(); 2583 bool UseLibcall = (Size != Align || Size > MaxInlineWidth); 2584 2585 llvm::Value *Ptr, *Order, *OrderFail = 0, *Val1 = 0, *Val2 = 0; 2586 Ptr = EmitScalarExpr(E->getPtr()); 2587 Order = EmitScalarExpr(E->getOrder()); 2588 if (E->isCmpXChg()) { 2589 Val1 = EmitScalarExpr(E->getVal1()); 2590 Val2 = EmitValToTemp(*this, E->getVal2()); 2591 OrderFail = EmitScalarExpr(E->getOrderFail()); 2592 (void)OrderFail; // OrderFail is unused at the moment 2593 } else if ((E->getOp() == AtomicExpr::Add || E->getOp() == AtomicExpr::Sub) && 2594 MemTy->isPointerType()) { 2595 // For pointers, we're required to do a bit of math: adding 1 to an int* 2596 // is not the same as adding 1 to a uintptr_t. 2597 QualType Val1Ty = E->getVal1()->getType(); 2598 llvm::Value *Val1Scalar = EmitScalarExpr(E->getVal1()); 2599 CharUnits PointeeIncAmt = 2600 getContext().getTypeSizeInChars(MemTy->getPointeeType()); 2601 Val1Scalar = Builder.CreateMul(Val1Scalar, CGM.getSize(PointeeIncAmt)); 2602 Val1 = CreateMemTemp(Val1Ty, ".atomictmp"); 2603 EmitStoreOfScalar(Val1Scalar, MakeAddrLValue(Val1, Val1Ty)); 2604 } else if (E->getOp() != AtomicExpr::Load) { 2605 Val1 = EmitValToTemp(*this, E->getVal1()); 2606 } 2607 2608 if (E->getOp() != AtomicExpr::Store && !Dest) 2609 Dest = CreateMemTemp(E->getType(), ".atomicdst"); 2610 2611 if (UseLibcall) { 2612 // FIXME: Finalize what the libcalls are actually supposed to look like. 2613 // See also http://gcc.gnu.org/wiki/Atomic/GCCMM/LIbrary . 2614 return EmitUnsupportedRValue(E, "atomic library call"); 2615 } 2616 #if 0 2617 if (UseLibcall) { 2618 const char* LibCallName; 2619 switch (E->getOp()) { 2620 case AtomicExpr::CmpXchgWeak: 2621 LibCallName = "__atomic_compare_exchange_generic"; break; 2622 case AtomicExpr::CmpXchgStrong: 2623 LibCallName = "__atomic_compare_exchange_generic"; break; 2624 case AtomicExpr::Add: LibCallName = "__atomic_fetch_add_generic"; break; 2625 case AtomicExpr::Sub: LibCallName = "__atomic_fetch_sub_generic"; break; 2626 case AtomicExpr::And: LibCallName = "__atomic_fetch_and_generic"; break; 2627 case AtomicExpr::Or: LibCallName = "__atomic_fetch_or_generic"; break; 2628 case AtomicExpr::Xor: LibCallName = "__atomic_fetch_xor_generic"; break; 2629 case AtomicExpr::Xchg: LibCallName = "__atomic_exchange_generic"; break; 2630 case AtomicExpr::Store: LibCallName = "__atomic_store_generic"; break; 2631 case AtomicExpr::Load: LibCallName = "__atomic_load_generic"; break; 2632 } 2633 llvm::SmallVector<QualType, 4> Params; 2634 CallArgList Args; 2635 QualType RetTy = getContext().VoidTy; 2636 if (E->getOp() != AtomicExpr::Store && !E->isCmpXChg()) 2637 Args.add(RValue::get(EmitCastToVoidPtr(Dest)), 2638 getContext().VoidPtrTy); 2639 Args.add(RValue::get(EmitCastToVoidPtr(Ptr)), 2640 getContext().VoidPtrTy); 2641 if (E->getOp() != AtomicExpr::Load) 2642 Args.add(RValue::get(EmitCastToVoidPtr(Val1)), 2643 getContext().VoidPtrTy); 2644 if (E->isCmpXChg()) { 2645 Args.add(RValue::get(EmitCastToVoidPtr(Val2)), 2646 getContext().VoidPtrTy); 2647 RetTy = getContext().IntTy; 2648 } 2649 Args.add(RValue::get(llvm::ConstantInt::get(SizeTy, Size)), 2650 getContext().getSizeType()); 2651 const CGFunctionInfo &FuncInfo = 2652 CGM.getTypes().getFunctionInfo(RetTy, Args, FunctionType::ExtInfo()); 2653 llvm::FunctionType *FTy = CGM.getTypes().GetFunctionType(FuncInfo, false); 2654 llvm::Constant *Func = CGM.CreateRuntimeFunction(FTy, LibCallName); 2655 RValue Res = EmitCall(FuncInfo, Func, ReturnValueSlot(), Args); 2656 if (E->isCmpXChg()) 2657 return Res; 2658 if (E->getOp() == AtomicExpr::Store) 2659 return RValue::get(0); 2660 return ConvertTempToRValue(*this, E->getType(), Dest); 2661 } 2662 #endif 2663 llvm::Type *IPtrTy = 2664 llvm::IntegerType::get(getLLVMContext(), Size * 8)->getPointerTo(); 2665 llvm::Value *OrigDest = Dest; 2666 Ptr = Builder.CreateBitCast(Ptr, IPtrTy); 2667 if (Val1) Val1 = Builder.CreateBitCast(Val1, IPtrTy); 2668 if (Val2) Val2 = Builder.CreateBitCast(Val2, IPtrTy); 2669 if (Dest && !E->isCmpXChg()) Dest = Builder.CreateBitCast(Dest, IPtrTy); 2670 2671 if (isa<llvm::ConstantInt>(Order)) { 2672 int ord = cast<llvm::ConstantInt>(Order)->getZExtValue(); 2673 switch (ord) { 2674 case 0: // memory_order_relaxed 2675 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2676 llvm::Monotonic); 2677 break; 2678 case 1: // memory_order_consume 2679 case 2: // memory_order_acquire 2680 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2681 llvm::Acquire); 2682 break; 2683 case 3: // memory_order_release 2684 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2685 llvm::Release); 2686 break; 2687 case 4: // memory_order_acq_rel 2688 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2689 llvm::AcquireRelease); 2690 break; 2691 case 5: // memory_order_seq_cst 2692 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2693 llvm::SequentiallyConsistent); 2694 break; 2695 default: // invalid order 2696 // We should not ever get here normally, but it's hard to 2697 // enforce that in general. 2698 break; 2699 } 2700 if (E->getOp() == AtomicExpr::Store) 2701 return RValue::get(0); 2702 return ConvertTempToRValue(*this, E->getType(), OrigDest); 2703 } 2704 2705 // Long case, when Order isn't obviously constant. 2706 2707 // Create all the relevant BB's 2708 llvm::BasicBlock *MonotonicBB = 0, *AcquireBB = 0, *ReleaseBB = 0, 2709 *AcqRelBB = 0, *SeqCstBB = 0; 2710 MonotonicBB = createBasicBlock("monotonic", CurFn); 2711 if (E->getOp() != AtomicExpr::Store) 2712 AcquireBB = createBasicBlock("acquire", CurFn); 2713 if (E->getOp() != AtomicExpr::Load) 2714 ReleaseBB = createBasicBlock("release", CurFn); 2715 if (E->getOp() != AtomicExpr::Load && E->getOp() != AtomicExpr::Store) 2716 AcqRelBB = createBasicBlock("acqrel", CurFn); 2717 SeqCstBB = createBasicBlock("seqcst", CurFn); 2718 llvm::BasicBlock *ContBB = createBasicBlock("atomic.continue", CurFn); 2719 2720 // Create the switch for the split 2721 // MonotonicBB is arbitrarily chosen as the default case; in practice, this 2722 // doesn't matter unless someone is crazy enough to use something that 2723 // doesn't fold to a constant for the ordering. 2724 Order = Builder.CreateIntCast(Order, Builder.getInt32Ty(), false); 2725 llvm::SwitchInst *SI = Builder.CreateSwitch(Order, MonotonicBB); 2726 2727 // Emit all the different atomics 2728 Builder.SetInsertPoint(MonotonicBB); 2729 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2730 llvm::Monotonic); 2731 Builder.CreateBr(ContBB); 2732 if (E->getOp() != AtomicExpr::Store) { 2733 Builder.SetInsertPoint(AcquireBB); 2734 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2735 llvm::Acquire); 2736 Builder.CreateBr(ContBB); 2737 SI->addCase(Builder.getInt32(1), AcquireBB); 2738 SI->addCase(Builder.getInt32(2), AcquireBB); 2739 } 2740 if (E->getOp() != AtomicExpr::Load) { 2741 Builder.SetInsertPoint(ReleaseBB); 2742 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2743 llvm::Release); 2744 Builder.CreateBr(ContBB); 2745 SI->addCase(Builder.getInt32(3), ReleaseBB); 2746 } 2747 if (E->getOp() != AtomicExpr::Load && E->getOp() != AtomicExpr::Store) { 2748 Builder.SetInsertPoint(AcqRelBB); 2749 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2750 llvm::AcquireRelease); 2751 Builder.CreateBr(ContBB); 2752 SI->addCase(Builder.getInt32(4), AcqRelBB); 2753 } 2754 Builder.SetInsertPoint(SeqCstBB); 2755 EmitAtomicOp(*this, E, Dest, Ptr, Val1, Val2, Size, Align, 2756 llvm::SequentiallyConsistent); 2757 Builder.CreateBr(ContBB); 2758 SI->addCase(Builder.getInt32(5), SeqCstBB); 2759 2760 // Cleanup and return 2761 Builder.SetInsertPoint(ContBB); 2762 if (E->getOp() == AtomicExpr::Store) 2763 return RValue::get(0); 2764 return ConvertTempToRValue(*this, E->getType(), OrigDest); 2765 } 2766 2767 void CodeGenFunction::SetFPAccuracy(llvm::Value *Val, unsigned AccuracyN, 2768 unsigned AccuracyD) { 2769 assert(Val->getType()->isFPOrFPVectorTy()); 2770 if (!AccuracyN || !isa<llvm::Instruction>(Val)) 2771 return; 2772 2773 llvm::Value *Vals[2]; 2774 Vals[0] = llvm::ConstantInt::get(Int32Ty, AccuracyN); 2775 Vals[1] = llvm::ConstantInt::get(Int32Ty, AccuracyD); 2776 llvm::MDNode *Node = llvm::MDNode::get(getLLVMContext(), Vals); 2777 2778 cast<llvm::Instruction>(Val)->setMetadata(llvm::LLVMContext::MD_fpaccuracy, 2779 Node); 2780 } 2781 2782 namespace { 2783 struct LValueOrRValue { 2784 LValue LV; 2785 RValue RV; 2786 }; 2787 } 2788 2789 static LValueOrRValue emitPseudoObjectExpr(CodeGenFunction &CGF, 2790 const PseudoObjectExpr *E, 2791 bool forLValue, 2792 AggValueSlot slot) { 2793 llvm::SmallVector<CodeGenFunction::OpaqueValueMappingData, 4> opaques; 2794 2795 // Find the result expression, if any. 2796 const Expr *resultExpr = E->getResultExpr(); 2797 LValueOrRValue result; 2798 2799 for (PseudoObjectExpr::const_semantics_iterator 2800 i = E->semantics_begin(), e = E->semantics_end(); i != e; ++i) { 2801 const Expr *semantic = *i; 2802 2803 // If this semantic expression is an opaque value, bind it 2804 // to the result of its source expression. 2805 if (const OpaqueValueExpr *ov = dyn_cast<OpaqueValueExpr>(semantic)) { 2806 2807 // If this is the result expression, we may need to evaluate 2808 // directly into the slot. 2809 typedef CodeGenFunction::OpaqueValueMappingData OVMA; 2810 OVMA opaqueData; 2811 if (ov == resultExpr && ov->isRValue() && !forLValue && 2812 CodeGenFunction::hasAggregateLLVMType(ov->getType()) && 2813 !ov->getType()->isAnyComplexType()) { 2814 CGF.EmitAggExpr(ov->getSourceExpr(), slot); 2815 2816 LValue LV = CGF.MakeAddrLValue(slot.getAddr(), ov->getType()); 2817 opaqueData = OVMA::bind(CGF, ov, LV); 2818 result.RV = slot.asRValue(); 2819 2820 // Otherwise, emit as normal. 2821 } else { 2822 opaqueData = OVMA::bind(CGF, ov, ov->getSourceExpr()); 2823 2824 // If this is the result, also evaluate the result now. 2825 if (ov == resultExpr) { 2826 if (forLValue) 2827 result.LV = CGF.EmitLValue(ov); 2828 else 2829 result.RV = CGF.EmitAnyExpr(ov, slot); 2830 } 2831 } 2832 2833 opaques.push_back(opaqueData); 2834 2835 // Otherwise, if the expression is the result, evaluate it 2836 // and remember the result. 2837 } else if (semantic == resultExpr) { 2838 if (forLValue) 2839 result.LV = CGF.EmitLValue(semantic); 2840 else 2841 result.RV = CGF.EmitAnyExpr(semantic, slot); 2842 2843 // Otherwise, evaluate the expression in an ignored context. 2844 } else { 2845 CGF.EmitIgnoredExpr(semantic); 2846 } 2847 } 2848 2849 // Unbind all the opaques now. 2850 for (unsigned i = 0, e = opaques.size(); i != e; ++i) 2851 opaques[i].unbind(CGF); 2852 2853 return result; 2854 } 2855 2856 RValue CodeGenFunction::EmitPseudoObjectRValue(const PseudoObjectExpr *E, 2857 AggValueSlot slot) { 2858 return emitPseudoObjectExpr(*this, E, false, slot).RV; 2859 } 2860 2861 LValue CodeGenFunction::EmitPseudoObjectLValue(const PseudoObjectExpr *E) { 2862 return emitPseudoObjectExpr(*this, E, true, AggValueSlot::ignored()).LV; 2863 } 2864