1 //===--- CGExprCXX.cpp - Emit LLVM Code for C++ expressions ---------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This contains code dealing with code generation of C++ expressions 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenFunction.h" 15 using namespace clang; 16 using namespace CodeGen; 17 18 RValue CodeGenFunction::EmitCXXMemberCall(const CXXMethodDecl *MD, 19 llvm::Value *Callee, 20 ReturnValueSlot ReturnValue, 21 llvm::Value *This, 22 llvm::Value *VTT, 23 CallExpr::const_arg_iterator ArgBeg, 24 CallExpr::const_arg_iterator ArgEnd) { 25 assert(MD->isInstance() && 26 "Trying to emit a member call expr on a static method!"); 27 28 const FunctionProtoType *FPT = MD->getType()->getAs<FunctionProtoType>(); 29 30 CallArgList Args; 31 32 // Push the this ptr. 33 Args.push_back(std::make_pair(RValue::get(This), 34 MD->getThisType(getContext()))); 35 36 // If there is a VTT parameter, emit it. 37 if (VTT) { 38 QualType T = getContext().getPointerType(getContext().VoidPtrTy); 39 Args.push_back(std::make_pair(RValue::get(VTT), T)); 40 } 41 42 // And the rest of the call args 43 EmitCallArgs(Args, FPT, ArgBeg, ArgEnd); 44 45 QualType ResultType = MD->getType()->getAs<FunctionType>()->getResultType(); 46 return EmitCall(CGM.getTypes().getFunctionInfo(ResultType, Args), Callee, 47 ReturnValue, Args, MD); 48 } 49 50 /// canDevirtualizeMemberFunctionCalls - Checks whether virtual calls on given 51 /// expr can be devirtualized. 52 static bool canDevirtualizeMemberFunctionCalls(const Expr *Base) { 53 if (const DeclRefExpr *DRE = dyn_cast<DeclRefExpr>(Base)) { 54 if (const VarDecl *VD = dyn_cast<VarDecl>(DRE->getDecl())) { 55 // This is a record decl. We know the type and can devirtualize it. 56 return VD->getType()->isRecordType(); 57 } 58 59 return false; 60 } 61 62 // We can always devirtualize calls on temporary object expressions. 63 if (isa<CXXTemporaryObjectExpr>(Base)) 64 return true; 65 66 // And calls on bound temporaries. 67 if (isa<CXXBindTemporaryExpr>(Base)) 68 return true; 69 70 // Check if this is a call expr that returns a record type. 71 if (const CallExpr *CE = dyn_cast<CallExpr>(Base)) 72 return CE->getCallReturnType()->isRecordType(); 73 74 // We can't devirtualize the call. 75 return false; 76 } 77 78 RValue CodeGenFunction::EmitCXXMemberCallExpr(const CXXMemberCallExpr *CE, 79 ReturnValueSlot ReturnValue) { 80 if (isa<BinaryOperator>(CE->getCallee()->IgnoreParens())) 81 return EmitCXXMemberPointerCallExpr(CE, ReturnValue); 82 83 const MemberExpr *ME = cast<MemberExpr>(CE->getCallee()->IgnoreParens()); 84 const CXXMethodDecl *MD = cast<CXXMethodDecl>(ME->getMemberDecl()); 85 86 if (MD->isStatic()) { 87 // The method is static, emit it as we would a regular call. 88 llvm::Value *Callee = CGM.GetAddrOfFunction(MD); 89 return EmitCall(getContext().getPointerType(MD->getType()), Callee, 90 ReturnValue, CE->arg_begin(), CE->arg_end()); 91 } 92 93 const FunctionProtoType *FPT = MD->getType()->getAs<FunctionProtoType>(); 94 95 const llvm::Type *Ty = 96 CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(MD), 97 FPT->isVariadic()); 98 llvm::Value *This; 99 100 if (ME->isArrow()) 101 This = EmitScalarExpr(ME->getBase()); 102 else { 103 LValue BaseLV = EmitLValue(ME->getBase()); 104 This = BaseLV.getAddress(); 105 } 106 107 if (MD->isCopyAssignment() && MD->isTrivial()) { 108 // We don't like to generate the trivial copy assignment operator when 109 // it isn't necessary; just produce the proper effect here. 110 llvm::Value *RHS = EmitLValue(*CE->arg_begin()).getAddress(); 111 EmitAggregateCopy(This, RHS, CE->getType()); 112 return RValue::get(This); 113 } 114 115 // C++ [class.virtual]p12: 116 // Explicit qualification with the scope operator (5.1) suppresses the 117 // virtual call mechanism. 118 // 119 // We also don't emit a virtual call if the base expression has a record type 120 // because then we know what the type is. 121 llvm::Value *Callee; 122 if (const CXXDestructorDecl *Destructor 123 = dyn_cast<CXXDestructorDecl>(MD)) { 124 if (Destructor->isTrivial()) 125 return RValue::get(0); 126 if (MD->isVirtual() && !ME->hasQualifier() && 127 !canDevirtualizeMemberFunctionCalls(ME->getBase())) { 128 Callee = BuildVirtualCall(Destructor, Dtor_Complete, This, Ty); 129 } else { 130 Callee = CGM.GetAddrOfFunction(GlobalDecl(Destructor, Dtor_Complete), Ty); 131 } 132 } else if (MD->isVirtual() && !ME->hasQualifier() && 133 !canDevirtualizeMemberFunctionCalls(ME->getBase())) { 134 Callee = BuildVirtualCall(MD, This, Ty); 135 } else { 136 Callee = CGM.GetAddrOfFunction(MD, Ty); 137 } 138 139 return EmitCXXMemberCall(MD, Callee, ReturnValue, This, /*VTT=*/0, 140 CE->arg_begin(), CE->arg_end()); 141 } 142 143 RValue 144 CodeGenFunction::EmitCXXMemberPointerCallExpr(const CXXMemberCallExpr *E, 145 ReturnValueSlot ReturnValue) { 146 const BinaryOperator *BO = 147 cast<BinaryOperator>(E->getCallee()->IgnoreParens()); 148 const Expr *BaseExpr = BO->getLHS(); 149 const Expr *MemFnExpr = BO->getRHS(); 150 151 const MemberPointerType *MPT = 152 MemFnExpr->getType()->getAs<MemberPointerType>(); 153 const FunctionProtoType *FPT = 154 MPT->getPointeeType()->getAs<FunctionProtoType>(); 155 const CXXRecordDecl *RD = 156 cast<CXXRecordDecl>(MPT->getClass()->getAs<RecordType>()->getDecl()); 157 158 const llvm::FunctionType *FTy = 159 CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(RD, FPT), 160 FPT->isVariadic()); 161 162 const llvm::Type *Int8PtrTy = 163 llvm::Type::getInt8Ty(VMContext)->getPointerTo(); 164 165 // Get the member function pointer. 166 llvm::Value *MemFnPtr = 167 CreateTempAlloca(ConvertType(MemFnExpr->getType()), "mem.fn"); 168 EmitAggExpr(MemFnExpr, MemFnPtr, /*VolatileDest=*/false); 169 170 // Emit the 'this' pointer. 171 llvm::Value *This; 172 173 if (BO->getOpcode() == BinaryOperator::PtrMemI) 174 This = EmitScalarExpr(BaseExpr); 175 else 176 This = EmitLValue(BaseExpr).getAddress(); 177 178 // Adjust it. 179 llvm::Value *Adj = Builder.CreateStructGEP(MemFnPtr, 1); 180 Adj = Builder.CreateLoad(Adj, "mem.fn.adj"); 181 182 llvm::Value *Ptr = Builder.CreateBitCast(This, Int8PtrTy, "ptr"); 183 Ptr = Builder.CreateGEP(Ptr, Adj, "adj"); 184 185 This = Builder.CreateBitCast(Ptr, This->getType(), "this"); 186 187 llvm::Value *FnPtr = Builder.CreateStructGEP(MemFnPtr, 0, "mem.fn.ptr"); 188 189 const llvm::Type *PtrDiffTy = ConvertType(getContext().getPointerDiffType()); 190 191 llvm::Value *FnAsInt = Builder.CreateLoad(FnPtr, "fn"); 192 193 // If the LSB in the function pointer is 1, the function pointer points to 194 // a virtual function. 195 llvm::Value *IsVirtual 196 = Builder.CreateAnd(FnAsInt, llvm::ConstantInt::get(PtrDiffTy, 1), 197 "and"); 198 199 IsVirtual = Builder.CreateTrunc(IsVirtual, 200 llvm::Type::getInt1Ty(VMContext)); 201 202 llvm::BasicBlock *FnVirtual = createBasicBlock("fn.virtual"); 203 llvm::BasicBlock *FnNonVirtual = createBasicBlock("fn.nonvirtual"); 204 llvm::BasicBlock *FnEnd = createBasicBlock("fn.end"); 205 206 Builder.CreateCondBr(IsVirtual, FnVirtual, FnNonVirtual); 207 EmitBlock(FnVirtual); 208 209 const llvm::Type *VTableTy = 210 FTy->getPointerTo()->getPointerTo()->getPointerTo(); 211 212 llvm::Value *VTable = Builder.CreateBitCast(This, VTableTy); 213 VTable = Builder.CreateLoad(VTable); 214 215 VTable = Builder.CreateGEP(VTable, FnAsInt, "fn"); 216 217 // Since the function pointer is 1 plus the virtual table offset, we 218 // subtract 1 by using a GEP. 219 VTable = Builder.CreateConstGEP1_64(VTable, (uint64_t)-1); 220 221 llvm::Value *VirtualFn = Builder.CreateLoad(VTable, "virtualfn"); 222 223 EmitBranch(FnEnd); 224 EmitBlock(FnNonVirtual); 225 226 // If the function is not virtual, just load the pointer. 227 llvm::Value *NonVirtualFn = Builder.CreateLoad(FnPtr, "fn"); 228 NonVirtualFn = Builder.CreateIntToPtr(NonVirtualFn, FTy->getPointerTo()); 229 230 EmitBlock(FnEnd); 231 232 llvm::PHINode *Callee = Builder.CreatePHI(FTy->getPointerTo()); 233 Callee->reserveOperandSpace(2); 234 Callee->addIncoming(VirtualFn, FnVirtual); 235 Callee->addIncoming(NonVirtualFn, FnNonVirtual); 236 237 CallArgList Args; 238 239 QualType ThisType = 240 getContext().getPointerType(getContext().getTagDeclType(RD)); 241 242 // Push the this ptr. 243 Args.push_back(std::make_pair(RValue::get(This), ThisType)); 244 245 // And the rest of the call args 246 EmitCallArgs(Args, FPT, E->arg_begin(), E->arg_end()); 247 QualType ResultType = BO->getType()->getAs<FunctionType>()->getResultType(); 248 return EmitCall(CGM.getTypes().getFunctionInfo(ResultType, Args), Callee, 249 ReturnValue, Args); 250 } 251 252 RValue 253 CodeGenFunction::EmitCXXOperatorMemberCallExpr(const CXXOperatorCallExpr *E, 254 const CXXMethodDecl *MD, 255 ReturnValueSlot ReturnValue) { 256 assert(MD->isInstance() && 257 "Trying to emit a member call expr on a static method!"); 258 259 if (MD->isCopyAssignment()) { 260 const CXXRecordDecl *ClassDecl = cast<CXXRecordDecl>(MD->getDeclContext()); 261 if (ClassDecl->hasTrivialCopyAssignment()) { 262 assert(!ClassDecl->hasUserDeclaredCopyAssignment() && 263 "EmitCXXOperatorMemberCallExpr - user declared copy assignment"); 264 llvm::Value *This = EmitLValue(E->getArg(0)).getAddress(); 265 llvm::Value *Src = EmitLValue(E->getArg(1)).getAddress(); 266 QualType Ty = E->getType(); 267 EmitAggregateCopy(This, Src, Ty); 268 return RValue::get(This); 269 } 270 } 271 272 const FunctionProtoType *FPT = MD->getType()->getAs<FunctionProtoType>(); 273 const llvm::Type *Ty = 274 CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(MD), 275 FPT->isVariadic()); 276 277 llvm::Value *This = EmitLValue(E->getArg(0)).getAddress(); 278 279 llvm::Value *Callee; 280 if (MD->isVirtual() && !canDevirtualizeMemberFunctionCalls(E->getArg(0))) 281 Callee = BuildVirtualCall(MD, This, Ty); 282 else 283 Callee = CGM.GetAddrOfFunction(MD, Ty); 284 285 return EmitCXXMemberCall(MD, Callee, ReturnValue, This, /*VTT=*/0, 286 E->arg_begin() + 1, E->arg_end()); 287 } 288 289 void 290 CodeGenFunction::EmitCXXConstructExpr(llvm::Value *Dest, 291 const CXXConstructExpr *E) { 292 assert(Dest && "Must have a destination!"); 293 const CXXConstructorDecl *CD = E->getConstructor(); 294 const ConstantArrayType *Array = 295 getContext().getAsConstantArrayType(E->getType()); 296 // For a copy constructor, even if it is trivial, must fall thru so 297 // its argument is code-gen'ed. 298 if (!CD->isCopyConstructor()) { 299 QualType InitType = E->getType(); 300 if (Array) 301 InitType = getContext().getBaseElementType(Array); 302 const CXXRecordDecl *RD = 303 cast<CXXRecordDecl>(InitType->getAs<RecordType>()->getDecl()); 304 if (RD->hasTrivialConstructor()) 305 return; 306 } 307 // Code gen optimization to eliminate copy constructor and return 308 // its first argument instead. 309 if (getContext().getLangOptions().ElideConstructors && E->isElidable()) { 310 const Expr *Arg = E->getArg(0); 311 312 if (const ImplicitCastExpr *ICE = dyn_cast<ImplicitCastExpr>(Arg)) { 313 assert((ICE->getCastKind() == CastExpr::CK_NoOp || 314 ICE->getCastKind() == CastExpr::CK_ConstructorConversion || 315 ICE->getCastKind() == CastExpr::CK_UserDefinedConversion) && 316 "Unknown implicit cast kind in constructor elision"); 317 Arg = ICE->getSubExpr(); 318 } 319 320 if (const CXXFunctionalCastExpr *FCE = dyn_cast<CXXFunctionalCastExpr>(Arg)) 321 Arg = FCE->getSubExpr(); 322 323 if (const CXXBindTemporaryExpr *BindExpr = 324 dyn_cast<CXXBindTemporaryExpr>(Arg)) 325 Arg = BindExpr->getSubExpr(); 326 327 EmitAggExpr(Arg, Dest, false); 328 return; 329 } 330 if (Array) { 331 QualType BaseElementTy = getContext().getBaseElementType(Array); 332 const llvm::Type *BasePtr = ConvertType(BaseElementTy); 333 BasePtr = llvm::PointerType::getUnqual(BasePtr); 334 llvm::Value *BaseAddrPtr = 335 Builder.CreateBitCast(Dest, BasePtr); 336 337 EmitCXXAggrConstructorCall(CD, Array, BaseAddrPtr, 338 E->arg_begin(), E->arg_end()); 339 } 340 else 341 // Call the constructor. 342 EmitCXXConstructorCall(CD, Ctor_Complete, Dest, 343 E->arg_begin(), E->arg_end()); 344 } 345 346 static CharUnits CalculateCookiePadding(ASTContext &Ctx, QualType ElementType) { 347 const RecordType *RT = ElementType->getAs<RecordType>(); 348 if (!RT) 349 return CharUnits::Zero(); 350 351 const CXXRecordDecl *RD = dyn_cast<CXXRecordDecl>(RT->getDecl()); 352 if (!RD) 353 return CharUnits::Zero(); 354 355 // Check if the class has a trivial destructor. 356 if (RD->hasTrivialDestructor()) { 357 // Check if the usual deallocation function takes two arguments. 358 const CXXMethodDecl *UsualDeallocationFunction = 0; 359 360 DeclarationName OpName = 361 Ctx.DeclarationNames.getCXXOperatorName(OO_Array_Delete); 362 DeclContext::lookup_const_iterator Op, OpEnd; 363 for (llvm::tie(Op, OpEnd) = RD->lookup(OpName); 364 Op != OpEnd; ++Op) { 365 const CXXMethodDecl *Delete = cast<CXXMethodDecl>(*Op); 366 367 if (Delete->isUsualDeallocationFunction()) { 368 UsualDeallocationFunction = Delete; 369 break; 370 } 371 } 372 373 // No usual deallocation function, we don't need a cookie. 374 if (!UsualDeallocationFunction) 375 return CharUnits::Zero(); 376 377 // The usual deallocation function doesn't take a size_t argument, so we 378 // don't need a cookie. 379 if (UsualDeallocationFunction->getNumParams() == 1) 380 return CharUnits::Zero(); 381 382 assert(UsualDeallocationFunction->getNumParams() == 2 && 383 "Unexpected deallocation function type!"); 384 } 385 386 // Padding is the maximum of sizeof(size_t) and alignof(ElementType) 387 return std::max(Ctx.getTypeSizeInChars(Ctx.getSizeType()), 388 Ctx.getTypeAlignInChars(ElementType)); 389 } 390 391 static CharUnits CalculateCookiePadding(ASTContext &Ctx, const CXXNewExpr *E) { 392 if (!E->isArray()) 393 return CharUnits::Zero(); 394 395 // No cookie is required if the new operator being used is 396 // ::operator new[](size_t, void*). 397 const FunctionDecl *OperatorNew = E->getOperatorNew(); 398 if (OperatorNew->getDeclContext()->getLookupContext()->isFileContext()) { 399 if (OperatorNew->getNumParams() == 2) { 400 CanQualType ParamType = 401 Ctx.getCanonicalType(OperatorNew->getParamDecl(1)->getType()); 402 403 if (ParamType == Ctx.VoidPtrTy) 404 return CharUnits::Zero(); 405 } 406 } 407 408 return CalculateCookiePadding(Ctx, E->getAllocatedType()); 409 } 410 411 static llvm::Value *EmitCXXNewAllocSize(CodeGenFunction &CGF, 412 const CXXNewExpr *E, 413 llvm::Value *& NumElements) { 414 QualType Type = E->getAllocatedType(); 415 CharUnits TypeSize = CGF.getContext().getTypeSizeInChars(Type); 416 const llvm::Type *SizeTy = CGF.ConvertType(CGF.getContext().getSizeType()); 417 418 if (!E->isArray()) 419 return llvm::ConstantInt::get(SizeTy, TypeSize.getQuantity()); 420 421 CharUnits CookiePadding = CalculateCookiePadding(CGF.getContext(), E); 422 423 Expr::EvalResult Result; 424 if (E->getArraySize()->Evaluate(Result, CGF.getContext()) && 425 !Result.HasSideEffects && Result.Val.isInt()) { 426 427 CharUnits AllocSize = 428 Result.Val.getInt().getZExtValue() * TypeSize + CookiePadding; 429 430 NumElements = 431 llvm::ConstantInt::get(SizeTy, Result.Val.getInt().getZExtValue()); 432 433 return llvm::ConstantInt::get(SizeTy, AllocSize.getQuantity()); 434 } 435 436 // Emit the array size expression. 437 NumElements = CGF.EmitScalarExpr(E->getArraySize()); 438 439 // Multiply with the type size. 440 llvm::Value *V = 441 CGF.Builder.CreateMul(NumElements, 442 llvm::ConstantInt::get(SizeTy, 443 TypeSize.getQuantity())); 444 445 // And add the cookie padding if necessary. 446 if (!CookiePadding.isZero()) 447 V = CGF.Builder.CreateAdd(V, 448 llvm::ConstantInt::get(SizeTy, CookiePadding.getQuantity())); 449 450 return V; 451 } 452 453 static void EmitNewInitializer(CodeGenFunction &CGF, const CXXNewExpr *E, 454 llvm::Value *NewPtr, 455 llvm::Value *NumElements) { 456 if (E->isArray()) { 457 if (CXXConstructorDecl *Ctor = E->getConstructor()) 458 CGF.EmitCXXAggrConstructorCall(Ctor, NumElements, NewPtr, 459 E->constructor_arg_begin(), 460 E->constructor_arg_end()); 461 return; 462 } 463 464 QualType AllocType = E->getAllocatedType(); 465 466 if (CXXConstructorDecl *Ctor = E->getConstructor()) { 467 CGF.EmitCXXConstructorCall(Ctor, Ctor_Complete, NewPtr, 468 E->constructor_arg_begin(), 469 E->constructor_arg_end()); 470 471 return; 472 } 473 474 // We have a POD type. 475 if (E->getNumConstructorArgs() == 0) 476 return; 477 478 assert(E->getNumConstructorArgs() == 1 && 479 "Can only have one argument to initializer of POD type."); 480 481 const Expr *Init = E->getConstructorArg(0); 482 483 if (!CGF.hasAggregateLLVMType(AllocType)) 484 CGF.EmitStoreOfScalar(CGF.EmitScalarExpr(Init), NewPtr, 485 AllocType.isVolatileQualified(), AllocType); 486 else if (AllocType->isAnyComplexType()) 487 CGF.EmitComplexExprIntoAddr(Init, NewPtr, 488 AllocType.isVolatileQualified()); 489 else 490 CGF.EmitAggExpr(Init, NewPtr, AllocType.isVolatileQualified()); 491 } 492 493 llvm::Value *CodeGenFunction::EmitCXXNewExpr(const CXXNewExpr *E) { 494 QualType AllocType = E->getAllocatedType(); 495 FunctionDecl *NewFD = E->getOperatorNew(); 496 const FunctionProtoType *NewFTy = NewFD->getType()->getAs<FunctionProtoType>(); 497 498 CallArgList NewArgs; 499 500 // The allocation size is the first argument. 501 QualType SizeTy = getContext().getSizeType(); 502 503 llvm::Value *NumElements = 0; 504 llvm::Value *AllocSize = EmitCXXNewAllocSize(*this, E, NumElements); 505 506 NewArgs.push_back(std::make_pair(RValue::get(AllocSize), SizeTy)); 507 508 // Emit the rest of the arguments. 509 // FIXME: Ideally, this should just use EmitCallArgs. 510 CXXNewExpr::const_arg_iterator NewArg = E->placement_arg_begin(); 511 512 // First, use the types from the function type. 513 // We start at 1 here because the first argument (the allocation size) 514 // has already been emitted. 515 for (unsigned i = 1, e = NewFTy->getNumArgs(); i != e; ++i, ++NewArg) { 516 QualType ArgType = NewFTy->getArgType(i); 517 518 assert(getContext().getCanonicalType(ArgType.getNonReferenceType()). 519 getTypePtr() == 520 getContext().getCanonicalType(NewArg->getType()).getTypePtr() && 521 "type mismatch in call argument!"); 522 523 NewArgs.push_back(std::make_pair(EmitCallArg(*NewArg, ArgType), 524 ArgType)); 525 526 } 527 528 // Either we've emitted all the call args, or we have a call to a 529 // variadic function. 530 assert((NewArg == E->placement_arg_end() || NewFTy->isVariadic()) && 531 "Extra arguments in non-variadic function!"); 532 533 // If we still have any arguments, emit them using the type of the argument. 534 for (CXXNewExpr::const_arg_iterator NewArgEnd = E->placement_arg_end(); 535 NewArg != NewArgEnd; ++NewArg) { 536 QualType ArgType = NewArg->getType(); 537 NewArgs.push_back(std::make_pair(EmitCallArg(*NewArg, ArgType), 538 ArgType)); 539 } 540 541 // Emit the call to new. 542 RValue RV = 543 EmitCall(CGM.getTypes().getFunctionInfo(NewFTy->getResultType(), NewArgs), 544 CGM.GetAddrOfFunction(NewFD), ReturnValueSlot(), NewArgs, NewFD); 545 546 // If an allocation function is declared with an empty exception specification 547 // it returns null to indicate failure to allocate storage. [expr.new]p13. 548 // (We don't need to check for null when there's no new initializer and 549 // we're allocating a POD type). 550 bool NullCheckResult = NewFTy->hasEmptyExceptionSpec() && 551 !(AllocType->isPODType() && !E->hasInitializer()); 552 553 llvm::BasicBlock *NewNull = 0; 554 llvm::BasicBlock *NewNotNull = 0; 555 llvm::BasicBlock *NewEnd = 0; 556 557 llvm::Value *NewPtr = RV.getScalarVal(); 558 559 if (NullCheckResult) { 560 NewNull = createBasicBlock("new.null"); 561 NewNotNull = createBasicBlock("new.notnull"); 562 NewEnd = createBasicBlock("new.end"); 563 564 llvm::Value *IsNull = 565 Builder.CreateICmpEQ(NewPtr, 566 llvm::Constant::getNullValue(NewPtr->getType()), 567 "isnull"); 568 569 Builder.CreateCondBr(IsNull, NewNull, NewNotNull); 570 EmitBlock(NewNotNull); 571 } 572 573 CharUnits CookiePadding = CalculateCookiePadding(getContext(), E); 574 if (!CookiePadding.isZero()) { 575 CharUnits CookieOffset = 576 CookiePadding - getContext().getTypeSizeInChars(SizeTy); 577 578 llvm::Value *NumElementsPtr = 579 Builder.CreateConstInBoundsGEP1_64(NewPtr, CookieOffset.getQuantity()); 580 581 NumElementsPtr = Builder.CreateBitCast(NumElementsPtr, 582 ConvertType(SizeTy)->getPointerTo()); 583 Builder.CreateStore(NumElements, NumElementsPtr); 584 585 // Now add the padding to the new ptr. 586 NewPtr = Builder.CreateConstInBoundsGEP1_64(NewPtr, 587 CookiePadding.getQuantity()); 588 } 589 590 NewPtr = Builder.CreateBitCast(NewPtr, ConvertType(E->getType())); 591 592 EmitNewInitializer(*this, E, NewPtr, NumElements); 593 594 if (NullCheckResult) { 595 Builder.CreateBr(NewEnd); 596 NewNotNull = Builder.GetInsertBlock(); 597 EmitBlock(NewNull); 598 Builder.CreateBr(NewEnd); 599 EmitBlock(NewEnd); 600 601 llvm::PHINode *PHI = Builder.CreatePHI(NewPtr->getType()); 602 PHI->reserveOperandSpace(2); 603 PHI->addIncoming(NewPtr, NewNotNull); 604 PHI->addIncoming(llvm::Constant::getNullValue(NewPtr->getType()), NewNull); 605 606 NewPtr = PHI; 607 } 608 609 return NewPtr; 610 } 611 612 static std::pair<llvm::Value *, llvm::Value *> 613 GetAllocatedObjectPtrAndNumElements(CodeGenFunction &CGF, 614 llvm::Value *Ptr, QualType DeleteTy) { 615 QualType SizeTy = CGF.getContext().getSizeType(); 616 const llvm::Type *SizeLTy = CGF.ConvertType(SizeTy); 617 618 CharUnits DeleteTypeAlign = CGF.getContext().getTypeAlignInChars(DeleteTy); 619 CharUnits CookiePadding = 620 std::max(CGF.getContext().getTypeSizeInChars(SizeTy), 621 DeleteTypeAlign); 622 assert(!CookiePadding.isZero() && "CookiePadding should not be 0."); 623 624 const llvm::Type *Int8PtrTy = llvm::Type::getInt8PtrTy(CGF.getLLVMContext()); 625 CharUnits CookieOffset = 626 CookiePadding - CGF.getContext().getTypeSizeInChars(SizeTy); 627 628 llvm::Value *AllocatedObjectPtr = CGF.Builder.CreateBitCast(Ptr, Int8PtrTy); 629 AllocatedObjectPtr = 630 CGF.Builder.CreateConstInBoundsGEP1_64(AllocatedObjectPtr, 631 -CookiePadding.getQuantity()); 632 633 llvm::Value *NumElementsPtr = 634 CGF.Builder.CreateConstInBoundsGEP1_64(AllocatedObjectPtr, 635 CookieOffset.getQuantity()); 636 NumElementsPtr = 637 CGF.Builder.CreateBitCast(NumElementsPtr, SizeLTy->getPointerTo()); 638 639 llvm::Value *NumElements = CGF.Builder.CreateLoad(NumElementsPtr); 640 NumElements = 641 CGF.Builder.CreateIntCast(NumElements, SizeLTy, /*isSigned=*/false); 642 643 return std::make_pair(AllocatedObjectPtr, NumElements); 644 } 645 646 void CodeGenFunction::EmitDeleteCall(const FunctionDecl *DeleteFD, 647 llvm::Value *Ptr, 648 QualType DeleteTy) { 649 const FunctionProtoType *DeleteFTy = 650 DeleteFD->getType()->getAs<FunctionProtoType>(); 651 652 CallArgList DeleteArgs; 653 654 // Check if we need to pass the size to the delete operator. 655 llvm::Value *Size = 0; 656 QualType SizeTy; 657 if (DeleteFTy->getNumArgs() == 2) { 658 SizeTy = DeleteFTy->getArgType(1); 659 CharUnits DeleteTypeSize = getContext().getTypeSizeInChars(DeleteTy); 660 Size = llvm::ConstantInt::get(ConvertType(SizeTy), 661 DeleteTypeSize.getQuantity()); 662 } 663 664 if (DeleteFD->getOverloadedOperator() == OO_Array_Delete && 665 !CalculateCookiePadding(getContext(), DeleteTy).isZero()) { 666 // We need to get the number of elements in the array from the cookie. 667 llvm::Value *AllocatedObjectPtr; 668 llvm::Value *NumElements; 669 llvm::tie(AllocatedObjectPtr, NumElements) = 670 GetAllocatedObjectPtrAndNumElements(*this, Ptr, DeleteTy); 671 672 // Multiply the size with the number of elements. 673 if (Size) 674 Size = Builder.CreateMul(NumElements, Size); 675 676 Ptr = AllocatedObjectPtr; 677 } 678 679 QualType ArgTy = DeleteFTy->getArgType(0); 680 llvm::Value *DeletePtr = Builder.CreateBitCast(Ptr, ConvertType(ArgTy)); 681 DeleteArgs.push_back(std::make_pair(RValue::get(DeletePtr), ArgTy)); 682 683 if (Size) 684 DeleteArgs.push_back(std::make_pair(RValue::get(Size), SizeTy)); 685 686 // Emit the call to delete. 687 EmitCall(CGM.getTypes().getFunctionInfo(DeleteFTy->getResultType(), 688 DeleteArgs), 689 CGM.GetAddrOfFunction(DeleteFD), ReturnValueSlot(), 690 DeleteArgs, DeleteFD); 691 } 692 693 void CodeGenFunction::EmitCXXDeleteExpr(const CXXDeleteExpr *E) { 694 695 // Get at the argument before we performed the implicit conversion 696 // to void*. 697 const Expr *Arg = E->getArgument(); 698 while (const ImplicitCastExpr *ICE = dyn_cast<ImplicitCastExpr>(Arg)) { 699 if (ICE->getCastKind() != CastExpr::CK_UserDefinedConversion && 700 ICE->getType()->isVoidPointerType()) 701 Arg = ICE->getSubExpr(); 702 else 703 break; 704 } 705 706 QualType DeleteTy = Arg->getType()->getAs<PointerType>()->getPointeeType(); 707 708 llvm::Value *Ptr = EmitScalarExpr(Arg); 709 710 // Null check the pointer. 711 llvm::BasicBlock *DeleteNotNull = createBasicBlock("delete.notnull"); 712 llvm::BasicBlock *DeleteEnd = createBasicBlock("delete.end"); 713 714 llvm::Value *IsNull = 715 Builder.CreateICmpEQ(Ptr, llvm::Constant::getNullValue(Ptr->getType()), 716 "isnull"); 717 718 Builder.CreateCondBr(IsNull, DeleteEnd, DeleteNotNull); 719 EmitBlock(DeleteNotNull); 720 721 bool ShouldCallDelete = true; 722 723 // Call the destructor if necessary. 724 if (const RecordType *RT = DeleteTy->getAs<RecordType>()) { 725 if (CXXRecordDecl *RD = dyn_cast<CXXRecordDecl>(RT->getDecl())) { 726 if (!RD->hasTrivialDestructor()) { 727 const CXXDestructorDecl *Dtor = RD->getDestructor(getContext()); 728 if (E->isArrayForm()) { 729 llvm::Value *AllocatedObjectPtr; 730 llvm::Value *NumElements; 731 llvm::tie(AllocatedObjectPtr, NumElements) = 732 GetAllocatedObjectPtrAndNumElements(*this, Ptr, DeleteTy); 733 734 EmitCXXAggrDestructorCall(Dtor, NumElements, Ptr); 735 } else if (Dtor->isVirtual()) { 736 const llvm::Type *Ty = 737 CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(Dtor), 738 /*isVariadic=*/false); 739 740 llvm::Value *Callee = BuildVirtualCall(Dtor, Dtor_Deleting, Ptr, Ty); 741 EmitCXXMemberCall(Dtor, Callee, ReturnValueSlot(), Ptr, /*VTT=*/0, 742 0, 0); 743 744 // The dtor took care of deleting the object. 745 ShouldCallDelete = false; 746 } else 747 EmitCXXDestructorCall(Dtor, Dtor_Complete, Ptr); 748 } 749 } 750 } 751 752 if (ShouldCallDelete) 753 EmitDeleteCall(E->getOperatorDelete(), Ptr, DeleteTy); 754 755 EmitBlock(DeleteEnd); 756 } 757 758 llvm::Value * CodeGenFunction::EmitCXXTypeidExpr(const CXXTypeidExpr *E) { 759 QualType Ty = E->getType(); 760 const llvm::Type *LTy = ConvertType(Ty)->getPointerTo(); 761 762 if (E->isTypeOperand()) { 763 llvm::Constant *TypeInfo = 764 CGM.GetAddrOfRTTIDescriptor(E->getTypeOperand()); 765 return Builder.CreateBitCast(TypeInfo, LTy); 766 } 767 768 Expr *subE = E->getExprOperand(); 769 Ty = subE->getType(); 770 CanQualType CanTy = CGM.getContext().getCanonicalType(Ty); 771 Ty = CanTy.getUnqualifiedType().getNonReferenceType(); 772 if (const RecordType *RT = Ty->getAs<RecordType>()) { 773 const CXXRecordDecl *RD = cast<CXXRecordDecl>(RT->getDecl()); 774 if (RD->isPolymorphic()) { 775 // FIXME: if subE is an lvalue do 776 LValue Obj = EmitLValue(subE); 777 llvm::Value *This = Obj.getAddress(); 778 LTy = LTy->getPointerTo()->getPointerTo(); 779 llvm::Value *V = Builder.CreateBitCast(This, LTy); 780 // We need to do a zero check for *p, unless it has NonNullAttr. 781 // FIXME: PointerType->hasAttr<NonNullAttr>() 782 bool CanBeZero = false; 783 if (UnaryOperator *UO = dyn_cast<UnaryOperator>(subE->IgnoreParens())) 784 if (UO->getOpcode() == UnaryOperator::Deref) 785 CanBeZero = true; 786 if (CanBeZero) { 787 llvm::BasicBlock *NonZeroBlock = createBasicBlock(); 788 llvm::BasicBlock *ZeroBlock = createBasicBlock(); 789 790 llvm::Value *Zero = llvm::Constant::getNullValue(LTy); 791 Builder.CreateCondBr(Builder.CreateICmpNE(V, Zero), 792 NonZeroBlock, ZeroBlock); 793 EmitBlock(ZeroBlock); 794 /// Call __cxa_bad_typeid 795 const llvm::Type *ResultType = llvm::Type::getVoidTy(VMContext); 796 const llvm::FunctionType *FTy; 797 FTy = llvm::FunctionType::get(ResultType, false); 798 llvm::Value *F = CGM.CreateRuntimeFunction(FTy, "__cxa_bad_typeid"); 799 Builder.CreateCall(F)->setDoesNotReturn(); 800 Builder.CreateUnreachable(); 801 EmitBlock(NonZeroBlock); 802 } 803 V = Builder.CreateLoad(V, "vtable"); 804 V = Builder.CreateConstInBoundsGEP1_64(V, -1ULL); 805 V = Builder.CreateLoad(V); 806 return V; 807 } 808 } 809 return Builder.CreateBitCast(CGM.GetAddrOfRTTIDescriptor(Ty), LTy); 810 } 811 812 llvm::Value *CodeGenFunction::EmitDynamicCast(llvm::Value *V, 813 const CXXDynamicCastExpr *DCE) { 814 QualType SrcTy = DCE->getSubExpr()->getType(); 815 QualType DestTy = DCE->getTypeAsWritten(); 816 QualType InnerType = DestTy->getPointeeType(); 817 818 const llvm::Type *LTy = ConvertType(DCE->getType()); 819 820 bool CanBeZero = false; 821 bool ToVoid = false; 822 bool ThrowOnBad = false; 823 if (DestTy->isPointerType()) { 824 // FIXME: if PointerType->hasAttr<NonNullAttr>(), we don't set this 825 CanBeZero = true; 826 if (InnerType->isVoidType()) 827 ToVoid = true; 828 } else { 829 LTy = LTy->getPointerTo(); 830 ThrowOnBad = true; 831 } 832 833 if (SrcTy->isPointerType() || SrcTy->isReferenceType()) 834 SrcTy = SrcTy->getPointeeType(); 835 SrcTy = SrcTy.getUnqualifiedType(); 836 837 if (DestTy->isPointerType() || DestTy->isReferenceType()) 838 DestTy = DestTy->getPointeeType(); 839 DestTy = DestTy.getUnqualifiedType(); 840 841 llvm::BasicBlock *ContBlock = createBasicBlock(); 842 llvm::BasicBlock *NullBlock = 0; 843 llvm::BasicBlock *NonZeroBlock = 0; 844 if (CanBeZero) { 845 NonZeroBlock = createBasicBlock(); 846 NullBlock = createBasicBlock(); 847 Builder.CreateCondBr(Builder.CreateIsNotNull(V), NonZeroBlock, NullBlock); 848 EmitBlock(NonZeroBlock); 849 } 850 851 llvm::BasicBlock *BadCastBlock = 0; 852 853 const llvm::Type *PtrDiffTy = ConvertType(getContext().getPointerDiffType()); 854 855 // See if this is a dynamic_cast(void*) 856 if (ToVoid) { 857 llvm::Value *This = V; 858 V = Builder.CreateBitCast(This, PtrDiffTy->getPointerTo()->getPointerTo()); 859 V = Builder.CreateLoad(V, "vtable"); 860 V = Builder.CreateConstInBoundsGEP1_64(V, -2ULL); 861 V = Builder.CreateLoad(V, "offset to top"); 862 This = Builder.CreateBitCast(This, llvm::Type::getInt8PtrTy(VMContext)); 863 V = Builder.CreateInBoundsGEP(This, V); 864 V = Builder.CreateBitCast(V, LTy); 865 } else { 866 /// Call __dynamic_cast 867 const llvm::Type *ResultType = llvm::Type::getInt8PtrTy(VMContext); 868 const llvm::FunctionType *FTy; 869 std::vector<const llvm::Type*> ArgTys; 870 const llvm::Type *PtrToInt8Ty 871 = llvm::Type::getInt8Ty(VMContext)->getPointerTo(); 872 ArgTys.push_back(PtrToInt8Ty); 873 ArgTys.push_back(PtrToInt8Ty); 874 ArgTys.push_back(PtrToInt8Ty); 875 ArgTys.push_back(PtrDiffTy); 876 FTy = llvm::FunctionType::get(ResultType, ArgTys, false); 877 878 // FIXME: Calculate better hint. 879 llvm::Value *hint = llvm::ConstantInt::get(PtrDiffTy, -1ULL); 880 881 assert(SrcTy->isRecordType() && "Src type must be record type!"); 882 assert(DestTy->isRecordType() && "Dest type must be record type!"); 883 884 llvm::Value *SrcArg 885 = CGM.GetAddrOfRTTIDescriptor(SrcTy.getUnqualifiedType()); 886 llvm::Value *DestArg 887 = CGM.GetAddrOfRTTIDescriptor(DestTy.getUnqualifiedType()); 888 889 V = Builder.CreateBitCast(V, PtrToInt8Ty); 890 V = Builder.CreateCall4(CGM.CreateRuntimeFunction(FTy, "__dynamic_cast"), 891 V, SrcArg, DestArg, hint); 892 V = Builder.CreateBitCast(V, LTy); 893 894 if (ThrowOnBad) { 895 BadCastBlock = createBasicBlock(); 896 897 Builder.CreateCondBr(Builder.CreateIsNotNull(V), ContBlock, BadCastBlock); 898 EmitBlock(BadCastBlock); 899 /// Call __cxa_bad_cast 900 ResultType = llvm::Type::getVoidTy(VMContext); 901 const llvm::FunctionType *FBadTy; 902 FBadTy = llvm::FunctionType::get(ResultType, false); 903 llvm::Value *F = CGM.CreateRuntimeFunction(FBadTy, "__cxa_bad_cast"); 904 Builder.CreateCall(F)->setDoesNotReturn(); 905 Builder.CreateUnreachable(); 906 } 907 } 908 909 if (CanBeZero) { 910 Builder.CreateBr(ContBlock); 911 EmitBlock(NullBlock); 912 Builder.CreateBr(ContBlock); 913 } 914 EmitBlock(ContBlock); 915 if (CanBeZero) { 916 llvm::PHINode *PHI = Builder.CreatePHI(LTy); 917 PHI->reserveOperandSpace(2); 918 PHI->addIncoming(V, NonZeroBlock); 919 PHI->addIncoming(llvm::Constant::getNullValue(LTy), NullBlock); 920 V = PHI; 921 } 922 923 return V; 924 } 925