1 //===----- CGOpenMPRuntime.cpp - Interface to OpenMP Runtimes -------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This provides a class for OpenMP runtime code generation. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CGCXXABI.h" 15 #include "CGCleanup.h" 16 #include "CGOpenMPRuntime.h" 17 #include "CodeGenFunction.h" 18 #include "clang/CodeGen/ConstantInitBuilder.h" 19 #include "clang/AST/Decl.h" 20 #include "clang/AST/StmtOpenMP.h" 21 #include "llvm/ADT/ArrayRef.h" 22 #include "llvm/ADT/BitmaskEnum.h" 23 #include "llvm/Bitcode/BitcodeReader.h" 24 #include "llvm/IR/CallSite.h" 25 #include "llvm/IR/DerivedTypes.h" 26 #include "llvm/IR/GlobalValue.h" 27 #include "llvm/IR/Value.h" 28 #include "llvm/Support/Format.h" 29 #include "llvm/Support/raw_ostream.h" 30 #include <cassert> 31 32 using namespace clang; 33 using namespace CodeGen; 34 35 namespace { 36 /// \brief Base class for handling code generation inside OpenMP regions. 37 class CGOpenMPRegionInfo : public CodeGenFunction::CGCapturedStmtInfo { 38 public: 39 /// \brief Kinds of OpenMP regions used in codegen. 40 enum CGOpenMPRegionKind { 41 /// \brief Region with outlined function for standalone 'parallel' 42 /// directive. 43 ParallelOutlinedRegion, 44 /// \brief Region with outlined function for standalone 'task' directive. 45 TaskOutlinedRegion, 46 /// \brief Region for constructs that do not require function outlining, 47 /// like 'for', 'sections', 'atomic' etc. directives. 48 InlinedRegion, 49 /// \brief Region with outlined function for standalone 'target' directive. 50 TargetRegion, 51 }; 52 53 CGOpenMPRegionInfo(const CapturedStmt &CS, 54 const CGOpenMPRegionKind RegionKind, 55 const RegionCodeGenTy &CodeGen, OpenMPDirectiveKind Kind, 56 bool HasCancel) 57 : CGCapturedStmtInfo(CS, CR_OpenMP), RegionKind(RegionKind), 58 CodeGen(CodeGen), Kind(Kind), HasCancel(HasCancel) {} 59 60 CGOpenMPRegionInfo(const CGOpenMPRegionKind RegionKind, 61 const RegionCodeGenTy &CodeGen, OpenMPDirectiveKind Kind, 62 bool HasCancel) 63 : CGCapturedStmtInfo(CR_OpenMP), RegionKind(RegionKind), CodeGen(CodeGen), 64 Kind(Kind), HasCancel(HasCancel) {} 65 66 /// \brief Get a variable or parameter for storing global thread id 67 /// inside OpenMP construct. 68 virtual const VarDecl *getThreadIDVariable() const = 0; 69 70 /// \brief Emit the captured statement body. 71 void EmitBody(CodeGenFunction &CGF, const Stmt *S) override; 72 73 /// \brief Get an LValue for the current ThreadID variable. 74 /// \return LValue for thread id variable. This LValue always has type int32*. 75 virtual LValue getThreadIDVariableLValue(CodeGenFunction &CGF); 76 77 virtual void emitUntiedSwitch(CodeGenFunction & /*CGF*/) {} 78 79 CGOpenMPRegionKind getRegionKind() const { return RegionKind; } 80 81 OpenMPDirectiveKind getDirectiveKind() const { return Kind; } 82 83 bool hasCancel() const { return HasCancel; } 84 85 static bool classof(const CGCapturedStmtInfo *Info) { 86 return Info->getKind() == CR_OpenMP; 87 } 88 89 ~CGOpenMPRegionInfo() override = default; 90 91 protected: 92 CGOpenMPRegionKind RegionKind; 93 RegionCodeGenTy CodeGen; 94 OpenMPDirectiveKind Kind; 95 bool HasCancel; 96 }; 97 98 /// \brief API for captured statement code generation in OpenMP constructs. 99 class CGOpenMPOutlinedRegionInfo final : public CGOpenMPRegionInfo { 100 public: 101 CGOpenMPOutlinedRegionInfo(const CapturedStmt &CS, const VarDecl *ThreadIDVar, 102 const RegionCodeGenTy &CodeGen, 103 OpenMPDirectiveKind Kind, bool HasCancel, 104 StringRef HelperName) 105 : CGOpenMPRegionInfo(CS, ParallelOutlinedRegion, CodeGen, Kind, 106 HasCancel), 107 ThreadIDVar(ThreadIDVar), HelperName(HelperName) { 108 assert(ThreadIDVar != nullptr && "No ThreadID in OpenMP region."); 109 } 110 111 /// \brief Get a variable or parameter for storing global thread id 112 /// inside OpenMP construct. 113 const VarDecl *getThreadIDVariable() const override { return ThreadIDVar; } 114 115 /// \brief Get the name of the capture helper. 116 StringRef getHelperName() const override { return HelperName; } 117 118 static bool classof(const CGCapturedStmtInfo *Info) { 119 return CGOpenMPRegionInfo::classof(Info) && 120 cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == 121 ParallelOutlinedRegion; 122 } 123 124 private: 125 /// \brief A variable or parameter storing global thread id for OpenMP 126 /// constructs. 127 const VarDecl *ThreadIDVar; 128 StringRef HelperName; 129 }; 130 131 /// \brief API for captured statement code generation in OpenMP constructs. 132 class CGOpenMPTaskOutlinedRegionInfo final : public CGOpenMPRegionInfo { 133 public: 134 class UntiedTaskActionTy final : public PrePostActionTy { 135 bool Untied; 136 const VarDecl *PartIDVar; 137 const RegionCodeGenTy UntiedCodeGen; 138 llvm::SwitchInst *UntiedSwitch = nullptr; 139 140 public: 141 UntiedTaskActionTy(bool Tied, const VarDecl *PartIDVar, 142 const RegionCodeGenTy &UntiedCodeGen) 143 : Untied(!Tied), PartIDVar(PartIDVar), UntiedCodeGen(UntiedCodeGen) {} 144 void Enter(CodeGenFunction &CGF) override { 145 if (Untied) { 146 // Emit task switching point. 147 auto PartIdLVal = CGF.EmitLoadOfPointerLValue( 148 CGF.GetAddrOfLocalVar(PartIDVar), 149 PartIDVar->getType()->castAs<PointerType>()); 150 auto *Res = CGF.EmitLoadOfScalar(PartIdLVal, PartIDVar->getLocation()); 151 auto *DoneBB = CGF.createBasicBlock(".untied.done."); 152 UntiedSwitch = CGF.Builder.CreateSwitch(Res, DoneBB); 153 CGF.EmitBlock(DoneBB); 154 CGF.EmitBranchThroughCleanup(CGF.ReturnBlock); 155 CGF.EmitBlock(CGF.createBasicBlock(".untied.jmp.")); 156 UntiedSwitch->addCase(CGF.Builder.getInt32(0), 157 CGF.Builder.GetInsertBlock()); 158 emitUntiedSwitch(CGF); 159 } 160 } 161 void emitUntiedSwitch(CodeGenFunction &CGF) const { 162 if (Untied) { 163 auto PartIdLVal = CGF.EmitLoadOfPointerLValue( 164 CGF.GetAddrOfLocalVar(PartIDVar), 165 PartIDVar->getType()->castAs<PointerType>()); 166 CGF.EmitStoreOfScalar(CGF.Builder.getInt32(UntiedSwitch->getNumCases()), 167 PartIdLVal); 168 UntiedCodeGen(CGF); 169 CodeGenFunction::JumpDest CurPoint = 170 CGF.getJumpDestInCurrentScope(".untied.next."); 171 CGF.EmitBranchThroughCleanup(CGF.ReturnBlock); 172 CGF.EmitBlock(CGF.createBasicBlock(".untied.jmp.")); 173 UntiedSwitch->addCase(CGF.Builder.getInt32(UntiedSwitch->getNumCases()), 174 CGF.Builder.GetInsertBlock()); 175 CGF.EmitBranchThroughCleanup(CurPoint); 176 CGF.EmitBlock(CurPoint.getBlock()); 177 } 178 } 179 unsigned getNumberOfParts() const { return UntiedSwitch->getNumCases(); } 180 }; 181 CGOpenMPTaskOutlinedRegionInfo(const CapturedStmt &CS, 182 const VarDecl *ThreadIDVar, 183 const RegionCodeGenTy &CodeGen, 184 OpenMPDirectiveKind Kind, bool HasCancel, 185 const UntiedTaskActionTy &Action) 186 : CGOpenMPRegionInfo(CS, TaskOutlinedRegion, CodeGen, Kind, HasCancel), 187 ThreadIDVar(ThreadIDVar), Action(Action) { 188 assert(ThreadIDVar != nullptr && "No ThreadID in OpenMP region."); 189 } 190 191 /// \brief Get a variable or parameter for storing global thread id 192 /// inside OpenMP construct. 193 const VarDecl *getThreadIDVariable() const override { return ThreadIDVar; } 194 195 /// \brief Get an LValue for the current ThreadID variable. 196 LValue getThreadIDVariableLValue(CodeGenFunction &CGF) override; 197 198 /// \brief Get the name of the capture helper. 199 StringRef getHelperName() const override { return ".omp_outlined."; } 200 201 void emitUntiedSwitch(CodeGenFunction &CGF) override { 202 Action.emitUntiedSwitch(CGF); 203 } 204 205 static bool classof(const CGCapturedStmtInfo *Info) { 206 return CGOpenMPRegionInfo::classof(Info) && 207 cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == 208 TaskOutlinedRegion; 209 } 210 211 private: 212 /// \brief A variable or parameter storing global thread id for OpenMP 213 /// constructs. 214 const VarDecl *ThreadIDVar; 215 /// Action for emitting code for untied tasks. 216 const UntiedTaskActionTy &Action; 217 }; 218 219 /// \brief API for inlined captured statement code generation in OpenMP 220 /// constructs. 221 class CGOpenMPInlinedRegionInfo : public CGOpenMPRegionInfo { 222 public: 223 CGOpenMPInlinedRegionInfo(CodeGenFunction::CGCapturedStmtInfo *OldCSI, 224 const RegionCodeGenTy &CodeGen, 225 OpenMPDirectiveKind Kind, bool HasCancel) 226 : CGOpenMPRegionInfo(InlinedRegion, CodeGen, Kind, HasCancel), 227 OldCSI(OldCSI), 228 OuterRegionInfo(dyn_cast_or_null<CGOpenMPRegionInfo>(OldCSI)) {} 229 230 // \brief Retrieve the value of the context parameter. 231 llvm::Value *getContextValue() const override { 232 if (OuterRegionInfo) 233 return OuterRegionInfo->getContextValue(); 234 llvm_unreachable("No context value for inlined OpenMP region"); 235 } 236 237 void setContextValue(llvm::Value *V) override { 238 if (OuterRegionInfo) { 239 OuterRegionInfo->setContextValue(V); 240 return; 241 } 242 llvm_unreachable("No context value for inlined OpenMP region"); 243 } 244 245 /// \brief Lookup the captured field decl for a variable. 246 const FieldDecl *lookup(const VarDecl *VD) const override { 247 if (OuterRegionInfo) 248 return OuterRegionInfo->lookup(VD); 249 // If there is no outer outlined region,no need to lookup in a list of 250 // captured variables, we can use the original one. 251 return nullptr; 252 } 253 254 FieldDecl *getThisFieldDecl() const override { 255 if (OuterRegionInfo) 256 return OuterRegionInfo->getThisFieldDecl(); 257 return nullptr; 258 } 259 260 /// \brief Get a variable or parameter for storing global thread id 261 /// inside OpenMP construct. 262 const VarDecl *getThreadIDVariable() const override { 263 if (OuterRegionInfo) 264 return OuterRegionInfo->getThreadIDVariable(); 265 return nullptr; 266 } 267 268 /// \brief Get an LValue for the current ThreadID variable. 269 LValue getThreadIDVariableLValue(CodeGenFunction &CGF) override { 270 if (OuterRegionInfo) 271 return OuterRegionInfo->getThreadIDVariableLValue(CGF); 272 llvm_unreachable("No LValue for inlined OpenMP construct"); 273 } 274 275 /// \brief Get the name of the capture helper. 276 StringRef getHelperName() const override { 277 if (auto *OuterRegionInfo = getOldCSI()) 278 return OuterRegionInfo->getHelperName(); 279 llvm_unreachable("No helper name for inlined OpenMP construct"); 280 } 281 282 void emitUntiedSwitch(CodeGenFunction &CGF) override { 283 if (OuterRegionInfo) 284 OuterRegionInfo->emitUntiedSwitch(CGF); 285 } 286 287 CodeGenFunction::CGCapturedStmtInfo *getOldCSI() const { return OldCSI; } 288 289 static bool classof(const CGCapturedStmtInfo *Info) { 290 return CGOpenMPRegionInfo::classof(Info) && 291 cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == InlinedRegion; 292 } 293 294 ~CGOpenMPInlinedRegionInfo() override = default; 295 296 private: 297 /// \brief CodeGen info about outer OpenMP region. 298 CodeGenFunction::CGCapturedStmtInfo *OldCSI; 299 CGOpenMPRegionInfo *OuterRegionInfo; 300 }; 301 302 /// \brief API for captured statement code generation in OpenMP target 303 /// constructs. For this captures, implicit parameters are used instead of the 304 /// captured fields. The name of the target region has to be unique in a given 305 /// application so it is provided by the client, because only the client has 306 /// the information to generate that. 307 class CGOpenMPTargetRegionInfo final : public CGOpenMPRegionInfo { 308 public: 309 CGOpenMPTargetRegionInfo(const CapturedStmt &CS, 310 const RegionCodeGenTy &CodeGen, StringRef HelperName) 311 : CGOpenMPRegionInfo(CS, TargetRegion, CodeGen, OMPD_target, 312 /*HasCancel=*/false), 313 HelperName(HelperName) {} 314 315 /// \brief This is unused for target regions because each starts executing 316 /// with a single thread. 317 const VarDecl *getThreadIDVariable() const override { return nullptr; } 318 319 /// \brief Get the name of the capture helper. 320 StringRef getHelperName() const override { return HelperName; } 321 322 static bool classof(const CGCapturedStmtInfo *Info) { 323 return CGOpenMPRegionInfo::classof(Info) && 324 cast<CGOpenMPRegionInfo>(Info)->getRegionKind() == TargetRegion; 325 } 326 327 private: 328 StringRef HelperName; 329 }; 330 331 static void EmptyCodeGen(CodeGenFunction &, PrePostActionTy &) { 332 llvm_unreachable("No codegen for expressions"); 333 } 334 /// \brief API for generation of expressions captured in a innermost OpenMP 335 /// region. 336 class CGOpenMPInnerExprInfo final : public CGOpenMPInlinedRegionInfo { 337 public: 338 CGOpenMPInnerExprInfo(CodeGenFunction &CGF, const CapturedStmt &CS) 339 : CGOpenMPInlinedRegionInfo(CGF.CapturedStmtInfo, EmptyCodeGen, 340 OMPD_unknown, 341 /*HasCancel=*/false), 342 PrivScope(CGF) { 343 // Make sure the globals captured in the provided statement are local by 344 // using the privatization logic. We assume the same variable is not 345 // captured more than once. 346 for (auto &C : CS.captures()) { 347 if (!C.capturesVariable() && !C.capturesVariableByCopy()) 348 continue; 349 350 const VarDecl *VD = C.getCapturedVar(); 351 if (VD->isLocalVarDeclOrParm()) 352 continue; 353 354 DeclRefExpr DRE(const_cast<VarDecl *>(VD), 355 /*RefersToEnclosingVariableOrCapture=*/false, 356 VD->getType().getNonReferenceType(), VK_LValue, 357 C.getLocation()); 358 PrivScope.addPrivate(VD, [&CGF, &DRE]() -> Address { 359 return CGF.EmitLValue(&DRE).getAddress(); 360 }); 361 } 362 (void)PrivScope.Privatize(); 363 } 364 365 /// \brief Lookup the captured field decl for a variable. 366 const FieldDecl *lookup(const VarDecl *VD) const override { 367 if (auto *FD = CGOpenMPInlinedRegionInfo::lookup(VD)) 368 return FD; 369 return nullptr; 370 } 371 372 /// \brief Emit the captured statement body. 373 void EmitBody(CodeGenFunction &CGF, const Stmt *S) override { 374 llvm_unreachable("No body for expressions"); 375 } 376 377 /// \brief Get a variable or parameter for storing global thread id 378 /// inside OpenMP construct. 379 const VarDecl *getThreadIDVariable() const override { 380 llvm_unreachable("No thread id for expressions"); 381 } 382 383 /// \brief Get the name of the capture helper. 384 StringRef getHelperName() const override { 385 llvm_unreachable("No helper name for expressions"); 386 } 387 388 static bool classof(const CGCapturedStmtInfo *Info) { return false; } 389 390 private: 391 /// Private scope to capture global variables. 392 CodeGenFunction::OMPPrivateScope PrivScope; 393 }; 394 395 /// \brief RAII for emitting code of OpenMP constructs. 396 class InlinedOpenMPRegionRAII { 397 CodeGenFunction &CGF; 398 llvm::DenseMap<const VarDecl *, FieldDecl *> LambdaCaptureFields; 399 FieldDecl *LambdaThisCaptureField = nullptr; 400 const CodeGen::CGBlockInfo *BlockInfo = nullptr; 401 402 public: 403 /// \brief Constructs region for combined constructs. 404 /// \param CodeGen Code generation sequence for combined directives. Includes 405 /// a list of functions used for code generation of implicitly inlined 406 /// regions. 407 InlinedOpenMPRegionRAII(CodeGenFunction &CGF, const RegionCodeGenTy &CodeGen, 408 OpenMPDirectiveKind Kind, bool HasCancel) 409 : CGF(CGF) { 410 // Start emission for the construct. 411 CGF.CapturedStmtInfo = new CGOpenMPInlinedRegionInfo( 412 CGF.CapturedStmtInfo, CodeGen, Kind, HasCancel); 413 std::swap(CGF.LambdaCaptureFields, LambdaCaptureFields); 414 LambdaThisCaptureField = CGF.LambdaThisCaptureField; 415 CGF.LambdaThisCaptureField = nullptr; 416 BlockInfo = CGF.BlockInfo; 417 CGF.BlockInfo = nullptr; 418 } 419 420 ~InlinedOpenMPRegionRAII() { 421 // Restore original CapturedStmtInfo only if we're done with code emission. 422 auto *OldCSI = 423 cast<CGOpenMPInlinedRegionInfo>(CGF.CapturedStmtInfo)->getOldCSI(); 424 delete CGF.CapturedStmtInfo; 425 CGF.CapturedStmtInfo = OldCSI; 426 std::swap(CGF.LambdaCaptureFields, LambdaCaptureFields); 427 CGF.LambdaThisCaptureField = LambdaThisCaptureField; 428 CGF.BlockInfo = BlockInfo; 429 } 430 }; 431 432 /// \brief Values for bit flags used in the ident_t to describe the fields. 433 /// All enumeric elements are named and described in accordance with the code 434 /// from http://llvm.org/svn/llvm-project/openmp/trunk/runtime/src/kmp.h 435 enum OpenMPLocationFlags : unsigned { 436 /// \brief Use trampoline for internal microtask. 437 OMP_IDENT_IMD = 0x01, 438 /// \brief Use c-style ident structure. 439 OMP_IDENT_KMPC = 0x02, 440 /// \brief Atomic reduction option for kmpc_reduce. 441 OMP_ATOMIC_REDUCE = 0x10, 442 /// \brief Explicit 'barrier' directive. 443 OMP_IDENT_BARRIER_EXPL = 0x20, 444 /// \brief Implicit barrier in code. 445 OMP_IDENT_BARRIER_IMPL = 0x40, 446 /// \brief Implicit barrier in 'for' directive. 447 OMP_IDENT_BARRIER_IMPL_FOR = 0x40, 448 /// \brief Implicit barrier in 'sections' directive. 449 OMP_IDENT_BARRIER_IMPL_SECTIONS = 0xC0, 450 /// \brief Implicit barrier in 'single' directive. 451 OMP_IDENT_BARRIER_IMPL_SINGLE = 0x140, 452 /// Call of __kmp_for_static_init for static loop. 453 OMP_IDENT_WORK_LOOP = 0x200, 454 /// Call of __kmp_for_static_init for sections. 455 OMP_IDENT_WORK_SECTIONS = 0x400, 456 /// Call of __kmp_for_static_init for distribute. 457 OMP_IDENT_WORK_DISTRIBUTE = 0x800, 458 LLVM_MARK_AS_BITMASK_ENUM(/*LargestValue=*/OMP_IDENT_WORK_DISTRIBUTE) 459 }; 460 461 /// \brief Describes ident structure that describes a source location. 462 /// All descriptions are taken from 463 /// http://llvm.org/svn/llvm-project/openmp/trunk/runtime/src/kmp.h 464 /// Original structure: 465 /// typedef struct ident { 466 /// kmp_int32 reserved_1; /**< might be used in Fortran; 467 /// see above */ 468 /// kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; 469 /// KMP_IDENT_KMPC identifies this union 470 /// member */ 471 /// kmp_int32 reserved_2; /**< not really used in Fortran any more; 472 /// see above */ 473 ///#if USE_ITT_BUILD 474 /// /* but currently used for storing 475 /// region-specific ITT */ 476 /// /* contextual information. */ 477 ///#endif /* USE_ITT_BUILD */ 478 /// kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for 479 /// C++ */ 480 /// char const *psource; /**< String describing the source location. 481 /// The string is composed of semi-colon separated 482 // fields which describe the source file, 483 /// the function and a pair of line numbers that 484 /// delimit the construct. 485 /// */ 486 /// } ident_t; 487 enum IdentFieldIndex { 488 /// \brief might be used in Fortran 489 IdentField_Reserved_1, 490 /// \brief OMP_IDENT_xxx flags; OMP_IDENT_KMPC identifies this union member. 491 IdentField_Flags, 492 /// \brief Not really used in Fortran any more 493 IdentField_Reserved_2, 494 /// \brief Source[4] in Fortran, do not use for C++ 495 IdentField_Reserved_3, 496 /// \brief String describing the source location. The string is composed of 497 /// semi-colon separated fields which describe the source file, the function 498 /// and a pair of line numbers that delimit the construct. 499 IdentField_PSource 500 }; 501 502 /// \brief Schedule types for 'omp for' loops (these enumerators are taken from 503 /// the enum sched_type in kmp.h). 504 enum OpenMPSchedType { 505 /// \brief Lower bound for default (unordered) versions. 506 OMP_sch_lower = 32, 507 OMP_sch_static_chunked = 33, 508 OMP_sch_static = 34, 509 OMP_sch_dynamic_chunked = 35, 510 OMP_sch_guided_chunked = 36, 511 OMP_sch_runtime = 37, 512 OMP_sch_auto = 38, 513 /// static with chunk adjustment (e.g., simd) 514 OMP_sch_static_balanced_chunked = 45, 515 /// \brief Lower bound for 'ordered' versions. 516 OMP_ord_lower = 64, 517 OMP_ord_static_chunked = 65, 518 OMP_ord_static = 66, 519 OMP_ord_dynamic_chunked = 67, 520 OMP_ord_guided_chunked = 68, 521 OMP_ord_runtime = 69, 522 OMP_ord_auto = 70, 523 OMP_sch_default = OMP_sch_static, 524 /// \brief dist_schedule types 525 OMP_dist_sch_static_chunked = 91, 526 OMP_dist_sch_static = 92, 527 /// Support for OpenMP 4.5 monotonic and nonmonotonic schedule modifiers. 528 /// Set if the monotonic schedule modifier was present. 529 OMP_sch_modifier_monotonic = (1 << 29), 530 /// Set if the nonmonotonic schedule modifier was present. 531 OMP_sch_modifier_nonmonotonic = (1 << 30), 532 }; 533 534 enum OpenMPRTLFunction { 535 /// \brief Call to void __kmpc_fork_call(ident_t *loc, kmp_int32 argc, 536 /// kmpc_micro microtask, ...); 537 OMPRTL__kmpc_fork_call, 538 /// \brief Call to void *__kmpc_threadprivate_cached(ident_t *loc, 539 /// kmp_int32 global_tid, void *data, size_t size, void ***cache); 540 OMPRTL__kmpc_threadprivate_cached, 541 /// \brief Call to void __kmpc_threadprivate_register( ident_t *, 542 /// void *data, kmpc_ctor ctor, kmpc_cctor cctor, kmpc_dtor dtor); 543 OMPRTL__kmpc_threadprivate_register, 544 // Call to __kmpc_int32 kmpc_global_thread_num(ident_t *loc); 545 OMPRTL__kmpc_global_thread_num, 546 // Call to void __kmpc_critical(ident_t *loc, kmp_int32 global_tid, 547 // kmp_critical_name *crit); 548 OMPRTL__kmpc_critical, 549 // Call to void __kmpc_critical_with_hint(ident_t *loc, kmp_int32 550 // global_tid, kmp_critical_name *crit, uintptr_t hint); 551 OMPRTL__kmpc_critical_with_hint, 552 // Call to void __kmpc_end_critical(ident_t *loc, kmp_int32 global_tid, 553 // kmp_critical_name *crit); 554 OMPRTL__kmpc_end_critical, 555 // Call to kmp_int32 __kmpc_cancel_barrier(ident_t *loc, kmp_int32 556 // global_tid); 557 OMPRTL__kmpc_cancel_barrier, 558 // Call to void __kmpc_barrier(ident_t *loc, kmp_int32 global_tid); 559 OMPRTL__kmpc_barrier, 560 // Call to void __kmpc_for_static_fini(ident_t *loc, kmp_int32 global_tid); 561 OMPRTL__kmpc_for_static_fini, 562 // Call to void __kmpc_serialized_parallel(ident_t *loc, kmp_int32 563 // global_tid); 564 OMPRTL__kmpc_serialized_parallel, 565 // Call to void __kmpc_end_serialized_parallel(ident_t *loc, kmp_int32 566 // global_tid); 567 OMPRTL__kmpc_end_serialized_parallel, 568 // Call to void __kmpc_push_num_threads(ident_t *loc, kmp_int32 global_tid, 569 // kmp_int32 num_threads); 570 OMPRTL__kmpc_push_num_threads, 571 // Call to void __kmpc_flush(ident_t *loc); 572 OMPRTL__kmpc_flush, 573 // Call to kmp_int32 __kmpc_master(ident_t *, kmp_int32 global_tid); 574 OMPRTL__kmpc_master, 575 // Call to void __kmpc_end_master(ident_t *, kmp_int32 global_tid); 576 OMPRTL__kmpc_end_master, 577 // Call to kmp_int32 __kmpc_omp_taskyield(ident_t *, kmp_int32 global_tid, 578 // int end_part); 579 OMPRTL__kmpc_omp_taskyield, 580 // Call to kmp_int32 __kmpc_single(ident_t *, kmp_int32 global_tid); 581 OMPRTL__kmpc_single, 582 // Call to void __kmpc_end_single(ident_t *, kmp_int32 global_tid); 583 OMPRTL__kmpc_end_single, 584 // Call to kmp_task_t * __kmpc_omp_task_alloc(ident_t *, kmp_int32 gtid, 585 // kmp_int32 flags, size_t sizeof_kmp_task_t, size_t sizeof_shareds, 586 // kmp_routine_entry_t *task_entry); 587 OMPRTL__kmpc_omp_task_alloc, 588 // Call to kmp_int32 __kmpc_omp_task(ident_t *, kmp_int32 gtid, kmp_task_t * 589 // new_task); 590 OMPRTL__kmpc_omp_task, 591 // Call to void __kmpc_copyprivate(ident_t *loc, kmp_int32 global_tid, 592 // size_t cpy_size, void *cpy_data, void(*cpy_func)(void *, void *), 593 // kmp_int32 didit); 594 OMPRTL__kmpc_copyprivate, 595 // Call to kmp_int32 __kmpc_reduce(ident_t *loc, kmp_int32 global_tid, 596 // kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void 597 // (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck); 598 OMPRTL__kmpc_reduce, 599 // Call to kmp_int32 __kmpc_reduce_nowait(ident_t *loc, kmp_int32 600 // global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, 601 // void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name 602 // *lck); 603 OMPRTL__kmpc_reduce_nowait, 604 // Call to void __kmpc_end_reduce(ident_t *loc, kmp_int32 global_tid, 605 // kmp_critical_name *lck); 606 OMPRTL__kmpc_end_reduce, 607 // Call to void __kmpc_end_reduce_nowait(ident_t *loc, kmp_int32 global_tid, 608 // kmp_critical_name *lck); 609 OMPRTL__kmpc_end_reduce_nowait, 610 // Call to void __kmpc_omp_task_begin_if0(ident_t *, kmp_int32 gtid, 611 // kmp_task_t * new_task); 612 OMPRTL__kmpc_omp_task_begin_if0, 613 // Call to void __kmpc_omp_task_complete_if0(ident_t *, kmp_int32 gtid, 614 // kmp_task_t * new_task); 615 OMPRTL__kmpc_omp_task_complete_if0, 616 // Call to void __kmpc_ordered(ident_t *loc, kmp_int32 global_tid); 617 OMPRTL__kmpc_ordered, 618 // Call to void __kmpc_end_ordered(ident_t *loc, kmp_int32 global_tid); 619 OMPRTL__kmpc_end_ordered, 620 // Call to kmp_int32 __kmpc_omp_taskwait(ident_t *loc, kmp_int32 621 // global_tid); 622 OMPRTL__kmpc_omp_taskwait, 623 // Call to void __kmpc_taskgroup(ident_t *loc, kmp_int32 global_tid); 624 OMPRTL__kmpc_taskgroup, 625 // Call to void __kmpc_end_taskgroup(ident_t *loc, kmp_int32 global_tid); 626 OMPRTL__kmpc_end_taskgroup, 627 // Call to void __kmpc_push_proc_bind(ident_t *loc, kmp_int32 global_tid, 628 // int proc_bind); 629 OMPRTL__kmpc_push_proc_bind, 630 // Call to kmp_int32 __kmpc_omp_task_with_deps(ident_t *loc_ref, kmp_int32 631 // gtid, kmp_task_t * new_task, kmp_int32 ndeps, kmp_depend_info_t 632 // *dep_list, kmp_int32 ndeps_noalias, kmp_depend_info_t *noalias_dep_list); 633 OMPRTL__kmpc_omp_task_with_deps, 634 // Call to void __kmpc_omp_wait_deps(ident_t *loc_ref, kmp_int32 635 // gtid, kmp_int32 ndeps, kmp_depend_info_t *dep_list, kmp_int32 636 // ndeps_noalias, kmp_depend_info_t *noalias_dep_list); 637 OMPRTL__kmpc_omp_wait_deps, 638 // Call to kmp_int32 __kmpc_cancellationpoint(ident_t *loc, kmp_int32 639 // global_tid, kmp_int32 cncl_kind); 640 OMPRTL__kmpc_cancellationpoint, 641 // Call to kmp_int32 __kmpc_cancel(ident_t *loc, kmp_int32 global_tid, 642 // kmp_int32 cncl_kind); 643 OMPRTL__kmpc_cancel, 644 // Call to void __kmpc_push_num_teams(ident_t *loc, kmp_int32 global_tid, 645 // kmp_int32 num_teams, kmp_int32 thread_limit); 646 OMPRTL__kmpc_push_num_teams, 647 // Call to void __kmpc_fork_teams(ident_t *loc, kmp_int32 argc, kmpc_micro 648 // microtask, ...); 649 OMPRTL__kmpc_fork_teams, 650 // Call to void __kmpc_taskloop(ident_t *loc, int gtid, kmp_task_t *task, int 651 // if_val, kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup, int 652 // sched, kmp_uint64 grainsize, void *task_dup); 653 OMPRTL__kmpc_taskloop, 654 // Call to void __kmpc_doacross_init(ident_t *loc, kmp_int32 gtid, kmp_int32 655 // num_dims, struct kmp_dim *dims); 656 OMPRTL__kmpc_doacross_init, 657 // Call to void __kmpc_doacross_fini(ident_t *loc, kmp_int32 gtid); 658 OMPRTL__kmpc_doacross_fini, 659 // Call to void __kmpc_doacross_post(ident_t *loc, kmp_int32 gtid, kmp_int64 660 // *vec); 661 OMPRTL__kmpc_doacross_post, 662 // Call to void __kmpc_doacross_wait(ident_t *loc, kmp_int32 gtid, kmp_int64 663 // *vec); 664 OMPRTL__kmpc_doacross_wait, 665 // Call to void *__kmpc_task_reduction_init(int gtid, int num_data, void 666 // *data); 667 OMPRTL__kmpc_task_reduction_init, 668 // Call to void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void 669 // *d); 670 OMPRTL__kmpc_task_reduction_get_th_data, 671 672 // 673 // Offloading related calls 674 // 675 // Call to int32_t __tgt_target(int64_t device_id, void *host_ptr, int32_t 676 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 677 // *arg_types); 678 OMPRTL__tgt_target, 679 // Call to int32_t __tgt_target_nowait(int64_t device_id, void *host_ptr, 680 // int32_t arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 681 // *arg_types); 682 OMPRTL__tgt_target_nowait, 683 // Call to int32_t __tgt_target_teams(int64_t device_id, void *host_ptr, 684 // int32_t arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 685 // *arg_types, int32_t num_teams, int32_t thread_limit); 686 OMPRTL__tgt_target_teams, 687 // Call to int32_t __tgt_target_teams_nowait(int64_t device_id, void 688 // *host_ptr, int32_t arg_num, void** args_base, void **args, size_t 689 // *arg_sizes, int64_t *arg_types, int32_t num_teams, int32_t thread_limit); 690 OMPRTL__tgt_target_teams_nowait, 691 // Call to void __tgt_register_lib(__tgt_bin_desc *desc); 692 OMPRTL__tgt_register_lib, 693 // Call to void __tgt_unregister_lib(__tgt_bin_desc *desc); 694 OMPRTL__tgt_unregister_lib, 695 // Call to void __tgt_target_data_begin(int64_t device_id, int32_t arg_num, 696 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 697 OMPRTL__tgt_target_data_begin, 698 // Call to void __tgt_target_data_begin_nowait(int64_t device_id, int32_t 699 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 700 // *arg_types); 701 OMPRTL__tgt_target_data_begin_nowait, 702 // Call to void __tgt_target_data_end(int64_t device_id, int32_t arg_num, 703 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 704 OMPRTL__tgt_target_data_end, 705 // Call to void __tgt_target_data_end_nowait(int64_t device_id, int32_t 706 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 707 // *arg_types); 708 OMPRTL__tgt_target_data_end_nowait, 709 // Call to void __tgt_target_data_update(int64_t device_id, int32_t arg_num, 710 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 711 OMPRTL__tgt_target_data_update, 712 // Call to void __tgt_target_data_update_nowait(int64_t device_id, int32_t 713 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 714 // *arg_types); 715 OMPRTL__tgt_target_data_update_nowait, 716 }; 717 718 /// A basic class for pre|post-action for advanced codegen sequence for OpenMP 719 /// region. 720 class CleanupTy final : public EHScopeStack::Cleanup { 721 PrePostActionTy *Action; 722 723 public: 724 explicit CleanupTy(PrePostActionTy *Action) : Action(Action) {} 725 void Emit(CodeGenFunction &CGF, Flags /*flags*/) override { 726 if (!CGF.HaveInsertPoint()) 727 return; 728 Action->Exit(CGF); 729 } 730 }; 731 732 } // anonymous namespace 733 734 void RegionCodeGenTy::operator()(CodeGenFunction &CGF) const { 735 CodeGenFunction::RunCleanupsScope Scope(CGF); 736 if (PrePostAction) { 737 CGF.EHStack.pushCleanup<CleanupTy>(NormalAndEHCleanup, PrePostAction); 738 Callback(CodeGen, CGF, *PrePostAction); 739 } else { 740 PrePostActionTy Action; 741 Callback(CodeGen, CGF, Action); 742 } 743 } 744 745 /// Check if the combiner is a call to UDR combiner and if it is so return the 746 /// UDR decl used for reduction. 747 static const OMPDeclareReductionDecl * 748 getReductionInit(const Expr *ReductionOp) { 749 if (auto *CE = dyn_cast<CallExpr>(ReductionOp)) 750 if (auto *OVE = dyn_cast<OpaqueValueExpr>(CE->getCallee())) 751 if (auto *DRE = 752 dyn_cast<DeclRefExpr>(OVE->getSourceExpr()->IgnoreImpCasts())) 753 if (auto *DRD = dyn_cast<OMPDeclareReductionDecl>(DRE->getDecl())) 754 return DRD; 755 return nullptr; 756 } 757 758 static void emitInitWithReductionInitializer(CodeGenFunction &CGF, 759 const OMPDeclareReductionDecl *DRD, 760 const Expr *InitOp, 761 Address Private, Address Original, 762 QualType Ty) { 763 if (DRD->getInitializer()) { 764 std::pair<llvm::Function *, llvm::Function *> Reduction = 765 CGF.CGM.getOpenMPRuntime().getUserDefinedReduction(DRD); 766 auto *CE = cast<CallExpr>(InitOp); 767 auto *OVE = cast<OpaqueValueExpr>(CE->getCallee()); 768 const Expr *LHS = CE->getArg(/*Arg=*/0)->IgnoreParenImpCasts(); 769 const Expr *RHS = CE->getArg(/*Arg=*/1)->IgnoreParenImpCasts(); 770 auto *LHSDRE = cast<DeclRefExpr>(cast<UnaryOperator>(LHS)->getSubExpr()); 771 auto *RHSDRE = cast<DeclRefExpr>(cast<UnaryOperator>(RHS)->getSubExpr()); 772 CodeGenFunction::OMPPrivateScope PrivateScope(CGF); 773 PrivateScope.addPrivate(cast<VarDecl>(LHSDRE->getDecl()), 774 [=]() -> Address { return Private; }); 775 PrivateScope.addPrivate(cast<VarDecl>(RHSDRE->getDecl()), 776 [=]() -> Address { return Original; }); 777 (void)PrivateScope.Privatize(); 778 RValue Func = RValue::get(Reduction.second); 779 CodeGenFunction::OpaqueValueMapping Map(CGF, OVE, Func); 780 CGF.EmitIgnoredExpr(InitOp); 781 } else { 782 llvm::Constant *Init = CGF.CGM.EmitNullConstant(Ty); 783 auto *GV = new llvm::GlobalVariable( 784 CGF.CGM.getModule(), Init->getType(), /*isConstant=*/true, 785 llvm::GlobalValue::PrivateLinkage, Init, ".init"); 786 LValue LV = CGF.MakeNaturalAlignAddrLValue(GV, Ty); 787 RValue InitRVal; 788 switch (CGF.getEvaluationKind(Ty)) { 789 case TEK_Scalar: 790 InitRVal = CGF.EmitLoadOfLValue(LV, DRD->getLocation()); 791 break; 792 case TEK_Complex: 793 InitRVal = 794 RValue::getComplex(CGF.EmitLoadOfComplex(LV, DRD->getLocation())); 795 break; 796 case TEK_Aggregate: 797 InitRVal = RValue::getAggregate(LV.getAddress()); 798 break; 799 } 800 OpaqueValueExpr OVE(DRD->getLocation(), Ty, VK_RValue); 801 CodeGenFunction::OpaqueValueMapping OpaqueMap(CGF, &OVE, InitRVal); 802 CGF.EmitAnyExprToMem(&OVE, Private, Ty.getQualifiers(), 803 /*IsInitializer=*/false); 804 } 805 } 806 807 /// \brief Emit initialization of arrays of complex types. 808 /// \param DestAddr Address of the array. 809 /// \param Type Type of array. 810 /// \param Init Initial expression of array. 811 /// \param SrcAddr Address of the original array. 812 static void EmitOMPAggregateInit(CodeGenFunction &CGF, Address DestAddr, 813 QualType Type, bool EmitDeclareReductionInit, 814 const Expr *Init, 815 const OMPDeclareReductionDecl *DRD, 816 Address SrcAddr = Address::invalid()) { 817 // Perform element-by-element initialization. 818 QualType ElementTy; 819 820 // Drill down to the base element type on both arrays. 821 auto ArrayTy = Type->getAsArrayTypeUnsafe(); 822 auto NumElements = CGF.emitArrayLength(ArrayTy, ElementTy, DestAddr); 823 DestAddr = 824 CGF.Builder.CreateElementBitCast(DestAddr, DestAddr.getElementType()); 825 if (DRD) 826 SrcAddr = 827 CGF.Builder.CreateElementBitCast(SrcAddr, DestAddr.getElementType()); 828 829 llvm::Value *SrcBegin = nullptr; 830 if (DRD) 831 SrcBegin = SrcAddr.getPointer(); 832 auto DestBegin = DestAddr.getPointer(); 833 // Cast from pointer to array type to pointer to single element. 834 auto DestEnd = CGF.Builder.CreateGEP(DestBegin, NumElements); 835 // The basic structure here is a while-do loop. 836 auto BodyBB = CGF.createBasicBlock("omp.arrayinit.body"); 837 auto DoneBB = CGF.createBasicBlock("omp.arrayinit.done"); 838 auto IsEmpty = 839 CGF.Builder.CreateICmpEQ(DestBegin, DestEnd, "omp.arrayinit.isempty"); 840 CGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB); 841 842 // Enter the loop body, making that address the current address. 843 auto EntryBB = CGF.Builder.GetInsertBlock(); 844 CGF.EmitBlock(BodyBB); 845 846 CharUnits ElementSize = CGF.getContext().getTypeSizeInChars(ElementTy); 847 848 llvm::PHINode *SrcElementPHI = nullptr; 849 Address SrcElementCurrent = Address::invalid(); 850 if (DRD) { 851 SrcElementPHI = CGF.Builder.CreatePHI(SrcBegin->getType(), 2, 852 "omp.arraycpy.srcElementPast"); 853 SrcElementPHI->addIncoming(SrcBegin, EntryBB); 854 SrcElementCurrent = 855 Address(SrcElementPHI, 856 SrcAddr.getAlignment().alignmentOfArrayElement(ElementSize)); 857 } 858 llvm::PHINode *DestElementPHI = CGF.Builder.CreatePHI( 859 DestBegin->getType(), 2, "omp.arraycpy.destElementPast"); 860 DestElementPHI->addIncoming(DestBegin, EntryBB); 861 Address DestElementCurrent = 862 Address(DestElementPHI, 863 DestAddr.getAlignment().alignmentOfArrayElement(ElementSize)); 864 865 // Emit copy. 866 { 867 CodeGenFunction::RunCleanupsScope InitScope(CGF); 868 if (EmitDeclareReductionInit) { 869 emitInitWithReductionInitializer(CGF, DRD, Init, DestElementCurrent, 870 SrcElementCurrent, ElementTy); 871 } else 872 CGF.EmitAnyExprToMem(Init, DestElementCurrent, ElementTy.getQualifiers(), 873 /*IsInitializer=*/false); 874 } 875 876 if (DRD) { 877 // Shift the address forward by one element. 878 auto SrcElementNext = CGF.Builder.CreateConstGEP1_32( 879 SrcElementPHI, /*Idx0=*/1, "omp.arraycpy.dest.element"); 880 SrcElementPHI->addIncoming(SrcElementNext, CGF.Builder.GetInsertBlock()); 881 } 882 883 // Shift the address forward by one element. 884 auto DestElementNext = CGF.Builder.CreateConstGEP1_32( 885 DestElementPHI, /*Idx0=*/1, "omp.arraycpy.dest.element"); 886 // Check whether we've reached the end. 887 auto Done = 888 CGF.Builder.CreateICmpEQ(DestElementNext, DestEnd, "omp.arraycpy.done"); 889 CGF.Builder.CreateCondBr(Done, DoneBB, BodyBB); 890 DestElementPHI->addIncoming(DestElementNext, CGF.Builder.GetInsertBlock()); 891 892 // Done. 893 CGF.EmitBlock(DoneBB, /*IsFinished=*/true); 894 } 895 896 LValue ReductionCodeGen::emitSharedLValue(CodeGenFunction &CGF, const Expr *E) { 897 return CGF.EmitOMPSharedLValue(E); 898 } 899 900 LValue ReductionCodeGen::emitSharedLValueUB(CodeGenFunction &CGF, 901 const Expr *E) { 902 if (const auto *OASE = dyn_cast<OMPArraySectionExpr>(E)) 903 return CGF.EmitOMPArraySectionExpr(OASE, /*IsLowerBound=*/false); 904 return LValue(); 905 } 906 907 void ReductionCodeGen::emitAggregateInitialization( 908 CodeGenFunction &CGF, unsigned N, Address PrivateAddr, LValue SharedLVal, 909 const OMPDeclareReductionDecl *DRD) { 910 // Emit VarDecl with copy init for arrays. 911 // Get the address of the original variable captured in current 912 // captured region. 913 auto *PrivateVD = 914 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 915 bool EmitDeclareReductionInit = 916 DRD && (DRD->getInitializer() || !PrivateVD->hasInit()); 917 EmitOMPAggregateInit(CGF, PrivateAddr, PrivateVD->getType(), 918 EmitDeclareReductionInit, 919 EmitDeclareReductionInit ? ClausesData[N].ReductionOp 920 : PrivateVD->getInit(), 921 DRD, SharedLVal.getAddress()); 922 } 923 924 ReductionCodeGen::ReductionCodeGen(ArrayRef<const Expr *> Shareds, 925 ArrayRef<const Expr *> Privates, 926 ArrayRef<const Expr *> ReductionOps) { 927 ClausesData.reserve(Shareds.size()); 928 SharedAddresses.reserve(Shareds.size()); 929 Sizes.reserve(Shareds.size()); 930 BaseDecls.reserve(Shareds.size()); 931 auto IPriv = Privates.begin(); 932 auto IRed = ReductionOps.begin(); 933 for (const auto *Ref : Shareds) { 934 ClausesData.emplace_back(Ref, *IPriv, *IRed); 935 std::advance(IPriv, 1); 936 std::advance(IRed, 1); 937 } 938 } 939 940 void ReductionCodeGen::emitSharedLValue(CodeGenFunction &CGF, unsigned N) { 941 assert(SharedAddresses.size() == N && 942 "Number of generated lvalues must be exactly N."); 943 LValue First = emitSharedLValue(CGF, ClausesData[N].Ref); 944 LValue Second = emitSharedLValueUB(CGF, ClausesData[N].Ref); 945 SharedAddresses.emplace_back(First, Second); 946 } 947 948 void ReductionCodeGen::emitAggregateType(CodeGenFunction &CGF, unsigned N) { 949 auto *PrivateVD = 950 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 951 QualType PrivateType = PrivateVD->getType(); 952 bool AsArraySection = isa<OMPArraySectionExpr>(ClausesData[N].Ref); 953 if (!PrivateType->isVariablyModifiedType()) { 954 Sizes.emplace_back( 955 CGF.getTypeSize( 956 SharedAddresses[N].first.getType().getNonReferenceType()), 957 nullptr); 958 return; 959 } 960 llvm::Value *Size; 961 llvm::Value *SizeInChars; 962 llvm::Type *ElemType = 963 cast<llvm::PointerType>(SharedAddresses[N].first.getPointer()->getType()) 964 ->getElementType(); 965 auto *ElemSizeOf = llvm::ConstantExpr::getSizeOf(ElemType); 966 if (AsArraySection) { 967 Size = CGF.Builder.CreatePtrDiff(SharedAddresses[N].second.getPointer(), 968 SharedAddresses[N].first.getPointer()); 969 Size = CGF.Builder.CreateNUWAdd( 970 Size, llvm::ConstantInt::get(Size->getType(), /*V=*/1)); 971 SizeInChars = CGF.Builder.CreateNUWMul(Size, ElemSizeOf); 972 } else { 973 SizeInChars = CGF.getTypeSize( 974 SharedAddresses[N].first.getType().getNonReferenceType()); 975 Size = CGF.Builder.CreateExactUDiv(SizeInChars, ElemSizeOf); 976 } 977 Sizes.emplace_back(SizeInChars, Size); 978 CodeGenFunction::OpaqueValueMapping OpaqueMap( 979 CGF, 980 cast<OpaqueValueExpr>( 981 CGF.getContext().getAsVariableArrayType(PrivateType)->getSizeExpr()), 982 RValue::get(Size)); 983 CGF.EmitVariablyModifiedType(PrivateType); 984 } 985 986 void ReductionCodeGen::emitAggregateType(CodeGenFunction &CGF, unsigned N, 987 llvm::Value *Size) { 988 auto *PrivateVD = 989 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 990 QualType PrivateType = PrivateVD->getType(); 991 if (!PrivateType->isVariablyModifiedType()) { 992 assert(!Size && !Sizes[N].second && 993 "Size should be nullptr for non-variably modified reduction " 994 "items."); 995 return; 996 } 997 CodeGenFunction::OpaqueValueMapping OpaqueMap( 998 CGF, 999 cast<OpaqueValueExpr>( 1000 CGF.getContext().getAsVariableArrayType(PrivateType)->getSizeExpr()), 1001 RValue::get(Size)); 1002 CGF.EmitVariablyModifiedType(PrivateType); 1003 } 1004 1005 void ReductionCodeGen::emitInitialization( 1006 CodeGenFunction &CGF, unsigned N, Address PrivateAddr, LValue SharedLVal, 1007 llvm::function_ref<bool(CodeGenFunction &)> DefaultInit) { 1008 assert(SharedAddresses.size() > N && "No variable was generated"); 1009 auto *PrivateVD = 1010 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 1011 auto *DRD = getReductionInit(ClausesData[N].ReductionOp); 1012 QualType PrivateType = PrivateVD->getType(); 1013 PrivateAddr = CGF.Builder.CreateElementBitCast( 1014 PrivateAddr, CGF.ConvertTypeForMem(PrivateType)); 1015 QualType SharedType = SharedAddresses[N].first.getType(); 1016 SharedLVal = CGF.MakeAddrLValue( 1017 CGF.Builder.CreateElementBitCast(SharedLVal.getAddress(), 1018 CGF.ConvertTypeForMem(SharedType)), 1019 SharedType, SharedAddresses[N].first.getBaseInfo(), 1020 CGF.CGM.getTBAAInfoForSubobject(SharedAddresses[N].first, SharedType)); 1021 if (CGF.getContext().getAsArrayType(PrivateVD->getType())) { 1022 emitAggregateInitialization(CGF, N, PrivateAddr, SharedLVal, DRD); 1023 } else if (DRD && (DRD->getInitializer() || !PrivateVD->hasInit())) { 1024 emitInitWithReductionInitializer(CGF, DRD, ClausesData[N].ReductionOp, 1025 PrivateAddr, SharedLVal.getAddress(), 1026 SharedLVal.getType()); 1027 } else if (!DefaultInit(CGF) && PrivateVD->hasInit() && 1028 !CGF.isTrivialInitializer(PrivateVD->getInit())) { 1029 CGF.EmitAnyExprToMem(PrivateVD->getInit(), PrivateAddr, 1030 PrivateVD->getType().getQualifiers(), 1031 /*IsInitializer=*/false); 1032 } 1033 } 1034 1035 bool ReductionCodeGen::needCleanups(unsigned N) { 1036 auto *PrivateVD = 1037 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 1038 QualType PrivateType = PrivateVD->getType(); 1039 QualType::DestructionKind DTorKind = PrivateType.isDestructedType(); 1040 return DTorKind != QualType::DK_none; 1041 } 1042 1043 void ReductionCodeGen::emitCleanups(CodeGenFunction &CGF, unsigned N, 1044 Address PrivateAddr) { 1045 auto *PrivateVD = 1046 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Private)->getDecl()); 1047 QualType PrivateType = PrivateVD->getType(); 1048 QualType::DestructionKind DTorKind = PrivateType.isDestructedType(); 1049 if (needCleanups(N)) { 1050 PrivateAddr = CGF.Builder.CreateElementBitCast( 1051 PrivateAddr, CGF.ConvertTypeForMem(PrivateType)); 1052 CGF.pushDestroy(DTorKind, PrivateAddr, PrivateType); 1053 } 1054 } 1055 1056 static LValue loadToBegin(CodeGenFunction &CGF, QualType BaseTy, QualType ElTy, 1057 LValue BaseLV) { 1058 BaseTy = BaseTy.getNonReferenceType(); 1059 while ((BaseTy->isPointerType() || BaseTy->isReferenceType()) && 1060 !CGF.getContext().hasSameType(BaseTy, ElTy)) { 1061 if (auto *PtrTy = BaseTy->getAs<PointerType>()) 1062 BaseLV = CGF.EmitLoadOfPointerLValue(BaseLV.getAddress(), PtrTy); 1063 else { 1064 LValue RefLVal = CGF.MakeAddrLValue(BaseLV.getAddress(), BaseTy); 1065 BaseLV = CGF.EmitLoadOfReferenceLValue(RefLVal); 1066 } 1067 BaseTy = BaseTy->getPointeeType(); 1068 } 1069 return CGF.MakeAddrLValue( 1070 CGF.Builder.CreateElementBitCast(BaseLV.getAddress(), 1071 CGF.ConvertTypeForMem(ElTy)), 1072 BaseLV.getType(), BaseLV.getBaseInfo(), 1073 CGF.CGM.getTBAAInfoForSubobject(BaseLV, BaseLV.getType())); 1074 } 1075 1076 static Address castToBase(CodeGenFunction &CGF, QualType BaseTy, QualType ElTy, 1077 llvm::Type *BaseLVType, CharUnits BaseLVAlignment, 1078 llvm::Value *Addr) { 1079 Address Tmp = Address::invalid(); 1080 Address TopTmp = Address::invalid(); 1081 Address MostTopTmp = Address::invalid(); 1082 BaseTy = BaseTy.getNonReferenceType(); 1083 while ((BaseTy->isPointerType() || BaseTy->isReferenceType()) && 1084 !CGF.getContext().hasSameType(BaseTy, ElTy)) { 1085 Tmp = CGF.CreateMemTemp(BaseTy); 1086 if (TopTmp.isValid()) 1087 CGF.Builder.CreateStore(Tmp.getPointer(), TopTmp); 1088 else 1089 MostTopTmp = Tmp; 1090 TopTmp = Tmp; 1091 BaseTy = BaseTy->getPointeeType(); 1092 } 1093 llvm::Type *Ty = BaseLVType; 1094 if (Tmp.isValid()) 1095 Ty = Tmp.getElementType(); 1096 Addr = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(Addr, Ty); 1097 if (Tmp.isValid()) { 1098 CGF.Builder.CreateStore(Addr, Tmp); 1099 return MostTopTmp; 1100 } 1101 return Address(Addr, BaseLVAlignment); 1102 } 1103 1104 Address ReductionCodeGen::adjustPrivateAddress(CodeGenFunction &CGF, unsigned N, 1105 Address PrivateAddr) { 1106 const DeclRefExpr *DE; 1107 const VarDecl *OrigVD = nullptr; 1108 if (auto *OASE = dyn_cast<OMPArraySectionExpr>(ClausesData[N].Ref)) { 1109 auto *Base = OASE->getBase()->IgnoreParenImpCasts(); 1110 while (auto *TempOASE = dyn_cast<OMPArraySectionExpr>(Base)) 1111 Base = TempOASE->getBase()->IgnoreParenImpCasts(); 1112 while (auto *TempASE = dyn_cast<ArraySubscriptExpr>(Base)) 1113 Base = TempASE->getBase()->IgnoreParenImpCasts(); 1114 DE = cast<DeclRefExpr>(Base); 1115 OrigVD = cast<VarDecl>(DE->getDecl()); 1116 } else if (auto *ASE = dyn_cast<ArraySubscriptExpr>(ClausesData[N].Ref)) { 1117 auto *Base = ASE->getBase()->IgnoreParenImpCasts(); 1118 while (auto *TempASE = dyn_cast<ArraySubscriptExpr>(Base)) 1119 Base = TempASE->getBase()->IgnoreParenImpCasts(); 1120 DE = cast<DeclRefExpr>(Base); 1121 OrigVD = cast<VarDecl>(DE->getDecl()); 1122 } 1123 if (OrigVD) { 1124 BaseDecls.emplace_back(OrigVD); 1125 auto OriginalBaseLValue = CGF.EmitLValue(DE); 1126 LValue BaseLValue = 1127 loadToBegin(CGF, OrigVD->getType(), SharedAddresses[N].first.getType(), 1128 OriginalBaseLValue); 1129 llvm::Value *Adjustment = CGF.Builder.CreatePtrDiff( 1130 BaseLValue.getPointer(), SharedAddresses[N].first.getPointer()); 1131 llvm::Value *PrivatePointer = 1132 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 1133 PrivateAddr.getPointer(), 1134 SharedAddresses[N].first.getAddress().getType()); 1135 llvm::Value *Ptr = CGF.Builder.CreateGEP(PrivatePointer, Adjustment); 1136 return castToBase(CGF, OrigVD->getType(), 1137 SharedAddresses[N].first.getType(), 1138 OriginalBaseLValue.getAddress().getType(), 1139 OriginalBaseLValue.getAlignment(), Ptr); 1140 } 1141 BaseDecls.emplace_back( 1142 cast<VarDecl>(cast<DeclRefExpr>(ClausesData[N].Ref)->getDecl())); 1143 return PrivateAddr; 1144 } 1145 1146 bool ReductionCodeGen::usesReductionInitializer(unsigned N) const { 1147 auto *DRD = getReductionInit(ClausesData[N].ReductionOp); 1148 return DRD && DRD->getInitializer(); 1149 } 1150 1151 LValue CGOpenMPRegionInfo::getThreadIDVariableLValue(CodeGenFunction &CGF) { 1152 return CGF.EmitLoadOfPointerLValue( 1153 CGF.GetAddrOfLocalVar(getThreadIDVariable()), 1154 getThreadIDVariable()->getType()->castAs<PointerType>()); 1155 } 1156 1157 void CGOpenMPRegionInfo::EmitBody(CodeGenFunction &CGF, const Stmt * /*S*/) { 1158 if (!CGF.HaveInsertPoint()) 1159 return; 1160 // 1.2.2 OpenMP Language Terminology 1161 // Structured block - An executable statement with a single entry at the 1162 // top and a single exit at the bottom. 1163 // The point of exit cannot be a branch out of the structured block. 1164 // longjmp() and throw() must not violate the entry/exit criteria. 1165 CGF.EHStack.pushTerminate(); 1166 CodeGen(CGF); 1167 CGF.EHStack.popTerminate(); 1168 } 1169 1170 LValue CGOpenMPTaskOutlinedRegionInfo::getThreadIDVariableLValue( 1171 CodeGenFunction &CGF) { 1172 return CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(getThreadIDVariable()), 1173 getThreadIDVariable()->getType(), 1174 AlignmentSource::Decl); 1175 } 1176 1177 CGOpenMPRuntime::CGOpenMPRuntime(CodeGenModule &CGM) 1178 : CGM(CGM), OffloadEntriesInfoManager(CGM) { 1179 IdentTy = llvm::StructType::create( 1180 "ident_t", CGM.Int32Ty /* reserved_1 */, CGM.Int32Ty /* flags */, 1181 CGM.Int32Ty /* reserved_2 */, CGM.Int32Ty /* reserved_3 */, 1182 CGM.Int8PtrTy /* psource */); 1183 KmpCriticalNameTy = llvm::ArrayType::get(CGM.Int32Ty, /*NumElements*/ 8); 1184 1185 loadOffloadInfoMetadata(); 1186 } 1187 1188 void CGOpenMPRuntime::clear() { 1189 InternalVars.clear(); 1190 } 1191 1192 static llvm::Function * 1193 emitCombinerOrInitializer(CodeGenModule &CGM, QualType Ty, 1194 const Expr *CombinerInitializer, const VarDecl *In, 1195 const VarDecl *Out, bool IsCombiner) { 1196 // void .omp_combiner.(Ty *in, Ty *out); 1197 auto &C = CGM.getContext(); 1198 QualType PtrTy = C.getPointerType(Ty).withRestrict(); 1199 FunctionArgList Args; 1200 ImplicitParamDecl OmpOutParm(C, /*DC=*/nullptr, Out->getLocation(), 1201 /*Id=*/nullptr, PtrTy, ImplicitParamDecl::Other); 1202 ImplicitParamDecl OmpInParm(C, /*DC=*/nullptr, In->getLocation(), 1203 /*Id=*/nullptr, PtrTy, ImplicitParamDecl::Other); 1204 Args.push_back(&OmpOutParm); 1205 Args.push_back(&OmpInParm); 1206 auto &FnInfo = 1207 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 1208 auto *FnTy = CGM.getTypes().GetFunctionType(FnInfo); 1209 auto *Fn = llvm::Function::Create( 1210 FnTy, llvm::GlobalValue::InternalLinkage, 1211 IsCombiner ? ".omp_combiner." : ".omp_initializer.", &CGM.getModule()); 1212 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, FnInfo); 1213 Fn->removeFnAttr(llvm::Attribute::NoInline); 1214 Fn->removeFnAttr(llvm::Attribute::OptimizeNone); 1215 Fn->addFnAttr(llvm::Attribute::AlwaysInline); 1216 CodeGenFunction CGF(CGM); 1217 // Map "T omp_in;" variable to "*omp_in_parm" value in all expressions. 1218 // Map "T omp_out;" variable to "*omp_out_parm" value in all expressions. 1219 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, In->getLocation(), 1220 Out->getLocation()); 1221 CodeGenFunction::OMPPrivateScope Scope(CGF); 1222 Address AddrIn = CGF.GetAddrOfLocalVar(&OmpInParm); 1223 Scope.addPrivate(In, [&CGF, AddrIn, PtrTy]() -> Address { 1224 return CGF.EmitLoadOfPointerLValue(AddrIn, PtrTy->castAs<PointerType>()) 1225 .getAddress(); 1226 }); 1227 Address AddrOut = CGF.GetAddrOfLocalVar(&OmpOutParm); 1228 Scope.addPrivate(Out, [&CGF, AddrOut, PtrTy]() -> Address { 1229 return CGF.EmitLoadOfPointerLValue(AddrOut, PtrTy->castAs<PointerType>()) 1230 .getAddress(); 1231 }); 1232 (void)Scope.Privatize(); 1233 if (!IsCombiner && Out->hasInit() && 1234 !CGF.isTrivialInitializer(Out->getInit())) { 1235 CGF.EmitAnyExprToMem(Out->getInit(), CGF.GetAddrOfLocalVar(Out), 1236 Out->getType().getQualifiers(), 1237 /*IsInitializer=*/true); 1238 } 1239 if (CombinerInitializer) 1240 CGF.EmitIgnoredExpr(CombinerInitializer); 1241 Scope.ForceCleanup(); 1242 CGF.FinishFunction(); 1243 return Fn; 1244 } 1245 1246 void CGOpenMPRuntime::emitUserDefinedReduction( 1247 CodeGenFunction *CGF, const OMPDeclareReductionDecl *D) { 1248 if (UDRMap.count(D) > 0) 1249 return; 1250 auto &C = CGM.getContext(); 1251 if (!In || !Out) { 1252 In = &C.Idents.get("omp_in"); 1253 Out = &C.Idents.get("omp_out"); 1254 } 1255 llvm::Function *Combiner = emitCombinerOrInitializer( 1256 CGM, D->getType(), D->getCombiner(), cast<VarDecl>(D->lookup(In).front()), 1257 cast<VarDecl>(D->lookup(Out).front()), 1258 /*IsCombiner=*/true); 1259 llvm::Function *Initializer = nullptr; 1260 if (auto *Init = D->getInitializer()) { 1261 if (!Priv || !Orig) { 1262 Priv = &C.Idents.get("omp_priv"); 1263 Orig = &C.Idents.get("omp_orig"); 1264 } 1265 Initializer = emitCombinerOrInitializer( 1266 CGM, D->getType(), 1267 D->getInitializerKind() == OMPDeclareReductionDecl::CallInit ? Init 1268 : nullptr, 1269 cast<VarDecl>(D->lookup(Orig).front()), 1270 cast<VarDecl>(D->lookup(Priv).front()), 1271 /*IsCombiner=*/false); 1272 } 1273 UDRMap.insert(std::make_pair(D, std::make_pair(Combiner, Initializer))); 1274 if (CGF) { 1275 auto &Decls = FunctionUDRMap.FindAndConstruct(CGF->CurFn); 1276 Decls.second.push_back(D); 1277 } 1278 } 1279 1280 std::pair<llvm::Function *, llvm::Function *> 1281 CGOpenMPRuntime::getUserDefinedReduction(const OMPDeclareReductionDecl *D) { 1282 auto I = UDRMap.find(D); 1283 if (I != UDRMap.end()) 1284 return I->second; 1285 emitUserDefinedReduction(/*CGF=*/nullptr, D); 1286 return UDRMap.lookup(D); 1287 } 1288 1289 // Layout information for ident_t. 1290 static CharUnits getIdentAlign(CodeGenModule &CGM) { 1291 return CGM.getPointerAlign(); 1292 } 1293 static CharUnits getIdentSize(CodeGenModule &CGM) { 1294 assert((4 * CGM.getPointerSize()).isMultipleOf(CGM.getPointerAlign())); 1295 return CharUnits::fromQuantity(16) + CGM.getPointerSize(); 1296 } 1297 static CharUnits getOffsetOfIdentField(IdentFieldIndex Field) { 1298 // All the fields except the last are i32, so this works beautifully. 1299 return unsigned(Field) * CharUnits::fromQuantity(4); 1300 } 1301 static Address createIdentFieldGEP(CodeGenFunction &CGF, Address Addr, 1302 IdentFieldIndex Field, 1303 const llvm::Twine &Name = "") { 1304 auto Offset = getOffsetOfIdentField(Field); 1305 return CGF.Builder.CreateStructGEP(Addr, Field, Offset, Name); 1306 } 1307 1308 static llvm::Value *emitParallelOrTeamsOutlinedFunction( 1309 CodeGenModule &CGM, const OMPExecutableDirective &D, const CapturedStmt *CS, 1310 const VarDecl *ThreadIDVar, OpenMPDirectiveKind InnermostKind, 1311 const StringRef OutlinedHelperName, const RegionCodeGenTy &CodeGen) { 1312 assert(ThreadIDVar->getType()->isPointerType() && 1313 "thread id variable must be of type kmp_int32 *"); 1314 CodeGenFunction CGF(CGM, true); 1315 bool HasCancel = false; 1316 if (auto *OPD = dyn_cast<OMPParallelDirective>(&D)) 1317 HasCancel = OPD->hasCancel(); 1318 else if (auto *OPSD = dyn_cast<OMPParallelSectionsDirective>(&D)) 1319 HasCancel = OPSD->hasCancel(); 1320 else if (auto *OPFD = dyn_cast<OMPParallelForDirective>(&D)) 1321 HasCancel = OPFD->hasCancel(); 1322 else if (auto *OPFD = dyn_cast<OMPTargetParallelForDirective>(&D)) 1323 HasCancel = OPFD->hasCancel(); 1324 else if (auto *OPFD = dyn_cast<OMPDistributeParallelForDirective>(&D)) 1325 HasCancel = OPFD->hasCancel(); 1326 else if (auto *OPFD = dyn_cast<OMPTeamsDistributeParallelForDirective>(&D)) 1327 HasCancel = OPFD->hasCancel(); 1328 else if (auto *OPFD = 1329 dyn_cast<OMPTargetTeamsDistributeParallelForDirective>(&D)) 1330 HasCancel = OPFD->hasCancel(); 1331 CGOpenMPOutlinedRegionInfo CGInfo(*CS, ThreadIDVar, CodeGen, InnermostKind, 1332 HasCancel, OutlinedHelperName); 1333 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo); 1334 return CGF.GenerateOpenMPCapturedStmtFunction(*CS); 1335 } 1336 1337 llvm::Value *CGOpenMPRuntime::emitParallelOutlinedFunction( 1338 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 1339 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen) { 1340 const CapturedStmt *CS = D.getCapturedStmt(OMPD_parallel); 1341 return emitParallelOrTeamsOutlinedFunction( 1342 CGM, D, CS, ThreadIDVar, InnermostKind, getOutlinedHelperName(), CodeGen); 1343 } 1344 1345 llvm::Value *CGOpenMPRuntime::emitTeamsOutlinedFunction( 1346 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 1347 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen) { 1348 const CapturedStmt *CS = D.getCapturedStmt(OMPD_teams); 1349 return emitParallelOrTeamsOutlinedFunction( 1350 CGM, D, CS, ThreadIDVar, InnermostKind, getOutlinedHelperName(), CodeGen); 1351 } 1352 1353 llvm::Value *CGOpenMPRuntime::emitTaskOutlinedFunction( 1354 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 1355 const VarDecl *PartIDVar, const VarDecl *TaskTVar, 1356 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen, 1357 bool Tied, unsigned &NumberOfParts) { 1358 auto &&UntiedCodeGen = [this, &D, TaskTVar](CodeGenFunction &CGF, 1359 PrePostActionTy &) { 1360 auto *ThreadID = getThreadID(CGF, D.getLocStart()); 1361 auto *UpLoc = emitUpdateLocation(CGF, D.getLocStart()); 1362 llvm::Value *TaskArgs[] = { 1363 UpLoc, ThreadID, 1364 CGF.EmitLoadOfPointerLValue(CGF.GetAddrOfLocalVar(TaskTVar), 1365 TaskTVar->getType()->castAs<PointerType>()) 1366 .getPointer()}; 1367 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_omp_task), TaskArgs); 1368 }; 1369 CGOpenMPTaskOutlinedRegionInfo::UntiedTaskActionTy Action(Tied, PartIDVar, 1370 UntiedCodeGen); 1371 CodeGen.setAction(Action); 1372 assert(!ThreadIDVar->getType()->isPointerType() && 1373 "thread id variable must be of type kmp_int32 for tasks"); 1374 const OpenMPDirectiveKind Region = 1375 isOpenMPTaskLoopDirective(D.getDirectiveKind()) ? OMPD_taskloop 1376 : OMPD_task; 1377 auto *CS = D.getCapturedStmt(Region); 1378 auto *TD = dyn_cast<OMPTaskDirective>(&D); 1379 CodeGenFunction CGF(CGM, true); 1380 CGOpenMPTaskOutlinedRegionInfo CGInfo(*CS, ThreadIDVar, CodeGen, 1381 InnermostKind, 1382 TD ? TD->hasCancel() : false, Action); 1383 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo); 1384 auto *Res = CGF.GenerateCapturedStmtFunction(*CS); 1385 if (!Tied) 1386 NumberOfParts = Action.getNumberOfParts(); 1387 return Res; 1388 } 1389 1390 Address CGOpenMPRuntime::getOrCreateDefaultLocation(unsigned Flags) { 1391 CharUnits Align = getIdentAlign(CGM); 1392 llvm::Value *Entry = OpenMPDefaultLocMap.lookup(Flags); 1393 if (!Entry) { 1394 if (!DefaultOpenMPPSource) { 1395 // Initialize default location for psource field of ident_t structure of 1396 // all ident_t objects. Format is ";file;function;line;column;;". 1397 // Taken from 1398 // http://llvm.org/svn/llvm-project/openmp/trunk/runtime/src/kmp_str.c 1399 DefaultOpenMPPSource = 1400 CGM.GetAddrOfConstantCString(";unknown;unknown;0;0;;").getPointer(); 1401 DefaultOpenMPPSource = 1402 llvm::ConstantExpr::getBitCast(DefaultOpenMPPSource, CGM.Int8PtrTy); 1403 } 1404 1405 ConstantInitBuilder builder(CGM); 1406 auto fields = builder.beginStruct(IdentTy); 1407 fields.addInt(CGM.Int32Ty, 0); 1408 fields.addInt(CGM.Int32Ty, Flags); 1409 fields.addInt(CGM.Int32Ty, 0); 1410 fields.addInt(CGM.Int32Ty, 0); 1411 fields.add(DefaultOpenMPPSource); 1412 auto DefaultOpenMPLocation = 1413 fields.finishAndCreateGlobal("", Align, /*isConstant*/ true, 1414 llvm::GlobalValue::PrivateLinkage); 1415 DefaultOpenMPLocation->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global); 1416 1417 OpenMPDefaultLocMap[Flags] = Entry = DefaultOpenMPLocation; 1418 } 1419 return Address(Entry, Align); 1420 } 1421 1422 llvm::Value *CGOpenMPRuntime::emitUpdateLocation(CodeGenFunction &CGF, 1423 SourceLocation Loc, 1424 unsigned Flags) { 1425 Flags |= OMP_IDENT_KMPC; 1426 // If no debug info is generated - return global default location. 1427 if (CGM.getCodeGenOpts().getDebugInfo() == codegenoptions::NoDebugInfo || 1428 Loc.isInvalid()) 1429 return getOrCreateDefaultLocation(Flags).getPointer(); 1430 1431 assert(CGF.CurFn && "No function in current CodeGenFunction."); 1432 1433 Address LocValue = Address::invalid(); 1434 auto I = OpenMPLocThreadIDMap.find(CGF.CurFn); 1435 if (I != OpenMPLocThreadIDMap.end()) 1436 LocValue = Address(I->second.DebugLoc, getIdentAlign(CGF.CGM)); 1437 1438 // OpenMPLocThreadIDMap may have null DebugLoc and non-null ThreadID, if 1439 // GetOpenMPThreadID was called before this routine. 1440 if (!LocValue.isValid()) { 1441 // Generate "ident_t .kmpc_loc.addr;" 1442 Address AI = CGF.CreateTempAlloca(IdentTy, getIdentAlign(CGF.CGM), 1443 ".kmpc_loc.addr"); 1444 auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn); 1445 Elem.second.DebugLoc = AI.getPointer(); 1446 LocValue = AI; 1447 1448 CGBuilderTy::InsertPointGuard IPG(CGF.Builder); 1449 CGF.Builder.SetInsertPoint(CGF.AllocaInsertPt); 1450 CGF.Builder.CreateMemCpy(LocValue, getOrCreateDefaultLocation(Flags), 1451 CGM.getSize(getIdentSize(CGF.CGM))); 1452 } 1453 1454 // char **psource = &.kmpc_loc_<flags>.addr.psource; 1455 Address PSource = createIdentFieldGEP(CGF, LocValue, IdentField_PSource); 1456 1457 auto OMPDebugLoc = OpenMPDebugLocMap.lookup(Loc.getRawEncoding()); 1458 if (OMPDebugLoc == nullptr) { 1459 SmallString<128> Buffer2; 1460 llvm::raw_svector_ostream OS2(Buffer2); 1461 // Build debug location 1462 PresumedLoc PLoc = CGF.getContext().getSourceManager().getPresumedLoc(Loc); 1463 OS2 << ";" << PLoc.getFilename() << ";"; 1464 if (const FunctionDecl *FD = 1465 dyn_cast_or_null<FunctionDecl>(CGF.CurFuncDecl)) { 1466 OS2 << FD->getQualifiedNameAsString(); 1467 } 1468 OS2 << ";" << PLoc.getLine() << ";" << PLoc.getColumn() << ";;"; 1469 OMPDebugLoc = CGF.Builder.CreateGlobalStringPtr(OS2.str()); 1470 OpenMPDebugLocMap[Loc.getRawEncoding()] = OMPDebugLoc; 1471 } 1472 // *psource = ";<File>;<Function>;<Line>;<Column>;;"; 1473 CGF.Builder.CreateStore(OMPDebugLoc, PSource); 1474 1475 // Our callers always pass this to a runtime function, so for 1476 // convenience, go ahead and return a naked pointer. 1477 return LocValue.getPointer(); 1478 } 1479 1480 llvm::Value *CGOpenMPRuntime::getThreadID(CodeGenFunction &CGF, 1481 SourceLocation Loc) { 1482 assert(CGF.CurFn && "No function in current CodeGenFunction."); 1483 1484 llvm::Value *ThreadID = nullptr; 1485 // Check whether we've already cached a load of the thread id in this 1486 // function. 1487 auto I = OpenMPLocThreadIDMap.find(CGF.CurFn); 1488 if (I != OpenMPLocThreadIDMap.end()) { 1489 ThreadID = I->second.ThreadID; 1490 if (ThreadID != nullptr) 1491 return ThreadID; 1492 } 1493 // If exceptions are enabled, do not use parameter to avoid possible crash. 1494 if (!CGF.EHStack.requiresLandingPad() || !CGF.getLangOpts().Exceptions || 1495 !CGF.getLangOpts().CXXExceptions || 1496 CGF.Builder.GetInsertBlock() == CGF.AllocaInsertPt->getParent()) { 1497 if (auto *OMPRegionInfo = 1498 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) { 1499 if (OMPRegionInfo->getThreadIDVariable()) { 1500 // Check if this an outlined function with thread id passed as argument. 1501 auto LVal = OMPRegionInfo->getThreadIDVariableLValue(CGF); 1502 ThreadID = CGF.EmitLoadOfScalar(LVal, Loc); 1503 // If value loaded in entry block, cache it and use it everywhere in 1504 // function. 1505 if (CGF.Builder.GetInsertBlock() == CGF.AllocaInsertPt->getParent()) { 1506 auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn); 1507 Elem.second.ThreadID = ThreadID; 1508 } 1509 return ThreadID; 1510 } 1511 } 1512 } 1513 1514 // This is not an outlined function region - need to call __kmpc_int32 1515 // kmpc_global_thread_num(ident_t *loc). 1516 // Generate thread id value and cache this value for use across the 1517 // function. 1518 CGBuilderTy::InsertPointGuard IPG(CGF.Builder); 1519 CGF.Builder.SetInsertPoint(CGF.AllocaInsertPt); 1520 auto *Call = CGF.Builder.CreateCall( 1521 createRuntimeFunction(OMPRTL__kmpc_global_thread_num), 1522 emitUpdateLocation(CGF, Loc)); 1523 Call->setCallingConv(CGF.getRuntimeCC()); 1524 auto &Elem = OpenMPLocThreadIDMap.FindAndConstruct(CGF.CurFn); 1525 Elem.second.ThreadID = Call; 1526 return Call; 1527 } 1528 1529 void CGOpenMPRuntime::functionFinished(CodeGenFunction &CGF) { 1530 assert(CGF.CurFn && "No function in current CodeGenFunction."); 1531 if (OpenMPLocThreadIDMap.count(CGF.CurFn)) 1532 OpenMPLocThreadIDMap.erase(CGF.CurFn); 1533 if (FunctionUDRMap.count(CGF.CurFn) > 0) { 1534 for(auto *D : FunctionUDRMap[CGF.CurFn]) { 1535 UDRMap.erase(D); 1536 } 1537 FunctionUDRMap.erase(CGF.CurFn); 1538 } 1539 } 1540 1541 llvm::Type *CGOpenMPRuntime::getIdentTyPointerTy() { 1542 if (!IdentTy) { 1543 } 1544 return llvm::PointerType::getUnqual(IdentTy); 1545 } 1546 1547 llvm::Type *CGOpenMPRuntime::getKmpc_MicroPointerTy() { 1548 if (!Kmpc_MicroTy) { 1549 // Build void (*kmpc_micro)(kmp_int32 *global_tid, kmp_int32 *bound_tid,...) 1550 llvm::Type *MicroParams[] = {llvm::PointerType::getUnqual(CGM.Int32Ty), 1551 llvm::PointerType::getUnqual(CGM.Int32Ty)}; 1552 Kmpc_MicroTy = llvm::FunctionType::get(CGM.VoidTy, MicroParams, true); 1553 } 1554 return llvm::PointerType::getUnqual(Kmpc_MicroTy); 1555 } 1556 1557 llvm::Constant * 1558 CGOpenMPRuntime::createRuntimeFunction(unsigned Function) { 1559 llvm::Constant *RTLFn = nullptr; 1560 switch (static_cast<OpenMPRTLFunction>(Function)) { 1561 case OMPRTL__kmpc_fork_call: { 1562 // Build void __kmpc_fork_call(ident_t *loc, kmp_int32 argc, kmpc_micro 1563 // microtask, ...); 1564 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1565 getKmpc_MicroPointerTy()}; 1566 llvm::FunctionType *FnTy = 1567 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ true); 1568 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_fork_call"); 1569 break; 1570 } 1571 case OMPRTL__kmpc_global_thread_num: { 1572 // Build kmp_int32 __kmpc_global_thread_num(ident_t *loc); 1573 llvm::Type *TypeParams[] = {getIdentTyPointerTy()}; 1574 llvm::FunctionType *FnTy = 1575 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 1576 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_global_thread_num"); 1577 break; 1578 } 1579 case OMPRTL__kmpc_threadprivate_cached: { 1580 // Build void *__kmpc_threadprivate_cached(ident_t *loc, 1581 // kmp_int32 global_tid, void *data, size_t size, void ***cache); 1582 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1583 CGM.VoidPtrTy, CGM.SizeTy, 1584 CGM.VoidPtrTy->getPointerTo()->getPointerTo()}; 1585 llvm::FunctionType *FnTy = 1586 llvm::FunctionType::get(CGM.VoidPtrTy, TypeParams, /*isVarArg*/ false); 1587 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_threadprivate_cached"); 1588 break; 1589 } 1590 case OMPRTL__kmpc_critical: { 1591 // Build void __kmpc_critical(ident_t *loc, kmp_int32 global_tid, 1592 // kmp_critical_name *crit); 1593 llvm::Type *TypeParams[] = { 1594 getIdentTyPointerTy(), CGM.Int32Ty, 1595 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1596 llvm::FunctionType *FnTy = 1597 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1598 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_critical"); 1599 break; 1600 } 1601 case OMPRTL__kmpc_critical_with_hint: { 1602 // Build void __kmpc_critical_with_hint(ident_t *loc, kmp_int32 global_tid, 1603 // kmp_critical_name *crit, uintptr_t hint); 1604 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1605 llvm::PointerType::getUnqual(KmpCriticalNameTy), 1606 CGM.IntPtrTy}; 1607 llvm::FunctionType *FnTy = 1608 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1609 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_critical_with_hint"); 1610 break; 1611 } 1612 case OMPRTL__kmpc_threadprivate_register: { 1613 // Build void __kmpc_threadprivate_register(ident_t *, void *data, 1614 // kmpc_ctor ctor, kmpc_cctor cctor, kmpc_dtor dtor); 1615 // typedef void *(*kmpc_ctor)(void *); 1616 auto KmpcCtorTy = 1617 llvm::FunctionType::get(CGM.VoidPtrTy, CGM.VoidPtrTy, 1618 /*isVarArg*/ false)->getPointerTo(); 1619 // typedef void *(*kmpc_cctor)(void *, void *); 1620 llvm::Type *KmpcCopyCtorTyArgs[] = {CGM.VoidPtrTy, CGM.VoidPtrTy}; 1621 auto KmpcCopyCtorTy = 1622 llvm::FunctionType::get(CGM.VoidPtrTy, KmpcCopyCtorTyArgs, 1623 /*isVarArg*/ false)->getPointerTo(); 1624 // typedef void (*kmpc_dtor)(void *); 1625 auto KmpcDtorTy = 1626 llvm::FunctionType::get(CGM.VoidTy, CGM.VoidPtrTy, /*isVarArg*/ false) 1627 ->getPointerTo(); 1628 llvm::Type *FnTyArgs[] = {getIdentTyPointerTy(), CGM.VoidPtrTy, KmpcCtorTy, 1629 KmpcCopyCtorTy, KmpcDtorTy}; 1630 auto FnTy = llvm::FunctionType::get(CGM.VoidTy, FnTyArgs, 1631 /*isVarArg*/ false); 1632 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_threadprivate_register"); 1633 break; 1634 } 1635 case OMPRTL__kmpc_end_critical: { 1636 // Build void __kmpc_end_critical(ident_t *loc, kmp_int32 global_tid, 1637 // kmp_critical_name *crit); 1638 llvm::Type *TypeParams[] = { 1639 getIdentTyPointerTy(), CGM.Int32Ty, 1640 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1641 llvm::FunctionType *FnTy = 1642 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1643 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_end_critical"); 1644 break; 1645 } 1646 case OMPRTL__kmpc_cancel_barrier: { 1647 // Build kmp_int32 __kmpc_cancel_barrier(ident_t *loc, kmp_int32 1648 // global_tid); 1649 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1650 llvm::FunctionType *FnTy = 1651 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 1652 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name*/ "__kmpc_cancel_barrier"); 1653 break; 1654 } 1655 case OMPRTL__kmpc_barrier: { 1656 // Build void __kmpc_barrier(ident_t *loc, kmp_int32 global_tid); 1657 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1658 llvm::FunctionType *FnTy = 1659 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1660 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name*/ "__kmpc_barrier"); 1661 break; 1662 } 1663 case OMPRTL__kmpc_for_static_fini: { 1664 // Build void __kmpc_for_static_fini(ident_t *loc, kmp_int32 global_tid); 1665 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1666 llvm::FunctionType *FnTy = 1667 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1668 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_for_static_fini"); 1669 break; 1670 } 1671 case OMPRTL__kmpc_push_num_threads: { 1672 // Build void __kmpc_push_num_threads(ident_t *loc, kmp_int32 global_tid, 1673 // kmp_int32 num_threads) 1674 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1675 CGM.Int32Ty}; 1676 llvm::FunctionType *FnTy = 1677 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1678 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_push_num_threads"); 1679 break; 1680 } 1681 case OMPRTL__kmpc_serialized_parallel: { 1682 // Build void __kmpc_serialized_parallel(ident_t *loc, kmp_int32 1683 // global_tid); 1684 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1685 llvm::FunctionType *FnTy = 1686 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1687 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_serialized_parallel"); 1688 break; 1689 } 1690 case OMPRTL__kmpc_end_serialized_parallel: { 1691 // Build void __kmpc_end_serialized_parallel(ident_t *loc, kmp_int32 1692 // global_tid); 1693 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1694 llvm::FunctionType *FnTy = 1695 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1696 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_end_serialized_parallel"); 1697 break; 1698 } 1699 case OMPRTL__kmpc_flush: { 1700 // Build void __kmpc_flush(ident_t *loc); 1701 llvm::Type *TypeParams[] = {getIdentTyPointerTy()}; 1702 llvm::FunctionType *FnTy = 1703 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1704 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_flush"); 1705 break; 1706 } 1707 case OMPRTL__kmpc_master: { 1708 // Build kmp_int32 __kmpc_master(ident_t *loc, kmp_int32 global_tid); 1709 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1710 llvm::FunctionType *FnTy = 1711 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1712 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_master"); 1713 break; 1714 } 1715 case OMPRTL__kmpc_end_master: { 1716 // Build void __kmpc_end_master(ident_t *loc, kmp_int32 global_tid); 1717 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1718 llvm::FunctionType *FnTy = 1719 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1720 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_end_master"); 1721 break; 1722 } 1723 case OMPRTL__kmpc_omp_taskyield: { 1724 // Build kmp_int32 __kmpc_omp_taskyield(ident_t *, kmp_int32 global_tid, 1725 // int end_part); 1726 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.IntTy}; 1727 llvm::FunctionType *FnTy = 1728 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1729 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_taskyield"); 1730 break; 1731 } 1732 case OMPRTL__kmpc_single: { 1733 // Build kmp_int32 __kmpc_single(ident_t *loc, kmp_int32 global_tid); 1734 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1735 llvm::FunctionType *FnTy = 1736 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1737 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_single"); 1738 break; 1739 } 1740 case OMPRTL__kmpc_end_single: { 1741 // Build void __kmpc_end_single(ident_t *loc, kmp_int32 global_tid); 1742 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1743 llvm::FunctionType *FnTy = 1744 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1745 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_end_single"); 1746 break; 1747 } 1748 case OMPRTL__kmpc_omp_task_alloc: { 1749 // Build kmp_task_t *__kmpc_omp_task_alloc(ident_t *, kmp_int32 gtid, 1750 // kmp_int32 flags, size_t sizeof_kmp_task_t, size_t sizeof_shareds, 1751 // kmp_routine_entry_t *task_entry); 1752 assert(KmpRoutineEntryPtrTy != nullptr && 1753 "Type kmp_routine_entry_t must be created."); 1754 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.Int32Ty, 1755 CGM.SizeTy, CGM.SizeTy, KmpRoutineEntryPtrTy}; 1756 // Return void * and then cast to particular kmp_task_t type. 1757 llvm::FunctionType *FnTy = 1758 llvm::FunctionType::get(CGM.VoidPtrTy, TypeParams, /*isVarArg=*/false); 1759 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_task_alloc"); 1760 break; 1761 } 1762 case OMPRTL__kmpc_omp_task: { 1763 // Build kmp_int32 __kmpc_omp_task(ident_t *, kmp_int32 gtid, kmp_task_t 1764 // *new_task); 1765 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1766 CGM.VoidPtrTy}; 1767 llvm::FunctionType *FnTy = 1768 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1769 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_task"); 1770 break; 1771 } 1772 case OMPRTL__kmpc_copyprivate: { 1773 // Build void __kmpc_copyprivate(ident_t *loc, kmp_int32 global_tid, 1774 // size_t cpy_size, void *cpy_data, void(*cpy_func)(void *, void *), 1775 // kmp_int32 didit); 1776 llvm::Type *CpyTypeParams[] = {CGM.VoidPtrTy, CGM.VoidPtrTy}; 1777 auto *CpyFnTy = 1778 llvm::FunctionType::get(CGM.VoidTy, CpyTypeParams, /*isVarArg=*/false); 1779 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.SizeTy, 1780 CGM.VoidPtrTy, CpyFnTy->getPointerTo(), 1781 CGM.Int32Ty}; 1782 llvm::FunctionType *FnTy = 1783 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1784 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_copyprivate"); 1785 break; 1786 } 1787 case OMPRTL__kmpc_reduce: { 1788 // Build kmp_int32 __kmpc_reduce(ident_t *loc, kmp_int32 global_tid, 1789 // kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void 1790 // (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck); 1791 llvm::Type *ReduceTypeParams[] = {CGM.VoidPtrTy, CGM.VoidPtrTy}; 1792 auto *ReduceFnTy = llvm::FunctionType::get(CGM.VoidTy, ReduceTypeParams, 1793 /*isVarArg=*/false); 1794 llvm::Type *TypeParams[] = { 1795 getIdentTyPointerTy(), CGM.Int32Ty, CGM.Int32Ty, CGM.SizeTy, 1796 CGM.VoidPtrTy, ReduceFnTy->getPointerTo(), 1797 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1798 llvm::FunctionType *FnTy = 1799 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1800 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_reduce"); 1801 break; 1802 } 1803 case OMPRTL__kmpc_reduce_nowait: { 1804 // Build kmp_int32 __kmpc_reduce_nowait(ident_t *loc, kmp_int32 1805 // global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, 1806 // void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name 1807 // *lck); 1808 llvm::Type *ReduceTypeParams[] = {CGM.VoidPtrTy, CGM.VoidPtrTy}; 1809 auto *ReduceFnTy = llvm::FunctionType::get(CGM.VoidTy, ReduceTypeParams, 1810 /*isVarArg=*/false); 1811 llvm::Type *TypeParams[] = { 1812 getIdentTyPointerTy(), CGM.Int32Ty, CGM.Int32Ty, CGM.SizeTy, 1813 CGM.VoidPtrTy, ReduceFnTy->getPointerTo(), 1814 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1815 llvm::FunctionType *FnTy = 1816 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1817 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_reduce_nowait"); 1818 break; 1819 } 1820 case OMPRTL__kmpc_end_reduce: { 1821 // Build void __kmpc_end_reduce(ident_t *loc, kmp_int32 global_tid, 1822 // kmp_critical_name *lck); 1823 llvm::Type *TypeParams[] = { 1824 getIdentTyPointerTy(), CGM.Int32Ty, 1825 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1826 llvm::FunctionType *FnTy = 1827 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1828 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_end_reduce"); 1829 break; 1830 } 1831 case OMPRTL__kmpc_end_reduce_nowait: { 1832 // Build __kmpc_end_reduce_nowait(ident_t *loc, kmp_int32 global_tid, 1833 // kmp_critical_name *lck); 1834 llvm::Type *TypeParams[] = { 1835 getIdentTyPointerTy(), CGM.Int32Ty, 1836 llvm::PointerType::getUnqual(KmpCriticalNameTy)}; 1837 llvm::FunctionType *FnTy = 1838 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1839 RTLFn = 1840 CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_end_reduce_nowait"); 1841 break; 1842 } 1843 case OMPRTL__kmpc_omp_task_begin_if0: { 1844 // Build void __kmpc_omp_task(ident_t *, kmp_int32 gtid, kmp_task_t 1845 // *new_task); 1846 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1847 CGM.VoidPtrTy}; 1848 llvm::FunctionType *FnTy = 1849 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1850 RTLFn = 1851 CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_task_begin_if0"); 1852 break; 1853 } 1854 case OMPRTL__kmpc_omp_task_complete_if0: { 1855 // Build void __kmpc_omp_task(ident_t *, kmp_int32 gtid, kmp_task_t 1856 // *new_task); 1857 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1858 CGM.VoidPtrTy}; 1859 llvm::FunctionType *FnTy = 1860 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1861 RTLFn = CGM.CreateRuntimeFunction(FnTy, 1862 /*Name=*/"__kmpc_omp_task_complete_if0"); 1863 break; 1864 } 1865 case OMPRTL__kmpc_ordered: { 1866 // Build void __kmpc_ordered(ident_t *loc, kmp_int32 global_tid); 1867 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1868 llvm::FunctionType *FnTy = 1869 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1870 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_ordered"); 1871 break; 1872 } 1873 case OMPRTL__kmpc_end_ordered: { 1874 // Build void __kmpc_end_ordered(ident_t *loc, kmp_int32 global_tid); 1875 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1876 llvm::FunctionType *FnTy = 1877 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1878 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_end_ordered"); 1879 break; 1880 } 1881 case OMPRTL__kmpc_omp_taskwait: { 1882 // Build kmp_int32 __kmpc_omp_taskwait(ident_t *loc, kmp_int32 global_tid); 1883 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1884 llvm::FunctionType *FnTy = 1885 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1886 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_omp_taskwait"); 1887 break; 1888 } 1889 case OMPRTL__kmpc_taskgroup: { 1890 // Build void __kmpc_taskgroup(ident_t *loc, kmp_int32 global_tid); 1891 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1892 llvm::FunctionType *FnTy = 1893 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1894 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_taskgroup"); 1895 break; 1896 } 1897 case OMPRTL__kmpc_end_taskgroup: { 1898 // Build void __kmpc_end_taskgroup(ident_t *loc, kmp_int32 global_tid); 1899 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 1900 llvm::FunctionType *FnTy = 1901 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1902 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_end_taskgroup"); 1903 break; 1904 } 1905 case OMPRTL__kmpc_push_proc_bind: { 1906 // Build void __kmpc_push_proc_bind(ident_t *loc, kmp_int32 global_tid, 1907 // int proc_bind) 1908 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.IntTy}; 1909 llvm::FunctionType *FnTy = 1910 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 1911 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_push_proc_bind"); 1912 break; 1913 } 1914 case OMPRTL__kmpc_omp_task_with_deps: { 1915 // Build kmp_int32 __kmpc_omp_task_with_deps(ident_t *, kmp_int32 gtid, 1916 // kmp_task_t *new_task, kmp_int32 ndeps, kmp_depend_info_t *dep_list, 1917 // kmp_int32 ndeps_noalias, kmp_depend_info_t *noalias_dep_list); 1918 llvm::Type *TypeParams[] = { 1919 getIdentTyPointerTy(), CGM.Int32Ty, CGM.VoidPtrTy, CGM.Int32Ty, 1920 CGM.VoidPtrTy, CGM.Int32Ty, CGM.VoidPtrTy}; 1921 llvm::FunctionType *FnTy = 1922 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg=*/false); 1923 RTLFn = 1924 CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_task_with_deps"); 1925 break; 1926 } 1927 case OMPRTL__kmpc_omp_wait_deps: { 1928 // Build void __kmpc_omp_wait_deps(ident_t *, kmp_int32 gtid, 1929 // kmp_int32 ndeps, kmp_depend_info_t *dep_list, kmp_int32 ndeps_noalias, 1930 // kmp_depend_info_t *noalias_dep_list); 1931 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1932 CGM.Int32Ty, CGM.VoidPtrTy, 1933 CGM.Int32Ty, CGM.VoidPtrTy}; 1934 llvm::FunctionType *FnTy = 1935 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1936 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_omp_wait_deps"); 1937 break; 1938 } 1939 case OMPRTL__kmpc_cancellationpoint: { 1940 // Build kmp_int32 __kmpc_cancellationpoint(ident_t *loc, kmp_int32 1941 // global_tid, kmp_int32 cncl_kind) 1942 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.IntTy}; 1943 llvm::FunctionType *FnTy = 1944 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 1945 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_cancellationpoint"); 1946 break; 1947 } 1948 case OMPRTL__kmpc_cancel: { 1949 // Build kmp_int32 __kmpc_cancel(ident_t *loc, kmp_int32 global_tid, 1950 // kmp_int32 cncl_kind) 1951 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.IntTy}; 1952 llvm::FunctionType *FnTy = 1953 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 1954 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_cancel"); 1955 break; 1956 } 1957 case OMPRTL__kmpc_push_num_teams: { 1958 // Build void kmpc_push_num_teams (ident_t loc, kmp_int32 global_tid, 1959 // kmp_int32 num_teams, kmp_int32 num_threads) 1960 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, CGM.Int32Ty, 1961 CGM.Int32Ty}; 1962 llvm::FunctionType *FnTy = 1963 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 1964 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_push_num_teams"); 1965 break; 1966 } 1967 case OMPRTL__kmpc_fork_teams: { 1968 // Build void __kmpc_fork_teams(ident_t *loc, kmp_int32 argc, kmpc_micro 1969 // microtask, ...); 1970 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 1971 getKmpc_MicroPointerTy()}; 1972 llvm::FunctionType *FnTy = 1973 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ true); 1974 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__kmpc_fork_teams"); 1975 break; 1976 } 1977 case OMPRTL__kmpc_taskloop: { 1978 // Build void __kmpc_taskloop(ident_t *loc, int gtid, kmp_task_t *task, int 1979 // if_val, kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup, int 1980 // sched, kmp_uint64 grainsize, void *task_dup); 1981 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), 1982 CGM.IntTy, 1983 CGM.VoidPtrTy, 1984 CGM.IntTy, 1985 CGM.Int64Ty->getPointerTo(), 1986 CGM.Int64Ty->getPointerTo(), 1987 CGM.Int64Ty, 1988 CGM.IntTy, 1989 CGM.IntTy, 1990 CGM.Int64Ty, 1991 CGM.VoidPtrTy}; 1992 llvm::FunctionType *FnTy = 1993 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 1994 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_taskloop"); 1995 break; 1996 } 1997 case OMPRTL__kmpc_doacross_init: { 1998 // Build void __kmpc_doacross_init(ident_t *loc, kmp_int32 gtid, kmp_int32 1999 // num_dims, struct kmp_dim *dims); 2000 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), 2001 CGM.Int32Ty, 2002 CGM.Int32Ty, 2003 CGM.VoidPtrTy}; 2004 llvm::FunctionType *FnTy = 2005 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2006 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_doacross_init"); 2007 break; 2008 } 2009 case OMPRTL__kmpc_doacross_fini: { 2010 // Build void __kmpc_doacross_fini(ident_t *loc, kmp_int32 gtid); 2011 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty}; 2012 llvm::FunctionType *FnTy = 2013 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2014 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_doacross_fini"); 2015 break; 2016 } 2017 case OMPRTL__kmpc_doacross_post: { 2018 // Build void __kmpc_doacross_post(ident_t *loc, kmp_int32 gtid, kmp_int64 2019 // *vec); 2020 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 2021 CGM.Int64Ty->getPointerTo()}; 2022 llvm::FunctionType *FnTy = 2023 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2024 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_doacross_post"); 2025 break; 2026 } 2027 case OMPRTL__kmpc_doacross_wait: { 2028 // Build void __kmpc_doacross_wait(ident_t *loc, kmp_int32 gtid, kmp_int64 2029 // *vec); 2030 llvm::Type *TypeParams[] = {getIdentTyPointerTy(), CGM.Int32Ty, 2031 CGM.Int64Ty->getPointerTo()}; 2032 llvm::FunctionType *FnTy = 2033 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2034 RTLFn = CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_doacross_wait"); 2035 break; 2036 } 2037 case OMPRTL__kmpc_task_reduction_init: { 2038 // Build void *__kmpc_task_reduction_init(int gtid, int num_data, void 2039 // *data); 2040 llvm::Type *TypeParams[] = {CGM.IntTy, CGM.IntTy, CGM.VoidPtrTy}; 2041 llvm::FunctionType *FnTy = 2042 llvm::FunctionType::get(CGM.VoidPtrTy, TypeParams, /*isVarArg=*/false); 2043 RTLFn = 2044 CGM.CreateRuntimeFunction(FnTy, /*Name=*/"__kmpc_task_reduction_init"); 2045 break; 2046 } 2047 case OMPRTL__kmpc_task_reduction_get_th_data: { 2048 // Build void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void 2049 // *d); 2050 llvm::Type *TypeParams[] = {CGM.IntTy, CGM.VoidPtrTy, CGM.VoidPtrTy}; 2051 llvm::FunctionType *FnTy = 2052 llvm::FunctionType::get(CGM.VoidPtrTy, TypeParams, /*isVarArg=*/false); 2053 RTLFn = CGM.CreateRuntimeFunction( 2054 FnTy, /*Name=*/"__kmpc_task_reduction_get_th_data"); 2055 break; 2056 } 2057 case OMPRTL__tgt_target: { 2058 // Build int32_t __tgt_target(int64_t device_id, void *host_ptr, int32_t 2059 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 2060 // *arg_types); 2061 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2062 CGM.VoidPtrTy, 2063 CGM.Int32Ty, 2064 CGM.VoidPtrPtrTy, 2065 CGM.VoidPtrPtrTy, 2066 CGM.SizeTy->getPointerTo(), 2067 CGM.Int64Ty->getPointerTo()}; 2068 llvm::FunctionType *FnTy = 2069 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2070 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target"); 2071 break; 2072 } 2073 case OMPRTL__tgt_target_nowait: { 2074 // Build int32_t __tgt_target_nowait(int64_t device_id, void *host_ptr, 2075 // int32_t arg_num, void** args_base, void **args, size_t *arg_sizes, 2076 // int64_t *arg_types); 2077 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2078 CGM.VoidPtrTy, 2079 CGM.Int32Ty, 2080 CGM.VoidPtrPtrTy, 2081 CGM.VoidPtrPtrTy, 2082 CGM.SizeTy->getPointerTo(), 2083 CGM.Int64Ty->getPointerTo()}; 2084 llvm::FunctionType *FnTy = 2085 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2086 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_nowait"); 2087 break; 2088 } 2089 case OMPRTL__tgt_target_teams: { 2090 // Build int32_t __tgt_target_teams(int64_t device_id, void *host_ptr, 2091 // int32_t arg_num, void** args_base, void **args, size_t *arg_sizes, 2092 // int64_t *arg_types, int32_t num_teams, int32_t thread_limit); 2093 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2094 CGM.VoidPtrTy, 2095 CGM.Int32Ty, 2096 CGM.VoidPtrPtrTy, 2097 CGM.VoidPtrPtrTy, 2098 CGM.SizeTy->getPointerTo(), 2099 CGM.Int64Ty->getPointerTo(), 2100 CGM.Int32Ty, 2101 CGM.Int32Ty}; 2102 llvm::FunctionType *FnTy = 2103 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2104 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_teams"); 2105 break; 2106 } 2107 case OMPRTL__tgt_target_teams_nowait: { 2108 // Build int32_t __tgt_target_teams_nowait(int64_t device_id, void 2109 // *host_ptr, int32_t arg_num, void** args_base, void **args, size_t 2110 // *arg_sizes, int64_t *arg_types, int32_t num_teams, int32_t thread_limit); 2111 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2112 CGM.VoidPtrTy, 2113 CGM.Int32Ty, 2114 CGM.VoidPtrPtrTy, 2115 CGM.VoidPtrPtrTy, 2116 CGM.SizeTy->getPointerTo(), 2117 CGM.Int64Ty->getPointerTo(), 2118 CGM.Int32Ty, 2119 CGM.Int32Ty}; 2120 llvm::FunctionType *FnTy = 2121 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2122 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_teams_nowait"); 2123 break; 2124 } 2125 case OMPRTL__tgt_register_lib: { 2126 // Build void __tgt_register_lib(__tgt_bin_desc *desc); 2127 QualType ParamTy = 2128 CGM.getContext().getPointerType(getTgtBinaryDescriptorQTy()); 2129 llvm::Type *TypeParams[] = {CGM.getTypes().ConvertTypeForMem(ParamTy)}; 2130 llvm::FunctionType *FnTy = 2131 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2132 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_register_lib"); 2133 break; 2134 } 2135 case OMPRTL__tgt_unregister_lib: { 2136 // Build void __tgt_unregister_lib(__tgt_bin_desc *desc); 2137 QualType ParamTy = 2138 CGM.getContext().getPointerType(getTgtBinaryDescriptorQTy()); 2139 llvm::Type *TypeParams[] = {CGM.getTypes().ConvertTypeForMem(ParamTy)}; 2140 llvm::FunctionType *FnTy = 2141 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2142 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_unregister_lib"); 2143 break; 2144 } 2145 case OMPRTL__tgt_target_data_begin: { 2146 // Build void __tgt_target_data_begin(int64_t device_id, int32_t arg_num, 2147 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 2148 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2149 CGM.Int32Ty, 2150 CGM.VoidPtrPtrTy, 2151 CGM.VoidPtrPtrTy, 2152 CGM.SizeTy->getPointerTo(), 2153 CGM.Int64Ty->getPointerTo()}; 2154 llvm::FunctionType *FnTy = 2155 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 2156 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_begin"); 2157 break; 2158 } 2159 case OMPRTL__tgt_target_data_begin_nowait: { 2160 // Build void __tgt_target_data_begin_nowait(int64_t device_id, int32_t 2161 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 2162 // *arg_types); 2163 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2164 CGM.Int32Ty, 2165 CGM.VoidPtrPtrTy, 2166 CGM.VoidPtrPtrTy, 2167 CGM.SizeTy->getPointerTo(), 2168 CGM.Int64Ty->getPointerTo()}; 2169 auto *FnTy = 2170 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2171 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_begin_nowait"); 2172 break; 2173 } 2174 case OMPRTL__tgt_target_data_end: { 2175 // Build void __tgt_target_data_end(int64_t device_id, int32_t arg_num, 2176 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 2177 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2178 CGM.Int32Ty, 2179 CGM.VoidPtrPtrTy, 2180 CGM.VoidPtrPtrTy, 2181 CGM.SizeTy->getPointerTo(), 2182 CGM.Int64Ty->getPointerTo()}; 2183 llvm::FunctionType *FnTy = 2184 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 2185 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_end"); 2186 break; 2187 } 2188 case OMPRTL__tgt_target_data_end_nowait: { 2189 // Build void __tgt_target_data_end_nowait(int64_t device_id, int32_t 2190 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 2191 // *arg_types); 2192 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2193 CGM.Int32Ty, 2194 CGM.VoidPtrPtrTy, 2195 CGM.VoidPtrPtrTy, 2196 CGM.SizeTy->getPointerTo(), 2197 CGM.Int64Ty->getPointerTo()}; 2198 auto *FnTy = 2199 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2200 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_end_nowait"); 2201 break; 2202 } 2203 case OMPRTL__tgt_target_data_update: { 2204 // Build void __tgt_target_data_update(int64_t device_id, int32_t arg_num, 2205 // void** args_base, void **args, size_t *arg_sizes, int64_t *arg_types); 2206 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2207 CGM.Int32Ty, 2208 CGM.VoidPtrPtrTy, 2209 CGM.VoidPtrPtrTy, 2210 CGM.SizeTy->getPointerTo(), 2211 CGM.Int64Ty->getPointerTo()}; 2212 llvm::FunctionType *FnTy = 2213 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 2214 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_update"); 2215 break; 2216 } 2217 case OMPRTL__tgt_target_data_update_nowait: { 2218 // Build void __tgt_target_data_update_nowait(int64_t device_id, int32_t 2219 // arg_num, void** args_base, void **args, size_t *arg_sizes, int64_t 2220 // *arg_types); 2221 llvm::Type *TypeParams[] = {CGM.Int64Ty, 2222 CGM.Int32Ty, 2223 CGM.VoidPtrPtrTy, 2224 CGM.VoidPtrPtrTy, 2225 CGM.SizeTy->getPointerTo(), 2226 CGM.Int64Ty->getPointerTo()}; 2227 auto *FnTy = 2228 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2229 RTLFn = CGM.CreateRuntimeFunction(FnTy, "__tgt_target_data_update_nowait"); 2230 break; 2231 } 2232 } 2233 assert(RTLFn && "Unable to find OpenMP runtime function"); 2234 return RTLFn; 2235 } 2236 2237 llvm::Constant *CGOpenMPRuntime::createForStaticInitFunction(unsigned IVSize, 2238 bool IVSigned) { 2239 assert((IVSize == 32 || IVSize == 64) && 2240 "IV size is not compatible with the omp runtime"); 2241 auto Name = IVSize == 32 ? (IVSigned ? "__kmpc_for_static_init_4" 2242 : "__kmpc_for_static_init_4u") 2243 : (IVSigned ? "__kmpc_for_static_init_8" 2244 : "__kmpc_for_static_init_8u"); 2245 auto ITy = IVSize == 32 ? CGM.Int32Ty : CGM.Int64Ty; 2246 auto PtrTy = llvm::PointerType::getUnqual(ITy); 2247 llvm::Type *TypeParams[] = { 2248 getIdentTyPointerTy(), // loc 2249 CGM.Int32Ty, // tid 2250 CGM.Int32Ty, // schedtype 2251 llvm::PointerType::getUnqual(CGM.Int32Ty), // p_lastiter 2252 PtrTy, // p_lower 2253 PtrTy, // p_upper 2254 PtrTy, // p_stride 2255 ITy, // incr 2256 ITy // chunk 2257 }; 2258 llvm::FunctionType *FnTy = 2259 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 2260 return CGM.CreateRuntimeFunction(FnTy, Name); 2261 } 2262 2263 llvm::Constant *CGOpenMPRuntime::createDispatchInitFunction(unsigned IVSize, 2264 bool IVSigned) { 2265 assert((IVSize == 32 || IVSize == 64) && 2266 "IV size is not compatible with the omp runtime"); 2267 auto Name = 2268 IVSize == 32 2269 ? (IVSigned ? "__kmpc_dispatch_init_4" : "__kmpc_dispatch_init_4u") 2270 : (IVSigned ? "__kmpc_dispatch_init_8" : "__kmpc_dispatch_init_8u"); 2271 auto ITy = IVSize == 32 ? CGM.Int32Ty : CGM.Int64Ty; 2272 llvm::Type *TypeParams[] = { getIdentTyPointerTy(), // loc 2273 CGM.Int32Ty, // tid 2274 CGM.Int32Ty, // schedtype 2275 ITy, // lower 2276 ITy, // upper 2277 ITy, // stride 2278 ITy // chunk 2279 }; 2280 llvm::FunctionType *FnTy = 2281 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg*/ false); 2282 return CGM.CreateRuntimeFunction(FnTy, Name); 2283 } 2284 2285 llvm::Constant *CGOpenMPRuntime::createDispatchFiniFunction(unsigned IVSize, 2286 bool IVSigned) { 2287 assert((IVSize == 32 || IVSize == 64) && 2288 "IV size is not compatible with the omp runtime"); 2289 auto Name = 2290 IVSize == 32 2291 ? (IVSigned ? "__kmpc_dispatch_fini_4" : "__kmpc_dispatch_fini_4u") 2292 : (IVSigned ? "__kmpc_dispatch_fini_8" : "__kmpc_dispatch_fini_8u"); 2293 llvm::Type *TypeParams[] = { 2294 getIdentTyPointerTy(), // loc 2295 CGM.Int32Ty, // tid 2296 }; 2297 llvm::FunctionType *FnTy = 2298 llvm::FunctionType::get(CGM.VoidTy, TypeParams, /*isVarArg=*/false); 2299 return CGM.CreateRuntimeFunction(FnTy, Name); 2300 } 2301 2302 llvm::Constant *CGOpenMPRuntime::createDispatchNextFunction(unsigned IVSize, 2303 bool IVSigned) { 2304 assert((IVSize == 32 || IVSize == 64) && 2305 "IV size is not compatible with the omp runtime"); 2306 auto Name = 2307 IVSize == 32 2308 ? (IVSigned ? "__kmpc_dispatch_next_4" : "__kmpc_dispatch_next_4u") 2309 : (IVSigned ? "__kmpc_dispatch_next_8" : "__kmpc_dispatch_next_8u"); 2310 auto ITy = IVSize == 32 ? CGM.Int32Ty : CGM.Int64Ty; 2311 auto PtrTy = llvm::PointerType::getUnqual(ITy); 2312 llvm::Type *TypeParams[] = { 2313 getIdentTyPointerTy(), // loc 2314 CGM.Int32Ty, // tid 2315 llvm::PointerType::getUnqual(CGM.Int32Ty), // p_lastiter 2316 PtrTy, // p_lower 2317 PtrTy, // p_upper 2318 PtrTy // p_stride 2319 }; 2320 llvm::FunctionType *FnTy = 2321 llvm::FunctionType::get(CGM.Int32Ty, TypeParams, /*isVarArg*/ false); 2322 return CGM.CreateRuntimeFunction(FnTy, Name); 2323 } 2324 2325 llvm::Constant * 2326 CGOpenMPRuntime::getOrCreateThreadPrivateCache(const VarDecl *VD) { 2327 assert(!CGM.getLangOpts().OpenMPUseTLS || 2328 !CGM.getContext().getTargetInfo().isTLSSupported()); 2329 // Lookup the entry, lazily creating it if necessary. 2330 return getOrCreateInternalVariable(CGM.Int8PtrPtrTy, 2331 Twine(CGM.getMangledName(VD)) + ".cache."); 2332 } 2333 2334 Address CGOpenMPRuntime::getAddrOfThreadPrivate(CodeGenFunction &CGF, 2335 const VarDecl *VD, 2336 Address VDAddr, 2337 SourceLocation Loc) { 2338 if (CGM.getLangOpts().OpenMPUseTLS && 2339 CGM.getContext().getTargetInfo().isTLSSupported()) 2340 return VDAddr; 2341 2342 auto VarTy = VDAddr.getElementType(); 2343 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 2344 CGF.Builder.CreatePointerCast(VDAddr.getPointer(), 2345 CGM.Int8PtrTy), 2346 CGM.getSize(CGM.GetTargetTypeStoreSize(VarTy)), 2347 getOrCreateThreadPrivateCache(VD)}; 2348 return Address(CGF.EmitRuntimeCall( 2349 createRuntimeFunction(OMPRTL__kmpc_threadprivate_cached), Args), 2350 VDAddr.getAlignment()); 2351 } 2352 2353 void CGOpenMPRuntime::emitThreadPrivateVarInit( 2354 CodeGenFunction &CGF, Address VDAddr, llvm::Value *Ctor, 2355 llvm::Value *CopyCtor, llvm::Value *Dtor, SourceLocation Loc) { 2356 // Call kmp_int32 __kmpc_global_thread_num(&loc) to init OpenMP runtime 2357 // library. 2358 auto OMPLoc = emitUpdateLocation(CGF, Loc); 2359 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_global_thread_num), 2360 OMPLoc); 2361 // Call __kmpc_threadprivate_register(&loc, &var, ctor, cctor/*NULL*/, dtor) 2362 // to register constructor/destructor for variable. 2363 llvm::Value *Args[] = {OMPLoc, 2364 CGF.Builder.CreatePointerCast(VDAddr.getPointer(), 2365 CGM.VoidPtrTy), 2366 Ctor, CopyCtor, Dtor}; 2367 CGF.EmitRuntimeCall( 2368 createRuntimeFunction(OMPRTL__kmpc_threadprivate_register), Args); 2369 } 2370 2371 llvm::Function *CGOpenMPRuntime::emitThreadPrivateVarDefinition( 2372 const VarDecl *VD, Address VDAddr, SourceLocation Loc, 2373 bool PerformInit, CodeGenFunction *CGF) { 2374 if (CGM.getLangOpts().OpenMPUseTLS && 2375 CGM.getContext().getTargetInfo().isTLSSupported()) 2376 return nullptr; 2377 2378 VD = VD->getDefinition(CGM.getContext()); 2379 if (VD && ThreadPrivateWithDefinition.count(VD) == 0) { 2380 ThreadPrivateWithDefinition.insert(VD); 2381 QualType ASTTy = VD->getType(); 2382 2383 llvm::Value *Ctor = nullptr, *CopyCtor = nullptr, *Dtor = nullptr; 2384 auto Init = VD->getAnyInitializer(); 2385 if (CGM.getLangOpts().CPlusPlus && PerformInit) { 2386 // Generate function that re-emits the declaration's initializer into the 2387 // threadprivate copy of the variable VD 2388 CodeGenFunction CtorCGF(CGM); 2389 FunctionArgList Args; 2390 ImplicitParamDecl Dst(CGM.getContext(), /*DC=*/nullptr, Loc, 2391 /*Id=*/nullptr, CGM.getContext().VoidPtrTy, 2392 ImplicitParamDecl::Other); 2393 Args.push_back(&Dst); 2394 2395 auto &FI = CGM.getTypes().arrangeBuiltinFunctionDeclaration( 2396 CGM.getContext().VoidPtrTy, Args); 2397 auto FTy = CGM.getTypes().GetFunctionType(FI); 2398 auto Fn = CGM.CreateGlobalInitOrDestructFunction( 2399 FTy, ".__kmpc_global_ctor_.", FI, Loc); 2400 CtorCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidPtrTy, Fn, FI, 2401 Args, Loc, Loc); 2402 auto ArgVal = CtorCGF.EmitLoadOfScalar( 2403 CtorCGF.GetAddrOfLocalVar(&Dst), /*Volatile=*/false, 2404 CGM.getContext().VoidPtrTy, Dst.getLocation()); 2405 Address Arg = Address(ArgVal, VDAddr.getAlignment()); 2406 Arg = CtorCGF.Builder.CreateElementBitCast( 2407 Arg, CtorCGF.ConvertTypeForMem(ASTTy)); 2408 CtorCGF.EmitAnyExprToMem(Init, Arg, Init->getType().getQualifiers(), 2409 /*IsInitializer=*/true); 2410 ArgVal = CtorCGF.EmitLoadOfScalar( 2411 CtorCGF.GetAddrOfLocalVar(&Dst), /*Volatile=*/false, 2412 CGM.getContext().VoidPtrTy, Dst.getLocation()); 2413 CtorCGF.Builder.CreateStore(ArgVal, CtorCGF.ReturnValue); 2414 CtorCGF.FinishFunction(); 2415 Ctor = Fn; 2416 } 2417 if (VD->getType().isDestructedType() != QualType::DK_none) { 2418 // Generate function that emits destructor call for the threadprivate copy 2419 // of the variable VD 2420 CodeGenFunction DtorCGF(CGM); 2421 FunctionArgList Args; 2422 ImplicitParamDecl Dst(CGM.getContext(), /*DC=*/nullptr, Loc, 2423 /*Id=*/nullptr, CGM.getContext().VoidPtrTy, 2424 ImplicitParamDecl::Other); 2425 Args.push_back(&Dst); 2426 2427 auto &FI = CGM.getTypes().arrangeBuiltinFunctionDeclaration( 2428 CGM.getContext().VoidTy, Args); 2429 auto FTy = CGM.getTypes().GetFunctionType(FI); 2430 auto Fn = CGM.CreateGlobalInitOrDestructFunction( 2431 FTy, ".__kmpc_global_dtor_.", FI, Loc); 2432 auto NL = ApplyDebugLocation::CreateEmpty(DtorCGF); 2433 DtorCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidTy, Fn, FI, Args, 2434 Loc, Loc); 2435 // Create a scope with an artificial location for the body of this function. 2436 auto AL = ApplyDebugLocation::CreateArtificial(DtorCGF); 2437 auto ArgVal = DtorCGF.EmitLoadOfScalar( 2438 DtorCGF.GetAddrOfLocalVar(&Dst), 2439 /*Volatile=*/false, CGM.getContext().VoidPtrTy, Dst.getLocation()); 2440 DtorCGF.emitDestroy(Address(ArgVal, VDAddr.getAlignment()), ASTTy, 2441 DtorCGF.getDestroyer(ASTTy.isDestructedType()), 2442 DtorCGF.needsEHCleanup(ASTTy.isDestructedType())); 2443 DtorCGF.FinishFunction(); 2444 Dtor = Fn; 2445 } 2446 // Do not emit init function if it is not required. 2447 if (!Ctor && !Dtor) 2448 return nullptr; 2449 2450 llvm::Type *CopyCtorTyArgs[] = {CGM.VoidPtrTy, CGM.VoidPtrTy}; 2451 auto CopyCtorTy = 2452 llvm::FunctionType::get(CGM.VoidPtrTy, CopyCtorTyArgs, 2453 /*isVarArg=*/false)->getPointerTo(); 2454 // Copying constructor for the threadprivate variable. 2455 // Must be NULL - reserved by runtime, but currently it requires that this 2456 // parameter is always NULL. Otherwise it fires assertion. 2457 CopyCtor = llvm::Constant::getNullValue(CopyCtorTy); 2458 if (Ctor == nullptr) { 2459 auto CtorTy = llvm::FunctionType::get(CGM.VoidPtrTy, CGM.VoidPtrTy, 2460 /*isVarArg=*/false)->getPointerTo(); 2461 Ctor = llvm::Constant::getNullValue(CtorTy); 2462 } 2463 if (Dtor == nullptr) { 2464 auto DtorTy = llvm::FunctionType::get(CGM.VoidTy, CGM.VoidPtrTy, 2465 /*isVarArg=*/false)->getPointerTo(); 2466 Dtor = llvm::Constant::getNullValue(DtorTy); 2467 } 2468 if (!CGF) { 2469 auto InitFunctionTy = 2470 llvm::FunctionType::get(CGM.VoidTy, /*isVarArg*/ false); 2471 auto InitFunction = CGM.CreateGlobalInitOrDestructFunction( 2472 InitFunctionTy, ".__omp_threadprivate_init_.", 2473 CGM.getTypes().arrangeNullaryFunction()); 2474 CodeGenFunction InitCGF(CGM); 2475 FunctionArgList ArgList; 2476 InitCGF.StartFunction(GlobalDecl(), CGM.getContext().VoidTy, InitFunction, 2477 CGM.getTypes().arrangeNullaryFunction(), ArgList, 2478 Loc, Loc); 2479 emitThreadPrivateVarInit(InitCGF, VDAddr, Ctor, CopyCtor, Dtor, Loc); 2480 InitCGF.FinishFunction(); 2481 return InitFunction; 2482 } 2483 emitThreadPrivateVarInit(*CGF, VDAddr, Ctor, CopyCtor, Dtor, Loc); 2484 } 2485 return nullptr; 2486 } 2487 2488 Address CGOpenMPRuntime::getAddrOfArtificialThreadPrivate(CodeGenFunction &CGF, 2489 QualType VarType, 2490 StringRef Name) { 2491 llvm::Twine VarName(Name, ".artificial."); 2492 llvm::Type *VarLVType = CGF.ConvertTypeForMem(VarType); 2493 llvm::Value *GAddr = getOrCreateInternalVariable(VarLVType, VarName); 2494 llvm::Value *Args[] = { 2495 emitUpdateLocation(CGF, SourceLocation()), 2496 getThreadID(CGF, SourceLocation()), 2497 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(GAddr, CGM.VoidPtrTy), 2498 CGF.Builder.CreateIntCast(CGF.getTypeSize(VarType), CGM.SizeTy, 2499 /*IsSigned=*/false), 2500 getOrCreateInternalVariable(CGM.VoidPtrPtrTy, VarName + ".cache.")}; 2501 return Address( 2502 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 2503 CGF.EmitRuntimeCall( 2504 createRuntimeFunction(OMPRTL__kmpc_threadprivate_cached), Args), 2505 VarLVType->getPointerTo(/*AddrSpace=*/0)), 2506 CGM.getPointerAlign()); 2507 } 2508 2509 /// \brief Emits code for OpenMP 'if' clause using specified \a CodeGen 2510 /// function. Here is the logic: 2511 /// if (Cond) { 2512 /// ThenGen(); 2513 /// } else { 2514 /// ElseGen(); 2515 /// } 2516 void CGOpenMPRuntime::emitOMPIfClause(CodeGenFunction &CGF, const Expr *Cond, 2517 const RegionCodeGenTy &ThenGen, 2518 const RegionCodeGenTy &ElseGen) { 2519 CodeGenFunction::LexicalScope ConditionScope(CGF, Cond->getSourceRange()); 2520 2521 // If the condition constant folds and can be elided, try to avoid emitting 2522 // the condition and the dead arm of the if/else. 2523 bool CondConstant; 2524 if (CGF.ConstantFoldsToSimpleInteger(Cond, CondConstant)) { 2525 if (CondConstant) 2526 ThenGen(CGF); 2527 else 2528 ElseGen(CGF); 2529 return; 2530 } 2531 2532 // Otherwise, the condition did not fold, or we couldn't elide it. Just 2533 // emit the conditional branch. 2534 auto ThenBlock = CGF.createBasicBlock("omp_if.then"); 2535 auto ElseBlock = CGF.createBasicBlock("omp_if.else"); 2536 auto ContBlock = CGF.createBasicBlock("omp_if.end"); 2537 CGF.EmitBranchOnBoolExpr(Cond, ThenBlock, ElseBlock, /*TrueCount=*/0); 2538 2539 // Emit the 'then' code. 2540 CGF.EmitBlock(ThenBlock); 2541 ThenGen(CGF); 2542 CGF.EmitBranch(ContBlock); 2543 // Emit the 'else' code if present. 2544 // There is no need to emit line number for unconditional branch. 2545 (void)ApplyDebugLocation::CreateEmpty(CGF); 2546 CGF.EmitBlock(ElseBlock); 2547 ElseGen(CGF); 2548 // There is no need to emit line number for unconditional branch. 2549 (void)ApplyDebugLocation::CreateEmpty(CGF); 2550 CGF.EmitBranch(ContBlock); 2551 // Emit the continuation block for code after the if. 2552 CGF.EmitBlock(ContBlock, /*IsFinished=*/true); 2553 } 2554 2555 void CGOpenMPRuntime::emitParallelCall(CodeGenFunction &CGF, SourceLocation Loc, 2556 llvm::Value *OutlinedFn, 2557 ArrayRef<llvm::Value *> CapturedVars, 2558 const Expr *IfCond) { 2559 if (!CGF.HaveInsertPoint()) 2560 return; 2561 auto *RTLoc = emitUpdateLocation(CGF, Loc); 2562 auto &&ThenGen = [OutlinedFn, CapturedVars, RTLoc](CodeGenFunction &CGF, 2563 PrePostActionTy &) { 2564 // Build call __kmpc_fork_call(loc, n, microtask, var1, .., varn); 2565 auto &RT = CGF.CGM.getOpenMPRuntime(); 2566 llvm::Value *Args[] = { 2567 RTLoc, 2568 CGF.Builder.getInt32(CapturedVars.size()), // Number of captured vars 2569 CGF.Builder.CreateBitCast(OutlinedFn, RT.getKmpc_MicroPointerTy())}; 2570 llvm::SmallVector<llvm::Value *, 16> RealArgs; 2571 RealArgs.append(std::begin(Args), std::end(Args)); 2572 RealArgs.append(CapturedVars.begin(), CapturedVars.end()); 2573 2574 auto RTLFn = RT.createRuntimeFunction(OMPRTL__kmpc_fork_call); 2575 CGF.EmitRuntimeCall(RTLFn, RealArgs); 2576 }; 2577 auto &&ElseGen = [OutlinedFn, CapturedVars, RTLoc, Loc](CodeGenFunction &CGF, 2578 PrePostActionTy &) { 2579 auto &RT = CGF.CGM.getOpenMPRuntime(); 2580 auto ThreadID = RT.getThreadID(CGF, Loc); 2581 // Build calls: 2582 // __kmpc_serialized_parallel(&Loc, GTid); 2583 llvm::Value *Args[] = {RTLoc, ThreadID}; 2584 CGF.EmitRuntimeCall( 2585 RT.createRuntimeFunction(OMPRTL__kmpc_serialized_parallel), Args); 2586 2587 // OutlinedFn(>id, &zero, CapturedStruct); 2588 auto ThreadIDAddr = RT.emitThreadIDAddress(CGF, Loc); 2589 Address ZeroAddr = 2590 CGF.CreateTempAlloca(CGF.Int32Ty, CharUnits::fromQuantity(4), 2591 /*Name*/ ".zero.addr"); 2592 CGF.InitTempAlloca(ZeroAddr, CGF.Builder.getInt32(/*C*/ 0)); 2593 llvm::SmallVector<llvm::Value *, 16> OutlinedFnArgs; 2594 OutlinedFnArgs.push_back(ThreadIDAddr.getPointer()); 2595 OutlinedFnArgs.push_back(ZeroAddr.getPointer()); 2596 OutlinedFnArgs.append(CapturedVars.begin(), CapturedVars.end()); 2597 RT.emitOutlinedFunctionCall(CGF, Loc, OutlinedFn, OutlinedFnArgs); 2598 2599 // __kmpc_end_serialized_parallel(&Loc, GTid); 2600 llvm::Value *EndArgs[] = {RT.emitUpdateLocation(CGF, Loc), ThreadID}; 2601 CGF.EmitRuntimeCall( 2602 RT.createRuntimeFunction(OMPRTL__kmpc_end_serialized_parallel), 2603 EndArgs); 2604 }; 2605 if (IfCond) 2606 emitOMPIfClause(CGF, IfCond, ThenGen, ElseGen); 2607 else { 2608 RegionCodeGenTy ThenRCG(ThenGen); 2609 ThenRCG(CGF); 2610 } 2611 } 2612 2613 // If we're inside an (outlined) parallel region, use the region info's 2614 // thread-ID variable (it is passed in a first argument of the outlined function 2615 // as "kmp_int32 *gtid"). Otherwise, if we're not inside parallel region, but in 2616 // regular serial code region, get thread ID by calling kmp_int32 2617 // kmpc_global_thread_num(ident_t *loc), stash this thread ID in a temporary and 2618 // return the address of that temp. 2619 Address CGOpenMPRuntime::emitThreadIDAddress(CodeGenFunction &CGF, 2620 SourceLocation Loc) { 2621 if (auto *OMPRegionInfo = 2622 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) 2623 if (OMPRegionInfo->getThreadIDVariable()) 2624 return OMPRegionInfo->getThreadIDVariableLValue(CGF).getAddress(); 2625 2626 auto ThreadID = getThreadID(CGF, Loc); 2627 auto Int32Ty = 2628 CGF.getContext().getIntTypeForBitwidth(/*DestWidth*/ 32, /*Signed*/ true); 2629 auto ThreadIDTemp = CGF.CreateMemTemp(Int32Ty, /*Name*/ ".threadid_temp."); 2630 CGF.EmitStoreOfScalar(ThreadID, 2631 CGF.MakeAddrLValue(ThreadIDTemp, Int32Ty)); 2632 2633 return ThreadIDTemp; 2634 } 2635 2636 llvm::Constant * 2637 CGOpenMPRuntime::getOrCreateInternalVariable(llvm::Type *Ty, 2638 const llvm::Twine &Name) { 2639 SmallString<256> Buffer; 2640 llvm::raw_svector_ostream Out(Buffer); 2641 Out << Name; 2642 auto RuntimeName = Out.str(); 2643 auto &Elem = *InternalVars.insert(std::make_pair(RuntimeName, nullptr)).first; 2644 if (Elem.second) { 2645 assert(Elem.second->getType()->getPointerElementType() == Ty && 2646 "OMP internal variable has different type than requested"); 2647 return &*Elem.second; 2648 } 2649 2650 return Elem.second = new llvm::GlobalVariable( 2651 CGM.getModule(), Ty, /*IsConstant*/ false, 2652 llvm::GlobalValue::CommonLinkage, llvm::Constant::getNullValue(Ty), 2653 Elem.first()); 2654 } 2655 2656 llvm::Value *CGOpenMPRuntime::getCriticalRegionLock(StringRef CriticalName) { 2657 llvm::Twine Name(".gomp_critical_user_", CriticalName); 2658 return getOrCreateInternalVariable(KmpCriticalNameTy, Name.concat(".var")); 2659 } 2660 2661 namespace { 2662 /// Common pre(post)-action for different OpenMP constructs. 2663 class CommonActionTy final : public PrePostActionTy { 2664 llvm::Value *EnterCallee; 2665 ArrayRef<llvm::Value *> EnterArgs; 2666 llvm::Value *ExitCallee; 2667 ArrayRef<llvm::Value *> ExitArgs; 2668 bool Conditional; 2669 llvm::BasicBlock *ContBlock = nullptr; 2670 2671 public: 2672 CommonActionTy(llvm::Value *EnterCallee, ArrayRef<llvm::Value *> EnterArgs, 2673 llvm::Value *ExitCallee, ArrayRef<llvm::Value *> ExitArgs, 2674 bool Conditional = false) 2675 : EnterCallee(EnterCallee), EnterArgs(EnterArgs), ExitCallee(ExitCallee), 2676 ExitArgs(ExitArgs), Conditional(Conditional) {} 2677 void Enter(CodeGenFunction &CGF) override { 2678 llvm::Value *EnterRes = CGF.EmitRuntimeCall(EnterCallee, EnterArgs); 2679 if (Conditional) { 2680 llvm::Value *CallBool = CGF.Builder.CreateIsNotNull(EnterRes); 2681 auto *ThenBlock = CGF.createBasicBlock("omp_if.then"); 2682 ContBlock = CGF.createBasicBlock("omp_if.end"); 2683 // Generate the branch (If-stmt) 2684 CGF.Builder.CreateCondBr(CallBool, ThenBlock, ContBlock); 2685 CGF.EmitBlock(ThenBlock); 2686 } 2687 } 2688 void Done(CodeGenFunction &CGF) { 2689 // Emit the rest of blocks/branches 2690 CGF.EmitBranch(ContBlock); 2691 CGF.EmitBlock(ContBlock, true); 2692 } 2693 void Exit(CodeGenFunction &CGF) override { 2694 CGF.EmitRuntimeCall(ExitCallee, ExitArgs); 2695 } 2696 }; 2697 } // anonymous namespace 2698 2699 void CGOpenMPRuntime::emitCriticalRegion(CodeGenFunction &CGF, 2700 StringRef CriticalName, 2701 const RegionCodeGenTy &CriticalOpGen, 2702 SourceLocation Loc, const Expr *Hint) { 2703 // __kmpc_critical[_with_hint](ident_t *, gtid, Lock[, hint]); 2704 // CriticalOpGen(); 2705 // __kmpc_end_critical(ident_t *, gtid, Lock); 2706 // Prepare arguments and build a call to __kmpc_critical 2707 if (!CGF.HaveInsertPoint()) 2708 return; 2709 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 2710 getCriticalRegionLock(CriticalName)}; 2711 llvm::SmallVector<llvm::Value *, 4> EnterArgs(std::begin(Args), 2712 std::end(Args)); 2713 if (Hint) { 2714 EnterArgs.push_back(CGF.Builder.CreateIntCast( 2715 CGF.EmitScalarExpr(Hint), CGM.IntPtrTy, /*isSigned=*/false)); 2716 } 2717 CommonActionTy Action( 2718 createRuntimeFunction(Hint ? OMPRTL__kmpc_critical_with_hint 2719 : OMPRTL__kmpc_critical), 2720 EnterArgs, createRuntimeFunction(OMPRTL__kmpc_end_critical), Args); 2721 CriticalOpGen.setAction(Action); 2722 emitInlinedDirective(CGF, OMPD_critical, CriticalOpGen); 2723 } 2724 2725 void CGOpenMPRuntime::emitMasterRegion(CodeGenFunction &CGF, 2726 const RegionCodeGenTy &MasterOpGen, 2727 SourceLocation Loc) { 2728 if (!CGF.HaveInsertPoint()) 2729 return; 2730 // if(__kmpc_master(ident_t *, gtid)) { 2731 // MasterOpGen(); 2732 // __kmpc_end_master(ident_t *, gtid); 2733 // } 2734 // Prepare arguments and build a call to __kmpc_master 2735 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 2736 CommonActionTy Action(createRuntimeFunction(OMPRTL__kmpc_master), Args, 2737 createRuntimeFunction(OMPRTL__kmpc_end_master), Args, 2738 /*Conditional=*/true); 2739 MasterOpGen.setAction(Action); 2740 emitInlinedDirective(CGF, OMPD_master, MasterOpGen); 2741 Action.Done(CGF); 2742 } 2743 2744 void CGOpenMPRuntime::emitTaskyieldCall(CodeGenFunction &CGF, 2745 SourceLocation Loc) { 2746 if (!CGF.HaveInsertPoint()) 2747 return; 2748 // Build call __kmpc_omp_taskyield(loc, thread_id, 0); 2749 llvm::Value *Args[] = { 2750 emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 2751 llvm::ConstantInt::get(CGM.IntTy, /*V=*/0, /*isSigned=*/true)}; 2752 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_omp_taskyield), Args); 2753 if (auto *Region = dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) 2754 Region->emitUntiedSwitch(CGF); 2755 } 2756 2757 void CGOpenMPRuntime::emitTaskgroupRegion(CodeGenFunction &CGF, 2758 const RegionCodeGenTy &TaskgroupOpGen, 2759 SourceLocation Loc) { 2760 if (!CGF.HaveInsertPoint()) 2761 return; 2762 // __kmpc_taskgroup(ident_t *, gtid); 2763 // TaskgroupOpGen(); 2764 // __kmpc_end_taskgroup(ident_t *, gtid); 2765 // Prepare arguments and build a call to __kmpc_taskgroup 2766 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 2767 CommonActionTy Action(createRuntimeFunction(OMPRTL__kmpc_taskgroup), Args, 2768 createRuntimeFunction(OMPRTL__kmpc_end_taskgroup), 2769 Args); 2770 TaskgroupOpGen.setAction(Action); 2771 emitInlinedDirective(CGF, OMPD_taskgroup, TaskgroupOpGen); 2772 } 2773 2774 /// Given an array of pointers to variables, project the address of a 2775 /// given variable. 2776 static Address emitAddrOfVarFromArray(CodeGenFunction &CGF, Address Array, 2777 unsigned Index, const VarDecl *Var) { 2778 // Pull out the pointer to the variable. 2779 Address PtrAddr = 2780 CGF.Builder.CreateConstArrayGEP(Array, Index, CGF.getPointerSize()); 2781 llvm::Value *Ptr = CGF.Builder.CreateLoad(PtrAddr); 2782 2783 Address Addr = Address(Ptr, CGF.getContext().getDeclAlign(Var)); 2784 Addr = CGF.Builder.CreateElementBitCast( 2785 Addr, CGF.ConvertTypeForMem(Var->getType())); 2786 return Addr; 2787 } 2788 2789 static llvm::Value *emitCopyprivateCopyFunction( 2790 CodeGenModule &CGM, llvm::Type *ArgsType, 2791 ArrayRef<const Expr *> CopyprivateVars, ArrayRef<const Expr *> DestExprs, 2792 ArrayRef<const Expr *> SrcExprs, ArrayRef<const Expr *> AssignmentOps, 2793 SourceLocation Loc) { 2794 auto &C = CGM.getContext(); 2795 // void copy_func(void *LHSArg, void *RHSArg); 2796 FunctionArgList Args; 2797 ImplicitParamDecl LHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 2798 ImplicitParamDecl::Other); 2799 ImplicitParamDecl RHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 2800 ImplicitParamDecl::Other); 2801 Args.push_back(&LHSArg); 2802 Args.push_back(&RHSArg); 2803 auto &CGFI = CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 2804 auto *Fn = llvm::Function::Create( 2805 CGM.getTypes().GetFunctionType(CGFI), llvm::GlobalValue::InternalLinkage, 2806 ".omp.copyprivate.copy_func", &CGM.getModule()); 2807 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, CGFI); 2808 CodeGenFunction CGF(CGM); 2809 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, CGFI, Args, Loc, Loc); 2810 // Dest = (void*[n])(LHSArg); 2811 // Src = (void*[n])(RHSArg); 2812 Address LHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 2813 CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&LHSArg)), 2814 ArgsType), CGF.getPointerAlign()); 2815 Address RHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 2816 CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&RHSArg)), 2817 ArgsType), CGF.getPointerAlign()); 2818 // *(Type0*)Dst[0] = *(Type0*)Src[0]; 2819 // *(Type1*)Dst[1] = *(Type1*)Src[1]; 2820 // ... 2821 // *(Typen*)Dst[n] = *(Typen*)Src[n]; 2822 for (unsigned I = 0, E = AssignmentOps.size(); I < E; ++I) { 2823 auto DestVar = cast<VarDecl>(cast<DeclRefExpr>(DestExprs[I])->getDecl()); 2824 Address DestAddr = emitAddrOfVarFromArray(CGF, LHS, I, DestVar); 2825 2826 auto SrcVar = cast<VarDecl>(cast<DeclRefExpr>(SrcExprs[I])->getDecl()); 2827 Address SrcAddr = emitAddrOfVarFromArray(CGF, RHS, I, SrcVar); 2828 2829 auto *VD = cast<DeclRefExpr>(CopyprivateVars[I])->getDecl(); 2830 QualType Type = VD->getType(); 2831 CGF.EmitOMPCopy(Type, DestAddr, SrcAddr, DestVar, SrcVar, AssignmentOps[I]); 2832 } 2833 CGF.FinishFunction(); 2834 return Fn; 2835 } 2836 2837 void CGOpenMPRuntime::emitSingleRegion(CodeGenFunction &CGF, 2838 const RegionCodeGenTy &SingleOpGen, 2839 SourceLocation Loc, 2840 ArrayRef<const Expr *> CopyprivateVars, 2841 ArrayRef<const Expr *> SrcExprs, 2842 ArrayRef<const Expr *> DstExprs, 2843 ArrayRef<const Expr *> AssignmentOps) { 2844 if (!CGF.HaveInsertPoint()) 2845 return; 2846 assert(CopyprivateVars.size() == SrcExprs.size() && 2847 CopyprivateVars.size() == DstExprs.size() && 2848 CopyprivateVars.size() == AssignmentOps.size()); 2849 auto &C = CGM.getContext(); 2850 // int32 did_it = 0; 2851 // if(__kmpc_single(ident_t *, gtid)) { 2852 // SingleOpGen(); 2853 // __kmpc_end_single(ident_t *, gtid); 2854 // did_it = 1; 2855 // } 2856 // call __kmpc_copyprivate(ident_t *, gtid, <buf_size>, <copyprivate list>, 2857 // <copy_func>, did_it); 2858 2859 Address DidIt = Address::invalid(); 2860 if (!CopyprivateVars.empty()) { 2861 // int32 did_it = 0; 2862 auto KmpInt32Ty = C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/1); 2863 DidIt = CGF.CreateMemTemp(KmpInt32Ty, ".omp.copyprivate.did_it"); 2864 CGF.Builder.CreateStore(CGF.Builder.getInt32(0), DidIt); 2865 } 2866 // Prepare arguments and build a call to __kmpc_single 2867 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 2868 CommonActionTy Action(createRuntimeFunction(OMPRTL__kmpc_single), Args, 2869 createRuntimeFunction(OMPRTL__kmpc_end_single), Args, 2870 /*Conditional=*/true); 2871 SingleOpGen.setAction(Action); 2872 emitInlinedDirective(CGF, OMPD_single, SingleOpGen); 2873 if (DidIt.isValid()) { 2874 // did_it = 1; 2875 CGF.Builder.CreateStore(CGF.Builder.getInt32(1), DidIt); 2876 } 2877 Action.Done(CGF); 2878 // call __kmpc_copyprivate(ident_t *, gtid, <buf_size>, <copyprivate list>, 2879 // <copy_func>, did_it); 2880 if (DidIt.isValid()) { 2881 llvm::APInt ArraySize(/*unsigned int numBits=*/32, CopyprivateVars.size()); 2882 auto CopyprivateArrayTy = 2883 C.getConstantArrayType(C.VoidPtrTy, ArraySize, ArrayType::Normal, 2884 /*IndexTypeQuals=*/0); 2885 // Create a list of all private variables for copyprivate. 2886 Address CopyprivateList = 2887 CGF.CreateMemTemp(CopyprivateArrayTy, ".omp.copyprivate.cpr_list"); 2888 for (unsigned I = 0, E = CopyprivateVars.size(); I < E; ++I) { 2889 Address Elem = CGF.Builder.CreateConstArrayGEP( 2890 CopyprivateList, I, CGF.getPointerSize()); 2891 CGF.Builder.CreateStore( 2892 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 2893 CGF.EmitLValue(CopyprivateVars[I]).getPointer(), CGF.VoidPtrTy), 2894 Elem); 2895 } 2896 // Build function that copies private values from single region to all other 2897 // threads in the corresponding parallel region. 2898 auto *CpyFn = emitCopyprivateCopyFunction( 2899 CGM, CGF.ConvertTypeForMem(CopyprivateArrayTy)->getPointerTo(), 2900 CopyprivateVars, SrcExprs, DstExprs, AssignmentOps, Loc); 2901 auto *BufSize = CGF.getTypeSize(CopyprivateArrayTy); 2902 Address CL = 2903 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(CopyprivateList, 2904 CGF.VoidPtrTy); 2905 auto *DidItVal = CGF.Builder.CreateLoad(DidIt); 2906 llvm::Value *Args[] = { 2907 emitUpdateLocation(CGF, Loc), // ident_t *<loc> 2908 getThreadID(CGF, Loc), // i32 <gtid> 2909 BufSize, // size_t <buf_size> 2910 CL.getPointer(), // void *<copyprivate list> 2911 CpyFn, // void (*) (void *, void *) <copy_func> 2912 DidItVal // i32 did_it 2913 }; 2914 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_copyprivate), Args); 2915 } 2916 } 2917 2918 void CGOpenMPRuntime::emitOrderedRegion(CodeGenFunction &CGF, 2919 const RegionCodeGenTy &OrderedOpGen, 2920 SourceLocation Loc, bool IsThreads) { 2921 if (!CGF.HaveInsertPoint()) 2922 return; 2923 // __kmpc_ordered(ident_t *, gtid); 2924 // OrderedOpGen(); 2925 // __kmpc_end_ordered(ident_t *, gtid); 2926 // Prepare arguments and build a call to __kmpc_ordered 2927 if (IsThreads) { 2928 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 2929 CommonActionTy Action(createRuntimeFunction(OMPRTL__kmpc_ordered), Args, 2930 createRuntimeFunction(OMPRTL__kmpc_end_ordered), 2931 Args); 2932 OrderedOpGen.setAction(Action); 2933 emitInlinedDirective(CGF, OMPD_ordered, OrderedOpGen); 2934 return; 2935 } 2936 emitInlinedDirective(CGF, OMPD_ordered, OrderedOpGen); 2937 } 2938 2939 void CGOpenMPRuntime::emitBarrierCall(CodeGenFunction &CGF, SourceLocation Loc, 2940 OpenMPDirectiveKind Kind, bool EmitChecks, 2941 bool ForceSimpleCall) { 2942 if (!CGF.HaveInsertPoint()) 2943 return; 2944 // Build call __kmpc_cancel_barrier(loc, thread_id); 2945 // Build call __kmpc_barrier(loc, thread_id); 2946 unsigned Flags; 2947 if (Kind == OMPD_for) 2948 Flags = OMP_IDENT_BARRIER_IMPL_FOR; 2949 else if (Kind == OMPD_sections) 2950 Flags = OMP_IDENT_BARRIER_IMPL_SECTIONS; 2951 else if (Kind == OMPD_single) 2952 Flags = OMP_IDENT_BARRIER_IMPL_SINGLE; 2953 else if (Kind == OMPD_barrier) 2954 Flags = OMP_IDENT_BARRIER_EXPL; 2955 else 2956 Flags = OMP_IDENT_BARRIER_IMPL; 2957 // Build call __kmpc_cancel_barrier(loc, thread_id) or __kmpc_barrier(loc, 2958 // thread_id); 2959 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc, Flags), 2960 getThreadID(CGF, Loc)}; 2961 if (auto *OMPRegionInfo = 2962 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) { 2963 if (!ForceSimpleCall && OMPRegionInfo->hasCancel()) { 2964 auto *Result = CGF.EmitRuntimeCall( 2965 createRuntimeFunction(OMPRTL__kmpc_cancel_barrier), Args); 2966 if (EmitChecks) { 2967 // if (__kmpc_cancel_barrier()) { 2968 // exit from construct; 2969 // } 2970 auto *ExitBB = CGF.createBasicBlock(".cancel.exit"); 2971 auto *ContBB = CGF.createBasicBlock(".cancel.continue"); 2972 auto *Cmp = CGF.Builder.CreateIsNotNull(Result); 2973 CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB); 2974 CGF.EmitBlock(ExitBB); 2975 // exit from construct; 2976 auto CancelDestination = 2977 CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind()); 2978 CGF.EmitBranchThroughCleanup(CancelDestination); 2979 CGF.EmitBlock(ContBB, /*IsFinished=*/true); 2980 } 2981 return; 2982 } 2983 } 2984 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_barrier), Args); 2985 } 2986 2987 /// \brief Map the OpenMP loop schedule to the runtime enumeration. 2988 static OpenMPSchedType getRuntimeSchedule(OpenMPScheduleClauseKind ScheduleKind, 2989 bool Chunked, bool Ordered) { 2990 switch (ScheduleKind) { 2991 case OMPC_SCHEDULE_static: 2992 return Chunked ? (Ordered ? OMP_ord_static_chunked : OMP_sch_static_chunked) 2993 : (Ordered ? OMP_ord_static : OMP_sch_static); 2994 case OMPC_SCHEDULE_dynamic: 2995 return Ordered ? OMP_ord_dynamic_chunked : OMP_sch_dynamic_chunked; 2996 case OMPC_SCHEDULE_guided: 2997 return Ordered ? OMP_ord_guided_chunked : OMP_sch_guided_chunked; 2998 case OMPC_SCHEDULE_runtime: 2999 return Ordered ? OMP_ord_runtime : OMP_sch_runtime; 3000 case OMPC_SCHEDULE_auto: 3001 return Ordered ? OMP_ord_auto : OMP_sch_auto; 3002 case OMPC_SCHEDULE_unknown: 3003 assert(!Chunked && "chunk was specified but schedule kind not known"); 3004 return Ordered ? OMP_ord_static : OMP_sch_static; 3005 } 3006 llvm_unreachable("Unexpected runtime schedule"); 3007 } 3008 3009 /// \brief Map the OpenMP distribute schedule to the runtime enumeration. 3010 static OpenMPSchedType 3011 getRuntimeSchedule(OpenMPDistScheduleClauseKind ScheduleKind, bool Chunked) { 3012 // only static is allowed for dist_schedule 3013 return Chunked ? OMP_dist_sch_static_chunked : OMP_dist_sch_static; 3014 } 3015 3016 bool CGOpenMPRuntime::isStaticNonchunked(OpenMPScheduleClauseKind ScheduleKind, 3017 bool Chunked) const { 3018 auto Schedule = getRuntimeSchedule(ScheduleKind, Chunked, /*Ordered=*/false); 3019 return Schedule == OMP_sch_static; 3020 } 3021 3022 bool CGOpenMPRuntime::isStaticNonchunked( 3023 OpenMPDistScheduleClauseKind ScheduleKind, bool Chunked) const { 3024 auto Schedule = getRuntimeSchedule(ScheduleKind, Chunked); 3025 return Schedule == OMP_dist_sch_static; 3026 } 3027 3028 3029 bool CGOpenMPRuntime::isDynamic(OpenMPScheduleClauseKind ScheduleKind) const { 3030 auto Schedule = 3031 getRuntimeSchedule(ScheduleKind, /*Chunked=*/false, /*Ordered=*/false); 3032 assert(Schedule != OMP_sch_static_chunked && "cannot be chunked here"); 3033 return Schedule != OMP_sch_static; 3034 } 3035 3036 static int addMonoNonMonoModifier(OpenMPSchedType Schedule, 3037 OpenMPScheduleClauseModifier M1, 3038 OpenMPScheduleClauseModifier M2) { 3039 int Modifier = 0; 3040 switch (M1) { 3041 case OMPC_SCHEDULE_MODIFIER_monotonic: 3042 Modifier = OMP_sch_modifier_monotonic; 3043 break; 3044 case OMPC_SCHEDULE_MODIFIER_nonmonotonic: 3045 Modifier = OMP_sch_modifier_nonmonotonic; 3046 break; 3047 case OMPC_SCHEDULE_MODIFIER_simd: 3048 if (Schedule == OMP_sch_static_chunked) 3049 Schedule = OMP_sch_static_balanced_chunked; 3050 break; 3051 case OMPC_SCHEDULE_MODIFIER_last: 3052 case OMPC_SCHEDULE_MODIFIER_unknown: 3053 break; 3054 } 3055 switch (M2) { 3056 case OMPC_SCHEDULE_MODIFIER_monotonic: 3057 Modifier = OMP_sch_modifier_monotonic; 3058 break; 3059 case OMPC_SCHEDULE_MODIFIER_nonmonotonic: 3060 Modifier = OMP_sch_modifier_nonmonotonic; 3061 break; 3062 case OMPC_SCHEDULE_MODIFIER_simd: 3063 if (Schedule == OMP_sch_static_chunked) 3064 Schedule = OMP_sch_static_balanced_chunked; 3065 break; 3066 case OMPC_SCHEDULE_MODIFIER_last: 3067 case OMPC_SCHEDULE_MODIFIER_unknown: 3068 break; 3069 } 3070 return Schedule | Modifier; 3071 } 3072 3073 void CGOpenMPRuntime::emitForDispatchInit( 3074 CodeGenFunction &CGF, SourceLocation Loc, 3075 const OpenMPScheduleTy &ScheduleKind, unsigned IVSize, bool IVSigned, 3076 bool Ordered, const DispatchRTInput &DispatchValues) { 3077 if (!CGF.HaveInsertPoint()) 3078 return; 3079 OpenMPSchedType Schedule = getRuntimeSchedule( 3080 ScheduleKind.Schedule, DispatchValues.Chunk != nullptr, Ordered); 3081 assert(Ordered || 3082 (Schedule != OMP_sch_static && Schedule != OMP_sch_static_chunked && 3083 Schedule != OMP_ord_static && Schedule != OMP_ord_static_chunked && 3084 Schedule != OMP_sch_static_balanced_chunked)); 3085 // Call __kmpc_dispatch_init( 3086 // ident_t *loc, kmp_int32 tid, kmp_int32 schedule, 3087 // kmp_int[32|64] lower, kmp_int[32|64] upper, 3088 // kmp_int[32|64] stride, kmp_int[32|64] chunk); 3089 3090 // If the Chunk was not specified in the clause - use default value 1. 3091 llvm::Value *Chunk = DispatchValues.Chunk ? DispatchValues.Chunk 3092 : CGF.Builder.getIntN(IVSize, 1); 3093 llvm::Value *Args[] = { 3094 emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 3095 CGF.Builder.getInt32(addMonoNonMonoModifier( 3096 Schedule, ScheduleKind.M1, ScheduleKind.M2)), // Schedule type 3097 DispatchValues.LB, // Lower 3098 DispatchValues.UB, // Upper 3099 CGF.Builder.getIntN(IVSize, 1), // Stride 3100 Chunk // Chunk 3101 }; 3102 CGF.EmitRuntimeCall(createDispatchInitFunction(IVSize, IVSigned), Args); 3103 } 3104 3105 static void emitForStaticInitCall( 3106 CodeGenFunction &CGF, llvm::Value *UpdateLocation, llvm::Value *ThreadId, 3107 llvm::Constant *ForStaticInitFunction, OpenMPSchedType Schedule, 3108 OpenMPScheduleClauseModifier M1, OpenMPScheduleClauseModifier M2, 3109 const CGOpenMPRuntime::StaticRTInput &Values) { 3110 if (!CGF.HaveInsertPoint()) 3111 return; 3112 3113 assert(!Values.Ordered); 3114 assert(Schedule == OMP_sch_static || Schedule == OMP_sch_static_chunked || 3115 Schedule == OMP_sch_static_balanced_chunked || 3116 Schedule == OMP_ord_static || Schedule == OMP_ord_static_chunked || 3117 Schedule == OMP_dist_sch_static || 3118 Schedule == OMP_dist_sch_static_chunked); 3119 3120 // Call __kmpc_for_static_init( 3121 // ident_t *loc, kmp_int32 tid, kmp_int32 schedtype, 3122 // kmp_int32 *p_lastiter, kmp_int[32|64] *p_lower, 3123 // kmp_int[32|64] *p_upper, kmp_int[32|64] *p_stride, 3124 // kmp_int[32|64] incr, kmp_int[32|64] chunk); 3125 llvm::Value *Chunk = Values.Chunk; 3126 if (Chunk == nullptr) { 3127 assert((Schedule == OMP_sch_static || Schedule == OMP_ord_static || 3128 Schedule == OMP_dist_sch_static) && 3129 "expected static non-chunked schedule"); 3130 // If the Chunk was not specified in the clause - use default value 1. 3131 Chunk = CGF.Builder.getIntN(Values.IVSize, 1); 3132 } else { 3133 assert((Schedule == OMP_sch_static_chunked || 3134 Schedule == OMP_sch_static_balanced_chunked || 3135 Schedule == OMP_ord_static_chunked || 3136 Schedule == OMP_dist_sch_static_chunked) && 3137 "expected static chunked schedule"); 3138 } 3139 llvm::Value *Args[] = { 3140 UpdateLocation, 3141 ThreadId, 3142 CGF.Builder.getInt32(addMonoNonMonoModifier(Schedule, M1, 3143 M2)), // Schedule type 3144 Values.IL.getPointer(), // &isLastIter 3145 Values.LB.getPointer(), // &LB 3146 Values.UB.getPointer(), // &UB 3147 Values.ST.getPointer(), // &Stride 3148 CGF.Builder.getIntN(Values.IVSize, 1), // Incr 3149 Chunk // Chunk 3150 }; 3151 CGF.EmitRuntimeCall(ForStaticInitFunction, Args); 3152 } 3153 3154 void CGOpenMPRuntime::emitForStaticInit(CodeGenFunction &CGF, 3155 SourceLocation Loc, 3156 OpenMPDirectiveKind DKind, 3157 const OpenMPScheduleTy &ScheduleKind, 3158 const StaticRTInput &Values) { 3159 OpenMPSchedType ScheduleNum = getRuntimeSchedule( 3160 ScheduleKind.Schedule, Values.Chunk != nullptr, Values.Ordered); 3161 assert(isOpenMPWorksharingDirective(DKind) && 3162 "Expected loop-based or sections-based directive."); 3163 auto *UpdatedLocation = emitUpdateLocation(CGF, Loc, 3164 isOpenMPLoopDirective(DKind) 3165 ? OMP_IDENT_WORK_LOOP 3166 : OMP_IDENT_WORK_SECTIONS); 3167 auto *ThreadId = getThreadID(CGF, Loc); 3168 auto *StaticInitFunction = 3169 createForStaticInitFunction(Values.IVSize, Values.IVSigned); 3170 emitForStaticInitCall(CGF, UpdatedLocation, ThreadId, StaticInitFunction, 3171 ScheduleNum, ScheduleKind.M1, ScheduleKind.M2, Values); 3172 } 3173 3174 void CGOpenMPRuntime::emitDistributeStaticInit( 3175 CodeGenFunction &CGF, SourceLocation Loc, 3176 OpenMPDistScheduleClauseKind SchedKind, 3177 const CGOpenMPRuntime::StaticRTInput &Values) { 3178 OpenMPSchedType ScheduleNum = 3179 getRuntimeSchedule(SchedKind, Values.Chunk != nullptr); 3180 auto *UpdatedLocation = 3181 emitUpdateLocation(CGF, Loc, OMP_IDENT_WORK_DISTRIBUTE); 3182 auto *ThreadId = getThreadID(CGF, Loc); 3183 auto *StaticInitFunction = 3184 createForStaticInitFunction(Values.IVSize, Values.IVSigned); 3185 emitForStaticInitCall(CGF, UpdatedLocation, ThreadId, StaticInitFunction, 3186 ScheduleNum, OMPC_SCHEDULE_MODIFIER_unknown, 3187 OMPC_SCHEDULE_MODIFIER_unknown, Values); 3188 } 3189 3190 void CGOpenMPRuntime::emitForStaticFinish(CodeGenFunction &CGF, 3191 SourceLocation Loc, 3192 OpenMPDirectiveKind DKind) { 3193 if (!CGF.HaveInsertPoint()) 3194 return; 3195 // Call __kmpc_for_static_fini(ident_t *loc, kmp_int32 tid); 3196 llvm::Value *Args[] = { 3197 emitUpdateLocation(CGF, Loc, 3198 isOpenMPDistributeDirective(DKind) 3199 ? OMP_IDENT_WORK_DISTRIBUTE 3200 : isOpenMPLoopDirective(DKind) 3201 ? OMP_IDENT_WORK_LOOP 3202 : OMP_IDENT_WORK_SECTIONS), 3203 getThreadID(CGF, Loc)}; 3204 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_for_static_fini), 3205 Args); 3206 } 3207 3208 void CGOpenMPRuntime::emitForOrderedIterationEnd(CodeGenFunction &CGF, 3209 SourceLocation Loc, 3210 unsigned IVSize, 3211 bool IVSigned) { 3212 if (!CGF.HaveInsertPoint()) 3213 return; 3214 // Call __kmpc_for_dynamic_fini_(4|8)[u](ident_t *loc, kmp_int32 tid); 3215 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 3216 CGF.EmitRuntimeCall(createDispatchFiniFunction(IVSize, IVSigned), Args); 3217 } 3218 3219 llvm::Value *CGOpenMPRuntime::emitForNext(CodeGenFunction &CGF, 3220 SourceLocation Loc, unsigned IVSize, 3221 bool IVSigned, Address IL, 3222 Address LB, Address UB, 3223 Address ST) { 3224 // Call __kmpc_dispatch_next( 3225 // ident_t *loc, kmp_int32 tid, kmp_int32 *p_lastiter, 3226 // kmp_int[32|64] *p_lower, kmp_int[32|64] *p_upper, 3227 // kmp_int[32|64] *p_stride); 3228 llvm::Value *Args[] = { 3229 emitUpdateLocation(CGF, Loc), 3230 getThreadID(CGF, Loc), 3231 IL.getPointer(), // &isLastIter 3232 LB.getPointer(), // &Lower 3233 UB.getPointer(), // &Upper 3234 ST.getPointer() // &Stride 3235 }; 3236 llvm::Value *Call = 3237 CGF.EmitRuntimeCall(createDispatchNextFunction(IVSize, IVSigned), Args); 3238 return CGF.EmitScalarConversion( 3239 Call, CGF.getContext().getIntTypeForBitwidth(32, /* Signed */ true), 3240 CGF.getContext().BoolTy, Loc); 3241 } 3242 3243 void CGOpenMPRuntime::emitNumThreadsClause(CodeGenFunction &CGF, 3244 llvm::Value *NumThreads, 3245 SourceLocation Loc) { 3246 if (!CGF.HaveInsertPoint()) 3247 return; 3248 // Build call __kmpc_push_num_threads(&loc, global_tid, num_threads) 3249 llvm::Value *Args[] = { 3250 emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 3251 CGF.Builder.CreateIntCast(NumThreads, CGF.Int32Ty, /*isSigned*/ true)}; 3252 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_push_num_threads), 3253 Args); 3254 } 3255 3256 void CGOpenMPRuntime::emitProcBindClause(CodeGenFunction &CGF, 3257 OpenMPProcBindClauseKind ProcBind, 3258 SourceLocation Loc) { 3259 if (!CGF.HaveInsertPoint()) 3260 return; 3261 // Constants for proc bind value accepted by the runtime. 3262 enum ProcBindTy { 3263 ProcBindFalse = 0, 3264 ProcBindTrue, 3265 ProcBindMaster, 3266 ProcBindClose, 3267 ProcBindSpread, 3268 ProcBindIntel, 3269 ProcBindDefault 3270 } RuntimeProcBind; 3271 switch (ProcBind) { 3272 case OMPC_PROC_BIND_master: 3273 RuntimeProcBind = ProcBindMaster; 3274 break; 3275 case OMPC_PROC_BIND_close: 3276 RuntimeProcBind = ProcBindClose; 3277 break; 3278 case OMPC_PROC_BIND_spread: 3279 RuntimeProcBind = ProcBindSpread; 3280 break; 3281 case OMPC_PROC_BIND_unknown: 3282 llvm_unreachable("Unsupported proc_bind value."); 3283 } 3284 // Build call __kmpc_push_proc_bind(&loc, global_tid, proc_bind) 3285 llvm::Value *Args[] = { 3286 emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 3287 llvm::ConstantInt::get(CGM.IntTy, RuntimeProcBind, /*isSigned=*/true)}; 3288 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_push_proc_bind), Args); 3289 } 3290 3291 void CGOpenMPRuntime::emitFlush(CodeGenFunction &CGF, ArrayRef<const Expr *>, 3292 SourceLocation Loc) { 3293 if (!CGF.HaveInsertPoint()) 3294 return; 3295 // Build call void __kmpc_flush(ident_t *loc) 3296 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_flush), 3297 emitUpdateLocation(CGF, Loc)); 3298 } 3299 3300 namespace { 3301 /// \brief Indexes of fields for type kmp_task_t. 3302 enum KmpTaskTFields { 3303 /// \brief List of shared variables. 3304 KmpTaskTShareds, 3305 /// \brief Task routine. 3306 KmpTaskTRoutine, 3307 /// \brief Partition id for the untied tasks. 3308 KmpTaskTPartId, 3309 /// Function with call of destructors for private variables. 3310 Data1, 3311 /// Task priority. 3312 Data2, 3313 /// (Taskloops only) Lower bound. 3314 KmpTaskTLowerBound, 3315 /// (Taskloops only) Upper bound. 3316 KmpTaskTUpperBound, 3317 /// (Taskloops only) Stride. 3318 KmpTaskTStride, 3319 /// (Taskloops only) Is last iteration flag. 3320 KmpTaskTLastIter, 3321 /// (Taskloops only) Reduction data. 3322 KmpTaskTReductions, 3323 }; 3324 } // anonymous namespace 3325 3326 bool CGOpenMPRuntime::OffloadEntriesInfoManagerTy::empty() const { 3327 // FIXME: Add other entries type when they become supported. 3328 return OffloadEntriesTargetRegion.empty(); 3329 } 3330 3331 /// \brief Initialize target region entry. 3332 void CGOpenMPRuntime::OffloadEntriesInfoManagerTy:: 3333 initializeTargetRegionEntryInfo(unsigned DeviceID, unsigned FileID, 3334 StringRef ParentName, unsigned LineNum, 3335 unsigned Order) { 3336 assert(CGM.getLangOpts().OpenMPIsDevice && "Initialization of entries is " 3337 "only required for the device " 3338 "code generation."); 3339 OffloadEntriesTargetRegion[DeviceID][FileID][ParentName][LineNum] = 3340 OffloadEntryInfoTargetRegion(Order, /*Addr=*/nullptr, /*ID=*/nullptr, 3341 /*Flags=*/0); 3342 ++OffloadingEntriesNum; 3343 } 3344 3345 void CGOpenMPRuntime::OffloadEntriesInfoManagerTy:: 3346 registerTargetRegionEntryInfo(unsigned DeviceID, unsigned FileID, 3347 StringRef ParentName, unsigned LineNum, 3348 llvm::Constant *Addr, llvm::Constant *ID, 3349 int32_t Flags) { 3350 // If we are emitting code for a target, the entry is already initialized, 3351 // only has to be registered. 3352 if (CGM.getLangOpts().OpenMPIsDevice) { 3353 assert(hasTargetRegionEntryInfo(DeviceID, FileID, ParentName, LineNum) && 3354 "Entry must exist."); 3355 auto &Entry = 3356 OffloadEntriesTargetRegion[DeviceID][FileID][ParentName][LineNum]; 3357 assert(Entry.isValid() && "Entry not initialized!"); 3358 Entry.setAddress(Addr); 3359 Entry.setID(ID); 3360 Entry.setFlags(Flags); 3361 return; 3362 } else { 3363 OffloadEntryInfoTargetRegion Entry(OffloadingEntriesNum++, Addr, ID, Flags); 3364 OffloadEntriesTargetRegion[DeviceID][FileID][ParentName][LineNum] = Entry; 3365 } 3366 } 3367 3368 bool CGOpenMPRuntime::OffloadEntriesInfoManagerTy::hasTargetRegionEntryInfo( 3369 unsigned DeviceID, unsigned FileID, StringRef ParentName, 3370 unsigned LineNum) const { 3371 auto PerDevice = OffloadEntriesTargetRegion.find(DeviceID); 3372 if (PerDevice == OffloadEntriesTargetRegion.end()) 3373 return false; 3374 auto PerFile = PerDevice->second.find(FileID); 3375 if (PerFile == PerDevice->second.end()) 3376 return false; 3377 auto PerParentName = PerFile->second.find(ParentName); 3378 if (PerParentName == PerFile->second.end()) 3379 return false; 3380 auto PerLine = PerParentName->second.find(LineNum); 3381 if (PerLine == PerParentName->second.end()) 3382 return false; 3383 // Fail if this entry is already registered. 3384 if (PerLine->second.getAddress() || PerLine->second.getID()) 3385 return false; 3386 return true; 3387 } 3388 3389 void CGOpenMPRuntime::OffloadEntriesInfoManagerTy::actOnTargetRegionEntriesInfo( 3390 const OffloadTargetRegionEntryInfoActTy &Action) { 3391 // Scan all target region entries and perform the provided action. 3392 for (auto &D : OffloadEntriesTargetRegion) 3393 for (auto &F : D.second) 3394 for (auto &P : F.second) 3395 for (auto &L : P.second) 3396 Action(D.first, F.first, P.first(), L.first, L.second); 3397 } 3398 3399 /// \brief Create a Ctor/Dtor-like function whose body is emitted through 3400 /// \a Codegen. This is used to emit the two functions that register and 3401 /// unregister the descriptor of the current compilation unit. 3402 static llvm::Function * 3403 createOffloadingBinaryDescriptorFunction(CodeGenModule &CGM, StringRef Name, 3404 const RegionCodeGenTy &Codegen) { 3405 auto &C = CGM.getContext(); 3406 FunctionArgList Args; 3407 ImplicitParamDecl DummyPtr(C, C.VoidPtrTy, ImplicitParamDecl::Other); 3408 Args.push_back(&DummyPtr); 3409 3410 CodeGenFunction CGF(CGM); 3411 // Disable debug info for global (de-)initializer because they are not part of 3412 // some particular construct. 3413 CGF.disableDebugInfo(); 3414 auto &FI = CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 3415 auto FTy = CGM.getTypes().GetFunctionType(FI); 3416 auto *Fn = CGM.CreateGlobalInitOrDestructFunction(FTy, Name, FI); 3417 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FI, Args); 3418 Codegen(CGF); 3419 CGF.FinishFunction(); 3420 return Fn; 3421 } 3422 3423 llvm::Function * 3424 CGOpenMPRuntime::createOffloadingBinaryDescriptorRegistration() { 3425 // If we don't have entries or if we are emitting code for the device, we 3426 // don't need to do anything. 3427 if (CGM.getLangOpts().OpenMPIsDevice || OffloadEntriesInfoManager.empty()) 3428 return nullptr; 3429 3430 auto &M = CGM.getModule(); 3431 auto &C = CGM.getContext(); 3432 3433 // Get list of devices we care about 3434 auto &Devices = CGM.getLangOpts().OMPTargetTriples; 3435 3436 // We should be creating an offloading descriptor only if there are devices 3437 // specified. 3438 assert(!Devices.empty() && "No OpenMP offloading devices??"); 3439 3440 // Create the external variables that will point to the begin and end of the 3441 // host entries section. These will be defined by the linker. 3442 auto *OffloadEntryTy = 3443 CGM.getTypes().ConvertTypeForMem(getTgtOffloadEntryQTy()); 3444 llvm::GlobalVariable *HostEntriesBegin = new llvm::GlobalVariable( 3445 M, OffloadEntryTy, /*isConstant=*/true, 3446 llvm::GlobalValue::ExternalLinkage, /*Initializer=*/nullptr, 3447 ".omp_offloading.entries_begin"); 3448 llvm::GlobalVariable *HostEntriesEnd = new llvm::GlobalVariable( 3449 M, OffloadEntryTy, /*isConstant=*/true, 3450 llvm::GlobalValue::ExternalLinkage, /*Initializer=*/nullptr, 3451 ".omp_offloading.entries_end"); 3452 3453 // Create all device images 3454 auto *DeviceImageTy = cast<llvm::StructType>( 3455 CGM.getTypes().ConvertTypeForMem(getTgtDeviceImageQTy())); 3456 ConstantInitBuilder DeviceImagesBuilder(CGM); 3457 auto DeviceImagesEntries = DeviceImagesBuilder.beginArray(DeviceImageTy); 3458 3459 for (unsigned i = 0; i < Devices.size(); ++i) { 3460 StringRef T = Devices[i].getTriple(); 3461 auto *ImgBegin = new llvm::GlobalVariable( 3462 M, CGM.Int8Ty, /*isConstant=*/true, llvm::GlobalValue::ExternalLinkage, 3463 /*Initializer=*/nullptr, 3464 Twine(".omp_offloading.img_start.") + Twine(T)); 3465 auto *ImgEnd = new llvm::GlobalVariable( 3466 M, CGM.Int8Ty, /*isConstant=*/true, llvm::GlobalValue::ExternalLinkage, 3467 /*Initializer=*/nullptr, Twine(".omp_offloading.img_end.") + Twine(T)); 3468 3469 auto Dev = DeviceImagesEntries.beginStruct(DeviceImageTy); 3470 Dev.add(ImgBegin); 3471 Dev.add(ImgEnd); 3472 Dev.add(HostEntriesBegin); 3473 Dev.add(HostEntriesEnd); 3474 Dev.finishAndAddTo(DeviceImagesEntries); 3475 } 3476 3477 // Create device images global array. 3478 llvm::GlobalVariable *DeviceImages = 3479 DeviceImagesEntries.finishAndCreateGlobal(".omp_offloading.device_images", 3480 CGM.getPointerAlign(), 3481 /*isConstant=*/true); 3482 DeviceImages->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global); 3483 3484 // This is a Zero array to be used in the creation of the constant expressions 3485 llvm::Constant *Index[] = {llvm::Constant::getNullValue(CGM.Int32Ty), 3486 llvm::Constant::getNullValue(CGM.Int32Ty)}; 3487 3488 // Create the target region descriptor. 3489 auto *BinaryDescriptorTy = cast<llvm::StructType>( 3490 CGM.getTypes().ConvertTypeForMem(getTgtBinaryDescriptorQTy())); 3491 ConstantInitBuilder DescBuilder(CGM); 3492 auto DescInit = DescBuilder.beginStruct(BinaryDescriptorTy); 3493 DescInit.addInt(CGM.Int32Ty, Devices.size()); 3494 DescInit.add(llvm::ConstantExpr::getGetElementPtr(DeviceImages->getValueType(), 3495 DeviceImages, 3496 Index)); 3497 DescInit.add(HostEntriesBegin); 3498 DescInit.add(HostEntriesEnd); 3499 3500 auto *Desc = DescInit.finishAndCreateGlobal(".omp_offloading.descriptor", 3501 CGM.getPointerAlign(), 3502 /*isConstant=*/true); 3503 3504 // Emit code to register or unregister the descriptor at execution 3505 // startup or closing, respectively. 3506 3507 // Create a variable to drive the registration and unregistration of the 3508 // descriptor, so we can reuse the logic that emits Ctors and Dtors. 3509 auto *IdentInfo = &C.Idents.get(".omp_offloading.reg_unreg_var"); 3510 ImplicitParamDecl RegUnregVar(C, C.getTranslationUnitDecl(), SourceLocation(), 3511 IdentInfo, C.CharTy, ImplicitParamDecl::Other); 3512 3513 auto *UnRegFn = createOffloadingBinaryDescriptorFunction( 3514 CGM, ".omp_offloading.descriptor_unreg", 3515 [&](CodeGenFunction &CGF, PrePostActionTy &) { 3516 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__tgt_unregister_lib), 3517 Desc); 3518 }); 3519 auto *RegFn = createOffloadingBinaryDescriptorFunction( 3520 CGM, ".omp_offloading.descriptor_reg", 3521 [&](CodeGenFunction &CGF, PrePostActionTy &) { 3522 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__tgt_register_lib), 3523 Desc); 3524 CGM.getCXXABI().registerGlobalDtor(CGF, RegUnregVar, UnRegFn, Desc); 3525 }); 3526 if (CGM.supportsCOMDAT()) { 3527 // It is sufficient to call registration function only once, so create a 3528 // COMDAT group for registration/unregistration functions and associated 3529 // data. That would reduce startup time and code size. Registration 3530 // function serves as a COMDAT group key. 3531 auto ComdatKey = M.getOrInsertComdat(RegFn->getName()); 3532 RegFn->setLinkage(llvm::GlobalValue::LinkOnceAnyLinkage); 3533 RegFn->setVisibility(llvm::GlobalValue::HiddenVisibility); 3534 RegFn->setComdat(ComdatKey); 3535 UnRegFn->setComdat(ComdatKey); 3536 DeviceImages->setComdat(ComdatKey); 3537 Desc->setComdat(ComdatKey); 3538 } 3539 return RegFn; 3540 } 3541 3542 void CGOpenMPRuntime::createOffloadEntry(llvm::Constant *ID, 3543 llvm::Constant *Addr, uint64_t Size, 3544 int32_t Flags) { 3545 StringRef Name = Addr->getName(); 3546 auto *TgtOffloadEntryType = cast<llvm::StructType>( 3547 CGM.getTypes().ConvertTypeForMem(getTgtOffloadEntryQTy())); 3548 llvm::LLVMContext &C = CGM.getModule().getContext(); 3549 llvm::Module &M = CGM.getModule(); 3550 3551 // Make sure the address has the right type. 3552 llvm::Constant *AddrPtr = llvm::ConstantExpr::getBitCast(ID, CGM.VoidPtrTy); 3553 3554 // Create constant string with the name. 3555 llvm::Constant *StrPtrInit = llvm::ConstantDataArray::getString(C, Name); 3556 3557 llvm::GlobalVariable *Str = 3558 new llvm::GlobalVariable(M, StrPtrInit->getType(), /*isConstant=*/true, 3559 llvm::GlobalValue::InternalLinkage, StrPtrInit, 3560 ".omp_offloading.entry_name"); 3561 Str->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global); 3562 llvm::Constant *StrPtr = llvm::ConstantExpr::getBitCast(Str, CGM.Int8PtrTy); 3563 3564 // We can't have any padding between symbols, so we need to have 1-byte 3565 // alignment. 3566 auto Align = CharUnits::fromQuantity(1); 3567 3568 // Create the entry struct. 3569 ConstantInitBuilder EntryBuilder(CGM); 3570 auto EntryInit = EntryBuilder.beginStruct(TgtOffloadEntryType); 3571 EntryInit.add(AddrPtr); 3572 EntryInit.add(StrPtr); 3573 EntryInit.addInt(CGM.SizeTy, Size); 3574 EntryInit.addInt(CGM.Int32Ty, Flags); 3575 EntryInit.addInt(CGM.Int32Ty, 0); 3576 llvm::GlobalVariable *Entry = EntryInit.finishAndCreateGlobal( 3577 Twine(".omp_offloading.entry.") + Name, Align, 3578 /*constant*/ true, llvm::GlobalValue::ExternalLinkage); 3579 3580 // The entry has to be created in the section the linker expects it to be. 3581 Entry->setSection(".omp_offloading.entries"); 3582 } 3583 3584 void CGOpenMPRuntime::createOffloadEntriesAndInfoMetadata() { 3585 // Emit the offloading entries and metadata so that the device codegen side 3586 // can easily figure out what to emit. The produced metadata looks like 3587 // this: 3588 // 3589 // !omp_offload.info = !{!1, ...} 3590 // 3591 // Right now we only generate metadata for function that contain target 3592 // regions. 3593 3594 // If we do not have entries, we dont need to do anything. 3595 if (OffloadEntriesInfoManager.empty()) 3596 return; 3597 3598 llvm::Module &M = CGM.getModule(); 3599 llvm::LLVMContext &C = M.getContext(); 3600 SmallVector<OffloadEntriesInfoManagerTy::OffloadEntryInfo *, 16> 3601 OrderedEntries(OffloadEntriesInfoManager.size()); 3602 3603 // Create the offloading info metadata node. 3604 llvm::NamedMDNode *MD = M.getOrInsertNamedMetadata("omp_offload.info"); 3605 3606 // Auxiliary methods to create metadata values and strings. 3607 auto getMDInt = [&](unsigned v) { 3608 return llvm::ConstantAsMetadata::get( 3609 llvm::ConstantInt::get(llvm::Type::getInt32Ty(C), v)); 3610 }; 3611 3612 auto getMDString = [&](StringRef v) { return llvm::MDString::get(C, v); }; 3613 3614 // Create function that emits metadata for each target region entry; 3615 auto &&TargetRegionMetadataEmitter = [&]( 3616 unsigned DeviceID, unsigned FileID, StringRef ParentName, unsigned Line, 3617 OffloadEntriesInfoManagerTy::OffloadEntryInfoTargetRegion &E) { 3618 llvm::SmallVector<llvm::Metadata *, 32> Ops; 3619 // Generate metadata for target regions. Each entry of this metadata 3620 // contains: 3621 // - Entry 0 -> Kind of this type of metadata (0). 3622 // - Entry 1 -> Device ID of the file where the entry was identified. 3623 // - Entry 2 -> File ID of the file where the entry was identified. 3624 // - Entry 3 -> Mangled name of the function where the entry was identified. 3625 // - Entry 4 -> Line in the file where the entry was identified. 3626 // - Entry 5 -> Order the entry was created. 3627 // The first element of the metadata node is the kind. 3628 Ops.push_back(getMDInt(E.getKind())); 3629 Ops.push_back(getMDInt(DeviceID)); 3630 Ops.push_back(getMDInt(FileID)); 3631 Ops.push_back(getMDString(ParentName)); 3632 Ops.push_back(getMDInt(Line)); 3633 Ops.push_back(getMDInt(E.getOrder())); 3634 3635 // Save this entry in the right position of the ordered entries array. 3636 OrderedEntries[E.getOrder()] = &E; 3637 3638 // Add metadata to the named metadata node. 3639 MD->addOperand(llvm::MDNode::get(C, Ops)); 3640 }; 3641 3642 OffloadEntriesInfoManager.actOnTargetRegionEntriesInfo( 3643 TargetRegionMetadataEmitter); 3644 3645 for (auto *E : OrderedEntries) { 3646 assert(E && "All ordered entries must exist!"); 3647 if (auto *CE = 3648 dyn_cast<OffloadEntriesInfoManagerTy::OffloadEntryInfoTargetRegion>( 3649 E)) { 3650 assert(CE->getID() && CE->getAddress() && 3651 "Entry ID and Addr are invalid!"); 3652 createOffloadEntry(CE->getID(), CE->getAddress(), /*Size=*/0); 3653 } else 3654 llvm_unreachable("Unsupported entry kind."); 3655 } 3656 } 3657 3658 /// \brief Loads all the offload entries information from the host IR 3659 /// metadata. 3660 void CGOpenMPRuntime::loadOffloadInfoMetadata() { 3661 // If we are in target mode, load the metadata from the host IR. This code has 3662 // to match the metadaata creation in createOffloadEntriesAndInfoMetadata(). 3663 3664 if (!CGM.getLangOpts().OpenMPIsDevice) 3665 return; 3666 3667 if (CGM.getLangOpts().OMPHostIRFile.empty()) 3668 return; 3669 3670 auto Buf = llvm::MemoryBuffer::getFile(CGM.getLangOpts().OMPHostIRFile); 3671 if (Buf.getError()) 3672 return; 3673 3674 llvm::LLVMContext C; 3675 auto ME = expectedToErrorOrAndEmitErrors( 3676 C, llvm::parseBitcodeFile(Buf.get()->getMemBufferRef(), C)); 3677 3678 if (ME.getError()) 3679 return; 3680 3681 llvm::NamedMDNode *MD = ME.get()->getNamedMetadata("omp_offload.info"); 3682 if (!MD) 3683 return; 3684 3685 for (auto I : MD->operands()) { 3686 llvm::MDNode *MN = cast<llvm::MDNode>(I); 3687 3688 auto getMDInt = [&](unsigned Idx) { 3689 llvm::ConstantAsMetadata *V = 3690 cast<llvm::ConstantAsMetadata>(MN->getOperand(Idx)); 3691 return cast<llvm::ConstantInt>(V->getValue())->getZExtValue(); 3692 }; 3693 3694 auto getMDString = [&](unsigned Idx) { 3695 llvm::MDString *V = cast<llvm::MDString>(MN->getOperand(Idx)); 3696 return V->getString(); 3697 }; 3698 3699 switch (getMDInt(0)) { 3700 default: 3701 llvm_unreachable("Unexpected metadata!"); 3702 break; 3703 case OffloadEntriesInfoManagerTy::OffloadEntryInfo:: 3704 OFFLOAD_ENTRY_INFO_TARGET_REGION: 3705 OffloadEntriesInfoManager.initializeTargetRegionEntryInfo( 3706 /*DeviceID=*/getMDInt(1), /*FileID=*/getMDInt(2), 3707 /*ParentName=*/getMDString(3), /*Line=*/getMDInt(4), 3708 /*Order=*/getMDInt(5)); 3709 break; 3710 } 3711 } 3712 } 3713 3714 void CGOpenMPRuntime::emitKmpRoutineEntryT(QualType KmpInt32Ty) { 3715 if (!KmpRoutineEntryPtrTy) { 3716 // Build typedef kmp_int32 (* kmp_routine_entry_t)(kmp_int32, void *); type. 3717 auto &C = CGM.getContext(); 3718 QualType KmpRoutineEntryTyArgs[] = {KmpInt32Ty, C.VoidPtrTy}; 3719 FunctionProtoType::ExtProtoInfo EPI; 3720 KmpRoutineEntryPtrQTy = C.getPointerType( 3721 C.getFunctionType(KmpInt32Ty, KmpRoutineEntryTyArgs, EPI)); 3722 KmpRoutineEntryPtrTy = CGM.getTypes().ConvertType(KmpRoutineEntryPtrQTy); 3723 } 3724 } 3725 3726 static FieldDecl *addFieldToRecordDecl(ASTContext &C, DeclContext *DC, 3727 QualType FieldTy) { 3728 auto *Field = FieldDecl::Create( 3729 C, DC, SourceLocation(), SourceLocation(), /*Id=*/nullptr, FieldTy, 3730 C.getTrivialTypeSourceInfo(FieldTy, SourceLocation()), 3731 /*BW=*/nullptr, /*Mutable=*/false, /*InitStyle=*/ICIS_NoInit); 3732 Field->setAccess(AS_public); 3733 DC->addDecl(Field); 3734 return Field; 3735 } 3736 3737 QualType CGOpenMPRuntime::getTgtOffloadEntryQTy() { 3738 3739 // Make sure the type of the entry is already created. This is the type we 3740 // have to create: 3741 // struct __tgt_offload_entry{ 3742 // void *addr; // Pointer to the offload entry info. 3743 // // (function or global) 3744 // char *name; // Name of the function or global. 3745 // size_t size; // Size of the entry info (0 if it a function). 3746 // int32_t flags; // Flags associated with the entry, e.g. 'link'. 3747 // int32_t reserved; // Reserved, to use by the runtime library. 3748 // }; 3749 if (TgtOffloadEntryQTy.isNull()) { 3750 ASTContext &C = CGM.getContext(); 3751 auto *RD = C.buildImplicitRecord("__tgt_offload_entry"); 3752 RD->startDefinition(); 3753 addFieldToRecordDecl(C, RD, C.VoidPtrTy); 3754 addFieldToRecordDecl(C, RD, C.getPointerType(C.CharTy)); 3755 addFieldToRecordDecl(C, RD, C.getSizeType()); 3756 addFieldToRecordDecl( 3757 C, RD, C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/true)); 3758 addFieldToRecordDecl( 3759 C, RD, C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/true)); 3760 RD->completeDefinition(); 3761 RD->addAttr(PackedAttr::CreateImplicit(C)); 3762 TgtOffloadEntryQTy = C.getRecordType(RD); 3763 } 3764 return TgtOffloadEntryQTy; 3765 } 3766 3767 QualType CGOpenMPRuntime::getTgtDeviceImageQTy() { 3768 // These are the types we need to build: 3769 // struct __tgt_device_image{ 3770 // void *ImageStart; // Pointer to the target code start. 3771 // void *ImageEnd; // Pointer to the target code end. 3772 // // We also add the host entries to the device image, as it may be useful 3773 // // for the target runtime to have access to that information. 3774 // __tgt_offload_entry *EntriesBegin; // Begin of the table with all 3775 // // the entries. 3776 // __tgt_offload_entry *EntriesEnd; // End of the table with all the 3777 // // entries (non inclusive). 3778 // }; 3779 if (TgtDeviceImageQTy.isNull()) { 3780 ASTContext &C = CGM.getContext(); 3781 auto *RD = C.buildImplicitRecord("__tgt_device_image"); 3782 RD->startDefinition(); 3783 addFieldToRecordDecl(C, RD, C.VoidPtrTy); 3784 addFieldToRecordDecl(C, RD, C.VoidPtrTy); 3785 addFieldToRecordDecl(C, RD, C.getPointerType(getTgtOffloadEntryQTy())); 3786 addFieldToRecordDecl(C, RD, C.getPointerType(getTgtOffloadEntryQTy())); 3787 RD->completeDefinition(); 3788 TgtDeviceImageQTy = C.getRecordType(RD); 3789 } 3790 return TgtDeviceImageQTy; 3791 } 3792 3793 QualType CGOpenMPRuntime::getTgtBinaryDescriptorQTy() { 3794 // struct __tgt_bin_desc{ 3795 // int32_t NumDevices; // Number of devices supported. 3796 // __tgt_device_image *DeviceImages; // Arrays of device images 3797 // // (one per device). 3798 // __tgt_offload_entry *EntriesBegin; // Begin of the table with all the 3799 // // entries. 3800 // __tgt_offload_entry *EntriesEnd; // End of the table with all the 3801 // // entries (non inclusive). 3802 // }; 3803 if (TgtBinaryDescriptorQTy.isNull()) { 3804 ASTContext &C = CGM.getContext(); 3805 auto *RD = C.buildImplicitRecord("__tgt_bin_desc"); 3806 RD->startDefinition(); 3807 addFieldToRecordDecl( 3808 C, RD, C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/true)); 3809 addFieldToRecordDecl(C, RD, C.getPointerType(getTgtDeviceImageQTy())); 3810 addFieldToRecordDecl(C, RD, C.getPointerType(getTgtOffloadEntryQTy())); 3811 addFieldToRecordDecl(C, RD, C.getPointerType(getTgtOffloadEntryQTy())); 3812 RD->completeDefinition(); 3813 TgtBinaryDescriptorQTy = C.getRecordType(RD); 3814 } 3815 return TgtBinaryDescriptorQTy; 3816 } 3817 3818 namespace { 3819 struct PrivateHelpersTy { 3820 PrivateHelpersTy(const VarDecl *Original, const VarDecl *PrivateCopy, 3821 const VarDecl *PrivateElemInit) 3822 : Original(Original), PrivateCopy(PrivateCopy), 3823 PrivateElemInit(PrivateElemInit) {} 3824 const VarDecl *Original; 3825 const VarDecl *PrivateCopy; 3826 const VarDecl *PrivateElemInit; 3827 }; 3828 typedef std::pair<CharUnits /*Align*/, PrivateHelpersTy> PrivateDataTy; 3829 } // anonymous namespace 3830 3831 static RecordDecl * 3832 createPrivatesRecordDecl(CodeGenModule &CGM, ArrayRef<PrivateDataTy> Privates) { 3833 if (!Privates.empty()) { 3834 auto &C = CGM.getContext(); 3835 // Build struct .kmp_privates_t. { 3836 // /* private vars */ 3837 // }; 3838 auto *RD = C.buildImplicitRecord(".kmp_privates.t"); 3839 RD->startDefinition(); 3840 for (auto &&Pair : Privates) { 3841 auto *VD = Pair.second.Original; 3842 auto Type = VD->getType(); 3843 Type = Type.getNonReferenceType(); 3844 auto *FD = addFieldToRecordDecl(C, RD, Type); 3845 if (VD->hasAttrs()) { 3846 for (specific_attr_iterator<AlignedAttr> I(VD->getAttrs().begin()), 3847 E(VD->getAttrs().end()); 3848 I != E; ++I) 3849 FD->addAttr(*I); 3850 } 3851 } 3852 RD->completeDefinition(); 3853 return RD; 3854 } 3855 return nullptr; 3856 } 3857 3858 static RecordDecl * 3859 createKmpTaskTRecordDecl(CodeGenModule &CGM, OpenMPDirectiveKind Kind, 3860 QualType KmpInt32Ty, 3861 QualType KmpRoutineEntryPointerQTy) { 3862 auto &C = CGM.getContext(); 3863 // Build struct kmp_task_t { 3864 // void * shareds; 3865 // kmp_routine_entry_t routine; 3866 // kmp_int32 part_id; 3867 // kmp_cmplrdata_t data1; 3868 // kmp_cmplrdata_t data2; 3869 // For taskloops additional fields: 3870 // kmp_uint64 lb; 3871 // kmp_uint64 ub; 3872 // kmp_int64 st; 3873 // kmp_int32 liter; 3874 // void * reductions; 3875 // }; 3876 auto *UD = C.buildImplicitRecord("kmp_cmplrdata_t", TTK_Union); 3877 UD->startDefinition(); 3878 addFieldToRecordDecl(C, UD, KmpInt32Ty); 3879 addFieldToRecordDecl(C, UD, KmpRoutineEntryPointerQTy); 3880 UD->completeDefinition(); 3881 QualType KmpCmplrdataTy = C.getRecordType(UD); 3882 auto *RD = C.buildImplicitRecord("kmp_task_t"); 3883 RD->startDefinition(); 3884 addFieldToRecordDecl(C, RD, C.VoidPtrTy); 3885 addFieldToRecordDecl(C, RD, KmpRoutineEntryPointerQTy); 3886 addFieldToRecordDecl(C, RD, KmpInt32Ty); 3887 addFieldToRecordDecl(C, RD, KmpCmplrdataTy); 3888 addFieldToRecordDecl(C, RD, KmpCmplrdataTy); 3889 if (isOpenMPTaskLoopDirective(Kind)) { 3890 QualType KmpUInt64Ty = 3891 CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/0); 3892 QualType KmpInt64Ty = 3893 CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/1); 3894 addFieldToRecordDecl(C, RD, KmpUInt64Ty); 3895 addFieldToRecordDecl(C, RD, KmpUInt64Ty); 3896 addFieldToRecordDecl(C, RD, KmpInt64Ty); 3897 addFieldToRecordDecl(C, RD, KmpInt32Ty); 3898 addFieldToRecordDecl(C, RD, C.VoidPtrTy); 3899 } 3900 RD->completeDefinition(); 3901 return RD; 3902 } 3903 3904 static RecordDecl * 3905 createKmpTaskTWithPrivatesRecordDecl(CodeGenModule &CGM, QualType KmpTaskTQTy, 3906 ArrayRef<PrivateDataTy> Privates) { 3907 auto &C = CGM.getContext(); 3908 // Build struct kmp_task_t_with_privates { 3909 // kmp_task_t task_data; 3910 // .kmp_privates_t. privates; 3911 // }; 3912 auto *RD = C.buildImplicitRecord("kmp_task_t_with_privates"); 3913 RD->startDefinition(); 3914 addFieldToRecordDecl(C, RD, KmpTaskTQTy); 3915 if (auto *PrivateRD = createPrivatesRecordDecl(CGM, Privates)) { 3916 addFieldToRecordDecl(C, RD, C.getRecordType(PrivateRD)); 3917 } 3918 RD->completeDefinition(); 3919 return RD; 3920 } 3921 3922 /// \brief Emit a proxy function which accepts kmp_task_t as the second 3923 /// argument. 3924 /// \code 3925 /// kmp_int32 .omp_task_entry.(kmp_int32 gtid, kmp_task_t *tt) { 3926 /// TaskFunction(gtid, tt->part_id, &tt->privates, task_privates_map, tt, 3927 /// For taskloops: 3928 /// tt->task_data.lb, tt->task_data.ub, tt->task_data.st, tt->task_data.liter, 3929 /// tt->reductions, tt->shareds); 3930 /// return 0; 3931 /// } 3932 /// \endcode 3933 static llvm::Value * 3934 emitProxyTaskFunction(CodeGenModule &CGM, SourceLocation Loc, 3935 OpenMPDirectiveKind Kind, QualType KmpInt32Ty, 3936 QualType KmpTaskTWithPrivatesPtrQTy, 3937 QualType KmpTaskTWithPrivatesQTy, QualType KmpTaskTQTy, 3938 QualType SharedsPtrTy, llvm::Value *TaskFunction, 3939 llvm::Value *TaskPrivatesMap) { 3940 auto &C = CGM.getContext(); 3941 FunctionArgList Args; 3942 ImplicitParamDecl GtidArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, KmpInt32Ty, 3943 ImplicitParamDecl::Other); 3944 ImplicitParamDecl TaskTypeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 3945 KmpTaskTWithPrivatesPtrQTy.withRestrict(), 3946 ImplicitParamDecl::Other); 3947 Args.push_back(&GtidArg); 3948 Args.push_back(&TaskTypeArg); 3949 auto &TaskEntryFnInfo = 3950 CGM.getTypes().arrangeBuiltinFunctionDeclaration(KmpInt32Ty, Args); 3951 auto *TaskEntryTy = CGM.getTypes().GetFunctionType(TaskEntryFnInfo); 3952 auto *TaskEntry = 3953 llvm::Function::Create(TaskEntryTy, llvm::GlobalValue::InternalLinkage, 3954 ".omp_task_entry.", &CGM.getModule()); 3955 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, TaskEntry, TaskEntryFnInfo); 3956 CodeGenFunction CGF(CGM); 3957 CGF.StartFunction(GlobalDecl(), KmpInt32Ty, TaskEntry, TaskEntryFnInfo, Args, 3958 Loc, Loc); 3959 3960 // TaskFunction(gtid, tt->task_data.part_id, &tt->privates, task_privates_map, 3961 // tt, 3962 // For taskloops: 3963 // tt->task_data.lb, tt->task_data.ub, tt->task_data.st, tt->task_data.liter, 3964 // tt->task_data.shareds); 3965 auto *GtidParam = CGF.EmitLoadOfScalar( 3966 CGF.GetAddrOfLocalVar(&GtidArg), /*Volatile=*/false, KmpInt32Ty, Loc); 3967 LValue TDBase = CGF.EmitLoadOfPointerLValue( 3968 CGF.GetAddrOfLocalVar(&TaskTypeArg), 3969 KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>()); 3970 auto *KmpTaskTWithPrivatesQTyRD = 3971 cast<RecordDecl>(KmpTaskTWithPrivatesQTy->getAsTagDecl()); 3972 LValue Base = 3973 CGF.EmitLValueForField(TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin()); 3974 auto *KmpTaskTQTyRD = cast<RecordDecl>(KmpTaskTQTy->getAsTagDecl()); 3975 auto PartIdFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTPartId); 3976 auto PartIdLVal = CGF.EmitLValueForField(Base, *PartIdFI); 3977 auto *PartidParam = PartIdLVal.getPointer(); 3978 3979 auto SharedsFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTShareds); 3980 auto SharedsLVal = CGF.EmitLValueForField(Base, *SharedsFI); 3981 auto *SharedsParam = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 3982 CGF.EmitLoadOfScalar(SharedsLVal, Loc), 3983 CGF.ConvertTypeForMem(SharedsPtrTy)); 3984 3985 auto PrivatesFI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin(), 1); 3986 llvm::Value *PrivatesParam; 3987 if (PrivatesFI != KmpTaskTWithPrivatesQTyRD->field_end()) { 3988 auto PrivatesLVal = CGF.EmitLValueForField(TDBase, *PrivatesFI); 3989 PrivatesParam = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 3990 PrivatesLVal.getPointer(), CGF.VoidPtrTy); 3991 } else 3992 PrivatesParam = llvm::ConstantPointerNull::get(CGF.VoidPtrTy); 3993 3994 llvm::Value *CommonArgs[] = {GtidParam, PartidParam, PrivatesParam, 3995 TaskPrivatesMap, 3996 CGF.Builder 3997 .CreatePointerBitCastOrAddrSpaceCast( 3998 TDBase.getAddress(), CGF.VoidPtrTy) 3999 .getPointer()}; 4000 SmallVector<llvm::Value *, 16> CallArgs(std::begin(CommonArgs), 4001 std::end(CommonArgs)); 4002 if (isOpenMPTaskLoopDirective(Kind)) { 4003 auto LBFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLowerBound); 4004 auto LBLVal = CGF.EmitLValueForField(Base, *LBFI); 4005 auto *LBParam = CGF.EmitLoadOfScalar(LBLVal, Loc); 4006 auto UBFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTUpperBound); 4007 auto UBLVal = CGF.EmitLValueForField(Base, *UBFI); 4008 auto *UBParam = CGF.EmitLoadOfScalar(UBLVal, Loc); 4009 auto StFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTStride); 4010 auto StLVal = CGF.EmitLValueForField(Base, *StFI); 4011 auto *StParam = CGF.EmitLoadOfScalar(StLVal, Loc); 4012 auto LIFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLastIter); 4013 auto LILVal = CGF.EmitLValueForField(Base, *LIFI); 4014 auto *LIParam = CGF.EmitLoadOfScalar(LILVal, Loc); 4015 auto RFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTReductions); 4016 auto RLVal = CGF.EmitLValueForField(Base, *RFI); 4017 auto *RParam = CGF.EmitLoadOfScalar(RLVal, Loc); 4018 CallArgs.push_back(LBParam); 4019 CallArgs.push_back(UBParam); 4020 CallArgs.push_back(StParam); 4021 CallArgs.push_back(LIParam); 4022 CallArgs.push_back(RParam); 4023 } 4024 CallArgs.push_back(SharedsParam); 4025 4026 CGM.getOpenMPRuntime().emitOutlinedFunctionCall(CGF, Loc, TaskFunction, 4027 CallArgs); 4028 CGF.EmitStoreThroughLValue( 4029 RValue::get(CGF.Builder.getInt32(/*C=*/0)), 4030 CGF.MakeAddrLValue(CGF.ReturnValue, KmpInt32Ty)); 4031 CGF.FinishFunction(); 4032 return TaskEntry; 4033 } 4034 4035 static llvm::Value *emitDestructorsFunction(CodeGenModule &CGM, 4036 SourceLocation Loc, 4037 QualType KmpInt32Ty, 4038 QualType KmpTaskTWithPrivatesPtrQTy, 4039 QualType KmpTaskTWithPrivatesQTy) { 4040 auto &C = CGM.getContext(); 4041 FunctionArgList Args; 4042 ImplicitParamDecl GtidArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, KmpInt32Ty, 4043 ImplicitParamDecl::Other); 4044 ImplicitParamDecl TaskTypeArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4045 KmpTaskTWithPrivatesPtrQTy.withRestrict(), 4046 ImplicitParamDecl::Other); 4047 Args.push_back(&GtidArg); 4048 Args.push_back(&TaskTypeArg); 4049 auto &DestructorFnInfo = 4050 CGM.getTypes().arrangeBuiltinFunctionDeclaration(KmpInt32Ty, Args); 4051 auto *DestructorFnTy = CGM.getTypes().GetFunctionType(DestructorFnInfo); 4052 auto *DestructorFn = 4053 llvm::Function::Create(DestructorFnTy, llvm::GlobalValue::InternalLinkage, 4054 ".omp_task_destructor.", &CGM.getModule()); 4055 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, DestructorFn, 4056 DestructorFnInfo); 4057 CodeGenFunction CGF(CGM); 4058 CGF.StartFunction(GlobalDecl(), KmpInt32Ty, DestructorFn, DestructorFnInfo, 4059 Args, Loc, Loc); 4060 4061 LValue Base = CGF.EmitLoadOfPointerLValue( 4062 CGF.GetAddrOfLocalVar(&TaskTypeArg), 4063 KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>()); 4064 auto *KmpTaskTWithPrivatesQTyRD = 4065 cast<RecordDecl>(KmpTaskTWithPrivatesQTy->getAsTagDecl()); 4066 auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin()); 4067 Base = CGF.EmitLValueForField(Base, *FI); 4068 for (auto *Field : 4069 cast<RecordDecl>(FI->getType()->getAsTagDecl())->fields()) { 4070 if (auto DtorKind = Field->getType().isDestructedType()) { 4071 auto FieldLValue = CGF.EmitLValueForField(Base, Field); 4072 CGF.pushDestroy(DtorKind, FieldLValue.getAddress(), Field->getType()); 4073 } 4074 } 4075 CGF.FinishFunction(); 4076 return DestructorFn; 4077 } 4078 4079 /// \brief Emit a privates mapping function for correct handling of private and 4080 /// firstprivate variables. 4081 /// \code 4082 /// void .omp_task_privates_map.(const .privates. *noalias privs, <ty1> 4083 /// **noalias priv1,..., <tyn> **noalias privn) { 4084 /// *priv1 = &.privates.priv1; 4085 /// ...; 4086 /// *privn = &.privates.privn; 4087 /// } 4088 /// \endcode 4089 static llvm::Value * 4090 emitTaskPrivateMappingFunction(CodeGenModule &CGM, SourceLocation Loc, 4091 ArrayRef<const Expr *> PrivateVars, 4092 ArrayRef<const Expr *> FirstprivateVars, 4093 ArrayRef<const Expr *> LastprivateVars, 4094 QualType PrivatesQTy, 4095 ArrayRef<PrivateDataTy> Privates) { 4096 auto &C = CGM.getContext(); 4097 FunctionArgList Args; 4098 ImplicitParamDecl TaskPrivatesArg( 4099 C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4100 C.getPointerType(PrivatesQTy).withConst().withRestrict(), 4101 ImplicitParamDecl::Other); 4102 Args.push_back(&TaskPrivatesArg); 4103 llvm::DenseMap<const VarDecl *, unsigned> PrivateVarsPos; 4104 unsigned Counter = 1; 4105 for (auto *E: PrivateVars) { 4106 Args.push_back(ImplicitParamDecl::Create( 4107 C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4108 C.getPointerType(C.getPointerType(E->getType())) 4109 .withConst() 4110 .withRestrict(), 4111 ImplicitParamDecl::Other)); 4112 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4113 PrivateVarsPos[VD] = Counter; 4114 ++Counter; 4115 } 4116 for (auto *E : FirstprivateVars) { 4117 Args.push_back(ImplicitParamDecl::Create( 4118 C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4119 C.getPointerType(C.getPointerType(E->getType())) 4120 .withConst() 4121 .withRestrict(), 4122 ImplicitParamDecl::Other)); 4123 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4124 PrivateVarsPos[VD] = Counter; 4125 ++Counter; 4126 } 4127 for (auto *E: LastprivateVars) { 4128 Args.push_back(ImplicitParamDecl::Create( 4129 C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4130 C.getPointerType(C.getPointerType(E->getType())) 4131 .withConst() 4132 .withRestrict(), 4133 ImplicitParamDecl::Other)); 4134 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4135 PrivateVarsPos[VD] = Counter; 4136 ++Counter; 4137 } 4138 auto &TaskPrivatesMapFnInfo = 4139 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 4140 auto *TaskPrivatesMapTy = 4141 CGM.getTypes().GetFunctionType(TaskPrivatesMapFnInfo); 4142 auto *TaskPrivatesMap = llvm::Function::Create( 4143 TaskPrivatesMapTy, llvm::GlobalValue::InternalLinkage, 4144 ".omp_task_privates_map.", &CGM.getModule()); 4145 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, TaskPrivatesMap, 4146 TaskPrivatesMapFnInfo); 4147 TaskPrivatesMap->removeFnAttr(llvm::Attribute::NoInline); 4148 TaskPrivatesMap->removeFnAttr(llvm::Attribute::OptimizeNone); 4149 TaskPrivatesMap->addFnAttr(llvm::Attribute::AlwaysInline); 4150 CodeGenFunction CGF(CGM); 4151 CGF.StartFunction(GlobalDecl(), C.VoidTy, TaskPrivatesMap, 4152 TaskPrivatesMapFnInfo, Args, Loc, Loc); 4153 4154 // *privi = &.privates.privi; 4155 LValue Base = CGF.EmitLoadOfPointerLValue( 4156 CGF.GetAddrOfLocalVar(&TaskPrivatesArg), 4157 TaskPrivatesArg.getType()->castAs<PointerType>()); 4158 auto *PrivatesQTyRD = cast<RecordDecl>(PrivatesQTy->getAsTagDecl()); 4159 Counter = 0; 4160 for (auto *Field : PrivatesQTyRD->fields()) { 4161 auto FieldLVal = CGF.EmitLValueForField(Base, Field); 4162 auto *VD = Args[PrivateVarsPos[Privates[Counter].second.Original]]; 4163 auto RefLVal = CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(VD), VD->getType()); 4164 auto RefLoadLVal = CGF.EmitLoadOfPointerLValue( 4165 RefLVal.getAddress(), RefLVal.getType()->castAs<PointerType>()); 4166 CGF.EmitStoreOfScalar(FieldLVal.getPointer(), RefLoadLVal); 4167 ++Counter; 4168 } 4169 CGF.FinishFunction(); 4170 return TaskPrivatesMap; 4171 } 4172 4173 static bool stable_sort_comparator(const PrivateDataTy P1, 4174 const PrivateDataTy P2) { 4175 return P1.first > P2.first; 4176 } 4177 4178 /// Emit initialization for private variables in task-based directives. 4179 static void emitPrivatesInit(CodeGenFunction &CGF, 4180 const OMPExecutableDirective &D, 4181 Address KmpTaskSharedsPtr, LValue TDBase, 4182 const RecordDecl *KmpTaskTWithPrivatesQTyRD, 4183 QualType SharedsTy, QualType SharedsPtrTy, 4184 const OMPTaskDataTy &Data, 4185 ArrayRef<PrivateDataTy> Privates, bool ForDup) { 4186 auto &C = CGF.getContext(); 4187 auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin()); 4188 LValue PrivatesBase = CGF.EmitLValueForField(TDBase, *FI); 4189 OpenMPDirectiveKind Kind = isOpenMPTaskLoopDirective(D.getDirectiveKind()) 4190 ? OMPD_taskloop 4191 : OMPD_task; 4192 const CapturedStmt &CS = *D.getCapturedStmt(Kind); 4193 CodeGenFunction::CGCapturedStmtInfo CapturesInfo(CS); 4194 LValue SrcBase; 4195 bool IsTargetTask = 4196 isOpenMPTargetDataManagementDirective(D.getDirectiveKind()) || 4197 isOpenMPTargetExecutionDirective(D.getDirectiveKind()); 4198 // For target-based directives skip 3 firstprivate arrays BasePointersArray, 4199 // PointersArray and SizesArray. The original variables for these arrays are 4200 // not captured and we get their addresses explicitly. 4201 if ((!IsTargetTask && !Data.FirstprivateVars.empty()) || 4202 (IsTargetTask && KmpTaskSharedsPtr.isValid())) { 4203 SrcBase = CGF.MakeAddrLValue( 4204 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4205 KmpTaskSharedsPtr, CGF.ConvertTypeForMem(SharedsPtrTy)), 4206 SharedsTy); 4207 } 4208 FI = cast<RecordDecl>(FI->getType()->getAsTagDecl())->field_begin(); 4209 for (auto &&Pair : Privates) { 4210 auto *VD = Pair.second.PrivateCopy; 4211 auto *Init = VD->getAnyInitializer(); 4212 if (Init && (!ForDup || (isa<CXXConstructExpr>(Init) && 4213 !CGF.isTrivialInitializer(Init)))) { 4214 LValue PrivateLValue = CGF.EmitLValueForField(PrivatesBase, *FI); 4215 if (auto *Elem = Pair.second.PrivateElemInit) { 4216 auto *OriginalVD = Pair.second.Original; 4217 // Check if the variable is the target-based BasePointersArray, 4218 // PointersArray or SizesArray. 4219 LValue SharedRefLValue; 4220 QualType Type = OriginalVD->getType(); 4221 auto *SharedField = CapturesInfo.lookup(OriginalVD); 4222 if (IsTargetTask && !SharedField) { 4223 assert(isa<ImplicitParamDecl>(OriginalVD) && 4224 isa<CapturedDecl>(OriginalVD->getDeclContext()) && 4225 cast<CapturedDecl>(OriginalVD->getDeclContext()) 4226 ->getNumParams() == 0 && 4227 isa<TranslationUnitDecl>( 4228 cast<CapturedDecl>(OriginalVD->getDeclContext()) 4229 ->getDeclContext()) && 4230 "Expected artificial target data variable."); 4231 SharedRefLValue = 4232 CGF.MakeAddrLValue(CGF.GetAddrOfLocalVar(OriginalVD), Type); 4233 } else { 4234 SharedRefLValue = CGF.EmitLValueForField(SrcBase, SharedField); 4235 SharedRefLValue = CGF.MakeAddrLValue( 4236 Address(SharedRefLValue.getPointer(), C.getDeclAlign(OriginalVD)), 4237 SharedRefLValue.getType(), LValueBaseInfo(AlignmentSource::Decl), 4238 SharedRefLValue.getTBAAInfo()); 4239 } 4240 if (Type->isArrayType()) { 4241 // Initialize firstprivate array. 4242 if (!isa<CXXConstructExpr>(Init) || CGF.isTrivialInitializer(Init)) { 4243 // Perform simple memcpy. 4244 CGF.EmitAggregateAssign(PrivateLValue.getAddress(), 4245 SharedRefLValue.getAddress(), Type); 4246 } else { 4247 // Initialize firstprivate array using element-by-element 4248 // initialization. 4249 CGF.EmitOMPAggregateAssign( 4250 PrivateLValue.getAddress(), SharedRefLValue.getAddress(), Type, 4251 [&CGF, Elem, Init, &CapturesInfo](Address DestElement, 4252 Address SrcElement) { 4253 // Clean up any temporaries needed by the initialization. 4254 CodeGenFunction::OMPPrivateScope InitScope(CGF); 4255 InitScope.addPrivate( 4256 Elem, [SrcElement]() -> Address { return SrcElement; }); 4257 (void)InitScope.Privatize(); 4258 // Emit initialization for single element. 4259 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII( 4260 CGF, &CapturesInfo); 4261 CGF.EmitAnyExprToMem(Init, DestElement, 4262 Init->getType().getQualifiers(), 4263 /*IsInitializer=*/false); 4264 }); 4265 } 4266 } else { 4267 CodeGenFunction::OMPPrivateScope InitScope(CGF); 4268 InitScope.addPrivate(Elem, [SharedRefLValue]() -> Address { 4269 return SharedRefLValue.getAddress(); 4270 }); 4271 (void)InitScope.Privatize(); 4272 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CapturesInfo); 4273 CGF.EmitExprAsInit(Init, VD, PrivateLValue, 4274 /*capturedByInit=*/false); 4275 } 4276 } else 4277 CGF.EmitExprAsInit(Init, VD, PrivateLValue, /*capturedByInit=*/false); 4278 } 4279 ++FI; 4280 } 4281 } 4282 4283 /// Check if duplication function is required for taskloops. 4284 static bool checkInitIsRequired(CodeGenFunction &CGF, 4285 ArrayRef<PrivateDataTy> Privates) { 4286 bool InitRequired = false; 4287 for (auto &&Pair : Privates) { 4288 auto *VD = Pair.second.PrivateCopy; 4289 auto *Init = VD->getAnyInitializer(); 4290 InitRequired = InitRequired || (Init && isa<CXXConstructExpr>(Init) && 4291 !CGF.isTrivialInitializer(Init)); 4292 } 4293 return InitRequired; 4294 } 4295 4296 4297 /// Emit task_dup function (for initialization of 4298 /// private/firstprivate/lastprivate vars and last_iter flag) 4299 /// \code 4300 /// void __task_dup_entry(kmp_task_t *task_dst, const kmp_task_t *task_src, int 4301 /// lastpriv) { 4302 /// // setup lastprivate flag 4303 /// task_dst->last = lastpriv; 4304 /// // could be constructor calls here... 4305 /// } 4306 /// \endcode 4307 static llvm::Value * 4308 emitTaskDupFunction(CodeGenModule &CGM, SourceLocation Loc, 4309 const OMPExecutableDirective &D, 4310 QualType KmpTaskTWithPrivatesPtrQTy, 4311 const RecordDecl *KmpTaskTWithPrivatesQTyRD, 4312 const RecordDecl *KmpTaskTQTyRD, QualType SharedsTy, 4313 QualType SharedsPtrTy, const OMPTaskDataTy &Data, 4314 ArrayRef<PrivateDataTy> Privates, bool WithLastIter) { 4315 auto &C = CGM.getContext(); 4316 FunctionArgList Args; 4317 ImplicitParamDecl DstArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4318 KmpTaskTWithPrivatesPtrQTy, 4319 ImplicitParamDecl::Other); 4320 ImplicitParamDecl SrcArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 4321 KmpTaskTWithPrivatesPtrQTy, 4322 ImplicitParamDecl::Other); 4323 ImplicitParamDecl LastprivArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.IntTy, 4324 ImplicitParamDecl::Other); 4325 Args.push_back(&DstArg); 4326 Args.push_back(&SrcArg); 4327 Args.push_back(&LastprivArg); 4328 auto &TaskDupFnInfo = 4329 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 4330 auto *TaskDupTy = CGM.getTypes().GetFunctionType(TaskDupFnInfo); 4331 auto *TaskDup = 4332 llvm::Function::Create(TaskDupTy, llvm::GlobalValue::InternalLinkage, 4333 ".omp_task_dup.", &CGM.getModule()); 4334 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, TaskDup, TaskDupFnInfo); 4335 CodeGenFunction CGF(CGM); 4336 CGF.StartFunction(GlobalDecl(), C.VoidTy, TaskDup, TaskDupFnInfo, Args, Loc, 4337 Loc); 4338 4339 LValue TDBase = CGF.EmitLoadOfPointerLValue( 4340 CGF.GetAddrOfLocalVar(&DstArg), 4341 KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>()); 4342 // task_dst->liter = lastpriv; 4343 if (WithLastIter) { 4344 auto LIFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTLastIter); 4345 LValue Base = CGF.EmitLValueForField( 4346 TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin()); 4347 LValue LILVal = CGF.EmitLValueForField(Base, *LIFI); 4348 llvm::Value *Lastpriv = CGF.EmitLoadOfScalar( 4349 CGF.GetAddrOfLocalVar(&LastprivArg), /*Volatile=*/false, C.IntTy, Loc); 4350 CGF.EmitStoreOfScalar(Lastpriv, LILVal); 4351 } 4352 4353 // Emit initial values for private copies (if any). 4354 assert(!Privates.empty()); 4355 Address KmpTaskSharedsPtr = Address::invalid(); 4356 if (!Data.FirstprivateVars.empty()) { 4357 LValue TDBase = CGF.EmitLoadOfPointerLValue( 4358 CGF.GetAddrOfLocalVar(&SrcArg), 4359 KmpTaskTWithPrivatesPtrQTy->castAs<PointerType>()); 4360 LValue Base = CGF.EmitLValueForField( 4361 TDBase, *KmpTaskTWithPrivatesQTyRD->field_begin()); 4362 KmpTaskSharedsPtr = Address( 4363 CGF.EmitLoadOfScalar(CGF.EmitLValueForField( 4364 Base, *std::next(KmpTaskTQTyRD->field_begin(), 4365 KmpTaskTShareds)), 4366 Loc), 4367 CGF.getNaturalTypeAlignment(SharedsTy)); 4368 } 4369 emitPrivatesInit(CGF, D, KmpTaskSharedsPtr, TDBase, KmpTaskTWithPrivatesQTyRD, 4370 SharedsTy, SharedsPtrTy, Data, Privates, /*ForDup=*/true); 4371 CGF.FinishFunction(); 4372 return TaskDup; 4373 } 4374 4375 /// Checks if destructor function is required to be generated. 4376 /// \return true if cleanups are required, false otherwise. 4377 static bool 4378 checkDestructorsRequired(const RecordDecl *KmpTaskTWithPrivatesQTyRD) { 4379 bool NeedsCleanup = false; 4380 auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin()); 4381 auto *PrivateRD = cast<RecordDecl>(FI->getType()->getAsTagDecl()); 4382 for (auto *FD : PrivateRD->fields()) { 4383 NeedsCleanup = NeedsCleanup || FD->getType().isDestructedType(); 4384 if (NeedsCleanup) 4385 break; 4386 } 4387 return NeedsCleanup; 4388 } 4389 4390 CGOpenMPRuntime::TaskResultTy 4391 CGOpenMPRuntime::emitTaskInit(CodeGenFunction &CGF, SourceLocation Loc, 4392 const OMPExecutableDirective &D, 4393 llvm::Value *TaskFunction, QualType SharedsTy, 4394 Address Shareds, const OMPTaskDataTy &Data) { 4395 auto &C = CGM.getContext(); 4396 llvm::SmallVector<PrivateDataTy, 4> Privates; 4397 // Aggregate privates and sort them by the alignment. 4398 auto I = Data.PrivateCopies.begin(); 4399 for (auto *E : Data.PrivateVars) { 4400 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4401 Privates.push_back(std::make_pair( 4402 C.getDeclAlign(VD), 4403 PrivateHelpersTy(VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()), 4404 /*PrivateElemInit=*/nullptr))); 4405 ++I; 4406 } 4407 I = Data.FirstprivateCopies.begin(); 4408 auto IElemInitRef = Data.FirstprivateInits.begin(); 4409 for (auto *E : Data.FirstprivateVars) { 4410 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4411 Privates.push_back(std::make_pair( 4412 C.getDeclAlign(VD), 4413 PrivateHelpersTy( 4414 VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()), 4415 cast<VarDecl>(cast<DeclRefExpr>(*IElemInitRef)->getDecl())))); 4416 ++I; 4417 ++IElemInitRef; 4418 } 4419 I = Data.LastprivateCopies.begin(); 4420 for (auto *E : Data.LastprivateVars) { 4421 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(E)->getDecl()); 4422 Privates.push_back(std::make_pair( 4423 C.getDeclAlign(VD), 4424 PrivateHelpersTy(VD, cast<VarDecl>(cast<DeclRefExpr>(*I)->getDecl()), 4425 /*PrivateElemInit=*/nullptr))); 4426 ++I; 4427 } 4428 std::stable_sort(Privates.begin(), Privates.end(), stable_sort_comparator); 4429 auto KmpInt32Ty = C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/1); 4430 // Build type kmp_routine_entry_t (if not built yet). 4431 emitKmpRoutineEntryT(KmpInt32Ty); 4432 // Build type kmp_task_t (if not built yet). 4433 if (isOpenMPTaskLoopDirective(D.getDirectiveKind())) { 4434 if (SavedKmpTaskloopTQTy.isNull()) { 4435 SavedKmpTaskloopTQTy = C.getRecordType(createKmpTaskTRecordDecl( 4436 CGM, D.getDirectiveKind(), KmpInt32Ty, KmpRoutineEntryPtrQTy)); 4437 } 4438 KmpTaskTQTy = SavedKmpTaskloopTQTy; 4439 } else { 4440 assert((D.getDirectiveKind() == OMPD_task || 4441 isOpenMPTargetExecutionDirective(D.getDirectiveKind()) || 4442 isOpenMPTargetDataManagementDirective(D.getDirectiveKind())) && 4443 "Expected taskloop, task or target directive"); 4444 if (SavedKmpTaskTQTy.isNull()) { 4445 SavedKmpTaskTQTy = C.getRecordType(createKmpTaskTRecordDecl( 4446 CGM, D.getDirectiveKind(), KmpInt32Ty, KmpRoutineEntryPtrQTy)); 4447 } 4448 KmpTaskTQTy = SavedKmpTaskTQTy; 4449 } 4450 auto *KmpTaskTQTyRD = cast<RecordDecl>(KmpTaskTQTy->getAsTagDecl()); 4451 // Build particular struct kmp_task_t for the given task. 4452 auto *KmpTaskTWithPrivatesQTyRD = 4453 createKmpTaskTWithPrivatesRecordDecl(CGM, KmpTaskTQTy, Privates); 4454 auto KmpTaskTWithPrivatesQTy = C.getRecordType(KmpTaskTWithPrivatesQTyRD); 4455 QualType KmpTaskTWithPrivatesPtrQTy = 4456 C.getPointerType(KmpTaskTWithPrivatesQTy); 4457 auto *KmpTaskTWithPrivatesTy = CGF.ConvertType(KmpTaskTWithPrivatesQTy); 4458 auto *KmpTaskTWithPrivatesPtrTy = KmpTaskTWithPrivatesTy->getPointerTo(); 4459 auto *KmpTaskTWithPrivatesTySize = CGF.getTypeSize(KmpTaskTWithPrivatesQTy); 4460 QualType SharedsPtrTy = C.getPointerType(SharedsTy); 4461 4462 // Emit initial values for private copies (if any). 4463 llvm::Value *TaskPrivatesMap = nullptr; 4464 auto *TaskPrivatesMapTy = 4465 std::next(cast<llvm::Function>(TaskFunction)->arg_begin(), 3)->getType(); 4466 if (!Privates.empty()) { 4467 auto FI = std::next(KmpTaskTWithPrivatesQTyRD->field_begin()); 4468 TaskPrivatesMap = emitTaskPrivateMappingFunction( 4469 CGM, Loc, Data.PrivateVars, Data.FirstprivateVars, Data.LastprivateVars, 4470 FI->getType(), Privates); 4471 TaskPrivatesMap = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4472 TaskPrivatesMap, TaskPrivatesMapTy); 4473 } else { 4474 TaskPrivatesMap = llvm::ConstantPointerNull::get( 4475 cast<llvm::PointerType>(TaskPrivatesMapTy)); 4476 } 4477 // Build a proxy function kmp_int32 .omp_task_entry.(kmp_int32 gtid, 4478 // kmp_task_t *tt); 4479 auto *TaskEntry = emitProxyTaskFunction( 4480 CGM, Loc, D.getDirectiveKind(), KmpInt32Ty, KmpTaskTWithPrivatesPtrQTy, 4481 KmpTaskTWithPrivatesQTy, KmpTaskTQTy, SharedsPtrTy, TaskFunction, 4482 TaskPrivatesMap); 4483 4484 // Build call kmp_task_t * __kmpc_omp_task_alloc(ident_t *, kmp_int32 gtid, 4485 // kmp_int32 flags, size_t sizeof_kmp_task_t, size_t sizeof_shareds, 4486 // kmp_routine_entry_t *task_entry); 4487 // Task flags. Format is taken from 4488 // http://llvm.org/svn/llvm-project/openmp/trunk/runtime/src/kmp.h, 4489 // description of kmp_tasking_flags struct. 4490 enum { 4491 TiedFlag = 0x1, 4492 FinalFlag = 0x2, 4493 DestructorsFlag = 0x8, 4494 PriorityFlag = 0x20 4495 }; 4496 unsigned Flags = Data.Tied ? TiedFlag : 0; 4497 bool NeedsCleanup = false; 4498 if (!Privates.empty()) { 4499 NeedsCleanup = checkDestructorsRequired(KmpTaskTWithPrivatesQTyRD); 4500 if (NeedsCleanup) 4501 Flags = Flags | DestructorsFlag; 4502 } 4503 if (Data.Priority.getInt()) 4504 Flags = Flags | PriorityFlag; 4505 auto *TaskFlags = 4506 Data.Final.getPointer() 4507 ? CGF.Builder.CreateSelect(Data.Final.getPointer(), 4508 CGF.Builder.getInt32(FinalFlag), 4509 CGF.Builder.getInt32(/*C=*/0)) 4510 : CGF.Builder.getInt32(Data.Final.getInt() ? FinalFlag : 0); 4511 TaskFlags = CGF.Builder.CreateOr(TaskFlags, CGF.Builder.getInt32(Flags)); 4512 auto *SharedsSize = CGM.getSize(C.getTypeSizeInChars(SharedsTy)); 4513 llvm::Value *AllocArgs[] = {emitUpdateLocation(CGF, Loc), 4514 getThreadID(CGF, Loc), TaskFlags, 4515 KmpTaskTWithPrivatesTySize, SharedsSize, 4516 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4517 TaskEntry, KmpRoutineEntryPtrTy)}; 4518 auto *NewTask = CGF.EmitRuntimeCall( 4519 createRuntimeFunction(OMPRTL__kmpc_omp_task_alloc), AllocArgs); 4520 auto *NewTaskNewTaskTTy = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4521 NewTask, KmpTaskTWithPrivatesPtrTy); 4522 LValue Base = CGF.MakeNaturalAlignAddrLValue(NewTaskNewTaskTTy, 4523 KmpTaskTWithPrivatesQTy); 4524 LValue TDBase = 4525 CGF.EmitLValueForField(Base, *KmpTaskTWithPrivatesQTyRD->field_begin()); 4526 // Fill the data in the resulting kmp_task_t record. 4527 // Copy shareds if there are any. 4528 Address KmpTaskSharedsPtr = Address::invalid(); 4529 if (!SharedsTy->getAsStructureType()->getDecl()->field_empty()) { 4530 KmpTaskSharedsPtr = 4531 Address(CGF.EmitLoadOfScalar( 4532 CGF.EmitLValueForField( 4533 TDBase, *std::next(KmpTaskTQTyRD->field_begin(), 4534 KmpTaskTShareds)), 4535 Loc), 4536 CGF.getNaturalTypeAlignment(SharedsTy)); 4537 CGF.EmitAggregateCopy(KmpTaskSharedsPtr, Shareds, SharedsTy); 4538 } 4539 // Emit initial values for private copies (if any). 4540 TaskResultTy Result; 4541 if (!Privates.empty()) { 4542 emitPrivatesInit(CGF, D, KmpTaskSharedsPtr, Base, KmpTaskTWithPrivatesQTyRD, 4543 SharedsTy, SharedsPtrTy, Data, Privates, 4544 /*ForDup=*/false); 4545 if (isOpenMPTaskLoopDirective(D.getDirectiveKind()) && 4546 (!Data.LastprivateVars.empty() || checkInitIsRequired(CGF, Privates))) { 4547 Result.TaskDupFn = emitTaskDupFunction( 4548 CGM, Loc, D, KmpTaskTWithPrivatesPtrQTy, KmpTaskTWithPrivatesQTyRD, 4549 KmpTaskTQTyRD, SharedsTy, SharedsPtrTy, Data, Privates, 4550 /*WithLastIter=*/!Data.LastprivateVars.empty()); 4551 } 4552 } 4553 // Fields of union "kmp_cmplrdata_t" for destructors and priority. 4554 enum { Priority = 0, Destructors = 1 }; 4555 // Provide pointer to function with destructors for privates. 4556 auto FI = std::next(KmpTaskTQTyRD->field_begin(), Data1); 4557 auto *KmpCmplrdataUD = (*FI)->getType()->getAsUnionType()->getDecl(); 4558 if (NeedsCleanup) { 4559 llvm::Value *DestructorFn = emitDestructorsFunction( 4560 CGM, Loc, KmpInt32Ty, KmpTaskTWithPrivatesPtrQTy, 4561 KmpTaskTWithPrivatesQTy); 4562 LValue Data1LV = CGF.EmitLValueForField(TDBase, *FI); 4563 LValue DestructorsLV = CGF.EmitLValueForField( 4564 Data1LV, *std::next(KmpCmplrdataUD->field_begin(), Destructors)); 4565 CGF.EmitStoreOfScalar(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4566 DestructorFn, KmpRoutineEntryPtrTy), 4567 DestructorsLV); 4568 } 4569 // Set priority. 4570 if (Data.Priority.getInt()) { 4571 LValue Data2LV = CGF.EmitLValueForField( 4572 TDBase, *std::next(KmpTaskTQTyRD->field_begin(), Data2)); 4573 LValue PriorityLV = CGF.EmitLValueForField( 4574 Data2LV, *std::next(KmpCmplrdataUD->field_begin(), Priority)); 4575 CGF.EmitStoreOfScalar(Data.Priority.getPointer(), PriorityLV); 4576 } 4577 Result.NewTask = NewTask; 4578 Result.TaskEntry = TaskEntry; 4579 Result.NewTaskNewTaskTTy = NewTaskNewTaskTTy; 4580 Result.TDBase = TDBase; 4581 Result.KmpTaskTQTyRD = KmpTaskTQTyRD; 4582 return Result; 4583 } 4584 4585 void CGOpenMPRuntime::emitTaskCall(CodeGenFunction &CGF, SourceLocation Loc, 4586 const OMPExecutableDirective &D, 4587 llvm::Value *TaskFunction, 4588 QualType SharedsTy, Address Shareds, 4589 const Expr *IfCond, 4590 const OMPTaskDataTy &Data) { 4591 if (!CGF.HaveInsertPoint()) 4592 return; 4593 4594 TaskResultTy Result = 4595 emitTaskInit(CGF, Loc, D, TaskFunction, SharedsTy, Shareds, Data); 4596 llvm::Value *NewTask = Result.NewTask; 4597 llvm::Value *TaskEntry = Result.TaskEntry; 4598 llvm::Value *NewTaskNewTaskTTy = Result.NewTaskNewTaskTTy; 4599 LValue TDBase = Result.TDBase; 4600 RecordDecl *KmpTaskTQTyRD = Result.KmpTaskTQTyRD; 4601 auto &C = CGM.getContext(); 4602 // Process list of dependences. 4603 Address DependenciesArray = Address::invalid(); 4604 unsigned NumDependencies = Data.Dependences.size(); 4605 if (NumDependencies) { 4606 // Dependence kind for RTL. 4607 enum RTLDependenceKindTy { DepIn = 0x01, DepInOut = 0x3 }; 4608 enum RTLDependInfoFieldsTy { BaseAddr, Len, Flags }; 4609 RecordDecl *KmpDependInfoRD; 4610 QualType FlagsTy = 4611 C.getIntTypeForBitwidth(C.getTypeSize(C.BoolTy), /*Signed=*/false); 4612 llvm::Type *LLVMFlagsTy = CGF.ConvertTypeForMem(FlagsTy); 4613 if (KmpDependInfoTy.isNull()) { 4614 KmpDependInfoRD = C.buildImplicitRecord("kmp_depend_info"); 4615 KmpDependInfoRD->startDefinition(); 4616 addFieldToRecordDecl(C, KmpDependInfoRD, C.getIntPtrType()); 4617 addFieldToRecordDecl(C, KmpDependInfoRD, C.getSizeType()); 4618 addFieldToRecordDecl(C, KmpDependInfoRD, FlagsTy); 4619 KmpDependInfoRD->completeDefinition(); 4620 KmpDependInfoTy = C.getRecordType(KmpDependInfoRD); 4621 } else 4622 KmpDependInfoRD = cast<RecordDecl>(KmpDependInfoTy->getAsTagDecl()); 4623 CharUnits DependencySize = C.getTypeSizeInChars(KmpDependInfoTy); 4624 // Define type kmp_depend_info[<Dependences.size()>]; 4625 QualType KmpDependInfoArrayTy = C.getConstantArrayType( 4626 KmpDependInfoTy, llvm::APInt(/*numBits=*/64, NumDependencies), 4627 ArrayType::Normal, /*IndexTypeQuals=*/0); 4628 // kmp_depend_info[<Dependences.size()>] deps; 4629 DependenciesArray = 4630 CGF.CreateMemTemp(KmpDependInfoArrayTy, ".dep.arr.addr"); 4631 for (unsigned i = 0; i < NumDependencies; ++i) { 4632 const Expr *E = Data.Dependences[i].second; 4633 auto Addr = CGF.EmitLValue(E); 4634 llvm::Value *Size; 4635 QualType Ty = E->getType(); 4636 if (auto *ASE = dyn_cast<OMPArraySectionExpr>(E->IgnoreParenImpCasts())) { 4637 LValue UpAddrLVal = 4638 CGF.EmitOMPArraySectionExpr(ASE, /*LowerBound=*/false); 4639 llvm::Value *UpAddr = 4640 CGF.Builder.CreateConstGEP1_32(UpAddrLVal.getPointer(), /*Idx0=*/1); 4641 llvm::Value *LowIntPtr = 4642 CGF.Builder.CreatePtrToInt(Addr.getPointer(), CGM.SizeTy); 4643 llvm::Value *UpIntPtr = CGF.Builder.CreatePtrToInt(UpAddr, CGM.SizeTy); 4644 Size = CGF.Builder.CreateNUWSub(UpIntPtr, LowIntPtr); 4645 } else 4646 Size = CGF.getTypeSize(Ty); 4647 auto Base = CGF.MakeAddrLValue( 4648 CGF.Builder.CreateConstArrayGEP(DependenciesArray, i, DependencySize), 4649 KmpDependInfoTy); 4650 // deps[i].base_addr = &<Dependences[i].second>; 4651 auto BaseAddrLVal = CGF.EmitLValueForField( 4652 Base, *std::next(KmpDependInfoRD->field_begin(), BaseAddr)); 4653 CGF.EmitStoreOfScalar( 4654 CGF.Builder.CreatePtrToInt(Addr.getPointer(), CGF.IntPtrTy), 4655 BaseAddrLVal); 4656 // deps[i].len = sizeof(<Dependences[i].second>); 4657 auto LenLVal = CGF.EmitLValueForField( 4658 Base, *std::next(KmpDependInfoRD->field_begin(), Len)); 4659 CGF.EmitStoreOfScalar(Size, LenLVal); 4660 // deps[i].flags = <Dependences[i].first>; 4661 RTLDependenceKindTy DepKind; 4662 switch (Data.Dependences[i].first) { 4663 case OMPC_DEPEND_in: 4664 DepKind = DepIn; 4665 break; 4666 // Out and InOut dependencies must use the same code. 4667 case OMPC_DEPEND_out: 4668 case OMPC_DEPEND_inout: 4669 DepKind = DepInOut; 4670 break; 4671 case OMPC_DEPEND_source: 4672 case OMPC_DEPEND_sink: 4673 case OMPC_DEPEND_unknown: 4674 llvm_unreachable("Unknown task dependence type"); 4675 } 4676 auto FlagsLVal = CGF.EmitLValueForField( 4677 Base, *std::next(KmpDependInfoRD->field_begin(), Flags)); 4678 CGF.EmitStoreOfScalar(llvm::ConstantInt::get(LLVMFlagsTy, DepKind), 4679 FlagsLVal); 4680 } 4681 DependenciesArray = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4682 CGF.Builder.CreateStructGEP(DependenciesArray, 0, CharUnits::Zero()), 4683 CGF.VoidPtrTy); 4684 } 4685 4686 // NOTE: routine and part_id fields are intialized by __kmpc_omp_task_alloc() 4687 // libcall. 4688 // Build kmp_int32 __kmpc_omp_task_with_deps(ident_t *, kmp_int32 gtid, 4689 // kmp_task_t *new_task, kmp_int32 ndeps, kmp_depend_info_t *dep_list, 4690 // kmp_int32 ndeps_noalias, kmp_depend_info_t *noalias_dep_list) if dependence 4691 // list is not empty 4692 auto *ThreadID = getThreadID(CGF, Loc); 4693 auto *UpLoc = emitUpdateLocation(CGF, Loc); 4694 llvm::Value *TaskArgs[] = { UpLoc, ThreadID, NewTask }; 4695 llvm::Value *DepTaskArgs[7]; 4696 if (NumDependencies) { 4697 DepTaskArgs[0] = UpLoc; 4698 DepTaskArgs[1] = ThreadID; 4699 DepTaskArgs[2] = NewTask; 4700 DepTaskArgs[3] = CGF.Builder.getInt32(NumDependencies); 4701 DepTaskArgs[4] = DependenciesArray.getPointer(); 4702 DepTaskArgs[5] = CGF.Builder.getInt32(0); 4703 DepTaskArgs[6] = llvm::ConstantPointerNull::get(CGF.VoidPtrTy); 4704 } 4705 auto &&ThenCodeGen = [this, &Data, TDBase, KmpTaskTQTyRD, NumDependencies, 4706 &TaskArgs, 4707 &DepTaskArgs](CodeGenFunction &CGF, PrePostActionTy &) { 4708 if (!Data.Tied) { 4709 auto PartIdFI = std::next(KmpTaskTQTyRD->field_begin(), KmpTaskTPartId); 4710 auto PartIdLVal = CGF.EmitLValueForField(TDBase, *PartIdFI); 4711 CGF.EmitStoreOfScalar(CGF.Builder.getInt32(0), PartIdLVal); 4712 } 4713 if (NumDependencies) { 4714 CGF.EmitRuntimeCall( 4715 createRuntimeFunction(OMPRTL__kmpc_omp_task_with_deps), DepTaskArgs); 4716 } else { 4717 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_omp_task), 4718 TaskArgs); 4719 } 4720 // Check if parent region is untied and build return for untied task; 4721 if (auto *Region = 4722 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) 4723 Region->emitUntiedSwitch(CGF); 4724 }; 4725 4726 llvm::Value *DepWaitTaskArgs[6]; 4727 if (NumDependencies) { 4728 DepWaitTaskArgs[0] = UpLoc; 4729 DepWaitTaskArgs[1] = ThreadID; 4730 DepWaitTaskArgs[2] = CGF.Builder.getInt32(NumDependencies); 4731 DepWaitTaskArgs[3] = DependenciesArray.getPointer(); 4732 DepWaitTaskArgs[4] = CGF.Builder.getInt32(0); 4733 DepWaitTaskArgs[5] = llvm::ConstantPointerNull::get(CGF.VoidPtrTy); 4734 } 4735 auto &&ElseCodeGen = [&TaskArgs, ThreadID, NewTaskNewTaskTTy, TaskEntry, 4736 NumDependencies, &DepWaitTaskArgs, 4737 Loc](CodeGenFunction &CGF, PrePostActionTy &) { 4738 auto &RT = CGF.CGM.getOpenMPRuntime(); 4739 CodeGenFunction::RunCleanupsScope LocalScope(CGF); 4740 // Build void __kmpc_omp_wait_deps(ident_t *, kmp_int32 gtid, 4741 // kmp_int32 ndeps, kmp_depend_info_t *dep_list, kmp_int32 4742 // ndeps_noalias, kmp_depend_info_t *noalias_dep_list); if dependence info 4743 // is specified. 4744 if (NumDependencies) 4745 CGF.EmitRuntimeCall(RT.createRuntimeFunction(OMPRTL__kmpc_omp_wait_deps), 4746 DepWaitTaskArgs); 4747 // Call proxy_task_entry(gtid, new_task); 4748 auto &&CodeGen = [TaskEntry, ThreadID, NewTaskNewTaskTTy, 4749 Loc](CodeGenFunction &CGF, PrePostActionTy &Action) { 4750 Action.Enter(CGF); 4751 llvm::Value *OutlinedFnArgs[] = {ThreadID, NewTaskNewTaskTTy}; 4752 CGF.CGM.getOpenMPRuntime().emitOutlinedFunctionCall(CGF, Loc, TaskEntry, 4753 OutlinedFnArgs); 4754 }; 4755 4756 // Build void __kmpc_omp_task_begin_if0(ident_t *, kmp_int32 gtid, 4757 // kmp_task_t *new_task); 4758 // Build void __kmpc_omp_task_complete_if0(ident_t *, kmp_int32 gtid, 4759 // kmp_task_t *new_task); 4760 RegionCodeGenTy RCG(CodeGen); 4761 CommonActionTy Action( 4762 RT.createRuntimeFunction(OMPRTL__kmpc_omp_task_begin_if0), TaskArgs, 4763 RT.createRuntimeFunction(OMPRTL__kmpc_omp_task_complete_if0), TaskArgs); 4764 RCG.setAction(Action); 4765 RCG(CGF); 4766 }; 4767 4768 if (IfCond) 4769 emitOMPIfClause(CGF, IfCond, ThenCodeGen, ElseCodeGen); 4770 else { 4771 RegionCodeGenTy ThenRCG(ThenCodeGen); 4772 ThenRCG(CGF); 4773 } 4774 } 4775 4776 void CGOpenMPRuntime::emitTaskLoopCall(CodeGenFunction &CGF, SourceLocation Loc, 4777 const OMPLoopDirective &D, 4778 llvm::Value *TaskFunction, 4779 QualType SharedsTy, Address Shareds, 4780 const Expr *IfCond, 4781 const OMPTaskDataTy &Data) { 4782 if (!CGF.HaveInsertPoint()) 4783 return; 4784 TaskResultTy Result = 4785 emitTaskInit(CGF, Loc, D, TaskFunction, SharedsTy, Shareds, Data); 4786 // NOTE: routine and part_id fields are intialized by __kmpc_omp_task_alloc() 4787 // libcall. 4788 // Call to void __kmpc_taskloop(ident_t *loc, int gtid, kmp_task_t *task, int 4789 // if_val, kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup, int 4790 // sched, kmp_uint64 grainsize, void *task_dup); 4791 llvm::Value *ThreadID = getThreadID(CGF, Loc); 4792 llvm::Value *UpLoc = emitUpdateLocation(CGF, Loc); 4793 llvm::Value *IfVal; 4794 if (IfCond) { 4795 IfVal = CGF.Builder.CreateIntCast(CGF.EvaluateExprAsBool(IfCond), CGF.IntTy, 4796 /*isSigned=*/true); 4797 } else 4798 IfVal = llvm::ConstantInt::getSigned(CGF.IntTy, /*V=*/1); 4799 4800 LValue LBLVal = CGF.EmitLValueForField( 4801 Result.TDBase, 4802 *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTLowerBound)); 4803 auto *LBVar = 4804 cast<VarDecl>(cast<DeclRefExpr>(D.getLowerBoundVariable())->getDecl()); 4805 CGF.EmitAnyExprToMem(LBVar->getInit(), LBLVal.getAddress(), LBLVal.getQuals(), 4806 /*IsInitializer=*/true); 4807 LValue UBLVal = CGF.EmitLValueForField( 4808 Result.TDBase, 4809 *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTUpperBound)); 4810 auto *UBVar = 4811 cast<VarDecl>(cast<DeclRefExpr>(D.getUpperBoundVariable())->getDecl()); 4812 CGF.EmitAnyExprToMem(UBVar->getInit(), UBLVal.getAddress(), UBLVal.getQuals(), 4813 /*IsInitializer=*/true); 4814 LValue StLVal = CGF.EmitLValueForField( 4815 Result.TDBase, 4816 *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTStride)); 4817 auto *StVar = 4818 cast<VarDecl>(cast<DeclRefExpr>(D.getStrideVariable())->getDecl()); 4819 CGF.EmitAnyExprToMem(StVar->getInit(), StLVal.getAddress(), StLVal.getQuals(), 4820 /*IsInitializer=*/true); 4821 // Store reductions address. 4822 LValue RedLVal = CGF.EmitLValueForField( 4823 Result.TDBase, 4824 *std::next(Result.KmpTaskTQTyRD->field_begin(), KmpTaskTReductions)); 4825 if (Data.Reductions) 4826 CGF.EmitStoreOfScalar(Data.Reductions, RedLVal); 4827 else { 4828 CGF.EmitNullInitialization(RedLVal.getAddress(), 4829 CGF.getContext().VoidPtrTy); 4830 } 4831 enum { NoSchedule = 0, Grainsize = 1, NumTasks = 2 }; 4832 llvm::Value *TaskArgs[] = { 4833 UpLoc, 4834 ThreadID, 4835 Result.NewTask, 4836 IfVal, 4837 LBLVal.getPointer(), 4838 UBLVal.getPointer(), 4839 CGF.EmitLoadOfScalar(StLVal, Loc), 4840 llvm::ConstantInt::getNullValue( 4841 CGF.IntTy), // Always 0 because taskgroup emitted by the compiler 4842 llvm::ConstantInt::getSigned( 4843 CGF.IntTy, Data.Schedule.getPointer() 4844 ? Data.Schedule.getInt() ? NumTasks : Grainsize 4845 : NoSchedule), 4846 Data.Schedule.getPointer() 4847 ? CGF.Builder.CreateIntCast(Data.Schedule.getPointer(), CGF.Int64Ty, 4848 /*isSigned=*/false) 4849 : llvm::ConstantInt::get(CGF.Int64Ty, /*V=*/0), 4850 Result.TaskDupFn ? CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4851 Result.TaskDupFn, CGF.VoidPtrTy) 4852 : llvm::ConstantPointerNull::get(CGF.VoidPtrTy)}; 4853 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_taskloop), TaskArgs); 4854 } 4855 4856 /// \brief Emit reduction operation for each element of array (required for 4857 /// array sections) LHS op = RHS. 4858 /// \param Type Type of array. 4859 /// \param LHSVar Variable on the left side of the reduction operation 4860 /// (references element of array in original variable). 4861 /// \param RHSVar Variable on the right side of the reduction operation 4862 /// (references element of array in original variable). 4863 /// \param RedOpGen Generator of reduction operation with use of LHSVar and 4864 /// RHSVar. 4865 static void EmitOMPAggregateReduction( 4866 CodeGenFunction &CGF, QualType Type, const VarDecl *LHSVar, 4867 const VarDecl *RHSVar, 4868 const llvm::function_ref<void(CodeGenFunction &CGF, const Expr *, 4869 const Expr *, const Expr *)> &RedOpGen, 4870 const Expr *XExpr = nullptr, const Expr *EExpr = nullptr, 4871 const Expr *UpExpr = nullptr) { 4872 // Perform element-by-element initialization. 4873 QualType ElementTy; 4874 Address LHSAddr = CGF.GetAddrOfLocalVar(LHSVar); 4875 Address RHSAddr = CGF.GetAddrOfLocalVar(RHSVar); 4876 4877 // Drill down to the base element type on both arrays. 4878 auto ArrayTy = Type->getAsArrayTypeUnsafe(); 4879 auto NumElements = CGF.emitArrayLength(ArrayTy, ElementTy, LHSAddr); 4880 4881 auto RHSBegin = RHSAddr.getPointer(); 4882 auto LHSBegin = LHSAddr.getPointer(); 4883 // Cast from pointer to array type to pointer to single element. 4884 auto LHSEnd = CGF.Builder.CreateGEP(LHSBegin, NumElements); 4885 // The basic structure here is a while-do loop. 4886 auto BodyBB = CGF.createBasicBlock("omp.arraycpy.body"); 4887 auto DoneBB = CGF.createBasicBlock("omp.arraycpy.done"); 4888 auto IsEmpty = 4889 CGF.Builder.CreateICmpEQ(LHSBegin, LHSEnd, "omp.arraycpy.isempty"); 4890 CGF.Builder.CreateCondBr(IsEmpty, DoneBB, BodyBB); 4891 4892 // Enter the loop body, making that address the current address. 4893 auto EntryBB = CGF.Builder.GetInsertBlock(); 4894 CGF.EmitBlock(BodyBB); 4895 4896 CharUnits ElementSize = CGF.getContext().getTypeSizeInChars(ElementTy); 4897 4898 llvm::PHINode *RHSElementPHI = CGF.Builder.CreatePHI( 4899 RHSBegin->getType(), 2, "omp.arraycpy.srcElementPast"); 4900 RHSElementPHI->addIncoming(RHSBegin, EntryBB); 4901 Address RHSElementCurrent = 4902 Address(RHSElementPHI, 4903 RHSAddr.getAlignment().alignmentOfArrayElement(ElementSize)); 4904 4905 llvm::PHINode *LHSElementPHI = CGF.Builder.CreatePHI( 4906 LHSBegin->getType(), 2, "omp.arraycpy.destElementPast"); 4907 LHSElementPHI->addIncoming(LHSBegin, EntryBB); 4908 Address LHSElementCurrent = 4909 Address(LHSElementPHI, 4910 LHSAddr.getAlignment().alignmentOfArrayElement(ElementSize)); 4911 4912 // Emit copy. 4913 CodeGenFunction::OMPPrivateScope Scope(CGF); 4914 Scope.addPrivate(LHSVar, [=]() -> Address { return LHSElementCurrent; }); 4915 Scope.addPrivate(RHSVar, [=]() -> Address { return RHSElementCurrent; }); 4916 Scope.Privatize(); 4917 RedOpGen(CGF, XExpr, EExpr, UpExpr); 4918 Scope.ForceCleanup(); 4919 4920 // Shift the address forward by one element. 4921 auto LHSElementNext = CGF.Builder.CreateConstGEP1_32( 4922 LHSElementPHI, /*Idx0=*/1, "omp.arraycpy.dest.element"); 4923 auto RHSElementNext = CGF.Builder.CreateConstGEP1_32( 4924 RHSElementPHI, /*Idx0=*/1, "omp.arraycpy.src.element"); 4925 // Check whether we've reached the end. 4926 auto Done = 4927 CGF.Builder.CreateICmpEQ(LHSElementNext, LHSEnd, "omp.arraycpy.done"); 4928 CGF.Builder.CreateCondBr(Done, DoneBB, BodyBB); 4929 LHSElementPHI->addIncoming(LHSElementNext, CGF.Builder.GetInsertBlock()); 4930 RHSElementPHI->addIncoming(RHSElementNext, CGF.Builder.GetInsertBlock()); 4931 4932 // Done. 4933 CGF.EmitBlock(DoneBB, /*IsFinished=*/true); 4934 } 4935 4936 /// Emit reduction combiner. If the combiner is a simple expression emit it as 4937 /// is, otherwise consider it as combiner of UDR decl and emit it as a call of 4938 /// UDR combiner function. 4939 static void emitReductionCombiner(CodeGenFunction &CGF, 4940 const Expr *ReductionOp) { 4941 if (auto *CE = dyn_cast<CallExpr>(ReductionOp)) 4942 if (auto *OVE = dyn_cast<OpaqueValueExpr>(CE->getCallee())) 4943 if (auto *DRE = 4944 dyn_cast<DeclRefExpr>(OVE->getSourceExpr()->IgnoreImpCasts())) 4945 if (auto *DRD = dyn_cast<OMPDeclareReductionDecl>(DRE->getDecl())) { 4946 std::pair<llvm::Function *, llvm::Function *> Reduction = 4947 CGF.CGM.getOpenMPRuntime().getUserDefinedReduction(DRD); 4948 RValue Func = RValue::get(Reduction.first); 4949 CodeGenFunction::OpaqueValueMapping Map(CGF, OVE, Func); 4950 CGF.EmitIgnoredExpr(ReductionOp); 4951 return; 4952 } 4953 CGF.EmitIgnoredExpr(ReductionOp); 4954 } 4955 4956 llvm::Value *CGOpenMPRuntime::emitReductionFunction( 4957 CodeGenModule &CGM, SourceLocation Loc, llvm::Type *ArgsType, 4958 ArrayRef<const Expr *> Privates, ArrayRef<const Expr *> LHSExprs, 4959 ArrayRef<const Expr *> RHSExprs, ArrayRef<const Expr *> ReductionOps) { 4960 auto &C = CGM.getContext(); 4961 4962 // void reduction_func(void *LHSArg, void *RHSArg); 4963 FunctionArgList Args; 4964 ImplicitParamDecl LHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 4965 ImplicitParamDecl::Other); 4966 ImplicitParamDecl RHSArg(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 4967 ImplicitParamDecl::Other); 4968 Args.push_back(&LHSArg); 4969 Args.push_back(&RHSArg); 4970 auto &CGFI = CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 4971 auto *Fn = llvm::Function::Create( 4972 CGM.getTypes().GetFunctionType(CGFI), llvm::GlobalValue::InternalLinkage, 4973 ".omp.reduction.reduction_func", &CGM.getModule()); 4974 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, CGFI); 4975 CodeGenFunction CGF(CGM); 4976 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, CGFI, Args, Loc, Loc); 4977 4978 // Dst = (void*[n])(LHSArg); 4979 // Src = (void*[n])(RHSArg); 4980 Address LHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4981 CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&LHSArg)), 4982 ArgsType), CGF.getPointerAlign()); 4983 Address RHS(CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 4984 CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(&RHSArg)), 4985 ArgsType), CGF.getPointerAlign()); 4986 4987 // ... 4988 // *(Type<i>*)lhs[i] = RedOp<i>(*(Type<i>*)lhs[i], *(Type<i>*)rhs[i]); 4989 // ... 4990 CodeGenFunction::OMPPrivateScope Scope(CGF); 4991 auto IPriv = Privates.begin(); 4992 unsigned Idx = 0; 4993 for (unsigned I = 0, E = ReductionOps.size(); I < E; ++I, ++IPriv, ++Idx) { 4994 auto RHSVar = cast<VarDecl>(cast<DeclRefExpr>(RHSExprs[I])->getDecl()); 4995 Scope.addPrivate(RHSVar, [&]() -> Address { 4996 return emitAddrOfVarFromArray(CGF, RHS, Idx, RHSVar); 4997 }); 4998 auto LHSVar = cast<VarDecl>(cast<DeclRefExpr>(LHSExprs[I])->getDecl()); 4999 Scope.addPrivate(LHSVar, [&]() -> Address { 5000 return emitAddrOfVarFromArray(CGF, LHS, Idx, LHSVar); 5001 }); 5002 QualType PrivTy = (*IPriv)->getType(); 5003 if (PrivTy->isVariablyModifiedType()) { 5004 // Get array size and emit VLA type. 5005 ++Idx; 5006 Address Elem = 5007 CGF.Builder.CreateConstArrayGEP(LHS, Idx, CGF.getPointerSize()); 5008 llvm::Value *Ptr = CGF.Builder.CreateLoad(Elem); 5009 auto *VLA = CGF.getContext().getAsVariableArrayType(PrivTy); 5010 auto *OVE = cast<OpaqueValueExpr>(VLA->getSizeExpr()); 5011 CodeGenFunction::OpaqueValueMapping OpaqueMap( 5012 CGF, OVE, RValue::get(CGF.Builder.CreatePtrToInt(Ptr, CGF.SizeTy))); 5013 CGF.EmitVariablyModifiedType(PrivTy); 5014 } 5015 } 5016 Scope.Privatize(); 5017 IPriv = Privates.begin(); 5018 auto ILHS = LHSExprs.begin(); 5019 auto IRHS = RHSExprs.begin(); 5020 for (auto *E : ReductionOps) { 5021 if ((*IPriv)->getType()->isArrayType()) { 5022 // Emit reduction for array section. 5023 auto *LHSVar = cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl()); 5024 auto *RHSVar = cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl()); 5025 EmitOMPAggregateReduction( 5026 CGF, (*IPriv)->getType(), LHSVar, RHSVar, 5027 [=](CodeGenFunction &CGF, const Expr *, const Expr *, const Expr *) { 5028 emitReductionCombiner(CGF, E); 5029 }); 5030 } else 5031 // Emit reduction for array subscript or single variable. 5032 emitReductionCombiner(CGF, E); 5033 ++IPriv; 5034 ++ILHS; 5035 ++IRHS; 5036 } 5037 Scope.ForceCleanup(); 5038 CGF.FinishFunction(); 5039 return Fn; 5040 } 5041 5042 void CGOpenMPRuntime::emitSingleReductionCombiner(CodeGenFunction &CGF, 5043 const Expr *ReductionOp, 5044 const Expr *PrivateRef, 5045 const DeclRefExpr *LHS, 5046 const DeclRefExpr *RHS) { 5047 if (PrivateRef->getType()->isArrayType()) { 5048 // Emit reduction for array section. 5049 auto *LHSVar = cast<VarDecl>(LHS->getDecl()); 5050 auto *RHSVar = cast<VarDecl>(RHS->getDecl()); 5051 EmitOMPAggregateReduction( 5052 CGF, PrivateRef->getType(), LHSVar, RHSVar, 5053 [=](CodeGenFunction &CGF, const Expr *, const Expr *, const Expr *) { 5054 emitReductionCombiner(CGF, ReductionOp); 5055 }); 5056 } else 5057 // Emit reduction for array subscript or single variable. 5058 emitReductionCombiner(CGF, ReductionOp); 5059 } 5060 5061 void CGOpenMPRuntime::emitReduction(CodeGenFunction &CGF, SourceLocation Loc, 5062 ArrayRef<const Expr *> Privates, 5063 ArrayRef<const Expr *> LHSExprs, 5064 ArrayRef<const Expr *> RHSExprs, 5065 ArrayRef<const Expr *> ReductionOps, 5066 ReductionOptionsTy Options) { 5067 if (!CGF.HaveInsertPoint()) 5068 return; 5069 5070 bool WithNowait = Options.WithNowait; 5071 bool SimpleReduction = Options.SimpleReduction; 5072 5073 // Next code should be emitted for reduction: 5074 // 5075 // static kmp_critical_name lock = { 0 }; 5076 // 5077 // void reduce_func(void *lhs[<n>], void *rhs[<n>]) { 5078 // *(Type0*)lhs[0] = ReductionOperation0(*(Type0*)lhs[0], *(Type0*)rhs[0]); 5079 // ... 5080 // *(Type<n>-1*)lhs[<n>-1] = ReductionOperation<n>-1(*(Type<n>-1*)lhs[<n>-1], 5081 // *(Type<n>-1*)rhs[<n>-1]); 5082 // } 5083 // 5084 // ... 5085 // void *RedList[<n>] = {&<RHSExprs>[0], ..., &<RHSExprs>[<n>-1]}; 5086 // switch (__kmpc_reduce{_nowait}(<loc>, <gtid>, <n>, sizeof(RedList), 5087 // RedList, reduce_func, &<lock>)) { 5088 // case 1: 5089 // ... 5090 // <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]); 5091 // ... 5092 // __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>); 5093 // break; 5094 // case 2: 5095 // ... 5096 // Atomic(<LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i])); 5097 // ... 5098 // [__kmpc_end_reduce(<loc>, <gtid>, &<lock>);] 5099 // break; 5100 // default:; 5101 // } 5102 // 5103 // if SimpleReduction is true, only the next code is generated: 5104 // ... 5105 // <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]); 5106 // ... 5107 5108 auto &C = CGM.getContext(); 5109 5110 if (SimpleReduction) { 5111 CodeGenFunction::RunCleanupsScope Scope(CGF); 5112 auto IPriv = Privates.begin(); 5113 auto ILHS = LHSExprs.begin(); 5114 auto IRHS = RHSExprs.begin(); 5115 for (auto *E : ReductionOps) { 5116 emitSingleReductionCombiner(CGF, E, *IPriv, cast<DeclRefExpr>(*ILHS), 5117 cast<DeclRefExpr>(*IRHS)); 5118 ++IPriv; 5119 ++ILHS; 5120 ++IRHS; 5121 } 5122 return; 5123 } 5124 5125 // 1. Build a list of reduction variables. 5126 // void *RedList[<n>] = {<ReductionVars>[0], ..., <ReductionVars>[<n>-1]}; 5127 auto Size = RHSExprs.size(); 5128 for (auto *E : Privates) { 5129 if (E->getType()->isVariablyModifiedType()) 5130 // Reserve place for array size. 5131 ++Size; 5132 } 5133 llvm::APInt ArraySize(/*unsigned int numBits=*/32, Size); 5134 QualType ReductionArrayTy = 5135 C.getConstantArrayType(C.VoidPtrTy, ArraySize, ArrayType::Normal, 5136 /*IndexTypeQuals=*/0); 5137 Address ReductionList = 5138 CGF.CreateMemTemp(ReductionArrayTy, ".omp.reduction.red_list"); 5139 auto IPriv = Privates.begin(); 5140 unsigned Idx = 0; 5141 for (unsigned I = 0, E = RHSExprs.size(); I < E; ++I, ++IPriv, ++Idx) { 5142 Address Elem = 5143 CGF.Builder.CreateConstArrayGEP(ReductionList, Idx, CGF.getPointerSize()); 5144 CGF.Builder.CreateStore( 5145 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 5146 CGF.EmitLValue(RHSExprs[I]).getPointer(), CGF.VoidPtrTy), 5147 Elem); 5148 if ((*IPriv)->getType()->isVariablyModifiedType()) { 5149 // Store array size. 5150 ++Idx; 5151 Elem = CGF.Builder.CreateConstArrayGEP(ReductionList, Idx, 5152 CGF.getPointerSize()); 5153 llvm::Value *Size = CGF.Builder.CreateIntCast( 5154 CGF.getVLASize( 5155 CGF.getContext().getAsVariableArrayType((*IPriv)->getType())) 5156 .first, 5157 CGF.SizeTy, /*isSigned=*/false); 5158 CGF.Builder.CreateStore(CGF.Builder.CreateIntToPtr(Size, CGF.VoidPtrTy), 5159 Elem); 5160 } 5161 } 5162 5163 // 2. Emit reduce_func(). 5164 auto *ReductionFn = emitReductionFunction( 5165 CGM, Loc, CGF.ConvertTypeForMem(ReductionArrayTy)->getPointerTo(), 5166 Privates, LHSExprs, RHSExprs, ReductionOps); 5167 5168 // 3. Create static kmp_critical_name lock = { 0 }; 5169 auto *Lock = getCriticalRegionLock(".reduction"); 5170 5171 // 4. Build res = __kmpc_reduce{_nowait}(<loc>, <gtid>, <n>, sizeof(RedList), 5172 // RedList, reduce_func, &<lock>); 5173 auto *IdentTLoc = emitUpdateLocation(CGF, Loc, OMP_ATOMIC_REDUCE); 5174 auto *ThreadId = getThreadID(CGF, Loc); 5175 auto *ReductionArrayTySize = CGF.getTypeSize(ReductionArrayTy); 5176 auto *RL = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 5177 ReductionList.getPointer(), CGF.VoidPtrTy); 5178 llvm::Value *Args[] = { 5179 IdentTLoc, // ident_t *<loc> 5180 ThreadId, // i32 <gtid> 5181 CGF.Builder.getInt32(RHSExprs.size()), // i32 <n> 5182 ReductionArrayTySize, // size_type sizeof(RedList) 5183 RL, // void *RedList 5184 ReductionFn, // void (*) (void *, void *) <reduce_func> 5185 Lock // kmp_critical_name *&<lock> 5186 }; 5187 auto Res = CGF.EmitRuntimeCall( 5188 createRuntimeFunction(WithNowait ? OMPRTL__kmpc_reduce_nowait 5189 : OMPRTL__kmpc_reduce), 5190 Args); 5191 5192 // 5. Build switch(res) 5193 auto *DefaultBB = CGF.createBasicBlock(".omp.reduction.default"); 5194 auto *SwInst = CGF.Builder.CreateSwitch(Res, DefaultBB, /*NumCases=*/2); 5195 5196 // 6. Build case 1: 5197 // ... 5198 // <LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i]); 5199 // ... 5200 // __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>); 5201 // break; 5202 auto *Case1BB = CGF.createBasicBlock(".omp.reduction.case1"); 5203 SwInst->addCase(CGF.Builder.getInt32(1), Case1BB); 5204 CGF.EmitBlock(Case1BB); 5205 5206 // Add emission of __kmpc_end_reduce{_nowait}(<loc>, <gtid>, &<lock>); 5207 llvm::Value *EndArgs[] = { 5208 IdentTLoc, // ident_t *<loc> 5209 ThreadId, // i32 <gtid> 5210 Lock // kmp_critical_name *&<lock> 5211 }; 5212 auto &&CodeGen = [&Privates, &LHSExprs, &RHSExprs, &ReductionOps]( 5213 CodeGenFunction &CGF, PrePostActionTy &Action) { 5214 auto &RT = CGF.CGM.getOpenMPRuntime(); 5215 auto IPriv = Privates.begin(); 5216 auto ILHS = LHSExprs.begin(); 5217 auto IRHS = RHSExprs.begin(); 5218 for (auto *E : ReductionOps) { 5219 RT.emitSingleReductionCombiner(CGF, E, *IPriv, cast<DeclRefExpr>(*ILHS), 5220 cast<DeclRefExpr>(*IRHS)); 5221 ++IPriv; 5222 ++ILHS; 5223 ++IRHS; 5224 } 5225 }; 5226 RegionCodeGenTy RCG(CodeGen); 5227 CommonActionTy Action( 5228 nullptr, llvm::None, 5229 createRuntimeFunction(WithNowait ? OMPRTL__kmpc_end_reduce_nowait 5230 : OMPRTL__kmpc_end_reduce), 5231 EndArgs); 5232 RCG.setAction(Action); 5233 RCG(CGF); 5234 5235 CGF.EmitBranch(DefaultBB); 5236 5237 // 7. Build case 2: 5238 // ... 5239 // Atomic(<LHSExprs>[i] = RedOp<i>(*<LHSExprs>[i], *<RHSExprs>[i])); 5240 // ... 5241 // break; 5242 auto *Case2BB = CGF.createBasicBlock(".omp.reduction.case2"); 5243 SwInst->addCase(CGF.Builder.getInt32(2), Case2BB); 5244 CGF.EmitBlock(Case2BB); 5245 5246 auto &&AtomicCodeGen = [Loc, &Privates, &LHSExprs, &RHSExprs, &ReductionOps]( 5247 CodeGenFunction &CGF, PrePostActionTy &Action) { 5248 auto ILHS = LHSExprs.begin(); 5249 auto IRHS = RHSExprs.begin(); 5250 auto IPriv = Privates.begin(); 5251 for (auto *E : ReductionOps) { 5252 const Expr *XExpr = nullptr; 5253 const Expr *EExpr = nullptr; 5254 const Expr *UpExpr = nullptr; 5255 BinaryOperatorKind BO = BO_Comma; 5256 if (auto *BO = dyn_cast<BinaryOperator>(E)) { 5257 if (BO->getOpcode() == BO_Assign) { 5258 XExpr = BO->getLHS(); 5259 UpExpr = BO->getRHS(); 5260 } 5261 } 5262 // Try to emit update expression as a simple atomic. 5263 auto *RHSExpr = UpExpr; 5264 if (RHSExpr) { 5265 // Analyze RHS part of the whole expression. 5266 if (auto *ACO = dyn_cast<AbstractConditionalOperator>( 5267 RHSExpr->IgnoreParenImpCasts())) { 5268 // If this is a conditional operator, analyze its condition for 5269 // min/max reduction operator. 5270 RHSExpr = ACO->getCond(); 5271 } 5272 if (auto *BORHS = 5273 dyn_cast<BinaryOperator>(RHSExpr->IgnoreParenImpCasts())) { 5274 EExpr = BORHS->getRHS(); 5275 BO = BORHS->getOpcode(); 5276 } 5277 } 5278 if (XExpr) { 5279 auto *VD = cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl()); 5280 auto &&AtomicRedGen = [BO, VD, 5281 Loc](CodeGenFunction &CGF, const Expr *XExpr, 5282 const Expr *EExpr, const Expr *UpExpr) { 5283 LValue X = CGF.EmitLValue(XExpr); 5284 RValue E; 5285 if (EExpr) 5286 E = CGF.EmitAnyExpr(EExpr); 5287 CGF.EmitOMPAtomicSimpleUpdateExpr( 5288 X, E, BO, /*IsXLHSInRHSPart=*/true, 5289 llvm::AtomicOrdering::Monotonic, Loc, 5290 [&CGF, UpExpr, VD, Loc](RValue XRValue) { 5291 CodeGenFunction::OMPPrivateScope PrivateScope(CGF); 5292 PrivateScope.addPrivate( 5293 VD, [&CGF, VD, XRValue, Loc]() -> Address { 5294 Address LHSTemp = CGF.CreateMemTemp(VD->getType()); 5295 CGF.emitOMPSimpleStore( 5296 CGF.MakeAddrLValue(LHSTemp, VD->getType()), XRValue, 5297 VD->getType().getNonReferenceType(), Loc); 5298 return LHSTemp; 5299 }); 5300 (void)PrivateScope.Privatize(); 5301 return CGF.EmitAnyExpr(UpExpr); 5302 }); 5303 }; 5304 if ((*IPriv)->getType()->isArrayType()) { 5305 // Emit atomic reduction for array section. 5306 auto *RHSVar = cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl()); 5307 EmitOMPAggregateReduction(CGF, (*IPriv)->getType(), VD, RHSVar, 5308 AtomicRedGen, XExpr, EExpr, UpExpr); 5309 } else 5310 // Emit atomic reduction for array subscript or single variable. 5311 AtomicRedGen(CGF, XExpr, EExpr, UpExpr); 5312 } else { 5313 // Emit as a critical region. 5314 auto &&CritRedGen = [E, Loc](CodeGenFunction &CGF, const Expr *, 5315 const Expr *, const Expr *) { 5316 auto &RT = CGF.CGM.getOpenMPRuntime(); 5317 RT.emitCriticalRegion( 5318 CGF, ".atomic_reduction", 5319 [=](CodeGenFunction &CGF, PrePostActionTy &Action) { 5320 Action.Enter(CGF); 5321 emitReductionCombiner(CGF, E); 5322 }, 5323 Loc); 5324 }; 5325 if ((*IPriv)->getType()->isArrayType()) { 5326 auto *LHSVar = cast<VarDecl>(cast<DeclRefExpr>(*ILHS)->getDecl()); 5327 auto *RHSVar = cast<VarDecl>(cast<DeclRefExpr>(*IRHS)->getDecl()); 5328 EmitOMPAggregateReduction(CGF, (*IPriv)->getType(), LHSVar, RHSVar, 5329 CritRedGen); 5330 } else 5331 CritRedGen(CGF, nullptr, nullptr, nullptr); 5332 } 5333 ++ILHS; 5334 ++IRHS; 5335 ++IPriv; 5336 } 5337 }; 5338 RegionCodeGenTy AtomicRCG(AtomicCodeGen); 5339 if (!WithNowait) { 5340 // Add emission of __kmpc_end_reduce(<loc>, <gtid>, &<lock>); 5341 llvm::Value *EndArgs[] = { 5342 IdentTLoc, // ident_t *<loc> 5343 ThreadId, // i32 <gtid> 5344 Lock // kmp_critical_name *&<lock> 5345 }; 5346 CommonActionTy Action(nullptr, llvm::None, 5347 createRuntimeFunction(OMPRTL__kmpc_end_reduce), 5348 EndArgs); 5349 AtomicRCG.setAction(Action); 5350 AtomicRCG(CGF); 5351 } else 5352 AtomicRCG(CGF); 5353 5354 CGF.EmitBranch(DefaultBB); 5355 CGF.EmitBlock(DefaultBB, /*IsFinished=*/true); 5356 } 5357 5358 /// Generates unique name for artificial threadprivate variables. 5359 /// Format is: <Prefix> "." <Loc_raw_encoding> "_" <N> 5360 static std::string generateUniqueName(StringRef Prefix, SourceLocation Loc, 5361 unsigned N) { 5362 SmallString<256> Buffer; 5363 llvm::raw_svector_ostream Out(Buffer); 5364 Out << Prefix << "." << Loc.getRawEncoding() << "_" << N; 5365 return Out.str(); 5366 } 5367 5368 /// Emits reduction initializer function: 5369 /// \code 5370 /// void @.red_init(void* %arg) { 5371 /// %0 = bitcast void* %arg to <type>* 5372 /// store <type> <init>, <type>* %0 5373 /// ret void 5374 /// } 5375 /// \endcode 5376 static llvm::Value *emitReduceInitFunction(CodeGenModule &CGM, 5377 SourceLocation Loc, 5378 ReductionCodeGen &RCG, unsigned N) { 5379 auto &C = CGM.getContext(); 5380 FunctionArgList Args; 5381 ImplicitParamDecl Param(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 5382 ImplicitParamDecl::Other); 5383 Args.emplace_back(&Param); 5384 auto &FnInfo = 5385 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 5386 auto *FnTy = CGM.getTypes().GetFunctionType(FnInfo); 5387 auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage, 5388 ".red_init.", &CGM.getModule()); 5389 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, FnInfo); 5390 CodeGenFunction CGF(CGM); 5391 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc); 5392 Address PrivateAddr = CGF.EmitLoadOfPointer( 5393 CGF.GetAddrOfLocalVar(&Param), 5394 C.getPointerType(C.VoidPtrTy).castAs<PointerType>()); 5395 llvm::Value *Size = nullptr; 5396 // If the size of the reduction item is non-constant, load it from global 5397 // threadprivate variable. 5398 if (RCG.getSizes(N).second) { 5399 Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate( 5400 CGF, CGM.getContext().getSizeType(), 5401 generateUniqueName("reduction_size", Loc, N)); 5402 Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false, 5403 CGM.getContext().getSizeType(), Loc); 5404 } 5405 RCG.emitAggregateType(CGF, N, Size); 5406 LValue SharedLVal; 5407 // If initializer uses initializer from declare reduction construct, emit a 5408 // pointer to the address of the original reduction item (reuired by reduction 5409 // initializer) 5410 if (RCG.usesReductionInitializer(N)) { 5411 Address SharedAddr = 5412 CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate( 5413 CGF, CGM.getContext().VoidPtrTy, 5414 generateUniqueName("reduction", Loc, N)); 5415 SharedLVal = CGF.MakeAddrLValue(SharedAddr, CGM.getContext().VoidPtrTy); 5416 } else { 5417 SharedLVal = CGF.MakeNaturalAlignAddrLValue( 5418 llvm::ConstantPointerNull::get(CGM.VoidPtrTy), 5419 CGM.getContext().VoidPtrTy); 5420 } 5421 // Emit the initializer: 5422 // %0 = bitcast void* %arg to <type>* 5423 // store <type> <init>, <type>* %0 5424 RCG.emitInitialization(CGF, N, PrivateAddr, SharedLVal, 5425 [](CodeGenFunction &) { return false; }); 5426 CGF.FinishFunction(); 5427 return Fn; 5428 } 5429 5430 /// Emits reduction combiner function: 5431 /// \code 5432 /// void @.red_comb(void* %arg0, void* %arg1) { 5433 /// %lhs = bitcast void* %arg0 to <type>* 5434 /// %rhs = bitcast void* %arg1 to <type>* 5435 /// %2 = <ReductionOp>(<type>* %lhs, <type>* %rhs) 5436 /// store <type> %2, <type>* %lhs 5437 /// ret void 5438 /// } 5439 /// \endcode 5440 static llvm::Value *emitReduceCombFunction(CodeGenModule &CGM, 5441 SourceLocation Loc, 5442 ReductionCodeGen &RCG, unsigned N, 5443 const Expr *ReductionOp, 5444 const Expr *LHS, const Expr *RHS, 5445 const Expr *PrivateRef) { 5446 auto &C = CGM.getContext(); 5447 auto *LHSVD = cast<VarDecl>(cast<DeclRefExpr>(LHS)->getDecl()); 5448 auto *RHSVD = cast<VarDecl>(cast<DeclRefExpr>(RHS)->getDecl()); 5449 FunctionArgList Args; 5450 ImplicitParamDecl ParamInOut(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, 5451 C.VoidPtrTy, ImplicitParamDecl::Other); 5452 ImplicitParamDecl ParamIn(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 5453 ImplicitParamDecl::Other); 5454 Args.emplace_back(&ParamInOut); 5455 Args.emplace_back(&ParamIn); 5456 auto &FnInfo = 5457 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 5458 auto *FnTy = CGM.getTypes().GetFunctionType(FnInfo); 5459 auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage, 5460 ".red_comb.", &CGM.getModule()); 5461 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, FnInfo); 5462 CodeGenFunction CGF(CGM); 5463 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc); 5464 llvm::Value *Size = nullptr; 5465 // If the size of the reduction item is non-constant, load it from global 5466 // threadprivate variable. 5467 if (RCG.getSizes(N).second) { 5468 Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate( 5469 CGF, CGM.getContext().getSizeType(), 5470 generateUniqueName("reduction_size", Loc, N)); 5471 Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false, 5472 CGM.getContext().getSizeType(), Loc); 5473 } 5474 RCG.emitAggregateType(CGF, N, Size); 5475 // Remap lhs and rhs variables to the addresses of the function arguments. 5476 // %lhs = bitcast void* %arg0 to <type>* 5477 // %rhs = bitcast void* %arg1 to <type>* 5478 CodeGenFunction::OMPPrivateScope PrivateScope(CGF); 5479 PrivateScope.addPrivate(LHSVD, [&C, &CGF, &ParamInOut, LHSVD]() -> Address { 5480 // Pull out the pointer to the variable. 5481 Address PtrAddr = CGF.EmitLoadOfPointer( 5482 CGF.GetAddrOfLocalVar(&ParamInOut), 5483 C.getPointerType(C.VoidPtrTy).castAs<PointerType>()); 5484 return CGF.Builder.CreateElementBitCast( 5485 PtrAddr, CGF.ConvertTypeForMem(LHSVD->getType())); 5486 }); 5487 PrivateScope.addPrivate(RHSVD, [&C, &CGF, &ParamIn, RHSVD]() -> Address { 5488 // Pull out the pointer to the variable. 5489 Address PtrAddr = CGF.EmitLoadOfPointer( 5490 CGF.GetAddrOfLocalVar(&ParamIn), 5491 C.getPointerType(C.VoidPtrTy).castAs<PointerType>()); 5492 return CGF.Builder.CreateElementBitCast( 5493 PtrAddr, CGF.ConvertTypeForMem(RHSVD->getType())); 5494 }); 5495 PrivateScope.Privatize(); 5496 // Emit the combiner body: 5497 // %2 = <ReductionOp>(<type> *%lhs, <type> *%rhs) 5498 // store <type> %2, <type>* %lhs 5499 CGM.getOpenMPRuntime().emitSingleReductionCombiner( 5500 CGF, ReductionOp, PrivateRef, cast<DeclRefExpr>(LHS), 5501 cast<DeclRefExpr>(RHS)); 5502 CGF.FinishFunction(); 5503 return Fn; 5504 } 5505 5506 /// Emits reduction finalizer function: 5507 /// \code 5508 /// void @.red_fini(void* %arg) { 5509 /// %0 = bitcast void* %arg to <type>* 5510 /// <destroy>(<type>* %0) 5511 /// ret void 5512 /// } 5513 /// \endcode 5514 static llvm::Value *emitReduceFiniFunction(CodeGenModule &CGM, 5515 SourceLocation Loc, 5516 ReductionCodeGen &RCG, unsigned N) { 5517 if (!RCG.needCleanups(N)) 5518 return nullptr; 5519 auto &C = CGM.getContext(); 5520 FunctionArgList Args; 5521 ImplicitParamDecl Param(C, /*DC=*/nullptr, Loc, /*Id=*/nullptr, C.VoidPtrTy, 5522 ImplicitParamDecl::Other); 5523 Args.emplace_back(&Param); 5524 auto &FnInfo = 5525 CGM.getTypes().arrangeBuiltinFunctionDeclaration(C.VoidTy, Args); 5526 auto *FnTy = CGM.getTypes().GetFunctionType(FnInfo); 5527 auto *Fn = llvm::Function::Create(FnTy, llvm::GlobalValue::InternalLinkage, 5528 ".red_fini.", &CGM.getModule()); 5529 CGM.SetInternalFunctionAttributes(/*D=*/nullptr, Fn, FnInfo); 5530 CodeGenFunction CGF(CGM); 5531 CGF.StartFunction(GlobalDecl(), C.VoidTy, Fn, FnInfo, Args, Loc, Loc); 5532 Address PrivateAddr = CGF.EmitLoadOfPointer( 5533 CGF.GetAddrOfLocalVar(&Param), 5534 C.getPointerType(C.VoidPtrTy).castAs<PointerType>()); 5535 llvm::Value *Size = nullptr; 5536 // If the size of the reduction item is non-constant, load it from global 5537 // threadprivate variable. 5538 if (RCG.getSizes(N).second) { 5539 Address SizeAddr = CGM.getOpenMPRuntime().getAddrOfArtificialThreadPrivate( 5540 CGF, CGM.getContext().getSizeType(), 5541 generateUniqueName("reduction_size", Loc, N)); 5542 Size = CGF.EmitLoadOfScalar(SizeAddr, /*Volatile=*/false, 5543 CGM.getContext().getSizeType(), Loc); 5544 } 5545 RCG.emitAggregateType(CGF, N, Size); 5546 // Emit the finalizer body: 5547 // <destroy>(<type>* %0) 5548 RCG.emitCleanups(CGF, N, PrivateAddr); 5549 CGF.FinishFunction(); 5550 return Fn; 5551 } 5552 5553 llvm::Value *CGOpenMPRuntime::emitTaskReductionInit( 5554 CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> LHSExprs, 5555 ArrayRef<const Expr *> RHSExprs, const OMPTaskDataTy &Data) { 5556 if (!CGF.HaveInsertPoint() || Data.ReductionVars.empty()) 5557 return nullptr; 5558 5559 // Build typedef struct: 5560 // kmp_task_red_input { 5561 // void *reduce_shar; // shared reduction item 5562 // size_t reduce_size; // size of data item 5563 // void *reduce_init; // data initialization routine 5564 // void *reduce_fini; // data finalization routine 5565 // void *reduce_comb; // data combiner routine 5566 // kmp_task_red_flags_t flags; // flags for additional info from compiler 5567 // } kmp_task_red_input_t; 5568 ASTContext &C = CGM.getContext(); 5569 auto *RD = C.buildImplicitRecord("kmp_task_red_input_t"); 5570 RD->startDefinition(); 5571 const FieldDecl *SharedFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy); 5572 const FieldDecl *SizeFD = addFieldToRecordDecl(C, RD, C.getSizeType()); 5573 const FieldDecl *InitFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy); 5574 const FieldDecl *FiniFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy); 5575 const FieldDecl *CombFD = addFieldToRecordDecl(C, RD, C.VoidPtrTy); 5576 const FieldDecl *FlagsFD = addFieldToRecordDecl( 5577 C, RD, C.getIntTypeForBitwidth(/*DestWidth=*/32, /*Signed=*/false)); 5578 RD->completeDefinition(); 5579 QualType RDType = C.getRecordType(RD); 5580 unsigned Size = Data.ReductionVars.size(); 5581 llvm::APInt ArraySize(/*numBits=*/64, Size); 5582 QualType ArrayRDType = C.getConstantArrayType( 5583 RDType, ArraySize, ArrayType::Normal, /*IndexTypeQuals=*/0); 5584 // kmp_task_red_input_t .rd_input.[Size]; 5585 Address TaskRedInput = CGF.CreateMemTemp(ArrayRDType, ".rd_input."); 5586 ReductionCodeGen RCG(Data.ReductionVars, Data.ReductionCopies, 5587 Data.ReductionOps); 5588 for (unsigned Cnt = 0; Cnt < Size; ++Cnt) { 5589 // kmp_task_red_input_t &ElemLVal = .rd_input.[Cnt]; 5590 llvm::Value *Idxs[] = {llvm::ConstantInt::get(CGM.SizeTy, /*V=*/0), 5591 llvm::ConstantInt::get(CGM.SizeTy, Cnt)}; 5592 llvm::Value *GEP = CGF.EmitCheckedInBoundsGEP( 5593 TaskRedInput.getPointer(), Idxs, 5594 /*SignedIndices=*/false, /*IsSubtraction=*/false, Loc, 5595 ".rd_input.gep."); 5596 LValue ElemLVal = CGF.MakeNaturalAlignAddrLValue(GEP, RDType); 5597 // ElemLVal.reduce_shar = &Shareds[Cnt]; 5598 LValue SharedLVal = CGF.EmitLValueForField(ElemLVal, SharedFD); 5599 RCG.emitSharedLValue(CGF, Cnt); 5600 llvm::Value *CastedShared = 5601 CGF.EmitCastToVoidPtr(RCG.getSharedLValue(Cnt).getPointer()); 5602 CGF.EmitStoreOfScalar(CastedShared, SharedLVal); 5603 RCG.emitAggregateType(CGF, Cnt); 5604 llvm::Value *SizeValInChars; 5605 llvm::Value *SizeVal; 5606 std::tie(SizeValInChars, SizeVal) = RCG.getSizes(Cnt); 5607 // We use delayed creation/initialization for VLAs, array sections and 5608 // custom reduction initializations. It is required because runtime does not 5609 // provide the way to pass the sizes of VLAs/array sections to 5610 // initializer/combiner/finalizer functions and does not pass the pointer to 5611 // original reduction item to the initializer. Instead threadprivate global 5612 // variables are used to store these values and use them in the functions. 5613 bool DelayedCreation = !!SizeVal; 5614 SizeValInChars = CGF.Builder.CreateIntCast(SizeValInChars, CGM.SizeTy, 5615 /*isSigned=*/false); 5616 LValue SizeLVal = CGF.EmitLValueForField(ElemLVal, SizeFD); 5617 CGF.EmitStoreOfScalar(SizeValInChars, SizeLVal); 5618 // ElemLVal.reduce_init = init; 5619 LValue InitLVal = CGF.EmitLValueForField(ElemLVal, InitFD); 5620 llvm::Value *InitAddr = 5621 CGF.EmitCastToVoidPtr(emitReduceInitFunction(CGM, Loc, RCG, Cnt)); 5622 CGF.EmitStoreOfScalar(InitAddr, InitLVal); 5623 DelayedCreation = DelayedCreation || RCG.usesReductionInitializer(Cnt); 5624 // ElemLVal.reduce_fini = fini; 5625 LValue FiniLVal = CGF.EmitLValueForField(ElemLVal, FiniFD); 5626 llvm::Value *Fini = emitReduceFiniFunction(CGM, Loc, RCG, Cnt); 5627 llvm::Value *FiniAddr = Fini 5628 ? CGF.EmitCastToVoidPtr(Fini) 5629 : llvm::ConstantPointerNull::get(CGM.VoidPtrTy); 5630 CGF.EmitStoreOfScalar(FiniAddr, FiniLVal); 5631 // ElemLVal.reduce_comb = comb; 5632 LValue CombLVal = CGF.EmitLValueForField(ElemLVal, CombFD); 5633 llvm::Value *CombAddr = CGF.EmitCastToVoidPtr(emitReduceCombFunction( 5634 CGM, Loc, RCG, Cnt, Data.ReductionOps[Cnt], LHSExprs[Cnt], 5635 RHSExprs[Cnt], Data.ReductionCopies[Cnt])); 5636 CGF.EmitStoreOfScalar(CombAddr, CombLVal); 5637 // ElemLVal.flags = 0; 5638 LValue FlagsLVal = CGF.EmitLValueForField(ElemLVal, FlagsFD); 5639 if (DelayedCreation) { 5640 CGF.EmitStoreOfScalar( 5641 llvm::ConstantInt::get(CGM.Int32Ty, /*V=*/1, /*IsSigned=*/true), 5642 FlagsLVal); 5643 } else 5644 CGF.EmitNullInitialization(FlagsLVal.getAddress(), FlagsLVal.getType()); 5645 } 5646 // Build call void *__kmpc_task_reduction_init(int gtid, int num_data, void 5647 // *data); 5648 llvm::Value *Args[] = { 5649 CGF.Builder.CreateIntCast(getThreadID(CGF, Loc), CGM.IntTy, 5650 /*isSigned=*/true), 5651 llvm::ConstantInt::get(CGM.IntTy, Size, /*isSigned=*/true), 5652 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(TaskRedInput.getPointer(), 5653 CGM.VoidPtrTy)}; 5654 return CGF.EmitRuntimeCall( 5655 createRuntimeFunction(OMPRTL__kmpc_task_reduction_init), Args); 5656 } 5657 5658 void CGOpenMPRuntime::emitTaskReductionFixups(CodeGenFunction &CGF, 5659 SourceLocation Loc, 5660 ReductionCodeGen &RCG, 5661 unsigned N) { 5662 auto Sizes = RCG.getSizes(N); 5663 // Emit threadprivate global variable if the type is non-constant 5664 // (Sizes.second = nullptr). 5665 if (Sizes.second) { 5666 llvm::Value *SizeVal = CGF.Builder.CreateIntCast(Sizes.second, CGM.SizeTy, 5667 /*isSigned=*/false); 5668 Address SizeAddr = getAddrOfArtificialThreadPrivate( 5669 CGF, CGM.getContext().getSizeType(), 5670 generateUniqueName("reduction_size", Loc, N)); 5671 CGF.Builder.CreateStore(SizeVal, SizeAddr, /*IsVolatile=*/false); 5672 } 5673 // Store address of the original reduction item if custom initializer is used. 5674 if (RCG.usesReductionInitializer(N)) { 5675 Address SharedAddr = getAddrOfArtificialThreadPrivate( 5676 CGF, CGM.getContext().VoidPtrTy, 5677 generateUniqueName("reduction", Loc, N)); 5678 CGF.Builder.CreateStore( 5679 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 5680 RCG.getSharedLValue(N).getPointer(), CGM.VoidPtrTy), 5681 SharedAddr, /*IsVolatile=*/false); 5682 } 5683 } 5684 5685 Address CGOpenMPRuntime::getTaskReductionItem(CodeGenFunction &CGF, 5686 SourceLocation Loc, 5687 llvm::Value *ReductionsPtr, 5688 LValue SharedLVal) { 5689 // Build call void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void 5690 // *d); 5691 llvm::Value *Args[] = { 5692 CGF.Builder.CreateIntCast(getThreadID(CGF, Loc), CGM.IntTy, 5693 /*isSigned=*/true), 5694 ReductionsPtr, 5695 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast(SharedLVal.getPointer(), 5696 CGM.VoidPtrTy)}; 5697 return Address( 5698 CGF.EmitRuntimeCall( 5699 createRuntimeFunction(OMPRTL__kmpc_task_reduction_get_th_data), Args), 5700 SharedLVal.getAlignment()); 5701 } 5702 5703 void CGOpenMPRuntime::emitTaskwaitCall(CodeGenFunction &CGF, 5704 SourceLocation Loc) { 5705 if (!CGF.HaveInsertPoint()) 5706 return; 5707 // Build call kmp_int32 __kmpc_omp_taskwait(ident_t *loc, kmp_int32 5708 // global_tid); 5709 llvm::Value *Args[] = {emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc)}; 5710 // Ignore return result until untied tasks are supported. 5711 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_omp_taskwait), Args); 5712 if (auto *Region = dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) 5713 Region->emitUntiedSwitch(CGF); 5714 } 5715 5716 void CGOpenMPRuntime::emitInlinedDirective(CodeGenFunction &CGF, 5717 OpenMPDirectiveKind InnerKind, 5718 const RegionCodeGenTy &CodeGen, 5719 bool HasCancel) { 5720 if (!CGF.HaveInsertPoint()) 5721 return; 5722 InlinedOpenMPRegionRAII Region(CGF, CodeGen, InnerKind, HasCancel); 5723 CGF.CapturedStmtInfo->EmitBody(CGF, /*S=*/nullptr); 5724 } 5725 5726 namespace { 5727 enum RTCancelKind { 5728 CancelNoreq = 0, 5729 CancelParallel = 1, 5730 CancelLoop = 2, 5731 CancelSections = 3, 5732 CancelTaskgroup = 4 5733 }; 5734 } // anonymous namespace 5735 5736 static RTCancelKind getCancellationKind(OpenMPDirectiveKind CancelRegion) { 5737 RTCancelKind CancelKind = CancelNoreq; 5738 if (CancelRegion == OMPD_parallel) 5739 CancelKind = CancelParallel; 5740 else if (CancelRegion == OMPD_for) 5741 CancelKind = CancelLoop; 5742 else if (CancelRegion == OMPD_sections) 5743 CancelKind = CancelSections; 5744 else { 5745 assert(CancelRegion == OMPD_taskgroup); 5746 CancelKind = CancelTaskgroup; 5747 } 5748 return CancelKind; 5749 } 5750 5751 void CGOpenMPRuntime::emitCancellationPointCall( 5752 CodeGenFunction &CGF, SourceLocation Loc, 5753 OpenMPDirectiveKind CancelRegion) { 5754 if (!CGF.HaveInsertPoint()) 5755 return; 5756 // Build call kmp_int32 __kmpc_cancellationpoint(ident_t *loc, kmp_int32 5757 // global_tid, kmp_int32 cncl_kind); 5758 if (auto *OMPRegionInfo = 5759 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) { 5760 // For 'cancellation point taskgroup', the task region info may not have a 5761 // cancel. This may instead happen in another adjacent task. 5762 if (CancelRegion == OMPD_taskgroup || OMPRegionInfo->hasCancel()) { 5763 llvm::Value *Args[] = { 5764 emitUpdateLocation(CGF, Loc), getThreadID(CGF, Loc), 5765 CGF.Builder.getInt32(getCancellationKind(CancelRegion))}; 5766 // Ignore return result until untied tasks are supported. 5767 auto *Result = CGF.EmitRuntimeCall( 5768 createRuntimeFunction(OMPRTL__kmpc_cancellationpoint), Args); 5769 // if (__kmpc_cancellationpoint()) { 5770 // exit from construct; 5771 // } 5772 auto *ExitBB = CGF.createBasicBlock(".cancel.exit"); 5773 auto *ContBB = CGF.createBasicBlock(".cancel.continue"); 5774 auto *Cmp = CGF.Builder.CreateIsNotNull(Result); 5775 CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB); 5776 CGF.EmitBlock(ExitBB); 5777 // exit from construct; 5778 auto CancelDest = 5779 CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind()); 5780 CGF.EmitBranchThroughCleanup(CancelDest); 5781 CGF.EmitBlock(ContBB, /*IsFinished=*/true); 5782 } 5783 } 5784 } 5785 5786 void CGOpenMPRuntime::emitCancelCall(CodeGenFunction &CGF, SourceLocation Loc, 5787 const Expr *IfCond, 5788 OpenMPDirectiveKind CancelRegion) { 5789 if (!CGF.HaveInsertPoint()) 5790 return; 5791 // Build call kmp_int32 __kmpc_cancel(ident_t *loc, kmp_int32 global_tid, 5792 // kmp_int32 cncl_kind); 5793 if (auto *OMPRegionInfo = 5794 dyn_cast_or_null<CGOpenMPRegionInfo>(CGF.CapturedStmtInfo)) { 5795 auto &&ThenGen = [Loc, CancelRegion, OMPRegionInfo](CodeGenFunction &CGF, 5796 PrePostActionTy &) { 5797 auto &RT = CGF.CGM.getOpenMPRuntime(); 5798 llvm::Value *Args[] = { 5799 RT.emitUpdateLocation(CGF, Loc), RT.getThreadID(CGF, Loc), 5800 CGF.Builder.getInt32(getCancellationKind(CancelRegion))}; 5801 // Ignore return result until untied tasks are supported. 5802 auto *Result = CGF.EmitRuntimeCall( 5803 RT.createRuntimeFunction(OMPRTL__kmpc_cancel), Args); 5804 // if (__kmpc_cancel()) { 5805 // exit from construct; 5806 // } 5807 auto *ExitBB = CGF.createBasicBlock(".cancel.exit"); 5808 auto *ContBB = CGF.createBasicBlock(".cancel.continue"); 5809 auto *Cmp = CGF.Builder.CreateIsNotNull(Result); 5810 CGF.Builder.CreateCondBr(Cmp, ExitBB, ContBB); 5811 CGF.EmitBlock(ExitBB); 5812 // exit from construct; 5813 auto CancelDest = 5814 CGF.getOMPCancelDestination(OMPRegionInfo->getDirectiveKind()); 5815 CGF.EmitBranchThroughCleanup(CancelDest); 5816 CGF.EmitBlock(ContBB, /*IsFinished=*/true); 5817 }; 5818 if (IfCond) 5819 emitOMPIfClause(CGF, IfCond, ThenGen, 5820 [](CodeGenFunction &, PrePostActionTy &) {}); 5821 else { 5822 RegionCodeGenTy ThenRCG(ThenGen); 5823 ThenRCG(CGF); 5824 } 5825 } 5826 } 5827 5828 /// \brief Obtain information that uniquely identifies a target entry. This 5829 /// consists of the file and device IDs as well as line number associated with 5830 /// the relevant entry source location. 5831 static void getTargetEntryUniqueInfo(ASTContext &C, SourceLocation Loc, 5832 unsigned &DeviceID, unsigned &FileID, 5833 unsigned &LineNum) { 5834 5835 auto &SM = C.getSourceManager(); 5836 5837 // The loc should be always valid and have a file ID (the user cannot use 5838 // #pragma directives in macros) 5839 5840 assert(Loc.isValid() && "Source location is expected to be always valid."); 5841 assert(Loc.isFileID() && "Source location is expected to refer to a file."); 5842 5843 PresumedLoc PLoc = SM.getPresumedLoc(Loc); 5844 assert(PLoc.isValid() && "Source location is expected to be always valid."); 5845 5846 llvm::sys::fs::UniqueID ID; 5847 if (llvm::sys::fs::getUniqueID(PLoc.getFilename(), ID)) 5848 llvm_unreachable("Source file with target region no longer exists!"); 5849 5850 DeviceID = ID.getDevice(); 5851 FileID = ID.getFile(); 5852 LineNum = PLoc.getLine(); 5853 } 5854 5855 void CGOpenMPRuntime::emitTargetOutlinedFunction( 5856 const OMPExecutableDirective &D, StringRef ParentName, 5857 llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID, 5858 bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) { 5859 assert(!ParentName.empty() && "Invalid target region parent name!"); 5860 5861 emitTargetOutlinedFunctionHelper(D, ParentName, OutlinedFn, OutlinedFnID, 5862 IsOffloadEntry, CodeGen); 5863 } 5864 5865 void CGOpenMPRuntime::emitTargetOutlinedFunctionHelper( 5866 const OMPExecutableDirective &D, StringRef ParentName, 5867 llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID, 5868 bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) { 5869 // Create a unique name for the entry function using the source location 5870 // information of the current target region. The name will be something like: 5871 // 5872 // __omp_offloading_DD_FFFF_PP_lBB 5873 // 5874 // where DD_FFFF is an ID unique to the file (device and file IDs), PP is the 5875 // mangled name of the function that encloses the target region and BB is the 5876 // line number of the target region. 5877 5878 unsigned DeviceID; 5879 unsigned FileID; 5880 unsigned Line; 5881 getTargetEntryUniqueInfo(CGM.getContext(), D.getLocStart(), DeviceID, FileID, 5882 Line); 5883 SmallString<64> EntryFnName; 5884 { 5885 llvm::raw_svector_ostream OS(EntryFnName); 5886 OS << "__omp_offloading" << llvm::format("_%x", DeviceID) 5887 << llvm::format("_%x_", FileID) << ParentName << "_l" << Line; 5888 } 5889 5890 const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target); 5891 5892 CodeGenFunction CGF(CGM, true); 5893 CGOpenMPTargetRegionInfo CGInfo(CS, CodeGen, EntryFnName); 5894 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo); 5895 5896 OutlinedFn = CGF.GenerateOpenMPCapturedStmtFunction(CS); 5897 5898 // If this target outline function is not an offload entry, we don't need to 5899 // register it. 5900 if (!IsOffloadEntry) 5901 return; 5902 5903 // The target region ID is used by the runtime library to identify the current 5904 // target region, so it only has to be unique and not necessarily point to 5905 // anything. It could be the pointer to the outlined function that implements 5906 // the target region, but we aren't using that so that the compiler doesn't 5907 // need to keep that, and could therefore inline the host function if proven 5908 // worthwhile during optimization. In the other hand, if emitting code for the 5909 // device, the ID has to be the function address so that it can retrieved from 5910 // the offloading entry and launched by the runtime library. We also mark the 5911 // outlined function to have external linkage in case we are emitting code for 5912 // the device, because these functions will be entry points to the device. 5913 5914 if (CGM.getLangOpts().OpenMPIsDevice) { 5915 OutlinedFnID = llvm::ConstantExpr::getBitCast(OutlinedFn, CGM.Int8PtrTy); 5916 OutlinedFn->setLinkage(llvm::GlobalValue::ExternalLinkage); 5917 OutlinedFn->setDSOLocal(false); 5918 } else 5919 OutlinedFnID = new llvm::GlobalVariable( 5920 CGM.getModule(), CGM.Int8Ty, /*isConstant=*/true, 5921 llvm::GlobalValue::PrivateLinkage, 5922 llvm::Constant::getNullValue(CGM.Int8Ty), ".omp_offload.region_id"); 5923 5924 // Register the information for the entry associated with this target region. 5925 OffloadEntriesInfoManager.registerTargetRegionEntryInfo( 5926 DeviceID, FileID, ParentName, Line, OutlinedFn, OutlinedFnID, 5927 /*Flags=*/0); 5928 } 5929 5930 /// discard all CompoundStmts intervening between two constructs 5931 static const Stmt *ignoreCompoundStmts(const Stmt *Body) { 5932 while (auto *CS = dyn_cast_or_null<CompoundStmt>(Body)) 5933 Body = CS->body_front(); 5934 5935 return Body; 5936 } 5937 5938 /// Emit the number of teams for a target directive. Inspect the num_teams 5939 /// clause associated with a teams construct combined or closely nested 5940 /// with the target directive. 5941 /// 5942 /// Emit a team of size one for directives such as 'target parallel' that 5943 /// have no associated teams construct. 5944 /// 5945 /// Otherwise, return nullptr. 5946 static llvm::Value * 5947 emitNumTeamsForTargetDirective(CGOpenMPRuntime &OMPRuntime, 5948 CodeGenFunction &CGF, 5949 const OMPExecutableDirective &D) { 5950 5951 assert(!CGF.getLangOpts().OpenMPIsDevice && "Clauses associated with the " 5952 "teams directive expected to be " 5953 "emitted only for the host!"); 5954 5955 auto &Bld = CGF.Builder; 5956 5957 // If the target directive is combined with a teams directive: 5958 // Return the value in the num_teams clause, if any. 5959 // Otherwise, return 0 to denote the runtime default. 5960 if (isOpenMPTeamsDirective(D.getDirectiveKind())) { 5961 if (const auto *NumTeamsClause = D.getSingleClause<OMPNumTeamsClause>()) { 5962 CodeGenFunction::RunCleanupsScope NumTeamsScope(CGF); 5963 auto NumTeams = CGF.EmitScalarExpr(NumTeamsClause->getNumTeams(), 5964 /*IgnoreResultAssign*/ true); 5965 return Bld.CreateIntCast(NumTeams, CGF.Int32Ty, 5966 /*IsSigned=*/true); 5967 } 5968 5969 // The default value is 0. 5970 return Bld.getInt32(0); 5971 } 5972 5973 // If the target directive is combined with a parallel directive but not a 5974 // teams directive, start one team. 5975 if (isOpenMPParallelDirective(D.getDirectiveKind())) 5976 return Bld.getInt32(1); 5977 5978 // If the current target region has a teams region enclosed, we need to get 5979 // the number of teams to pass to the runtime function call. This is done 5980 // by generating the expression in a inlined region. This is required because 5981 // the expression is captured in the enclosing target environment when the 5982 // teams directive is not combined with target. 5983 5984 const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target); 5985 5986 if (auto *TeamsDir = dyn_cast_or_null<OMPExecutableDirective>( 5987 ignoreCompoundStmts(CS.getCapturedStmt()))) { 5988 if (isOpenMPTeamsDirective(TeamsDir->getDirectiveKind())) { 5989 if (auto *NTE = TeamsDir->getSingleClause<OMPNumTeamsClause>()) { 5990 CGOpenMPInnerExprInfo CGInfo(CGF, CS); 5991 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo); 5992 llvm::Value *NumTeams = CGF.EmitScalarExpr(NTE->getNumTeams()); 5993 return Bld.CreateIntCast(NumTeams, CGF.Int32Ty, 5994 /*IsSigned=*/true); 5995 } 5996 5997 // If we have an enclosed teams directive but no num_teams clause we use 5998 // the default value 0. 5999 return Bld.getInt32(0); 6000 } 6001 } 6002 6003 // No teams associated with the directive. 6004 return nullptr; 6005 } 6006 6007 /// Emit the number of threads for a target directive. Inspect the 6008 /// thread_limit clause associated with a teams construct combined or closely 6009 /// nested with the target directive. 6010 /// 6011 /// Emit the num_threads clause for directives such as 'target parallel' that 6012 /// have no associated teams construct. 6013 /// 6014 /// Otherwise, return nullptr. 6015 static llvm::Value * 6016 emitNumThreadsForTargetDirective(CGOpenMPRuntime &OMPRuntime, 6017 CodeGenFunction &CGF, 6018 const OMPExecutableDirective &D) { 6019 6020 assert(!CGF.getLangOpts().OpenMPIsDevice && "Clauses associated with the " 6021 "teams directive expected to be " 6022 "emitted only for the host!"); 6023 6024 auto &Bld = CGF.Builder; 6025 6026 // 6027 // If the target directive is combined with a teams directive: 6028 // Return the value in the thread_limit clause, if any. 6029 // 6030 // If the target directive is combined with a parallel directive: 6031 // Return the value in the num_threads clause, if any. 6032 // 6033 // If both clauses are set, select the minimum of the two. 6034 // 6035 // If neither teams or parallel combined directives set the number of threads 6036 // in a team, return 0 to denote the runtime default. 6037 // 6038 // If this is not a teams directive return nullptr. 6039 6040 if (isOpenMPTeamsDirective(D.getDirectiveKind()) || 6041 isOpenMPParallelDirective(D.getDirectiveKind())) { 6042 llvm::Value *DefaultThreadLimitVal = Bld.getInt32(0); 6043 llvm::Value *NumThreadsVal = nullptr; 6044 llvm::Value *ThreadLimitVal = nullptr; 6045 6046 if (const auto *ThreadLimitClause = 6047 D.getSingleClause<OMPThreadLimitClause>()) { 6048 CodeGenFunction::RunCleanupsScope ThreadLimitScope(CGF); 6049 auto ThreadLimit = CGF.EmitScalarExpr(ThreadLimitClause->getThreadLimit(), 6050 /*IgnoreResultAssign*/ true); 6051 ThreadLimitVal = Bld.CreateIntCast(ThreadLimit, CGF.Int32Ty, 6052 /*IsSigned=*/true); 6053 } 6054 6055 if (const auto *NumThreadsClause = 6056 D.getSingleClause<OMPNumThreadsClause>()) { 6057 CodeGenFunction::RunCleanupsScope NumThreadsScope(CGF); 6058 llvm::Value *NumThreads = 6059 CGF.EmitScalarExpr(NumThreadsClause->getNumThreads(), 6060 /*IgnoreResultAssign*/ true); 6061 NumThreadsVal = 6062 Bld.CreateIntCast(NumThreads, CGF.Int32Ty, /*IsSigned=*/true); 6063 } 6064 6065 // Select the lesser of thread_limit and num_threads. 6066 if (NumThreadsVal) 6067 ThreadLimitVal = ThreadLimitVal 6068 ? Bld.CreateSelect(Bld.CreateICmpSLT(NumThreadsVal, 6069 ThreadLimitVal), 6070 NumThreadsVal, ThreadLimitVal) 6071 : NumThreadsVal; 6072 6073 // Set default value passed to the runtime if either teams or a target 6074 // parallel type directive is found but no clause is specified. 6075 if (!ThreadLimitVal) 6076 ThreadLimitVal = DefaultThreadLimitVal; 6077 6078 return ThreadLimitVal; 6079 } 6080 6081 // If the current target region has a teams region enclosed, we need to get 6082 // the thread limit to pass to the runtime function call. This is done 6083 // by generating the expression in a inlined region. This is required because 6084 // the expression is captured in the enclosing target environment when the 6085 // teams directive is not combined with target. 6086 6087 const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target); 6088 6089 if (auto *TeamsDir = dyn_cast_or_null<OMPExecutableDirective>( 6090 ignoreCompoundStmts(CS.getCapturedStmt()))) { 6091 if (isOpenMPTeamsDirective(TeamsDir->getDirectiveKind())) { 6092 if (auto *TLE = TeamsDir->getSingleClause<OMPThreadLimitClause>()) { 6093 CGOpenMPInnerExprInfo CGInfo(CGF, CS); 6094 CodeGenFunction::CGCapturedStmtRAII CapInfoRAII(CGF, &CGInfo); 6095 llvm::Value *ThreadLimit = CGF.EmitScalarExpr(TLE->getThreadLimit()); 6096 return CGF.Builder.CreateIntCast(ThreadLimit, CGF.Int32Ty, 6097 /*IsSigned=*/true); 6098 } 6099 6100 // If we have an enclosed teams directive but no thread_limit clause we 6101 // use the default value 0. 6102 return CGF.Builder.getInt32(0); 6103 } 6104 } 6105 6106 // No teams associated with the directive. 6107 return nullptr; 6108 } 6109 6110 namespace { 6111 // \brief Utility to handle information from clauses associated with a given 6112 // construct that use mappable expressions (e.g. 'map' clause, 'to' clause). 6113 // It provides a convenient interface to obtain the information and generate 6114 // code for that information. 6115 class MappableExprsHandler { 6116 public: 6117 /// \brief Values for bit flags used to specify the mapping type for 6118 /// offloading. 6119 enum OpenMPOffloadMappingFlags { 6120 /// \brief Allocate memory on the device and move data from host to device. 6121 OMP_MAP_TO = 0x01, 6122 /// \brief Allocate memory on the device and move data from device to host. 6123 OMP_MAP_FROM = 0x02, 6124 /// \brief Always perform the requested mapping action on the element, even 6125 /// if it was already mapped before. 6126 OMP_MAP_ALWAYS = 0x04, 6127 /// \brief Delete the element from the device environment, ignoring the 6128 /// current reference count associated with the element. 6129 OMP_MAP_DELETE = 0x08, 6130 /// \brief The element being mapped is a pointer-pointee pair; both the 6131 /// pointer and the pointee should be mapped. 6132 OMP_MAP_PTR_AND_OBJ = 0x10, 6133 /// \brief This flags signals that the base address of an entry should be 6134 /// passed to the target kernel as an argument. 6135 OMP_MAP_TARGET_PARAM = 0x20, 6136 /// \brief Signal that the runtime library has to return the device pointer 6137 /// in the current position for the data being mapped. Used when we have the 6138 /// use_device_ptr clause. 6139 OMP_MAP_RETURN_PARAM = 0x40, 6140 /// \brief This flag signals that the reference being passed is a pointer to 6141 /// private data. 6142 OMP_MAP_PRIVATE = 0x80, 6143 /// \brief Pass the element to the device by value. 6144 OMP_MAP_LITERAL = 0x100, 6145 /// Implicit map 6146 OMP_MAP_IMPLICIT = 0x200, 6147 }; 6148 6149 /// Class that associates information with a base pointer to be passed to the 6150 /// runtime library. 6151 class BasePointerInfo { 6152 /// The base pointer. 6153 llvm::Value *Ptr = nullptr; 6154 /// The base declaration that refers to this device pointer, or null if 6155 /// there is none. 6156 const ValueDecl *DevPtrDecl = nullptr; 6157 6158 public: 6159 BasePointerInfo(llvm::Value *Ptr, const ValueDecl *DevPtrDecl = nullptr) 6160 : Ptr(Ptr), DevPtrDecl(DevPtrDecl) {} 6161 llvm::Value *operator*() const { return Ptr; } 6162 const ValueDecl *getDevicePtrDecl() const { return DevPtrDecl; } 6163 void setDevicePtrDecl(const ValueDecl *D) { DevPtrDecl = D; } 6164 }; 6165 6166 typedef SmallVector<BasePointerInfo, 16> MapBaseValuesArrayTy; 6167 typedef SmallVector<llvm::Value *, 16> MapValuesArrayTy; 6168 typedef SmallVector<uint64_t, 16> MapFlagsArrayTy; 6169 6170 private: 6171 /// \brief Directive from where the map clauses were extracted. 6172 const OMPExecutableDirective &CurDir; 6173 6174 /// \brief Function the directive is being generated for. 6175 CodeGenFunction &CGF; 6176 6177 /// \brief Set of all first private variables in the current directive. 6178 llvm::SmallPtrSet<const VarDecl *, 8> FirstPrivateDecls; 6179 /// Set of all reduction variables in the current directive. 6180 llvm::SmallPtrSet<const VarDecl *, 8> ReductionDecls; 6181 6182 /// Map between device pointer declarations and their expression components. 6183 /// The key value for declarations in 'this' is null. 6184 llvm::DenseMap< 6185 const ValueDecl *, 6186 SmallVector<OMPClauseMappableExprCommon::MappableExprComponentListRef, 4>> 6187 DevPointersMap; 6188 6189 llvm::Value *getExprTypeSize(const Expr *E) const { 6190 auto ExprTy = E->getType().getCanonicalType(); 6191 6192 // Reference types are ignored for mapping purposes. 6193 if (auto *RefTy = ExprTy->getAs<ReferenceType>()) 6194 ExprTy = RefTy->getPointeeType().getCanonicalType(); 6195 6196 // Given that an array section is considered a built-in type, we need to 6197 // do the calculation based on the length of the section instead of relying 6198 // on CGF.getTypeSize(E->getType()). 6199 if (const auto *OAE = dyn_cast<OMPArraySectionExpr>(E)) { 6200 QualType BaseTy = OMPArraySectionExpr::getBaseOriginalType( 6201 OAE->getBase()->IgnoreParenImpCasts()) 6202 .getCanonicalType(); 6203 6204 // If there is no length associated with the expression, that means we 6205 // are using the whole length of the base. 6206 if (!OAE->getLength() && OAE->getColonLoc().isValid()) 6207 return CGF.getTypeSize(BaseTy); 6208 6209 llvm::Value *ElemSize; 6210 if (auto *PTy = BaseTy->getAs<PointerType>()) 6211 ElemSize = CGF.getTypeSize(PTy->getPointeeType().getCanonicalType()); 6212 else { 6213 auto *ATy = cast<ArrayType>(BaseTy.getTypePtr()); 6214 assert(ATy && "Expecting array type if not a pointer type."); 6215 ElemSize = CGF.getTypeSize(ATy->getElementType().getCanonicalType()); 6216 } 6217 6218 // If we don't have a length at this point, that is because we have an 6219 // array section with a single element. 6220 if (!OAE->getLength()) 6221 return ElemSize; 6222 6223 auto *LengthVal = CGF.EmitScalarExpr(OAE->getLength()); 6224 LengthVal = 6225 CGF.Builder.CreateIntCast(LengthVal, CGF.SizeTy, /*isSigned=*/false); 6226 return CGF.Builder.CreateNUWMul(LengthVal, ElemSize); 6227 } 6228 return CGF.getTypeSize(ExprTy); 6229 } 6230 6231 /// \brief Return the corresponding bits for a given map clause modifier. Add 6232 /// a flag marking the map as a pointer if requested. Add a flag marking the 6233 /// map as the first one of a series of maps that relate to the same map 6234 /// expression. 6235 uint64_t getMapTypeBits(OpenMPMapClauseKind MapType, 6236 OpenMPMapClauseKind MapTypeModifier, bool AddPtrFlag, 6237 bool AddIsTargetParamFlag) const { 6238 uint64_t Bits = 0u; 6239 switch (MapType) { 6240 case OMPC_MAP_alloc: 6241 case OMPC_MAP_release: 6242 // alloc and release is the default behavior in the runtime library, i.e. 6243 // if we don't pass any bits alloc/release that is what the runtime is 6244 // going to do. Therefore, we don't need to signal anything for these two 6245 // type modifiers. 6246 break; 6247 case OMPC_MAP_to: 6248 Bits = OMP_MAP_TO; 6249 break; 6250 case OMPC_MAP_from: 6251 Bits = OMP_MAP_FROM; 6252 break; 6253 case OMPC_MAP_tofrom: 6254 Bits = OMP_MAP_TO | OMP_MAP_FROM; 6255 break; 6256 case OMPC_MAP_delete: 6257 Bits = OMP_MAP_DELETE; 6258 break; 6259 default: 6260 llvm_unreachable("Unexpected map type!"); 6261 break; 6262 } 6263 if (AddPtrFlag) 6264 Bits |= OMP_MAP_PTR_AND_OBJ; 6265 if (AddIsTargetParamFlag) 6266 Bits |= OMP_MAP_TARGET_PARAM; 6267 if (MapTypeModifier == OMPC_MAP_always) 6268 Bits |= OMP_MAP_ALWAYS; 6269 return Bits; 6270 } 6271 6272 /// \brief Return true if the provided expression is a final array section. A 6273 /// final array section, is one whose length can't be proved to be one. 6274 bool isFinalArraySectionExpression(const Expr *E) const { 6275 auto *OASE = dyn_cast<OMPArraySectionExpr>(E); 6276 6277 // It is not an array section and therefore not a unity-size one. 6278 if (!OASE) 6279 return false; 6280 6281 // An array section with no colon always refer to a single element. 6282 if (OASE->getColonLoc().isInvalid()) 6283 return false; 6284 6285 auto *Length = OASE->getLength(); 6286 6287 // If we don't have a length we have to check if the array has size 1 6288 // for this dimension. Also, we should always expect a length if the 6289 // base type is pointer. 6290 if (!Length) { 6291 auto BaseQTy = OMPArraySectionExpr::getBaseOriginalType( 6292 OASE->getBase()->IgnoreParenImpCasts()) 6293 .getCanonicalType(); 6294 if (auto *ATy = dyn_cast<ConstantArrayType>(BaseQTy.getTypePtr())) 6295 return ATy->getSize().getSExtValue() != 1; 6296 // If we don't have a constant dimension length, we have to consider 6297 // the current section as having any size, so it is not necessarily 6298 // unitary. If it happen to be unity size, that's user fault. 6299 return true; 6300 } 6301 6302 // Check if the length evaluates to 1. 6303 llvm::APSInt ConstLength; 6304 if (!Length->EvaluateAsInt(ConstLength, CGF.getContext())) 6305 return true; // Can have more that size 1. 6306 6307 return ConstLength.getSExtValue() != 1; 6308 } 6309 6310 /// \brief Generate the base pointers, section pointers, sizes and map type 6311 /// bits for the provided map type, map modifier, and expression components. 6312 /// \a IsFirstComponent should be set to true if the provided set of 6313 /// components is the first associated with a capture. 6314 void generateInfoForComponentList( 6315 OpenMPMapClauseKind MapType, OpenMPMapClauseKind MapTypeModifier, 6316 OMPClauseMappableExprCommon::MappableExprComponentListRef Components, 6317 MapBaseValuesArrayTy &BasePointers, MapValuesArrayTy &Pointers, 6318 MapValuesArrayTy &Sizes, MapFlagsArrayTy &Types, 6319 bool IsFirstComponentList, bool IsImplicit) const { 6320 6321 // The following summarizes what has to be generated for each map and the 6322 // types bellow. The generated information is expressed in this order: 6323 // base pointer, section pointer, size, flags 6324 // (to add to the ones that come from the map type and modifier). 6325 // 6326 // double d; 6327 // int i[100]; 6328 // float *p; 6329 // 6330 // struct S1 { 6331 // int i; 6332 // float f[50]; 6333 // } 6334 // struct S2 { 6335 // int i; 6336 // float f[50]; 6337 // S1 s; 6338 // double *p; 6339 // struct S2 *ps; 6340 // } 6341 // S2 s; 6342 // S2 *ps; 6343 // 6344 // map(d) 6345 // &d, &d, sizeof(double), noflags 6346 // 6347 // map(i) 6348 // &i, &i, 100*sizeof(int), noflags 6349 // 6350 // map(i[1:23]) 6351 // &i(=&i[0]), &i[1], 23*sizeof(int), noflags 6352 // 6353 // map(p) 6354 // &p, &p, sizeof(float*), noflags 6355 // 6356 // map(p[1:24]) 6357 // p, &p[1], 24*sizeof(float), noflags 6358 // 6359 // map(s) 6360 // &s, &s, sizeof(S2), noflags 6361 // 6362 // map(s.i) 6363 // &s, &(s.i), sizeof(int), noflags 6364 // 6365 // map(s.s.f) 6366 // &s, &(s.i.f), 50*sizeof(int), noflags 6367 // 6368 // map(s.p) 6369 // &s, &(s.p), sizeof(double*), noflags 6370 // 6371 // map(s.p[:22], s.a s.b) 6372 // &s, &(s.p), sizeof(double*), noflags 6373 // &(s.p), &(s.p[0]), 22*sizeof(double), ptr_flag 6374 // 6375 // map(s.ps) 6376 // &s, &(s.ps), sizeof(S2*), noflags 6377 // 6378 // map(s.ps->s.i) 6379 // &s, &(s.ps), sizeof(S2*), noflags 6380 // &(s.ps), &(s.ps->s.i), sizeof(int), ptr_flag 6381 // 6382 // map(s.ps->ps) 6383 // &s, &(s.ps), sizeof(S2*), noflags 6384 // &(s.ps), &(s.ps->ps), sizeof(S2*), ptr_flag 6385 // 6386 // map(s.ps->ps->ps) 6387 // &s, &(s.ps), sizeof(S2*), noflags 6388 // &(s.ps), &(s.ps->ps), sizeof(S2*), ptr_flag 6389 // &(s.ps->ps), &(s.ps->ps->ps), sizeof(S2*), ptr_flag 6390 // 6391 // map(s.ps->ps->s.f[:22]) 6392 // &s, &(s.ps), sizeof(S2*), noflags 6393 // &(s.ps), &(s.ps->ps), sizeof(S2*), ptr_flag 6394 // &(s.ps->ps), &(s.ps->ps->s.f[0]), 22*sizeof(float), ptr_flag 6395 // 6396 // map(ps) 6397 // &ps, &ps, sizeof(S2*), noflags 6398 // 6399 // map(ps->i) 6400 // ps, &(ps->i), sizeof(int), noflags 6401 // 6402 // map(ps->s.f) 6403 // ps, &(ps->s.f[0]), 50*sizeof(float), noflags 6404 // 6405 // map(ps->p) 6406 // ps, &(ps->p), sizeof(double*), noflags 6407 // 6408 // map(ps->p[:22]) 6409 // ps, &(ps->p), sizeof(double*), noflags 6410 // &(ps->p), &(ps->p[0]), 22*sizeof(double), ptr_flag 6411 // 6412 // map(ps->ps) 6413 // ps, &(ps->ps), sizeof(S2*), noflags 6414 // 6415 // map(ps->ps->s.i) 6416 // ps, &(ps->ps), sizeof(S2*), noflags 6417 // &(ps->ps), &(ps->ps->s.i), sizeof(int), ptr_flag 6418 // 6419 // map(ps->ps->ps) 6420 // ps, &(ps->ps), sizeof(S2*), noflags 6421 // &(ps->ps), &(ps->ps->ps), sizeof(S2*), ptr_flag 6422 // 6423 // map(ps->ps->ps->ps) 6424 // ps, &(ps->ps), sizeof(S2*), noflags 6425 // &(ps->ps), &(ps->ps->ps), sizeof(S2*), ptr_flag 6426 // &(ps->ps->ps), &(ps->ps->ps->ps), sizeof(S2*), ptr_flag 6427 // 6428 // map(ps->ps->ps->s.f[:22]) 6429 // ps, &(ps->ps), sizeof(S2*), noflags 6430 // &(ps->ps), &(ps->ps->ps), sizeof(S2*), ptr_flag 6431 // &(ps->ps->ps), &(ps->ps->ps->s.f[0]), 22*sizeof(float), ptr_flag 6432 6433 // Track if the map information being generated is the first for a capture. 6434 bool IsCaptureFirstInfo = IsFirstComponentList; 6435 6436 // Scan the components from the base to the complete expression. 6437 auto CI = Components.rbegin(); 6438 auto CE = Components.rend(); 6439 auto I = CI; 6440 6441 // Track if the map information being generated is the first for a list of 6442 // components. 6443 bool IsExpressionFirstInfo = true; 6444 llvm::Value *BP = nullptr; 6445 6446 if (auto *ME = dyn_cast<MemberExpr>(I->getAssociatedExpression())) { 6447 // The base is the 'this' pointer. The content of the pointer is going 6448 // to be the base of the field being mapped. 6449 BP = CGF.EmitScalarExpr(ME->getBase()); 6450 } else { 6451 // The base is the reference to the variable. 6452 // BP = &Var. 6453 BP = CGF.EmitOMPSharedLValue(I->getAssociatedExpression()).getPointer(); 6454 6455 // If the variable is a pointer and is being dereferenced (i.e. is not 6456 // the last component), the base has to be the pointer itself, not its 6457 // reference. References are ignored for mapping purposes. 6458 QualType Ty = 6459 I->getAssociatedDeclaration()->getType().getNonReferenceType(); 6460 if (Ty->isAnyPointerType() && std::next(I) != CE) { 6461 auto PtrAddr = CGF.MakeNaturalAlignAddrLValue(BP, Ty); 6462 BP = CGF.EmitLoadOfPointerLValue(PtrAddr.getAddress(), 6463 Ty->castAs<PointerType>()) 6464 .getPointer(); 6465 6466 // We do not need to generate individual map information for the 6467 // pointer, it can be associated with the combined storage. 6468 ++I; 6469 } 6470 } 6471 6472 uint64_t DefaultFlags = IsImplicit ? OMP_MAP_IMPLICIT : 0; 6473 for (; I != CE; ++I) { 6474 auto Next = std::next(I); 6475 6476 // We need to generate the addresses and sizes if this is the last 6477 // component, if the component is a pointer or if it is an array section 6478 // whose length can't be proved to be one. If this is a pointer, it 6479 // becomes the base address for the following components. 6480 6481 // A final array section, is one whose length can't be proved to be one. 6482 bool IsFinalArraySection = 6483 isFinalArraySectionExpression(I->getAssociatedExpression()); 6484 6485 // Get information on whether the element is a pointer. Have to do a 6486 // special treatment for array sections given that they are built-in 6487 // types. 6488 const auto *OASE = 6489 dyn_cast<OMPArraySectionExpr>(I->getAssociatedExpression()); 6490 bool IsPointer = 6491 (OASE && 6492 OMPArraySectionExpr::getBaseOriginalType(OASE) 6493 .getCanonicalType() 6494 ->isAnyPointerType()) || 6495 I->getAssociatedExpression()->getType()->isAnyPointerType(); 6496 6497 if (Next == CE || IsPointer || IsFinalArraySection) { 6498 6499 // If this is not the last component, we expect the pointer to be 6500 // associated with an array expression or member expression. 6501 assert((Next == CE || 6502 isa<MemberExpr>(Next->getAssociatedExpression()) || 6503 isa<ArraySubscriptExpr>(Next->getAssociatedExpression()) || 6504 isa<OMPArraySectionExpr>(Next->getAssociatedExpression())) && 6505 "Unexpected expression"); 6506 6507 llvm::Value *LB = 6508 CGF.EmitOMPSharedLValue(I->getAssociatedExpression()).getPointer(); 6509 auto *Size = getExprTypeSize(I->getAssociatedExpression()); 6510 6511 // If we have a member expression and the current component is a 6512 // reference, we have to map the reference too. Whenever we have a 6513 // reference, the section that reference refers to is going to be a 6514 // load instruction from the storage assigned to the reference. 6515 if (isa<MemberExpr>(I->getAssociatedExpression()) && 6516 I->getAssociatedDeclaration()->getType()->isReferenceType()) { 6517 auto *LI = cast<llvm::LoadInst>(LB); 6518 auto *RefAddr = LI->getPointerOperand(); 6519 6520 BasePointers.push_back(BP); 6521 Pointers.push_back(RefAddr); 6522 Sizes.push_back(CGF.getTypeSize(CGF.getContext().VoidPtrTy)); 6523 Types.push_back(DefaultFlags | 6524 getMapTypeBits( 6525 /*MapType*/ OMPC_MAP_alloc, 6526 /*MapTypeModifier=*/OMPC_MAP_unknown, 6527 !IsExpressionFirstInfo, IsCaptureFirstInfo)); 6528 IsExpressionFirstInfo = false; 6529 IsCaptureFirstInfo = false; 6530 // The reference will be the next base address. 6531 BP = RefAddr; 6532 } 6533 6534 BasePointers.push_back(BP); 6535 Pointers.push_back(LB); 6536 Sizes.push_back(Size); 6537 6538 // We need to add a pointer flag for each map that comes from the 6539 // same expression except for the first one. We also need to signal 6540 // this map is the first one that relates with the current capture 6541 // (there is a set of entries for each capture). 6542 Types.push_back(DefaultFlags | getMapTypeBits(MapType, MapTypeModifier, 6543 !IsExpressionFirstInfo, 6544 IsCaptureFirstInfo)); 6545 6546 // If we have a final array section, we are done with this expression. 6547 if (IsFinalArraySection) 6548 break; 6549 6550 // The pointer becomes the base for the next element. 6551 if (Next != CE) 6552 BP = LB; 6553 6554 IsExpressionFirstInfo = false; 6555 IsCaptureFirstInfo = false; 6556 } 6557 } 6558 } 6559 6560 /// \brief Return the adjusted map modifiers if the declaration a capture 6561 /// refers to appears in a first-private clause. This is expected to be used 6562 /// only with directives that start with 'target'. 6563 unsigned adjustMapModifiersForPrivateClauses(const CapturedStmt::Capture &Cap, 6564 unsigned CurrentModifiers) { 6565 assert(Cap.capturesVariable() && "Expected capture by reference only!"); 6566 6567 // A first private variable captured by reference will use only the 6568 // 'private ptr' and 'map to' flag. Return the right flags if the captured 6569 // declaration is known as first-private in this handler. 6570 if (FirstPrivateDecls.count(Cap.getCapturedVar())) 6571 return MappableExprsHandler::OMP_MAP_PRIVATE | 6572 MappableExprsHandler::OMP_MAP_TO; 6573 // Reduction variable will use only the 'private ptr' and 'map to_from' 6574 // flag. 6575 if (ReductionDecls.count(Cap.getCapturedVar())) { 6576 return MappableExprsHandler::OMP_MAP_TO | 6577 MappableExprsHandler::OMP_MAP_FROM; 6578 } 6579 6580 // We didn't modify anything. 6581 return CurrentModifiers; 6582 } 6583 6584 public: 6585 MappableExprsHandler(const OMPExecutableDirective &Dir, CodeGenFunction &CGF) 6586 : CurDir(Dir), CGF(CGF) { 6587 // Extract firstprivate clause information. 6588 for (const auto *C : Dir.getClausesOfKind<OMPFirstprivateClause>()) 6589 for (const auto *D : C->varlists()) 6590 FirstPrivateDecls.insert( 6591 cast<VarDecl>(cast<DeclRefExpr>(D)->getDecl())->getCanonicalDecl()); 6592 for (const auto *C : Dir.getClausesOfKind<OMPReductionClause>()) { 6593 for (const auto *D : C->varlists()) { 6594 ReductionDecls.insert( 6595 cast<VarDecl>(cast<DeclRefExpr>(D)->getDecl())->getCanonicalDecl()); 6596 } 6597 } 6598 // Extract device pointer clause information. 6599 for (const auto *C : Dir.getClausesOfKind<OMPIsDevicePtrClause>()) 6600 for (auto L : C->component_lists()) 6601 DevPointersMap[L.first].push_back(L.second); 6602 } 6603 6604 /// \brief Generate all the base pointers, section pointers, sizes and map 6605 /// types for the extracted mappable expressions. Also, for each item that 6606 /// relates with a device pointer, a pair of the relevant declaration and 6607 /// index where it occurs is appended to the device pointers info array. 6608 void generateAllInfo(MapBaseValuesArrayTy &BasePointers, 6609 MapValuesArrayTy &Pointers, MapValuesArrayTy &Sizes, 6610 MapFlagsArrayTy &Types) const { 6611 BasePointers.clear(); 6612 Pointers.clear(); 6613 Sizes.clear(); 6614 Types.clear(); 6615 6616 struct MapInfo { 6617 /// Kind that defines how a device pointer has to be returned. 6618 enum ReturnPointerKind { 6619 // Don't have to return any pointer. 6620 RPK_None, 6621 // Pointer is the base of the declaration. 6622 RPK_Base, 6623 // Pointer is a member of the base declaration - 'this' 6624 RPK_Member, 6625 // Pointer is a reference and a member of the base declaration - 'this' 6626 RPK_MemberReference, 6627 }; 6628 OMPClauseMappableExprCommon::MappableExprComponentListRef Components; 6629 OpenMPMapClauseKind MapType = OMPC_MAP_unknown; 6630 OpenMPMapClauseKind MapTypeModifier = OMPC_MAP_unknown; 6631 ReturnPointerKind ReturnDevicePointer = RPK_None; 6632 bool IsImplicit = false; 6633 6634 MapInfo() = default; 6635 MapInfo( 6636 OMPClauseMappableExprCommon::MappableExprComponentListRef Components, 6637 OpenMPMapClauseKind MapType, OpenMPMapClauseKind MapTypeModifier, 6638 ReturnPointerKind ReturnDevicePointer, bool IsImplicit) 6639 : Components(Components), MapType(MapType), 6640 MapTypeModifier(MapTypeModifier), 6641 ReturnDevicePointer(ReturnDevicePointer), IsImplicit(IsImplicit) {} 6642 }; 6643 6644 // We have to process the component lists that relate with the same 6645 // declaration in a single chunk so that we can generate the map flags 6646 // correctly. Therefore, we organize all lists in a map. 6647 llvm::MapVector<const ValueDecl *, SmallVector<MapInfo, 8>> Info; 6648 6649 // Helper function to fill the information map for the different supported 6650 // clauses. 6651 auto &&InfoGen = [&Info]( 6652 const ValueDecl *D, 6653 OMPClauseMappableExprCommon::MappableExprComponentListRef L, 6654 OpenMPMapClauseKind MapType, OpenMPMapClauseKind MapModifier, 6655 MapInfo::ReturnPointerKind ReturnDevicePointer, bool IsImplicit) { 6656 const ValueDecl *VD = 6657 D ? cast<ValueDecl>(D->getCanonicalDecl()) : nullptr; 6658 Info[VD].emplace_back(L, MapType, MapModifier, ReturnDevicePointer, 6659 IsImplicit); 6660 }; 6661 6662 // FIXME: MSVC 2013 seems to require this-> to find member CurDir. 6663 for (auto *C : this->CurDir.getClausesOfKind<OMPMapClause>()) 6664 for (auto L : C->component_lists()) { 6665 InfoGen(L.first, L.second, C->getMapType(), C->getMapTypeModifier(), 6666 MapInfo::RPK_None, C->isImplicit()); 6667 } 6668 for (auto *C : this->CurDir.getClausesOfKind<OMPToClause>()) 6669 for (auto L : C->component_lists()) { 6670 InfoGen(L.first, L.second, OMPC_MAP_to, OMPC_MAP_unknown, 6671 MapInfo::RPK_None, C->isImplicit()); 6672 } 6673 for (auto *C : this->CurDir.getClausesOfKind<OMPFromClause>()) 6674 for (auto L : C->component_lists()) { 6675 InfoGen(L.first, L.second, OMPC_MAP_from, OMPC_MAP_unknown, 6676 MapInfo::RPK_None, C->isImplicit()); 6677 } 6678 6679 // Look at the use_device_ptr clause information and mark the existing map 6680 // entries as such. If there is no map information for an entry in the 6681 // use_device_ptr list, we create one with map type 'alloc' and zero size 6682 // section. It is the user fault if that was not mapped before. 6683 // FIXME: MSVC 2013 seems to require this-> to find member CurDir. 6684 for (auto *C : this->CurDir.getClausesOfKind<OMPUseDevicePtrClause>()) 6685 for (auto L : C->component_lists()) { 6686 assert(!L.second.empty() && "Not expecting empty list of components!"); 6687 const ValueDecl *VD = L.second.back().getAssociatedDeclaration(); 6688 VD = cast<ValueDecl>(VD->getCanonicalDecl()); 6689 auto *IE = L.second.back().getAssociatedExpression(); 6690 // If the first component is a member expression, we have to look into 6691 // 'this', which maps to null in the map of map information. Otherwise 6692 // look directly for the information. 6693 auto It = Info.find(isa<MemberExpr>(IE) ? nullptr : VD); 6694 6695 // We potentially have map information for this declaration already. 6696 // Look for the first set of components that refer to it. 6697 if (It != Info.end()) { 6698 auto CI = std::find_if( 6699 It->second.begin(), It->second.end(), [VD](const MapInfo &MI) { 6700 return MI.Components.back().getAssociatedDeclaration() == VD; 6701 }); 6702 // If we found a map entry, signal that the pointer has to be returned 6703 // and move on to the next declaration. 6704 if (CI != It->second.end()) { 6705 CI->ReturnDevicePointer = isa<MemberExpr>(IE) 6706 ? (VD->getType()->isReferenceType() 6707 ? MapInfo::RPK_MemberReference 6708 : MapInfo::RPK_Member) 6709 : MapInfo::RPK_Base; 6710 continue; 6711 } 6712 } 6713 6714 // We didn't find any match in our map information - generate a zero 6715 // size array section. 6716 // FIXME: MSVC 2013 seems to require this-> to find member CGF. 6717 llvm::Value *Ptr = this->CGF.EmitLoadOfScalar(this->CGF.EmitLValue(IE), 6718 IE->getExprLoc()); 6719 BasePointers.push_back({Ptr, VD}); 6720 Pointers.push_back(Ptr); 6721 Sizes.push_back(llvm::Constant::getNullValue(this->CGF.SizeTy)); 6722 Types.push_back(OMP_MAP_RETURN_PARAM | OMP_MAP_TARGET_PARAM); 6723 } 6724 6725 for (auto &M : Info) { 6726 // We need to know when we generate information for the first component 6727 // associated with a capture, because the mapping flags depend on it. 6728 bool IsFirstComponentList = true; 6729 for (MapInfo &L : M.second) { 6730 assert(!L.Components.empty() && 6731 "Not expecting declaration with no component lists."); 6732 6733 // Remember the current base pointer index. 6734 unsigned CurrentBasePointersIdx = BasePointers.size(); 6735 // FIXME: MSVC 2013 seems to require this-> to find the member method. 6736 this->generateInfoForComponentList( 6737 L.MapType, L.MapTypeModifier, L.Components, BasePointers, Pointers, 6738 Sizes, Types, IsFirstComponentList, L.IsImplicit); 6739 6740 // If this entry relates with a device pointer, set the relevant 6741 // declaration and add the 'return pointer' flag. 6742 if (IsFirstComponentList && 6743 L.ReturnDevicePointer != MapInfo::RPK_None) { 6744 // If the pointer is not the base of the map, we need to skip the 6745 // base. If it is a reference in a member field, we also need to skip 6746 // the map of the reference. 6747 if (L.ReturnDevicePointer != MapInfo::RPK_Base) { 6748 ++CurrentBasePointersIdx; 6749 if (L.ReturnDevicePointer == MapInfo::RPK_MemberReference) 6750 ++CurrentBasePointersIdx; 6751 } 6752 assert(BasePointers.size() > CurrentBasePointersIdx && 6753 "Unexpected number of mapped base pointers."); 6754 6755 auto *RelevantVD = L.Components.back().getAssociatedDeclaration(); 6756 assert(RelevantVD && 6757 "No relevant declaration related with device pointer??"); 6758 6759 BasePointers[CurrentBasePointersIdx].setDevicePtrDecl(RelevantVD); 6760 Types[CurrentBasePointersIdx] |= OMP_MAP_RETURN_PARAM; 6761 } 6762 IsFirstComponentList = false; 6763 } 6764 } 6765 } 6766 6767 /// \brief Generate the base pointers, section pointers, sizes and map types 6768 /// associated to a given capture. 6769 void generateInfoForCapture(const CapturedStmt::Capture *Cap, 6770 llvm::Value *Arg, 6771 MapBaseValuesArrayTy &BasePointers, 6772 MapValuesArrayTy &Pointers, 6773 MapValuesArrayTy &Sizes, 6774 MapFlagsArrayTy &Types) const { 6775 assert(!Cap->capturesVariableArrayType() && 6776 "Not expecting to generate map info for a variable array type!"); 6777 6778 BasePointers.clear(); 6779 Pointers.clear(); 6780 Sizes.clear(); 6781 Types.clear(); 6782 6783 // We need to know when we generating information for the first component 6784 // associated with a capture, because the mapping flags depend on it. 6785 bool IsFirstComponentList = true; 6786 6787 const ValueDecl *VD = 6788 Cap->capturesThis() 6789 ? nullptr 6790 : cast<ValueDecl>(Cap->getCapturedVar()->getCanonicalDecl()); 6791 6792 // If this declaration appears in a is_device_ptr clause we just have to 6793 // pass the pointer by value. If it is a reference to a declaration, we just 6794 // pass its value, otherwise, if it is a member expression, we need to map 6795 // 'to' the field. 6796 if (!VD) { 6797 auto It = DevPointersMap.find(VD); 6798 if (It != DevPointersMap.end()) { 6799 for (auto L : It->second) { 6800 generateInfoForComponentList( 6801 /*MapType=*/OMPC_MAP_to, /*MapTypeModifier=*/OMPC_MAP_unknown, L, 6802 BasePointers, Pointers, Sizes, Types, IsFirstComponentList, 6803 /*IsImplicit=*/false); 6804 IsFirstComponentList = false; 6805 } 6806 return; 6807 } 6808 } else if (DevPointersMap.count(VD)) { 6809 BasePointers.push_back({Arg, VD}); 6810 Pointers.push_back(Arg); 6811 Sizes.push_back(CGF.getTypeSize(CGF.getContext().VoidPtrTy)); 6812 Types.push_back(OMP_MAP_LITERAL | OMP_MAP_TARGET_PARAM); 6813 return; 6814 } 6815 6816 // FIXME: MSVC 2013 seems to require this-> to find member CurDir. 6817 for (auto *C : this->CurDir.getClausesOfKind<OMPMapClause>()) 6818 for (auto L : C->decl_component_lists(VD)) { 6819 assert(L.first == VD && 6820 "We got information for the wrong declaration??"); 6821 assert(!L.second.empty() && 6822 "Not expecting declaration with no component lists."); 6823 generateInfoForComponentList( 6824 C->getMapType(), C->getMapTypeModifier(), L.second, BasePointers, 6825 Pointers, Sizes, Types, IsFirstComponentList, C->isImplicit()); 6826 IsFirstComponentList = false; 6827 } 6828 6829 return; 6830 } 6831 6832 /// \brief Generate the default map information for a given capture \a CI, 6833 /// record field declaration \a RI and captured value \a CV. 6834 void generateDefaultMapInfo(const CapturedStmt::Capture &CI, 6835 const FieldDecl &RI, llvm::Value *CV, 6836 MapBaseValuesArrayTy &CurBasePointers, 6837 MapValuesArrayTy &CurPointers, 6838 MapValuesArrayTy &CurSizes, 6839 MapFlagsArrayTy &CurMapTypes) { 6840 6841 // Do the default mapping. 6842 if (CI.capturesThis()) { 6843 CurBasePointers.push_back(CV); 6844 CurPointers.push_back(CV); 6845 const PointerType *PtrTy = cast<PointerType>(RI.getType().getTypePtr()); 6846 CurSizes.push_back(CGF.getTypeSize(PtrTy->getPointeeType())); 6847 // Default map type. 6848 CurMapTypes.push_back(OMP_MAP_TO | OMP_MAP_FROM); 6849 } else if (CI.capturesVariableByCopy()) { 6850 CurBasePointers.push_back(CV); 6851 CurPointers.push_back(CV); 6852 if (!RI.getType()->isAnyPointerType()) { 6853 // We have to signal to the runtime captures passed by value that are 6854 // not pointers. 6855 CurMapTypes.push_back(OMP_MAP_LITERAL); 6856 CurSizes.push_back(CGF.getTypeSize(RI.getType())); 6857 } else { 6858 // Pointers are implicitly mapped with a zero size and no flags 6859 // (other than first map that is added for all implicit maps). 6860 CurMapTypes.push_back(0u); 6861 CurSizes.push_back(llvm::Constant::getNullValue(CGF.SizeTy)); 6862 } 6863 } else { 6864 assert(CI.capturesVariable() && "Expected captured reference."); 6865 CurBasePointers.push_back(CV); 6866 CurPointers.push_back(CV); 6867 6868 const ReferenceType *PtrTy = 6869 cast<ReferenceType>(RI.getType().getTypePtr()); 6870 QualType ElementType = PtrTy->getPointeeType(); 6871 CurSizes.push_back(CGF.getTypeSize(ElementType)); 6872 // The default map type for a scalar/complex type is 'to' because by 6873 // default the value doesn't have to be retrieved. For an aggregate 6874 // type, the default is 'tofrom'. 6875 CurMapTypes.emplace_back(adjustMapModifiersForPrivateClauses( 6876 CI, ElementType->isAggregateType() ? (OMP_MAP_TO | OMP_MAP_FROM) 6877 : OMP_MAP_TO)); 6878 } 6879 // Every default map produces a single argument which is a target parameter. 6880 CurMapTypes.back() |= OMP_MAP_TARGET_PARAM; 6881 } 6882 }; 6883 6884 enum OpenMPOffloadingReservedDeviceIDs { 6885 /// \brief Device ID if the device was not defined, runtime should get it 6886 /// from environment variables in the spec. 6887 OMP_DEVICEID_UNDEF = -1, 6888 }; 6889 } // anonymous namespace 6890 6891 /// \brief Emit the arrays used to pass the captures and map information to the 6892 /// offloading runtime library. If there is no map or capture information, 6893 /// return nullptr by reference. 6894 static void 6895 emitOffloadingArrays(CodeGenFunction &CGF, 6896 MappableExprsHandler::MapBaseValuesArrayTy &BasePointers, 6897 MappableExprsHandler::MapValuesArrayTy &Pointers, 6898 MappableExprsHandler::MapValuesArrayTy &Sizes, 6899 MappableExprsHandler::MapFlagsArrayTy &MapTypes, 6900 CGOpenMPRuntime::TargetDataInfo &Info) { 6901 auto &CGM = CGF.CGM; 6902 auto &Ctx = CGF.getContext(); 6903 6904 // Reset the array information. 6905 Info.clearArrayInfo(); 6906 Info.NumberOfPtrs = BasePointers.size(); 6907 6908 if (Info.NumberOfPtrs) { 6909 // Detect if we have any capture size requiring runtime evaluation of the 6910 // size so that a constant array could be eventually used. 6911 bool hasRuntimeEvaluationCaptureSize = false; 6912 for (auto *S : Sizes) 6913 if (!isa<llvm::Constant>(S)) { 6914 hasRuntimeEvaluationCaptureSize = true; 6915 break; 6916 } 6917 6918 llvm::APInt PointerNumAP(32, Info.NumberOfPtrs, /*isSigned=*/true); 6919 QualType PointerArrayType = 6920 Ctx.getConstantArrayType(Ctx.VoidPtrTy, PointerNumAP, ArrayType::Normal, 6921 /*IndexTypeQuals=*/0); 6922 6923 Info.BasePointersArray = 6924 CGF.CreateMemTemp(PointerArrayType, ".offload_baseptrs").getPointer(); 6925 Info.PointersArray = 6926 CGF.CreateMemTemp(PointerArrayType, ".offload_ptrs").getPointer(); 6927 6928 // If we don't have any VLA types or other types that require runtime 6929 // evaluation, we can use a constant array for the map sizes, otherwise we 6930 // need to fill up the arrays as we do for the pointers. 6931 if (hasRuntimeEvaluationCaptureSize) { 6932 QualType SizeArrayType = Ctx.getConstantArrayType( 6933 Ctx.getSizeType(), PointerNumAP, ArrayType::Normal, 6934 /*IndexTypeQuals=*/0); 6935 Info.SizesArray = 6936 CGF.CreateMemTemp(SizeArrayType, ".offload_sizes").getPointer(); 6937 } else { 6938 // We expect all the sizes to be constant, so we collect them to create 6939 // a constant array. 6940 SmallVector<llvm::Constant *, 16> ConstSizes; 6941 for (auto S : Sizes) 6942 ConstSizes.push_back(cast<llvm::Constant>(S)); 6943 6944 auto *SizesArrayInit = llvm::ConstantArray::get( 6945 llvm::ArrayType::get(CGM.SizeTy, ConstSizes.size()), ConstSizes); 6946 auto *SizesArrayGbl = new llvm::GlobalVariable( 6947 CGM.getModule(), SizesArrayInit->getType(), 6948 /*isConstant=*/true, llvm::GlobalValue::PrivateLinkage, 6949 SizesArrayInit, ".offload_sizes"); 6950 SizesArrayGbl->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global); 6951 Info.SizesArray = SizesArrayGbl; 6952 } 6953 6954 // The map types are always constant so we don't need to generate code to 6955 // fill arrays. Instead, we create an array constant. 6956 llvm::Constant *MapTypesArrayInit = 6957 llvm::ConstantDataArray::get(CGF.Builder.getContext(), MapTypes); 6958 auto *MapTypesArrayGbl = new llvm::GlobalVariable( 6959 CGM.getModule(), MapTypesArrayInit->getType(), 6960 /*isConstant=*/true, llvm::GlobalValue::PrivateLinkage, 6961 MapTypesArrayInit, ".offload_maptypes"); 6962 MapTypesArrayGbl->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global); 6963 Info.MapTypesArray = MapTypesArrayGbl; 6964 6965 for (unsigned i = 0; i < Info.NumberOfPtrs; ++i) { 6966 llvm::Value *BPVal = *BasePointers[i]; 6967 llvm::Value *BP = CGF.Builder.CreateConstInBoundsGEP2_32( 6968 llvm::ArrayType::get(CGM.VoidPtrTy, Info.NumberOfPtrs), 6969 Info.BasePointersArray, 0, i); 6970 BP = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 6971 BP, BPVal->getType()->getPointerTo(/*AddrSpace=*/0)); 6972 Address BPAddr(BP, Ctx.getTypeAlignInChars(Ctx.VoidPtrTy)); 6973 CGF.Builder.CreateStore(BPVal, BPAddr); 6974 6975 if (Info.requiresDevicePointerInfo()) 6976 if (auto *DevVD = BasePointers[i].getDevicePtrDecl()) 6977 Info.CaptureDeviceAddrMap.insert(std::make_pair(DevVD, BPAddr)); 6978 6979 llvm::Value *PVal = Pointers[i]; 6980 llvm::Value *P = CGF.Builder.CreateConstInBoundsGEP2_32( 6981 llvm::ArrayType::get(CGM.VoidPtrTy, Info.NumberOfPtrs), 6982 Info.PointersArray, 0, i); 6983 P = CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 6984 P, PVal->getType()->getPointerTo(/*AddrSpace=*/0)); 6985 Address PAddr(P, Ctx.getTypeAlignInChars(Ctx.VoidPtrTy)); 6986 CGF.Builder.CreateStore(PVal, PAddr); 6987 6988 if (hasRuntimeEvaluationCaptureSize) { 6989 llvm::Value *S = CGF.Builder.CreateConstInBoundsGEP2_32( 6990 llvm::ArrayType::get(CGM.SizeTy, Info.NumberOfPtrs), 6991 Info.SizesArray, 6992 /*Idx0=*/0, 6993 /*Idx1=*/i); 6994 Address SAddr(S, Ctx.getTypeAlignInChars(Ctx.getSizeType())); 6995 CGF.Builder.CreateStore( 6996 CGF.Builder.CreateIntCast(Sizes[i], CGM.SizeTy, /*isSigned=*/true), 6997 SAddr); 6998 } 6999 } 7000 } 7001 } 7002 /// \brief Emit the arguments to be passed to the runtime library based on the 7003 /// arrays of pointers, sizes and map types. 7004 static void emitOffloadingArraysArgument( 7005 CodeGenFunction &CGF, llvm::Value *&BasePointersArrayArg, 7006 llvm::Value *&PointersArrayArg, llvm::Value *&SizesArrayArg, 7007 llvm::Value *&MapTypesArrayArg, CGOpenMPRuntime::TargetDataInfo &Info) { 7008 auto &CGM = CGF.CGM; 7009 if (Info.NumberOfPtrs) { 7010 BasePointersArrayArg = CGF.Builder.CreateConstInBoundsGEP2_32( 7011 llvm::ArrayType::get(CGM.VoidPtrTy, Info.NumberOfPtrs), 7012 Info.BasePointersArray, 7013 /*Idx0=*/0, /*Idx1=*/0); 7014 PointersArrayArg = CGF.Builder.CreateConstInBoundsGEP2_32( 7015 llvm::ArrayType::get(CGM.VoidPtrTy, Info.NumberOfPtrs), 7016 Info.PointersArray, 7017 /*Idx0=*/0, 7018 /*Idx1=*/0); 7019 SizesArrayArg = CGF.Builder.CreateConstInBoundsGEP2_32( 7020 llvm::ArrayType::get(CGM.SizeTy, Info.NumberOfPtrs), Info.SizesArray, 7021 /*Idx0=*/0, /*Idx1=*/0); 7022 MapTypesArrayArg = CGF.Builder.CreateConstInBoundsGEP2_32( 7023 llvm::ArrayType::get(CGM.Int64Ty, Info.NumberOfPtrs), 7024 Info.MapTypesArray, 7025 /*Idx0=*/0, 7026 /*Idx1=*/0); 7027 } else { 7028 BasePointersArrayArg = llvm::ConstantPointerNull::get(CGM.VoidPtrPtrTy); 7029 PointersArrayArg = llvm::ConstantPointerNull::get(CGM.VoidPtrPtrTy); 7030 SizesArrayArg = llvm::ConstantPointerNull::get(CGM.SizeTy->getPointerTo()); 7031 MapTypesArrayArg = 7032 llvm::ConstantPointerNull::get(CGM.Int64Ty->getPointerTo()); 7033 } 7034 } 7035 7036 void CGOpenMPRuntime::emitTargetCall(CodeGenFunction &CGF, 7037 const OMPExecutableDirective &D, 7038 llvm::Value *OutlinedFn, 7039 llvm::Value *OutlinedFnID, 7040 const Expr *IfCond, const Expr *Device) { 7041 if (!CGF.HaveInsertPoint()) 7042 return; 7043 7044 assert(OutlinedFn && "Invalid outlined function!"); 7045 7046 const bool RequiresOuterTask = D.hasClausesOfKind<OMPDependClause>(); 7047 llvm::SmallVector<llvm::Value *, 16> CapturedVars; 7048 const CapturedStmt &CS = *D.getCapturedStmt(OMPD_target); 7049 auto &&ArgsCodegen = [&CS, &CapturedVars](CodeGenFunction &CGF, 7050 PrePostActionTy &) { 7051 CGF.GenerateOpenMPCapturedVars(CS, CapturedVars); 7052 }; 7053 emitInlinedDirective(CGF, OMPD_unknown, ArgsCodegen); 7054 7055 CodeGenFunction::OMPTargetDataInfo InputInfo; 7056 llvm::Value *MapTypesArray = nullptr; 7057 // Fill up the pointer arrays and transfer execution to the device. 7058 auto &&ThenGen = [this, Device, OutlinedFn, OutlinedFnID, &D, &InputInfo, 7059 &MapTypesArray, &CS, RequiresOuterTask, 7060 &CapturedVars](CodeGenFunction &CGF, PrePostActionTy &) { 7061 // On top of the arrays that were filled up, the target offloading call 7062 // takes as arguments the device id as well as the host pointer. The host 7063 // pointer is used by the runtime library to identify the current target 7064 // region, so it only has to be unique and not necessarily point to 7065 // anything. It could be the pointer to the outlined function that 7066 // implements the target region, but we aren't using that so that the 7067 // compiler doesn't need to keep that, and could therefore inline the host 7068 // function if proven worthwhile during optimization. 7069 7070 // From this point on, we need to have an ID of the target region defined. 7071 assert(OutlinedFnID && "Invalid outlined function ID!"); 7072 7073 // Emit device ID if any. 7074 llvm::Value *DeviceID; 7075 if (Device) { 7076 DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device), 7077 CGF.Int64Ty, /*isSigned=*/true); 7078 } else { 7079 DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF); 7080 } 7081 7082 // Emit the number of elements in the offloading arrays. 7083 llvm::Value *PointerNum = 7084 CGF.Builder.getInt32(InputInfo.NumberOfTargetItems); 7085 7086 // Return value of the runtime offloading call. 7087 llvm::Value *Return; 7088 7089 auto *NumTeams = emitNumTeamsForTargetDirective(*this, CGF, D); 7090 auto *NumThreads = emitNumThreadsForTargetDirective(*this, CGF, D); 7091 7092 bool HasNowait = D.hasClausesOfKind<OMPNowaitClause>(); 7093 // The target region is an outlined function launched by the runtime 7094 // via calls __tgt_target() or __tgt_target_teams(). 7095 // 7096 // __tgt_target() launches a target region with one team and one thread, 7097 // executing a serial region. This master thread may in turn launch 7098 // more threads within its team upon encountering a parallel region, 7099 // however, no additional teams can be launched on the device. 7100 // 7101 // __tgt_target_teams() launches a target region with one or more teams, 7102 // each with one or more threads. This call is required for target 7103 // constructs such as: 7104 // 'target teams' 7105 // 'target' / 'teams' 7106 // 'target teams distribute parallel for' 7107 // 'target parallel' 7108 // and so on. 7109 // 7110 // Note that on the host and CPU targets, the runtime implementation of 7111 // these calls simply call the outlined function without forking threads. 7112 // The outlined functions themselves have runtime calls to 7113 // __kmpc_fork_teams() and __kmpc_fork() for this purpose, codegen'd by 7114 // the compiler in emitTeamsCall() and emitParallelCall(). 7115 // 7116 // In contrast, on the NVPTX target, the implementation of 7117 // __tgt_target_teams() launches a GPU kernel with the requested number 7118 // of teams and threads so no additional calls to the runtime are required. 7119 if (NumTeams) { 7120 // If we have NumTeams defined this means that we have an enclosed teams 7121 // region. Therefore we also expect to have NumThreads defined. These two 7122 // values should be defined in the presence of a teams directive, 7123 // regardless of having any clauses associated. If the user is using teams 7124 // but no clauses, these two values will be the default that should be 7125 // passed to the runtime library - a 32-bit integer with the value zero. 7126 assert(NumThreads && "Thread limit expression should be available along " 7127 "with number of teams."); 7128 llvm::Value *OffloadingArgs[] = {DeviceID, 7129 OutlinedFnID, 7130 PointerNum, 7131 InputInfo.BasePointersArray.getPointer(), 7132 InputInfo.PointersArray.getPointer(), 7133 InputInfo.SizesArray.getPointer(), 7134 MapTypesArray, 7135 NumTeams, 7136 NumThreads}; 7137 Return = CGF.EmitRuntimeCall( 7138 createRuntimeFunction(HasNowait ? OMPRTL__tgt_target_teams_nowait 7139 : OMPRTL__tgt_target_teams), 7140 OffloadingArgs); 7141 } else { 7142 llvm::Value *OffloadingArgs[] = {DeviceID, 7143 OutlinedFnID, 7144 PointerNum, 7145 InputInfo.BasePointersArray.getPointer(), 7146 InputInfo.PointersArray.getPointer(), 7147 InputInfo.SizesArray.getPointer(), 7148 MapTypesArray}; 7149 Return = CGF.EmitRuntimeCall( 7150 createRuntimeFunction(HasNowait ? OMPRTL__tgt_target_nowait 7151 : OMPRTL__tgt_target), 7152 OffloadingArgs); 7153 } 7154 7155 // Check the error code and execute the host version if required. 7156 llvm::BasicBlock *OffloadFailedBlock = 7157 CGF.createBasicBlock("omp_offload.failed"); 7158 llvm::BasicBlock *OffloadContBlock = 7159 CGF.createBasicBlock("omp_offload.cont"); 7160 llvm::Value *Failed = CGF.Builder.CreateIsNotNull(Return); 7161 CGF.Builder.CreateCondBr(Failed, OffloadFailedBlock, OffloadContBlock); 7162 7163 CGF.EmitBlock(OffloadFailedBlock); 7164 if (RequiresOuterTask) { 7165 CapturedVars.clear(); 7166 CGF.GenerateOpenMPCapturedVars(CS, CapturedVars); 7167 } 7168 emitOutlinedFunctionCall(CGF, D.getLocStart(), OutlinedFn, CapturedVars); 7169 CGF.EmitBranch(OffloadContBlock); 7170 7171 CGF.EmitBlock(OffloadContBlock, /*IsFinished=*/true); 7172 }; 7173 7174 // Notify that the host version must be executed. 7175 auto &&ElseGen = [this, &D, OutlinedFn, &CS, &CapturedVars, 7176 RequiresOuterTask](CodeGenFunction &CGF, 7177 PrePostActionTy &) { 7178 if (RequiresOuterTask) { 7179 CapturedVars.clear(); 7180 CGF.GenerateOpenMPCapturedVars(CS, CapturedVars); 7181 } 7182 emitOutlinedFunctionCall(CGF, D.getLocStart(), OutlinedFn, CapturedVars); 7183 }; 7184 7185 auto &&TargetThenGen = [this, &ThenGen, &D, &InputInfo, &MapTypesArray, 7186 &CapturedVars, RequiresOuterTask, 7187 &CS](CodeGenFunction &CGF, PrePostActionTy &) { 7188 // Fill up the arrays with all the captured variables. 7189 MappableExprsHandler::MapBaseValuesArrayTy BasePointers; 7190 MappableExprsHandler::MapValuesArrayTy Pointers; 7191 MappableExprsHandler::MapValuesArrayTy Sizes; 7192 MappableExprsHandler::MapFlagsArrayTy MapTypes; 7193 7194 MappableExprsHandler::MapBaseValuesArrayTy CurBasePointers; 7195 MappableExprsHandler::MapValuesArrayTy CurPointers; 7196 MappableExprsHandler::MapValuesArrayTy CurSizes; 7197 MappableExprsHandler::MapFlagsArrayTy CurMapTypes; 7198 7199 // Get mappable expression information. 7200 MappableExprsHandler MEHandler(D, CGF); 7201 7202 auto RI = CS.getCapturedRecordDecl()->field_begin(); 7203 auto CV = CapturedVars.begin(); 7204 for (CapturedStmt::const_capture_iterator CI = CS.capture_begin(), 7205 CE = CS.capture_end(); 7206 CI != CE; ++CI, ++RI, ++CV) { 7207 CurBasePointers.clear(); 7208 CurPointers.clear(); 7209 CurSizes.clear(); 7210 CurMapTypes.clear(); 7211 7212 // VLA sizes are passed to the outlined region by copy and do not have map 7213 // information associated. 7214 if (CI->capturesVariableArrayType()) { 7215 CurBasePointers.push_back(*CV); 7216 CurPointers.push_back(*CV); 7217 CurSizes.push_back(CGF.getTypeSize(RI->getType())); 7218 // Copy to the device as an argument. No need to retrieve it. 7219 CurMapTypes.push_back(MappableExprsHandler::OMP_MAP_LITERAL | 7220 MappableExprsHandler::OMP_MAP_TARGET_PARAM); 7221 } else { 7222 // If we have any information in the map clause, we use it, otherwise we 7223 // just do a default mapping. 7224 MEHandler.generateInfoForCapture(CI, *CV, CurBasePointers, CurPointers, 7225 CurSizes, CurMapTypes); 7226 if (CurBasePointers.empty()) 7227 MEHandler.generateDefaultMapInfo(*CI, **RI, *CV, CurBasePointers, 7228 CurPointers, CurSizes, CurMapTypes); 7229 } 7230 // We expect to have at least an element of information for this capture. 7231 assert(!CurBasePointers.empty() && 7232 "Non-existing map pointer for capture!"); 7233 assert(CurBasePointers.size() == CurPointers.size() && 7234 CurBasePointers.size() == CurSizes.size() && 7235 CurBasePointers.size() == CurMapTypes.size() && 7236 "Inconsistent map information sizes!"); 7237 7238 // We need to append the results of this capture to what we already have. 7239 BasePointers.append(CurBasePointers.begin(), CurBasePointers.end()); 7240 Pointers.append(CurPointers.begin(), CurPointers.end()); 7241 Sizes.append(CurSizes.begin(), CurSizes.end()); 7242 MapTypes.append(CurMapTypes.begin(), CurMapTypes.end()); 7243 } 7244 7245 TargetDataInfo Info; 7246 // Fill up the arrays and create the arguments. 7247 emitOffloadingArrays(CGF, BasePointers, Pointers, Sizes, MapTypes, Info); 7248 emitOffloadingArraysArgument(CGF, Info.BasePointersArray, 7249 Info.PointersArray, Info.SizesArray, 7250 Info.MapTypesArray, Info); 7251 InputInfo.NumberOfTargetItems = Info.NumberOfPtrs; 7252 InputInfo.BasePointersArray = 7253 Address(Info.BasePointersArray, CGM.getPointerAlign()); 7254 InputInfo.PointersArray = 7255 Address(Info.PointersArray, CGM.getPointerAlign()); 7256 InputInfo.SizesArray = Address(Info.SizesArray, CGM.getPointerAlign()); 7257 MapTypesArray = Info.MapTypesArray; 7258 if (RequiresOuterTask) 7259 CGF.EmitOMPTargetTaskBasedDirective(D, ThenGen, InputInfo); 7260 else 7261 emitInlinedDirective(CGF, D.getDirectiveKind(), ThenGen); 7262 }; 7263 7264 auto &&TargetElseGen = [this, &ElseGen, &D, RequiresOuterTask]( 7265 CodeGenFunction &CGF, PrePostActionTy &) { 7266 if (RequiresOuterTask) { 7267 CodeGenFunction::OMPTargetDataInfo InputInfo; 7268 CGF.EmitOMPTargetTaskBasedDirective(D, ElseGen, InputInfo); 7269 } else { 7270 emitInlinedDirective(CGF, D.getDirectiveKind(), ElseGen); 7271 } 7272 }; 7273 7274 // If we have a target function ID it means that we need to support 7275 // offloading, otherwise, just execute on the host. We need to execute on host 7276 // regardless of the conditional in the if clause if, e.g., the user do not 7277 // specify target triples. 7278 if (OutlinedFnID) { 7279 if (IfCond) { 7280 emitOMPIfClause(CGF, IfCond, TargetThenGen, TargetElseGen); 7281 } else { 7282 RegionCodeGenTy ThenRCG(TargetThenGen); 7283 ThenRCG(CGF); 7284 } 7285 } else { 7286 RegionCodeGenTy ElseRCG(TargetElseGen); 7287 ElseRCG(CGF); 7288 } 7289 } 7290 7291 void CGOpenMPRuntime::scanForTargetRegionsFunctions(const Stmt *S, 7292 StringRef ParentName) { 7293 if (!S) 7294 return; 7295 7296 // Codegen OMP target directives that offload compute to the device. 7297 bool requiresDeviceCodegen = 7298 isa<OMPExecutableDirective>(S) && 7299 isOpenMPTargetExecutionDirective( 7300 cast<OMPExecutableDirective>(S)->getDirectiveKind()); 7301 7302 if (requiresDeviceCodegen) { 7303 auto &E = *cast<OMPExecutableDirective>(S); 7304 unsigned DeviceID; 7305 unsigned FileID; 7306 unsigned Line; 7307 getTargetEntryUniqueInfo(CGM.getContext(), E.getLocStart(), DeviceID, 7308 FileID, Line); 7309 7310 // Is this a target region that should not be emitted as an entry point? If 7311 // so just signal we are done with this target region. 7312 if (!OffloadEntriesInfoManager.hasTargetRegionEntryInfo(DeviceID, FileID, 7313 ParentName, Line)) 7314 return; 7315 7316 switch (S->getStmtClass()) { 7317 case Stmt::OMPTargetDirectiveClass: 7318 CodeGenFunction::EmitOMPTargetDeviceFunction( 7319 CGM, ParentName, cast<OMPTargetDirective>(*S)); 7320 break; 7321 case Stmt::OMPTargetParallelDirectiveClass: 7322 CodeGenFunction::EmitOMPTargetParallelDeviceFunction( 7323 CGM, ParentName, cast<OMPTargetParallelDirective>(*S)); 7324 break; 7325 case Stmt::OMPTargetTeamsDirectiveClass: 7326 CodeGenFunction::EmitOMPTargetTeamsDeviceFunction( 7327 CGM, ParentName, cast<OMPTargetTeamsDirective>(*S)); 7328 break; 7329 case Stmt::OMPTargetTeamsDistributeDirectiveClass: 7330 CodeGenFunction::EmitOMPTargetTeamsDistributeDeviceFunction( 7331 CGM, ParentName, cast<OMPTargetTeamsDistributeDirective>(*S)); 7332 break; 7333 case Stmt::OMPTargetTeamsDistributeSimdDirectiveClass: 7334 CodeGenFunction::EmitOMPTargetTeamsDistributeSimdDeviceFunction( 7335 CGM, ParentName, cast<OMPTargetTeamsDistributeSimdDirective>(*S)); 7336 break; 7337 case Stmt::OMPTargetParallelForDirectiveClass: 7338 CodeGenFunction::EmitOMPTargetParallelForDeviceFunction( 7339 CGM, ParentName, cast<OMPTargetParallelForDirective>(*S)); 7340 break; 7341 case Stmt::OMPTargetParallelForSimdDirectiveClass: 7342 CodeGenFunction::EmitOMPTargetParallelForSimdDeviceFunction( 7343 CGM, ParentName, cast<OMPTargetParallelForSimdDirective>(*S)); 7344 break; 7345 case Stmt::OMPTargetSimdDirectiveClass: 7346 CodeGenFunction::EmitOMPTargetSimdDeviceFunction( 7347 CGM, ParentName, cast<OMPTargetSimdDirective>(*S)); 7348 break; 7349 case Stmt::OMPTargetTeamsDistributeParallelForDirectiveClass: 7350 CodeGenFunction::EmitOMPTargetTeamsDistributeParallelForDeviceFunction( 7351 CGM, ParentName, 7352 cast<OMPTargetTeamsDistributeParallelForDirective>(*S)); 7353 break; 7354 case Stmt::OMPTargetTeamsDistributeParallelForSimdDirectiveClass: 7355 CodeGenFunction:: 7356 EmitOMPTargetTeamsDistributeParallelForSimdDeviceFunction( 7357 CGM, ParentName, 7358 cast<OMPTargetTeamsDistributeParallelForSimdDirective>(*S)); 7359 break; 7360 default: 7361 llvm_unreachable("Unknown target directive for OpenMP device codegen."); 7362 } 7363 return; 7364 } 7365 7366 if (const OMPExecutableDirective *E = dyn_cast<OMPExecutableDirective>(S)) { 7367 if (!E->hasAssociatedStmt() || !E->getAssociatedStmt()) 7368 return; 7369 7370 scanForTargetRegionsFunctions( 7371 E->getInnermostCapturedStmt()->getCapturedStmt(), ParentName); 7372 return; 7373 } 7374 7375 // If this is a lambda function, look into its body. 7376 if (auto *L = dyn_cast<LambdaExpr>(S)) 7377 S = L->getBody(); 7378 7379 // Keep looking for target regions recursively. 7380 for (auto *II : S->children()) 7381 scanForTargetRegionsFunctions(II, ParentName); 7382 } 7383 7384 bool CGOpenMPRuntime::emitTargetFunctions(GlobalDecl GD) { 7385 auto &FD = *cast<FunctionDecl>(GD.getDecl()); 7386 7387 // If emitting code for the host, we do not process FD here. Instead we do 7388 // the normal code generation. 7389 if (!CGM.getLangOpts().OpenMPIsDevice) 7390 return false; 7391 7392 // Try to detect target regions in the function. 7393 scanForTargetRegionsFunctions(FD.getBody(), CGM.getMangledName(GD)); 7394 7395 // We should not emit any function other that the ones created during the 7396 // scanning. Therefore, we signal that this function is completely dealt 7397 // with. 7398 return true; 7399 } 7400 7401 bool CGOpenMPRuntime::emitTargetGlobalVariable(GlobalDecl GD) { 7402 if (!CGM.getLangOpts().OpenMPIsDevice) 7403 return false; 7404 7405 // Check if there are Ctors/Dtors in this declaration and look for target 7406 // regions in it. We use the complete variant to produce the kernel name 7407 // mangling. 7408 QualType RDTy = cast<VarDecl>(GD.getDecl())->getType(); 7409 if (auto *RD = RDTy->getBaseElementTypeUnsafe()->getAsCXXRecordDecl()) { 7410 for (auto *Ctor : RD->ctors()) { 7411 StringRef ParentName = 7412 CGM.getMangledName(GlobalDecl(Ctor, Ctor_Complete)); 7413 scanForTargetRegionsFunctions(Ctor->getBody(), ParentName); 7414 } 7415 auto *Dtor = RD->getDestructor(); 7416 if (Dtor) { 7417 StringRef ParentName = 7418 CGM.getMangledName(GlobalDecl(Dtor, Dtor_Complete)); 7419 scanForTargetRegionsFunctions(Dtor->getBody(), ParentName); 7420 } 7421 } 7422 7423 // If we are in target mode, we do not emit any global (declare target is not 7424 // implemented yet). Therefore we signal that GD was processed in this case. 7425 return true; 7426 } 7427 7428 bool CGOpenMPRuntime::emitTargetGlobal(GlobalDecl GD) { 7429 auto *VD = GD.getDecl(); 7430 if (isa<FunctionDecl>(VD)) 7431 return emitTargetFunctions(GD); 7432 7433 return emitTargetGlobalVariable(GD); 7434 } 7435 7436 llvm::Function *CGOpenMPRuntime::emitRegistrationFunction() { 7437 // If we have offloading in the current module, we need to emit the entries 7438 // now and register the offloading descriptor. 7439 createOffloadEntriesAndInfoMetadata(); 7440 7441 // Create and register the offloading binary descriptors. This is the main 7442 // entity that captures all the information about offloading in the current 7443 // compilation unit. 7444 return createOffloadingBinaryDescriptorRegistration(); 7445 } 7446 7447 void CGOpenMPRuntime::emitTeamsCall(CodeGenFunction &CGF, 7448 const OMPExecutableDirective &D, 7449 SourceLocation Loc, 7450 llvm::Value *OutlinedFn, 7451 ArrayRef<llvm::Value *> CapturedVars) { 7452 if (!CGF.HaveInsertPoint()) 7453 return; 7454 7455 auto *RTLoc = emitUpdateLocation(CGF, Loc); 7456 CodeGenFunction::RunCleanupsScope Scope(CGF); 7457 7458 // Build call __kmpc_fork_teams(loc, n, microtask, var1, .., varn); 7459 llvm::Value *Args[] = { 7460 RTLoc, 7461 CGF.Builder.getInt32(CapturedVars.size()), // Number of captured vars 7462 CGF.Builder.CreateBitCast(OutlinedFn, getKmpc_MicroPointerTy())}; 7463 llvm::SmallVector<llvm::Value *, 16> RealArgs; 7464 RealArgs.append(std::begin(Args), std::end(Args)); 7465 RealArgs.append(CapturedVars.begin(), CapturedVars.end()); 7466 7467 auto RTLFn = createRuntimeFunction(OMPRTL__kmpc_fork_teams); 7468 CGF.EmitRuntimeCall(RTLFn, RealArgs); 7469 } 7470 7471 void CGOpenMPRuntime::emitNumTeamsClause(CodeGenFunction &CGF, 7472 const Expr *NumTeams, 7473 const Expr *ThreadLimit, 7474 SourceLocation Loc) { 7475 if (!CGF.HaveInsertPoint()) 7476 return; 7477 7478 auto *RTLoc = emitUpdateLocation(CGF, Loc); 7479 7480 llvm::Value *NumTeamsVal = 7481 (NumTeams) 7482 ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(NumTeams), 7483 CGF.CGM.Int32Ty, /* isSigned = */ true) 7484 : CGF.Builder.getInt32(0); 7485 7486 llvm::Value *ThreadLimitVal = 7487 (ThreadLimit) 7488 ? CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(ThreadLimit), 7489 CGF.CGM.Int32Ty, /* isSigned = */ true) 7490 : CGF.Builder.getInt32(0); 7491 7492 // Build call __kmpc_push_num_teamss(&loc, global_tid, num_teams, thread_limit) 7493 llvm::Value *PushNumTeamsArgs[] = {RTLoc, getThreadID(CGF, Loc), NumTeamsVal, 7494 ThreadLimitVal}; 7495 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__kmpc_push_num_teams), 7496 PushNumTeamsArgs); 7497 } 7498 7499 void CGOpenMPRuntime::emitTargetDataCalls( 7500 CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond, 7501 const Expr *Device, const RegionCodeGenTy &CodeGen, TargetDataInfo &Info) { 7502 if (!CGF.HaveInsertPoint()) 7503 return; 7504 7505 // Action used to replace the default codegen action and turn privatization 7506 // off. 7507 PrePostActionTy NoPrivAction; 7508 7509 // Generate the code for the opening of the data environment. Capture all the 7510 // arguments of the runtime call by reference because they are used in the 7511 // closing of the region. 7512 auto &&BeginThenGen = [this, &D, Device, &Info, 7513 &CodeGen](CodeGenFunction &CGF, PrePostActionTy &) { 7514 // Fill up the arrays with all the mapped variables. 7515 MappableExprsHandler::MapBaseValuesArrayTy BasePointers; 7516 MappableExprsHandler::MapValuesArrayTy Pointers; 7517 MappableExprsHandler::MapValuesArrayTy Sizes; 7518 MappableExprsHandler::MapFlagsArrayTy MapTypes; 7519 7520 // Get map clause information. 7521 MappableExprsHandler MCHandler(D, CGF); 7522 MCHandler.generateAllInfo(BasePointers, Pointers, Sizes, MapTypes); 7523 7524 // Fill up the arrays and create the arguments. 7525 emitOffloadingArrays(CGF, BasePointers, Pointers, Sizes, MapTypes, Info); 7526 7527 llvm::Value *BasePointersArrayArg = nullptr; 7528 llvm::Value *PointersArrayArg = nullptr; 7529 llvm::Value *SizesArrayArg = nullptr; 7530 llvm::Value *MapTypesArrayArg = nullptr; 7531 emitOffloadingArraysArgument(CGF, BasePointersArrayArg, PointersArrayArg, 7532 SizesArrayArg, MapTypesArrayArg, Info); 7533 7534 // Emit device ID if any. 7535 llvm::Value *DeviceID = nullptr; 7536 if (Device) { 7537 DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device), 7538 CGF.Int64Ty, /*isSigned=*/true); 7539 } else { 7540 DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF); 7541 } 7542 7543 // Emit the number of elements in the offloading arrays. 7544 auto *PointerNum = CGF.Builder.getInt32(Info.NumberOfPtrs); 7545 7546 llvm::Value *OffloadingArgs[] = { 7547 DeviceID, PointerNum, BasePointersArrayArg, 7548 PointersArrayArg, SizesArrayArg, MapTypesArrayArg}; 7549 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__tgt_target_data_begin), 7550 OffloadingArgs); 7551 7552 // If device pointer privatization is required, emit the body of the region 7553 // here. It will have to be duplicated: with and without privatization. 7554 if (!Info.CaptureDeviceAddrMap.empty()) 7555 CodeGen(CGF); 7556 }; 7557 7558 // Generate code for the closing of the data region. 7559 auto &&EndThenGen = [this, Device, &Info](CodeGenFunction &CGF, 7560 PrePostActionTy &) { 7561 assert(Info.isValid() && "Invalid data environment closing arguments."); 7562 7563 llvm::Value *BasePointersArrayArg = nullptr; 7564 llvm::Value *PointersArrayArg = nullptr; 7565 llvm::Value *SizesArrayArg = nullptr; 7566 llvm::Value *MapTypesArrayArg = nullptr; 7567 emitOffloadingArraysArgument(CGF, BasePointersArrayArg, PointersArrayArg, 7568 SizesArrayArg, MapTypesArrayArg, Info); 7569 7570 // Emit device ID if any. 7571 llvm::Value *DeviceID = nullptr; 7572 if (Device) { 7573 DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device), 7574 CGF.Int64Ty, /*isSigned=*/true); 7575 } else { 7576 DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF); 7577 } 7578 7579 // Emit the number of elements in the offloading arrays. 7580 auto *PointerNum = CGF.Builder.getInt32(Info.NumberOfPtrs); 7581 7582 llvm::Value *OffloadingArgs[] = { 7583 DeviceID, PointerNum, BasePointersArrayArg, 7584 PointersArrayArg, SizesArrayArg, MapTypesArrayArg}; 7585 CGF.EmitRuntimeCall(createRuntimeFunction(OMPRTL__tgt_target_data_end), 7586 OffloadingArgs); 7587 }; 7588 7589 // If we need device pointer privatization, we need to emit the body of the 7590 // region with no privatization in the 'else' branch of the conditional. 7591 // Otherwise, we don't have to do anything. 7592 auto &&BeginElseGen = [&Info, &CodeGen, &NoPrivAction](CodeGenFunction &CGF, 7593 PrePostActionTy &) { 7594 if (!Info.CaptureDeviceAddrMap.empty()) { 7595 CodeGen.setAction(NoPrivAction); 7596 CodeGen(CGF); 7597 } 7598 }; 7599 7600 // We don't have to do anything to close the region if the if clause evaluates 7601 // to false. 7602 auto &&EndElseGen = [](CodeGenFunction &CGF, PrePostActionTy &) {}; 7603 7604 if (IfCond) { 7605 emitOMPIfClause(CGF, IfCond, BeginThenGen, BeginElseGen); 7606 } else { 7607 RegionCodeGenTy RCG(BeginThenGen); 7608 RCG(CGF); 7609 } 7610 7611 // If we don't require privatization of device pointers, we emit the body in 7612 // between the runtime calls. This avoids duplicating the body code. 7613 if (Info.CaptureDeviceAddrMap.empty()) { 7614 CodeGen.setAction(NoPrivAction); 7615 CodeGen(CGF); 7616 } 7617 7618 if (IfCond) { 7619 emitOMPIfClause(CGF, IfCond, EndThenGen, EndElseGen); 7620 } else { 7621 RegionCodeGenTy RCG(EndThenGen); 7622 RCG(CGF); 7623 } 7624 } 7625 7626 void CGOpenMPRuntime::emitTargetDataStandAloneCall( 7627 CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond, 7628 const Expr *Device) { 7629 if (!CGF.HaveInsertPoint()) 7630 return; 7631 7632 assert((isa<OMPTargetEnterDataDirective>(D) || 7633 isa<OMPTargetExitDataDirective>(D) || 7634 isa<OMPTargetUpdateDirective>(D)) && 7635 "Expecting either target enter, exit data, or update directives."); 7636 7637 CodeGenFunction::OMPTargetDataInfo InputInfo; 7638 llvm::Value *MapTypesArray = nullptr; 7639 // Generate the code for the opening of the data environment. 7640 auto &&ThenGen = [this, &D, Device, &InputInfo, 7641 &MapTypesArray](CodeGenFunction &CGF, PrePostActionTy &) { 7642 // Emit device ID if any. 7643 llvm::Value *DeviceID = nullptr; 7644 if (Device) { 7645 DeviceID = CGF.Builder.CreateIntCast(CGF.EmitScalarExpr(Device), 7646 CGF.Int64Ty, /*isSigned=*/true); 7647 } else { 7648 DeviceID = CGF.Builder.getInt64(OMP_DEVICEID_UNDEF); 7649 } 7650 7651 // Emit the number of elements in the offloading arrays. 7652 llvm::Constant *PointerNum = 7653 CGF.Builder.getInt32(InputInfo.NumberOfTargetItems); 7654 7655 llvm::Value *OffloadingArgs[] = {DeviceID, 7656 PointerNum, 7657 InputInfo.BasePointersArray.getPointer(), 7658 InputInfo.PointersArray.getPointer(), 7659 InputInfo.SizesArray.getPointer(), 7660 MapTypesArray}; 7661 7662 // Select the right runtime function call for each expected standalone 7663 // directive. 7664 const bool HasNowait = D.hasClausesOfKind<OMPNowaitClause>(); 7665 OpenMPRTLFunction RTLFn; 7666 switch (D.getDirectiveKind()) { 7667 default: 7668 llvm_unreachable("Unexpected standalone target data directive."); 7669 break; 7670 case OMPD_target_enter_data: 7671 RTLFn = HasNowait ? OMPRTL__tgt_target_data_begin_nowait 7672 : OMPRTL__tgt_target_data_begin; 7673 break; 7674 case OMPD_target_exit_data: 7675 RTLFn = HasNowait ? OMPRTL__tgt_target_data_end_nowait 7676 : OMPRTL__tgt_target_data_end; 7677 break; 7678 case OMPD_target_update: 7679 RTLFn = HasNowait ? OMPRTL__tgt_target_data_update_nowait 7680 : OMPRTL__tgt_target_data_update; 7681 break; 7682 } 7683 CGF.EmitRuntimeCall(createRuntimeFunction(RTLFn), OffloadingArgs); 7684 }; 7685 7686 auto &&TargetThenGen = [this, &ThenGen, &D, &InputInfo, &MapTypesArray]( 7687 CodeGenFunction &CGF, PrePostActionTy &) { 7688 // Fill up the arrays with all the mapped variables. 7689 MappableExprsHandler::MapBaseValuesArrayTy BasePointers; 7690 MappableExprsHandler::MapValuesArrayTy Pointers; 7691 MappableExprsHandler::MapValuesArrayTy Sizes; 7692 MappableExprsHandler::MapFlagsArrayTy MapTypes; 7693 7694 // Get map clause information. 7695 MappableExprsHandler MEHandler(D, CGF); 7696 MEHandler.generateAllInfo(BasePointers, Pointers, Sizes, MapTypes); 7697 7698 TargetDataInfo Info; 7699 // Fill up the arrays and create the arguments. 7700 emitOffloadingArrays(CGF, BasePointers, Pointers, Sizes, MapTypes, Info); 7701 emitOffloadingArraysArgument(CGF, Info.BasePointersArray, 7702 Info.PointersArray, Info.SizesArray, 7703 Info.MapTypesArray, Info); 7704 InputInfo.NumberOfTargetItems = Info.NumberOfPtrs; 7705 InputInfo.BasePointersArray = 7706 Address(Info.BasePointersArray, CGM.getPointerAlign()); 7707 InputInfo.PointersArray = 7708 Address(Info.PointersArray, CGM.getPointerAlign()); 7709 InputInfo.SizesArray = 7710 Address(Info.SizesArray, CGM.getPointerAlign()); 7711 MapTypesArray = Info.MapTypesArray; 7712 if (D.hasClausesOfKind<OMPDependClause>()) 7713 CGF.EmitOMPTargetTaskBasedDirective(D, ThenGen, InputInfo); 7714 else 7715 emitInlinedDirective(CGF, D.getDirectiveKind(), ThenGen); 7716 }; 7717 7718 if (IfCond) 7719 emitOMPIfClause(CGF, IfCond, TargetThenGen, 7720 [](CodeGenFunction &CGF, PrePostActionTy &) {}); 7721 else { 7722 RegionCodeGenTy ThenRCG(TargetThenGen); 7723 ThenRCG(CGF); 7724 } 7725 } 7726 7727 namespace { 7728 /// Kind of parameter in a function with 'declare simd' directive. 7729 enum ParamKindTy { LinearWithVarStride, Linear, Uniform, Vector }; 7730 /// Attribute set of the parameter. 7731 struct ParamAttrTy { 7732 ParamKindTy Kind = Vector; 7733 llvm::APSInt StrideOrArg; 7734 llvm::APSInt Alignment; 7735 }; 7736 } // namespace 7737 7738 static unsigned evaluateCDTSize(const FunctionDecl *FD, 7739 ArrayRef<ParamAttrTy> ParamAttrs) { 7740 // Every vector variant of a SIMD-enabled function has a vector length (VLEN). 7741 // If OpenMP clause "simdlen" is used, the VLEN is the value of the argument 7742 // of that clause. The VLEN value must be power of 2. 7743 // In other case the notion of the function`s "characteristic data type" (CDT) 7744 // is used to compute the vector length. 7745 // CDT is defined in the following order: 7746 // a) For non-void function, the CDT is the return type. 7747 // b) If the function has any non-uniform, non-linear parameters, then the 7748 // CDT is the type of the first such parameter. 7749 // c) If the CDT determined by a) or b) above is struct, union, or class 7750 // type which is pass-by-value (except for the type that maps to the 7751 // built-in complex data type), the characteristic data type is int. 7752 // d) If none of the above three cases is applicable, the CDT is int. 7753 // The VLEN is then determined based on the CDT and the size of vector 7754 // register of that ISA for which current vector version is generated. The 7755 // VLEN is computed using the formula below: 7756 // VLEN = sizeof(vector_register) / sizeof(CDT), 7757 // where vector register size specified in section 3.2.1 Registers and the 7758 // Stack Frame of original AMD64 ABI document. 7759 QualType RetType = FD->getReturnType(); 7760 if (RetType.isNull()) 7761 return 0; 7762 ASTContext &C = FD->getASTContext(); 7763 QualType CDT; 7764 if (!RetType.isNull() && !RetType->isVoidType()) 7765 CDT = RetType; 7766 else { 7767 unsigned Offset = 0; 7768 if (auto *MD = dyn_cast<CXXMethodDecl>(FD)) { 7769 if (ParamAttrs[Offset].Kind == Vector) 7770 CDT = C.getPointerType(C.getRecordType(MD->getParent())); 7771 ++Offset; 7772 } 7773 if (CDT.isNull()) { 7774 for (unsigned I = 0, E = FD->getNumParams(); I < E; ++I) { 7775 if (ParamAttrs[I + Offset].Kind == Vector) { 7776 CDT = FD->getParamDecl(I)->getType(); 7777 break; 7778 } 7779 } 7780 } 7781 } 7782 if (CDT.isNull()) 7783 CDT = C.IntTy; 7784 CDT = CDT->getCanonicalTypeUnqualified(); 7785 if (CDT->isRecordType() || CDT->isUnionType()) 7786 CDT = C.IntTy; 7787 return C.getTypeSize(CDT); 7788 } 7789 7790 static void 7791 emitX86DeclareSimdFunction(const FunctionDecl *FD, llvm::Function *Fn, 7792 const llvm::APSInt &VLENVal, 7793 ArrayRef<ParamAttrTy> ParamAttrs, 7794 OMPDeclareSimdDeclAttr::BranchStateTy State) { 7795 struct ISADataTy { 7796 char ISA; 7797 unsigned VecRegSize; 7798 }; 7799 ISADataTy ISAData[] = { 7800 { 7801 'b', 128 7802 }, // SSE 7803 { 7804 'c', 256 7805 }, // AVX 7806 { 7807 'd', 256 7808 }, // AVX2 7809 { 7810 'e', 512 7811 }, // AVX512 7812 }; 7813 llvm::SmallVector<char, 2> Masked; 7814 switch (State) { 7815 case OMPDeclareSimdDeclAttr::BS_Undefined: 7816 Masked.push_back('N'); 7817 Masked.push_back('M'); 7818 break; 7819 case OMPDeclareSimdDeclAttr::BS_Notinbranch: 7820 Masked.push_back('N'); 7821 break; 7822 case OMPDeclareSimdDeclAttr::BS_Inbranch: 7823 Masked.push_back('M'); 7824 break; 7825 } 7826 for (auto Mask : Masked) { 7827 for (auto &Data : ISAData) { 7828 SmallString<256> Buffer; 7829 llvm::raw_svector_ostream Out(Buffer); 7830 Out << "_ZGV" << Data.ISA << Mask; 7831 if (!VLENVal) { 7832 Out << llvm::APSInt::getUnsigned(Data.VecRegSize / 7833 evaluateCDTSize(FD, ParamAttrs)); 7834 } else 7835 Out << VLENVal; 7836 for (auto &ParamAttr : ParamAttrs) { 7837 switch (ParamAttr.Kind){ 7838 case LinearWithVarStride: 7839 Out << 's' << ParamAttr.StrideOrArg; 7840 break; 7841 case Linear: 7842 Out << 'l'; 7843 if (!!ParamAttr.StrideOrArg) 7844 Out << ParamAttr.StrideOrArg; 7845 break; 7846 case Uniform: 7847 Out << 'u'; 7848 break; 7849 case Vector: 7850 Out << 'v'; 7851 break; 7852 } 7853 if (!!ParamAttr.Alignment) 7854 Out << 'a' << ParamAttr.Alignment; 7855 } 7856 Out << '_' << Fn->getName(); 7857 Fn->addFnAttr(Out.str()); 7858 } 7859 } 7860 } 7861 7862 void CGOpenMPRuntime::emitDeclareSimdFunction(const FunctionDecl *FD, 7863 llvm::Function *Fn) { 7864 ASTContext &C = CGM.getContext(); 7865 FD = FD->getCanonicalDecl(); 7866 // Map params to their positions in function decl. 7867 llvm::DenseMap<const Decl *, unsigned> ParamPositions; 7868 if (isa<CXXMethodDecl>(FD)) 7869 ParamPositions.insert({FD, 0}); 7870 unsigned ParamPos = ParamPositions.size(); 7871 for (auto *P : FD->parameters()) { 7872 ParamPositions.insert({P->getCanonicalDecl(), ParamPos}); 7873 ++ParamPos; 7874 } 7875 for (auto *Attr : FD->specific_attrs<OMPDeclareSimdDeclAttr>()) { 7876 llvm::SmallVector<ParamAttrTy, 8> ParamAttrs(ParamPositions.size()); 7877 // Mark uniform parameters. 7878 for (auto *E : Attr->uniforms()) { 7879 E = E->IgnoreParenImpCasts(); 7880 unsigned Pos; 7881 if (isa<CXXThisExpr>(E)) 7882 Pos = ParamPositions[FD]; 7883 else { 7884 auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl()) 7885 ->getCanonicalDecl(); 7886 Pos = ParamPositions[PVD]; 7887 } 7888 ParamAttrs[Pos].Kind = Uniform; 7889 } 7890 // Get alignment info. 7891 auto NI = Attr->alignments_begin(); 7892 for (auto *E : Attr->aligneds()) { 7893 E = E->IgnoreParenImpCasts(); 7894 unsigned Pos; 7895 QualType ParmTy; 7896 if (isa<CXXThisExpr>(E)) { 7897 Pos = ParamPositions[FD]; 7898 ParmTy = E->getType(); 7899 } else { 7900 auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl()) 7901 ->getCanonicalDecl(); 7902 Pos = ParamPositions[PVD]; 7903 ParmTy = PVD->getType(); 7904 } 7905 ParamAttrs[Pos].Alignment = 7906 (*NI) ? (*NI)->EvaluateKnownConstInt(C) 7907 : llvm::APSInt::getUnsigned( 7908 C.toCharUnitsFromBits(C.getOpenMPDefaultSimdAlign(ParmTy)) 7909 .getQuantity()); 7910 ++NI; 7911 } 7912 // Mark linear parameters. 7913 auto SI = Attr->steps_begin(); 7914 auto MI = Attr->modifiers_begin(); 7915 for (auto *E : Attr->linears()) { 7916 E = E->IgnoreParenImpCasts(); 7917 unsigned Pos; 7918 if (isa<CXXThisExpr>(E)) 7919 Pos = ParamPositions[FD]; 7920 else { 7921 auto *PVD = cast<ParmVarDecl>(cast<DeclRefExpr>(E)->getDecl()) 7922 ->getCanonicalDecl(); 7923 Pos = ParamPositions[PVD]; 7924 } 7925 auto &ParamAttr = ParamAttrs[Pos]; 7926 ParamAttr.Kind = Linear; 7927 if (*SI) { 7928 if (!(*SI)->EvaluateAsInt(ParamAttr.StrideOrArg, C, 7929 Expr::SE_AllowSideEffects)) { 7930 if (auto *DRE = cast<DeclRefExpr>((*SI)->IgnoreParenImpCasts())) { 7931 if (auto *StridePVD = cast<ParmVarDecl>(DRE->getDecl())) { 7932 ParamAttr.Kind = LinearWithVarStride; 7933 ParamAttr.StrideOrArg = llvm::APSInt::getUnsigned( 7934 ParamPositions[StridePVD->getCanonicalDecl()]); 7935 } 7936 } 7937 } 7938 } 7939 ++SI; 7940 ++MI; 7941 } 7942 llvm::APSInt VLENVal; 7943 if (const Expr *VLEN = Attr->getSimdlen()) 7944 VLENVal = VLEN->EvaluateKnownConstInt(C); 7945 OMPDeclareSimdDeclAttr::BranchStateTy State = Attr->getBranchState(); 7946 if (CGM.getTriple().getArch() == llvm::Triple::x86 || 7947 CGM.getTriple().getArch() == llvm::Triple::x86_64) 7948 emitX86DeclareSimdFunction(FD, Fn, VLENVal, ParamAttrs, State); 7949 } 7950 } 7951 7952 namespace { 7953 /// Cleanup action for doacross support. 7954 class DoacrossCleanupTy final : public EHScopeStack::Cleanup { 7955 public: 7956 static const int DoacrossFinArgs = 2; 7957 7958 private: 7959 llvm::Value *RTLFn; 7960 llvm::Value *Args[DoacrossFinArgs]; 7961 7962 public: 7963 DoacrossCleanupTy(llvm::Value *RTLFn, ArrayRef<llvm::Value *> CallArgs) 7964 : RTLFn(RTLFn) { 7965 assert(CallArgs.size() == DoacrossFinArgs); 7966 std::copy(CallArgs.begin(), CallArgs.end(), std::begin(Args)); 7967 } 7968 void Emit(CodeGenFunction &CGF, Flags /*flags*/) override { 7969 if (!CGF.HaveInsertPoint()) 7970 return; 7971 CGF.EmitRuntimeCall(RTLFn, Args); 7972 } 7973 }; 7974 } // namespace 7975 7976 void CGOpenMPRuntime::emitDoacrossInit(CodeGenFunction &CGF, 7977 const OMPLoopDirective &D) { 7978 if (!CGF.HaveInsertPoint()) 7979 return; 7980 7981 ASTContext &C = CGM.getContext(); 7982 QualType Int64Ty = C.getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/true); 7983 RecordDecl *RD; 7984 if (KmpDimTy.isNull()) { 7985 // Build struct kmp_dim { // loop bounds info casted to kmp_int64 7986 // kmp_int64 lo; // lower 7987 // kmp_int64 up; // upper 7988 // kmp_int64 st; // stride 7989 // }; 7990 RD = C.buildImplicitRecord("kmp_dim"); 7991 RD->startDefinition(); 7992 addFieldToRecordDecl(C, RD, Int64Ty); 7993 addFieldToRecordDecl(C, RD, Int64Ty); 7994 addFieldToRecordDecl(C, RD, Int64Ty); 7995 RD->completeDefinition(); 7996 KmpDimTy = C.getRecordType(RD); 7997 } else 7998 RD = cast<RecordDecl>(KmpDimTy->getAsTagDecl()); 7999 8000 Address DimsAddr = CGF.CreateMemTemp(KmpDimTy, "dims"); 8001 CGF.EmitNullInitialization(DimsAddr, KmpDimTy); 8002 enum { LowerFD = 0, UpperFD, StrideFD }; 8003 // Fill dims with data. 8004 LValue DimsLVal = CGF.MakeAddrLValue(DimsAddr, KmpDimTy); 8005 // dims.upper = num_iterations; 8006 LValue UpperLVal = 8007 CGF.EmitLValueForField(DimsLVal, *std::next(RD->field_begin(), UpperFD)); 8008 llvm::Value *NumIterVal = CGF.EmitScalarConversion( 8009 CGF.EmitScalarExpr(D.getNumIterations()), D.getNumIterations()->getType(), 8010 Int64Ty, D.getNumIterations()->getExprLoc()); 8011 CGF.EmitStoreOfScalar(NumIterVal, UpperLVal); 8012 // dims.stride = 1; 8013 LValue StrideLVal = 8014 CGF.EmitLValueForField(DimsLVal, *std::next(RD->field_begin(), StrideFD)); 8015 CGF.EmitStoreOfScalar(llvm::ConstantInt::getSigned(CGM.Int64Ty, /*V=*/1), 8016 StrideLVal); 8017 8018 // Build call void __kmpc_doacross_init(ident_t *loc, kmp_int32 gtid, 8019 // kmp_int32 num_dims, struct kmp_dim * dims); 8020 llvm::Value *Args[] = {emitUpdateLocation(CGF, D.getLocStart()), 8021 getThreadID(CGF, D.getLocStart()), 8022 llvm::ConstantInt::getSigned(CGM.Int32Ty, 1), 8023 CGF.Builder.CreatePointerBitCastOrAddrSpaceCast( 8024 DimsAddr.getPointer(), CGM.VoidPtrTy)}; 8025 8026 llvm::Value *RTLFn = createRuntimeFunction(OMPRTL__kmpc_doacross_init); 8027 CGF.EmitRuntimeCall(RTLFn, Args); 8028 llvm::Value *FiniArgs[DoacrossCleanupTy::DoacrossFinArgs] = { 8029 emitUpdateLocation(CGF, D.getLocEnd()), getThreadID(CGF, D.getLocEnd())}; 8030 llvm::Value *FiniRTLFn = createRuntimeFunction(OMPRTL__kmpc_doacross_fini); 8031 CGF.EHStack.pushCleanup<DoacrossCleanupTy>(NormalAndEHCleanup, FiniRTLFn, 8032 llvm::makeArrayRef(FiniArgs)); 8033 } 8034 8035 void CGOpenMPRuntime::emitDoacrossOrdered(CodeGenFunction &CGF, 8036 const OMPDependClause *C) { 8037 QualType Int64Ty = 8038 CGM.getContext().getIntTypeForBitwidth(/*DestWidth=*/64, /*Signed=*/1); 8039 const Expr *CounterVal = C->getCounterValue(); 8040 assert(CounterVal); 8041 llvm::Value *CntVal = CGF.EmitScalarConversion(CGF.EmitScalarExpr(CounterVal), 8042 CounterVal->getType(), Int64Ty, 8043 CounterVal->getExprLoc()); 8044 Address CntAddr = CGF.CreateMemTemp(Int64Ty, ".cnt.addr"); 8045 CGF.EmitStoreOfScalar(CntVal, CntAddr, /*Volatile=*/false, Int64Ty); 8046 llvm::Value *Args[] = {emitUpdateLocation(CGF, C->getLocStart()), 8047 getThreadID(CGF, C->getLocStart()), 8048 CntAddr.getPointer()}; 8049 llvm::Value *RTLFn; 8050 if (C->getDependencyKind() == OMPC_DEPEND_source) 8051 RTLFn = createRuntimeFunction(OMPRTL__kmpc_doacross_post); 8052 else { 8053 assert(C->getDependencyKind() == OMPC_DEPEND_sink); 8054 RTLFn = createRuntimeFunction(OMPRTL__kmpc_doacross_wait); 8055 } 8056 CGF.EmitRuntimeCall(RTLFn, Args); 8057 } 8058 8059 void CGOpenMPRuntime::emitCall(CodeGenFunction &CGF, llvm::Value *Callee, 8060 ArrayRef<llvm::Value *> Args, 8061 SourceLocation Loc) const { 8062 auto DL = ApplyDebugLocation::CreateDefaultArtificial(CGF, Loc); 8063 8064 if (auto *Fn = dyn_cast<llvm::Function>(Callee)) { 8065 if (Fn->doesNotThrow()) { 8066 CGF.EmitNounwindRuntimeCall(Fn, Args); 8067 return; 8068 } 8069 } 8070 CGF.EmitRuntimeCall(Callee, Args); 8071 } 8072 8073 void CGOpenMPRuntime::emitOutlinedFunctionCall( 8074 CodeGenFunction &CGF, SourceLocation Loc, llvm::Value *OutlinedFn, 8075 ArrayRef<llvm::Value *> Args) const { 8076 assert(Loc.isValid() && "Outlined function call location must be valid."); 8077 emitCall(CGF, OutlinedFn, Args, Loc); 8078 } 8079 8080 Address CGOpenMPRuntime::getParameterAddress(CodeGenFunction &CGF, 8081 const VarDecl *NativeParam, 8082 const VarDecl *TargetParam) const { 8083 return CGF.GetAddrOfLocalVar(NativeParam); 8084 } 8085 8086 llvm::Value *CGOpenMPSIMDRuntime::emitParallelOutlinedFunction( 8087 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 8088 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen) { 8089 llvm_unreachable("Not supported in SIMD-only mode"); 8090 } 8091 8092 llvm::Value *CGOpenMPSIMDRuntime::emitTeamsOutlinedFunction( 8093 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 8094 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen) { 8095 llvm_unreachable("Not supported in SIMD-only mode"); 8096 } 8097 8098 llvm::Value *CGOpenMPSIMDRuntime::emitTaskOutlinedFunction( 8099 const OMPExecutableDirective &D, const VarDecl *ThreadIDVar, 8100 const VarDecl *PartIDVar, const VarDecl *TaskTVar, 8101 OpenMPDirectiveKind InnermostKind, const RegionCodeGenTy &CodeGen, 8102 bool Tied, unsigned &NumberOfParts) { 8103 llvm_unreachable("Not supported in SIMD-only mode"); 8104 } 8105 8106 void CGOpenMPSIMDRuntime::emitParallelCall(CodeGenFunction &CGF, 8107 SourceLocation Loc, 8108 llvm::Value *OutlinedFn, 8109 ArrayRef<llvm::Value *> CapturedVars, 8110 const Expr *IfCond) { 8111 llvm_unreachable("Not supported in SIMD-only mode"); 8112 } 8113 8114 void CGOpenMPSIMDRuntime::emitCriticalRegion( 8115 CodeGenFunction &CGF, StringRef CriticalName, 8116 const RegionCodeGenTy &CriticalOpGen, SourceLocation Loc, 8117 const Expr *Hint) { 8118 llvm_unreachable("Not supported in SIMD-only mode"); 8119 } 8120 8121 void CGOpenMPSIMDRuntime::emitMasterRegion(CodeGenFunction &CGF, 8122 const RegionCodeGenTy &MasterOpGen, 8123 SourceLocation Loc) { 8124 llvm_unreachable("Not supported in SIMD-only mode"); 8125 } 8126 8127 void CGOpenMPSIMDRuntime::emitTaskyieldCall(CodeGenFunction &CGF, 8128 SourceLocation Loc) { 8129 llvm_unreachable("Not supported in SIMD-only mode"); 8130 } 8131 8132 void CGOpenMPSIMDRuntime::emitTaskgroupRegion( 8133 CodeGenFunction &CGF, const RegionCodeGenTy &TaskgroupOpGen, 8134 SourceLocation Loc) { 8135 llvm_unreachable("Not supported in SIMD-only mode"); 8136 } 8137 8138 void CGOpenMPSIMDRuntime::emitSingleRegion( 8139 CodeGenFunction &CGF, const RegionCodeGenTy &SingleOpGen, 8140 SourceLocation Loc, ArrayRef<const Expr *> CopyprivateVars, 8141 ArrayRef<const Expr *> DestExprs, ArrayRef<const Expr *> SrcExprs, 8142 ArrayRef<const Expr *> AssignmentOps) { 8143 llvm_unreachable("Not supported in SIMD-only mode"); 8144 } 8145 8146 void CGOpenMPSIMDRuntime::emitOrderedRegion(CodeGenFunction &CGF, 8147 const RegionCodeGenTy &OrderedOpGen, 8148 SourceLocation Loc, 8149 bool IsThreads) { 8150 llvm_unreachable("Not supported in SIMD-only mode"); 8151 } 8152 8153 void CGOpenMPSIMDRuntime::emitBarrierCall(CodeGenFunction &CGF, 8154 SourceLocation Loc, 8155 OpenMPDirectiveKind Kind, 8156 bool EmitChecks, 8157 bool ForceSimpleCall) { 8158 llvm_unreachable("Not supported in SIMD-only mode"); 8159 } 8160 8161 void CGOpenMPSIMDRuntime::emitForDispatchInit( 8162 CodeGenFunction &CGF, SourceLocation Loc, 8163 const OpenMPScheduleTy &ScheduleKind, unsigned IVSize, bool IVSigned, 8164 bool Ordered, const DispatchRTInput &DispatchValues) { 8165 llvm_unreachable("Not supported in SIMD-only mode"); 8166 } 8167 8168 void CGOpenMPSIMDRuntime::emitForStaticInit( 8169 CodeGenFunction &CGF, SourceLocation Loc, OpenMPDirectiveKind DKind, 8170 const OpenMPScheduleTy &ScheduleKind, const StaticRTInput &Values) { 8171 llvm_unreachable("Not supported in SIMD-only mode"); 8172 } 8173 8174 void CGOpenMPSIMDRuntime::emitDistributeStaticInit( 8175 CodeGenFunction &CGF, SourceLocation Loc, 8176 OpenMPDistScheduleClauseKind SchedKind, const StaticRTInput &Values) { 8177 llvm_unreachable("Not supported in SIMD-only mode"); 8178 } 8179 8180 void CGOpenMPSIMDRuntime::emitForOrderedIterationEnd(CodeGenFunction &CGF, 8181 SourceLocation Loc, 8182 unsigned IVSize, 8183 bool IVSigned) { 8184 llvm_unreachable("Not supported in SIMD-only mode"); 8185 } 8186 8187 void CGOpenMPSIMDRuntime::emitForStaticFinish(CodeGenFunction &CGF, 8188 SourceLocation Loc, 8189 OpenMPDirectiveKind DKind) { 8190 llvm_unreachable("Not supported in SIMD-only mode"); 8191 } 8192 8193 llvm::Value *CGOpenMPSIMDRuntime::emitForNext(CodeGenFunction &CGF, 8194 SourceLocation Loc, 8195 unsigned IVSize, bool IVSigned, 8196 Address IL, Address LB, 8197 Address UB, Address ST) { 8198 llvm_unreachable("Not supported in SIMD-only mode"); 8199 } 8200 8201 void CGOpenMPSIMDRuntime::emitNumThreadsClause(CodeGenFunction &CGF, 8202 llvm::Value *NumThreads, 8203 SourceLocation Loc) { 8204 llvm_unreachable("Not supported in SIMD-only mode"); 8205 } 8206 8207 void CGOpenMPSIMDRuntime::emitProcBindClause(CodeGenFunction &CGF, 8208 OpenMPProcBindClauseKind ProcBind, 8209 SourceLocation Loc) { 8210 llvm_unreachable("Not supported in SIMD-only mode"); 8211 } 8212 8213 Address CGOpenMPSIMDRuntime::getAddrOfThreadPrivate(CodeGenFunction &CGF, 8214 const VarDecl *VD, 8215 Address VDAddr, 8216 SourceLocation Loc) { 8217 llvm_unreachable("Not supported in SIMD-only mode"); 8218 } 8219 8220 llvm::Function *CGOpenMPSIMDRuntime::emitThreadPrivateVarDefinition( 8221 const VarDecl *VD, Address VDAddr, SourceLocation Loc, bool PerformInit, 8222 CodeGenFunction *CGF) { 8223 llvm_unreachable("Not supported in SIMD-only mode"); 8224 } 8225 8226 Address CGOpenMPSIMDRuntime::getAddrOfArtificialThreadPrivate( 8227 CodeGenFunction &CGF, QualType VarType, StringRef Name) { 8228 llvm_unreachable("Not supported in SIMD-only mode"); 8229 } 8230 8231 void CGOpenMPSIMDRuntime::emitFlush(CodeGenFunction &CGF, 8232 ArrayRef<const Expr *> Vars, 8233 SourceLocation Loc) { 8234 llvm_unreachable("Not supported in SIMD-only mode"); 8235 } 8236 8237 void CGOpenMPSIMDRuntime::emitTaskCall(CodeGenFunction &CGF, SourceLocation Loc, 8238 const OMPExecutableDirective &D, 8239 llvm::Value *TaskFunction, 8240 QualType SharedsTy, Address Shareds, 8241 const Expr *IfCond, 8242 const OMPTaskDataTy &Data) { 8243 llvm_unreachable("Not supported in SIMD-only mode"); 8244 } 8245 8246 void CGOpenMPSIMDRuntime::emitTaskLoopCall( 8247 CodeGenFunction &CGF, SourceLocation Loc, const OMPLoopDirective &D, 8248 llvm::Value *TaskFunction, QualType SharedsTy, Address Shareds, 8249 const Expr *IfCond, const OMPTaskDataTy &Data) { 8250 llvm_unreachable("Not supported in SIMD-only mode"); 8251 } 8252 8253 void CGOpenMPSIMDRuntime::emitReduction( 8254 CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> Privates, 8255 ArrayRef<const Expr *> LHSExprs, ArrayRef<const Expr *> RHSExprs, 8256 ArrayRef<const Expr *> ReductionOps, ReductionOptionsTy Options) { 8257 assert(Options.SimpleReduction && "Only simple reduction is expected."); 8258 CGOpenMPRuntime::emitReduction(CGF, Loc, Privates, LHSExprs, RHSExprs, 8259 ReductionOps, Options); 8260 } 8261 8262 llvm::Value *CGOpenMPSIMDRuntime::emitTaskReductionInit( 8263 CodeGenFunction &CGF, SourceLocation Loc, ArrayRef<const Expr *> LHSExprs, 8264 ArrayRef<const Expr *> RHSExprs, const OMPTaskDataTy &Data) { 8265 llvm_unreachable("Not supported in SIMD-only mode"); 8266 } 8267 8268 void CGOpenMPSIMDRuntime::emitTaskReductionFixups(CodeGenFunction &CGF, 8269 SourceLocation Loc, 8270 ReductionCodeGen &RCG, 8271 unsigned N) { 8272 llvm_unreachable("Not supported in SIMD-only mode"); 8273 } 8274 8275 Address CGOpenMPSIMDRuntime::getTaskReductionItem(CodeGenFunction &CGF, 8276 SourceLocation Loc, 8277 llvm::Value *ReductionsPtr, 8278 LValue SharedLVal) { 8279 llvm_unreachable("Not supported in SIMD-only mode"); 8280 } 8281 8282 void CGOpenMPSIMDRuntime::emitTaskwaitCall(CodeGenFunction &CGF, 8283 SourceLocation Loc) { 8284 llvm_unreachable("Not supported in SIMD-only mode"); 8285 } 8286 8287 void CGOpenMPSIMDRuntime::emitCancellationPointCall( 8288 CodeGenFunction &CGF, SourceLocation Loc, 8289 OpenMPDirectiveKind CancelRegion) { 8290 llvm_unreachable("Not supported in SIMD-only mode"); 8291 } 8292 8293 void CGOpenMPSIMDRuntime::emitCancelCall(CodeGenFunction &CGF, 8294 SourceLocation Loc, const Expr *IfCond, 8295 OpenMPDirectiveKind CancelRegion) { 8296 llvm_unreachable("Not supported in SIMD-only mode"); 8297 } 8298 8299 void CGOpenMPSIMDRuntime::emitTargetOutlinedFunction( 8300 const OMPExecutableDirective &D, StringRef ParentName, 8301 llvm::Function *&OutlinedFn, llvm::Constant *&OutlinedFnID, 8302 bool IsOffloadEntry, const RegionCodeGenTy &CodeGen) { 8303 llvm_unreachable("Not supported in SIMD-only mode"); 8304 } 8305 8306 void CGOpenMPSIMDRuntime::emitTargetCall(CodeGenFunction &CGF, 8307 const OMPExecutableDirective &D, 8308 llvm::Value *OutlinedFn, 8309 llvm::Value *OutlinedFnID, 8310 const Expr *IfCond, const Expr *Device) { 8311 llvm_unreachable("Not supported in SIMD-only mode"); 8312 } 8313 8314 bool CGOpenMPSIMDRuntime::emitTargetFunctions(GlobalDecl GD) { 8315 llvm_unreachable("Not supported in SIMD-only mode"); 8316 } 8317 8318 bool CGOpenMPSIMDRuntime::emitTargetGlobalVariable(GlobalDecl GD) { 8319 llvm_unreachable("Not supported in SIMD-only mode"); 8320 } 8321 8322 bool CGOpenMPSIMDRuntime::emitTargetGlobal(GlobalDecl GD) { 8323 return false; 8324 } 8325 8326 llvm::Function *CGOpenMPSIMDRuntime::emitRegistrationFunction() { 8327 return nullptr; 8328 } 8329 8330 void CGOpenMPSIMDRuntime::emitTeamsCall(CodeGenFunction &CGF, 8331 const OMPExecutableDirective &D, 8332 SourceLocation Loc, 8333 llvm::Value *OutlinedFn, 8334 ArrayRef<llvm::Value *> CapturedVars) { 8335 llvm_unreachable("Not supported in SIMD-only mode"); 8336 } 8337 8338 void CGOpenMPSIMDRuntime::emitNumTeamsClause(CodeGenFunction &CGF, 8339 const Expr *NumTeams, 8340 const Expr *ThreadLimit, 8341 SourceLocation Loc) { 8342 llvm_unreachable("Not supported in SIMD-only mode"); 8343 } 8344 8345 void CGOpenMPSIMDRuntime::emitTargetDataCalls( 8346 CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond, 8347 const Expr *Device, const RegionCodeGenTy &CodeGen, TargetDataInfo &Info) { 8348 llvm_unreachable("Not supported in SIMD-only mode"); 8349 } 8350 8351 void CGOpenMPSIMDRuntime::emitTargetDataStandAloneCall( 8352 CodeGenFunction &CGF, const OMPExecutableDirective &D, const Expr *IfCond, 8353 const Expr *Device) { 8354 llvm_unreachable("Not supported in SIMD-only mode"); 8355 } 8356 8357 void CGOpenMPSIMDRuntime::emitDoacrossInit(CodeGenFunction &CGF, 8358 const OMPLoopDirective &D) { 8359 llvm_unreachable("Not supported in SIMD-only mode"); 8360 } 8361 8362 void CGOpenMPSIMDRuntime::emitDoacrossOrdered(CodeGenFunction &CGF, 8363 const OMPDependClause *C) { 8364 llvm_unreachable("Not supported in SIMD-only mode"); 8365 } 8366 8367 const VarDecl * 8368 CGOpenMPSIMDRuntime::translateParameter(const FieldDecl *FD, 8369 const VarDecl *NativeParam) const { 8370 llvm_unreachable("Not supported in SIMD-only mode"); 8371 } 8372 8373 Address 8374 CGOpenMPSIMDRuntime::getParameterAddress(CodeGenFunction &CGF, 8375 const VarDecl *NativeParam, 8376 const VarDecl *TargetParam) const { 8377 llvm_unreachable("Not supported in SIMD-only mode"); 8378 } 8379 8380