1 //===------ CodeGeneration.cpp - Code generate the Scops. -----------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // The CodeGeneration pass takes a Scop created by ScopInfo and translates it 11 // back to LLVM-IR using Cloog. 12 // 13 // The Scop describes the high level memory behaviour of a control flow region. 14 // Transformation passes can update the schedule (execution order) of statements 15 // in the Scop. Cloog is used to generate an abstract syntax tree (clast) that 16 // reflects the updated execution order. This clast is used to create new 17 // LLVM-IR that is computational equivalent to the original control flow region, 18 // but executes its code in the new execution order defined by the changed 19 // scattering. 20 // 21 //===----------------------------------------------------------------------===// 22 23 #include "polly/CodeGen/Cloog.h" 24 #ifdef CLOOG_FOUND 25 26 #include "polly/Dependences.h" 27 #include "polly/LinkAllPasses.h" 28 #include "polly/Options.h" 29 #include "polly/ScopInfo.h" 30 #include "polly/TempScopInfo.h" 31 #include "polly/CodeGen/CodeGeneration.h" 32 #include "polly/CodeGen/BlockGenerators.h" 33 #include "polly/CodeGen/LoopGenerators.h" 34 #include "polly/CodeGen/PTXGenerator.h" 35 #include "polly/CodeGen/Utils.h" 36 #include "polly/Support/GICHelper.h" 37 #include "polly/Support/ScopHelper.h" 38 39 #include "llvm/IR/Module.h" 40 #include "llvm/ADT/SetVector.h" 41 #include "llvm/ADT/PostOrderIterator.h" 42 #include "llvm/Analysis/LoopInfo.h" 43 #include "llvm/Analysis/ScalarEvolutionExpander.h" 44 #include "llvm/Support/Debug.h" 45 #include "llvm/IR/DataLayout.h" 46 #include "llvm/Transforms/Utils/BasicBlockUtils.h" 47 48 #define CLOOG_INT_GMP 1 49 #include "cloog/cloog.h" 50 #include "cloog/isl/cloog.h" 51 52 #include "isl/aff.h" 53 54 #include <vector> 55 #include <utility> 56 57 using namespace polly; 58 using namespace llvm; 59 60 #define DEBUG_TYPE "polly-codegen" 61 62 struct isl_set; 63 64 namespace polly { 65 static cl::opt<bool> 66 OpenMP("enable-polly-openmp", cl::desc("Generate OpenMP parallel code"), 67 cl::value_desc("OpenMP code generation enabled if true"), 68 cl::init(false), cl::ZeroOrMore, cl::cat(PollyCategory)); 69 70 #ifdef GPU_CODEGEN 71 static cl::opt<bool> 72 GPGPU("enable-polly-gpgpu", cl::desc("Generate GPU parallel code"), 73 cl::Hidden, cl::value_desc("GPGPU code generation enabled if true"), 74 cl::init(false), cl::ZeroOrMore, cl::cat(PollyCategory)); 75 76 static cl::opt<std::string> 77 GPUTriple("polly-gpgpu-triple", 78 cl::desc("Target triple for GPU code generation"), cl::Hidden, 79 cl::init(""), cl::cat(PollyCategory)); 80 #endif /* GPU_CODEGEN */ 81 82 typedef DenseMap<const char *, Value *> CharMapT; 83 84 /// Class to generate LLVM-IR that calculates the value of a clast_expr. 85 class ClastExpCodeGen { 86 PollyIRBuilder &Builder; 87 const CharMapT &IVS; 88 89 Value *codegen(const clast_name *e, Type *Ty); 90 Value *codegen(const clast_term *e, Type *Ty); 91 Value *codegen(const clast_binary *e, Type *Ty); 92 Value *codegen(const clast_reduction *r, Type *Ty); 93 94 public: 95 // A generator for clast expressions. 96 // 97 // @param B The IRBuilder that defines where the code to calculate the 98 // clast expressions should be inserted. 99 // @param IVMAP A Map that translates strings describing the induction 100 // variables to the Values* that represent these variables 101 // on the LLVM side. 102 ClastExpCodeGen(PollyIRBuilder &B, CharMapT &IVMap); 103 104 // Generates code to calculate a given clast expression. 105 // 106 // @param e The expression to calculate. 107 // @return The Value that holds the result. 108 Value *codegen(const clast_expr *e, Type *Ty); 109 }; 110 111 Value *ClastExpCodeGen::codegen(const clast_name *e, Type *Ty) { 112 CharMapT::const_iterator I = IVS.find(e->name); 113 114 assert(I != IVS.end() && "Clast name not found"); 115 116 return Builder.CreateSExtOrBitCast(I->second, Ty); 117 } 118 119 static APInt APInt_from_MPZ(const mpz_t mpz) { 120 uint64_t *p = nullptr; 121 size_t sz; 122 123 p = (uint64_t *)mpz_export(p, &sz, -1, sizeof(uint64_t), 0, 0, mpz); 124 125 if (p) { 126 APInt A((unsigned)mpz_sizeinbase(mpz, 2), (unsigned)sz, p); 127 A = A.zext(A.getBitWidth() + 1); 128 free(p); 129 130 if (mpz_sgn(mpz) == -1) 131 return -A; 132 else 133 return A; 134 } else { 135 uint64_t val = 0; 136 return APInt(1, 1, &val); 137 } 138 } 139 140 Value *ClastExpCodeGen::codegen(const clast_term *e, Type *Ty) { 141 APInt a = APInt_from_MPZ(e->val); 142 143 Value *ConstOne = ConstantInt::get(Builder.getContext(), a); 144 ConstOne = Builder.CreateSExtOrBitCast(ConstOne, Ty); 145 146 if (!e->var) 147 return ConstOne; 148 149 Value *var = codegen(e->var, Ty); 150 return Builder.CreateMul(ConstOne, var); 151 } 152 153 Value *ClastExpCodeGen::codegen(const clast_binary *e, Type *Ty) { 154 Value *LHS = codegen(e->LHS, Ty); 155 156 APInt RHS_AP = APInt_from_MPZ(e->RHS); 157 158 Value *RHS = ConstantInt::get(Builder.getContext(), RHS_AP); 159 RHS = Builder.CreateSExtOrBitCast(RHS, Ty); 160 161 switch (e->type) { 162 case clast_bin_mod: 163 return Builder.CreateSRem(LHS, RHS); 164 case clast_bin_fdiv: { 165 // floord(n,d) ((n < 0) ? (n - d + 1) : n) / d 166 Value *One = ConstantInt::get(Ty, 1); 167 Value *Zero = ConstantInt::get(Ty, 0); 168 Value *Sum1 = Builder.CreateSub(LHS, RHS); 169 Value *Sum2 = Builder.CreateAdd(Sum1, One); 170 Value *isNegative = Builder.CreateICmpSLT(LHS, Zero); 171 Value *Dividend = Builder.CreateSelect(isNegative, Sum2, LHS); 172 return Builder.CreateSDiv(Dividend, RHS); 173 } 174 case clast_bin_cdiv: { 175 // ceild(n,d) ((n < 0) ? n : (n + d - 1)) / d 176 Value *One = ConstantInt::get(Ty, 1); 177 Value *Zero = ConstantInt::get(Ty, 0); 178 Value *Sum1 = Builder.CreateAdd(LHS, RHS); 179 Value *Sum2 = Builder.CreateSub(Sum1, One); 180 Value *isNegative = Builder.CreateICmpSLT(LHS, Zero); 181 Value *Dividend = Builder.CreateSelect(isNegative, LHS, Sum2); 182 return Builder.CreateSDiv(Dividend, RHS); 183 } 184 case clast_bin_div: 185 return Builder.CreateSDiv(LHS, RHS); 186 } 187 188 llvm_unreachable("Unknown clast binary expression type"); 189 } 190 191 Value *ClastExpCodeGen::codegen(const clast_reduction *r, Type *Ty) { 192 assert((r->type == clast_red_min || r->type == clast_red_max || 193 r->type == clast_red_sum) && 194 "Clast reduction type not supported"); 195 Value *old = codegen(r->elts[0], Ty); 196 197 for (int i = 1; i < r->n; ++i) { 198 Value *exprValue = codegen(r->elts[i], Ty); 199 200 switch (r->type) { 201 case clast_red_min: { 202 Value *cmp = Builder.CreateICmpSLT(old, exprValue); 203 old = Builder.CreateSelect(cmp, old, exprValue); 204 break; 205 } 206 case clast_red_max: { 207 Value *cmp = Builder.CreateICmpSGT(old, exprValue); 208 old = Builder.CreateSelect(cmp, old, exprValue); 209 break; 210 } 211 case clast_red_sum: 212 old = Builder.CreateAdd(old, exprValue); 213 break; 214 } 215 } 216 217 return old; 218 } 219 220 ClastExpCodeGen::ClastExpCodeGen(PollyIRBuilder &B, CharMapT &IVMap) 221 : Builder(B), IVS(IVMap) {} 222 223 Value *ClastExpCodeGen::codegen(const clast_expr *e, Type *Ty) { 224 switch (e->type) { 225 case clast_expr_name: 226 return codegen((const clast_name *)e, Ty); 227 case clast_expr_term: 228 return codegen((const clast_term *)e, Ty); 229 case clast_expr_bin: 230 return codegen((const clast_binary *)e, Ty); 231 case clast_expr_red: 232 return codegen((const clast_reduction *)e, Ty); 233 } 234 235 llvm_unreachable("Unknown clast expression!"); 236 } 237 238 class ClastStmtCodeGen { 239 public: 240 const std::vector<std::string> &getParallelLoops(); 241 242 private: 243 // The Scop we code generate. 244 Scop *S; 245 Pass *P; 246 LoopInfo &LI; 247 ScalarEvolution &SE; 248 DominatorTree &DT; 249 const DataLayout &DL; 250 251 // The Builder specifies the current location to code generate at. 252 PollyIRBuilder &Builder; 253 254 // Map the Values from the old code to their counterparts in the new code. 255 ValueMapT ValueMap; 256 257 // Map the loops from the old code to expressions function of the induction 258 // variables in the new code. For example, when the code generator produces 259 // this AST: 260 // 261 // for (int c1 = 0; c1 <= 1023; c1 += 1) 262 // for (int c2 = 0; c2 <= 1023; c2 += 1) 263 // Stmt(c2 + 3, c1); 264 // 265 // LoopToScev is a map associating: 266 // "outer loop in the old loop nest" -> SCEV("c2 + 3"), 267 // "inner loop in the old loop nest" -> SCEV("c1"). 268 LoopToScevMapT LoopToScev; 269 270 // clastVars maps from the textual representation of a clast variable to its 271 // current *Value. clast variables are scheduling variables, original 272 // induction variables or parameters. They are used either in loop bounds or 273 // to define the statement instance that is executed. 274 // 275 // for (s = 0; s < n + 3; ++i) 276 // for (t = s; t < m; ++j) 277 // Stmt(i = s + 3 * m, j = t); 278 // 279 // {s,t,i,j,n,m} is the set of clast variables in this clast. 280 CharMapT ClastVars; 281 282 // Codegenerator for clast expressions. 283 ClastExpCodeGen ExpGen; 284 285 // Do we currently generate parallel code? 286 bool parallelCodeGeneration; 287 288 std::vector<std::string> parallelLoops; 289 290 void codegen(const clast_assignment *a); 291 292 void codegen(const clast_assignment *a, ScopStmt *Statement, 293 unsigned Dimension, int vectorDim, 294 std::vector<ValueMapT> *VectorVMap = 0, 295 std::vector<LoopToScevMapT> *VLTS = 0); 296 297 void codegenSubstitutions(const clast_stmt *Assignment, ScopStmt *Statement, 298 int vectorDim = 0, 299 std::vector<ValueMapT> *VectorVMap = 0, 300 std::vector<LoopToScevMapT> *VLTS = 0); 301 302 void codegen(const clast_user_stmt *u, std::vector<Value *> *IVS = nullptr, 303 const char *iterator = nullptr, 304 __isl_take isl_set *scatteringDomain = 0); 305 306 void codegen(const clast_block *b); 307 308 /// @brief Create a classical sequential loop. 309 void codegenForSequential(const clast_for *f); 310 311 /// @brief Create OpenMP structure values. 312 /// 313 /// Create a list of values that has to be stored into the OpenMP subfuncition 314 /// structure. 315 SetVector<Value *> getOMPValues(const clast_stmt *Body); 316 317 /// @brief Update ClastVars and ValueMap according to a value map. 318 /// 319 /// @param VMap A map from old to new values. 320 void updateWithValueMap(ParallelLoopGenerator::ValueToValueMapTy &VMap); 321 322 /// @brief Create an OpenMP parallel for loop. 323 /// 324 /// This loop reflects a loop as if it would have been created by an OpenMP 325 /// statement. 326 void codegenForOpenMP(const clast_for *f); 327 328 #ifdef GPU_CODEGEN 329 /// @brief Create GPGPU device memory access values. 330 /// 331 /// Create a list of values that will be set to be parameters of the GPGPU 332 /// subfunction. These parameters represent device memory base addresses 333 /// and the size in bytes. 334 SetVector<Value *> getGPUValues(unsigned &OutputBytes); 335 336 /// @brief Create a GPU parallel for loop. 337 /// 338 /// This loop reflects a loop as if it would have been created by a GPU 339 /// statement. 340 void codegenForGPGPU(const clast_for *F); 341 342 /// @brief Get innermost for loop. 343 const clast_stmt *getScheduleInfo(const clast_for *F, 344 std::vector<int> &NumIters, 345 unsigned &LoopDepth, 346 unsigned &NonPLoopDepth); 347 #endif /* GPU_CODEGEN */ 348 349 /// @brief Check if a loop is parallel 350 /// 351 /// Detect if a clast_for loop can be executed in parallel. 352 /// 353 /// @param For The clast for loop to check. 354 /// 355 /// @return bool Returns true if the incoming clast_for statement can 356 /// execute in parallel. 357 bool isParallelFor(const clast_for *For); 358 359 bool isInnermostLoop(const clast_for *f); 360 361 /// @brief Get the number of loop iterations for this loop. 362 /// @param f The clast for loop to check. 363 int getNumberOfIterations(const clast_for *f); 364 365 /// @brief Create vector instructions for this loop. 366 void codegenForVector(const clast_for *f); 367 368 void codegen(const clast_for *f); 369 370 Value *codegen(const clast_equation *eq); 371 372 void codegen(const clast_guard *g); 373 374 void codegen(const clast_stmt *stmt); 375 376 void addParameters(const CloogNames *names); 377 378 IntegerType *getIntPtrTy(); 379 380 public: 381 void codegen(const clast_root *r); 382 383 ClastStmtCodeGen(Scop *scop, PollyIRBuilder &B, Pass *P); 384 }; 385 } 386 387 IntegerType *ClastStmtCodeGen::getIntPtrTy() { 388 return P->getAnalysis<DataLayoutPass>().getDataLayout().getIntPtrType( 389 Builder.getContext()); 390 } 391 392 const std::vector<std::string> &ClastStmtCodeGen::getParallelLoops() { 393 return parallelLoops; 394 } 395 396 void ClastStmtCodeGen::codegen(const clast_assignment *a) { 397 Value *V = ExpGen.codegen(a->RHS, getIntPtrTy()); 398 ClastVars[a->LHS] = V; 399 } 400 401 void ClastStmtCodeGen::codegen(const clast_assignment *A, ScopStmt *Stmt, 402 unsigned Dim, int VectorDim, 403 std::vector<ValueMapT> *VectorVMap, 404 std::vector<LoopToScevMapT> *VLTS) { 405 Value *RHS; 406 407 assert(!A->LHS && "Statement assignments do not have left hand side"); 408 409 RHS = ExpGen.codegen(A->RHS, Builder.getInt64Ty()); 410 411 const llvm::SCEV *URHS = S->getSE()->getUnknown(RHS); 412 if (VLTS) 413 (*VLTS)[VectorDim][Stmt->getLoopForDimension(Dim)] = URHS; 414 LoopToScev[Stmt->getLoopForDimension(Dim)] = URHS; 415 } 416 417 void ClastStmtCodeGen::codegenSubstitutions(const clast_stmt *Assignment, 418 ScopStmt *Statement, int vectorDim, 419 std::vector<ValueMapT> *VectorVMap, 420 std::vector<LoopToScevMapT> *VLTS) { 421 int Dimension = 0; 422 423 while (Assignment) { 424 assert(CLAST_STMT_IS_A(Assignment, stmt_ass) && 425 "Substitions are expected to be assignments"); 426 codegen((const clast_assignment *)Assignment, Statement, Dimension, 427 vectorDim, VectorVMap, VLTS); 428 Assignment = Assignment->next; 429 Dimension++; 430 } 431 } 432 433 // Takes the cloog specific domain and translates it into a map Statement -> 434 // PartialSchedule, where the PartialSchedule contains all the dimensions that 435 // have been code generated up to this point. 436 static __isl_give isl_map *extractPartialSchedule(ScopStmt *Statement, 437 __isl_take isl_set *Domain) { 438 isl_map *Schedule = Statement->getScattering(); 439 int ScheduledDimensions = isl_set_dim(Domain, isl_dim_set); 440 int UnscheduledDimensions = 441 isl_map_dim(Schedule, isl_dim_out) - ScheduledDimensions; 442 443 isl_set_free(Domain); 444 445 return isl_map_project_out(Schedule, isl_dim_out, ScheduledDimensions, 446 UnscheduledDimensions); 447 } 448 449 void ClastStmtCodeGen::codegen(const clast_user_stmt *u, 450 std::vector<Value *> *IVS, const char *iterator, 451 __isl_take isl_set *Domain) { 452 ScopStmt *Statement = (ScopStmt *)u->statement->usr; 453 454 if (u->substitutions) 455 codegenSubstitutions(u->substitutions, Statement); 456 457 int VectorDimensions = IVS ? IVS->size() : 1; 458 459 if (VectorDimensions == 1) { 460 BlockGenerator::generate(Builder, *Statement, ValueMap, LoopToScev, P, LI, 461 SE); 462 isl_set_free(Domain); 463 return; 464 } 465 466 VectorValueMapT VectorMap(VectorDimensions); 467 std::vector<LoopToScevMapT> VLTS(VectorDimensions); 468 469 if (IVS) { 470 assert(u->substitutions && "Substitutions expected!"); 471 int i = 0; 472 for (Value *IV : *IVS) { 473 ClastVars[iterator] = IV; 474 codegenSubstitutions(u->substitutions, Statement, i, &VectorMap, &VLTS); 475 i++; 476 } 477 } 478 479 // Copy the current value map into all vector maps if the key wasn't 480 // available yet. This is needed in case vector codegen is performed in 481 // OpenMP subfunctions. 482 for (const auto &KV : ValueMap) 483 for (int i = 0; i < VectorDimensions; ++i) 484 VectorMap[i].insert(KV); 485 486 isl_map *Schedule = extractPartialSchedule(Statement, Domain); 487 VectorBlockGenerator::generate(Builder, *Statement, VectorMap, VLTS, Schedule, 488 P, LI, SE); 489 isl_map_free(Schedule); 490 } 491 492 void ClastStmtCodeGen::codegen(const clast_block *b) { 493 if (b->body) 494 codegen(b->body); 495 } 496 497 void ClastStmtCodeGen::codegenForSequential(const clast_for *f) { 498 Value *LowerBound, *UpperBound, *IV, *Stride; 499 BasicBlock *ExitBlock; 500 Type *IntPtrTy = getIntPtrTy(); 501 502 LowerBound = ExpGen.codegen(f->LB, IntPtrTy); 503 UpperBound = ExpGen.codegen(f->UB, IntPtrTy); 504 Stride = Builder.getInt(APInt_from_MPZ(f->stride)); 505 506 IV = createLoop(LowerBound, UpperBound, Stride, Builder, P, LI, DT, ExitBlock, 507 CmpInst::ICMP_SLE); 508 509 // Add loop iv to symbols. 510 ClastVars[f->iterator] = IV; 511 512 if (f->body) 513 codegen(f->body); 514 515 // Loop is finished, so remove its iv from the live symbols. 516 ClastVars.erase(f->iterator); 517 Builder.SetInsertPoint(ExitBlock->begin()); 518 } 519 520 // Helper class to determine all scalar parameters used in the basic blocks of a 521 // clast. Scalar parameters are scalar variables defined outside of the SCoP. 522 class ParameterVisitor : public ClastVisitor { 523 std::set<Value *> Values; 524 525 public: 526 ParameterVisitor() : ClastVisitor(), Values() {} 527 528 void visitUser(const clast_user_stmt *Stmt) { 529 const ScopStmt *S = static_cast<const ScopStmt *>(Stmt->statement->usr); 530 const BasicBlock *BB = S->getBasicBlock(); 531 532 // Check all the operands of instructions in the basic block. 533 for (const Instruction &Inst : *BB) { 534 for (Value *SrcVal : Inst.operands()) { 535 if (Instruction *OpInst = dyn_cast<Instruction>(SrcVal)) 536 if (S->getParent()->getRegion().contains(OpInst)) 537 continue; 538 539 if (isa<Instruction>(SrcVal) || isa<Argument>(SrcVal)) 540 Values.insert(SrcVal); 541 } 542 } 543 } 544 545 // Iterator to iterate over the values found. 546 typedef std::set<Value *>::const_iterator const_iterator; 547 inline const_iterator begin() const { return Values.begin(); } 548 inline const_iterator end() const { return Values.end(); } 549 }; 550 551 SetVector<Value *> ClastStmtCodeGen::getOMPValues(const clast_stmt *Body) { 552 SetVector<Value *> Values; 553 554 // The clast variables 555 for (const auto &I : ClastVars) 556 Values.insert(I.second); 557 558 // Find the temporaries that are referenced in the clast statements' 559 // basic blocks but are not defined by these blocks (e.g., references 560 // to function arguments or temporaries defined before the start of 561 // the SCoP). 562 ParameterVisitor Params; 563 Params.visit(Body); 564 565 for (Value *V : Params) { 566 Values.insert(V); 567 DEBUG(dbgs() << "Adding temporary for OMP copy-in: " << *V << "\n"); 568 } 569 570 return Values; 571 } 572 573 void ClastStmtCodeGen::updateWithValueMap( 574 ParallelLoopGenerator::ValueToValueMapTy &VMap) { 575 std::set<Value *> Inserted; 576 577 for (const auto &I : ClastVars) { 578 ClastVars[I.first] = VMap[I.second]; 579 Inserted.insert(I.second); 580 } 581 582 for (const auto &I : VMap) { 583 if (Inserted.count(I.first)) 584 continue; 585 586 ValueMap[I.first] = I.second; 587 } 588 } 589 590 static void clearDomtree(Function *F, DominatorTree &DT) { 591 DomTreeNode *N = DT.getNode(&F->getEntryBlock()); 592 std::vector<BasicBlock *> Nodes; 593 for (po_iterator<DomTreeNode *> I = po_begin(N), E = po_end(N); I != E; ++I) 594 Nodes.push_back(I->getBlock()); 595 596 for (BasicBlock *BB : Nodes) 597 DT.eraseNode(BB); 598 } 599 600 void ClastStmtCodeGen::codegenForOpenMP(const clast_for *For) { 601 Value *Stride, *LB, *UB, *IV; 602 BasicBlock::iterator LoopBody; 603 IntegerType *IntPtrTy = getIntPtrTy(); 604 SetVector<Value *> Values; 605 ParallelLoopGenerator::ValueToValueMapTy VMap; 606 ParallelLoopGenerator OMPGen(Builder, P, LI, DT, DL); 607 608 Stride = Builder.getInt(APInt_from_MPZ(For->stride)); 609 Stride = Builder.CreateSExtOrBitCast(Stride, IntPtrTy); 610 LB = ExpGen.codegen(For->LB, IntPtrTy); 611 UB = ExpGen.codegen(For->UB, IntPtrTy); 612 613 Values = getOMPValues(For->body); 614 615 IV = OMPGen.createParallelLoop(LB, UB, Stride, Values, VMap, &LoopBody); 616 BasicBlock::iterator AfterLoop = Builder.GetInsertPoint(); 617 Builder.SetInsertPoint(LoopBody); 618 619 // Save the current values. 620 const ValueMapT ValueMapCopy = ValueMap; 621 const CharMapT ClastVarsCopy = ClastVars; 622 623 updateWithValueMap(VMap); 624 ClastVars[For->iterator] = IV; 625 626 if (For->body) 627 codegen(For->body); 628 629 // Restore the original values. 630 ValueMap = ValueMapCopy; 631 ClastVars = ClastVarsCopy; 632 633 clearDomtree((*LoopBody).getParent()->getParent(), 634 P->getAnalysis<DominatorTreeWrapperPass>().getDomTree()); 635 636 Builder.SetInsertPoint(AfterLoop); 637 } 638 639 #ifdef GPU_CODEGEN 640 static unsigned getArraySizeInBytes(const ArrayType *AT) { 641 unsigned Bytes = AT->getNumElements(); 642 if (const ArrayType *T = dyn_cast<ArrayType>(AT->getElementType())) 643 Bytes *= getArraySizeInBytes(T); 644 else 645 Bytes *= AT->getElementType()->getPrimitiveSizeInBits() / 8; 646 647 return Bytes; 648 } 649 650 SetVector<Value *> ClastStmtCodeGen::getGPUValues(unsigned &OutputBytes) { 651 SetVector<Value *> Values; 652 OutputBytes = 0; 653 654 // Record the memory reference base addresses. 655 for (ScopStmt *Stmt : *S) { 656 for (MemoryAccess *MA : *Stmt) { 657 Value *BaseAddr = MA->getBaseAddr(); 658 Values.insert((BaseAddr)); 659 660 // FIXME: we assume that there is one and only one array to be written 661 // in a SCoP. 662 int NumWrites = 0; 663 if (MA->isWrite()) { 664 ++NumWrites; 665 assert(NumWrites <= 1 && 666 "We support at most one array to be written in a SCoP."); 667 if (const PointerType *PT = 668 dyn_cast<PointerType>(BaseAddr->getType())) { 669 Type *T = PT->getArrayElementType(); 670 const ArrayType *ATy = dyn_cast<ArrayType>(T); 671 OutputBytes = getArraySizeInBytes(ATy); 672 } 673 } 674 } 675 } 676 677 return Values; 678 } 679 680 const clast_stmt *ClastStmtCodeGen::getScheduleInfo(const clast_for *F, 681 std::vector<int> &NumIters, 682 unsigned &LoopDepth, 683 unsigned &NonPLoopDepth) { 684 clast_stmt *Stmt = (clast_stmt *)F; 685 const clast_for *Result; 686 bool NonParaFlag = false; 687 LoopDepth = 0; 688 NonPLoopDepth = 0; 689 690 while (Stmt) { 691 if (CLAST_STMT_IS_A(Stmt, stmt_for)) { 692 const clast_for *T = (clast_for *)Stmt; 693 if (isParallelFor(T)) { 694 if (!NonParaFlag) { 695 NumIters.push_back(getNumberOfIterations(T)); 696 Result = T; 697 } 698 } else 699 NonParaFlag = true; 700 701 Stmt = T->body; 702 LoopDepth++; 703 continue; 704 } 705 Stmt = Stmt->next; 706 } 707 708 assert(NumIters.size() == 4 && 709 "The loops should be tiled into 4-depth parallel loops and an " 710 "innermost non-parallel one (if exist)."); 711 NonPLoopDepth = LoopDepth - NumIters.size(); 712 assert(NonPLoopDepth <= 1 && 713 "We support only one innermost non-parallel loop currently."); 714 return (const clast_stmt *)Result->body; 715 } 716 717 void ClastStmtCodeGen::codegenForGPGPU(const clast_for *F) { 718 BasicBlock::iterator LoopBody; 719 SetVector<Value *> Values; 720 SetVector<Value *> IVS; 721 std::vector<int> NumIterations; 722 PTXGenerator::ValueToValueMapTy VMap; 723 724 assert(!GPUTriple.empty() && 725 "Target triple should be set properly for GPGPU code generation."); 726 PTXGenerator PTXGen(Builder, P, GPUTriple); 727 728 // Get original IVS and ScopStmt 729 unsigned TiledLoopDepth, NonPLoopDepth; 730 const clast_stmt *InnerStmt = 731 getScheduleInfo(F, NumIterations, TiledLoopDepth, NonPLoopDepth); 732 const clast_stmt *TmpStmt; 733 const clast_user_stmt *U; 734 const clast_for *InnerFor; 735 if (CLAST_STMT_IS_A(InnerStmt, stmt_for)) { 736 InnerFor = (const clast_for *)InnerStmt; 737 TmpStmt = InnerFor->body; 738 } else 739 TmpStmt = InnerStmt; 740 U = (const clast_user_stmt *)TmpStmt; 741 ScopStmt *Statement = (ScopStmt *)U->statement->usr; 742 for (unsigned i = 0; i < Statement->getNumIterators() - NonPLoopDepth; i++) { 743 const Value *IV = Statement->getInductionVariableForDimension(i); 744 IVS.insert(const_cast<Value *>(IV)); 745 } 746 747 unsigned OutBytes; 748 Values = getGPUValues(OutBytes); 749 PTXGen.setOutputBytes(OutBytes); 750 PTXGen.startGeneration(Values, IVS, VMap, &LoopBody); 751 752 BasicBlock::iterator AfterLoop = Builder.GetInsertPoint(); 753 Builder.SetInsertPoint(LoopBody); 754 755 BasicBlock *AfterBB = nullptr; 756 if (NonPLoopDepth) { 757 Value *LowerBound, *UpperBound, *IV, *Stride; 758 Type *IntPtrTy = getIntPtrTy(); 759 LowerBound = ExpGen.codegen(InnerFor->LB, IntPtrTy); 760 UpperBound = ExpGen.codegen(InnerFor->UB, IntPtrTy); 761 Stride = Builder.getInt(APInt_from_MPZ(InnerFor->stride)); 762 IV = createLoop(LowerBound, UpperBound, Stride, Builder, P, LI, DT, AfterBB, 763 CmpInst::ICMP_SLE); 764 const Value *OldIV_ = Statement->getInductionVariableForDimension(2); 765 Value *OldIV = const_cast<Value *>(OldIV_); 766 VMap.insert(std::make_pair(OldIV, IV)); 767 } 768 769 updateWithValueMap(VMap); 770 771 BlockGenerator::generate(Builder, *Statement, ValueMap, LoopToScev, P, LI, 772 SE); 773 774 if (AfterBB) 775 Builder.SetInsertPoint(AfterBB->begin()); 776 777 // FIXME: The replacement of the host base address with the parameter of ptx 778 // subfunction should have been done by updateWithValueMap. We use the 779 // following codes to avoid affecting other parts of Polly. This should be 780 // fixed later. 781 Function *FN = Builder.GetInsertBlock()->getParent(); 782 for (Value *BaseAddr : Values) 783 for (BasicBlock *BB : *FN) 784 for (Instruction *Inst : *BB) 785 Inst->replaceUsesOfWith(BaseAddr, ValueMap[BaseAddr]); 786 Builder.SetInsertPoint(AfterLoop); 787 PTXGen.setLaunchingParameters(NumIterations[0], NumIterations[1], 788 NumIterations[2], NumIterations[3]); 789 PTXGen.finishGeneration(FN); 790 } 791 #endif 792 793 bool ClastStmtCodeGen::isInnermostLoop(const clast_for *f) { 794 const clast_stmt *stmt = f->body; 795 796 while (stmt) { 797 if (!CLAST_STMT_IS_A(stmt, stmt_user)) 798 return false; 799 800 stmt = stmt->next; 801 } 802 803 return true; 804 } 805 806 int ClastStmtCodeGen::getNumberOfIterations(const clast_for *For) { 807 isl_set *LoopDomain = isl_set_copy(isl_set_from_cloog_domain(For->domain)); 808 int NumberOfIterations = polly::getNumberOfIterations(LoopDomain); 809 if (NumberOfIterations == -1) 810 return -1; 811 return NumberOfIterations / mpz_get_si(For->stride) + 1; 812 } 813 814 void ClastStmtCodeGen::codegenForVector(const clast_for *F) { 815 DEBUG(dbgs() << "Vectorizing loop '" << F->iterator << "'\n"); 816 int VectorWidth = getNumberOfIterations(F); 817 818 Value *LB = ExpGen.codegen(F->LB, getIntPtrTy()); 819 820 APInt Stride = APInt_from_MPZ(F->stride); 821 IntegerType *LoopIVType = dyn_cast<IntegerType>(LB->getType()); 822 Stride = Stride.zext(LoopIVType->getBitWidth()); 823 Value *StrideValue = ConstantInt::get(LoopIVType, Stride); 824 825 std::vector<Value *> IVS(VectorWidth); 826 IVS[0] = LB; 827 828 for (int i = 1; i < VectorWidth; i++) 829 IVS[i] = Builder.CreateAdd(IVS[i - 1], StrideValue, "p_vector_iv"); 830 831 isl_set *Domain = isl_set_copy(isl_set_from_cloog_domain(F->domain)); 832 833 // Add loop iv to symbols. 834 ClastVars[F->iterator] = LB; 835 836 const clast_stmt *Stmt = F->body; 837 838 while (Stmt) { 839 codegen((const clast_user_stmt *)Stmt, &IVS, F->iterator, 840 isl_set_copy(Domain)); 841 Stmt = Stmt->next; 842 } 843 844 // Loop is finished, so remove its iv from the live symbols. 845 isl_set_free(Domain); 846 ClastVars.erase(F->iterator); 847 } 848 849 static isl_union_map *getCombinedScheduleForSpace(Scop *S, unsigned dimLevel) { 850 isl_space *Space = S->getParamSpace(); 851 isl_union_map *schedule = isl_union_map_empty(Space); 852 853 for (ScopStmt *Stmt : *S) { 854 unsigned remainingDimensions = Stmt->getNumScattering() - dimLevel; 855 isl_map *Scattering = isl_map_project_out( 856 Stmt->getScattering(), isl_dim_out, dimLevel, remainingDimensions); 857 schedule = isl_union_map_add_map(schedule, Scattering); 858 } 859 860 return schedule; 861 } 862 863 bool ClastStmtCodeGen::isParallelFor(const clast_for *f) { 864 isl_set *Domain = isl_set_copy(isl_set_from_cloog_domain(f->domain)); 865 assert(Domain && "Cannot access domain of loop"); 866 867 Dependences &D = P->getAnalysis<Dependences>(); 868 isl_union_map *Deps = 869 D.getDependences(Dependences::TYPE_RAW | Dependences::TYPE_WAW | 870 Dependences::TYPE_WAR | Dependences::TYPE_RAW); 871 isl_union_map *Schedule = 872 getCombinedScheduleForSpace(S, isl_set_n_dim(Domain)); 873 Schedule = 874 isl_union_map_intersect_range(Schedule, isl_union_set_from_set(Domain)); 875 bool IsParallel = D.isParallel(Schedule, Deps); 876 isl_union_map_free(Schedule); 877 return IsParallel; 878 } 879 880 void ClastStmtCodeGen::codegen(const clast_for *f) { 881 bool Vector = PollyVectorizerChoice != VECTORIZER_NONE; 882 if ((Vector || OpenMP) && isParallelFor(f)) { 883 if (Vector && isInnermostLoop(f) && (-1 != getNumberOfIterations(f)) && 884 (getNumberOfIterations(f) <= 16)) { 885 codegenForVector(f); 886 return; 887 } 888 889 if (OpenMP && !parallelCodeGeneration) { 890 parallelCodeGeneration = true; 891 parallelLoops.push_back(f->iterator); 892 codegenForOpenMP(f); 893 parallelCodeGeneration = false; 894 return; 895 } 896 } 897 898 #ifdef GPU_CODEGEN 899 if (GPGPU && isParallelFor(f)) { 900 if (!parallelCodeGeneration) { 901 parallelCodeGeneration = true; 902 parallelLoops.push_back(f->iterator); 903 codegenForGPGPU(f); 904 parallelCodeGeneration = false; 905 return; 906 } 907 } 908 #endif 909 910 codegenForSequential(f); 911 } 912 913 Value *ClastStmtCodeGen::codegen(const clast_equation *eq) { 914 Value *LHS = ExpGen.codegen(eq->LHS, getIntPtrTy()); 915 Value *RHS = ExpGen.codegen(eq->RHS, getIntPtrTy()); 916 CmpInst::Predicate P; 917 918 if (eq->sign == 0) 919 P = ICmpInst::ICMP_EQ; 920 else if (eq->sign > 0) 921 P = ICmpInst::ICMP_SGE; 922 else 923 P = ICmpInst::ICMP_SLE; 924 925 return Builder.CreateICmp(P, LHS, RHS); 926 } 927 928 void ClastStmtCodeGen::codegen(const clast_guard *g) { 929 Function *F = Builder.GetInsertBlock()->getParent(); 930 LLVMContext &Context = F->getContext(); 931 932 BasicBlock *CondBB = 933 SplitBlock(Builder.GetInsertBlock(), Builder.GetInsertPoint(), P); 934 CondBB->setName("polly.cond"); 935 BasicBlock *MergeBB = SplitBlock(CondBB, CondBB->begin(), P); 936 MergeBB->setName("polly.merge"); 937 BasicBlock *ThenBB = BasicBlock::Create(Context, "polly.then", F); 938 939 DominatorTree &DT = P->getAnalysis<DominatorTreeWrapperPass>().getDomTree(); 940 DT.addNewBlock(ThenBB, CondBB); 941 DT.changeImmediateDominator(MergeBB, CondBB); 942 943 CondBB->getTerminator()->eraseFromParent(); 944 945 Builder.SetInsertPoint(CondBB); 946 947 Value *Predicate = codegen(&(g->eq[0])); 948 949 for (int i = 1; i < g->n; ++i) { 950 Value *TmpPredicate = codegen(&(g->eq[i])); 951 Predicate = Builder.CreateAnd(Predicate, TmpPredicate); 952 } 953 954 Builder.CreateCondBr(Predicate, ThenBB, MergeBB); 955 Builder.SetInsertPoint(ThenBB); 956 Builder.CreateBr(MergeBB); 957 Builder.SetInsertPoint(ThenBB->begin()); 958 959 LoopInfo &LI = P->getAnalysis<LoopInfo>(); 960 Loop *L = LI.getLoopFor(CondBB); 961 if (L) 962 L->addBasicBlockToLoop(ThenBB, LI.getBase()); 963 964 codegen(g->then); 965 966 Builder.SetInsertPoint(MergeBB->begin()); 967 } 968 969 void ClastStmtCodeGen::codegen(const clast_stmt *stmt) { 970 if (CLAST_STMT_IS_A(stmt, stmt_root)) 971 assert(false && "No second root statement expected"); 972 else if (CLAST_STMT_IS_A(stmt, stmt_ass)) 973 codegen((const clast_assignment *)stmt); 974 else if (CLAST_STMT_IS_A(stmt, stmt_user)) 975 codegen((const clast_user_stmt *)stmt); 976 else if (CLAST_STMT_IS_A(stmt, stmt_block)) 977 codegen((const clast_block *)stmt); 978 else if (CLAST_STMT_IS_A(stmt, stmt_for)) 979 codegen((const clast_for *)stmt); 980 else if (CLAST_STMT_IS_A(stmt, stmt_guard)) 981 codegen((const clast_guard *)stmt); 982 983 if (stmt->next) 984 codegen(stmt->next); 985 } 986 987 void ClastStmtCodeGen::addParameters(const CloogNames *names) { 988 SCEVExpander Rewriter(P->getAnalysis<ScalarEvolution>(), "polly"); 989 990 int i = 0; 991 for (Scop::param_iterator PI = S->param_begin(), PE = S->param_end(); 992 PI != PE; ++PI) { 993 assert(i < names->nb_parameters && "Not enough parameter names"); 994 995 const SCEV *Param = *PI; 996 Type *Ty = Param->getType(); 997 998 Instruction *insertLocation = --(Builder.GetInsertBlock()->end()); 999 Value *V = Rewriter.expandCodeFor(Param, Ty, insertLocation); 1000 ClastVars[names->parameters[i]] = V; 1001 1002 ++i; 1003 } 1004 } 1005 1006 void ClastStmtCodeGen::codegen(const clast_root *r) { 1007 addParameters(r->names); 1008 1009 parallelCodeGeneration = false; 1010 1011 const clast_stmt *stmt = (const clast_stmt *)r; 1012 if (stmt->next) 1013 codegen(stmt->next); 1014 } 1015 1016 ClastStmtCodeGen::ClastStmtCodeGen(Scop *scop, PollyIRBuilder &B, Pass *P) 1017 : S(scop), P(P), LI(P->getAnalysis<LoopInfo>()), 1018 SE(P->getAnalysis<ScalarEvolution>()), 1019 DT(P->getAnalysis<DominatorTreeWrapperPass>().getDomTree()), 1020 DL(P->getAnalysis<DataLayoutPass>().getDataLayout()), Builder(B), 1021 ExpGen(Builder, ClastVars) {} 1022 1023 namespace { 1024 class CodeGeneration : public ScopPass { 1025 std::vector<std::string> ParallelLoops; 1026 1027 public: 1028 static char ID; 1029 1030 CodeGeneration() : ScopPass(ID) {} 1031 1032 bool runOnScop(Scop &S) { 1033 ParallelLoops.clear(); 1034 1035 assert(!S.getRegion().isTopLevelRegion() && 1036 "Top level regions are not supported"); 1037 1038 simplifyRegion(&S, this); 1039 1040 Value *RTC = ConstantInt::getTrue(S.getSE()->getContext()); 1041 BasicBlock *StartBlock = executeScopConditionally(S, this, RTC); 1042 1043 PollyIRBuilder Builder(StartBlock->begin()); 1044 1045 ClastStmtCodeGen CodeGen(&S, Builder, this); 1046 CloogInfo &C = getAnalysis<CloogInfo>(); 1047 CodeGen.codegen(C.getClast()); 1048 1049 ParallelLoops.insert(ParallelLoops.begin(), 1050 CodeGen.getParallelLoops().begin(), 1051 CodeGen.getParallelLoops().end()); 1052 return true; 1053 } 1054 1055 virtual void printScop(raw_ostream &OS) const { 1056 for (const auto &PI : ParallelLoops) 1057 OS << "Parallel loop with iterator '" << PI << "' generated\n"; 1058 } 1059 1060 virtual void getAnalysisUsage(AnalysisUsage &AU) const { 1061 AU.addRequired<CloogInfo>(); 1062 AU.addRequired<Dependences>(); 1063 AU.addRequired<DominatorTreeWrapperPass>(); 1064 AU.addRequired<RegionInfoPass>(); 1065 AU.addRequired<ScalarEvolution>(); 1066 AU.addRequired<ScopDetection>(); 1067 AU.addRequired<ScopInfo>(); 1068 AU.addRequired<DataLayoutPass>(); 1069 AU.addRequired<DataLayoutPass>(); 1070 AU.addRequired<LoopInfo>(); 1071 1072 AU.addPreserved<CloogInfo>(); 1073 AU.addPreserved<DataLayoutPass>(); 1074 AU.addPreserved<Dependences>(); 1075 AU.addPreserved<LoopInfo>(); 1076 AU.addPreserved<DominatorTreeWrapperPass>(); 1077 AU.addPreserved<ScopDetection>(); 1078 AU.addPreserved<ScalarEvolution>(); 1079 1080 // FIXME: We do not yet add regions for the newly generated code to the 1081 // region tree. 1082 AU.addPreserved<RegionInfoPass>(); 1083 AU.addPreserved<TempScopInfo>(); 1084 AU.addPreserved<ScopInfo>(); 1085 AU.addPreservedID(IndependentBlocksID); 1086 } 1087 }; 1088 } 1089 1090 char CodeGeneration::ID = 1; 1091 1092 Pass *polly::createCodeGenerationPass() { return new CodeGeneration(); } 1093 1094 INITIALIZE_PASS_BEGIN(CodeGeneration, "polly-codegen", 1095 "Polly - Create LLVM-IR from SCoPs", false, false); 1096 INITIALIZE_PASS_DEPENDENCY(CloogInfo); 1097 INITIALIZE_PASS_DEPENDENCY(Dependences); 1098 INITIALIZE_PASS_DEPENDENCY(DominatorTreeWrapperPass); 1099 INITIALIZE_PASS_DEPENDENCY(RegionInfoPass); 1100 INITIALIZE_PASS_DEPENDENCY(DataLayoutPass); 1101 INITIALIZE_PASS_DEPENDENCY(ScalarEvolution); 1102 INITIALIZE_PASS_DEPENDENCY(ScopDetection); 1103 INITIALIZE_PASS_DEPENDENCY(DataLayoutPass); 1104 INITIALIZE_PASS_END(CodeGeneration, "polly-codegen", 1105 "Polly - Create LLVM-IR from SCoPs", false, false) 1106 1107 #endif // CLOOG_FOUND 1108