1 //===------ CodeGeneration.cpp - Code generate the Scops. -----------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // The CodeGeneration pass takes a Scop created by ScopInfo and translates it 11 // back to LLVM-IR using Cloog. 12 // 13 // The Scop describes the high level memory behaviour of a control flow region. 14 // Transformation passes can update the schedule (execution order) of statements 15 // in the Scop. Cloog is used to generate an abstract syntax tree (clast) that 16 // reflects the updated execution order. This clast is used to create new 17 // LLVM-IR that is computational equivalent to the original control flow region, 18 // but executes its code in the new execution order defined by the changed 19 // scattering. 20 // 21 //===----------------------------------------------------------------------===// 22 23 #include "polly/CodeGen/Cloog.h" 24 #ifdef CLOOG_FOUND 25 26 #include "polly/Dependences.h" 27 #include "polly/LinkAllPasses.h" 28 #include "polly/Options.h" 29 #include "polly/ScopInfo.h" 30 #include "polly/TempScopInfo.h" 31 #include "polly/CodeGen/CodeGeneration.h" 32 #include "polly/CodeGen/BlockGenerators.h" 33 #include "polly/CodeGen/LoopGenerators.h" 34 #include "polly/CodeGen/PTXGenerator.h" 35 #include "polly/CodeGen/Utils.h" 36 #include "polly/Support/GICHelper.h" 37 #include "polly/Support/ScopHelper.h" 38 39 #include "llvm/IR/Module.h" 40 #include "llvm/ADT/SetVector.h" 41 #include "llvm/ADT/PostOrderIterator.h" 42 #include "llvm/Analysis/LoopInfo.h" 43 #include "llvm/Analysis/ScalarEvolutionExpander.h" 44 #include "llvm/Support/Debug.h" 45 #include "llvm/IR/DataLayout.h" 46 #include "llvm/Transforms/Utils/BasicBlockUtils.h" 47 48 #define CLOOG_INT_GMP 1 49 #include "cloog/cloog.h" 50 #include "cloog/isl/cloog.h" 51 52 #include "isl/aff.h" 53 54 #include <vector> 55 #include <utility> 56 57 using namespace polly; 58 using namespace llvm; 59 60 #define DEBUG_TYPE "polly-codegen" 61 62 struct isl_set; 63 64 namespace polly { 65 static cl::opt<bool> 66 OpenMP("enable-polly-openmp", cl::desc("Generate OpenMP parallel code"), 67 cl::value_desc("OpenMP code generation enabled if true"), 68 cl::init(false), cl::ZeroOrMore, cl::cat(PollyCategory)); 69 70 #ifdef GPU_CODEGEN 71 static cl::opt<bool> 72 GPGPU("enable-polly-gpgpu", cl::desc("Generate GPU parallel code"), 73 cl::Hidden, cl::value_desc("GPGPU code generation enabled if true"), 74 cl::init(false), cl::ZeroOrMore, cl::cat(PollyCategory)); 75 76 static cl::opt<std::string> 77 GPUTriple("polly-gpgpu-triple", 78 cl::desc("Target triple for GPU code generation"), cl::Hidden, 79 cl::init(""), cl::cat(PollyCategory)); 80 #endif /* GPU_CODEGEN */ 81 82 typedef DenseMap<const char *, Value *> CharMapT; 83 84 /// Class to generate LLVM-IR that calculates the value of a clast_expr. 85 class ClastExpCodeGen { 86 PollyIRBuilder &Builder; 87 const CharMapT &IVS; 88 89 Value *codegen(const clast_name *e, Type *Ty); 90 Value *codegen(const clast_term *e, Type *Ty); 91 Value *codegen(const clast_binary *e, Type *Ty); 92 Value *codegen(const clast_reduction *r, Type *Ty); 93 94 public: 95 // A generator for clast expressions. 96 // 97 // @param B The IRBuilder that defines where the code to calculate the 98 // clast expressions should be inserted. 99 // @param IVMAP A Map that translates strings describing the induction 100 // variables to the Values* that represent these variables 101 // on the LLVM side. 102 ClastExpCodeGen(PollyIRBuilder &B, CharMapT &IVMap); 103 104 // Generates code to calculate a given clast expression. 105 // 106 // @param e The expression to calculate. 107 // @return The Value that holds the result. 108 Value *codegen(const clast_expr *e, Type *Ty); 109 }; 110 111 Value *ClastExpCodeGen::codegen(const clast_name *e, Type *Ty) { 112 CharMapT::const_iterator I = IVS.find(e->name); 113 114 assert(I != IVS.end() && "Clast name not found"); 115 116 return Builder.CreateSExtOrBitCast(I->second, Ty); 117 } 118 119 static APInt APInt_from_MPZ(const mpz_t mpz) { 120 uint64_t *p = nullptr; 121 size_t sz; 122 123 p = (uint64_t *)mpz_export(p, &sz, -1, sizeof(uint64_t), 0, 0, mpz); 124 125 if (p) { 126 APInt A((unsigned)mpz_sizeinbase(mpz, 2), (unsigned)sz, p); 127 A = A.zext(A.getBitWidth() + 1); 128 free(p); 129 130 if (mpz_sgn(mpz) == -1) 131 return -A; 132 else 133 return A; 134 } else { 135 uint64_t val = 0; 136 return APInt(1, 1, &val); 137 } 138 } 139 140 Value *ClastExpCodeGen::codegen(const clast_term *e, Type *Ty) { 141 APInt a = APInt_from_MPZ(e->val); 142 143 Value *ConstOne = ConstantInt::get(Builder.getContext(), a); 144 ConstOne = Builder.CreateSExtOrBitCast(ConstOne, Ty); 145 146 if (!e->var) 147 return ConstOne; 148 149 Value *var = codegen(e->var, Ty); 150 return Builder.CreateMul(ConstOne, var); 151 } 152 153 Value *ClastExpCodeGen::codegen(const clast_binary *e, Type *Ty) { 154 Value *LHS = codegen(e->LHS, Ty); 155 156 APInt RHS_AP = APInt_from_MPZ(e->RHS); 157 158 Value *RHS = ConstantInt::get(Builder.getContext(), RHS_AP); 159 RHS = Builder.CreateSExtOrBitCast(RHS, Ty); 160 161 switch (e->type) { 162 case clast_bin_mod: 163 return Builder.CreateSRem(LHS, RHS); 164 case clast_bin_fdiv: { 165 // floord(n,d) ((n < 0) ? (n - d + 1) : n) / d 166 Value *One = ConstantInt::get(Ty, 1); 167 Value *Zero = ConstantInt::get(Ty, 0); 168 Value *Sum1 = Builder.CreateSub(LHS, RHS); 169 Value *Sum2 = Builder.CreateAdd(Sum1, One); 170 Value *isNegative = Builder.CreateICmpSLT(LHS, Zero); 171 Value *Dividend = Builder.CreateSelect(isNegative, Sum2, LHS); 172 return Builder.CreateSDiv(Dividend, RHS); 173 } 174 case clast_bin_cdiv: { 175 // ceild(n,d) ((n < 0) ? n : (n + d - 1)) / d 176 Value *One = ConstantInt::get(Ty, 1); 177 Value *Zero = ConstantInt::get(Ty, 0); 178 Value *Sum1 = Builder.CreateAdd(LHS, RHS); 179 Value *Sum2 = Builder.CreateSub(Sum1, One); 180 Value *isNegative = Builder.CreateICmpSLT(LHS, Zero); 181 Value *Dividend = Builder.CreateSelect(isNegative, LHS, Sum2); 182 return Builder.CreateSDiv(Dividend, RHS); 183 } 184 case clast_bin_div: 185 return Builder.CreateSDiv(LHS, RHS); 186 } 187 188 llvm_unreachable("Unknown clast binary expression type"); 189 } 190 191 Value *ClastExpCodeGen::codegen(const clast_reduction *r, Type *Ty) { 192 assert((r->type == clast_red_min || r->type == clast_red_max || 193 r->type == clast_red_sum) && 194 "Clast reduction type not supported"); 195 Value *old = codegen(r->elts[0], Ty); 196 197 for (int i = 1; i < r->n; ++i) { 198 Value *exprValue = codegen(r->elts[i], Ty); 199 200 switch (r->type) { 201 case clast_red_min: { 202 Value *cmp = Builder.CreateICmpSLT(old, exprValue); 203 old = Builder.CreateSelect(cmp, old, exprValue); 204 break; 205 } 206 case clast_red_max: { 207 Value *cmp = Builder.CreateICmpSGT(old, exprValue); 208 old = Builder.CreateSelect(cmp, old, exprValue); 209 break; 210 } 211 case clast_red_sum: 212 old = Builder.CreateAdd(old, exprValue); 213 break; 214 } 215 } 216 217 return old; 218 } 219 220 ClastExpCodeGen::ClastExpCodeGen(PollyIRBuilder &B, CharMapT &IVMap) 221 : Builder(B), IVS(IVMap) {} 222 223 Value *ClastExpCodeGen::codegen(const clast_expr *e, Type *Ty) { 224 switch (e->type) { 225 case clast_expr_name: 226 return codegen((const clast_name *)e, Ty); 227 case clast_expr_term: 228 return codegen((const clast_term *)e, Ty); 229 case clast_expr_bin: 230 return codegen((const clast_binary *)e, Ty); 231 case clast_expr_red: 232 return codegen((const clast_reduction *)e, Ty); 233 } 234 235 llvm_unreachable("Unknown clast expression!"); 236 } 237 238 class ClastStmtCodeGen { 239 public: 240 const std::vector<std::string> &getParallelLoops(); 241 242 private: 243 // The Scop we code generate. 244 Scop *S; 245 Pass *P; 246 LoopInfo &LI; 247 ScalarEvolution &SE; 248 DominatorTree &DT; 249 250 // The Builder specifies the current location to code generate at. 251 PollyIRBuilder &Builder; 252 253 // Map the Values from the old code to their counterparts in the new code. 254 ValueMapT ValueMap; 255 256 // Map the loops from the old code to expressions function of the induction 257 // variables in the new code. For example, when the code generator produces 258 // this AST: 259 // 260 // for (int c1 = 0; c1 <= 1023; c1 += 1) 261 // for (int c2 = 0; c2 <= 1023; c2 += 1) 262 // Stmt(c2 + 3, c1); 263 // 264 // LoopToScev is a map associating: 265 // "outer loop in the old loop nest" -> SCEV("c2 + 3"), 266 // "inner loop in the old loop nest" -> SCEV("c1"). 267 LoopToScevMapT LoopToScev; 268 269 // clastVars maps from the textual representation of a clast variable to its 270 // current *Value. clast variables are scheduling variables, original 271 // induction variables or parameters. They are used either in loop bounds or 272 // to define the statement instance that is executed. 273 // 274 // for (s = 0; s < n + 3; ++i) 275 // for (t = s; t < m; ++j) 276 // Stmt(i = s + 3 * m, j = t); 277 // 278 // {s,t,i,j,n,m} is the set of clast variables in this clast. 279 CharMapT ClastVars; 280 281 // Codegenerator for clast expressions. 282 ClastExpCodeGen ExpGen; 283 284 // Do we currently generate parallel code? 285 bool parallelCodeGeneration; 286 287 std::vector<std::string> parallelLoops; 288 289 void codegen(const clast_assignment *a); 290 291 void codegen(const clast_assignment *a, ScopStmt *Statement, 292 unsigned Dimension, int vectorDim, 293 std::vector<ValueMapT> *VectorVMap = 0, 294 std::vector<LoopToScevMapT> *VLTS = 0); 295 296 void codegenSubstitutions(const clast_stmt *Assignment, ScopStmt *Statement, 297 int vectorDim = 0, 298 std::vector<ValueMapT> *VectorVMap = 0, 299 std::vector<LoopToScevMapT> *VLTS = 0); 300 301 void codegen(const clast_user_stmt *u, std::vector<Value *> *IVS = nullptr, 302 const char *iterator = nullptr, 303 __isl_take isl_set *scatteringDomain = 0); 304 305 void codegen(const clast_block *b); 306 307 /// @brief Create a classical sequential loop. 308 void codegenForSequential(const clast_for *f); 309 310 /// @brief Create OpenMP structure values. 311 /// 312 /// Create a list of values that has to be stored into the OpenMP subfuncition 313 /// structure. 314 SetVector<Value *> getOMPValues(const clast_stmt *Body); 315 316 /// @brief Update ClastVars and ValueMap according to a value map. 317 /// 318 /// @param VMap A map from old to new values. 319 void updateWithValueMap(OMPGenerator::ValueToValueMapTy &VMap); 320 321 /// @brief Create an OpenMP parallel for loop. 322 /// 323 /// This loop reflects a loop as if it would have been created by an OpenMP 324 /// statement. 325 void codegenForOpenMP(const clast_for *f); 326 327 #ifdef GPU_CODEGEN 328 /// @brief Create GPGPU device memory access values. 329 /// 330 /// Create a list of values that will be set to be parameters of the GPGPU 331 /// subfunction. These parameters represent device memory base addresses 332 /// and the size in bytes. 333 SetVector<Value *> getGPUValues(unsigned &OutputBytes); 334 335 /// @brief Create a GPU parallel for loop. 336 /// 337 /// This loop reflects a loop as if it would have been created by a GPU 338 /// statement. 339 void codegenForGPGPU(const clast_for *F); 340 341 /// @brief Get innermost for loop. 342 const clast_stmt *getScheduleInfo(const clast_for *F, 343 std::vector<int> &NumIters, 344 unsigned &LoopDepth, 345 unsigned &NonPLoopDepth); 346 #endif /* GPU_CODEGEN */ 347 348 /// @brief Check if a loop is parallel 349 /// 350 /// Detect if a clast_for loop can be executed in parallel. 351 /// 352 /// @param For The clast for loop to check. 353 /// 354 /// @return bool Returns true if the incoming clast_for statement can 355 /// execute in parallel. 356 bool isParallelFor(const clast_for *For); 357 358 bool isInnermostLoop(const clast_for *f); 359 360 /// @brief Get the number of loop iterations for this loop. 361 /// @param f The clast for loop to check. 362 int getNumberOfIterations(const clast_for *f); 363 364 /// @brief Create vector instructions for this loop. 365 void codegenForVector(const clast_for *f); 366 367 void codegen(const clast_for *f); 368 369 Value *codegen(const clast_equation *eq); 370 371 void codegen(const clast_guard *g); 372 373 void codegen(const clast_stmt *stmt); 374 375 void addParameters(const CloogNames *names); 376 377 IntegerType *getIntPtrTy(); 378 379 public: 380 void codegen(const clast_root *r); 381 382 ClastStmtCodeGen(Scop *scop, PollyIRBuilder &B, Pass *P); 383 }; 384 } 385 386 IntegerType *ClastStmtCodeGen::getIntPtrTy() { 387 return P->getAnalysis<DataLayoutPass>().getDataLayout().getIntPtrType( 388 Builder.getContext()); 389 } 390 391 const std::vector<std::string> &ClastStmtCodeGen::getParallelLoops() { 392 return parallelLoops; 393 } 394 395 void ClastStmtCodeGen::codegen(const clast_assignment *a) { 396 Value *V = ExpGen.codegen(a->RHS, getIntPtrTy()); 397 ClastVars[a->LHS] = V; 398 } 399 400 void ClastStmtCodeGen::codegen(const clast_assignment *A, ScopStmt *Stmt, 401 unsigned Dim, int VectorDim, 402 std::vector<ValueMapT> *VectorVMap, 403 std::vector<LoopToScevMapT> *VLTS) { 404 Value *RHS; 405 406 assert(!A->LHS && "Statement assignments do not have left hand side"); 407 408 RHS = ExpGen.codegen(A->RHS, Builder.getInt64Ty()); 409 410 const llvm::SCEV *URHS = S->getSE()->getUnknown(RHS); 411 if (VLTS) 412 (*VLTS)[VectorDim][Stmt->getLoopForDimension(Dim)] = URHS; 413 LoopToScev[Stmt->getLoopForDimension(Dim)] = URHS; 414 415 const PHINode *PN = Stmt->getInductionVariableForDimension(Dim); 416 if (PN) { 417 RHS = Builder.CreateTruncOrBitCast(RHS, PN->getType()); 418 419 if (VectorVMap) 420 (*VectorVMap)[VectorDim][PN] = RHS; 421 422 ValueMap[PN] = RHS; 423 } 424 } 425 426 void ClastStmtCodeGen::codegenSubstitutions(const clast_stmt *Assignment, 427 ScopStmt *Statement, int vectorDim, 428 std::vector<ValueMapT> *VectorVMap, 429 std::vector<LoopToScevMapT> *VLTS) { 430 int Dimension = 0; 431 432 while (Assignment) { 433 assert(CLAST_STMT_IS_A(Assignment, stmt_ass) && 434 "Substitions are expected to be assignments"); 435 codegen((const clast_assignment *)Assignment, Statement, Dimension, 436 vectorDim, VectorVMap, VLTS); 437 Assignment = Assignment->next; 438 Dimension++; 439 } 440 } 441 442 // Takes the cloog specific domain and translates it into a map Statement -> 443 // PartialSchedule, where the PartialSchedule contains all the dimensions that 444 // have been code generated up to this point. 445 static __isl_give isl_map *extractPartialSchedule(ScopStmt *Statement, 446 __isl_take isl_set *Domain) { 447 isl_map *Schedule = Statement->getScattering(); 448 int ScheduledDimensions = isl_set_dim(Domain, isl_dim_set); 449 int UnscheduledDimensions = 450 isl_map_dim(Schedule, isl_dim_out) - ScheduledDimensions; 451 452 isl_set_free(Domain); 453 454 return isl_map_project_out(Schedule, isl_dim_out, ScheduledDimensions, 455 UnscheduledDimensions); 456 } 457 458 void ClastStmtCodeGen::codegen(const clast_user_stmt *u, 459 std::vector<Value *> *IVS, const char *iterator, 460 __isl_take isl_set *Domain) { 461 ScopStmt *Statement = (ScopStmt *)u->statement->usr; 462 463 if (u->substitutions) 464 codegenSubstitutions(u->substitutions, Statement); 465 466 int VectorDimensions = IVS ? IVS->size() : 1; 467 468 if (VectorDimensions == 1) { 469 BlockGenerator::generate(Builder, *Statement, ValueMap, LoopToScev, P, LI, 470 SE); 471 isl_set_free(Domain); 472 return; 473 } 474 475 VectorValueMapT VectorMap(VectorDimensions); 476 std::vector<LoopToScevMapT> VLTS(VectorDimensions); 477 478 if (IVS) { 479 assert(u->substitutions && "Substitutions expected!"); 480 int i = 0; 481 for (Value *IV : *IVS) { 482 ClastVars[iterator] = IV; 483 codegenSubstitutions(u->substitutions, Statement, i, &VectorMap, &VLTS); 484 i++; 485 } 486 } 487 488 // Copy the current value map into all vector maps if the key wasn't 489 // available yet. This is needed in case vector codegen is performed in 490 // OpenMP subfunctions. 491 for (auto KV : ValueMap) 492 for (int i = 0; i < VectorDimensions; ++i) 493 VectorMap[i].insert(KV); 494 495 isl_map *Schedule = extractPartialSchedule(Statement, Domain); 496 VectorBlockGenerator::generate(Builder, *Statement, VectorMap, VLTS, Schedule, 497 P, LI, SE); 498 isl_map_free(Schedule); 499 } 500 501 void ClastStmtCodeGen::codegen(const clast_block *b) { 502 if (b->body) 503 codegen(b->body); 504 } 505 506 void ClastStmtCodeGen::codegenForSequential(const clast_for *f) { 507 Value *LowerBound, *UpperBound, *IV, *Stride; 508 BasicBlock *ExitBlock; 509 Type *IntPtrTy = getIntPtrTy(); 510 511 LowerBound = ExpGen.codegen(f->LB, IntPtrTy); 512 UpperBound = ExpGen.codegen(f->UB, IntPtrTy); 513 Stride = Builder.getInt(APInt_from_MPZ(f->stride)); 514 515 IV = createLoop(LowerBound, UpperBound, Stride, Builder, P, LI, DT, ExitBlock, 516 CmpInst::ICMP_SLE); 517 518 // Add loop iv to symbols. 519 ClastVars[f->iterator] = IV; 520 521 if (f->body) 522 codegen(f->body); 523 524 // Loop is finished, so remove its iv from the live symbols. 525 ClastVars.erase(f->iterator); 526 Builder.SetInsertPoint(ExitBlock->begin()); 527 } 528 529 // Helper class to determine all scalar parameters used in the basic blocks of a 530 // clast. Scalar parameters are scalar variables defined outside of the SCoP. 531 class ParameterVisitor : public ClastVisitor { 532 std::set<Value *> Values; 533 534 public: 535 ParameterVisitor() : ClastVisitor(), Values() {} 536 537 void visitUser(const clast_user_stmt *Stmt) { 538 const ScopStmt *S = static_cast<const ScopStmt *>(Stmt->statement->usr); 539 const BasicBlock *BB = S->getBasicBlock(); 540 541 // Check all the operands of instructions in the basic block. 542 for (const Instruction &Inst : *BB) { 543 for (Value *SrcVal : Inst.operands()) { 544 if (Instruction *OpInst = dyn_cast<Instruction>(SrcVal)) 545 if (S->getParent()->getRegion().contains(OpInst)) 546 continue; 547 548 if (isa<Instruction>(SrcVal) || isa<Argument>(SrcVal)) 549 Values.insert(SrcVal); 550 } 551 } 552 } 553 554 // Iterator to iterate over the values found. 555 typedef std::set<Value *>::const_iterator const_iterator; 556 inline const_iterator begin() const { return Values.begin(); } 557 inline const_iterator end() const { return Values.end(); } 558 }; 559 560 SetVector<Value *> ClastStmtCodeGen::getOMPValues(const clast_stmt *Body) { 561 SetVector<Value *> Values; 562 563 // The clast variables 564 for (auto I : ClastVars) 565 Values.insert(I.second); 566 567 // Find the temporaries that are referenced in the clast statements' 568 // basic blocks but are not defined by these blocks (e.g., references 569 // to function arguments or temporaries defined before the start of 570 // the SCoP). 571 ParameterVisitor Params; 572 Params.visit(Body); 573 574 for (Value *V : Params) { 575 Values.insert(V); 576 DEBUG(dbgs() << "Adding temporary for OMP copy-in: " << *V << "\n"); 577 } 578 579 return Values; 580 } 581 582 void 583 ClastStmtCodeGen::updateWithValueMap(OMPGenerator::ValueToValueMapTy &VMap) { 584 std::set<Value *> Inserted; 585 586 for (auto I : ClastVars) { 587 ClastVars[I.first] = VMap[I.second]; 588 Inserted.insert(I.second); 589 } 590 591 for (auto I : VMap) { 592 if (Inserted.count(I.first)) 593 continue; 594 595 ValueMap[I.first] = I.second; 596 } 597 } 598 599 static void clearDomtree(Function *F, DominatorTree &DT) { 600 DomTreeNode *N = DT.getNode(&F->getEntryBlock()); 601 std::vector<BasicBlock *> Nodes; 602 for (po_iterator<DomTreeNode *> I = po_begin(N), E = po_end(N); I != E; ++I) 603 Nodes.push_back(I->getBlock()); 604 605 for (BasicBlock *BB : Nodes) 606 DT.eraseNode(BB); 607 } 608 609 void ClastStmtCodeGen::codegenForOpenMP(const clast_for *For) { 610 Value *Stride, *LB, *UB, *IV; 611 BasicBlock::iterator LoopBody; 612 IntegerType *IntPtrTy = getIntPtrTy(); 613 SetVector<Value *> Values; 614 OMPGenerator::ValueToValueMapTy VMap; 615 OMPGenerator OMPGen(Builder, P); 616 617 Stride = Builder.getInt(APInt_from_MPZ(For->stride)); 618 Stride = Builder.CreateSExtOrBitCast(Stride, IntPtrTy); 619 LB = ExpGen.codegen(For->LB, IntPtrTy); 620 UB = ExpGen.codegen(For->UB, IntPtrTy); 621 622 Values = getOMPValues(For->body); 623 624 IV = OMPGen.createParallelLoop(LB, UB, Stride, Values, VMap, &LoopBody); 625 BasicBlock::iterator AfterLoop = Builder.GetInsertPoint(); 626 Builder.SetInsertPoint(LoopBody); 627 628 // Save the current values. 629 const ValueMapT ValueMapCopy = ValueMap; 630 const CharMapT ClastVarsCopy = ClastVars; 631 632 updateWithValueMap(VMap); 633 ClastVars[For->iterator] = IV; 634 635 if (For->body) 636 codegen(For->body); 637 638 // Restore the original values. 639 ValueMap = ValueMapCopy; 640 ClastVars = ClastVarsCopy; 641 642 clearDomtree((*LoopBody).getParent()->getParent(), 643 P->getAnalysis<DominatorTreeWrapperPass>().getDomTree()); 644 645 Builder.SetInsertPoint(AfterLoop); 646 } 647 648 #ifdef GPU_CODEGEN 649 static unsigned getArraySizeInBytes(const ArrayType *AT) { 650 unsigned Bytes = AT->getNumElements(); 651 if (const ArrayType *T = dyn_cast<ArrayType>(AT->getElementType())) 652 Bytes *= getArraySizeInBytes(T); 653 else 654 Bytes *= AT->getElementType()->getPrimitiveSizeInBits() / 8; 655 656 return Bytes; 657 } 658 659 SetVector<Value *> ClastStmtCodeGen::getGPUValues(unsigned &OutputBytes) { 660 SetVector<Value *> Values; 661 OutputBytes = 0; 662 663 // Record the memory reference base addresses. 664 for (ScopStmt *Stmt : *S) { 665 for (MemoryAccess *MA : *Stmt) { 666 Value *BaseAddr = MA->getBaseAddr(); 667 Values.insert((BaseAddr)); 668 669 // FIXME: we assume that there is one and only one array to be written 670 // in a SCoP. 671 int NumWrites = 0; 672 if (MA->isWrite()) { 673 ++NumWrites; 674 assert(NumWrites <= 1 && 675 "We support at most one array to be written in a SCoP."); 676 if (const PointerType *PT = 677 dyn_cast<PointerType>(BaseAddr->getType())) { 678 Type *T = PT->getArrayElementType(); 679 const ArrayType *ATy = dyn_cast<ArrayType>(T); 680 OutputBytes = getArraySizeInBytes(ATy); 681 } 682 } 683 } 684 } 685 686 return Values; 687 } 688 689 const clast_stmt *ClastStmtCodeGen::getScheduleInfo(const clast_for *F, 690 std::vector<int> &NumIters, 691 unsigned &LoopDepth, 692 unsigned &NonPLoopDepth) { 693 clast_stmt *Stmt = (clast_stmt *)F; 694 const clast_for *Result; 695 bool NonParaFlag = false; 696 LoopDepth = 0; 697 NonPLoopDepth = 0; 698 699 while (Stmt) { 700 if (CLAST_STMT_IS_A(Stmt, stmt_for)) { 701 const clast_for *T = (clast_for *)Stmt; 702 if (isParallelFor(T)) { 703 if (!NonParaFlag) { 704 NumIters.push_back(getNumberOfIterations(T)); 705 Result = T; 706 } 707 } else 708 NonParaFlag = true; 709 710 Stmt = T->body; 711 LoopDepth++; 712 continue; 713 } 714 Stmt = Stmt->next; 715 } 716 717 assert(NumIters.size() == 4 && 718 "The loops should be tiled into 4-depth parallel loops and an " 719 "innermost non-parallel one (if exist)."); 720 NonPLoopDepth = LoopDepth - NumIters.size(); 721 assert(NonPLoopDepth <= 1 && 722 "We support only one innermost non-parallel loop currently."); 723 return (const clast_stmt *)Result->body; 724 } 725 726 void ClastStmtCodeGen::codegenForGPGPU(const clast_for *F) { 727 BasicBlock::iterator LoopBody; 728 SetVector<Value *> Values; 729 SetVector<Value *> IVS; 730 std::vector<int> NumIterations; 731 PTXGenerator::ValueToValueMapTy VMap; 732 733 assert(!GPUTriple.empty() && 734 "Target triple should be set properly for GPGPU code generation."); 735 PTXGenerator PTXGen(Builder, P, GPUTriple); 736 737 // Get original IVS and ScopStmt 738 unsigned TiledLoopDepth, NonPLoopDepth; 739 const clast_stmt *InnerStmt = 740 getScheduleInfo(F, NumIterations, TiledLoopDepth, NonPLoopDepth); 741 const clast_stmt *TmpStmt; 742 const clast_user_stmt *U; 743 const clast_for *InnerFor; 744 if (CLAST_STMT_IS_A(InnerStmt, stmt_for)) { 745 InnerFor = (const clast_for *)InnerStmt; 746 TmpStmt = InnerFor->body; 747 } else 748 TmpStmt = InnerStmt; 749 U = (const clast_user_stmt *)TmpStmt; 750 ScopStmt *Statement = (ScopStmt *)U->statement->usr; 751 for (unsigned i = 0; i < Statement->getNumIterators() - NonPLoopDepth; i++) { 752 const Value *IV = Statement->getInductionVariableForDimension(i); 753 IVS.insert(const_cast<Value *>(IV)); 754 } 755 756 unsigned OutBytes; 757 Values = getGPUValues(OutBytes); 758 PTXGen.setOutputBytes(OutBytes); 759 PTXGen.startGeneration(Values, IVS, VMap, &LoopBody); 760 761 BasicBlock::iterator AfterLoop = Builder.GetInsertPoint(); 762 Builder.SetInsertPoint(LoopBody); 763 764 BasicBlock *AfterBB = 0; 765 if (NonPLoopDepth) { 766 Value *LowerBound, *UpperBound, *IV, *Stride; 767 Type *IntPtrTy = getIntPtrTy(); 768 LowerBound = ExpGen.codegen(InnerFor->LB, IntPtrTy); 769 UpperBound = ExpGen.codegen(InnerFor->UB, IntPtrTy); 770 Stride = Builder.getInt(APInt_from_MPZ(InnerFor->stride)); 771 IV = createLoop(LowerBound, UpperBound, Stride, Builder, P, LI, DT, AfterBB, 772 CmpInst::ICMP_SLE); 773 const Value *OldIV_ = Statement->getInductionVariableForDimension(2); 774 Value *OldIV = const_cast<Value *>(OldIV_); 775 VMap.insert(std::make_pair(OldIV, IV)); 776 } 777 778 updateWithValueMap(VMap); 779 780 BlockGenerator::generate(Builder, *Statement, ValueMap, LoopToScev, P, LI, 781 SE); 782 783 if (AfterBB) 784 Builder.SetInsertPoint(AfterBB->begin()); 785 786 // FIXME: The replacement of the host base address with the parameter of ptx 787 // subfunction should have been done by updateWithValueMap. We use the 788 // following codes to avoid affecting other parts of Polly. This should be 789 // fixed later. 790 Function *FN = Builder.GetInsertBlock()->getParent(); 791 for (Value *BaseAddr : Values) 792 for (BasicBlock *BB : *FN) 793 for (Instruction *Inst : *BB) 794 Inst->replaceUsesOfWith(BaseAddr, ValueMap[BaseAddr]); 795 Builder.SetInsertPoint(AfterLoop); 796 PTXGen.setLaunchingParameters(NumIterations[0], NumIterations[1], 797 NumIterations[2], NumIterations[3]); 798 PTXGen.finishGeneration(FN); 799 } 800 #endif 801 802 bool ClastStmtCodeGen::isInnermostLoop(const clast_for *f) { 803 const clast_stmt *stmt = f->body; 804 805 while (stmt) { 806 if (!CLAST_STMT_IS_A(stmt, stmt_user)) 807 return false; 808 809 stmt = stmt->next; 810 } 811 812 return true; 813 } 814 815 int ClastStmtCodeGen::getNumberOfIterations(const clast_for *For) { 816 isl_set *LoopDomain = isl_set_copy(isl_set_from_cloog_domain(For->domain)); 817 int NumberOfIterations = polly::getNumberOfIterations(LoopDomain); 818 if (NumberOfIterations == -1) 819 return -1; 820 return NumberOfIterations / mpz_get_si(For->stride) + 1; 821 } 822 823 void ClastStmtCodeGen::codegenForVector(const clast_for *F) { 824 DEBUG(dbgs() << "Vectorizing loop '" << F->iterator << "'\n";); 825 int VectorWidth = getNumberOfIterations(F); 826 827 Value *LB = ExpGen.codegen(F->LB, getIntPtrTy()); 828 829 APInt Stride = APInt_from_MPZ(F->stride); 830 IntegerType *LoopIVType = dyn_cast<IntegerType>(LB->getType()); 831 Stride = Stride.zext(LoopIVType->getBitWidth()); 832 Value *StrideValue = ConstantInt::get(LoopIVType, Stride); 833 834 std::vector<Value *> IVS(VectorWidth); 835 IVS[0] = LB; 836 837 for (int i = 1; i < VectorWidth; i++) 838 IVS[i] = Builder.CreateAdd(IVS[i - 1], StrideValue, "p_vector_iv"); 839 840 isl_set *Domain = isl_set_copy(isl_set_from_cloog_domain(F->domain)); 841 842 // Add loop iv to symbols. 843 ClastVars[F->iterator] = LB; 844 845 const clast_stmt *Stmt = F->body; 846 847 while (Stmt) { 848 codegen((const clast_user_stmt *)Stmt, &IVS, F->iterator, 849 isl_set_copy(Domain)); 850 Stmt = Stmt->next; 851 } 852 853 // Loop is finished, so remove its iv from the live symbols. 854 isl_set_free(Domain); 855 ClastVars.erase(F->iterator); 856 } 857 858 static isl_union_map *getCombinedScheduleForSpace(Scop *S, unsigned dimLevel) { 859 isl_space *Space = S->getParamSpace(); 860 isl_union_map *schedule = isl_union_map_empty(Space); 861 862 for (ScopStmt *Stmt : *S) { 863 unsigned remainingDimensions = Stmt->getNumScattering() - dimLevel; 864 isl_map *Scattering = isl_map_project_out( 865 Stmt->getScattering(), isl_dim_out, dimLevel, remainingDimensions); 866 schedule = isl_union_map_add_map(schedule, Scattering); 867 } 868 869 return schedule; 870 } 871 872 bool ClastStmtCodeGen::isParallelFor(const clast_for *f) { 873 isl_set *Domain = isl_set_copy(isl_set_from_cloog_domain(f->domain)); 874 assert(Domain && "Cannot access domain of loop"); 875 876 Dependences &D = P->getAnalysis<Dependences>(); 877 isl_union_map *Deps = 878 D.getDependences(Dependences::TYPE_RAW | Dependences::TYPE_WAW | 879 Dependences::TYPE_WAR | Dependences::TYPE_RAW); 880 isl_union_map *Schedule = 881 getCombinedScheduleForSpace(S, isl_set_n_dim(Domain)); 882 Schedule = 883 isl_union_map_intersect_range(Schedule, isl_union_set_from_set(Domain)); 884 bool IsParallel = D.isParallel(Schedule, Deps); 885 isl_union_map_free(Schedule); 886 return IsParallel; 887 } 888 889 void ClastStmtCodeGen::codegen(const clast_for *f) { 890 bool Vector = PollyVectorizerChoice != VECTORIZER_NONE; 891 if ((Vector || OpenMP) && isParallelFor(f)) { 892 if (Vector && isInnermostLoop(f) && (-1 != getNumberOfIterations(f)) && 893 (getNumberOfIterations(f) <= 16)) { 894 codegenForVector(f); 895 return; 896 } 897 898 if (OpenMP && !parallelCodeGeneration) { 899 parallelCodeGeneration = true; 900 parallelLoops.push_back(f->iterator); 901 codegenForOpenMP(f); 902 parallelCodeGeneration = false; 903 return; 904 } 905 } 906 907 #ifdef GPU_CODEGEN 908 if (GPGPU && isParallelFor(f)) { 909 if (!parallelCodeGeneration) { 910 parallelCodeGeneration = true; 911 parallelLoops.push_back(f->iterator); 912 codegenForGPGPU(f); 913 parallelCodeGeneration = false; 914 return; 915 } 916 } 917 #endif 918 919 codegenForSequential(f); 920 } 921 922 Value *ClastStmtCodeGen::codegen(const clast_equation *eq) { 923 Value *LHS = ExpGen.codegen(eq->LHS, getIntPtrTy()); 924 Value *RHS = ExpGen.codegen(eq->RHS, getIntPtrTy()); 925 CmpInst::Predicate P; 926 927 if (eq->sign == 0) 928 P = ICmpInst::ICMP_EQ; 929 else if (eq->sign > 0) 930 P = ICmpInst::ICMP_SGE; 931 else 932 P = ICmpInst::ICMP_SLE; 933 934 return Builder.CreateICmp(P, LHS, RHS); 935 } 936 937 void ClastStmtCodeGen::codegen(const clast_guard *g) { 938 Function *F = Builder.GetInsertBlock()->getParent(); 939 LLVMContext &Context = F->getContext(); 940 941 BasicBlock *CondBB = 942 SplitBlock(Builder.GetInsertBlock(), Builder.GetInsertPoint(), P); 943 CondBB->setName("polly.cond"); 944 BasicBlock *MergeBB = SplitBlock(CondBB, CondBB->begin(), P); 945 MergeBB->setName("polly.merge"); 946 BasicBlock *ThenBB = BasicBlock::Create(Context, "polly.then", F); 947 948 DominatorTree &DT = P->getAnalysis<DominatorTreeWrapperPass>().getDomTree(); 949 DT.addNewBlock(ThenBB, CondBB); 950 DT.changeImmediateDominator(MergeBB, CondBB); 951 952 CondBB->getTerminator()->eraseFromParent(); 953 954 Builder.SetInsertPoint(CondBB); 955 956 Value *Predicate = codegen(&(g->eq[0])); 957 958 for (int i = 1; i < g->n; ++i) { 959 Value *TmpPredicate = codegen(&(g->eq[i])); 960 Predicate = Builder.CreateAnd(Predicate, TmpPredicate); 961 } 962 963 Builder.CreateCondBr(Predicate, ThenBB, MergeBB); 964 Builder.SetInsertPoint(ThenBB); 965 Builder.CreateBr(MergeBB); 966 Builder.SetInsertPoint(ThenBB->begin()); 967 968 LoopInfo &LI = P->getAnalysis<LoopInfo>(); 969 Loop *L = LI.getLoopFor(CondBB); 970 if (L) 971 L->addBasicBlockToLoop(ThenBB, LI.getBase()); 972 973 codegen(g->then); 974 975 Builder.SetInsertPoint(MergeBB->begin()); 976 } 977 978 void ClastStmtCodeGen::codegen(const clast_stmt *stmt) { 979 if (CLAST_STMT_IS_A(stmt, stmt_root)) 980 assert(false && "No second root statement expected"); 981 else if (CLAST_STMT_IS_A(stmt, stmt_ass)) 982 codegen((const clast_assignment *)stmt); 983 else if (CLAST_STMT_IS_A(stmt, stmt_user)) 984 codegen((const clast_user_stmt *)stmt); 985 else if (CLAST_STMT_IS_A(stmt, stmt_block)) 986 codegen((const clast_block *)stmt); 987 else if (CLAST_STMT_IS_A(stmt, stmt_for)) 988 codegen((const clast_for *)stmt); 989 else if (CLAST_STMT_IS_A(stmt, stmt_guard)) 990 codegen((const clast_guard *)stmt); 991 992 if (stmt->next) 993 codegen(stmt->next); 994 } 995 996 void ClastStmtCodeGen::addParameters(const CloogNames *names) { 997 SCEVExpander Rewriter(P->getAnalysis<ScalarEvolution>(), "polly"); 998 999 int i = 0; 1000 for (Scop::param_iterator PI = S->param_begin(), PE = S->param_end(); 1001 PI != PE; ++PI) { 1002 assert(i < names->nb_parameters && "Not enough parameter names"); 1003 1004 const SCEV *Param = *PI; 1005 Type *Ty = Param->getType(); 1006 1007 Instruction *insertLocation = --(Builder.GetInsertBlock()->end()); 1008 Value *V = Rewriter.expandCodeFor(Param, Ty, insertLocation); 1009 ClastVars[names->parameters[i]] = V; 1010 1011 ++i; 1012 } 1013 } 1014 1015 void ClastStmtCodeGen::codegen(const clast_root *r) { 1016 addParameters(r->names); 1017 1018 parallelCodeGeneration = false; 1019 1020 const clast_stmt *stmt = (const clast_stmt *)r; 1021 if (stmt->next) 1022 codegen(stmt->next); 1023 } 1024 1025 ClastStmtCodeGen::ClastStmtCodeGen(Scop *scop, PollyIRBuilder &B, Pass *P) 1026 : S(scop), P(P), LI(P->getAnalysis<LoopInfo>()), 1027 SE(P->getAnalysis<ScalarEvolution>()), 1028 DT(P->getAnalysis<DominatorTreeWrapperPass>().getDomTree()), Builder(B), 1029 ExpGen(Builder, ClastVars) {} 1030 1031 namespace { 1032 class CodeGeneration : public ScopPass { 1033 std::vector<std::string> ParallelLoops; 1034 1035 public: 1036 static char ID; 1037 1038 CodeGeneration() : ScopPass(ID) {} 1039 1040 bool runOnScop(Scop &S) { 1041 ParallelLoops.clear(); 1042 1043 assert(!S.getRegion().isTopLevelRegion() && 1044 "Top level regions are not supported"); 1045 1046 simplifyRegion(&S, this); 1047 1048 BasicBlock *StartBlock = executeScopConditionally(S, this); 1049 1050 PollyIRBuilder Builder(StartBlock->begin()); 1051 1052 ClastStmtCodeGen CodeGen(&S, Builder, this); 1053 CloogInfo &C = getAnalysis<CloogInfo>(); 1054 CodeGen.codegen(C.getClast()); 1055 1056 ParallelLoops.insert(ParallelLoops.begin(), 1057 CodeGen.getParallelLoops().begin(), 1058 CodeGen.getParallelLoops().end()); 1059 return true; 1060 } 1061 1062 virtual void printScop(raw_ostream &OS) const { 1063 for (auto PI : ParallelLoops) 1064 OS << "Parallel loop with iterator '" << PI << "' generated\n"; 1065 } 1066 1067 virtual void getAnalysisUsage(AnalysisUsage &AU) const { 1068 AU.addRequired<CloogInfo>(); 1069 AU.addRequired<Dependences>(); 1070 AU.addRequired<DominatorTreeWrapperPass>(); 1071 AU.addRequired<RegionInfoPass>(); 1072 AU.addRequired<ScalarEvolution>(); 1073 AU.addRequired<ScopDetection>(); 1074 AU.addRequired<ScopInfo>(); 1075 AU.addRequired<DataLayoutPass>(); 1076 AU.addRequired<LoopInfo>(); 1077 1078 AU.addPreserved<CloogInfo>(); 1079 AU.addPreserved<Dependences>(); 1080 AU.addPreserved<LoopInfo>(); 1081 AU.addPreserved<DominatorTreeWrapperPass>(); 1082 AU.addPreserved<ScopDetection>(); 1083 AU.addPreserved<ScalarEvolution>(); 1084 1085 // FIXME: We do not yet add regions for the newly generated code to the 1086 // region tree. 1087 AU.addPreserved<RegionInfoPass>(); 1088 AU.addPreserved<TempScopInfo>(); 1089 AU.addPreserved<ScopInfo>(); 1090 AU.addPreservedID(IndependentBlocksID); 1091 } 1092 }; 1093 } 1094 1095 char CodeGeneration::ID = 1; 1096 1097 Pass *polly::createCodeGenerationPass() { return new CodeGeneration(); } 1098 1099 INITIALIZE_PASS_BEGIN(CodeGeneration, "polly-codegen", 1100 "Polly - Create LLVM-IR from SCoPs", false, false); 1101 INITIALIZE_PASS_DEPENDENCY(CloogInfo); 1102 INITIALIZE_PASS_DEPENDENCY(Dependences); 1103 INITIALIZE_PASS_DEPENDENCY(DominatorTreeWrapperPass); 1104 INITIALIZE_PASS_DEPENDENCY(RegionInfoPass); 1105 INITIALIZE_PASS_DEPENDENCY(ScalarEvolution); 1106 INITIALIZE_PASS_DEPENDENCY(ScopDetection); 1107 INITIALIZE_PASS_DEPENDENCY(DataLayoutPass); 1108 INITIALIZE_PASS_END(CodeGeneration, "polly-codegen", 1109 "Polly - Create LLVM-IR from SCoPs", false, false) 1110 1111 #endif // CLOOG_FOUND 1112