1 //===- CodeGenSchedule.cpp - Scheduling MachineModels ---------------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file defines structures to encapsulate the machine model as described in 10 // the target description. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenSchedule.h" 15 #include "CodeGenInstruction.h" 16 #include "CodeGenTarget.h" 17 #include "llvm/ADT/MapVector.h" 18 #include "llvm/ADT/STLExtras.h" 19 #include "llvm/ADT/SmallPtrSet.h" 20 #include "llvm/ADT/SmallSet.h" 21 #include "llvm/ADT/SmallVector.h" 22 #include "llvm/Support/Casting.h" 23 #include "llvm/Support/Debug.h" 24 #include "llvm/Support/Regex.h" 25 #include "llvm/Support/raw_ostream.h" 26 #include "llvm/TableGen/Error.h" 27 #include <algorithm> 28 #include <iterator> 29 #include <utility> 30 31 using namespace llvm; 32 33 #define DEBUG_TYPE "subtarget-emitter" 34 35 #ifndef NDEBUG 36 static void dumpIdxVec(ArrayRef<unsigned> V) { 37 for (unsigned Idx : V) 38 dbgs() << Idx << ", "; 39 } 40 #endif 41 42 namespace { 43 44 // (instrs a, b, ...) Evaluate and union all arguments. Identical to AddOp. 45 struct InstrsOp : public SetTheory::Operator { 46 void apply(SetTheory &ST, DagInit *Expr, SetTheory::RecSet &Elts, 47 ArrayRef<SMLoc> Loc) override { 48 ST.evaluate(Expr->arg_begin(), Expr->arg_end(), Elts, Loc); 49 } 50 }; 51 52 // (instregex "OpcPat",...) Find all instructions matching an opcode pattern. 53 struct InstRegexOp : public SetTheory::Operator { 54 const CodeGenTarget &Target; 55 InstRegexOp(const CodeGenTarget &t): Target(t) {} 56 57 /// Remove any text inside of parentheses from S. 58 static std::string removeParens(llvm::StringRef S) { 59 std::string Result; 60 unsigned Paren = 0; 61 // NB: We don't care about escaped parens here. 62 for (char C : S) { 63 switch (C) { 64 case '(': 65 ++Paren; 66 break; 67 case ')': 68 --Paren; 69 break; 70 default: 71 if (Paren == 0) 72 Result += C; 73 } 74 } 75 return Result; 76 } 77 78 void apply(SetTheory &ST, DagInit *Expr, SetTheory::RecSet &Elts, 79 ArrayRef<SMLoc> Loc) override { 80 ArrayRef<const CodeGenInstruction *> Instructions = 81 Target.getInstructionsByEnumValue(); 82 83 unsigned NumGeneric = Target.getNumFixedInstructions(); 84 unsigned NumPseudos = Target.getNumPseudoInstructions(); 85 auto Generics = Instructions.slice(0, NumGeneric); 86 auto Pseudos = Instructions.slice(NumGeneric, NumPseudos); 87 auto NonPseudos = Instructions.slice(NumGeneric + NumPseudos); 88 89 for (Init *Arg : make_range(Expr->arg_begin(), Expr->arg_end())) { 90 StringInit *SI = dyn_cast<StringInit>(Arg); 91 if (!SI) 92 PrintFatalError(Loc, "instregex requires pattern string: " + 93 Expr->getAsString()); 94 StringRef Original = SI->getValue(); 95 96 // Extract a prefix that we can binary search on. 97 static const char RegexMetachars[] = "()^$|*+?.[]\\{}"; 98 auto FirstMeta = Original.find_first_of(RegexMetachars); 99 100 // Look for top-level | or ?. We cannot optimize them to binary search. 101 if (removeParens(Original).find_first_of("|?") != std::string::npos) 102 FirstMeta = 0; 103 104 Optional<Regex> Regexpr = None; 105 StringRef Prefix = Original.substr(0, FirstMeta); 106 StringRef PatStr = Original.substr(FirstMeta); 107 if (!PatStr.empty()) { 108 // For the rest use a python-style prefix match. 109 std::string pat = std::string(PatStr); 110 if (pat[0] != '^') { 111 pat.insert(0, "^("); 112 pat.insert(pat.end(), ')'); 113 } 114 Regexpr = Regex(pat); 115 } 116 117 int NumMatches = 0; 118 119 // The generic opcodes are unsorted, handle them manually. 120 for (auto *Inst : Generics) { 121 StringRef InstName = Inst->TheDef->getName(); 122 if (InstName.startswith(Prefix) && 123 (!Regexpr || Regexpr->match(InstName.substr(Prefix.size())))) { 124 Elts.insert(Inst->TheDef); 125 NumMatches++; 126 } 127 } 128 129 // Target instructions are split into two ranges: pseudo instructions 130 // first, than non-pseudos. Each range is in lexicographical order 131 // sorted by name. Find the sub-ranges that start with our prefix. 132 struct Comp { 133 bool operator()(const CodeGenInstruction *LHS, StringRef RHS) { 134 return LHS->TheDef->getName() < RHS; 135 } 136 bool operator()(StringRef LHS, const CodeGenInstruction *RHS) { 137 return LHS < RHS->TheDef->getName() && 138 !RHS->TheDef->getName().startswith(LHS); 139 } 140 }; 141 auto Range1 = 142 std::equal_range(Pseudos.begin(), Pseudos.end(), Prefix, Comp()); 143 auto Range2 = std::equal_range(NonPseudos.begin(), NonPseudos.end(), 144 Prefix, Comp()); 145 146 // For these ranges we know that instruction names start with the prefix. 147 // Check if there's a regex that needs to be checked. 148 const auto HandleNonGeneric = [&](const CodeGenInstruction *Inst) { 149 StringRef InstName = Inst->TheDef->getName(); 150 if (!Regexpr || Regexpr->match(InstName.substr(Prefix.size()))) { 151 Elts.insert(Inst->TheDef); 152 NumMatches++; 153 } 154 }; 155 std::for_each(Range1.first, Range1.second, HandleNonGeneric); 156 std::for_each(Range2.first, Range2.second, HandleNonGeneric); 157 158 if (0 == NumMatches) 159 PrintFatalError(Loc, "instregex has no matches: " + Original); 160 } 161 } 162 }; 163 164 } // end anonymous namespace 165 166 /// CodeGenModels ctor interprets machine model records and populates maps. 167 CodeGenSchedModels::CodeGenSchedModels(RecordKeeper &RK, 168 const CodeGenTarget &TGT): 169 Records(RK), Target(TGT) { 170 171 Sets.addFieldExpander("InstRW", "Instrs"); 172 173 // Allow Set evaluation to recognize the dags used in InstRW records: 174 // (instrs Op1, Op1...) 175 Sets.addOperator("instrs", std::make_unique<InstrsOp>()); 176 Sets.addOperator("instregex", std::make_unique<InstRegexOp>(Target)); 177 178 // Instantiate a CodeGenProcModel for each SchedMachineModel with the values 179 // that are explicitly referenced in tablegen records. Resources associated 180 // with each processor will be derived later. Populate ProcModelMap with the 181 // CodeGenProcModel instances. 182 collectProcModels(); 183 184 // Instantiate a CodeGenSchedRW for each SchedReadWrite record explicitly 185 // defined, and populate SchedReads and SchedWrites vectors. Implicit 186 // SchedReadWrites that represent sequences derived from expanded variant will 187 // be inferred later. 188 collectSchedRW(); 189 190 // Instantiate a CodeGenSchedClass for each unique SchedRW signature directly 191 // required by an instruction definition, and populate SchedClassIdxMap. Set 192 // NumItineraryClasses to the number of explicit itinerary classes referenced 193 // by instructions. Set NumInstrSchedClasses to the number of itinerary 194 // classes plus any classes implied by instructions that derive from class 195 // Sched and provide SchedRW list. This does not infer any new classes from 196 // SchedVariant. 197 collectSchedClasses(); 198 199 // Find instruction itineraries for each processor. Sort and populate 200 // CodeGenProcModel::ItinDefList. (Cycle-to-cycle itineraries). This requires 201 // all itinerary classes to be discovered. 202 collectProcItins(); 203 204 // Find ItinRW records for each processor and itinerary class. 205 // (For per-operand resources mapped to itinerary classes). 206 collectProcItinRW(); 207 208 // Find UnsupportedFeatures records for each processor. 209 // (For per-operand resources mapped to itinerary classes). 210 collectProcUnsupportedFeatures(); 211 212 // Infer new SchedClasses from SchedVariant. 213 inferSchedClasses(); 214 215 // Populate each CodeGenProcModel's WriteResDefs, ReadAdvanceDefs, and 216 // ProcResourceDefs. 217 LLVM_DEBUG( 218 dbgs() << "\n+++ RESOURCE DEFINITIONS (collectProcResources) +++\n"); 219 collectProcResources(); 220 221 // Collect optional processor description. 222 collectOptionalProcessorInfo(); 223 224 // Check MCInstPredicate definitions. 225 checkMCInstPredicates(); 226 227 // Check STIPredicate definitions. 228 checkSTIPredicates(); 229 230 // Find STIPredicate definitions for each processor model, and construct 231 // STIPredicateFunction objects. 232 collectSTIPredicates(); 233 234 checkCompleteness(); 235 } 236 237 void CodeGenSchedModels::checkSTIPredicates() const { 238 DenseMap<StringRef, const Record *> Declarations; 239 240 // There cannot be multiple declarations with the same name. 241 const RecVec Decls = Records.getAllDerivedDefinitions("STIPredicateDecl"); 242 for (const Record *R : Decls) { 243 StringRef Name = R->getValueAsString("Name"); 244 const auto It = Declarations.find(Name); 245 if (It == Declarations.end()) { 246 Declarations[Name] = R; 247 continue; 248 } 249 250 PrintError(R->getLoc(), "STIPredicate " + Name + " multiply declared."); 251 PrintFatalNote(It->second->getLoc(), "Previous declaration was here."); 252 } 253 254 // Disallow InstructionEquivalenceClasses with an empty instruction list. 255 const RecVec Defs = 256 Records.getAllDerivedDefinitions("InstructionEquivalenceClass"); 257 for (const Record *R : Defs) { 258 RecVec Opcodes = R->getValueAsListOfDefs("Opcodes"); 259 if (Opcodes.empty()) { 260 PrintFatalError(R->getLoc(), "Invalid InstructionEquivalenceClass " 261 "defined with an empty opcode list."); 262 } 263 } 264 } 265 266 // Used by function `processSTIPredicate` to construct a mask of machine 267 // instruction operands. 268 static APInt constructOperandMask(ArrayRef<int64_t> Indices) { 269 APInt OperandMask; 270 if (Indices.empty()) 271 return OperandMask; 272 273 int64_t MaxIndex = *std::max_element(Indices.begin(), Indices.end()); 274 assert(MaxIndex >= 0 && "Invalid negative indices in input!"); 275 OperandMask = OperandMask.zext(MaxIndex + 1); 276 for (const int64_t Index : Indices) { 277 assert(Index >= 0 && "Invalid negative indices!"); 278 OperandMask.setBit(Index); 279 } 280 281 return OperandMask; 282 } 283 284 static void 285 processSTIPredicate(STIPredicateFunction &Fn, 286 const DenseMap<Record *, unsigned> &ProcModelMap) { 287 DenseMap<const Record *, unsigned> Opcode2Index; 288 using OpcodeMapPair = std::pair<const Record *, OpcodeInfo>; 289 std::vector<OpcodeMapPair> OpcodeMappings; 290 std::vector<std::pair<APInt, APInt>> OpcodeMasks; 291 292 DenseMap<const Record *, unsigned> Predicate2Index; 293 unsigned NumUniquePredicates = 0; 294 295 // Number unique predicates and opcodes used by InstructionEquivalenceClass 296 // definitions. Each unique opcode will be associated with an OpcodeInfo 297 // object. 298 for (const Record *Def : Fn.getDefinitions()) { 299 RecVec Classes = Def->getValueAsListOfDefs("Classes"); 300 for (const Record *EC : Classes) { 301 const Record *Pred = EC->getValueAsDef("Predicate"); 302 if (Predicate2Index.find(Pred) == Predicate2Index.end()) 303 Predicate2Index[Pred] = NumUniquePredicates++; 304 305 RecVec Opcodes = EC->getValueAsListOfDefs("Opcodes"); 306 for (const Record *Opcode : Opcodes) { 307 if (Opcode2Index.find(Opcode) == Opcode2Index.end()) { 308 Opcode2Index[Opcode] = OpcodeMappings.size(); 309 OpcodeMappings.emplace_back(Opcode, OpcodeInfo()); 310 } 311 } 312 } 313 } 314 315 // Initialize vector `OpcodeMasks` with default values. We want to keep track 316 // of which processors "use" which opcodes. We also want to be able to 317 // identify predicates that are used by different processors for a same 318 // opcode. 319 // This information is used later on by this algorithm to sort OpcodeMapping 320 // elements based on their processor and predicate sets. 321 OpcodeMasks.resize(OpcodeMappings.size()); 322 APInt DefaultProcMask(ProcModelMap.size(), 0); 323 APInt DefaultPredMask(NumUniquePredicates, 0); 324 for (std::pair<APInt, APInt> &MaskPair : OpcodeMasks) 325 MaskPair = std::make_pair(DefaultProcMask, DefaultPredMask); 326 327 // Construct a OpcodeInfo object for every unique opcode declared by an 328 // InstructionEquivalenceClass definition. 329 for (const Record *Def : Fn.getDefinitions()) { 330 RecVec Classes = Def->getValueAsListOfDefs("Classes"); 331 const Record *SchedModel = Def->getValueAsDef("SchedModel"); 332 unsigned ProcIndex = ProcModelMap.find(SchedModel)->second; 333 APInt ProcMask(ProcModelMap.size(), 0); 334 ProcMask.setBit(ProcIndex); 335 336 for (const Record *EC : Classes) { 337 RecVec Opcodes = EC->getValueAsListOfDefs("Opcodes"); 338 339 std::vector<int64_t> OpIndices = 340 EC->getValueAsListOfInts("OperandIndices"); 341 APInt OperandMask = constructOperandMask(OpIndices); 342 343 const Record *Pred = EC->getValueAsDef("Predicate"); 344 APInt PredMask(NumUniquePredicates, 0); 345 PredMask.setBit(Predicate2Index[Pred]); 346 347 for (const Record *Opcode : Opcodes) { 348 unsigned OpcodeIdx = Opcode2Index[Opcode]; 349 if (OpcodeMasks[OpcodeIdx].first[ProcIndex]) { 350 std::string Message = 351 "Opcode " + Opcode->getName().str() + 352 " used by multiple InstructionEquivalenceClass definitions."; 353 PrintFatalError(EC->getLoc(), Message); 354 } 355 OpcodeMasks[OpcodeIdx].first |= ProcMask; 356 OpcodeMasks[OpcodeIdx].second |= PredMask; 357 OpcodeInfo &OI = OpcodeMappings[OpcodeIdx].second; 358 359 OI.addPredicateForProcModel(ProcMask, OperandMask, Pred); 360 } 361 } 362 } 363 364 // Sort OpcodeMappings elements based on their CPU and predicate masks. 365 // As a last resort, order elements by opcode identifier. 366 llvm::sort(OpcodeMappings, 367 [&](const OpcodeMapPair &Lhs, const OpcodeMapPair &Rhs) { 368 unsigned LhsIdx = Opcode2Index[Lhs.first]; 369 unsigned RhsIdx = Opcode2Index[Rhs.first]; 370 const std::pair<APInt, APInt> &LhsMasks = OpcodeMasks[LhsIdx]; 371 const std::pair<APInt, APInt> &RhsMasks = OpcodeMasks[RhsIdx]; 372 373 auto LessThan = [](const APInt &Lhs, const APInt &Rhs) { 374 unsigned LhsCountPopulation = Lhs.countPopulation(); 375 unsigned RhsCountPopulation = Rhs.countPopulation(); 376 return ((LhsCountPopulation < RhsCountPopulation) || 377 ((LhsCountPopulation == RhsCountPopulation) && 378 (Lhs.countLeadingZeros() > Rhs.countLeadingZeros()))); 379 }; 380 381 if (LhsMasks.first != RhsMasks.first) 382 return LessThan(LhsMasks.first, RhsMasks.first); 383 384 if (LhsMasks.second != RhsMasks.second) 385 return LessThan(LhsMasks.second, RhsMasks.second); 386 387 return LhsIdx < RhsIdx; 388 }); 389 390 // Now construct opcode groups. Groups are used by the SubtargetEmitter when 391 // expanding the body of a STIPredicate function. In particular, each opcode 392 // group is expanded into a sequence of labels in a switch statement. 393 // It identifies opcodes for which different processors define same predicates 394 // and same opcode masks. 395 for (OpcodeMapPair &Info : OpcodeMappings) 396 Fn.addOpcode(Info.first, std::move(Info.second)); 397 } 398 399 void CodeGenSchedModels::collectSTIPredicates() { 400 // Map STIPredicateDecl records to elements of vector 401 // CodeGenSchedModels::STIPredicates. 402 DenseMap<const Record *, unsigned> Decl2Index; 403 404 RecVec RV = Records.getAllDerivedDefinitions("STIPredicate"); 405 for (const Record *R : RV) { 406 const Record *Decl = R->getValueAsDef("Declaration"); 407 408 const auto It = Decl2Index.find(Decl); 409 if (It == Decl2Index.end()) { 410 Decl2Index[Decl] = STIPredicates.size(); 411 STIPredicateFunction Predicate(Decl); 412 Predicate.addDefinition(R); 413 STIPredicates.emplace_back(std::move(Predicate)); 414 continue; 415 } 416 417 STIPredicateFunction &PreviousDef = STIPredicates[It->second]; 418 PreviousDef.addDefinition(R); 419 } 420 421 for (STIPredicateFunction &Fn : STIPredicates) 422 processSTIPredicate(Fn, ProcModelMap); 423 } 424 425 void OpcodeInfo::addPredicateForProcModel(const llvm::APInt &CpuMask, 426 const llvm::APInt &OperandMask, 427 const Record *Predicate) { 428 auto It = llvm::find_if( 429 Predicates, [&OperandMask, &Predicate](const PredicateInfo &P) { 430 return P.Predicate == Predicate && P.OperandMask == OperandMask; 431 }); 432 if (It == Predicates.end()) { 433 Predicates.emplace_back(CpuMask, OperandMask, Predicate); 434 return; 435 } 436 It->ProcModelMask |= CpuMask; 437 } 438 439 void CodeGenSchedModels::checkMCInstPredicates() const { 440 RecVec MCPredicates = Records.getAllDerivedDefinitions("TIIPredicate"); 441 if (MCPredicates.empty()) 442 return; 443 444 // A target cannot have multiple TIIPredicate definitions with a same name. 445 llvm::StringMap<const Record *> TIIPredicates(MCPredicates.size()); 446 for (const Record *TIIPred : MCPredicates) { 447 StringRef Name = TIIPred->getValueAsString("FunctionName"); 448 StringMap<const Record *>::const_iterator It = TIIPredicates.find(Name); 449 if (It == TIIPredicates.end()) { 450 TIIPredicates[Name] = TIIPred; 451 continue; 452 } 453 454 PrintError(TIIPred->getLoc(), 455 "TIIPredicate " + Name + " is multiply defined."); 456 PrintFatalNote(It->second->getLoc(), 457 " Previous definition of " + Name + " was here."); 458 } 459 } 460 461 void CodeGenSchedModels::collectRetireControlUnits() { 462 RecVec Units = Records.getAllDerivedDefinitions("RetireControlUnit"); 463 464 for (Record *RCU : Units) { 465 CodeGenProcModel &PM = getProcModel(RCU->getValueAsDef("SchedModel")); 466 if (PM.RetireControlUnit) { 467 PrintError(RCU->getLoc(), 468 "Expected a single RetireControlUnit definition"); 469 PrintNote(PM.RetireControlUnit->getLoc(), 470 "Previous definition of RetireControlUnit was here"); 471 } 472 PM.RetireControlUnit = RCU; 473 } 474 } 475 476 void CodeGenSchedModels::collectLoadStoreQueueInfo() { 477 RecVec Queues = Records.getAllDerivedDefinitions("MemoryQueue"); 478 479 for (Record *Queue : Queues) { 480 CodeGenProcModel &PM = getProcModel(Queue->getValueAsDef("SchedModel")); 481 if (Queue->isSubClassOf("LoadQueue")) { 482 if (PM.LoadQueue) { 483 PrintError(Queue->getLoc(), 484 "Expected a single LoadQueue definition"); 485 PrintNote(PM.LoadQueue->getLoc(), 486 "Previous definition of LoadQueue was here"); 487 } 488 489 PM.LoadQueue = Queue; 490 } 491 492 if (Queue->isSubClassOf("StoreQueue")) { 493 if (PM.StoreQueue) { 494 PrintError(Queue->getLoc(), 495 "Expected a single StoreQueue definition"); 496 PrintNote(PM.LoadQueue->getLoc(), 497 "Previous definition of StoreQueue was here"); 498 } 499 500 PM.StoreQueue = Queue; 501 } 502 } 503 } 504 505 /// Collect optional processor information. 506 void CodeGenSchedModels::collectOptionalProcessorInfo() { 507 // Find register file definitions for each processor. 508 collectRegisterFiles(); 509 510 // Collect processor RetireControlUnit descriptors if available. 511 collectRetireControlUnits(); 512 513 // Collect information about load/store queues. 514 collectLoadStoreQueueInfo(); 515 516 checkCompleteness(); 517 } 518 519 /// Gather all processor models. 520 void CodeGenSchedModels::collectProcModels() { 521 RecVec ProcRecords = Records.getAllDerivedDefinitions("Processor"); 522 llvm::sort(ProcRecords, LessRecordFieldName()); 523 524 // Reserve space because we can. Reallocation would be ok. 525 ProcModels.reserve(ProcRecords.size()+1); 526 527 // Use idx=0 for NoModel/NoItineraries. 528 Record *NoModelDef = Records.getDef("NoSchedModel"); 529 Record *NoItinsDef = Records.getDef("NoItineraries"); 530 ProcModels.emplace_back(0, "NoSchedModel", NoModelDef, NoItinsDef); 531 ProcModelMap[NoModelDef] = 0; 532 533 // For each processor, find a unique machine model. 534 LLVM_DEBUG(dbgs() << "+++ PROCESSOR MODELs (addProcModel) +++\n"); 535 for (Record *ProcRecord : ProcRecords) 536 addProcModel(ProcRecord); 537 } 538 539 /// Get a unique processor model based on the defined MachineModel and 540 /// ProcessorItineraries. 541 void CodeGenSchedModels::addProcModel(Record *ProcDef) { 542 Record *ModelKey = getModelOrItinDef(ProcDef); 543 if (!ProcModelMap.insert(std::make_pair(ModelKey, ProcModels.size())).second) 544 return; 545 546 std::string Name = std::string(ModelKey->getName()); 547 if (ModelKey->isSubClassOf("SchedMachineModel")) { 548 Record *ItinsDef = ModelKey->getValueAsDef("Itineraries"); 549 ProcModels.emplace_back(ProcModels.size(), Name, ModelKey, ItinsDef); 550 } 551 else { 552 // An itinerary is defined without a machine model. Infer a new model. 553 if (!ModelKey->getValueAsListOfDefs("IID").empty()) 554 Name = Name + "Model"; 555 ProcModels.emplace_back(ProcModels.size(), Name, 556 ProcDef->getValueAsDef("SchedModel"), ModelKey); 557 } 558 LLVM_DEBUG(ProcModels.back().dump()); 559 } 560 561 // Recursively find all reachable SchedReadWrite records. 562 static void scanSchedRW(Record *RWDef, RecVec &RWDefs, 563 SmallPtrSet<Record*, 16> &RWSet) { 564 if (!RWSet.insert(RWDef).second) 565 return; 566 RWDefs.push_back(RWDef); 567 // Reads don't currently have sequence records, but it can be added later. 568 if (RWDef->isSubClassOf("WriteSequence")) { 569 RecVec Seq = RWDef->getValueAsListOfDefs("Writes"); 570 for (Record *WSRec : Seq) 571 scanSchedRW(WSRec, RWDefs, RWSet); 572 } 573 else if (RWDef->isSubClassOf("SchedVariant")) { 574 // Visit each variant (guarded by a different predicate). 575 RecVec Vars = RWDef->getValueAsListOfDefs("Variants"); 576 for (Record *Variant : Vars) { 577 // Visit each RW in the sequence selected by the current variant. 578 RecVec Selected = Variant->getValueAsListOfDefs("Selected"); 579 for (Record *SelDef : Selected) 580 scanSchedRW(SelDef, RWDefs, RWSet); 581 } 582 } 583 } 584 585 // Collect and sort all SchedReadWrites reachable via tablegen records. 586 // More may be inferred later when inferring new SchedClasses from variants. 587 void CodeGenSchedModels::collectSchedRW() { 588 // Reserve idx=0 for invalid writes/reads. 589 SchedWrites.resize(1); 590 SchedReads.resize(1); 591 592 SmallPtrSet<Record*, 16> RWSet; 593 594 // Find all SchedReadWrites referenced by instruction defs. 595 RecVec SWDefs, SRDefs; 596 for (const CodeGenInstruction *Inst : Target.getInstructionsByEnumValue()) { 597 Record *SchedDef = Inst->TheDef; 598 if (SchedDef->isValueUnset("SchedRW")) 599 continue; 600 RecVec RWs = SchedDef->getValueAsListOfDefs("SchedRW"); 601 for (Record *RW : RWs) { 602 if (RW->isSubClassOf("SchedWrite")) 603 scanSchedRW(RW, SWDefs, RWSet); 604 else { 605 assert(RW->isSubClassOf("SchedRead") && "Unknown SchedReadWrite"); 606 scanSchedRW(RW, SRDefs, RWSet); 607 } 608 } 609 } 610 // Find all ReadWrites referenced by InstRW. 611 RecVec InstRWDefs = Records.getAllDerivedDefinitions("InstRW"); 612 for (Record *InstRWDef : InstRWDefs) { 613 // For all OperandReadWrites. 614 RecVec RWDefs = InstRWDef->getValueAsListOfDefs("OperandReadWrites"); 615 for (Record *RWDef : RWDefs) { 616 if (RWDef->isSubClassOf("SchedWrite")) 617 scanSchedRW(RWDef, SWDefs, RWSet); 618 else { 619 assert(RWDef->isSubClassOf("SchedRead") && "Unknown SchedReadWrite"); 620 scanSchedRW(RWDef, SRDefs, RWSet); 621 } 622 } 623 } 624 // Find all ReadWrites referenced by ItinRW. 625 RecVec ItinRWDefs = Records.getAllDerivedDefinitions("ItinRW"); 626 for (Record *ItinRWDef : ItinRWDefs) { 627 // For all OperandReadWrites. 628 RecVec RWDefs = ItinRWDef->getValueAsListOfDefs("OperandReadWrites"); 629 for (Record *RWDef : RWDefs) { 630 if (RWDef->isSubClassOf("SchedWrite")) 631 scanSchedRW(RWDef, SWDefs, RWSet); 632 else { 633 assert(RWDef->isSubClassOf("SchedRead") && "Unknown SchedReadWrite"); 634 scanSchedRW(RWDef, SRDefs, RWSet); 635 } 636 } 637 } 638 // Find all ReadWrites referenced by SchedAlias. AliasDefs needs to be sorted 639 // for the loop below that initializes Alias vectors. 640 RecVec AliasDefs = Records.getAllDerivedDefinitions("SchedAlias"); 641 llvm::sort(AliasDefs, LessRecord()); 642 for (Record *ADef : AliasDefs) { 643 Record *MatchDef = ADef->getValueAsDef("MatchRW"); 644 Record *AliasDef = ADef->getValueAsDef("AliasRW"); 645 if (MatchDef->isSubClassOf("SchedWrite")) { 646 if (!AliasDef->isSubClassOf("SchedWrite")) 647 PrintFatalError(ADef->getLoc(), "SchedWrite Alias must be SchedWrite"); 648 scanSchedRW(AliasDef, SWDefs, RWSet); 649 } 650 else { 651 assert(MatchDef->isSubClassOf("SchedRead") && "Unknown SchedReadWrite"); 652 if (!AliasDef->isSubClassOf("SchedRead")) 653 PrintFatalError(ADef->getLoc(), "SchedRead Alias must be SchedRead"); 654 scanSchedRW(AliasDef, SRDefs, RWSet); 655 } 656 } 657 // Sort and add the SchedReadWrites directly referenced by instructions or 658 // itinerary resources. Index reads and writes in separate domains. 659 llvm::sort(SWDefs, LessRecord()); 660 for (Record *SWDef : SWDefs) { 661 assert(!getSchedRWIdx(SWDef, /*IsRead=*/false) && "duplicate SchedWrite"); 662 SchedWrites.emplace_back(SchedWrites.size(), SWDef); 663 } 664 llvm::sort(SRDefs, LessRecord()); 665 for (Record *SRDef : SRDefs) { 666 assert(!getSchedRWIdx(SRDef, /*IsRead-*/true) && "duplicate SchedWrite"); 667 SchedReads.emplace_back(SchedReads.size(), SRDef); 668 } 669 // Initialize WriteSequence vectors. 670 for (CodeGenSchedRW &CGRW : SchedWrites) { 671 if (!CGRW.IsSequence) 672 continue; 673 findRWs(CGRW.TheDef->getValueAsListOfDefs("Writes"), CGRW.Sequence, 674 /*IsRead=*/false); 675 } 676 // Initialize Aliases vectors. 677 for (Record *ADef : AliasDefs) { 678 Record *AliasDef = ADef->getValueAsDef("AliasRW"); 679 getSchedRW(AliasDef).IsAlias = true; 680 Record *MatchDef = ADef->getValueAsDef("MatchRW"); 681 CodeGenSchedRW &RW = getSchedRW(MatchDef); 682 if (RW.IsAlias) 683 PrintFatalError(ADef->getLoc(), "Cannot Alias an Alias"); 684 RW.Aliases.push_back(ADef); 685 } 686 LLVM_DEBUG( 687 dbgs() << "\n+++ SCHED READS and WRITES (collectSchedRW) +++\n"; 688 for (unsigned WIdx = 0, WEnd = SchedWrites.size(); WIdx != WEnd; ++WIdx) { 689 dbgs() << WIdx << ": "; 690 SchedWrites[WIdx].dump(); 691 dbgs() << '\n'; 692 } for (unsigned RIdx = 0, REnd = SchedReads.size(); RIdx != REnd; 693 ++RIdx) { 694 dbgs() << RIdx << ": "; 695 SchedReads[RIdx].dump(); 696 dbgs() << '\n'; 697 } RecVec RWDefs = Records.getAllDerivedDefinitions("SchedReadWrite"); 698 for (Record *RWDef 699 : RWDefs) { 700 if (!getSchedRWIdx(RWDef, RWDef->isSubClassOf("SchedRead"))) { 701 StringRef Name = RWDef->getName(); 702 if (Name != "NoWrite" && Name != "ReadDefault") 703 dbgs() << "Unused SchedReadWrite " << Name << '\n'; 704 } 705 }); 706 } 707 708 /// Compute a SchedWrite name from a sequence of writes. 709 std::string CodeGenSchedModels::genRWName(ArrayRef<unsigned> Seq, bool IsRead) { 710 std::string Name("("); 711 for (auto I = Seq.begin(), E = Seq.end(); I != E; ++I) { 712 if (I != Seq.begin()) 713 Name += '_'; 714 Name += getSchedRW(*I, IsRead).Name; 715 } 716 Name += ')'; 717 return Name; 718 } 719 720 unsigned CodeGenSchedModels::getSchedRWIdx(const Record *Def, 721 bool IsRead) const { 722 const std::vector<CodeGenSchedRW> &RWVec = IsRead ? SchedReads : SchedWrites; 723 const auto I = find_if( 724 RWVec, [Def](const CodeGenSchedRW &RW) { return RW.TheDef == Def; }); 725 return I == RWVec.end() ? 0 : std::distance(RWVec.begin(), I); 726 } 727 728 bool CodeGenSchedModels::hasReadOfWrite(Record *WriteDef) const { 729 for (const CodeGenSchedRW &Read : SchedReads) { 730 Record *ReadDef = Read.TheDef; 731 if (!ReadDef || !ReadDef->isSubClassOf("ProcReadAdvance")) 732 continue; 733 734 RecVec ValidWrites = ReadDef->getValueAsListOfDefs("ValidWrites"); 735 if (is_contained(ValidWrites, WriteDef)) { 736 return true; 737 } 738 } 739 return false; 740 } 741 742 static void splitSchedReadWrites(const RecVec &RWDefs, 743 RecVec &WriteDefs, RecVec &ReadDefs) { 744 for (Record *RWDef : RWDefs) { 745 if (RWDef->isSubClassOf("SchedWrite")) 746 WriteDefs.push_back(RWDef); 747 else { 748 assert(RWDef->isSubClassOf("SchedRead") && "unknown SchedReadWrite"); 749 ReadDefs.push_back(RWDef); 750 } 751 } 752 } 753 754 // Split the SchedReadWrites defs and call findRWs for each list. 755 void CodeGenSchedModels::findRWs(const RecVec &RWDefs, 756 IdxVec &Writes, IdxVec &Reads) const { 757 RecVec WriteDefs; 758 RecVec ReadDefs; 759 splitSchedReadWrites(RWDefs, WriteDefs, ReadDefs); 760 findRWs(WriteDefs, Writes, false); 761 findRWs(ReadDefs, Reads, true); 762 } 763 764 // Call getSchedRWIdx for all elements in a sequence of SchedRW defs. 765 void CodeGenSchedModels::findRWs(const RecVec &RWDefs, IdxVec &RWs, 766 bool IsRead) const { 767 for (Record *RWDef : RWDefs) { 768 unsigned Idx = getSchedRWIdx(RWDef, IsRead); 769 assert(Idx && "failed to collect SchedReadWrite"); 770 RWs.push_back(Idx); 771 } 772 } 773 774 void CodeGenSchedModels::expandRWSequence(unsigned RWIdx, IdxVec &RWSeq, 775 bool IsRead) const { 776 const CodeGenSchedRW &SchedRW = getSchedRW(RWIdx, IsRead); 777 if (!SchedRW.IsSequence) { 778 RWSeq.push_back(RWIdx); 779 return; 780 } 781 int Repeat = 782 SchedRW.TheDef ? SchedRW.TheDef->getValueAsInt("Repeat") : 1; 783 for (int i = 0; i < Repeat; ++i) { 784 for (unsigned I : SchedRW.Sequence) { 785 expandRWSequence(I, RWSeq, IsRead); 786 } 787 } 788 } 789 790 // Expand a SchedWrite as a sequence following any aliases that coincide with 791 // the given processor model. 792 void CodeGenSchedModels::expandRWSeqForProc( 793 unsigned RWIdx, IdxVec &RWSeq, bool IsRead, 794 const CodeGenProcModel &ProcModel) const { 795 796 const CodeGenSchedRW &SchedWrite = getSchedRW(RWIdx, IsRead); 797 Record *AliasDef = nullptr; 798 for (const Record *Rec : SchedWrite.Aliases) { 799 const CodeGenSchedRW &AliasRW = getSchedRW(Rec->getValueAsDef("AliasRW")); 800 if (Rec->getValueInit("SchedModel")->isComplete()) { 801 Record *ModelDef = Rec->getValueAsDef("SchedModel"); 802 if (&getProcModel(ModelDef) != &ProcModel) 803 continue; 804 } 805 if (AliasDef) 806 PrintFatalError(AliasRW.TheDef->getLoc(), "Multiple aliases " 807 "defined for processor " + ProcModel.ModelName + 808 " Ensure only one SchedAlias exists per RW."); 809 AliasDef = AliasRW.TheDef; 810 } 811 if (AliasDef) { 812 expandRWSeqForProc(getSchedRWIdx(AliasDef, IsRead), 813 RWSeq, IsRead,ProcModel); 814 return; 815 } 816 if (!SchedWrite.IsSequence) { 817 RWSeq.push_back(RWIdx); 818 return; 819 } 820 int Repeat = 821 SchedWrite.TheDef ? SchedWrite.TheDef->getValueAsInt("Repeat") : 1; 822 for (int I = 0, E = Repeat; I < E; ++I) { 823 for (unsigned Idx : SchedWrite.Sequence) { 824 expandRWSeqForProc(Idx, RWSeq, IsRead, ProcModel); 825 } 826 } 827 } 828 829 // Find the existing SchedWrite that models this sequence of writes. 830 unsigned CodeGenSchedModels::findRWForSequence(ArrayRef<unsigned> Seq, 831 bool IsRead) { 832 std::vector<CodeGenSchedRW> &RWVec = IsRead ? SchedReads : SchedWrites; 833 834 auto I = find_if(RWVec, [Seq](CodeGenSchedRW &RW) { 835 return makeArrayRef(RW.Sequence) == Seq; 836 }); 837 // Index zero reserved for invalid RW. 838 return I == RWVec.end() ? 0 : std::distance(RWVec.begin(), I); 839 } 840 841 /// Add this ReadWrite if it doesn't already exist. 842 unsigned CodeGenSchedModels::findOrInsertRW(ArrayRef<unsigned> Seq, 843 bool IsRead) { 844 assert(!Seq.empty() && "cannot insert empty sequence"); 845 if (Seq.size() == 1) 846 return Seq.back(); 847 848 unsigned Idx = findRWForSequence(Seq, IsRead); 849 if (Idx) 850 return Idx; 851 852 std::vector<CodeGenSchedRW> &RWVec = IsRead ? SchedReads : SchedWrites; 853 unsigned RWIdx = RWVec.size(); 854 CodeGenSchedRW SchedRW(RWIdx, IsRead, Seq, genRWName(Seq, IsRead)); 855 RWVec.push_back(SchedRW); 856 return RWIdx; 857 } 858 859 /// Visit all the instruction definitions for this target to gather and 860 /// enumerate the itinerary classes. These are the explicitly specified 861 /// SchedClasses. More SchedClasses may be inferred. 862 void CodeGenSchedModels::collectSchedClasses() { 863 864 // NoItinerary is always the first class at Idx=0 865 assert(SchedClasses.empty() && "Expected empty sched class"); 866 SchedClasses.emplace_back(0, "NoInstrModel", 867 Records.getDef("NoItinerary")); 868 SchedClasses.back().ProcIndices.push_back(0); 869 870 // Create a SchedClass for each unique combination of itinerary class and 871 // SchedRW list. 872 for (const CodeGenInstruction *Inst : Target.getInstructionsByEnumValue()) { 873 Record *ItinDef = Inst->TheDef->getValueAsDef("Itinerary"); 874 IdxVec Writes, Reads; 875 if (!Inst->TheDef->isValueUnset("SchedRW")) 876 findRWs(Inst->TheDef->getValueAsListOfDefs("SchedRW"), Writes, Reads); 877 878 // ProcIdx == 0 indicates the class applies to all processors. 879 unsigned SCIdx = addSchedClass(ItinDef, Writes, Reads, /*ProcIndices*/{0}); 880 InstrClassMap[Inst->TheDef] = SCIdx; 881 } 882 // Create classes for InstRW defs. 883 RecVec InstRWDefs = Records.getAllDerivedDefinitions("InstRW"); 884 llvm::sort(InstRWDefs, LessRecord()); 885 LLVM_DEBUG(dbgs() << "\n+++ SCHED CLASSES (createInstRWClass) +++\n"); 886 for (Record *RWDef : InstRWDefs) 887 createInstRWClass(RWDef); 888 889 NumInstrSchedClasses = SchedClasses.size(); 890 891 bool EnableDump = false; 892 LLVM_DEBUG(EnableDump = true); 893 if (!EnableDump) 894 return; 895 896 LLVM_DEBUG( 897 dbgs() 898 << "\n+++ ITINERARIES and/or MACHINE MODELS (collectSchedClasses) +++\n"); 899 for (const CodeGenInstruction *Inst : Target.getInstructionsByEnumValue()) { 900 StringRef InstName = Inst->TheDef->getName(); 901 unsigned SCIdx = getSchedClassIdx(*Inst); 902 if (!SCIdx) { 903 LLVM_DEBUG({ 904 if (!Inst->hasNoSchedulingInfo) 905 dbgs() << "No machine model for " << Inst->TheDef->getName() << '\n'; 906 }); 907 continue; 908 } 909 CodeGenSchedClass &SC = getSchedClass(SCIdx); 910 if (SC.ProcIndices[0] != 0) 911 PrintFatalError(Inst->TheDef->getLoc(), "Instruction's sched class " 912 "must not be subtarget specific."); 913 914 IdxVec ProcIndices; 915 if (SC.ItinClassDef->getName() != "NoItinerary") { 916 ProcIndices.push_back(0); 917 dbgs() << "Itinerary for " << InstName << ": " 918 << SC.ItinClassDef->getName() << '\n'; 919 } 920 if (!SC.Writes.empty()) { 921 ProcIndices.push_back(0); 922 LLVM_DEBUG({ 923 dbgs() << "SchedRW machine model for " << InstName; 924 for (IdxIter WI = SC.Writes.begin(), WE = SC.Writes.end(); WI != WE; 925 ++WI) 926 dbgs() << " " << SchedWrites[*WI].Name; 927 for (IdxIter RI = SC.Reads.begin(), RE = SC.Reads.end(); RI != RE; ++RI) 928 dbgs() << " " << SchedReads[*RI].Name; 929 dbgs() << '\n'; 930 }); 931 } 932 const RecVec &RWDefs = SchedClasses[SCIdx].InstRWs; 933 for (Record *RWDef : RWDefs) { 934 const CodeGenProcModel &ProcModel = 935 getProcModel(RWDef->getValueAsDef("SchedModel")); 936 ProcIndices.push_back(ProcModel.Index); 937 LLVM_DEBUG(dbgs() << "InstRW on " << ProcModel.ModelName << " for " 938 << InstName); 939 IdxVec Writes; 940 IdxVec Reads; 941 findRWs(RWDef->getValueAsListOfDefs("OperandReadWrites"), 942 Writes, Reads); 943 LLVM_DEBUG({ 944 for (unsigned WIdx : Writes) 945 dbgs() << " " << SchedWrites[WIdx].Name; 946 for (unsigned RIdx : Reads) 947 dbgs() << " " << SchedReads[RIdx].Name; 948 dbgs() << '\n'; 949 }); 950 } 951 // If ProcIndices contains zero, the class applies to all processors. 952 LLVM_DEBUG({ 953 if (!std::count(ProcIndices.begin(), ProcIndices.end(), 0)) { 954 for (const CodeGenProcModel &PM : ProcModels) { 955 if (!std::count(ProcIndices.begin(), ProcIndices.end(), PM.Index)) 956 dbgs() << "No machine model for " << Inst->TheDef->getName() 957 << " on processor " << PM.ModelName << '\n'; 958 } 959 } 960 }); 961 } 962 } 963 964 // Get the SchedClass index for an instruction. 965 unsigned 966 CodeGenSchedModels::getSchedClassIdx(const CodeGenInstruction &Inst) const { 967 return InstrClassMap.lookup(Inst.TheDef); 968 } 969 970 std::string 971 CodeGenSchedModels::createSchedClassName(Record *ItinClassDef, 972 ArrayRef<unsigned> OperWrites, 973 ArrayRef<unsigned> OperReads) { 974 975 std::string Name; 976 if (ItinClassDef && ItinClassDef->getName() != "NoItinerary") 977 Name = std::string(ItinClassDef->getName()); 978 for (unsigned Idx : OperWrites) { 979 if (!Name.empty()) 980 Name += '_'; 981 Name += SchedWrites[Idx].Name; 982 } 983 for (unsigned Idx : OperReads) { 984 Name += '_'; 985 Name += SchedReads[Idx].Name; 986 } 987 return Name; 988 } 989 990 std::string CodeGenSchedModels::createSchedClassName(const RecVec &InstDefs) { 991 992 std::string Name; 993 for (RecIter I = InstDefs.begin(), E = InstDefs.end(); I != E; ++I) { 994 if (I != InstDefs.begin()) 995 Name += '_'; 996 Name += (*I)->getName(); 997 } 998 return Name; 999 } 1000 1001 /// Add an inferred sched class from an itinerary class and per-operand list of 1002 /// SchedWrites and SchedReads. ProcIndices contains the set of IDs of 1003 /// processors that may utilize this class. 1004 unsigned CodeGenSchedModels::addSchedClass(Record *ItinClassDef, 1005 ArrayRef<unsigned> OperWrites, 1006 ArrayRef<unsigned> OperReads, 1007 ArrayRef<unsigned> ProcIndices) { 1008 assert(!ProcIndices.empty() && "expect at least one ProcIdx"); 1009 1010 auto IsKeyEqual = [=](const CodeGenSchedClass &SC) { 1011 return SC.isKeyEqual(ItinClassDef, OperWrites, OperReads); 1012 }; 1013 1014 auto I = find_if(make_range(schedClassBegin(), schedClassEnd()), IsKeyEqual); 1015 unsigned Idx = I == schedClassEnd() ? 0 : std::distance(schedClassBegin(), I); 1016 if (Idx || SchedClasses[0].isKeyEqual(ItinClassDef, OperWrites, OperReads)) { 1017 IdxVec PI; 1018 std::set_union(SchedClasses[Idx].ProcIndices.begin(), 1019 SchedClasses[Idx].ProcIndices.end(), 1020 ProcIndices.begin(), ProcIndices.end(), 1021 std::back_inserter(PI)); 1022 SchedClasses[Idx].ProcIndices = std::move(PI); 1023 return Idx; 1024 } 1025 Idx = SchedClasses.size(); 1026 SchedClasses.emplace_back(Idx, 1027 createSchedClassName(ItinClassDef, OperWrites, 1028 OperReads), 1029 ItinClassDef); 1030 CodeGenSchedClass &SC = SchedClasses.back(); 1031 SC.Writes = OperWrites; 1032 SC.Reads = OperReads; 1033 SC.ProcIndices = ProcIndices; 1034 1035 return Idx; 1036 } 1037 1038 // Create classes for each set of opcodes that are in the same InstReadWrite 1039 // definition across all processors. 1040 void CodeGenSchedModels::createInstRWClass(Record *InstRWDef) { 1041 // ClassInstrs will hold an entry for each subset of Instrs in InstRWDef that 1042 // intersects with an existing class via a previous InstRWDef. Instrs that do 1043 // not intersect with an existing class refer back to their former class as 1044 // determined from ItinDef or SchedRW. 1045 SmallMapVector<unsigned, SmallVector<Record *, 8>, 4> ClassInstrs; 1046 // Sort Instrs into sets. 1047 const RecVec *InstDefs = Sets.expand(InstRWDef); 1048 if (InstDefs->empty()) 1049 PrintFatalError(InstRWDef->getLoc(), "No matching instruction opcodes"); 1050 1051 for (Record *InstDef : *InstDefs) { 1052 InstClassMapTy::const_iterator Pos = InstrClassMap.find(InstDef); 1053 if (Pos == InstrClassMap.end()) 1054 PrintFatalError(InstDef->getLoc(), "No sched class for instruction."); 1055 unsigned SCIdx = Pos->second; 1056 ClassInstrs[SCIdx].push_back(InstDef); 1057 } 1058 // For each set of Instrs, create a new class if necessary, and map or remap 1059 // the Instrs to it. 1060 for (auto &Entry : ClassInstrs) { 1061 unsigned OldSCIdx = Entry.first; 1062 ArrayRef<Record*> InstDefs = Entry.second; 1063 // If the all instrs in the current class are accounted for, then leave 1064 // them mapped to their old class. 1065 if (OldSCIdx) { 1066 const RecVec &RWDefs = SchedClasses[OldSCIdx].InstRWs; 1067 if (!RWDefs.empty()) { 1068 const RecVec *OrigInstDefs = Sets.expand(RWDefs[0]); 1069 unsigned OrigNumInstrs = 1070 count_if(*OrigInstDefs, [&](Record *OIDef) { 1071 return InstrClassMap[OIDef] == OldSCIdx; 1072 }); 1073 if (OrigNumInstrs == InstDefs.size()) { 1074 assert(SchedClasses[OldSCIdx].ProcIndices[0] == 0 && 1075 "expected a generic SchedClass"); 1076 Record *RWModelDef = InstRWDef->getValueAsDef("SchedModel"); 1077 // Make sure we didn't already have a InstRW containing this 1078 // instruction on this model. 1079 for (Record *RWD : RWDefs) { 1080 if (RWD->getValueAsDef("SchedModel") == RWModelDef && 1081 RWModelDef->getValueAsBit("FullInstRWOverlapCheck")) { 1082 assert(!InstDefs.empty()); // Checked at function start. 1083 PrintError( 1084 InstRWDef->getLoc(), 1085 "Overlapping InstRW definition for \"" + 1086 InstDefs.front()->getName() + 1087 "\" also matches previous \"" + 1088 RWD->getValue("Instrs")->getValue()->getAsString() + 1089 "\"."); 1090 PrintFatalNote(RWD->getLoc(), "Previous match was here."); 1091 } 1092 } 1093 LLVM_DEBUG(dbgs() << "InstRW: Reuse SC " << OldSCIdx << ":" 1094 << SchedClasses[OldSCIdx].Name << " on " 1095 << RWModelDef->getName() << "\n"); 1096 SchedClasses[OldSCIdx].InstRWs.push_back(InstRWDef); 1097 continue; 1098 } 1099 } 1100 } 1101 unsigned SCIdx = SchedClasses.size(); 1102 SchedClasses.emplace_back(SCIdx, createSchedClassName(InstDefs), nullptr); 1103 CodeGenSchedClass &SC = SchedClasses.back(); 1104 LLVM_DEBUG(dbgs() << "InstRW: New SC " << SCIdx << ":" << SC.Name << " on " 1105 << InstRWDef->getValueAsDef("SchedModel")->getName() 1106 << "\n"); 1107 1108 // Preserve ItinDef and Writes/Reads for processors without an InstRW entry. 1109 SC.ItinClassDef = SchedClasses[OldSCIdx].ItinClassDef; 1110 SC.Writes = SchedClasses[OldSCIdx].Writes; 1111 SC.Reads = SchedClasses[OldSCIdx].Reads; 1112 SC.ProcIndices.push_back(0); 1113 // If we had an old class, copy it's InstRWs to this new class. 1114 if (OldSCIdx) { 1115 Record *RWModelDef = InstRWDef->getValueAsDef("SchedModel"); 1116 for (Record *OldRWDef : SchedClasses[OldSCIdx].InstRWs) { 1117 if (OldRWDef->getValueAsDef("SchedModel") == RWModelDef) { 1118 assert(!InstDefs.empty()); // Checked at function start. 1119 PrintError( 1120 InstRWDef->getLoc(), 1121 "Overlapping InstRW definition for \"" + 1122 InstDefs.front()->getName() + "\" also matches previous \"" + 1123 OldRWDef->getValue("Instrs")->getValue()->getAsString() + 1124 "\"."); 1125 PrintFatalNote(OldRWDef->getLoc(), "Previous match was here."); 1126 } 1127 assert(OldRWDef != InstRWDef && 1128 "SchedClass has duplicate InstRW def"); 1129 SC.InstRWs.push_back(OldRWDef); 1130 } 1131 } 1132 // Map each Instr to this new class. 1133 for (Record *InstDef : InstDefs) 1134 InstrClassMap[InstDef] = SCIdx; 1135 SC.InstRWs.push_back(InstRWDef); 1136 } 1137 } 1138 1139 // True if collectProcItins found anything. 1140 bool CodeGenSchedModels::hasItineraries() const { 1141 for (const CodeGenProcModel &PM : make_range(procModelBegin(),procModelEnd())) 1142 if (PM.hasItineraries()) 1143 return true; 1144 return false; 1145 } 1146 1147 // Gather the processor itineraries. 1148 void CodeGenSchedModels::collectProcItins() { 1149 LLVM_DEBUG(dbgs() << "\n+++ PROBLEM ITINERARIES (collectProcItins) +++\n"); 1150 for (CodeGenProcModel &ProcModel : ProcModels) { 1151 if (!ProcModel.hasItineraries()) 1152 continue; 1153 1154 RecVec ItinRecords = ProcModel.ItinsDef->getValueAsListOfDefs("IID"); 1155 assert(!ItinRecords.empty() && "ProcModel.hasItineraries is incorrect"); 1156 1157 // Populate ItinDefList with Itinerary records. 1158 ProcModel.ItinDefList.resize(NumInstrSchedClasses); 1159 1160 // Insert each itinerary data record in the correct position within 1161 // the processor model's ItinDefList. 1162 for (Record *ItinData : ItinRecords) { 1163 const Record *ItinDef = ItinData->getValueAsDef("TheClass"); 1164 bool FoundClass = false; 1165 1166 for (const CodeGenSchedClass &SC : 1167 make_range(schedClassBegin(), schedClassEnd())) { 1168 // Multiple SchedClasses may share an itinerary. Update all of them. 1169 if (SC.ItinClassDef == ItinDef) { 1170 ProcModel.ItinDefList[SC.Index] = ItinData; 1171 FoundClass = true; 1172 } 1173 } 1174 if (!FoundClass) { 1175 LLVM_DEBUG(dbgs() << ProcModel.ItinsDef->getName() 1176 << " missing class for itinerary " 1177 << ItinDef->getName() << '\n'); 1178 } 1179 } 1180 // Check for missing itinerary entries. 1181 assert(!ProcModel.ItinDefList[0] && "NoItinerary class can't have rec"); 1182 LLVM_DEBUG( 1183 for (unsigned i = 1, N = ProcModel.ItinDefList.size(); i < N; ++i) { 1184 if (!ProcModel.ItinDefList[i]) 1185 dbgs() << ProcModel.ItinsDef->getName() 1186 << " missing itinerary for class " << SchedClasses[i].Name 1187 << '\n'; 1188 }); 1189 } 1190 } 1191 1192 // Gather the read/write types for each itinerary class. 1193 void CodeGenSchedModels::collectProcItinRW() { 1194 RecVec ItinRWDefs = Records.getAllDerivedDefinitions("ItinRW"); 1195 llvm::sort(ItinRWDefs, LessRecord()); 1196 for (Record *RWDef : ItinRWDefs) { 1197 if (!RWDef->getValueInit("SchedModel")->isComplete()) 1198 PrintFatalError(RWDef->getLoc(), "SchedModel is undefined"); 1199 Record *ModelDef = RWDef->getValueAsDef("SchedModel"); 1200 ProcModelMapTy::const_iterator I = ProcModelMap.find(ModelDef); 1201 if (I == ProcModelMap.end()) { 1202 PrintFatalError(RWDef->getLoc(), "Undefined SchedMachineModel " 1203 + ModelDef->getName()); 1204 } 1205 ProcModels[I->second].ItinRWDefs.push_back(RWDef); 1206 } 1207 } 1208 1209 // Gather the unsupported features for processor models. 1210 void CodeGenSchedModels::collectProcUnsupportedFeatures() { 1211 for (CodeGenProcModel &ProcModel : ProcModels) { 1212 for (Record *Pred : ProcModel.ModelDef->getValueAsListOfDefs("UnsupportedFeatures")) { 1213 ProcModel.UnsupportedFeaturesDefs.push_back(Pred); 1214 } 1215 } 1216 } 1217 1218 /// Infer new classes from existing classes. In the process, this may create new 1219 /// SchedWrites from sequences of existing SchedWrites. 1220 void CodeGenSchedModels::inferSchedClasses() { 1221 LLVM_DEBUG( 1222 dbgs() << "\n+++ INFERRING SCHED CLASSES (inferSchedClasses) +++\n"); 1223 LLVM_DEBUG(dbgs() << NumInstrSchedClasses << " instr sched classes.\n"); 1224 1225 // Visit all existing classes and newly created classes. 1226 for (unsigned Idx = 0; Idx != SchedClasses.size(); ++Idx) { 1227 assert(SchedClasses[Idx].Index == Idx && "bad SCIdx"); 1228 1229 if (SchedClasses[Idx].ItinClassDef) 1230 inferFromItinClass(SchedClasses[Idx].ItinClassDef, Idx); 1231 if (!SchedClasses[Idx].InstRWs.empty()) 1232 inferFromInstRWs(Idx); 1233 if (!SchedClasses[Idx].Writes.empty()) { 1234 inferFromRW(SchedClasses[Idx].Writes, SchedClasses[Idx].Reads, 1235 Idx, SchedClasses[Idx].ProcIndices); 1236 } 1237 assert(SchedClasses.size() < (NumInstrSchedClasses*6) && 1238 "too many SchedVariants"); 1239 } 1240 } 1241 1242 /// Infer classes from per-processor itinerary resources. 1243 void CodeGenSchedModels::inferFromItinClass(Record *ItinClassDef, 1244 unsigned FromClassIdx) { 1245 for (unsigned PIdx = 0, PEnd = ProcModels.size(); PIdx != PEnd; ++PIdx) { 1246 const CodeGenProcModel &PM = ProcModels[PIdx]; 1247 // For all ItinRW entries. 1248 bool HasMatch = false; 1249 for (const Record *Rec : PM.ItinRWDefs) { 1250 RecVec Matched = Rec->getValueAsListOfDefs("MatchedItinClasses"); 1251 if (!std::count(Matched.begin(), Matched.end(), ItinClassDef)) 1252 continue; 1253 if (HasMatch) 1254 PrintFatalError(Rec->getLoc(), "Duplicate itinerary class " 1255 + ItinClassDef->getName() 1256 + " in ItinResources for " + PM.ModelName); 1257 HasMatch = true; 1258 IdxVec Writes, Reads; 1259 findRWs(Rec->getValueAsListOfDefs("OperandReadWrites"), Writes, Reads); 1260 inferFromRW(Writes, Reads, FromClassIdx, PIdx); 1261 } 1262 } 1263 } 1264 1265 /// Infer classes from per-processor InstReadWrite definitions. 1266 void CodeGenSchedModels::inferFromInstRWs(unsigned SCIdx) { 1267 for (unsigned I = 0, E = SchedClasses[SCIdx].InstRWs.size(); I != E; ++I) { 1268 assert(SchedClasses[SCIdx].InstRWs.size() == E && "InstrRWs was mutated!"); 1269 Record *Rec = SchedClasses[SCIdx].InstRWs[I]; 1270 const RecVec *InstDefs = Sets.expand(Rec); 1271 RecIter II = InstDefs->begin(), IE = InstDefs->end(); 1272 for (; II != IE; ++II) { 1273 if (InstrClassMap[*II] == SCIdx) 1274 break; 1275 } 1276 // If this class no longer has any instructions mapped to it, it has become 1277 // irrelevant. 1278 if (II == IE) 1279 continue; 1280 IdxVec Writes, Reads; 1281 findRWs(Rec->getValueAsListOfDefs("OperandReadWrites"), Writes, Reads); 1282 unsigned PIdx = getProcModel(Rec->getValueAsDef("SchedModel")).Index; 1283 inferFromRW(Writes, Reads, SCIdx, PIdx); // May mutate SchedClasses. 1284 SchedClasses[SCIdx].InstRWProcIndices.insert(PIdx); 1285 } 1286 } 1287 1288 namespace { 1289 1290 // Helper for substituteVariantOperand. 1291 struct TransVariant { 1292 Record *VarOrSeqDef; // Variant or sequence. 1293 unsigned RWIdx; // Index of this variant or sequence's matched type. 1294 unsigned ProcIdx; // Processor model index or zero for any. 1295 unsigned TransVecIdx; // Index into PredTransitions::TransVec. 1296 1297 TransVariant(Record *def, unsigned rwi, unsigned pi, unsigned ti): 1298 VarOrSeqDef(def), RWIdx(rwi), ProcIdx(pi), TransVecIdx(ti) {} 1299 }; 1300 1301 // Associate a predicate with the SchedReadWrite that it guards. 1302 // RWIdx is the index of the read/write variant. 1303 struct PredCheck { 1304 bool IsRead; 1305 unsigned RWIdx; 1306 Record *Predicate; 1307 1308 PredCheck(bool r, unsigned w, Record *p): IsRead(r), RWIdx(w), Predicate(p) {} 1309 }; 1310 1311 // A Predicate transition is a list of RW sequences guarded by a PredTerm. 1312 struct PredTransition { 1313 // A predicate term is a conjunction of PredChecks. 1314 SmallVector<PredCheck, 4> PredTerm; 1315 SmallVector<SmallVector<unsigned,4>, 16> WriteSequences; 1316 SmallVector<SmallVector<unsigned,4>, 16> ReadSequences; 1317 SmallVector<unsigned, 4> ProcIndices; 1318 1319 PredTransition() = default; 1320 PredTransition(ArrayRef<PredCheck> PT) { 1321 PredTerm.assign(PT.begin(), PT.end()); 1322 ProcIndices.assign(1, 0); 1323 } 1324 PredTransition(ArrayRef<PredCheck> PT, ArrayRef<unsigned> PIds) { 1325 PredTerm.assign(PT.begin(), PT.end()); 1326 ProcIndices.assign(PIds.begin(), PIds.end()); 1327 } 1328 }; 1329 1330 // Encapsulate a set of partially constructed transitions. 1331 // The results are built by repeated calls to substituteVariants. 1332 class PredTransitions { 1333 CodeGenSchedModels &SchedModels; 1334 1335 public: 1336 std::vector<PredTransition> TransVec; 1337 1338 PredTransitions(CodeGenSchedModels &sm): SchedModels(sm) {} 1339 1340 void substituteVariantOperand(const SmallVectorImpl<unsigned> &RWSeq, 1341 bool IsRead, bool IsForAnyCPU, 1342 unsigned StartIdx); 1343 1344 void substituteVariants(const PredTransition &Trans); 1345 1346 #ifndef NDEBUG 1347 void dump() const; 1348 #endif 1349 1350 private: 1351 bool mutuallyExclusive(Record *PredDef, ArrayRef<PredCheck> Term); 1352 void getIntersectingVariants( 1353 const CodeGenSchedRW &SchedRW, unsigned TransIdx, 1354 std::vector<TransVariant> &IntersectingVariants); 1355 void pushVariant(const TransVariant &VInfo, bool IsRead); 1356 }; 1357 1358 } // end anonymous namespace 1359 1360 // Return true if this predicate is mutually exclusive with a PredTerm. This 1361 // degenerates into checking if the predicate is mutually exclusive with any 1362 // predicate in the Term's conjunction. 1363 // 1364 // All predicates associated with a given SchedRW are considered mutually 1365 // exclusive. This should work even if the conditions expressed by the 1366 // predicates are not exclusive because the predicates for a given SchedWrite 1367 // are always checked in the order they are defined in the .td file. Later 1368 // conditions implicitly negate any prior condition. 1369 bool PredTransitions::mutuallyExclusive(Record *PredDef, 1370 ArrayRef<PredCheck> Term) { 1371 for (const PredCheck &PC: Term) { 1372 if (PC.Predicate == PredDef) 1373 return false; 1374 1375 const CodeGenSchedRW &SchedRW = SchedModels.getSchedRW(PC.RWIdx, PC.IsRead); 1376 assert(SchedRW.HasVariants && "PredCheck must refer to a SchedVariant"); 1377 RecVec Variants = SchedRW.TheDef->getValueAsListOfDefs("Variants"); 1378 if (any_of(Variants, [PredDef](const Record *R) { 1379 return R->getValueAsDef("Predicate") == PredDef; 1380 })) 1381 return true; 1382 } 1383 return false; 1384 } 1385 1386 static bool hasAliasedVariants(const CodeGenSchedRW &RW, 1387 CodeGenSchedModels &SchedModels) { 1388 if (RW.HasVariants) 1389 return true; 1390 1391 for (Record *Alias : RW.Aliases) { 1392 const CodeGenSchedRW &AliasRW = 1393 SchedModels.getSchedRW(Alias->getValueAsDef("AliasRW")); 1394 if (AliasRW.HasVariants) 1395 return true; 1396 if (AliasRW.IsSequence) { 1397 IdxVec ExpandedRWs; 1398 SchedModels.expandRWSequence(AliasRW.Index, ExpandedRWs, AliasRW.IsRead); 1399 for (unsigned SI : ExpandedRWs) { 1400 if (hasAliasedVariants(SchedModels.getSchedRW(SI, AliasRW.IsRead), 1401 SchedModels)) 1402 return true; 1403 } 1404 } 1405 } 1406 return false; 1407 } 1408 1409 static bool hasVariant(ArrayRef<PredTransition> Transitions, 1410 CodeGenSchedModels &SchedModels) { 1411 for (const PredTransition &PTI : Transitions) { 1412 for (const SmallVectorImpl<unsigned> &WSI : PTI.WriteSequences) 1413 for (unsigned WI : WSI) 1414 if (hasAliasedVariants(SchedModels.getSchedWrite(WI), SchedModels)) 1415 return true; 1416 1417 for (const SmallVectorImpl<unsigned> &RSI : PTI.ReadSequences) 1418 for (unsigned RI : RSI) 1419 if (hasAliasedVariants(SchedModels.getSchedRead(RI), SchedModels)) 1420 return true; 1421 } 1422 return false; 1423 } 1424 1425 // Populate IntersectingVariants with any variants or aliased sequences of the 1426 // given SchedRW whose processor indices and predicates are not mutually 1427 // exclusive with the given transition. 1428 void PredTransitions::getIntersectingVariants( 1429 const CodeGenSchedRW &SchedRW, unsigned TransIdx, 1430 std::vector<TransVariant> &IntersectingVariants) { 1431 1432 bool GenericRW = false; 1433 1434 std::vector<TransVariant> Variants; 1435 if (SchedRW.HasVariants) { 1436 unsigned VarProcIdx = 0; 1437 if (SchedRW.TheDef->getValueInit("SchedModel")->isComplete()) { 1438 Record *ModelDef = SchedRW.TheDef->getValueAsDef("SchedModel"); 1439 VarProcIdx = SchedModels.getProcModel(ModelDef).Index; 1440 } 1441 // Push each variant. Assign TransVecIdx later. 1442 const RecVec VarDefs = SchedRW.TheDef->getValueAsListOfDefs("Variants"); 1443 for (Record *VarDef : VarDefs) 1444 Variants.emplace_back(VarDef, SchedRW.Index, VarProcIdx, 0); 1445 if (VarProcIdx == 0) 1446 GenericRW = true; 1447 } 1448 for (RecIter AI = SchedRW.Aliases.begin(), AE = SchedRW.Aliases.end(); 1449 AI != AE; ++AI) { 1450 // If either the SchedAlias itself or the SchedReadWrite that it aliases 1451 // to is defined within a processor model, constrain all variants to 1452 // that processor. 1453 unsigned AliasProcIdx = 0; 1454 if ((*AI)->getValueInit("SchedModel")->isComplete()) { 1455 Record *ModelDef = (*AI)->getValueAsDef("SchedModel"); 1456 AliasProcIdx = SchedModels.getProcModel(ModelDef).Index; 1457 } 1458 const CodeGenSchedRW &AliasRW = 1459 SchedModels.getSchedRW((*AI)->getValueAsDef("AliasRW")); 1460 1461 if (AliasRW.HasVariants) { 1462 const RecVec VarDefs = AliasRW.TheDef->getValueAsListOfDefs("Variants"); 1463 for (Record *VD : VarDefs) 1464 Variants.emplace_back(VD, AliasRW.Index, AliasProcIdx, 0); 1465 } 1466 if (AliasRW.IsSequence) 1467 Variants.emplace_back(AliasRW.TheDef, SchedRW.Index, AliasProcIdx, 0); 1468 if (AliasProcIdx == 0) 1469 GenericRW = true; 1470 } 1471 for (TransVariant &Variant : Variants) { 1472 // Don't expand variants if the processor models don't intersect. 1473 // A zero processor index means any processor. 1474 SmallVectorImpl<unsigned> &ProcIndices = TransVec[TransIdx].ProcIndices; 1475 if (ProcIndices[0] && Variant.ProcIdx) { 1476 unsigned Cnt = std::count(ProcIndices.begin(), ProcIndices.end(), 1477 Variant.ProcIdx); 1478 if (!Cnt) 1479 continue; 1480 if (Cnt > 1) { 1481 const CodeGenProcModel &PM = 1482 *(SchedModels.procModelBegin() + Variant.ProcIdx); 1483 PrintFatalError(Variant.VarOrSeqDef->getLoc(), 1484 "Multiple variants defined for processor " + 1485 PM.ModelName + 1486 " Ensure only one SchedAlias exists per RW."); 1487 } 1488 } 1489 if (Variant.VarOrSeqDef->isSubClassOf("SchedVar")) { 1490 Record *PredDef = Variant.VarOrSeqDef->getValueAsDef("Predicate"); 1491 if (mutuallyExclusive(PredDef, TransVec[TransIdx].PredTerm)) 1492 continue; 1493 } 1494 if (IntersectingVariants.empty()) { 1495 // The first variant builds on the existing transition. 1496 Variant.TransVecIdx = TransIdx; 1497 IntersectingVariants.push_back(Variant); 1498 } 1499 else { 1500 // Push another copy of the current transition for more variants. 1501 Variant.TransVecIdx = TransVec.size(); 1502 IntersectingVariants.push_back(Variant); 1503 TransVec.push_back(TransVec[TransIdx]); 1504 } 1505 } 1506 if (GenericRW && IntersectingVariants.empty()) { 1507 PrintFatalError(SchedRW.TheDef->getLoc(), "No variant of this type has " 1508 "a matching predicate on any processor"); 1509 } 1510 } 1511 1512 // Push the Reads/Writes selected by this variant onto the PredTransition 1513 // specified by VInfo. 1514 void PredTransitions:: 1515 pushVariant(const TransVariant &VInfo, bool IsRead) { 1516 PredTransition &Trans = TransVec[VInfo.TransVecIdx]; 1517 1518 // If this operand transition is reached through a processor-specific alias, 1519 // then the whole transition is specific to this processor. 1520 if (VInfo.ProcIdx != 0) 1521 Trans.ProcIndices.assign(1, VInfo.ProcIdx); 1522 1523 IdxVec SelectedRWs; 1524 if (VInfo.VarOrSeqDef->isSubClassOf("SchedVar")) { 1525 Record *PredDef = VInfo.VarOrSeqDef->getValueAsDef("Predicate"); 1526 Trans.PredTerm.emplace_back(IsRead, VInfo.RWIdx,PredDef); 1527 RecVec SelectedDefs = VInfo.VarOrSeqDef->getValueAsListOfDefs("Selected"); 1528 SchedModels.findRWs(SelectedDefs, SelectedRWs, IsRead); 1529 } 1530 else { 1531 assert(VInfo.VarOrSeqDef->isSubClassOf("WriteSequence") && 1532 "variant must be a SchedVariant or aliased WriteSequence"); 1533 SelectedRWs.push_back(SchedModels.getSchedRWIdx(VInfo.VarOrSeqDef, IsRead)); 1534 } 1535 1536 const CodeGenSchedRW &SchedRW = SchedModels.getSchedRW(VInfo.RWIdx, IsRead); 1537 1538 SmallVectorImpl<SmallVector<unsigned,4>> &RWSequences = IsRead 1539 ? Trans.ReadSequences : Trans.WriteSequences; 1540 if (SchedRW.IsVariadic) { 1541 unsigned OperIdx = RWSequences.size()-1; 1542 // Make N-1 copies of this transition's last sequence. 1543 RWSequences.insert(RWSequences.end(), SelectedRWs.size() - 1, 1544 RWSequences[OperIdx]); 1545 // Push each of the N elements of the SelectedRWs onto a copy of the last 1546 // sequence (split the current operand into N operands). 1547 // Note that write sequences should be expanded within this loop--the entire 1548 // sequence belongs to a single operand. 1549 for (IdxIter RWI = SelectedRWs.begin(), RWE = SelectedRWs.end(); 1550 RWI != RWE; ++RWI, ++OperIdx) { 1551 IdxVec ExpandedRWs; 1552 if (IsRead) 1553 ExpandedRWs.push_back(*RWI); 1554 else 1555 SchedModels.expandRWSequence(*RWI, ExpandedRWs, IsRead); 1556 RWSequences[OperIdx].insert(RWSequences[OperIdx].end(), 1557 ExpandedRWs.begin(), ExpandedRWs.end()); 1558 } 1559 assert(OperIdx == RWSequences.size() && "missed a sequence"); 1560 } 1561 else { 1562 // Push this transition's expanded sequence onto this transition's last 1563 // sequence (add to the current operand's sequence). 1564 SmallVectorImpl<unsigned> &Seq = RWSequences.back(); 1565 IdxVec ExpandedRWs; 1566 for (IdxIter RWI = SelectedRWs.begin(), RWE = SelectedRWs.end(); 1567 RWI != RWE; ++RWI) { 1568 if (IsRead) 1569 ExpandedRWs.push_back(*RWI); 1570 else 1571 SchedModels.expandRWSequence(*RWI, ExpandedRWs, IsRead); 1572 } 1573 Seq.insert(Seq.end(), ExpandedRWs.begin(), ExpandedRWs.end()); 1574 } 1575 } 1576 1577 // RWSeq is a sequence of all Reads or all Writes for the next read or write 1578 // operand. StartIdx is an index into TransVec where partial results 1579 // starts. RWSeq must be applied to all transitions between StartIdx and the end 1580 // of TransVec. 1581 void PredTransitions::substituteVariantOperand( 1582 const SmallVectorImpl<unsigned> &RWSeq, bool IsRead, bool IsForAnyCPU, 1583 unsigned StartIdx) { 1584 1585 auto CollectAndAddVariants = [&](unsigned TransIdx, 1586 const CodeGenSchedRW &SchedRW) { 1587 // Distribute this partial PredTransition across intersecting variants. 1588 // This will push a copies of TransVec[TransIdx] on the back of TransVec. 1589 std::vector<TransVariant> IntersectingVariants; 1590 getIntersectingVariants(SchedRW, TransIdx, IntersectingVariants); 1591 // Now expand each variant on top of its copy of the transition. 1592 for (const TransVariant &IV : IntersectingVariants) 1593 pushVariant(IV, IsRead); 1594 return !IntersectingVariants.empty(); 1595 }; 1596 1597 // Visit each original RW within the current sequence. 1598 for (SmallVectorImpl<unsigned>::const_iterator 1599 RWI = RWSeq.begin(), RWE = RWSeq.end(); RWI != RWE; ++RWI) { 1600 const CodeGenSchedRW &SchedRW = SchedModels.getSchedRW(*RWI, IsRead); 1601 // Push this RW on all partial PredTransitions or distribute variants. 1602 // New PredTransitions may be pushed within this loop which should not be 1603 // revisited (TransEnd must be loop invariant). 1604 bool HasAliases = false, WasPushed = false; 1605 for (unsigned TransIdx = StartIdx, TransEnd = TransVec.size(); 1606 TransIdx != TransEnd; ++TransIdx) { 1607 // In the common case, push RW onto the current operand's sequence. 1608 if (!hasAliasedVariants(SchedRW, SchedModels)) { 1609 if (IsRead) 1610 TransVec[TransIdx].ReadSequences.back().push_back(*RWI); 1611 else 1612 TransVec[TransIdx].WriteSequences.back().push_back(*RWI); 1613 continue; 1614 } 1615 HasAliases = true; 1616 WasPushed |= CollectAndAddVariants(TransIdx, SchedRW); 1617 } 1618 if (IsRead && IsForAnyCPU && HasAliases && !WasPushed) { 1619 // If we're here this means that in some sched class: 1620 // a) We have read variant for CPU A 1621 // b) We have write variant for CPU B 1622 // b) We don't have write variant for CPU A 1623 // d) We must expand all read/write variants (IsForAnyCPU is true) 1624 // e) We couldn't expand SchedRW because TransVec doesn't have 1625 // any transition with compatible CPU ID. 1626 // In such case we create new empty transition with zero (AnyCPU) 1627 // index. 1628 TransVec.emplace_back(TransVec[StartIdx].PredTerm); 1629 TransVec.back().ReadSequences.emplace_back(); 1630 CollectAndAddVariants(TransVec.size() - 1, SchedRW); 1631 } 1632 } 1633 } 1634 1635 // For each variant of a Read/Write in Trans, substitute the sequence of 1636 // Read/Writes guarded by the variant. This is exponential in the number of 1637 // variant Read/Writes, but in practice detection of mutually exclusive 1638 // predicates should result in linear growth in the total number variants. 1639 // 1640 // This is one step in a breadth-first search of nested variants. 1641 void PredTransitions::substituteVariants(const PredTransition &Trans) { 1642 // Build up a set of partial results starting at the back of 1643 // PredTransitions. Remember the first new transition. 1644 unsigned StartIdx = TransVec.size(); 1645 TransVec.emplace_back(Trans.PredTerm, Trans.ProcIndices); 1646 1647 bool IsForAnyCPU = llvm::count(Trans.ProcIndices, 0); 1648 // Visit each original write sequence. 1649 for (SmallVectorImpl<SmallVector<unsigned,4>>::const_iterator 1650 WSI = Trans.WriteSequences.begin(), WSE = Trans.WriteSequences.end(); 1651 WSI != WSE; ++WSI) { 1652 // Push a new (empty) write sequence onto all partial Transitions. 1653 for (std::vector<PredTransition>::iterator I = 1654 TransVec.begin() + StartIdx, E = TransVec.end(); I != E; ++I) { 1655 I->WriteSequences.emplace_back(); 1656 } 1657 substituteVariantOperand(*WSI, /*IsRead=*/false, IsForAnyCPU, StartIdx); 1658 } 1659 // Visit each original read sequence. 1660 for (SmallVectorImpl<SmallVector<unsigned,4>>::const_iterator 1661 RSI = Trans.ReadSequences.begin(), RSE = Trans.ReadSequences.end(); 1662 RSI != RSE; ++RSI) { 1663 // Push a new (empty) read sequence onto all partial Transitions. 1664 for (std::vector<PredTransition>::iterator I = 1665 TransVec.begin() + StartIdx, E = TransVec.end(); I != E; ++I) { 1666 I->ReadSequences.emplace_back(); 1667 } 1668 substituteVariantOperand(*RSI, /*IsRead=*/true, IsForAnyCPU, StartIdx); 1669 } 1670 } 1671 1672 static void addSequences(CodeGenSchedModels &SchedModels, 1673 const SmallVectorImpl<SmallVector<unsigned, 4>> &Seqs, 1674 IdxVec &Result, bool IsRead) { 1675 for (const auto &S : Seqs) 1676 if (!S.empty()) 1677 Result.push_back(SchedModels.findOrInsertRW(S, IsRead)); 1678 } 1679 1680 static void dumpTransition(const CodeGenSchedModels &SchedModels, 1681 const CodeGenSchedClass &FromSC, 1682 const CodeGenSchedTransition &SCTrans) { 1683 LLVM_DEBUG(dbgs() << "Adding transition from " << FromSC.Name << "(" 1684 << FromSC.Index << ") to " 1685 << SchedModels.getSchedClass(SCTrans.ToClassIdx).Name << "(" 1686 << SCTrans.ToClassIdx << ")" 1687 << " on processor indices: ("; 1688 dumpIdxVec(SCTrans.ProcIndices); dbgs() << ")\n"); 1689 } 1690 // Create a new SchedClass for each variant found by inferFromRW. Pass 1691 static void inferFromTransitions(ArrayRef<PredTransition> LastTransitions, 1692 unsigned FromClassIdx, 1693 CodeGenSchedModels &SchedModels) { 1694 // For each PredTransition, create a new CodeGenSchedTransition, which usually 1695 // requires creating a new SchedClass. 1696 for (ArrayRef<PredTransition>::iterator 1697 I = LastTransitions.begin(), E = LastTransitions.end(); I != E; ++I) { 1698 IdxVec OperWritesVariant, OperReadsVariant; 1699 addSequences(SchedModels, I->WriteSequences, OperWritesVariant, false); 1700 addSequences(SchedModels, I->ReadSequences, OperReadsVariant, true); 1701 CodeGenSchedTransition SCTrans; 1702 1703 // Transition should not contain processor indices already assigned to 1704 // InstRWs in this scheduling class. 1705 const CodeGenSchedClass &FromSC = SchedModels.getSchedClass(FromClassIdx); 1706 llvm::copy_if(I->ProcIndices, std::back_inserter(SCTrans.ProcIndices), 1707 [&FromSC](unsigned PIdx) { 1708 return !FromSC.InstRWProcIndices.count(PIdx); 1709 }); 1710 if (SCTrans.ProcIndices.empty()) 1711 continue; 1712 SCTrans.ToClassIdx = 1713 SchedModels.addSchedClass(/*ItinClassDef=*/nullptr, OperWritesVariant, 1714 OperReadsVariant, I->ProcIndices); 1715 dumpTransition(SchedModels, FromSC, SCTrans); 1716 // The final PredTerm is unique set of predicates guarding the transition. 1717 RecVec Preds; 1718 transform(I->PredTerm, std::back_inserter(Preds), 1719 [](const PredCheck &P) { 1720 return P.Predicate; 1721 }); 1722 Preds.erase(std::unique(Preds.begin(), Preds.end()), Preds.end()); 1723 SCTrans.PredTerm = std::move(Preds); 1724 SchedModels.getSchedClass(FromClassIdx) 1725 .Transitions.push_back(std::move(SCTrans)); 1726 } 1727 } 1728 1729 // Create new SchedClasses for the given ReadWrite list. If any of the 1730 // ReadWrites refers to a SchedVariant, create a new SchedClass for each variant 1731 // of the ReadWrite list, following Aliases if necessary. 1732 void CodeGenSchedModels::inferFromRW(ArrayRef<unsigned> OperWrites, 1733 ArrayRef<unsigned> OperReads, 1734 unsigned FromClassIdx, 1735 ArrayRef<unsigned> ProcIndices) { 1736 LLVM_DEBUG(dbgs() << "INFER RW proc("; dumpIdxVec(ProcIndices); 1737 dbgs() << ") "); 1738 // Create a seed transition with an empty PredTerm and the expanded sequences 1739 // of SchedWrites for the current SchedClass. 1740 std::vector<PredTransition> LastTransitions; 1741 LastTransitions.emplace_back(); 1742 LastTransitions.back().ProcIndices.append(ProcIndices.begin(), 1743 ProcIndices.end()); 1744 1745 for (unsigned WriteIdx : OperWrites) { 1746 IdxVec WriteSeq; 1747 expandRWSequence(WriteIdx, WriteSeq, /*IsRead=*/false); 1748 LastTransitions[0].WriteSequences.emplace_back(); 1749 SmallVectorImpl<unsigned> &Seq = LastTransitions[0].WriteSequences.back(); 1750 Seq.append(WriteSeq.begin(), WriteSeq.end()); 1751 LLVM_DEBUG(dbgs() << "("; dumpIdxVec(Seq); dbgs() << ") "); 1752 } 1753 LLVM_DEBUG(dbgs() << " Reads: "); 1754 for (unsigned ReadIdx : OperReads) { 1755 IdxVec ReadSeq; 1756 expandRWSequence(ReadIdx, ReadSeq, /*IsRead=*/true); 1757 LastTransitions[0].ReadSequences.emplace_back(); 1758 SmallVectorImpl<unsigned> &Seq = LastTransitions[0].ReadSequences.back(); 1759 Seq.append(ReadSeq.begin(), ReadSeq.end()); 1760 LLVM_DEBUG(dbgs() << "("; dumpIdxVec(Seq); dbgs() << ") "); 1761 } 1762 LLVM_DEBUG(dbgs() << '\n'); 1763 1764 // Collect all PredTransitions for individual operands. 1765 // Iterate until no variant writes remain. 1766 while (hasVariant(LastTransitions, *this)) { 1767 PredTransitions Transitions(*this); 1768 for (const PredTransition &Trans : LastTransitions) 1769 Transitions.substituteVariants(Trans); 1770 LLVM_DEBUG(Transitions.dump()); 1771 LastTransitions.swap(Transitions.TransVec); 1772 } 1773 // If the first transition has no variants, nothing to do. 1774 if (LastTransitions[0].PredTerm.empty()) 1775 return; 1776 1777 // WARNING: We are about to mutate the SchedClasses vector. Do not refer to 1778 // OperWrites, OperReads, or ProcIndices after calling inferFromTransitions. 1779 inferFromTransitions(LastTransitions, FromClassIdx, *this); 1780 } 1781 1782 // Check if any processor resource group contains all resource records in 1783 // SubUnits. 1784 bool CodeGenSchedModels::hasSuperGroup(RecVec &SubUnits, CodeGenProcModel &PM) { 1785 for (unsigned i = 0, e = PM.ProcResourceDefs.size(); i < e; ++i) { 1786 if (!PM.ProcResourceDefs[i]->isSubClassOf("ProcResGroup")) 1787 continue; 1788 RecVec SuperUnits = 1789 PM.ProcResourceDefs[i]->getValueAsListOfDefs("Resources"); 1790 RecIter RI = SubUnits.begin(), RE = SubUnits.end(); 1791 for ( ; RI != RE; ++RI) { 1792 if (!is_contained(SuperUnits, *RI)) { 1793 break; 1794 } 1795 } 1796 if (RI == RE) 1797 return true; 1798 } 1799 return false; 1800 } 1801 1802 // Verify that overlapping groups have a common supergroup. 1803 void CodeGenSchedModels::verifyProcResourceGroups(CodeGenProcModel &PM) { 1804 for (unsigned i = 0, e = PM.ProcResourceDefs.size(); i < e; ++i) { 1805 if (!PM.ProcResourceDefs[i]->isSubClassOf("ProcResGroup")) 1806 continue; 1807 RecVec CheckUnits = 1808 PM.ProcResourceDefs[i]->getValueAsListOfDefs("Resources"); 1809 for (unsigned j = i+1; j < e; ++j) { 1810 if (!PM.ProcResourceDefs[j]->isSubClassOf("ProcResGroup")) 1811 continue; 1812 RecVec OtherUnits = 1813 PM.ProcResourceDefs[j]->getValueAsListOfDefs("Resources"); 1814 if (std::find_first_of(CheckUnits.begin(), CheckUnits.end(), 1815 OtherUnits.begin(), OtherUnits.end()) 1816 != CheckUnits.end()) { 1817 // CheckUnits and OtherUnits overlap 1818 OtherUnits.insert(OtherUnits.end(), CheckUnits.begin(), 1819 CheckUnits.end()); 1820 if (!hasSuperGroup(OtherUnits, PM)) { 1821 PrintFatalError((PM.ProcResourceDefs[i])->getLoc(), 1822 "proc resource group overlaps with " 1823 + PM.ProcResourceDefs[j]->getName() 1824 + " but no supergroup contains both."); 1825 } 1826 } 1827 } 1828 } 1829 } 1830 1831 // Collect all the RegisterFile definitions available in this target. 1832 void CodeGenSchedModels::collectRegisterFiles() { 1833 RecVec RegisterFileDefs = Records.getAllDerivedDefinitions("RegisterFile"); 1834 1835 // RegisterFiles is the vector of CodeGenRegisterFile. 1836 for (Record *RF : RegisterFileDefs) { 1837 // For each register file definition, construct a CodeGenRegisterFile object 1838 // and add it to the appropriate scheduling model. 1839 CodeGenProcModel &PM = getProcModel(RF->getValueAsDef("SchedModel")); 1840 PM.RegisterFiles.emplace_back(CodeGenRegisterFile(RF->getName(),RF)); 1841 CodeGenRegisterFile &CGRF = PM.RegisterFiles.back(); 1842 CGRF.MaxMovesEliminatedPerCycle = 1843 RF->getValueAsInt("MaxMovesEliminatedPerCycle"); 1844 CGRF.AllowZeroMoveEliminationOnly = 1845 RF->getValueAsBit("AllowZeroMoveEliminationOnly"); 1846 1847 // Now set the number of physical registers as well as the cost of registers 1848 // in each register class. 1849 CGRF.NumPhysRegs = RF->getValueAsInt("NumPhysRegs"); 1850 if (!CGRF.NumPhysRegs) { 1851 PrintFatalError(RF->getLoc(), 1852 "Invalid RegisterFile with zero physical registers"); 1853 } 1854 1855 RecVec RegisterClasses = RF->getValueAsListOfDefs("RegClasses"); 1856 std::vector<int64_t> RegisterCosts = RF->getValueAsListOfInts("RegCosts"); 1857 ListInit *MoveElimInfo = RF->getValueAsListInit("AllowMoveElimination"); 1858 for (unsigned I = 0, E = RegisterClasses.size(); I < E; ++I) { 1859 int Cost = RegisterCosts.size() > I ? RegisterCosts[I] : 1; 1860 1861 bool AllowMoveElim = false; 1862 if (MoveElimInfo->size() > I) { 1863 BitInit *Val = cast<BitInit>(MoveElimInfo->getElement(I)); 1864 AllowMoveElim = Val->getValue(); 1865 } 1866 1867 CGRF.Costs.emplace_back(RegisterClasses[I], Cost, AllowMoveElim); 1868 } 1869 } 1870 } 1871 1872 // Collect and sort WriteRes, ReadAdvance, and ProcResources. 1873 void CodeGenSchedModels::collectProcResources() { 1874 ProcResourceDefs = Records.getAllDerivedDefinitions("ProcResourceUnits"); 1875 ProcResGroups = Records.getAllDerivedDefinitions("ProcResGroup"); 1876 1877 // Add any subtarget-specific SchedReadWrites that are directly associated 1878 // with processor resources. Refer to the parent SchedClass's ProcIndices to 1879 // determine which processors they apply to. 1880 for (const CodeGenSchedClass &SC : 1881 make_range(schedClassBegin(), schedClassEnd())) { 1882 if (SC.ItinClassDef) { 1883 collectItinProcResources(SC.ItinClassDef); 1884 continue; 1885 } 1886 1887 // This class may have a default ReadWrite list which can be overriden by 1888 // InstRW definitions. 1889 for (Record *RW : SC.InstRWs) { 1890 Record *RWModelDef = RW->getValueAsDef("SchedModel"); 1891 unsigned PIdx = getProcModel(RWModelDef).Index; 1892 IdxVec Writes, Reads; 1893 findRWs(RW->getValueAsListOfDefs("OperandReadWrites"), Writes, Reads); 1894 collectRWResources(Writes, Reads, PIdx); 1895 } 1896 1897 collectRWResources(SC.Writes, SC.Reads, SC.ProcIndices); 1898 } 1899 // Add resources separately defined by each subtarget. 1900 RecVec WRDefs = Records.getAllDerivedDefinitions("WriteRes"); 1901 for (Record *WR : WRDefs) { 1902 Record *ModelDef = WR->getValueAsDef("SchedModel"); 1903 addWriteRes(WR, getProcModel(ModelDef).Index); 1904 } 1905 RecVec SWRDefs = Records.getAllDerivedDefinitions("SchedWriteRes"); 1906 for (Record *SWR : SWRDefs) { 1907 Record *ModelDef = SWR->getValueAsDef("SchedModel"); 1908 addWriteRes(SWR, getProcModel(ModelDef).Index); 1909 } 1910 RecVec RADefs = Records.getAllDerivedDefinitions("ReadAdvance"); 1911 for (Record *RA : RADefs) { 1912 Record *ModelDef = RA->getValueAsDef("SchedModel"); 1913 addReadAdvance(RA, getProcModel(ModelDef).Index); 1914 } 1915 RecVec SRADefs = Records.getAllDerivedDefinitions("SchedReadAdvance"); 1916 for (Record *SRA : SRADefs) { 1917 if (SRA->getValueInit("SchedModel")->isComplete()) { 1918 Record *ModelDef = SRA->getValueAsDef("SchedModel"); 1919 addReadAdvance(SRA, getProcModel(ModelDef).Index); 1920 } 1921 } 1922 // Add ProcResGroups that are defined within this processor model, which may 1923 // not be directly referenced but may directly specify a buffer size. 1924 RecVec ProcResGroups = Records.getAllDerivedDefinitions("ProcResGroup"); 1925 for (Record *PRG : ProcResGroups) { 1926 if (!PRG->getValueInit("SchedModel")->isComplete()) 1927 continue; 1928 CodeGenProcModel &PM = getProcModel(PRG->getValueAsDef("SchedModel")); 1929 if (!is_contained(PM.ProcResourceDefs, PRG)) 1930 PM.ProcResourceDefs.push_back(PRG); 1931 } 1932 // Add ProcResourceUnits unconditionally. 1933 for (Record *PRU : Records.getAllDerivedDefinitions("ProcResourceUnits")) { 1934 if (!PRU->getValueInit("SchedModel")->isComplete()) 1935 continue; 1936 CodeGenProcModel &PM = getProcModel(PRU->getValueAsDef("SchedModel")); 1937 if (!is_contained(PM.ProcResourceDefs, PRU)) 1938 PM.ProcResourceDefs.push_back(PRU); 1939 } 1940 // Finalize each ProcModel by sorting the record arrays. 1941 for (CodeGenProcModel &PM : ProcModels) { 1942 llvm::sort(PM.WriteResDefs, LessRecord()); 1943 llvm::sort(PM.ReadAdvanceDefs, LessRecord()); 1944 llvm::sort(PM.ProcResourceDefs, LessRecord()); 1945 LLVM_DEBUG( 1946 PM.dump(); 1947 dbgs() << "WriteResDefs: "; for (RecIter RI = PM.WriteResDefs.begin(), 1948 RE = PM.WriteResDefs.end(); 1949 RI != RE; ++RI) { 1950 if ((*RI)->isSubClassOf("WriteRes")) 1951 dbgs() << (*RI)->getValueAsDef("WriteType")->getName() << " "; 1952 else 1953 dbgs() << (*RI)->getName() << " "; 1954 } dbgs() << "\nReadAdvanceDefs: "; 1955 for (RecIter RI = PM.ReadAdvanceDefs.begin(), 1956 RE = PM.ReadAdvanceDefs.end(); 1957 RI != RE; ++RI) { 1958 if ((*RI)->isSubClassOf("ReadAdvance")) 1959 dbgs() << (*RI)->getValueAsDef("ReadType")->getName() << " "; 1960 else 1961 dbgs() << (*RI)->getName() << " "; 1962 } dbgs() 1963 << "\nProcResourceDefs: "; 1964 for (RecIter RI = PM.ProcResourceDefs.begin(), 1965 RE = PM.ProcResourceDefs.end(); 1966 RI != RE; ++RI) { dbgs() << (*RI)->getName() << " "; } dbgs() 1967 << '\n'); 1968 verifyProcResourceGroups(PM); 1969 } 1970 1971 ProcResourceDefs.clear(); 1972 ProcResGroups.clear(); 1973 } 1974 1975 void CodeGenSchedModels::checkCompleteness() { 1976 bool Complete = true; 1977 bool HadCompleteModel = false; 1978 for (const CodeGenProcModel &ProcModel : procModels()) { 1979 const bool HasItineraries = ProcModel.hasItineraries(); 1980 if (!ProcModel.ModelDef->getValueAsBit("CompleteModel")) 1981 continue; 1982 for (const CodeGenInstruction *Inst : Target.getInstructionsByEnumValue()) { 1983 if (Inst->hasNoSchedulingInfo) 1984 continue; 1985 if (ProcModel.isUnsupported(*Inst)) 1986 continue; 1987 unsigned SCIdx = getSchedClassIdx(*Inst); 1988 if (!SCIdx) { 1989 if (Inst->TheDef->isValueUnset("SchedRW") && !HadCompleteModel) { 1990 PrintError(Inst->TheDef->getLoc(), 1991 "No schedule information for instruction '" + 1992 Inst->TheDef->getName() + "' in SchedMachineModel '" + 1993 ProcModel.ModelDef->getName() + "'"); 1994 Complete = false; 1995 } 1996 continue; 1997 } 1998 1999 const CodeGenSchedClass &SC = getSchedClass(SCIdx); 2000 if (!SC.Writes.empty()) 2001 continue; 2002 if (HasItineraries && SC.ItinClassDef != nullptr && 2003 SC.ItinClassDef->getName() != "NoItinerary") 2004 continue; 2005 2006 const RecVec &InstRWs = SC.InstRWs; 2007 auto I = find_if(InstRWs, [&ProcModel](const Record *R) { 2008 return R->getValueAsDef("SchedModel") == ProcModel.ModelDef; 2009 }); 2010 if (I == InstRWs.end()) { 2011 PrintError(Inst->TheDef->getLoc(), "'" + ProcModel.ModelName + 2012 "' lacks information for '" + 2013 Inst->TheDef->getName() + "'"); 2014 Complete = false; 2015 } 2016 } 2017 HadCompleteModel = true; 2018 } 2019 if (!Complete) { 2020 errs() << "\n\nIncomplete schedule models found.\n" 2021 << "- Consider setting 'CompleteModel = 0' while developing new models.\n" 2022 << "- Pseudo instructions can be marked with 'hasNoSchedulingInfo = 1'.\n" 2023 << "- Instructions should usually have Sched<[...]> as a superclass, " 2024 "you may temporarily use an empty list.\n" 2025 << "- Instructions related to unsupported features can be excluded with " 2026 "list<Predicate> UnsupportedFeatures = [HasA,..,HasY]; in the " 2027 "processor model.\n\n"; 2028 PrintFatalError("Incomplete schedule model"); 2029 } 2030 } 2031 2032 // Collect itinerary class resources for each processor. 2033 void CodeGenSchedModels::collectItinProcResources(Record *ItinClassDef) { 2034 for (unsigned PIdx = 0, PEnd = ProcModels.size(); PIdx != PEnd; ++PIdx) { 2035 const CodeGenProcModel &PM = ProcModels[PIdx]; 2036 // For all ItinRW entries. 2037 bool HasMatch = false; 2038 for (RecIter II = PM.ItinRWDefs.begin(), IE = PM.ItinRWDefs.end(); 2039 II != IE; ++II) { 2040 RecVec Matched = (*II)->getValueAsListOfDefs("MatchedItinClasses"); 2041 if (!std::count(Matched.begin(), Matched.end(), ItinClassDef)) 2042 continue; 2043 if (HasMatch) 2044 PrintFatalError((*II)->getLoc(), "Duplicate itinerary class " 2045 + ItinClassDef->getName() 2046 + " in ItinResources for " + PM.ModelName); 2047 HasMatch = true; 2048 IdxVec Writes, Reads; 2049 findRWs((*II)->getValueAsListOfDefs("OperandReadWrites"), Writes, Reads); 2050 collectRWResources(Writes, Reads, PIdx); 2051 } 2052 } 2053 } 2054 2055 void CodeGenSchedModels::collectRWResources(unsigned RWIdx, bool IsRead, 2056 ArrayRef<unsigned> ProcIndices) { 2057 const CodeGenSchedRW &SchedRW = getSchedRW(RWIdx, IsRead); 2058 if (SchedRW.TheDef) { 2059 if (!IsRead && SchedRW.TheDef->isSubClassOf("SchedWriteRes")) { 2060 for (unsigned Idx : ProcIndices) 2061 addWriteRes(SchedRW.TheDef, Idx); 2062 } 2063 else if (IsRead && SchedRW.TheDef->isSubClassOf("SchedReadAdvance")) { 2064 for (unsigned Idx : ProcIndices) 2065 addReadAdvance(SchedRW.TheDef, Idx); 2066 } 2067 } 2068 for (RecIter AI = SchedRW.Aliases.begin(), AE = SchedRW.Aliases.end(); 2069 AI != AE; ++AI) { 2070 IdxVec AliasProcIndices; 2071 if ((*AI)->getValueInit("SchedModel")->isComplete()) { 2072 AliasProcIndices.push_back( 2073 getProcModel((*AI)->getValueAsDef("SchedModel")).Index); 2074 } 2075 else 2076 AliasProcIndices = ProcIndices; 2077 const CodeGenSchedRW &AliasRW = getSchedRW((*AI)->getValueAsDef("AliasRW")); 2078 assert(AliasRW.IsRead == IsRead && "cannot alias reads to writes"); 2079 2080 IdxVec ExpandedRWs; 2081 expandRWSequence(AliasRW.Index, ExpandedRWs, IsRead); 2082 for (IdxIter SI = ExpandedRWs.begin(), SE = ExpandedRWs.end(); 2083 SI != SE; ++SI) { 2084 collectRWResources(*SI, IsRead, AliasProcIndices); 2085 } 2086 } 2087 } 2088 2089 // Collect resources for a set of read/write types and processor indices. 2090 void CodeGenSchedModels::collectRWResources(ArrayRef<unsigned> Writes, 2091 ArrayRef<unsigned> Reads, 2092 ArrayRef<unsigned> ProcIndices) { 2093 for (unsigned Idx : Writes) 2094 collectRWResources(Idx, /*IsRead=*/false, ProcIndices); 2095 2096 for (unsigned Idx : Reads) 2097 collectRWResources(Idx, /*IsRead=*/true, ProcIndices); 2098 } 2099 2100 // Find the processor's resource units for this kind of resource. 2101 Record *CodeGenSchedModels::findProcResUnits(Record *ProcResKind, 2102 const CodeGenProcModel &PM, 2103 ArrayRef<SMLoc> Loc) const { 2104 if (ProcResKind->isSubClassOf("ProcResourceUnits")) 2105 return ProcResKind; 2106 2107 Record *ProcUnitDef = nullptr; 2108 assert(!ProcResourceDefs.empty()); 2109 assert(!ProcResGroups.empty()); 2110 2111 for (Record *ProcResDef : ProcResourceDefs) { 2112 if (ProcResDef->getValueAsDef("Kind") == ProcResKind 2113 && ProcResDef->getValueAsDef("SchedModel") == PM.ModelDef) { 2114 if (ProcUnitDef) { 2115 PrintFatalError(Loc, 2116 "Multiple ProcessorResourceUnits associated with " 2117 + ProcResKind->getName()); 2118 } 2119 ProcUnitDef = ProcResDef; 2120 } 2121 } 2122 for (Record *ProcResGroup : ProcResGroups) { 2123 if (ProcResGroup == ProcResKind 2124 && ProcResGroup->getValueAsDef("SchedModel") == PM.ModelDef) { 2125 if (ProcUnitDef) { 2126 PrintFatalError(Loc, 2127 "Multiple ProcessorResourceUnits associated with " 2128 + ProcResKind->getName()); 2129 } 2130 ProcUnitDef = ProcResGroup; 2131 } 2132 } 2133 if (!ProcUnitDef) { 2134 PrintFatalError(Loc, 2135 "No ProcessorResources associated with " 2136 + ProcResKind->getName()); 2137 } 2138 return ProcUnitDef; 2139 } 2140 2141 // Iteratively add a resource and its super resources. 2142 void CodeGenSchedModels::addProcResource(Record *ProcResKind, 2143 CodeGenProcModel &PM, 2144 ArrayRef<SMLoc> Loc) { 2145 while (true) { 2146 Record *ProcResUnits = findProcResUnits(ProcResKind, PM, Loc); 2147 2148 // See if this ProcResource is already associated with this processor. 2149 if (is_contained(PM.ProcResourceDefs, ProcResUnits)) 2150 return; 2151 2152 PM.ProcResourceDefs.push_back(ProcResUnits); 2153 if (ProcResUnits->isSubClassOf("ProcResGroup")) 2154 return; 2155 2156 if (!ProcResUnits->getValueInit("Super")->isComplete()) 2157 return; 2158 2159 ProcResKind = ProcResUnits->getValueAsDef("Super"); 2160 } 2161 } 2162 2163 // Add resources for a SchedWrite to this processor if they don't exist. 2164 void CodeGenSchedModels::addWriteRes(Record *ProcWriteResDef, unsigned PIdx) { 2165 assert(PIdx && "don't add resources to an invalid Processor model"); 2166 2167 RecVec &WRDefs = ProcModels[PIdx].WriteResDefs; 2168 if (is_contained(WRDefs, ProcWriteResDef)) 2169 return; 2170 WRDefs.push_back(ProcWriteResDef); 2171 2172 // Visit ProcResourceKinds referenced by the newly discovered WriteRes. 2173 RecVec ProcResDefs = ProcWriteResDef->getValueAsListOfDefs("ProcResources"); 2174 for (RecIter WritePRI = ProcResDefs.begin(), WritePRE = ProcResDefs.end(); 2175 WritePRI != WritePRE; ++WritePRI) { 2176 addProcResource(*WritePRI, ProcModels[PIdx], ProcWriteResDef->getLoc()); 2177 } 2178 } 2179 2180 // Add resources for a ReadAdvance to this processor if they don't exist. 2181 void CodeGenSchedModels::addReadAdvance(Record *ProcReadAdvanceDef, 2182 unsigned PIdx) { 2183 RecVec &RADefs = ProcModels[PIdx].ReadAdvanceDefs; 2184 if (is_contained(RADefs, ProcReadAdvanceDef)) 2185 return; 2186 RADefs.push_back(ProcReadAdvanceDef); 2187 } 2188 2189 unsigned CodeGenProcModel::getProcResourceIdx(Record *PRDef) const { 2190 RecIter PRPos = find(ProcResourceDefs, PRDef); 2191 if (PRPos == ProcResourceDefs.end()) 2192 PrintFatalError(PRDef->getLoc(), "ProcResource def is not included in " 2193 "the ProcResources list for " + ModelName); 2194 // Idx=0 is reserved for invalid. 2195 return 1 + (PRPos - ProcResourceDefs.begin()); 2196 } 2197 2198 bool CodeGenProcModel::isUnsupported(const CodeGenInstruction &Inst) const { 2199 for (const Record *TheDef : UnsupportedFeaturesDefs) { 2200 for (const Record *PredDef : Inst.TheDef->getValueAsListOfDefs("Predicates")) { 2201 if (TheDef->getName() == PredDef->getName()) 2202 return true; 2203 } 2204 } 2205 return false; 2206 } 2207 2208 #ifndef NDEBUG 2209 void CodeGenProcModel::dump() const { 2210 dbgs() << Index << ": " << ModelName << " " 2211 << (ModelDef ? ModelDef->getName() : "inferred") << " " 2212 << (ItinsDef ? ItinsDef->getName() : "no itinerary") << '\n'; 2213 } 2214 2215 void CodeGenSchedRW::dump() const { 2216 dbgs() << Name << (IsVariadic ? " (V) " : " "); 2217 if (IsSequence) { 2218 dbgs() << "("; 2219 dumpIdxVec(Sequence); 2220 dbgs() << ")"; 2221 } 2222 } 2223 2224 void CodeGenSchedClass::dump(const CodeGenSchedModels* SchedModels) const { 2225 dbgs() << "SCHEDCLASS " << Index << ":" << Name << '\n' 2226 << " Writes: "; 2227 for (unsigned i = 0, N = Writes.size(); i < N; ++i) { 2228 SchedModels->getSchedWrite(Writes[i]).dump(); 2229 if (i < N-1) { 2230 dbgs() << '\n'; 2231 dbgs().indent(10); 2232 } 2233 } 2234 dbgs() << "\n Reads: "; 2235 for (unsigned i = 0, N = Reads.size(); i < N; ++i) { 2236 SchedModels->getSchedRead(Reads[i]).dump(); 2237 if (i < N-1) { 2238 dbgs() << '\n'; 2239 dbgs().indent(10); 2240 } 2241 } 2242 dbgs() << "\n ProcIdx: "; dumpIdxVec(ProcIndices); 2243 if (!Transitions.empty()) { 2244 dbgs() << "\n Transitions for Proc "; 2245 for (const CodeGenSchedTransition &Transition : Transitions) { 2246 dumpIdxVec(Transition.ProcIndices); 2247 } 2248 } 2249 dbgs() << '\n'; 2250 } 2251 2252 void PredTransitions::dump() const { 2253 dbgs() << "Expanded Variants:\n"; 2254 for (std::vector<PredTransition>::const_iterator 2255 TI = TransVec.begin(), TE = TransVec.end(); TI != TE; ++TI) { 2256 dbgs() << "{"; 2257 for (SmallVectorImpl<PredCheck>::const_iterator 2258 PCI = TI->PredTerm.begin(), PCE = TI->PredTerm.end(); 2259 PCI != PCE; ++PCI) { 2260 if (PCI != TI->PredTerm.begin()) 2261 dbgs() << ", "; 2262 dbgs() << SchedModels.getSchedRW(PCI->RWIdx, PCI->IsRead).Name 2263 << ":" << PCI->Predicate->getName(); 2264 } 2265 dbgs() << "},\n => {"; 2266 for (SmallVectorImpl<SmallVector<unsigned,4>>::const_iterator 2267 WSI = TI->WriteSequences.begin(), WSE = TI->WriteSequences.end(); 2268 WSI != WSE; ++WSI) { 2269 dbgs() << "("; 2270 for (SmallVectorImpl<unsigned>::const_iterator 2271 WI = WSI->begin(), WE = WSI->end(); WI != WE; ++WI) { 2272 if (WI != WSI->begin()) 2273 dbgs() << ", "; 2274 dbgs() << SchedModels.getSchedWrite(*WI).Name; 2275 } 2276 dbgs() << "),"; 2277 } 2278 dbgs() << "}\n"; 2279 } 2280 } 2281 #endif // NDEBUG 2282