1 //===- RegisterCoalescer.cpp - Generic Register Coalescing Interface -------==// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This file implements the generic RegisterCoalescer interface which 11 // is used as the common interface used by all clients and 12 // implementations of register coalescing. 13 // 14 //===----------------------------------------------------------------------===// 15 16 #define DEBUG_TYPE "regcoalescing" 17 #include "RegisterCoalescer.h" 18 #include "LiveDebugVariables.h" 19 #include "RegisterClassInfo.h" 20 #include "VirtRegMap.h" 21 22 #include "llvm/Pass.h" 23 #include "llvm/Value.h" 24 #include "llvm/CodeGen/LiveIntervalAnalysis.h" 25 #include "llvm/CodeGen/MachineInstr.h" 26 #include "llvm/CodeGen/MachineRegisterInfo.h" 27 #include "llvm/Target/TargetInstrInfo.h" 28 #include "llvm/Target/TargetRegisterInfo.h" 29 #include "llvm/CodeGen/LiveIntervalAnalysis.h" 30 #include "llvm/Analysis/AliasAnalysis.h" 31 #include "llvm/CodeGen/MachineFrameInfo.h" 32 #include "llvm/CodeGen/MachineInstr.h" 33 #include "llvm/CodeGen/MachineLoopInfo.h" 34 #include "llvm/CodeGen/MachineRegisterInfo.h" 35 #include "llvm/CodeGen/Passes.h" 36 #include "llvm/Target/TargetInstrInfo.h" 37 #include "llvm/Target/TargetMachine.h" 38 #include "llvm/Target/TargetOptions.h" 39 #include "llvm/Support/CommandLine.h" 40 #include "llvm/Support/Debug.h" 41 #include "llvm/Support/ErrorHandling.h" 42 #include "llvm/Support/raw_ostream.h" 43 #include "llvm/ADT/OwningPtr.h" 44 #include "llvm/ADT/SmallSet.h" 45 #include "llvm/ADT/Statistic.h" 46 #include "llvm/ADT/STLExtras.h" 47 #include <algorithm> 48 #include <cmath> 49 using namespace llvm; 50 51 STATISTIC(numJoins , "Number of interval joins performed"); 52 STATISTIC(numCrossRCs , "Number of cross class joins performed"); 53 STATISTIC(numCommutes , "Number of instruction commuting performed"); 54 STATISTIC(numExtends , "Number of copies extended"); 55 STATISTIC(NumReMats , "Number of instructions re-materialized"); 56 STATISTIC(numPeep , "Number of identity moves eliminated after coalescing"); 57 STATISTIC(numAborts , "Number of times interval joining aborted"); 58 STATISTIC(NumInflated , "Number of register classes inflated"); 59 60 static cl::opt<bool> 61 EnableJoining("join-liveintervals", 62 cl::desc("Coalesce copies (default=true)"), 63 cl::init(true)); 64 65 static cl::opt<bool> 66 DisableCrossClassJoin("disable-cross-class-join", 67 cl::desc("Avoid coalescing cross register class copies"), 68 cl::init(false), cl::Hidden); 69 70 static cl::opt<bool> 71 EnablePhysicalJoin("join-physregs", 72 cl::desc("Join physical register copies"), 73 cl::init(false), cl::Hidden); 74 75 static cl::opt<bool> 76 VerifyCoalescing("verify-coalescing", 77 cl::desc("Verify machine instrs before and after register coalescing"), 78 cl::Hidden); 79 80 namespace { 81 class RegisterCoalescer : public MachineFunctionPass { 82 MachineFunction* MF; 83 MachineRegisterInfo* MRI; 84 const TargetMachine* TM; 85 const TargetRegisterInfo* TRI; 86 const TargetInstrInfo* TII; 87 LiveIntervals *LIS; 88 LiveDebugVariables *LDV; 89 const MachineLoopInfo* Loops; 90 AliasAnalysis *AA; 91 RegisterClassInfo RegClassInfo; 92 93 /// JoinedCopies - Keep track of copies eliminated due to coalescing. 94 /// 95 SmallPtrSet<MachineInstr*, 32> JoinedCopies; 96 97 /// ReMatCopies - Keep track of copies eliminated due to remat. 98 /// 99 SmallPtrSet<MachineInstr*, 32> ReMatCopies; 100 101 /// ReMatDefs - Keep track of definition instructions which have 102 /// been remat'ed. 103 SmallPtrSet<MachineInstr*, 8> ReMatDefs; 104 105 /// joinIntervals - join compatible live intervals 106 void joinIntervals(); 107 108 /// CopyCoalesceInMBB - Coalesce copies in the specified MBB, putting 109 /// copies that cannot yet be coalesced into the "TryAgain" list. 110 void CopyCoalesceInMBB(MachineBasicBlock *MBB, 111 std::vector<MachineInstr*> &TryAgain); 112 113 /// JoinCopy - Attempt to join intervals corresponding to SrcReg/DstReg, 114 /// which are the src/dst of the copy instruction CopyMI. This returns 115 /// true if the copy was successfully coalesced away. If it is not 116 /// currently possible to coalesce this interval, but it may be possible if 117 /// other things get coalesced, then it returns true by reference in 118 /// 'Again'. 119 bool JoinCopy(MachineInstr *TheCopy, bool &Again); 120 121 /// JoinIntervals - Attempt to join these two intervals. On failure, this 122 /// returns false. The output "SrcInt" will not have been modified, so we 123 /// can use this information below to update aliases. 124 bool JoinIntervals(CoalescerPair &CP); 125 126 /// AdjustCopiesBackFrom - We found a non-trivially-coalescable copy. If 127 /// the source value number is defined by a copy from the destination reg 128 /// see if we can merge these two destination reg valno# into a single 129 /// value number, eliminating a copy. 130 bool AdjustCopiesBackFrom(const CoalescerPair &CP, MachineInstr *CopyMI); 131 132 /// HasOtherReachingDefs - Return true if there are definitions of IntB 133 /// other than BValNo val# that can reach uses of AValno val# of IntA. 134 bool HasOtherReachingDefs(LiveInterval &IntA, LiveInterval &IntB, 135 VNInfo *AValNo, VNInfo *BValNo); 136 137 /// RemoveCopyByCommutingDef - We found a non-trivially-coalescable copy. 138 /// If the source value number is defined by a commutable instruction and 139 /// its other operand is coalesced to the copy dest register, see if we 140 /// can transform the copy into a noop by commuting the definition. 141 bool RemoveCopyByCommutingDef(const CoalescerPair &CP,MachineInstr *CopyMI); 142 143 /// ReMaterializeTrivialDef - If the source of a copy is defined by a 144 /// trivial computation, replace the copy by rematerialize the definition. 145 /// If PreserveSrcInt is true, make sure SrcInt is valid after the call. 146 bool ReMaterializeTrivialDef(LiveInterval &SrcInt, bool PreserveSrcInt, 147 unsigned DstReg, unsigned DstSubIdx, 148 MachineInstr *CopyMI); 149 150 /// shouldJoinPhys - Return true if a physreg copy should be joined. 151 bool shouldJoinPhys(CoalescerPair &CP); 152 153 /// isWinToJoinCrossClass - Return true if it's profitable to coalesce 154 /// two virtual registers from different register classes. 155 bool isWinToJoinCrossClass(unsigned SrcReg, 156 unsigned DstReg, 157 const TargetRegisterClass *SrcRC, 158 const TargetRegisterClass *DstRC, 159 const TargetRegisterClass *NewRC); 160 161 /// UpdateRegDefsUses - Replace all defs and uses of SrcReg to DstReg and 162 /// update the subregister number if it is not zero. If DstReg is a 163 /// physical register and the existing subregister number of the def / use 164 /// being updated is not zero, make sure to set it to the correct physical 165 /// subregister. 166 void UpdateRegDefsUses(const CoalescerPair &CP); 167 168 /// RemoveDeadDef - If a def of a live interval is now determined dead, 169 /// remove the val# it defines. If the live interval becomes empty, remove 170 /// it as well. 171 bool RemoveDeadDef(LiveInterval &li, MachineInstr *DefMI); 172 173 /// RemoveCopyFlag - If DstReg is no longer defined by CopyMI, clear the 174 /// VNInfo copy flag for DstReg and all aliases. 175 void RemoveCopyFlag(unsigned DstReg, const MachineInstr *CopyMI); 176 177 /// markAsJoined - Remember that CopyMI has already been joined. 178 void markAsJoined(MachineInstr *CopyMI); 179 180 /// eliminateUndefCopy - Handle copies of undef values. 181 bool eliminateUndefCopy(MachineInstr *CopyMI, const CoalescerPair &CP); 182 183 public: 184 static char ID; // Class identification, replacement for typeinfo 185 RegisterCoalescer() : MachineFunctionPass(ID) { 186 initializeRegisterCoalescerPass(*PassRegistry::getPassRegistry()); 187 } 188 189 virtual void getAnalysisUsage(AnalysisUsage &AU) const; 190 191 virtual void releaseMemory(); 192 193 /// runOnMachineFunction - pass entry point 194 virtual bool runOnMachineFunction(MachineFunction&); 195 196 /// print - Implement the dump method. 197 virtual void print(raw_ostream &O, const Module* = 0) const; 198 }; 199 } /// end anonymous namespace 200 201 char &llvm::RegisterCoalescerPassID = RegisterCoalescer::ID; 202 203 INITIALIZE_PASS_BEGIN(RegisterCoalescer, "simple-register-coalescing", 204 "Simple Register Coalescing", false, false) 205 INITIALIZE_PASS_DEPENDENCY(LiveIntervals) 206 INITIALIZE_PASS_DEPENDENCY(LiveDebugVariables) 207 INITIALIZE_PASS_DEPENDENCY(SlotIndexes) 208 INITIALIZE_PASS_DEPENDENCY(MachineLoopInfo) 209 INITIALIZE_PASS_DEPENDENCY(StrongPHIElimination) 210 INITIALIZE_PASS_DEPENDENCY(PHIElimination) 211 INITIALIZE_PASS_DEPENDENCY(TwoAddressInstructionPass) 212 INITIALIZE_AG_DEPENDENCY(AliasAnalysis) 213 INITIALIZE_PASS_END(RegisterCoalescer, "simple-register-coalescing", 214 "Simple Register Coalescing", false, false) 215 216 char RegisterCoalescer::ID = 0; 217 218 static unsigned compose(const TargetRegisterInfo &tri, unsigned a, unsigned b) { 219 if (!a) return b; 220 if (!b) return a; 221 return tri.composeSubRegIndices(a, b); 222 } 223 224 static bool isMoveInstr(const TargetRegisterInfo &tri, const MachineInstr *MI, 225 unsigned &Src, unsigned &Dst, 226 unsigned &SrcSub, unsigned &DstSub) { 227 if (MI->isCopy()) { 228 Dst = MI->getOperand(0).getReg(); 229 DstSub = MI->getOperand(0).getSubReg(); 230 Src = MI->getOperand(1).getReg(); 231 SrcSub = MI->getOperand(1).getSubReg(); 232 } else if (MI->isSubregToReg()) { 233 Dst = MI->getOperand(0).getReg(); 234 DstSub = compose(tri, MI->getOperand(0).getSubReg(), 235 MI->getOperand(3).getImm()); 236 Src = MI->getOperand(2).getReg(); 237 SrcSub = MI->getOperand(2).getSubReg(); 238 } else 239 return false; 240 return true; 241 } 242 243 bool CoalescerPair::setRegisters(const MachineInstr *MI) { 244 SrcReg = DstReg = SubIdx = 0; 245 NewRC = 0; 246 Flipped = CrossClass = false; 247 248 unsigned Src, Dst, SrcSub, DstSub; 249 if (!isMoveInstr(TRI, MI, Src, Dst, SrcSub, DstSub)) 250 return false; 251 Partial = SrcSub || DstSub; 252 253 // If one register is a physreg, it must be Dst. 254 if (TargetRegisterInfo::isPhysicalRegister(Src)) { 255 if (TargetRegisterInfo::isPhysicalRegister(Dst)) 256 return false; 257 std::swap(Src, Dst); 258 std::swap(SrcSub, DstSub); 259 Flipped = true; 260 } 261 262 const MachineRegisterInfo &MRI = MI->getParent()->getParent()->getRegInfo(); 263 264 if (TargetRegisterInfo::isPhysicalRegister(Dst)) { 265 // Eliminate DstSub on a physreg. 266 if (DstSub) { 267 Dst = TRI.getSubReg(Dst, DstSub); 268 if (!Dst) return false; 269 DstSub = 0; 270 } 271 272 // Eliminate SrcSub by picking a corresponding Dst superregister. 273 if (SrcSub) { 274 Dst = TRI.getMatchingSuperReg(Dst, SrcSub, MRI.getRegClass(Src)); 275 if (!Dst) return false; 276 SrcSub = 0; 277 } else if (!MRI.getRegClass(Src)->contains(Dst)) { 278 return false; 279 } 280 } else { 281 // Both registers are virtual. 282 283 // Both registers have subreg indices. 284 if (SrcSub && DstSub) { 285 // For now we only handle the case of identical indices in commensurate 286 // registers: Dreg:ssub_1 + Dreg:ssub_1 -> Dreg 287 // FIXME: Handle Qreg:ssub_3 + Dreg:ssub_1 as QReg:dsub_1 + Dreg. 288 if (SrcSub != DstSub) 289 return false; 290 const TargetRegisterClass *SrcRC = MRI.getRegClass(Src); 291 const TargetRegisterClass *DstRC = MRI.getRegClass(Dst); 292 if (!getCommonSubClass(DstRC, SrcRC)) 293 return false; 294 SrcSub = DstSub = 0; 295 } 296 297 // There can be no SrcSub. 298 if (SrcSub) { 299 std::swap(Src, Dst); 300 DstSub = SrcSub; 301 SrcSub = 0; 302 assert(!Flipped && "Unexpected flip"); 303 Flipped = true; 304 } 305 306 // Find the new register class. 307 const TargetRegisterClass *SrcRC = MRI.getRegClass(Src); 308 const TargetRegisterClass *DstRC = MRI.getRegClass(Dst); 309 if (DstSub) 310 NewRC = TRI.getMatchingSuperRegClass(DstRC, SrcRC, DstSub); 311 else 312 NewRC = getCommonSubClass(DstRC, SrcRC); 313 if (!NewRC) 314 return false; 315 CrossClass = NewRC != DstRC || NewRC != SrcRC; 316 } 317 // Check our invariants 318 assert(TargetRegisterInfo::isVirtualRegister(Src) && "Src must be virtual"); 319 assert(!(TargetRegisterInfo::isPhysicalRegister(Dst) && DstSub) && 320 "Cannot have a physical SubIdx"); 321 SrcReg = Src; 322 DstReg = Dst; 323 SubIdx = DstSub; 324 return true; 325 } 326 327 bool CoalescerPair::flip() { 328 if (SubIdx || TargetRegisterInfo::isPhysicalRegister(DstReg)) 329 return false; 330 std::swap(SrcReg, DstReg); 331 Flipped = !Flipped; 332 return true; 333 } 334 335 bool CoalescerPair::isCoalescable(const MachineInstr *MI) const { 336 if (!MI) 337 return false; 338 unsigned Src, Dst, SrcSub, DstSub; 339 if (!isMoveInstr(TRI, MI, Src, Dst, SrcSub, DstSub)) 340 return false; 341 342 // Find the virtual register that is SrcReg. 343 if (Dst == SrcReg) { 344 std::swap(Src, Dst); 345 std::swap(SrcSub, DstSub); 346 } else if (Src != SrcReg) { 347 return false; 348 } 349 350 // Now check that Dst matches DstReg. 351 if (TargetRegisterInfo::isPhysicalRegister(DstReg)) { 352 if (!TargetRegisterInfo::isPhysicalRegister(Dst)) 353 return false; 354 assert(!SubIdx && "Inconsistent CoalescerPair state."); 355 // DstSub could be set for a physreg from INSERT_SUBREG. 356 if (DstSub) 357 Dst = TRI.getSubReg(Dst, DstSub); 358 // Full copy of Src. 359 if (!SrcSub) 360 return DstReg == Dst; 361 // This is a partial register copy. Check that the parts match. 362 return TRI.getSubReg(DstReg, SrcSub) == Dst; 363 } else { 364 // DstReg is virtual. 365 if (DstReg != Dst) 366 return false; 367 // Registers match, do the subregisters line up? 368 return compose(TRI, SubIdx, SrcSub) == DstSub; 369 } 370 } 371 372 void RegisterCoalescer::getAnalysisUsage(AnalysisUsage &AU) const { 373 AU.setPreservesCFG(); 374 AU.addRequired<AliasAnalysis>(); 375 AU.addRequired<LiveIntervals>(); 376 AU.addPreserved<LiveIntervals>(); 377 AU.addRequired<LiveDebugVariables>(); 378 AU.addPreserved<LiveDebugVariables>(); 379 AU.addPreserved<SlotIndexes>(); 380 AU.addRequired<MachineLoopInfo>(); 381 AU.addPreserved<MachineLoopInfo>(); 382 AU.addPreservedID(MachineDominatorsID); 383 AU.addPreservedID(StrongPHIEliminationID); 384 AU.addPreservedID(PHIEliminationID); 385 AU.addPreservedID(TwoAddressInstructionPassID); 386 MachineFunctionPass::getAnalysisUsage(AU); 387 } 388 389 void RegisterCoalescer::markAsJoined(MachineInstr *CopyMI) { 390 /// Joined copies are not deleted immediately, but kept in JoinedCopies. 391 JoinedCopies.insert(CopyMI); 392 393 /// Mark all register operands of CopyMI as <undef> so they won't affect dead 394 /// code elimination. 395 for (MachineInstr::mop_iterator I = CopyMI->operands_begin(), 396 E = CopyMI->operands_end(); I != E; ++I) 397 if (I->isReg()) 398 I->setIsUndef(true); 399 } 400 401 /// AdjustCopiesBackFrom - We found a non-trivially-coalescable copy with IntA 402 /// being the source and IntB being the dest, thus this defines a value number 403 /// in IntB. If the source value number (in IntA) is defined by a copy from B, 404 /// see if we can merge these two pieces of B into a single value number, 405 /// eliminating a copy. For example: 406 /// 407 /// A3 = B0 408 /// ... 409 /// B1 = A3 <- this copy 410 /// 411 /// In this case, B0 can be extended to where the B1 copy lives, allowing the B1 412 /// value number to be replaced with B0 (which simplifies the B liveinterval). 413 /// 414 /// This returns true if an interval was modified. 415 /// 416 bool RegisterCoalescer::AdjustCopiesBackFrom(const CoalescerPair &CP, 417 MachineInstr *CopyMI) { 418 // Bail if there is no dst interval - can happen when merging physical subreg 419 // operations. 420 if (!LIS->hasInterval(CP.getDstReg())) 421 return false; 422 423 LiveInterval &IntA = 424 LIS->getInterval(CP.isFlipped() ? CP.getDstReg() : CP.getSrcReg()); 425 LiveInterval &IntB = 426 LIS->getInterval(CP.isFlipped() ? CP.getSrcReg() : CP.getDstReg()); 427 SlotIndex CopyIdx = LIS->getInstructionIndex(CopyMI).getDefIndex(); 428 429 // BValNo is a value number in B that is defined by a copy from A. 'B3' in 430 // the example above. 431 LiveInterval::iterator BLR = IntB.FindLiveRangeContaining(CopyIdx); 432 if (BLR == IntB.end()) return false; 433 VNInfo *BValNo = BLR->valno; 434 435 // Get the location that B is defined at. Two options: either this value has 436 // an unknown definition point or it is defined at CopyIdx. If unknown, we 437 // can't process it. 438 if (!BValNo->isDefByCopy()) return false; 439 assert(BValNo->def == CopyIdx && "Copy doesn't define the value?"); 440 441 // AValNo is the value number in A that defines the copy, A3 in the example. 442 SlotIndex CopyUseIdx = CopyIdx.getUseIndex(); 443 LiveInterval::iterator ALR = IntA.FindLiveRangeContaining(CopyUseIdx); 444 // The live range might not exist after fun with physreg coalescing. 445 if (ALR == IntA.end()) return false; 446 VNInfo *AValNo = ALR->valno; 447 // If it's re-defined by an early clobber somewhere in the live range, then 448 // it's not safe to eliminate the copy. FIXME: This is a temporary workaround. 449 // See PR3149: 450 // 172 %ECX<def> = MOV32rr %reg1039<kill> 451 // 180 INLINEASM <es:subl $5,$1 452 // sbbl $3,$0>, 10, %EAX<def>, 14, %ECX<earlyclobber,def>, 9, 453 // %EAX<kill>, 454 // 36, <fi#0>, 1, %reg0, 0, 9, %ECX<kill>, 36, <fi#1>, 1, %reg0, 0 455 // 188 %EAX<def> = MOV32rr %EAX<kill> 456 // 196 %ECX<def> = MOV32rr %ECX<kill> 457 // 204 %ECX<def> = MOV32rr %ECX<kill> 458 // 212 %EAX<def> = MOV32rr %EAX<kill> 459 // 220 %EAX<def> = MOV32rr %EAX 460 // 228 %reg1039<def> = MOV32rr %ECX<kill> 461 // The early clobber operand ties ECX input to the ECX def. 462 // 463 // The live interval of ECX is represented as this: 464 // %reg20,inf = [46,47:1)[174,230:0) 0@174-(230) 1@46-(47) 465 // The coalescer has no idea there was a def in the middle of [174,230]. 466 if (AValNo->hasRedefByEC()) 467 return false; 468 469 // If AValNo is defined as a copy from IntB, we can potentially process this. 470 // Get the instruction that defines this value number. 471 if (!CP.isCoalescable(AValNo->getCopy())) 472 return false; 473 474 // Get the LiveRange in IntB that this value number starts with. 475 LiveInterval::iterator ValLR = 476 IntB.FindLiveRangeContaining(AValNo->def.getPrevSlot()); 477 if (ValLR == IntB.end()) 478 return false; 479 480 // Make sure that the end of the live range is inside the same block as 481 // CopyMI. 482 MachineInstr *ValLREndInst = 483 LIS->getInstructionFromIndex(ValLR->end.getPrevSlot()); 484 if (!ValLREndInst || ValLREndInst->getParent() != CopyMI->getParent()) 485 return false; 486 487 // Okay, we now know that ValLR ends in the same block that the CopyMI 488 // live-range starts. If there are no intervening live ranges between them in 489 // IntB, we can merge them. 490 if (ValLR+1 != BLR) return false; 491 492 // If a live interval is a physical register, conservatively check if any 493 // of its aliases is overlapping the live interval of the virtual register. 494 // If so, do not coalesce. 495 if (TargetRegisterInfo::isPhysicalRegister(IntB.reg)) { 496 for (const unsigned *AS = TRI->getAliasSet(IntB.reg); *AS; ++AS) 497 if (LIS->hasInterval(*AS) && IntA.overlaps(LIS->getInterval(*AS))) { 498 DEBUG({ 499 dbgs() << "\t\tInterfere with alias "; 500 LIS->getInterval(*AS).print(dbgs(), TRI); 501 }); 502 return false; 503 } 504 } 505 506 DEBUG({ 507 dbgs() << "Extending: "; 508 IntB.print(dbgs(), TRI); 509 }); 510 511 SlotIndex FillerStart = ValLR->end, FillerEnd = BLR->start; 512 // We are about to delete CopyMI, so need to remove it as the 'instruction 513 // that defines this value #'. Update the valnum with the new defining 514 // instruction #. 515 BValNo->def = FillerStart; 516 BValNo->setCopy(0); 517 518 // Okay, we can merge them. We need to insert a new liverange: 519 // [ValLR.end, BLR.begin) of either value number, then we merge the 520 // two value numbers. 521 IntB.addRange(LiveRange(FillerStart, FillerEnd, BValNo)); 522 523 // If the IntB live range is assigned to a physical register, and if that 524 // physreg has sub-registers, update their live intervals as well. 525 if (TargetRegisterInfo::isPhysicalRegister(IntB.reg)) { 526 for (const unsigned *SR = TRI->getSubRegisters(IntB.reg); *SR; ++SR) { 527 if (!LIS->hasInterval(*SR)) 528 continue; 529 LiveInterval &SRLI = LIS->getInterval(*SR); 530 SRLI.addRange(LiveRange(FillerStart, FillerEnd, 531 SRLI.getNextValue(FillerStart, 0, 532 LIS->getVNInfoAllocator()))); 533 } 534 } 535 536 // Okay, merge "B1" into the same value number as "B0". 537 if (BValNo != ValLR->valno) { 538 // If B1 is killed by a PHI, then the merged live range must also be killed 539 // by the same PHI, as B0 and B1 can not overlap. 540 bool HasPHIKill = BValNo->hasPHIKill(); 541 IntB.MergeValueNumberInto(BValNo, ValLR->valno); 542 if (HasPHIKill) 543 ValLR->valno->setHasPHIKill(true); 544 } 545 DEBUG({ 546 dbgs() << " result = "; 547 IntB.print(dbgs(), TRI); 548 dbgs() << "\n"; 549 }); 550 551 // If the source instruction was killing the source register before the 552 // merge, unset the isKill marker given the live range has been extended. 553 int UIdx = ValLREndInst->findRegisterUseOperandIdx(IntB.reg, true); 554 if (UIdx != -1) { 555 ValLREndInst->getOperand(UIdx).setIsKill(false); 556 } 557 558 // If the copy instruction was killing the destination register before the 559 // merge, find the last use and trim the live range. That will also add the 560 // isKill marker. 561 if (ALR->end == CopyIdx) 562 LIS->shrinkToUses(&IntA); 563 564 ++numExtends; 565 return true; 566 } 567 568 /// HasOtherReachingDefs - Return true if there are definitions of IntB 569 /// other than BValNo val# that can reach uses of AValno val# of IntA. 570 bool RegisterCoalescer::HasOtherReachingDefs(LiveInterval &IntA, 571 LiveInterval &IntB, 572 VNInfo *AValNo, 573 VNInfo *BValNo) { 574 for (LiveInterval::iterator AI = IntA.begin(), AE = IntA.end(); 575 AI != AE; ++AI) { 576 if (AI->valno != AValNo) continue; 577 LiveInterval::Ranges::iterator BI = 578 std::upper_bound(IntB.ranges.begin(), IntB.ranges.end(), AI->start); 579 if (BI != IntB.ranges.begin()) 580 --BI; 581 for (; BI != IntB.ranges.end() && AI->end >= BI->start; ++BI) { 582 if (BI->valno == BValNo) 583 continue; 584 if (BI->start <= AI->start && BI->end > AI->start) 585 return true; 586 if (BI->start > AI->start && BI->start < AI->end) 587 return true; 588 } 589 } 590 return false; 591 } 592 593 /// RemoveCopyByCommutingDef - We found a non-trivially-coalescable copy with 594 /// IntA being the source and IntB being the dest, thus this defines a value 595 /// number in IntB. If the source value number (in IntA) is defined by a 596 /// commutable instruction and its other operand is coalesced to the copy dest 597 /// register, see if we can transform the copy into a noop by commuting the 598 /// definition. For example, 599 /// 600 /// A3 = op A2 B0<kill> 601 /// ... 602 /// B1 = A3 <- this copy 603 /// ... 604 /// = op A3 <- more uses 605 /// 606 /// ==> 607 /// 608 /// B2 = op B0 A2<kill> 609 /// ... 610 /// B1 = B2 <- now an identify copy 611 /// ... 612 /// = op B2 <- more uses 613 /// 614 /// This returns true if an interval was modified. 615 /// 616 bool RegisterCoalescer::RemoveCopyByCommutingDef(const CoalescerPair &CP, 617 MachineInstr *CopyMI) { 618 // FIXME: For now, only eliminate the copy by commuting its def when the 619 // source register is a virtual register. We want to guard against cases 620 // where the copy is a back edge copy and commuting the def lengthen the 621 // live interval of the source register to the entire loop. 622 if (CP.isPhys() && CP.isFlipped()) 623 return false; 624 625 // Bail if there is no dst interval. 626 if (!LIS->hasInterval(CP.getDstReg())) 627 return false; 628 629 SlotIndex CopyIdx = LIS->getInstructionIndex(CopyMI).getDefIndex(); 630 631 LiveInterval &IntA = 632 LIS->getInterval(CP.isFlipped() ? CP.getDstReg() : CP.getSrcReg()); 633 LiveInterval &IntB = 634 LIS->getInterval(CP.isFlipped() ? CP.getSrcReg() : CP.getDstReg()); 635 636 // BValNo is a value number in B that is defined by a copy from A. 'B3' in 637 // the example above. 638 VNInfo *BValNo = IntB.getVNInfoAt(CopyIdx); 639 if (!BValNo || !BValNo->isDefByCopy()) 640 return false; 641 642 assert(BValNo->def == CopyIdx && "Copy doesn't define the value?"); 643 644 // AValNo is the value number in A that defines the copy, A3 in the example. 645 VNInfo *AValNo = IntA.getVNInfoAt(CopyIdx.getUseIndex()); 646 assert(AValNo && "COPY source not live"); 647 648 // If other defs can reach uses of this def, then it's not safe to perform 649 // the optimization. 650 if (AValNo->isPHIDef() || AValNo->isUnused() || AValNo->hasPHIKill()) 651 return false; 652 MachineInstr *DefMI = LIS->getInstructionFromIndex(AValNo->def); 653 if (!DefMI) 654 return false; 655 const MCInstrDesc &MCID = DefMI->getDesc(); 656 if (!MCID.isCommutable()) 657 return false; 658 // If DefMI is a two-address instruction then commuting it will change the 659 // destination register. 660 int DefIdx = DefMI->findRegisterDefOperandIdx(IntA.reg); 661 assert(DefIdx != -1); 662 unsigned UseOpIdx; 663 if (!DefMI->isRegTiedToUseOperand(DefIdx, &UseOpIdx)) 664 return false; 665 unsigned Op1, Op2, NewDstIdx; 666 if (!TII->findCommutedOpIndices(DefMI, Op1, Op2)) 667 return false; 668 if (Op1 == UseOpIdx) 669 NewDstIdx = Op2; 670 else if (Op2 == UseOpIdx) 671 NewDstIdx = Op1; 672 else 673 return false; 674 675 MachineOperand &NewDstMO = DefMI->getOperand(NewDstIdx); 676 unsigned NewReg = NewDstMO.getReg(); 677 if (NewReg != IntB.reg || !NewDstMO.isKill()) 678 return false; 679 680 // Make sure there are no other definitions of IntB that would reach the 681 // uses which the new definition can reach. 682 if (HasOtherReachingDefs(IntA, IntB, AValNo, BValNo)) 683 return false; 684 685 // Abort if the aliases of IntB.reg have values that are not simply the 686 // clobbers from the superreg. 687 if (TargetRegisterInfo::isPhysicalRegister(IntB.reg)) 688 for (const unsigned *AS = TRI->getAliasSet(IntB.reg); *AS; ++AS) 689 if (LIS->hasInterval(*AS) && 690 HasOtherReachingDefs(IntA, LIS->getInterval(*AS), AValNo, 0)) 691 return false; 692 693 // If some of the uses of IntA.reg is already coalesced away, return false. 694 // It's not possible to determine whether it's safe to perform the coalescing. 695 for (MachineRegisterInfo::use_nodbg_iterator UI = 696 MRI->use_nodbg_begin(IntA.reg), 697 UE = MRI->use_nodbg_end(); UI != UE; ++UI) { 698 MachineInstr *UseMI = &*UI; 699 SlotIndex UseIdx = LIS->getInstructionIndex(UseMI); 700 LiveInterval::iterator ULR = IntA.FindLiveRangeContaining(UseIdx); 701 if (ULR == IntA.end()) 702 continue; 703 if (ULR->valno == AValNo && JoinedCopies.count(UseMI)) 704 return false; 705 } 706 707 DEBUG(dbgs() << "\tRemoveCopyByCommutingDef: " << AValNo->def << '\t' 708 << *DefMI); 709 710 // At this point we have decided that it is legal to do this 711 // transformation. Start by commuting the instruction. 712 MachineBasicBlock *MBB = DefMI->getParent(); 713 MachineInstr *NewMI = TII->commuteInstruction(DefMI); 714 if (!NewMI) 715 return false; 716 if (TargetRegisterInfo::isVirtualRegister(IntA.reg) && 717 TargetRegisterInfo::isVirtualRegister(IntB.reg) && 718 !MRI->constrainRegClass(IntB.reg, MRI->getRegClass(IntA.reg))) 719 return false; 720 if (NewMI != DefMI) { 721 LIS->ReplaceMachineInstrInMaps(DefMI, NewMI); 722 MBB->insert(DefMI, NewMI); 723 MBB->erase(DefMI); 724 } 725 unsigned OpIdx = NewMI->findRegisterUseOperandIdx(IntA.reg, false); 726 NewMI->getOperand(OpIdx).setIsKill(); 727 728 // If ALR and BLR overlaps and end of BLR extends beyond end of ALR, e.g. 729 // A = or A, B 730 // ... 731 // B = A 732 // ... 733 // C = A<kill> 734 // ... 735 // = B 736 737 // Update uses of IntA of the specific Val# with IntB. 738 for (MachineRegisterInfo::use_iterator UI = MRI->use_begin(IntA.reg), 739 UE = MRI->use_end(); UI != UE;) { 740 MachineOperand &UseMO = UI.getOperand(); 741 MachineInstr *UseMI = &*UI; 742 ++UI; 743 if (JoinedCopies.count(UseMI)) 744 continue; 745 if (UseMI->isDebugValue()) { 746 // FIXME These don't have an instruction index. Not clear we have enough 747 // info to decide whether to do this replacement or not. For now do it. 748 UseMO.setReg(NewReg); 749 continue; 750 } 751 SlotIndex UseIdx = LIS->getInstructionIndex(UseMI).getUseIndex(); 752 LiveInterval::iterator ULR = IntA.FindLiveRangeContaining(UseIdx); 753 if (ULR == IntA.end() || ULR->valno != AValNo) 754 continue; 755 if (TargetRegisterInfo::isPhysicalRegister(NewReg)) 756 UseMO.substPhysReg(NewReg, *TRI); 757 else 758 UseMO.setReg(NewReg); 759 if (UseMI == CopyMI) 760 continue; 761 if (!UseMI->isCopy()) 762 continue; 763 if (UseMI->getOperand(0).getReg() != IntB.reg || 764 UseMI->getOperand(0).getSubReg()) 765 continue; 766 767 // This copy will become a noop. If it's defining a new val#, merge it into 768 // BValNo. 769 SlotIndex DefIdx = UseIdx.getDefIndex(); 770 VNInfo *DVNI = IntB.getVNInfoAt(DefIdx); 771 if (!DVNI) 772 continue; 773 DEBUG(dbgs() << "\t\tnoop: " << DefIdx << '\t' << *UseMI); 774 assert(DVNI->def == DefIdx); 775 BValNo = IntB.MergeValueNumberInto(BValNo, DVNI); 776 markAsJoined(UseMI); 777 } 778 779 // Extend BValNo by merging in IntA live ranges of AValNo. Val# definition 780 // is updated. 781 VNInfo *ValNo = BValNo; 782 ValNo->def = AValNo->def; 783 ValNo->setCopy(0); 784 for (LiveInterval::iterator AI = IntA.begin(), AE = IntA.end(); 785 AI != AE; ++AI) { 786 if (AI->valno != AValNo) continue; 787 IntB.addRange(LiveRange(AI->start, AI->end, ValNo)); 788 } 789 DEBUG(dbgs() << "\t\textended: " << IntB << '\n'); 790 791 IntA.removeValNo(AValNo); 792 DEBUG(dbgs() << "\t\ttrimmed: " << IntA << '\n'); 793 ++numCommutes; 794 return true; 795 } 796 797 /// ReMaterializeTrivialDef - If the source of a copy is defined by a trivial 798 /// computation, replace the copy by rematerialize the definition. 799 bool RegisterCoalescer::ReMaterializeTrivialDef(LiveInterval &SrcInt, 800 bool preserveSrcInt, 801 unsigned DstReg, 802 unsigned DstSubIdx, 803 MachineInstr *CopyMI) { 804 SlotIndex CopyIdx = LIS->getInstructionIndex(CopyMI).getUseIndex(); 805 LiveInterval::iterator SrcLR = SrcInt.FindLiveRangeContaining(CopyIdx); 806 assert(SrcLR != SrcInt.end() && "Live range not found!"); 807 VNInfo *ValNo = SrcLR->valno; 808 if (ValNo->isPHIDef() || ValNo->isUnused()) 809 return false; 810 MachineInstr *DefMI = LIS->getInstructionFromIndex(ValNo->def); 811 if (!DefMI) 812 return false; 813 assert(DefMI && "Defining instruction disappeared"); 814 const MCInstrDesc &MCID = DefMI->getDesc(); 815 if (!MCID.isAsCheapAsAMove()) 816 return false; 817 if (!TII->isTriviallyReMaterializable(DefMI, AA)) 818 return false; 819 bool SawStore = false; 820 if (!DefMI->isSafeToMove(TII, AA, SawStore)) 821 return false; 822 if (MCID.getNumDefs() != 1) 823 return false; 824 if (!DefMI->isImplicitDef()) { 825 // Make sure the copy destination register class fits the instruction 826 // definition register class. The mismatch can happen as a result of earlier 827 // extract_subreg, insert_subreg, subreg_to_reg coalescing. 828 const TargetRegisterClass *RC = TII->getRegClass(MCID, 0, TRI); 829 if (TargetRegisterInfo::isVirtualRegister(DstReg)) { 830 if (MRI->getRegClass(DstReg) != RC) 831 return false; 832 } else if (!RC->contains(DstReg)) 833 return false; 834 } 835 836 // If destination register has a sub-register index on it, make sure it 837 // matches the instruction register class. 838 if (DstSubIdx) { 839 const MCInstrDesc &MCID = DefMI->getDesc(); 840 if (MCID.getNumDefs() != 1) 841 return false; 842 const TargetRegisterClass *DstRC = MRI->getRegClass(DstReg); 843 const TargetRegisterClass *DstSubRC = 844 DstRC->getSubRegisterRegClass(DstSubIdx); 845 const TargetRegisterClass *DefRC = TII->getRegClass(MCID, 0, TRI); 846 if (DefRC == DstRC) 847 DstSubIdx = 0; 848 else if (DefRC != DstSubRC) 849 return false; 850 } 851 852 RemoveCopyFlag(DstReg, CopyMI); 853 854 MachineBasicBlock *MBB = CopyMI->getParent(); 855 MachineBasicBlock::iterator MII = 856 llvm::next(MachineBasicBlock::iterator(CopyMI)); 857 TII->reMaterialize(*MBB, MII, DstReg, DstSubIdx, DefMI, *TRI); 858 MachineInstr *NewMI = prior(MII); 859 860 // CopyMI may have implicit operands, transfer them over to the newly 861 // rematerialized instruction. And update implicit def interval valnos. 862 for (unsigned i = CopyMI->getDesc().getNumOperands(), 863 e = CopyMI->getNumOperands(); i != e; ++i) { 864 MachineOperand &MO = CopyMI->getOperand(i); 865 if (MO.isReg() && MO.isImplicit()) 866 NewMI->addOperand(MO); 867 if (MO.isDef()) 868 RemoveCopyFlag(MO.getReg(), CopyMI); 869 } 870 871 NewMI->copyImplicitOps(CopyMI); 872 LIS->ReplaceMachineInstrInMaps(CopyMI, NewMI); 873 CopyMI->eraseFromParent(); 874 ReMatCopies.insert(CopyMI); 875 ReMatDefs.insert(DefMI); 876 DEBUG(dbgs() << "Remat: " << *NewMI); 877 ++NumReMats; 878 879 // The source interval can become smaller because we removed a use. 880 if (preserveSrcInt) 881 LIS->shrinkToUses(&SrcInt); 882 883 return true; 884 } 885 886 /// eliminateUndefCopy - ProcessImpicitDefs may leave some copies of <undef> 887 /// values, it only removes local variables. When we have a copy like: 888 /// 889 /// %vreg1 = COPY %vreg2<undef> 890 /// 891 /// We delete the copy and remove the corresponding value number from %vreg1. 892 /// Any uses of that value number are marked as <undef>. 893 bool RegisterCoalescer::eliminateUndefCopy(MachineInstr *CopyMI, 894 const CoalescerPair &CP) { 895 SlotIndex Idx = LIS->getInstructionIndex(CopyMI); 896 LiveInterval *SrcInt = &LIS->getInterval(CP.getSrcReg()); 897 if (SrcInt->liveAt(Idx)) 898 return false; 899 LiveInterval *DstInt = &LIS->getInterval(CP.getDstReg()); 900 if (DstInt->liveAt(Idx)) 901 return false; 902 903 // No intervals are live-in to CopyMI - it is undef. 904 if (CP.isFlipped()) 905 DstInt = SrcInt; 906 SrcInt = 0; 907 908 VNInfo *DeadVNI = DstInt->getVNInfoAt(Idx.getDefIndex()); 909 assert(DeadVNI && "No value defined in DstInt"); 910 DstInt->removeValNo(DeadVNI); 911 912 // Find new undef uses. 913 for (MachineRegisterInfo::reg_nodbg_iterator 914 I = MRI->reg_nodbg_begin(DstInt->reg), E = MRI->reg_nodbg_end(); 915 I != E; ++I) { 916 MachineOperand &MO = I.getOperand(); 917 if (MO.isDef() || MO.isUndef()) 918 continue; 919 MachineInstr *MI = MO.getParent(); 920 SlotIndex Idx = LIS->getInstructionIndex(MI); 921 if (DstInt->liveAt(Idx)) 922 continue; 923 MO.setIsUndef(true); 924 DEBUG(dbgs() << "\tnew undef: " << Idx << '\t' << *MI); 925 } 926 return true; 927 } 928 929 /// UpdateRegDefsUses - Replace all defs and uses of SrcReg to DstReg and 930 /// update the subregister number if it is not zero. If DstReg is a 931 /// physical register and the existing subregister number of the def / use 932 /// being updated is not zero, make sure to set it to the correct physical 933 /// subregister. 934 void 935 RegisterCoalescer::UpdateRegDefsUses(const CoalescerPair &CP) { 936 bool DstIsPhys = CP.isPhys(); 937 unsigned SrcReg = CP.getSrcReg(); 938 unsigned DstReg = CP.getDstReg(); 939 unsigned SubIdx = CP.getSubIdx(); 940 941 // Update LiveDebugVariables. 942 LDV->renameRegister(SrcReg, DstReg, SubIdx); 943 944 for (MachineRegisterInfo::reg_iterator I = MRI->reg_begin(SrcReg); 945 MachineInstr *UseMI = I.skipInstruction();) { 946 // A PhysReg copy that won't be coalesced can perhaps be rematerialized 947 // instead. 948 if (DstIsPhys) { 949 if (UseMI->isFullCopy() && 950 UseMI->getOperand(1).getReg() == SrcReg && 951 UseMI->getOperand(0).getReg() != SrcReg && 952 UseMI->getOperand(0).getReg() != DstReg && 953 !JoinedCopies.count(UseMI) && 954 ReMaterializeTrivialDef(LIS->getInterval(SrcReg), false, 955 UseMI->getOperand(0).getReg(), 0, UseMI)) 956 continue; 957 } 958 959 SmallVector<unsigned,8> Ops; 960 bool Reads, Writes; 961 tie(Reads, Writes) = UseMI->readsWritesVirtualRegister(SrcReg, &Ops); 962 bool Kills = false, Deads = false; 963 964 // Replace SrcReg with DstReg in all UseMI operands. 965 for (unsigned i = 0, e = Ops.size(); i != e; ++i) { 966 MachineOperand &MO = UseMI->getOperand(Ops[i]); 967 Kills |= MO.isKill(); 968 Deads |= MO.isDead(); 969 970 if (DstIsPhys) 971 MO.substPhysReg(DstReg, *TRI); 972 else 973 MO.substVirtReg(DstReg, SubIdx, *TRI); 974 } 975 976 // This instruction is a copy that will be removed. 977 if (JoinedCopies.count(UseMI)) 978 continue; 979 980 if (SubIdx) { 981 // If UseMI was a simple SrcReg def, make sure we didn't turn it into a 982 // read-modify-write of DstReg. 983 if (Deads) 984 UseMI->addRegisterDead(DstReg, TRI); 985 else if (!Reads && Writes) 986 UseMI->addRegisterDefined(DstReg, TRI); 987 988 // Kill flags apply to the whole physical register. 989 if (DstIsPhys && Kills) 990 UseMI->addRegisterKilled(DstReg, TRI); 991 } 992 993 DEBUG({ 994 dbgs() << "\t\tupdated: "; 995 if (!UseMI->isDebugValue()) 996 dbgs() << LIS->getInstructionIndex(UseMI) << "\t"; 997 dbgs() << *UseMI; 998 }); 999 } 1000 } 1001 1002 /// removeIntervalIfEmpty - Check if the live interval of a physical register 1003 /// is empty, if so remove it and also remove the empty intervals of its 1004 /// sub-registers. Return true if live interval is removed. 1005 static bool removeIntervalIfEmpty(LiveInterval &li, LiveIntervals *LIS, 1006 const TargetRegisterInfo *TRI) { 1007 if (li.empty()) { 1008 if (TargetRegisterInfo::isPhysicalRegister(li.reg)) 1009 for (const unsigned* SR = TRI->getSubRegisters(li.reg); *SR; ++SR) { 1010 if (!LIS->hasInterval(*SR)) 1011 continue; 1012 LiveInterval &sli = LIS->getInterval(*SR); 1013 if (sli.empty()) 1014 LIS->removeInterval(*SR); 1015 } 1016 LIS->removeInterval(li.reg); 1017 return true; 1018 } 1019 return false; 1020 } 1021 1022 /// RemoveDeadDef - If a def of a live interval is now determined dead, remove 1023 /// the val# it defines. If the live interval becomes empty, remove it as well. 1024 bool RegisterCoalescer::RemoveDeadDef(LiveInterval &li, 1025 MachineInstr *DefMI) { 1026 SlotIndex DefIdx = LIS->getInstructionIndex(DefMI).getDefIndex(); 1027 LiveInterval::iterator MLR = li.FindLiveRangeContaining(DefIdx); 1028 if (DefIdx != MLR->valno->def) 1029 return false; 1030 li.removeValNo(MLR->valno); 1031 return removeIntervalIfEmpty(li, LIS, TRI); 1032 } 1033 1034 void RegisterCoalescer::RemoveCopyFlag(unsigned DstReg, 1035 const MachineInstr *CopyMI) { 1036 SlotIndex DefIdx = LIS->getInstructionIndex(CopyMI).getDefIndex(); 1037 if (LIS->hasInterval(DstReg)) { 1038 LiveInterval &LI = LIS->getInterval(DstReg); 1039 if (const LiveRange *LR = LI.getLiveRangeContaining(DefIdx)) 1040 if (LR->valno->def == DefIdx) 1041 LR->valno->setCopy(0); 1042 } 1043 if (!TargetRegisterInfo::isPhysicalRegister(DstReg)) 1044 return; 1045 for (const unsigned* AS = TRI->getAliasSet(DstReg); *AS; ++AS) { 1046 if (!LIS->hasInterval(*AS)) 1047 continue; 1048 LiveInterval &LI = LIS->getInterval(*AS); 1049 if (const LiveRange *LR = LI.getLiveRangeContaining(DefIdx)) 1050 if (LR->valno->def == DefIdx) 1051 LR->valno->setCopy(0); 1052 } 1053 } 1054 1055 /// shouldJoinPhys - Return true if a copy involving a physreg should be joined. 1056 /// We need to be careful about coalescing a source physical register with a 1057 /// virtual register. Once the coalescing is done, it cannot be broken and these 1058 /// are not spillable! If the destination interval uses are far away, think 1059 /// twice about coalescing them! 1060 bool RegisterCoalescer::shouldJoinPhys(CoalescerPair &CP) { 1061 bool Allocatable = LIS->isAllocatable(CP.getDstReg()); 1062 LiveInterval &JoinVInt = LIS->getInterval(CP.getSrcReg()); 1063 1064 /// Always join simple intervals that are defined by a single copy from a 1065 /// reserved register. This doesn't increase register pressure, so it is 1066 /// always beneficial. 1067 if (!Allocatable && CP.isFlipped() && JoinVInt.containsOneValue()) 1068 return true; 1069 1070 if (!EnablePhysicalJoin) { 1071 DEBUG(dbgs() << "\tPhysreg joins disabled.\n"); 1072 return false; 1073 } 1074 1075 // Only coalesce to allocatable physreg, we don't want to risk modifying 1076 // reserved registers. 1077 if (!Allocatable) { 1078 DEBUG(dbgs() << "\tRegister is an unallocatable physreg.\n"); 1079 return false; // Not coalescable. 1080 } 1081 1082 // Don't join with physregs that have a ridiculous number of live 1083 // ranges. The data structure performance is really bad when that 1084 // happens. 1085 if (LIS->hasInterval(CP.getDstReg()) && 1086 LIS->getInterval(CP.getDstReg()).ranges.size() > 1000) { 1087 ++numAborts; 1088 DEBUG(dbgs() 1089 << "\tPhysical register live interval too complicated, abort!\n"); 1090 return false; 1091 } 1092 1093 // FIXME: Why are we skipping this test for partial copies? 1094 // CodeGen/X86/phys_subreg_coalesce-3.ll needs it. 1095 if (!CP.isPartial()) { 1096 const TargetRegisterClass *RC = MRI->getRegClass(CP.getSrcReg()); 1097 unsigned Threshold = RegClassInfo.getNumAllocatableRegs(RC) * 2; 1098 unsigned Length = LIS->getApproximateInstructionCount(JoinVInt); 1099 if (Length > Threshold) { 1100 ++numAborts; 1101 DEBUG(dbgs() << "\tMay tie down a physical register, abort!\n"); 1102 return false; 1103 } 1104 } 1105 return true; 1106 } 1107 1108 /// isWinToJoinCrossClass - Return true if it's profitable to coalesce 1109 /// two virtual registers from different register classes. 1110 bool 1111 RegisterCoalescer::isWinToJoinCrossClass(unsigned SrcReg, 1112 unsigned DstReg, 1113 const TargetRegisterClass *SrcRC, 1114 const TargetRegisterClass *DstRC, 1115 const TargetRegisterClass *NewRC) { 1116 unsigned NewRCCount = RegClassInfo.getNumAllocatableRegs(NewRC); 1117 // This heuristics is good enough in practice, but it's obviously not *right*. 1118 // 4 is a magic number that works well enough for x86, ARM, etc. It filter 1119 // out all but the most restrictive register classes. 1120 if (NewRCCount > 4 || 1121 // Early exit if the function is fairly small, coalesce aggressively if 1122 // that's the case. For really special register classes with 3 or 1123 // fewer registers, be a bit more careful. 1124 (LIS->getFuncInstructionCount() / NewRCCount) < 8) 1125 return true; 1126 LiveInterval &SrcInt = LIS->getInterval(SrcReg); 1127 LiveInterval &DstInt = LIS->getInterval(DstReg); 1128 unsigned SrcSize = LIS->getApproximateInstructionCount(SrcInt); 1129 unsigned DstSize = LIS->getApproximateInstructionCount(DstInt); 1130 1131 // Coalesce aggressively if the intervals are small compared to the number of 1132 // registers in the new class. The number 4 is fairly arbitrary, chosen to be 1133 // less aggressive than the 8 used for the whole function size. 1134 const unsigned ThresSize = 4 * NewRCCount; 1135 if (SrcSize <= ThresSize && DstSize <= ThresSize) 1136 return true; 1137 1138 // Estimate *register use density*. If it doubles or more, abort. 1139 unsigned SrcUses = std::distance(MRI->use_nodbg_begin(SrcReg), 1140 MRI->use_nodbg_end()); 1141 unsigned DstUses = std::distance(MRI->use_nodbg_begin(DstReg), 1142 MRI->use_nodbg_end()); 1143 unsigned NewUses = SrcUses + DstUses; 1144 unsigned NewSize = SrcSize + DstSize; 1145 if (SrcRC != NewRC && SrcSize > ThresSize) { 1146 unsigned SrcRCCount = RegClassInfo.getNumAllocatableRegs(SrcRC); 1147 if (NewUses*SrcSize*SrcRCCount > 2*SrcUses*NewSize*NewRCCount) 1148 return false; 1149 } 1150 if (DstRC != NewRC && DstSize > ThresSize) { 1151 unsigned DstRCCount = RegClassInfo.getNumAllocatableRegs(DstRC); 1152 if (NewUses*DstSize*DstRCCount > 2*DstUses*NewSize*NewRCCount) 1153 return false; 1154 } 1155 return true; 1156 } 1157 1158 1159 /// JoinCopy - Attempt to join intervals corresponding to SrcReg/DstReg, 1160 /// which are the src/dst of the copy instruction CopyMI. This returns true 1161 /// if the copy was successfully coalesced away. If it is not currently 1162 /// possible to coalesce this interval, but it may be possible if other 1163 /// things get coalesced, then it returns true by reference in 'Again'. 1164 bool RegisterCoalescer::JoinCopy(MachineInstr *CopyMI, bool &Again) { 1165 1166 Again = false; 1167 if (JoinedCopies.count(CopyMI) || ReMatCopies.count(CopyMI)) 1168 return false; // Already done. 1169 1170 DEBUG(dbgs() << LIS->getInstructionIndex(CopyMI) << '\t' << *CopyMI); 1171 1172 CoalescerPair CP(*TII, *TRI); 1173 if (!CP.setRegisters(CopyMI)) { 1174 DEBUG(dbgs() << "\tNot coalescable.\n"); 1175 return false; 1176 } 1177 1178 // If they are already joined we continue. 1179 if (CP.getSrcReg() == CP.getDstReg()) { 1180 markAsJoined(CopyMI); 1181 DEBUG(dbgs() << "\tCopy already coalesced.\n"); 1182 return false; // Not coalescable. 1183 } 1184 1185 // Eliminate undefs. 1186 if (!CP.isPhys() && eliminateUndefCopy(CopyMI, CP)) { 1187 markAsJoined(CopyMI); 1188 DEBUG(dbgs() << "\tEliminated copy of <undef> value.\n"); 1189 return false; // Not coalescable. 1190 } 1191 1192 DEBUG(dbgs() << "\tConsidering merging " << PrintReg(CP.getSrcReg(), TRI) 1193 << " with " << PrintReg(CP.getDstReg(), TRI, CP.getSubIdx()) 1194 << "\n"); 1195 1196 // Enforce policies. 1197 if (CP.isPhys()) { 1198 if (!shouldJoinPhys(CP)) { 1199 // Before giving up coalescing, if definition of source is defined by 1200 // trivial computation, try rematerializing it. 1201 if (!CP.isFlipped() && 1202 ReMaterializeTrivialDef(LIS->getInterval(CP.getSrcReg()), true, 1203 CP.getDstReg(), 0, CopyMI)) 1204 return true; 1205 return false; 1206 } 1207 } else { 1208 // Avoid constraining virtual register regclass too much. 1209 if (CP.isCrossClass()) { 1210 DEBUG(dbgs() << "\tCross-class to " << CP.getNewRC()->getName() << ".\n"); 1211 if (DisableCrossClassJoin) { 1212 DEBUG(dbgs() << "\tCross-class joins disabled.\n"); 1213 return false; 1214 } 1215 if (!isWinToJoinCrossClass(CP.getSrcReg(), CP.getDstReg(), 1216 MRI->getRegClass(CP.getSrcReg()), 1217 MRI->getRegClass(CP.getDstReg()), 1218 CP.getNewRC())) { 1219 DEBUG(dbgs() << "\tAvoid coalescing to constrained register class.\n"); 1220 Again = true; // May be possible to coalesce later. 1221 return false; 1222 } 1223 } 1224 1225 // When possible, let DstReg be the larger interval. 1226 if (!CP.getSubIdx() && LIS->getInterval(CP.getSrcReg()).ranges.size() > 1227 LIS->getInterval(CP.getDstReg()).ranges.size()) 1228 CP.flip(); 1229 } 1230 1231 // Okay, attempt to join these two intervals. On failure, this returns false. 1232 // Otherwise, if one of the intervals being joined is a physreg, this method 1233 // always canonicalizes DstInt to be it. The output "SrcInt" will not have 1234 // been modified, so we can use this information below to update aliases. 1235 if (!JoinIntervals(CP)) { 1236 // Coalescing failed. 1237 1238 // If definition of source is defined by trivial computation, try 1239 // rematerializing it. 1240 if (!CP.isFlipped() && 1241 ReMaterializeTrivialDef(LIS->getInterval(CP.getSrcReg()), true, 1242 CP.getDstReg(), 0, CopyMI)) 1243 return true; 1244 1245 // If we can eliminate the copy without merging the live ranges, do so now. 1246 if (!CP.isPartial()) { 1247 if (AdjustCopiesBackFrom(CP, CopyMI) || 1248 RemoveCopyByCommutingDef(CP, CopyMI)) { 1249 markAsJoined(CopyMI); 1250 DEBUG(dbgs() << "\tTrivial!\n"); 1251 return true; 1252 } 1253 } 1254 1255 // Otherwise, we are unable to join the intervals. 1256 DEBUG(dbgs() << "\tInterference!\n"); 1257 Again = true; // May be possible to coalesce later. 1258 return false; 1259 } 1260 1261 // Coalescing to a virtual register that is of a sub-register class of the 1262 // other. Make sure the resulting register is set to the right register class. 1263 if (CP.isCrossClass()) { 1264 ++numCrossRCs; 1265 MRI->setRegClass(CP.getDstReg(), CP.getNewRC()); 1266 } 1267 1268 // Remember to delete the copy instruction. 1269 markAsJoined(CopyMI); 1270 1271 UpdateRegDefsUses(CP); 1272 1273 // If we have extended the live range of a physical register, make sure we 1274 // update live-in lists as well. 1275 if (CP.isPhys()) { 1276 SmallVector<MachineBasicBlock*, 16> BlockSeq; 1277 // JoinIntervals invalidates the VNInfos in SrcInt, but we only need the 1278 // ranges for this, and they are preserved. 1279 LiveInterval &SrcInt = LIS->getInterval(CP.getSrcReg()); 1280 for (LiveInterval::const_iterator I = SrcInt.begin(), E = SrcInt.end(); 1281 I != E; ++I ) { 1282 LIS->findLiveInMBBs(I->start, I->end, BlockSeq); 1283 for (unsigned idx = 0, size = BlockSeq.size(); idx != size; ++idx) { 1284 MachineBasicBlock &block = *BlockSeq[idx]; 1285 if (!block.isLiveIn(CP.getDstReg())) 1286 block.addLiveIn(CP.getDstReg()); 1287 } 1288 BlockSeq.clear(); 1289 } 1290 } 1291 1292 // SrcReg is guarateed to be the register whose live interval that is 1293 // being merged. 1294 LIS->removeInterval(CP.getSrcReg()); 1295 1296 // Update regalloc hint. 1297 TRI->UpdateRegAllocHint(CP.getSrcReg(), CP.getDstReg(), *MF); 1298 1299 DEBUG({ 1300 LiveInterval &DstInt = LIS->getInterval(CP.getDstReg()); 1301 dbgs() << "\tJoined. Result = "; 1302 DstInt.print(dbgs(), TRI); 1303 dbgs() << "\n"; 1304 }); 1305 1306 ++numJoins; 1307 return true; 1308 } 1309 1310 /// ComputeUltimateVN - Assuming we are going to join two live intervals, 1311 /// compute what the resultant value numbers for each value in the input two 1312 /// ranges will be. This is complicated by copies between the two which can 1313 /// and will commonly cause multiple value numbers to be merged into one. 1314 /// 1315 /// VN is the value number that we're trying to resolve. InstDefiningValue 1316 /// keeps track of the new InstDefiningValue assignment for the result 1317 /// LiveInterval. ThisFromOther/OtherFromThis are sets that keep track of 1318 /// whether a value in this or other is a copy from the opposite set. 1319 /// ThisValNoAssignments/OtherValNoAssignments keep track of value #'s that have 1320 /// already been assigned. 1321 /// 1322 /// ThisFromOther[x] - If x is defined as a copy from the other interval, this 1323 /// contains the value number the copy is from. 1324 /// 1325 static unsigned ComputeUltimateVN(VNInfo *VNI, 1326 SmallVector<VNInfo*, 16> &NewVNInfo, 1327 DenseMap<VNInfo*, VNInfo*> &ThisFromOther, 1328 DenseMap<VNInfo*, VNInfo*> &OtherFromThis, 1329 SmallVector<int, 16> &ThisValNoAssignments, 1330 SmallVector<int, 16> &OtherValNoAssignments) { 1331 unsigned VN = VNI->id; 1332 1333 // If the VN has already been computed, just return it. 1334 if (ThisValNoAssignments[VN] >= 0) 1335 return ThisValNoAssignments[VN]; 1336 assert(ThisValNoAssignments[VN] != -2 && "Cyclic value numbers"); 1337 1338 // If this val is not a copy from the other val, then it must be a new value 1339 // number in the destination. 1340 DenseMap<VNInfo*, VNInfo*>::iterator I = ThisFromOther.find(VNI); 1341 if (I == ThisFromOther.end()) { 1342 NewVNInfo.push_back(VNI); 1343 return ThisValNoAssignments[VN] = NewVNInfo.size()-1; 1344 } 1345 VNInfo *OtherValNo = I->second; 1346 1347 // Otherwise, this *is* a copy from the RHS. If the other side has already 1348 // been computed, return it. 1349 if (OtherValNoAssignments[OtherValNo->id] >= 0) 1350 return ThisValNoAssignments[VN] = OtherValNoAssignments[OtherValNo->id]; 1351 1352 // Mark this value number as currently being computed, then ask what the 1353 // ultimate value # of the other value is. 1354 ThisValNoAssignments[VN] = -2; 1355 unsigned UltimateVN = 1356 ComputeUltimateVN(OtherValNo, NewVNInfo, OtherFromThis, ThisFromOther, 1357 OtherValNoAssignments, ThisValNoAssignments); 1358 return ThisValNoAssignments[VN] = UltimateVN; 1359 } 1360 1361 1362 // Find out if we have something like 1363 // A = X 1364 // B = X 1365 // if so, we can pretend this is actually 1366 // A = X 1367 // B = A 1368 // which allows us to coalesce A and B. 1369 // VNI is the definition of B. LR is the life range of A that includes 1370 // the slot just before B. If we return true, we add "B = X" to DupCopies. 1371 // This implies that A dominates B. 1372 static bool RegistersDefinedFromSameValue(LiveIntervals &li, 1373 const TargetRegisterInfo &tri, 1374 CoalescerPair &CP, 1375 VNInfo *VNI, 1376 LiveRange *LR, 1377 SmallVector<MachineInstr*, 8> &DupCopies) { 1378 // FIXME: This is very conservative. For example, we don't handle 1379 // physical registers. 1380 1381 MachineInstr *MI = VNI->getCopy(); 1382 1383 if (!MI->isFullCopy() || CP.isPartial() || CP.isPhys()) 1384 return false; 1385 1386 unsigned Dst = MI->getOperand(0).getReg(); 1387 unsigned Src = MI->getOperand(1).getReg(); 1388 1389 if (!TargetRegisterInfo::isVirtualRegister(Src) || 1390 !TargetRegisterInfo::isVirtualRegister(Dst)) 1391 return false; 1392 1393 unsigned A = CP.getDstReg(); 1394 unsigned B = CP.getSrcReg(); 1395 1396 if (B == Dst) 1397 std::swap(A, B); 1398 assert(Dst == A); 1399 1400 VNInfo *Other = LR->valno; 1401 if (!Other->isDefByCopy()) 1402 return false; 1403 const MachineInstr *OtherMI = Other->getCopy(); 1404 1405 if (!OtherMI->isFullCopy()) 1406 return false; 1407 1408 unsigned OtherDst = OtherMI->getOperand(0).getReg(); 1409 unsigned OtherSrc = OtherMI->getOperand(1).getReg(); 1410 1411 if (!TargetRegisterInfo::isVirtualRegister(OtherSrc) || 1412 !TargetRegisterInfo::isVirtualRegister(OtherDst)) 1413 return false; 1414 1415 assert(OtherDst == B); 1416 1417 if (Src != OtherSrc) 1418 return false; 1419 1420 // If the copies use two different value numbers of X, we cannot merge 1421 // A and B. 1422 LiveInterval &SrcInt = li.getInterval(Src); 1423 // getVNInfoBefore returns NULL for undef copies. In this case, the 1424 // optimization is still safe. 1425 if (SrcInt.getVNInfoBefore(Other->def) != SrcInt.getVNInfoBefore(VNI->def)) 1426 return false; 1427 1428 DupCopies.push_back(MI); 1429 1430 return true; 1431 } 1432 1433 /// JoinIntervals - Attempt to join these two intervals. On failure, this 1434 /// returns false. 1435 bool RegisterCoalescer::JoinIntervals(CoalescerPair &CP) { 1436 LiveInterval &RHS = LIS->getInterval(CP.getSrcReg()); 1437 DEBUG({ dbgs() << "\t\tRHS = "; RHS.print(dbgs(), TRI); dbgs() << "\n"; }); 1438 1439 // If a live interval is a physical register, check for interference with any 1440 // aliases. The interference check implemented here is a bit more conservative 1441 // than the full interfeence check below. We allow overlapping live ranges 1442 // only when one is a copy of the other. 1443 if (CP.isPhys()) { 1444 for (const unsigned *AS = TRI->getAliasSet(CP.getDstReg()); *AS; ++AS){ 1445 if (!LIS->hasInterval(*AS)) 1446 continue; 1447 const LiveInterval &LHS = LIS->getInterval(*AS); 1448 LiveInterval::const_iterator LI = LHS.begin(); 1449 for (LiveInterval::const_iterator RI = RHS.begin(), RE = RHS.end(); 1450 RI != RE; ++RI) { 1451 LI = std::lower_bound(LI, LHS.end(), RI->start); 1452 // Does LHS have an overlapping live range starting before RI? 1453 if ((LI != LHS.begin() && LI[-1].end > RI->start) && 1454 (RI->start != RI->valno->def || 1455 !CP.isCoalescable(LIS->getInstructionFromIndex(RI->start)))) { 1456 DEBUG({ 1457 dbgs() << "\t\tInterference from alias: "; 1458 LHS.print(dbgs(), TRI); 1459 dbgs() << "\n\t\tOverlap at " << RI->start << " and no copy.\n"; 1460 }); 1461 return false; 1462 } 1463 1464 // Check that LHS ranges beginning in this range are copies. 1465 for (; LI != LHS.end() && LI->start < RI->end; ++LI) { 1466 if (LI->start != LI->valno->def || 1467 !CP.isCoalescable(LIS->getInstructionFromIndex(LI->start))) { 1468 DEBUG({ 1469 dbgs() << "\t\tInterference from alias: "; 1470 LHS.print(dbgs(), TRI); 1471 dbgs() << "\n\t\tDef at " << LI->start << " is not a copy.\n"; 1472 }); 1473 return false; 1474 } 1475 } 1476 } 1477 } 1478 } 1479 1480 // Compute the final value assignment, assuming that the live ranges can be 1481 // coalesced. 1482 SmallVector<int, 16> LHSValNoAssignments; 1483 SmallVector<int, 16> RHSValNoAssignments; 1484 DenseMap<VNInfo*, VNInfo*> LHSValsDefinedFromRHS; 1485 DenseMap<VNInfo*, VNInfo*> RHSValsDefinedFromLHS; 1486 SmallVector<VNInfo*, 16> NewVNInfo; 1487 1488 SmallVector<MachineInstr*, 8> DupCopies; 1489 1490 LiveInterval &LHS = LIS->getOrCreateInterval(CP.getDstReg()); 1491 DEBUG({ dbgs() << "\t\tLHS = "; LHS.print(dbgs(), TRI); dbgs() << "\n"; }); 1492 1493 // Loop over the value numbers of the LHS, seeing if any are defined from 1494 // the RHS. 1495 for (LiveInterval::vni_iterator i = LHS.vni_begin(), e = LHS.vni_end(); 1496 i != e; ++i) { 1497 VNInfo *VNI = *i; 1498 if (VNI->isUnused() || !VNI->isDefByCopy()) // Src not defined by a copy? 1499 continue; 1500 1501 // Never join with a register that has EarlyClobber redefs. 1502 if (VNI->hasRedefByEC()) 1503 return false; 1504 1505 // Figure out the value # from the RHS. 1506 LiveRange *lr = RHS.getLiveRangeContaining(VNI->def.getPrevSlot()); 1507 // The copy could be to an aliased physreg. 1508 if (!lr) continue; 1509 1510 // DstReg is known to be a register in the LHS interval. If the src is 1511 // from the RHS interval, we can use its value #. 1512 MachineInstr *MI = VNI->getCopy(); 1513 if (!CP.isCoalescable(MI) && 1514 !RegistersDefinedFromSameValue(*LIS, *TRI, CP, VNI, lr, DupCopies)) 1515 continue; 1516 1517 LHSValsDefinedFromRHS[VNI] = lr->valno; 1518 } 1519 1520 // Loop over the value numbers of the RHS, seeing if any are defined from 1521 // the LHS. 1522 for (LiveInterval::vni_iterator i = RHS.vni_begin(), e = RHS.vni_end(); 1523 i != e; ++i) { 1524 VNInfo *VNI = *i; 1525 if (VNI->isUnused() || !VNI->isDefByCopy()) // Src not defined by a copy? 1526 continue; 1527 1528 // Never join with a register that has EarlyClobber redefs. 1529 if (VNI->hasRedefByEC()) 1530 return false; 1531 1532 // Figure out the value # from the LHS. 1533 LiveRange *lr = LHS.getLiveRangeContaining(VNI->def.getPrevSlot()); 1534 // The copy could be to an aliased physreg. 1535 if (!lr) continue; 1536 1537 // DstReg is known to be a register in the RHS interval. If the src is 1538 // from the LHS interval, we can use its value #. 1539 MachineInstr *MI = VNI->getCopy(); 1540 if (!CP.isCoalescable(MI) && 1541 !RegistersDefinedFromSameValue(*LIS, *TRI, CP, VNI, lr, DupCopies)) 1542 continue; 1543 1544 RHSValsDefinedFromLHS[VNI] = lr->valno; 1545 } 1546 1547 LHSValNoAssignments.resize(LHS.getNumValNums(), -1); 1548 RHSValNoAssignments.resize(RHS.getNumValNums(), -1); 1549 NewVNInfo.reserve(LHS.getNumValNums() + RHS.getNumValNums()); 1550 1551 for (LiveInterval::vni_iterator i = LHS.vni_begin(), e = LHS.vni_end(); 1552 i != e; ++i) { 1553 VNInfo *VNI = *i; 1554 unsigned VN = VNI->id; 1555 if (LHSValNoAssignments[VN] >= 0 || VNI->isUnused()) 1556 continue; 1557 ComputeUltimateVN(VNI, NewVNInfo, 1558 LHSValsDefinedFromRHS, RHSValsDefinedFromLHS, 1559 LHSValNoAssignments, RHSValNoAssignments); 1560 } 1561 for (LiveInterval::vni_iterator i = RHS.vni_begin(), e = RHS.vni_end(); 1562 i != e; ++i) { 1563 VNInfo *VNI = *i; 1564 unsigned VN = VNI->id; 1565 if (RHSValNoAssignments[VN] >= 0 || VNI->isUnused()) 1566 continue; 1567 // If this value number isn't a copy from the LHS, it's a new number. 1568 if (RHSValsDefinedFromLHS.find(VNI) == RHSValsDefinedFromLHS.end()) { 1569 NewVNInfo.push_back(VNI); 1570 RHSValNoAssignments[VN] = NewVNInfo.size()-1; 1571 continue; 1572 } 1573 1574 ComputeUltimateVN(VNI, NewVNInfo, 1575 RHSValsDefinedFromLHS, LHSValsDefinedFromRHS, 1576 RHSValNoAssignments, LHSValNoAssignments); 1577 } 1578 1579 // Armed with the mappings of LHS/RHS values to ultimate values, walk the 1580 // interval lists to see if these intervals are coalescable. 1581 LiveInterval::const_iterator I = LHS.begin(); 1582 LiveInterval::const_iterator IE = LHS.end(); 1583 LiveInterval::const_iterator J = RHS.begin(); 1584 LiveInterval::const_iterator JE = RHS.end(); 1585 1586 // Skip ahead until the first place of potential sharing. 1587 if (I != IE && J != JE) { 1588 if (I->start < J->start) { 1589 I = std::upper_bound(I, IE, J->start); 1590 if (I != LHS.begin()) --I; 1591 } else if (J->start < I->start) { 1592 J = std::upper_bound(J, JE, I->start); 1593 if (J != RHS.begin()) --J; 1594 } 1595 } 1596 1597 while (I != IE && J != JE) { 1598 // Determine if these two live ranges overlap. 1599 bool Overlaps; 1600 if (I->start < J->start) { 1601 Overlaps = I->end > J->start; 1602 } else { 1603 Overlaps = J->end > I->start; 1604 } 1605 1606 // If so, check value # info to determine if they are really different. 1607 if (Overlaps) { 1608 // If the live range overlap will map to the same value number in the 1609 // result liverange, we can still coalesce them. If not, we can't. 1610 if (LHSValNoAssignments[I->valno->id] != 1611 RHSValNoAssignments[J->valno->id]) 1612 return false; 1613 // If it's re-defined by an early clobber somewhere in the live range, 1614 // then conservatively abort coalescing. 1615 if (NewVNInfo[LHSValNoAssignments[I->valno->id]]->hasRedefByEC()) 1616 return false; 1617 } 1618 1619 if (I->end < J->end) 1620 ++I; 1621 else 1622 ++J; 1623 } 1624 1625 // Update kill info. Some live ranges are extended due to copy coalescing. 1626 for (DenseMap<VNInfo*, VNInfo*>::iterator I = LHSValsDefinedFromRHS.begin(), 1627 E = LHSValsDefinedFromRHS.end(); I != E; ++I) { 1628 VNInfo *VNI = I->first; 1629 unsigned LHSValID = LHSValNoAssignments[VNI->id]; 1630 if (VNI->hasPHIKill()) 1631 NewVNInfo[LHSValID]->setHasPHIKill(true); 1632 } 1633 1634 // Update kill info. Some live ranges are extended due to copy coalescing. 1635 for (DenseMap<VNInfo*, VNInfo*>::iterator I = RHSValsDefinedFromLHS.begin(), 1636 E = RHSValsDefinedFromLHS.end(); I != E; ++I) { 1637 VNInfo *VNI = I->first; 1638 unsigned RHSValID = RHSValNoAssignments[VNI->id]; 1639 if (VNI->hasPHIKill()) 1640 NewVNInfo[RHSValID]->setHasPHIKill(true); 1641 } 1642 1643 if (LHSValNoAssignments.empty()) 1644 LHSValNoAssignments.push_back(-1); 1645 if (RHSValNoAssignments.empty()) 1646 RHSValNoAssignments.push_back(-1); 1647 1648 SmallVector<unsigned, 8> SourceRegisters; 1649 for (SmallVector<MachineInstr*, 8>::iterator I = DupCopies.begin(), 1650 E = DupCopies.end(); I != E; ++I) { 1651 MachineInstr *MI = *I; 1652 1653 // We have pretended that the assignment to B in 1654 // A = X 1655 // B = X 1656 // was actually a copy from A. Now that we decided to coalesce A and B, 1657 // transform the code into 1658 // A = X 1659 // X = X 1660 // and mark the X as coalesced to keep the illusion. 1661 unsigned Src = MI->getOperand(1).getReg(); 1662 SourceRegisters.push_back(Src); 1663 MI->getOperand(0).substVirtReg(Src, 0, *TRI); 1664 1665 markAsJoined(MI); 1666 } 1667 1668 // If B = X was the last use of X in a liverange, we have to shrink it now 1669 // that B = X is gone. 1670 for (SmallVector<unsigned, 8>::iterator I = SourceRegisters.begin(), 1671 E = SourceRegisters.end(); I != E; ++I) { 1672 LIS->shrinkToUses(&LIS->getInterval(*I)); 1673 } 1674 1675 // If we get here, we know that we can coalesce the live ranges. Ask the 1676 // intervals to coalesce themselves now. 1677 LHS.join(RHS, &LHSValNoAssignments[0], &RHSValNoAssignments[0], NewVNInfo, 1678 MRI); 1679 return true; 1680 } 1681 1682 namespace { 1683 // DepthMBBCompare - Comparison predicate that sort first based on the loop 1684 // depth of the basic block (the unsigned), and then on the MBB number. 1685 struct DepthMBBCompare { 1686 typedef std::pair<unsigned, MachineBasicBlock*> DepthMBBPair; 1687 bool operator()(const DepthMBBPair &LHS, const DepthMBBPair &RHS) const { 1688 // Deeper loops first 1689 if (LHS.first != RHS.first) 1690 return LHS.first > RHS.first; 1691 1692 // Prefer blocks that are more connected in the CFG. This takes care of 1693 // the most difficult copies first while intervals are short. 1694 unsigned cl = LHS.second->pred_size() + LHS.second->succ_size(); 1695 unsigned cr = RHS.second->pred_size() + RHS.second->succ_size(); 1696 if (cl != cr) 1697 return cl > cr; 1698 1699 // As a last resort, sort by block number. 1700 return LHS.second->getNumber() < RHS.second->getNumber(); 1701 } 1702 }; 1703 } 1704 1705 void RegisterCoalescer::CopyCoalesceInMBB(MachineBasicBlock *MBB, 1706 std::vector<MachineInstr*> &TryAgain) { 1707 DEBUG(dbgs() << MBB->getName() << ":\n"); 1708 1709 SmallVector<MachineInstr*, 8> VirtCopies; 1710 SmallVector<MachineInstr*, 8> PhysCopies; 1711 SmallVector<MachineInstr*, 8> ImpDefCopies; 1712 for (MachineBasicBlock::iterator MII = MBB->begin(), E = MBB->end(); 1713 MII != E;) { 1714 MachineInstr *Inst = MII++; 1715 1716 // If this isn't a copy nor a extract_subreg, we can't join intervals. 1717 unsigned SrcReg, DstReg; 1718 if (Inst->isCopy()) { 1719 DstReg = Inst->getOperand(0).getReg(); 1720 SrcReg = Inst->getOperand(1).getReg(); 1721 } else if (Inst->isSubregToReg()) { 1722 DstReg = Inst->getOperand(0).getReg(); 1723 SrcReg = Inst->getOperand(2).getReg(); 1724 } else 1725 continue; 1726 1727 bool SrcIsPhys = TargetRegisterInfo::isPhysicalRegister(SrcReg); 1728 bool DstIsPhys = TargetRegisterInfo::isPhysicalRegister(DstReg); 1729 if (LIS->hasInterval(SrcReg) && LIS->getInterval(SrcReg).empty()) 1730 ImpDefCopies.push_back(Inst); 1731 else if (SrcIsPhys || DstIsPhys) 1732 PhysCopies.push_back(Inst); 1733 else 1734 VirtCopies.push_back(Inst); 1735 } 1736 1737 // Try coalescing implicit copies and insert_subreg <undef> first, 1738 // followed by copies to / from physical registers, then finally copies 1739 // from virtual registers to virtual registers. 1740 for (unsigned i = 0, e = ImpDefCopies.size(); i != e; ++i) { 1741 MachineInstr *TheCopy = ImpDefCopies[i]; 1742 bool Again = false; 1743 if (!JoinCopy(TheCopy, Again)) 1744 if (Again) 1745 TryAgain.push_back(TheCopy); 1746 } 1747 for (unsigned i = 0, e = PhysCopies.size(); i != e; ++i) { 1748 MachineInstr *TheCopy = PhysCopies[i]; 1749 bool Again = false; 1750 if (!JoinCopy(TheCopy, Again)) 1751 if (Again) 1752 TryAgain.push_back(TheCopy); 1753 } 1754 for (unsigned i = 0, e = VirtCopies.size(); i != e; ++i) { 1755 MachineInstr *TheCopy = VirtCopies[i]; 1756 bool Again = false; 1757 if (!JoinCopy(TheCopy, Again)) 1758 if (Again) 1759 TryAgain.push_back(TheCopy); 1760 } 1761 } 1762 1763 void RegisterCoalescer::joinIntervals() { 1764 DEBUG(dbgs() << "********** JOINING INTERVALS ***********\n"); 1765 1766 std::vector<MachineInstr*> TryAgainList; 1767 if (Loops->empty()) { 1768 // If there are no loops in the function, join intervals in function order. 1769 for (MachineFunction::iterator I = MF->begin(), E = MF->end(); 1770 I != E; ++I) 1771 CopyCoalesceInMBB(I, TryAgainList); 1772 } else { 1773 // Otherwise, join intervals in inner loops before other intervals. 1774 // Unfortunately we can't just iterate over loop hierarchy here because 1775 // there may be more MBB's than BB's. Collect MBB's for sorting. 1776 1777 // Join intervals in the function prolog first. We want to join physical 1778 // registers with virtual registers before the intervals got too long. 1779 std::vector<std::pair<unsigned, MachineBasicBlock*> > MBBs; 1780 for (MachineFunction::iterator I = MF->begin(), E = MF->end();I != E;++I){ 1781 MachineBasicBlock *MBB = I; 1782 MBBs.push_back(std::make_pair(Loops->getLoopDepth(MBB), I)); 1783 } 1784 1785 // Sort by loop depth. 1786 std::sort(MBBs.begin(), MBBs.end(), DepthMBBCompare()); 1787 1788 // Finally, join intervals in loop nest order. 1789 for (unsigned i = 0, e = MBBs.size(); i != e; ++i) 1790 CopyCoalesceInMBB(MBBs[i].second, TryAgainList); 1791 } 1792 1793 // Joining intervals can allow other intervals to be joined. Iteratively join 1794 // until we make no progress. 1795 bool ProgressMade = true; 1796 while (ProgressMade) { 1797 ProgressMade = false; 1798 1799 for (unsigned i = 0, e = TryAgainList.size(); i != e; ++i) { 1800 MachineInstr *&TheCopy = TryAgainList[i]; 1801 if (!TheCopy) 1802 continue; 1803 1804 bool Again = false; 1805 bool Success = JoinCopy(TheCopy, Again); 1806 if (Success || !Again) { 1807 TheCopy= 0; // Mark this one as done. 1808 ProgressMade = true; 1809 } 1810 } 1811 } 1812 } 1813 1814 void RegisterCoalescer::releaseMemory() { 1815 JoinedCopies.clear(); 1816 ReMatCopies.clear(); 1817 ReMatDefs.clear(); 1818 } 1819 1820 bool RegisterCoalescer::runOnMachineFunction(MachineFunction &fn) { 1821 MF = &fn; 1822 MRI = &fn.getRegInfo(); 1823 TM = &fn.getTarget(); 1824 TRI = TM->getRegisterInfo(); 1825 TII = TM->getInstrInfo(); 1826 LIS = &getAnalysis<LiveIntervals>(); 1827 LDV = &getAnalysis<LiveDebugVariables>(); 1828 AA = &getAnalysis<AliasAnalysis>(); 1829 Loops = &getAnalysis<MachineLoopInfo>(); 1830 1831 DEBUG(dbgs() << "********** SIMPLE REGISTER COALESCING **********\n" 1832 << "********** Function: " 1833 << ((Value*)MF->getFunction())->getName() << '\n'); 1834 1835 if (VerifyCoalescing) 1836 MF->verify(this, "Before register coalescing"); 1837 1838 RegClassInfo.runOnMachineFunction(fn); 1839 1840 // Join (coalesce) intervals if requested. 1841 if (EnableJoining) { 1842 joinIntervals(); 1843 DEBUG({ 1844 dbgs() << "********** INTERVALS POST JOINING **********\n"; 1845 for (LiveIntervals::iterator I = LIS->begin(), E = LIS->end(); 1846 I != E; ++I){ 1847 I->second->print(dbgs(), TRI); 1848 dbgs() << "\n"; 1849 } 1850 }); 1851 } 1852 1853 // Perform a final pass over the instructions and compute spill weights 1854 // and remove identity moves. 1855 SmallVector<unsigned, 4> DeadDefs, InflateRegs; 1856 for (MachineFunction::iterator mbbi = MF->begin(), mbbe = MF->end(); 1857 mbbi != mbbe; ++mbbi) { 1858 MachineBasicBlock* mbb = mbbi; 1859 for (MachineBasicBlock::iterator mii = mbb->begin(), mie = mbb->end(); 1860 mii != mie; ) { 1861 MachineInstr *MI = mii; 1862 if (JoinedCopies.count(MI)) { 1863 // Delete all coalesced copies. 1864 bool DoDelete = true; 1865 assert(MI->isCopyLike() && "Unrecognized copy instruction"); 1866 unsigned SrcReg = MI->getOperand(MI->isSubregToReg() ? 2 : 1).getReg(); 1867 unsigned DstReg = MI->getOperand(0).getReg(); 1868 1869 // Collect candidates for register class inflation. 1870 if (TargetRegisterInfo::isVirtualRegister(SrcReg) && 1871 RegClassInfo.isProperSubClass(MRI->getRegClass(SrcReg))) 1872 InflateRegs.push_back(SrcReg); 1873 if (TargetRegisterInfo::isVirtualRegister(DstReg) && 1874 RegClassInfo.isProperSubClass(MRI->getRegClass(DstReg))) 1875 InflateRegs.push_back(DstReg); 1876 1877 if (TargetRegisterInfo::isPhysicalRegister(SrcReg) && 1878 MI->getNumOperands() > 2) 1879 // Do not delete extract_subreg, insert_subreg of physical 1880 // registers unless the definition is dead. e.g. 1881 // %DO<def> = INSERT_SUBREG %D0<undef>, %S0<kill>, 1 1882 // or else the scavenger may complain. LowerSubregs will 1883 // delete them later. 1884 DoDelete = false; 1885 1886 if (MI->allDefsAreDead()) { 1887 if (TargetRegisterInfo::isVirtualRegister(SrcReg) && 1888 LIS->hasInterval(SrcReg)) 1889 LIS->shrinkToUses(&LIS->getInterval(SrcReg)); 1890 DoDelete = true; 1891 } 1892 if (!DoDelete) { 1893 // We need the instruction to adjust liveness, so make it a KILL. 1894 if (MI->isSubregToReg()) { 1895 MI->RemoveOperand(3); 1896 MI->RemoveOperand(1); 1897 } 1898 MI->setDesc(TII->get(TargetOpcode::KILL)); 1899 mii = llvm::next(mii); 1900 } else { 1901 LIS->RemoveMachineInstrFromMaps(MI); 1902 mii = mbbi->erase(mii); 1903 ++numPeep; 1904 } 1905 continue; 1906 } 1907 1908 // Now check if this is a remat'ed def instruction which is now dead. 1909 if (ReMatDefs.count(MI)) { 1910 bool isDead = true; 1911 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 1912 const MachineOperand &MO = MI->getOperand(i); 1913 if (!MO.isReg()) 1914 continue; 1915 unsigned Reg = MO.getReg(); 1916 if (!Reg) 1917 continue; 1918 if (TargetRegisterInfo::isVirtualRegister(Reg)) { 1919 DeadDefs.push_back(Reg); 1920 // Remat may also enable register class inflation. 1921 if (RegClassInfo.isProperSubClass(MRI->getRegClass(Reg))) 1922 InflateRegs.push_back(Reg); 1923 } 1924 if (MO.isDead()) 1925 continue; 1926 if (TargetRegisterInfo::isPhysicalRegister(Reg) || 1927 !MRI->use_nodbg_empty(Reg)) { 1928 isDead = false; 1929 break; 1930 } 1931 } 1932 if (isDead) { 1933 while (!DeadDefs.empty()) { 1934 unsigned DeadDef = DeadDefs.back(); 1935 DeadDefs.pop_back(); 1936 RemoveDeadDef(LIS->getInterval(DeadDef), MI); 1937 } 1938 LIS->RemoveMachineInstrFromMaps(mii); 1939 mii = mbbi->erase(mii); 1940 continue; 1941 } else 1942 DeadDefs.clear(); 1943 } 1944 1945 ++mii; 1946 1947 // Check for now unnecessary kill flags. 1948 if (LIS->isNotInMIMap(MI)) continue; 1949 SlotIndex DefIdx = LIS->getInstructionIndex(MI).getDefIndex(); 1950 for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) { 1951 MachineOperand &MO = MI->getOperand(i); 1952 if (!MO.isReg() || !MO.isKill()) continue; 1953 unsigned reg = MO.getReg(); 1954 if (!reg || !LIS->hasInterval(reg)) continue; 1955 if (!LIS->getInterval(reg).killedAt(DefIdx)) { 1956 MO.setIsKill(false); 1957 continue; 1958 } 1959 // When leaving a kill flag on a physreg, check if any subregs should 1960 // remain alive. 1961 if (!TargetRegisterInfo::isPhysicalRegister(reg)) 1962 continue; 1963 for (const unsigned *SR = TRI->getSubRegisters(reg); 1964 unsigned S = *SR; ++SR) 1965 if (LIS->hasInterval(S) && LIS->getInterval(S).liveAt(DefIdx)) 1966 MI->addRegisterDefined(S, TRI); 1967 } 1968 } 1969 } 1970 1971 // After deleting a lot of copies, register classes may be less constrained. 1972 // Removing sub-register opreands may alow GR32_ABCD -> GR32 and DPR_VFP2 -> 1973 // DPR inflation. 1974 array_pod_sort(InflateRegs.begin(), InflateRegs.end()); 1975 InflateRegs.erase(std::unique(InflateRegs.begin(), InflateRegs.end()), 1976 InflateRegs.end()); 1977 DEBUG(dbgs() << "Trying to inflate " << InflateRegs.size() << " regs.\n"); 1978 for (unsigned i = 0, e = InflateRegs.size(); i != e; ++i) { 1979 unsigned Reg = InflateRegs[i]; 1980 if (MRI->reg_nodbg_empty(Reg)) 1981 continue; 1982 if (MRI->recomputeRegClass(Reg, *TM)) { 1983 DEBUG(dbgs() << PrintReg(Reg) << " inflated to " 1984 << MRI->getRegClass(Reg)->getName() << '\n'); 1985 ++NumInflated; 1986 } 1987 } 1988 1989 DEBUG(dump()); 1990 DEBUG(LDV->dump()); 1991 if (VerifyCoalescing) 1992 MF->verify(this, "After register coalescing"); 1993 return true; 1994 } 1995 1996 /// print - Implement the dump method. 1997 void RegisterCoalescer::print(raw_ostream &O, const Module* m) const { 1998 LIS->print(O, m); 1999 } 2000