1 //===-- SIInstrInfo.cpp - SI Instruction Information ---------------------===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 /// \file 11 /// \brief SI Implementation of TargetInstrInfo. 12 // 13 //===----------------------------------------------------------------------===// 14 15 #include "SIInstrInfo.h" 16 #include "AMDGPUTargetMachine.h" 17 #include "GCNHazardRecognizer.h" 18 #include "SIDefines.h" 19 #include "SIMachineFunctionInfo.h" 20 #include "llvm/CodeGen/MachineFrameInfo.h" 21 #include "llvm/CodeGen/MachineInstrBuilder.h" 22 #include "llvm/CodeGen/MachineRegisterInfo.h" 23 #include "llvm/CodeGen/ScheduleDAG.h" 24 #include "llvm/IR/Function.h" 25 #include "llvm/CodeGen/RegisterScavenging.h" 26 #include "llvm/MC/MCInstrDesc.h" 27 #include "llvm/Support/Debug.h" 28 29 using namespace llvm; 30 31 SIInstrInfo::SIInstrInfo(const SISubtarget &ST) 32 : AMDGPUInstrInfo(ST), RI(), ST(ST) {} 33 34 //===----------------------------------------------------------------------===// 35 // TargetInstrInfo callbacks 36 //===----------------------------------------------------------------------===// 37 38 static unsigned getNumOperandsNoGlue(SDNode *Node) { 39 unsigned N = Node->getNumOperands(); 40 while (N && Node->getOperand(N - 1).getValueType() == MVT::Glue) 41 --N; 42 return N; 43 } 44 45 static SDValue findChainOperand(SDNode *Load) { 46 SDValue LastOp = Load->getOperand(getNumOperandsNoGlue(Load) - 1); 47 assert(LastOp.getValueType() == MVT::Other && "Chain missing from load node"); 48 return LastOp; 49 } 50 51 /// \brief Returns true if both nodes have the same value for the given 52 /// operand \p Op, or if both nodes do not have this operand. 53 static bool nodesHaveSameOperandValue(SDNode *N0, SDNode* N1, unsigned OpName) { 54 unsigned Opc0 = N0->getMachineOpcode(); 55 unsigned Opc1 = N1->getMachineOpcode(); 56 57 int Op0Idx = AMDGPU::getNamedOperandIdx(Opc0, OpName); 58 int Op1Idx = AMDGPU::getNamedOperandIdx(Opc1, OpName); 59 60 if (Op0Idx == -1 && Op1Idx == -1) 61 return true; 62 63 64 if ((Op0Idx == -1 && Op1Idx != -1) || 65 (Op1Idx == -1 && Op0Idx != -1)) 66 return false; 67 68 // getNamedOperandIdx returns the index for the MachineInstr's operands, 69 // which includes the result as the first operand. We are indexing into the 70 // MachineSDNode's operands, so we need to skip the result operand to get 71 // the real index. 72 --Op0Idx; 73 --Op1Idx; 74 75 return N0->getOperand(Op0Idx) == N1->getOperand(Op1Idx); 76 } 77 78 bool SIInstrInfo::isReallyTriviallyReMaterializable(const MachineInstr &MI, 79 AliasAnalysis *AA) const { 80 // TODO: The generic check fails for VALU instructions that should be 81 // rematerializable due to implicit reads of exec. We really want all of the 82 // generic logic for this except for this. 83 switch (MI.getOpcode()) { 84 case AMDGPU::V_MOV_B32_e32: 85 case AMDGPU::V_MOV_B32_e64: 86 case AMDGPU::V_MOV_B64_PSEUDO: 87 return true; 88 default: 89 return false; 90 } 91 } 92 93 bool SIInstrInfo::areLoadsFromSameBasePtr(SDNode *Load0, SDNode *Load1, 94 int64_t &Offset0, 95 int64_t &Offset1) const { 96 if (!Load0->isMachineOpcode() || !Load1->isMachineOpcode()) 97 return false; 98 99 unsigned Opc0 = Load0->getMachineOpcode(); 100 unsigned Opc1 = Load1->getMachineOpcode(); 101 102 // Make sure both are actually loads. 103 if (!get(Opc0).mayLoad() || !get(Opc1).mayLoad()) 104 return false; 105 106 if (isDS(Opc0) && isDS(Opc1)) { 107 108 // FIXME: Handle this case: 109 if (getNumOperandsNoGlue(Load0) != getNumOperandsNoGlue(Load1)) 110 return false; 111 112 // Check base reg. 113 if (Load0->getOperand(1) != Load1->getOperand(1)) 114 return false; 115 116 // Check chain. 117 if (findChainOperand(Load0) != findChainOperand(Load1)) 118 return false; 119 120 // Skip read2 / write2 variants for simplicity. 121 // TODO: We should report true if the used offsets are adjacent (excluded 122 // st64 versions). 123 if (AMDGPU::getNamedOperandIdx(Opc0, AMDGPU::OpName::data1) != -1 || 124 AMDGPU::getNamedOperandIdx(Opc1, AMDGPU::OpName::data1) != -1) 125 return false; 126 127 Offset0 = cast<ConstantSDNode>(Load0->getOperand(2))->getZExtValue(); 128 Offset1 = cast<ConstantSDNode>(Load1->getOperand(2))->getZExtValue(); 129 return true; 130 } 131 132 if (isSMRD(Opc0) && isSMRD(Opc1)) { 133 assert(getNumOperandsNoGlue(Load0) == getNumOperandsNoGlue(Load1)); 134 135 // Check base reg. 136 if (Load0->getOperand(0) != Load1->getOperand(0)) 137 return false; 138 139 const ConstantSDNode *Load0Offset = 140 dyn_cast<ConstantSDNode>(Load0->getOperand(1)); 141 const ConstantSDNode *Load1Offset = 142 dyn_cast<ConstantSDNode>(Load1->getOperand(1)); 143 144 if (!Load0Offset || !Load1Offset) 145 return false; 146 147 // Check chain. 148 if (findChainOperand(Load0) != findChainOperand(Load1)) 149 return false; 150 151 Offset0 = Load0Offset->getZExtValue(); 152 Offset1 = Load1Offset->getZExtValue(); 153 return true; 154 } 155 156 // MUBUF and MTBUF can access the same addresses. 157 if ((isMUBUF(Opc0) || isMTBUF(Opc0)) && (isMUBUF(Opc1) || isMTBUF(Opc1))) { 158 159 // MUBUF and MTBUF have vaddr at different indices. 160 if (!nodesHaveSameOperandValue(Load0, Load1, AMDGPU::OpName::soffset) || 161 findChainOperand(Load0) != findChainOperand(Load1) || 162 !nodesHaveSameOperandValue(Load0, Load1, AMDGPU::OpName::vaddr) || 163 !nodesHaveSameOperandValue(Load0, Load1, AMDGPU::OpName::srsrc)) 164 return false; 165 166 int OffIdx0 = AMDGPU::getNamedOperandIdx(Opc0, AMDGPU::OpName::offset); 167 int OffIdx1 = AMDGPU::getNamedOperandIdx(Opc1, AMDGPU::OpName::offset); 168 169 if (OffIdx0 == -1 || OffIdx1 == -1) 170 return false; 171 172 // getNamedOperandIdx returns the index for MachineInstrs. Since they 173 // inlcude the output in the operand list, but SDNodes don't, we need to 174 // subtract the index by one. 175 --OffIdx0; 176 --OffIdx1; 177 178 SDValue Off0 = Load0->getOperand(OffIdx0); 179 SDValue Off1 = Load1->getOperand(OffIdx1); 180 181 // The offset might be a FrameIndexSDNode. 182 if (!isa<ConstantSDNode>(Off0) || !isa<ConstantSDNode>(Off1)) 183 return false; 184 185 Offset0 = cast<ConstantSDNode>(Off0)->getZExtValue(); 186 Offset1 = cast<ConstantSDNode>(Off1)->getZExtValue(); 187 return true; 188 } 189 190 return false; 191 } 192 193 static bool isStride64(unsigned Opc) { 194 switch (Opc) { 195 case AMDGPU::DS_READ2ST64_B32: 196 case AMDGPU::DS_READ2ST64_B64: 197 case AMDGPU::DS_WRITE2ST64_B32: 198 case AMDGPU::DS_WRITE2ST64_B64: 199 return true; 200 default: 201 return false; 202 } 203 } 204 205 bool SIInstrInfo::getMemOpBaseRegImmOfs(MachineInstr &LdSt, unsigned &BaseReg, 206 int64_t &Offset, 207 const TargetRegisterInfo *TRI) const { 208 unsigned Opc = LdSt.getOpcode(); 209 210 if (isDS(LdSt)) { 211 const MachineOperand *OffsetImm = 212 getNamedOperand(LdSt, AMDGPU::OpName::offset); 213 if (OffsetImm) { 214 // Normal, single offset LDS instruction. 215 const MachineOperand *AddrReg = 216 getNamedOperand(LdSt, AMDGPU::OpName::addr); 217 218 BaseReg = AddrReg->getReg(); 219 Offset = OffsetImm->getImm(); 220 return true; 221 } 222 223 // The 2 offset instructions use offset0 and offset1 instead. We can treat 224 // these as a load with a single offset if the 2 offsets are consecutive. We 225 // will use this for some partially aligned loads. 226 const MachineOperand *Offset0Imm = 227 getNamedOperand(LdSt, AMDGPU::OpName::offset0); 228 const MachineOperand *Offset1Imm = 229 getNamedOperand(LdSt, AMDGPU::OpName::offset1); 230 231 uint8_t Offset0 = Offset0Imm->getImm(); 232 uint8_t Offset1 = Offset1Imm->getImm(); 233 234 if (Offset1 > Offset0 && Offset1 - Offset0 == 1) { 235 // Each of these offsets is in element sized units, so we need to convert 236 // to bytes of the individual reads. 237 238 unsigned EltSize; 239 if (LdSt.mayLoad()) 240 EltSize = getOpRegClass(LdSt, 0)->getSize() / 2; 241 else { 242 assert(LdSt.mayStore()); 243 int Data0Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::data0); 244 EltSize = getOpRegClass(LdSt, Data0Idx)->getSize(); 245 } 246 247 if (isStride64(Opc)) 248 EltSize *= 64; 249 250 const MachineOperand *AddrReg = 251 getNamedOperand(LdSt, AMDGPU::OpName::addr); 252 BaseReg = AddrReg->getReg(); 253 Offset = EltSize * Offset0; 254 return true; 255 } 256 257 return false; 258 } 259 260 if (isMUBUF(LdSt) || isMTBUF(LdSt)) { 261 if (AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::soffset) != -1) 262 return false; 263 264 const MachineOperand *AddrReg = 265 getNamedOperand(LdSt, AMDGPU::OpName::vaddr); 266 if (!AddrReg) 267 return false; 268 269 const MachineOperand *OffsetImm = 270 getNamedOperand(LdSt, AMDGPU::OpName::offset); 271 BaseReg = AddrReg->getReg(); 272 Offset = OffsetImm->getImm(); 273 return true; 274 } 275 276 if (isSMRD(LdSt)) { 277 const MachineOperand *OffsetImm = 278 getNamedOperand(LdSt, AMDGPU::OpName::offset); 279 if (!OffsetImm) 280 return false; 281 282 const MachineOperand *SBaseReg = 283 getNamedOperand(LdSt, AMDGPU::OpName::sbase); 284 BaseReg = SBaseReg->getReg(); 285 Offset = OffsetImm->getImm(); 286 return true; 287 } 288 289 if (isFLAT(LdSt)) { 290 const MachineOperand *AddrReg = getNamedOperand(LdSt, AMDGPU::OpName::addr); 291 BaseReg = AddrReg->getReg(); 292 Offset = 0; 293 return true; 294 } 295 296 return false; 297 } 298 299 bool SIInstrInfo::shouldClusterMemOps(MachineInstr &FirstLdSt, 300 MachineInstr &SecondLdSt, 301 unsigned NumLoads) const { 302 const MachineOperand *FirstDst = nullptr; 303 const MachineOperand *SecondDst = nullptr; 304 305 if (isDS(FirstLdSt) && isDS(SecondLdSt)) { 306 FirstDst = getNamedOperand(FirstLdSt, AMDGPU::OpName::vdst); 307 SecondDst = getNamedOperand(SecondLdSt, AMDGPU::OpName::vdst); 308 } 309 310 if (isSMRD(FirstLdSt) && isSMRD(SecondLdSt)) { 311 FirstDst = getNamedOperand(FirstLdSt, AMDGPU::OpName::sdst); 312 SecondDst = getNamedOperand(SecondLdSt, AMDGPU::OpName::sdst); 313 } 314 315 if ((isMUBUF(FirstLdSt) && isMUBUF(SecondLdSt)) || 316 (isMTBUF(FirstLdSt) && isMTBUF(SecondLdSt))) { 317 FirstDst = getNamedOperand(FirstLdSt, AMDGPU::OpName::vdata); 318 SecondDst = getNamedOperand(SecondLdSt, AMDGPU::OpName::vdata); 319 } 320 321 if (!FirstDst || !SecondDst) 322 return false; 323 324 // Try to limit clustering based on the total number of bytes loaded 325 // rather than the number of instructions. This is done to help reduce 326 // register pressure. The method used is somewhat inexact, though, 327 // because it assumes that all loads in the cluster will load the 328 // same number of bytes as FirstLdSt. 329 330 // The unit of this value is bytes. 331 // FIXME: This needs finer tuning. 332 unsigned LoadClusterThreshold = 16; 333 334 const MachineRegisterInfo &MRI = 335 FirstLdSt.getParent()->getParent()->getRegInfo(); 336 const TargetRegisterClass *DstRC = MRI.getRegClass(FirstDst->getReg()); 337 338 return (NumLoads * DstRC->getSize()) <= LoadClusterThreshold; 339 } 340 341 void SIInstrInfo::copyPhysReg(MachineBasicBlock &MBB, 342 MachineBasicBlock::iterator MI, 343 const DebugLoc &DL, unsigned DestReg, 344 unsigned SrcReg, bool KillSrc) const { 345 346 // If we are trying to copy to or from SCC, there is a bug somewhere else in 347 // the backend. While it may be theoretically possible to do this, it should 348 // never be necessary. 349 assert(DestReg != AMDGPU::SCC && SrcReg != AMDGPU::SCC); 350 351 static const int16_t Sub0_15[] = { 352 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3, 353 AMDGPU::sub4, AMDGPU::sub5, AMDGPU::sub6, AMDGPU::sub7, 354 AMDGPU::sub8, AMDGPU::sub9, AMDGPU::sub10, AMDGPU::sub11, 355 AMDGPU::sub12, AMDGPU::sub13, AMDGPU::sub14, AMDGPU::sub15, 356 }; 357 358 static const int16_t Sub0_15_64[] = { 359 AMDGPU::sub0_sub1, AMDGPU::sub2_sub3, 360 AMDGPU::sub4_sub5, AMDGPU::sub6_sub7, 361 AMDGPU::sub8_sub9, AMDGPU::sub10_sub11, 362 AMDGPU::sub12_sub13, AMDGPU::sub14_sub15, 363 }; 364 365 static const int16_t Sub0_7[] = { 366 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3, 367 AMDGPU::sub4, AMDGPU::sub5, AMDGPU::sub6, AMDGPU::sub7, 368 }; 369 370 static const int16_t Sub0_7_64[] = { 371 AMDGPU::sub0_sub1, AMDGPU::sub2_sub3, 372 AMDGPU::sub4_sub5, AMDGPU::sub6_sub7, 373 }; 374 375 static const int16_t Sub0_3[] = { 376 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, AMDGPU::sub3, 377 }; 378 379 static const int16_t Sub0_3_64[] = { 380 AMDGPU::sub0_sub1, AMDGPU::sub2_sub3, 381 }; 382 383 static const int16_t Sub0_2[] = { 384 AMDGPU::sub0, AMDGPU::sub1, AMDGPU::sub2, 385 }; 386 387 static const int16_t Sub0_1[] = { 388 AMDGPU::sub0, AMDGPU::sub1, 389 }; 390 391 unsigned Opcode; 392 ArrayRef<int16_t> SubIndices; 393 394 if (AMDGPU::SReg_32RegClass.contains(DestReg)) { 395 assert(AMDGPU::SReg_32RegClass.contains(SrcReg)); 396 BuildMI(MBB, MI, DL, get(AMDGPU::S_MOV_B32), DestReg) 397 .addReg(SrcReg, getKillRegState(KillSrc)); 398 return; 399 400 } else if (AMDGPU::SReg_64RegClass.contains(DestReg)) { 401 if (DestReg == AMDGPU::VCC) { 402 if (AMDGPU::SReg_64RegClass.contains(SrcReg)) { 403 BuildMI(MBB, MI, DL, get(AMDGPU::S_MOV_B64), AMDGPU::VCC) 404 .addReg(SrcReg, getKillRegState(KillSrc)); 405 } else { 406 // FIXME: Hack until VReg_1 removed. 407 assert(AMDGPU::VGPR_32RegClass.contains(SrcReg)); 408 BuildMI(MBB, MI, DL, get(AMDGPU::V_CMP_NE_I32_e32)) 409 .addImm(0) 410 .addReg(SrcReg, getKillRegState(KillSrc)); 411 } 412 413 return; 414 } 415 416 assert(AMDGPU::SReg_64RegClass.contains(SrcReg)); 417 BuildMI(MBB, MI, DL, get(AMDGPU::S_MOV_B64), DestReg) 418 .addReg(SrcReg, getKillRegState(KillSrc)); 419 return; 420 421 } else if (AMDGPU::SReg_128RegClass.contains(DestReg)) { 422 assert(AMDGPU::SReg_128RegClass.contains(SrcReg)); 423 Opcode = AMDGPU::S_MOV_B64; 424 SubIndices = Sub0_3_64; 425 426 } else if (AMDGPU::SReg_256RegClass.contains(DestReg)) { 427 assert(AMDGPU::SReg_256RegClass.contains(SrcReg)); 428 Opcode = AMDGPU::S_MOV_B64; 429 SubIndices = Sub0_7_64; 430 431 } else if (AMDGPU::SReg_512RegClass.contains(DestReg)) { 432 assert(AMDGPU::SReg_512RegClass.contains(SrcReg)); 433 Opcode = AMDGPU::S_MOV_B64; 434 SubIndices = Sub0_15_64; 435 436 } else if (AMDGPU::VGPR_32RegClass.contains(DestReg)) { 437 assert(AMDGPU::VGPR_32RegClass.contains(SrcReg) || 438 AMDGPU::SReg_32RegClass.contains(SrcReg)); 439 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DestReg) 440 .addReg(SrcReg, getKillRegState(KillSrc)); 441 return; 442 443 } else if (AMDGPU::VReg_64RegClass.contains(DestReg)) { 444 assert(AMDGPU::VReg_64RegClass.contains(SrcReg) || 445 AMDGPU::SReg_64RegClass.contains(SrcReg)); 446 Opcode = AMDGPU::V_MOV_B32_e32; 447 SubIndices = Sub0_1; 448 449 } else if (AMDGPU::VReg_96RegClass.contains(DestReg)) { 450 assert(AMDGPU::VReg_96RegClass.contains(SrcReg)); 451 Opcode = AMDGPU::V_MOV_B32_e32; 452 SubIndices = Sub0_2; 453 454 } else if (AMDGPU::VReg_128RegClass.contains(DestReg)) { 455 assert(AMDGPU::VReg_128RegClass.contains(SrcReg) || 456 AMDGPU::SReg_128RegClass.contains(SrcReg)); 457 Opcode = AMDGPU::V_MOV_B32_e32; 458 SubIndices = Sub0_3; 459 460 } else if (AMDGPU::VReg_256RegClass.contains(DestReg)) { 461 assert(AMDGPU::VReg_256RegClass.contains(SrcReg) || 462 AMDGPU::SReg_256RegClass.contains(SrcReg)); 463 Opcode = AMDGPU::V_MOV_B32_e32; 464 SubIndices = Sub0_7; 465 466 } else if (AMDGPU::VReg_512RegClass.contains(DestReg)) { 467 assert(AMDGPU::VReg_512RegClass.contains(SrcReg) || 468 AMDGPU::SReg_512RegClass.contains(SrcReg)); 469 Opcode = AMDGPU::V_MOV_B32_e32; 470 SubIndices = Sub0_15; 471 472 } else { 473 llvm_unreachable("Can't copy register!"); 474 } 475 476 bool Forward = RI.getHWRegIndex(DestReg) <= RI.getHWRegIndex(SrcReg); 477 478 for (unsigned Idx = 0; Idx < SubIndices.size(); ++Idx) { 479 unsigned SubIdx; 480 if (Forward) 481 SubIdx = SubIndices[Idx]; 482 else 483 SubIdx = SubIndices[SubIndices.size() - Idx - 1]; 484 485 MachineInstrBuilder Builder = BuildMI(MBB, MI, DL, 486 get(Opcode), RI.getSubReg(DestReg, SubIdx)); 487 488 Builder.addReg(RI.getSubReg(SrcReg, SubIdx)); 489 490 if (Idx == SubIndices.size() - 1) 491 Builder.addReg(SrcReg, getKillRegState(KillSrc) | RegState::Implicit); 492 493 if (Idx == 0) 494 Builder.addReg(DestReg, RegState::Define | RegState::Implicit); 495 496 Builder.addReg(SrcReg, RegState::Implicit); 497 } 498 } 499 500 int SIInstrInfo::commuteOpcode(const MachineInstr &MI) const { 501 const unsigned Opcode = MI.getOpcode(); 502 503 int NewOpc; 504 505 // Try to map original to commuted opcode 506 NewOpc = AMDGPU::getCommuteRev(Opcode); 507 if (NewOpc != -1) 508 // Check if the commuted (REV) opcode exists on the target. 509 return pseudoToMCOpcode(NewOpc) != -1 ? NewOpc : -1; 510 511 // Try to map commuted to original opcode 512 NewOpc = AMDGPU::getCommuteOrig(Opcode); 513 if (NewOpc != -1) 514 // Check if the original (non-REV) opcode exists on the target. 515 return pseudoToMCOpcode(NewOpc) != -1 ? NewOpc : -1; 516 517 return Opcode; 518 } 519 520 unsigned SIInstrInfo::getMovOpcode(const TargetRegisterClass *DstRC) const { 521 522 if (DstRC->getSize() == 4) { 523 return RI.isSGPRClass(DstRC) ? AMDGPU::S_MOV_B32 : AMDGPU::V_MOV_B32_e32; 524 } else if (DstRC->getSize() == 8 && RI.isSGPRClass(DstRC)) { 525 return AMDGPU::S_MOV_B64; 526 } else if (DstRC->getSize() == 8 && !RI.isSGPRClass(DstRC)) { 527 return AMDGPU::V_MOV_B64_PSEUDO; 528 } 529 return AMDGPU::COPY; 530 } 531 532 static unsigned getSGPRSpillSaveOpcode(unsigned Size) { 533 switch (Size) { 534 case 4: 535 return AMDGPU::SI_SPILL_S32_SAVE; 536 case 8: 537 return AMDGPU::SI_SPILL_S64_SAVE; 538 case 16: 539 return AMDGPU::SI_SPILL_S128_SAVE; 540 case 32: 541 return AMDGPU::SI_SPILL_S256_SAVE; 542 case 64: 543 return AMDGPU::SI_SPILL_S512_SAVE; 544 default: 545 llvm_unreachable("unknown register size"); 546 } 547 } 548 549 static unsigned getVGPRSpillSaveOpcode(unsigned Size) { 550 switch (Size) { 551 case 4: 552 return AMDGPU::SI_SPILL_V32_SAVE; 553 case 8: 554 return AMDGPU::SI_SPILL_V64_SAVE; 555 case 12: 556 return AMDGPU::SI_SPILL_V96_SAVE; 557 case 16: 558 return AMDGPU::SI_SPILL_V128_SAVE; 559 case 32: 560 return AMDGPU::SI_SPILL_V256_SAVE; 561 case 64: 562 return AMDGPU::SI_SPILL_V512_SAVE; 563 default: 564 llvm_unreachable("unknown register size"); 565 } 566 } 567 568 void SIInstrInfo::storeRegToStackSlot(MachineBasicBlock &MBB, 569 MachineBasicBlock::iterator MI, 570 unsigned SrcReg, bool isKill, 571 int FrameIndex, 572 const TargetRegisterClass *RC, 573 const TargetRegisterInfo *TRI) const { 574 MachineFunction *MF = MBB.getParent(); 575 SIMachineFunctionInfo *MFI = MF->getInfo<SIMachineFunctionInfo>(); 576 MachineFrameInfo *FrameInfo = MF->getFrameInfo(); 577 DebugLoc DL = MBB.findDebugLoc(MI); 578 579 unsigned Size = FrameInfo->getObjectSize(FrameIndex); 580 unsigned Align = FrameInfo->getObjectAlignment(FrameIndex); 581 MachinePointerInfo PtrInfo 582 = MachinePointerInfo::getFixedStack(*MF, FrameIndex); 583 MachineMemOperand *MMO 584 = MF->getMachineMemOperand(PtrInfo, MachineMemOperand::MOStore, 585 Size, Align); 586 587 if (RI.isSGPRClass(RC)) { 588 MFI->setHasSpilledSGPRs(); 589 590 if (TargetRegisterInfo::isVirtualRegister(SrcReg) && RC->getSize() == 4) { 591 // m0 may not be allowed for readlane. 592 MachineRegisterInfo &MRI = MF->getRegInfo(); 593 MRI.constrainRegClass(SrcReg, &AMDGPU::SReg_32_XM0RegClass); 594 } 595 596 // We are only allowed to create one new instruction when spilling 597 // registers, so we need to use pseudo instruction for spilling 598 // SGPRs. 599 unsigned Opcode = getSGPRSpillSaveOpcode(RC->getSize()); 600 BuildMI(MBB, MI, DL, get(Opcode)) 601 .addReg(SrcReg, getKillRegState(isKill)) // src 602 .addFrameIndex(FrameIndex) // frame_idx 603 .addMemOperand(MMO); 604 605 return; 606 } 607 608 if (!ST.isVGPRSpillingEnabled(*MF->getFunction())) { 609 LLVMContext &Ctx = MF->getFunction()->getContext(); 610 Ctx.emitError("SIInstrInfo::storeRegToStackSlot - Do not know how to" 611 " spill register"); 612 BuildMI(MBB, MI, DL, get(AMDGPU::KILL)) 613 .addReg(SrcReg); 614 615 return; 616 } 617 618 assert(RI.hasVGPRs(RC) && "Only VGPR spilling expected"); 619 620 unsigned Opcode = getVGPRSpillSaveOpcode(RC->getSize()); 621 MFI->setHasSpilledVGPRs(); 622 BuildMI(MBB, MI, DL, get(Opcode)) 623 .addReg(SrcReg, getKillRegState(isKill)) // src 624 .addFrameIndex(FrameIndex) // frame_idx 625 .addReg(MFI->getScratchRSrcReg()) // scratch_rsrc 626 .addReg(MFI->getScratchWaveOffsetReg()) // scratch_offset 627 .addImm(0) // offset 628 .addMemOperand(MMO); 629 } 630 631 static unsigned getSGPRSpillRestoreOpcode(unsigned Size) { 632 switch (Size) { 633 case 4: 634 return AMDGPU::SI_SPILL_S32_RESTORE; 635 case 8: 636 return AMDGPU::SI_SPILL_S64_RESTORE; 637 case 16: 638 return AMDGPU::SI_SPILL_S128_RESTORE; 639 case 32: 640 return AMDGPU::SI_SPILL_S256_RESTORE; 641 case 64: 642 return AMDGPU::SI_SPILL_S512_RESTORE; 643 default: 644 llvm_unreachable("unknown register size"); 645 } 646 } 647 648 static unsigned getVGPRSpillRestoreOpcode(unsigned Size) { 649 switch (Size) { 650 case 4: 651 return AMDGPU::SI_SPILL_V32_RESTORE; 652 case 8: 653 return AMDGPU::SI_SPILL_V64_RESTORE; 654 case 12: 655 return AMDGPU::SI_SPILL_V96_RESTORE; 656 case 16: 657 return AMDGPU::SI_SPILL_V128_RESTORE; 658 case 32: 659 return AMDGPU::SI_SPILL_V256_RESTORE; 660 case 64: 661 return AMDGPU::SI_SPILL_V512_RESTORE; 662 default: 663 llvm_unreachable("unknown register size"); 664 } 665 } 666 667 void SIInstrInfo::loadRegFromStackSlot(MachineBasicBlock &MBB, 668 MachineBasicBlock::iterator MI, 669 unsigned DestReg, int FrameIndex, 670 const TargetRegisterClass *RC, 671 const TargetRegisterInfo *TRI) const { 672 MachineFunction *MF = MBB.getParent(); 673 const SIMachineFunctionInfo *MFI = MF->getInfo<SIMachineFunctionInfo>(); 674 MachineFrameInfo *FrameInfo = MF->getFrameInfo(); 675 DebugLoc DL = MBB.findDebugLoc(MI); 676 unsigned Align = FrameInfo->getObjectAlignment(FrameIndex); 677 unsigned Size = FrameInfo->getObjectSize(FrameIndex); 678 679 MachinePointerInfo PtrInfo 680 = MachinePointerInfo::getFixedStack(*MF, FrameIndex); 681 682 MachineMemOperand *MMO = MF->getMachineMemOperand( 683 PtrInfo, MachineMemOperand::MOLoad, Size, Align); 684 685 if (RI.isSGPRClass(RC)) { 686 // FIXME: Maybe this should not include a memoperand because it will be 687 // lowered to non-memory instructions. 688 unsigned Opcode = getSGPRSpillRestoreOpcode(RC->getSize()); 689 690 if (TargetRegisterInfo::isVirtualRegister(DestReg) && RC->getSize() == 4) { 691 // m0 may not be allowed for readlane. 692 MachineRegisterInfo &MRI = MF->getRegInfo(); 693 MRI.constrainRegClass(DestReg, &AMDGPU::SReg_32_XM0RegClass); 694 } 695 696 BuildMI(MBB, MI, DL, get(Opcode), DestReg) 697 .addFrameIndex(FrameIndex) // frame_idx 698 .addMemOperand(MMO); 699 700 return; 701 } 702 703 if (!ST.isVGPRSpillingEnabled(*MF->getFunction())) { 704 LLVMContext &Ctx = MF->getFunction()->getContext(); 705 Ctx.emitError("SIInstrInfo::loadRegFromStackSlot - Do not know how to" 706 " restore register"); 707 BuildMI(MBB, MI, DL, get(AMDGPU::IMPLICIT_DEF), DestReg); 708 709 return; 710 } 711 712 assert(RI.hasVGPRs(RC) && "Only VGPR spilling expected"); 713 714 unsigned Opcode = getVGPRSpillRestoreOpcode(RC->getSize()); 715 BuildMI(MBB, MI, DL, get(Opcode), DestReg) 716 .addFrameIndex(FrameIndex) // frame_idx 717 .addReg(MFI->getScratchRSrcReg()) // scratch_rsrc 718 .addReg(MFI->getScratchWaveOffsetReg()) // scratch_offset 719 .addImm(0) // offset 720 .addMemOperand(MMO); 721 } 722 723 /// \param @Offset Offset in bytes of the FrameIndex being spilled 724 unsigned SIInstrInfo::calculateLDSSpillAddress( 725 MachineBasicBlock &MBB, MachineInstr &MI, RegScavenger *RS, unsigned TmpReg, 726 unsigned FrameOffset, unsigned Size) const { 727 MachineFunction *MF = MBB.getParent(); 728 SIMachineFunctionInfo *MFI = MF->getInfo<SIMachineFunctionInfo>(); 729 const SISubtarget &ST = MF->getSubtarget<SISubtarget>(); 730 const SIRegisterInfo *TRI = ST.getRegisterInfo(); 731 DebugLoc DL = MBB.findDebugLoc(MI); 732 unsigned WorkGroupSize = MFI->getMaximumWorkGroupSize(*MF); 733 unsigned WavefrontSize = ST.getWavefrontSize(); 734 735 unsigned TIDReg = MFI->getTIDReg(); 736 if (!MFI->hasCalculatedTID()) { 737 MachineBasicBlock &Entry = MBB.getParent()->front(); 738 MachineBasicBlock::iterator Insert = Entry.front(); 739 DebugLoc DL = Insert->getDebugLoc(); 740 741 TIDReg = RI.findUnusedRegister(MF->getRegInfo(), &AMDGPU::VGPR_32RegClass); 742 if (TIDReg == AMDGPU::NoRegister) 743 return TIDReg; 744 745 if (!AMDGPU::isShader(MF->getFunction()->getCallingConv()) && 746 WorkGroupSize > WavefrontSize) { 747 748 unsigned TIDIGXReg 749 = TRI->getPreloadedValue(*MF, SIRegisterInfo::WORKGROUP_ID_X); 750 unsigned TIDIGYReg 751 = TRI->getPreloadedValue(*MF, SIRegisterInfo::WORKGROUP_ID_Y); 752 unsigned TIDIGZReg 753 = TRI->getPreloadedValue(*MF, SIRegisterInfo::WORKGROUP_ID_Z); 754 unsigned InputPtrReg = 755 TRI->getPreloadedValue(*MF, SIRegisterInfo::KERNARG_SEGMENT_PTR); 756 for (unsigned Reg : {TIDIGXReg, TIDIGYReg, TIDIGZReg}) { 757 if (!Entry.isLiveIn(Reg)) 758 Entry.addLiveIn(Reg); 759 } 760 761 RS->enterBasicBlock(Entry); 762 // FIXME: Can we scavenge an SReg_64 and access the subregs? 763 unsigned STmp0 = RS->scavengeRegister(&AMDGPU::SGPR_32RegClass, 0); 764 unsigned STmp1 = RS->scavengeRegister(&AMDGPU::SGPR_32RegClass, 0); 765 BuildMI(Entry, Insert, DL, get(AMDGPU::S_LOAD_DWORD_IMM), STmp0) 766 .addReg(InputPtrReg) 767 .addImm(SI::KernelInputOffsets::NGROUPS_Z); 768 BuildMI(Entry, Insert, DL, get(AMDGPU::S_LOAD_DWORD_IMM), STmp1) 769 .addReg(InputPtrReg) 770 .addImm(SI::KernelInputOffsets::NGROUPS_Y); 771 772 // NGROUPS.X * NGROUPS.Y 773 BuildMI(Entry, Insert, DL, get(AMDGPU::S_MUL_I32), STmp1) 774 .addReg(STmp1) 775 .addReg(STmp0); 776 // (NGROUPS.X * NGROUPS.Y) * TIDIG.X 777 BuildMI(Entry, Insert, DL, get(AMDGPU::V_MUL_U32_U24_e32), TIDReg) 778 .addReg(STmp1) 779 .addReg(TIDIGXReg); 780 // NGROUPS.Z * TIDIG.Y + (NGROUPS.X * NGROPUS.Y * TIDIG.X) 781 BuildMI(Entry, Insert, DL, get(AMDGPU::V_MAD_U32_U24), TIDReg) 782 .addReg(STmp0) 783 .addReg(TIDIGYReg) 784 .addReg(TIDReg); 785 // (NGROUPS.Z * TIDIG.Y + (NGROUPS.X * NGROPUS.Y * TIDIG.X)) + TIDIG.Z 786 BuildMI(Entry, Insert, DL, get(AMDGPU::V_ADD_I32_e32), TIDReg) 787 .addReg(TIDReg) 788 .addReg(TIDIGZReg); 789 } else { 790 // Get the wave id 791 BuildMI(Entry, Insert, DL, get(AMDGPU::V_MBCNT_LO_U32_B32_e64), 792 TIDReg) 793 .addImm(-1) 794 .addImm(0); 795 796 BuildMI(Entry, Insert, DL, get(AMDGPU::V_MBCNT_HI_U32_B32_e64), 797 TIDReg) 798 .addImm(-1) 799 .addReg(TIDReg); 800 } 801 802 BuildMI(Entry, Insert, DL, get(AMDGPU::V_LSHLREV_B32_e32), 803 TIDReg) 804 .addImm(2) 805 .addReg(TIDReg); 806 MFI->setTIDReg(TIDReg); 807 } 808 809 // Add FrameIndex to LDS offset 810 unsigned LDSOffset = MFI->LDSSize + (FrameOffset * WorkGroupSize); 811 BuildMI(MBB, MI, DL, get(AMDGPU::V_ADD_I32_e32), TmpReg) 812 .addImm(LDSOffset) 813 .addReg(TIDReg); 814 815 return TmpReg; 816 } 817 818 void SIInstrInfo::insertWaitStates(MachineBasicBlock &MBB, 819 MachineBasicBlock::iterator MI, 820 int Count) const { 821 DebugLoc DL = MBB.findDebugLoc(MI); 822 while (Count > 0) { 823 int Arg; 824 if (Count >= 8) 825 Arg = 7; 826 else 827 Arg = Count - 1; 828 Count -= 8; 829 BuildMI(MBB, MI, DL, get(AMDGPU::S_NOP)) 830 .addImm(Arg); 831 } 832 } 833 834 void SIInstrInfo::insertNoop(MachineBasicBlock &MBB, 835 MachineBasicBlock::iterator MI) const { 836 insertWaitStates(MBB, MI, 1); 837 } 838 839 unsigned SIInstrInfo::getNumWaitStates(const MachineInstr &MI) const { 840 switch (MI.getOpcode()) { 841 default: return 1; // FIXME: Do wait states equal cycles? 842 843 case AMDGPU::S_NOP: 844 return MI.getOperand(0).getImm() + 1; 845 } 846 } 847 848 bool SIInstrInfo::expandPostRAPseudo(MachineInstr &MI) const { 849 MachineBasicBlock &MBB = *MI.getParent(); 850 DebugLoc DL = MBB.findDebugLoc(MI); 851 switch (MI.getOpcode()) { 852 default: return AMDGPUInstrInfo::expandPostRAPseudo(MI); 853 854 case AMDGPU::V_MOV_B64_PSEUDO: { 855 unsigned Dst = MI.getOperand(0).getReg(); 856 unsigned DstLo = RI.getSubReg(Dst, AMDGPU::sub0); 857 unsigned DstHi = RI.getSubReg(Dst, AMDGPU::sub1); 858 859 const MachineOperand &SrcOp = MI.getOperand(1); 860 // FIXME: Will this work for 64-bit floating point immediates? 861 assert(!SrcOp.isFPImm()); 862 if (SrcOp.isImm()) { 863 APInt Imm(64, SrcOp.getImm()); 864 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DstLo) 865 .addImm(Imm.getLoBits(32).getZExtValue()) 866 .addReg(Dst, RegState::Implicit | RegState::Define); 867 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DstHi) 868 .addImm(Imm.getHiBits(32).getZExtValue()) 869 .addReg(Dst, RegState::Implicit | RegState::Define); 870 } else { 871 assert(SrcOp.isReg()); 872 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DstLo) 873 .addReg(RI.getSubReg(SrcOp.getReg(), AMDGPU::sub0)) 874 .addReg(Dst, RegState::Implicit | RegState::Define); 875 BuildMI(MBB, MI, DL, get(AMDGPU::V_MOV_B32_e32), DstHi) 876 .addReg(RI.getSubReg(SrcOp.getReg(), AMDGPU::sub1)) 877 .addReg(Dst, RegState::Implicit | RegState::Define); 878 } 879 MI.eraseFromParent(); 880 break; 881 } 882 883 case AMDGPU::V_CNDMASK_B64_PSEUDO: { 884 unsigned Dst = MI.getOperand(0).getReg(); 885 unsigned DstLo = RI.getSubReg(Dst, AMDGPU::sub0); 886 unsigned DstHi = RI.getSubReg(Dst, AMDGPU::sub1); 887 unsigned Src0 = MI.getOperand(1).getReg(); 888 unsigned Src1 = MI.getOperand(2).getReg(); 889 const MachineOperand &SrcCond = MI.getOperand(3); 890 891 BuildMI(MBB, MI, DL, get(AMDGPU::V_CNDMASK_B32_e64), DstLo) 892 .addReg(RI.getSubReg(Src0, AMDGPU::sub0)) 893 .addReg(RI.getSubReg(Src1, AMDGPU::sub0)) 894 .addReg(SrcCond.getReg()) 895 .addReg(Dst, RegState::Implicit | RegState::Define); 896 BuildMI(MBB, MI, DL, get(AMDGPU::V_CNDMASK_B32_e64), DstHi) 897 .addReg(RI.getSubReg(Src0, AMDGPU::sub1)) 898 .addReg(RI.getSubReg(Src1, AMDGPU::sub1)) 899 .addReg(SrcCond.getReg(), getKillRegState(SrcCond.isKill())) 900 .addReg(Dst, RegState::Implicit | RegState::Define); 901 MI.eraseFromParent(); 902 break; 903 } 904 905 case AMDGPU::SI_PC_ADD_REL_OFFSET: { 906 const SIRegisterInfo *TRI 907 = static_cast<const SIRegisterInfo *>(ST.getRegisterInfo()); 908 MachineFunction &MF = *MBB.getParent(); 909 unsigned Reg = MI.getOperand(0).getReg(); 910 unsigned RegLo = TRI->getSubReg(Reg, AMDGPU::sub0); 911 unsigned RegHi = TRI->getSubReg(Reg, AMDGPU::sub1); 912 913 // Create a bundle so these instructions won't be re-ordered by the 914 // post-RA scheduler. 915 MIBundleBuilder Bundler(MBB, MI); 916 Bundler.append(BuildMI(MF, DL, get(AMDGPU::S_GETPC_B64), Reg)); 917 918 // Add 32-bit offset from this instruction to the start of the 919 // constant data. 920 Bundler.append(BuildMI(MF, DL, get(AMDGPU::S_ADD_U32), RegLo) 921 .addReg(RegLo) 922 .addOperand(MI.getOperand(1))); 923 Bundler.append(BuildMI(MF, DL, get(AMDGPU::S_ADDC_U32), RegHi) 924 .addReg(RegHi) 925 .addImm(0)); 926 927 llvm::finalizeBundle(MBB, Bundler.begin()); 928 929 MI.eraseFromParent(); 930 break; 931 } 932 } 933 return true; 934 } 935 936 /// Commutes the operands in the given instruction. 937 /// The commutable operands are specified by their indices OpIdx0 and OpIdx1. 938 /// 939 /// Do not call this method for a non-commutable instruction or for 940 /// non-commutable pair of operand indices OpIdx0 and OpIdx1. 941 /// Even though the instruction is commutable, the method may still 942 /// fail to commute the operands, null pointer is returned in such cases. 943 MachineInstr *SIInstrInfo::commuteInstructionImpl(MachineInstr &MI, bool NewMI, 944 unsigned OpIdx0, 945 unsigned OpIdx1) const { 946 int CommutedOpcode = commuteOpcode(MI); 947 if (CommutedOpcode == -1) 948 return nullptr; 949 950 int Src0Idx = 951 AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::src0); 952 MachineOperand &Src0 = MI.getOperand(Src0Idx); 953 if (!Src0.isReg()) 954 return nullptr; 955 956 int Src1Idx = 957 AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::src1); 958 959 if ((OpIdx0 != static_cast<unsigned>(Src0Idx) || 960 OpIdx1 != static_cast<unsigned>(Src1Idx)) && 961 (OpIdx0 != static_cast<unsigned>(Src1Idx) || 962 OpIdx1 != static_cast<unsigned>(Src0Idx))) 963 return nullptr; 964 965 MachineOperand &Src1 = MI.getOperand(Src1Idx); 966 967 if (isVOP2(MI) || isVOPC(MI)) { 968 const MCInstrDesc &InstrDesc = MI.getDesc(); 969 // For VOP2 and VOPC instructions, any operand type is valid to use for 970 // src0. Make sure we can use the src0 as src1. 971 // 972 // We could be stricter here and only allow commuting if there is a reason 973 // to do so. i.e. if both operands are VGPRs there is no real benefit, 974 // although MachineCSE attempts to find matches by commuting. 975 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 976 if (!isLegalRegOperand(MRI, InstrDesc.OpInfo[Src1Idx], Src0)) 977 return nullptr; 978 } 979 980 MachineInstr *CommutedMI = &MI; 981 if (!Src1.isReg()) { 982 // Allow commuting instructions with Imm operands. 983 if (NewMI || !Src1.isImm() || (!isVOP2(MI) && !isVOP3(MI))) { 984 return nullptr; 985 } 986 // Be sure to copy the source modifiers to the right place. 987 if (MachineOperand *Src0Mods = 988 getNamedOperand(MI, AMDGPU::OpName::src0_modifiers)) { 989 MachineOperand *Src1Mods = 990 getNamedOperand(MI, AMDGPU::OpName::src1_modifiers); 991 992 int Src0ModsVal = Src0Mods->getImm(); 993 if (!Src1Mods && Src0ModsVal != 0) 994 return nullptr; 995 996 // XXX - This assert might be a lie. It might be useful to have a neg 997 // modifier with 0.0. 998 int Src1ModsVal = Src1Mods->getImm(); 999 assert((Src1ModsVal == 0) && "Not expecting modifiers with immediates"); 1000 1001 Src1Mods->setImm(Src0ModsVal); 1002 Src0Mods->setImm(Src1ModsVal); 1003 } 1004 1005 unsigned Reg = Src0.getReg(); 1006 unsigned SubReg = Src0.getSubReg(); 1007 if (Src1.isImm()) 1008 Src0.ChangeToImmediate(Src1.getImm()); 1009 else 1010 llvm_unreachable("Should only have immediates"); 1011 1012 Src1.ChangeToRegister(Reg, false); 1013 Src1.setSubReg(SubReg); 1014 } else { 1015 CommutedMI = 1016 TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx0, OpIdx1); 1017 } 1018 1019 if (CommutedMI) 1020 CommutedMI->setDesc(get(CommutedOpcode)); 1021 1022 return CommutedMI; 1023 } 1024 1025 // This needs to be implemented because the source modifiers may be inserted 1026 // between the true commutable operands, and the base 1027 // TargetInstrInfo::commuteInstruction uses it. 1028 bool SIInstrInfo::findCommutedOpIndices(MachineInstr &MI, unsigned &SrcOpIdx0, 1029 unsigned &SrcOpIdx1) const { 1030 const MCInstrDesc &MCID = MI.getDesc(); 1031 if (!MCID.isCommutable()) 1032 return false; 1033 1034 unsigned Opc = MI.getOpcode(); 1035 int Src0Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src0); 1036 if (Src0Idx == -1) 1037 return false; 1038 1039 // FIXME: Workaround TargetInstrInfo::commuteInstruction asserting on 1040 // immediate. Also, immediate src0 operand is not handled in 1041 // SIInstrInfo::commuteInstruction(); 1042 if (!MI.getOperand(Src0Idx).isReg()) 1043 return false; 1044 1045 int Src1Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src1); 1046 if (Src1Idx == -1) 1047 return false; 1048 1049 MachineOperand &Src1 = MI.getOperand(Src1Idx); 1050 if (Src1.isImm()) { 1051 // SIInstrInfo::commuteInstruction() does support commuting the immediate 1052 // operand src1 in 2 and 3 operand instructions. 1053 if (!isVOP2(MI.getOpcode()) && !isVOP3(MI.getOpcode())) 1054 return false; 1055 } else if (Src1.isReg()) { 1056 // If any source modifiers are set, the generic instruction commuting won't 1057 // understand how to copy the source modifiers. 1058 if (hasModifiersSet(MI, AMDGPU::OpName::src0_modifiers) || 1059 hasModifiersSet(MI, AMDGPU::OpName::src1_modifiers)) 1060 return false; 1061 } else 1062 return false; 1063 1064 return fixCommutedOpIndices(SrcOpIdx0, SrcOpIdx1, Src0Idx, Src1Idx); 1065 } 1066 1067 unsigned SIInstrInfo::getBranchOpcode(SIInstrInfo::BranchPredicate Cond) { 1068 switch (Cond) { 1069 case SIInstrInfo::SCC_TRUE: 1070 return AMDGPU::S_CBRANCH_SCC1; 1071 case SIInstrInfo::SCC_FALSE: 1072 return AMDGPU::S_CBRANCH_SCC0; 1073 case SIInstrInfo::VCCNZ: 1074 return AMDGPU::S_CBRANCH_VCCNZ; 1075 case SIInstrInfo::VCCZ: 1076 return AMDGPU::S_CBRANCH_VCCZ; 1077 case SIInstrInfo::EXECNZ: 1078 return AMDGPU::S_CBRANCH_EXECNZ; 1079 case SIInstrInfo::EXECZ: 1080 return AMDGPU::S_CBRANCH_EXECZ; 1081 default: 1082 llvm_unreachable("invalid branch predicate"); 1083 } 1084 } 1085 1086 SIInstrInfo::BranchPredicate SIInstrInfo::getBranchPredicate(unsigned Opcode) { 1087 switch (Opcode) { 1088 case AMDGPU::S_CBRANCH_SCC0: 1089 return SCC_FALSE; 1090 case AMDGPU::S_CBRANCH_SCC1: 1091 return SCC_TRUE; 1092 case AMDGPU::S_CBRANCH_VCCNZ: 1093 return VCCNZ; 1094 case AMDGPU::S_CBRANCH_VCCZ: 1095 return VCCZ; 1096 case AMDGPU::S_CBRANCH_EXECNZ: 1097 return EXECNZ; 1098 case AMDGPU::S_CBRANCH_EXECZ: 1099 return EXECZ; 1100 default: 1101 return INVALID_BR; 1102 } 1103 } 1104 1105 bool SIInstrInfo::analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, 1106 MachineBasicBlock *&FBB, 1107 SmallVectorImpl<MachineOperand> &Cond, 1108 bool AllowModify) const { 1109 MachineBasicBlock::iterator I = MBB.getFirstTerminator(); 1110 1111 if (I == MBB.end()) 1112 return false; 1113 1114 if (I->getOpcode() == AMDGPU::S_BRANCH) { 1115 // Unconditional Branch 1116 TBB = I->getOperand(0).getMBB(); 1117 return false; 1118 } 1119 1120 BranchPredicate Pred = getBranchPredicate(I->getOpcode()); 1121 if (Pred == INVALID_BR) 1122 return true; 1123 1124 MachineBasicBlock *CondBB = I->getOperand(0).getMBB(); 1125 Cond.push_back(MachineOperand::CreateImm(Pred)); 1126 1127 ++I; 1128 1129 if (I == MBB.end()) { 1130 // Conditional branch followed by fall-through. 1131 TBB = CondBB; 1132 return false; 1133 } 1134 1135 if (I->getOpcode() == AMDGPU::S_BRANCH) { 1136 TBB = CondBB; 1137 FBB = I->getOperand(0).getMBB(); 1138 return false; 1139 } 1140 1141 return true; 1142 } 1143 1144 unsigned SIInstrInfo::RemoveBranch(MachineBasicBlock &MBB) const { 1145 MachineBasicBlock::iterator I = MBB.getFirstTerminator(); 1146 1147 unsigned Count = 0; 1148 while (I != MBB.end()) { 1149 MachineBasicBlock::iterator Next = std::next(I); 1150 I->eraseFromParent(); 1151 ++Count; 1152 I = Next; 1153 } 1154 1155 return Count; 1156 } 1157 1158 unsigned SIInstrInfo::InsertBranch(MachineBasicBlock &MBB, 1159 MachineBasicBlock *TBB, 1160 MachineBasicBlock *FBB, 1161 ArrayRef<MachineOperand> Cond, 1162 const DebugLoc &DL) const { 1163 1164 if (!FBB && Cond.empty()) { 1165 BuildMI(&MBB, DL, get(AMDGPU::S_BRANCH)) 1166 .addMBB(TBB); 1167 return 1; 1168 } 1169 1170 assert(TBB && Cond[0].isImm()); 1171 1172 unsigned Opcode 1173 = getBranchOpcode(static_cast<BranchPredicate>(Cond[0].getImm())); 1174 1175 if (!FBB) { 1176 BuildMI(&MBB, DL, get(Opcode)) 1177 .addMBB(TBB); 1178 return 1; 1179 } 1180 1181 assert(TBB && FBB); 1182 1183 BuildMI(&MBB, DL, get(Opcode)) 1184 .addMBB(TBB); 1185 BuildMI(&MBB, DL, get(AMDGPU::S_BRANCH)) 1186 .addMBB(FBB); 1187 1188 return 2; 1189 } 1190 1191 bool SIInstrInfo::ReverseBranchCondition( 1192 SmallVectorImpl<MachineOperand> &Cond) const { 1193 assert(Cond.size() == 1); 1194 Cond[0].setImm(-Cond[0].getImm()); 1195 return false; 1196 } 1197 1198 static void removeModOperands(MachineInstr &MI) { 1199 unsigned Opc = MI.getOpcode(); 1200 int Src0ModIdx = AMDGPU::getNamedOperandIdx(Opc, 1201 AMDGPU::OpName::src0_modifiers); 1202 int Src1ModIdx = AMDGPU::getNamedOperandIdx(Opc, 1203 AMDGPU::OpName::src1_modifiers); 1204 int Src2ModIdx = AMDGPU::getNamedOperandIdx(Opc, 1205 AMDGPU::OpName::src2_modifiers); 1206 1207 MI.RemoveOperand(Src2ModIdx); 1208 MI.RemoveOperand(Src1ModIdx); 1209 MI.RemoveOperand(Src0ModIdx); 1210 } 1211 1212 // TODO: Maybe this should be removed this and custom fold everything in 1213 // SIFoldOperands? 1214 bool SIInstrInfo::FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, 1215 unsigned Reg, MachineRegisterInfo *MRI) const { 1216 if (!MRI->hasOneNonDBGUse(Reg)) 1217 return false; 1218 1219 unsigned Opc = UseMI.getOpcode(); 1220 if (Opc == AMDGPU::V_MAD_F32 || Opc == AMDGPU::V_MAC_F32_e64) { 1221 // Don't fold if we are using source modifiers. The new VOP2 instructions 1222 // don't have them. 1223 if (hasModifiersSet(UseMI, AMDGPU::OpName::src0_modifiers) || 1224 hasModifiersSet(UseMI, AMDGPU::OpName::src1_modifiers) || 1225 hasModifiersSet(UseMI, AMDGPU::OpName::src2_modifiers)) { 1226 return false; 1227 } 1228 1229 const MachineOperand &ImmOp = DefMI.getOperand(1); 1230 1231 // If this is a free constant, there's no reason to do this. 1232 // TODO: We could fold this here instead of letting SIFoldOperands do it 1233 // later. 1234 if (isInlineConstant(ImmOp, 4)) 1235 return false; 1236 1237 MachineOperand *Src0 = getNamedOperand(UseMI, AMDGPU::OpName::src0); 1238 MachineOperand *Src1 = getNamedOperand(UseMI, AMDGPU::OpName::src1); 1239 MachineOperand *Src2 = getNamedOperand(UseMI, AMDGPU::OpName::src2); 1240 1241 // Multiplied part is the constant: Use v_madmk_f32 1242 // We should only expect these to be on src0 due to canonicalizations. 1243 if (Src0->isReg() && Src0->getReg() == Reg) { 1244 if (!Src1->isReg() || RI.isSGPRClass(MRI->getRegClass(Src1->getReg()))) 1245 return false; 1246 1247 if (!Src2->isReg() || RI.isSGPRClass(MRI->getRegClass(Src2->getReg()))) 1248 return false; 1249 1250 // We need to swap operands 0 and 1 since madmk constant is at operand 1. 1251 1252 const int64_t Imm = DefMI.getOperand(1).getImm(); 1253 1254 // FIXME: This would be a lot easier if we could return a new instruction 1255 // instead of having to modify in place. 1256 1257 // Remove these first since they are at the end. 1258 UseMI.RemoveOperand( 1259 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::omod)); 1260 UseMI.RemoveOperand( 1261 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::clamp)); 1262 1263 unsigned Src1Reg = Src1->getReg(); 1264 unsigned Src1SubReg = Src1->getSubReg(); 1265 Src0->setReg(Src1Reg); 1266 Src0->setSubReg(Src1SubReg); 1267 Src0->setIsKill(Src1->isKill()); 1268 1269 if (Opc == AMDGPU::V_MAC_F32_e64) { 1270 UseMI.untieRegOperand( 1271 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src2)); 1272 } 1273 1274 Src1->ChangeToImmediate(Imm); 1275 1276 removeModOperands(UseMI); 1277 UseMI.setDesc(get(AMDGPU::V_MADMK_F32)); 1278 1279 bool DeleteDef = MRI->hasOneNonDBGUse(Reg); 1280 if (DeleteDef) 1281 DefMI.eraseFromParent(); 1282 1283 return true; 1284 } 1285 1286 // Added part is the constant: Use v_madak_f32 1287 if (Src2->isReg() && Src2->getReg() == Reg) { 1288 // Not allowed to use constant bus for another operand. 1289 // We can however allow an inline immediate as src0. 1290 if (!Src0->isImm() && 1291 (Src0->isReg() && RI.isSGPRClass(MRI->getRegClass(Src0->getReg())))) 1292 return false; 1293 1294 if (!Src1->isReg() || RI.isSGPRClass(MRI->getRegClass(Src1->getReg()))) 1295 return false; 1296 1297 const int64_t Imm = DefMI.getOperand(1).getImm(); 1298 1299 // FIXME: This would be a lot easier if we could return a new instruction 1300 // instead of having to modify in place. 1301 1302 // Remove these first since they are at the end. 1303 UseMI.RemoveOperand( 1304 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::omod)); 1305 UseMI.RemoveOperand( 1306 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::clamp)); 1307 1308 if (Opc == AMDGPU::V_MAC_F32_e64) { 1309 UseMI.untieRegOperand( 1310 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src2)); 1311 } 1312 1313 // ChangingToImmediate adds Src2 back to the instruction. 1314 Src2->ChangeToImmediate(Imm); 1315 1316 // These come before src2. 1317 removeModOperands(UseMI); 1318 UseMI.setDesc(get(AMDGPU::V_MADAK_F32)); 1319 1320 bool DeleteDef = MRI->hasOneNonDBGUse(Reg); 1321 if (DeleteDef) 1322 DefMI.eraseFromParent(); 1323 1324 return true; 1325 } 1326 } 1327 1328 return false; 1329 } 1330 1331 static bool offsetsDoNotOverlap(int WidthA, int OffsetA, 1332 int WidthB, int OffsetB) { 1333 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB; 1334 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA; 1335 int LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB; 1336 return LowOffset + LowWidth <= HighOffset; 1337 } 1338 1339 bool SIInstrInfo::checkInstOffsetsDoNotOverlap(MachineInstr &MIa, 1340 MachineInstr &MIb) const { 1341 unsigned BaseReg0, BaseReg1; 1342 int64_t Offset0, Offset1; 1343 1344 if (getMemOpBaseRegImmOfs(MIa, BaseReg0, Offset0, &RI) && 1345 getMemOpBaseRegImmOfs(MIb, BaseReg1, Offset1, &RI)) { 1346 1347 if (!MIa.hasOneMemOperand() || !MIb.hasOneMemOperand()) { 1348 // FIXME: Handle ds_read2 / ds_write2. 1349 return false; 1350 } 1351 unsigned Width0 = (*MIa.memoperands_begin())->getSize(); 1352 unsigned Width1 = (*MIb.memoperands_begin())->getSize(); 1353 if (BaseReg0 == BaseReg1 && 1354 offsetsDoNotOverlap(Width0, Offset0, Width1, Offset1)) { 1355 return true; 1356 } 1357 } 1358 1359 return false; 1360 } 1361 1362 bool SIInstrInfo::areMemAccessesTriviallyDisjoint(MachineInstr &MIa, 1363 MachineInstr &MIb, 1364 AliasAnalysis *AA) const { 1365 assert((MIa.mayLoad() || MIa.mayStore()) && 1366 "MIa must load from or modify a memory location"); 1367 assert((MIb.mayLoad() || MIb.mayStore()) && 1368 "MIb must load from or modify a memory location"); 1369 1370 if (MIa.hasUnmodeledSideEffects() || MIb.hasUnmodeledSideEffects()) 1371 return false; 1372 1373 // XXX - Can we relax this between address spaces? 1374 if (MIa.hasOrderedMemoryRef() || MIb.hasOrderedMemoryRef()) 1375 return false; 1376 1377 // TODO: Should we check the address space from the MachineMemOperand? That 1378 // would allow us to distinguish objects we know don't alias based on the 1379 // underlying address space, even if it was lowered to a different one, 1380 // e.g. private accesses lowered to use MUBUF instructions on a scratch 1381 // buffer. 1382 if (isDS(MIa)) { 1383 if (isDS(MIb)) 1384 return checkInstOffsetsDoNotOverlap(MIa, MIb); 1385 1386 return !isFLAT(MIb); 1387 } 1388 1389 if (isMUBUF(MIa) || isMTBUF(MIa)) { 1390 if (isMUBUF(MIb) || isMTBUF(MIb)) 1391 return checkInstOffsetsDoNotOverlap(MIa, MIb); 1392 1393 return !isFLAT(MIb) && !isSMRD(MIb); 1394 } 1395 1396 if (isSMRD(MIa)) { 1397 if (isSMRD(MIb)) 1398 return checkInstOffsetsDoNotOverlap(MIa, MIb); 1399 1400 return !isFLAT(MIb) && !isMUBUF(MIa) && !isMTBUF(MIa); 1401 } 1402 1403 if (isFLAT(MIa)) { 1404 if (isFLAT(MIb)) 1405 return checkInstOffsetsDoNotOverlap(MIa, MIb); 1406 1407 return false; 1408 } 1409 1410 return false; 1411 } 1412 1413 MachineInstr *SIInstrInfo::convertToThreeAddress(MachineFunction::iterator &MBB, 1414 MachineInstr &MI, 1415 LiveVariables *LV) const { 1416 1417 switch (MI.getOpcode()) { 1418 default: 1419 return nullptr; 1420 case AMDGPU::V_MAC_F32_e64: 1421 break; 1422 case AMDGPU::V_MAC_F32_e32: { 1423 const MachineOperand *Src0 = getNamedOperand(MI, AMDGPU::OpName::src0); 1424 if (Src0->isImm() && !isInlineConstant(*Src0, 4)) 1425 return nullptr; 1426 break; 1427 } 1428 } 1429 1430 const MachineOperand *Dst = getNamedOperand(MI, AMDGPU::OpName::vdst); 1431 const MachineOperand *Src0 = getNamedOperand(MI, AMDGPU::OpName::src0); 1432 const MachineOperand *Src1 = getNamedOperand(MI, AMDGPU::OpName::src1); 1433 const MachineOperand *Src2 = getNamedOperand(MI, AMDGPU::OpName::src2); 1434 1435 return BuildMI(*MBB, MI, MI.getDebugLoc(), get(AMDGPU::V_MAD_F32)) 1436 .addOperand(*Dst) 1437 .addImm(0) // Src0 mods 1438 .addOperand(*Src0) 1439 .addImm(0) // Src1 mods 1440 .addOperand(*Src1) 1441 .addImm(0) // Src mods 1442 .addOperand(*Src2) 1443 .addImm(0) // clamp 1444 .addImm(0); // omod 1445 } 1446 1447 bool SIInstrInfo::isSchedulingBoundary(const MachineInstr &MI, 1448 const MachineBasicBlock *MBB, 1449 const MachineFunction &MF) const { 1450 // XXX - Do we want the SP check in the base implementation? 1451 1452 // Target-independent instructions do not have an implicit-use of EXEC, even 1453 // when they operate on VGPRs. Treating EXEC modifications as scheduling 1454 // boundaries prevents incorrect movements of such instructions. 1455 return TargetInstrInfo::isSchedulingBoundary(MI, MBB, MF) || 1456 MI.modifiesRegister(AMDGPU::EXEC, &RI); 1457 } 1458 1459 bool SIInstrInfo::isInlineConstant(const APInt &Imm) const { 1460 int64_t SVal = Imm.getSExtValue(); 1461 if (SVal >= -16 && SVal <= 64) 1462 return true; 1463 1464 if (Imm.getBitWidth() == 64) { 1465 uint64_t Val = Imm.getZExtValue(); 1466 return (DoubleToBits(0.0) == Val) || 1467 (DoubleToBits(1.0) == Val) || 1468 (DoubleToBits(-1.0) == Val) || 1469 (DoubleToBits(0.5) == Val) || 1470 (DoubleToBits(-0.5) == Val) || 1471 (DoubleToBits(2.0) == Val) || 1472 (DoubleToBits(-2.0) == Val) || 1473 (DoubleToBits(4.0) == Val) || 1474 (DoubleToBits(-4.0) == Val); 1475 } 1476 1477 // The actual type of the operand does not seem to matter as long 1478 // as the bits match one of the inline immediate values. For example: 1479 // 1480 // -nan has the hexadecimal encoding of 0xfffffffe which is -2 in decimal, 1481 // so it is a legal inline immediate. 1482 // 1483 // 1065353216 has the hexadecimal encoding 0x3f800000 which is 1.0f in 1484 // floating-point, so it is a legal inline immediate. 1485 uint32_t Val = Imm.getZExtValue(); 1486 1487 return (FloatToBits(0.0f) == Val) || 1488 (FloatToBits(1.0f) == Val) || 1489 (FloatToBits(-1.0f) == Val) || 1490 (FloatToBits(0.5f) == Val) || 1491 (FloatToBits(-0.5f) == Val) || 1492 (FloatToBits(2.0f) == Val) || 1493 (FloatToBits(-2.0f) == Val) || 1494 (FloatToBits(4.0f) == Val) || 1495 (FloatToBits(-4.0f) == Val); 1496 } 1497 1498 bool SIInstrInfo::isInlineConstant(const MachineOperand &MO, 1499 unsigned OpSize) const { 1500 if (MO.isImm()) { 1501 // MachineOperand provides no way to tell the true operand size, since it 1502 // only records a 64-bit value. We need to know the size to determine if a 1503 // 32-bit floating point immediate bit pattern is legal for an integer 1504 // immediate. It would be for any 32-bit integer operand, but would not be 1505 // for a 64-bit one. 1506 1507 unsigned BitSize = 8 * OpSize; 1508 return isInlineConstant(APInt(BitSize, MO.getImm(), true)); 1509 } 1510 1511 return false; 1512 } 1513 1514 bool SIInstrInfo::isLiteralConstant(const MachineOperand &MO, 1515 unsigned OpSize) const { 1516 return MO.isImm() && !isInlineConstant(MO, OpSize); 1517 } 1518 1519 static bool compareMachineOp(const MachineOperand &Op0, 1520 const MachineOperand &Op1) { 1521 if (Op0.getType() != Op1.getType()) 1522 return false; 1523 1524 switch (Op0.getType()) { 1525 case MachineOperand::MO_Register: 1526 return Op0.getReg() == Op1.getReg(); 1527 case MachineOperand::MO_Immediate: 1528 return Op0.getImm() == Op1.getImm(); 1529 default: 1530 llvm_unreachable("Didn't expect to be comparing these operand types"); 1531 } 1532 } 1533 1534 bool SIInstrInfo::isImmOperandLegal(const MachineInstr &MI, unsigned OpNo, 1535 const MachineOperand &MO) const { 1536 const MCOperandInfo &OpInfo = get(MI.getOpcode()).OpInfo[OpNo]; 1537 1538 assert(MO.isImm() || MO.isTargetIndex() || MO.isFI()); 1539 1540 if (OpInfo.OperandType == MCOI::OPERAND_IMMEDIATE) 1541 return true; 1542 1543 if (OpInfo.RegClass < 0) 1544 return false; 1545 1546 unsigned OpSize = RI.getRegClass(OpInfo.RegClass)->getSize(); 1547 if (isLiteralConstant(MO, OpSize)) 1548 return RI.opCanUseLiteralConstant(OpInfo.OperandType); 1549 1550 return RI.opCanUseInlineConstant(OpInfo.OperandType); 1551 } 1552 1553 bool SIInstrInfo::hasVALU32BitEncoding(unsigned Opcode) const { 1554 int Op32 = AMDGPU::getVOPe32(Opcode); 1555 if (Op32 == -1) 1556 return false; 1557 1558 return pseudoToMCOpcode(Op32) != -1; 1559 } 1560 1561 bool SIInstrInfo::hasModifiers(unsigned Opcode) const { 1562 // The src0_modifier operand is present on all instructions 1563 // that have modifiers. 1564 1565 return AMDGPU::getNamedOperandIdx(Opcode, 1566 AMDGPU::OpName::src0_modifiers) != -1; 1567 } 1568 1569 bool SIInstrInfo::hasModifiersSet(const MachineInstr &MI, 1570 unsigned OpName) const { 1571 const MachineOperand *Mods = getNamedOperand(MI, OpName); 1572 return Mods && Mods->getImm(); 1573 } 1574 1575 bool SIInstrInfo::usesConstantBus(const MachineRegisterInfo &MRI, 1576 const MachineOperand &MO, 1577 unsigned OpSize) const { 1578 // Literal constants use the constant bus. 1579 if (isLiteralConstant(MO, OpSize)) 1580 return true; 1581 1582 if (!MO.isReg() || !MO.isUse()) 1583 return false; 1584 1585 if (TargetRegisterInfo::isVirtualRegister(MO.getReg())) 1586 return RI.isSGPRClass(MRI.getRegClass(MO.getReg())); 1587 1588 // FLAT_SCR is just an SGPR pair. 1589 if (!MO.isImplicit() && (MO.getReg() == AMDGPU::FLAT_SCR)) 1590 return true; 1591 1592 // EXEC register uses the constant bus. 1593 if (!MO.isImplicit() && MO.getReg() == AMDGPU::EXEC) 1594 return true; 1595 1596 // SGPRs use the constant bus 1597 return (MO.getReg() == AMDGPU::VCC || MO.getReg() == AMDGPU::M0 || 1598 (!MO.isImplicit() && 1599 (AMDGPU::SGPR_32RegClass.contains(MO.getReg()) || 1600 AMDGPU::SGPR_64RegClass.contains(MO.getReg())))); 1601 } 1602 1603 static unsigned findImplicitSGPRRead(const MachineInstr &MI) { 1604 for (const MachineOperand &MO : MI.implicit_operands()) { 1605 // We only care about reads. 1606 if (MO.isDef()) 1607 continue; 1608 1609 switch (MO.getReg()) { 1610 case AMDGPU::VCC: 1611 case AMDGPU::M0: 1612 case AMDGPU::FLAT_SCR: 1613 return MO.getReg(); 1614 1615 default: 1616 break; 1617 } 1618 } 1619 1620 return AMDGPU::NoRegister; 1621 } 1622 1623 static bool shouldReadExec(const MachineInstr &MI) { 1624 if (SIInstrInfo::isVALU(MI)) { 1625 switch (MI.getOpcode()) { 1626 case AMDGPU::V_READLANE_B32: 1627 case AMDGPU::V_READLANE_B32_si: 1628 case AMDGPU::V_READLANE_B32_vi: 1629 case AMDGPU::V_WRITELANE_B32: 1630 case AMDGPU::V_WRITELANE_B32_si: 1631 case AMDGPU::V_WRITELANE_B32_vi: 1632 return false; 1633 } 1634 1635 return true; 1636 } 1637 1638 if (SIInstrInfo::isGenericOpcode(MI.getOpcode()) || 1639 SIInstrInfo::isSALU(MI) || 1640 SIInstrInfo::isSMRD(MI)) 1641 return false; 1642 1643 return true; 1644 } 1645 1646 bool SIInstrInfo::verifyInstruction(const MachineInstr &MI, 1647 StringRef &ErrInfo) const { 1648 uint16_t Opcode = MI.getOpcode(); 1649 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 1650 int Src0Idx = AMDGPU::getNamedOperandIdx(Opcode, AMDGPU::OpName::src0); 1651 int Src1Idx = AMDGPU::getNamedOperandIdx(Opcode, AMDGPU::OpName::src1); 1652 int Src2Idx = AMDGPU::getNamedOperandIdx(Opcode, AMDGPU::OpName::src2); 1653 1654 // Make sure the number of operands is correct. 1655 const MCInstrDesc &Desc = get(Opcode); 1656 if (!Desc.isVariadic() && 1657 Desc.getNumOperands() != MI.getNumExplicitOperands()) { 1658 ErrInfo = "Instruction has wrong number of operands."; 1659 return false; 1660 } 1661 1662 // Make sure the register classes are correct. 1663 for (int i = 0, e = Desc.getNumOperands(); i != e; ++i) { 1664 if (MI.getOperand(i).isFPImm()) { 1665 ErrInfo = "FPImm Machine Operands are not supported. ISel should bitcast " 1666 "all fp values to integers."; 1667 return false; 1668 } 1669 1670 int RegClass = Desc.OpInfo[i].RegClass; 1671 1672 switch (Desc.OpInfo[i].OperandType) { 1673 case MCOI::OPERAND_REGISTER: 1674 if (MI.getOperand(i).isImm()) { 1675 ErrInfo = "Illegal immediate value for operand."; 1676 return false; 1677 } 1678 break; 1679 case AMDGPU::OPERAND_REG_IMM32: 1680 break; 1681 case AMDGPU::OPERAND_REG_INLINE_C: 1682 if (isLiteralConstant(MI.getOperand(i), 1683 RI.getRegClass(RegClass)->getSize())) { 1684 ErrInfo = "Illegal immediate value for operand."; 1685 return false; 1686 } 1687 break; 1688 case MCOI::OPERAND_IMMEDIATE: 1689 case AMDGPU::OPERAND_KIMM32: 1690 // Check if this operand is an immediate. 1691 // FrameIndex operands will be replaced by immediates, so they are 1692 // allowed. 1693 if (!MI.getOperand(i).isImm() && !MI.getOperand(i).isFI()) { 1694 ErrInfo = "Expected immediate, but got non-immediate"; 1695 return false; 1696 } 1697 // Fall-through 1698 default: 1699 continue; 1700 } 1701 1702 if (!MI.getOperand(i).isReg()) 1703 continue; 1704 1705 if (RegClass != -1) { 1706 unsigned Reg = MI.getOperand(i).getReg(); 1707 if (Reg == AMDGPU::NoRegister || 1708 TargetRegisterInfo::isVirtualRegister(Reg)) 1709 continue; 1710 1711 const TargetRegisterClass *RC = RI.getRegClass(RegClass); 1712 if (!RC->contains(Reg)) { 1713 ErrInfo = "Operand has incorrect register class."; 1714 return false; 1715 } 1716 } 1717 } 1718 1719 // Verify VOP* 1720 if (isVOP1(MI) || isVOP2(MI) || isVOP3(MI) || isVOPC(MI)) { 1721 // Only look at the true operands. Only a real operand can use the constant 1722 // bus, and we don't want to check pseudo-operands like the source modifier 1723 // flags. 1724 const int OpIndices[] = { Src0Idx, Src1Idx, Src2Idx }; 1725 1726 unsigned ConstantBusCount = 0; 1727 1728 if (AMDGPU::getNamedOperandIdx(Opcode, AMDGPU::OpName::imm) != -1) 1729 ++ConstantBusCount; 1730 1731 unsigned SGPRUsed = findImplicitSGPRRead(MI); 1732 if (SGPRUsed != AMDGPU::NoRegister) 1733 ++ConstantBusCount; 1734 1735 for (int OpIdx : OpIndices) { 1736 if (OpIdx == -1) 1737 break; 1738 const MachineOperand &MO = MI.getOperand(OpIdx); 1739 if (usesConstantBus(MRI, MO, getOpSize(Opcode, OpIdx))) { 1740 if (MO.isReg()) { 1741 if (MO.getReg() != SGPRUsed) 1742 ++ConstantBusCount; 1743 SGPRUsed = MO.getReg(); 1744 } else { 1745 ++ConstantBusCount; 1746 } 1747 } 1748 } 1749 if (ConstantBusCount > 1) { 1750 ErrInfo = "VOP* instruction uses the constant bus more than once"; 1751 return false; 1752 } 1753 } 1754 1755 // Verify misc. restrictions on specific instructions. 1756 if (Desc.getOpcode() == AMDGPU::V_DIV_SCALE_F32 || 1757 Desc.getOpcode() == AMDGPU::V_DIV_SCALE_F64) { 1758 const MachineOperand &Src0 = MI.getOperand(Src0Idx); 1759 const MachineOperand &Src1 = MI.getOperand(Src1Idx); 1760 const MachineOperand &Src2 = MI.getOperand(Src2Idx); 1761 if (Src0.isReg() && Src1.isReg() && Src2.isReg()) { 1762 if (!compareMachineOp(Src0, Src1) && 1763 !compareMachineOp(Src0, Src2)) { 1764 ErrInfo = "v_div_scale_{f32|f64} require src0 = src1 or src2"; 1765 return false; 1766 } 1767 } 1768 } 1769 1770 // Make sure we aren't losing exec uses in the td files. This mostly requires 1771 // being careful when using let Uses to try to add other use registers. 1772 if (shouldReadExec(MI)) { 1773 if (!MI.hasRegisterImplicitUseOperand(AMDGPU::EXEC)) { 1774 ErrInfo = "VALU instruction does not implicitly read exec mask"; 1775 return false; 1776 } 1777 } 1778 1779 return true; 1780 } 1781 1782 unsigned SIInstrInfo::getVALUOp(const MachineInstr &MI) { 1783 switch (MI.getOpcode()) { 1784 default: return AMDGPU::INSTRUCTION_LIST_END; 1785 case AMDGPU::REG_SEQUENCE: return AMDGPU::REG_SEQUENCE; 1786 case AMDGPU::COPY: return AMDGPU::COPY; 1787 case AMDGPU::PHI: return AMDGPU::PHI; 1788 case AMDGPU::INSERT_SUBREG: return AMDGPU::INSERT_SUBREG; 1789 case AMDGPU::S_MOV_B32: 1790 return MI.getOperand(1).isReg() ? 1791 AMDGPU::COPY : AMDGPU::V_MOV_B32_e32; 1792 case AMDGPU::S_ADD_I32: 1793 case AMDGPU::S_ADD_U32: return AMDGPU::V_ADD_I32_e32; 1794 case AMDGPU::S_ADDC_U32: return AMDGPU::V_ADDC_U32_e32; 1795 case AMDGPU::S_SUB_I32: 1796 case AMDGPU::S_SUB_U32: return AMDGPU::V_SUB_I32_e32; 1797 case AMDGPU::S_SUBB_U32: return AMDGPU::V_SUBB_U32_e32; 1798 case AMDGPU::S_MUL_I32: return AMDGPU::V_MUL_LO_I32; 1799 case AMDGPU::S_AND_B32: return AMDGPU::V_AND_B32_e32; 1800 case AMDGPU::S_OR_B32: return AMDGPU::V_OR_B32_e32; 1801 case AMDGPU::S_XOR_B32: return AMDGPU::V_XOR_B32_e32; 1802 case AMDGPU::S_MIN_I32: return AMDGPU::V_MIN_I32_e32; 1803 case AMDGPU::S_MIN_U32: return AMDGPU::V_MIN_U32_e32; 1804 case AMDGPU::S_MAX_I32: return AMDGPU::V_MAX_I32_e32; 1805 case AMDGPU::S_MAX_U32: return AMDGPU::V_MAX_U32_e32; 1806 case AMDGPU::S_ASHR_I32: return AMDGPU::V_ASHR_I32_e32; 1807 case AMDGPU::S_ASHR_I64: return AMDGPU::V_ASHR_I64; 1808 case AMDGPU::S_LSHL_B32: return AMDGPU::V_LSHL_B32_e32; 1809 case AMDGPU::S_LSHL_B64: return AMDGPU::V_LSHL_B64; 1810 case AMDGPU::S_LSHR_B32: return AMDGPU::V_LSHR_B32_e32; 1811 case AMDGPU::S_LSHR_B64: return AMDGPU::V_LSHR_B64; 1812 case AMDGPU::S_SEXT_I32_I8: return AMDGPU::V_BFE_I32; 1813 case AMDGPU::S_SEXT_I32_I16: return AMDGPU::V_BFE_I32; 1814 case AMDGPU::S_BFE_U32: return AMDGPU::V_BFE_U32; 1815 case AMDGPU::S_BFE_I32: return AMDGPU::V_BFE_I32; 1816 case AMDGPU::S_BFM_B32: return AMDGPU::V_BFM_B32_e64; 1817 case AMDGPU::S_BREV_B32: return AMDGPU::V_BFREV_B32_e32; 1818 case AMDGPU::S_NOT_B32: return AMDGPU::V_NOT_B32_e32; 1819 case AMDGPU::S_NOT_B64: return AMDGPU::V_NOT_B32_e32; 1820 case AMDGPU::S_CMP_EQ_I32: return AMDGPU::V_CMP_EQ_I32_e32; 1821 case AMDGPU::S_CMP_LG_I32: return AMDGPU::V_CMP_NE_I32_e32; 1822 case AMDGPU::S_CMP_GT_I32: return AMDGPU::V_CMP_GT_I32_e32; 1823 case AMDGPU::S_CMP_GE_I32: return AMDGPU::V_CMP_GE_I32_e32; 1824 case AMDGPU::S_CMP_LT_I32: return AMDGPU::V_CMP_LT_I32_e32; 1825 case AMDGPU::S_CMP_LE_I32: return AMDGPU::V_CMP_LE_I32_e32; 1826 case AMDGPU::S_CMP_EQ_U32: return AMDGPU::V_CMP_EQ_U32_e32; 1827 case AMDGPU::S_CMP_LG_U32: return AMDGPU::V_CMP_NE_U32_e32; 1828 case AMDGPU::S_CMP_GT_U32: return AMDGPU::V_CMP_GT_U32_e32; 1829 case AMDGPU::S_CMP_GE_U32: return AMDGPU::V_CMP_GE_U32_e32; 1830 case AMDGPU::S_CMP_LT_U32: return AMDGPU::V_CMP_LT_U32_e32; 1831 case AMDGPU::S_CMP_LE_U32: return AMDGPU::V_CMP_LE_U32_e32; 1832 case AMDGPU::S_BCNT1_I32_B32: return AMDGPU::V_BCNT_U32_B32_e64; 1833 case AMDGPU::S_FF1_I32_B32: return AMDGPU::V_FFBL_B32_e32; 1834 case AMDGPU::S_FLBIT_I32_B32: return AMDGPU::V_FFBH_U32_e32; 1835 case AMDGPU::S_FLBIT_I32: return AMDGPU::V_FFBH_I32_e64; 1836 case AMDGPU::S_CBRANCH_SCC0: return AMDGPU::S_CBRANCH_VCCZ; 1837 case AMDGPU::S_CBRANCH_SCC1: return AMDGPU::S_CBRANCH_VCCNZ; 1838 } 1839 } 1840 1841 bool SIInstrInfo::isSALUOpSupportedOnVALU(const MachineInstr &MI) const { 1842 return getVALUOp(MI) != AMDGPU::INSTRUCTION_LIST_END; 1843 } 1844 1845 const TargetRegisterClass *SIInstrInfo::getOpRegClass(const MachineInstr &MI, 1846 unsigned OpNo) const { 1847 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 1848 const MCInstrDesc &Desc = get(MI.getOpcode()); 1849 if (MI.isVariadic() || OpNo >= Desc.getNumOperands() || 1850 Desc.OpInfo[OpNo].RegClass == -1) { 1851 unsigned Reg = MI.getOperand(OpNo).getReg(); 1852 1853 if (TargetRegisterInfo::isVirtualRegister(Reg)) 1854 return MRI.getRegClass(Reg); 1855 return RI.getPhysRegClass(Reg); 1856 } 1857 1858 unsigned RCID = Desc.OpInfo[OpNo].RegClass; 1859 return RI.getRegClass(RCID); 1860 } 1861 1862 bool SIInstrInfo::canReadVGPR(const MachineInstr &MI, unsigned OpNo) const { 1863 switch (MI.getOpcode()) { 1864 case AMDGPU::COPY: 1865 case AMDGPU::REG_SEQUENCE: 1866 case AMDGPU::PHI: 1867 case AMDGPU::INSERT_SUBREG: 1868 return RI.hasVGPRs(getOpRegClass(MI, 0)); 1869 default: 1870 return RI.hasVGPRs(getOpRegClass(MI, OpNo)); 1871 } 1872 } 1873 1874 void SIInstrInfo::legalizeOpWithMove(MachineInstr &MI, unsigned OpIdx) const { 1875 MachineBasicBlock::iterator I = MI; 1876 MachineBasicBlock *MBB = MI.getParent(); 1877 MachineOperand &MO = MI.getOperand(OpIdx); 1878 MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo(); 1879 unsigned RCID = get(MI.getOpcode()).OpInfo[OpIdx].RegClass; 1880 const TargetRegisterClass *RC = RI.getRegClass(RCID); 1881 unsigned Opcode = AMDGPU::V_MOV_B32_e32; 1882 if (MO.isReg()) 1883 Opcode = AMDGPU::COPY; 1884 else if (RI.isSGPRClass(RC)) 1885 Opcode = AMDGPU::S_MOV_B32; 1886 1887 const TargetRegisterClass *VRC = RI.getEquivalentVGPRClass(RC); 1888 if (RI.getCommonSubClass(&AMDGPU::VReg_64RegClass, VRC)) 1889 VRC = &AMDGPU::VReg_64RegClass; 1890 else 1891 VRC = &AMDGPU::VGPR_32RegClass; 1892 1893 unsigned Reg = MRI.createVirtualRegister(VRC); 1894 DebugLoc DL = MBB->findDebugLoc(I); 1895 BuildMI(*MI.getParent(), I, DL, get(Opcode), Reg).addOperand(MO); 1896 MO.ChangeToRegister(Reg, false); 1897 } 1898 1899 unsigned SIInstrInfo::buildExtractSubReg(MachineBasicBlock::iterator MI, 1900 MachineRegisterInfo &MRI, 1901 MachineOperand &SuperReg, 1902 const TargetRegisterClass *SuperRC, 1903 unsigned SubIdx, 1904 const TargetRegisterClass *SubRC) 1905 const { 1906 MachineBasicBlock *MBB = MI->getParent(); 1907 DebugLoc DL = MI->getDebugLoc(); 1908 unsigned SubReg = MRI.createVirtualRegister(SubRC); 1909 1910 if (SuperReg.getSubReg() == AMDGPU::NoSubRegister) { 1911 BuildMI(*MBB, MI, DL, get(TargetOpcode::COPY), SubReg) 1912 .addReg(SuperReg.getReg(), 0, SubIdx); 1913 return SubReg; 1914 } 1915 1916 // Just in case the super register is itself a sub-register, copy it to a new 1917 // value so we don't need to worry about merging its subreg index with the 1918 // SubIdx passed to this function. The register coalescer should be able to 1919 // eliminate this extra copy. 1920 unsigned NewSuperReg = MRI.createVirtualRegister(SuperRC); 1921 1922 BuildMI(*MBB, MI, DL, get(TargetOpcode::COPY), NewSuperReg) 1923 .addReg(SuperReg.getReg(), 0, SuperReg.getSubReg()); 1924 1925 BuildMI(*MBB, MI, DL, get(TargetOpcode::COPY), SubReg) 1926 .addReg(NewSuperReg, 0, SubIdx); 1927 1928 return SubReg; 1929 } 1930 1931 MachineOperand SIInstrInfo::buildExtractSubRegOrImm( 1932 MachineBasicBlock::iterator MII, 1933 MachineRegisterInfo &MRI, 1934 MachineOperand &Op, 1935 const TargetRegisterClass *SuperRC, 1936 unsigned SubIdx, 1937 const TargetRegisterClass *SubRC) const { 1938 if (Op.isImm()) { 1939 // XXX - Is there a better way to do this? 1940 if (SubIdx == AMDGPU::sub0) 1941 return MachineOperand::CreateImm(Op.getImm() & 0xFFFFFFFF); 1942 if (SubIdx == AMDGPU::sub1) 1943 return MachineOperand::CreateImm(Op.getImm() >> 32); 1944 1945 llvm_unreachable("Unhandled register index for immediate"); 1946 } 1947 1948 unsigned SubReg = buildExtractSubReg(MII, MRI, Op, SuperRC, 1949 SubIdx, SubRC); 1950 return MachineOperand::CreateReg(SubReg, false); 1951 } 1952 1953 // Change the order of operands from (0, 1, 2) to (0, 2, 1) 1954 void SIInstrInfo::swapOperands(MachineInstr &Inst) const { 1955 assert(Inst.getNumExplicitOperands() == 3); 1956 MachineOperand Op1 = Inst.getOperand(1); 1957 Inst.RemoveOperand(1); 1958 Inst.addOperand(Op1); 1959 } 1960 1961 bool SIInstrInfo::isLegalRegOperand(const MachineRegisterInfo &MRI, 1962 const MCOperandInfo &OpInfo, 1963 const MachineOperand &MO) const { 1964 if (!MO.isReg()) 1965 return false; 1966 1967 unsigned Reg = MO.getReg(); 1968 const TargetRegisterClass *RC = 1969 TargetRegisterInfo::isVirtualRegister(Reg) ? 1970 MRI.getRegClass(Reg) : 1971 RI.getPhysRegClass(Reg); 1972 1973 const SIRegisterInfo *TRI = 1974 static_cast<const SIRegisterInfo*>(MRI.getTargetRegisterInfo()); 1975 RC = TRI->getSubRegClass(RC, MO.getSubReg()); 1976 1977 // In order to be legal, the common sub-class must be equal to the 1978 // class of the current operand. For example: 1979 // 1980 // v_mov_b32 s0 ; Operand defined as vsrc_32 1981 // ; RI.getCommonSubClass(s0,vsrc_32) = sgpr ; LEGAL 1982 // 1983 // s_sendmsg 0, s0 ; Operand defined as m0reg 1984 // ; RI.getCommonSubClass(s0,m0reg) = m0reg ; NOT LEGAL 1985 1986 return RI.getCommonSubClass(RC, RI.getRegClass(OpInfo.RegClass)) == RC; 1987 } 1988 1989 bool SIInstrInfo::isLegalVSrcOperand(const MachineRegisterInfo &MRI, 1990 const MCOperandInfo &OpInfo, 1991 const MachineOperand &MO) const { 1992 if (MO.isReg()) 1993 return isLegalRegOperand(MRI, OpInfo, MO); 1994 1995 // Handle non-register types that are treated like immediates. 1996 assert(MO.isImm() || MO.isTargetIndex() || MO.isFI()); 1997 return true; 1998 } 1999 2000 bool SIInstrInfo::isOperandLegal(const MachineInstr &MI, unsigned OpIdx, 2001 const MachineOperand *MO) const { 2002 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 2003 const MCInstrDesc &InstDesc = MI.getDesc(); 2004 const MCOperandInfo &OpInfo = InstDesc.OpInfo[OpIdx]; 2005 const TargetRegisterClass *DefinedRC = 2006 OpInfo.RegClass != -1 ? RI.getRegClass(OpInfo.RegClass) : nullptr; 2007 if (!MO) 2008 MO = &MI.getOperand(OpIdx); 2009 2010 if (isVALU(MI) && usesConstantBus(MRI, *MO, DefinedRC->getSize())) { 2011 2012 RegSubRegPair SGPRUsed; 2013 if (MO->isReg()) 2014 SGPRUsed = RegSubRegPair(MO->getReg(), MO->getSubReg()); 2015 2016 for (unsigned i = 0, e = MI.getNumOperands(); i != e; ++i) { 2017 if (i == OpIdx) 2018 continue; 2019 const MachineOperand &Op = MI.getOperand(i); 2020 if (Op.isReg()) { 2021 if ((Op.getReg() != SGPRUsed.Reg || Op.getSubReg() != SGPRUsed.SubReg) && 2022 usesConstantBus(MRI, Op, getOpSize(MI, i))) { 2023 return false; 2024 } 2025 } else if (InstDesc.OpInfo[i].OperandType == AMDGPU::OPERAND_KIMM32) { 2026 return false; 2027 } 2028 } 2029 } 2030 2031 if (MO->isReg()) { 2032 assert(DefinedRC); 2033 return isLegalRegOperand(MRI, OpInfo, *MO); 2034 } 2035 2036 // Handle non-register types that are treated like immediates. 2037 assert(MO->isImm() || MO->isTargetIndex() || MO->isFI()); 2038 2039 if (!DefinedRC) { 2040 // This operand expects an immediate. 2041 return true; 2042 } 2043 2044 return isImmOperandLegal(MI, OpIdx, *MO); 2045 } 2046 2047 void SIInstrInfo::legalizeOperandsVOP2(MachineRegisterInfo &MRI, 2048 MachineInstr &MI) const { 2049 unsigned Opc = MI.getOpcode(); 2050 const MCInstrDesc &InstrDesc = get(Opc); 2051 2052 int Src1Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src1); 2053 MachineOperand &Src1 = MI.getOperand(Src1Idx); 2054 2055 // If there is an implicit SGPR use such as VCC use for v_addc_u32/v_subb_u32 2056 // we need to only have one constant bus use. 2057 // 2058 // Note we do not need to worry about literal constants here. They are 2059 // disabled for the operand type for instructions because they will always 2060 // violate the one constant bus use rule. 2061 bool HasImplicitSGPR = findImplicitSGPRRead(MI) != AMDGPU::NoRegister; 2062 if (HasImplicitSGPR) { 2063 int Src0Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src0); 2064 MachineOperand &Src0 = MI.getOperand(Src0Idx); 2065 2066 if (Src0.isReg() && RI.isSGPRReg(MRI, Src0.getReg())) 2067 legalizeOpWithMove(MI, Src0Idx); 2068 } 2069 2070 // VOP2 src0 instructions support all operand types, so we don't need to check 2071 // their legality. If src1 is already legal, we don't need to do anything. 2072 if (isLegalRegOperand(MRI, InstrDesc.OpInfo[Src1Idx], Src1)) 2073 return; 2074 2075 // We do not use commuteInstruction here because it is too aggressive and will 2076 // commute if it is possible. We only want to commute here if it improves 2077 // legality. This can be called a fairly large number of times so don't waste 2078 // compile time pointlessly swapping and checking legality again. 2079 if (HasImplicitSGPR || !MI.isCommutable()) { 2080 legalizeOpWithMove(MI, Src1Idx); 2081 return; 2082 } 2083 2084 int Src0Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src0); 2085 MachineOperand &Src0 = MI.getOperand(Src0Idx); 2086 2087 // If src0 can be used as src1, commuting will make the operands legal. 2088 // Otherwise we have to give up and insert a move. 2089 // 2090 // TODO: Other immediate-like operand kinds could be commuted if there was a 2091 // MachineOperand::ChangeTo* for them. 2092 if ((!Src1.isImm() && !Src1.isReg()) || 2093 !isLegalRegOperand(MRI, InstrDesc.OpInfo[Src1Idx], Src0)) { 2094 legalizeOpWithMove(MI, Src1Idx); 2095 return; 2096 } 2097 2098 int CommutedOpc = commuteOpcode(MI); 2099 if (CommutedOpc == -1) { 2100 legalizeOpWithMove(MI, Src1Idx); 2101 return; 2102 } 2103 2104 MI.setDesc(get(CommutedOpc)); 2105 2106 unsigned Src0Reg = Src0.getReg(); 2107 unsigned Src0SubReg = Src0.getSubReg(); 2108 bool Src0Kill = Src0.isKill(); 2109 2110 if (Src1.isImm()) 2111 Src0.ChangeToImmediate(Src1.getImm()); 2112 else if (Src1.isReg()) { 2113 Src0.ChangeToRegister(Src1.getReg(), false, false, Src1.isKill()); 2114 Src0.setSubReg(Src1.getSubReg()); 2115 } else 2116 llvm_unreachable("Should only have register or immediate operands"); 2117 2118 Src1.ChangeToRegister(Src0Reg, false, false, Src0Kill); 2119 Src1.setSubReg(Src0SubReg); 2120 } 2121 2122 // Legalize VOP3 operands. Because all operand types are supported for any 2123 // operand, and since literal constants are not allowed and should never be 2124 // seen, we only need to worry about inserting copies if we use multiple SGPR 2125 // operands. 2126 void SIInstrInfo::legalizeOperandsVOP3(MachineRegisterInfo &MRI, 2127 MachineInstr &MI) const { 2128 unsigned Opc = MI.getOpcode(); 2129 2130 int VOP3Idx[3] = { 2131 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src0), 2132 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src1), 2133 AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src2) 2134 }; 2135 2136 // Find the one SGPR operand we are allowed to use. 2137 unsigned SGPRReg = findUsedSGPR(MI, VOP3Idx); 2138 2139 for (unsigned i = 0; i < 3; ++i) { 2140 int Idx = VOP3Idx[i]; 2141 if (Idx == -1) 2142 break; 2143 MachineOperand &MO = MI.getOperand(Idx); 2144 2145 // We should never see a VOP3 instruction with an illegal immediate operand. 2146 if (!MO.isReg()) 2147 continue; 2148 2149 if (!RI.isSGPRClass(MRI.getRegClass(MO.getReg()))) 2150 continue; // VGPRs are legal 2151 2152 if (SGPRReg == AMDGPU::NoRegister || SGPRReg == MO.getReg()) { 2153 SGPRReg = MO.getReg(); 2154 // We can use one SGPR in each VOP3 instruction. 2155 continue; 2156 } 2157 2158 // If we make it this far, then the operand is not legal and we must 2159 // legalize it. 2160 legalizeOpWithMove(MI, Idx); 2161 } 2162 } 2163 2164 unsigned SIInstrInfo::readlaneVGPRToSGPR(unsigned SrcReg, MachineInstr &UseMI, 2165 MachineRegisterInfo &MRI) const { 2166 const TargetRegisterClass *VRC = MRI.getRegClass(SrcReg); 2167 const TargetRegisterClass *SRC = RI.getEquivalentSGPRClass(VRC); 2168 unsigned DstReg = MRI.createVirtualRegister(SRC); 2169 unsigned SubRegs = VRC->getSize() / 4; 2170 2171 SmallVector<unsigned, 8> SRegs; 2172 for (unsigned i = 0; i < SubRegs; ++i) { 2173 unsigned SGPR = MRI.createVirtualRegister(&AMDGPU::SGPR_32RegClass); 2174 BuildMI(*UseMI.getParent(), UseMI, UseMI.getDebugLoc(), 2175 get(AMDGPU::V_READFIRSTLANE_B32), SGPR) 2176 .addReg(SrcReg, 0, RI.getSubRegFromChannel(i)); 2177 SRegs.push_back(SGPR); 2178 } 2179 2180 MachineInstrBuilder MIB = 2181 BuildMI(*UseMI.getParent(), UseMI, UseMI.getDebugLoc(), 2182 get(AMDGPU::REG_SEQUENCE), DstReg); 2183 for (unsigned i = 0; i < SubRegs; ++i) { 2184 MIB.addReg(SRegs[i]); 2185 MIB.addImm(RI.getSubRegFromChannel(i)); 2186 } 2187 return DstReg; 2188 } 2189 2190 void SIInstrInfo::legalizeOperandsSMRD(MachineRegisterInfo &MRI, 2191 MachineInstr &MI) const { 2192 2193 // If the pointer is store in VGPRs, then we need to move them to 2194 // SGPRs using v_readfirstlane. This is safe because we only select 2195 // loads with uniform pointers to SMRD instruction so we know the 2196 // pointer value is uniform. 2197 MachineOperand *SBase = getNamedOperand(MI, AMDGPU::OpName::sbase); 2198 if (SBase && !RI.isSGPRClass(MRI.getRegClass(SBase->getReg()))) { 2199 unsigned SGPR = readlaneVGPRToSGPR(SBase->getReg(), MI, MRI); 2200 SBase->setReg(SGPR); 2201 } 2202 } 2203 2204 void SIInstrInfo::legalizeOperands(MachineInstr &MI) const { 2205 MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 2206 2207 // Legalize VOP2 2208 if (isVOP2(MI) || isVOPC(MI)) { 2209 legalizeOperandsVOP2(MRI, MI); 2210 return; 2211 } 2212 2213 // Legalize VOP3 2214 if (isVOP3(MI)) { 2215 legalizeOperandsVOP3(MRI, MI); 2216 return; 2217 } 2218 2219 // Legalize SMRD 2220 if (isSMRD(MI)) { 2221 legalizeOperandsSMRD(MRI, MI); 2222 return; 2223 } 2224 2225 // Legalize REG_SEQUENCE and PHI 2226 // The register class of the operands much be the same type as the register 2227 // class of the output. 2228 if (MI.getOpcode() == AMDGPU::PHI) { 2229 const TargetRegisterClass *RC = nullptr, *SRC = nullptr, *VRC = nullptr; 2230 for (unsigned i = 1, e = MI.getNumOperands(); i != e; i += 2) { 2231 if (!MI.getOperand(i).isReg() || 2232 !TargetRegisterInfo::isVirtualRegister(MI.getOperand(i).getReg())) 2233 continue; 2234 const TargetRegisterClass *OpRC = 2235 MRI.getRegClass(MI.getOperand(i).getReg()); 2236 if (RI.hasVGPRs(OpRC)) { 2237 VRC = OpRC; 2238 } else { 2239 SRC = OpRC; 2240 } 2241 } 2242 2243 // If any of the operands are VGPR registers, then they all most be 2244 // otherwise we will create illegal VGPR->SGPR copies when legalizing 2245 // them. 2246 if (VRC || !RI.isSGPRClass(getOpRegClass(MI, 0))) { 2247 if (!VRC) { 2248 assert(SRC); 2249 VRC = RI.getEquivalentVGPRClass(SRC); 2250 } 2251 RC = VRC; 2252 } else { 2253 RC = SRC; 2254 } 2255 2256 // Update all the operands so they have the same type. 2257 for (unsigned I = 1, E = MI.getNumOperands(); I != E; I += 2) { 2258 MachineOperand &Op = MI.getOperand(I); 2259 if (!Op.isReg() || !TargetRegisterInfo::isVirtualRegister(Op.getReg())) 2260 continue; 2261 unsigned DstReg = MRI.createVirtualRegister(RC); 2262 2263 // MI is a PHI instruction. 2264 MachineBasicBlock *InsertBB = MI.getOperand(I + 1).getMBB(); 2265 MachineBasicBlock::iterator Insert = InsertBB->getFirstTerminator(); 2266 2267 BuildMI(*InsertBB, Insert, MI.getDebugLoc(), get(AMDGPU::COPY), DstReg) 2268 .addOperand(Op); 2269 Op.setReg(DstReg); 2270 } 2271 } 2272 2273 // REG_SEQUENCE doesn't really require operand legalization, but if one has a 2274 // VGPR dest type and SGPR sources, insert copies so all operands are 2275 // VGPRs. This seems to help operand folding / the register coalescer. 2276 if (MI.getOpcode() == AMDGPU::REG_SEQUENCE) { 2277 MachineBasicBlock *MBB = MI.getParent(); 2278 const TargetRegisterClass *DstRC = getOpRegClass(MI, 0); 2279 if (RI.hasVGPRs(DstRC)) { 2280 // Update all the operands so they are VGPR register classes. These may 2281 // not be the same register class because REG_SEQUENCE supports mixing 2282 // subregister index types e.g. sub0_sub1 + sub2 + sub3 2283 for (unsigned I = 1, E = MI.getNumOperands(); I != E; I += 2) { 2284 MachineOperand &Op = MI.getOperand(I); 2285 if (!Op.isReg() || !TargetRegisterInfo::isVirtualRegister(Op.getReg())) 2286 continue; 2287 2288 const TargetRegisterClass *OpRC = MRI.getRegClass(Op.getReg()); 2289 const TargetRegisterClass *VRC = RI.getEquivalentVGPRClass(OpRC); 2290 if (VRC == OpRC) 2291 continue; 2292 2293 unsigned DstReg = MRI.createVirtualRegister(VRC); 2294 2295 BuildMI(*MBB, MI, MI.getDebugLoc(), get(AMDGPU::COPY), DstReg) 2296 .addOperand(Op); 2297 2298 Op.setReg(DstReg); 2299 Op.setIsKill(); 2300 } 2301 } 2302 2303 return; 2304 } 2305 2306 // Legalize INSERT_SUBREG 2307 // src0 must have the same register class as dst 2308 if (MI.getOpcode() == AMDGPU::INSERT_SUBREG) { 2309 unsigned Dst = MI.getOperand(0).getReg(); 2310 unsigned Src0 = MI.getOperand(1).getReg(); 2311 const TargetRegisterClass *DstRC = MRI.getRegClass(Dst); 2312 const TargetRegisterClass *Src0RC = MRI.getRegClass(Src0); 2313 if (DstRC != Src0RC) { 2314 MachineBasicBlock &MBB = *MI.getParent(); 2315 unsigned NewSrc0 = MRI.createVirtualRegister(DstRC); 2316 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::COPY), NewSrc0) 2317 .addReg(Src0); 2318 MI.getOperand(1).setReg(NewSrc0); 2319 } 2320 return; 2321 } 2322 2323 // Legalize MIMG 2324 if (isMIMG(MI)) { 2325 MachineOperand *SRsrc = getNamedOperand(MI, AMDGPU::OpName::srsrc); 2326 if (SRsrc && !RI.isSGPRClass(MRI.getRegClass(SRsrc->getReg()))) { 2327 unsigned SGPR = readlaneVGPRToSGPR(SRsrc->getReg(), MI, MRI); 2328 SRsrc->setReg(SGPR); 2329 } 2330 2331 MachineOperand *SSamp = getNamedOperand(MI, AMDGPU::OpName::ssamp); 2332 if (SSamp && !RI.isSGPRClass(MRI.getRegClass(SSamp->getReg()))) { 2333 unsigned SGPR = readlaneVGPRToSGPR(SSamp->getReg(), MI, MRI); 2334 SSamp->setReg(SGPR); 2335 } 2336 return; 2337 } 2338 2339 // Legalize MUBUF* instructions 2340 // FIXME: If we start using the non-addr64 instructions for compute, we 2341 // may need to legalize them here. 2342 int SRsrcIdx = 2343 AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::srsrc); 2344 if (SRsrcIdx != -1) { 2345 // We have an MUBUF instruction 2346 MachineOperand *SRsrc = &MI.getOperand(SRsrcIdx); 2347 unsigned SRsrcRC = get(MI.getOpcode()).OpInfo[SRsrcIdx].RegClass; 2348 if (RI.getCommonSubClass(MRI.getRegClass(SRsrc->getReg()), 2349 RI.getRegClass(SRsrcRC))) { 2350 // The operands are legal. 2351 // FIXME: We may need to legalize operands besided srsrc. 2352 return; 2353 } 2354 2355 MachineBasicBlock &MBB = *MI.getParent(); 2356 2357 // Extract the ptr from the resource descriptor. 2358 unsigned SRsrcPtr = buildExtractSubReg(MI, MRI, *SRsrc, 2359 &AMDGPU::VReg_128RegClass, AMDGPU::sub0_sub1, &AMDGPU::VReg_64RegClass); 2360 2361 // Create an empty resource descriptor 2362 unsigned Zero64 = MRI.createVirtualRegister(&AMDGPU::SReg_64RegClass); 2363 unsigned SRsrcFormatLo = MRI.createVirtualRegister(&AMDGPU::SGPR_32RegClass); 2364 unsigned SRsrcFormatHi = MRI.createVirtualRegister(&AMDGPU::SGPR_32RegClass); 2365 unsigned NewSRsrc = MRI.createVirtualRegister(&AMDGPU::SReg_128RegClass); 2366 uint64_t RsrcDataFormat = getDefaultRsrcDataFormat(); 2367 2368 // Zero64 = 0 2369 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::S_MOV_B64), Zero64) 2370 .addImm(0); 2371 2372 // SRsrcFormatLo = RSRC_DATA_FORMAT{31-0} 2373 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::S_MOV_B32), SRsrcFormatLo) 2374 .addImm(RsrcDataFormat & 0xFFFFFFFF); 2375 2376 // SRsrcFormatHi = RSRC_DATA_FORMAT{63-32} 2377 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::S_MOV_B32), SRsrcFormatHi) 2378 .addImm(RsrcDataFormat >> 32); 2379 2380 // NewSRsrc = {Zero64, SRsrcFormat} 2381 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::REG_SEQUENCE), NewSRsrc) 2382 .addReg(Zero64) 2383 .addImm(AMDGPU::sub0_sub1) 2384 .addReg(SRsrcFormatLo) 2385 .addImm(AMDGPU::sub2) 2386 .addReg(SRsrcFormatHi) 2387 .addImm(AMDGPU::sub3); 2388 2389 MachineOperand *VAddr = getNamedOperand(MI, AMDGPU::OpName::vaddr); 2390 unsigned NewVAddr = MRI.createVirtualRegister(&AMDGPU::VReg_64RegClass); 2391 if (VAddr) { 2392 // This is already an ADDR64 instruction so we need to add the pointer 2393 // extracted from the resource descriptor to the current value of VAddr. 2394 unsigned NewVAddrLo = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2395 unsigned NewVAddrHi = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2396 2397 // NewVaddrLo = SRsrcPtr:sub0 + VAddr:sub0 2398 DebugLoc DL = MI.getDebugLoc(); 2399 BuildMI(MBB, MI, DL, get(AMDGPU::V_ADD_I32_e32), NewVAddrLo) 2400 .addReg(SRsrcPtr, 0, AMDGPU::sub0) 2401 .addReg(VAddr->getReg(), 0, AMDGPU::sub0); 2402 2403 // NewVaddrHi = SRsrcPtr:sub1 + VAddr:sub1 2404 BuildMI(MBB, MI, DL, get(AMDGPU::V_ADDC_U32_e32), NewVAddrHi) 2405 .addReg(SRsrcPtr, 0, AMDGPU::sub1) 2406 .addReg(VAddr->getReg(), 0, AMDGPU::sub1); 2407 2408 // NewVaddr = {NewVaddrHi, NewVaddrLo} 2409 BuildMI(MBB, MI, MI.getDebugLoc(), get(AMDGPU::REG_SEQUENCE), NewVAddr) 2410 .addReg(NewVAddrLo) 2411 .addImm(AMDGPU::sub0) 2412 .addReg(NewVAddrHi) 2413 .addImm(AMDGPU::sub1); 2414 } else { 2415 // This instructions is the _OFFSET variant, so we need to convert it to 2416 // ADDR64. 2417 assert(MBB.getParent()->getSubtarget<SISubtarget>().getGeneration() 2418 < SISubtarget::VOLCANIC_ISLANDS && 2419 "FIXME: Need to emit flat atomics here"); 2420 2421 MachineOperand *VData = getNamedOperand(MI, AMDGPU::OpName::vdata); 2422 MachineOperand *Offset = getNamedOperand(MI, AMDGPU::OpName::offset); 2423 MachineOperand *SOffset = getNamedOperand(MI, AMDGPU::OpName::soffset); 2424 unsigned Addr64Opcode = AMDGPU::getAddr64Inst(MI.getOpcode()); 2425 2426 // Atomics rith return have have an additional tied operand and are 2427 // missing some of the special bits. 2428 MachineOperand *VDataIn = getNamedOperand(MI, AMDGPU::OpName::vdata_in); 2429 MachineInstr *Addr64; 2430 2431 if (!VDataIn) { 2432 // Regular buffer load / store. 2433 MachineInstrBuilder MIB = 2434 BuildMI(MBB, MI, MI.getDebugLoc(), get(Addr64Opcode)) 2435 .addOperand(*VData) 2436 .addReg(AMDGPU::NoRegister) // Dummy value for vaddr. 2437 // This will be replaced later 2438 // with the new value of vaddr. 2439 .addOperand(*SRsrc) 2440 .addOperand(*SOffset) 2441 .addOperand(*Offset); 2442 2443 // Atomics do not have this operand. 2444 if (const MachineOperand *GLC = 2445 getNamedOperand(MI, AMDGPU::OpName::glc)) { 2446 MIB.addImm(GLC->getImm()); 2447 } 2448 2449 MIB.addImm(getNamedImmOperand(MI, AMDGPU::OpName::slc)); 2450 2451 if (const MachineOperand *TFE = 2452 getNamedOperand(MI, AMDGPU::OpName::tfe)) { 2453 MIB.addImm(TFE->getImm()); 2454 } 2455 2456 MIB.setMemRefs(MI.memoperands_begin(), MI.memoperands_end()); 2457 Addr64 = MIB; 2458 } else { 2459 // Atomics with return. 2460 Addr64 = BuildMI(MBB, MI, MI.getDebugLoc(), get(Addr64Opcode)) 2461 .addOperand(*VData) 2462 .addOperand(*VDataIn) 2463 .addReg(AMDGPU::NoRegister) // Dummy value for vaddr. 2464 // This will be replaced later 2465 // with the new value of vaddr. 2466 .addOperand(*SRsrc) 2467 .addOperand(*SOffset) 2468 .addOperand(*Offset) 2469 .addImm(getNamedImmOperand(MI, AMDGPU::OpName::slc)) 2470 .setMemRefs(MI.memoperands_begin(), MI.memoperands_end()); 2471 } 2472 2473 MI.removeFromParent(); 2474 2475 // NewVaddr = {NewVaddrHi, NewVaddrLo} 2476 BuildMI(MBB, Addr64, Addr64->getDebugLoc(), get(AMDGPU::REG_SEQUENCE), 2477 NewVAddr) 2478 .addReg(SRsrcPtr, 0, AMDGPU::sub0) 2479 .addImm(AMDGPU::sub0) 2480 .addReg(SRsrcPtr, 0, AMDGPU::sub1) 2481 .addImm(AMDGPU::sub1); 2482 2483 VAddr = getNamedOperand(*Addr64, AMDGPU::OpName::vaddr); 2484 SRsrc = getNamedOperand(*Addr64, AMDGPU::OpName::srsrc); 2485 } 2486 2487 // Update the instruction to use NewVaddr 2488 VAddr->setReg(NewVAddr); 2489 // Update the instruction to use NewSRsrc 2490 SRsrc->setReg(NewSRsrc); 2491 } 2492 } 2493 2494 void SIInstrInfo::moveToVALU(MachineInstr &TopInst) const { 2495 SmallVector<MachineInstr *, 128> Worklist; 2496 Worklist.push_back(&TopInst); 2497 2498 while (!Worklist.empty()) { 2499 MachineInstr &Inst = *Worklist.pop_back_val(); 2500 MachineBasicBlock *MBB = Inst.getParent(); 2501 MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo(); 2502 2503 unsigned Opcode = Inst.getOpcode(); 2504 unsigned NewOpcode = getVALUOp(Inst); 2505 2506 // Handle some special cases 2507 switch (Opcode) { 2508 default: 2509 break; 2510 case AMDGPU::S_AND_B64: 2511 splitScalar64BitBinaryOp(Worklist, Inst, AMDGPU::V_AND_B32_e64); 2512 Inst.eraseFromParent(); 2513 continue; 2514 2515 case AMDGPU::S_OR_B64: 2516 splitScalar64BitBinaryOp(Worklist, Inst, AMDGPU::V_OR_B32_e64); 2517 Inst.eraseFromParent(); 2518 continue; 2519 2520 case AMDGPU::S_XOR_B64: 2521 splitScalar64BitBinaryOp(Worklist, Inst, AMDGPU::V_XOR_B32_e64); 2522 Inst.eraseFromParent(); 2523 continue; 2524 2525 case AMDGPU::S_NOT_B64: 2526 splitScalar64BitUnaryOp(Worklist, Inst, AMDGPU::V_NOT_B32_e32); 2527 Inst.eraseFromParent(); 2528 continue; 2529 2530 case AMDGPU::S_BCNT1_I32_B64: 2531 splitScalar64BitBCNT(Worklist, Inst); 2532 Inst.eraseFromParent(); 2533 continue; 2534 2535 case AMDGPU::S_BFE_I64: { 2536 splitScalar64BitBFE(Worklist, Inst); 2537 Inst.eraseFromParent(); 2538 continue; 2539 } 2540 2541 case AMDGPU::S_LSHL_B32: 2542 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2543 NewOpcode = AMDGPU::V_LSHLREV_B32_e64; 2544 swapOperands(Inst); 2545 } 2546 break; 2547 case AMDGPU::S_ASHR_I32: 2548 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2549 NewOpcode = AMDGPU::V_ASHRREV_I32_e64; 2550 swapOperands(Inst); 2551 } 2552 break; 2553 case AMDGPU::S_LSHR_B32: 2554 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2555 NewOpcode = AMDGPU::V_LSHRREV_B32_e64; 2556 swapOperands(Inst); 2557 } 2558 break; 2559 case AMDGPU::S_LSHL_B64: 2560 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2561 NewOpcode = AMDGPU::V_LSHLREV_B64; 2562 swapOperands(Inst); 2563 } 2564 break; 2565 case AMDGPU::S_ASHR_I64: 2566 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2567 NewOpcode = AMDGPU::V_ASHRREV_I64; 2568 swapOperands(Inst); 2569 } 2570 break; 2571 case AMDGPU::S_LSHR_B64: 2572 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) { 2573 NewOpcode = AMDGPU::V_LSHRREV_B64; 2574 swapOperands(Inst); 2575 } 2576 break; 2577 2578 case AMDGPU::S_ABS_I32: 2579 lowerScalarAbs(Worklist, Inst); 2580 Inst.eraseFromParent(); 2581 continue; 2582 2583 case AMDGPU::S_CBRANCH_SCC0: 2584 case AMDGPU::S_CBRANCH_SCC1: 2585 // Clear unused bits of vcc 2586 BuildMI(*MBB, Inst, Inst.getDebugLoc(), get(AMDGPU::S_AND_B64), 2587 AMDGPU::VCC) 2588 .addReg(AMDGPU::EXEC) 2589 .addReg(AMDGPU::VCC); 2590 break; 2591 2592 case AMDGPU::S_BFE_U64: 2593 case AMDGPU::S_BFM_B64: 2594 llvm_unreachable("Moving this op to VALU not implemented"); 2595 } 2596 2597 if (NewOpcode == AMDGPU::INSTRUCTION_LIST_END) { 2598 // We cannot move this instruction to the VALU, so we should try to 2599 // legalize its operands instead. 2600 legalizeOperands(Inst); 2601 continue; 2602 } 2603 2604 // Use the new VALU Opcode. 2605 const MCInstrDesc &NewDesc = get(NewOpcode); 2606 Inst.setDesc(NewDesc); 2607 2608 // Remove any references to SCC. Vector instructions can't read from it, and 2609 // We're just about to add the implicit use / defs of VCC, and we don't want 2610 // both. 2611 for (unsigned i = Inst.getNumOperands() - 1; i > 0; --i) { 2612 MachineOperand &Op = Inst.getOperand(i); 2613 if (Op.isReg() && Op.getReg() == AMDGPU::SCC) { 2614 Inst.RemoveOperand(i); 2615 addSCCDefUsersToVALUWorklist(Inst, Worklist); 2616 } 2617 } 2618 2619 if (Opcode == AMDGPU::S_SEXT_I32_I8 || Opcode == AMDGPU::S_SEXT_I32_I16) { 2620 // We are converting these to a BFE, so we need to add the missing 2621 // operands for the size and offset. 2622 unsigned Size = (Opcode == AMDGPU::S_SEXT_I32_I8) ? 8 : 16; 2623 Inst.addOperand(MachineOperand::CreateImm(0)); 2624 Inst.addOperand(MachineOperand::CreateImm(Size)); 2625 2626 } else if (Opcode == AMDGPU::S_BCNT1_I32_B32) { 2627 // The VALU version adds the second operand to the result, so insert an 2628 // extra 0 operand. 2629 Inst.addOperand(MachineOperand::CreateImm(0)); 2630 } 2631 2632 Inst.addImplicitDefUseOperands(*Inst.getParent()->getParent()); 2633 2634 if (Opcode == AMDGPU::S_BFE_I32 || Opcode == AMDGPU::S_BFE_U32) { 2635 const MachineOperand &OffsetWidthOp = Inst.getOperand(2); 2636 // If we need to move this to VGPRs, we need to unpack the second operand 2637 // back into the 2 separate ones for bit offset and width. 2638 assert(OffsetWidthOp.isImm() && 2639 "Scalar BFE is only implemented for constant width and offset"); 2640 uint32_t Imm = OffsetWidthOp.getImm(); 2641 2642 uint32_t Offset = Imm & 0x3f; // Extract bits [5:0]. 2643 uint32_t BitWidth = (Imm & 0x7f0000) >> 16; // Extract bits [22:16]. 2644 Inst.RemoveOperand(2); // Remove old immediate. 2645 Inst.addOperand(MachineOperand::CreateImm(Offset)); 2646 Inst.addOperand(MachineOperand::CreateImm(BitWidth)); 2647 } 2648 2649 bool HasDst = Inst.getOperand(0).isReg() && Inst.getOperand(0).isDef(); 2650 unsigned NewDstReg = AMDGPU::NoRegister; 2651 if (HasDst) { 2652 // Update the destination register class. 2653 const TargetRegisterClass *NewDstRC = getDestEquivalentVGPRClass(Inst); 2654 if (!NewDstRC) 2655 continue; 2656 2657 unsigned DstReg = Inst.getOperand(0).getReg(); 2658 NewDstReg = MRI.createVirtualRegister(NewDstRC); 2659 MRI.replaceRegWith(DstReg, NewDstReg); 2660 } 2661 2662 // Legalize the operands 2663 legalizeOperands(Inst); 2664 2665 if (HasDst) 2666 addUsersToMoveToVALUWorklist(NewDstReg, MRI, Worklist); 2667 } 2668 } 2669 2670 void SIInstrInfo::lowerScalarAbs(SmallVectorImpl<MachineInstr *> &Worklist, 2671 MachineInstr &Inst) const { 2672 MachineBasicBlock &MBB = *Inst.getParent(); 2673 MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo(); 2674 MachineBasicBlock::iterator MII = Inst; 2675 DebugLoc DL = Inst.getDebugLoc(); 2676 2677 MachineOperand &Dest = Inst.getOperand(0); 2678 MachineOperand &Src = Inst.getOperand(1); 2679 unsigned TmpReg = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2680 unsigned ResultReg = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2681 2682 BuildMI(MBB, MII, DL, get(AMDGPU::V_SUB_I32_e32), TmpReg) 2683 .addImm(0) 2684 .addReg(Src.getReg()); 2685 2686 BuildMI(MBB, MII, DL, get(AMDGPU::V_MAX_I32_e64), ResultReg) 2687 .addReg(Src.getReg()) 2688 .addReg(TmpReg); 2689 2690 MRI.replaceRegWith(Dest.getReg(), ResultReg); 2691 addUsersToMoveToVALUWorklist(ResultReg, MRI, Worklist); 2692 } 2693 2694 void SIInstrInfo::splitScalar64BitUnaryOp( 2695 SmallVectorImpl<MachineInstr *> &Worklist, MachineInstr &Inst, 2696 unsigned Opcode) const { 2697 MachineBasicBlock &MBB = *Inst.getParent(); 2698 MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo(); 2699 2700 MachineOperand &Dest = Inst.getOperand(0); 2701 MachineOperand &Src0 = Inst.getOperand(1); 2702 DebugLoc DL = Inst.getDebugLoc(); 2703 2704 MachineBasicBlock::iterator MII = Inst; 2705 2706 const MCInstrDesc &InstDesc = get(Opcode); 2707 const TargetRegisterClass *Src0RC = Src0.isReg() ? 2708 MRI.getRegClass(Src0.getReg()) : 2709 &AMDGPU::SGPR_32RegClass; 2710 2711 const TargetRegisterClass *Src0SubRC = RI.getSubRegClass(Src0RC, AMDGPU::sub0); 2712 2713 MachineOperand SrcReg0Sub0 = buildExtractSubRegOrImm(MII, MRI, Src0, Src0RC, 2714 AMDGPU::sub0, Src0SubRC); 2715 2716 const TargetRegisterClass *DestRC = MRI.getRegClass(Dest.getReg()); 2717 const TargetRegisterClass *NewDestRC = RI.getEquivalentVGPRClass(DestRC); 2718 const TargetRegisterClass *NewDestSubRC = RI.getSubRegClass(NewDestRC, AMDGPU::sub0); 2719 2720 unsigned DestSub0 = MRI.createVirtualRegister(NewDestSubRC); 2721 BuildMI(MBB, MII, DL, InstDesc, DestSub0) 2722 .addOperand(SrcReg0Sub0); 2723 2724 MachineOperand SrcReg0Sub1 = buildExtractSubRegOrImm(MII, MRI, Src0, Src0RC, 2725 AMDGPU::sub1, Src0SubRC); 2726 2727 unsigned DestSub1 = MRI.createVirtualRegister(NewDestSubRC); 2728 BuildMI(MBB, MII, DL, InstDesc, DestSub1) 2729 .addOperand(SrcReg0Sub1); 2730 2731 unsigned FullDestReg = MRI.createVirtualRegister(NewDestRC); 2732 BuildMI(MBB, MII, DL, get(TargetOpcode::REG_SEQUENCE), FullDestReg) 2733 .addReg(DestSub0) 2734 .addImm(AMDGPU::sub0) 2735 .addReg(DestSub1) 2736 .addImm(AMDGPU::sub1); 2737 2738 MRI.replaceRegWith(Dest.getReg(), FullDestReg); 2739 2740 // We don't need to legalizeOperands here because for a single operand, src0 2741 // will support any kind of input. 2742 2743 // Move all users of this moved value. 2744 addUsersToMoveToVALUWorklist(FullDestReg, MRI, Worklist); 2745 } 2746 2747 void SIInstrInfo::splitScalar64BitBinaryOp( 2748 SmallVectorImpl<MachineInstr *> &Worklist, MachineInstr &Inst, 2749 unsigned Opcode) const { 2750 MachineBasicBlock &MBB = *Inst.getParent(); 2751 MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo(); 2752 2753 MachineOperand &Dest = Inst.getOperand(0); 2754 MachineOperand &Src0 = Inst.getOperand(1); 2755 MachineOperand &Src1 = Inst.getOperand(2); 2756 DebugLoc DL = Inst.getDebugLoc(); 2757 2758 MachineBasicBlock::iterator MII = Inst; 2759 2760 const MCInstrDesc &InstDesc = get(Opcode); 2761 const TargetRegisterClass *Src0RC = Src0.isReg() ? 2762 MRI.getRegClass(Src0.getReg()) : 2763 &AMDGPU::SGPR_32RegClass; 2764 2765 const TargetRegisterClass *Src0SubRC = RI.getSubRegClass(Src0RC, AMDGPU::sub0); 2766 const TargetRegisterClass *Src1RC = Src1.isReg() ? 2767 MRI.getRegClass(Src1.getReg()) : 2768 &AMDGPU::SGPR_32RegClass; 2769 2770 const TargetRegisterClass *Src1SubRC = RI.getSubRegClass(Src1RC, AMDGPU::sub0); 2771 2772 MachineOperand SrcReg0Sub0 = buildExtractSubRegOrImm(MII, MRI, Src0, Src0RC, 2773 AMDGPU::sub0, Src0SubRC); 2774 MachineOperand SrcReg1Sub0 = buildExtractSubRegOrImm(MII, MRI, Src1, Src1RC, 2775 AMDGPU::sub0, Src1SubRC); 2776 2777 const TargetRegisterClass *DestRC = MRI.getRegClass(Dest.getReg()); 2778 const TargetRegisterClass *NewDestRC = RI.getEquivalentVGPRClass(DestRC); 2779 const TargetRegisterClass *NewDestSubRC = RI.getSubRegClass(NewDestRC, AMDGPU::sub0); 2780 2781 unsigned DestSub0 = MRI.createVirtualRegister(NewDestSubRC); 2782 MachineInstr &LoHalf = *BuildMI(MBB, MII, DL, InstDesc, DestSub0) 2783 .addOperand(SrcReg0Sub0) 2784 .addOperand(SrcReg1Sub0); 2785 2786 MachineOperand SrcReg0Sub1 = buildExtractSubRegOrImm(MII, MRI, Src0, Src0RC, 2787 AMDGPU::sub1, Src0SubRC); 2788 MachineOperand SrcReg1Sub1 = buildExtractSubRegOrImm(MII, MRI, Src1, Src1RC, 2789 AMDGPU::sub1, Src1SubRC); 2790 2791 unsigned DestSub1 = MRI.createVirtualRegister(NewDestSubRC); 2792 MachineInstr &HiHalf = *BuildMI(MBB, MII, DL, InstDesc, DestSub1) 2793 .addOperand(SrcReg0Sub1) 2794 .addOperand(SrcReg1Sub1); 2795 2796 unsigned FullDestReg = MRI.createVirtualRegister(NewDestRC); 2797 BuildMI(MBB, MII, DL, get(TargetOpcode::REG_SEQUENCE), FullDestReg) 2798 .addReg(DestSub0) 2799 .addImm(AMDGPU::sub0) 2800 .addReg(DestSub1) 2801 .addImm(AMDGPU::sub1); 2802 2803 MRI.replaceRegWith(Dest.getReg(), FullDestReg); 2804 2805 // Try to legalize the operands in case we need to swap the order to keep it 2806 // valid. 2807 legalizeOperands(LoHalf); 2808 legalizeOperands(HiHalf); 2809 2810 // Move all users of this moved vlaue. 2811 addUsersToMoveToVALUWorklist(FullDestReg, MRI, Worklist); 2812 } 2813 2814 void SIInstrInfo::splitScalar64BitBCNT( 2815 SmallVectorImpl<MachineInstr *> &Worklist, MachineInstr &Inst) const { 2816 MachineBasicBlock &MBB = *Inst.getParent(); 2817 MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo(); 2818 2819 MachineBasicBlock::iterator MII = Inst; 2820 DebugLoc DL = Inst.getDebugLoc(); 2821 2822 MachineOperand &Dest = Inst.getOperand(0); 2823 MachineOperand &Src = Inst.getOperand(1); 2824 2825 const MCInstrDesc &InstDesc = get(AMDGPU::V_BCNT_U32_B32_e64); 2826 const TargetRegisterClass *SrcRC = Src.isReg() ? 2827 MRI.getRegClass(Src.getReg()) : 2828 &AMDGPU::SGPR_32RegClass; 2829 2830 unsigned MidReg = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2831 unsigned ResultReg = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2832 2833 const TargetRegisterClass *SrcSubRC = RI.getSubRegClass(SrcRC, AMDGPU::sub0); 2834 2835 MachineOperand SrcRegSub0 = buildExtractSubRegOrImm(MII, MRI, Src, SrcRC, 2836 AMDGPU::sub0, SrcSubRC); 2837 MachineOperand SrcRegSub1 = buildExtractSubRegOrImm(MII, MRI, Src, SrcRC, 2838 AMDGPU::sub1, SrcSubRC); 2839 2840 BuildMI(MBB, MII, DL, InstDesc, MidReg) 2841 .addOperand(SrcRegSub0) 2842 .addImm(0); 2843 2844 BuildMI(MBB, MII, DL, InstDesc, ResultReg) 2845 .addOperand(SrcRegSub1) 2846 .addReg(MidReg); 2847 2848 MRI.replaceRegWith(Dest.getReg(), ResultReg); 2849 2850 // We don't need to legalize operands here. src0 for etiher instruction can be 2851 // an SGPR, and the second input is unused or determined here. 2852 addUsersToMoveToVALUWorklist(ResultReg, MRI, Worklist); 2853 } 2854 2855 void SIInstrInfo::splitScalar64BitBFE(SmallVectorImpl<MachineInstr *> &Worklist, 2856 MachineInstr &Inst) const { 2857 MachineBasicBlock &MBB = *Inst.getParent(); 2858 MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo(); 2859 MachineBasicBlock::iterator MII = Inst; 2860 DebugLoc DL = Inst.getDebugLoc(); 2861 2862 MachineOperand &Dest = Inst.getOperand(0); 2863 uint32_t Imm = Inst.getOperand(2).getImm(); 2864 uint32_t Offset = Imm & 0x3f; // Extract bits [5:0]. 2865 uint32_t BitWidth = (Imm & 0x7f0000) >> 16; // Extract bits [22:16]. 2866 2867 (void) Offset; 2868 2869 // Only sext_inreg cases handled. 2870 assert(Inst.getOpcode() == AMDGPU::S_BFE_I64 && BitWidth <= 32 && 2871 Offset == 0 && "Not implemented"); 2872 2873 if (BitWidth < 32) { 2874 unsigned MidRegLo = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2875 unsigned MidRegHi = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2876 unsigned ResultReg = MRI.createVirtualRegister(&AMDGPU::VReg_64RegClass); 2877 2878 BuildMI(MBB, MII, DL, get(AMDGPU::V_BFE_I32), MidRegLo) 2879 .addReg(Inst.getOperand(1).getReg(), 0, AMDGPU::sub0) 2880 .addImm(0) 2881 .addImm(BitWidth); 2882 2883 BuildMI(MBB, MII, DL, get(AMDGPU::V_ASHRREV_I32_e32), MidRegHi) 2884 .addImm(31) 2885 .addReg(MidRegLo); 2886 2887 BuildMI(MBB, MII, DL, get(TargetOpcode::REG_SEQUENCE), ResultReg) 2888 .addReg(MidRegLo) 2889 .addImm(AMDGPU::sub0) 2890 .addReg(MidRegHi) 2891 .addImm(AMDGPU::sub1); 2892 2893 MRI.replaceRegWith(Dest.getReg(), ResultReg); 2894 addUsersToMoveToVALUWorklist(ResultReg, MRI, Worklist); 2895 return; 2896 } 2897 2898 MachineOperand &Src = Inst.getOperand(1); 2899 unsigned TmpReg = MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass); 2900 unsigned ResultReg = MRI.createVirtualRegister(&AMDGPU::VReg_64RegClass); 2901 2902 BuildMI(MBB, MII, DL, get(AMDGPU::V_ASHRREV_I32_e64), TmpReg) 2903 .addImm(31) 2904 .addReg(Src.getReg(), 0, AMDGPU::sub0); 2905 2906 BuildMI(MBB, MII, DL, get(TargetOpcode::REG_SEQUENCE), ResultReg) 2907 .addReg(Src.getReg(), 0, AMDGPU::sub0) 2908 .addImm(AMDGPU::sub0) 2909 .addReg(TmpReg) 2910 .addImm(AMDGPU::sub1); 2911 2912 MRI.replaceRegWith(Dest.getReg(), ResultReg); 2913 addUsersToMoveToVALUWorklist(ResultReg, MRI, Worklist); 2914 } 2915 2916 void SIInstrInfo::addUsersToMoveToVALUWorklist( 2917 unsigned DstReg, 2918 MachineRegisterInfo &MRI, 2919 SmallVectorImpl<MachineInstr *> &Worklist) const { 2920 for (MachineRegisterInfo::use_iterator I = MRI.use_begin(DstReg), 2921 E = MRI.use_end(); I != E; ++I) { 2922 MachineInstr &UseMI = *I->getParent(); 2923 if (!canReadVGPR(UseMI, I.getOperandNo())) { 2924 Worklist.push_back(&UseMI); 2925 } 2926 } 2927 } 2928 2929 void SIInstrInfo::addSCCDefUsersToVALUWorklist( 2930 MachineInstr &SCCDefInst, SmallVectorImpl<MachineInstr *> &Worklist) const { 2931 // This assumes that all the users of SCC are in the same block 2932 // as the SCC def. 2933 for (MachineInstr &MI : 2934 llvm::make_range(MachineBasicBlock::iterator(SCCDefInst), 2935 SCCDefInst.getParent()->end())) { 2936 // Exit if we find another SCC def. 2937 if (MI.findRegisterDefOperandIdx(AMDGPU::SCC) != -1) 2938 return; 2939 2940 if (MI.findRegisterUseOperandIdx(AMDGPU::SCC) != -1) 2941 Worklist.push_back(&MI); 2942 } 2943 } 2944 2945 const TargetRegisterClass *SIInstrInfo::getDestEquivalentVGPRClass( 2946 const MachineInstr &Inst) const { 2947 const TargetRegisterClass *NewDstRC = getOpRegClass(Inst, 0); 2948 2949 switch (Inst.getOpcode()) { 2950 // For target instructions, getOpRegClass just returns the virtual register 2951 // class associated with the operand, so we need to find an equivalent VGPR 2952 // register class in order to move the instruction to the VALU. 2953 case AMDGPU::COPY: 2954 case AMDGPU::PHI: 2955 case AMDGPU::REG_SEQUENCE: 2956 case AMDGPU::INSERT_SUBREG: 2957 if (RI.hasVGPRs(NewDstRC)) 2958 return nullptr; 2959 2960 NewDstRC = RI.getEquivalentVGPRClass(NewDstRC); 2961 if (!NewDstRC) 2962 return nullptr; 2963 return NewDstRC; 2964 default: 2965 return NewDstRC; 2966 } 2967 } 2968 2969 // Find the one SGPR operand we are allowed to use. 2970 unsigned SIInstrInfo::findUsedSGPR(const MachineInstr &MI, 2971 int OpIndices[3]) const { 2972 const MCInstrDesc &Desc = MI.getDesc(); 2973 2974 // Find the one SGPR operand we are allowed to use. 2975 // 2976 // First we need to consider the instruction's operand requirements before 2977 // legalizing. Some operands are required to be SGPRs, such as implicit uses 2978 // of VCC, but we are still bound by the constant bus requirement to only use 2979 // one. 2980 // 2981 // If the operand's class is an SGPR, we can never move it. 2982 2983 unsigned SGPRReg = findImplicitSGPRRead(MI); 2984 if (SGPRReg != AMDGPU::NoRegister) 2985 return SGPRReg; 2986 2987 unsigned UsedSGPRs[3] = { AMDGPU::NoRegister }; 2988 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo(); 2989 2990 for (unsigned i = 0; i < 3; ++i) { 2991 int Idx = OpIndices[i]; 2992 if (Idx == -1) 2993 break; 2994 2995 const MachineOperand &MO = MI.getOperand(Idx); 2996 if (!MO.isReg()) 2997 continue; 2998 2999 // Is this operand statically required to be an SGPR based on the operand 3000 // constraints? 3001 const TargetRegisterClass *OpRC = RI.getRegClass(Desc.OpInfo[Idx].RegClass); 3002 bool IsRequiredSGPR = RI.isSGPRClass(OpRC); 3003 if (IsRequiredSGPR) 3004 return MO.getReg(); 3005 3006 // If this could be a VGPR or an SGPR, Check the dynamic register class. 3007 unsigned Reg = MO.getReg(); 3008 const TargetRegisterClass *RegRC = MRI.getRegClass(Reg); 3009 if (RI.isSGPRClass(RegRC)) 3010 UsedSGPRs[i] = Reg; 3011 } 3012 3013 // We don't have a required SGPR operand, so we have a bit more freedom in 3014 // selecting operands to move. 3015 3016 // Try to select the most used SGPR. If an SGPR is equal to one of the 3017 // others, we choose that. 3018 // 3019 // e.g. 3020 // V_FMA_F32 v0, s0, s0, s0 -> No moves 3021 // V_FMA_F32 v0, s0, s1, s0 -> Move s1 3022 3023 // TODO: If some of the operands are 64-bit SGPRs and some 32, we should 3024 // prefer those. 3025 3026 if (UsedSGPRs[0] != AMDGPU::NoRegister) { 3027 if (UsedSGPRs[0] == UsedSGPRs[1] || UsedSGPRs[0] == UsedSGPRs[2]) 3028 SGPRReg = UsedSGPRs[0]; 3029 } 3030 3031 if (SGPRReg == AMDGPU::NoRegister && UsedSGPRs[1] != AMDGPU::NoRegister) { 3032 if (UsedSGPRs[1] == UsedSGPRs[2]) 3033 SGPRReg = UsedSGPRs[1]; 3034 } 3035 3036 return SGPRReg; 3037 } 3038 3039 MachineOperand *SIInstrInfo::getNamedOperand(MachineInstr &MI, 3040 unsigned OperandName) const { 3041 int Idx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), OperandName); 3042 if (Idx == -1) 3043 return nullptr; 3044 3045 return &MI.getOperand(Idx); 3046 } 3047 3048 uint64_t SIInstrInfo::getDefaultRsrcDataFormat() const { 3049 uint64_t RsrcDataFormat = AMDGPU::RSRC_DATA_FORMAT; 3050 if (ST.isAmdHsaOS()) { 3051 RsrcDataFormat |= (1ULL << 56); 3052 3053 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) 3054 // Set MTYPE = 2 3055 RsrcDataFormat |= (2ULL << 59); 3056 } 3057 3058 return RsrcDataFormat; 3059 } 3060 3061 uint64_t SIInstrInfo::getScratchRsrcWords23() const { 3062 uint64_t Rsrc23 = getDefaultRsrcDataFormat() | 3063 AMDGPU::RSRC_TID_ENABLE | 3064 0xffffffff; // Size; 3065 3066 uint64_t EltSizeValue = Log2_32(ST.getMaxPrivateElementSize()) - 1; 3067 3068 Rsrc23 |= (EltSizeValue << AMDGPU::RSRC_ELEMENT_SIZE_SHIFT) | 3069 // IndexStride = 64 3070 (UINT64_C(3) << AMDGPU::RSRC_INDEX_STRIDE_SHIFT); 3071 3072 // If TID_ENABLE is set, DATA_FORMAT specifies stride bits [14:17]. 3073 // Clear them unless we want a huge stride. 3074 if (ST.getGeneration() >= SISubtarget::VOLCANIC_ISLANDS) 3075 Rsrc23 &= ~AMDGPU::RSRC_DATA_FORMAT; 3076 3077 return Rsrc23; 3078 } 3079 3080 bool SIInstrInfo::isLowLatencyInstruction(const MachineInstr &MI) const { 3081 unsigned Opc = MI.getOpcode(); 3082 3083 return isSMRD(Opc); 3084 } 3085 3086 bool SIInstrInfo::isHighLatencyInstruction(const MachineInstr &MI) const { 3087 unsigned Opc = MI.getOpcode(); 3088 3089 return isMUBUF(Opc) || isMTBUF(Opc) || isMIMG(Opc); 3090 } 3091 3092 unsigned SIInstrInfo::getInstSizeInBytes(const MachineInstr &MI) const { 3093 unsigned Opc = MI.getOpcode(); 3094 const MCInstrDesc &Desc = getMCOpcodeFromPseudo(Opc); 3095 unsigned DescSize = Desc.getSize(); 3096 3097 // If we have a definitive size, we can use it. Otherwise we need to inspect 3098 // the operands to know the size. 3099 if (DescSize == 8 || DescSize == 4) 3100 return DescSize; 3101 3102 assert(DescSize == 0); 3103 3104 // 4-byte instructions may have a 32-bit literal encoded after them. Check 3105 // operands that coud ever be literals. 3106 if (isVALU(MI) || isSALU(MI)) { 3107 int Src0Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src0); 3108 if (Src0Idx == -1) 3109 return 4; // No operands. 3110 3111 if (isLiteralConstant(MI.getOperand(Src0Idx), getOpSize(MI, Src0Idx))) 3112 return 8; 3113 3114 int Src1Idx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::src1); 3115 if (Src1Idx == -1) 3116 return 4; 3117 3118 if (isLiteralConstant(MI.getOperand(Src1Idx), getOpSize(MI, Src1Idx))) 3119 return 8; 3120 3121 return 4; 3122 } 3123 3124 switch (Opc) { 3125 case TargetOpcode::IMPLICIT_DEF: 3126 case TargetOpcode::KILL: 3127 case TargetOpcode::DBG_VALUE: 3128 case TargetOpcode::BUNDLE: 3129 case TargetOpcode::EH_LABEL: 3130 return 0; 3131 case TargetOpcode::INLINEASM: { 3132 const MachineFunction *MF = MI.getParent()->getParent(); 3133 const char *AsmStr = MI.getOperand(0).getSymbolName(); 3134 return getInlineAsmLength(AsmStr, *MF->getTarget().getMCAsmInfo()); 3135 } 3136 default: 3137 llvm_unreachable("unable to find instruction size"); 3138 } 3139 } 3140 3141 ArrayRef<std::pair<int, const char *>> 3142 SIInstrInfo::getSerializableTargetIndices() const { 3143 static const std::pair<int, const char *> TargetIndices[] = { 3144 {AMDGPU::TI_CONSTDATA_START, "amdgpu-constdata-start"}, 3145 {AMDGPU::TI_SCRATCH_RSRC_DWORD0, "amdgpu-scratch-rsrc-dword0"}, 3146 {AMDGPU::TI_SCRATCH_RSRC_DWORD1, "amdgpu-scratch-rsrc-dword1"}, 3147 {AMDGPU::TI_SCRATCH_RSRC_DWORD2, "amdgpu-scratch-rsrc-dword2"}, 3148 {AMDGPU::TI_SCRATCH_RSRC_DWORD3, "amdgpu-scratch-rsrc-dword3"}}; 3149 return makeArrayRef(TargetIndices); 3150 } 3151 3152 /// This is used by the post-RA scheduler (SchedulePostRAList.cpp). The 3153 /// post-RA version of misched uses CreateTargetMIHazardRecognizer. 3154 ScheduleHazardRecognizer * 3155 SIInstrInfo::CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II, 3156 const ScheduleDAG *DAG) const { 3157 return new GCNHazardRecognizer(DAG->MF); 3158 } 3159 3160 /// This is the hazard recognizer used at -O0 by the PostRAHazardRecognizer 3161 /// pass. 3162 ScheduleHazardRecognizer * 3163 SIInstrInfo::CreateTargetPostRAHazardRecognizer(const MachineFunction &MF) const { 3164 return new GCNHazardRecognizer(MF); 3165 } 3166