1 //===-- SelectionDAG.cpp - Implement the SelectionDAG data structures -----===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // This implements the SelectionDAG class. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "llvm/CodeGen/SelectionDAG.h" 15 #include "SDNodeOrdering.h" 16 #include "SDNodeDbgValue.h" 17 #include "llvm/Constants.h" 18 #include "llvm/Analysis/DebugInfo.h" 19 #include "llvm/Analysis/ValueTracking.h" 20 #include "llvm/Function.h" 21 #include "llvm/GlobalAlias.h" 22 #include "llvm/GlobalVariable.h" 23 #include "llvm/Intrinsics.h" 24 #include "llvm/DerivedTypes.h" 25 #include "llvm/Assembly/Writer.h" 26 #include "llvm/CallingConv.h" 27 #include "llvm/CodeGen/MachineBasicBlock.h" 28 #include "llvm/CodeGen/MachineConstantPool.h" 29 #include "llvm/CodeGen/MachineFrameInfo.h" 30 #include "llvm/CodeGen/MachineModuleInfo.h" 31 #include "llvm/Target/TargetRegisterInfo.h" 32 #include "llvm/Target/TargetData.h" 33 #include "llvm/Target/TargetLowering.h" 34 #include "llvm/Target/TargetSelectionDAGInfo.h" 35 #include "llvm/Target/TargetOptions.h" 36 #include "llvm/Target/TargetInstrInfo.h" 37 #include "llvm/Target/TargetIntrinsicInfo.h" 38 #include "llvm/Target/TargetMachine.h" 39 #include "llvm/Support/CommandLine.h" 40 #include "llvm/Support/Debug.h" 41 #include "llvm/Support/ErrorHandling.h" 42 #include "llvm/Support/ManagedStatic.h" 43 #include "llvm/Support/MathExtras.h" 44 #include "llvm/Support/raw_ostream.h" 45 #include "llvm/Support/Mutex.h" 46 #include "llvm/ADT/SetVector.h" 47 #include "llvm/ADT/SmallPtrSet.h" 48 #include "llvm/ADT/SmallSet.h" 49 #include "llvm/ADT/SmallVector.h" 50 #include "llvm/ADT/StringExtras.h" 51 #include <algorithm> 52 #include <cmath> 53 using namespace llvm; 54 55 /// makeVTList - Return an instance of the SDVTList struct initialized with the 56 /// specified members. 57 static SDVTList makeVTList(const EVT *VTs, unsigned NumVTs) { 58 SDVTList Res = {VTs, NumVTs}; 59 return Res; 60 } 61 62 static const fltSemantics *EVTToAPFloatSemantics(EVT VT) { 63 switch (VT.getSimpleVT().SimpleTy) { 64 default: llvm_unreachable("Unknown FP format"); 65 case MVT::f32: return &APFloat::IEEEsingle; 66 case MVT::f64: return &APFloat::IEEEdouble; 67 case MVT::f80: return &APFloat::x87DoubleExtended; 68 case MVT::f128: return &APFloat::IEEEquad; 69 case MVT::ppcf128: return &APFloat::PPCDoubleDouble; 70 } 71 } 72 73 SelectionDAG::DAGUpdateListener::~DAGUpdateListener() {} 74 75 //===----------------------------------------------------------------------===// 76 // ConstantFPSDNode Class 77 //===----------------------------------------------------------------------===// 78 79 /// isExactlyValue - We don't rely on operator== working on double values, as 80 /// it returns true for things that are clearly not equal, like -0.0 and 0.0. 81 /// As such, this method can be used to do an exact bit-for-bit comparison of 82 /// two floating point values. 83 bool ConstantFPSDNode::isExactlyValue(const APFloat& V) const { 84 return getValueAPF().bitwiseIsEqual(V); 85 } 86 87 bool ConstantFPSDNode::isValueValidForType(EVT VT, 88 const APFloat& Val) { 89 assert(VT.isFloatingPoint() && "Can only convert between FP types"); 90 91 // PPC long double cannot be converted to any other type. 92 if (VT == MVT::ppcf128 || 93 &Val.getSemantics() == &APFloat::PPCDoubleDouble) 94 return false; 95 96 // convert modifies in place, so make a copy. 97 APFloat Val2 = APFloat(Val); 98 bool losesInfo; 99 (void) Val2.convert(*EVTToAPFloatSemantics(VT), APFloat::rmNearestTiesToEven, 100 &losesInfo); 101 return !losesInfo; 102 } 103 104 //===----------------------------------------------------------------------===// 105 // ISD Namespace 106 //===----------------------------------------------------------------------===// 107 108 /// isBuildVectorAllOnes - Return true if the specified node is a 109 /// BUILD_VECTOR where all of the elements are ~0 or undef. 110 bool ISD::isBuildVectorAllOnes(const SDNode *N) { 111 // Look through a bit convert. 112 if (N->getOpcode() == ISD::BITCAST) 113 N = N->getOperand(0).getNode(); 114 115 if (N->getOpcode() != ISD::BUILD_VECTOR) return false; 116 117 unsigned i = 0, e = N->getNumOperands(); 118 119 // Skip over all of the undef values. 120 while (i != e && N->getOperand(i).getOpcode() == ISD::UNDEF) 121 ++i; 122 123 // Do not accept an all-undef vector. 124 if (i == e) return false; 125 126 // Do not accept build_vectors that aren't all constants or which have non-~0 127 // elements. 128 SDValue NotZero = N->getOperand(i); 129 if (isa<ConstantSDNode>(NotZero)) { 130 if (!cast<ConstantSDNode>(NotZero)->isAllOnesValue()) 131 return false; 132 } else if (isa<ConstantFPSDNode>(NotZero)) { 133 if (!cast<ConstantFPSDNode>(NotZero)->getValueAPF(). 134 bitcastToAPInt().isAllOnesValue()) 135 return false; 136 } else 137 return false; 138 139 // Okay, we have at least one ~0 value, check to see if the rest match or are 140 // undefs. 141 for (++i; i != e; ++i) 142 if (N->getOperand(i) != NotZero && 143 N->getOperand(i).getOpcode() != ISD::UNDEF) 144 return false; 145 return true; 146 } 147 148 149 /// isBuildVectorAllZeros - Return true if the specified node is a 150 /// BUILD_VECTOR where all of the elements are 0 or undef. 151 bool ISD::isBuildVectorAllZeros(const SDNode *N) { 152 // Look through a bit convert. 153 if (N->getOpcode() == ISD::BITCAST) 154 N = N->getOperand(0).getNode(); 155 156 if (N->getOpcode() != ISD::BUILD_VECTOR) return false; 157 158 unsigned i = 0, e = N->getNumOperands(); 159 160 // Skip over all of the undef values. 161 while (i != e && N->getOperand(i).getOpcode() == ISD::UNDEF) 162 ++i; 163 164 // Do not accept an all-undef vector. 165 if (i == e) return false; 166 167 // Do not accept build_vectors that aren't all constants or which have non-0 168 // elements. 169 SDValue Zero = N->getOperand(i); 170 if (isa<ConstantSDNode>(Zero)) { 171 if (!cast<ConstantSDNode>(Zero)->isNullValue()) 172 return false; 173 } else if (isa<ConstantFPSDNode>(Zero)) { 174 if (!cast<ConstantFPSDNode>(Zero)->getValueAPF().isPosZero()) 175 return false; 176 } else 177 return false; 178 179 // Okay, we have at least one 0 value, check to see if the rest match or are 180 // undefs. 181 for (++i; i != e; ++i) 182 if (N->getOperand(i) != Zero && 183 N->getOperand(i).getOpcode() != ISD::UNDEF) 184 return false; 185 return true; 186 } 187 188 /// isScalarToVector - Return true if the specified node is a 189 /// ISD::SCALAR_TO_VECTOR node or a BUILD_VECTOR node where only the low 190 /// element is not an undef. 191 bool ISD::isScalarToVector(const SDNode *N) { 192 if (N->getOpcode() == ISD::SCALAR_TO_VECTOR) 193 return true; 194 195 if (N->getOpcode() != ISD::BUILD_VECTOR) 196 return false; 197 if (N->getOperand(0).getOpcode() == ISD::UNDEF) 198 return false; 199 unsigned NumElems = N->getNumOperands(); 200 if (NumElems == 1) 201 return false; 202 for (unsigned i = 1; i < NumElems; ++i) { 203 SDValue V = N->getOperand(i); 204 if (V.getOpcode() != ISD::UNDEF) 205 return false; 206 } 207 return true; 208 } 209 210 /// getSetCCSwappedOperands - Return the operation corresponding to (Y op X) 211 /// when given the operation for (X op Y). 212 ISD::CondCode ISD::getSetCCSwappedOperands(ISD::CondCode Operation) { 213 // To perform this operation, we just need to swap the L and G bits of the 214 // operation. 215 unsigned OldL = (Operation >> 2) & 1; 216 unsigned OldG = (Operation >> 1) & 1; 217 return ISD::CondCode((Operation & ~6) | // Keep the N, U, E bits 218 (OldL << 1) | // New G bit 219 (OldG << 2)); // New L bit. 220 } 221 222 /// getSetCCInverse - Return the operation corresponding to !(X op Y), where 223 /// 'op' is a valid SetCC operation. 224 ISD::CondCode ISD::getSetCCInverse(ISD::CondCode Op, bool isInteger) { 225 unsigned Operation = Op; 226 if (isInteger) 227 Operation ^= 7; // Flip L, G, E bits, but not U. 228 else 229 Operation ^= 15; // Flip all of the condition bits. 230 231 if (Operation > ISD::SETTRUE2) 232 Operation &= ~8; // Don't let N and U bits get set. 233 234 return ISD::CondCode(Operation); 235 } 236 237 238 /// isSignedOp - For an integer comparison, return 1 if the comparison is a 239 /// signed operation and 2 if the result is an unsigned comparison. Return zero 240 /// if the operation does not depend on the sign of the input (setne and seteq). 241 static int isSignedOp(ISD::CondCode Opcode) { 242 switch (Opcode) { 243 default: llvm_unreachable("Illegal integer setcc operation!"); 244 case ISD::SETEQ: 245 case ISD::SETNE: return 0; 246 case ISD::SETLT: 247 case ISD::SETLE: 248 case ISD::SETGT: 249 case ISD::SETGE: return 1; 250 case ISD::SETULT: 251 case ISD::SETULE: 252 case ISD::SETUGT: 253 case ISD::SETUGE: return 2; 254 } 255 } 256 257 /// getSetCCOrOperation - Return the result of a logical OR between different 258 /// comparisons of identical values: ((X op1 Y) | (X op2 Y)). This function 259 /// returns SETCC_INVALID if it is not possible to represent the resultant 260 /// comparison. 261 ISD::CondCode ISD::getSetCCOrOperation(ISD::CondCode Op1, ISD::CondCode Op2, 262 bool isInteger) { 263 if (isInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3) 264 // Cannot fold a signed integer setcc with an unsigned integer setcc. 265 return ISD::SETCC_INVALID; 266 267 unsigned Op = Op1 | Op2; // Combine all of the condition bits. 268 269 // If the N and U bits get set then the resultant comparison DOES suddenly 270 // care about orderedness, and is true when ordered. 271 if (Op > ISD::SETTRUE2) 272 Op &= ~16; // Clear the U bit if the N bit is set. 273 274 // Canonicalize illegal integer setcc's. 275 if (isInteger && Op == ISD::SETUNE) // e.g. SETUGT | SETULT 276 Op = ISD::SETNE; 277 278 return ISD::CondCode(Op); 279 } 280 281 /// getSetCCAndOperation - Return the result of a logical AND between different 282 /// comparisons of identical values: ((X op1 Y) & (X op2 Y)). This 283 /// function returns zero if it is not possible to represent the resultant 284 /// comparison. 285 ISD::CondCode ISD::getSetCCAndOperation(ISD::CondCode Op1, ISD::CondCode Op2, 286 bool isInteger) { 287 if (isInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3) 288 // Cannot fold a signed setcc with an unsigned setcc. 289 return ISD::SETCC_INVALID; 290 291 // Combine all of the condition bits. 292 ISD::CondCode Result = ISD::CondCode(Op1 & Op2); 293 294 // Canonicalize illegal integer setcc's. 295 if (isInteger) { 296 switch (Result) { 297 default: break; 298 case ISD::SETUO : Result = ISD::SETFALSE; break; // SETUGT & SETULT 299 case ISD::SETOEQ: // SETEQ & SETU[LG]E 300 case ISD::SETUEQ: Result = ISD::SETEQ ; break; // SETUGE & SETULE 301 case ISD::SETOLT: Result = ISD::SETULT ; break; // SETULT & SETNE 302 case ISD::SETOGT: Result = ISD::SETUGT ; break; // SETUGT & SETNE 303 } 304 } 305 306 return Result; 307 } 308 309 //===----------------------------------------------------------------------===// 310 // SDNode Profile Support 311 //===----------------------------------------------------------------------===// 312 313 /// AddNodeIDOpcode - Add the node opcode to the NodeID data. 314 /// 315 static void AddNodeIDOpcode(FoldingSetNodeID &ID, unsigned OpC) { 316 ID.AddInteger(OpC); 317 } 318 319 /// AddNodeIDValueTypes - Value type lists are intern'd so we can represent them 320 /// solely with their pointer. 321 static void AddNodeIDValueTypes(FoldingSetNodeID &ID, SDVTList VTList) { 322 ID.AddPointer(VTList.VTs); 323 } 324 325 /// AddNodeIDOperands - Various routines for adding operands to the NodeID data. 326 /// 327 static void AddNodeIDOperands(FoldingSetNodeID &ID, 328 const SDValue *Ops, unsigned NumOps) { 329 for (; NumOps; --NumOps, ++Ops) { 330 ID.AddPointer(Ops->getNode()); 331 ID.AddInteger(Ops->getResNo()); 332 } 333 } 334 335 /// AddNodeIDOperands - Various routines for adding operands to the NodeID data. 336 /// 337 static void AddNodeIDOperands(FoldingSetNodeID &ID, 338 const SDUse *Ops, unsigned NumOps) { 339 for (; NumOps; --NumOps, ++Ops) { 340 ID.AddPointer(Ops->getNode()); 341 ID.AddInteger(Ops->getResNo()); 342 } 343 } 344 345 static void AddNodeIDNode(FoldingSetNodeID &ID, 346 unsigned short OpC, SDVTList VTList, 347 const SDValue *OpList, unsigned N) { 348 AddNodeIDOpcode(ID, OpC); 349 AddNodeIDValueTypes(ID, VTList); 350 AddNodeIDOperands(ID, OpList, N); 351 } 352 353 /// AddNodeIDCustom - If this is an SDNode with special info, add this info to 354 /// the NodeID data. 355 static void AddNodeIDCustom(FoldingSetNodeID &ID, const SDNode *N) { 356 switch (N->getOpcode()) { 357 case ISD::TargetExternalSymbol: 358 case ISD::ExternalSymbol: 359 llvm_unreachable("Should only be used on nodes with operands"); 360 default: break; // Normal nodes don't need extra info. 361 case ISD::TargetConstant: 362 case ISD::Constant: 363 ID.AddPointer(cast<ConstantSDNode>(N)->getConstantIntValue()); 364 break; 365 case ISD::TargetConstantFP: 366 case ISD::ConstantFP: { 367 ID.AddPointer(cast<ConstantFPSDNode>(N)->getConstantFPValue()); 368 break; 369 } 370 case ISD::TargetGlobalAddress: 371 case ISD::GlobalAddress: 372 case ISD::TargetGlobalTLSAddress: 373 case ISD::GlobalTLSAddress: { 374 const GlobalAddressSDNode *GA = cast<GlobalAddressSDNode>(N); 375 ID.AddPointer(GA->getGlobal()); 376 ID.AddInteger(GA->getOffset()); 377 ID.AddInteger(GA->getTargetFlags()); 378 break; 379 } 380 case ISD::BasicBlock: 381 ID.AddPointer(cast<BasicBlockSDNode>(N)->getBasicBlock()); 382 break; 383 case ISD::Register: 384 ID.AddInteger(cast<RegisterSDNode>(N)->getReg()); 385 break; 386 387 case ISD::SRCVALUE: 388 ID.AddPointer(cast<SrcValueSDNode>(N)->getValue()); 389 break; 390 case ISD::FrameIndex: 391 case ISD::TargetFrameIndex: 392 ID.AddInteger(cast<FrameIndexSDNode>(N)->getIndex()); 393 break; 394 case ISD::JumpTable: 395 case ISD::TargetJumpTable: 396 ID.AddInteger(cast<JumpTableSDNode>(N)->getIndex()); 397 ID.AddInteger(cast<JumpTableSDNode>(N)->getTargetFlags()); 398 break; 399 case ISD::ConstantPool: 400 case ISD::TargetConstantPool: { 401 const ConstantPoolSDNode *CP = cast<ConstantPoolSDNode>(N); 402 ID.AddInteger(CP->getAlignment()); 403 ID.AddInteger(CP->getOffset()); 404 if (CP->isMachineConstantPoolEntry()) 405 CP->getMachineCPVal()->addSelectionDAGCSEId(ID); 406 else 407 ID.AddPointer(CP->getConstVal()); 408 ID.AddInteger(CP->getTargetFlags()); 409 break; 410 } 411 case ISD::LOAD: { 412 const LoadSDNode *LD = cast<LoadSDNode>(N); 413 ID.AddInteger(LD->getMemoryVT().getRawBits()); 414 ID.AddInteger(LD->getRawSubclassData()); 415 break; 416 } 417 case ISD::STORE: { 418 const StoreSDNode *ST = cast<StoreSDNode>(N); 419 ID.AddInteger(ST->getMemoryVT().getRawBits()); 420 ID.AddInteger(ST->getRawSubclassData()); 421 break; 422 } 423 case ISD::ATOMIC_CMP_SWAP: 424 case ISD::ATOMIC_SWAP: 425 case ISD::ATOMIC_LOAD_ADD: 426 case ISD::ATOMIC_LOAD_SUB: 427 case ISD::ATOMIC_LOAD_AND: 428 case ISD::ATOMIC_LOAD_OR: 429 case ISD::ATOMIC_LOAD_XOR: 430 case ISD::ATOMIC_LOAD_NAND: 431 case ISD::ATOMIC_LOAD_MIN: 432 case ISD::ATOMIC_LOAD_MAX: 433 case ISD::ATOMIC_LOAD_UMIN: 434 case ISD::ATOMIC_LOAD_UMAX: 435 case ISD::ATOMIC_LOAD: 436 case ISD::ATOMIC_STORE: { 437 const AtomicSDNode *AT = cast<AtomicSDNode>(N); 438 ID.AddInteger(AT->getMemoryVT().getRawBits()); 439 ID.AddInteger(AT->getRawSubclassData()); 440 break; 441 } 442 case ISD::VECTOR_SHUFFLE: { 443 const ShuffleVectorSDNode *SVN = cast<ShuffleVectorSDNode>(N); 444 for (unsigned i = 0, e = N->getValueType(0).getVectorNumElements(); 445 i != e; ++i) 446 ID.AddInteger(SVN->getMaskElt(i)); 447 break; 448 } 449 case ISD::TargetBlockAddress: 450 case ISD::BlockAddress: { 451 ID.AddPointer(cast<BlockAddressSDNode>(N)->getBlockAddress()); 452 ID.AddInteger(cast<BlockAddressSDNode>(N)->getTargetFlags()); 453 break; 454 } 455 } // end switch (N->getOpcode()) 456 } 457 458 /// AddNodeIDNode - Generic routine for adding a nodes info to the NodeID 459 /// data. 460 static void AddNodeIDNode(FoldingSetNodeID &ID, const SDNode *N) { 461 AddNodeIDOpcode(ID, N->getOpcode()); 462 // Add the return value info. 463 AddNodeIDValueTypes(ID, N->getVTList()); 464 // Add the operand info. 465 AddNodeIDOperands(ID, N->op_begin(), N->getNumOperands()); 466 467 // Handle SDNode leafs with special info. 468 AddNodeIDCustom(ID, N); 469 } 470 471 /// encodeMemSDNodeFlags - Generic routine for computing a value for use in 472 /// the CSE map that carries volatility, temporalness, indexing mode, and 473 /// extension/truncation information. 474 /// 475 static inline unsigned 476 encodeMemSDNodeFlags(int ConvType, ISD::MemIndexedMode AM, bool isVolatile, 477 bool isNonTemporal, bool isInvariant) { 478 assert((ConvType & 3) == ConvType && 479 "ConvType may not require more than 2 bits!"); 480 assert((AM & 7) == AM && 481 "AM may not require more than 3 bits!"); 482 return ConvType | 483 (AM << 2) | 484 (isVolatile << 5) | 485 (isNonTemporal << 6) | 486 (isInvariant << 7); 487 } 488 489 //===----------------------------------------------------------------------===// 490 // SelectionDAG Class 491 //===----------------------------------------------------------------------===// 492 493 /// doNotCSE - Return true if CSE should not be performed for this node. 494 static bool doNotCSE(SDNode *N) { 495 if (N->getValueType(0) == MVT::Glue) 496 return true; // Never CSE anything that produces a flag. 497 498 switch (N->getOpcode()) { 499 default: break; 500 case ISD::HANDLENODE: 501 case ISD::EH_LABEL: 502 return true; // Never CSE these nodes. 503 } 504 505 // Check that remaining values produced are not flags. 506 for (unsigned i = 1, e = N->getNumValues(); i != e; ++i) 507 if (N->getValueType(i) == MVT::Glue) 508 return true; // Never CSE anything that produces a flag. 509 510 return false; 511 } 512 513 /// RemoveDeadNodes - This method deletes all unreachable nodes in the 514 /// SelectionDAG. 515 void SelectionDAG::RemoveDeadNodes() { 516 // Create a dummy node (which is not added to allnodes), that adds a reference 517 // to the root node, preventing it from being deleted. 518 HandleSDNode Dummy(getRoot()); 519 520 SmallVector<SDNode*, 128> DeadNodes; 521 522 // Add all obviously-dead nodes to the DeadNodes worklist. 523 for (allnodes_iterator I = allnodes_begin(), E = allnodes_end(); I != E; ++I) 524 if (I->use_empty()) 525 DeadNodes.push_back(I); 526 527 RemoveDeadNodes(DeadNodes); 528 529 // If the root changed (e.g. it was a dead load, update the root). 530 setRoot(Dummy.getValue()); 531 } 532 533 /// RemoveDeadNodes - This method deletes the unreachable nodes in the 534 /// given list, and any nodes that become unreachable as a result. 535 void SelectionDAG::RemoveDeadNodes(SmallVectorImpl<SDNode *> &DeadNodes, 536 DAGUpdateListener *UpdateListener) { 537 538 // Process the worklist, deleting the nodes and adding their uses to the 539 // worklist. 540 while (!DeadNodes.empty()) { 541 SDNode *N = DeadNodes.pop_back_val(); 542 543 if (UpdateListener) 544 UpdateListener->NodeDeleted(N, 0); 545 546 // Take the node out of the appropriate CSE map. 547 RemoveNodeFromCSEMaps(N); 548 549 // Next, brutally remove the operand list. This is safe to do, as there are 550 // no cycles in the graph. 551 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) { 552 SDUse &Use = *I++; 553 SDNode *Operand = Use.getNode(); 554 Use.set(SDValue()); 555 556 // Now that we removed this operand, see if there are no uses of it left. 557 if (Operand->use_empty()) 558 DeadNodes.push_back(Operand); 559 } 560 561 DeallocateNode(N); 562 } 563 } 564 565 void SelectionDAG::RemoveDeadNode(SDNode *N, DAGUpdateListener *UpdateListener){ 566 SmallVector<SDNode*, 16> DeadNodes(1, N); 567 568 // Create a dummy node that adds a reference to the root node, preventing 569 // it from being deleted. (This matters if the root is an operand of the 570 // dead node.) 571 HandleSDNode Dummy(getRoot()); 572 573 RemoveDeadNodes(DeadNodes, UpdateListener); 574 } 575 576 void SelectionDAG::DeleteNode(SDNode *N) { 577 // First take this out of the appropriate CSE map. 578 RemoveNodeFromCSEMaps(N); 579 580 // Finally, remove uses due to operands of this node, remove from the 581 // AllNodes list, and delete the node. 582 DeleteNodeNotInCSEMaps(N); 583 } 584 585 void SelectionDAG::DeleteNodeNotInCSEMaps(SDNode *N) { 586 assert(N != AllNodes.begin() && "Cannot delete the entry node!"); 587 assert(N->use_empty() && "Cannot delete a node that is not dead!"); 588 589 // Drop all of the operands and decrement used node's use counts. 590 N->DropOperands(); 591 592 DeallocateNode(N); 593 } 594 595 void SelectionDAG::DeallocateNode(SDNode *N) { 596 if (N->OperandsNeedDelete) 597 delete[] N->OperandList; 598 599 // Set the opcode to DELETED_NODE to help catch bugs when node 600 // memory is reallocated. 601 N->NodeType = ISD::DELETED_NODE; 602 603 NodeAllocator.Deallocate(AllNodes.remove(N)); 604 605 // Remove the ordering of this node. 606 Ordering->remove(N); 607 608 // If any of the SDDbgValue nodes refer to this SDNode, invalidate them. 609 ArrayRef<SDDbgValue*> DbgVals = DbgInfo->getSDDbgValues(N); 610 for (unsigned i = 0, e = DbgVals.size(); i != e; ++i) 611 DbgVals[i]->setIsInvalidated(); 612 } 613 614 /// RemoveNodeFromCSEMaps - Take the specified node out of the CSE map that 615 /// correspond to it. This is useful when we're about to delete or repurpose 616 /// the node. We don't want future request for structurally identical nodes 617 /// to return N anymore. 618 bool SelectionDAG::RemoveNodeFromCSEMaps(SDNode *N) { 619 bool Erased = false; 620 switch (N->getOpcode()) { 621 case ISD::HANDLENODE: return false; // noop. 622 case ISD::CONDCODE: 623 assert(CondCodeNodes[cast<CondCodeSDNode>(N)->get()] && 624 "Cond code doesn't exist!"); 625 Erased = CondCodeNodes[cast<CondCodeSDNode>(N)->get()] != 0; 626 CondCodeNodes[cast<CondCodeSDNode>(N)->get()] = 0; 627 break; 628 case ISD::ExternalSymbol: 629 Erased = ExternalSymbols.erase(cast<ExternalSymbolSDNode>(N)->getSymbol()); 630 break; 631 case ISD::TargetExternalSymbol: { 632 ExternalSymbolSDNode *ESN = cast<ExternalSymbolSDNode>(N); 633 Erased = TargetExternalSymbols.erase( 634 std::pair<std::string,unsigned char>(ESN->getSymbol(), 635 ESN->getTargetFlags())); 636 break; 637 } 638 case ISD::VALUETYPE: { 639 EVT VT = cast<VTSDNode>(N)->getVT(); 640 if (VT.isExtended()) { 641 Erased = ExtendedValueTypeNodes.erase(VT); 642 } else { 643 Erased = ValueTypeNodes[VT.getSimpleVT().SimpleTy] != 0; 644 ValueTypeNodes[VT.getSimpleVT().SimpleTy] = 0; 645 } 646 break; 647 } 648 default: 649 // Remove it from the CSE Map. 650 assert(N->getOpcode() != ISD::DELETED_NODE && "DELETED_NODE in CSEMap!"); 651 assert(N->getOpcode() != ISD::EntryToken && "EntryToken in CSEMap!"); 652 Erased = CSEMap.RemoveNode(N); 653 break; 654 } 655 #ifndef NDEBUG 656 // Verify that the node was actually in one of the CSE maps, unless it has a 657 // flag result (which cannot be CSE'd) or is one of the special cases that are 658 // not subject to CSE. 659 if (!Erased && N->getValueType(N->getNumValues()-1) != MVT::Glue && 660 !N->isMachineOpcode() && !doNotCSE(N)) { 661 N->dump(this); 662 dbgs() << "\n"; 663 llvm_unreachable("Node is not in map!"); 664 } 665 #endif 666 return Erased; 667 } 668 669 /// AddModifiedNodeToCSEMaps - The specified node has been removed from the CSE 670 /// maps and modified in place. Add it back to the CSE maps, unless an identical 671 /// node already exists, in which case transfer all its users to the existing 672 /// node. This transfer can potentially trigger recursive merging. 673 /// 674 void 675 SelectionDAG::AddModifiedNodeToCSEMaps(SDNode *N, 676 DAGUpdateListener *UpdateListener) { 677 // For node types that aren't CSE'd, just act as if no identical node 678 // already exists. 679 if (!doNotCSE(N)) { 680 SDNode *Existing = CSEMap.GetOrInsertNode(N); 681 if (Existing != N) { 682 // If there was already an existing matching node, use ReplaceAllUsesWith 683 // to replace the dead one with the existing one. This can cause 684 // recursive merging of other unrelated nodes down the line. 685 ReplaceAllUsesWith(N, Existing, UpdateListener); 686 687 // N is now dead. Inform the listener if it exists and delete it. 688 if (UpdateListener) 689 UpdateListener->NodeDeleted(N, Existing); 690 DeleteNodeNotInCSEMaps(N); 691 return; 692 } 693 } 694 695 // If the node doesn't already exist, we updated it. Inform a listener if 696 // it exists. 697 if (UpdateListener) 698 UpdateListener->NodeUpdated(N); 699 } 700 701 /// FindModifiedNodeSlot - Find a slot for the specified node if its operands 702 /// were replaced with those specified. If this node is never memoized, 703 /// return null, otherwise return a pointer to the slot it would take. If a 704 /// node already exists with these operands, the slot will be non-null. 705 SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, SDValue Op, 706 void *&InsertPos) { 707 if (doNotCSE(N)) 708 return 0; 709 710 SDValue Ops[] = { Op }; 711 FoldingSetNodeID ID; 712 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops, 1); 713 AddNodeIDCustom(ID, N); 714 SDNode *Node = CSEMap.FindNodeOrInsertPos(ID, InsertPos); 715 return Node; 716 } 717 718 /// FindModifiedNodeSlot - Find a slot for the specified node if its operands 719 /// were replaced with those specified. If this node is never memoized, 720 /// return null, otherwise return a pointer to the slot it would take. If a 721 /// node already exists with these operands, the slot will be non-null. 722 SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, 723 SDValue Op1, SDValue Op2, 724 void *&InsertPos) { 725 if (doNotCSE(N)) 726 return 0; 727 728 SDValue Ops[] = { Op1, Op2 }; 729 FoldingSetNodeID ID; 730 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops, 2); 731 AddNodeIDCustom(ID, N); 732 SDNode *Node = CSEMap.FindNodeOrInsertPos(ID, InsertPos); 733 return Node; 734 } 735 736 737 /// FindModifiedNodeSlot - Find a slot for the specified node if its operands 738 /// were replaced with those specified. If this node is never memoized, 739 /// return null, otherwise return a pointer to the slot it would take. If a 740 /// node already exists with these operands, the slot will be non-null. 741 SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, 742 const SDValue *Ops,unsigned NumOps, 743 void *&InsertPos) { 744 if (doNotCSE(N)) 745 return 0; 746 747 FoldingSetNodeID ID; 748 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops, NumOps); 749 AddNodeIDCustom(ID, N); 750 SDNode *Node = CSEMap.FindNodeOrInsertPos(ID, InsertPos); 751 return Node; 752 } 753 754 #ifndef NDEBUG 755 /// VerifyNodeCommon - Sanity check the given node. Aborts if it is invalid. 756 static void VerifyNodeCommon(SDNode *N) { 757 switch (N->getOpcode()) { 758 default: 759 break; 760 case ISD::BUILD_PAIR: { 761 EVT VT = N->getValueType(0); 762 assert(N->getNumValues() == 1 && "Too many results!"); 763 assert(!VT.isVector() && (VT.isInteger() || VT.isFloatingPoint()) && 764 "Wrong return type!"); 765 assert(N->getNumOperands() == 2 && "Wrong number of operands!"); 766 assert(N->getOperand(0).getValueType() == N->getOperand(1).getValueType() && 767 "Mismatched operand types!"); 768 assert(N->getOperand(0).getValueType().isInteger() == VT.isInteger() && 769 "Wrong operand type!"); 770 assert(VT.getSizeInBits() == 2 * N->getOperand(0).getValueSizeInBits() && 771 "Wrong return type size"); 772 break; 773 } 774 case ISD::BUILD_VECTOR: { 775 assert(N->getNumValues() == 1 && "Too many results!"); 776 assert(N->getValueType(0).isVector() && "Wrong return type!"); 777 assert(N->getNumOperands() == N->getValueType(0).getVectorNumElements() && 778 "Wrong number of operands!"); 779 EVT EltVT = N->getValueType(0).getVectorElementType(); 780 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ++I) { 781 assert((I->getValueType() == EltVT || 782 (EltVT.isInteger() && I->getValueType().isInteger() && 783 EltVT.bitsLE(I->getValueType()))) && 784 "Wrong operand type!"); 785 assert(I->getValueType() == N->getOperand(0).getValueType() && 786 "Operands must all have the same type"); 787 } 788 break; 789 } 790 } 791 } 792 793 /// VerifySDNode - Sanity check the given SDNode. Aborts if it is invalid. 794 static void VerifySDNode(SDNode *N) { 795 // The SDNode allocators cannot be used to allocate nodes with fields that are 796 // not present in an SDNode! 797 assert(!isa<MemSDNode>(N) && "Bad MemSDNode!"); 798 assert(!isa<ShuffleVectorSDNode>(N) && "Bad ShuffleVectorSDNode!"); 799 assert(!isa<ConstantSDNode>(N) && "Bad ConstantSDNode!"); 800 assert(!isa<ConstantFPSDNode>(N) && "Bad ConstantFPSDNode!"); 801 assert(!isa<GlobalAddressSDNode>(N) && "Bad GlobalAddressSDNode!"); 802 assert(!isa<FrameIndexSDNode>(N) && "Bad FrameIndexSDNode!"); 803 assert(!isa<JumpTableSDNode>(N) && "Bad JumpTableSDNode!"); 804 assert(!isa<ConstantPoolSDNode>(N) && "Bad ConstantPoolSDNode!"); 805 assert(!isa<BasicBlockSDNode>(N) && "Bad BasicBlockSDNode!"); 806 assert(!isa<SrcValueSDNode>(N) && "Bad SrcValueSDNode!"); 807 assert(!isa<MDNodeSDNode>(N) && "Bad MDNodeSDNode!"); 808 assert(!isa<RegisterSDNode>(N) && "Bad RegisterSDNode!"); 809 assert(!isa<BlockAddressSDNode>(N) && "Bad BlockAddressSDNode!"); 810 assert(!isa<EHLabelSDNode>(N) && "Bad EHLabelSDNode!"); 811 assert(!isa<ExternalSymbolSDNode>(N) && "Bad ExternalSymbolSDNode!"); 812 assert(!isa<CondCodeSDNode>(N) && "Bad CondCodeSDNode!"); 813 assert(!isa<CvtRndSatSDNode>(N) && "Bad CvtRndSatSDNode!"); 814 assert(!isa<VTSDNode>(N) && "Bad VTSDNode!"); 815 assert(!isa<MachineSDNode>(N) && "Bad MachineSDNode!"); 816 817 VerifyNodeCommon(N); 818 } 819 820 /// VerifyMachineNode - Sanity check the given MachineNode. Aborts if it is 821 /// invalid. 822 static void VerifyMachineNode(SDNode *N) { 823 // The MachineNode allocators cannot be used to allocate nodes with fields 824 // that are not present in a MachineNode! 825 // Currently there are no such nodes. 826 827 VerifyNodeCommon(N); 828 } 829 #endif // NDEBUG 830 831 /// getEVTAlignment - Compute the default alignment value for the 832 /// given type. 833 /// 834 unsigned SelectionDAG::getEVTAlignment(EVT VT) const { 835 Type *Ty = VT == MVT::iPTR ? 836 PointerType::get(Type::getInt8Ty(*getContext()), 0) : 837 VT.getTypeForEVT(*getContext()); 838 839 return TLI.getTargetData()->getABITypeAlignment(Ty); 840 } 841 842 // EntryNode could meaningfully have debug info if we can find it... 843 SelectionDAG::SelectionDAG(const TargetMachine &tm, CodeGenOpt::Level OL) 844 : TM(tm), TLI(*tm.getTargetLowering()), TSI(*tm.getSelectionDAGInfo()), 845 OptLevel(OL), EntryNode(ISD::EntryToken, DebugLoc(), getVTList(MVT::Other)), 846 Root(getEntryNode()), Ordering(0) { 847 AllNodes.push_back(&EntryNode); 848 Ordering = new SDNodeOrdering(); 849 DbgInfo = new SDDbgInfo(); 850 } 851 852 void SelectionDAG::init(MachineFunction &mf) { 853 MF = &mf; 854 Context = &mf.getFunction()->getContext(); 855 } 856 857 SelectionDAG::~SelectionDAG() { 858 allnodes_clear(); 859 delete Ordering; 860 delete DbgInfo; 861 } 862 863 void SelectionDAG::allnodes_clear() { 864 assert(&*AllNodes.begin() == &EntryNode); 865 AllNodes.remove(AllNodes.begin()); 866 while (!AllNodes.empty()) 867 DeallocateNode(AllNodes.begin()); 868 } 869 870 void SelectionDAG::clear() { 871 allnodes_clear(); 872 OperandAllocator.Reset(); 873 CSEMap.clear(); 874 875 ExtendedValueTypeNodes.clear(); 876 ExternalSymbols.clear(); 877 TargetExternalSymbols.clear(); 878 std::fill(CondCodeNodes.begin(), CondCodeNodes.end(), 879 static_cast<CondCodeSDNode*>(0)); 880 std::fill(ValueTypeNodes.begin(), ValueTypeNodes.end(), 881 static_cast<SDNode*>(0)); 882 883 EntryNode.UseList = 0; 884 AllNodes.push_back(&EntryNode); 885 Root = getEntryNode(); 886 Ordering->clear(); 887 DbgInfo->clear(); 888 } 889 890 SDValue SelectionDAG::getAnyExtOrTrunc(SDValue Op, DebugLoc DL, EVT VT) { 891 return VT.bitsGT(Op.getValueType()) ? 892 getNode(ISD::ANY_EXTEND, DL, VT, Op) : 893 getNode(ISD::TRUNCATE, DL, VT, Op); 894 } 895 896 SDValue SelectionDAG::getSExtOrTrunc(SDValue Op, DebugLoc DL, EVT VT) { 897 return VT.bitsGT(Op.getValueType()) ? 898 getNode(ISD::SIGN_EXTEND, DL, VT, Op) : 899 getNode(ISD::TRUNCATE, DL, VT, Op); 900 } 901 902 SDValue SelectionDAG::getZExtOrTrunc(SDValue Op, DebugLoc DL, EVT VT) { 903 return VT.bitsGT(Op.getValueType()) ? 904 getNode(ISD::ZERO_EXTEND, DL, VT, Op) : 905 getNode(ISD::TRUNCATE, DL, VT, Op); 906 } 907 908 SDValue SelectionDAG::getZeroExtendInReg(SDValue Op, DebugLoc DL, EVT VT) { 909 assert(!VT.isVector() && 910 "getZeroExtendInReg should use the vector element type instead of " 911 "the vector type!"); 912 if (Op.getValueType() == VT) return Op; 913 unsigned BitWidth = Op.getValueType().getScalarType().getSizeInBits(); 914 APInt Imm = APInt::getLowBitsSet(BitWidth, 915 VT.getSizeInBits()); 916 return getNode(ISD::AND, DL, Op.getValueType(), Op, 917 getConstant(Imm, Op.getValueType())); 918 } 919 920 /// getNOT - Create a bitwise NOT operation as (XOR Val, -1). 921 /// 922 SDValue SelectionDAG::getNOT(DebugLoc DL, SDValue Val, EVT VT) { 923 EVT EltVT = VT.getScalarType(); 924 SDValue NegOne = 925 getConstant(APInt::getAllOnesValue(EltVT.getSizeInBits()), VT); 926 return getNode(ISD::XOR, DL, VT, Val, NegOne); 927 } 928 929 SDValue SelectionDAG::getConstant(uint64_t Val, EVT VT, bool isT) { 930 EVT EltVT = VT.getScalarType(); 931 assert((EltVT.getSizeInBits() >= 64 || 932 (uint64_t)((int64_t)Val >> EltVT.getSizeInBits()) + 1 < 2) && 933 "getConstant with a uint64_t value that doesn't fit in the type!"); 934 return getConstant(APInt(EltVT.getSizeInBits(), Val), VT, isT); 935 } 936 937 SDValue SelectionDAG::getConstant(const APInt &Val, EVT VT, bool isT) { 938 return getConstant(*ConstantInt::get(*Context, Val), VT, isT); 939 } 940 941 SDValue SelectionDAG::getConstant(const ConstantInt &Val, EVT VT, bool isT) { 942 assert(VT.isInteger() && "Cannot create FP integer constant!"); 943 944 EVT EltVT = VT.getScalarType(); 945 const ConstantInt *Elt = &Val; 946 947 // In some cases the vector type is legal but the element type is illegal and 948 // needs to be promoted, for example v8i8 on ARM. In this case, promote the 949 // inserted value (the type does not need to match the vector element type). 950 // Any extra bits introduced will be truncated away. 951 if (VT.isVector() && TLI.getTypeAction(*getContext(), EltVT) == 952 TargetLowering::TypePromoteInteger) { 953 EltVT = TLI.getTypeToTransformTo(*getContext(), EltVT); 954 APInt NewVal = Elt->getValue().zext(EltVT.getSizeInBits()); 955 Elt = ConstantInt::get(*getContext(), NewVal); 956 } 957 958 assert(Elt->getBitWidth() == EltVT.getSizeInBits() && 959 "APInt size does not match type size!"); 960 unsigned Opc = isT ? ISD::TargetConstant : ISD::Constant; 961 FoldingSetNodeID ID; 962 AddNodeIDNode(ID, Opc, getVTList(EltVT), 0, 0); 963 ID.AddPointer(Elt); 964 void *IP = 0; 965 SDNode *N = NULL; 966 if ((N = CSEMap.FindNodeOrInsertPos(ID, IP))) 967 if (!VT.isVector()) 968 return SDValue(N, 0); 969 970 if (!N) { 971 N = new (NodeAllocator) ConstantSDNode(isT, Elt, EltVT); 972 CSEMap.InsertNode(N, IP); 973 AllNodes.push_back(N); 974 } 975 976 SDValue Result(N, 0); 977 if (VT.isVector()) { 978 SmallVector<SDValue, 8> Ops; 979 Ops.assign(VT.getVectorNumElements(), Result); 980 Result = getNode(ISD::BUILD_VECTOR, DebugLoc(), VT, &Ops[0], Ops.size()); 981 } 982 return Result; 983 } 984 985 SDValue SelectionDAG::getIntPtrConstant(uint64_t Val, bool isTarget) { 986 return getConstant(Val, TLI.getPointerTy(), isTarget); 987 } 988 989 990 SDValue SelectionDAG::getConstantFP(const APFloat& V, EVT VT, bool isTarget) { 991 return getConstantFP(*ConstantFP::get(*getContext(), V), VT, isTarget); 992 } 993 994 SDValue SelectionDAG::getConstantFP(const ConstantFP& V, EVT VT, bool isTarget){ 995 assert(VT.isFloatingPoint() && "Cannot create integer FP constant!"); 996 997 EVT EltVT = VT.getScalarType(); 998 999 // Do the map lookup using the actual bit pattern for the floating point 1000 // value, so that we don't have problems with 0.0 comparing equal to -0.0, and 1001 // we don't have issues with SNANs. 1002 unsigned Opc = isTarget ? ISD::TargetConstantFP : ISD::ConstantFP; 1003 FoldingSetNodeID ID; 1004 AddNodeIDNode(ID, Opc, getVTList(EltVT), 0, 0); 1005 ID.AddPointer(&V); 1006 void *IP = 0; 1007 SDNode *N = NULL; 1008 if ((N = CSEMap.FindNodeOrInsertPos(ID, IP))) 1009 if (!VT.isVector()) 1010 return SDValue(N, 0); 1011 1012 if (!N) { 1013 N = new (NodeAllocator) ConstantFPSDNode(isTarget, &V, EltVT); 1014 CSEMap.InsertNode(N, IP); 1015 AllNodes.push_back(N); 1016 } 1017 1018 SDValue Result(N, 0); 1019 if (VT.isVector()) { 1020 SmallVector<SDValue, 8> Ops; 1021 Ops.assign(VT.getVectorNumElements(), Result); 1022 // FIXME DebugLoc info might be appropriate here 1023 Result = getNode(ISD::BUILD_VECTOR, DebugLoc(), VT, &Ops[0], Ops.size()); 1024 } 1025 return Result; 1026 } 1027 1028 SDValue SelectionDAG::getConstantFP(double Val, EVT VT, bool isTarget) { 1029 EVT EltVT = VT.getScalarType(); 1030 if (EltVT==MVT::f32) 1031 return getConstantFP(APFloat((float)Val), VT, isTarget); 1032 else if (EltVT==MVT::f64) 1033 return getConstantFP(APFloat(Val), VT, isTarget); 1034 else if (EltVT==MVT::f80 || EltVT==MVT::f128) { 1035 bool ignored; 1036 APFloat apf = APFloat(Val); 1037 apf.convert(*EVTToAPFloatSemantics(EltVT), APFloat::rmNearestTiesToEven, 1038 &ignored); 1039 return getConstantFP(apf, VT, isTarget); 1040 } else { 1041 assert(0 && "Unsupported type in getConstantFP"); 1042 return SDValue(); 1043 } 1044 } 1045 1046 SDValue SelectionDAG::getGlobalAddress(const GlobalValue *GV, DebugLoc DL, 1047 EVT VT, int64_t Offset, 1048 bool isTargetGA, 1049 unsigned char TargetFlags) { 1050 assert((TargetFlags == 0 || isTargetGA) && 1051 "Cannot set target flags on target-independent globals"); 1052 1053 // Truncate (with sign-extension) the offset value to the pointer size. 1054 EVT PTy = TLI.getPointerTy(); 1055 unsigned BitWidth = PTy.getSizeInBits(); 1056 if (BitWidth < 64) 1057 Offset = (Offset << (64 - BitWidth) >> (64 - BitWidth)); 1058 1059 const GlobalVariable *GVar = dyn_cast<GlobalVariable>(GV); 1060 if (!GVar) { 1061 // If GV is an alias then use the aliasee for determining thread-localness. 1062 if (const GlobalAlias *GA = dyn_cast<GlobalAlias>(GV)) 1063 GVar = dyn_cast_or_null<GlobalVariable>(GA->resolveAliasedGlobal(false)); 1064 } 1065 1066 unsigned Opc; 1067 if (GVar && GVar->isThreadLocal()) 1068 Opc = isTargetGA ? ISD::TargetGlobalTLSAddress : ISD::GlobalTLSAddress; 1069 else 1070 Opc = isTargetGA ? ISD::TargetGlobalAddress : ISD::GlobalAddress; 1071 1072 FoldingSetNodeID ID; 1073 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1074 ID.AddPointer(GV); 1075 ID.AddInteger(Offset); 1076 ID.AddInteger(TargetFlags); 1077 void *IP = 0; 1078 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1079 return SDValue(E, 0); 1080 1081 SDNode *N = new (NodeAllocator) GlobalAddressSDNode(Opc, DL, GV, VT, 1082 Offset, TargetFlags); 1083 CSEMap.InsertNode(N, IP); 1084 AllNodes.push_back(N); 1085 return SDValue(N, 0); 1086 } 1087 1088 SDValue SelectionDAG::getFrameIndex(int FI, EVT VT, bool isTarget) { 1089 unsigned Opc = isTarget ? ISD::TargetFrameIndex : ISD::FrameIndex; 1090 FoldingSetNodeID ID; 1091 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1092 ID.AddInteger(FI); 1093 void *IP = 0; 1094 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1095 return SDValue(E, 0); 1096 1097 SDNode *N = new (NodeAllocator) FrameIndexSDNode(FI, VT, isTarget); 1098 CSEMap.InsertNode(N, IP); 1099 AllNodes.push_back(N); 1100 return SDValue(N, 0); 1101 } 1102 1103 SDValue SelectionDAG::getJumpTable(int JTI, EVT VT, bool isTarget, 1104 unsigned char TargetFlags) { 1105 assert((TargetFlags == 0 || isTarget) && 1106 "Cannot set target flags on target-independent jump tables"); 1107 unsigned Opc = isTarget ? ISD::TargetJumpTable : ISD::JumpTable; 1108 FoldingSetNodeID ID; 1109 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1110 ID.AddInteger(JTI); 1111 ID.AddInteger(TargetFlags); 1112 void *IP = 0; 1113 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1114 return SDValue(E, 0); 1115 1116 SDNode *N = new (NodeAllocator) JumpTableSDNode(JTI, VT, isTarget, 1117 TargetFlags); 1118 CSEMap.InsertNode(N, IP); 1119 AllNodes.push_back(N); 1120 return SDValue(N, 0); 1121 } 1122 1123 SDValue SelectionDAG::getConstantPool(const Constant *C, EVT VT, 1124 unsigned Alignment, int Offset, 1125 bool isTarget, 1126 unsigned char TargetFlags) { 1127 assert((TargetFlags == 0 || isTarget) && 1128 "Cannot set target flags on target-independent globals"); 1129 if (Alignment == 0) 1130 Alignment = TLI.getTargetData()->getPrefTypeAlignment(C->getType()); 1131 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool; 1132 FoldingSetNodeID ID; 1133 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1134 ID.AddInteger(Alignment); 1135 ID.AddInteger(Offset); 1136 ID.AddPointer(C); 1137 ID.AddInteger(TargetFlags); 1138 void *IP = 0; 1139 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1140 return SDValue(E, 0); 1141 1142 SDNode *N = new (NodeAllocator) ConstantPoolSDNode(isTarget, C, VT, Offset, 1143 Alignment, TargetFlags); 1144 CSEMap.InsertNode(N, IP); 1145 AllNodes.push_back(N); 1146 return SDValue(N, 0); 1147 } 1148 1149 1150 SDValue SelectionDAG::getConstantPool(MachineConstantPoolValue *C, EVT VT, 1151 unsigned Alignment, int Offset, 1152 bool isTarget, 1153 unsigned char TargetFlags) { 1154 assert((TargetFlags == 0 || isTarget) && 1155 "Cannot set target flags on target-independent globals"); 1156 if (Alignment == 0) 1157 Alignment = TLI.getTargetData()->getPrefTypeAlignment(C->getType()); 1158 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool; 1159 FoldingSetNodeID ID; 1160 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1161 ID.AddInteger(Alignment); 1162 ID.AddInteger(Offset); 1163 C->addSelectionDAGCSEId(ID); 1164 ID.AddInteger(TargetFlags); 1165 void *IP = 0; 1166 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1167 return SDValue(E, 0); 1168 1169 SDNode *N = new (NodeAllocator) ConstantPoolSDNode(isTarget, C, VT, Offset, 1170 Alignment, TargetFlags); 1171 CSEMap.InsertNode(N, IP); 1172 AllNodes.push_back(N); 1173 return SDValue(N, 0); 1174 } 1175 1176 SDValue SelectionDAG::getBasicBlock(MachineBasicBlock *MBB) { 1177 FoldingSetNodeID ID; 1178 AddNodeIDNode(ID, ISD::BasicBlock, getVTList(MVT::Other), 0, 0); 1179 ID.AddPointer(MBB); 1180 void *IP = 0; 1181 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1182 return SDValue(E, 0); 1183 1184 SDNode *N = new (NodeAllocator) BasicBlockSDNode(MBB); 1185 CSEMap.InsertNode(N, IP); 1186 AllNodes.push_back(N); 1187 return SDValue(N, 0); 1188 } 1189 1190 SDValue SelectionDAG::getValueType(EVT VT) { 1191 if (VT.isSimple() && (unsigned)VT.getSimpleVT().SimpleTy >= 1192 ValueTypeNodes.size()) 1193 ValueTypeNodes.resize(VT.getSimpleVT().SimpleTy+1); 1194 1195 SDNode *&N = VT.isExtended() ? 1196 ExtendedValueTypeNodes[VT] : ValueTypeNodes[VT.getSimpleVT().SimpleTy]; 1197 1198 if (N) return SDValue(N, 0); 1199 N = new (NodeAllocator) VTSDNode(VT); 1200 AllNodes.push_back(N); 1201 return SDValue(N, 0); 1202 } 1203 1204 SDValue SelectionDAG::getExternalSymbol(const char *Sym, EVT VT) { 1205 SDNode *&N = ExternalSymbols[Sym]; 1206 if (N) return SDValue(N, 0); 1207 N = new (NodeAllocator) ExternalSymbolSDNode(false, Sym, 0, VT); 1208 AllNodes.push_back(N); 1209 return SDValue(N, 0); 1210 } 1211 1212 SDValue SelectionDAG::getTargetExternalSymbol(const char *Sym, EVT VT, 1213 unsigned char TargetFlags) { 1214 SDNode *&N = 1215 TargetExternalSymbols[std::pair<std::string,unsigned char>(Sym, 1216 TargetFlags)]; 1217 if (N) return SDValue(N, 0); 1218 N = new (NodeAllocator) ExternalSymbolSDNode(true, Sym, TargetFlags, VT); 1219 AllNodes.push_back(N); 1220 return SDValue(N, 0); 1221 } 1222 1223 SDValue SelectionDAG::getCondCode(ISD::CondCode Cond) { 1224 if ((unsigned)Cond >= CondCodeNodes.size()) 1225 CondCodeNodes.resize(Cond+1); 1226 1227 if (CondCodeNodes[Cond] == 0) { 1228 CondCodeSDNode *N = new (NodeAllocator) CondCodeSDNode(Cond); 1229 CondCodeNodes[Cond] = N; 1230 AllNodes.push_back(N); 1231 } 1232 1233 return SDValue(CondCodeNodes[Cond], 0); 1234 } 1235 1236 // commuteShuffle - swaps the values of N1 and N2, and swaps all indices in 1237 // the shuffle mask M that point at N1 to point at N2, and indices that point 1238 // N2 to point at N1. 1239 static void commuteShuffle(SDValue &N1, SDValue &N2, SmallVectorImpl<int> &M) { 1240 std::swap(N1, N2); 1241 int NElts = M.size(); 1242 for (int i = 0; i != NElts; ++i) { 1243 if (M[i] >= NElts) 1244 M[i] -= NElts; 1245 else if (M[i] >= 0) 1246 M[i] += NElts; 1247 } 1248 } 1249 1250 SDValue SelectionDAG::getVectorShuffle(EVT VT, DebugLoc dl, SDValue N1, 1251 SDValue N2, const int *Mask) { 1252 assert(N1.getValueType() == N2.getValueType() && "Invalid VECTOR_SHUFFLE"); 1253 assert(VT.isVector() && N1.getValueType().isVector() && 1254 "Vector Shuffle VTs must be a vectors"); 1255 assert(VT.getVectorElementType() == N1.getValueType().getVectorElementType() 1256 && "Vector Shuffle VTs must have same element type"); 1257 1258 // Canonicalize shuffle undef, undef -> undef 1259 if (N1.getOpcode() == ISD::UNDEF && N2.getOpcode() == ISD::UNDEF) 1260 return getUNDEF(VT); 1261 1262 // Validate that all indices in Mask are within the range of the elements 1263 // input to the shuffle. 1264 unsigned NElts = VT.getVectorNumElements(); 1265 SmallVector<int, 8> MaskVec; 1266 for (unsigned i = 0; i != NElts; ++i) { 1267 assert(Mask[i] < (int)(NElts * 2) && "Index out of range"); 1268 MaskVec.push_back(Mask[i]); 1269 } 1270 1271 // Canonicalize shuffle v, v -> v, undef 1272 if (N1 == N2) { 1273 N2 = getUNDEF(VT); 1274 for (unsigned i = 0; i != NElts; ++i) 1275 if (MaskVec[i] >= (int)NElts) MaskVec[i] -= NElts; 1276 } 1277 1278 // Canonicalize shuffle undef, v -> v, undef. Commute the shuffle mask. 1279 if (N1.getOpcode() == ISD::UNDEF) 1280 commuteShuffle(N1, N2, MaskVec); 1281 1282 // Canonicalize all index into lhs, -> shuffle lhs, undef 1283 // Canonicalize all index into rhs, -> shuffle rhs, undef 1284 bool AllLHS = true, AllRHS = true; 1285 bool N2Undef = N2.getOpcode() == ISD::UNDEF; 1286 for (unsigned i = 0; i != NElts; ++i) { 1287 if (MaskVec[i] >= (int)NElts) { 1288 if (N2Undef) 1289 MaskVec[i] = -1; 1290 else 1291 AllLHS = false; 1292 } else if (MaskVec[i] >= 0) { 1293 AllRHS = false; 1294 } 1295 } 1296 if (AllLHS && AllRHS) 1297 return getUNDEF(VT); 1298 if (AllLHS && !N2Undef) 1299 N2 = getUNDEF(VT); 1300 if (AllRHS) { 1301 N1 = getUNDEF(VT); 1302 commuteShuffle(N1, N2, MaskVec); 1303 } 1304 1305 // If Identity shuffle, or all shuffle in to undef, return that node. 1306 bool AllUndef = true; 1307 bool Identity = true; 1308 for (unsigned i = 0; i != NElts; ++i) { 1309 if (MaskVec[i] >= 0 && MaskVec[i] != (int)i) Identity = false; 1310 if (MaskVec[i] >= 0) AllUndef = false; 1311 } 1312 if (Identity && NElts == N1.getValueType().getVectorNumElements()) 1313 return N1; 1314 if (AllUndef) 1315 return getUNDEF(VT); 1316 1317 FoldingSetNodeID ID; 1318 SDValue Ops[2] = { N1, N2 }; 1319 AddNodeIDNode(ID, ISD::VECTOR_SHUFFLE, getVTList(VT), Ops, 2); 1320 for (unsigned i = 0; i != NElts; ++i) 1321 ID.AddInteger(MaskVec[i]); 1322 1323 void* IP = 0; 1324 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1325 return SDValue(E, 0); 1326 1327 // Allocate the mask array for the node out of the BumpPtrAllocator, since 1328 // SDNode doesn't have access to it. This memory will be "leaked" when 1329 // the node is deallocated, but recovered when the NodeAllocator is released. 1330 int *MaskAlloc = OperandAllocator.Allocate<int>(NElts); 1331 memcpy(MaskAlloc, &MaskVec[0], NElts * sizeof(int)); 1332 1333 ShuffleVectorSDNode *N = 1334 new (NodeAllocator) ShuffleVectorSDNode(VT, dl, N1, N2, MaskAlloc); 1335 CSEMap.InsertNode(N, IP); 1336 AllNodes.push_back(N); 1337 return SDValue(N, 0); 1338 } 1339 1340 SDValue SelectionDAG::getConvertRndSat(EVT VT, DebugLoc dl, 1341 SDValue Val, SDValue DTy, 1342 SDValue STy, SDValue Rnd, SDValue Sat, 1343 ISD::CvtCode Code) { 1344 // If the src and dest types are the same and the conversion is between 1345 // integer types of the same sign or two floats, no conversion is necessary. 1346 if (DTy == STy && 1347 (Code == ISD::CVT_UU || Code == ISD::CVT_SS || Code == ISD::CVT_FF)) 1348 return Val; 1349 1350 FoldingSetNodeID ID; 1351 SDValue Ops[] = { Val, DTy, STy, Rnd, Sat }; 1352 AddNodeIDNode(ID, ISD::CONVERT_RNDSAT, getVTList(VT), &Ops[0], 5); 1353 void* IP = 0; 1354 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1355 return SDValue(E, 0); 1356 1357 CvtRndSatSDNode *N = new (NodeAllocator) CvtRndSatSDNode(VT, dl, Ops, 5, 1358 Code); 1359 CSEMap.InsertNode(N, IP); 1360 AllNodes.push_back(N); 1361 return SDValue(N, 0); 1362 } 1363 1364 SDValue SelectionDAG::getRegister(unsigned RegNo, EVT VT) { 1365 FoldingSetNodeID ID; 1366 AddNodeIDNode(ID, ISD::Register, getVTList(VT), 0, 0); 1367 ID.AddInteger(RegNo); 1368 void *IP = 0; 1369 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1370 return SDValue(E, 0); 1371 1372 SDNode *N = new (NodeAllocator) RegisterSDNode(RegNo, VT); 1373 CSEMap.InsertNode(N, IP); 1374 AllNodes.push_back(N); 1375 return SDValue(N, 0); 1376 } 1377 1378 SDValue SelectionDAG::getEHLabel(DebugLoc dl, SDValue Root, MCSymbol *Label) { 1379 FoldingSetNodeID ID; 1380 SDValue Ops[] = { Root }; 1381 AddNodeIDNode(ID, ISD::EH_LABEL, getVTList(MVT::Other), &Ops[0], 1); 1382 ID.AddPointer(Label); 1383 void *IP = 0; 1384 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1385 return SDValue(E, 0); 1386 1387 SDNode *N = new (NodeAllocator) EHLabelSDNode(dl, Root, Label); 1388 CSEMap.InsertNode(N, IP); 1389 AllNodes.push_back(N); 1390 return SDValue(N, 0); 1391 } 1392 1393 1394 SDValue SelectionDAG::getBlockAddress(const BlockAddress *BA, EVT VT, 1395 bool isTarget, 1396 unsigned char TargetFlags) { 1397 unsigned Opc = isTarget ? ISD::TargetBlockAddress : ISD::BlockAddress; 1398 1399 FoldingSetNodeID ID; 1400 AddNodeIDNode(ID, Opc, getVTList(VT), 0, 0); 1401 ID.AddPointer(BA); 1402 ID.AddInteger(TargetFlags); 1403 void *IP = 0; 1404 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1405 return SDValue(E, 0); 1406 1407 SDNode *N = new (NodeAllocator) BlockAddressSDNode(Opc, VT, BA, TargetFlags); 1408 CSEMap.InsertNode(N, IP); 1409 AllNodes.push_back(N); 1410 return SDValue(N, 0); 1411 } 1412 1413 SDValue SelectionDAG::getSrcValue(const Value *V) { 1414 assert((!V || V->getType()->isPointerTy()) && 1415 "SrcValue is not a pointer?"); 1416 1417 FoldingSetNodeID ID; 1418 AddNodeIDNode(ID, ISD::SRCVALUE, getVTList(MVT::Other), 0, 0); 1419 ID.AddPointer(V); 1420 1421 void *IP = 0; 1422 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1423 return SDValue(E, 0); 1424 1425 SDNode *N = new (NodeAllocator) SrcValueSDNode(V); 1426 CSEMap.InsertNode(N, IP); 1427 AllNodes.push_back(N); 1428 return SDValue(N, 0); 1429 } 1430 1431 /// getMDNode - Return an MDNodeSDNode which holds an MDNode. 1432 SDValue SelectionDAG::getMDNode(const MDNode *MD) { 1433 FoldingSetNodeID ID; 1434 AddNodeIDNode(ID, ISD::MDNODE_SDNODE, getVTList(MVT::Other), 0, 0); 1435 ID.AddPointer(MD); 1436 1437 void *IP = 0; 1438 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 1439 return SDValue(E, 0); 1440 1441 SDNode *N = new (NodeAllocator) MDNodeSDNode(MD); 1442 CSEMap.InsertNode(N, IP); 1443 AllNodes.push_back(N); 1444 return SDValue(N, 0); 1445 } 1446 1447 1448 /// getShiftAmountOperand - Return the specified value casted to 1449 /// the target's desired shift amount type. 1450 SDValue SelectionDAG::getShiftAmountOperand(EVT LHSTy, SDValue Op) { 1451 EVT OpTy = Op.getValueType(); 1452 MVT ShTy = TLI.getShiftAmountTy(LHSTy); 1453 if (OpTy == ShTy || OpTy.isVector()) return Op; 1454 1455 ISD::NodeType Opcode = OpTy.bitsGT(ShTy) ? ISD::TRUNCATE : ISD::ZERO_EXTEND; 1456 return getNode(Opcode, Op.getDebugLoc(), ShTy, Op); 1457 } 1458 1459 /// CreateStackTemporary - Create a stack temporary, suitable for holding the 1460 /// specified value type. 1461 SDValue SelectionDAG::CreateStackTemporary(EVT VT, unsigned minAlign) { 1462 MachineFrameInfo *FrameInfo = getMachineFunction().getFrameInfo(); 1463 unsigned ByteSize = VT.getStoreSize(); 1464 Type *Ty = VT.getTypeForEVT(*getContext()); 1465 unsigned StackAlign = 1466 std::max((unsigned)TLI.getTargetData()->getPrefTypeAlignment(Ty), minAlign); 1467 1468 int FrameIdx = FrameInfo->CreateStackObject(ByteSize, StackAlign, false); 1469 return getFrameIndex(FrameIdx, TLI.getPointerTy()); 1470 } 1471 1472 /// CreateStackTemporary - Create a stack temporary suitable for holding 1473 /// either of the specified value types. 1474 SDValue SelectionDAG::CreateStackTemporary(EVT VT1, EVT VT2) { 1475 unsigned Bytes = std::max(VT1.getStoreSizeInBits(), 1476 VT2.getStoreSizeInBits())/8; 1477 Type *Ty1 = VT1.getTypeForEVT(*getContext()); 1478 Type *Ty2 = VT2.getTypeForEVT(*getContext()); 1479 const TargetData *TD = TLI.getTargetData(); 1480 unsigned Align = std::max(TD->getPrefTypeAlignment(Ty1), 1481 TD->getPrefTypeAlignment(Ty2)); 1482 1483 MachineFrameInfo *FrameInfo = getMachineFunction().getFrameInfo(); 1484 int FrameIdx = FrameInfo->CreateStackObject(Bytes, Align, false); 1485 return getFrameIndex(FrameIdx, TLI.getPointerTy()); 1486 } 1487 1488 SDValue SelectionDAG::FoldSetCC(EVT VT, SDValue N1, 1489 SDValue N2, ISD::CondCode Cond, DebugLoc dl) { 1490 // These setcc operations always fold. 1491 switch (Cond) { 1492 default: break; 1493 case ISD::SETFALSE: 1494 case ISD::SETFALSE2: return getConstant(0, VT); 1495 case ISD::SETTRUE: 1496 case ISD::SETTRUE2: return getConstant(1, VT); 1497 1498 case ISD::SETOEQ: 1499 case ISD::SETOGT: 1500 case ISD::SETOGE: 1501 case ISD::SETOLT: 1502 case ISD::SETOLE: 1503 case ISD::SETONE: 1504 case ISD::SETO: 1505 case ISD::SETUO: 1506 case ISD::SETUEQ: 1507 case ISD::SETUNE: 1508 assert(!N1.getValueType().isInteger() && "Illegal setcc for integer!"); 1509 break; 1510 } 1511 1512 if (ConstantSDNode *N2C = dyn_cast<ConstantSDNode>(N2.getNode())) { 1513 const APInt &C2 = N2C->getAPIntValue(); 1514 if (ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.getNode())) { 1515 const APInt &C1 = N1C->getAPIntValue(); 1516 1517 switch (Cond) { 1518 default: llvm_unreachable("Unknown integer setcc!"); 1519 case ISD::SETEQ: return getConstant(C1 == C2, VT); 1520 case ISD::SETNE: return getConstant(C1 != C2, VT); 1521 case ISD::SETULT: return getConstant(C1.ult(C2), VT); 1522 case ISD::SETUGT: return getConstant(C1.ugt(C2), VT); 1523 case ISD::SETULE: return getConstant(C1.ule(C2), VT); 1524 case ISD::SETUGE: return getConstant(C1.uge(C2), VT); 1525 case ISD::SETLT: return getConstant(C1.slt(C2), VT); 1526 case ISD::SETGT: return getConstant(C1.sgt(C2), VT); 1527 case ISD::SETLE: return getConstant(C1.sle(C2), VT); 1528 case ISD::SETGE: return getConstant(C1.sge(C2), VT); 1529 } 1530 } 1531 } 1532 if (ConstantFPSDNode *N1C = dyn_cast<ConstantFPSDNode>(N1.getNode())) { 1533 if (ConstantFPSDNode *N2C = dyn_cast<ConstantFPSDNode>(N2.getNode())) { 1534 // No compile time operations on this type yet. 1535 if (N1C->getValueType(0) == MVT::ppcf128) 1536 return SDValue(); 1537 1538 APFloat::cmpResult R = N1C->getValueAPF().compare(N2C->getValueAPF()); 1539 switch (Cond) { 1540 default: break; 1541 case ISD::SETEQ: if (R==APFloat::cmpUnordered) 1542 return getUNDEF(VT); 1543 // fall through 1544 case ISD::SETOEQ: return getConstant(R==APFloat::cmpEqual, VT); 1545 case ISD::SETNE: if (R==APFloat::cmpUnordered) 1546 return getUNDEF(VT); 1547 // fall through 1548 case ISD::SETONE: return getConstant(R==APFloat::cmpGreaterThan || 1549 R==APFloat::cmpLessThan, VT); 1550 case ISD::SETLT: if (R==APFloat::cmpUnordered) 1551 return getUNDEF(VT); 1552 // fall through 1553 case ISD::SETOLT: return getConstant(R==APFloat::cmpLessThan, VT); 1554 case ISD::SETGT: if (R==APFloat::cmpUnordered) 1555 return getUNDEF(VT); 1556 // fall through 1557 case ISD::SETOGT: return getConstant(R==APFloat::cmpGreaterThan, VT); 1558 case ISD::SETLE: if (R==APFloat::cmpUnordered) 1559 return getUNDEF(VT); 1560 // fall through 1561 case ISD::SETOLE: return getConstant(R==APFloat::cmpLessThan || 1562 R==APFloat::cmpEqual, VT); 1563 case ISD::SETGE: if (R==APFloat::cmpUnordered) 1564 return getUNDEF(VT); 1565 // fall through 1566 case ISD::SETOGE: return getConstant(R==APFloat::cmpGreaterThan || 1567 R==APFloat::cmpEqual, VT); 1568 case ISD::SETO: return getConstant(R!=APFloat::cmpUnordered, VT); 1569 case ISD::SETUO: return getConstant(R==APFloat::cmpUnordered, VT); 1570 case ISD::SETUEQ: return getConstant(R==APFloat::cmpUnordered || 1571 R==APFloat::cmpEqual, VT); 1572 case ISD::SETUNE: return getConstant(R!=APFloat::cmpEqual, VT); 1573 case ISD::SETULT: return getConstant(R==APFloat::cmpUnordered || 1574 R==APFloat::cmpLessThan, VT); 1575 case ISD::SETUGT: return getConstant(R==APFloat::cmpGreaterThan || 1576 R==APFloat::cmpUnordered, VT); 1577 case ISD::SETULE: return getConstant(R!=APFloat::cmpGreaterThan, VT); 1578 case ISD::SETUGE: return getConstant(R!=APFloat::cmpLessThan, VT); 1579 } 1580 } else { 1581 // Ensure that the constant occurs on the RHS. 1582 return getSetCC(dl, VT, N2, N1, ISD::getSetCCSwappedOperands(Cond)); 1583 } 1584 } 1585 1586 // Could not fold it. 1587 return SDValue(); 1588 } 1589 1590 /// SignBitIsZero - Return true if the sign bit of Op is known to be zero. We 1591 /// use this predicate to simplify operations downstream. 1592 bool SelectionDAG::SignBitIsZero(SDValue Op, unsigned Depth) const { 1593 // This predicate is not safe for vector operations. 1594 if (Op.getValueType().isVector()) 1595 return false; 1596 1597 unsigned BitWidth = Op.getValueType().getScalarType().getSizeInBits(); 1598 return MaskedValueIsZero(Op, APInt::getSignBit(BitWidth), Depth); 1599 } 1600 1601 /// MaskedValueIsZero - Return true if 'V & Mask' is known to be zero. We use 1602 /// this predicate to simplify operations downstream. Mask is known to be zero 1603 /// for bits that V cannot have. 1604 bool SelectionDAG::MaskedValueIsZero(SDValue Op, const APInt &Mask, 1605 unsigned Depth) const { 1606 APInt KnownZero, KnownOne; 1607 ComputeMaskedBits(Op, Mask, KnownZero, KnownOne, Depth); 1608 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1609 return (KnownZero & Mask) == Mask; 1610 } 1611 1612 /// ComputeMaskedBits - Determine which of the bits specified in Mask are 1613 /// known to be either zero or one and return them in the KnownZero/KnownOne 1614 /// bitsets. This code only analyzes bits in Mask, in order to short-circuit 1615 /// processing. 1616 void SelectionDAG::ComputeMaskedBits(SDValue Op, const APInt &Mask, 1617 APInt &KnownZero, APInt &KnownOne, 1618 unsigned Depth) const { 1619 unsigned BitWidth = Mask.getBitWidth(); 1620 assert(BitWidth == Op.getValueType().getScalarType().getSizeInBits() && 1621 "Mask size mismatches value type size!"); 1622 1623 KnownZero = KnownOne = APInt(BitWidth, 0); // Don't know anything. 1624 if (Depth == 6 || Mask == 0) 1625 return; // Limit search depth. 1626 1627 APInt KnownZero2, KnownOne2; 1628 1629 switch (Op.getOpcode()) { 1630 case ISD::Constant: 1631 // We know all of the bits for a constant! 1632 KnownOne = cast<ConstantSDNode>(Op)->getAPIntValue() & Mask; 1633 KnownZero = ~KnownOne & Mask; 1634 return; 1635 case ISD::AND: 1636 // If either the LHS or the RHS are Zero, the result is zero. 1637 ComputeMaskedBits(Op.getOperand(1), Mask, KnownZero, KnownOne, Depth+1); 1638 ComputeMaskedBits(Op.getOperand(0), Mask & ~KnownZero, 1639 KnownZero2, KnownOne2, Depth+1); 1640 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1641 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1642 1643 // Output known-1 bits are only known if set in both the LHS & RHS. 1644 KnownOne &= KnownOne2; 1645 // Output known-0 are known to be clear if zero in either the LHS | RHS. 1646 KnownZero |= KnownZero2; 1647 return; 1648 case ISD::OR: 1649 ComputeMaskedBits(Op.getOperand(1), Mask, KnownZero, KnownOne, Depth+1); 1650 ComputeMaskedBits(Op.getOperand(0), Mask & ~KnownOne, 1651 KnownZero2, KnownOne2, Depth+1); 1652 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1653 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1654 1655 // Output known-0 bits are only known if clear in both the LHS & RHS. 1656 KnownZero &= KnownZero2; 1657 // Output known-1 are known to be set if set in either the LHS | RHS. 1658 KnownOne |= KnownOne2; 1659 return; 1660 case ISD::XOR: { 1661 ComputeMaskedBits(Op.getOperand(1), Mask, KnownZero, KnownOne, Depth+1); 1662 ComputeMaskedBits(Op.getOperand(0), Mask, KnownZero2, KnownOne2, Depth+1); 1663 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1664 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1665 1666 // Output known-0 bits are known if clear or set in both the LHS & RHS. 1667 APInt KnownZeroOut = (KnownZero & KnownZero2) | (KnownOne & KnownOne2); 1668 // Output known-1 are known to be set if set in only one of the LHS, RHS. 1669 KnownOne = (KnownZero & KnownOne2) | (KnownOne & KnownZero2); 1670 KnownZero = KnownZeroOut; 1671 return; 1672 } 1673 case ISD::MUL: { 1674 APInt Mask2 = APInt::getAllOnesValue(BitWidth); 1675 ComputeMaskedBits(Op.getOperand(1), Mask2, KnownZero, KnownOne, Depth+1); 1676 ComputeMaskedBits(Op.getOperand(0), Mask2, KnownZero2, KnownOne2, Depth+1); 1677 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1678 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1679 1680 // If low bits are zero in either operand, output low known-0 bits. 1681 // Also compute a conserative estimate for high known-0 bits. 1682 // More trickiness is possible, but this is sufficient for the 1683 // interesting case of alignment computation. 1684 KnownOne.clearAllBits(); 1685 unsigned TrailZ = KnownZero.countTrailingOnes() + 1686 KnownZero2.countTrailingOnes(); 1687 unsigned LeadZ = std::max(KnownZero.countLeadingOnes() + 1688 KnownZero2.countLeadingOnes(), 1689 BitWidth) - BitWidth; 1690 1691 TrailZ = std::min(TrailZ, BitWidth); 1692 LeadZ = std::min(LeadZ, BitWidth); 1693 KnownZero = APInt::getLowBitsSet(BitWidth, TrailZ) | 1694 APInt::getHighBitsSet(BitWidth, LeadZ); 1695 KnownZero &= Mask; 1696 return; 1697 } 1698 case ISD::UDIV: { 1699 // For the purposes of computing leading zeros we can conservatively 1700 // treat a udiv as a logical right shift by the power of 2 known to 1701 // be less than the denominator. 1702 APInt AllOnes = APInt::getAllOnesValue(BitWidth); 1703 ComputeMaskedBits(Op.getOperand(0), 1704 AllOnes, KnownZero2, KnownOne2, Depth+1); 1705 unsigned LeadZ = KnownZero2.countLeadingOnes(); 1706 1707 KnownOne2.clearAllBits(); 1708 KnownZero2.clearAllBits(); 1709 ComputeMaskedBits(Op.getOperand(1), 1710 AllOnes, KnownZero2, KnownOne2, Depth+1); 1711 unsigned RHSUnknownLeadingOnes = KnownOne2.countLeadingZeros(); 1712 if (RHSUnknownLeadingOnes != BitWidth) 1713 LeadZ = std::min(BitWidth, 1714 LeadZ + BitWidth - RHSUnknownLeadingOnes - 1); 1715 1716 KnownZero = APInt::getHighBitsSet(BitWidth, LeadZ) & Mask; 1717 return; 1718 } 1719 case ISD::SELECT: 1720 ComputeMaskedBits(Op.getOperand(2), Mask, KnownZero, KnownOne, Depth+1); 1721 ComputeMaskedBits(Op.getOperand(1), Mask, KnownZero2, KnownOne2, Depth+1); 1722 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1723 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1724 1725 // Only known if known in both the LHS and RHS. 1726 KnownOne &= KnownOne2; 1727 KnownZero &= KnownZero2; 1728 return; 1729 case ISD::SELECT_CC: 1730 ComputeMaskedBits(Op.getOperand(3), Mask, KnownZero, KnownOne, Depth+1); 1731 ComputeMaskedBits(Op.getOperand(2), Mask, KnownZero2, KnownOne2, Depth+1); 1732 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1733 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1734 1735 // Only known if known in both the LHS and RHS. 1736 KnownOne &= KnownOne2; 1737 KnownZero &= KnownZero2; 1738 return; 1739 case ISD::SADDO: 1740 case ISD::UADDO: 1741 case ISD::SSUBO: 1742 case ISD::USUBO: 1743 case ISD::SMULO: 1744 case ISD::UMULO: 1745 if (Op.getResNo() != 1) 1746 return; 1747 // The boolean result conforms to getBooleanContents. Fall through. 1748 case ISD::SETCC: 1749 // If we know the result of a setcc has the top bits zero, use this info. 1750 if (TLI.getBooleanContents(Op.getValueType().isVector()) == 1751 TargetLowering::ZeroOrOneBooleanContent && BitWidth > 1) 1752 KnownZero |= APInt::getHighBitsSet(BitWidth, BitWidth - 1); 1753 return; 1754 case ISD::SHL: 1755 // (shl X, C1) & C2 == 0 iff (X & C2 >>u C1) == 0 1756 if (ConstantSDNode *SA = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 1757 unsigned ShAmt = SA->getZExtValue(); 1758 1759 // If the shift count is an invalid immediate, don't do anything. 1760 if (ShAmt >= BitWidth) 1761 return; 1762 1763 ComputeMaskedBits(Op.getOperand(0), Mask.lshr(ShAmt), 1764 KnownZero, KnownOne, Depth+1); 1765 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1766 KnownZero <<= ShAmt; 1767 KnownOne <<= ShAmt; 1768 // low bits known zero. 1769 KnownZero |= APInt::getLowBitsSet(BitWidth, ShAmt); 1770 } 1771 return; 1772 case ISD::SRL: 1773 // (ushr X, C1) & C2 == 0 iff (-1 >> C1) & C2 == 0 1774 if (ConstantSDNode *SA = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 1775 unsigned ShAmt = SA->getZExtValue(); 1776 1777 // If the shift count is an invalid immediate, don't do anything. 1778 if (ShAmt >= BitWidth) 1779 return; 1780 1781 ComputeMaskedBits(Op.getOperand(0), (Mask << ShAmt), 1782 KnownZero, KnownOne, Depth+1); 1783 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1784 KnownZero = KnownZero.lshr(ShAmt); 1785 KnownOne = KnownOne.lshr(ShAmt); 1786 1787 APInt HighBits = APInt::getHighBitsSet(BitWidth, ShAmt) & Mask; 1788 KnownZero |= HighBits; // High bits known zero. 1789 } 1790 return; 1791 case ISD::SRA: 1792 if (ConstantSDNode *SA = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 1793 unsigned ShAmt = SA->getZExtValue(); 1794 1795 // If the shift count is an invalid immediate, don't do anything. 1796 if (ShAmt >= BitWidth) 1797 return; 1798 1799 APInt InDemandedMask = (Mask << ShAmt); 1800 // If any of the demanded bits are produced by the sign extension, we also 1801 // demand the input sign bit. 1802 APInt HighBits = APInt::getHighBitsSet(BitWidth, ShAmt) & Mask; 1803 if (HighBits.getBoolValue()) 1804 InDemandedMask |= APInt::getSignBit(BitWidth); 1805 1806 ComputeMaskedBits(Op.getOperand(0), InDemandedMask, KnownZero, KnownOne, 1807 Depth+1); 1808 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1809 KnownZero = KnownZero.lshr(ShAmt); 1810 KnownOne = KnownOne.lshr(ShAmt); 1811 1812 // Handle the sign bits. 1813 APInt SignBit = APInt::getSignBit(BitWidth); 1814 SignBit = SignBit.lshr(ShAmt); // Adjust to where it is now in the mask. 1815 1816 if (KnownZero.intersects(SignBit)) { 1817 KnownZero |= HighBits; // New bits are known zero. 1818 } else if (KnownOne.intersects(SignBit)) { 1819 KnownOne |= HighBits; // New bits are known one. 1820 } 1821 } 1822 return; 1823 case ISD::SIGN_EXTEND_INREG: { 1824 EVT EVT = cast<VTSDNode>(Op.getOperand(1))->getVT(); 1825 unsigned EBits = EVT.getScalarType().getSizeInBits(); 1826 1827 // Sign extension. Compute the demanded bits in the result that are not 1828 // present in the input. 1829 APInt NewBits = APInt::getHighBitsSet(BitWidth, BitWidth - EBits) & Mask; 1830 1831 APInt InSignBit = APInt::getSignBit(EBits); 1832 APInt InputDemandedBits = Mask & APInt::getLowBitsSet(BitWidth, EBits); 1833 1834 // If the sign extended bits are demanded, we know that the sign 1835 // bit is demanded. 1836 InSignBit = InSignBit.zext(BitWidth); 1837 if (NewBits.getBoolValue()) 1838 InputDemandedBits |= InSignBit; 1839 1840 ComputeMaskedBits(Op.getOperand(0), InputDemandedBits, 1841 KnownZero, KnownOne, Depth+1); 1842 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1843 1844 // If the sign bit of the input is known set or clear, then we know the 1845 // top bits of the result. 1846 if (KnownZero.intersects(InSignBit)) { // Input sign bit known clear 1847 KnownZero |= NewBits; 1848 KnownOne &= ~NewBits; 1849 } else if (KnownOne.intersects(InSignBit)) { // Input sign bit known set 1850 KnownOne |= NewBits; 1851 KnownZero &= ~NewBits; 1852 } else { // Input sign bit unknown 1853 KnownZero &= ~NewBits; 1854 KnownOne &= ~NewBits; 1855 } 1856 return; 1857 } 1858 case ISD::CTTZ: 1859 case ISD::CTTZ_ZERO_UNDEF: 1860 case ISD::CTLZ: 1861 case ISD::CTLZ_ZERO_UNDEF: 1862 case ISD::CTPOP: { 1863 unsigned LowBits = Log2_32(BitWidth)+1; 1864 KnownZero = APInt::getHighBitsSet(BitWidth, BitWidth - LowBits); 1865 KnownOne.clearAllBits(); 1866 return; 1867 } 1868 case ISD::LOAD: { 1869 if (ISD::isZEXTLoad(Op.getNode())) { 1870 LoadSDNode *LD = cast<LoadSDNode>(Op); 1871 EVT VT = LD->getMemoryVT(); 1872 unsigned MemBits = VT.getScalarType().getSizeInBits(); 1873 KnownZero |= APInt::getHighBitsSet(BitWidth, BitWidth - MemBits) & Mask; 1874 } 1875 return; 1876 } 1877 case ISD::ZERO_EXTEND: { 1878 EVT InVT = Op.getOperand(0).getValueType(); 1879 unsigned InBits = InVT.getScalarType().getSizeInBits(); 1880 APInt NewBits = APInt::getHighBitsSet(BitWidth, BitWidth - InBits) & Mask; 1881 APInt InMask = Mask.trunc(InBits); 1882 KnownZero = KnownZero.trunc(InBits); 1883 KnownOne = KnownOne.trunc(InBits); 1884 ComputeMaskedBits(Op.getOperand(0), InMask, KnownZero, KnownOne, Depth+1); 1885 KnownZero = KnownZero.zext(BitWidth); 1886 KnownOne = KnownOne.zext(BitWidth); 1887 KnownZero |= NewBits; 1888 return; 1889 } 1890 case ISD::SIGN_EXTEND: { 1891 EVT InVT = Op.getOperand(0).getValueType(); 1892 unsigned InBits = InVT.getScalarType().getSizeInBits(); 1893 APInt InSignBit = APInt::getSignBit(InBits); 1894 APInt NewBits = APInt::getHighBitsSet(BitWidth, BitWidth - InBits) & Mask; 1895 APInt InMask = Mask.trunc(InBits); 1896 1897 // If any of the sign extended bits are demanded, we know that the sign 1898 // bit is demanded. Temporarily set this bit in the mask for our callee. 1899 if (NewBits.getBoolValue()) 1900 InMask |= InSignBit; 1901 1902 KnownZero = KnownZero.trunc(InBits); 1903 KnownOne = KnownOne.trunc(InBits); 1904 ComputeMaskedBits(Op.getOperand(0), InMask, KnownZero, KnownOne, Depth+1); 1905 1906 // Note if the sign bit is known to be zero or one. 1907 bool SignBitKnownZero = KnownZero.isNegative(); 1908 bool SignBitKnownOne = KnownOne.isNegative(); 1909 assert(!(SignBitKnownZero && SignBitKnownOne) && 1910 "Sign bit can't be known to be both zero and one!"); 1911 1912 // If the sign bit wasn't actually demanded by our caller, we don't 1913 // want it set in the KnownZero and KnownOne result values. Reset the 1914 // mask and reapply it to the result values. 1915 InMask = Mask.trunc(InBits); 1916 KnownZero &= InMask; 1917 KnownOne &= InMask; 1918 1919 KnownZero = KnownZero.zext(BitWidth); 1920 KnownOne = KnownOne.zext(BitWidth); 1921 1922 // If the sign bit is known zero or one, the top bits match. 1923 if (SignBitKnownZero) 1924 KnownZero |= NewBits; 1925 else if (SignBitKnownOne) 1926 KnownOne |= NewBits; 1927 return; 1928 } 1929 case ISD::ANY_EXTEND: { 1930 EVT InVT = Op.getOperand(0).getValueType(); 1931 unsigned InBits = InVT.getScalarType().getSizeInBits(); 1932 APInt InMask = Mask.trunc(InBits); 1933 KnownZero = KnownZero.trunc(InBits); 1934 KnownOne = KnownOne.trunc(InBits); 1935 ComputeMaskedBits(Op.getOperand(0), InMask, KnownZero, KnownOne, Depth+1); 1936 KnownZero = KnownZero.zext(BitWidth); 1937 KnownOne = KnownOne.zext(BitWidth); 1938 return; 1939 } 1940 case ISD::TRUNCATE: { 1941 EVT InVT = Op.getOperand(0).getValueType(); 1942 unsigned InBits = InVT.getScalarType().getSizeInBits(); 1943 APInt InMask = Mask.zext(InBits); 1944 KnownZero = KnownZero.zext(InBits); 1945 KnownOne = KnownOne.zext(InBits); 1946 ComputeMaskedBits(Op.getOperand(0), InMask, KnownZero, KnownOne, Depth+1); 1947 assert((KnownZero & KnownOne) == 0 && "Bits known to be one AND zero?"); 1948 KnownZero = KnownZero.trunc(BitWidth); 1949 KnownOne = KnownOne.trunc(BitWidth); 1950 break; 1951 } 1952 case ISD::AssertZext: { 1953 EVT VT = cast<VTSDNode>(Op.getOperand(1))->getVT(); 1954 APInt InMask = APInt::getLowBitsSet(BitWidth, VT.getSizeInBits()); 1955 ComputeMaskedBits(Op.getOperand(0), Mask & InMask, KnownZero, 1956 KnownOne, Depth+1); 1957 KnownZero |= (~InMask) & Mask; 1958 return; 1959 } 1960 case ISD::FGETSIGN: 1961 // All bits are zero except the low bit. 1962 KnownZero = APInt::getHighBitsSet(BitWidth, BitWidth - 1); 1963 return; 1964 1965 case ISD::SUB: { 1966 if (ConstantSDNode *CLHS = dyn_cast<ConstantSDNode>(Op.getOperand(0))) { 1967 // We know that the top bits of C-X are clear if X contains less bits 1968 // than C (i.e. no wrap-around can happen). For example, 20-X is 1969 // positive if we can prove that X is >= 0 and < 16. 1970 if (CLHS->getAPIntValue().isNonNegative()) { 1971 unsigned NLZ = (CLHS->getAPIntValue()+1).countLeadingZeros(); 1972 // NLZ can't be BitWidth with no sign bit 1973 APInt MaskV = APInt::getHighBitsSet(BitWidth, NLZ+1); 1974 ComputeMaskedBits(Op.getOperand(1), MaskV, KnownZero2, KnownOne2, 1975 Depth+1); 1976 1977 // If all of the MaskV bits are known to be zero, then we know the 1978 // output top bits are zero, because we now know that the output is 1979 // from [0-C]. 1980 if ((KnownZero2 & MaskV) == MaskV) { 1981 unsigned NLZ2 = CLHS->getAPIntValue().countLeadingZeros(); 1982 // Top bits known zero. 1983 KnownZero = APInt::getHighBitsSet(BitWidth, NLZ2) & Mask; 1984 } 1985 } 1986 } 1987 } 1988 // fall through 1989 case ISD::ADD: 1990 case ISD::ADDE: { 1991 // Output known-0 bits are known if clear or set in both the low clear bits 1992 // common to both LHS & RHS. For example, 8+(X<<3) is known to have the 1993 // low 3 bits clear. 1994 APInt Mask2 = APInt::getLowBitsSet(BitWidth, 1995 BitWidth - Mask.countLeadingZeros()); 1996 ComputeMaskedBits(Op.getOperand(0), Mask2, KnownZero2, KnownOne2, Depth+1); 1997 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 1998 unsigned KnownZeroOut = KnownZero2.countTrailingOnes(); 1999 2000 ComputeMaskedBits(Op.getOperand(1), Mask2, KnownZero2, KnownOne2, Depth+1); 2001 assert((KnownZero2 & KnownOne2) == 0 && "Bits known to be one AND zero?"); 2002 KnownZeroOut = std::min(KnownZeroOut, 2003 KnownZero2.countTrailingOnes()); 2004 2005 if (Op.getOpcode() == ISD::ADD) { 2006 KnownZero |= APInt::getLowBitsSet(BitWidth, KnownZeroOut); 2007 return; 2008 } 2009 2010 // With ADDE, a carry bit may be added in, so we can only use this 2011 // information if we know (at least) that the low two bits are clear. We 2012 // then return to the caller that the low bit is unknown but that other bits 2013 // are known zero. 2014 if (KnownZeroOut >= 2) // ADDE 2015 KnownZero |= APInt::getBitsSet(BitWidth, 1, KnownZeroOut); 2016 return; 2017 } 2018 case ISD::SREM: 2019 if (ConstantSDNode *Rem = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 2020 const APInt &RA = Rem->getAPIntValue().abs(); 2021 if (RA.isPowerOf2()) { 2022 APInt LowBits = RA - 1; 2023 APInt Mask2 = LowBits | APInt::getSignBit(BitWidth); 2024 ComputeMaskedBits(Op.getOperand(0), Mask2,KnownZero2,KnownOne2,Depth+1); 2025 2026 // The low bits of the first operand are unchanged by the srem. 2027 KnownZero = KnownZero2 & LowBits; 2028 KnownOne = KnownOne2 & LowBits; 2029 2030 // If the first operand is non-negative or has all low bits zero, then 2031 // the upper bits are all zero. 2032 if (KnownZero2[BitWidth-1] || ((KnownZero2 & LowBits) == LowBits)) 2033 KnownZero |= ~LowBits; 2034 2035 // If the first operand is negative and not all low bits are zero, then 2036 // the upper bits are all one. 2037 if (KnownOne2[BitWidth-1] && ((KnownOne2 & LowBits) != 0)) 2038 KnownOne |= ~LowBits; 2039 2040 KnownZero &= Mask; 2041 KnownOne &= Mask; 2042 2043 assert((KnownZero & KnownOne) == 0&&"Bits known to be one AND zero?"); 2044 } 2045 } 2046 return; 2047 case ISD::UREM: { 2048 if (ConstantSDNode *Rem = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 2049 const APInt &RA = Rem->getAPIntValue(); 2050 if (RA.isPowerOf2()) { 2051 APInt LowBits = (RA - 1); 2052 APInt Mask2 = LowBits & Mask; 2053 KnownZero |= ~LowBits & Mask; 2054 ComputeMaskedBits(Op.getOperand(0), Mask2, KnownZero, KnownOne,Depth+1); 2055 assert((KnownZero & KnownOne) == 0&&"Bits known to be one AND zero?"); 2056 break; 2057 } 2058 } 2059 2060 // Since the result is less than or equal to either operand, any leading 2061 // zero bits in either operand must also exist in the result. 2062 APInt AllOnes = APInt::getAllOnesValue(BitWidth); 2063 ComputeMaskedBits(Op.getOperand(0), AllOnes, KnownZero, KnownOne, 2064 Depth+1); 2065 ComputeMaskedBits(Op.getOperand(1), AllOnes, KnownZero2, KnownOne2, 2066 Depth+1); 2067 2068 uint32_t Leaders = std::max(KnownZero.countLeadingOnes(), 2069 KnownZero2.countLeadingOnes()); 2070 KnownOne.clearAllBits(); 2071 KnownZero = APInt::getHighBitsSet(BitWidth, Leaders) & Mask; 2072 return; 2073 } 2074 case ISD::FrameIndex: 2075 case ISD::TargetFrameIndex: 2076 if (unsigned Align = InferPtrAlignment(Op)) { 2077 // The low bits are known zero if the pointer is aligned. 2078 KnownZero = APInt::getLowBitsSet(BitWidth, Log2_32(Align)); 2079 return; 2080 } 2081 break; 2082 2083 default: 2084 if (Op.getOpcode() < ISD::BUILTIN_OP_END) 2085 break; 2086 // Fallthrough 2087 case ISD::INTRINSIC_WO_CHAIN: 2088 case ISD::INTRINSIC_W_CHAIN: 2089 case ISD::INTRINSIC_VOID: 2090 // Allow the target to implement this method for its nodes. 2091 TLI.computeMaskedBitsForTargetNode(Op, Mask, KnownZero, KnownOne, *this, 2092 Depth); 2093 return; 2094 } 2095 } 2096 2097 /// ComputeNumSignBits - Return the number of times the sign bit of the 2098 /// register is replicated into the other bits. We know that at least 1 bit 2099 /// is always equal to the sign bit (itself), but other cases can give us 2100 /// information. For example, immediately after an "SRA X, 2", we know that 2101 /// the top 3 bits are all equal to each other, so we return 3. 2102 unsigned SelectionDAG::ComputeNumSignBits(SDValue Op, unsigned Depth) const{ 2103 EVT VT = Op.getValueType(); 2104 assert(VT.isInteger() && "Invalid VT!"); 2105 unsigned VTBits = VT.getScalarType().getSizeInBits(); 2106 unsigned Tmp, Tmp2; 2107 unsigned FirstAnswer = 1; 2108 2109 if (Depth == 6) 2110 return 1; // Limit search depth. 2111 2112 switch (Op.getOpcode()) { 2113 default: break; 2114 case ISD::AssertSext: 2115 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits(); 2116 return VTBits-Tmp+1; 2117 case ISD::AssertZext: 2118 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits(); 2119 return VTBits-Tmp; 2120 2121 case ISD::Constant: { 2122 const APInt &Val = cast<ConstantSDNode>(Op)->getAPIntValue(); 2123 return Val.getNumSignBits(); 2124 } 2125 2126 case ISD::SIGN_EXTEND: 2127 Tmp = VTBits-Op.getOperand(0).getValueType().getScalarType().getSizeInBits(); 2128 return ComputeNumSignBits(Op.getOperand(0), Depth+1) + Tmp; 2129 2130 case ISD::SIGN_EXTEND_INREG: 2131 // Max of the input and what this extends. 2132 Tmp = 2133 cast<VTSDNode>(Op.getOperand(1))->getVT().getScalarType().getSizeInBits(); 2134 Tmp = VTBits-Tmp+1; 2135 2136 Tmp2 = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2137 return std::max(Tmp, Tmp2); 2138 2139 case ISD::SRA: 2140 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2141 // SRA X, C -> adds C sign bits. 2142 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 2143 Tmp += C->getZExtValue(); 2144 if (Tmp > VTBits) Tmp = VTBits; 2145 } 2146 return Tmp; 2147 case ISD::SHL: 2148 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 2149 // shl destroys sign bits. 2150 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2151 if (C->getZExtValue() >= VTBits || // Bad shift. 2152 C->getZExtValue() >= Tmp) break; // Shifted all sign bits out. 2153 return Tmp - C->getZExtValue(); 2154 } 2155 break; 2156 case ISD::AND: 2157 case ISD::OR: 2158 case ISD::XOR: // NOT is handled here. 2159 // Logical binary ops preserve the number of sign bits at the worst. 2160 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2161 if (Tmp != 1) { 2162 Tmp2 = ComputeNumSignBits(Op.getOperand(1), Depth+1); 2163 FirstAnswer = std::min(Tmp, Tmp2); 2164 // We computed what we know about the sign bits as our first 2165 // answer. Now proceed to the generic code that uses 2166 // ComputeMaskedBits, and pick whichever answer is better. 2167 } 2168 break; 2169 2170 case ISD::SELECT: 2171 Tmp = ComputeNumSignBits(Op.getOperand(1), Depth+1); 2172 if (Tmp == 1) return 1; // Early out. 2173 Tmp2 = ComputeNumSignBits(Op.getOperand(2), Depth+1); 2174 return std::min(Tmp, Tmp2); 2175 2176 case ISD::SADDO: 2177 case ISD::UADDO: 2178 case ISD::SSUBO: 2179 case ISD::USUBO: 2180 case ISD::SMULO: 2181 case ISD::UMULO: 2182 if (Op.getResNo() != 1) 2183 break; 2184 // The boolean result conforms to getBooleanContents. Fall through. 2185 case ISD::SETCC: 2186 // If setcc returns 0/-1, all bits are sign bits. 2187 if (TLI.getBooleanContents(Op.getValueType().isVector()) == 2188 TargetLowering::ZeroOrNegativeOneBooleanContent) 2189 return VTBits; 2190 break; 2191 case ISD::ROTL: 2192 case ISD::ROTR: 2193 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Op.getOperand(1))) { 2194 unsigned RotAmt = C->getZExtValue() & (VTBits-1); 2195 2196 // Handle rotate right by N like a rotate left by 32-N. 2197 if (Op.getOpcode() == ISD::ROTR) 2198 RotAmt = (VTBits-RotAmt) & (VTBits-1); 2199 2200 // If we aren't rotating out all of the known-in sign bits, return the 2201 // number that are left. This handles rotl(sext(x), 1) for example. 2202 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2203 if (Tmp > RotAmt+1) return Tmp-RotAmt; 2204 } 2205 break; 2206 case ISD::ADD: 2207 // Add can have at most one carry bit. Thus we know that the output 2208 // is, at worst, one more bit than the inputs. 2209 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2210 if (Tmp == 1) return 1; // Early out. 2211 2212 // Special case decrementing a value (ADD X, -1): 2213 if (ConstantSDNode *CRHS = dyn_cast<ConstantSDNode>(Op.getOperand(1))) 2214 if (CRHS->isAllOnesValue()) { 2215 APInt KnownZero, KnownOne; 2216 APInt Mask = APInt::getAllOnesValue(VTBits); 2217 ComputeMaskedBits(Op.getOperand(0), Mask, KnownZero, KnownOne, Depth+1); 2218 2219 // If the input is known to be 0 or 1, the output is 0/-1, which is all 2220 // sign bits set. 2221 if ((KnownZero | APInt(VTBits, 1)) == Mask) 2222 return VTBits; 2223 2224 // If we are subtracting one from a positive number, there is no carry 2225 // out of the result. 2226 if (KnownZero.isNegative()) 2227 return Tmp; 2228 } 2229 2230 Tmp2 = ComputeNumSignBits(Op.getOperand(1), Depth+1); 2231 if (Tmp2 == 1) return 1; 2232 return std::min(Tmp, Tmp2)-1; 2233 break; 2234 2235 case ISD::SUB: 2236 Tmp2 = ComputeNumSignBits(Op.getOperand(1), Depth+1); 2237 if (Tmp2 == 1) return 1; 2238 2239 // Handle NEG. 2240 if (ConstantSDNode *CLHS = dyn_cast<ConstantSDNode>(Op.getOperand(0))) 2241 if (CLHS->isNullValue()) { 2242 APInt KnownZero, KnownOne; 2243 APInt Mask = APInt::getAllOnesValue(VTBits); 2244 ComputeMaskedBits(Op.getOperand(1), Mask, KnownZero, KnownOne, Depth+1); 2245 // If the input is known to be 0 or 1, the output is 0/-1, which is all 2246 // sign bits set. 2247 if ((KnownZero | APInt(VTBits, 1)) == Mask) 2248 return VTBits; 2249 2250 // If the input is known to be positive (the sign bit is known clear), 2251 // the output of the NEG has the same number of sign bits as the input. 2252 if (KnownZero.isNegative()) 2253 return Tmp2; 2254 2255 // Otherwise, we treat this like a SUB. 2256 } 2257 2258 // Sub can have at most one carry bit. Thus we know that the output 2259 // is, at worst, one more bit than the inputs. 2260 Tmp = ComputeNumSignBits(Op.getOperand(0), Depth+1); 2261 if (Tmp == 1) return 1; // Early out. 2262 return std::min(Tmp, Tmp2)-1; 2263 break; 2264 case ISD::TRUNCATE: 2265 // FIXME: it's tricky to do anything useful for this, but it is an important 2266 // case for targets like X86. 2267 break; 2268 } 2269 2270 // Handle LOADX separately here. EXTLOAD case will fallthrough. 2271 if (Op.getOpcode() == ISD::LOAD) { 2272 LoadSDNode *LD = cast<LoadSDNode>(Op); 2273 unsigned ExtType = LD->getExtensionType(); 2274 switch (ExtType) { 2275 default: break; 2276 case ISD::SEXTLOAD: // '17' bits known 2277 Tmp = LD->getMemoryVT().getScalarType().getSizeInBits(); 2278 return VTBits-Tmp+1; 2279 case ISD::ZEXTLOAD: // '16' bits known 2280 Tmp = LD->getMemoryVT().getScalarType().getSizeInBits(); 2281 return VTBits-Tmp; 2282 } 2283 } 2284 2285 // Allow the target to implement this method for its nodes. 2286 if (Op.getOpcode() >= ISD::BUILTIN_OP_END || 2287 Op.getOpcode() == ISD::INTRINSIC_WO_CHAIN || 2288 Op.getOpcode() == ISD::INTRINSIC_W_CHAIN || 2289 Op.getOpcode() == ISD::INTRINSIC_VOID) { 2290 unsigned NumBits = TLI.ComputeNumSignBitsForTargetNode(Op, Depth); 2291 if (NumBits > 1) FirstAnswer = std::max(FirstAnswer, NumBits); 2292 } 2293 2294 // Finally, if we can prove that the top bits of the result are 0's or 1's, 2295 // use this information. 2296 APInt KnownZero, KnownOne; 2297 APInt Mask = APInt::getAllOnesValue(VTBits); 2298 ComputeMaskedBits(Op, Mask, KnownZero, KnownOne, Depth); 2299 2300 if (KnownZero.isNegative()) { // sign bit is 0 2301 Mask = KnownZero; 2302 } else if (KnownOne.isNegative()) { // sign bit is 1; 2303 Mask = KnownOne; 2304 } else { 2305 // Nothing known. 2306 return FirstAnswer; 2307 } 2308 2309 // Okay, we know that the sign bit in Mask is set. Use CLZ to determine 2310 // the number of identical bits in the top of the input value. 2311 Mask = ~Mask; 2312 Mask <<= Mask.getBitWidth()-VTBits; 2313 // Return # leading zeros. We use 'min' here in case Val was zero before 2314 // shifting. We don't want to return '64' as for an i32 "0". 2315 return std::max(FirstAnswer, std::min(VTBits, Mask.countLeadingZeros())); 2316 } 2317 2318 /// isBaseWithConstantOffset - Return true if the specified operand is an 2319 /// ISD::ADD with a ConstantSDNode on the right-hand side, or if it is an 2320 /// ISD::OR with a ConstantSDNode that is guaranteed to have the same 2321 /// semantics as an ADD. This handles the equivalence: 2322 /// X|Cst == X+Cst iff X&Cst = 0. 2323 bool SelectionDAG::isBaseWithConstantOffset(SDValue Op) const { 2324 if ((Op.getOpcode() != ISD::ADD && Op.getOpcode() != ISD::OR) || 2325 !isa<ConstantSDNode>(Op.getOperand(1))) 2326 return false; 2327 2328 if (Op.getOpcode() == ISD::OR && 2329 !MaskedValueIsZero(Op.getOperand(0), 2330 cast<ConstantSDNode>(Op.getOperand(1))->getAPIntValue())) 2331 return false; 2332 2333 return true; 2334 } 2335 2336 2337 bool SelectionDAG::isKnownNeverNaN(SDValue Op) const { 2338 // If we're told that NaNs won't happen, assume they won't. 2339 if (getTarget().Options.NoNaNsFPMath) 2340 return true; 2341 2342 // If the value is a constant, we can obviously see if it is a NaN or not. 2343 if (const ConstantFPSDNode *C = dyn_cast<ConstantFPSDNode>(Op)) 2344 return !C->getValueAPF().isNaN(); 2345 2346 // TODO: Recognize more cases here. 2347 2348 return false; 2349 } 2350 2351 bool SelectionDAG::isKnownNeverZero(SDValue Op) const { 2352 // If the value is a constant, we can obviously see if it is a zero or not. 2353 if (const ConstantFPSDNode *C = dyn_cast<ConstantFPSDNode>(Op)) 2354 return !C->isZero(); 2355 2356 // TODO: Recognize more cases here. 2357 switch (Op.getOpcode()) { 2358 default: break; 2359 case ISD::OR: 2360 if (const ConstantSDNode *C = dyn_cast<ConstantSDNode>(Op.getOperand(1))) 2361 return !C->isNullValue(); 2362 break; 2363 } 2364 2365 return false; 2366 } 2367 2368 bool SelectionDAG::isEqualTo(SDValue A, SDValue B) const { 2369 // Check the obvious case. 2370 if (A == B) return true; 2371 2372 // For for negative and positive zero. 2373 if (const ConstantFPSDNode *CA = dyn_cast<ConstantFPSDNode>(A)) 2374 if (const ConstantFPSDNode *CB = dyn_cast<ConstantFPSDNode>(B)) 2375 if (CA->isZero() && CB->isZero()) return true; 2376 2377 // Otherwise they may not be equal. 2378 return false; 2379 } 2380 2381 /// getNode - Gets or creates the specified node. 2382 /// 2383 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT) { 2384 FoldingSetNodeID ID; 2385 AddNodeIDNode(ID, Opcode, getVTList(VT), 0, 0); 2386 void *IP = 0; 2387 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 2388 return SDValue(E, 0); 2389 2390 SDNode *N = new (NodeAllocator) SDNode(Opcode, DL, getVTList(VT)); 2391 CSEMap.InsertNode(N, IP); 2392 2393 AllNodes.push_back(N); 2394 #ifndef NDEBUG 2395 VerifySDNode(N); 2396 #endif 2397 return SDValue(N, 0); 2398 } 2399 2400 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, 2401 EVT VT, SDValue Operand) { 2402 // Constant fold unary operations with an integer constant operand. 2403 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Operand.getNode())) { 2404 const APInt &Val = C->getAPIntValue(); 2405 switch (Opcode) { 2406 default: break; 2407 case ISD::SIGN_EXTEND: 2408 return getConstant(Val.sextOrTrunc(VT.getSizeInBits()), VT); 2409 case ISD::ANY_EXTEND: 2410 case ISD::ZERO_EXTEND: 2411 case ISD::TRUNCATE: 2412 return getConstant(Val.zextOrTrunc(VT.getSizeInBits()), VT); 2413 case ISD::UINT_TO_FP: 2414 case ISD::SINT_TO_FP: { 2415 // No compile time operations on ppcf128. 2416 if (VT == MVT::ppcf128) break; 2417 APFloat apf(APInt::getNullValue(VT.getSizeInBits())); 2418 (void)apf.convertFromAPInt(Val, 2419 Opcode==ISD::SINT_TO_FP, 2420 APFloat::rmNearestTiesToEven); 2421 return getConstantFP(apf, VT); 2422 } 2423 case ISD::BITCAST: 2424 if (VT == MVT::f32 && C->getValueType(0) == MVT::i32) 2425 return getConstantFP(Val.bitsToFloat(), VT); 2426 else if (VT == MVT::f64 && C->getValueType(0) == MVT::i64) 2427 return getConstantFP(Val.bitsToDouble(), VT); 2428 break; 2429 case ISD::BSWAP: 2430 return getConstant(Val.byteSwap(), VT); 2431 case ISD::CTPOP: 2432 return getConstant(Val.countPopulation(), VT); 2433 case ISD::CTLZ: 2434 case ISD::CTLZ_ZERO_UNDEF: 2435 return getConstant(Val.countLeadingZeros(), VT); 2436 case ISD::CTTZ: 2437 case ISD::CTTZ_ZERO_UNDEF: 2438 return getConstant(Val.countTrailingZeros(), VT); 2439 } 2440 } 2441 2442 // Constant fold unary operations with a floating point constant operand. 2443 if (ConstantFPSDNode *C = dyn_cast<ConstantFPSDNode>(Operand.getNode())) { 2444 APFloat V = C->getValueAPF(); // make copy 2445 if (VT != MVT::ppcf128 && Operand.getValueType() != MVT::ppcf128) { 2446 switch (Opcode) { 2447 case ISD::FNEG: 2448 V.changeSign(); 2449 return getConstantFP(V, VT); 2450 case ISD::FABS: 2451 V.clearSign(); 2452 return getConstantFP(V, VT); 2453 case ISD::FP_ROUND: 2454 case ISD::FP_EXTEND: { 2455 bool ignored; 2456 // This can return overflow, underflow, or inexact; we don't care. 2457 // FIXME need to be more flexible about rounding mode. 2458 (void)V.convert(*EVTToAPFloatSemantics(VT), 2459 APFloat::rmNearestTiesToEven, &ignored); 2460 return getConstantFP(V, VT); 2461 } 2462 case ISD::FP_TO_SINT: 2463 case ISD::FP_TO_UINT: { 2464 integerPart x[2]; 2465 bool ignored; 2466 assert(integerPartWidth >= 64); 2467 // FIXME need to be more flexible about rounding mode. 2468 APFloat::opStatus s = V.convertToInteger(x, VT.getSizeInBits(), 2469 Opcode==ISD::FP_TO_SINT, 2470 APFloat::rmTowardZero, &ignored); 2471 if (s==APFloat::opInvalidOp) // inexact is OK, in fact usual 2472 break; 2473 APInt api(VT.getSizeInBits(), x); 2474 return getConstant(api, VT); 2475 } 2476 case ISD::BITCAST: 2477 if (VT == MVT::i32 && C->getValueType(0) == MVT::f32) 2478 return getConstant((uint32_t)V.bitcastToAPInt().getZExtValue(), VT); 2479 else if (VT == MVT::i64 && C->getValueType(0) == MVT::f64) 2480 return getConstant(V.bitcastToAPInt().getZExtValue(), VT); 2481 break; 2482 } 2483 } 2484 } 2485 2486 unsigned OpOpcode = Operand.getNode()->getOpcode(); 2487 switch (Opcode) { 2488 case ISD::TokenFactor: 2489 case ISD::MERGE_VALUES: 2490 case ISD::CONCAT_VECTORS: 2491 return Operand; // Factor, merge or concat of one node? No need. 2492 case ISD::FP_ROUND: llvm_unreachable("Invalid method to make FP_ROUND node"); 2493 case ISD::FP_EXTEND: 2494 assert(VT.isFloatingPoint() && 2495 Operand.getValueType().isFloatingPoint() && "Invalid FP cast!"); 2496 if (Operand.getValueType() == VT) return Operand; // noop conversion. 2497 assert((!VT.isVector() || 2498 VT.getVectorNumElements() == 2499 Operand.getValueType().getVectorNumElements()) && 2500 "Vector element count mismatch!"); 2501 if (Operand.getOpcode() == ISD::UNDEF) 2502 return getUNDEF(VT); 2503 break; 2504 case ISD::SIGN_EXTEND: 2505 assert(VT.isInteger() && Operand.getValueType().isInteger() && 2506 "Invalid SIGN_EXTEND!"); 2507 if (Operand.getValueType() == VT) return Operand; // noop extension 2508 assert(Operand.getValueType().getScalarType().bitsLT(VT.getScalarType()) && 2509 "Invalid sext node, dst < src!"); 2510 assert((!VT.isVector() || 2511 VT.getVectorNumElements() == 2512 Operand.getValueType().getVectorNumElements()) && 2513 "Vector element count mismatch!"); 2514 if (OpOpcode == ISD::SIGN_EXTEND || OpOpcode == ISD::ZERO_EXTEND) 2515 return getNode(OpOpcode, DL, VT, Operand.getNode()->getOperand(0)); 2516 else if (OpOpcode == ISD::UNDEF) 2517 // sext(undef) = 0, because the top bits will all be the same. 2518 return getConstant(0, VT); 2519 break; 2520 case ISD::ZERO_EXTEND: 2521 assert(VT.isInteger() && Operand.getValueType().isInteger() && 2522 "Invalid ZERO_EXTEND!"); 2523 if (Operand.getValueType() == VT) return Operand; // noop extension 2524 assert(Operand.getValueType().getScalarType().bitsLT(VT.getScalarType()) && 2525 "Invalid zext node, dst < src!"); 2526 assert((!VT.isVector() || 2527 VT.getVectorNumElements() == 2528 Operand.getValueType().getVectorNumElements()) && 2529 "Vector element count mismatch!"); 2530 if (OpOpcode == ISD::ZERO_EXTEND) // (zext (zext x)) -> (zext x) 2531 return getNode(ISD::ZERO_EXTEND, DL, VT, 2532 Operand.getNode()->getOperand(0)); 2533 else if (OpOpcode == ISD::UNDEF) 2534 // zext(undef) = 0, because the top bits will be zero. 2535 return getConstant(0, VT); 2536 break; 2537 case ISD::ANY_EXTEND: 2538 assert(VT.isInteger() && Operand.getValueType().isInteger() && 2539 "Invalid ANY_EXTEND!"); 2540 if (Operand.getValueType() == VT) return Operand; // noop extension 2541 assert(Operand.getValueType().getScalarType().bitsLT(VT.getScalarType()) && 2542 "Invalid anyext node, dst < src!"); 2543 assert((!VT.isVector() || 2544 VT.getVectorNumElements() == 2545 Operand.getValueType().getVectorNumElements()) && 2546 "Vector element count mismatch!"); 2547 2548 if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND || 2549 OpOpcode == ISD::ANY_EXTEND) 2550 // (ext (zext x)) -> (zext x) and (ext (sext x)) -> (sext x) 2551 return getNode(OpOpcode, DL, VT, Operand.getNode()->getOperand(0)); 2552 else if (OpOpcode == ISD::UNDEF) 2553 return getUNDEF(VT); 2554 2555 // (ext (trunx x)) -> x 2556 if (OpOpcode == ISD::TRUNCATE) { 2557 SDValue OpOp = Operand.getNode()->getOperand(0); 2558 if (OpOp.getValueType() == VT) 2559 return OpOp; 2560 } 2561 break; 2562 case ISD::TRUNCATE: 2563 assert(VT.isInteger() && Operand.getValueType().isInteger() && 2564 "Invalid TRUNCATE!"); 2565 if (Operand.getValueType() == VT) return Operand; // noop truncate 2566 assert(Operand.getValueType().getScalarType().bitsGT(VT.getScalarType()) && 2567 "Invalid truncate node, src < dst!"); 2568 assert((!VT.isVector() || 2569 VT.getVectorNumElements() == 2570 Operand.getValueType().getVectorNumElements()) && 2571 "Vector element count mismatch!"); 2572 if (OpOpcode == ISD::TRUNCATE) 2573 return getNode(ISD::TRUNCATE, DL, VT, Operand.getNode()->getOperand(0)); 2574 else if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND || 2575 OpOpcode == ISD::ANY_EXTEND) { 2576 // If the source is smaller than the dest, we still need an extend. 2577 if (Operand.getNode()->getOperand(0).getValueType().getScalarType() 2578 .bitsLT(VT.getScalarType())) 2579 return getNode(OpOpcode, DL, VT, Operand.getNode()->getOperand(0)); 2580 else if (Operand.getNode()->getOperand(0).getValueType().bitsGT(VT)) 2581 return getNode(ISD::TRUNCATE, DL, VT, Operand.getNode()->getOperand(0)); 2582 else 2583 return Operand.getNode()->getOperand(0); 2584 } 2585 break; 2586 case ISD::BITCAST: 2587 // Basic sanity checking. 2588 assert(VT.getSizeInBits() == Operand.getValueType().getSizeInBits() 2589 && "Cannot BITCAST between types of different sizes!"); 2590 if (VT == Operand.getValueType()) return Operand; // noop conversion. 2591 if (OpOpcode == ISD::BITCAST) // bitconv(bitconv(x)) -> bitconv(x) 2592 return getNode(ISD::BITCAST, DL, VT, Operand.getOperand(0)); 2593 if (OpOpcode == ISD::UNDEF) 2594 return getUNDEF(VT); 2595 break; 2596 case ISD::SCALAR_TO_VECTOR: 2597 assert(VT.isVector() && !Operand.getValueType().isVector() && 2598 (VT.getVectorElementType() == Operand.getValueType() || 2599 (VT.getVectorElementType().isInteger() && 2600 Operand.getValueType().isInteger() && 2601 VT.getVectorElementType().bitsLE(Operand.getValueType()))) && 2602 "Illegal SCALAR_TO_VECTOR node!"); 2603 if (OpOpcode == ISD::UNDEF) 2604 return getUNDEF(VT); 2605 // scalar_to_vector(extract_vector_elt V, 0) -> V, top bits are undefined. 2606 if (OpOpcode == ISD::EXTRACT_VECTOR_ELT && 2607 isa<ConstantSDNode>(Operand.getOperand(1)) && 2608 Operand.getConstantOperandVal(1) == 0 && 2609 Operand.getOperand(0).getValueType() == VT) 2610 return Operand.getOperand(0); 2611 break; 2612 case ISD::FNEG: 2613 // -(X-Y) -> (Y-X) is unsafe because when X==Y, -0.0 != +0.0 2614 if (getTarget().Options.UnsafeFPMath && OpOpcode == ISD::FSUB) 2615 return getNode(ISD::FSUB, DL, VT, Operand.getNode()->getOperand(1), 2616 Operand.getNode()->getOperand(0)); 2617 if (OpOpcode == ISD::FNEG) // --X -> X 2618 return Operand.getNode()->getOperand(0); 2619 break; 2620 case ISD::FABS: 2621 if (OpOpcode == ISD::FNEG) // abs(-X) -> abs(X) 2622 return getNode(ISD::FABS, DL, VT, Operand.getNode()->getOperand(0)); 2623 break; 2624 } 2625 2626 SDNode *N; 2627 SDVTList VTs = getVTList(VT); 2628 if (VT != MVT::Glue) { // Don't CSE flag producing nodes 2629 FoldingSetNodeID ID; 2630 SDValue Ops[1] = { Operand }; 2631 AddNodeIDNode(ID, Opcode, VTs, Ops, 1); 2632 void *IP = 0; 2633 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 2634 return SDValue(E, 0); 2635 2636 N = new (NodeAllocator) UnarySDNode(Opcode, DL, VTs, Operand); 2637 CSEMap.InsertNode(N, IP); 2638 } else { 2639 N = new (NodeAllocator) UnarySDNode(Opcode, DL, VTs, Operand); 2640 } 2641 2642 AllNodes.push_back(N); 2643 #ifndef NDEBUG 2644 VerifySDNode(N); 2645 #endif 2646 return SDValue(N, 0); 2647 } 2648 2649 SDValue SelectionDAG::FoldConstantArithmetic(unsigned Opcode, 2650 EVT VT, 2651 ConstantSDNode *Cst1, 2652 ConstantSDNode *Cst2) { 2653 const APInt &C1 = Cst1->getAPIntValue(), &C2 = Cst2->getAPIntValue(); 2654 2655 switch (Opcode) { 2656 case ISD::ADD: return getConstant(C1 + C2, VT); 2657 case ISD::SUB: return getConstant(C1 - C2, VT); 2658 case ISD::MUL: return getConstant(C1 * C2, VT); 2659 case ISD::UDIV: 2660 if (C2.getBoolValue()) return getConstant(C1.udiv(C2), VT); 2661 break; 2662 case ISD::UREM: 2663 if (C2.getBoolValue()) return getConstant(C1.urem(C2), VT); 2664 break; 2665 case ISD::SDIV: 2666 if (C2.getBoolValue()) return getConstant(C1.sdiv(C2), VT); 2667 break; 2668 case ISD::SREM: 2669 if (C2.getBoolValue()) return getConstant(C1.srem(C2), VT); 2670 break; 2671 case ISD::AND: return getConstant(C1 & C2, VT); 2672 case ISD::OR: return getConstant(C1 | C2, VT); 2673 case ISD::XOR: return getConstant(C1 ^ C2, VT); 2674 case ISD::SHL: return getConstant(C1 << C2, VT); 2675 case ISD::SRL: return getConstant(C1.lshr(C2), VT); 2676 case ISD::SRA: return getConstant(C1.ashr(C2), VT); 2677 case ISD::ROTL: return getConstant(C1.rotl(C2), VT); 2678 case ISD::ROTR: return getConstant(C1.rotr(C2), VT); 2679 default: break; 2680 } 2681 2682 return SDValue(); 2683 } 2684 2685 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 2686 SDValue N1, SDValue N2) { 2687 ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.getNode()); 2688 ConstantSDNode *N2C = dyn_cast<ConstantSDNode>(N2.getNode()); 2689 switch (Opcode) { 2690 default: break; 2691 case ISD::TokenFactor: 2692 assert(VT == MVT::Other && N1.getValueType() == MVT::Other && 2693 N2.getValueType() == MVT::Other && "Invalid token factor!"); 2694 // Fold trivial token factors. 2695 if (N1.getOpcode() == ISD::EntryToken) return N2; 2696 if (N2.getOpcode() == ISD::EntryToken) return N1; 2697 if (N1 == N2) return N1; 2698 break; 2699 case ISD::CONCAT_VECTORS: 2700 // A CONCAT_VECTOR with all operands BUILD_VECTOR can be simplified to 2701 // one big BUILD_VECTOR. 2702 if (N1.getOpcode() == ISD::BUILD_VECTOR && 2703 N2.getOpcode() == ISD::BUILD_VECTOR) { 2704 SmallVector<SDValue, 16> Elts(N1.getNode()->op_begin(), 2705 N1.getNode()->op_end()); 2706 Elts.append(N2.getNode()->op_begin(), N2.getNode()->op_end()); 2707 return getNode(ISD::BUILD_VECTOR, DL, VT, &Elts[0], Elts.size()); 2708 } 2709 break; 2710 case ISD::AND: 2711 assert(VT.isInteger() && "This operator does not apply to FP types!"); 2712 assert(N1.getValueType() == N2.getValueType() && 2713 N1.getValueType() == VT && "Binary operator types must match!"); 2714 // (X & 0) -> 0. This commonly occurs when legalizing i64 values, so it's 2715 // worth handling here. 2716 if (N2C && N2C->isNullValue()) 2717 return N2; 2718 if (N2C && N2C->isAllOnesValue()) // X & -1 -> X 2719 return N1; 2720 break; 2721 case ISD::OR: 2722 case ISD::XOR: 2723 case ISD::ADD: 2724 case ISD::SUB: 2725 assert(VT.isInteger() && "This operator does not apply to FP types!"); 2726 assert(N1.getValueType() == N2.getValueType() && 2727 N1.getValueType() == VT && "Binary operator types must match!"); 2728 // (X ^|+- 0) -> X. This commonly occurs when legalizing i64 values, so 2729 // it's worth handling here. 2730 if (N2C && N2C->isNullValue()) 2731 return N1; 2732 break; 2733 case ISD::UDIV: 2734 case ISD::UREM: 2735 case ISD::MULHU: 2736 case ISD::MULHS: 2737 case ISD::MUL: 2738 case ISD::SDIV: 2739 case ISD::SREM: 2740 assert(VT.isInteger() && "This operator does not apply to FP types!"); 2741 assert(N1.getValueType() == N2.getValueType() && 2742 N1.getValueType() == VT && "Binary operator types must match!"); 2743 break; 2744 case ISD::FADD: 2745 case ISD::FSUB: 2746 case ISD::FMUL: 2747 case ISD::FDIV: 2748 case ISD::FREM: 2749 if (getTarget().Options.UnsafeFPMath) { 2750 if (Opcode == ISD::FADD) { 2751 // 0+x --> x 2752 if (ConstantFPSDNode *CFP = dyn_cast<ConstantFPSDNode>(N1)) 2753 if (CFP->getValueAPF().isZero()) 2754 return N2; 2755 // x+0 --> x 2756 if (ConstantFPSDNode *CFP = dyn_cast<ConstantFPSDNode>(N2)) 2757 if (CFP->getValueAPF().isZero()) 2758 return N1; 2759 } else if (Opcode == ISD::FSUB) { 2760 // x-0 --> x 2761 if (ConstantFPSDNode *CFP = dyn_cast<ConstantFPSDNode>(N2)) 2762 if (CFP->getValueAPF().isZero()) 2763 return N1; 2764 } 2765 } 2766 assert(VT.isFloatingPoint() && "This operator only applies to FP types!"); 2767 assert(N1.getValueType() == N2.getValueType() && 2768 N1.getValueType() == VT && "Binary operator types must match!"); 2769 break; 2770 case ISD::FCOPYSIGN: // N1 and result must match. N1/N2 need not match. 2771 assert(N1.getValueType() == VT && 2772 N1.getValueType().isFloatingPoint() && 2773 N2.getValueType().isFloatingPoint() && 2774 "Invalid FCOPYSIGN!"); 2775 break; 2776 case ISD::SHL: 2777 case ISD::SRA: 2778 case ISD::SRL: 2779 case ISD::ROTL: 2780 case ISD::ROTR: 2781 assert(VT == N1.getValueType() && 2782 "Shift operators return type must be the same as their first arg"); 2783 assert(VT.isInteger() && N2.getValueType().isInteger() && 2784 "Shifts only work on integers"); 2785 // Verify that the shift amount VT is bit enough to hold valid shift 2786 // amounts. This catches things like trying to shift an i1024 value by an 2787 // i8, which is easy to fall into in generic code that uses 2788 // TLI.getShiftAmount(). 2789 assert(N2.getValueType().getSizeInBits() >= 2790 Log2_32_Ceil(N1.getValueType().getSizeInBits()) && 2791 "Invalid use of small shift amount with oversized value!"); 2792 2793 // Always fold shifts of i1 values so the code generator doesn't need to 2794 // handle them. Since we know the size of the shift has to be less than the 2795 // size of the value, the shift/rotate count is guaranteed to be zero. 2796 if (VT == MVT::i1) 2797 return N1; 2798 if (N2C && N2C->isNullValue()) 2799 return N1; 2800 break; 2801 case ISD::FP_ROUND_INREG: { 2802 EVT EVT = cast<VTSDNode>(N2)->getVT(); 2803 assert(VT == N1.getValueType() && "Not an inreg round!"); 2804 assert(VT.isFloatingPoint() && EVT.isFloatingPoint() && 2805 "Cannot FP_ROUND_INREG integer types"); 2806 assert(EVT.isVector() == VT.isVector() && 2807 "FP_ROUND_INREG type should be vector iff the operand " 2808 "type is vector!"); 2809 assert((!EVT.isVector() || 2810 EVT.getVectorNumElements() == VT.getVectorNumElements()) && 2811 "Vector element counts must match in FP_ROUND_INREG"); 2812 assert(EVT.bitsLE(VT) && "Not rounding down!"); 2813 (void)EVT; 2814 if (cast<VTSDNode>(N2)->getVT() == VT) return N1; // Not actually rounding. 2815 break; 2816 } 2817 case ISD::FP_ROUND: 2818 assert(VT.isFloatingPoint() && 2819 N1.getValueType().isFloatingPoint() && 2820 VT.bitsLE(N1.getValueType()) && 2821 isa<ConstantSDNode>(N2) && "Invalid FP_ROUND!"); 2822 if (N1.getValueType() == VT) return N1; // noop conversion. 2823 break; 2824 case ISD::AssertSext: 2825 case ISD::AssertZext: { 2826 EVT EVT = cast<VTSDNode>(N2)->getVT(); 2827 assert(VT == N1.getValueType() && "Not an inreg extend!"); 2828 assert(VT.isInteger() && EVT.isInteger() && 2829 "Cannot *_EXTEND_INREG FP types"); 2830 assert(!EVT.isVector() && 2831 "AssertSExt/AssertZExt type should be the vector element type " 2832 "rather than the vector type!"); 2833 assert(EVT.bitsLE(VT) && "Not extending!"); 2834 if (VT == EVT) return N1; // noop assertion. 2835 break; 2836 } 2837 case ISD::SIGN_EXTEND_INREG: { 2838 EVT EVT = cast<VTSDNode>(N2)->getVT(); 2839 assert(VT == N1.getValueType() && "Not an inreg extend!"); 2840 assert(VT.isInteger() && EVT.isInteger() && 2841 "Cannot *_EXTEND_INREG FP types"); 2842 assert(EVT.isVector() == VT.isVector() && 2843 "SIGN_EXTEND_INREG type should be vector iff the operand " 2844 "type is vector!"); 2845 assert((!EVT.isVector() || 2846 EVT.getVectorNumElements() == VT.getVectorNumElements()) && 2847 "Vector element counts must match in SIGN_EXTEND_INREG"); 2848 assert(EVT.bitsLE(VT) && "Not extending!"); 2849 if (EVT == VT) return N1; // Not actually extending 2850 2851 if (N1C) { 2852 APInt Val = N1C->getAPIntValue(); 2853 unsigned FromBits = EVT.getScalarType().getSizeInBits(); 2854 Val <<= Val.getBitWidth()-FromBits; 2855 Val = Val.ashr(Val.getBitWidth()-FromBits); 2856 return getConstant(Val, VT); 2857 } 2858 break; 2859 } 2860 case ISD::EXTRACT_VECTOR_ELT: 2861 // EXTRACT_VECTOR_ELT of an UNDEF is an UNDEF. 2862 if (N1.getOpcode() == ISD::UNDEF) 2863 return getUNDEF(VT); 2864 2865 // EXTRACT_VECTOR_ELT of CONCAT_VECTORS is often formed while lowering is 2866 // expanding copies of large vectors from registers. 2867 if (N2C && 2868 N1.getOpcode() == ISD::CONCAT_VECTORS && 2869 N1.getNumOperands() > 0) { 2870 unsigned Factor = 2871 N1.getOperand(0).getValueType().getVectorNumElements(); 2872 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, 2873 N1.getOperand(N2C->getZExtValue() / Factor), 2874 getConstant(N2C->getZExtValue() % Factor, 2875 N2.getValueType())); 2876 } 2877 2878 // EXTRACT_VECTOR_ELT of BUILD_VECTOR is often formed while lowering is 2879 // expanding large vector constants. 2880 if (N2C && N1.getOpcode() == ISD::BUILD_VECTOR) { 2881 SDValue Elt = N1.getOperand(N2C->getZExtValue()); 2882 EVT VEltTy = N1.getValueType().getVectorElementType(); 2883 if (Elt.getValueType() != VEltTy) { 2884 // If the vector element type is not legal, the BUILD_VECTOR operands 2885 // are promoted and implicitly truncated. Make that explicit here. 2886 Elt = getNode(ISD::TRUNCATE, DL, VEltTy, Elt); 2887 } 2888 if (VT != VEltTy) { 2889 // If the vector element type is not legal, the EXTRACT_VECTOR_ELT 2890 // result is implicitly extended. 2891 Elt = getNode(ISD::ANY_EXTEND, DL, VT, Elt); 2892 } 2893 return Elt; 2894 } 2895 2896 // EXTRACT_VECTOR_ELT of INSERT_VECTOR_ELT is often formed when vector 2897 // operations are lowered to scalars. 2898 if (N1.getOpcode() == ISD::INSERT_VECTOR_ELT) { 2899 // If the indices are the same, return the inserted element else 2900 // if the indices are known different, extract the element from 2901 // the original vector. 2902 SDValue N1Op2 = N1.getOperand(2); 2903 ConstantSDNode *N1Op2C = dyn_cast<ConstantSDNode>(N1Op2.getNode()); 2904 2905 if (N1Op2C && N2C) { 2906 if (N1Op2C->getZExtValue() == N2C->getZExtValue()) { 2907 if (VT == N1.getOperand(1).getValueType()) 2908 return N1.getOperand(1); 2909 else 2910 return getSExtOrTrunc(N1.getOperand(1), DL, VT); 2911 } 2912 2913 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, N1.getOperand(0), N2); 2914 } 2915 } 2916 break; 2917 case ISD::EXTRACT_ELEMENT: 2918 assert(N2C && (unsigned)N2C->getZExtValue() < 2 && "Bad EXTRACT_ELEMENT!"); 2919 assert(!N1.getValueType().isVector() && !VT.isVector() && 2920 (N1.getValueType().isInteger() == VT.isInteger()) && 2921 N1.getValueType() != VT && 2922 "Wrong types for EXTRACT_ELEMENT!"); 2923 2924 // EXTRACT_ELEMENT of BUILD_PAIR is often formed while legalize is expanding 2925 // 64-bit integers into 32-bit parts. Instead of building the extract of 2926 // the BUILD_PAIR, only to have legalize rip it apart, just do it now. 2927 if (N1.getOpcode() == ISD::BUILD_PAIR) 2928 return N1.getOperand(N2C->getZExtValue()); 2929 2930 // EXTRACT_ELEMENT of a constant int is also very common. 2931 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(N1)) { 2932 unsigned ElementSize = VT.getSizeInBits(); 2933 unsigned Shift = ElementSize * N2C->getZExtValue(); 2934 APInt ShiftedVal = C->getAPIntValue().lshr(Shift); 2935 return getConstant(ShiftedVal.trunc(ElementSize), VT); 2936 } 2937 break; 2938 case ISD::EXTRACT_SUBVECTOR: { 2939 SDValue Index = N2; 2940 if (VT.isSimple() && N1.getValueType().isSimple()) { 2941 assert(VT.isVector() && N1.getValueType().isVector() && 2942 "Extract subvector VTs must be a vectors!"); 2943 assert(VT.getVectorElementType() == N1.getValueType().getVectorElementType() && 2944 "Extract subvector VTs must have the same element type!"); 2945 assert(VT.getSimpleVT() <= N1.getValueType().getSimpleVT() && 2946 "Extract subvector must be from larger vector to smaller vector!"); 2947 2948 if (isa<ConstantSDNode>(Index.getNode())) { 2949 assert((VT.getVectorNumElements() + 2950 cast<ConstantSDNode>(Index.getNode())->getZExtValue() 2951 <= N1.getValueType().getVectorNumElements()) 2952 && "Extract subvector overflow!"); 2953 } 2954 2955 // Trivial extraction. 2956 if (VT.getSimpleVT() == N1.getValueType().getSimpleVT()) 2957 return N1; 2958 } 2959 break; 2960 } 2961 } 2962 2963 if (N1C) { 2964 if (N2C) { 2965 SDValue SV = FoldConstantArithmetic(Opcode, VT, N1C, N2C); 2966 if (SV.getNode()) return SV; 2967 } else { // Cannonicalize constant to RHS if commutative 2968 if (isCommutativeBinOp(Opcode)) { 2969 std::swap(N1C, N2C); 2970 std::swap(N1, N2); 2971 } 2972 } 2973 } 2974 2975 // Constant fold FP operations. 2976 ConstantFPSDNode *N1CFP = dyn_cast<ConstantFPSDNode>(N1.getNode()); 2977 ConstantFPSDNode *N2CFP = dyn_cast<ConstantFPSDNode>(N2.getNode()); 2978 if (N1CFP) { 2979 if (!N2CFP && isCommutativeBinOp(Opcode)) { 2980 // Cannonicalize constant to RHS if commutative 2981 std::swap(N1CFP, N2CFP); 2982 std::swap(N1, N2); 2983 } else if (N2CFP && VT != MVT::ppcf128) { 2984 APFloat V1 = N1CFP->getValueAPF(), V2 = N2CFP->getValueAPF(); 2985 APFloat::opStatus s; 2986 switch (Opcode) { 2987 case ISD::FADD: 2988 s = V1.add(V2, APFloat::rmNearestTiesToEven); 2989 if (s != APFloat::opInvalidOp) 2990 return getConstantFP(V1, VT); 2991 break; 2992 case ISD::FSUB: 2993 s = V1.subtract(V2, APFloat::rmNearestTiesToEven); 2994 if (s!=APFloat::opInvalidOp) 2995 return getConstantFP(V1, VT); 2996 break; 2997 case ISD::FMUL: 2998 s = V1.multiply(V2, APFloat::rmNearestTiesToEven); 2999 if (s!=APFloat::opInvalidOp) 3000 return getConstantFP(V1, VT); 3001 break; 3002 case ISD::FDIV: 3003 s = V1.divide(V2, APFloat::rmNearestTiesToEven); 3004 if (s!=APFloat::opInvalidOp && s!=APFloat::opDivByZero) 3005 return getConstantFP(V1, VT); 3006 break; 3007 case ISD::FREM : 3008 s = V1.mod(V2, APFloat::rmNearestTiesToEven); 3009 if (s!=APFloat::opInvalidOp && s!=APFloat::opDivByZero) 3010 return getConstantFP(V1, VT); 3011 break; 3012 case ISD::FCOPYSIGN: 3013 V1.copySign(V2); 3014 return getConstantFP(V1, VT); 3015 default: break; 3016 } 3017 } 3018 } 3019 3020 // Canonicalize an UNDEF to the RHS, even over a constant. 3021 if (N1.getOpcode() == ISD::UNDEF) { 3022 if (isCommutativeBinOp(Opcode)) { 3023 std::swap(N1, N2); 3024 } else { 3025 switch (Opcode) { 3026 case ISD::FP_ROUND_INREG: 3027 case ISD::SIGN_EXTEND_INREG: 3028 case ISD::SUB: 3029 case ISD::FSUB: 3030 case ISD::FDIV: 3031 case ISD::FREM: 3032 case ISD::SRA: 3033 return N1; // fold op(undef, arg2) -> undef 3034 case ISD::UDIV: 3035 case ISD::SDIV: 3036 case ISD::UREM: 3037 case ISD::SREM: 3038 case ISD::SRL: 3039 case ISD::SHL: 3040 if (!VT.isVector()) 3041 return getConstant(0, VT); // fold op(undef, arg2) -> 0 3042 // For vectors, we can't easily build an all zero vector, just return 3043 // the LHS. 3044 return N2; 3045 } 3046 } 3047 } 3048 3049 // Fold a bunch of operators when the RHS is undef. 3050 if (N2.getOpcode() == ISD::UNDEF) { 3051 switch (Opcode) { 3052 case ISD::XOR: 3053 if (N1.getOpcode() == ISD::UNDEF) 3054 // Handle undef ^ undef -> 0 special case. This is a common 3055 // idiom (misuse). 3056 return getConstant(0, VT); 3057 // fallthrough 3058 case ISD::ADD: 3059 case ISD::ADDC: 3060 case ISD::ADDE: 3061 case ISD::SUB: 3062 case ISD::UDIV: 3063 case ISD::SDIV: 3064 case ISD::UREM: 3065 case ISD::SREM: 3066 return N2; // fold op(arg1, undef) -> undef 3067 case ISD::FADD: 3068 case ISD::FSUB: 3069 case ISD::FMUL: 3070 case ISD::FDIV: 3071 case ISD::FREM: 3072 if (getTarget().Options.UnsafeFPMath) 3073 return N2; 3074 break; 3075 case ISD::MUL: 3076 case ISD::AND: 3077 case ISD::SRL: 3078 case ISD::SHL: 3079 if (!VT.isVector()) 3080 return getConstant(0, VT); // fold op(arg1, undef) -> 0 3081 // For vectors, we can't easily build an all zero vector, just return 3082 // the LHS. 3083 return N1; 3084 case ISD::OR: 3085 if (!VT.isVector()) 3086 return getConstant(APInt::getAllOnesValue(VT.getSizeInBits()), VT); 3087 // For vectors, we can't easily build an all one vector, just return 3088 // the LHS. 3089 return N1; 3090 case ISD::SRA: 3091 return N1; 3092 } 3093 } 3094 3095 // Memoize this node if possible. 3096 SDNode *N; 3097 SDVTList VTs = getVTList(VT); 3098 if (VT != MVT::Glue) { 3099 SDValue Ops[] = { N1, N2 }; 3100 FoldingSetNodeID ID; 3101 AddNodeIDNode(ID, Opcode, VTs, Ops, 2); 3102 void *IP = 0; 3103 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 3104 return SDValue(E, 0); 3105 3106 N = new (NodeAllocator) BinarySDNode(Opcode, DL, VTs, N1, N2); 3107 CSEMap.InsertNode(N, IP); 3108 } else { 3109 N = new (NodeAllocator) BinarySDNode(Opcode, DL, VTs, N1, N2); 3110 } 3111 3112 AllNodes.push_back(N); 3113 #ifndef NDEBUG 3114 VerifySDNode(N); 3115 #endif 3116 return SDValue(N, 0); 3117 } 3118 3119 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 3120 SDValue N1, SDValue N2, SDValue N3) { 3121 // Perform various simplifications. 3122 ConstantSDNode *N1C = dyn_cast<ConstantSDNode>(N1.getNode()); 3123 switch (Opcode) { 3124 case ISD::CONCAT_VECTORS: 3125 // A CONCAT_VECTOR with all operands BUILD_VECTOR can be simplified to 3126 // one big BUILD_VECTOR. 3127 if (N1.getOpcode() == ISD::BUILD_VECTOR && 3128 N2.getOpcode() == ISD::BUILD_VECTOR && 3129 N3.getOpcode() == ISD::BUILD_VECTOR) { 3130 SmallVector<SDValue, 16> Elts(N1.getNode()->op_begin(), 3131 N1.getNode()->op_end()); 3132 Elts.append(N2.getNode()->op_begin(), N2.getNode()->op_end()); 3133 Elts.append(N3.getNode()->op_begin(), N3.getNode()->op_end()); 3134 return getNode(ISD::BUILD_VECTOR, DL, VT, &Elts[0], Elts.size()); 3135 } 3136 break; 3137 case ISD::SETCC: { 3138 // Use FoldSetCC to simplify SETCC's. 3139 SDValue Simp = FoldSetCC(VT, N1, N2, cast<CondCodeSDNode>(N3)->get(), DL); 3140 if (Simp.getNode()) return Simp; 3141 break; 3142 } 3143 case ISD::SELECT: 3144 if (N1C) { 3145 if (N1C->getZExtValue()) 3146 return N2; // select true, X, Y -> X 3147 else 3148 return N3; // select false, X, Y -> Y 3149 } 3150 3151 if (N2 == N3) return N2; // select C, X, X -> X 3152 break; 3153 case ISD::VECTOR_SHUFFLE: 3154 llvm_unreachable("should use getVectorShuffle constructor!"); 3155 break; 3156 case ISD::INSERT_SUBVECTOR: { 3157 SDValue Index = N3; 3158 if (VT.isSimple() && N1.getValueType().isSimple() 3159 && N2.getValueType().isSimple()) { 3160 assert(VT.isVector() && N1.getValueType().isVector() && 3161 N2.getValueType().isVector() && 3162 "Insert subvector VTs must be a vectors"); 3163 assert(VT == N1.getValueType() && 3164 "Dest and insert subvector source types must match!"); 3165 assert(N2.getValueType().getSimpleVT() <= N1.getValueType().getSimpleVT() && 3166 "Insert subvector must be from smaller vector to larger vector!"); 3167 if (isa<ConstantSDNode>(Index.getNode())) { 3168 assert((N2.getValueType().getVectorNumElements() + 3169 cast<ConstantSDNode>(Index.getNode())->getZExtValue() 3170 <= VT.getVectorNumElements()) 3171 && "Insert subvector overflow!"); 3172 } 3173 3174 // Trivial insertion. 3175 if (VT.getSimpleVT() == N2.getValueType().getSimpleVT()) 3176 return N2; 3177 } 3178 break; 3179 } 3180 case ISD::BITCAST: 3181 // Fold bit_convert nodes from a type to themselves. 3182 if (N1.getValueType() == VT) 3183 return N1; 3184 break; 3185 } 3186 3187 // Memoize node if it doesn't produce a flag. 3188 SDNode *N; 3189 SDVTList VTs = getVTList(VT); 3190 if (VT != MVT::Glue) { 3191 SDValue Ops[] = { N1, N2, N3 }; 3192 FoldingSetNodeID ID; 3193 AddNodeIDNode(ID, Opcode, VTs, Ops, 3); 3194 void *IP = 0; 3195 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 3196 return SDValue(E, 0); 3197 3198 N = new (NodeAllocator) TernarySDNode(Opcode, DL, VTs, N1, N2, N3); 3199 CSEMap.InsertNode(N, IP); 3200 } else { 3201 N = new (NodeAllocator) TernarySDNode(Opcode, DL, VTs, N1, N2, N3); 3202 } 3203 3204 AllNodes.push_back(N); 3205 #ifndef NDEBUG 3206 VerifySDNode(N); 3207 #endif 3208 return SDValue(N, 0); 3209 } 3210 3211 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 3212 SDValue N1, SDValue N2, SDValue N3, 3213 SDValue N4) { 3214 SDValue Ops[] = { N1, N2, N3, N4 }; 3215 return getNode(Opcode, DL, VT, Ops, 4); 3216 } 3217 3218 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 3219 SDValue N1, SDValue N2, SDValue N3, 3220 SDValue N4, SDValue N5) { 3221 SDValue Ops[] = { N1, N2, N3, N4, N5 }; 3222 return getNode(Opcode, DL, VT, Ops, 5); 3223 } 3224 3225 /// getStackArgumentTokenFactor - Compute a TokenFactor to force all 3226 /// the incoming stack arguments to be loaded from the stack. 3227 SDValue SelectionDAG::getStackArgumentTokenFactor(SDValue Chain) { 3228 SmallVector<SDValue, 8> ArgChains; 3229 3230 // Include the original chain at the beginning of the list. When this is 3231 // used by target LowerCall hooks, this helps legalize find the 3232 // CALLSEQ_BEGIN node. 3233 ArgChains.push_back(Chain); 3234 3235 // Add a chain value for each stack argument. 3236 for (SDNode::use_iterator U = getEntryNode().getNode()->use_begin(), 3237 UE = getEntryNode().getNode()->use_end(); U != UE; ++U) 3238 if (LoadSDNode *L = dyn_cast<LoadSDNode>(*U)) 3239 if (FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(L->getBasePtr())) 3240 if (FI->getIndex() < 0) 3241 ArgChains.push_back(SDValue(L, 1)); 3242 3243 // Build a tokenfactor for all the chains. 3244 return getNode(ISD::TokenFactor, Chain.getDebugLoc(), MVT::Other, 3245 &ArgChains[0], ArgChains.size()); 3246 } 3247 3248 /// SplatByte - Distribute ByteVal over NumBits bits. 3249 static APInt SplatByte(unsigned NumBits, uint8_t ByteVal) { 3250 APInt Val = APInt(NumBits, ByteVal); 3251 unsigned Shift = 8; 3252 for (unsigned i = NumBits; i > 8; i >>= 1) { 3253 Val = (Val << Shift) | Val; 3254 Shift <<= 1; 3255 } 3256 return Val; 3257 } 3258 3259 /// getMemsetValue - Vectorized representation of the memset value 3260 /// operand. 3261 static SDValue getMemsetValue(SDValue Value, EVT VT, SelectionDAG &DAG, 3262 DebugLoc dl) { 3263 assert(Value.getOpcode() != ISD::UNDEF); 3264 3265 unsigned NumBits = VT.getScalarType().getSizeInBits(); 3266 if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(Value)) { 3267 APInt Val = SplatByte(NumBits, C->getZExtValue() & 255); 3268 if (VT.isInteger()) 3269 return DAG.getConstant(Val, VT); 3270 return DAG.getConstantFP(APFloat(Val), VT); 3271 } 3272 3273 Value = DAG.getNode(ISD::ZERO_EXTEND, dl, VT, Value); 3274 if (NumBits > 8) { 3275 // Use a multiplication with 0x010101... to extend the input to the 3276 // required length. 3277 APInt Magic = SplatByte(NumBits, 0x01); 3278 Value = DAG.getNode(ISD::MUL, dl, VT, Value, DAG.getConstant(Magic, VT)); 3279 } 3280 3281 return Value; 3282 } 3283 3284 /// getMemsetStringVal - Similar to getMemsetValue. Except this is only 3285 /// used when a memcpy is turned into a memset when the source is a constant 3286 /// string ptr. 3287 static SDValue getMemsetStringVal(EVT VT, DebugLoc dl, SelectionDAG &DAG, 3288 const TargetLowering &TLI, 3289 std::string &Str, unsigned Offset) { 3290 // Handle vector with all elements zero. 3291 if (Str.empty()) { 3292 if (VT.isInteger()) 3293 return DAG.getConstant(0, VT); 3294 else if (VT == MVT::f32 || VT == MVT::f64) 3295 return DAG.getConstantFP(0.0, VT); 3296 else if (VT.isVector()) { 3297 unsigned NumElts = VT.getVectorNumElements(); 3298 MVT EltVT = (VT.getVectorElementType() == MVT::f32) ? MVT::i32 : MVT::i64; 3299 return DAG.getNode(ISD::BITCAST, dl, VT, 3300 DAG.getConstant(0, EVT::getVectorVT(*DAG.getContext(), 3301 EltVT, NumElts))); 3302 } else 3303 llvm_unreachable("Expected type!"); 3304 } 3305 3306 assert(!VT.isVector() && "Can't handle vector type here!"); 3307 unsigned NumBits = VT.getSizeInBits(); 3308 unsigned MSB = NumBits / 8; 3309 uint64_t Val = 0; 3310 if (TLI.isLittleEndian()) 3311 Offset = Offset + MSB - 1; 3312 for (unsigned i = 0; i != MSB; ++i) { 3313 Val = (Val << 8) | (unsigned char)Str[Offset]; 3314 Offset += TLI.isLittleEndian() ? -1 : 1; 3315 } 3316 return DAG.getConstant(Val, VT); 3317 } 3318 3319 /// getMemBasePlusOffset - Returns base and offset node for the 3320 /// 3321 static SDValue getMemBasePlusOffset(SDValue Base, unsigned Offset, 3322 SelectionDAG &DAG) { 3323 EVT VT = Base.getValueType(); 3324 return DAG.getNode(ISD::ADD, Base.getDebugLoc(), 3325 VT, Base, DAG.getConstant(Offset, VT)); 3326 } 3327 3328 /// isMemSrcFromString - Returns true if memcpy source is a string constant. 3329 /// 3330 static bool isMemSrcFromString(SDValue Src, std::string &Str) { 3331 unsigned SrcDelta = 0; 3332 GlobalAddressSDNode *G = NULL; 3333 if (Src.getOpcode() == ISD::GlobalAddress) 3334 G = cast<GlobalAddressSDNode>(Src); 3335 else if (Src.getOpcode() == ISD::ADD && 3336 Src.getOperand(0).getOpcode() == ISD::GlobalAddress && 3337 Src.getOperand(1).getOpcode() == ISD::Constant) { 3338 G = cast<GlobalAddressSDNode>(Src.getOperand(0)); 3339 SrcDelta = cast<ConstantSDNode>(Src.getOperand(1))->getZExtValue(); 3340 } 3341 if (!G) 3342 return false; 3343 3344 const GlobalVariable *GV = dyn_cast<GlobalVariable>(G->getGlobal()); 3345 if (GV && GetConstantStringInfo(GV, Str, SrcDelta, false)) 3346 return true; 3347 3348 return false; 3349 } 3350 3351 /// FindOptimalMemOpLowering - Determines the optimial series memory ops 3352 /// to replace the memset / memcpy. Return true if the number of memory ops 3353 /// is below the threshold. It returns the types of the sequence of 3354 /// memory ops to perform memset / memcpy by reference. 3355 static bool FindOptimalMemOpLowering(std::vector<EVT> &MemOps, 3356 unsigned Limit, uint64_t Size, 3357 unsigned DstAlign, unsigned SrcAlign, 3358 bool IsZeroVal, 3359 bool MemcpyStrSrc, 3360 SelectionDAG &DAG, 3361 const TargetLowering &TLI) { 3362 assert((SrcAlign == 0 || SrcAlign >= DstAlign) && 3363 "Expecting memcpy / memset source to meet alignment requirement!"); 3364 // If 'SrcAlign' is zero, that means the memory operation does not need to 3365 // load the value, i.e. memset or memcpy from constant string. Otherwise, 3366 // it's the inferred alignment of the source. 'DstAlign', on the other hand, 3367 // is the specified alignment of the memory operation. If it is zero, that 3368 // means it's possible to change the alignment of the destination. 3369 // 'MemcpyStrSrc' indicates whether the memcpy source is constant so it does 3370 // not need to be loaded. 3371 EVT VT = TLI.getOptimalMemOpType(Size, DstAlign, SrcAlign, 3372 IsZeroVal, MemcpyStrSrc, 3373 DAG.getMachineFunction()); 3374 3375 if (VT == MVT::Other) { 3376 if (DstAlign >= TLI.getTargetData()->getPointerPrefAlignment() || 3377 TLI.allowsUnalignedMemoryAccesses(VT)) { 3378 VT = TLI.getPointerTy(); 3379 } else { 3380 switch (DstAlign & 7) { 3381 case 0: VT = MVT::i64; break; 3382 case 4: VT = MVT::i32; break; 3383 case 2: VT = MVT::i16; break; 3384 default: VT = MVT::i8; break; 3385 } 3386 } 3387 3388 MVT LVT = MVT::i64; 3389 while (!TLI.isTypeLegal(LVT)) 3390 LVT = (MVT::SimpleValueType)(LVT.SimpleTy - 1); 3391 assert(LVT.isInteger()); 3392 3393 if (VT.bitsGT(LVT)) 3394 VT = LVT; 3395 } 3396 3397 unsigned NumMemOps = 0; 3398 while (Size != 0) { 3399 unsigned VTSize = VT.getSizeInBits() / 8; 3400 while (VTSize > Size) { 3401 // For now, only use non-vector load / store's for the left-over pieces. 3402 if (VT.isVector() || VT.isFloatingPoint()) { 3403 VT = MVT::i64; 3404 while (!TLI.isTypeLegal(VT)) 3405 VT = (MVT::SimpleValueType)(VT.getSimpleVT().SimpleTy - 1); 3406 VTSize = VT.getSizeInBits() / 8; 3407 } else { 3408 // This can result in a type that is not legal on the target, e.g. 3409 // 1 or 2 bytes on PPC. 3410 VT = (MVT::SimpleValueType)(VT.getSimpleVT().SimpleTy - 1); 3411 VTSize >>= 1; 3412 } 3413 } 3414 3415 if (++NumMemOps > Limit) 3416 return false; 3417 MemOps.push_back(VT); 3418 Size -= VTSize; 3419 } 3420 3421 return true; 3422 } 3423 3424 static SDValue getMemcpyLoadsAndStores(SelectionDAG &DAG, DebugLoc dl, 3425 SDValue Chain, SDValue Dst, 3426 SDValue Src, uint64_t Size, 3427 unsigned Align, bool isVol, 3428 bool AlwaysInline, 3429 MachinePointerInfo DstPtrInfo, 3430 MachinePointerInfo SrcPtrInfo) { 3431 // Turn a memcpy of undef to nop. 3432 if (Src.getOpcode() == ISD::UNDEF) 3433 return Chain; 3434 3435 // Expand memcpy to a series of load and store ops if the size operand falls 3436 // below a certain threshold. 3437 // TODO: In the AlwaysInline case, if the size is big then generate a loop 3438 // rather than maybe a humongous number of loads and stores. 3439 const TargetLowering &TLI = DAG.getTargetLoweringInfo(); 3440 std::vector<EVT> MemOps; 3441 bool DstAlignCanChange = false; 3442 MachineFunction &MF = DAG.getMachineFunction(); 3443 MachineFrameInfo *MFI = MF.getFrameInfo(); 3444 bool OptSize = MF.getFunction()->hasFnAttr(Attribute::OptimizeForSize); 3445 FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Dst); 3446 if (FI && !MFI->isFixedObjectIndex(FI->getIndex())) 3447 DstAlignCanChange = true; 3448 unsigned SrcAlign = DAG.InferPtrAlignment(Src); 3449 if (Align > SrcAlign) 3450 SrcAlign = Align; 3451 std::string Str; 3452 bool CopyFromStr = isMemSrcFromString(Src, Str); 3453 bool isZeroStr = CopyFromStr && Str.empty(); 3454 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemcpy(OptSize); 3455 3456 if (!FindOptimalMemOpLowering(MemOps, Limit, Size, 3457 (DstAlignCanChange ? 0 : Align), 3458 (isZeroStr ? 0 : SrcAlign), 3459 true, CopyFromStr, DAG, TLI)) 3460 return SDValue(); 3461 3462 if (DstAlignCanChange) { 3463 Type *Ty = MemOps[0].getTypeForEVT(*DAG.getContext()); 3464 unsigned NewAlign = (unsigned) TLI.getTargetData()->getABITypeAlignment(Ty); 3465 if (NewAlign > Align) { 3466 // Give the stack frame object a larger alignment if needed. 3467 if (MFI->getObjectAlignment(FI->getIndex()) < NewAlign) 3468 MFI->setObjectAlignment(FI->getIndex(), NewAlign); 3469 Align = NewAlign; 3470 } 3471 } 3472 3473 SmallVector<SDValue, 8> OutChains; 3474 unsigned NumMemOps = MemOps.size(); 3475 uint64_t SrcOff = 0, DstOff = 0; 3476 for (unsigned i = 0; i != NumMemOps; ++i) { 3477 EVT VT = MemOps[i]; 3478 unsigned VTSize = VT.getSizeInBits() / 8; 3479 SDValue Value, Store; 3480 3481 if (CopyFromStr && 3482 (isZeroStr || (VT.isInteger() && !VT.isVector()))) { 3483 // It's unlikely a store of a vector immediate can be done in a single 3484 // instruction. It would require a load from a constantpool first. 3485 // We only handle zero vectors here. 3486 // FIXME: Handle other cases where store of vector immediate is done in 3487 // a single instruction. 3488 Value = getMemsetStringVal(VT, dl, DAG, TLI, Str, SrcOff); 3489 Store = DAG.getStore(Chain, dl, Value, 3490 getMemBasePlusOffset(Dst, DstOff, DAG), 3491 DstPtrInfo.getWithOffset(DstOff), isVol, 3492 false, Align); 3493 } else { 3494 // The type might not be legal for the target. This should only happen 3495 // if the type is smaller than a legal type, as on PPC, so the right 3496 // thing to do is generate a LoadExt/StoreTrunc pair. These simplify 3497 // to Load/Store if NVT==VT. 3498 // FIXME does the case above also need this? 3499 EVT NVT = TLI.getTypeToTransformTo(*DAG.getContext(), VT); 3500 assert(NVT.bitsGE(VT)); 3501 Value = DAG.getExtLoad(ISD::EXTLOAD, dl, NVT, Chain, 3502 getMemBasePlusOffset(Src, SrcOff, DAG), 3503 SrcPtrInfo.getWithOffset(SrcOff), VT, isVol, false, 3504 MinAlign(SrcAlign, SrcOff)); 3505 Store = DAG.getTruncStore(Chain, dl, Value, 3506 getMemBasePlusOffset(Dst, DstOff, DAG), 3507 DstPtrInfo.getWithOffset(DstOff), VT, isVol, 3508 false, Align); 3509 } 3510 OutChains.push_back(Store); 3511 SrcOff += VTSize; 3512 DstOff += VTSize; 3513 } 3514 3515 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, 3516 &OutChains[0], OutChains.size()); 3517 } 3518 3519 static SDValue getMemmoveLoadsAndStores(SelectionDAG &DAG, DebugLoc dl, 3520 SDValue Chain, SDValue Dst, 3521 SDValue Src, uint64_t Size, 3522 unsigned Align, bool isVol, 3523 bool AlwaysInline, 3524 MachinePointerInfo DstPtrInfo, 3525 MachinePointerInfo SrcPtrInfo) { 3526 // Turn a memmove of undef to nop. 3527 if (Src.getOpcode() == ISD::UNDEF) 3528 return Chain; 3529 3530 // Expand memmove to a series of load and store ops if the size operand falls 3531 // below a certain threshold. 3532 const TargetLowering &TLI = DAG.getTargetLoweringInfo(); 3533 std::vector<EVT> MemOps; 3534 bool DstAlignCanChange = false; 3535 MachineFunction &MF = DAG.getMachineFunction(); 3536 MachineFrameInfo *MFI = MF.getFrameInfo(); 3537 bool OptSize = MF.getFunction()->hasFnAttr(Attribute::OptimizeForSize); 3538 FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Dst); 3539 if (FI && !MFI->isFixedObjectIndex(FI->getIndex())) 3540 DstAlignCanChange = true; 3541 unsigned SrcAlign = DAG.InferPtrAlignment(Src); 3542 if (Align > SrcAlign) 3543 SrcAlign = Align; 3544 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemmove(OptSize); 3545 3546 if (!FindOptimalMemOpLowering(MemOps, Limit, Size, 3547 (DstAlignCanChange ? 0 : Align), 3548 SrcAlign, true, false, DAG, TLI)) 3549 return SDValue(); 3550 3551 if (DstAlignCanChange) { 3552 Type *Ty = MemOps[0].getTypeForEVT(*DAG.getContext()); 3553 unsigned NewAlign = (unsigned) TLI.getTargetData()->getABITypeAlignment(Ty); 3554 if (NewAlign > Align) { 3555 // Give the stack frame object a larger alignment if needed. 3556 if (MFI->getObjectAlignment(FI->getIndex()) < NewAlign) 3557 MFI->setObjectAlignment(FI->getIndex(), NewAlign); 3558 Align = NewAlign; 3559 } 3560 } 3561 3562 uint64_t SrcOff = 0, DstOff = 0; 3563 SmallVector<SDValue, 8> LoadValues; 3564 SmallVector<SDValue, 8> LoadChains; 3565 SmallVector<SDValue, 8> OutChains; 3566 unsigned NumMemOps = MemOps.size(); 3567 for (unsigned i = 0; i < NumMemOps; i++) { 3568 EVT VT = MemOps[i]; 3569 unsigned VTSize = VT.getSizeInBits() / 8; 3570 SDValue Value, Store; 3571 3572 Value = DAG.getLoad(VT, dl, Chain, 3573 getMemBasePlusOffset(Src, SrcOff, DAG), 3574 SrcPtrInfo.getWithOffset(SrcOff), isVol, 3575 false, false, SrcAlign); 3576 LoadValues.push_back(Value); 3577 LoadChains.push_back(Value.getValue(1)); 3578 SrcOff += VTSize; 3579 } 3580 Chain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, 3581 &LoadChains[0], LoadChains.size()); 3582 OutChains.clear(); 3583 for (unsigned i = 0; i < NumMemOps; i++) { 3584 EVT VT = MemOps[i]; 3585 unsigned VTSize = VT.getSizeInBits() / 8; 3586 SDValue Value, Store; 3587 3588 Store = DAG.getStore(Chain, dl, LoadValues[i], 3589 getMemBasePlusOffset(Dst, DstOff, DAG), 3590 DstPtrInfo.getWithOffset(DstOff), isVol, false, Align); 3591 OutChains.push_back(Store); 3592 DstOff += VTSize; 3593 } 3594 3595 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, 3596 &OutChains[0], OutChains.size()); 3597 } 3598 3599 static SDValue getMemsetStores(SelectionDAG &DAG, DebugLoc dl, 3600 SDValue Chain, SDValue Dst, 3601 SDValue Src, uint64_t Size, 3602 unsigned Align, bool isVol, 3603 MachinePointerInfo DstPtrInfo) { 3604 // Turn a memset of undef to nop. 3605 if (Src.getOpcode() == ISD::UNDEF) 3606 return Chain; 3607 3608 // Expand memset to a series of load/store ops if the size operand 3609 // falls below a certain threshold. 3610 const TargetLowering &TLI = DAG.getTargetLoweringInfo(); 3611 std::vector<EVT> MemOps; 3612 bool DstAlignCanChange = false; 3613 MachineFunction &MF = DAG.getMachineFunction(); 3614 MachineFrameInfo *MFI = MF.getFrameInfo(); 3615 bool OptSize = MF.getFunction()->hasFnAttr(Attribute::OptimizeForSize); 3616 FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Dst); 3617 if (FI && !MFI->isFixedObjectIndex(FI->getIndex())) 3618 DstAlignCanChange = true; 3619 bool IsZeroVal = 3620 isa<ConstantSDNode>(Src) && cast<ConstantSDNode>(Src)->isNullValue(); 3621 if (!FindOptimalMemOpLowering(MemOps, TLI.getMaxStoresPerMemset(OptSize), 3622 Size, (DstAlignCanChange ? 0 : Align), 0, 3623 IsZeroVal, false, DAG, TLI)) 3624 return SDValue(); 3625 3626 if (DstAlignCanChange) { 3627 Type *Ty = MemOps[0].getTypeForEVT(*DAG.getContext()); 3628 unsigned NewAlign = (unsigned) TLI.getTargetData()->getABITypeAlignment(Ty); 3629 if (NewAlign > Align) { 3630 // Give the stack frame object a larger alignment if needed. 3631 if (MFI->getObjectAlignment(FI->getIndex()) < NewAlign) 3632 MFI->setObjectAlignment(FI->getIndex(), NewAlign); 3633 Align = NewAlign; 3634 } 3635 } 3636 3637 SmallVector<SDValue, 8> OutChains; 3638 uint64_t DstOff = 0; 3639 unsigned NumMemOps = MemOps.size(); 3640 3641 // Find the largest store and generate the bit pattern for it. 3642 EVT LargestVT = MemOps[0]; 3643 for (unsigned i = 1; i < NumMemOps; i++) 3644 if (MemOps[i].bitsGT(LargestVT)) 3645 LargestVT = MemOps[i]; 3646 SDValue MemSetValue = getMemsetValue(Src, LargestVT, DAG, dl); 3647 3648 for (unsigned i = 0; i < NumMemOps; i++) { 3649 EVT VT = MemOps[i]; 3650 3651 // If this store is smaller than the largest store see whether we can get 3652 // the smaller value for free with a truncate. 3653 SDValue Value = MemSetValue; 3654 if (VT.bitsLT(LargestVT)) { 3655 if (!LargestVT.isVector() && !VT.isVector() && 3656 TLI.isTruncateFree(LargestVT, VT)) 3657 Value = DAG.getNode(ISD::TRUNCATE, dl, VT, MemSetValue); 3658 else 3659 Value = getMemsetValue(Src, VT, DAG, dl); 3660 } 3661 assert(Value.getValueType() == VT && "Value with wrong type."); 3662 SDValue Store = DAG.getStore(Chain, dl, Value, 3663 getMemBasePlusOffset(Dst, DstOff, DAG), 3664 DstPtrInfo.getWithOffset(DstOff), 3665 isVol, false, Align); 3666 OutChains.push_back(Store); 3667 DstOff += VT.getSizeInBits() / 8; 3668 } 3669 3670 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, 3671 &OutChains[0], OutChains.size()); 3672 } 3673 3674 SDValue SelectionDAG::getMemcpy(SDValue Chain, DebugLoc dl, SDValue Dst, 3675 SDValue Src, SDValue Size, 3676 unsigned Align, bool isVol, bool AlwaysInline, 3677 MachinePointerInfo DstPtrInfo, 3678 MachinePointerInfo SrcPtrInfo) { 3679 3680 // Check to see if we should lower the memcpy to loads and stores first. 3681 // For cases within the target-specified limits, this is the best choice. 3682 ConstantSDNode *ConstantSize = dyn_cast<ConstantSDNode>(Size); 3683 if (ConstantSize) { 3684 // Memcpy with size zero? Just return the original chain. 3685 if (ConstantSize->isNullValue()) 3686 return Chain; 3687 3688 SDValue Result = getMemcpyLoadsAndStores(*this, dl, Chain, Dst, Src, 3689 ConstantSize->getZExtValue(),Align, 3690 isVol, false, DstPtrInfo, SrcPtrInfo); 3691 if (Result.getNode()) 3692 return Result; 3693 } 3694 3695 // Then check to see if we should lower the memcpy with target-specific 3696 // code. If the target chooses to do this, this is the next best. 3697 SDValue Result = 3698 TSI.EmitTargetCodeForMemcpy(*this, dl, Chain, Dst, Src, Size, Align, 3699 isVol, AlwaysInline, 3700 DstPtrInfo, SrcPtrInfo); 3701 if (Result.getNode()) 3702 return Result; 3703 3704 // If we really need inline code and the target declined to provide it, 3705 // use a (potentially long) sequence of loads and stores. 3706 if (AlwaysInline) { 3707 assert(ConstantSize && "AlwaysInline requires a constant size!"); 3708 return getMemcpyLoadsAndStores(*this, dl, Chain, Dst, Src, 3709 ConstantSize->getZExtValue(), Align, isVol, 3710 true, DstPtrInfo, SrcPtrInfo); 3711 } 3712 3713 // FIXME: If the memcpy is volatile (isVol), lowering it to a plain libc 3714 // memcpy is not guaranteed to be safe. libc memcpys aren't required to 3715 // respect volatile, so they may do things like read or write memory 3716 // beyond the given memory regions. But fixing this isn't easy, and most 3717 // people don't care. 3718 3719 // Emit a library call. 3720 TargetLowering::ArgListTy Args; 3721 TargetLowering::ArgListEntry Entry; 3722 Entry.Ty = TLI.getTargetData()->getIntPtrType(*getContext()); 3723 Entry.Node = Dst; Args.push_back(Entry); 3724 Entry.Node = Src; Args.push_back(Entry); 3725 Entry.Node = Size; Args.push_back(Entry); 3726 // FIXME: pass in DebugLoc 3727 std::pair<SDValue,SDValue> CallResult = 3728 TLI.LowerCallTo(Chain, Type::getVoidTy(*getContext()), 3729 false, false, false, false, 0, 3730 TLI.getLibcallCallingConv(RTLIB::MEMCPY), false, 3731 /*isReturnValueUsed=*/false, 3732 getExternalSymbol(TLI.getLibcallName(RTLIB::MEMCPY), 3733 TLI.getPointerTy()), 3734 Args, *this, dl); 3735 return CallResult.second; 3736 } 3737 3738 SDValue SelectionDAG::getMemmove(SDValue Chain, DebugLoc dl, SDValue Dst, 3739 SDValue Src, SDValue Size, 3740 unsigned Align, bool isVol, 3741 MachinePointerInfo DstPtrInfo, 3742 MachinePointerInfo SrcPtrInfo) { 3743 3744 // Check to see if we should lower the memmove to loads and stores first. 3745 // For cases within the target-specified limits, this is the best choice. 3746 ConstantSDNode *ConstantSize = dyn_cast<ConstantSDNode>(Size); 3747 if (ConstantSize) { 3748 // Memmove with size zero? Just return the original chain. 3749 if (ConstantSize->isNullValue()) 3750 return Chain; 3751 3752 SDValue Result = 3753 getMemmoveLoadsAndStores(*this, dl, Chain, Dst, Src, 3754 ConstantSize->getZExtValue(), Align, isVol, 3755 false, DstPtrInfo, SrcPtrInfo); 3756 if (Result.getNode()) 3757 return Result; 3758 } 3759 3760 // Then check to see if we should lower the memmove with target-specific 3761 // code. If the target chooses to do this, this is the next best. 3762 SDValue Result = 3763 TSI.EmitTargetCodeForMemmove(*this, dl, Chain, Dst, Src, Size, Align, isVol, 3764 DstPtrInfo, SrcPtrInfo); 3765 if (Result.getNode()) 3766 return Result; 3767 3768 // FIXME: If the memmove is volatile, lowering it to plain libc memmove may 3769 // not be safe. See memcpy above for more details. 3770 3771 // Emit a library call. 3772 TargetLowering::ArgListTy Args; 3773 TargetLowering::ArgListEntry Entry; 3774 Entry.Ty = TLI.getTargetData()->getIntPtrType(*getContext()); 3775 Entry.Node = Dst; Args.push_back(Entry); 3776 Entry.Node = Src; Args.push_back(Entry); 3777 Entry.Node = Size; Args.push_back(Entry); 3778 // FIXME: pass in DebugLoc 3779 std::pair<SDValue,SDValue> CallResult = 3780 TLI.LowerCallTo(Chain, Type::getVoidTy(*getContext()), 3781 false, false, false, false, 0, 3782 TLI.getLibcallCallingConv(RTLIB::MEMMOVE), false, 3783 /*isReturnValueUsed=*/false, 3784 getExternalSymbol(TLI.getLibcallName(RTLIB::MEMMOVE), 3785 TLI.getPointerTy()), 3786 Args, *this, dl); 3787 return CallResult.second; 3788 } 3789 3790 SDValue SelectionDAG::getMemset(SDValue Chain, DebugLoc dl, SDValue Dst, 3791 SDValue Src, SDValue Size, 3792 unsigned Align, bool isVol, 3793 MachinePointerInfo DstPtrInfo) { 3794 3795 // Check to see if we should lower the memset to stores first. 3796 // For cases within the target-specified limits, this is the best choice. 3797 ConstantSDNode *ConstantSize = dyn_cast<ConstantSDNode>(Size); 3798 if (ConstantSize) { 3799 // Memset with size zero? Just return the original chain. 3800 if (ConstantSize->isNullValue()) 3801 return Chain; 3802 3803 SDValue Result = 3804 getMemsetStores(*this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), 3805 Align, isVol, DstPtrInfo); 3806 3807 if (Result.getNode()) 3808 return Result; 3809 } 3810 3811 // Then check to see if we should lower the memset with target-specific 3812 // code. If the target chooses to do this, this is the next best. 3813 SDValue Result = 3814 TSI.EmitTargetCodeForMemset(*this, dl, Chain, Dst, Src, Size, Align, isVol, 3815 DstPtrInfo); 3816 if (Result.getNode()) 3817 return Result; 3818 3819 // Emit a library call. 3820 Type *IntPtrTy = TLI.getTargetData()->getIntPtrType(*getContext()); 3821 TargetLowering::ArgListTy Args; 3822 TargetLowering::ArgListEntry Entry; 3823 Entry.Node = Dst; Entry.Ty = IntPtrTy; 3824 Args.push_back(Entry); 3825 // Extend or truncate the argument to be an i32 value for the call. 3826 if (Src.getValueType().bitsGT(MVT::i32)) 3827 Src = getNode(ISD::TRUNCATE, dl, MVT::i32, Src); 3828 else 3829 Src = getNode(ISD::ZERO_EXTEND, dl, MVT::i32, Src); 3830 Entry.Node = Src; 3831 Entry.Ty = Type::getInt32Ty(*getContext()); 3832 Entry.isSExt = true; 3833 Args.push_back(Entry); 3834 Entry.Node = Size; 3835 Entry.Ty = IntPtrTy; 3836 Entry.isSExt = false; 3837 Args.push_back(Entry); 3838 // FIXME: pass in DebugLoc 3839 std::pair<SDValue,SDValue> CallResult = 3840 TLI.LowerCallTo(Chain, Type::getVoidTy(*getContext()), 3841 false, false, false, false, 0, 3842 TLI.getLibcallCallingConv(RTLIB::MEMSET), false, 3843 /*isReturnValueUsed=*/false, 3844 getExternalSymbol(TLI.getLibcallName(RTLIB::MEMSET), 3845 TLI.getPointerTy()), 3846 Args, *this, dl); 3847 return CallResult.second; 3848 } 3849 3850 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 3851 SDValue Chain, SDValue Ptr, SDValue Cmp, 3852 SDValue Swp, MachinePointerInfo PtrInfo, 3853 unsigned Alignment, 3854 AtomicOrdering Ordering, 3855 SynchronizationScope SynchScope) { 3856 if (Alignment == 0) // Ensure that codegen never sees alignment 0 3857 Alignment = getEVTAlignment(MemVT); 3858 3859 MachineFunction &MF = getMachineFunction(); 3860 unsigned Flags = MachineMemOperand::MOLoad | MachineMemOperand::MOStore; 3861 3862 // For now, atomics are considered to be volatile always. 3863 // FIXME: Volatile isn't really correct; we should keep track of atomic 3864 // orderings in the memoperand. 3865 Flags |= MachineMemOperand::MOVolatile; 3866 3867 MachineMemOperand *MMO = 3868 MF.getMachineMemOperand(PtrInfo, Flags, MemVT.getStoreSize(), Alignment); 3869 3870 return getAtomic(Opcode, dl, MemVT, Chain, Ptr, Cmp, Swp, MMO, 3871 Ordering, SynchScope); 3872 } 3873 3874 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 3875 SDValue Chain, 3876 SDValue Ptr, SDValue Cmp, 3877 SDValue Swp, MachineMemOperand *MMO, 3878 AtomicOrdering Ordering, 3879 SynchronizationScope SynchScope) { 3880 assert(Opcode == ISD::ATOMIC_CMP_SWAP && "Invalid Atomic Op"); 3881 assert(Cmp.getValueType() == Swp.getValueType() && "Invalid Atomic Op Types"); 3882 3883 EVT VT = Cmp.getValueType(); 3884 3885 SDVTList VTs = getVTList(VT, MVT::Other); 3886 FoldingSetNodeID ID; 3887 ID.AddInteger(MemVT.getRawBits()); 3888 SDValue Ops[] = {Chain, Ptr, Cmp, Swp}; 3889 AddNodeIDNode(ID, Opcode, VTs, Ops, 4); 3890 void* IP = 0; 3891 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 3892 cast<AtomicSDNode>(E)->refineAlignment(MMO); 3893 return SDValue(E, 0); 3894 } 3895 SDNode *N = new (NodeAllocator) AtomicSDNode(Opcode, dl, VTs, MemVT, Chain, 3896 Ptr, Cmp, Swp, MMO, Ordering, 3897 SynchScope); 3898 CSEMap.InsertNode(N, IP); 3899 AllNodes.push_back(N); 3900 return SDValue(N, 0); 3901 } 3902 3903 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 3904 SDValue Chain, 3905 SDValue Ptr, SDValue Val, 3906 const Value* PtrVal, 3907 unsigned Alignment, 3908 AtomicOrdering Ordering, 3909 SynchronizationScope SynchScope) { 3910 if (Alignment == 0) // Ensure that codegen never sees alignment 0 3911 Alignment = getEVTAlignment(MemVT); 3912 3913 MachineFunction &MF = getMachineFunction(); 3914 // A monotonic store does not load; a release store "loads" in the sense 3915 // that other stores cannot be sunk past it. 3916 // (An atomicrmw obviously both loads and stores.) 3917 unsigned Flags = MachineMemOperand::MOStore; 3918 if (Opcode != ISD::ATOMIC_STORE || Ordering > Monotonic) 3919 Flags |= MachineMemOperand::MOLoad; 3920 3921 // For now, atomics are considered to be volatile always. 3922 // FIXME: Volatile isn't really correct; we should keep track of atomic 3923 // orderings in the memoperand. 3924 Flags |= MachineMemOperand::MOVolatile; 3925 3926 MachineMemOperand *MMO = 3927 MF.getMachineMemOperand(MachinePointerInfo(PtrVal), Flags, 3928 MemVT.getStoreSize(), Alignment); 3929 3930 return getAtomic(Opcode, dl, MemVT, Chain, Ptr, Val, MMO, 3931 Ordering, SynchScope); 3932 } 3933 3934 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 3935 SDValue Chain, 3936 SDValue Ptr, SDValue Val, 3937 MachineMemOperand *MMO, 3938 AtomicOrdering Ordering, 3939 SynchronizationScope SynchScope) { 3940 assert((Opcode == ISD::ATOMIC_LOAD_ADD || 3941 Opcode == ISD::ATOMIC_LOAD_SUB || 3942 Opcode == ISD::ATOMIC_LOAD_AND || 3943 Opcode == ISD::ATOMIC_LOAD_OR || 3944 Opcode == ISD::ATOMIC_LOAD_XOR || 3945 Opcode == ISD::ATOMIC_LOAD_NAND || 3946 Opcode == ISD::ATOMIC_LOAD_MIN || 3947 Opcode == ISD::ATOMIC_LOAD_MAX || 3948 Opcode == ISD::ATOMIC_LOAD_UMIN || 3949 Opcode == ISD::ATOMIC_LOAD_UMAX || 3950 Opcode == ISD::ATOMIC_SWAP || 3951 Opcode == ISD::ATOMIC_STORE) && 3952 "Invalid Atomic Op"); 3953 3954 EVT VT = Val.getValueType(); 3955 3956 SDVTList VTs = Opcode == ISD::ATOMIC_STORE ? getVTList(MVT::Other) : 3957 getVTList(VT, MVT::Other); 3958 FoldingSetNodeID ID; 3959 ID.AddInteger(MemVT.getRawBits()); 3960 SDValue Ops[] = {Chain, Ptr, Val}; 3961 AddNodeIDNode(ID, Opcode, VTs, Ops, 3); 3962 void* IP = 0; 3963 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 3964 cast<AtomicSDNode>(E)->refineAlignment(MMO); 3965 return SDValue(E, 0); 3966 } 3967 SDNode *N = new (NodeAllocator) AtomicSDNode(Opcode, dl, VTs, MemVT, Chain, 3968 Ptr, Val, MMO, 3969 Ordering, SynchScope); 3970 CSEMap.InsertNode(N, IP); 3971 AllNodes.push_back(N); 3972 return SDValue(N, 0); 3973 } 3974 3975 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 3976 EVT VT, SDValue Chain, 3977 SDValue Ptr, 3978 const Value* PtrVal, 3979 unsigned Alignment, 3980 AtomicOrdering Ordering, 3981 SynchronizationScope SynchScope) { 3982 if (Alignment == 0) // Ensure that codegen never sees alignment 0 3983 Alignment = getEVTAlignment(MemVT); 3984 3985 MachineFunction &MF = getMachineFunction(); 3986 // A monotonic load does not store; an acquire load "stores" in the sense 3987 // that other loads cannot be hoisted past it. 3988 unsigned Flags = MachineMemOperand::MOLoad; 3989 if (Ordering > Monotonic) 3990 Flags |= MachineMemOperand::MOStore; 3991 3992 // For now, atomics are considered to be volatile always. 3993 // FIXME: Volatile isn't really correct; we should keep track of atomic 3994 // orderings in the memoperand. 3995 Flags |= MachineMemOperand::MOVolatile; 3996 3997 MachineMemOperand *MMO = 3998 MF.getMachineMemOperand(MachinePointerInfo(PtrVal), Flags, 3999 MemVT.getStoreSize(), Alignment); 4000 4001 return getAtomic(Opcode, dl, MemVT, VT, Chain, Ptr, MMO, 4002 Ordering, SynchScope); 4003 } 4004 4005 SDValue SelectionDAG::getAtomic(unsigned Opcode, DebugLoc dl, EVT MemVT, 4006 EVT VT, SDValue Chain, 4007 SDValue Ptr, 4008 MachineMemOperand *MMO, 4009 AtomicOrdering Ordering, 4010 SynchronizationScope SynchScope) { 4011 assert(Opcode == ISD::ATOMIC_LOAD && "Invalid Atomic Op"); 4012 4013 SDVTList VTs = getVTList(VT, MVT::Other); 4014 FoldingSetNodeID ID; 4015 ID.AddInteger(MemVT.getRawBits()); 4016 SDValue Ops[] = {Chain, Ptr}; 4017 AddNodeIDNode(ID, Opcode, VTs, Ops, 2); 4018 void* IP = 0; 4019 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 4020 cast<AtomicSDNode>(E)->refineAlignment(MMO); 4021 return SDValue(E, 0); 4022 } 4023 SDNode *N = new (NodeAllocator) AtomicSDNode(Opcode, dl, VTs, MemVT, Chain, 4024 Ptr, MMO, Ordering, SynchScope); 4025 CSEMap.InsertNode(N, IP); 4026 AllNodes.push_back(N); 4027 return SDValue(N, 0); 4028 } 4029 4030 /// getMergeValues - Create a MERGE_VALUES node from the given operands. 4031 SDValue SelectionDAG::getMergeValues(const SDValue *Ops, unsigned NumOps, 4032 DebugLoc dl) { 4033 if (NumOps == 1) 4034 return Ops[0]; 4035 4036 SmallVector<EVT, 4> VTs; 4037 VTs.reserve(NumOps); 4038 for (unsigned i = 0; i < NumOps; ++i) 4039 VTs.push_back(Ops[i].getValueType()); 4040 return getNode(ISD::MERGE_VALUES, dl, getVTList(&VTs[0], NumOps), 4041 Ops, NumOps); 4042 } 4043 4044 SDValue 4045 SelectionDAG::getMemIntrinsicNode(unsigned Opcode, DebugLoc dl, 4046 const EVT *VTs, unsigned NumVTs, 4047 const SDValue *Ops, unsigned NumOps, 4048 EVT MemVT, MachinePointerInfo PtrInfo, 4049 unsigned Align, bool Vol, 4050 bool ReadMem, bool WriteMem) { 4051 return getMemIntrinsicNode(Opcode, dl, makeVTList(VTs, NumVTs), Ops, NumOps, 4052 MemVT, PtrInfo, Align, Vol, 4053 ReadMem, WriteMem); 4054 } 4055 4056 SDValue 4057 SelectionDAG::getMemIntrinsicNode(unsigned Opcode, DebugLoc dl, SDVTList VTList, 4058 const SDValue *Ops, unsigned NumOps, 4059 EVT MemVT, MachinePointerInfo PtrInfo, 4060 unsigned Align, bool Vol, 4061 bool ReadMem, bool WriteMem) { 4062 if (Align == 0) // Ensure that codegen never sees alignment 0 4063 Align = getEVTAlignment(MemVT); 4064 4065 MachineFunction &MF = getMachineFunction(); 4066 unsigned Flags = 0; 4067 if (WriteMem) 4068 Flags |= MachineMemOperand::MOStore; 4069 if (ReadMem) 4070 Flags |= MachineMemOperand::MOLoad; 4071 if (Vol) 4072 Flags |= MachineMemOperand::MOVolatile; 4073 MachineMemOperand *MMO = 4074 MF.getMachineMemOperand(PtrInfo, Flags, MemVT.getStoreSize(), Align); 4075 4076 return getMemIntrinsicNode(Opcode, dl, VTList, Ops, NumOps, MemVT, MMO); 4077 } 4078 4079 SDValue 4080 SelectionDAG::getMemIntrinsicNode(unsigned Opcode, DebugLoc dl, SDVTList VTList, 4081 const SDValue *Ops, unsigned NumOps, 4082 EVT MemVT, MachineMemOperand *MMO) { 4083 assert((Opcode == ISD::INTRINSIC_VOID || 4084 Opcode == ISD::INTRINSIC_W_CHAIN || 4085 Opcode == ISD::PREFETCH || 4086 (Opcode <= INT_MAX && 4087 (int)Opcode >= ISD::FIRST_TARGET_MEMORY_OPCODE)) && 4088 "Opcode is not a memory-accessing opcode!"); 4089 4090 // Memoize the node unless it returns a flag. 4091 MemIntrinsicSDNode *N; 4092 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) { 4093 FoldingSetNodeID ID; 4094 AddNodeIDNode(ID, Opcode, VTList, Ops, NumOps); 4095 void *IP = 0; 4096 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 4097 cast<MemIntrinsicSDNode>(E)->refineAlignment(MMO); 4098 return SDValue(E, 0); 4099 } 4100 4101 N = new (NodeAllocator) MemIntrinsicSDNode(Opcode, dl, VTList, Ops, NumOps, 4102 MemVT, MMO); 4103 CSEMap.InsertNode(N, IP); 4104 } else { 4105 N = new (NodeAllocator) MemIntrinsicSDNode(Opcode, dl, VTList, Ops, NumOps, 4106 MemVT, MMO); 4107 } 4108 AllNodes.push_back(N); 4109 return SDValue(N, 0); 4110 } 4111 4112 /// InferPointerInfo - If the specified ptr/offset is a frame index, infer a 4113 /// MachinePointerInfo record from it. This is particularly useful because the 4114 /// code generator has many cases where it doesn't bother passing in a 4115 /// MachinePointerInfo to getLoad or getStore when it has "FI+Cst". 4116 static MachinePointerInfo InferPointerInfo(SDValue Ptr, int64_t Offset = 0) { 4117 // If this is FI+Offset, we can model it. 4118 if (const FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Ptr)) 4119 return MachinePointerInfo::getFixedStack(FI->getIndex(), Offset); 4120 4121 // If this is (FI+Offset1)+Offset2, we can model it. 4122 if (Ptr.getOpcode() != ISD::ADD || 4123 !isa<ConstantSDNode>(Ptr.getOperand(1)) || 4124 !isa<FrameIndexSDNode>(Ptr.getOperand(0))) 4125 return MachinePointerInfo(); 4126 4127 int FI = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex(); 4128 return MachinePointerInfo::getFixedStack(FI, Offset+ 4129 cast<ConstantSDNode>(Ptr.getOperand(1))->getSExtValue()); 4130 } 4131 4132 /// InferPointerInfo - If the specified ptr/offset is a frame index, infer a 4133 /// MachinePointerInfo record from it. This is particularly useful because the 4134 /// code generator has many cases where it doesn't bother passing in a 4135 /// MachinePointerInfo to getLoad or getStore when it has "FI+Cst". 4136 static MachinePointerInfo InferPointerInfo(SDValue Ptr, SDValue OffsetOp) { 4137 // If the 'Offset' value isn't a constant, we can't handle this. 4138 if (ConstantSDNode *OffsetNode = dyn_cast<ConstantSDNode>(OffsetOp)) 4139 return InferPointerInfo(Ptr, OffsetNode->getSExtValue()); 4140 if (OffsetOp.getOpcode() == ISD::UNDEF) 4141 return InferPointerInfo(Ptr); 4142 return MachinePointerInfo(); 4143 } 4144 4145 4146 SDValue 4147 SelectionDAG::getLoad(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, 4148 EVT VT, DebugLoc dl, SDValue Chain, 4149 SDValue Ptr, SDValue Offset, 4150 MachinePointerInfo PtrInfo, EVT MemVT, 4151 bool isVolatile, bool isNonTemporal, bool isInvariant, 4152 unsigned Alignment, const MDNode *TBAAInfo) { 4153 assert(Chain.getValueType() == MVT::Other && 4154 "Invalid chain type"); 4155 if (Alignment == 0) // Ensure that codegen never sees alignment 0 4156 Alignment = getEVTAlignment(VT); 4157 4158 unsigned Flags = MachineMemOperand::MOLoad; 4159 if (isVolatile) 4160 Flags |= MachineMemOperand::MOVolatile; 4161 if (isNonTemporal) 4162 Flags |= MachineMemOperand::MONonTemporal; 4163 if (isInvariant) 4164 Flags |= MachineMemOperand::MOInvariant; 4165 4166 // If we don't have a PtrInfo, infer the trivial frame index case to simplify 4167 // clients. 4168 if (PtrInfo.V == 0) 4169 PtrInfo = InferPointerInfo(Ptr, Offset); 4170 4171 MachineFunction &MF = getMachineFunction(); 4172 MachineMemOperand *MMO = 4173 MF.getMachineMemOperand(PtrInfo, Flags, MemVT.getStoreSize(), Alignment, 4174 TBAAInfo); 4175 return getLoad(AM, ExtType, VT, dl, Chain, Ptr, Offset, MemVT, MMO); 4176 } 4177 4178 SDValue 4179 SelectionDAG::getLoad(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, 4180 EVT VT, DebugLoc dl, SDValue Chain, 4181 SDValue Ptr, SDValue Offset, EVT MemVT, 4182 MachineMemOperand *MMO) { 4183 if (VT == MemVT) { 4184 ExtType = ISD::NON_EXTLOAD; 4185 } else if (ExtType == ISD::NON_EXTLOAD) { 4186 assert(VT == MemVT && "Non-extending load from different memory type!"); 4187 } else { 4188 // Extending load. 4189 assert(MemVT.getScalarType().bitsLT(VT.getScalarType()) && 4190 "Should only be an extending load, not truncating!"); 4191 assert(VT.isInteger() == MemVT.isInteger() && 4192 "Cannot convert from FP to Int or Int -> FP!"); 4193 assert(VT.isVector() == MemVT.isVector() && 4194 "Cannot use trunc store to convert to or from a vector!"); 4195 assert((!VT.isVector() || 4196 VT.getVectorNumElements() == MemVT.getVectorNumElements()) && 4197 "Cannot use trunc store to change the number of vector elements!"); 4198 } 4199 4200 bool Indexed = AM != ISD::UNINDEXED; 4201 assert((Indexed || Offset.getOpcode() == ISD::UNDEF) && 4202 "Unindexed load with an offset!"); 4203 4204 SDVTList VTs = Indexed ? 4205 getVTList(VT, Ptr.getValueType(), MVT::Other) : getVTList(VT, MVT::Other); 4206 SDValue Ops[] = { Chain, Ptr, Offset }; 4207 FoldingSetNodeID ID; 4208 AddNodeIDNode(ID, ISD::LOAD, VTs, Ops, 3); 4209 ID.AddInteger(MemVT.getRawBits()); 4210 ID.AddInteger(encodeMemSDNodeFlags(ExtType, AM, MMO->isVolatile(), 4211 MMO->isNonTemporal(), 4212 MMO->isInvariant())); 4213 void *IP = 0; 4214 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 4215 cast<LoadSDNode>(E)->refineAlignment(MMO); 4216 return SDValue(E, 0); 4217 } 4218 SDNode *N = new (NodeAllocator) LoadSDNode(Ops, dl, VTs, AM, ExtType, 4219 MemVT, MMO); 4220 CSEMap.InsertNode(N, IP); 4221 AllNodes.push_back(N); 4222 return SDValue(N, 0); 4223 } 4224 4225 SDValue SelectionDAG::getLoad(EVT VT, DebugLoc dl, 4226 SDValue Chain, SDValue Ptr, 4227 MachinePointerInfo PtrInfo, 4228 bool isVolatile, bool isNonTemporal, 4229 bool isInvariant, unsigned Alignment, 4230 const MDNode *TBAAInfo) { 4231 SDValue Undef = getUNDEF(Ptr.getValueType()); 4232 return getLoad(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef, 4233 PtrInfo, VT, isVolatile, isNonTemporal, isInvariant, Alignment, 4234 TBAAInfo); 4235 } 4236 4237 SDValue SelectionDAG::getExtLoad(ISD::LoadExtType ExtType, DebugLoc dl, EVT VT, 4238 SDValue Chain, SDValue Ptr, 4239 MachinePointerInfo PtrInfo, EVT MemVT, 4240 bool isVolatile, bool isNonTemporal, 4241 unsigned Alignment, const MDNode *TBAAInfo) { 4242 SDValue Undef = getUNDEF(Ptr.getValueType()); 4243 return getLoad(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, 4244 PtrInfo, MemVT, isVolatile, isNonTemporal, false, Alignment, 4245 TBAAInfo); 4246 } 4247 4248 4249 SDValue 4250 SelectionDAG::getIndexedLoad(SDValue OrigLoad, DebugLoc dl, SDValue Base, 4251 SDValue Offset, ISD::MemIndexedMode AM) { 4252 LoadSDNode *LD = cast<LoadSDNode>(OrigLoad); 4253 assert(LD->getOffset().getOpcode() == ISD::UNDEF && 4254 "Load is already a indexed load!"); 4255 return getLoad(AM, LD->getExtensionType(), OrigLoad.getValueType(), dl, 4256 LD->getChain(), Base, Offset, LD->getPointerInfo(), 4257 LD->getMemoryVT(), LD->isVolatile(), LD->isNonTemporal(), 4258 false, LD->getAlignment()); 4259 } 4260 4261 SDValue SelectionDAG::getStore(SDValue Chain, DebugLoc dl, SDValue Val, 4262 SDValue Ptr, MachinePointerInfo PtrInfo, 4263 bool isVolatile, bool isNonTemporal, 4264 unsigned Alignment, const MDNode *TBAAInfo) { 4265 assert(Chain.getValueType() == MVT::Other && 4266 "Invalid chain type"); 4267 if (Alignment == 0) // Ensure that codegen never sees alignment 0 4268 Alignment = getEVTAlignment(Val.getValueType()); 4269 4270 unsigned Flags = MachineMemOperand::MOStore; 4271 if (isVolatile) 4272 Flags |= MachineMemOperand::MOVolatile; 4273 if (isNonTemporal) 4274 Flags |= MachineMemOperand::MONonTemporal; 4275 4276 if (PtrInfo.V == 0) 4277 PtrInfo = InferPointerInfo(Ptr); 4278 4279 MachineFunction &MF = getMachineFunction(); 4280 MachineMemOperand *MMO = 4281 MF.getMachineMemOperand(PtrInfo, Flags, 4282 Val.getValueType().getStoreSize(), Alignment, 4283 TBAAInfo); 4284 4285 return getStore(Chain, dl, Val, Ptr, MMO); 4286 } 4287 4288 SDValue SelectionDAG::getStore(SDValue Chain, DebugLoc dl, SDValue Val, 4289 SDValue Ptr, MachineMemOperand *MMO) { 4290 assert(Chain.getValueType() == MVT::Other && 4291 "Invalid chain type"); 4292 EVT VT = Val.getValueType(); 4293 SDVTList VTs = getVTList(MVT::Other); 4294 SDValue Undef = getUNDEF(Ptr.getValueType()); 4295 SDValue Ops[] = { Chain, Val, Ptr, Undef }; 4296 FoldingSetNodeID ID; 4297 AddNodeIDNode(ID, ISD::STORE, VTs, Ops, 4); 4298 ID.AddInteger(VT.getRawBits()); 4299 ID.AddInteger(encodeMemSDNodeFlags(false, ISD::UNINDEXED, MMO->isVolatile(), 4300 MMO->isNonTemporal(), MMO->isInvariant())); 4301 void *IP = 0; 4302 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 4303 cast<StoreSDNode>(E)->refineAlignment(MMO); 4304 return SDValue(E, 0); 4305 } 4306 SDNode *N = new (NodeAllocator) StoreSDNode(Ops, dl, VTs, ISD::UNINDEXED, 4307 false, VT, MMO); 4308 CSEMap.InsertNode(N, IP); 4309 AllNodes.push_back(N); 4310 return SDValue(N, 0); 4311 } 4312 4313 SDValue SelectionDAG::getTruncStore(SDValue Chain, DebugLoc dl, SDValue Val, 4314 SDValue Ptr, MachinePointerInfo PtrInfo, 4315 EVT SVT,bool isVolatile, bool isNonTemporal, 4316 unsigned Alignment, 4317 const MDNode *TBAAInfo) { 4318 assert(Chain.getValueType() == MVT::Other && 4319 "Invalid chain type"); 4320 if (Alignment == 0) // Ensure that codegen never sees alignment 0 4321 Alignment = getEVTAlignment(SVT); 4322 4323 unsigned Flags = MachineMemOperand::MOStore; 4324 if (isVolatile) 4325 Flags |= MachineMemOperand::MOVolatile; 4326 if (isNonTemporal) 4327 Flags |= MachineMemOperand::MONonTemporal; 4328 4329 if (PtrInfo.V == 0) 4330 PtrInfo = InferPointerInfo(Ptr); 4331 4332 MachineFunction &MF = getMachineFunction(); 4333 MachineMemOperand *MMO = 4334 MF.getMachineMemOperand(PtrInfo, Flags, SVT.getStoreSize(), Alignment, 4335 TBAAInfo); 4336 4337 return getTruncStore(Chain, dl, Val, Ptr, SVT, MMO); 4338 } 4339 4340 SDValue SelectionDAG::getTruncStore(SDValue Chain, DebugLoc dl, SDValue Val, 4341 SDValue Ptr, EVT SVT, 4342 MachineMemOperand *MMO) { 4343 EVT VT = Val.getValueType(); 4344 4345 assert(Chain.getValueType() == MVT::Other && 4346 "Invalid chain type"); 4347 if (VT == SVT) 4348 return getStore(Chain, dl, Val, Ptr, MMO); 4349 4350 assert(SVT.getScalarType().bitsLT(VT.getScalarType()) && 4351 "Should only be a truncating store, not extending!"); 4352 assert(VT.isInteger() == SVT.isInteger() && 4353 "Can't do FP-INT conversion!"); 4354 assert(VT.isVector() == SVT.isVector() && 4355 "Cannot use trunc store to convert to or from a vector!"); 4356 assert((!VT.isVector() || 4357 VT.getVectorNumElements() == SVT.getVectorNumElements()) && 4358 "Cannot use trunc store to change the number of vector elements!"); 4359 4360 SDVTList VTs = getVTList(MVT::Other); 4361 SDValue Undef = getUNDEF(Ptr.getValueType()); 4362 SDValue Ops[] = { Chain, Val, Ptr, Undef }; 4363 FoldingSetNodeID ID; 4364 AddNodeIDNode(ID, ISD::STORE, VTs, Ops, 4); 4365 ID.AddInteger(SVT.getRawBits()); 4366 ID.AddInteger(encodeMemSDNodeFlags(true, ISD::UNINDEXED, MMO->isVolatile(), 4367 MMO->isNonTemporal(), MMO->isInvariant())); 4368 void *IP = 0; 4369 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 4370 cast<StoreSDNode>(E)->refineAlignment(MMO); 4371 return SDValue(E, 0); 4372 } 4373 SDNode *N = new (NodeAllocator) StoreSDNode(Ops, dl, VTs, ISD::UNINDEXED, 4374 true, SVT, MMO); 4375 CSEMap.InsertNode(N, IP); 4376 AllNodes.push_back(N); 4377 return SDValue(N, 0); 4378 } 4379 4380 SDValue 4381 SelectionDAG::getIndexedStore(SDValue OrigStore, DebugLoc dl, SDValue Base, 4382 SDValue Offset, ISD::MemIndexedMode AM) { 4383 StoreSDNode *ST = cast<StoreSDNode>(OrigStore); 4384 assert(ST->getOffset().getOpcode() == ISD::UNDEF && 4385 "Store is already a indexed store!"); 4386 SDVTList VTs = getVTList(Base.getValueType(), MVT::Other); 4387 SDValue Ops[] = { ST->getChain(), ST->getValue(), Base, Offset }; 4388 FoldingSetNodeID ID; 4389 AddNodeIDNode(ID, ISD::STORE, VTs, Ops, 4); 4390 ID.AddInteger(ST->getMemoryVT().getRawBits()); 4391 ID.AddInteger(ST->getRawSubclassData()); 4392 void *IP = 0; 4393 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 4394 return SDValue(E, 0); 4395 4396 SDNode *N = new (NodeAllocator) StoreSDNode(Ops, dl, VTs, AM, 4397 ST->isTruncatingStore(), 4398 ST->getMemoryVT(), 4399 ST->getMemOperand()); 4400 CSEMap.InsertNode(N, IP); 4401 AllNodes.push_back(N); 4402 return SDValue(N, 0); 4403 } 4404 4405 SDValue SelectionDAG::getVAArg(EVT VT, DebugLoc dl, 4406 SDValue Chain, SDValue Ptr, 4407 SDValue SV, 4408 unsigned Align) { 4409 SDValue Ops[] = { Chain, Ptr, SV, getTargetConstant(Align, MVT::i32) }; 4410 return getNode(ISD::VAARG, dl, getVTList(VT, MVT::Other), Ops, 4); 4411 } 4412 4413 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 4414 const SDUse *Ops, unsigned NumOps) { 4415 switch (NumOps) { 4416 case 0: return getNode(Opcode, DL, VT); 4417 case 1: return getNode(Opcode, DL, VT, Ops[0]); 4418 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1]); 4419 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2]); 4420 default: break; 4421 } 4422 4423 // Copy from an SDUse array into an SDValue array for use with 4424 // the regular getNode logic. 4425 SmallVector<SDValue, 8> NewOps(Ops, Ops + NumOps); 4426 return getNode(Opcode, DL, VT, &NewOps[0], NumOps); 4427 } 4428 4429 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, EVT VT, 4430 const SDValue *Ops, unsigned NumOps) { 4431 switch (NumOps) { 4432 case 0: return getNode(Opcode, DL, VT); 4433 case 1: return getNode(Opcode, DL, VT, Ops[0]); 4434 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1]); 4435 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2]); 4436 default: break; 4437 } 4438 4439 switch (Opcode) { 4440 default: break; 4441 case ISD::SELECT_CC: { 4442 assert(NumOps == 5 && "SELECT_CC takes 5 operands!"); 4443 assert(Ops[0].getValueType() == Ops[1].getValueType() && 4444 "LHS and RHS of condition must have same type!"); 4445 assert(Ops[2].getValueType() == Ops[3].getValueType() && 4446 "True and False arms of SelectCC must have same type!"); 4447 assert(Ops[2].getValueType() == VT && 4448 "select_cc node must be of same type as true and false value!"); 4449 break; 4450 } 4451 case ISD::BR_CC: { 4452 assert(NumOps == 5 && "BR_CC takes 5 operands!"); 4453 assert(Ops[2].getValueType() == Ops[3].getValueType() && 4454 "LHS/RHS of comparison should match types!"); 4455 break; 4456 } 4457 } 4458 4459 // Memoize nodes. 4460 SDNode *N; 4461 SDVTList VTs = getVTList(VT); 4462 4463 if (VT != MVT::Glue) { 4464 FoldingSetNodeID ID; 4465 AddNodeIDNode(ID, Opcode, VTs, Ops, NumOps); 4466 void *IP = 0; 4467 4468 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 4469 return SDValue(E, 0); 4470 4471 N = new (NodeAllocator) SDNode(Opcode, DL, VTs, Ops, NumOps); 4472 CSEMap.InsertNode(N, IP); 4473 } else { 4474 N = new (NodeAllocator) SDNode(Opcode, DL, VTs, Ops, NumOps); 4475 } 4476 4477 AllNodes.push_back(N); 4478 #ifndef NDEBUG 4479 VerifySDNode(N); 4480 #endif 4481 return SDValue(N, 0); 4482 } 4483 4484 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, 4485 const std::vector<EVT> &ResultTys, 4486 const SDValue *Ops, unsigned NumOps) { 4487 return getNode(Opcode, DL, getVTList(&ResultTys[0], ResultTys.size()), 4488 Ops, NumOps); 4489 } 4490 4491 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, 4492 const EVT *VTs, unsigned NumVTs, 4493 const SDValue *Ops, unsigned NumOps) { 4494 if (NumVTs == 1) 4495 return getNode(Opcode, DL, VTs[0], Ops, NumOps); 4496 return getNode(Opcode, DL, makeVTList(VTs, NumVTs), Ops, NumOps); 4497 } 4498 4499 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4500 const SDValue *Ops, unsigned NumOps) { 4501 if (VTList.NumVTs == 1) 4502 return getNode(Opcode, DL, VTList.VTs[0], Ops, NumOps); 4503 4504 #if 0 4505 switch (Opcode) { 4506 // FIXME: figure out how to safely handle things like 4507 // int foo(int x) { return 1 << (x & 255); } 4508 // int bar() { return foo(256); } 4509 case ISD::SRA_PARTS: 4510 case ISD::SRL_PARTS: 4511 case ISD::SHL_PARTS: 4512 if (N3.getOpcode() == ISD::SIGN_EXTEND_INREG && 4513 cast<VTSDNode>(N3.getOperand(1))->getVT() != MVT::i1) 4514 return getNode(Opcode, DL, VT, N1, N2, N3.getOperand(0)); 4515 else if (N3.getOpcode() == ISD::AND) 4516 if (ConstantSDNode *AndRHS = dyn_cast<ConstantSDNode>(N3.getOperand(1))) { 4517 // If the and is only masking out bits that cannot effect the shift, 4518 // eliminate the and. 4519 unsigned NumBits = VT.getScalarType().getSizeInBits()*2; 4520 if ((AndRHS->getValue() & (NumBits-1)) == NumBits-1) 4521 return getNode(Opcode, DL, VT, N1, N2, N3.getOperand(0)); 4522 } 4523 break; 4524 } 4525 #endif 4526 4527 // Memoize the node unless it returns a flag. 4528 SDNode *N; 4529 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) { 4530 FoldingSetNodeID ID; 4531 AddNodeIDNode(ID, Opcode, VTList, Ops, NumOps); 4532 void *IP = 0; 4533 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 4534 return SDValue(E, 0); 4535 4536 if (NumOps == 1) { 4537 N = new (NodeAllocator) UnarySDNode(Opcode, DL, VTList, Ops[0]); 4538 } else if (NumOps == 2) { 4539 N = new (NodeAllocator) BinarySDNode(Opcode, DL, VTList, Ops[0], Ops[1]); 4540 } else if (NumOps == 3) { 4541 N = new (NodeAllocator) TernarySDNode(Opcode, DL, VTList, Ops[0], Ops[1], 4542 Ops[2]); 4543 } else { 4544 N = new (NodeAllocator) SDNode(Opcode, DL, VTList, Ops, NumOps); 4545 } 4546 CSEMap.InsertNode(N, IP); 4547 } else { 4548 if (NumOps == 1) { 4549 N = new (NodeAllocator) UnarySDNode(Opcode, DL, VTList, Ops[0]); 4550 } else if (NumOps == 2) { 4551 N = new (NodeAllocator) BinarySDNode(Opcode, DL, VTList, Ops[0], Ops[1]); 4552 } else if (NumOps == 3) { 4553 N = new (NodeAllocator) TernarySDNode(Opcode, DL, VTList, Ops[0], Ops[1], 4554 Ops[2]); 4555 } else { 4556 N = new (NodeAllocator) SDNode(Opcode, DL, VTList, Ops, NumOps); 4557 } 4558 } 4559 AllNodes.push_back(N); 4560 #ifndef NDEBUG 4561 VerifySDNode(N); 4562 #endif 4563 return SDValue(N, 0); 4564 } 4565 4566 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList) { 4567 return getNode(Opcode, DL, VTList, 0, 0); 4568 } 4569 4570 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4571 SDValue N1) { 4572 SDValue Ops[] = { N1 }; 4573 return getNode(Opcode, DL, VTList, Ops, 1); 4574 } 4575 4576 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4577 SDValue N1, SDValue N2) { 4578 SDValue Ops[] = { N1, N2 }; 4579 return getNode(Opcode, DL, VTList, Ops, 2); 4580 } 4581 4582 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4583 SDValue N1, SDValue N2, SDValue N3) { 4584 SDValue Ops[] = { N1, N2, N3 }; 4585 return getNode(Opcode, DL, VTList, Ops, 3); 4586 } 4587 4588 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4589 SDValue N1, SDValue N2, SDValue N3, 4590 SDValue N4) { 4591 SDValue Ops[] = { N1, N2, N3, N4 }; 4592 return getNode(Opcode, DL, VTList, Ops, 4); 4593 } 4594 4595 SDValue SelectionDAG::getNode(unsigned Opcode, DebugLoc DL, SDVTList VTList, 4596 SDValue N1, SDValue N2, SDValue N3, 4597 SDValue N4, SDValue N5) { 4598 SDValue Ops[] = { N1, N2, N3, N4, N5 }; 4599 return getNode(Opcode, DL, VTList, Ops, 5); 4600 } 4601 4602 SDVTList SelectionDAG::getVTList(EVT VT) { 4603 return makeVTList(SDNode::getValueTypeList(VT), 1); 4604 } 4605 4606 SDVTList SelectionDAG::getVTList(EVT VT1, EVT VT2) { 4607 for (std::vector<SDVTList>::reverse_iterator I = VTList.rbegin(), 4608 E = VTList.rend(); I != E; ++I) 4609 if (I->NumVTs == 2 && I->VTs[0] == VT1 && I->VTs[1] == VT2) 4610 return *I; 4611 4612 EVT *Array = Allocator.Allocate<EVT>(2); 4613 Array[0] = VT1; 4614 Array[1] = VT2; 4615 SDVTList Result = makeVTList(Array, 2); 4616 VTList.push_back(Result); 4617 return Result; 4618 } 4619 4620 SDVTList SelectionDAG::getVTList(EVT VT1, EVT VT2, EVT VT3) { 4621 for (std::vector<SDVTList>::reverse_iterator I = VTList.rbegin(), 4622 E = VTList.rend(); I != E; ++I) 4623 if (I->NumVTs == 3 && I->VTs[0] == VT1 && I->VTs[1] == VT2 && 4624 I->VTs[2] == VT3) 4625 return *I; 4626 4627 EVT *Array = Allocator.Allocate<EVT>(3); 4628 Array[0] = VT1; 4629 Array[1] = VT2; 4630 Array[2] = VT3; 4631 SDVTList Result = makeVTList(Array, 3); 4632 VTList.push_back(Result); 4633 return Result; 4634 } 4635 4636 SDVTList SelectionDAG::getVTList(EVT VT1, EVT VT2, EVT VT3, EVT VT4) { 4637 for (std::vector<SDVTList>::reverse_iterator I = VTList.rbegin(), 4638 E = VTList.rend(); I != E; ++I) 4639 if (I->NumVTs == 4 && I->VTs[0] == VT1 && I->VTs[1] == VT2 && 4640 I->VTs[2] == VT3 && I->VTs[3] == VT4) 4641 return *I; 4642 4643 EVT *Array = Allocator.Allocate<EVT>(4); 4644 Array[0] = VT1; 4645 Array[1] = VT2; 4646 Array[2] = VT3; 4647 Array[3] = VT4; 4648 SDVTList Result = makeVTList(Array, 4); 4649 VTList.push_back(Result); 4650 return Result; 4651 } 4652 4653 SDVTList SelectionDAG::getVTList(const EVT *VTs, unsigned NumVTs) { 4654 switch (NumVTs) { 4655 case 0: llvm_unreachable("Cannot have nodes without results!"); 4656 case 1: return getVTList(VTs[0]); 4657 case 2: return getVTList(VTs[0], VTs[1]); 4658 case 3: return getVTList(VTs[0], VTs[1], VTs[2]); 4659 case 4: return getVTList(VTs[0], VTs[1], VTs[2], VTs[3]); 4660 default: break; 4661 } 4662 4663 for (std::vector<SDVTList>::reverse_iterator I = VTList.rbegin(), 4664 E = VTList.rend(); I != E; ++I) { 4665 if (I->NumVTs != NumVTs || VTs[0] != I->VTs[0] || VTs[1] != I->VTs[1]) 4666 continue; 4667 4668 bool NoMatch = false; 4669 for (unsigned i = 2; i != NumVTs; ++i) 4670 if (VTs[i] != I->VTs[i]) { 4671 NoMatch = true; 4672 break; 4673 } 4674 if (!NoMatch) 4675 return *I; 4676 } 4677 4678 EVT *Array = Allocator.Allocate<EVT>(NumVTs); 4679 std::copy(VTs, VTs+NumVTs, Array); 4680 SDVTList Result = makeVTList(Array, NumVTs); 4681 VTList.push_back(Result); 4682 return Result; 4683 } 4684 4685 4686 /// UpdateNodeOperands - *Mutate* the specified node in-place to have the 4687 /// specified operands. If the resultant node already exists in the DAG, 4688 /// this does not modify the specified node, instead it returns the node that 4689 /// already exists. If the resultant node does not exist in the DAG, the 4690 /// input node is returned. As a degenerate case, if you specify the same 4691 /// input operands as the node already has, the input node is returned. 4692 SDNode *SelectionDAG::UpdateNodeOperands(SDNode *N, SDValue Op) { 4693 assert(N->getNumOperands() == 1 && "Update with wrong number of operands"); 4694 4695 // Check to see if there is no change. 4696 if (Op == N->getOperand(0)) return N; 4697 4698 // See if the modified node already exists. 4699 void *InsertPos = 0; 4700 if (SDNode *Existing = FindModifiedNodeSlot(N, Op, InsertPos)) 4701 return Existing; 4702 4703 // Nope it doesn't. Remove the node from its current place in the maps. 4704 if (InsertPos) 4705 if (!RemoveNodeFromCSEMaps(N)) 4706 InsertPos = 0; 4707 4708 // Now we update the operands. 4709 N->OperandList[0].set(Op); 4710 4711 // If this gets put into a CSE map, add it. 4712 if (InsertPos) CSEMap.InsertNode(N, InsertPos); 4713 return N; 4714 } 4715 4716 SDNode *SelectionDAG::UpdateNodeOperands(SDNode *N, SDValue Op1, SDValue Op2) { 4717 assert(N->getNumOperands() == 2 && "Update with wrong number of operands"); 4718 4719 // Check to see if there is no change. 4720 if (Op1 == N->getOperand(0) && Op2 == N->getOperand(1)) 4721 return N; // No operands changed, just return the input node. 4722 4723 // See if the modified node already exists. 4724 void *InsertPos = 0; 4725 if (SDNode *Existing = FindModifiedNodeSlot(N, Op1, Op2, InsertPos)) 4726 return Existing; 4727 4728 // Nope it doesn't. Remove the node from its current place in the maps. 4729 if (InsertPos) 4730 if (!RemoveNodeFromCSEMaps(N)) 4731 InsertPos = 0; 4732 4733 // Now we update the operands. 4734 if (N->OperandList[0] != Op1) 4735 N->OperandList[0].set(Op1); 4736 if (N->OperandList[1] != Op2) 4737 N->OperandList[1].set(Op2); 4738 4739 // If this gets put into a CSE map, add it. 4740 if (InsertPos) CSEMap.InsertNode(N, InsertPos); 4741 return N; 4742 } 4743 4744 SDNode *SelectionDAG:: 4745 UpdateNodeOperands(SDNode *N, SDValue Op1, SDValue Op2, SDValue Op3) { 4746 SDValue Ops[] = { Op1, Op2, Op3 }; 4747 return UpdateNodeOperands(N, Ops, 3); 4748 } 4749 4750 SDNode *SelectionDAG:: 4751 UpdateNodeOperands(SDNode *N, SDValue Op1, SDValue Op2, 4752 SDValue Op3, SDValue Op4) { 4753 SDValue Ops[] = { Op1, Op2, Op3, Op4 }; 4754 return UpdateNodeOperands(N, Ops, 4); 4755 } 4756 4757 SDNode *SelectionDAG:: 4758 UpdateNodeOperands(SDNode *N, SDValue Op1, SDValue Op2, 4759 SDValue Op3, SDValue Op4, SDValue Op5) { 4760 SDValue Ops[] = { Op1, Op2, Op3, Op4, Op5 }; 4761 return UpdateNodeOperands(N, Ops, 5); 4762 } 4763 4764 SDNode *SelectionDAG:: 4765 UpdateNodeOperands(SDNode *N, const SDValue *Ops, unsigned NumOps) { 4766 assert(N->getNumOperands() == NumOps && 4767 "Update with wrong number of operands"); 4768 4769 // Check to see if there is no change. 4770 bool AnyChange = false; 4771 for (unsigned i = 0; i != NumOps; ++i) { 4772 if (Ops[i] != N->getOperand(i)) { 4773 AnyChange = true; 4774 break; 4775 } 4776 } 4777 4778 // No operands changed, just return the input node. 4779 if (!AnyChange) return N; 4780 4781 // See if the modified node already exists. 4782 void *InsertPos = 0; 4783 if (SDNode *Existing = FindModifiedNodeSlot(N, Ops, NumOps, InsertPos)) 4784 return Existing; 4785 4786 // Nope it doesn't. Remove the node from its current place in the maps. 4787 if (InsertPos) 4788 if (!RemoveNodeFromCSEMaps(N)) 4789 InsertPos = 0; 4790 4791 // Now we update the operands. 4792 for (unsigned i = 0; i != NumOps; ++i) 4793 if (N->OperandList[i] != Ops[i]) 4794 N->OperandList[i].set(Ops[i]); 4795 4796 // If this gets put into a CSE map, add it. 4797 if (InsertPos) CSEMap.InsertNode(N, InsertPos); 4798 return N; 4799 } 4800 4801 /// DropOperands - Release the operands and set this node to have 4802 /// zero operands. 4803 void SDNode::DropOperands() { 4804 // Unlike the code in MorphNodeTo that does this, we don't need to 4805 // watch for dead nodes here. 4806 for (op_iterator I = op_begin(), E = op_end(); I != E; ) { 4807 SDUse &Use = *I++; 4808 Use.set(SDValue()); 4809 } 4810 } 4811 4812 /// SelectNodeTo - These are wrappers around MorphNodeTo that accept a 4813 /// machine opcode. 4814 /// 4815 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4816 EVT VT) { 4817 SDVTList VTs = getVTList(VT); 4818 return SelectNodeTo(N, MachineOpc, VTs, 0, 0); 4819 } 4820 4821 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4822 EVT VT, SDValue Op1) { 4823 SDVTList VTs = getVTList(VT); 4824 SDValue Ops[] = { Op1 }; 4825 return SelectNodeTo(N, MachineOpc, VTs, Ops, 1); 4826 } 4827 4828 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4829 EVT VT, SDValue Op1, 4830 SDValue Op2) { 4831 SDVTList VTs = getVTList(VT); 4832 SDValue Ops[] = { Op1, Op2 }; 4833 return SelectNodeTo(N, MachineOpc, VTs, Ops, 2); 4834 } 4835 4836 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4837 EVT VT, SDValue Op1, 4838 SDValue Op2, SDValue Op3) { 4839 SDVTList VTs = getVTList(VT); 4840 SDValue Ops[] = { Op1, Op2, Op3 }; 4841 return SelectNodeTo(N, MachineOpc, VTs, Ops, 3); 4842 } 4843 4844 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4845 EVT VT, const SDValue *Ops, 4846 unsigned NumOps) { 4847 SDVTList VTs = getVTList(VT); 4848 return SelectNodeTo(N, MachineOpc, VTs, Ops, NumOps); 4849 } 4850 4851 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4852 EVT VT1, EVT VT2, const SDValue *Ops, 4853 unsigned NumOps) { 4854 SDVTList VTs = getVTList(VT1, VT2); 4855 return SelectNodeTo(N, MachineOpc, VTs, Ops, NumOps); 4856 } 4857 4858 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4859 EVT VT1, EVT VT2) { 4860 SDVTList VTs = getVTList(VT1, VT2); 4861 return SelectNodeTo(N, MachineOpc, VTs, (SDValue *)0, 0); 4862 } 4863 4864 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4865 EVT VT1, EVT VT2, EVT VT3, 4866 const SDValue *Ops, unsigned NumOps) { 4867 SDVTList VTs = getVTList(VT1, VT2, VT3); 4868 return SelectNodeTo(N, MachineOpc, VTs, Ops, NumOps); 4869 } 4870 4871 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4872 EVT VT1, EVT VT2, EVT VT3, EVT VT4, 4873 const SDValue *Ops, unsigned NumOps) { 4874 SDVTList VTs = getVTList(VT1, VT2, VT3, VT4); 4875 return SelectNodeTo(N, MachineOpc, VTs, Ops, NumOps); 4876 } 4877 4878 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4879 EVT VT1, EVT VT2, 4880 SDValue Op1) { 4881 SDVTList VTs = getVTList(VT1, VT2); 4882 SDValue Ops[] = { Op1 }; 4883 return SelectNodeTo(N, MachineOpc, VTs, Ops, 1); 4884 } 4885 4886 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4887 EVT VT1, EVT VT2, 4888 SDValue Op1, SDValue Op2) { 4889 SDVTList VTs = getVTList(VT1, VT2); 4890 SDValue Ops[] = { Op1, Op2 }; 4891 return SelectNodeTo(N, MachineOpc, VTs, Ops, 2); 4892 } 4893 4894 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4895 EVT VT1, EVT VT2, 4896 SDValue Op1, SDValue Op2, 4897 SDValue Op3) { 4898 SDVTList VTs = getVTList(VT1, VT2); 4899 SDValue Ops[] = { Op1, Op2, Op3 }; 4900 return SelectNodeTo(N, MachineOpc, VTs, Ops, 3); 4901 } 4902 4903 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4904 EVT VT1, EVT VT2, EVT VT3, 4905 SDValue Op1, SDValue Op2, 4906 SDValue Op3) { 4907 SDVTList VTs = getVTList(VT1, VT2, VT3); 4908 SDValue Ops[] = { Op1, Op2, Op3 }; 4909 return SelectNodeTo(N, MachineOpc, VTs, Ops, 3); 4910 } 4911 4912 SDNode *SelectionDAG::SelectNodeTo(SDNode *N, unsigned MachineOpc, 4913 SDVTList VTs, const SDValue *Ops, 4914 unsigned NumOps) { 4915 N = MorphNodeTo(N, ~MachineOpc, VTs, Ops, NumOps); 4916 // Reset the NodeID to -1. 4917 N->setNodeId(-1); 4918 return N; 4919 } 4920 4921 /// UpdadeDebugLocOnMergedSDNode - If the opt level is -O0 then it throws away 4922 /// the line number information on the merged node since it is not possible to 4923 /// preserve the information that operation is associated with multiple lines. 4924 /// This will make the debugger working better at -O0, were there is a higher 4925 /// probability having other instructions associated with that line. 4926 /// 4927 SDNode *SelectionDAG::UpdadeDebugLocOnMergedSDNode(SDNode *N, DebugLoc OLoc) { 4928 DebugLoc NLoc = N->getDebugLoc(); 4929 if (!(NLoc.isUnknown()) && (OptLevel == CodeGenOpt::None) && (OLoc != NLoc)) { 4930 N->setDebugLoc(DebugLoc()); 4931 } 4932 return N; 4933 } 4934 4935 /// MorphNodeTo - This *mutates* the specified node to have the specified 4936 /// return type, opcode, and operands. 4937 /// 4938 /// Note that MorphNodeTo returns the resultant node. If there is already a 4939 /// node of the specified opcode and operands, it returns that node instead of 4940 /// the current one. Note that the DebugLoc need not be the same. 4941 /// 4942 /// Using MorphNodeTo is faster than creating a new node and swapping it in 4943 /// with ReplaceAllUsesWith both because it often avoids allocating a new 4944 /// node, and because it doesn't require CSE recalculation for any of 4945 /// the node's users. 4946 /// 4947 SDNode *SelectionDAG::MorphNodeTo(SDNode *N, unsigned Opc, 4948 SDVTList VTs, const SDValue *Ops, 4949 unsigned NumOps) { 4950 // If an identical node already exists, use it. 4951 void *IP = 0; 4952 if (VTs.VTs[VTs.NumVTs-1] != MVT::Glue) { 4953 FoldingSetNodeID ID; 4954 AddNodeIDNode(ID, Opc, VTs, Ops, NumOps); 4955 if (SDNode *ON = CSEMap.FindNodeOrInsertPos(ID, IP)) 4956 return UpdadeDebugLocOnMergedSDNode(ON, N->getDebugLoc()); 4957 } 4958 4959 if (!RemoveNodeFromCSEMaps(N)) 4960 IP = 0; 4961 4962 // Start the morphing. 4963 N->NodeType = Opc; 4964 N->ValueList = VTs.VTs; 4965 N->NumValues = VTs.NumVTs; 4966 4967 // Clear the operands list, updating used nodes to remove this from their 4968 // use list. Keep track of any operands that become dead as a result. 4969 SmallPtrSet<SDNode*, 16> DeadNodeSet; 4970 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) { 4971 SDUse &Use = *I++; 4972 SDNode *Used = Use.getNode(); 4973 Use.set(SDValue()); 4974 if (Used->use_empty()) 4975 DeadNodeSet.insert(Used); 4976 } 4977 4978 if (MachineSDNode *MN = dyn_cast<MachineSDNode>(N)) { 4979 // Initialize the memory references information. 4980 MN->setMemRefs(0, 0); 4981 // If NumOps is larger than the # of operands we can have in a 4982 // MachineSDNode, reallocate the operand list. 4983 if (NumOps > MN->NumOperands || !MN->OperandsNeedDelete) { 4984 if (MN->OperandsNeedDelete) 4985 delete[] MN->OperandList; 4986 if (NumOps > array_lengthof(MN->LocalOperands)) 4987 // We're creating a final node that will live unmorphed for the 4988 // remainder of the current SelectionDAG iteration, so we can allocate 4989 // the operands directly out of a pool with no recycling metadata. 4990 MN->InitOperands(OperandAllocator.Allocate<SDUse>(NumOps), 4991 Ops, NumOps); 4992 else 4993 MN->InitOperands(MN->LocalOperands, Ops, NumOps); 4994 MN->OperandsNeedDelete = false; 4995 } else 4996 MN->InitOperands(MN->OperandList, Ops, NumOps); 4997 } else { 4998 // If NumOps is larger than the # of operands we currently have, reallocate 4999 // the operand list. 5000 if (NumOps > N->NumOperands) { 5001 if (N->OperandsNeedDelete) 5002 delete[] N->OperandList; 5003 N->InitOperands(new SDUse[NumOps], Ops, NumOps); 5004 N->OperandsNeedDelete = true; 5005 } else 5006 N->InitOperands(N->OperandList, Ops, NumOps); 5007 } 5008 5009 // Delete any nodes that are still dead after adding the uses for the 5010 // new operands. 5011 if (!DeadNodeSet.empty()) { 5012 SmallVector<SDNode *, 16> DeadNodes; 5013 for (SmallPtrSet<SDNode *, 16>::iterator I = DeadNodeSet.begin(), 5014 E = DeadNodeSet.end(); I != E; ++I) 5015 if ((*I)->use_empty()) 5016 DeadNodes.push_back(*I); 5017 RemoveDeadNodes(DeadNodes); 5018 } 5019 5020 if (IP) 5021 CSEMap.InsertNode(N, IP); // Memoize the new node. 5022 return N; 5023 } 5024 5025 5026 /// getMachineNode - These are used for target selectors to create a new node 5027 /// with specified return type(s), MachineInstr opcode, and operands. 5028 /// 5029 /// Note that getMachineNode returns the resultant node. If there is already a 5030 /// node of the specified opcode and operands, it returns that node instead of 5031 /// the current one. 5032 MachineSDNode * 5033 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT) { 5034 SDVTList VTs = getVTList(VT); 5035 return getMachineNode(Opcode, dl, VTs, 0, 0); 5036 } 5037 5038 MachineSDNode * 5039 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT, SDValue Op1) { 5040 SDVTList VTs = getVTList(VT); 5041 SDValue Ops[] = { Op1 }; 5042 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5043 } 5044 5045 MachineSDNode * 5046 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT, 5047 SDValue Op1, SDValue Op2) { 5048 SDVTList VTs = getVTList(VT); 5049 SDValue Ops[] = { Op1, Op2 }; 5050 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5051 } 5052 5053 MachineSDNode * 5054 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT, 5055 SDValue Op1, SDValue Op2, SDValue Op3) { 5056 SDVTList VTs = getVTList(VT); 5057 SDValue Ops[] = { Op1, Op2, Op3 }; 5058 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5059 } 5060 5061 MachineSDNode * 5062 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT, 5063 const SDValue *Ops, unsigned NumOps) { 5064 SDVTList VTs = getVTList(VT); 5065 return getMachineNode(Opcode, dl, VTs, Ops, NumOps); 5066 } 5067 5068 MachineSDNode * 5069 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT1, EVT VT2) { 5070 SDVTList VTs = getVTList(VT1, VT2); 5071 return getMachineNode(Opcode, dl, VTs, 0, 0); 5072 } 5073 5074 MachineSDNode * 5075 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5076 EVT VT1, EVT VT2, SDValue Op1) { 5077 SDVTList VTs = getVTList(VT1, VT2); 5078 SDValue Ops[] = { Op1 }; 5079 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5080 } 5081 5082 MachineSDNode * 5083 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5084 EVT VT1, EVT VT2, SDValue Op1, SDValue Op2) { 5085 SDVTList VTs = getVTList(VT1, VT2); 5086 SDValue Ops[] = { Op1, Op2 }; 5087 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5088 } 5089 5090 MachineSDNode * 5091 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5092 EVT VT1, EVT VT2, SDValue Op1, 5093 SDValue Op2, SDValue Op3) { 5094 SDVTList VTs = getVTList(VT1, VT2); 5095 SDValue Ops[] = { Op1, Op2, Op3 }; 5096 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5097 } 5098 5099 MachineSDNode * 5100 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5101 EVT VT1, EVT VT2, 5102 const SDValue *Ops, unsigned NumOps) { 5103 SDVTList VTs = getVTList(VT1, VT2); 5104 return getMachineNode(Opcode, dl, VTs, Ops, NumOps); 5105 } 5106 5107 MachineSDNode * 5108 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5109 EVT VT1, EVT VT2, EVT VT3, 5110 SDValue Op1, SDValue Op2) { 5111 SDVTList VTs = getVTList(VT1, VT2, VT3); 5112 SDValue Ops[] = { Op1, Op2 }; 5113 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5114 } 5115 5116 MachineSDNode * 5117 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5118 EVT VT1, EVT VT2, EVT VT3, 5119 SDValue Op1, SDValue Op2, SDValue Op3) { 5120 SDVTList VTs = getVTList(VT1, VT2, VT3); 5121 SDValue Ops[] = { Op1, Op2, Op3 }; 5122 return getMachineNode(Opcode, dl, VTs, Ops, array_lengthof(Ops)); 5123 } 5124 5125 MachineSDNode * 5126 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5127 EVT VT1, EVT VT2, EVT VT3, 5128 const SDValue *Ops, unsigned NumOps) { 5129 SDVTList VTs = getVTList(VT1, VT2, VT3); 5130 return getMachineNode(Opcode, dl, VTs, Ops, NumOps); 5131 } 5132 5133 MachineSDNode * 5134 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, EVT VT1, 5135 EVT VT2, EVT VT3, EVT VT4, 5136 const SDValue *Ops, unsigned NumOps) { 5137 SDVTList VTs = getVTList(VT1, VT2, VT3, VT4); 5138 return getMachineNode(Opcode, dl, VTs, Ops, NumOps); 5139 } 5140 5141 MachineSDNode * 5142 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc dl, 5143 const std::vector<EVT> &ResultTys, 5144 const SDValue *Ops, unsigned NumOps) { 5145 SDVTList VTs = getVTList(&ResultTys[0], ResultTys.size()); 5146 return getMachineNode(Opcode, dl, VTs, Ops, NumOps); 5147 } 5148 5149 MachineSDNode * 5150 SelectionDAG::getMachineNode(unsigned Opcode, DebugLoc DL, SDVTList VTs, 5151 const SDValue *Ops, unsigned NumOps) { 5152 bool DoCSE = VTs.VTs[VTs.NumVTs-1] != MVT::Glue; 5153 MachineSDNode *N; 5154 void *IP = 0; 5155 5156 if (DoCSE) { 5157 FoldingSetNodeID ID; 5158 AddNodeIDNode(ID, ~Opcode, VTs, Ops, NumOps); 5159 IP = 0; 5160 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) { 5161 return cast<MachineSDNode>(UpdadeDebugLocOnMergedSDNode(E, DL)); 5162 } 5163 } 5164 5165 // Allocate a new MachineSDNode. 5166 N = new (NodeAllocator) MachineSDNode(~Opcode, DL, VTs); 5167 5168 // Initialize the operands list. 5169 if (NumOps > array_lengthof(N->LocalOperands)) 5170 // We're creating a final node that will live unmorphed for the 5171 // remainder of the current SelectionDAG iteration, so we can allocate 5172 // the operands directly out of a pool with no recycling metadata. 5173 N->InitOperands(OperandAllocator.Allocate<SDUse>(NumOps), 5174 Ops, NumOps); 5175 else 5176 N->InitOperands(N->LocalOperands, Ops, NumOps); 5177 N->OperandsNeedDelete = false; 5178 5179 if (DoCSE) 5180 CSEMap.InsertNode(N, IP); 5181 5182 AllNodes.push_back(N); 5183 #ifndef NDEBUG 5184 VerifyMachineNode(N); 5185 #endif 5186 return N; 5187 } 5188 5189 /// getTargetExtractSubreg - A convenience function for creating 5190 /// TargetOpcode::EXTRACT_SUBREG nodes. 5191 SDValue 5192 SelectionDAG::getTargetExtractSubreg(int SRIdx, DebugLoc DL, EVT VT, 5193 SDValue Operand) { 5194 SDValue SRIdxVal = getTargetConstant(SRIdx, MVT::i32); 5195 SDNode *Subreg = getMachineNode(TargetOpcode::EXTRACT_SUBREG, DL, 5196 VT, Operand, SRIdxVal); 5197 return SDValue(Subreg, 0); 5198 } 5199 5200 /// getTargetInsertSubreg - A convenience function for creating 5201 /// TargetOpcode::INSERT_SUBREG nodes. 5202 SDValue 5203 SelectionDAG::getTargetInsertSubreg(int SRIdx, DebugLoc DL, EVT VT, 5204 SDValue Operand, SDValue Subreg) { 5205 SDValue SRIdxVal = getTargetConstant(SRIdx, MVT::i32); 5206 SDNode *Result = getMachineNode(TargetOpcode::INSERT_SUBREG, DL, 5207 VT, Operand, Subreg, SRIdxVal); 5208 return SDValue(Result, 0); 5209 } 5210 5211 /// getNodeIfExists - Get the specified node if it's already available, or 5212 /// else return NULL. 5213 SDNode *SelectionDAG::getNodeIfExists(unsigned Opcode, SDVTList VTList, 5214 const SDValue *Ops, unsigned NumOps) { 5215 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) { 5216 FoldingSetNodeID ID; 5217 AddNodeIDNode(ID, Opcode, VTList, Ops, NumOps); 5218 void *IP = 0; 5219 if (SDNode *E = CSEMap.FindNodeOrInsertPos(ID, IP)) 5220 return E; 5221 } 5222 return NULL; 5223 } 5224 5225 /// getDbgValue - Creates a SDDbgValue node. 5226 /// 5227 SDDbgValue * 5228 SelectionDAG::getDbgValue(MDNode *MDPtr, SDNode *N, unsigned R, uint64_t Off, 5229 DebugLoc DL, unsigned O) { 5230 return new (Allocator) SDDbgValue(MDPtr, N, R, Off, DL, O); 5231 } 5232 5233 SDDbgValue * 5234 SelectionDAG::getDbgValue(MDNode *MDPtr, const Value *C, uint64_t Off, 5235 DebugLoc DL, unsigned O) { 5236 return new (Allocator) SDDbgValue(MDPtr, C, Off, DL, O); 5237 } 5238 5239 SDDbgValue * 5240 SelectionDAG::getDbgValue(MDNode *MDPtr, unsigned FI, uint64_t Off, 5241 DebugLoc DL, unsigned O) { 5242 return new (Allocator) SDDbgValue(MDPtr, FI, Off, DL, O); 5243 } 5244 5245 namespace { 5246 5247 /// RAUWUpdateListener - Helper for ReplaceAllUsesWith - When the node 5248 /// pointed to by a use iterator is deleted, increment the use iterator 5249 /// so that it doesn't dangle. 5250 /// 5251 /// This class also manages a "downlink" DAGUpdateListener, to forward 5252 /// messages to ReplaceAllUsesWith's callers. 5253 /// 5254 class RAUWUpdateListener : public SelectionDAG::DAGUpdateListener { 5255 SelectionDAG::DAGUpdateListener *DownLink; 5256 SDNode::use_iterator &UI; 5257 SDNode::use_iterator &UE; 5258 5259 virtual void NodeDeleted(SDNode *N, SDNode *E) { 5260 // Increment the iterator as needed. 5261 while (UI != UE && N == *UI) 5262 ++UI; 5263 5264 // Then forward the message. 5265 if (DownLink) DownLink->NodeDeleted(N, E); 5266 } 5267 5268 virtual void NodeUpdated(SDNode *N) { 5269 // Just forward the message. 5270 if (DownLink) DownLink->NodeUpdated(N); 5271 } 5272 5273 public: 5274 RAUWUpdateListener(SelectionDAG::DAGUpdateListener *dl, 5275 SDNode::use_iterator &ui, 5276 SDNode::use_iterator &ue) 5277 : DownLink(dl), UI(ui), UE(ue) {} 5278 }; 5279 5280 } 5281 5282 /// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead. 5283 /// This can cause recursive merging of nodes in the DAG. 5284 /// 5285 /// This version assumes From has a single result value. 5286 /// 5287 void SelectionDAG::ReplaceAllUsesWith(SDValue FromN, SDValue To, 5288 DAGUpdateListener *UpdateListener) { 5289 SDNode *From = FromN.getNode(); 5290 assert(From->getNumValues() == 1 && FromN.getResNo() == 0 && 5291 "Cannot replace with this method!"); 5292 assert(From != To.getNode() && "Cannot replace uses of with self"); 5293 5294 // Iterate over all the existing uses of From. New uses will be added 5295 // to the beginning of the use list, which we avoid visiting. 5296 // This specifically avoids visiting uses of From that arise while the 5297 // replacement is happening, because any such uses would be the result 5298 // of CSE: If an existing node looks like From after one of its operands 5299 // is replaced by To, we don't want to replace of all its users with To 5300 // too. See PR3018 for more info. 5301 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end(); 5302 RAUWUpdateListener Listener(UpdateListener, UI, UE); 5303 while (UI != UE) { 5304 SDNode *User = *UI; 5305 5306 // This node is about to morph, remove its old self from the CSE maps. 5307 RemoveNodeFromCSEMaps(User); 5308 5309 // A user can appear in a use list multiple times, and when this 5310 // happens the uses are usually next to each other in the list. 5311 // To help reduce the number of CSE recomputations, process all 5312 // the uses of this user that we can find this way. 5313 do { 5314 SDUse &Use = UI.getUse(); 5315 ++UI; 5316 Use.set(To); 5317 } while (UI != UE && *UI == User); 5318 5319 // Now that we have modified User, add it back to the CSE maps. If it 5320 // already exists there, recursively merge the results together. 5321 AddModifiedNodeToCSEMaps(User, &Listener); 5322 } 5323 5324 // If we just RAUW'd the root, take note. 5325 if (FromN == getRoot()) 5326 setRoot(To); 5327 } 5328 5329 /// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead. 5330 /// This can cause recursive merging of nodes in the DAG. 5331 /// 5332 /// This version assumes that for each value of From, there is a 5333 /// corresponding value in To in the same position with the same type. 5334 /// 5335 void SelectionDAG::ReplaceAllUsesWith(SDNode *From, SDNode *To, 5336 DAGUpdateListener *UpdateListener) { 5337 #ifndef NDEBUG 5338 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i) 5339 assert((!From->hasAnyUseOfValue(i) || 5340 From->getValueType(i) == To->getValueType(i)) && 5341 "Cannot use this version of ReplaceAllUsesWith!"); 5342 #endif 5343 5344 // Handle the trivial case. 5345 if (From == To) 5346 return; 5347 5348 // Iterate over just the existing users of From. See the comments in 5349 // the ReplaceAllUsesWith above. 5350 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end(); 5351 RAUWUpdateListener Listener(UpdateListener, UI, UE); 5352 while (UI != UE) { 5353 SDNode *User = *UI; 5354 5355 // This node is about to morph, remove its old self from the CSE maps. 5356 RemoveNodeFromCSEMaps(User); 5357 5358 // A user can appear in a use list multiple times, and when this 5359 // happens the uses are usually next to each other in the list. 5360 // To help reduce the number of CSE recomputations, process all 5361 // the uses of this user that we can find this way. 5362 do { 5363 SDUse &Use = UI.getUse(); 5364 ++UI; 5365 Use.setNode(To); 5366 } while (UI != UE && *UI == User); 5367 5368 // Now that we have modified User, add it back to the CSE maps. If it 5369 // already exists there, recursively merge the results together. 5370 AddModifiedNodeToCSEMaps(User, &Listener); 5371 } 5372 5373 // If we just RAUW'd the root, take note. 5374 if (From == getRoot().getNode()) 5375 setRoot(SDValue(To, getRoot().getResNo())); 5376 } 5377 5378 /// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead. 5379 /// This can cause recursive merging of nodes in the DAG. 5380 /// 5381 /// This version can replace From with any result values. To must match the 5382 /// number and types of values returned by From. 5383 void SelectionDAG::ReplaceAllUsesWith(SDNode *From, 5384 const SDValue *To, 5385 DAGUpdateListener *UpdateListener) { 5386 if (From->getNumValues() == 1) // Handle the simple case efficiently. 5387 return ReplaceAllUsesWith(SDValue(From, 0), To[0], UpdateListener); 5388 5389 // Iterate over just the existing users of From. See the comments in 5390 // the ReplaceAllUsesWith above. 5391 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end(); 5392 RAUWUpdateListener Listener(UpdateListener, UI, UE); 5393 while (UI != UE) { 5394 SDNode *User = *UI; 5395 5396 // This node is about to morph, remove its old self from the CSE maps. 5397 RemoveNodeFromCSEMaps(User); 5398 5399 // A user can appear in a use list multiple times, and when this 5400 // happens the uses are usually next to each other in the list. 5401 // To help reduce the number of CSE recomputations, process all 5402 // the uses of this user that we can find this way. 5403 do { 5404 SDUse &Use = UI.getUse(); 5405 const SDValue &ToOp = To[Use.getResNo()]; 5406 ++UI; 5407 Use.set(ToOp); 5408 } while (UI != UE && *UI == User); 5409 5410 // Now that we have modified User, add it back to the CSE maps. If it 5411 // already exists there, recursively merge the results together. 5412 AddModifiedNodeToCSEMaps(User, &Listener); 5413 } 5414 5415 // If we just RAUW'd the root, take note. 5416 if (From == getRoot().getNode()) 5417 setRoot(SDValue(To[getRoot().getResNo()])); 5418 } 5419 5420 /// ReplaceAllUsesOfValueWith - Replace any uses of From with To, leaving 5421 /// uses of other values produced by From.getNode() alone. The Deleted 5422 /// vector is handled the same way as for ReplaceAllUsesWith. 5423 void SelectionDAG::ReplaceAllUsesOfValueWith(SDValue From, SDValue To, 5424 DAGUpdateListener *UpdateListener){ 5425 // Handle the really simple, really trivial case efficiently. 5426 if (From == To) return; 5427 5428 // Handle the simple, trivial, case efficiently. 5429 if (From.getNode()->getNumValues() == 1) { 5430 ReplaceAllUsesWith(From, To, UpdateListener); 5431 return; 5432 } 5433 5434 // Iterate over just the existing users of From. See the comments in 5435 // the ReplaceAllUsesWith above. 5436 SDNode::use_iterator UI = From.getNode()->use_begin(), 5437 UE = From.getNode()->use_end(); 5438 RAUWUpdateListener Listener(UpdateListener, UI, UE); 5439 while (UI != UE) { 5440 SDNode *User = *UI; 5441 bool UserRemovedFromCSEMaps = false; 5442 5443 // A user can appear in a use list multiple times, and when this 5444 // happens the uses are usually next to each other in the list. 5445 // To help reduce the number of CSE recomputations, process all 5446 // the uses of this user that we can find this way. 5447 do { 5448 SDUse &Use = UI.getUse(); 5449 5450 // Skip uses of different values from the same node. 5451 if (Use.getResNo() != From.getResNo()) { 5452 ++UI; 5453 continue; 5454 } 5455 5456 // If this node hasn't been modified yet, it's still in the CSE maps, 5457 // so remove its old self from the CSE maps. 5458 if (!UserRemovedFromCSEMaps) { 5459 RemoveNodeFromCSEMaps(User); 5460 UserRemovedFromCSEMaps = true; 5461 } 5462 5463 ++UI; 5464 Use.set(To); 5465 } while (UI != UE && *UI == User); 5466 5467 // We are iterating over all uses of the From node, so if a use 5468 // doesn't use the specific value, no changes are made. 5469 if (!UserRemovedFromCSEMaps) 5470 continue; 5471 5472 // Now that we have modified User, add it back to the CSE maps. If it 5473 // already exists there, recursively merge the results together. 5474 AddModifiedNodeToCSEMaps(User, &Listener); 5475 } 5476 5477 // If we just RAUW'd the root, take note. 5478 if (From == getRoot()) 5479 setRoot(To); 5480 } 5481 5482 namespace { 5483 /// UseMemo - This class is used by SelectionDAG::ReplaceAllUsesOfValuesWith 5484 /// to record information about a use. 5485 struct UseMemo { 5486 SDNode *User; 5487 unsigned Index; 5488 SDUse *Use; 5489 }; 5490 5491 /// operator< - Sort Memos by User. 5492 bool operator<(const UseMemo &L, const UseMemo &R) { 5493 return (intptr_t)L.User < (intptr_t)R.User; 5494 } 5495 } 5496 5497 /// ReplaceAllUsesOfValuesWith - Replace any uses of From with To, leaving 5498 /// uses of other values produced by From.getNode() alone. The same value 5499 /// may appear in both the From and To list. The Deleted vector is 5500 /// handled the same way as for ReplaceAllUsesWith. 5501 void SelectionDAG::ReplaceAllUsesOfValuesWith(const SDValue *From, 5502 const SDValue *To, 5503 unsigned Num, 5504 DAGUpdateListener *UpdateListener){ 5505 // Handle the simple, trivial case efficiently. 5506 if (Num == 1) 5507 return ReplaceAllUsesOfValueWith(*From, *To, UpdateListener); 5508 5509 // Read up all the uses and make records of them. This helps 5510 // processing new uses that are introduced during the 5511 // replacement process. 5512 SmallVector<UseMemo, 4> Uses; 5513 for (unsigned i = 0; i != Num; ++i) { 5514 unsigned FromResNo = From[i].getResNo(); 5515 SDNode *FromNode = From[i].getNode(); 5516 for (SDNode::use_iterator UI = FromNode->use_begin(), 5517 E = FromNode->use_end(); UI != E; ++UI) { 5518 SDUse &Use = UI.getUse(); 5519 if (Use.getResNo() == FromResNo) { 5520 UseMemo Memo = { *UI, i, &Use }; 5521 Uses.push_back(Memo); 5522 } 5523 } 5524 } 5525 5526 // Sort the uses, so that all the uses from a given User are together. 5527 std::sort(Uses.begin(), Uses.end()); 5528 5529 for (unsigned UseIndex = 0, UseIndexEnd = Uses.size(); 5530 UseIndex != UseIndexEnd; ) { 5531 // We know that this user uses some value of From. If it is the right 5532 // value, update it. 5533 SDNode *User = Uses[UseIndex].User; 5534 5535 // This node is about to morph, remove its old self from the CSE maps. 5536 RemoveNodeFromCSEMaps(User); 5537 5538 // The Uses array is sorted, so all the uses for a given User 5539 // are next to each other in the list. 5540 // To help reduce the number of CSE recomputations, process all 5541 // the uses of this user that we can find this way. 5542 do { 5543 unsigned i = Uses[UseIndex].Index; 5544 SDUse &Use = *Uses[UseIndex].Use; 5545 ++UseIndex; 5546 5547 Use.set(To[i]); 5548 } while (UseIndex != UseIndexEnd && Uses[UseIndex].User == User); 5549 5550 // Now that we have modified User, add it back to the CSE maps. If it 5551 // already exists there, recursively merge the results together. 5552 AddModifiedNodeToCSEMaps(User, UpdateListener); 5553 } 5554 } 5555 5556 /// AssignTopologicalOrder - Assign a unique node id for each node in the DAG 5557 /// based on their topological order. It returns the maximum id and a vector 5558 /// of the SDNodes* in assigned order by reference. 5559 unsigned SelectionDAG::AssignTopologicalOrder() { 5560 5561 unsigned DAGSize = 0; 5562 5563 // SortedPos tracks the progress of the algorithm. Nodes before it are 5564 // sorted, nodes after it are unsorted. When the algorithm completes 5565 // it is at the end of the list. 5566 allnodes_iterator SortedPos = allnodes_begin(); 5567 5568 // Visit all the nodes. Move nodes with no operands to the front of 5569 // the list immediately. Annotate nodes that do have operands with their 5570 // operand count. Before we do this, the Node Id fields of the nodes 5571 // may contain arbitrary values. After, the Node Id fields for nodes 5572 // before SortedPos will contain the topological sort index, and the 5573 // Node Id fields for nodes At SortedPos and after will contain the 5574 // count of outstanding operands. 5575 for (allnodes_iterator I = allnodes_begin(),E = allnodes_end(); I != E; ) { 5576 SDNode *N = I++; 5577 checkForCycles(N); 5578 unsigned Degree = N->getNumOperands(); 5579 if (Degree == 0) { 5580 // A node with no uses, add it to the result array immediately. 5581 N->setNodeId(DAGSize++); 5582 allnodes_iterator Q = N; 5583 if (Q != SortedPos) 5584 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(Q)); 5585 assert(SortedPos != AllNodes.end() && "Overran node list"); 5586 ++SortedPos; 5587 } else { 5588 // Temporarily use the Node Id as scratch space for the degree count. 5589 N->setNodeId(Degree); 5590 } 5591 } 5592 5593 // Visit all the nodes. As we iterate, moves nodes into sorted order, 5594 // such that by the time the end is reached all nodes will be sorted. 5595 for (allnodes_iterator I = allnodes_begin(),E = allnodes_end(); I != E; ++I) { 5596 SDNode *N = I; 5597 checkForCycles(N); 5598 // N is in sorted position, so all its uses have one less operand 5599 // that needs to be sorted. 5600 for (SDNode::use_iterator UI = N->use_begin(), UE = N->use_end(); 5601 UI != UE; ++UI) { 5602 SDNode *P = *UI; 5603 unsigned Degree = P->getNodeId(); 5604 assert(Degree != 0 && "Invalid node degree"); 5605 --Degree; 5606 if (Degree == 0) { 5607 // All of P's operands are sorted, so P may sorted now. 5608 P->setNodeId(DAGSize++); 5609 if (P != SortedPos) 5610 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(P)); 5611 assert(SortedPos != AllNodes.end() && "Overran node list"); 5612 ++SortedPos; 5613 } else { 5614 // Update P's outstanding operand count. 5615 P->setNodeId(Degree); 5616 } 5617 } 5618 if (I == SortedPos) { 5619 #ifndef NDEBUG 5620 SDNode *S = ++I; 5621 dbgs() << "Overran sorted position:\n"; 5622 S->dumprFull(); 5623 #endif 5624 llvm_unreachable(0); 5625 } 5626 } 5627 5628 assert(SortedPos == AllNodes.end() && 5629 "Topological sort incomplete!"); 5630 assert(AllNodes.front().getOpcode() == ISD::EntryToken && 5631 "First node in topological sort is not the entry token!"); 5632 assert(AllNodes.front().getNodeId() == 0 && 5633 "First node in topological sort has non-zero id!"); 5634 assert(AllNodes.front().getNumOperands() == 0 && 5635 "First node in topological sort has operands!"); 5636 assert(AllNodes.back().getNodeId() == (int)DAGSize-1 && 5637 "Last node in topologic sort has unexpected id!"); 5638 assert(AllNodes.back().use_empty() && 5639 "Last node in topologic sort has users!"); 5640 assert(DAGSize == allnodes_size() && "Node count mismatch!"); 5641 return DAGSize; 5642 } 5643 5644 /// AssignOrdering - Assign an order to the SDNode. 5645 void SelectionDAG::AssignOrdering(const SDNode *SD, unsigned Order) { 5646 assert(SD && "Trying to assign an order to a null node!"); 5647 Ordering->add(SD, Order); 5648 } 5649 5650 /// GetOrdering - Get the order for the SDNode. 5651 unsigned SelectionDAG::GetOrdering(const SDNode *SD) const { 5652 assert(SD && "Trying to get the order of a null node!"); 5653 return Ordering->getOrder(SD); 5654 } 5655 5656 /// AddDbgValue - Add a dbg_value SDNode. If SD is non-null that means the 5657 /// value is produced by SD. 5658 void SelectionDAG::AddDbgValue(SDDbgValue *DB, SDNode *SD, bool isParameter) { 5659 DbgInfo->add(DB, SD, isParameter); 5660 if (SD) 5661 SD->setHasDebugValue(true); 5662 } 5663 5664 /// TransferDbgValues - Transfer SDDbgValues. 5665 void SelectionDAG::TransferDbgValues(SDValue From, SDValue To) { 5666 if (From == To || !From.getNode()->getHasDebugValue()) 5667 return; 5668 SDNode *FromNode = From.getNode(); 5669 SDNode *ToNode = To.getNode(); 5670 ArrayRef<SDDbgValue *> DVs = GetDbgValues(FromNode); 5671 SmallVector<SDDbgValue *, 2> ClonedDVs; 5672 for (ArrayRef<SDDbgValue *>::iterator I = DVs.begin(), E = DVs.end(); 5673 I != E; ++I) { 5674 SDDbgValue *Dbg = *I; 5675 if (Dbg->getKind() == SDDbgValue::SDNODE) { 5676 SDDbgValue *Clone = getDbgValue(Dbg->getMDPtr(), ToNode, To.getResNo(), 5677 Dbg->getOffset(), Dbg->getDebugLoc(), 5678 Dbg->getOrder()); 5679 ClonedDVs.push_back(Clone); 5680 } 5681 } 5682 for (SmallVector<SDDbgValue *, 2>::iterator I = ClonedDVs.begin(), 5683 E = ClonedDVs.end(); I != E; ++I) 5684 AddDbgValue(*I, ToNode, false); 5685 } 5686 5687 //===----------------------------------------------------------------------===// 5688 // SDNode Class 5689 //===----------------------------------------------------------------------===// 5690 5691 HandleSDNode::~HandleSDNode() { 5692 DropOperands(); 5693 } 5694 5695 GlobalAddressSDNode::GlobalAddressSDNode(unsigned Opc, DebugLoc DL, 5696 const GlobalValue *GA, 5697 EVT VT, int64_t o, unsigned char TF) 5698 : SDNode(Opc, DL, getSDVTList(VT)), Offset(o), TargetFlags(TF) { 5699 TheGlobal = GA; 5700 } 5701 5702 MemSDNode::MemSDNode(unsigned Opc, DebugLoc dl, SDVTList VTs, EVT memvt, 5703 MachineMemOperand *mmo) 5704 : SDNode(Opc, dl, VTs), MemoryVT(memvt), MMO(mmo) { 5705 SubclassData = encodeMemSDNodeFlags(0, ISD::UNINDEXED, MMO->isVolatile(), 5706 MMO->isNonTemporal(), MMO->isInvariant()); 5707 assert(isVolatile() == MMO->isVolatile() && "Volatile encoding error!"); 5708 assert(isNonTemporal() == MMO->isNonTemporal() && 5709 "Non-temporal encoding error!"); 5710 assert(memvt.getStoreSize() == MMO->getSize() && "Size mismatch!"); 5711 } 5712 5713 MemSDNode::MemSDNode(unsigned Opc, DebugLoc dl, SDVTList VTs, 5714 const SDValue *Ops, unsigned NumOps, EVT memvt, 5715 MachineMemOperand *mmo) 5716 : SDNode(Opc, dl, VTs, Ops, NumOps), 5717 MemoryVT(memvt), MMO(mmo) { 5718 SubclassData = encodeMemSDNodeFlags(0, ISD::UNINDEXED, MMO->isVolatile(), 5719 MMO->isNonTemporal(), MMO->isInvariant()); 5720 assert(isVolatile() == MMO->isVolatile() && "Volatile encoding error!"); 5721 assert(memvt.getStoreSize() == MMO->getSize() && "Size mismatch!"); 5722 } 5723 5724 /// Profile - Gather unique data for the node. 5725 /// 5726 void SDNode::Profile(FoldingSetNodeID &ID) const { 5727 AddNodeIDNode(ID, this); 5728 } 5729 5730 namespace { 5731 struct EVTArray { 5732 std::vector<EVT> VTs; 5733 5734 EVTArray() { 5735 VTs.reserve(MVT::LAST_VALUETYPE); 5736 for (unsigned i = 0; i < MVT::LAST_VALUETYPE; ++i) 5737 VTs.push_back(MVT((MVT::SimpleValueType)i)); 5738 } 5739 }; 5740 } 5741 5742 static ManagedStatic<std::set<EVT, EVT::compareRawBits> > EVTs; 5743 static ManagedStatic<EVTArray> SimpleVTArray; 5744 static ManagedStatic<sys::SmartMutex<true> > VTMutex; 5745 5746 /// getValueTypeList - Return a pointer to the specified value type. 5747 /// 5748 const EVT *SDNode::getValueTypeList(EVT VT) { 5749 if (VT.isExtended()) { 5750 sys::SmartScopedLock<true> Lock(*VTMutex); 5751 return &(*EVTs->insert(VT).first); 5752 } else { 5753 assert(VT.getSimpleVT() < MVT::LAST_VALUETYPE && 5754 "Value type out of range!"); 5755 return &SimpleVTArray->VTs[VT.getSimpleVT().SimpleTy]; 5756 } 5757 } 5758 5759 /// hasNUsesOfValue - Return true if there are exactly NUSES uses of the 5760 /// indicated value. This method ignores uses of other values defined by this 5761 /// operation. 5762 bool SDNode::hasNUsesOfValue(unsigned NUses, unsigned Value) const { 5763 assert(Value < getNumValues() && "Bad value!"); 5764 5765 // TODO: Only iterate over uses of a given value of the node 5766 for (SDNode::use_iterator UI = use_begin(), E = use_end(); UI != E; ++UI) { 5767 if (UI.getUse().getResNo() == Value) { 5768 if (NUses == 0) 5769 return false; 5770 --NUses; 5771 } 5772 } 5773 5774 // Found exactly the right number of uses? 5775 return NUses == 0; 5776 } 5777 5778 5779 /// hasAnyUseOfValue - Return true if there are any use of the indicated 5780 /// value. This method ignores uses of other values defined by this operation. 5781 bool SDNode::hasAnyUseOfValue(unsigned Value) const { 5782 assert(Value < getNumValues() && "Bad value!"); 5783 5784 for (SDNode::use_iterator UI = use_begin(), E = use_end(); UI != E; ++UI) 5785 if (UI.getUse().getResNo() == Value) 5786 return true; 5787 5788 return false; 5789 } 5790 5791 5792 /// isOnlyUserOf - Return true if this node is the only use of N. 5793 /// 5794 bool SDNode::isOnlyUserOf(SDNode *N) const { 5795 bool Seen = false; 5796 for (SDNode::use_iterator I = N->use_begin(), E = N->use_end(); I != E; ++I) { 5797 SDNode *User = *I; 5798 if (User == this) 5799 Seen = true; 5800 else 5801 return false; 5802 } 5803 5804 return Seen; 5805 } 5806 5807 /// isOperand - Return true if this node is an operand of N. 5808 /// 5809 bool SDValue::isOperandOf(SDNode *N) const { 5810 for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i) 5811 if (*this == N->getOperand(i)) 5812 return true; 5813 return false; 5814 } 5815 5816 bool SDNode::isOperandOf(SDNode *N) const { 5817 for (unsigned i = 0, e = N->NumOperands; i != e; ++i) 5818 if (this == N->OperandList[i].getNode()) 5819 return true; 5820 return false; 5821 } 5822 5823 /// reachesChainWithoutSideEffects - Return true if this operand (which must 5824 /// be a chain) reaches the specified operand without crossing any 5825 /// side-effecting instructions on any chain path. In practice, this looks 5826 /// through token factors and non-volatile loads. In order to remain efficient, 5827 /// this only looks a couple of nodes in, it does not do an exhaustive search. 5828 bool SDValue::reachesChainWithoutSideEffects(SDValue Dest, 5829 unsigned Depth) const { 5830 if (*this == Dest) return true; 5831 5832 // Don't search too deeply, we just want to be able to see through 5833 // TokenFactor's etc. 5834 if (Depth == 0) return false; 5835 5836 // If this is a token factor, all inputs to the TF happen in parallel. If any 5837 // of the operands of the TF does not reach dest, then we cannot do the xform. 5838 if (getOpcode() == ISD::TokenFactor) { 5839 for (unsigned i = 0, e = getNumOperands(); i != e; ++i) 5840 if (!getOperand(i).reachesChainWithoutSideEffects(Dest, Depth-1)) 5841 return false; 5842 return true; 5843 } 5844 5845 // Loads don't have side effects, look through them. 5846 if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(*this)) { 5847 if (!Ld->isVolatile()) 5848 return Ld->getChain().reachesChainWithoutSideEffects(Dest, Depth-1); 5849 } 5850 return false; 5851 } 5852 5853 /// hasPredecessor - Return true if N is a predecessor of this node. 5854 /// N is either an operand of this node, or can be reached by recursively 5855 /// traversing up the operands. 5856 /// NOTE: This is an expensive method. Use it carefully. 5857 bool SDNode::hasPredecessor(const SDNode *N) const { 5858 SmallPtrSet<const SDNode *, 32> Visited; 5859 SmallVector<const SDNode *, 16> Worklist; 5860 return hasPredecessorHelper(N, Visited, Worklist); 5861 } 5862 5863 bool SDNode::hasPredecessorHelper(const SDNode *N, 5864 SmallPtrSet<const SDNode *, 32> &Visited, 5865 SmallVector<const SDNode *, 16> &Worklist) const { 5866 if (Visited.empty()) { 5867 Worklist.push_back(this); 5868 } else { 5869 // Take a look in the visited set. If we've already encountered this node 5870 // we needn't search further. 5871 if (Visited.count(N)) 5872 return true; 5873 } 5874 5875 // Haven't visited N yet. Continue the search. 5876 while (!Worklist.empty()) { 5877 const SDNode *M = Worklist.pop_back_val(); 5878 for (unsigned i = 0, e = M->getNumOperands(); i != e; ++i) { 5879 SDNode *Op = M->getOperand(i).getNode(); 5880 if (Visited.insert(Op)) 5881 Worklist.push_back(Op); 5882 if (Op == N) 5883 return true; 5884 } 5885 } 5886 5887 return false; 5888 } 5889 5890 uint64_t SDNode::getConstantOperandVal(unsigned Num) const { 5891 assert(Num < NumOperands && "Invalid child # of SDNode!"); 5892 return cast<ConstantSDNode>(OperandList[Num])->getZExtValue(); 5893 } 5894 5895 std::string SDNode::getOperationName(const SelectionDAG *G) const { 5896 switch (getOpcode()) { 5897 default: 5898 if (getOpcode() < ISD::BUILTIN_OP_END) 5899 return "<<Unknown DAG Node>>"; 5900 if (isMachineOpcode()) { 5901 if (G) 5902 if (const TargetInstrInfo *TII = G->getTarget().getInstrInfo()) 5903 if (getMachineOpcode() < TII->getNumOpcodes()) 5904 return TII->get(getMachineOpcode()).getName(); 5905 return "<<Unknown Machine Node #" + utostr(getOpcode()) + ">>"; 5906 } 5907 if (G) { 5908 const TargetLowering &TLI = G->getTargetLoweringInfo(); 5909 const char *Name = TLI.getTargetNodeName(getOpcode()); 5910 if (Name) return Name; 5911 return "<<Unknown Target Node #" + utostr(getOpcode()) + ">>"; 5912 } 5913 return "<<Unknown Node #" + utostr(getOpcode()) + ">>"; 5914 5915 #ifndef NDEBUG 5916 case ISD::DELETED_NODE: 5917 return "<<Deleted Node!>>"; 5918 #endif 5919 case ISD::PREFETCH: return "Prefetch"; 5920 case ISD::MEMBARRIER: return "MemBarrier"; 5921 case ISD::ATOMIC_FENCE: return "AtomicFence"; 5922 case ISD::ATOMIC_CMP_SWAP: return "AtomicCmpSwap"; 5923 case ISD::ATOMIC_SWAP: return "AtomicSwap"; 5924 case ISD::ATOMIC_LOAD_ADD: return "AtomicLoadAdd"; 5925 case ISD::ATOMIC_LOAD_SUB: return "AtomicLoadSub"; 5926 case ISD::ATOMIC_LOAD_AND: return "AtomicLoadAnd"; 5927 case ISD::ATOMIC_LOAD_OR: return "AtomicLoadOr"; 5928 case ISD::ATOMIC_LOAD_XOR: return "AtomicLoadXor"; 5929 case ISD::ATOMIC_LOAD_NAND: return "AtomicLoadNand"; 5930 case ISD::ATOMIC_LOAD_MIN: return "AtomicLoadMin"; 5931 case ISD::ATOMIC_LOAD_MAX: return "AtomicLoadMax"; 5932 case ISD::ATOMIC_LOAD_UMIN: return "AtomicLoadUMin"; 5933 case ISD::ATOMIC_LOAD_UMAX: return "AtomicLoadUMax"; 5934 case ISD::ATOMIC_LOAD: return "AtomicLoad"; 5935 case ISD::ATOMIC_STORE: return "AtomicStore"; 5936 case ISD::PCMARKER: return "PCMarker"; 5937 case ISD::READCYCLECOUNTER: return "ReadCycleCounter"; 5938 case ISD::SRCVALUE: return "SrcValue"; 5939 case ISD::MDNODE_SDNODE: return "MDNode"; 5940 case ISD::EntryToken: return "EntryToken"; 5941 case ISD::TokenFactor: return "TokenFactor"; 5942 case ISD::AssertSext: return "AssertSext"; 5943 case ISD::AssertZext: return "AssertZext"; 5944 5945 case ISD::BasicBlock: return "BasicBlock"; 5946 case ISD::VALUETYPE: return "ValueType"; 5947 case ISD::Register: return "Register"; 5948 5949 case ISD::Constant: return "Constant"; 5950 case ISD::ConstantFP: return "ConstantFP"; 5951 case ISD::GlobalAddress: return "GlobalAddress"; 5952 case ISD::GlobalTLSAddress: return "GlobalTLSAddress"; 5953 case ISD::FrameIndex: return "FrameIndex"; 5954 case ISD::JumpTable: return "JumpTable"; 5955 case ISD::GLOBAL_OFFSET_TABLE: return "GLOBAL_OFFSET_TABLE"; 5956 case ISD::RETURNADDR: return "RETURNADDR"; 5957 case ISD::FRAMEADDR: return "FRAMEADDR"; 5958 case ISD::FRAME_TO_ARGS_OFFSET: return "FRAME_TO_ARGS_OFFSET"; 5959 case ISD::EXCEPTIONADDR: return "EXCEPTIONADDR"; 5960 case ISD::LSDAADDR: return "LSDAADDR"; 5961 case ISD::EHSELECTION: return "EHSELECTION"; 5962 case ISD::EH_RETURN: return "EH_RETURN"; 5963 case ISD::EH_SJLJ_SETJMP: return "EH_SJLJ_SETJMP"; 5964 case ISD::EH_SJLJ_LONGJMP: return "EH_SJLJ_LONGJMP"; 5965 case ISD::ConstantPool: return "ConstantPool"; 5966 case ISD::ExternalSymbol: return "ExternalSymbol"; 5967 case ISD::BlockAddress: return "BlockAddress"; 5968 case ISD::INTRINSIC_WO_CHAIN: 5969 case ISD::INTRINSIC_VOID: 5970 case ISD::INTRINSIC_W_CHAIN: { 5971 unsigned OpNo = getOpcode() == ISD::INTRINSIC_WO_CHAIN ? 0 : 1; 5972 unsigned IID = cast<ConstantSDNode>(getOperand(OpNo))->getZExtValue(); 5973 if (IID < Intrinsic::num_intrinsics) 5974 return Intrinsic::getName((Intrinsic::ID)IID); 5975 else if (const TargetIntrinsicInfo *TII = G->getTarget().getIntrinsicInfo()) 5976 return TII->getName(IID); 5977 llvm_unreachable("Invalid intrinsic ID"); 5978 } 5979 5980 case ISD::BUILD_VECTOR: return "BUILD_VECTOR"; 5981 case ISD::TargetConstant: return "TargetConstant"; 5982 case ISD::TargetConstantFP:return "TargetConstantFP"; 5983 case ISD::TargetGlobalAddress: return "TargetGlobalAddress"; 5984 case ISD::TargetGlobalTLSAddress: return "TargetGlobalTLSAddress"; 5985 case ISD::TargetFrameIndex: return "TargetFrameIndex"; 5986 case ISD::TargetJumpTable: return "TargetJumpTable"; 5987 case ISD::TargetConstantPool: return "TargetConstantPool"; 5988 case ISD::TargetExternalSymbol: return "TargetExternalSymbol"; 5989 case ISD::TargetBlockAddress: return "TargetBlockAddress"; 5990 5991 case ISD::CopyToReg: return "CopyToReg"; 5992 case ISD::CopyFromReg: return "CopyFromReg"; 5993 case ISD::UNDEF: return "undef"; 5994 case ISD::MERGE_VALUES: return "merge_values"; 5995 case ISD::INLINEASM: return "inlineasm"; 5996 case ISD::EH_LABEL: return "eh_label"; 5997 case ISD::HANDLENODE: return "handlenode"; 5998 5999 // Unary operators 6000 case ISD::FABS: return "fabs"; 6001 case ISD::FNEG: return "fneg"; 6002 case ISD::FSQRT: return "fsqrt"; 6003 case ISD::FSIN: return "fsin"; 6004 case ISD::FCOS: return "fcos"; 6005 case ISD::FTRUNC: return "ftrunc"; 6006 case ISD::FFLOOR: return "ffloor"; 6007 case ISD::FCEIL: return "fceil"; 6008 case ISD::FRINT: return "frint"; 6009 case ISD::FNEARBYINT: return "fnearbyint"; 6010 case ISD::FEXP: return "fexp"; 6011 case ISD::FEXP2: return "fexp2"; 6012 case ISD::FLOG: return "flog"; 6013 case ISD::FLOG2: return "flog2"; 6014 case ISD::FLOG10: return "flog10"; 6015 6016 // Binary operators 6017 case ISD::ADD: return "add"; 6018 case ISD::SUB: return "sub"; 6019 case ISD::MUL: return "mul"; 6020 case ISD::MULHU: return "mulhu"; 6021 case ISD::MULHS: return "mulhs"; 6022 case ISD::SDIV: return "sdiv"; 6023 case ISD::UDIV: return "udiv"; 6024 case ISD::SREM: return "srem"; 6025 case ISD::UREM: return "urem"; 6026 case ISD::SMUL_LOHI: return "smul_lohi"; 6027 case ISD::UMUL_LOHI: return "umul_lohi"; 6028 case ISD::SDIVREM: return "sdivrem"; 6029 case ISD::UDIVREM: return "udivrem"; 6030 case ISD::AND: return "and"; 6031 case ISD::OR: return "or"; 6032 case ISD::XOR: return "xor"; 6033 case ISD::SHL: return "shl"; 6034 case ISD::SRA: return "sra"; 6035 case ISD::SRL: return "srl"; 6036 case ISD::ROTL: return "rotl"; 6037 case ISD::ROTR: return "rotr"; 6038 case ISD::FADD: return "fadd"; 6039 case ISD::FSUB: return "fsub"; 6040 case ISD::FMUL: return "fmul"; 6041 case ISD::FDIV: return "fdiv"; 6042 case ISD::FMA: return "fma"; 6043 case ISD::FREM: return "frem"; 6044 case ISD::FCOPYSIGN: return "fcopysign"; 6045 case ISD::FGETSIGN: return "fgetsign"; 6046 case ISD::FPOW: return "fpow"; 6047 6048 case ISD::FPOWI: return "fpowi"; 6049 case ISD::SETCC: return "setcc"; 6050 case ISD::SELECT: return "select"; 6051 case ISD::VSELECT: return "vselect"; 6052 case ISD::SELECT_CC: return "select_cc"; 6053 case ISD::INSERT_VECTOR_ELT: return "insert_vector_elt"; 6054 case ISD::EXTRACT_VECTOR_ELT: return "extract_vector_elt"; 6055 case ISD::CONCAT_VECTORS: return "concat_vectors"; 6056 case ISD::INSERT_SUBVECTOR: return "insert_subvector"; 6057 case ISD::EXTRACT_SUBVECTOR: return "extract_subvector"; 6058 case ISD::SCALAR_TO_VECTOR: return "scalar_to_vector"; 6059 case ISD::VECTOR_SHUFFLE: return "vector_shuffle"; 6060 case ISD::CARRY_FALSE: return "carry_false"; 6061 case ISD::ADDC: return "addc"; 6062 case ISD::ADDE: return "adde"; 6063 case ISD::SADDO: return "saddo"; 6064 case ISD::UADDO: return "uaddo"; 6065 case ISD::SSUBO: return "ssubo"; 6066 case ISD::USUBO: return "usubo"; 6067 case ISD::SMULO: return "smulo"; 6068 case ISD::UMULO: return "umulo"; 6069 case ISD::SUBC: return "subc"; 6070 case ISD::SUBE: return "sube"; 6071 case ISD::SHL_PARTS: return "shl_parts"; 6072 case ISD::SRA_PARTS: return "sra_parts"; 6073 case ISD::SRL_PARTS: return "srl_parts"; 6074 6075 // Conversion operators. 6076 case ISD::SIGN_EXTEND: return "sign_extend"; 6077 case ISD::ZERO_EXTEND: return "zero_extend"; 6078 case ISD::ANY_EXTEND: return "any_extend"; 6079 case ISD::SIGN_EXTEND_INREG: return "sign_extend_inreg"; 6080 case ISD::TRUNCATE: return "truncate"; 6081 case ISD::FP_ROUND: return "fp_round"; 6082 case ISD::FLT_ROUNDS_: return "flt_rounds"; 6083 case ISD::FP_ROUND_INREG: return "fp_round_inreg"; 6084 case ISD::FP_EXTEND: return "fp_extend"; 6085 6086 case ISD::SINT_TO_FP: return "sint_to_fp"; 6087 case ISD::UINT_TO_FP: return "uint_to_fp"; 6088 case ISD::FP_TO_SINT: return "fp_to_sint"; 6089 case ISD::FP_TO_UINT: return "fp_to_uint"; 6090 case ISD::BITCAST: return "bitcast"; 6091 case ISD::FP16_TO_FP32: return "fp16_to_fp32"; 6092 case ISD::FP32_TO_FP16: return "fp32_to_fp16"; 6093 6094 case ISD::CONVERT_RNDSAT: { 6095 switch (cast<CvtRndSatSDNode>(this)->getCvtCode()) { 6096 default: llvm_unreachable("Unknown cvt code!"); 6097 case ISD::CVT_FF: return "cvt_ff"; 6098 case ISD::CVT_FS: return "cvt_fs"; 6099 case ISD::CVT_FU: return "cvt_fu"; 6100 case ISD::CVT_SF: return "cvt_sf"; 6101 case ISD::CVT_UF: return "cvt_uf"; 6102 case ISD::CVT_SS: return "cvt_ss"; 6103 case ISD::CVT_SU: return "cvt_su"; 6104 case ISD::CVT_US: return "cvt_us"; 6105 case ISD::CVT_UU: return "cvt_uu"; 6106 } 6107 } 6108 6109 // Control flow instructions 6110 case ISD::BR: return "br"; 6111 case ISD::BRIND: return "brind"; 6112 case ISD::BR_JT: return "br_jt"; 6113 case ISD::BRCOND: return "brcond"; 6114 case ISD::BR_CC: return "br_cc"; 6115 case ISD::CALLSEQ_START: return "callseq_start"; 6116 case ISD::CALLSEQ_END: return "callseq_end"; 6117 6118 // Other operators 6119 case ISD::LOAD: return "load"; 6120 case ISD::STORE: return "store"; 6121 case ISD::VAARG: return "vaarg"; 6122 case ISD::VACOPY: return "vacopy"; 6123 case ISD::VAEND: return "vaend"; 6124 case ISD::VASTART: return "vastart"; 6125 case ISD::DYNAMIC_STACKALLOC: return "dynamic_stackalloc"; 6126 case ISD::EXTRACT_ELEMENT: return "extract_element"; 6127 case ISD::BUILD_PAIR: return "build_pair"; 6128 case ISD::STACKSAVE: return "stacksave"; 6129 case ISD::STACKRESTORE: return "stackrestore"; 6130 case ISD::TRAP: return "trap"; 6131 6132 // Bit manipulation 6133 case ISD::BSWAP: return "bswap"; 6134 case ISD::CTPOP: return "ctpop"; 6135 case ISD::CTTZ: return "cttz"; 6136 case ISD::CTTZ_ZERO_UNDEF: return "cttz_zero_undef"; 6137 case ISD::CTLZ: return "ctlz"; 6138 case ISD::CTLZ_ZERO_UNDEF: return "ctlz_zero_undef"; 6139 6140 // Trampolines 6141 case ISD::INIT_TRAMPOLINE: return "init_trampoline"; 6142 case ISD::ADJUST_TRAMPOLINE: return "adjust_trampoline"; 6143 6144 case ISD::CONDCODE: 6145 switch (cast<CondCodeSDNode>(this)->get()) { 6146 default: llvm_unreachable("Unknown setcc condition!"); 6147 case ISD::SETOEQ: return "setoeq"; 6148 case ISD::SETOGT: return "setogt"; 6149 case ISD::SETOGE: return "setoge"; 6150 case ISD::SETOLT: return "setolt"; 6151 case ISD::SETOLE: return "setole"; 6152 case ISD::SETONE: return "setone"; 6153 6154 case ISD::SETO: return "seto"; 6155 case ISD::SETUO: return "setuo"; 6156 case ISD::SETUEQ: return "setue"; 6157 case ISD::SETUGT: return "setugt"; 6158 case ISD::SETUGE: return "setuge"; 6159 case ISD::SETULT: return "setult"; 6160 case ISD::SETULE: return "setule"; 6161 case ISD::SETUNE: return "setune"; 6162 6163 case ISD::SETEQ: return "seteq"; 6164 case ISD::SETGT: return "setgt"; 6165 case ISD::SETGE: return "setge"; 6166 case ISD::SETLT: return "setlt"; 6167 case ISD::SETLE: return "setle"; 6168 case ISD::SETNE: return "setne"; 6169 6170 case ISD::SETTRUE: return "settrue"; 6171 case ISD::SETTRUE2: return "settrue2"; 6172 case ISD::SETFALSE: return "setfalse"; 6173 case ISD::SETFALSE2: return "setfalse2"; 6174 } 6175 } 6176 } 6177 6178 const char *SDNode::getIndexedModeName(ISD::MemIndexedMode AM) { 6179 switch (AM) { 6180 default: 6181 return ""; 6182 case ISD::PRE_INC: 6183 return "<pre-inc>"; 6184 case ISD::PRE_DEC: 6185 return "<pre-dec>"; 6186 case ISD::POST_INC: 6187 return "<post-inc>"; 6188 case ISD::POST_DEC: 6189 return "<post-dec>"; 6190 } 6191 } 6192 6193 std::string ISD::ArgFlagsTy::getArgFlagsString() { 6194 std::string S = "< "; 6195 6196 if (isZExt()) 6197 S += "zext "; 6198 if (isSExt()) 6199 S += "sext "; 6200 if (isInReg()) 6201 S += "inreg "; 6202 if (isSRet()) 6203 S += "sret "; 6204 if (isByVal()) 6205 S += "byval "; 6206 if (isNest()) 6207 S += "nest "; 6208 if (getByValAlign()) 6209 S += "byval-align:" + utostr(getByValAlign()) + " "; 6210 if (getOrigAlign()) 6211 S += "orig-align:" + utostr(getOrigAlign()) + " "; 6212 if (getByValSize()) 6213 S += "byval-size:" + utostr(getByValSize()) + " "; 6214 return S + ">"; 6215 } 6216 6217 void SDNode::dump() const { dump(0); } 6218 void SDNode::dump(const SelectionDAG *G) const { 6219 print(dbgs(), G); 6220 dbgs() << '\n'; 6221 } 6222 6223 void SDNode::print_types(raw_ostream &OS, const SelectionDAG *G) const { 6224 OS << (void*)this << ": "; 6225 6226 for (unsigned i = 0, e = getNumValues(); i != e; ++i) { 6227 if (i) OS << ","; 6228 if (getValueType(i) == MVT::Other) 6229 OS << "ch"; 6230 else 6231 OS << getValueType(i).getEVTString(); 6232 } 6233 OS << " = " << getOperationName(G); 6234 } 6235 6236 void SDNode::print_details(raw_ostream &OS, const SelectionDAG *G) const { 6237 if (const MachineSDNode *MN = dyn_cast<MachineSDNode>(this)) { 6238 if (!MN->memoperands_empty()) { 6239 OS << "<"; 6240 OS << "Mem:"; 6241 for (MachineSDNode::mmo_iterator i = MN->memoperands_begin(), 6242 e = MN->memoperands_end(); i != e; ++i) { 6243 OS << **i; 6244 if (llvm::next(i) != e) 6245 OS << " "; 6246 } 6247 OS << ">"; 6248 } 6249 } else if (const ShuffleVectorSDNode *SVN = 6250 dyn_cast<ShuffleVectorSDNode>(this)) { 6251 OS << "<"; 6252 for (unsigned i = 0, e = ValueList[0].getVectorNumElements(); i != e; ++i) { 6253 int Idx = SVN->getMaskElt(i); 6254 if (i) OS << ","; 6255 if (Idx < 0) 6256 OS << "u"; 6257 else 6258 OS << Idx; 6259 } 6260 OS << ">"; 6261 } else if (const ConstantSDNode *CSDN = dyn_cast<ConstantSDNode>(this)) { 6262 OS << '<' << CSDN->getAPIntValue() << '>'; 6263 } else if (const ConstantFPSDNode *CSDN = dyn_cast<ConstantFPSDNode>(this)) { 6264 if (&CSDN->getValueAPF().getSemantics()==&APFloat::IEEEsingle) 6265 OS << '<' << CSDN->getValueAPF().convertToFloat() << '>'; 6266 else if (&CSDN->getValueAPF().getSemantics()==&APFloat::IEEEdouble) 6267 OS << '<' << CSDN->getValueAPF().convertToDouble() << '>'; 6268 else { 6269 OS << "<APFloat("; 6270 CSDN->getValueAPF().bitcastToAPInt().dump(); 6271 OS << ")>"; 6272 } 6273 } else if (const GlobalAddressSDNode *GADN = 6274 dyn_cast<GlobalAddressSDNode>(this)) { 6275 int64_t offset = GADN->getOffset(); 6276 OS << '<'; 6277 WriteAsOperand(OS, GADN->getGlobal()); 6278 OS << '>'; 6279 if (offset > 0) 6280 OS << " + " << offset; 6281 else 6282 OS << " " << offset; 6283 if (unsigned int TF = GADN->getTargetFlags()) 6284 OS << " [TF=" << TF << ']'; 6285 } else if (const FrameIndexSDNode *FIDN = dyn_cast<FrameIndexSDNode>(this)) { 6286 OS << "<" << FIDN->getIndex() << ">"; 6287 } else if (const JumpTableSDNode *JTDN = dyn_cast<JumpTableSDNode>(this)) { 6288 OS << "<" << JTDN->getIndex() << ">"; 6289 if (unsigned int TF = JTDN->getTargetFlags()) 6290 OS << " [TF=" << TF << ']'; 6291 } else if (const ConstantPoolSDNode *CP = dyn_cast<ConstantPoolSDNode>(this)){ 6292 int offset = CP->getOffset(); 6293 if (CP->isMachineConstantPoolEntry()) 6294 OS << "<" << *CP->getMachineCPVal() << ">"; 6295 else 6296 OS << "<" << *CP->getConstVal() << ">"; 6297 if (offset > 0) 6298 OS << " + " << offset; 6299 else 6300 OS << " " << offset; 6301 if (unsigned int TF = CP->getTargetFlags()) 6302 OS << " [TF=" << TF << ']'; 6303 } else if (const BasicBlockSDNode *BBDN = dyn_cast<BasicBlockSDNode>(this)) { 6304 OS << "<"; 6305 const Value *LBB = (const Value*)BBDN->getBasicBlock()->getBasicBlock(); 6306 if (LBB) 6307 OS << LBB->getName() << " "; 6308 OS << (const void*)BBDN->getBasicBlock() << ">"; 6309 } else if (const RegisterSDNode *R = dyn_cast<RegisterSDNode>(this)) { 6310 OS << ' ' << PrintReg(R->getReg(), G ? G->getTarget().getRegisterInfo() :0); 6311 } else if (const ExternalSymbolSDNode *ES = 6312 dyn_cast<ExternalSymbolSDNode>(this)) { 6313 OS << "'" << ES->getSymbol() << "'"; 6314 if (unsigned int TF = ES->getTargetFlags()) 6315 OS << " [TF=" << TF << ']'; 6316 } else if (const SrcValueSDNode *M = dyn_cast<SrcValueSDNode>(this)) { 6317 if (M->getValue()) 6318 OS << "<" << M->getValue() << ">"; 6319 else 6320 OS << "<null>"; 6321 } else if (const MDNodeSDNode *MD = dyn_cast<MDNodeSDNode>(this)) { 6322 if (MD->getMD()) 6323 OS << "<" << MD->getMD() << ">"; 6324 else 6325 OS << "<null>"; 6326 } else if (const VTSDNode *N = dyn_cast<VTSDNode>(this)) { 6327 OS << ":" << N->getVT().getEVTString(); 6328 } 6329 else if (const LoadSDNode *LD = dyn_cast<LoadSDNode>(this)) { 6330 OS << "<" << *LD->getMemOperand(); 6331 6332 bool doExt = true; 6333 switch (LD->getExtensionType()) { 6334 default: doExt = false; break; 6335 case ISD::EXTLOAD: OS << ", anyext"; break; 6336 case ISD::SEXTLOAD: OS << ", sext"; break; 6337 case ISD::ZEXTLOAD: OS << ", zext"; break; 6338 } 6339 if (doExt) 6340 OS << " from " << LD->getMemoryVT().getEVTString(); 6341 6342 const char *AM = getIndexedModeName(LD->getAddressingMode()); 6343 if (*AM) 6344 OS << ", " << AM; 6345 6346 OS << ">"; 6347 } else if (const StoreSDNode *ST = dyn_cast<StoreSDNode>(this)) { 6348 OS << "<" << *ST->getMemOperand(); 6349 6350 if (ST->isTruncatingStore()) 6351 OS << ", trunc to " << ST->getMemoryVT().getEVTString(); 6352 6353 const char *AM = getIndexedModeName(ST->getAddressingMode()); 6354 if (*AM) 6355 OS << ", " << AM; 6356 6357 OS << ">"; 6358 } else if (const MemSDNode* M = dyn_cast<MemSDNode>(this)) { 6359 OS << "<" << *M->getMemOperand() << ">"; 6360 } else if (const BlockAddressSDNode *BA = 6361 dyn_cast<BlockAddressSDNode>(this)) { 6362 OS << "<"; 6363 WriteAsOperand(OS, BA->getBlockAddress()->getFunction(), false); 6364 OS << ", "; 6365 WriteAsOperand(OS, BA->getBlockAddress()->getBasicBlock(), false); 6366 OS << ">"; 6367 if (unsigned int TF = BA->getTargetFlags()) 6368 OS << " [TF=" << TF << ']'; 6369 } 6370 6371 if (G) 6372 if (unsigned Order = G->GetOrdering(this)) 6373 OS << " [ORD=" << Order << ']'; 6374 6375 if (getNodeId() != -1) 6376 OS << " [ID=" << getNodeId() << ']'; 6377 6378 DebugLoc dl = getDebugLoc(); 6379 if (G && !dl.isUnknown()) { 6380 DIScope 6381 Scope(dl.getScope(G->getMachineFunction().getFunction()->getContext())); 6382 OS << " dbg:"; 6383 // Omit the directory, since it's usually long and uninteresting. 6384 if (Scope.Verify()) 6385 OS << Scope.getFilename(); 6386 else 6387 OS << "<unknown>"; 6388 OS << ':' << dl.getLine(); 6389 if (dl.getCol() != 0) 6390 OS << ':' << dl.getCol(); 6391 } 6392 } 6393 6394 void SDNode::print(raw_ostream &OS, const SelectionDAG *G) const { 6395 print_types(OS, G); 6396 for (unsigned i = 0, e = getNumOperands(); i != e; ++i) { 6397 if (i) OS << ", "; else OS << " "; 6398 OS << (void*)getOperand(i).getNode(); 6399 if (unsigned RN = getOperand(i).getResNo()) 6400 OS << ":" << RN; 6401 } 6402 print_details(OS, G); 6403 } 6404 6405 static void printrWithDepthHelper(raw_ostream &OS, const SDNode *N, 6406 const SelectionDAG *G, unsigned depth, 6407 unsigned indent) { 6408 if (depth == 0) 6409 return; 6410 6411 OS.indent(indent); 6412 6413 N->print(OS, G); 6414 6415 if (depth < 1) 6416 return; 6417 6418 for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i) { 6419 // Don't follow chain operands. 6420 if (N->getOperand(i).getValueType() == MVT::Other) 6421 continue; 6422 OS << '\n'; 6423 printrWithDepthHelper(OS, N->getOperand(i).getNode(), G, depth-1, indent+2); 6424 } 6425 } 6426 6427 void SDNode::printrWithDepth(raw_ostream &OS, const SelectionDAG *G, 6428 unsigned depth) const { 6429 printrWithDepthHelper(OS, this, G, depth, 0); 6430 } 6431 6432 void SDNode::printrFull(raw_ostream &OS, const SelectionDAG *G) const { 6433 // Don't print impossibly deep things. 6434 printrWithDepth(OS, G, 10); 6435 } 6436 6437 void SDNode::dumprWithDepth(const SelectionDAG *G, unsigned depth) const { 6438 printrWithDepth(dbgs(), G, depth); 6439 } 6440 6441 void SDNode::dumprFull(const SelectionDAG *G) const { 6442 // Don't print impossibly deep things. 6443 dumprWithDepth(G, 10); 6444 } 6445 6446 static void DumpNodes(const SDNode *N, unsigned indent, const SelectionDAG *G) { 6447 for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i) 6448 if (N->getOperand(i).getNode()->hasOneUse()) 6449 DumpNodes(N->getOperand(i).getNode(), indent+2, G); 6450 else 6451 dbgs() << "\n" << std::string(indent+2, ' ') 6452 << (void*)N->getOperand(i).getNode() << ": <multiple use>"; 6453 6454 6455 dbgs() << "\n"; 6456 dbgs().indent(indent); 6457 N->dump(G); 6458 } 6459 6460 SDValue SelectionDAG::UnrollVectorOp(SDNode *N, unsigned ResNE) { 6461 assert(N->getNumValues() == 1 && 6462 "Can't unroll a vector with multiple results!"); 6463 6464 EVT VT = N->getValueType(0); 6465 unsigned NE = VT.getVectorNumElements(); 6466 EVT EltVT = VT.getVectorElementType(); 6467 DebugLoc dl = N->getDebugLoc(); 6468 6469 SmallVector<SDValue, 8> Scalars; 6470 SmallVector<SDValue, 4> Operands(N->getNumOperands()); 6471 6472 // If ResNE is 0, fully unroll the vector op. 6473 if (ResNE == 0) 6474 ResNE = NE; 6475 else if (NE > ResNE) 6476 NE = ResNE; 6477 6478 unsigned i; 6479 for (i= 0; i != NE; ++i) { 6480 for (unsigned j = 0, e = N->getNumOperands(); j != e; ++j) { 6481 SDValue Operand = N->getOperand(j); 6482 EVT OperandVT = Operand.getValueType(); 6483 if (OperandVT.isVector()) { 6484 // A vector operand; extract a single element. 6485 EVT OperandEltVT = OperandVT.getVectorElementType(); 6486 Operands[j] = getNode(ISD::EXTRACT_VECTOR_ELT, dl, 6487 OperandEltVT, 6488 Operand, 6489 getConstant(i, TLI.getPointerTy())); 6490 } else { 6491 // A scalar operand; just use it as is. 6492 Operands[j] = Operand; 6493 } 6494 } 6495 6496 switch (N->getOpcode()) { 6497 default: 6498 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, 6499 &Operands[0], Operands.size())); 6500 break; 6501 case ISD::VSELECT: 6502 Scalars.push_back(getNode(ISD::SELECT, dl, EltVT, 6503 &Operands[0], Operands.size())); 6504 break; 6505 case ISD::SHL: 6506 case ISD::SRA: 6507 case ISD::SRL: 6508 case ISD::ROTL: 6509 case ISD::ROTR: 6510 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, Operands[0], 6511 getShiftAmountOperand(Operands[0].getValueType(), 6512 Operands[1]))); 6513 break; 6514 case ISD::SIGN_EXTEND_INREG: 6515 case ISD::FP_ROUND_INREG: { 6516 EVT ExtVT = cast<VTSDNode>(Operands[1])->getVT().getVectorElementType(); 6517 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, 6518 Operands[0], 6519 getValueType(ExtVT))); 6520 } 6521 } 6522 } 6523 6524 for (; i < ResNE; ++i) 6525 Scalars.push_back(getUNDEF(EltVT)); 6526 6527 return getNode(ISD::BUILD_VECTOR, dl, 6528 EVT::getVectorVT(*getContext(), EltVT, ResNE), 6529 &Scalars[0], Scalars.size()); 6530 } 6531 6532 6533 /// isConsecutiveLoad - Return true if LD is loading 'Bytes' bytes from a 6534 /// location that is 'Dist' units away from the location that the 'Base' load 6535 /// is loading from. 6536 bool SelectionDAG::isConsecutiveLoad(LoadSDNode *LD, LoadSDNode *Base, 6537 unsigned Bytes, int Dist) const { 6538 if (LD->getChain() != Base->getChain()) 6539 return false; 6540 EVT VT = LD->getValueType(0); 6541 if (VT.getSizeInBits() / 8 != Bytes) 6542 return false; 6543 6544 SDValue Loc = LD->getOperand(1); 6545 SDValue BaseLoc = Base->getOperand(1); 6546 if (Loc.getOpcode() == ISD::FrameIndex) { 6547 if (BaseLoc.getOpcode() != ISD::FrameIndex) 6548 return false; 6549 const MachineFrameInfo *MFI = getMachineFunction().getFrameInfo(); 6550 int FI = cast<FrameIndexSDNode>(Loc)->getIndex(); 6551 int BFI = cast<FrameIndexSDNode>(BaseLoc)->getIndex(); 6552 int FS = MFI->getObjectSize(FI); 6553 int BFS = MFI->getObjectSize(BFI); 6554 if (FS != BFS || FS != (int)Bytes) return false; 6555 return MFI->getObjectOffset(FI) == (MFI->getObjectOffset(BFI) + Dist*Bytes); 6556 } 6557 6558 // Handle X+C 6559 if (isBaseWithConstantOffset(Loc) && Loc.getOperand(0) == BaseLoc && 6560 cast<ConstantSDNode>(Loc.getOperand(1))->getSExtValue() == Dist*Bytes) 6561 return true; 6562 6563 const GlobalValue *GV1 = NULL; 6564 const GlobalValue *GV2 = NULL; 6565 int64_t Offset1 = 0; 6566 int64_t Offset2 = 0; 6567 bool isGA1 = TLI.isGAPlusOffset(Loc.getNode(), GV1, Offset1); 6568 bool isGA2 = TLI.isGAPlusOffset(BaseLoc.getNode(), GV2, Offset2); 6569 if (isGA1 && isGA2 && GV1 == GV2) 6570 return Offset1 == (Offset2 + Dist*Bytes); 6571 return false; 6572 } 6573 6574 6575 /// InferPtrAlignment - Infer alignment of a load / store address. Return 0 if 6576 /// it cannot be inferred. 6577 unsigned SelectionDAG::InferPtrAlignment(SDValue Ptr) const { 6578 // If this is a GlobalAddress + cst, return the alignment. 6579 const GlobalValue *GV; 6580 int64_t GVOffset = 0; 6581 if (TLI.isGAPlusOffset(Ptr.getNode(), GV, GVOffset)) { 6582 unsigned PtrWidth = TLI.getPointerTy().getSizeInBits(); 6583 APInt AllOnes = APInt::getAllOnesValue(PtrWidth); 6584 APInt KnownZero(PtrWidth, 0), KnownOne(PtrWidth, 0); 6585 llvm::ComputeMaskedBits(const_cast<GlobalValue*>(GV), AllOnes, 6586 KnownZero, KnownOne, TLI.getTargetData()); 6587 unsigned AlignBits = KnownZero.countTrailingOnes(); 6588 unsigned Align = AlignBits ? 1 << std::min(31U, AlignBits) : 0; 6589 if (Align) 6590 return MinAlign(Align, GVOffset); 6591 } 6592 6593 // If this is a direct reference to a stack slot, use information about the 6594 // stack slot's alignment. 6595 int FrameIdx = 1 << 31; 6596 int64_t FrameOffset = 0; 6597 if (FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Ptr)) { 6598 FrameIdx = FI->getIndex(); 6599 } else if (isBaseWithConstantOffset(Ptr) && 6600 isa<FrameIndexSDNode>(Ptr.getOperand(0))) { 6601 // Handle FI+Cst 6602 FrameIdx = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex(); 6603 FrameOffset = Ptr.getConstantOperandVal(1); 6604 } 6605 6606 if (FrameIdx != (1 << 31)) { 6607 const MachineFrameInfo &MFI = *getMachineFunction().getFrameInfo(); 6608 unsigned FIInfoAlign = MinAlign(MFI.getObjectAlignment(FrameIdx), 6609 FrameOffset); 6610 return FIInfoAlign; 6611 } 6612 6613 return 0; 6614 } 6615 6616 void SelectionDAG::dump() const { 6617 dbgs() << "SelectionDAG has " << AllNodes.size() << " nodes:"; 6618 6619 for (allnodes_const_iterator I = allnodes_begin(), E = allnodes_end(); 6620 I != E; ++I) { 6621 const SDNode *N = I; 6622 if (!N->hasOneUse() && N != getRoot().getNode()) 6623 DumpNodes(N, 2, this); 6624 } 6625 6626 if (getRoot().getNode()) DumpNodes(getRoot().getNode(), 2, this); 6627 6628 dbgs() << "\n\n"; 6629 } 6630 6631 void SDNode::printr(raw_ostream &OS, const SelectionDAG *G) const { 6632 print_types(OS, G); 6633 print_details(OS, G); 6634 } 6635 6636 typedef SmallPtrSet<const SDNode *, 128> VisitedSDNodeSet; 6637 static void DumpNodesr(raw_ostream &OS, const SDNode *N, unsigned indent, 6638 const SelectionDAG *G, VisitedSDNodeSet &once) { 6639 if (!once.insert(N)) // If we've been here before, return now. 6640 return; 6641 6642 // Dump the current SDNode, but don't end the line yet. 6643 OS.indent(indent); 6644 N->printr(OS, G); 6645 6646 // Having printed this SDNode, walk the children: 6647 for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i) { 6648 const SDNode *child = N->getOperand(i).getNode(); 6649 6650 if (i) OS << ","; 6651 OS << " "; 6652 6653 if (child->getNumOperands() == 0) { 6654 // This child has no grandchildren; print it inline right here. 6655 child->printr(OS, G); 6656 once.insert(child); 6657 } else { // Just the address. FIXME: also print the child's opcode. 6658 OS << (void*)child; 6659 if (unsigned RN = N->getOperand(i).getResNo()) 6660 OS << ":" << RN; 6661 } 6662 } 6663 6664 OS << "\n"; 6665 6666 // Dump children that have grandchildren on their own line(s). 6667 for (unsigned i = 0, e = N->getNumOperands(); i != e; ++i) { 6668 const SDNode *child = N->getOperand(i).getNode(); 6669 DumpNodesr(OS, child, indent+2, G, once); 6670 } 6671 } 6672 6673 void SDNode::dumpr() const { 6674 VisitedSDNodeSet once; 6675 DumpNodesr(dbgs(), this, 0, 0, once); 6676 } 6677 6678 void SDNode::dumpr(const SelectionDAG *G) const { 6679 VisitedSDNodeSet once; 6680 DumpNodesr(dbgs(), this, 0, G, once); 6681 } 6682 6683 6684 // getAddressSpace - Return the address space this GlobalAddress belongs to. 6685 unsigned GlobalAddressSDNode::getAddressSpace() const { 6686 return getGlobal()->getType()->getAddressSpace(); 6687 } 6688 6689 6690 Type *ConstantPoolSDNode::getType() const { 6691 if (isMachineConstantPoolEntry()) 6692 return Val.MachineCPVal->getType(); 6693 return Val.ConstVal->getType(); 6694 } 6695 6696 bool BuildVectorSDNode::isConstantSplat(APInt &SplatValue, 6697 APInt &SplatUndef, 6698 unsigned &SplatBitSize, 6699 bool &HasAnyUndefs, 6700 unsigned MinSplatBits, 6701 bool isBigEndian) { 6702 EVT VT = getValueType(0); 6703 assert(VT.isVector() && "Expected a vector type"); 6704 unsigned sz = VT.getSizeInBits(); 6705 if (MinSplatBits > sz) 6706 return false; 6707 6708 SplatValue = APInt(sz, 0); 6709 SplatUndef = APInt(sz, 0); 6710 6711 // Get the bits. Bits with undefined values (when the corresponding element 6712 // of the vector is an ISD::UNDEF value) are set in SplatUndef and cleared 6713 // in SplatValue. If any of the values are not constant, give up and return 6714 // false. 6715 unsigned int nOps = getNumOperands(); 6716 assert(nOps > 0 && "isConstantSplat has 0-size build vector"); 6717 unsigned EltBitSize = VT.getVectorElementType().getSizeInBits(); 6718 6719 for (unsigned j = 0; j < nOps; ++j) { 6720 unsigned i = isBigEndian ? nOps-1-j : j; 6721 SDValue OpVal = getOperand(i); 6722 unsigned BitPos = j * EltBitSize; 6723 6724 if (OpVal.getOpcode() == ISD::UNDEF) 6725 SplatUndef |= APInt::getBitsSet(sz, BitPos, BitPos + EltBitSize); 6726 else if (ConstantSDNode *CN = dyn_cast<ConstantSDNode>(OpVal)) 6727 SplatValue |= CN->getAPIntValue().zextOrTrunc(EltBitSize). 6728 zextOrTrunc(sz) << BitPos; 6729 else if (ConstantFPSDNode *CN = dyn_cast<ConstantFPSDNode>(OpVal)) 6730 SplatValue |= CN->getValueAPF().bitcastToAPInt().zextOrTrunc(sz) <<BitPos; 6731 else 6732 return false; 6733 } 6734 6735 // The build_vector is all constants or undefs. Find the smallest element 6736 // size that splats the vector. 6737 6738 HasAnyUndefs = (SplatUndef != 0); 6739 while (sz > 8) { 6740 6741 unsigned HalfSize = sz / 2; 6742 APInt HighValue = SplatValue.lshr(HalfSize).trunc(HalfSize); 6743 APInt LowValue = SplatValue.trunc(HalfSize); 6744 APInt HighUndef = SplatUndef.lshr(HalfSize).trunc(HalfSize); 6745 APInt LowUndef = SplatUndef.trunc(HalfSize); 6746 6747 // If the two halves do not match (ignoring undef bits), stop here. 6748 if ((HighValue & ~LowUndef) != (LowValue & ~HighUndef) || 6749 MinSplatBits > HalfSize) 6750 break; 6751 6752 SplatValue = HighValue | LowValue; 6753 SplatUndef = HighUndef & LowUndef; 6754 6755 sz = HalfSize; 6756 } 6757 6758 SplatBitSize = sz; 6759 return true; 6760 } 6761 6762 bool ShuffleVectorSDNode::isSplatMask(const int *Mask, EVT VT) { 6763 // Find the first non-undef value in the shuffle mask. 6764 unsigned i, e; 6765 for (i = 0, e = VT.getVectorNumElements(); i != e && Mask[i] < 0; ++i) 6766 /* search */; 6767 6768 assert(i != e && "VECTOR_SHUFFLE node with all undef indices!"); 6769 6770 // Make sure all remaining elements are either undef or the same as the first 6771 // non-undef value. 6772 for (int Idx = Mask[i]; i != e; ++i) 6773 if (Mask[i] >= 0 && Mask[i] != Idx) 6774 return false; 6775 return true; 6776 } 6777 6778 #ifdef XDEBUG 6779 static void checkForCyclesHelper(const SDNode *N, 6780 SmallPtrSet<const SDNode*, 32> &Visited, 6781 SmallPtrSet<const SDNode*, 32> &Checked) { 6782 // If this node has already been checked, don't check it again. 6783 if (Checked.count(N)) 6784 return; 6785 6786 // If a node has already been visited on this depth-first walk, reject it as 6787 // a cycle. 6788 if (!Visited.insert(N)) { 6789 dbgs() << "Offending node:\n"; 6790 N->dumprFull(); 6791 errs() << "Detected cycle in SelectionDAG\n"; 6792 abort(); 6793 } 6794 6795 for(unsigned i = 0, e = N->getNumOperands(); i != e; ++i) 6796 checkForCyclesHelper(N->getOperand(i).getNode(), Visited, Checked); 6797 6798 Checked.insert(N); 6799 Visited.erase(N); 6800 } 6801 #endif 6802 6803 void llvm::checkForCycles(const llvm::SDNode *N) { 6804 #ifdef XDEBUG 6805 assert(N && "Checking nonexistant SDNode"); 6806 SmallPtrSet<const SDNode*, 32> visited; 6807 SmallPtrSet<const SDNode*, 32> checked; 6808 checkForCyclesHelper(N, visited, checked); 6809 #endif 6810 } 6811 6812 void llvm::checkForCycles(const llvm::SelectionDAG *DAG) { 6813 checkForCycles(DAG->getRoot().getNode()); 6814 } 6815