1 //===-- RuntimeDyldELF.cpp - Run-time dynamic linker for MC-JIT -*- C++ -*-===// 2 // 3 // The LLVM Compiler Infrastructure 4 // 5 // This file is distributed under the University of Illinois Open Source 6 // License. See LICENSE.TXT for details. 7 // 8 //===----------------------------------------------------------------------===// 9 // 10 // Implementation of ELF support for the MC-JIT runtime dynamic linker. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #define DEBUG_TYPE "dyld" 15 #include "RuntimeDyldELF.h" 16 #include "JITRegistrar.h" 17 #include "ObjectImageCommon.h" 18 #include "llvm/ADT/IntervalMap.h" 19 #include "llvm/ADT/OwningPtr.h" 20 #include "llvm/ADT/STLExtras.h" 21 #include "llvm/ADT/StringRef.h" 22 #include "llvm/ADT/Triple.h" 23 #include "llvm/ExecutionEngine/ObjectBuffer.h" 24 #include "llvm/ExecutionEngine/ObjectImage.h" 25 #include "llvm/Object/ELFObjectFile.h" 26 #include "llvm/Object/ObjectFile.h" 27 #include "llvm/Support/ELF.h" 28 #include "llvm/Support/MemoryBuffer.h" 29 30 using namespace llvm; 31 using namespace llvm::object; 32 33 namespace { 34 35 static inline 36 error_code check(error_code Err) { 37 if (Err) { 38 report_fatal_error(Err.message()); 39 } 40 return Err; 41 } 42 43 template<class ELFT> 44 class DyldELFObject 45 : public ELFObjectFile<ELFT> { 46 LLVM_ELF_IMPORT_TYPES_ELFT(ELFT) 47 48 typedef Elf_Shdr_Impl<ELFT> Elf_Shdr; 49 typedef Elf_Sym_Impl<ELFT> Elf_Sym; 50 typedef 51 Elf_Rel_Impl<ELFT, false> Elf_Rel; 52 typedef 53 Elf_Rel_Impl<ELFT, true> Elf_Rela; 54 55 typedef Elf_Ehdr_Impl<ELFT> Elf_Ehdr; 56 57 typedef typename ELFDataTypeTypedefHelper< 58 ELFT>::value_type addr_type; 59 60 public: 61 DyldELFObject(MemoryBuffer *Wrapper, error_code &ec); 62 63 void updateSectionAddress(const SectionRef &Sec, uint64_t Addr); 64 void updateSymbolAddress(const SymbolRef &Sym, uint64_t Addr); 65 66 // Methods for type inquiry through isa, cast and dyn_cast 67 static inline bool classof(const Binary *v) { 68 return (isa<ELFObjectFile<ELFT> >(v) 69 && classof(cast<ELFObjectFile 70 <ELFT> >(v))); 71 } 72 static inline bool classof( 73 const ELFObjectFile<ELFT> *v) { 74 return v->isDyldType(); 75 } 76 }; 77 78 template<class ELFT> 79 class ELFObjectImage : public ObjectImageCommon { 80 protected: 81 DyldELFObject<ELFT> *DyldObj; 82 bool Registered; 83 84 public: 85 ELFObjectImage(ObjectBuffer *Input, 86 DyldELFObject<ELFT> *Obj) 87 : ObjectImageCommon(Input, Obj), 88 DyldObj(Obj), 89 Registered(false) {} 90 91 virtual ~ELFObjectImage() { 92 if (Registered) 93 deregisterWithDebugger(); 94 } 95 96 // Subclasses can override these methods to update the image with loaded 97 // addresses for sections and common symbols 98 virtual void updateSectionAddress(const SectionRef &Sec, uint64_t Addr) 99 { 100 DyldObj->updateSectionAddress(Sec, Addr); 101 } 102 103 virtual void updateSymbolAddress(const SymbolRef &Sym, uint64_t Addr) 104 { 105 DyldObj->updateSymbolAddress(Sym, Addr); 106 } 107 108 virtual void registerWithDebugger() 109 { 110 JITRegistrar::getGDBRegistrar().registerObject(*Buffer); 111 Registered = true; 112 } 113 virtual void deregisterWithDebugger() 114 { 115 JITRegistrar::getGDBRegistrar().deregisterObject(*Buffer); 116 } 117 }; 118 119 // The MemoryBuffer passed into this constructor is just a wrapper around the 120 // actual memory. Ultimately, the Binary parent class will take ownership of 121 // this MemoryBuffer object but not the underlying memory. 122 template<class ELFT> 123 DyldELFObject<ELFT>::DyldELFObject(MemoryBuffer *Wrapper, error_code &ec) 124 : ELFObjectFile<ELFT>(Wrapper, ec) { 125 this->isDyldELFObject = true; 126 } 127 128 template<class ELFT> 129 void DyldELFObject<ELFT>::updateSectionAddress(const SectionRef &Sec, 130 uint64_t Addr) { 131 DataRefImpl ShdrRef = Sec.getRawDataRefImpl(); 132 Elf_Shdr *shdr = const_cast<Elf_Shdr*>( 133 reinterpret_cast<const Elf_Shdr *>(ShdrRef.p)); 134 135 // This assumes the address passed in matches the target address bitness 136 // The template-based type cast handles everything else. 137 shdr->sh_addr = static_cast<addr_type>(Addr); 138 } 139 140 template<class ELFT> 141 void DyldELFObject<ELFT>::updateSymbolAddress(const SymbolRef &SymRef, 142 uint64_t Addr) { 143 144 Elf_Sym *sym = const_cast<Elf_Sym*>( 145 ELFObjectFile<ELFT>::getSymbol(SymRef.getRawDataRefImpl())); 146 147 // This assumes the address passed in matches the target address bitness 148 // The template-based type cast handles everything else. 149 sym->st_value = static_cast<addr_type>(Addr); 150 } 151 152 } // namespace 153 154 namespace llvm { 155 156 void RuntimeDyldELF::registerEHFrames() { 157 if (!MemMgr) 158 return; 159 for (int i = 0, e = UnregisteredEHFrameSections.size(); i != e; ++i) { 160 SID EHFrameSID = UnregisteredEHFrameSections[i]; 161 uint8_t *EHFrameAddr = Sections[EHFrameSID].Address; 162 uint64_t EHFrameLoadAddr = Sections[EHFrameSID].LoadAddress; 163 size_t EHFrameSize = Sections[EHFrameSID].Size; 164 MemMgr->registerEHFrames(EHFrameAddr, EHFrameLoadAddr, EHFrameSize); 165 RegisteredEHFrameSections.push_back(EHFrameSID); 166 } 167 UnregisteredEHFrameSections.clear(); 168 } 169 170 void RuntimeDyldELF::deregisterEHFrames() { 171 if (!MemMgr) 172 return; 173 for (int i = 0, e = RegisteredEHFrameSections.size(); i != e; ++i) { 174 SID EHFrameSID = RegisteredEHFrameSections[i]; 175 uint8_t *EHFrameAddr = Sections[EHFrameSID].Address; 176 uint64_t EHFrameLoadAddr = Sections[EHFrameSID].LoadAddress; 177 size_t EHFrameSize = Sections[EHFrameSID].Size; 178 MemMgr->deregisterEHFrames(EHFrameAddr, EHFrameLoadAddr, EHFrameSize); 179 } 180 RegisteredEHFrameSections.clear(); 181 } 182 183 ObjectImage *RuntimeDyldELF::createObjectImageFromFile(object::ObjectFile *ObjFile) { 184 if (!ObjFile) 185 return NULL; 186 187 error_code ec; 188 MemoryBuffer* Buffer = MemoryBuffer::getMemBuffer(ObjFile->getData(), 189 "", 190 false); 191 192 if (ObjFile->getBytesInAddress() == 4 && ObjFile->isLittleEndian()) { 193 DyldELFObject<ELFType<support::little, 2, false> > *Obj = 194 new DyldELFObject<ELFType<support::little, 2, false> >(Buffer, ec); 195 return new ELFObjectImage<ELFType<support::little, 2, false> >(NULL, Obj); 196 } 197 else if (ObjFile->getBytesInAddress() == 4 && !ObjFile->isLittleEndian()) { 198 DyldELFObject<ELFType<support::big, 2, false> > *Obj = 199 new DyldELFObject<ELFType<support::big, 2, false> >(Buffer, ec); 200 return new ELFObjectImage<ELFType<support::big, 2, false> >(NULL, Obj); 201 } 202 else if (ObjFile->getBytesInAddress() == 8 && !ObjFile->isLittleEndian()) { 203 DyldELFObject<ELFType<support::big, 2, true> > *Obj = 204 new DyldELFObject<ELFType<support::big, 2, true> >(Buffer, ec); 205 return new ELFObjectImage<ELFType<support::big, 2, true> >(NULL, Obj); 206 } 207 else if (ObjFile->getBytesInAddress() == 8 && ObjFile->isLittleEndian()) { 208 DyldELFObject<ELFType<support::little, 2, true> > *Obj = 209 new DyldELFObject<ELFType<support::little, 2, true> >(Buffer, ec); 210 return new ELFObjectImage<ELFType<support::little, 2, true> >(NULL, Obj); 211 } 212 else 213 llvm_unreachable("Unexpected ELF format"); 214 } 215 216 ObjectImage *RuntimeDyldELF::createObjectImage(ObjectBuffer *Buffer) { 217 if (Buffer->getBufferSize() < ELF::EI_NIDENT) 218 llvm_unreachable("Unexpected ELF object size"); 219 std::pair<unsigned char, unsigned char> Ident = std::make_pair( 220 (uint8_t)Buffer->getBufferStart()[ELF::EI_CLASS], 221 (uint8_t)Buffer->getBufferStart()[ELF::EI_DATA]); 222 error_code ec; 223 224 if (Ident.first == ELF::ELFCLASS32 && Ident.second == ELF::ELFDATA2LSB) { 225 DyldELFObject<ELFType<support::little, 4, false> > *Obj = 226 new DyldELFObject<ELFType<support::little, 4, false> >( 227 Buffer->getMemBuffer(), ec); 228 return new ELFObjectImage<ELFType<support::little, 4, false> >(Buffer, Obj); 229 } 230 else if (Ident.first == ELF::ELFCLASS32 && Ident.second == ELF::ELFDATA2MSB) { 231 DyldELFObject<ELFType<support::big, 4, false> > *Obj = 232 new DyldELFObject<ELFType<support::big, 4, false> >( 233 Buffer->getMemBuffer(), ec); 234 return new ELFObjectImage<ELFType<support::big, 4, false> >(Buffer, Obj); 235 } 236 else if (Ident.first == ELF::ELFCLASS64 && Ident.second == ELF::ELFDATA2MSB) { 237 DyldELFObject<ELFType<support::big, 8, true> > *Obj = 238 new DyldELFObject<ELFType<support::big, 8, true> >( 239 Buffer->getMemBuffer(), ec); 240 return new ELFObjectImage<ELFType<support::big, 8, true> >(Buffer, Obj); 241 } 242 else if (Ident.first == ELF::ELFCLASS64 && Ident.second == ELF::ELFDATA2LSB) { 243 DyldELFObject<ELFType<support::little, 8, true> > *Obj = 244 new DyldELFObject<ELFType<support::little, 8, true> >( 245 Buffer->getMemBuffer(), ec); 246 return new ELFObjectImage<ELFType<support::little, 8, true> >(Buffer, Obj); 247 } 248 else 249 llvm_unreachable("Unexpected ELF format"); 250 } 251 252 RuntimeDyldELF::~RuntimeDyldELF() { 253 } 254 255 void RuntimeDyldELF::resolveX86_64Relocation(const SectionEntry &Section, 256 uint64_t Offset, 257 uint64_t Value, 258 uint32_t Type, 259 int64_t Addend, 260 uint64_t SymOffset) { 261 switch (Type) { 262 default: 263 llvm_unreachable("Relocation type not implemented yet!"); 264 break; 265 case ELF::R_X86_64_64: { 266 uint64_t *Target = reinterpret_cast<uint64_t*>(Section.Address + Offset); 267 *Target = Value + Addend; 268 DEBUG(dbgs() << "Writing " << format("%p", (Value + Addend)) 269 << " at " << format("%p\n",Target)); 270 break; 271 } 272 case ELF::R_X86_64_32: 273 case ELF::R_X86_64_32S: { 274 Value += Addend; 275 assert((Type == ELF::R_X86_64_32 && (Value <= UINT32_MAX)) || 276 (Type == ELF::R_X86_64_32S && 277 ((int64_t)Value <= INT32_MAX && (int64_t)Value >= INT32_MIN))); 278 uint32_t TruncatedAddr = (Value & 0xFFFFFFFF); 279 uint32_t *Target = reinterpret_cast<uint32_t*>(Section.Address + Offset); 280 *Target = TruncatedAddr; 281 DEBUG(dbgs() << "Writing " << format("%p", TruncatedAddr) 282 << " at " << format("%p\n",Target)); 283 break; 284 } 285 case ELF::R_X86_64_GOTPCREL: { 286 // findGOTEntry returns the 'G + GOT' part of the relocation calculation 287 // based on the load/target address of the GOT (not the current/local addr). 288 uint64_t GOTAddr = findGOTEntry(Value, SymOffset); 289 uint32_t *Target = reinterpret_cast<uint32_t*>(Section.Address + Offset); 290 uint64_t FinalAddress = Section.LoadAddress + Offset; 291 // The processRelocationRef method combines the symbol offset and the addend 292 // and in most cases that's what we want. For this relocation type, we need 293 // the raw addend, so we subtract the symbol offset to get it. 294 int64_t RealOffset = GOTAddr + Addend - SymOffset - FinalAddress; 295 assert(RealOffset <= INT32_MAX && RealOffset >= INT32_MIN); 296 int32_t TruncOffset = (RealOffset & 0xFFFFFFFF); 297 *Target = TruncOffset; 298 break; 299 } 300 case ELF::R_X86_64_PC32: { 301 // Get the placeholder value from the generated object since 302 // a previous relocation attempt may have overwritten the loaded version 303 uint32_t *Placeholder = reinterpret_cast<uint32_t*>(Section.ObjAddress 304 + Offset); 305 uint32_t *Target = reinterpret_cast<uint32_t*>(Section.Address + Offset); 306 uint64_t FinalAddress = Section.LoadAddress + Offset; 307 int64_t RealOffset = *Placeholder + Value + Addend - FinalAddress; 308 assert(RealOffset <= INT32_MAX && RealOffset >= INT32_MIN); 309 int32_t TruncOffset = (RealOffset & 0xFFFFFFFF); 310 *Target = TruncOffset; 311 break; 312 } 313 case ELF::R_X86_64_PC64: { 314 // Get the placeholder value from the generated object since 315 // a previous relocation attempt may have overwritten the loaded version 316 uint64_t *Placeholder = reinterpret_cast<uint64_t*>(Section.ObjAddress 317 + Offset); 318 uint64_t *Target = reinterpret_cast<uint64_t*>(Section.Address + Offset); 319 uint64_t FinalAddress = Section.LoadAddress + Offset; 320 *Target = *Placeholder + Value + Addend - FinalAddress; 321 break; 322 } 323 } 324 } 325 326 void RuntimeDyldELF::resolveX86Relocation(const SectionEntry &Section, 327 uint64_t Offset, 328 uint32_t Value, 329 uint32_t Type, 330 int32_t Addend) { 331 switch (Type) { 332 case ELF::R_386_32: { 333 // Get the placeholder value from the generated object since 334 // a previous relocation attempt may have overwritten the loaded version 335 uint32_t *Placeholder = reinterpret_cast<uint32_t*>(Section.ObjAddress 336 + Offset); 337 uint32_t *Target = reinterpret_cast<uint32_t*>(Section.Address + Offset); 338 *Target = *Placeholder + Value + Addend; 339 break; 340 } 341 case ELF::R_386_PC32: { 342 // Get the placeholder value from the generated object since 343 // a previous relocation attempt may have overwritten the loaded version 344 uint32_t *Placeholder = reinterpret_cast<uint32_t*>(Section.ObjAddress 345 + Offset); 346 uint32_t *Target = reinterpret_cast<uint32_t*>(Section.Address + Offset); 347 uint32_t FinalAddress = ((Section.LoadAddress + Offset) & 0xFFFFFFFF); 348 uint32_t RealOffset = *Placeholder + Value + Addend - FinalAddress; 349 *Target = RealOffset; 350 break; 351 } 352 default: 353 // There are other relocation types, but it appears these are the 354 // only ones currently used by the LLVM ELF object writer 355 llvm_unreachable("Relocation type not implemented yet!"); 356 break; 357 } 358 } 359 360 void RuntimeDyldELF::resolveAArch64Relocation(const SectionEntry &Section, 361 uint64_t Offset, 362 uint64_t Value, 363 uint32_t Type, 364 int64_t Addend) { 365 uint32_t *TargetPtr = reinterpret_cast<uint32_t*>(Section.Address + Offset); 366 uint64_t FinalAddress = Section.LoadAddress + Offset; 367 368 DEBUG(dbgs() << "resolveAArch64Relocation, LocalAddress: 0x" 369 << format("%llx", Section.Address + Offset) 370 << " FinalAddress: 0x" << format("%llx",FinalAddress) 371 << " Value: 0x" << format("%llx",Value) 372 << " Type: 0x" << format("%x",Type) 373 << " Addend: 0x" << format("%llx",Addend) 374 << "\n"); 375 376 switch (Type) { 377 default: 378 llvm_unreachable("Relocation type not implemented yet!"); 379 break; 380 case ELF::R_AARCH64_ABS64: { 381 uint64_t *TargetPtr = reinterpret_cast<uint64_t*>(Section.Address + Offset); 382 *TargetPtr = Value + Addend; 383 break; 384 } 385 case ELF::R_AARCH64_PREL32: { 386 uint64_t Result = Value + Addend - FinalAddress; 387 assert(static_cast<int64_t>(Result) >= INT32_MIN && 388 static_cast<int64_t>(Result) <= UINT32_MAX); 389 *TargetPtr = static_cast<uint32_t>(Result & 0xffffffffU); 390 break; 391 } 392 case ELF::R_AARCH64_CALL26: // fallthrough 393 case ELF::R_AARCH64_JUMP26: { 394 // Operation: S+A-P. Set Call or B immediate value to bits fff_fffc of the 395 // calculation. 396 uint64_t BranchImm = Value + Addend - FinalAddress; 397 398 // "Check that -2^27 <= result < 2^27". 399 assert(-(1LL << 27) <= static_cast<int64_t>(BranchImm) && 400 static_cast<int64_t>(BranchImm) < (1LL << 27)); 401 402 // AArch64 code is emitted with .rela relocations. The data already in any 403 // bits affected by the relocation on entry is garbage. 404 *TargetPtr &= 0xfc000000U; 405 // Immediate goes in bits 25:0 of B and BL. 406 *TargetPtr |= static_cast<uint32_t>(BranchImm & 0xffffffcU) >> 2; 407 break; 408 } 409 case ELF::R_AARCH64_MOVW_UABS_G3: { 410 uint64_t Result = Value + Addend; 411 412 // AArch64 code is emitted with .rela relocations. The data already in any 413 // bits affected by the relocation on entry is garbage. 414 *TargetPtr &= 0xffe0001fU; 415 // Immediate goes in bits 20:5 of MOVZ/MOVK instruction 416 *TargetPtr |= Result >> (48 - 5); 417 // Shift must be "lsl #48", in bits 22:21 418 assert((*TargetPtr >> 21 & 0x3) == 3 && "invalid shift for relocation"); 419 break; 420 } 421 case ELF::R_AARCH64_MOVW_UABS_G2_NC: { 422 uint64_t Result = Value + Addend; 423 424 // AArch64 code is emitted with .rela relocations. The data already in any 425 // bits affected by the relocation on entry is garbage. 426 *TargetPtr &= 0xffe0001fU; 427 // Immediate goes in bits 20:5 of MOVZ/MOVK instruction 428 *TargetPtr |= ((Result & 0xffff00000000ULL) >> (32 - 5)); 429 // Shift must be "lsl #32", in bits 22:21 430 assert((*TargetPtr >> 21 & 0x3) == 2 && "invalid shift for relocation"); 431 break; 432 } 433 case ELF::R_AARCH64_MOVW_UABS_G1_NC: { 434 uint64_t Result = Value + Addend; 435 436 // AArch64 code is emitted with .rela relocations. The data already in any 437 // bits affected by the relocation on entry is garbage. 438 *TargetPtr &= 0xffe0001fU; 439 // Immediate goes in bits 20:5 of MOVZ/MOVK instruction 440 *TargetPtr |= ((Result & 0xffff0000U) >> (16 - 5)); 441 // Shift must be "lsl #16", in bits 22:2 442 assert((*TargetPtr >> 21 & 0x3) == 1 && "invalid shift for relocation"); 443 break; 444 } 445 case ELF::R_AARCH64_MOVW_UABS_G0_NC: { 446 uint64_t Result = Value + Addend; 447 448 // AArch64 code is emitted with .rela relocations. The data already in any 449 // bits affected by the relocation on entry is garbage. 450 *TargetPtr &= 0xffe0001fU; 451 // Immediate goes in bits 20:5 of MOVZ/MOVK instruction 452 *TargetPtr |= ((Result & 0xffffU) << 5); 453 // Shift must be "lsl #0", in bits 22:21. 454 assert((*TargetPtr >> 21 & 0x3) == 0 && "invalid shift for relocation"); 455 break; 456 } 457 case ELF::R_AARCH64_ADR_PREL_PG_HI21: { 458 // Operation: Page(S+A) - Page(P) 459 uint64_t Result = ((Value + Addend) & ~0xfffULL) - (FinalAddress & ~0xfffULL); 460 461 // Check that -2^32 <= X < 2^32 462 assert(static_cast<int64_t>(Result) >= (-1LL << 32) && 463 static_cast<int64_t>(Result) < (1LL << 32) && 464 "overflow check failed for relocation"); 465 466 // AArch64 code is emitted with .rela relocations. The data already in any 467 // bits affected by the relocation on entry is garbage. 468 *TargetPtr &= 0x9f00001fU; 469 // Immediate goes in bits 30:29 + 5:23 of ADRP instruction, taken 470 // from bits 32:12 of X. 471 *TargetPtr |= ((Result & 0x3000U) << (29 - 12)); 472 *TargetPtr |= ((Result & 0x1ffffc000ULL) >> (14 - 5)); 473 break; 474 } 475 case ELF::R_AARCH64_LDST32_ABS_LO12_NC: { 476 // Operation: S + A 477 uint64_t Result = Value + Addend; 478 479 // AArch64 code is emitted with .rela relocations. The data already in any 480 // bits affected by the relocation on entry is garbage. 481 *TargetPtr &= 0xffc003ffU; 482 // Immediate goes in bits 21:10 of LD/ST instruction, taken 483 // from bits 11:2 of X 484 *TargetPtr |= ((Result & 0xffc) << (10 - 2)); 485 break; 486 } 487 case ELF::R_AARCH64_LDST64_ABS_LO12_NC: { 488 // Operation: S + A 489 uint64_t Result = Value + Addend; 490 491 // AArch64 code is emitted with .rela relocations. The data already in any 492 // bits affected by the relocation on entry is garbage. 493 *TargetPtr &= 0xffc003ffU; 494 // Immediate goes in bits 21:10 of LD/ST instruction, taken 495 // from bits 11:3 of X 496 *TargetPtr |= ((Result & 0xff8) << (10 - 3)); 497 break; 498 } 499 } 500 } 501 502 void RuntimeDyldELF::resolveARMRelocation(const SectionEntry &Section, 503 uint64_t Offset, 504 uint32_t Value, 505 uint32_t Type, 506 int32_t Addend) { 507 // TODO: Add Thumb relocations. 508 uint32_t *Placeholder = reinterpret_cast<uint32_t*>(Section.ObjAddress + 509 Offset); 510 uint32_t* TargetPtr = (uint32_t*)(Section.Address + Offset); 511 uint32_t FinalAddress = ((Section.LoadAddress + Offset) & 0xFFFFFFFF); 512 Value += Addend; 513 514 DEBUG(dbgs() << "resolveARMRelocation, LocalAddress: " 515 << Section.Address + Offset 516 << " FinalAddress: " << format("%p",FinalAddress) 517 << " Value: " << format("%x",Value) 518 << " Type: " << format("%x",Type) 519 << " Addend: " << format("%x",Addend) 520 << "\n"); 521 522 switch(Type) { 523 default: 524 llvm_unreachable("Not implemented relocation type!"); 525 526 case ELF::R_ARM_NONE: 527 break; 528 // Write a 32bit value to relocation address, taking into account the 529 // implicit addend encoded in the target. 530 case ELF::R_ARM_PREL31: 531 case ELF::R_ARM_TARGET1: 532 case ELF::R_ARM_ABS32: 533 *TargetPtr = *Placeholder + Value; 534 break; 535 // Write first 16 bit of 32 bit value to the mov instruction. 536 // Last 4 bit should be shifted. 537 case ELF::R_ARM_MOVW_ABS_NC: 538 // We are not expecting any other addend in the relocation address. 539 // Using 0x000F0FFF because MOVW has its 16 bit immediate split into 2 540 // non-contiguous fields. 541 assert((*Placeholder & 0x000F0FFF) == 0); 542 Value = Value & 0xFFFF; 543 *TargetPtr = *Placeholder | (Value & 0xFFF); 544 *TargetPtr |= ((Value >> 12) & 0xF) << 16; 545 break; 546 // Write last 16 bit of 32 bit value to the mov instruction. 547 // Last 4 bit should be shifted. 548 case ELF::R_ARM_MOVT_ABS: 549 // We are not expecting any other addend in the relocation address. 550 // Use 0x000F0FFF for the same reason as R_ARM_MOVW_ABS_NC. 551 assert((*Placeholder & 0x000F0FFF) == 0); 552 553 Value = (Value >> 16) & 0xFFFF; 554 *TargetPtr = *Placeholder | (Value & 0xFFF); 555 *TargetPtr |= ((Value >> 12) & 0xF) << 16; 556 break; 557 // Write 24 bit relative value to the branch instruction. 558 case ELF::R_ARM_PC24 : // Fall through. 559 case ELF::R_ARM_CALL : // Fall through. 560 case ELF::R_ARM_JUMP24: { 561 int32_t RelValue = static_cast<int32_t>(Value - FinalAddress - 8); 562 RelValue = (RelValue & 0x03FFFFFC) >> 2; 563 assert((*TargetPtr & 0xFFFFFF) == 0xFFFFFE); 564 *TargetPtr &= 0xFF000000; 565 *TargetPtr |= RelValue; 566 break; 567 } 568 case ELF::R_ARM_PRIVATE_0: 569 // This relocation is reserved by the ARM ELF ABI for internal use. We 570 // appropriate it here to act as an R_ARM_ABS32 without any addend for use 571 // in the stubs created during JIT (which can't put an addend into the 572 // original object file). 573 *TargetPtr = Value; 574 break; 575 } 576 } 577 578 void RuntimeDyldELF::resolveMIPSRelocation(const SectionEntry &Section, 579 uint64_t Offset, 580 uint32_t Value, 581 uint32_t Type, 582 int32_t Addend) { 583 uint32_t *Placeholder = reinterpret_cast<uint32_t*>(Section.ObjAddress + 584 Offset); 585 uint32_t* TargetPtr = (uint32_t*)(Section.Address + Offset); 586 Value += Addend; 587 588 DEBUG(dbgs() << "resolveMipselocation, LocalAddress: " 589 << Section.Address + Offset 590 << " FinalAddress: " 591 << format("%p",Section.LoadAddress + Offset) 592 << " Value: " << format("%x",Value) 593 << " Type: " << format("%x",Type) 594 << " Addend: " << format("%x",Addend) 595 << "\n"); 596 597 switch(Type) { 598 default: 599 llvm_unreachable("Not implemented relocation type!"); 600 break; 601 case ELF::R_MIPS_32: 602 *TargetPtr = Value + (*Placeholder); 603 break; 604 case ELF::R_MIPS_26: 605 *TargetPtr = ((*Placeholder) & 0xfc000000) | (( Value & 0x0fffffff) >> 2); 606 break; 607 case ELF::R_MIPS_HI16: 608 // Get the higher 16-bits. Also add 1 if bit 15 is 1. 609 Value += ((*Placeholder) & 0x0000ffff) << 16; 610 *TargetPtr = ((*Placeholder) & 0xffff0000) | 611 (((Value + 0x8000) >> 16) & 0xffff); 612 break; 613 case ELF::R_MIPS_LO16: 614 Value += ((*Placeholder) & 0x0000ffff); 615 *TargetPtr = ((*Placeholder) & 0xffff0000) | (Value & 0xffff); 616 break; 617 case ELF::R_MIPS_UNUSED1: 618 // Similar to ELF::R_ARM_PRIVATE_0, R_MIPS_UNUSED1 and R_MIPS_UNUSED2 619 // are used for internal JIT purpose. These relocations are similar to 620 // R_MIPS_HI16 and R_MIPS_LO16, but they do not take any addend into 621 // account. 622 *TargetPtr = ((*TargetPtr) & 0xffff0000) | 623 (((Value + 0x8000) >> 16) & 0xffff); 624 break; 625 case ELF::R_MIPS_UNUSED2: 626 *TargetPtr = ((*TargetPtr) & 0xffff0000) | (Value & 0xffff); 627 break; 628 } 629 } 630 631 // Return the .TOC. section address to R_PPC64_TOC relocations. 632 uint64_t RuntimeDyldELF::findPPC64TOC() const { 633 // The TOC consists of sections .got, .toc, .tocbss, .plt in that 634 // order. The TOC starts where the first of these sections starts. 635 SectionList::const_iterator it = Sections.begin(); 636 SectionList::const_iterator ite = Sections.end(); 637 for (; it != ite; ++it) { 638 if (it->Name == ".got" || 639 it->Name == ".toc" || 640 it->Name == ".tocbss" || 641 it->Name == ".plt") 642 break; 643 } 644 if (it == ite) { 645 // This may happen for 646 // * references to TOC base base (sym@toc, .odp relocation) without 647 // a .toc directive. 648 // In this case just use the first section (which is usually 649 // the .odp) since the code won't reference the .toc base 650 // directly. 651 it = Sections.begin(); 652 } 653 assert (it != ite); 654 // Per the ppc64-elf-linux ABI, The TOC base is TOC value plus 0x8000 655 // thus permitting a full 64 Kbytes segment. 656 return it->LoadAddress + 0x8000; 657 } 658 659 // Returns the sections and offset associated with the ODP entry referenced 660 // by Symbol. 661 void RuntimeDyldELF::findOPDEntrySection(ObjectImage &Obj, 662 ObjSectionToIDMap &LocalSections, 663 RelocationValueRef &Rel) { 664 // Get the ELF symbol value (st_value) to compare with Relocation offset in 665 // .opd entries 666 for (section_iterator si = Obj.begin_sections(), se = Obj.end_sections(); 667 si != se; ++si) { 668 section_iterator RelSecI = si->getRelocatedSection(); 669 if (RelSecI == Obj.end_sections()) 670 continue; 671 672 StringRef RelSectionName; 673 check(RelSecI->getName(RelSectionName)); 674 if (RelSectionName != ".opd") 675 continue; 676 677 for (relocation_iterator i = si->relocation_begin(), 678 e = si->relocation_end(); i != e;) { 679 // The R_PPC64_ADDR64 relocation indicates the first field 680 // of a .opd entry 681 uint64_t TypeFunc; 682 check(i->getType(TypeFunc)); 683 if (TypeFunc != ELF::R_PPC64_ADDR64) { 684 ++i; 685 continue; 686 } 687 688 uint64_t TargetSymbolOffset; 689 symbol_iterator TargetSymbol = i->getSymbol(); 690 check(i->getOffset(TargetSymbolOffset)); 691 int64_t Addend; 692 check(getELFRelocationAddend(*i, Addend)); 693 694 ++i; 695 if (i == e) 696 break; 697 698 // Just check if following relocation is a R_PPC64_TOC 699 uint64_t TypeTOC; 700 check(i->getType(TypeTOC)); 701 if (TypeTOC != ELF::R_PPC64_TOC) 702 continue; 703 704 // Finally compares the Symbol value and the target symbol offset 705 // to check if this .opd entry refers to the symbol the relocation 706 // points to. 707 if (Rel.Addend != (int64_t)TargetSymbolOffset) 708 continue; 709 710 section_iterator tsi(Obj.end_sections()); 711 check(TargetSymbol->getSection(tsi)); 712 Rel.SectionID = findOrEmitSection(Obj, (*tsi), true, LocalSections); 713 Rel.Addend = (intptr_t)Addend; 714 return; 715 } 716 } 717 llvm_unreachable("Attempting to get address of ODP entry!"); 718 } 719 720 // Relocation masks following the #lo(value), #hi(value), #higher(value), 721 // and #highest(value) macros defined in section 4.5.1. Relocation Types 722 // in PPC-elf64abi document. 723 // 724 static inline 725 uint16_t applyPPClo (uint64_t value) 726 { 727 return value & 0xffff; 728 } 729 730 static inline 731 uint16_t applyPPChi (uint64_t value) 732 { 733 return (value >> 16) & 0xffff; 734 } 735 736 static inline 737 uint16_t applyPPChigher (uint64_t value) 738 { 739 return (value >> 32) & 0xffff; 740 } 741 742 static inline 743 uint16_t applyPPChighest (uint64_t value) 744 { 745 return (value >> 48) & 0xffff; 746 } 747 748 void RuntimeDyldELF::resolvePPC64Relocation(const SectionEntry &Section, 749 uint64_t Offset, 750 uint64_t Value, 751 uint32_t Type, 752 int64_t Addend) { 753 uint8_t* LocalAddress = Section.Address + Offset; 754 switch (Type) { 755 default: 756 llvm_unreachable("Relocation type not implemented yet!"); 757 break; 758 case ELF::R_PPC64_ADDR16_LO : 759 writeInt16BE(LocalAddress, applyPPClo (Value + Addend)); 760 break; 761 case ELF::R_PPC64_ADDR16_HI : 762 writeInt16BE(LocalAddress, applyPPChi (Value + Addend)); 763 break; 764 case ELF::R_PPC64_ADDR16_HIGHER : 765 writeInt16BE(LocalAddress, applyPPChigher (Value + Addend)); 766 break; 767 case ELF::R_PPC64_ADDR16_HIGHEST : 768 writeInt16BE(LocalAddress, applyPPChighest (Value + Addend)); 769 break; 770 case ELF::R_PPC64_ADDR14 : { 771 assert(((Value + Addend) & 3) == 0); 772 // Preserve the AA/LK bits in the branch instruction 773 uint8_t aalk = *(LocalAddress+3); 774 writeInt16BE(LocalAddress + 2, (aalk & 3) | ((Value + Addend) & 0xfffc)); 775 } break; 776 case ELF::R_PPC64_ADDR32 : { 777 int32_t Result = static_cast<int32_t>(Value + Addend); 778 if (SignExtend32<32>(Result) != Result) 779 llvm_unreachable("Relocation R_PPC64_ADDR32 overflow"); 780 writeInt32BE(LocalAddress, Result); 781 } break; 782 case ELF::R_PPC64_REL24 : { 783 uint64_t FinalAddress = (Section.LoadAddress + Offset); 784 int32_t delta = static_cast<int32_t>(Value - FinalAddress + Addend); 785 if (SignExtend32<24>(delta) != delta) 786 llvm_unreachable("Relocation R_PPC64_REL24 overflow"); 787 // Generates a 'bl <address>' instruction 788 writeInt32BE(LocalAddress, 0x48000001 | (delta & 0x03FFFFFC)); 789 } break; 790 case ELF::R_PPC64_REL32 : { 791 uint64_t FinalAddress = (Section.LoadAddress + Offset); 792 int32_t delta = static_cast<int32_t>(Value - FinalAddress + Addend); 793 if (SignExtend32<32>(delta) != delta) 794 llvm_unreachable("Relocation R_PPC64_REL32 overflow"); 795 writeInt32BE(LocalAddress, delta); 796 } break; 797 case ELF::R_PPC64_REL64: { 798 uint64_t FinalAddress = (Section.LoadAddress + Offset); 799 uint64_t Delta = Value - FinalAddress + Addend; 800 writeInt64BE(LocalAddress, Delta); 801 } break; 802 case ELF::R_PPC64_ADDR64 : 803 writeInt64BE(LocalAddress, Value + Addend); 804 break; 805 case ELF::R_PPC64_TOC : 806 writeInt64BE(LocalAddress, findPPC64TOC()); 807 break; 808 case ELF::R_PPC64_TOC16 : { 809 uint64_t TOCStart = findPPC64TOC(); 810 Value = applyPPClo((Value + Addend) - TOCStart); 811 writeInt16BE(LocalAddress, applyPPClo(Value)); 812 } break; 813 case ELF::R_PPC64_TOC16_DS : { 814 uint64_t TOCStart = findPPC64TOC(); 815 Value = ((Value + Addend) - TOCStart); 816 writeInt16BE(LocalAddress, applyPPClo(Value)); 817 } break; 818 } 819 } 820 821 void RuntimeDyldELF::resolveSystemZRelocation(const SectionEntry &Section, 822 uint64_t Offset, 823 uint64_t Value, 824 uint32_t Type, 825 int64_t Addend) { 826 uint8_t *LocalAddress = Section.Address + Offset; 827 switch (Type) { 828 default: 829 llvm_unreachable("Relocation type not implemented yet!"); 830 break; 831 case ELF::R_390_PC16DBL: 832 case ELF::R_390_PLT16DBL: { 833 int64_t Delta = (Value + Addend) - (Section.LoadAddress + Offset); 834 assert(int16_t(Delta / 2) * 2 == Delta && "R_390_PC16DBL overflow"); 835 writeInt16BE(LocalAddress, Delta / 2); 836 break; 837 } 838 case ELF::R_390_PC32DBL: 839 case ELF::R_390_PLT32DBL: { 840 int64_t Delta = (Value + Addend) - (Section.LoadAddress + Offset); 841 assert(int32_t(Delta / 2) * 2 == Delta && "R_390_PC32DBL overflow"); 842 writeInt32BE(LocalAddress, Delta / 2); 843 break; 844 } 845 case ELF::R_390_PC32: { 846 int64_t Delta = (Value + Addend) - (Section.LoadAddress + Offset); 847 assert(int32_t(Delta) == Delta && "R_390_PC32 overflow"); 848 writeInt32BE(LocalAddress, Delta); 849 break; 850 } 851 case ELF::R_390_64: 852 writeInt64BE(LocalAddress, Value + Addend); 853 break; 854 } 855 } 856 857 // The target location for the relocation is described by RE.SectionID and 858 // RE.Offset. RE.SectionID can be used to find the SectionEntry. Each 859 // SectionEntry has three members describing its location. 860 // SectionEntry::Address is the address at which the section has been loaded 861 // into memory in the current (host) process. SectionEntry::LoadAddress is the 862 // address that the section will have in the target process. 863 // SectionEntry::ObjAddress is the address of the bits for this section in the 864 // original emitted object image (also in the current address space). 865 // 866 // Relocations will be applied as if the section were loaded at 867 // SectionEntry::LoadAddress, but they will be applied at an address based 868 // on SectionEntry::Address. SectionEntry::ObjAddress will be used to refer to 869 // Target memory contents if they are required for value calculations. 870 // 871 // The Value parameter here is the load address of the symbol for the 872 // relocation to be applied. For relocations which refer to symbols in the 873 // current object Value will be the LoadAddress of the section in which 874 // the symbol resides (RE.Addend provides additional information about the 875 // symbol location). For external symbols, Value will be the address of the 876 // symbol in the target address space. 877 void RuntimeDyldELF::resolveRelocation(const RelocationEntry &RE, 878 uint64_t Value) { 879 const SectionEntry &Section = Sections[RE.SectionID]; 880 return resolveRelocation(Section, RE.Offset, Value, RE.RelType, RE.Addend, 881 RE.SymOffset); 882 } 883 884 void RuntimeDyldELF::resolveRelocation(const SectionEntry &Section, 885 uint64_t Offset, 886 uint64_t Value, 887 uint32_t Type, 888 int64_t Addend, 889 uint64_t SymOffset) { 890 switch (Arch) { 891 case Triple::x86_64: 892 resolveX86_64Relocation(Section, Offset, Value, Type, Addend, SymOffset); 893 break; 894 case Triple::x86: 895 resolveX86Relocation(Section, Offset, 896 (uint32_t)(Value & 0xffffffffL), Type, 897 (uint32_t)(Addend & 0xffffffffL)); 898 break; 899 case Triple::aarch64: 900 resolveAArch64Relocation(Section, Offset, Value, Type, Addend); 901 break; 902 case Triple::arm: // Fall through. 903 case Triple::thumb: 904 resolveARMRelocation(Section, Offset, 905 (uint32_t)(Value & 0xffffffffL), Type, 906 (uint32_t)(Addend & 0xffffffffL)); 907 break; 908 case Triple::mips: // Fall through. 909 case Triple::mipsel: 910 resolveMIPSRelocation(Section, Offset, 911 (uint32_t)(Value & 0xffffffffL), Type, 912 (uint32_t)(Addend & 0xffffffffL)); 913 break; 914 case Triple::ppc64: // Fall through. 915 case Triple::ppc64le: 916 resolvePPC64Relocation(Section, Offset, Value, Type, Addend); 917 break; 918 case Triple::systemz: 919 resolveSystemZRelocation(Section, Offset, Value, Type, Addend); 920 break; 921 default: llvm_unreachable("Unsupported CPU type!"); 922 } 923 } 924 925 void RuntimeDyldELF::processRelocationRef(unsigned SectionID, 926 RelocationRef RelI, 927 ObjectImage &Obj, 928 ObjSectionToIDMap &ObjSectionToID, 929 const SymbolTableMap &Symbols, 930 StubMap &Stubs) { 931 uint64_t RelType; 932 Check(RelI.getType(RelType)); 933 int64_t Addend; 934 Check(getELFRelocationAddend(RelI, Addend)); 935 symbol_iterator Symbol = RelI.getSymbol(); 936 937 // Obtain the symbol name which is referenced in the relocation 938 StringRef TargetName; 939 if (Symbol != Obj.end_symbols()) 940 Symbol->getName(TargetName); 941 DEBUG(dbgs() << "\t\tRelType: " << RelType 942 << " Addend: " << Addend 943 << " TargetName: " << TargetName 944 << "\n"); 945 RelocationValueRef Value; 946 // First search for the symbol in the local symbol table 947 SymbolTableMap::const_iterator lsi = Symbols.end(); 948 SymbolRef::Type SymType = SymbolRef::ST_Unknown; 949 if (Symbol != Obj.end_symbols()) { 950 lsi = Symbols.find(TargetName.data()); 951 Symbol->getType(SymType); 952 } 953 if (lsi != Symbols.end()) { 954 Value.SectionID = lsi->second.first; 955 Value.Offset = lsi->second.second; 956 Value.Addend = lsi->second.second + Addend; 957 } else { 958 // Search for the symbol in the global symbol table 959 SymbolTableMap::const_iterator gsi = GlobalSymbolTable.end(); 960 if (Symbol != Obj.end_symbols()) 961 gsi = GlobalSymbolTable.find(TargetName.data()); 962 if (gsi != GlobalSymbolTable.end()) { 963 Value.SectionID = gsi->second.first; 964 Value.Offset = gsi->second.second; 965 Value.Addend = gsi->second.second + Addend; 966 } else { 967 switch (SymType) { 968 case SymbolRef::ST_Debug: { 969 // TODO: Now ELF SymbolRef::ST_Debug = STT_SECTION, it's not obviously 970 // and can be changed by another developers. Maybe best way is add 971 // a new symbol type ST_Section to SymbolRef and use it. 972 section_iterator si(Obj.end_sections()); 973 Symbol->getSection(si); 974 if (si == Obj.end_sections()) 975 llvm_unreachable("Symbol section not found, bad object file format!"); 976 DEBUG(dbgs() << "\t\tThis is section symbol\n"); 977 // Default to 'true' in case isText fails (though it never does). 978 bool isCode = true; 979 si->isText(isCode); 980 Value.SectionID = findOrEmitSection(Obj, 981 (*si), 982 isCode, 983 ObjSectionToID); 984 Value.Addend = Addend; 985 break; 986 } 987 case SymbolRef::ST_Data: 988 case SymbolRef::ST_Unknown: { 989 Value.SymbolName = TargetName.data(); 990 Value.Addend = Addend; 991 992 // Absolute relocations will have a zero symbol ID (STN_UNDEF), which 993 // will manifest here as a NULL symbol name. 994 // We can set this as a valid (but empty) symbol name, and rely 995 // on addRelocationForSymbol to handle this. 996 if (!Value.SymbolName) 997 Value.SymbolName = ""; 998 break; 999 } 1000 default: 1001 llvm_unreachable("Unresolved symbol type!"); 1002 break; 1003 } 1004 } 1005 } 1006 uint64_t Offset; 1007 Check(RelI.getOffset(Offset)); 1008 1009 DEBUG(dbgs() << "\t\tSectionID: " << SectionID 1010 << " Offset: " << Offset 1011 << "\n"); 1012 if (Arch == Triple::aarch64 && 1013 (RelType == ELF::R_AARCH64_CALL26 || 1014 RelType == ELF::R_AARCH64_JUMP26)) { 1015 // This is an AArch64 branch relocation, need to use a stub function. 1016 DEBUG(dbgs() << "\t\tThis is an AArch64 branch relocation."); 1017 SectionEntry &Section = Sections[SectionID]; 1018 1019 // Look for an existing stub. 1020 StubMap::const_iterator i = Stubs.find(Value); 1021 if (i != Stubs.end()) { 1022 resolveRelocation(Section, Offset, 1023 (uint64_t)Section.Address + i->second, RelType, 0); 1024 DEBUG(dbgs() << " Stub function found\n"); 1025 } else { 1026 // Create a new stub function. 1027 DEBUG(dbgs() << " Create a new stub function\n"); 1028 Stubs[Value] = Section.StubOffset; 1029 uint8_t *StubTargetAddr = createStubFunction(Section.Address + 1030 Section.StubOffset); 1031 1032 RelocationEntry REmovz_g3(SectionID, 1033 StubTargetAddr - Section.Address, 1034 ELF::R_AARCH64_MOVW_UABS_G3, Value.Addend); 1035 RelocationEntry REmovk_g2(SectionID, 1036 StubTargetAddr - Section.Address + 4, 1037 ELF::R_AARCH64_MOVW_UABS_G2_NC, Value.Addend); 1038 RelocationEntry REmovk_g1(SectionID, 1039 StubTargetAddr - Section.Address + 8, 1040 ELF::R_AARCH64_MOVW_UABS_G1_NC, Value.Addend); 1041 RelocationEntry REmovk_g0(SectionID, 1042 StubTargetAddr - Section.Address + 12, 1043 ELF::R_AARCH64_MOVW_UABS_G0_NC, Value.Addend); 1044 1045 if (Value.SymbolName) { 1046 addRelocationForSymbol(REmovz_g3, Value.SymbolName); 1047 addRelocationForSymbol(REmovk_g2, Value.SymbolName); 1048 addRelocationForSymbol(REmovk_g1, Value.SymbolName); 1049 addRelocationForSymbol(REmovk_g0, Value.SymbolName); 1050 } else { 1051 addRelocationForSection(REmovz_g3, Value.SectionID); 1052 addRelocationForSection(REmovk_g2, Value.SectionID); 1053 addRelocationForSection(REmovk_g1, Value.SectionID); 1054 addRelocationForSection(REmovk_g0, Value.SectionID); 1055 } 1056 resolveRelocation(Section, Offset, 1057 (uint64_t)Section.Address + Section.StubOffset, 1058 RelType, 0); 1059 Section.StubOffset += getMaxStubSize(); 1060 } 1061 } else if (Arch == Triple::arm && 1062 (RelType == ELF::R_ARM_PC24 || 1063 RelType == ELF::R_ARM_CALL || 1064 RelType == ELF::R_ARM_JUMP24)) { 1065 // This is an ARM branch relocation, need to use a stub function. 1066 DEBUG(dbgs() << "\t\tThis is an ARM branch relocation."); 1067 SectionEntry &Section = Sections[SectionID]; 1068 1069 // Look for an existing stub. 1070 StubMap::const_iterator i = Stubs.find(Value); 1071 if (i != Stubs.end()) { 1072 resolveRelocation(Section, Offset, 1073 (uint64_t)Section.Address + i->second, RelType, 0); 1074 DEBUG(dbgs() << " Stub function found\n"); 1075 } else { 1076 // Create a new stub function. 1077 DEBUG(dbgs() << " Create a new stub function\n"); 1078 Stubs[Value] = Section.StubOffset; 1079 uint8_t *StubTargetAddr = createStubFunction(Section.Address + 1080 Section.StubOffset); 1081 RelocationEntry RE(SectionID, StubTargetAddr - Section.Address, 1082 ELF::R_ARM_PRIVATE_0, Value.Addend); 1083 if (Value.SymbolName) 1084 addRelocationForSymbol(RE, Value.SymbolName); 1085 else 1086 addRelocationForSection(RE, Value.SectionID); 1087 1088 resolveRelocation(Section, Offset, 1089 (uint64_t)Section.Address + Section.StubOffset, 1090 RelType, 0); 1091 Section.StubOffset += getMaxStubSize(); 1092 } 1093 } else if ((Arch == Triple::mipsel || Arch == Triple::mips) && 1094 RelType == ELF::R_MIPS_26) { 1095 // This is an Mips branch relocation, need to use a stub function. 1096 DEBUG(dbgs() << "\t\tThis is a Mips branch relocation."); 1097 SectionEntry &Section = Sections[SectionID]; 1098 uint8_t *Target = Section.Address + Offset; 1099 uint32_t *TargetAddress = (uint32_t *)Target; 1100 1101 // Extract the addend from the instruction. 1102 uint32_t Addend = ((*TargetAddress) & 0x03ffffff) << 2; 1103 1104 Value.Addend += Addend; 1105 1106 // Look up for existing stub. 1107 StubMap::const_iterator i = Stubs.find(Value); 1108 if (i != Stubs.end()) { 1109 RelocationEntry RE(SectionID, Offset, RelType, i->second); 1110 addRelocationForSection(RE, SectionID); 1111 DEBUG(dbgs() << " Stub function found\n"); 1112 } else { 1113 // Create a new stub function. 1114 DEBUG(dbgs() << " Create a new stub function\n"); 1115 Stubs[Value] = Section.StubOffset; 1116 uint8_t *StubTargetAddr = createStubFunction(Section.Address + 1117 Section.StubOffset); 1118 1119 // Creating Hi and Lo relocations for the filled stub instructions. 1120 RelocationEntry REHi(SectionID, 1121 StubTargetAddr - Section.Address, 1122 ELF::R_MIPS_UNUSED1, Value.Addend); 1123 RelocationEntry RELo(SectionID, 1124 StubTargetAddr - Section.Address + 4, 1125 ELF::R_MIPS_UNUSED2, Value.Addend); 1126 1127 if (Value.SymbolName) { 1128 addRelocationForSymbol(REHi, Value.SymbolName); 1129 addRelocationForSymbol(RELo, Value.SymbolName); 1130 } else { 1131 addRelocationForSection(REHi, Value.SectionID); 1132 addRelocationForSection(RELo, Value.SectionID); 1133 } 1134 1135 RelocationEntry RE(SectionID, Offset, RelType, Section.StubOffset); 1136 addRelocationForSection(RE, SectionID); 1137 Section.StubOffset += getMaxStubSize(); 1138 } 1139 } else if (Arch == Triple::ppc64 || Arch == Triple::ppc64le) { 1140 if (RelType == ELF::R_PPC64_REL24) { 1141 // A PPC branch relocation will need a stub function if the target is 1142 // an external symbol (Symbol::ST_Unknown) or if the target address 1143 // is not within the signed 24-bits branch address. 1144 SectionEntry &Section = Sections[SectionID]; 1145 uint8_t *Target = Section.Address + Offset; 1146 bool RangeOverflow = false; 1147 if (SymType != SymbolRef::ST_Unknown) { 1148 // A function call may points to the .opd entry, so the final symbol value 1149 // in calculated based in the relocation values in .opd section. 1150 findOPDEntrySection(Obj, ObjSectionToID, Value); 1151 uint8_t *RelocTarget = Sections[Value.SectionID].Address + Value.Addend; 1152 int32_t delta = static_cast<int32_t>(Target - RelocTarget); 1153 // If it is within 24-bits branch range, just set the branch target 1154 if (SignExtend32<24>(delta) == delta) { 1155 RelocationEntry RE(SectionID, Offset, RelType, Value.Addend); 1156 if (Value.SymbolName) 1157 addRelocationForSymbol(RE, Value.SymbolName); 1158 else 1159 addRelocationForSection(RE, Value.SectionID); 1160 } else { 1161 RangeOverflow = true; 1162 } 1163 } 1164 if (SymType == SymbolRef::ST_Unknown || RangeOverflow == true) { 1165 // It is an external symbol (SymbolRef::ST_Unknown) or within a range 1166 // larger than 24-bits. 1167 StubMap::const_iterator i = Stubs.find(Value); 1168 if (i != Stubs.end()) { 1169 // Symbol function stub already created, just relocate to it 1170 resolveRelocation(Section, Offset, 1171 (uint64_t)Section.Address + i->second, RelType, 0); 1172 DEBUG(dbgs() << " Stub function found\n"); 1173 } else { 1174 // Create a new stub function. 1175 DEBUG(dbgs() << " Create a new stub function\n"); 1176 Stubs[Value] = Section.StubOffset; 1177 uint8_t *StubTargetAddr = createStubFunction(Section.Address + 1178 Section.StubOffset); 1179 RelocationEntry RE(SectionID, StubTargetAddr - Section.Address, 1180 ELF::R_PPC64_ADDR64, Value.Addend); 1181 1182 // Generates the 64-bits address loads as exemplified in section 1183 // 4.5.1 in PPC64 ELF ABI. 1184 RelocationEntry REhst(SectionID, 1185 StubTargetAddr - Section.Address + 2, 1186 ELF::R_PPC64_ADDR16_HIGHEST, Value.Addend); 1187 RelocationEntry REhr(SectionID, 1188 StubTargetAddr - Section.Address + 6, 1189 ELF::R_PPC64_ADDR16_HIGHER, Value.Addend); 1190 RelocationEntry REh(SectionID, 1191 StubTargetAddr - Section.Address + 14, 1192 ELF::R_PPC64_ADDR16_HI, Value.Addend); 1193 RelocationEntry REl(SectionID, 1194 StubTargetAddr - Section.Address + 18, 1195 ELF::R_PPC64_ADDR16_LO, Value.Addend); 1196 1197 if (Value.SymbolName) { 1198 addRelocationForSymbol(REhst, Value.SymbolName); 1199 addRelocationForSymbol(REhr, Value.SymbolName); 1200 addRelocationForSymbol(REh, Value.SymbolName); 1201 addRelocationForSymbol(REl, Value.SymbolName); 1202 } else { 1203 addRelocationForSection(REhst, Value.SectionID); 1204 addRelocationForSection(REhr, Value.SectionID); 1205 addRelocationForSection(REh, Value.SectionID); 1206 addRelocationForSection(REl, Value.SectionID); 1207 } 1208 1209 resolveRelocation(Section, Offset, 1210 (uint64_t)Section.Address + Section.StubOffset, 1211 RelType, 0); 1212 if (SymType == SymbolRef::ST_Unknown) 1213 // Restore the TOC for external calls 1214 writeInt32BE(Target+4, 0xE8410028); // ld r2,40(r1) 1215 Section.StubOffset += getMaxStubSize(); 1216 } 1217 } 1218 } else { 1219 RelocationEntry RE(SectionID, Offset, RelType, Value.Addend); 1220 // Extra check to avoid relocation againt empty symbols (usually 1221 // the R_PPC64_TOC). 1222 if (SymType != SymbolRef::ST_Unknown && TargetName.empty()) 1223 Value.SymbolName = NULL; 1224 1225 if (Value.SymbolName) 1226 addRelocationForSymbol(RE, Value.SymbolName); 1227 else 1228 addRelocationForSection(RE, Value.SectionID); 1229 } 1230 } else if (Arch == Triple::systemz && 1231 (RelType == ELF::R_390_PLT32DBL || 1232 RelType == ELF::R_390_GOTENT)) { 1233 // Create function stubs for both PLT and GOT references, regardless of 1234 // whether the GOT reference is to data or code. The stub contains the 1235 // full address of the symbol, as needed by GOT references, and the 1236 // executable part only adds an overhead of 8 bytes. 1237 // 1238 // We could try to conserve space by allocating the code and data 1239 // parts of the stub separately. However, as things stand, we allocate 1240 // a stub for every relocation, so using a GOT in JIT code should be 1241 // no less space efficient than using an explicit constant pool. 1242 DEBUG(dbgs() << "\t\tThis is a SystemZ indirect relocation."); 1243 SectionEntry &Section = Sections[SectionID]; 1244 1245 // Look for an existing stub. 1246 StubMap::const_iterator i = Stubs.find(Value); 1247 uintptr_t StubAddress; 1248 if (i != Stubs.end()) { 1249 StubAddress = uintptr_t(Section.Address) + i->second; 1250 DEBUG(dbgs() << " Stub function found\n"); 1251 } else { 1252 // Create a new stub function. 1253 DEBUG(dbgs() << " Create a new stub function\n"); 1254 1255 uintptr_t BaseAddress = uintptr_t(Section.Address); 1256 uintptr_t StubAlignment = getStubAlignment(); 1257 StubAddress = (BaseAddress + Section.StubOffset + 1258 StubAlignment - 1) & -StubAlignment; 1259 unsigned StubOffset = StubAddress - BaseAddress; 1260 1261 Stubs[Value] = StubOffset; 1262 createStubFunction((uint8_t *)StubAddress); 1263 RelocationEntry RE(SectionID, StubOffset + 8, 1264 ELF::R_390_64, Value.Addend - Addend); 1265 if (Value.SymbolName) 1266 addRelocationForSymbol(RE, Value.SymbolName); 1267 else 1268 addRelocationForSection(RE, Value.SectionID); 1269 Section.StubOffset = StubOffset + getMaxStubSize(); 1270 } 1271 1272 if (RelType == ELF::R_390_GOTENT) 1273 resolveRelocation(Section, Offset, StubAddress + 8, 1274 ELF::R_390_PC32DBL, Addend); 1275 else 1276 resolveRelocation(Section, Offset, StubAddress, RelType, Addend); 1277 } else if (Arch == Triple::x86_64 && RelType == ELF::R_X86_64_PLT32) { 1278 // The way the PLT relocations normally work is that the linker allocates the 1279 // PLT and this relocation makes a PC-relative call into the PLT. The PLT 1280 // entry will then jump to an address provided by the GOT. On first call, the 1281 // GOT address will point back into PLT code that resolves the symbol. After 1282 // the first call, the GOT entry points to the actual function. 1283 // 1284 // For local functions we're ignoring all of that here and just replacing 1285 // the PLT32 relocation type with PC32, which will translate the relocation 1286 // into a PC-relative call directly to the function. For external symbols we 1287 // can't be sure the function will be within 2^32 bytes of the call site, so 1288 // we need to create a stub, which calls into the GOT. This case is 1289 // equivalent to the usual PLT implementation except that we use the stub 1290 // mechanism in RuntimeDyld (which puts stubs at the end of the section) 1291 // rather than allocating a PLT section. 1292 if (Value.SymbolName) { 1293 // This is a call to an external function. 1294 // Look for an existing stub. 1295 SectionEntry &Section = Sections[SectionID]; 1296 StubMap::const_iterator i = Stubs.find(Value); 1297 uintptr_t StubAddress; 1298 if (i != Stubs.end()) { 1299 StubAddress = uintptr_t(Section.Address) + i->second; 1300 DEBUG(dbgs() << " Stub function found\n"); 1301 } else { 1302 // Create a new stub function (equivalent to a PLT entry). 1303 DEBUG(dbgs() << " Create a new stub function\n"); 1304 1305 uintptr_t BaseAddress = uintptr_t(Section.Address); 1306 uintptr_t StubAlignment = getStubAlignment(); 1307 StubAddress = (BaseAddress + Section.StubOffset + 1308 StubAlignment - 1) & -StubAlignment; 1309 unsigned StubOffset = StubAddress - BaseAddress; 1310 Stubs[Value] = StubOffset; 1311 createStubFunction((uint8_t *)StubAddress); 1312 1313 // Create a GOT entry for the external function. 1314 GOTEntries.push_back(Value); 1315 1316 // Make our stub function a relative call to the GOT entry. 1317 RelocationEntry RE(SectionID, StubOffset + 2, 1318 ELF::R_X86_64_GOTPCREL, -4); 1319 addRelocationForSymbol(RE, Value.SymbolName); 1320 1321 // Bump our stub offset counter 1322 Section.StubOffset = StubOffset + getMaxStubSize(); 1323 } 1324 1325 // Make the target call a call into the stub table. 1326 resolveRelocation(Section, Offset, StubAddress, 1327 ELF::R_X86_64_PC32, Addend); 1328 } else { 1329 RelocationEntry RE(SectionID, Offset, ELF::R_X86_64_PC32, Value.Addend, 1330 Value.Offset); 1331 addRelocationForSection(RE, Value.SectionID); 1332 } 1333 } else { 1334 if (Arch == Triple::x86_64 && RelType == ELF::R_X86_64_GOTPCREL) { 1335 GOTEntries.push_back(Value); 1336 } 1337 RelocationEntry RE(SectionID, Offset, RelType, Value.Addend, Value.Offset); 1338 if (Value.SymbolName) 1339 addRelocationForSymbol(RE, Value.SymbolName); 1340 else 1341 addRelocationForSection(RE, Value.SectionID); 1342 } 1343 } 1344 1345 void RuntimeDyldELF::updateGOTEntries(StringRef Name, uint64_t Addr) { 1346 1347 SmallVectorImpl<std::pair<SID, GOTRelocations> >::iterator it; 1348 SmallVectorImpl<std::pair<SID, GOTRelocations> >::iterator end = GOTs.end(); 1349 1350 for (it = GOTs.begin(); it != end; ++it) { 1351 GOTRelocations &GOTEntries = it->second; 1352 for (int i = 0, e = GOTEntries.size(); i != e; ++i) { 1353 if (GOTEntries[i].SymbolName != 0 && GOTEntries[i].SymbolName == Name) { 1354 GOTEntries[i].Offset = Addr; 1355 } 1356 } 1357 } 1358 } 1359 1360 size_t RuntimeDyldELF::getGOTEntrySize() { 1361 // We don't use the GOT in all of these cases, but it's essentially free 1362 // to put them all here. 1363 size_t Result = 0; 1364 switch (Arch) { 1365 case Triple::x86_64: 1366 case Triple::aarch64: 1367 case Triple::ppc64: 1368 case Triple::ppc64le: 1369 case Triple::systemz: 1370 Result = sizeof(uint64_t); 1371 break; 1372 case Triple::x86: 1373 case Triple::arm: 1374 case Triple::thumb: 1375 case Triple::mips: 1376 case Triple::mipsel: 1377 Result = sizeof(uint32_t); 1378 break; 1379 default: llvm_unreachable("Unsupported CPU type!"); 1380 } 1381 return Result; 1382 } 1383 1384 uint64_t RuntimeDyldELF::findGOTEntry(uint64_t LoadAddress, 1385 uint64_t Offset) { 1386 1387 const size_t GOTEntrySize = getGOTEntrySize(); 1388 1389 SmallVectorImpl<std::pair<SID, GOTRelocations> >::const_iterator it; 1390 SmallVectorImpl<std::pair<SID, GOTRelocations> >::const_iterator end = GOTs.end(); 1391 1392 int GOTIndex = -1; 1393 for (it = GOTs.begin(); it != end; ++it) { 1394 SID GOTSectionID = it->first; 1395 const GOTRelocations &GOTEntries = it->second; 1396 1397 // Find the matching entry in our vector. 1398 uint64_t SymbolOffset = 0; 1399 for (int i = 0, e = GOTEntries.size(); i != e; ++i) { 1400 if (GOTEntries[i].SymbolName == 0) { 1401 if (getSectionLoadAddress(GOTEntries[i].SectionID) == LoadAddress && 1402 GOTEntries[i].Offset == Offset) { 1403 GOTIndex = i; 1404 SymbolOffset = GOTEntries[i].Offset; 1405 break; 1406 } 1407 } else { 1408 // GOT entries for external symbols use the addend as the address when 1409 // the external symbol has been resolved. 1410 if (GOTEntries[i].Offset == LoadAddress) { 1411 GOTIndex = i; 1412 // Don't use the Addend here. The relocation handler will use it. 1413 break; 1414 } 1415 } 1416 } 1417 1418 if (GOTIndex != -1) { 1419 if (GOTEntrySize == sizeof(uint64_t)) { 1420 uint64_t *LocalGOTAddr = (uint64_t*)getSectionAddress(GOTSectionID); 1421 // Fill in this entry with the address of the symbol being referenced. 1422 LocalGOTAddr[GOTIndex] = LoadAddress + SymbolOffset; 1423 } else { 1424 uint32_t *LocalGOTAddr = (uint32_t*)getSectionAddress(GOTSectionID); 1425 // Fill in this entry with the address of the symbol being referenced. 1426 LocalGOTAddr[GOTIndex] = (uint32_t)(LoadAddress + SymbolOffset); 1427 } 1428 1429 // Calculate the load address of this entry 1430 return getSectionLoadAddress(GOTSectionID) + (GOTIndex * GOTEntrySize); 1431 } 1432 } 1433 1434 assert(GOTIndex != -1 && "Unable to find requested GOT entry."); 1435 return 0; 1436 } 1437 1438 void RuntimeDyldELF::finalizeLoad(ObjSectionToIDMap &SectionMap) { 1439 // If necessary, allocate the global offset table 1440 if (MemMgr) { 1441 // Allocate the GOT if necessary 1442 size_t numGOTEntries = GOTEntries.size(); 1443 if (numGOTEntries != 0) { 1444 // Allocate memory for the section 1445 unsigned SectionID = Sections.size(); 1446 size_t TotalSize = numGOTEntries * getGOTEntrySize(); 1447 uint8_t *Addr = MemMgr->allocateDataSection(TotalSize, getGOTEntrySize(), 1448 SectionID, ".got", false); 1449 if (!Addr) 1450 report_fatal_error("Unable to allocate memory for GOT!"); 1451 1452 GOTs.push_back(std::make_pair(SectionID, GOTEntries)); 1453 Sections.push_back(SectionEntry(".got", Addr, TotalSize, 0)); 1454 // For now, initialize all GOT entries to zero. We'll fill them in as 1455 // needed when GOT-based relocations are applied. 1456 memset(Addr, 0, TotalSize); 1457 } 1458 } 1459 else { 1460 report_fatal_error("Unable to allocate memory for GOT!"); 1461 } 1462 1463 // Look for and record the EH frame section. 1464 ObjSectionToIDMap::iterator i, e; 1465 for (i = SectionMap.begin(), e = SectionMap.end(); i != e; ++i) { 1466 const SectionRef &Section = i->first; 1467 StringRef Name; 1468 Section.getName(Name); 1469 if (Name == ".eh_frame") { 1470 UnregisteredEHFrameSections.push_back(i->second); 1471 break; 1472 } 1473 } 1474 } 1475 1476 bool RuntimeDyldELF::isCompatibleFormat(const ObjectBuffer *Buffer) const { 1477 if (Buffer->getBufferSize() < strlen(ELF::ElfMagic)) 1478 return false; 1479 return (memcmp(Buffer->getBufferStart(), ELF::ElfMagic, strlen(ELF::ElfMagic))) == 0; 1480 } 1481 1482 bool RuntimeDyldELF::isCompatibleFile(const object::ObjectFile *Obj) const { 1483 return Obj->isELF(); 1484 } 1485 1486 } // namespace llvm 1487