RuntimeDyldMachOAArch64.h revision 274968
1//===-- RuntimeDyldMachOAArch64.h -- MachO/AArch64 specific code. -*- C++ -*-=//
2//
3//                     The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9
10#ifndef LLVM_RUNTIMEDYLDMACHOAARCH64_H
11#define LLVM_RUNTIMEDYLDMACHOAARCH64_H
12
13#include "../RuntimeDyldMachO.h"
14
15#define DEBUG_TYPE "dyld"
16
17namespace llvm {
18
19class RuntimeDyldMachOAArch64
20    : public RuntimeDyldMachOCRTPBase<RuntimeDyldMachOAArch64> {
21public:
22  RuntimeDyldMachOAArch64(RTDyldMemoryManager *MM)
23      : RuntimeDyldMachOCRTPBase(MM) {}
24
25  unsigned getMaxStubSize() override { return 8; }
26
27  unsigned getStubAlignment() override { return 8; }
28
29  relocation_iterator
30  processRelocationRef(unsigned SectionID, relocation_iterator RelI,
31                       ObjectImage &ObjImg, ObjSectionToIDMap &ObjSectionToID,
32                       const SymbolTableMap &Symbols, StubMap &Stubs) override {
33    const MachOObjectFile &Obj =
34        static_cast<const MachOObjectFile &>(*ObjImg.getObjectFile());
35    MachO::any_relocation_info RelInfo =
36        Obj.getRelocation(RelI->getRawDataRefImpl());
37
38    assert(!Obj.isRelocationScattered(RelInfo) && "");
39
40    // ARM64 has an ARM64_RELOC_ADDEND relocation type that carries an explicit
41    // addend for the following relocation. If found: (1) store the associated
42    // addend, (2) consume the next relocation, and (3) use the stored addend to
43    // override the addend.
44    bool HasExplicitAddend = false;
45    int64_t ExplicitAddend = 0;
46    if (Obj.getAnyRelocationType(RelInfo) == MachO::ARM64_RELOC_ADDEND) {
47      assert(!Obj.getPlainRelocationExternal(RelInfo));
48      assert(!Obj.getAnyRelocationPCRel(RelInfo));
49      assert(Obj.getAnyRelocationLength(RelInfo) == 2);
50      HasExplicitAddend = true;
51      int64_t RawAddend = Obj.getPlainRelocationSymbolNum(RelInfo);
52      // Sign-extend the 24-bit to 64-bit.
53      ExplicitAddend = (RawAddend << 40) >> 40;
54      ++RelI;
55      RelInfo = Obj.getRelocation(RelI->getRawDataRefImpl());
56    }
57
58    RelocationEntry RE(getBasicRelocationEntry(SectionID, ObjImg, RelI));
59    RelocationValueRef Value(
60        getRelocationValueRef(ObjImg, RelI, RE, ObjSectionToID, Symbols));
61
62    if (HasExplicitAddend) {
63      RE.Addend = ExplicitAddend;
64      Value.Addend = ExplicitAddend;
65    }
66
67    bool IsExtern = Obj.getPlainRelocationExternal(RelInfo);
68    if (!IsExtern && RE.IsPCRel)
69      makeValueAddendPCRel(Value, ObjImg, RelI);
70
71    RE.Addend = Value.Addend;
72
73    if (RE.RelType == MachO::ARM64_RELOC_GOT_LOAD_PAGE21 ||
74        RE.RelType == MachO::ARM64_RELOC_GOT_LOAD_PAGEOFF12)
75      processGOTRelocation(RE, Value, Stubs);
76    else {
77      if (Value.SymbolName)
78        addRelocationForSymbol(RE, Value.SymbolName);
79      else
80        addRelocationForSection(RE, Value.SectionID);
81    }
82
83    return ++RelI;
84  }
85
86  void resolveRelocation(const RelocationEntry &RE, uint64_t Value) {
87    DEBUG(dumpRelocationToResolve(RE, Value));
88
89    const SectionEntry &Section = Sections[RE.SectionID];
90    uint8_t *LocalAddress = Section.Address + RE.Offset;
91
92    switch (RE.RelType) {
93    default:
94      llvm_unreachable("Invalid relocation type!");
95    case MachO::ARM64_RELOC_UNSIGNED: {
96      assert(!RE.IsPCRel && "PCRel and ARM64_RELOC_UNSIGNED not supported");
97      // Mask in the target value a byte at a time (we don't have an alignment
98      // guarantee for the target address, so this is safest).
99      if (RE.Size < 2)
100        llvm_unreachable("Invalid size for ARM64_RELOC_UNSIGNED");
101
102      writeBytesUnaligned(LocalAddress, Value + RE.Addend, 1 << RE.Size);
103      break;
104    }
105    case MachO::ARM64_RELOC_BRANCH26: {
106      assert(RE.IsPCRel && "not PCRel and ARM64_RELOC_BRANCH26 not supported");
107      // Mask the value into the target address. We know instructions are
108      // 32-bit aligned, so we can do it all at once.
109      uint32_t *p = (uint32_t *)LocalAddress;
110      // Check if the addend is encoded in the instruction.
111      uint32_t EncodedAddend = *p & 0x03FFFFFF;
112      if (EncodedAddend != 0) {
113        if (RE.Addend == 0)
114          llvm_unreachable("branch26 instruction has embedded addend.");
115        else
116          llvm_unreachable("branch26 instruction has embedded addend and"
117                           "ARM64_RELOC_ADDEND.");
118      }
119      // Check if branch is in range.
120      uint64_t FinalAddress = Section.LoadAddress + RE.Offset;
121      uint64_t PCRelVal = Value - FinalAddress + RE.Addend;
122      assert(isInt<26>(PCRelVal) && "Branch target out of range!");
123      // Insert the value into the instruction.
124      *p = (*p & 0xFC000000) | ((uint32_t)(PCRelVal >> 2) & 0x03FFFFFF);
125      break;
126    }
127    case MachO::ARM64_RELOC_GOT_LOAD_PAGE21:
128    case MachO::ARM64_RELOC_PAGE21: {
129      assert(RE.IsPCRel && "not PCRel and ARM64_RELOC_PAGE21 not supported");
130      // Mask the value into the target address. We know instructions are
131      // 32-bit aligned, so we can do it all at once.
132      uint32_t *p = (uint32_t *)LocalAddress;
133      // Check if the addend is encoded in the instruction.
134      uint32_t EncodedAddend =
135          ((*p & 0x60000000) >> 29) | ((*p & 0x01FFFFE0) >> 3);
136      if (EncodedAddend != 0) {
137        if (RE.Addend == 0)
138          llvm_unreachable("adrp instruction has embedded addend.");
139        else
140          llvm_unreachable("adrp instruction has embedded addend and"
141                           "ARM64_RELOC_ADDEND.");
142      }
143      // Adjust for PC-relative relocation and offset.
144      uint64_t FinalAddress = Section.LoadAddress + RE.Offset;
145      uint64_t PCRelVal =
146          ((Value + RE.Addend) & (-4096)) - (FinalAddress & (-4096));
147      // Check that the value fits into 21 bits (+ 12 lower bits).
148      assert(isInt<33>(PCRelVal) && "Invalid page reloc value!");
149      // Insert the value into the instruction.
150      uint32_t ImmLoValue = (uint32_t)(PCRelVal << 17) & 0x60000000;
151      uint32_t ImmHiValue = (uint32_t)(PCRelVal >> 9) & 0x00FFFFE0;
152      *p = (*p & 0x9F00001F) | ImmHiValue | ImmLoValue;
153      break;
154    }
155    case MachO::ARM64_RELOC_GOT_LOAD_PAGEOFF12:
156    case MachO::ARM64_RELOC_PAGEOFF12: {
157      assert(!RE.IsPCRel && "PCRel and ARM64_RELOC_PAGEOFF21 not supported");
158      // Mask the value into the target address. We know instructions are
159      // 32-bit aligned, so we can do it all at once.
160      uint32_t *p = (uint32_t *)LocalAddress;
161      // Check if the addend is encoded in the instruction.
162      uint32_t EncodedAddend = *p & 0x003FFC00;
163      if (EncodedAddend != 0) {
164        if (RE.Addend == 0)
165          llvm_unreachable("adrp instruction has embedded addend.");
166        else
167          llvm_unreachable("adrp instruction has embedded addend and"
168                           "ARM64_RELOC_ADDEND.");
169      }
170      // Add the offset from the symbol.
171      Value += RE.Addend;
172      // Mask out the page address and only use the lower 12 bits.
173      Value &= 0xFFF;
174      // Check which instruction we are updating to obtain the implicit shift
175      // factor from LDR/STR instructions.
176      if (*p & 0x08000000) {
177        uint32_t ImplicitShift = ((*p >> 30) & 0x3);
178        switch (ImplicitShift) {
179        case 0:
180          // Check if this a vector op.
181          if ((*p & 0x04800000) == 0x04800000) {
182            ImplicitShift = 4;
183            assert(((Value & 0xF) == 0) &&
184                   "128-bit LDR/STR not 16-byte aligned.");
185          }
186          break;
187        case 1:
188          assert(((Value & 0x1) == 0) && "16-bit LDR/STR not 2-byte aligned.");
189        case 2:
190          assert(((Value & 0x3) == 0) && "32-bit LDR/STR not 4-byte aligned.");
191        case 3:
192          assert(((Value & 0x7) == 0) && "64-bit LDR/STR not 8-byte aligned.");
193        }
194        // Compensate for implicit shift.
195        Value >>= ImplicitShift;
196      }
197      // Insert the value into the instruction.
198      *p = (*p & 0xFFC003FF) | ((uint32_t)(Value << 10) & 0x003FFC00);
199      break;
200    }
201    case MachO::ARM64_RELOC_SUBTRACTOR:
202    case MachO::ARM64_RELOC_POINTER_TO_GOT:
203    case MachO::ARM64_RELOC_TLVP_LOAD_PAGE21:
204    case MachO::ARM64_RELOC_TLVP_LOAD_PAGEOFF12:
205      llvm_unreachable("Relocation type not implemented yet!");
206    case MachO::ARM64_RELOC_ADDEND:
207      llvm_unreachable("ARM64_RELOC_ADDEND should have been handeled by "
208                       "processRelocationRef!");
209    }
210  }
211
212  void finalizeSection(ObjectImage &ObjImg, unsigned SectionID,
213                       const SectionRef &Section) {}
214
215private:
216  void processGOTRelocation(const RelocationEntry &RE,
217                            RelocationValueRef &Value, StubMap &Stubs) {
218    assert(RE.Size == 2);
219    SectionEntry &Section = Sections[RE.SectionID];
220    StubMap::const_iterator i = Stubs.find(Value);
221    uint8_t *Addr;
222    if (i != Stubs.end())
223      Addr = Section.Address + i->second;
224    else {
225      // FIXME: There must be a better way to do this then to check and fix the
226      // alignment every time!!!
227      uintptr_t BaseAddress = uintptr_t(Section.Address);
228      uintptr_t StubAlignment = getStubAlignment();
229      uintptr_t StubAddress =
230          (BaseAddress + Section.StubOffset + StubAlignment - 1) &
231          -StubAlignment;
232      unsigned StubOffset = StubAddress - BaseAddress;
233      Stubs[Value] = StubOffset;
234      assert(((StubAddress % getStubAlignment()) == 0) &&
235             "GOT entry not aligned");
236      RelocationEntry GOTRE(RE.SectionID, StubOffset,
237                            MachO::ARM64_RELOC_UNSIGNED, Value.Addend,
238                            /*IsPCRel=*/false, /*Size=*/3);
239      if (Value.SymbolName)
240        addRelocationForSymbol(GOTRE, Value.SymbolName);
241      else
242        addRelocationForSection(GOTRE, Value.SectionID);
243      Section.StubOffset = StubOffset + getMaxStubSize();
244      Addr = (uint8_t *)StubAddress;
245    }
246    RelocationEntry TargetRE(RE.SectionID, RE.Offset, RE.RelType, /*Addend=*/0,
247                             RE.IsPCRel, RE.Size);
248    resolveRelocation(TargetRE, (uint64_t)Addr);
249  }
250};
251}
252
253#undef DEBUG_TYPE
254
255#endif // LLVM_RUNTIMEDYLDMACHOAARCH64_H
256