1 /*
2  * Copyright (C) 2014 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM64_H_
18 #define ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM64_H_
19 
20 #include "arch/arm64/quick_method_frame_info_arm64.h"
21 #include "code_generator.h"
22 #include "common_arm64.h"
23 #include "dex/compiler_enums.h"
24 #include "driver/compiler_options.h"
25 #include "nodes.h"
26 #include "parallel_move_resolver.h"
27 #include "utils/arm64/assembler_arm64.h"
28 #include "utils/string_reference.h"
29 #include "vixl/a64/disasm-a64.h"
30 #include "vixl/a64/macro-assembler-a64.h"
31 
32 namespace art {
33 namespace arm64 {
34 
35 class CodeGeneratorARM64;
36 
37 // Use a local definition to prevent copying mistakes.
38 static constexpr size_t kArm64WordSize = kArm64PointerSize;
39 
40 static const vixl::Register kParameterCoreRegisters[] = {
41   vixl::x1, vixl::x2, vixl::x3, vixl::x4, vixl::x5, vixl::x6, vixl::x7
42 };
43 static constexpr size_t kParameterCoreRegistersLength = arraysize(kParameterCoreRegisters);
44 static const vixl::FPRegister kParameterFPRegisters[] = {
45   vixl::d0, vixl::d1, vixl::d2, vixl::d3, vixl::d4, vixl::d5, vixl::d6, vixl::d7
46 };
47 static constexpr size_t kParameterFPRegistersLength = arraysize(kParameterFPRegisters);
48 
49 const vixl::Register tr = vixl::x19;                        // Thread Register
50 static const vixl::Register kArtMethodRegister = vixl::x0;  // Method register on invoke.
51 
52 const vixl::CPURegList vixl_reserved_core_registers(vixl::ip0, vixl::ip1);
53 const vixl::CPURegList vixl_reserved_fp_registers(vixl::d31);
54 
55 const vixl::CPURegList runtime_reserved_core_registers(tr, vixl::lr);
56 
57 // Callee-saved registers AAPCS64 (without x19 - Thread Register)
58 const vixl::CPURegList callee_saved_core_registers(vixl::CPURegister::kRegister,
59                                                    vixl::kXRegSize,
60                                                    vixl::x20.code(),
61                                                    vixl::x30.code());
62 const vixl::CPURegList callee_saved_fp_registers(vixl::CPURegister::kFPRegister,
63                                                  vixl::kDRegSize,
64                                                  vixl::d8.code(),
65                                                  vixl::d15.code());
66 Location ARM64ReturnLocation(Primitive::Type return_type);
67 
68 class SlowPathCodeARM64 : public SlowPathCode {
69  public:
SlowPathCodeARM64(HInstruction * instruction)70   explicit SlowPathCodeARM64(HInstruction* instruction)
71       : SlowPathCode(instruction), entry_label_(), exit_label_() {}
72 
GetEntryLabel()73   vixl::Label* GetEntryLabel() { return &entry_label_; }
GetExitLabel()74   vixl::Label* GetExitLabel() { return &exit_label_; }
75 
76   void SaveLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE;
77   void RestoreLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE;
78 
79  private:
80   vixl::Label entry_label_;
81   vixl::Label exit_label_;
82 
83   DISALLOW_COPY_AND_ASSIGN(SlowPathCodeARM64);
84 };
85 
86 class JumpTableARM64 : public DeletableArenaObject<kArenaAllocSwitchTable> {
87  public:
JumpTableARM64(HPackedSwitch * switch_instr)88   explicit JumpTableARM64(HPackedSwitch* switch_instr)
89     : switch_instr_(switch_instr), table_start_() {}
90 
GetTableStartLabel()91   vixl::Label* GetTableStartLabel() { return &table_start_; }
92 
93   void EmitTable(CodeGeneratorARM64* codegen);
94 
95  private:
96   HPackedSwitch* const switch_instr_;
97   vixl::Label table_start_;
98 
99   DISALLOW_COPY_AND_ASSIGN(JumpTableARM64);
100 };
101 
102 static const vixl::Register kRuntimeParameterCoreRegisters[] =
103     { vixl::x0, vixl::x1, vixl::x2, vixl::x3, vixl::x4, vixl::x5, vixl::x6, vixl::x7 };
104 static constexpr size_t kRuntimeParameterCoreRegistersLength =
105     arraysize(kRuntimeParameterCoreRegisters);
106 static const vixl::FPRegister kRuntimeParameterFpuRegisters[] =
107     { vixl::d0, vixl::d1, vixl::d2, vixl::d3, vixl::d4, vixl::d5, vixl::d6, vixl::d7 };
108 static constexpr size_t kRuntimeParameterFpuRegistersLength =
109     arraysize(kRuntimeParameterCoreRegisters);
110 
111 class InvokeRuntimeCallingConvention : public CallingConvention<vixl::Register, vixl::FPRegister> {
112  public:
113   static constexpr size_t kParameterCoreRegistersLength = arraysize(kParameterCoreRegisters);
114 
InvokeRuntimeCallingConvention()115   InvokeRuntimeCallingConvention()
116       : CallingConvention(kRuntimeParameterCoreRegisters,
117                           kRuntimeParameterCoreRegistersLength,
118                           kRuntimeParameterFpuRegisters,
119                           kRuntimeParameterFpuRegistersLength,
120                           kArm64PointerSize) {}
121 
122   Location GetReturnLocation(Primitive::Type return_type);
123 
124  private:
125   DISALLOW_COPY_AND_ASSIGN(InvokeRuntimeCallingConvention);
126 };
127 
128 class InvokeDexCallingConvention : public CallingConvention<vixl::Register, vixl::FPRegister> {
129  public:
InvokeDexCallingConvention()130   InvokeDexCallingConvention()
131       : CallingConvention(kParameterCoreRegisters,
132                           kParameterCoreRegistersLength,
133                           kParameterFPRegisters,
134                           kParameterFPRegistersLength,
135                           kArm64PointerSize) {}
136 
GetReturnLocation(Primitive::Type return_type)137   Location GetReturnLocation(Primitive::Type return_type) const {
138     return ARM64ReturnLocation(return_type);
139   }
140 
141 
142  private:
143   DISALLOW_COPY_AND_ASSIGN(InvokeDexCallingConvention);
144 };
145 
146 class InvokeDexCallingConventionVisitorARM64 : public InvokeDexCallingConventionVisitor {
147  public:
InvokeDexCallingConventionVisitorARM64()148   InvokeDexCallingConventionVisitorARM64() {}
~InvokeDexCallingConventionVisitorARM64()149   virtual ~InvokeDexCallingConventionVisitorARM64() {}
150 
151   Location GetNextLocation(Primitive::Type type) OVERRIDE;
GetReturnLocation(Primitive::Type return_type)152   Location GetReturnLocation(Primitive::Type return_type) const OVERRIDE {
153     return calling_convention.GetReturnLocation(return_type);
154   }
155   Location GetMethodLocation() const OVERRIDE;
156 
157  private:
158   InvokeDexCallingConvention calling_convention;
159 
160   DISALLOW_COPY_AND_ASSIGN(InvokeDexCallingConventionVisitorARM64);
161 };
162 
163 class FieldAccessCallingConventionARM64 : public FieldAccessCallingConvention {
164  public:
FieldAccessCallingConventionARM64()165   FieldAccessCallingConventionARM64() {}
166 
GetObjectLocation()167   Location GetObjectLocation() const OVERRIDE {
168     return helpers::LocationFrom(vixl::x1);
169   }
GetFieldIndexLocation()170   Location GetFieldIndexLocation() const OVERRIDE {
171     return helpers::LocationFrom(vixl::x0);
172   }
GetReturnLocation(Primitive::Type type ATTRIBUTE_UNUSED)173   Location GetReturnLocation(Primitive::Type type ATTRIBUTE_UNUSED) const OVERRIDE {
174     return helpers::LocationFrom(vixl::x0);
175   }
GetSetValueLocation(Primitive::Type type,bool is_instance)176   Location GetSetValueLocation(Primitive::Type type, bool is_instance) const OVERRIDE {
177     return Primitive::Is64BitType(type)
178         ? helpers::LocationFrom(vixl::x2)
179         : (is_instance
180             ? helpers::LocationFrom(vixl::x2)
181             : helpers::LocationFrom(vixl::x1));
182   }
GetFpuLocation(Primitive::Type type ATTRIBUTE_UNUSED)183   Location GetFpuLocation(Primitive::Type type ATTRIBUTE_UNUSED) const OVERRIDE {
184     return helpers::LocationFrom(vixl::d0);
185   }
186 
187  private:
188   DISALLOW_COPY_AND_ASSIGN(FieldAccessCallingConventionARM64);
189 };
190 
191 class InstructionCodeGeneratorARM64 : public InstructionCodeGenerator {
192  public:
193   InstructionCodeGeneratorARM64(HGraph* graph, CodeGeneratorARM64* codegen);
194 
195 #define DECLARE_VISIT_INSTRUCTION(name, super) \
196   void Visit##name(H##name* instr) OVERRIDE;
197 
198   FOR_EACH_CONCRETE_INSTRUCTION_COMMON(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_CONCRETE_INSTRUCTION_ARM64(DECLARE_VISIT_INSTRUCTION)199   FOR_EACH_CONCRETE_INSTRUCTION_ARM64(DECLARE_VISIT_INSTRUCTION)
200   FOR_EACH_CONCRETE_INSTRUCTION_SHARED(DECLARE_VISIT_INSTRUCTION)
201 
202 #undef DECLARE_VISIT_INSTRUCTION
203 
204   void VisitInstruction(HInstruction* instruction) OVERRIDE {
205     LOG(FATAL) << "Unreachable instruction " << instruction->DebugName()
206                << " (id " << instruction->GetId() << ")";
207   }
208 
GetAssembler()209   Arm64Assembler* GetAssembler() const { return assembler_; }
GetVIXLAssembler()210   vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; }
211 
212  private:
213   void GenerateClassInitializationCheck(SlowPathCodeARM64* slow_path, vixl::Register class_reg);
214   void GenerateSuspendCheck(HSuspendCheck* instruction, HBasicBlock* successor);
215   void HandleBinaryOp(HBinaryOperation* instr);
216 
217   void HandleFieldSet(HInstruction* instruction,
218                       const FieldInfo& field_info,
219                       bool value_can_be_null);
220   void HandleFieldGet(HInstruction* instruction, const FieldInfo& field_info);
221   void HandleCondition(HCondition* instruction);
222 
223   // Generate a heap reference load using one register `out`:
224   //
225   //   out <- *(out + offset)
226   //
227   // while honoring heap poisoning and/or read barriers (if any).
228   //
229   // Location `maybe_temp` is used when generating a read barrier and
230   // shall be a register in that case; it may be an invalid location
231   // otherwise.
232   void GenerateReferenceLoadOneRegister(HInstruction* instruction,
233                                         Location out,
234                                         uint32_t offset,
235                                         Location maybe_temp);
236   // Generate a heap reference load using two different registers
237   // `out` and `obj`:
238   //
239   //   out <- *(obj + offset)
240   //
241   // while honoring heap poisoning and/or read barriers (if any).
242   //
243   // Location `maybe_temp` is used when generating a Baker's (fast
244   // path) read barrier and shall be a register in that case; it may
245   // be an invalid location otherwise.
246   void GenerateReferenceLoadTwoRegisters(HInstruction* instruction,
247                                          Location out,
248                                          Location obj,
249                                          uint32_t offset,
250                                          Location maybe_temp);
251   // Generate a GC root reference load:
252   //
253   //   root <- *(obj + offset)
254   //
255   // while honoring read barriers (if any).
256   void GenerateGcRootFieldLoad(HInstruction* instruction,
257                                Location root,
258                                vixl::Register obj,
259                                uint32_t offset,
260                                vixl::Label* fixup_label = nullptr);
261 
262   // Generate a floating-point comparison.
263   void GenerateFcmp(HInstruction* instruction);
264 
265   void HandleShift(HBinaryOperation* instr);
266   void GenerateTestAndBranch(HInstruction* instruction,
267                              size_t condition_input_index,
268                              vixl::Label* true_target,
269                              vixl::Label* false_target);
270   void DivRemOneOrMinusOne(HBinaryOperation* instruction);
271   void DivRemByPowerOfTwo(HBinaryOperation* instruction);
272   void GenerateDivRemWithAnyConstant(HBinaryOperation* instruction);
273   void GenerateDivRemIntegral(HBinaryOperation* instruction);
274   void HandleGoto(HInstruction* got, HBasicBlock* successor);
275 
276   Arm64Assembler* const assembler_;
277   CodeGeneratorARM64* const codegen_;
278 
279   DISALLOW_COPY_AND_ASSIGN(InstructionCodeGeneratorARM64);
280 };
281 
282 class LocationsBuilderARM64 : public HGraphVisitor {
283  public:
LocationsBuilderARM64(HGraph * graph,CodeGeneratorARM64 * codegen)284   LocationsBuilderARM64(HGraph* graph, CodeGeneratorARM64* codegen)
285       : HGraphVisitor(graph), codegen_(codegen) {}
286 
287 #define DECLARE_VISIT_INSTRUCTION(name, super) \
288   void Visit##name(H##name* instr) OVERRIDE;
289 
290   FOR_EACH_CONCRETE_INSTRUCTION_COMMON(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_CONCRETE_INSTRUCTION_ARM64(DECLARE_VISIT_INSTRUCTION)291   FOR_EACH_CONCRETE_INSTRUCTION_ARM64(DECLARE_VISIT_INSTRUCTION)
292   FOR_EACH_CONCRETE_INSTRUCTION_SHARED(DECLARE_VISIT_INSTRUCTION)
293 
294 #undef DECLARE_VISIT_INSTRUCTION
295 
296   void VisitInstruction(HInstruction* instruction) OVERRIDE {
297     LOG(FATAL) << "Unreachable instruction " << instruction->DebugName()
298                << " (id " << instruction->GetId() << ")";
299   }
300 
301  private:
302   void HandleBinaryOp(HBinaryOperation* instr);
303   void HandleFieldSet(HInstruction* instruction);
304   void HandleFieldGet(HInstruction* instruction);
305   void HandleInvoke(HInvoke* instr);
306   void HandleCondition(HCondition* instruction);
307   void HandleShift(HBinaryOperation* instr);
308 
309   CodeGeneratorARM64* const codegen_;
310   InvokeDexCallingConventionVisitorARM64 parameter_visitor_;
311 
312   DISALLOW_COPY_AND_ASSIGN(LocationsBuilderARM64);
313 };
314 
315 class ParallelMoveResolverARM64 : public ParallelMoveResolverNoSwap {
316  public:
ParallelMoveResolverARM64(ArenaAllocator * allocator,CodeGeneratorARM64 * codegen)317   ParallelMoveResolverARM64(ArenaAllocator* allocator, CodeGeneratorARM64* codegen)
318       : ParallelMoveResolverNoSwap(allocator), codegen_(codegen), vixl_temps_() {}
319 
320  protected:
321   void PrepareForEmitNativeCode() OVERRIDE;
322   void FinishEmitNativeCode() OVERRIDE;
323   Location AllocateScratchLocationFor(Location::Kind kind) OVERRIDE;
324   void FreeScratchLocation(Location loc) OVERRIDE;
325   void EmitMove(size_t index) OVERRIDE;
326 
327  private:
328   Arm64Assembler* GetAssembler() const;
GetVIXLAssembler()329   vixl::MacroAssembler* GetVIXLAssembler() const {
330     return GetAssembler()->vixl_masm_;
331   }
332 
333   CodeGeneratorARM64* const codegen_;
334   vixl::UseScratchRegisterScope vixl_temps_;
335 
336   DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolverARM64);
337 };
338 
339 class CodeGeneratorARM64 : public CodeGenerator {
340  public:
341   CodeGeneratorARM64(HGraph* graph,
342                      const Arm64InstructionSetFeatures& isa_features,
343                      const CompilerOptions& compiler_options,
344                      OptimizingCompilerStats* stats = nullptr);
~CodeGeneratorARM64()345   virtual ~CodeGeneratorARM64() {}
346 
347   void GenerateFrameEntry() OVERRIDE;
348   void GenerateFrameExit() OVERRIDE;
349 
350   vixl::CPURegList GetFramePreservedCoreRegisters() const;
351   vixl::CPURegList GetFramePreservedFPRegisters() const;
352 
353   void Bind(HBasicBlock* block) OVERRIDE;
354 
GetLabelOf(HBasicBlock * block)355   vixl::Label* GetLabelOf(HBasicBlock* block) {
356     block = FirstNonEmptyBlock(block);
357     return &(block_labels_[block->GetBlockId()]);
358   }
359 
GetWordSize()360   size_t GetWordSize() const OVERRIDE {
361     return kArm64WordSize;
362   }
363 
GetFloatingPointSpillSlotSize()364   size_t GetFloatingPointSpillSlotSize() const OVERRIDE {
365     // Allocated in D registers, which are word sized.
366     return kArm64WordSize;
367   }
368 
GetAddressOf(HBasicBlock * block)369   uintptr_t GetAddressOf(HBasicBlock* block) OVERRIDE {
370     vixl::Label* block_entry_label = GetLabelOf(block);
371     DCHECK(block_entry_label->IsBound());
372     return block_entry_label->location();
373   }
374 
GetLocationBuilder()375   HGraphVisitor* GetLocationBuilder() OVERRIDE { return &location_builder_; }
GetInstructionVisitor()376   HGraphVisitor* GetInstructionVisitor() OVERRIDE { return &instruction_visitor_; }
GetAssembler()377   Arm64Assembler* GetAssembler() OVERRIDE { return &assembler_; }
GetAssembler()378   const Arm64Assembler& GetAssembler() const OVERRIDE { return assembler_; }
GetVIXLAssembler()379   vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; }
380 
381   // Emit a write barrier.
382   void MarkGCCard(vixl::Register object, vixl::Register value, bool value_can_be_null);
383 
384   void GenerateMemoryBarrier(MemBarrierKind kind);
385 
386   // Register allocation.
387 
388   void SetupBlockedRegisters() const OVERRIDE;
389 
390   size_t SaveCoreRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
391   size_t RestoreCoreRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
392   size_t SaveFloatingPointRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
393   size_t RestoreFloatingPointRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
394 
395   // The number of registers that can be allocated. The register allocator may
396   // decide to reserve and not use a few of them.
397   // We do not consider registers sp, xzr, wzr. They are either not allocatable
398   // (xzr, wzr), or make for poor allocatable registers (sp alignment
399   // requirements, etc.). This also facilitates our task as all other registers
400   // can easily be mapped via to or from their type and index or code.
401   static const int kNumberOfAllocatableRegisters = vixl::kNumberOfRegisters - 1;
402   static const int kNumberOfAllocatableFPRegisters = vixl::kNumberOfFPRegisters;
403   static constexpr int kNumberOfAllocatableRegisterPairs = 0;
404 
405   void DumpCoreRegister(std::ostream& stream, int reg) const OVERRIDE;
406   void DumpFloatingPointRegister(std::ostream& stream, int reg) const OVERRIDE;
407 
GetInstructionSet()408   InstructionSet GetInstructionSet() const OVERRIDE {
409     return InstructionSet::kArm64;
410   }
411 
GetInstructionSetFeatures()412   const Arm64InstructionSetFeatures& GetInstructionSetFeatures() const {
413     return isa_features_;
414   }
415 
Initialize()416   void Initialize() OVERRIDE {
417     block_labels_.resize(GetGraph()->GetBlocks().size());
418   }
419 
CreateJumpTable(HPackedSwitch * switch_instr)420   JumpTableARM64* CreateJumpTable(HPackedSwitch* switch_instr) {
421     jump_tables_.emplace_back(new (GetGraph()->GetArena()) JumpTableARM64(switch_instr));
422     return jump_tables_.back().get();
423   }
424 
425   void Finalize(CodeAllocator* allocator) OVERRIDE;
426 
427   // Code generation helpers.
428   void MoveConstant(vixl::CPURegister destination, HConstant* constant);
429   void MoveConstant(Location destination, int32_t value) OVERRIDE;
430   void MoveLocation(Location dst, Location src, Primitive::Type dst_type) OVERRIDE;
431   void AddLocationAsTemp(Location location, LocationSummary* locations) OVERRIDE;
432 
433   void Load(Primitive::Type type, vixl::CPURegister dst, const vixl::MemOperand& src);
434   void Store(Primitive::Type type, vixl::CPURegister src, const vixl::MemOperand& dst);
435   void LoadAcquire(HInstruction* instruction,
436                    vixl::CPURegister dst,
437                    const vixl::MemOperand& src,
438                    bool needs_null_check);
439   void StoreRelease(Primitive::Type type, vixl::CPURegister src, const vixl::MemOperand& dst);
440 
441   // Generate code to invoke a runtime entry point.
442   void InvokeRuntime(QuickEntrypointEnum entrypoint,
443                      HInstruction* instruction,
444                      uint32_t dex_pc,
445                      SlowPathCode* slow_path) OVERRIDE;
446 
447   void InvokeRuntime(int32_t offset,
448                      HInstruction* instruction,
449                      uint32_t dex_pc,
450                      SlowPathCode* slow_path);
451 
GetMoveResolver()452   ParallelMoveResolverARM64* GetMoveResolver() OVERRIDE { return &move_resolver_; }
453 
NeedsTwoRegisters(Primitive::Type type ATTRIBUTE_UNUSED)454   bool NeedsTwoRegisters(Primitive::Type type ATTRIBUTE_UNUSED) const OVERRIDE {
455     return false;
456   }
457 
458   // Check if the desired_string_load_kind is supported. If it is, return it,
459   // otherwise return a fall-back kind that should be used instead.
460   HLoadString::LoadKind GetSupportedLoadStringKind(
461       HLoadString::LoadKind desired_string_load_kind) OVERRIDE;
462 
463   // Check if the desired_dispatch_info is supported. If it is, return it,
464   // otherwise return a fall-back info that should be used instead.
465   HInvokeStaticOrDirect::DispatchInfo GetSupportedInvokeStaticOrDirectDispatch(
466       const HInvokeStaticOrDirect::DispatchInfo& desired_dispatch_info,
467       MethodReference target_method) OVERRIDE;
468 
469   void GenerateStaticOrDirectCall(HInvokeStaticOrDirect* invoke, Location temp) OVERRIDE;
470   void GenerateVirtualCall(HInvokeVirtual* invoke, Location temp) OVERRIDE;
471 
MoveFromReturnRegister(Location trg ATTRIBUTE_UNUSED,Primitive::Type type ATTRIBUTE_UNUSED)472   void MoveFromReturnRegister(Location trg ATTRIBUTE_UNUSED,
473                               Primitive::Type type ATTRIBUTE_UNUSED) OVERRIDE {
474     UNIMPLEMENTED(FATAL);
475   }
476 
477   // Add a new PC-relative string patch for an instruction and return the label
478   // to be bound before the instruction. The instruction will be either the
479   // ADRP (pass `adrp_label = null`) or the ADD (pass `adrp_label` pointing
480   // to the associated ADRP patch label).
481   vixl::Label* NewPcRelativeStringPatch(const DexFile& dex_file,
482                                         uint32_t string_index,
483                                         vixl::Label* adrp_label = nullptr);
484 
485   // Add a new PC-relative dex cache array patch for an instruction and return
486   // the label to be bound before the instruction. The instruction will be
487   // either the ADRP (pass `adrp_label = null`) or the LDR (pass `adrp_label`
488   // pointing to the associated ADRP patch label).
489   vixl::Label* NewPcRelativeDexCacheArrayPatch(const DexFile& dex_file,
490                                                uint32_t element_offset,
491                                                vixl::Label* adrp_label = nullptr);
492 
493   vixl::Literal<uint32_t>* DeduplicateBootImageStringLiteral(const DexFile& dex_file,
494                                                              uint32_t string_index);
495   vixl::Literal<uint32_t>* DeduplicateBootImageAddressLiteral(uint64_t address);
496   vixl::Literal<uint64_t>* DeduplicateDexCacheAddressLiteral(uint64_t address);
497 
498   void EmitLinkerPatches(ArenaVector<LinkerPatch>* linker_patches) OVERRIDE;
499 
500   // Fast path implementation of ReadBarrier::Barrier for a heap
501   // reference field load when Baker's read barriers are used.
502   void GenerateFieldLoadWithBakerReadBarrier(HInstruction* instruction,
503                                              Location ref,
504                                              vixl::Register obj,
505                                              uint32_t offset,
506                                              vixl::Register temp,
507                                              bool needs_null_check,
508                                              bool use_load_acquire);
509   // Fast path implementation of ReadBarrier::Barrier for a heap
510   // reference array load when Baker's read barriers are used.
511   void GenerateArrayLoadWithBakerReadBarrier(HInstruction* instruction,
512                                              Location ref,
513                                              vixl::Register obj,
514                                              uint32_t data_offset,
515                                              Location index,
516                                              vixl::Register temp,
517                                              bool needs_null_check);
518 
519   // Generate a read barrier for a heap reference within `instruction`
520   // using a slow path.
521   //
522   // A read barrier for an object reference read from the heap is
523   // implemented as a call to the artReadBarrierSlow runtime entry
524   // point, which is passed the values in locations `ref`, `obj`, and
525   // `offset`:
526   //
527   //   mirror::Object* artReadBarrierSlow(mirror::Object* ref,
528   //                                      mirror::Object* obj,
529   //                                      uint32_t offset);
530   //
531   // The `out` location contains the value returned by
532   // artReadBarrierSlow.
533   //
534   // When `index` is provided (i.e. for array accesses), the offset
535   // value passed to artReadBarrierSlow is adjusted to take `index`
536   // into account.
537   void GenerateReadBarrierSlow(HInstruction* instruction,
538                                Location out,
539                                Location ref,
540                                Location obj,
541                                uint32_t offset,
542                                Location index = Location::NoLocation());
543 
544   // If read barriers are enabled, generate a read barrier for a heap
545   // reference using a slow path. If heap poisoning is enabled, also
546   // unpoison the reference in `out`.
547   void MaybeGenerateReadBarrierSlow(HInstruction* instruction,
548                                     Location out,
549                                     Location ref,
550                                     Location obj,
551                                     uint32_t offset,
552                                     Location index = Location::NoLocation());
553 
554   // Generate a read barrier for a GC root within `instruction` using
555   // a slow path.
556   //
557   // A read barrier for an object reference GC root is implemented as
558   // a call to the artReadBarrierForRootSlow runtime entry point,
559   // which is passed the value in location `root`:
560   //
561   //   mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root);
562   //
563   // The `out` location contains the value returned by
564   // artReadBarrierForRootSlow.
565   void GenerateReadBarrierForRootSlow(HInstruction* instruction, Location out, Location root);
566 
567   void GenerateNop();
568 
569   void GenerateImplicitNullCheck(HNullCheck* instruction);
570   void GenerateExplicitNullCheck(HNullCheck* instruction);
571 
572  private:
573   // Factored implementation of GenerateFieldLoadWithBakerReadBarrier
574   // and GenerateArrayLoadWithBakerReadBarrier.
575   void GenerateReferenceLoadWithBakerReadBarrier(HInstruction* instruction,
576                                                  Location ref,
577                                                  vixl::Register obj,
578                                                  uint32_t offset,
579                                                  Location index,
580                                                  vixl::Register temp,
581                                                  bool needs_null_check,
582                                                  bool use_load_acquire);
583 
584   using Uint64ToLiteralMap = ArenaSafeMap<uint64_t, vixl::Literal<uint64_t>*>;
585   using Uint32ToLiteralMap = ArenaSafeMap<uint32_t, vixl::Literal<uint32_t>*>;
586   using MethodToLiteralMap = ArenaSafeMap<MethodReference,
587                                           vixl::Literal<uint64_t>*,
588                                           MethodReferenceComparator>;
589   using BootStringToLiteralMap = ArenaSafeMap<StringReference,
590                                               vixl::Literal<uint32_t>*,
591                                               StringReferenceValueComparator>;
592 
593   vixl::Literal<uint32_t>* DeduplicateUint32Literal(uint32_t value, Uint32ToLiteralMap* map);
594   vixl::Literal<uint64_t>* DeduplicateUint64Literal(uint64_t value);
595   vixl::Literal<uint64_t>* DeduplicateMethodLiteral(MethodReference target_method,
596                                                     MethodToLiteralMap* map);
597   vixl::Literal<uint64_t>* DeduplicateMethodAddressLiteral(MethodReference target_method);
598   vixl::Literal<uint64_t>* DeduplicateMethodCodeLiteral(MethodReference target_method);
599 
600   // The PcRelativePatchInfo is used for PC-relative addressing of dex cache arrays
601   // and boot image strings. The only difference is the interpretation of the offset_or_index.
602   struct PcRelativePatchInfo {
PcRelativePatchInfoPcRelativePatchInfo603     PcRelativePatchInfo(const DexFile& dex_file, uint32_t off_or_idx)
604         : target_dex_file(dex_file), offset_or_index(off_or_idx), label(), pc_insn_label() { }
605 
606     const DexFile& target_dex_file;
607     // Either the dex cache array element offset or the string index.
608     uint32_t offset_or_index;
609     vixl::Label label;
610     vixl::Label* pc_insn_label;
611   };
612 
613   vixl::Label* NewPcRelativePatch(const DexFile& dex_file,
614                                   uint32_t offset_or_index,
615                                   vixl::Label* adrp_label,
616                                   ArenaDeque<PcRelativePatchInfo>* patches);
617 
618   void EmitJumpTables();
619 
620   // Labels for each block that will be compiled.
621   // We use a deque so that the `vixl::Label` objects do not move in memory.
622   ArenaDeque<vixl::Label> block_labels_;  // Indexed by block id.
623   vixl::Label frame_entry_label_;
624   ArenaVector<std::unique_ptr<JumpTableARM64>> jump_tables_;
625 
626   LocationsBuilderARM64 location_builder_;
627   InstructionCodeGeneratorARM64 instruction_visitor_;
628   ParallelMoveResolverARM64 move_resolver_;
629   Arm64Assembler assembler_;
630   const Arm64InstructionSetFeatures& isa_features_;
631 
632   // Deduplication map for 32-bit literals, used for non-patchable boot image addresses.
633   Uint32ToLiteralMap uint32_literals_;
634   // Deduplication map for 64-bit literals, used for non-patchable method address, method code
635   // or string dex cache address.
636   Uint64ToLiteralMap uint64_literals_;
637   // Method patch info, map MethodReference to a literal for method address and method code.
638   MethodToLiteralMap method_patches_;
639   MethodToLiteralMap call_patches_;
640   // Relative call patch info.
641   // Using ArenaDeque<> which retains element addresses on push/emplace_back().
642   ArenaDeque<MethodPatchInfo<vixl::Label>> relative_call_patches_;
643   // PC-relative DexCache access info.
644   ArenaDeque<PcRelativePatchInfo> pc_relative_dex_cache_patches_;
645   // Deduplication map for boot string literals for kBootImageLinkTimeAddress.
646   BootStringToLiteralMap boot_image_string_patches_;
647   // PC-relative String patch info.
648   ArenaDeque<PcRelativePatchInfo> pc_relative_string_patches_;
649   // Deduplication map for patchable boot image addresses.
650   Uint32ToLiteralMap boot_image_address_patches_;
651 
652   DISALLOW_COPY_AND_ASSIGN(CodeGeneratorARM64);
653 };
654 
GetAssembler()655 inline Arm64Assembler* ParallelMoveResolverARM64::GetAssembler() const {
656   return codegen_->GetAssembler();
657 }
658 
659 }  // namespace arm64
660 }  // namespace art
661 
662 #endif  // ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM64_H_
663