1 /*
2  * Copyright (C) 2016 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef ART_COMPILER_UTILS_ARM64_JNI_MACRO_ASSEMBLER_ARM64_H_
18 #define ART_COMPILER_UTILS_ARM64_JNI_MACRO_ASSEMBLER_ARM64_H_
19 
20 #include <stdint.h>
21 #include <memory>
22 #include <vector>
23 
24 #include <android-base/logging.h>
25 
26 #include "assembler_arm64.h"
27 #include "base/arena_containers.h"
28 #include "base/enums.h"
29 #include "base/macros.h"
30 #include "offsets.h"
31 #include "utils/assembler.h"
32 #include "utils/jni_macro_assembler.h"
33 
34 // TODO(VIXL): Make VIXL compile with -Wshadow.
35 #pragma GCC diagnostic push
36 #pragma GCC diagnostic ignored "-Wshadow"
37 #include "aarch64/macro-assembler-aarch64.h"
38 #pragma GCC diagnostic pop
39 
40 namespace art {
41 namespace arm64 {
42 
43 class Arm64JNIMacroAssembler final : public JNIMacroAssemblerFwd<Arm64Assembler, PointerSize::k64> {
44  public:
Arm64JNIMacroAssembler(ArenaAllocator * allocator)45   explicit Arm64JNIMacroAssembler(ArenaAllocator* allocator)
46       : JNIMacroAssemblerFwd(allocator),
47         exception_blocks_(allocator->Adapter(kArenaAllocAssembler)) {}
48 
49   ~Arm64JNIMacroAssembler();
50 
51   // Finalize the code.
52   void FinalizeCode() override;
53 
54   // Emit code that will create an activation on the stack.
55   void BuildFrame(size_t frame_size,
56                   ManagedRegister method_reg,
57                   ArrayRef<const ManagedRegister> callee_save_regs) override;
58 
59   // Emit code that will remove an activation from the stack.
60   void RemoveFrame(size_t frame_size,
61                    ArrayRef<const ManagedRegister> callee_save_regs,
62                    bool may_suspend) override;
63 
64   void IncreaseFrameSize(size_t adjust) override;
65   void DecreaseFrameSize(size_t adjust) override;
66 
67   // Store routines.
68   void Store(FrameOffset offs, ManagedRegister src, size_t size) override;
69   void StoreRef(FrameOffset dest, ManagedRegister src) override;
70   void StoreRawPtr(FrameOffset dest, ManagedRegister src) override;
71   void StoreImmediateToFrame(FrameOffset dest, uint32_t imm) override;
72   void StoreStackOffsetToThread(ThreadOffset64 thr_offs, FrameOffset fr_offs) override;
73   void StoreStackPointerToThread(ThreadOffset64 thr_offs) override;
74   void StoreSpanning(FrameOffset dest, ManagedRegister src, FrameOffset in_off) override;
75 
76   // Load routines.
77   void Load(ManagedRegister dest, FrameOffset src, size_t size) override;
78   void LoadFromThread(ManagedRegister dest, ThreadOffset64 src, size_t size) override;
79   void LoadRef(ManagedRegister dest, FrameOffset src) override;
80   void LoadRef(ManagedRegister dest,
81                ManagedRegister base,
82                MemberOffset offs,
83                bool unpoison_reference) override;
84   void LoadRawPtr(ManagedRegister dest, ManagedRegister base, Offset offs) override;
85   void LoadRawPtrFromThread(ManagedRegister dest, ThreadOffset64 offs) override;
86 
87   // Copying routines.
88   void MoveArguments(ArrayRef<ArgumentLocation> dests, ArrayRef<ArgumentLocation> srcs) override;
89   void Move(ManagedRegister dest, ManagedRegister src, size_t size) override;
90   void CopyRawPtrFromThread(FrameOffset fr_offs, ThreadOffset64 thr_offs) override;
91   void CopyRawPtrToThread(ThreadOffset64 thr_offs, FrameOffset fr_offs, ManagedRegister scratch)
92       override;
93   void CopyRef(FrameOffset dest, FrameOffset src) override;
94   void CopyRef(FrameOffset dest,
95                ManagedRegister base,
96                MemberOffset offs,
97                bool unpoison_reference) override;
98   void Copy(FrameOffset dest, FrameOffset src, size_t size) override;
99   void Copy(FrameOffset dest,
100             ManagedRegister src_base,
101             Offset src_offset,
102             ManagedRegister scratch,
103             size_t size) override;
104   void Copy(ManagedRegister dest_base,
105             Offset dest_offset,
106             FrameOffset src,
107             ManagedRegister scratch,
108             size_t size) override;
109   void Copy(FrameOffset dest,
110             FrameOffset src_base,
111             Offset src_offset,
112             ManagedRegister scratch,
113             size_t size) override;
114   void Copy(ManagedRegister dest,
115             Offset dest_offset,
116             ManagedRegister src,
117             Offset src_offset,
118             ManagedRegister scratch,
119             size_t size) override;
120   void Copy(FrameOffset dest,
121             Offset dest_offset,
122             FrameOffset src,
123             Offset src_offset,
124             ManagedRegister scratch,
125             size_t size) override;
126   void MemoryBarrier(ManagedRegister scratch) override;
127 
128   // Sign extension.
129   void SignExtend(ManagedRegister mreg, size_t size) override;
130 
131   // Zero extension.
132   void ZeroExtend(ManagedRegister mreg, size_t size) override;
133 
134   // Exploit fast access in managed code to Thread::Current().
135   void GetCurrentThread(ManagedRegister dest) override;
136   void GetCurrentThread(FrameOffset dest_offset) override;
137 
138   // Set up `out_reg` to hold a `jobject` (`StackReference<Object>*` to a spilled value),
139   // or to be null if the value is null and `null_allowed`. `in_reg` holds a possibly
140   // stale reference that can be used to avoid loading the spilled value to
141   // see if the value is null.
142   void CreateJObject(ManagedRegister out_reg,
143                      FrameOffset spilled_reference_offset,
144                      ManagedRegister in_reg,
145                      bool null_allowed) override;
146 
147   // Set up `out_off` to hold a `jobject` (`StackReference<Object>*` to a spilled value),
148   // or to be null if the value is null and `null_allowed`.
149   void CreateJObject(FrameOffset out_off,
150                      FrameOffset spilled_reference_offset,
151                      bool null_allowed) override;
152 
153   // Heap::VerifyObject on src. In some cases (such as a reference to this) we
154   // know that src may not be null.
155   void VerifyObject(ManagedRegister src, bool could_be_null) override;
156   void VerifyObject(FrameOffset src, bool could_be_null) override;
157 
158   // Jump to address held at [base+offset] (used for tail calls).
159   void Jump(ManagedRegister base, Offset offset) override;
160 
161   // Call to address held at [base+offset].
162   void Call(ManagedRegister base, Offset offset) override;
163   void Call(FrameOffset base, Offset offset) override;
164   void CallFromThread(ThreadOffset64 offset) override;
165 
166   // Generate code to check if Thread::Current()->exception_ is non-null
167   // and branch to a ExceptionSlowPath if it is.
168   void ExceptionPoll(size_t stack_adjust) override;
169 
170   // Create a new label that can be used with Jump/Bind calls.
171   std::unique_ptr<JNIMacroLabel> CreateLabel() override;
172   // Emit an unconditional jump to the label.
173   void Jump(JNIMacroLabel* label) override;
174   // Emit a conditional jump to the label by applying a unary condition test to the GC marking flag.
175   void TestGcMarking(JNIMacroLabel* label, JNIMacroUnaryCondition cond) override;
176   // Code at this offset will serve as the target for the Jump call.
177   void Bind(JNIMacroLabel* label) override;
178 
179  private:
180   class Arm64Exception {
181    public:
Arm64Exception(vixl::aarch64::Register scratch,size_t stack_adjust)182     Arm64Exception(vixl::aarch64::Register scratch, size_t stack_adjust)
183         : scratch_(scratch), stack_adjust_(stack_adjust) {}
184 
Entry()185     vixl::aarch64::Label* Entry() { return &exception_entry_; }
186 
187     // Register used for passing Thread::Current()->exception_ .
188     const vixl::aarch64::Register scratch_;
189 
190     // Stack adjust for ExceptionPool.
191     const size_t stack_adjust_;
192 
193     vixl::aarch64::Label exception_entry_;
194 
195    private:
196     DISALLOW_COPY_AND_ASSIGN(Arm64Exception);
197   };
198 
199   // Emits Exception block.
200   void EmitExceptionPoll(Arm64Exception *exception);
201 
202   void StoreWToOffset(StoreOperandType type,
203                       WRegister source,
204                       XRegister base,
205                       int32_t offset);
206   void StoreToOffset(XRegister source, XRegister base, int32_t offset);
207   void StoreSToOffset(SRegister source, XRegister base, int32_t offset);
208   void StoreDToOffset(DRegister source, XRegister base, int32_t offset);
209 
210   void LoadImmediate(XRegister dest,
211                      int32_t value,
212                      vixl::aarch64::Condition cond = vixl::aarch64::al);
213   void Load(Arm64ManagedRegister dst, XRegister src, int32_t src_offset, size_t size);
214   void LoadWFromOffset(LoadOperandType type,
215                        WRegister dest,
216                        XRegister base,
217                        int32_t offset);
218   void LoadFromOffset(XRegister dest, XRegister base, int32_t offset);
219   void LoadSFromOffset(SRegister dest, XRegister base, int32_t offset);
220   void LoadDFromOffset(DRegister dest, XRegister base, int32_t offset);
221   void AddConstant(XRegister rd,
222                    int32_t value,
223                    vixl::aarch64::Condition cond = vixl::aarch64::al);
224   void AddConstant(XRegister rd,
225                    XRegister rn,
226                    int32_t value,
227                    vixl::aarch64::Condition cond = vixl::aarch64::al);
228 
229   // List of exception blocks to generate at the end of the code cache.
230   ArenaVector<std::unique_ptr<Arm64Exception>> exception_blocks_;
231 };
232 
233 class Arm64JNIMacroLabel final
234     : public JNIMacroLabelCommon<Arm64JNIMacroLabel,
235                                  vixl::aarch64::Label,
236                                  InstructionSet::kArm64> {
237  public:
AsArm64()238   vixl::aarch64::Label* AsArm64() {
239     return AsPlatformLabel();
240   }
241 };
242 
243 }  // namespace arm64
244 }  // namespace art
245 
246 #endif  // ART_COMPILER_UTILS_ARM64_JNI_MACRO_ASSEMBLER_ARM64_H_
247