/art/compiler/utils/arm/ |
D | jni_macro_assembler_arm_vixl.h | 191 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust); 210 ArmException(ArmManagedRegister scratch, size_t stack_adjust) in ArmException() argument 211 : scratch_(scratch), stack_adjust_(stack_adjust) {} in ArmException()
|
D | jni_macro_assembler_arm_vixl.cc | 614 void ArmVIXLJNIMacroAssembler::ExceptionPoll(ManagedRegister m_scratch, size_t stack_adjust) { in ExceptionPoll() argument 615 CHECK_ALIGNED(stack_adjust, kStackAlignment); in ExceptionPoll() 620 new ArmVIXLJNIMacroAssembler::ArmException(scratch, stack_adjust)); in ExceptionPoll()
|
/art/compiler/utils/arm64/ |
D | jni_macro_assembler_arm64.h | 172 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) OVERRIDE; 186 Arm64Exception(Arm64ManagedRegister scratch, size_t stack_adjust) in Arm64Exception() argument 187 : scratch_(scratch), stack_adjust_(stack_adjust) {} in Arm64Exception()
|
D | jni_macro_assembler_arm64.cc | 622 void Arm64JNIMacroAssembler::ExceptionPoll(ManagedRegister m_scratch, size_t stack_adjust) { in ExceptionPoll() argument 623 CHECK_ALIGNED(stack_adjust, kStackAlignment); in ExceptionPoll() 625 exception_blocks_.emplace_back(new Arm64Exception(scratch, stack_adjust)); in ExceptionPoll()
|
/art/compiler/utils/x86/ |
D | jni_macro_assembler_x86.cc | 30 explicit X86ExceptionSlowPath(size_t stack_adjust) : stack_adjust_(stack_adjust) {} in X86ExceptionSlowPath() argument 520 void X86JNIMacroAssembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) { in ExceptionPoll() argument 521 X86ExceptionSlowPath* slow = new (__ GetAllocator()) X86ExceptionSlowPath(stack_adjust); in ExceptionPoll()
|
D | jni_macro_assembler_x86.h | 156 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) OVERRIDE;
|
/art/compiler/utils/x86_64/ |
D | jni_macro_assembler_x86_64.cc | 580 explicit X86_64ExceptionSlowPath(size_t stack_adjust) : stack_adjust_(stack_adjust) {} in X86_64ExceptionSlowPath() argument 586 void X86_64JNIMacroAssembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) { in ExceptionPoll() argument 587 X86_64ExceptionSlowPath* slow = new (__ GetAllocator()) X86_64ExceptionSlowPath(stack_adjust); in ExceptionPoll()
|
D | jni_macro_assembler_x86_64.h | 182 void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) OVERRIDE;
|
/art/compiler/utils/ |
D | jni_macro_assembler.h | 207 virtual void ExceptionPoll(ManagedRegister scratch, size_t stack_adjust) = 0;
|
/art/compiler/optimizing/ |
D | intrinsics_x86.cc | 1781 int stack_adjust = kX86WordSize; in VisitStringGetCharsNoCheck() local 1782 __ cfi().AdjustCFAOffset(stack_adjust); in VisitStringGetCharsNoCheck() 1798 __ cfi().AdjustCFAOffset(stack_adjust); in VisitStringGetCharsNoCheck() 1835 __ cfi().AdjustCFAOffset(-stack_adjust); in VisitStringGetCharsNoCheck() 1839 __ cfi().AdjustCFAOffset(-stack_adjust); in VisitStringGetCharsNoCheck()
|
/art/compiler/utils/mips/ |
D | assembler_mips.h | 248 explicit MipsExceptionSlowPath(MipsManagedRegister scratch, size_t stack_adjust) in MipsExceptionSlowPath() argument 249 : scratch_(scratch), stack_adjust_(stack_adjust) {} in MipsExceptionSlowPath() 1364 void ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) OVERRIDE;
|
/art/compiler/utils/mips64/ |
D | assembler_mips64.h | 399 explicit Mips64ExceptionSlowPath(Mips64ManagedRegister scratch, size_t stack_adjust) in Mips64ExceptionSlowPath() argument 400 : scratch_(scratch), stack_adjust_(stack_adjust) {} in Mips64ExceptionSlowPath() 1429 void ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) OVERRIDE;
|
D | assembler_mips64.cc | 4050 void Mips64Assembler::ExceptionPoll(ManagedRegister mscratch, size_t stack_adjust) { in ExceptionPoll() argument 4052 exception_blocks_.emplace_back(scratch, stack_adjust); in ExceptionPoll()
|