/* * Copyright (C) 2012 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "asm_support_x86.S" #include "interpreter/cfi_asm_support.h" #include "arch/quick_alloc_entrypoints.S" // For x86, the CFA is esp+4, the address above the pushed return address on the stack. /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveAllCalleeSaves) */ MACRO2(SETUP_SAVE_ALL_CALLEE_SAVES_FRAME, got_reg, temp_reg) PUSH edi // Save callee saves (ebx is saved/restored by the upcall) PUSH esi PUSH ebp subl MACRO_LITERAL(12), %esp // Grow stack by 3 words. CFI_ADJUST_CFA_OFFSET(12) SETUP_GOT_NOSAVE RAW_VAR(got_reg) // Load Runtime::instance_ from GOT. movl SYMBOL(_ZN3art7Runtime9instance_E)@GOT(REG_VAR(got_reg)), REG_VAR(temp_reg) movl (REG_VAR(temp_reg)), REG_VAR(temp_reg) // Push save all callee-save method. pushl RUNTIME_SAVE_ALL_CALLEE_SAVES_METHOD_OFFSET(REG_VAR(temp_reg)) CFI_ADJUST_CFA_OFFSET(4) // Store esp as the top quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET // Ugly compile-time check, but we only have the preprocessor. // Last +4: implicit return address pushed on stack when caller made call. #if (FRAME_SIZE_SAVE_ALL_CALLEE_SAVES != 3*4 + 16 + 4) #error "FRAME_SIZE_SAVE_ALL_CALLEE_SAVES(X86) size not as expected." #endif END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveRefsOnly) */ MACRO2(SETUP_SAVE_REFS_ONLY_FRAME, got_reg, temp_reg) PUSH edi // Save callee saves (ebx is saved/restored by the upcall) PUSH esi PUSH ebp subl MACRO_LITERAL(12), %esp // Grow stack by 3 words. CFI_ADJUST_CFA_OFFSET(12) SETUP_GOT_NOSAVE RAW_VAR(got_reg) // Load Runtime::instance_ from GOT. movl SYMBOL(_ZN3art7Runtime9instance_E)@GOT(REG_VAR(got_reg)), REG_VAR(temp_reg) movl (REG_VAR(temp_reg)), REG_VAR(temp_reg) // Push save all callee-save method. pushl RUNTIME_SAVE_REFS_ONLY_METHOD_OFFSET(REG_VAR(temp_reg)) CFI_ADJUST_CFA_OFFSET(4) // Store esp as the top quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET // Ugly compile-time check, but we only have the preprocessor. // Last +4: implicit return address pushed on stack when caller made call. #if (FRAME_SIZE_SAVE_REFS_ONLY != 3*4 + 16 + 4) #error "FRAME_SIZE_SAVE_REFS_ONLY(X86) size not as expected." #endif END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveRefsOnly) * and preserves the value of got_reg at entry. */ MACRO2(SETUP_SAVE_REFS_ONLY_FRAME_PRESERVE_GOT_REG, got_reg, temp_reg) PUSH edi // Save callee saves (ebx is saved/restored by the upcall) PUSH esi PUSH ebp PUSH RAW_VAR(got_reg) // Save got_reg subl MACRO_LITERAL(8), %esp // Grow stack by 2 words. CFI_ADJUST_CFA_OFFSET(8) SETUP_GOT_NOSAVE RAW_VAR(got_reg) // Load Runtime::instance_ from GOT. movl SYMBOL(_ZN3art7Runtime9instance_E)@GOT(REG_VAR(got_reg)), REG_VAR(temp_reg) movl (REG_VAR(temp_reg)), REG_VAR(temp_reg) // Push save all callee-save method. pushl RUNTIME_SAVE_REFS_ONLY_METHOD_OFFSET(REG_VAR(temp_reg)) CFI_ADJUST_CFA_OFFSET(4) // Store esp as the top quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET // Restore got_reg. movl 12(%esp), REG_VAR(got_reg) CFI_RESTORE(RAW_VAR(got_reg)) // Ugly compile-time check, but we only have the preprocessor. // Last +4: implicit return address pushed on stack when caller made call. #if (FRAME_SIZE_SAVE_REFS_ONLY != 3*4 + 16 + 4) #error "FRAME_SIZE_SAVE_REFS_ONLY(X86) size not as expected." #endif END_MACRO MACRO0(RESTORE_SAVE_REFS_ONLY_FRAME) addl MACRO_LITERAL(16), %esp // Unwind stack up to saved values CFI_ADJUST_CFA_OFFSET(-16) POP ebp // Restore callee saves (ebx is saved/restored by the upcall) POP esi POP edi END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveRefsAndArgs) */ MACRO2(SETUP_SAVE_REFS_AND_ARGS_FRAME, got_reg, temp_reg) SETUP_SAVE_REFS_AND_ARGS_FRAME_REGISTERS_ONLY SETUP_GOT_NOSAVE RAW_VAR(got_reg) // Load Runtime::instance_ from GOT. movl SYMBOL(_ZN3art7Runtime9instance_E)@GOT(REG_VAR(got_reg)), REG_VAR(temp_reg) movl (REG_VAR(temp_reg)), REG_VAR(temp_reg) // Push save all callee-save method. pushl RUNTIME_SAVE_REFS_AND_ARGS_METHOD_OFFSET(REG_VAR(temp_reg)) CFI_ADJUST_CFA_OFFSET(4) // Store esp as the stop quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveRefsAndArgs) where the method is passed in EAX. */ MACRO0(SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX) SETUP_SAVE_REFS_AND_ARGS_FRAME_REGISTERS_ONLY pushl %eax // Store the ArtMethod reference at the bottom of the stack. CFI_ADJUST_CFA_OFFSET(4) // Store esp as the stop quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET END_MACRO // Restore register and jump to routine // Inputs: EDI contains pointer to code. // Notes: Need to pop EAX too (restores Method*) MACRO0(RESTORE_SAVE_REFS_AND_ARGS_FRAME_AND_JUMP) POP eax // Restore Method* // Restore FPRs. movsd 0(%esp), %xmm0 movsd 8(%esp), %xmm1 movsd 16(%esp), %xmm2 movsd 24(%esp), %xmm3 addl MACRO_LITERAL(32), %esp // Remove FPRs. CFI_ADJUST_CFA_OFFSET(-32) POP ecx // Restore args except eax POP edx POP ebx POP ebp // Restore callee saves POP esi xchgl 0(%esp),%edi // restore EDI and place code pointer as only value on stack ret END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveEverything) * when EDI and ESI are already saved. */ MACRO3(SETUP_SAVE_EVERYTHING_FRAME_EDI_ESI_SAVED, got_reg, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET) // Save core registers from highest to lowest to agree with core spills bitmap. // EDI and ESI, or at least placeholders for them, are already on the stack. PUSH ebp PUSH ebx PUSH edx PUSH ecx PUSH eax // Create space for FPR registers and stack alignment padding. subl MACRO_LITERAL(12 + 8 * 8), %esp CFI_ADJUST_CFA_OFFSET(12 + 8 * 8) // Save FPRs. movsd %xmm0, 12(%esp) movsd %xmm1, 20(%esp) movsd %xmm2, 28(%esp) movsd %xmm3, 36(%esp) movsd %xmm4, 44(%esp) movsd %xmm5, 52(%esp) movsd %xmm6, 60(%esp) movsd %xmm7, 68(%esp) SETUP_GOT_NOSAVE RAW_VAR(got_reg) // Load Runtime::instance_ from GOT. movl SYMBOL(_ZN3art7Runtime9instance_E)@GOT(REG_VAR(got_reg)), REG_VAR(temp_reg) movl (REG_VAR(temp_reg)), REG_VAR(temp_reg) // Push save everything callee-save method. pushl \runtime_method_offset(REG_VAR(temp_reg)) CFI_ADJUST_CFA_OFFSET(4) // Store esp as the stop quick frame. movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET // Ugly compile-time check, but we only have the preprocessor. // Last +4: implicit return address pushed on stack when caller made call. #if (FRAME_SIZE_SAVE_EVERYTHING != 7*4 + 8*8 + 12 + 4 + 4) #error "FRAME_SIZE_SAVE_EVERYTHING(X86) size not as expected." #endif END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveEverything) * when EDI is already saved. */ MACRO3(SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED, got_reg, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET) // Save core registers from highest to lowest to agree with core spills bitmap. // EDI, or at least a placeholder for it, is already on the stack. PUSH esi SETUP_SAVE_EVERYTHING_FRAME_EDI_ESI_SAVED RAW_VAR(got_reg), RAW_VAR(temp_reg), \runtime_method_offset END_MACRO /* * Macro that sets up the callee save frame to conform with * Runtime::CreateCalleeSaveMethod(kSaveEverything) */ MACRO3(SETUP_SAVE_EVERYTHING_FRAME, got_reg, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET) PUSH edi SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED RAW_VAR(got_reg), RAW_VAR(temp_reg), \runtime_method_offset END_MACRO MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_FRPS) // Restore FPRs. Method and padding is still on the stack. movsd 16(%esp), %xmm0 movsd 24(%esp), %xmm1 movsd 32(%esp), %xmm2 movsd 40(%esp), %xmm3 movsd 48(%esp), %xmm4 movsd 56(%esp), %xmm5 movsd 64(%esp), %xmm6 movsd 72(%esp), %xmm7 END_MACRO MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX) // Restore core registers (except eax). POP ecx POP edx POP ebx POP ebp POP esi POP edi END_MACRO MACRO0(RESTORE_SAVE_EVERYTHING_FRAME) RESTORE_SAVE_EVERYTHING_FRAME_FRPS // Remove save everything callee save method, stack alignment padding and FPRs. addl MACRO_LITERAL(16 + 8 * 8), %esp CFI_ADJUST_CFA_OFFSET(-(16 + 8 * 8)) POP eax RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX END_MACRO MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_KEEP_EAX) RESTORE_SAVE_EVERYTHING_FRAME_FRPS // Remove save everything callee save method, stack alignment padding and FPRs, skip EAX. addl MACRO_LITERAL(16 + 8 * 8 + 4), %esp CFI_ADJUST_CFA_OFFSET(-(16 + 8 * 8 + 4)) RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX END_MACRO /* * Macro that calls through to artDeliverPendingExceptionFromCode, where the pending * exception is Thread::Current()->exception_. */ MACRO0(DELIVER_PENDING_EXCEPTION) SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx, ebx // save callee saves for throw DELIVER_PENDING_EXCEPTION_FRAME_READY END_MACRO MACRO2(NO_ARG_RUNTIME_EXCEPTION, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up subl MACRO_LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) call CALLVAR(cxx_name) // cxx_name(Thread*) UNREACHABLE END_FUNCTION VAR(c_name) END_MACRO MACRO2(NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_EVERYTHING_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up subl MACRO_LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) call CALLVAR(cxx_name) // cxx_name(Thread*) UNREACHABLE END_FUNCTION VAR(c_name) END_MACRO MACRO2(ONE_ARG_RUNTIME_EXCEPTION, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up subl MACRO_LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, Thread*) UNREACHABLE END_FUNCTION VAR(c_name) END_MACRO MACRO2(TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_EVERYTHING_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up PUSH eax // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*) UNREACHABLE END_FUNCTION VAR(c_name) END_MACRO /* * Called by managed code to create and deliver a NullPointerException. */ NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_null_pointer_exception, artThrowNullPointerExceptionFromCode /* * Call installed by a signal handler to create and deliver a NullPointerException. */ DEFINE_FUNCTION_CUSTOM_CFA art_quick_throw_null_pointer_exception_from_signal, 2 * __SIZEOF_POINTER__ // Fault address and return address were saved by the fault handler. // Save all registers as basis for long jump context; EDI will replace fault address later. SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED ebx, ebx // Retrieve fault address and save EDI. movl (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp), %eax movl %edi, (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp) CFI_REL_OFFSET(%edi, (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)) // Outgoing argument set up subl MACRO_LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass arg1 call SYMBOL(artThrowNullPointerExceptionFromSignal) // (addr, self) UNREACHABLE END_FUNCTION art_quick_throw_null_pointer_exception /* * Called by managed code to create and deliver an ArithmeticException. */ NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_div_zero, artThrowDivZeroFromCode /* * Called by managed code to create and deliver a StackOverflowError. */ NO_ARG_RUNTIME_EXCEPTION art_quick_throw_stack_overflow, artThrowStackOverflowFromCode /* * Called by managed code, saves callee saves and then calls artThrowException * that will place a mock Method* at the bottom of the stack. Arg1 holds the exception. */ ONE_ARG_RUNTIME_EXCEPTION art_quick_deliver_exception, artDeliverExceptionFromCode /* * Called by managed code to create and deliver an ArrayIndexOutOfBoundsException. Arg1 holds * index, arg2 holds limit. */ TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_array_bounds, artThrowArrayBoundsFromCode /* * Called by managed code to create and deliver a StringIndexOutOfBoundsException * as if thrown from a call to String.charAt(). Arg1 holds index, arg2 holds limit. */ TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_string_bounds, artThrowStringBoundsFromCode /* * All generated callsites for interface invokes and invocation slow paths will load arguments * as usual - except instead of loading arg0/r0 with the target Method*, arg0/r0 will contain * the method_idx. This wrapper will save arg1-arg3 and call the appropriate C helper. * NOTE: "this" is first visible argument of the target, and so can be found in arg1/r1. * * The helper will attempt to locate the target and return a 64-bit result in r0/r1 consisting * of the target Method* in r0 and method->code_ in r1. * * If unsuccessful, the helper will return null/null and there will be a pending exception in the * thread and we branch to another stub to deliver it. * * On success this wrapper will restore arguments and *jump* to the target, leaving the lr * pointing back to the original caller. */ MACRO1(INVOKE_TRAMPOLINE_BODY, cxx_name) SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, ebx movl %esp, %edx // remember SP // Outgoing argument set up PUSH edx // pass SP pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*, SP) movl %edx, %edi // save code pointer in EDI addl MACRO_LITERAL(20), %esp // Pop arguments skip eax CFI_ADJUST_CFA_OFFSET(-20) // Restore FPRs. movsd 0(%esp), %xmm0 movsd 8(%esp), %xmm1 movsd 16(%esp), %xmm2 movsd 24(%esp), %xmm3 // Remove space for FPR args. addl MACRO_LITERAL(4 * 8), %esp CFI_ADJUST_CFA_OFFSET(-4 * 8) POP ecx // Restore args except eax POP edx POP ebx POP ebp // Restore callee saves POP esi // Swap EDI callee save with code pointer. xchgl %edi, (%esp) testl %eax, %eax // Branch forward if exception pending. jz 1f // Tail call to intended method. ret 1: addl MACRO_LITERAL(4), %esp // Pop code pointer off stack CFI_ADJUST_CFA_OFFSET(-4) DELIVER_PENDING_EXCEPTION END_MACRO MACRO2(INVOKE_TRAMPOLINE, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) INVOKE_TRAMPOLINE_BODY RAW_VAR(cxx_name) END_FUNCTION VAR(c_name) END_MACRO INVOKE_TRAMPOLINE art_quick_invoke_interface_trampoline_with_access_check, artInvokeInterfaceTrampolineWithAccessCheck INVOKE_TRAMPOLINE art_quick_invoke_static_trampoline_with_access_check, artInvokeStaticTrampolineWithAccessCheck INVOKE_TRAMPOLINE art_quick_invoke_direct_trampoline_with_access_check, artInvokeDirectTrampolineWithAccessCheck INVOKE_TRAMPOLINE art_quick_invoke_super_trampoline_with_access_check, artInvokeSuperTrampolineWithAccessCheck INVOKE_TRAMPOLINE art_quick_invoke_virtual_trampoline_with_access_check, artInvokeVirtualTrampolineWithAccessCheck /* * Helper for quick invocation stub to set up XMM registers. * Increments shorty and arg_array and clobbers temp_char. * Branches to finished if it encounters the end of the shorty. */ MACRO5(LOOP_OVER_SHORTY_LOADING_XMMS, xmm_reg, shorty, arg_array, temp_char, finished) 1: // LOOP movb (REG_VAR(shorty)), REG_VAR(temp_char) // temp_char := *shorty addl MACRO_LITERAL(1), REG_VAR(shorty) // shorty++ cmpb MACRO_LITERAL(0), REG_VAR(temp_char) // if (temp_char == '\0') je VAR(finished) // goto finished cmpb MACRO_LITERAL(68), REG_VAR(temp_char) // if (temp_char == 'D') je 2f // goto FOUND_DOUBLE cmpb MACRO_LITERAL(70), REG_VAR(temp_char) // if (temp_char == 'F') je 3f // goto FOUND_FLOAT addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++ // Handle extra space in arg array taken by a long. cmpb MACRO_LITERAL(74), REG_VAR(temp_char) // if (temp_char != 'J') jne 1b // goto LOOP addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++ jmp 1b // goto LOOP 2: // FOUND_DOUBLE movsd (REG_VAR(arg_array)), REG_VAR(xmm_reg) addl MACRO_LITERAL(8), REG_VAR(arg_array) // arg_array+=2 jmp 4f 3: // FOUND_FLOAT movss (REG_VAR(arg_array)), REG_VAR(xmm_reg) addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++ 4: END_MACRO /* * Helper for quick invocation stub to set up GPR registers. * Increments shorty and arg_array, and returns the current short character in * temp_char. Branches to finished if it encounters the end of the shorty. */ MACRO4(SKIP_OVER_FLOATS, shorty, arg_array, temp_char, finished) 1: // LOOP: movb (REG_VAR(shorty)), REG_VAR(temp_char) // temp_char := *shorty addl MACRO_LITERAL(1), REG_VAR(shorty) // shorty++ cmpb MACRO_LITERAL(0), REG_VAR(temp_char) // if (temp_char == '\0') je VAR(finished) // goto finished cmpb MACRO_LITERAL(70), REG_VAR(temp_char) // if (temp_char == 'F') je 3f // goto SKIP_FLOAT cmpb MACRO_LITERAL(68), REG_VAR(temp_char) // if (temp_char == 'D') je 4f // goto SKIP_DOUBLE jmp 5f // goto end 3: // SKIP_FLOAT addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++ jmp 1b // goto LOOP 4: // SKIP_DOUBLE addl MACRO_LITERAL(8), REG_VAR(arg_array) // arg_array+=2 jmp 1b // goto LOOP 5: END_MACRO /* * Quick invocation stub (non-static). * On entry: * [sp] = return address * [sp + 4] = method pointer * [sp + 8] = argument array or null for no argument methods * [sp + 12] = size of argument array in bytes * [sp + 16] = (managed) thread pointer * [sp + 20] = JValue* result * [sp + 24] = shorty */ DEFINE_FUNCTION art_quick_invoke_stub // Save the non-volatiles. PUSH ebp // save ebp PUSH ebx // save ebx PUSH esi // save esi PUSH edi // save edi // Set up argument XMM registers. mov 24+16(%esp), %esi // ESI := shorty + 1 ; ie skip return arg character. addl LITERAL(1), %esi mov 8+16(%esp), %edi // EDI := arg_array + 4 ; ie skip this pointer. addl LITERAL(4), %edi // Clobbers ESI, EDI, EAX. LOOP_OVER_SHORTY_LOADING_XMMS xmm0, esi, edi, al, .Lxmm_setup_finished LOOP_OVER_SHORTY_LOADING_XMMS xmm1, esi, edi, al, .Lxmm_setup_finished LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished LOOP_OVER_SHORTY_LOADING_XMMS xmm3, esi, edi, al, .Lxmm_setup_finished .balign 16 .Lxmm_setup_finished: mov %esp, %ebp // copy value of stack pointer into base pointer CFI_DEF_CFA_REGISTER(ebp) mov 28(%ebp), %ebx // get arg array size // reserve space for return addr, method*, ebx, ebp, esi, and edi in frame addl LITERAL(36), %ebx // align frame size to 16 bytes andl LITERAL(0xFFFFFFF0), %ebx subl LITERAL(20), %ebx // remove space for return address, ebx, ebp, esi and edi subl %ebx, %esp // reserve stack space for argument array movl LITERAL(0), (%esp) // store null for method* // Copy arg array into stack. movl 28(%ebp), %ecx // ECX = size of args movl 24(%ebp), %esi // ESI = argument array leal 4(%esp), %edi // EDI = just after Method* in stack arguments rep movsb // while (ecx--) { *edi++ = *esi++ } mov 40(%ebp), %esi // ESI := shorty + 1 ; ie skip return arg character. addl LITERAL(1), %esi mov 24(%ebp), %edi // EDI := arg_array mov 0(%edi), %ecx // ECX := this pointer addl LITERAL(4), %edi // EDI := arg_array + 4 ; ie skip this pointer. // Enumerate the possible cases for loading GPRS. // edx (and maybe ebx): SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG je .LfirstLong // Must be an integer value. movl (%edi), %edx addl LITERAL(4), %edi // arg_array++ // Now check ebx SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished // Must be first word of a long, or an integer. First word of long doesn't // go into EBX, but can be loaded there anyways, as it is harmless. movl (%edi), %ebx jmp .Lgpr_setup_finished .LfirstLong: movl (%edi), %edx movl 4(%edi), %ebx // Nothing left to load. .Lgpr_setup_finished: mov 20(%ebp), %eax // move method pointer into eax call *ART_METHOD_QUICK_CODE_OFFSET_32(%eax) // call the method mov %ebp, %esp // restore stack pointer CFI_DEF_CFA_REGISTER(esp) POP edi // pop edi POP esi // pop esi POP ebx // pop ebx POP ebp // pop ebp mov 20(%esp), %ecx // get result pointer mov %eax, (%ecx) // store the result assuming its a long, int or Object* mov %edx, 4(%ecx) // store the other half of the result mov 24(%esp), %edx // get the shorty cmpb LITERAL(68), (%edx) // test if result type char == 'D' je .Lreturn_double_quick cmpb LITERAL(70), (%edx) // test if result type char == 'F' je .Lreturn_float_quick ret .Lreturn_double_quick: movsd %xmm0, (%ecx) // store the floating point result ret .Lreturn_float_quick: movss %xmm0, (%ecx) // store the floating point result ret END_FUNCTION art_quick_invoke_stub /* * Quick invocation stub (static). * On entry: * [sp] = return address * [sp + 4] = method pointer * [sp + 8] = argument array or null for no argument methods * [sp + 12] = size of argument array in bytes * [sp + 16] = (managed) thread pointer * [sp + 20] = JValue* result * [sp + 24] = shorty */ DEFINE_FUNCTION art_quick_invoke_static_stub // Save the non-volatiles. PUSH ebp // save ebp PUSH ebx // save ebx PUSH esi // save esi PUSH edi // save edi // Set up argument XMM registers. mov 24+16(%esp), %esi // ESI := shorty + 1 ; ie skip return arg character. addl LITERAL(1), %esi mov 8+16(%esp), %edi // EDI := arg_array // Clobbers ESI, EDI, EAX. LOOP_OVER_SHORTY_LOADING_XMMS xmm0, esi, edi, al, .Lxmm_setup_finished2 LOOP_OVER_SHORTY_LOADING_XMMS xmm1, esi, edi, al, .Lxmm_setup_finished2 LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished2 LOOP_OVER_SHORTY_LOADING_XMMS xmm3, esi, edi, al, .Lxmm_setup_finished2 .balign 16 .Lxmm_setup_finished2: mov %esp, %ebp // copy value of stack pointer into base pointer CFI_DEF_CFA_REGISTER(ebp) mov 28(%ebp), %ebx // get arg array size // reserve space for return addr, method*, ebx, ebp, esi, and edi in frame addl LITERAL(36), %ebx // align frame size to 16 bytes andl LITERAL(0xFFFFFFF0), %ebx subl LITERAL(20), %ebx // remove space for return address, ebx, ebp, esi and edi subl %ebx, %esp // reserve stack space for argument array movl LITERAL(0), (%esp) // store null for method* // Copy arg array into stack. movl 28(%ebp), %ecx // ECX = size of args movl 24(%ebp), %esi // ESI = argument array leal 4(%esp), %edi // EDI = just after Method* in stack arguments rep movsb // while (ecx--) { *edi++ = *esi++ } mov 40(%ebp), %esi // ESI := shorty + 1 ; ie skip return arg character. addl LITERAL(1), %esi mov 24(%ebp), %edi // EDI := arg_array // Enumerate the possible cases for loading GPRS. // ecx (and maybe edx) SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2 cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG je .LfirstLong2 // Must be an integer value. Load into ECX. movl (%edi), %ecx addl LITERAL(4), %edi // arg_array++ // Now check edx (and maybe ebx). SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2 cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG je .LSecondLong2 // Must be an integer. Load into EDX. movl (%edi), %edx addl LITERAL(4), %edi // arg_array++ // Is there anything for ebx? SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2 // Must be first word of a long, or an integer. First word of long doesn't // go into EBX, but can be loaded there anyways, as it is harmless. movl (%edi), %ebx jmp .Lgpr_setup_finished2 .LSecondLong2: // EDX:EBX is long. That is all. movl (%edi), %edx movl 4(%edi), %ebx jmp .Lgpr_setup_finished2 .LfirstLong2: // ECX:EDX is a long movl (%edi), %ecx movl 4(%edi), %edx addl LITERAL(8), %edi // arg_array += 2 // Anything for EBX? SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2 // Must be first word of a long, or an integer. First word of long doesn't // go into EBX, but can be loaded there anyways, as it is harmless. movl (%edi), %ebx jmp .Lgpr_setup_finished2 // Nothing left to load. .Lgpr_setup_finished2: mov 20(%ebp), %eax // move method pointer into eax call *ART_METHOD_QUICK_CODE_OFFSET_32(%eax) // call the method mov %ebp, %esp // restore stack pointer CFI_DEF_CFA_REGISTER(esp) POP edi // pop edi POP esi // pop esi POP ebx // pop ebx POP ebp // pop ebp mov 20(%esp), %ecx // get result pointer mov %eax, (%ecx) // store the result assuming its a long, int or Object* mov %edx, 4(%ecx) // store the other half of the result mov 24(%esp), %edx // get the shorty cmpb LITERAL(68), (%edx) // test if result type char == 'D' je .Lreturn_double_quick2 cmpb LITERAL(70), (%edx) // test if result type char == 'F' je .Lreturn_float_quick2 ret .Lreturn_double_quick2: movsd %xmm0, (%ecx) // store the floating point result ret .Lreturn_float_quick2: movss %xmm0, (%ecx) // store the floating point result ret END_FUNCTION art_quick_invoke_static_stub MACRO3(ONE_ARG_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl MACRO_LITERAL(8), %esp // push padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(TWO_ARG_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up PUSH eax // push padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(THREE_ARG_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH edx // pass arg3 PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(FOUR_ARG_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME_PRESERVE_GOT_REG ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl MACRO_LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ebx // pass arg4 PUSH edx // pass arg3 PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, arg4, Thread*) addl MACRO_LITERAL(32), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-32) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(ONE_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl MACRO_LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(TWO_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up PUSH eax // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, referrer, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO MACRO3(THREE_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH edx // pass arg3 PUSH ecx // pass arg2 PUSH eax // pass arg1 call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-32) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address CALL_MACRO(return_macro) // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO /* * Macro for resolution and initialization of indexed DEX file * constants such as classes and strings. */ MACRO3(ONE_ARG_SAVE_EVERYTHING_DOWNCALL, c_name, cxx_name, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET) DEFINE_FUNCTION VAR(c_name) SETUP_SAVE_EVERYTHING_FRAME ebx, ebx, \runtime_method_offset // save ref containing registers for GC // Outgoing argument set up subl MACRO_LITERAL(8), %esp // push padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass the index of the constant as arg1 call CALLVAR(cxx_name) // cxx_name(arg1, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) testl %eax, %eax // If result is null, deliver the OOME. jz 1f CFI_REMEMBER_STATE RESTORE_SAVE_EVERYTHING_FRAME_KEEP_EAX // restore frame up to return address ret // return CFI_RESTORE_STATE_AND_DEF_CFA(esp, FRAME_SIZE_SAVE_EVERYTHING) 1: DELIVER_PENDING_EXCEPTION_FRAME_READY END_FUNCTION VAR(c_name) END_MACRO MACRO2(ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT, c_name, cxx_name) ONE_ARG_SAVE_EVERYTHING_DOWNCALL \c_name, \cxx_name, RUNTIME_SAVE_EVERYTHING_FOR_CLINIT_METHOD_OFFSET END_MACRO MACRO0(RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER) testl %eax, %eax // eax == 0 ? jz 1f // if eax == 0 goto 1 ret // return 1: // deliver exception on current thread DELIVER_PENDING_EXCEPTION END_MACRO MACRO0(RETURN_IF_EAX_ZERO) testl %eax, %eax // eax == 0 ? jnz 1f // if eax != 0 goto 1 ret // return 1: // deliver exception on current thread DELIVER_PENDING_EXCEPTION END_MACRO MACRO0(RETURN_OR_DELIVER_PENDING_EXCEPTION) cmpl MACRO_LITERAL(0),%fs:THREAD_EXCEPTION_OFFSET // exception field == 0 ? jne 1f // if exception field != 0 goto 1 ret // return 1: // deliver exception on current thread DELIVER_PENDING_EXCEPTION END_MACRO // Generate the allocation entrypoints for each allocator. GENERATE_ALLOC_ENTRYPOINTS_FOR_NON_TLAB_ALLOCATORS // Comment out allocators that have x86 specific asm. // Region TLAB: // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_INITIALIZED(_region_tlab, RegionTLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_WITH_ACCESS_CHECK(_region_tlab, RegionTLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_OBJECT(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED8(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED16(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED32(_region_tlab, RegionTLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED64(_region_tlab, RegionTLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_BYTES(_region_tlab, RegionTLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_CHARS(_region_tlab, RegionTLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_STRING(_region_tlab, RegionTLAB) // Normal TLAB: // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_INITIALIZED(_tlab, TLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_WITH_ACCESS_CHECK(_tlab, TLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_OBJECT(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED8(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED16(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED32(_tlab, TLAB) // GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED64(_tlab, TLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_BYTES(_tlab, TLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_CHARS(_tlab, TLAB) GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_STRING(_tlab, TLAB) // A hand-written override for GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_rosalloc, RosAlloc). MACRO2(ART_QUICK_ALLOC_OBJECT_ROSALLOC, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) // Fast path rosalloc allocation. // eax: type/return value // ecx, ebx, edx: free movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread // Check if the thread local allocation // stack has room movl THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx), %ecx cmpl THREAD_LOCAL_ALLOC_STACK_END_OFFSET(%ebx), %ecx jae .Lslow_path\c_name movl MIRROR_CLASS_OBJECT_SIZE_ALLOC_FAST_PATH_OFFSET(%eax), %ecx // Load the object size (ecx) // Check if the size is for a thread // local allocation. Also does the // finalizable and initialization check. cmpl LITERAL(ROSALLOC_MAX_THREAD_LOCAL_BRACKET_SIZE), %ecx ja .Lslow_path\c_name shrl LITERAL(ROSALLOC_BRACKET_QUANTUM_SIZE_SHIFT), %ecx // Calculate the rosalloc bracket index // from object size. // Load thread local rosalloc run (ebx) // Subtract __SIZEOF_POINTER__ to subtract // one from edi as there is no 0 byte run // and the size is already aligned. movl (THREAD_ROSALLOC_RUNS_OFFSET - __SIZEOF_POINTER__)(%ebx, %ecx, __SIZEOF_POINTER__), %ebx // Load free_list head (edi), // this will be the return value. movl (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_HEAD_OFFSET)(%ebx), %ecx jecxz .Lslow_path\c_name // Point of no slow path. Won't go to // the slow path from here on. // Load the next pointer of the head // and update head of free list with // next pointer movl ROSALLOC_SLOT_NEXT_OFFSET(%ecx), %edx movl %edx, (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_HEAD_OFFSET)(%ebx) // Decrement size of free list by 1 decl (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_SIZE_OFFSET)(%ebx) // Store the class pointer in the // header. This also overwrites the // next pointer. The offsets are // asserted to match. #if ROSALLOC_SLOT_NEXT_OFFSET != MIRROR_OBJECT_CLASS_OFFSET #error "Class pointer needs to overwrite next pointer." #endif POISON_HEAP_REF eax movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%ecx) movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread // Push the new object onto the thread // local allocation stack and // increment the thread local // allocation stack top. movl THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx), %eax movl %ecx, (%eax) addl LITERAL(COMPRESSED_REFERENCE_SIZE), %eax movl %eax, THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx) // No fence needed for x86. movl %ecx, %eax // Move object to return register ret .Lslow_path\c_name: SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax call SYMBOL(artAllocObjectFromCodeResolvedRosAlloc) // cxx_name(arg0, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception END_FUNCTION VAR(c_name) END_MACRO ART_QUICK_ALLOC_OBJECT_ROSALLOC art_quick_alloc_object_resolved_rosalloc, artAllocObjectFromCodeResolvedRosAlloc ART_QUICK_ALLOC_OBJECT_ROSALLOC art_quick_alloc_object_initialized_rosalloc, artAllocObjectFromCodeInitializedRosAlloc // The common fast path code for art_quick_alloc_object_resolved/initialized_tlab // and art_quick_alloc_object_resolved/initialized_region_tlab. // // EAX: type/return_value MACRO1(ALLOC_OBJECT_RESOLVED_TLAB_FAST_PATH, slowPathLabel) movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread movl THREAD_LOCAL_END_OFFSET(%ebx), %edi // Load thread_local_end. subl THREAD_LOCAL_POS_OFFSET(%ebx), %edi // Compute the remaining buffer size. movl MIRROR_CLASS_OBJECT_SIZE_ALLOC_FAST_PATH_OFFSET(%eax), %ecx // Load the object size. cmpl %edi, %ecx // Check if it fits. ja VAR(slowPathLabel) movl THREAD_LOCAL_POS_OFFSET(%ebx), %edx // Load thread_local_pos // as allocated object. addl %edx, %ecx // Add the object size. movl %ecx, THREAD_LOCAL_POS_OFFSET(%ebx) // Update thread_local_pos. incl THREAD_LOCAL_OBJECTS_OFFSET(%ebx) // Increase thread_local_objects. // Store the class pointer in the header. // No fence needed for x86. POISON_HEAP_REF eax movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%edx) movl %edx, %eax POP edi ret // Fast path succeeded. END_MACRO // The common slow path code for art_quick_alloc_object_resolved/initialized_tlab // and art_quick_alloc_object_resolved/initialized_region_tlab. MACRO1(ALLOC_OBJECT_RESOLVED_TLAB_SLOW_PATH, cxx_name) POP edi SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax call CALLVAR(cxx_name) // cxx_name(arg0, Thread*) addl LITERAL(16), %esp CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception END_MACRO MACRO2(ART_QUICK_ALLOC_OBJECT_TLAB, c_name, cxx_name) DEFINE_FUNCTION VAR(c_name) // Fast path tlab allocation. // EAX: type // EBX, ECX, EDX: free. PUSH edi ALLOC_OBJECT_RESOLVED_TLAB_FAST_PATH .Lslow_path\c_name .Lslow_path\c_name: ALLOC_OBJECT_RESOLVED_TLAB_SLOW_PATH RAW_VAR(cxx_name) END_FUNCTION VAR(c_name) END_MACRO ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_resolved_tlab, artAllocObjectFromCodeResolvedTLAB ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_initialized_tlab, artAllocObjectFromCodeInitializedTLAB ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_resolved_region_tlab, artAllocObjectFromCodeResolvedRegionTLAB ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_initialized_region_tlab, artAllocObjectFromCodeInitializedRegionTLAB // The fast path code for art_quick_alloc_array_region_tlab. // Inputs: EAX: the class, ECX: int32_t component_count, EDX: total_size // Free temp: EBX // Output: EAX: return value. MACRO1(ALLOC_ARRAY_TLAB_FAST_PATH_RESOLVED_WITH_SIZE, slowPathLabel) mov %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread // Mask out the unaligned part to make sure we are 8 byte aligned. andl LITERAL(OBJECT_ALIGNMENT_MASK_TOGGLED), %edx movl THREAD_LOCAL_END_OFFSET(%ebx), %edi subl THREAD_LOCAL_POS_OFFSET(%ebx), %edi cmpl %edi, %edx // Check if it fits. ja RAW_VAR(slowPathLabel) movl THREAD_LOCAL_POS_OFFSET(%ebx), %edi addl %edi, %edx // Add the object size. movl %edx, THREAD_LOCAL_POS_OFFSET(%ebx) // Update thread_local_pos_ addl LITERAL(1), THREAD_LOCAL_OBJECTS_OFFSET(%ebx) // Increase thread_local_objects. // Store the class pointer in the // header. // No fence needed for x86. POISON_HEAP_REF eax movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%edi) movl %ecx, MIRROR_ARRAY_LENGTH_OFFSET(%edi) movl %edi, %eax POP edi ret // Fast path succeeded. END_MACRO MACRO1(COMPUTE_ARRAY_SIZE_UNKNOWN, slow_path) // We should never enter here. Code is provided for reference. int3 // Possibly a large object, go slow. // Also does negative array size check. cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_WIDE_ARRAY_DATA_OFFSET) / 8), %ecx ja RAW_VAR(slow_path) PUSH ecx movl %ecx, %edx movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%eax), %ecx // Load component type. UNPOISON_HEAP_REF ecx movl MIRROR_CLASS_OBJECT_PRIMITIVE_TYPE_OFFSET(%ecx), %ecx // Load primitive type. shr MACRO_LITERAL(PRIMITIVE_TYPE_SIZE_SHIFT_SHIFT), %ecx // Get component size shift. sall %cl, %edx // Calculate array count shifted. // Add array header + alignment rounding. add MACRO_LITERAL(MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK), %edx // Add 4 extra bytes if we are doing a long array. add MACRO_LITERAL(1), %ecx and MACRO_LITERAL(4), %ecx #if MIRROR_WIDE_ARRAY_DATA_OFFSET != MIRROR_INT_ARRAY_DATA_OFFSET + 4 #error Long array data offset must be 4 greater than int array data offset. #endif addl %ecx, %edx POP ecx END_MACRO MACRO1(COMPUTE_ARRAY_SIZE_8, slow_path) // EAX: mirror::Class* klass, ECX: int32_t component_count // Possibly a large object, go slow. // Also does negative array size check. cmpl LITERAL(MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET), %ecx ja RAW_VAR(slow_path) // Add array header + alignment rounding. leal (MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK)(%ecx), %edx END_MACRO MACRO1(COMPUTE_ARRAY_SIZE_16, slow_path) // EAX: mirror::Class* klass, ECX: int32_t component_count // Possibly a large object, go slow. // Also does negative array size check. cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET) / 2), %ecx ja RAW_VAR(slow_path) // Add array header + alignment rounding. leal ((MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 2)(%ecx), %edx sall MACRO_LITERAL(1), %edx END_MACRO MACRO1(COMPUTE_ARRAY_SIZE_32, slow_path) // EAX: mirror::Class* klass, ECX: int32_t component_count // Possibly a large object, go slow. // Also does negative array size check. cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET) / 4), %ecx ja RAW_VAR(slow_path) // Add array header + alignment rounding. leal ((MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 4)(%ecx), %edx sall MACRO_LITERAL(2), %edx END_MACRO MACRO1(COMPUTE_ARRAY_SIZE_64, slow_path) // EAX: mirror::Class* klass, ECX: int32_t component_count // Possibly a large object, go slow. // Also does negative array size check. cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_WIDE_ARRAY_DATA_OFFSET) / 8), %ecx ja RAW_VAR(slow_path) // Add array header + alignment rounding. leal ((MIRROR_WIDE_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 8)(%ecx), %edx sall MACRO_LITERAL(3), %edx END_MACRO MACRO3(GENERATE_ALLOC_ARRAY_TLAB, c_entrypoint, cxx_name, size_setup) DEFINE_FUNCTION VAR(c_entrypoint) // EAX: mirror::Class* klass, ECX: int32_t component_count PUSH edi CALL_MACRO(size_setup) .Lslow_path\c_entrypoint ALLOC_ARRAY_TLAB_FAST_PATH_RESOLVED_WITH_SIZE .Lslow_path\c_entrypoint .Lslow_path\c_entrypoint: POP edi SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up PUSH eax // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx PUSH eax call CALLVAR(cxx_name) // cxx_name(arg0, arg1, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception END_FUNCTION VAR(c_entrypoint) END_MACRO GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_UNKNOWN GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved8_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_8 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved16_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_16 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved32_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_32 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved64_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_64 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_UNKNOWN GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved8_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_8 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved16_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_16 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved32_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_32 GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved64_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_64 ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT art_quick_initialize_static_storage, artInitializeStaticStorageFromCode ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT art_quick_resolve_type, artResolveTypeFromCode ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_type_and_verify_access, artResolveTypeAndVerifyAccessFromCode ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_method_handle, artResolveMethodHandleFromCode ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_method_type, artResolveMethodTypeFromCode ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_string, artResolveStringFromCode TWO_ARG_REF_DOWNCALL art_quick_handle_fill_data, artHandleFillArrayDataFromCode, RETURN_IF_EAX_ZERO DEFINE_FUNCTION art_quick_lock_object testl %eax, %eax // null check object/eax jz .Lslow_lock .Lretry_lock: movl MIRROR_OBJECT_LOCK_WORD_OFFSET(%eax), %ecx // ecx := lock word test LITERAL(LOCK_WORD_STATE_MASK_SHIFTED), %ecx // test the 2 high bits. jne .Lslow_lock // slow path if either of the two high bits are set. movl %ecx, %edx // save lock word (edx) to keep read barrier bits. andl LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED), %ecx // zero the gc bits. test %ecx, %ecx jnz .Lalready_thin // lock word contains a thin lock // unlocked case - edx: original lock word, eax: obj. movl %eax, %ecx // remember object in case of retry movl %edx, %eax // eax: lock word zero except for read barrier bits. movl %fs:THREAD_ID_OFFSET, %edx // load thread id. or %eax, %edx // edx: thread id with count of 0 + read barrier bits. lock cmpxchg %edx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%ecx) // eax: old val, edx: new val. jnz .Llock_cmpxchg_fail // cmpxchg failed retry ret .Lalready_thin: // edx: lock word (with high 2 bits zero and original rb bits), eax: obj. movl %fs:THREAD_ID_OFFSET, %ecx // ecx := thread id cmpw %cx, %dx // do we hold the lock already? jne .Lslow_lock movl %edx, %ecx // copy the lock word to check count overflow. andl LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED), %ecx // zero the read barrier bits. addl LITERAL(LOCK_WORD_THIN_LOCK_COUNT_ONE), %ecx // increment recursion count for overflow check. test LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED), %ecx // overflowed if the first gc state bit is set. jne .Lslow_lock // count overflowed so go slow movl %eax, %ecx // save obj to use eax for cmpxchg. movl %edx, %eax // copy the lock word as the old val for cmpxchg. addl LITERAL(LOCK_WORD_THIN_LOCK_COUNT_ONE), %edx // increment recursion count again for real. // update lockword, cmpxchg necessary for read barrier bits. lock cmpxchg %edx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%ecx) // eax: old val, edx: new val. jnz .Llock_cmpxchg_fail // cmpxchg failed retry ret .Llock_cmpxchg_fail: movl %ecx, %eax // restore eax jmp .Lretry_lock .Lslow_lock: SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass object call SYMBOL(artLockObjectFromCode) // artLockObjectFromCode(object, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_EAX_ZERO END_FUNCTION art_quick_lock_object DEFINE_FUNCTION art_quick_lock_object_no_inline SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass object call SYMBOL(artLockObjectFromCode) // artLockObjectFromCode(object, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_EAX_ZERO END_FUNCTION art_quick_lock_object_no_inline DEFINE_FUNCTION art_quick_unlock_object testl %eax, %eax // null check object/eax jz .Lslow_unlock .Lretry_unlock: movl MIRROR_OBJECT_LOCK_WORD_OFFSET(%eax), %ecx // ecx := lock word movl %fs:THREAD_ID_OFFSET, %edx // edx := thread id test LITERAL(LOCK_WORD_STATE_MASK_SHIFTED), %ecx jnz .Lslow_unlock // lock word contains a monitor cmpw %cx, %dx // does the thread id match? jne .Lslow_unlock movl %ecx, %edx // copy the lock word to detect new count of 0. andl LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED), %edx // zero the gc bits. cmpl LITERAL(LOCK_WORD_THIN_LOCK_COUNT_ONE), %edx jae .Lrecursive_thin_unlock // update lockword, cmpxchg necessary for read barrier bits. movl %eax, %edx // edx: obj movl %ecx, %eax // eax: old lock word. andl LITERAL(LOCK_WORD_GC_STATE_MASK_SHIFTED), %ecx // ecx: new lock word zero except original rb bits. #ifndef USE_READ_BARRIER movl %ecx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%edx) #else lock cmpxchg %ecx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%edx) // eax: old val, ecx: new val. jnz .Lunlock_cmpxchg_fail // cmpxchg failed retry #endif ret .Lrecursive_thin_unlock: // ecx: original lock word, eax: obj // update lockword, cmpxchg necessary for read barrier bits. movl %eax, %edx // edx: obj movl %ecx, %eax // eax: old lock word. subl LITERAL(LOCK_WORD_THIN_LOCK_COUNT_ONE), %ecx // ecx: new lock word with decremented count. #ifndef USE_READ_BARRIER mov %ecx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%edx) #else lock cmpxchg %ecx, MIRROR_OBJECT_LOCK_WORD_OFFSET(%edx) // eax: old val, ecx: new val. jnz .Lunlock_cmpxchg_fail // cmpxchg failed retry #endif ret .Lunlock_cmpxchg_fail: // edx: obj movl %edx, %eax // restore eax jmp .Lretry_unlock .Lslow_unlock: SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass object call SYMBOL(artUnlockObjectFromCode) // artUnlockObjectFromCode(object, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_EAX_ZERO END_FUNCTION art_quick_unlock_object DEFINE_FUNCTION art_quick_unlock_object_no_inline SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass object call SYMBOL(artUnlockObjectFromCode) // artUnlockObjectFromCode(object, Thread*) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_EAX_ZERO END_FUNCTION art_quick_unlock_object_no_inline DEFINE_FUNCTION art_quick_instance_of PUSH eax // alignment padding PUSH ecx // pass arg2 - obj->klass PUSH eax // pass arg1 - checked class call SYMBOL(artInstanceOfFromCode) // (Object* obj, Class* ref_klass) addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_instance_of DEFINE_FUNCTION art_quick_check_instance_of // Type check using the bit string passes null as the target class. In that case just throw. testl %ecx, %ecx jz .Lthrow_class_cast_exception_for_bitstring_check PUSH eax // alignment padding PUSH ecx // pass arg2 - checked class PUSH eax // pass arg1 - obj call SYMBOL(artInstanceOfFromCode) // (Object* obj, Class* ref_klass) testl %eax, %eax jz .Lthrow_class_cast_exception // jump forward if not assignable addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret CFI_ADJUST_CFA_OFFSET(12) // Reset unwind info so following code unwinds. .Lthrow_class_cast_exception: POP eax // pop arguments POP ecx addl LITERAL(4), %esp CFI_ADJUST_CFA_OFFSET(-4) .Lthrow_class_cast_exception_for_bitstring_check: SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up PUSH eax // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass arg2 PUSH eax // pass arg1 call SYMBOL(artThrowClassCastExceptionForObject) // (Object* src, Class* dest, Thread*) UNREACHABLE END_FUNCTION art_quick_check_instance_of // Restore reg's value if reg is not the same as exclude_reg, otherwise just adjust stack. MACRO2(POP_REG_NE, reg, exclude_reg) .ifc RAW_VAR(reg), RAW_VAR(exclude_reg) addl MACRO_LITERAL(4), %esp CFI_ADJUST_CFA_OFFSET(-4) .else POP RAW_VAR(reg) .endif END_MACRO /* * Macro to insert read barrier, only used in art_quick_aput_obj. * obj_reg and dest_reg are registers, offset is a defined literal such as * MIRROR_OBJECT_CLASS_OFFSET. * pop_eax is a boolean flag, indicating if eax is popped after the call. * TODO: When read barrier has a fast path, add heap unpoisoning support for the fast path. */ MACRO4(READ_BARRIER, obj_reg, offset, dest_reg, pop_eax) #ifdef USE_READ_BARRIER PUSH eax // save registers used in art_quick_aput_obj PUSH ebx PUSH edx PUSH ecx // Outgoing argument set up pushl MACRO_LITERAL((RAW_VAR(offset))) // pass offset, double parentheses are necessary CFI_ADJUST_CFA_OFFSET(4) PUSH RAW_VAR(obj_reg) // pass obj_reg PUSH eax // pass ref, just pass eax for now since parameter ref is unused call SYMBOL(artReadBarrierSlow) // artReadBarrierSlow(ref, obj_reg, offset) // No need to unpoison return value in eax, artReadBarrierSlow() would do the unpoisoning. .ifnc RAW_VAR(dest_reg), eax movl %eax, REG_VAR(dest_reg) // save loaded ref in dest_reg .endif addl MACRO_LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) POP_REG_NE ecx, RAW_VAR(dest_reg) // Restore args except dest_reg POP_REG_NE edx, RAW_VAR(dest_reg) POP_REG_NE ebx, RAW_VAR(dest_reg) .ifc RAW_VAR(pop_eax), true POP_REG_NE eax, RAW_VAR(dest_reg) .endif #else movl RAW_VAR(offset)(REG_VAR(obj_reg)), REG_VAR(dest_reg) UNPOISON_HEAP_REF RAW_VAR(dest_reg) #endif // USE_READ_BARRIER END_MACRO DEFINE_FUNCTION art_quick_aput_obj test %edx, %edx // store of null jz .Ldo_aput_null READ_BARRIER eax, MIRROR_OBJECT_CLASS_OFFSET, ebx, true READ_BARRIER ebx, MIRROR_CLASS_COMPONENT_TYPE_OFFSET, ebx, true // value's type == array's component type - trivial assignability #if defined(USE_READ_BARRIER) READ_BARRIER edx, MIRROR_OBJECT_CLASS_OFFSET, eax, false cmpl %eax, %ebx POP eax // restore eax from the push in the beginning of READ_BARRIER macro // This asymmetric push/pop saves a push of eax and maintains stack alignment. #elif defined(USE_HEAP_POISONING) PUSH eax // save eax movl MIRROR_OBJECT_CLASS_OFFSET(%edx), %eax UNPOISON_HEAP_REF eax cmpl %eax, %ebx POP eax // restore eax #else cmpl MIRROR_OBJECT_CLASS_OFFSET(%edx), %ebx #endif jne .Lcheck_assignability .Ldo_aput: POISON_HEAP_REF edx movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4) movl %fs:THREAD_CARD_TABLE_OFFSET, %edx shrl LITERAL(CARD_TABLE_CARD_SHIFT), %eax movb %dl, (%edx, %eax) ret .Ldo_aput_null: movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4) ret .Lcheck_assignability: PUSH eax // save arguments PUSH ecx PUSH edx #if defined(USE_READ_BARRIER) subl LITERAL(4), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(4) READ_BARRIER edx, MIRROR_OBJECT_CLASS_OFFSET, eax, true subl LITERAL(4), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass arg2 - type of the value to be stored #elif defined(USE_HEAP_POISONING) subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) movl MIRROR_OBJECT_CLASS_OFFSET(%edx), %eax UNPOISON_HEAP_REF eax PUSH eax // pass arg2 - type of the value to be stored #else subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl MIRROR_OBJECT_CLASS_OFFSET(%edx) // pass arg2 - type of the value to be stored CFI_ADJUST_CFA_OFFSET(4) #endif PUSH ebx // pass arg1 - component type of the array call SYMBOL(artIsAssignableFromCode) // (Class* a, Class* b) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) testl %eax, %eax jz .Lthrow_array_store_exception POP edx POP ecx POP eax POISON_HEAP_REF edx movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4) // do the aput movl %fs:THREAD_CARD_TABLE_OFFSET, %edx shrl LITERAL(CARD_TABLE_CARD_SHIFT), %eax movb %dl, (%edx, %eax) ret CFI_ADJUST_CFA_OFFSET(12) // 3 POP after the jz for unwinding. .Lthrow_array_store_exception: POP edx POP ecx POP eax SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx, ebx // save all registers as basis for long jump context // Outgoing argument set up PUSH eax // alignment padding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH edx // pass arg2 - value PUSH eax // pass arg1 - array call SYMBOL(artThrowArrayStoreException) // (array, value, Thread*) UNREACHABLE END_FUNCTION art_quick_aput_obj DEFINE_FUNCTION art_quick_memcpy SETUP_GOT_NOSAVE ebx // clobbers EBX PUSH edx // pass arg3 PUSH ecx // pass arg2 PUSH eax // pass arg1 call PLT_SYMBOL(memcpy) // (void*, const void*, size_t) addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_memcpy DEFINE_FUNCTION art_quick_test_suspend SETUP_SAVE_EVERYTHING_FRAME ebx, ebx, RUNTIME_SAVE_EVERYTHING_FOR_SUSPEND_CHECK_METHOD_OFFSET // save everything for GC // Outgoing argument set up subl MACRO_LITERAL(12), %esp // push padding CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artTestSuspendFromCode) // (Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_EVERYTHING_FRAME // restore frame up to return address ret // return END_FUNCTION art_quick_test_suspend DEFINE_FUNCTION art_quick_d2l subl LITERAL(12), %esp // alignment padding, room for argument CFI_ADJUST_CFA_OFFSET(12) movsd %xmm0, 0(%esp) // arg a call SYMBOL(art_d2l) // (jdouble a) addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_d2l DEFINE_FUNCTION art_quick_f2l subl LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) movss %xmm0, 0(%esp) // arg a call SYMBOL(art_f2l) // (jfloat a) addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_f2l DEFINE_FUNCTION art_quick_ldiv subl LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) PUSH ebx // pass arg4 b.hi PUSH edx // pass arg3 b.lo PUSH ecx // pass arg2 a.hi PUSH eax // pass arg1 a.lo call SYMBOL(artLdiv) // (jlong a, jlong b) addl LITERAL(28), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-28) ret END_FUNCTION art_quick_ldiv DEFINE_FUNCTION art_quick_lmod subl LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) PUSH ebx // pass arg4 b.hi PUSH edx // pass arg3 b.lo PUSH ecx // pass arg2 a.hi PUSH eax // pass arg1 a.lo call SYMBOL(artLmod) // (jlong a, jlong b) addl LITERAL(28), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-28) ret END_FUNCTION art_quick_lmod DEFINE_FUNCTION art_quick_lmul imul %eax, %ebx // ebx = a.lo(eax) * b.hi(ebx) imul %edx, %ecx // ecx = b.lo(edx) * a.hi(ecx) mul %edx // edx:eax = a.lo(eax) * b.lo(edx) add %ebx, %ecx add %ecx, %edx // edx += (a.lo * b.hi) + (b.lo * a.hi) ret END_FUNCTION art_quick_lmul DEFINE_FUNCTION art_quick_lshl // ecx:eax << edx xchg %edx, %ecx shld %cl,%eax,%edx shl %cl,%eax test LITERAL(32), %cl jz 1f mov %eax, %edx xor %eax, %eax 1: ret END_FUNCTION art_quick_lshl DEFINE_FUNCTION art_quick_lshr // ecx:eax >> edx xchg %edx, %ecx shrd %cl,%edx,%eax sar %cl,%edx test LITERAL(32),%cl jz 1f mov %edx, %eax sar LITERAL(31), %edx 1: ret END_FUNCTION art_quick_lshr DEFINE_FUNCTION art_quick_lushr // ecx:eax >>> edx xchg %edx, %ecx shrd %cl,%edx,%eax shr %cl,%edx test LITERAL(32),%cl jz 1f mov %edx, %eax xor %edx, %edx 1: ret END_FUNCTION art_quick_lushr // Note: Functions `art{Get,Set}{Static,Instance}FromCompiledCode` are // defined with a macro in runtime/entrypoints/quick/quick_field_entrypoints.cc. ONE_ARG_REF_DOWNCALL art_quick_get_boolean_static, artGetBooleanStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get_byte_static, artGetByteStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get_char_static, artGetCharStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get_short_static, artGetShortStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get32_static, artGet32StaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get64_static, artGet64StaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION ONE_ARG_REF_DOWNCALL art_quick_get_obj_static, artGetObjStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get_boolean_instance, artGetBooleanInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get_byte_instance, artGetByteInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get_char_instance, artGetCharInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get_short_instance, artGetShortInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get32_instance, artGet32InstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get64_instance, artGet64InstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_get_obj_instance, artGetObjInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION TWO_ARG_REF_DOWNCALL art_quick_set8_static, artSet8StaticFromCompiledCode, RETURN_IF_EAX_ZERO TWO_ARG_REF_DOWNCALL art_quick_set16_static, artSet16StaticFromCompiledCode, RETURN_IF_EAX_ZERO TWO_ARG_REF_DOWNCALL art_quick_set32_static, artSet32StaticFromCompiledCode, RETURN_IF_EAX_ZERO TWO_ARG_REF_DOWNCALL art_quick_set_obj_static, artSetObjStaticFromCompiledCode, RETURN_IF_EAX_ZERO THREE_ARG_REF_DOWNCALL art_quick_set64_static, artSet64StaticFromCompiledCode, RETURN_IF_EAX_ZERO THREE_ARG_REF_DOWNCALL art_quick_set8_instance, artSet8InstanceFromCompiledCode, RETURN_IF_EAX_ZERO THREE_ARG_REF_DOWNCALL art_quick_set16_instance, artSet16InstanceFromCompiledCode, RETURN_IF_EAX_ZERO THREE_ARG_REF_DOWNCALL art_quick_set32_instance, artSet32InstanceFromCompiledCode, RETURN_IF_EAX_ZERO THREE_ARG_REF_DOWNCALL art_quick_set_obj_instance, artSetObjInstanceFromCompiledCode, RETURN_IF_EAX_ZERO // Call artSet64InstanceFromCode with 4 word size arguments. DEFINE_FUNCTION art_quick_set64_instance movd %ebx, %xmm0 SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC movd %xmm0, %ebx // Outgoing argument set up subl LITERAL(12), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ebx // pass high half of new_val PUSH edx // pass low half of new_val PUSH ecx // pass object PUSH eax // pass field_idx call SYMBOL(artSet64InstanceFromCompiledCode) // (field_idx, Object*, new_val, Thread*) addl LITERAL(32), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-32) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_EAX_ZERO // return or deliver exception END_FUNCTION art_quick_set64_instance DEFINE_FUNCTION art_quick_proxy_invoke_handler SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX PUSH esp // pass SP pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass receiver PUSH eax // pass proxy method call SYMBOL(artQuickProxyInvokeHandler) // (proxy method, receiver, Thread*, SP) movd %eax, %xmm0 // place return value also into floating point return value movd %edx, %xmm1 punpckldq %xmm1, %xmm0 addl LITERAL(16 + FRAME_SIZE_SAVE_REFS_AND_ARGS - FRAME_SIZE_SAVE_REFS_ONLY), %esp CFI_ADJUST_CFA_OFFSET(-(16 + FRAME_SIZE_SAVE_REFS_AND_ARGS - FRAME_SIZE_SAVE_REFS_ONLY)) RESTORE_SAVE_REFS_ONLY_FRAME RETURN_OR_DELIVER_PENDING_EXCEPTION // return or deliver exception END_FUNCTION art_quick_proxy_invoke_handler /* * Called to resolve an imt conflict. * eax is the conflict ArtMethod. * xmm7 is a hidden argument that holds the target interface method's dex method index. * * Note that this stub writes to eax. * Because of lack of free registers, it also saves and restores edi. */ DEFINE_FUNCTION art_quick_imt_conflict_trampoline PUSH EDI PUSH ESI PUSH EDX movl 16(%esp), %edi // Load referrer. movd %xmm7, %esi // Get target method index stored in xmm7, remember it in ESI. // If the method is obsolete, just go through the dex cache miss slow path. // The obsolete flag is set with suspended threads, so we do not need an acquire operation here. testl LITERAL(ACC_OBSOLETE_METHOD), ART_METHOD_ACCESS_FLAGS_OFFSET(%edi) jnz .Limt_conflict_trampoline_dex_cache_miss movl ART_METHOD_DECLARING_CLASS_OFFSET(%edi), %edi // Load declaring class (no read barrier). movl MIRROR_CLASS_DEX_CACHE_OFFSET(%edi), %edi // Load the DexCache (without read barrier). UNPOISON_HEAP_REF edi movl MIRROR_DEX_CACHE_RESOLVED_METHODS_OFFSET(%edi), %edi // Load the resolved methods. pushl ART_METHOD_JNI_OFFSET_32(%eax) // Push ImtConflictTable. CFI_ADJUST_CFA_OFFSET(4) movl %esi, %eax // Copy the method index from ESI. andl LITERAL(METHOD_DEX_CACHE_SIZE_MINUS_ONE), %eax // Calculate DexCache method slot index. leal 0(%edi, %eax, 2 * __SIZEOF_POINTER__), %edi // Load DexCache method slot address. mov %ecx, %edx // Make EDX:EAX == ECX:EBX so that LOCK CMPXCHG8B makes no changes. mov %ebx, %eax // (The actual value does not matter.) lock cmpxchg8b (%edi) // Relaxed atomic load EDX:EAX from the dex cache slot. popl %edi // Pop ImtConflictTable. CFI_ADJUST_CFA_OFFSET(-4) cmp %edx, %esi // Compare method index to see if we had a DexCache method hit. jne .Limt_conflict_trampoline_dex_cache_miss .Limt_table_iterate: cmpl %eax, 0(%edi) jne .Limt_table_next_entry // We successfully hit an entry in the table. Load the target method // and jump to it. movl __SIZEOF_POINTER__(%edi), %eax CFI_REMEMBER_STATE POP EDX POP ESI POP EDI jmp *ART_METHOD_QUICK_CODE_OFFSET_32(%eax) CFI_RESTORE_STATE_AND_DEF_CFA(esp, 16) .Limt_table_next_entry: // If the entry is null, the interface method is not in the ImtConflictTable. cmpl LITERAL(0), 0(%edi) jz .Lconflict_trampoline // Iterate over the entries of the ImtConflictTable. addl LITERAL(2 * __SIZEOF_POINTER__), %edi jmp .Limt_table_iterate .Lconflict_trampoline: // Call the runtime stub to populate the ImtConflictTable and jump to the // resolved method. CFI_REMEMBER_STATE POP EDX POP ESI POP EDI INVOKE_TRAMPOLINE_BODY artInvokeInterfaceTrampoline CFI_RESTORE_STATE_AND_DEF_CFA(esp, 16) .Limt_conflict_trampoline_dex_cache_miss: // We're not creating a proper runtime method frame here, // artLookupResolvedMethod() is not allowed to walk the stack. // Save core register args; EDX is already saved. PUSH ebx PUSH ecx // Save FPR args. subl MACRO_LITERAL(32), %esp CFI_ADJUST_CFA_OFFSET(32) movsd %xmm0, 0(%esp) movsd %xmm1, 8(%esp) movsd %xmm2, 16(%esp) movsd %xmm3, 24(%esp) pushl 32+8+16(%esp) // Pass referrer. CFI_ADJUST_CFA_OFFSET(4) pushl %esi // Pass method index. CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artLookupResolvedMethod) // (uint32_t method_index, ArtMethod* referrer) addl LITERAL(8), %esp // Pop arguments. CFI_ADJUST_CFA_OFFSET(-8) // Restore FPR args. movsd 0(%esp), %xmm0 movsd 8(%esp), %xmm1 movsd 16(%esp), %xmm2 movsd 24(%esp), %xmm3 addl MACRO_LITERAL(32), %esp CFI_ADJUST_CFA_OFFSET(-32) // Restore core register args. POP ecx POP ebx cmp LITERAL(0), %eax // If the method wasn't resolved, je .Lconflict_trampoline // skip the lookup and go to artInvokeInterfaceTrampoline(). jmp .Limt_table_iterate END_FUNCTION art_quick_imt_conflict_trampoline DEFINE_FUNCTION art_quick_resolution_trampoline SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, ebx movl %esp, %edi PUSH EDI // pass SP. do not just PUSH ESP; that messes up unwinding pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // pass receiver PUSH eax // pass method call SYMBOL(artQuickResolutionTrampoline) // (Method* called, receiver, Thread*, SP) movl %eax, %edi // remember code pointer in EDI addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) test %eax, %eax // if code pointer is null goto deliver pending exception jz 1f RESTORE_SAVE_REFS_AND_ARGS_FRAME_AND_JUMP 1: RESTORE_SAVE_REFS_AND_ARGS_FRAME DELIVER_PENDING_EXCEPTION END_FUNCTION art_quick_resolution_trampoline DEFINE_FUNCTION art_quick_generic_jni_trampoline SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX movl %esp, %ebp // save SP at callee-save frame CFI_DEF_CFA_REGISTER(ebp) subl LITERAL(5120), %esp // prepare for artQuickGenericJniTrampoline call // (Thread*, managed_sp, reserved_area) // (esp) 4(esp) 8(esp) <= C calling convention // fs:... ebp esp <= where they are movl %esp, %eax subl LITERAL(4), %esp // Padding for 16B alignment. pushl %eax // Pass reserved area. pushl %ebp // Pass managed frame SP. pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current(). call SYMBOL(artQuickGenericJniTrampoline) // (Thread*, sp) // The C call will have registered the complete save-frame on success. // The result of the call is: // eax: pointer to native code, 0 on error. // The bottom of the reserved area contains values for arg registers, // hidden arg register and SP for out args for the call. // Check for error (class init check or locking for synchronized native method can throw). test %eax, %eax jz .Lexception_in_native // On x86 there are no registers passed, so no native call args to pop here. // Save code pointer in EDX. movl %eax, %edx // Load hidden arg (EAX) for @CriticalNative. movl 16(%esp), %eax // Load SP for out args, releasing unneeded reserved area. movl 20(%esp), %esp // Native call. call *%edx // result sign extension is handled in C code // prepare for artQuickGenericJniEndTrampoline call // (Thread*, result, result_f) // (esp) 4(esp) 12(esp) <= C calling convention // fs:... eax:edx fp0 <= where they are subl LITERAL(20), %esp // Padding & pass float result. fstpl (%esp) pushl %edx // Pass int result. pushl %eax pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current(). call SYMBOL(artQuickGenericJniEndTrampoline) // Pending exceptions possible. mov %fs:THREAD_EXCEPTION_OFFSET, %ebx testl %ebx, %ebx jnz .Lexception_in_native // Tear down the alloca. movl %ebp, %esp CFI_DEF_CFA_REGISTER(esp) // Tear down the callee-save frame. // Remove space for FPR args and EAX addl LITERAL(4 + 4 * 8), %esp CFI_ADJUST_CFA_OFFSET(-(4 + 4 * 8)) POP ecx addl LITERAL(4), %esp // Avoid edx, as it may be part of the result. CFI_ADJUST_CFA_OFFSET(-4) POP ebx POP ebp // Restore callee saves POP esi POP edi // Quick expects the return value to be in xmm0. movd %eax, %xmm0 movd %edx, %xmm1 punpckldq %xmm1, %xmm0 ret .Lexception_in_native: pushl %fs:THREAD_TOP_QUICK_FRAME_OFFSET addl LITERAL(-1), (%esp) // Remove the GenericJNI tag. movl (%esp), %esp // Do a call to push a new save-all frame required by the runtime. call .Lexception_call .Lexception_call: DELIVER_PENDING_EXCEPTION END_FUNCTION art_quick_generic_jni_trampoline DEFINE_FUNCTION art_quick_to_interpreter_bridge SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, ebx // save frame mov %esp, %edx // remember SP PUSH eax // alignment padding PUSH edx // pass SP pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) PUSH eax // pass method call SYMBOL(artQuickToInterpreterBridge) // (method, Thread*, SP) addl LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) // Return eax:edx in xmm0 also. movd %eax, %xmm0 movd %edx, %xmm1 punpckldq %xmm1, %xmm0 addl LITERAL(48), %esp // Remove FPRs and EAX, ECX, EDX, EBX. CFI_ADJUST_CFA_OFFSET(-48) POP ebp // Restore callee saves POP esi POP edi RETURN_OR_DELIVER_PENDING_EXCEPTION // return or deliver exception END_FUNCTION art_quick_to_interpreter_bridge /* * Called by managed code, saves callee saves and then calls artInvokeObsoleteMethod */ ONE_ARG_RUNTIME_EXCEPTION art_invoke_obsolete_method_stub, artInvokeObsoleteMethod /* * Routine that intercepts method calls and returns. */ DEFINE_FUNCTION art_quick_instrumentation_entry SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, edx PUSH eax // Save eax which will be clobbered by the callee-save method. subl LITERAL(16), %esp // Align stack (12 bytes) and reserve space for the SP argument CFI_ADJUST_CFA_OFFSET(16) // (4 bytes). We lack the scratch registers to calculate the SP // right now, so we will just fill it in later. pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current(). CFI_ADJUST_CFA_OFFSET(4) PUSH ecx // Pass receiver. PUSH eax // Pass Method*. leal 32(%esp), %eax // Put original SP into eax movl %eax, 12(%esp) // set SP call SYMBOL(artInstrumentationMethodEntryFromCode) // (Method*, Object*, Thread*, SP) addl LITERAL(28), %esp // Pop arguments upto saved Method*. CFI_ADJUST_CFA_OFFSET(-28) testl %eax, %eax jz 1f // Test for null return (indicating exception) and handle it. movl 60(%esp), %edi // Restore edi. movl %eax, 60(%esp) // Place code* over edi, just under return pc. movl SYMBOL(art_quick_instrumentation_exit)@GOT(%ebx), %ebx // Place instrumentation exit as return pc. ebx holds the GOT computed on entry. movl %ebx, 64(%esp) movl 0(%esp), %eax // Restore eax. // Restore FPRs (extra 4 bytes of offset due to EAX push at top). movsd 8(%esp), %xmm0 movsd 16(%esp), %xmm1 movsd 24(%esp), %xmm2 movsd 32(%esp), %xmm3 // Restore GPRs. movl 40(%esp), %ecx // Restore ecx. movl 44(%esp), %edx // Restore edx. movl 48(%esp), %ebx // Restore ebx. movl 52(%esp), %ebp // Restore ebp. movl 56(%esp), %esi // Restore esi. addl LITERAL(60), %esp // Wind stack back upto code*. CFI_ADJUST_CFA_OFFSET(-60) ret // Call method (and pop). 1: // Make caller handle exception addl LITERAL(4), %esp CFI_ADJUST_CFA_OFFSET(-4) RESTORE_SAVE_REFS_AND_ARGS_FRAME DELIVER_PENDING_EXCEPTION END_FUNCTION art_quick_instrumentation_entry DEFINE_FUNCTION_CUSTOM_CFA art_quick_instrumentation_exit, 0 pushl LITERAL(0) // Push a fake return PC as there will be none on the stack. CFI_ADJUST_CFA_OFFSET(4) SETUP_SAVE_EVERYTHING_FRAME ebx, ebx movl %esp, %ecx // Remember SP subl LITERAL(8), %esp // Align stack. CFI_ADJUST_CFA_OFFSET(8) PUSH edx // Save gpr return value. edx and eax need to be together, // which isn't the case in kSaveEverything frame. PUSH eax leal 32(%esp), %eax // Get pointer to fpr_result, in kSaveEverything frame movl %esp, %edx // Get pointer to gpr_result PUSH eax // Pass fpr_result PUSH edx // Pass gpr_result PUSH ecx // Pass SP pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current. CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artInstrumentationMethodExitFromCode) // (Thread*, SP, gpr_result*, fpr_result*) // Return result could have been changed if it's a reference. movl 16(%esp), %ecx movl %ecx, (80+32)(%esp) addl LITERAL(32), %esp // Pop arguments and grp_result. CFI_ADJUST_CFA_OFFSET(-32) testl %eax, %eax // Check if we returned error. jz .Ldo_deliver_instrumentation_exception testl %edx, %edx jnz .Ldeoptimize // Normal return. movl %eax, FRAME_SIZE_SAVE_EVERYTHING-4(%esp) // Set return pc. RESTORE_SAVE_EVERYTHING_FRAME ret .Ldeoptimize: mov %edx, (FRAME_SIZE_SAVE_EVERYTHING-4)(%esp) // Set return pc. RESTORE_SAVE_EVERYTHING_FRAME jmp SYMBOL(art_quick_deoptimize) .Ldo_deliver_instrumentation_exception: DELIVER_PENDING_EXCEPTION_FRAME_READY END_FUNCTION art_quick_instrumentation_exit /* * Instrumentation has requested that we deoptimize into the interpreter. The deoptimization * will long jump to the upcall with a special exception of -1. */ DEFINE_FUNCTION art_quick_deoptimize SETUP_SAVE_EVERYTHING_FRAME ebx, ebx subl LITERAL(12), %esp // Align stack. CFI_ADJUST_CFA_OFFSET(12) pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current(). CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artDeoptimize) // (Thread*) UNREACHABLE END_FUNCTION art_quick_deoptimize /* * Compiled code has requested that we deoptimize into the interpreter. The deoptimization * will long jump to the interpreter bridge. */ DEFINE_FUNCTION art_quick_deoptimize_from_compiled_code SETUP_SAVE_EVERYTHING_FRAME ebx, ebx subl LITERAL(8), %esp // Align stack. CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current(). CFI_ADJUST_CFA_OFFSET(4) PUSH eax call SYMBOL(artDeoptimizeFromCompiledCode) // (DeoptimizationKind, Thread*) UNREACHABLE END_FUNCTION art_quick_deoptimize_from_compiled_code /* * String's compareTo. * * On entry: * eax: this string object (known non-null) * ecx: comp string object (known non-null) */ DEFINE_FUNCTION art_quick_string_compareto PUSH esi // push callee save reg PUSH edi // push callee save reg mov MIRROR_STRING_COUNT_OFFSET(%eax), %edx mov MIRROR_STRING_COUNT_OFFSET(%ecx), %ebx lea MIRROR_STRING_VALUE_OFFSET(%eax), %esi lea MIRROR_STRING_VALUE_OFFSET(%ecx), %edi #if (STRING_COMPRESSION_FEATURE) /* Differ cases */ shrl LITERAL(1), %edx jnc .Lstring_compareto_this_is_compressed shrl LITERAL(1), %ebx jnc .Lstring_compareto_that_is_compressed jmp .Lstring_compareto_both_not_compressed .Lstring_compareto_this_is_compressed: shrl LITERAL(1), %ebx jnc .Lstring_compareto_both_compressed /* If (this->IsCompressed() && that->IsCompressed() == false) */ mov %edx, %eax subl %ebx, %eax mov %edx, %ecx cmovg %ebx, %ecx /* Going into loop to compare each character */ jecxz .Lstring_compareto_keep_length // check loop counter (if 0, don't compare) .Lstring_compareto_loop_comparison_this_compressed: movzbl (%esi), %edx // move *(this_cur_char) byte to long movzwl (%edi), %ebx // move *(that_cur_char) word to long addl LITERAL(1), %esi // ++this_cur_char (8-bit) addl LITERAL(2), %edi // ++that_cur_char (16-bit) subl %ebx, %edx loope .Lstring_compareto_loop_comparison_this_compressed cmovne %edx, %eax // return eax = *(this_cur_char) - *(that_cur_char) jmp .Lstring_compareto_return .Lstring_compareto_that_is_compressed: mov %edx, %eax subl %ebx, %eax mov %edx, %ecx cmovg %ebx, %ecx /* If (this->IsCompressed() == false && that->IsCompressed()) */ jecxz .Lstring_compareto_keep_length // check loop counter, if 0, don't compare .Lstring_compareto_loop_comparison_that_compressed: movzwl (%esi), %edx // move *(this_cur_char) word to long movzbl (%edi), %ebx // move *(that_cur_char) byte to long addl LITERAL(2), %esi // ++this_cur_char (16-bit) addl LITERAL(1), %edi // ++that_cur_char (8-bit) subl %ebx, %edx loope .Lstring_compareto_loop_comparison_that_compressed cmovne %edx, %eax jmp .Lstring_compareto_return // return eax = *(this_cur_char) - *(that_cur_char) .Lstring_compareto_both_compressed: /* Calculate min length and count diff */ mov %edx, %ecx mov %edx, %eax subl %ebx, %eax cmovg %ebx, %ecx jecxz .Lstring_compareto_keep_length repe cmpsb je .Lstring_compareto_keep_length movzbl -1(%esi), %eax // get last compared char from this string (8-bit) movzbl -1(%edi), %ecx // get last compared char from comp string (8-bit) jmp .Lstring_compareto_count_difference #endif // STRING_COMPRESSION_FEATURE .Lstring_compareto_both_not_compressed: /* Calculate min length and count diff */ mov %edx, %ecx mov %edx, %eax subl %ebx, %eax cmovg %ebx, %ecx /* * At this point we have: * eax: value to return if first part of strings are equal * ecx: minimum among the lengths of the two strings * esi: pointer to this string data * edi: pointer to comp string data */ jecxz .Lstring_compareto_keep_length repe cmpsw // find nonmatching chars in [%esi] and [%edi], up to length %ecx je .Lstring_compareto_keep_length movzwl -2(%esi), %eax // get last compared char from this string (16-bit) movzwl -2(%edi), %ecx // get last compared char from comp string (16-bit) .Lstring_compareto_count_difference: subl %ecx, %eax .Lstring_compareto_keep_length: .Lstring_compareto_return: POP edi // pop callee save reg POP esi // pop callee save reg ret END_FUNCTION art_quick_string_compareto DEFINE_FUNCTION art_quick_string_builder_append SETUP_SAVE_REFS_ONLY_FRAME ebx, ebx // save ref containing registers for GC // Outgoing argument set up leal FRAME_SIZE_SAVE_REFS_ONLY + __SIZEOF_POINTER__(%esp), %edi // prepare args push %eax // push padding CFI_ADJUST_CFA_OFFSET(4) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) push %edi // pass args CFI_ADJUST_CFA_OFFSET(4) push %eax // pass format CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artStringBuilderAppend) // (uint32_t, const unit32_t*, Thread*) addl MACRO_LITERAL(16), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception END_FUNCTION art_quick_string_builder_append // Create a function `name` calling the ReadBarrier::Mark routine, // getting its argument and returning its result through register // `reg`, saving and restoring all caller-save registers. // // If `reg` is different from `eax`, the generated function follows a // non-standard runtime calling convention: // - register `reg` is used to pass the (sole) argument of this function // (instead of EAX); // - register `reg` is used to return the result of this function // (instead of EAX); // - EAX is treated like a normal (non-argument) caller-save register; // - everything else is the same as in the standard runtime calling // convention (e.g. standard callee-save registers are preserved). MACRO2(READ_BARRIER_MARK_REG, name, reg) DEFINE_FUNCTION VAR(name) // Null check so that we can load the lock word. test REG_VAR(reg), REG_VAR(reg) jz .Lret_rb_\name .Lnot_null_\name: // Check the mark bit, if it is 1 return. testl LITERAL(LOCK_WORD_MARK_BIT_MASK_SHIFTED), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(reg)) jz .Lslow_rb_\name ret .Lslow_rb_\name: PUSH eax mov MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(reg)), %eax add LITERAL(LOCK_WORD_STATE_FORWARDING_ADDRESS_OVERFLOW), %eax // Jump if overflow, the only case where it overflows should be the forwarding address one. // Taken ~25% of the time. jnae .Lret_forwarding_address\name // Save all potentially live caller-save core registers. mov 0(%esp), %eax PUSH ecx PUSH edx PUSH ebx // 8-byte align the stack to improve (8-byte) XMM register saving and restoring. // and create space for caller-save floating-point registers. subl MACRO_LITERAL(4 + 8 * 8), %esp CFI_ADJUST_CFA_OFFSET(4 + 8 * 8) // Save all potentially live caller-save floating-point registers. movsd %xmm0, 0(%esp) movsd %xmm1, 8(%esp) movsd %xmm2, 16(%esp) movsd %xmm3, 24(%esp) movsd %xmm4, 32(%esp) movsd %xmm5, 40(%esp) movsd %xmm6, 48(%esp) movsd %xmm7, 56(%esp) subl LITERAL(4), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(4) PUSH RAW_VAR(reg) // pass arg1 - obj from `reg` call SYMBOL(artReadBarrierMark) // artReadBarrierMark(obj) .ifnc RAW_VAR(reg), eax movl %eax, REG_VAR(reg) // return result into `reg` .endif addl LITERAL(8), %esp // pop argument and remove padding CFI_ADJUST_CFA_OFFSET(-8) // Restore floating-point registers. movsd 0(%esp), %xmm0 movsd 8(%esp), %xmm1 movsd 16(%esp), %xmm2 movsd 24(%esp), %xmm3 movsd 32(%esp), %xmm4 movsd 40(%esp), %xmm5 movsd 48(%esp), %xmm6 movsd 56(%esp), %xmm7 // Remove floating-point registers and padding. addl MACRO_LITERAL(8 * 8 + 4), %esp CFI_ADJUST_CFA_OFFSET(-(8 * 8 + 4)) // Restore core regs, except `reg`, as it is used to return the // result of this function (simply remove it from the stack instead). POP_REG_NE ebx, RAW_VAR(reg) POP_REG_NE edx, RAW_VAR(reg) POP_REG_NE ecx, RAW_VAR(reg) POP_REG_NE eax, RAW_VAR(reg) .Lret_rb_\name: ret .Lret_forwarding_address\name: // The overflow cleared the top bits. sall LITERAL(LOCK_WORD_STATE_FORWARDING_ADDRESS_SHIFT), %eax mov %eax, REG_VAR(reg) POP_REG_NE eax, RAW_VAR(reg) ret END_FUNCTION VAR(name) END_MACRO READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg00, eax READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg01, ecx READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg02, edx READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg03, ebx READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg05, ebp // Note: There is no art_quick_read_barrier_mark_reg04, as register 4 (ESP) // cannot be used to pass arguments. READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg06, esi READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg07, edi DEFINE_FUNCTION art_quick_read_barrier_slow PUSH edx // pass arg3 - offset PUSH ecx // pass arg2 - obj PUSH eax // pass arg1 - ref call SYMBOL(artReadBarrierSlow) // artReadBarrierSlow(ref, obj, offset) addl LITERAL(12), %esp // pop arguments CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_read_barrier_slow DEFINE_FUNCTION art_quick_read_barrier_for_root_slow subl LITERAL(8), %esp // alignment padding CFI_ADJUST_CFA_OFFSET(8) PUSH eax // pass arg1 - root call SYMBOL(artReadBarrierForRootSlow) // artReadBarrierForRootSlow(root) addl LITERAL(12), %esp // pop argument and remove padding CFI_ADJUST_CFA_OFFSET(-12) ret END_FUNCTION art_quick_read_barrier_for_root_slow /* * On stack replacement stub. * On entry: * [sp] = return address * [sp + 4] = stack to copy * [sp + 8] = size of stack * [sp + 12] = pc to call * [sp + 16] = JValue* result * [sp + 20] = shorty * [sp + 24] = thread */ DEFINE_FUNCTION art_quick_osr_stub // Save native callee saves. PUSH ebp PUSH ebx PUSH esi PUSH edi SAVE_SIZE=20 // 4 registers and the return address mov 4+16(%esp), %esi // ESI = argument array mov 8+16(%esp), %ecx // ECX = size of args mov 12+16(%esp), %ebx // EBX = pc to call mov %esp, %ebp // Save stack pointer CFI_DEF_CFA(ebp, SAVE_SIZE) // CFA = ebp + SAVE_SIZE CFI_REMEMBER_STATE andl LITERAL(0xFFFFFFF0), %esp // Align stack pushl %ebp // Save old stack pointer subl LITERAL(12), %esp // Align stack movl LITERAL(0), (%esp) // Store null for ArtMethod* slot // ebp isn't properly spilled in the osr method, so we need use DWARF expression. // NB: the CFI must be before the call since this is the address gdb will lookup. // NB: gdb expects that cfa_expression returns the CFA value (not address to it). CFI_ESCAPE( /* cfa = [sp + 12] + SAVE_SIZE */ \ 0x0f, 6, /* DW_CFA_def_cfa_expression(len) */ \ 0x92, 4, 12, /* DW_OP_bregx(reg,offset) */ \ 0x06, /* DW_OP_deref */ \ 0x23, SAVE_SIZE) /* DW_OP_plus_uconst(val) */ call .Losr_entry mov 12(%esp), %esp // Restore stack pointer. CFI_DEF_CFA(esp, SAVE_SIZE) // CFA = esp + SAVE_SIZE // Restore callee saves. POP edi POP esi POP ebx POP ebp mov 16(%esp), %ecx // Get JValue result mov %eax, (%ecx) // Store the result. mov %edx, 4(%ecx) // Store the other half of the result. ret .Losr_entry: CFI_RESTORE_STATE_AND_DEF_CFA(ebp, SAVE_SIZE) // CFA = ebp + SAVE_SIZE subl LITERAL(4), %ecx // Given stack size contains pushed frame pointer, substract it. subl %ecx, %esp mov %esp, %edi // EDI = beginning of stack rep movsb // while (ecx--) { *edi++ = *esi++ } jmp *%ebx END_FUNCTION art_quick_osr_stub DEFINE_FUNCTION art_quick_invoke_polymorphic // On entry: EAX := unused, ECX := receiver SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, ebx // Save frame. mov %esp, %edx // Remember SP sub LITERAL(4), %esp // Alignment padding CFI_ADJUST_CFA_OFFSET(4) push %edx // Push SP CFI_ADJUST_CFA_OFFSET(4) pushl %fs:THREAD_SELF_OFFSET // Push Thread::Current() CFI_ADJUST_CFA_OFFSET(4) push %ecx // Push receiver (method handle) CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artInvokePolymorphic) // invoke with (receiver, thread, SP) addl LITERAL(16), %esp // Pop arguments. CFI_ADJUST_CFA_OFFSET(-16) mov %eax, 4(%esp) // Result is in EAX:EDX. Copy to saved FP state. mov %edx, 8(%esp) mov %edx, 40(%esp) // Copy EDX to saved context RESTORE_SAVE_REFS_AND_ARGS_FRAME RETURN_OR_DELIVER_PENDING_EXCEPTION END_FUNCTION art_quick_invoke_polymorphic DEFINE_FUNCTION art_quick_invoke_custom SETUP_SAVE_REFS_AND_ARGS_FRAME ebx, ebx // Save frame. // EAX := call_site_index mov %esp, %ecx // Remember SP. subl LITERAL(4), %esp // Alignment padding. CFI_ADJUST_CFA_OFFSET(4) push %ecx // pass SP CFI_ADJUST_CFA_OFFSET(4) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) push %eax // pass call_site_index CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artInvokeCustom) // artInvokeCustom(call_site_index, Thread*, SP) addl LITERAL(16), %esp // Pop arguments. CFI_ADJUST_CFA_OFFSET(-16) mov %eax, 4(%esp) // Result is in EAX:EDX. Copy to saved FP state. mov %edx, 8(%esp) mov %edx, 40(%esp) // Copy EDX to saved context RESTORE_SAVE_REFS_AND_ARGS_FRAME RETURN_OR_DELIVER_PENDING_EXCEPTION END_FUNCTION art_quick_invoke_custom // Wrap ExecuteSwitchImpl in assembly method which specifies DEX PC for unwinding. // Argument 0: ESP+4: The context pointer for ExecuteSwitchImpl. // Argument 1: ESP+8: Pointer to the templated ExecuteSwitchImpl to call. // Argument 2: ESP+12: The value of DEX PC (memory address of the methods bytecode). DEFINE_FUNCTION ExecuteSwitchImplAsm PUSH ebx // Spill EBX; Increments ESP, so arg0 is at ESP+8 now. mov 12(%esp), %eax // EAX = C++ templated interpreter function mov 16(%esp), %ebx // EBX = DEX PC (callee save register) mov 8(%esp), %ecx // ECX = Context argument for the function CFI_DEFINE_DEX_PC_WITH_OFFSET(0 /* EAX */, 3 /* EBX */, 0) sub LITERAL(4), %esp // Alignment padding CFI_ADJUST_CFA_OFFSET(4) push %ecx // Push argument CFI_ADJUST_CFA_OFFSET(4) call *%eax // Call the wrapped function addl LITERAL(8), %esp CFI_ADJUST_CFA_OFFSET(-8) POP ebx // Restore EBX ret END_FUNCTION ExecuteSwitchImplAsm // On entry: eax is the class, ebp is the inline cache. DEFINE_FUNCTION art_quick_update_inline_cache #if (INLINE_CACHE_SIZE != 5) #error "INLINE_CACHE_SIZE not as expected." #endif // Don't update the cache if we are marking. cmpl LITERAL(0), %fs:THREAD_IS_GC_MARKING_OFFSET jnz .Lret PUSH ecx movl %eax, %ecx // eax will be used for cmpxchg .Lentry1: movl INLINE_CACHE_CLASSES_OFFSET(%ebp), %eax cmpl %ecx, %eax je .Ldone cmpl LITERAL(0), %eax jne .Lentry2 lock cmpxchg %ecx, INLINE_CACHE_CLASSES_OFFSET(%ebp) jz .Ldone jmp .Lentry1 .Lentry2: movl (INLINE_CACHE_CLASSES_OFFSET+4)(%ebp), %eax cmpl %ecx, %eax je .Ldone cmpl LITERAL(0), %eax jne .Lentry3 lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+4)(%ebp) jz .Ldone jmp .Lentry2 .Lentry3: movl (INLINE_CACHE_CLASSES_OFFSET+8)(%ebp), %eax cmpl %ecx, %eax je .Ldone cmpl LITERAL(0), %eax jne .Lentry4 lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+8)(%ebp) jz .Ldone jmp .Lentry3 .Lentry4: movl (INLINE_CACHE_CLASSES_OFFSET+12)(%ebp), %eax cmpl %ecx, %eax je .Ldone cmpl LITERAL(0), %eax jne .Lentry5 lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+12)(%ebp) jz .Ldone jmp .Lentry4 .Lentry5: // Unconditionally store, the cache is megamorphic. movl %ecx, (INLINE_CACHE_CLASSES_OFFSET+16)(%ebp) .Ldone: // Restore registers movl %ecx, %eax POP ecx .Lret: ret END_FUNCTION art_quick_update_inline_cache // TODO: implement these! UNIMPLEMENTED art_quick_memcmp16 // On entry, the method is at the bottom of the stack. DEFINE_FUNCTION art_quick_compile_optimized SETUP_SAVE_EVERYTHING_FRAME ebx, ebx mov FRAME_SIZE_SAVE_EVERYTHING(%esp), %eax // Fetch ArtMethod sub LITERAL(8), %esp // Alignment padding CFI_ADJUST_CFA_OFFSET(8) pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current() CFI_ADJUST_CFA_OFFSET(4) pushl %eax CFI_ADJUST_CFA_OFFSET(4) call SYMBOL(artCompileOptimized) // (ArtMethod*, Thread*) addl LITERAL(16), %esp // Pop arguments. CFI_ADJUST_CFA_OFFSET(-16) RESTORE_SAVE_EVERYTHING_FRAME ret END_FUNCTION art_quick_compile_optimized