/* * Copyright (C) 2011 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include "compiler.h" #include "compiler_internals.h" #include "driver/compiler_driver.h" #include "driver/compiler_options.h" #include "dataflow_iterator-inl.h" #include "leb128.h" #include "mirror/object.h" #include "pass_driver_me_opts.h" #include "runtime.h" #include "base/logging.h" #include "base/timing_logger.h" #include "driver/compiler_options.h" #include "dex/quick/dex_file_to_method_inliner_map.h" namespace art { extern "C" void ArtInitQuickCompilerContext(art::CompilerDriver* driver) { CHECK(driver->GetCompilerContext() == nullptr); } extern "C" void ArtUnInitQuickCompilerContext(art::CompilerDriver* driver) { CHECK(driver->GetCompilerContext() == nullptr); } /* Default optimizer/debug setting for the compiler. */ static uint32_t kCompilerOptimizerDisableFlags = 0 | // Disable specific optimizations (1 << kLoadStoreElimination) | // (1 << kLoadHoisting) | // (1 << kSuppressLoads) | // (1 << kNullCheckElimination) | // (1 << kClassInitCheckElimination) | (1 << kGlobalValueNumbering) | // (1 << kPromoteRegs) | // (1 << kTrackLiveTemps) | // (1 << kSafeOptimizations) | // (1 << kBBOpt) | // (1 << kMatch) | // (1 << kPromoteCompilerTemps) | // (1 << kSuppressExceptionEdges) | // (1 << kSuppressMethodInlining) | 0; static uint32_t kCompilerDebugFlags = 0 | // Enable debug/testing modes // (1 << kDebugDisplayMissingTargets) | // (1 << kDebugVerbose) | // (1 << kDebugDumpCFG) | // (1 << kDebugSlowFieldPath) | // (1 << kDebugSlowInvokePath) | // (1 << kDebugSlowStringPath) | // (1 << kDebugSlowestFieldPath) | // (1 << kDebugSlowestStringPath) | // (1 << kDebugExerciseResolveMethod) | // (1 << kDebugVerifyDataflow) | // (1 << kDebugShowMemoryUsage) | // (1 << kDebugShowNops) | // (1 << kDebugCountOpcodes) | // (1 << kDebugDumpCheckStats) | // (1 << kDebugDumpBitcodeFile) | // (1 << kDebugVerifyBitcode) | // (1 << kDebugShowSummaryMemoryUsage) | // (1 << kDebugShowFilterStats) | // (1 << kDebugTimings) | // (1 << kDebugCodegenDump) | 0; COMPILE_ASSERT(0U == static_cast(kNone), kNone_not_0); COMPILE_ASSERT(1U == static_cast(kArm), kArm_not_1); COMPILE_ASSERT(2U == static_cast(kArm64), kArm64_not_2); COMPILE_ASSERT(3U == static_cast(kThumb2), kThumb2_not_3); COMPILE_ASSERT(4U == static_cast(kX86), kX86_not_4); COMPILE_ASSERT(5U == static_cast(kX86_64), kX86_64_not_5); COMPILE_ASSERT(6U == static_cast(kMips), kMips_not_6); COMPILE_ASSERT(7U == static_cast(kMips64), kMips64_not_7); // Additional disabled optimizations (over generally disabled) per instruction set. static constexpr uint32_t kDisabledOptimizationsPerISA[] = { // 0 = kNone. ~0U, // 1 = kArm, unused (will use kThumb2). ~0U, // 2 = kArm64. 0, // 3 = kThumb2. 0, // 4 = kX86. (1 << kLoadStoreElimination) | 0, // 5 = kX86_64. (1 << kLoadStoreElimination) | 0, // 6 = kMips. (1 << kLoadStoreElimination) | (1 << kLoadHoisting) | (1 << kSuppressLoads) | (1 << kNullCheckElimination) | (1 << kPromoteRegs) | (1 << kTrackLiveTemps) | (1 << kSafeOptimizations) | (1 << kBBOpt) | (1 << kMatch) | (1 << kPromoteCompilerTemps) | 0, // 7 = kMips64. ~0U }; COMPILE_ASSERT(sizeof(kDisabledOptimizationsPerISA) == 8 * sizeof(uint32_t), kDisabledOpts_unexp); // Supported shorty types per instruction set. nullptr means that all are available. // Z : boolean // B : byte // S : short // C : char // I : int // J : long // F : float // D : double // L : reference(object, array) // V : void static const char* kSupportedTypes[] = { // 0 = kNone. "", // 1 = kArm, unused (will use kThumb2). "", // 2 = kArm64. nullptr, // 3 = kThumb2. nullptr, // 4 = kX86. nullptr, // 5 = kX86_64. nullptr, // 6 = kMips. nullptr, // 7 = kMips64. "" }; COMPILE_ASSERT(sizeof(kSupportedTypes) == 8 * sizeof(char*), kSupportedTypes_unexp); static int kAllOpcodes[] = { Instruction::NOP, Instruction::MOVE, Instruction::MOVE_FROM16, Instruction::MOVE_16, Instruction::MOVE_WIDE, Instruction::MOVE_WIDE_FROM16, Instruction::MOVE_WIDE_16, Instruction::MOVE_OBJECT, Instruction::MOVE_OBJECT_FROM16, Instruction::MOVE_OBJECT_16, Instruction::MOVE_RESULT, Instruction::MOVE_RESULT_WIDE, Instruction::MOVE_RESULT_OBJECT, Instruction::MOVE_EXCEPTION, Instruction::RETURN_VOID, Instruction::RETURN, Instruction::RETURN_WIDE, Instruction::RETURN_OBJECT, Instruction::CONST_4, Instruction::CONST_16, Instruction::CONST, Instruction::CONST_HIGH16, Instruction::CONST_WIDE_16, Instruction::CONST_WIDE_32, Instruction::CONST_WIDE, Instruction::CONST_WIDE_HIGH16, Instruction::CONST_STRING, Instruction::CONST_STRING_JUMBO, Instruction::CONST_CLASS, Instruction::MONITOR_ENTER, Instruction::MONITOR_EXIT, Instruction::CHECK_CAST, Instruction::INSTANCE_OF, Instruction::ARRAY_LENGTH, Instruction::NEW_INSTANCE, Instruction::NEW_ARRAY, Instruction::FILLED_NEW_ARRAY, Instruction::FILLED_NEW_ARRAY_RANGE, Instruction::FILL_ARRAY_DATA, Instruction::THROW, Instruction::GOTO, Instruction::GOTO_16, Instruction::GOTO_32, Instruction::PACKED_SWITCH, Instruction::SPARSE_SWITCH, Instruction::CMPL_FLOAT, Instruction::CMPG_FLOAT, Instruction::CMPL_DOUBLE, Instruction::CMPG_DOUBLE, Instruction::CMP_LONG, Instruction::IF_EQ, Instruction::IF_NE, Instruction::IF_LT, Instruction::IF_GE, Instruction::IF_GT, Instruction::IF_LE, Instruction::IF_EQZ, Instruction::IF_NEZ, Instruction::IF_LTZ, Instruction::IF_GEZ, Instruction::IF_GTZ, Instruction::IF_LEZ, Instruction::UNUSED_3E, Instruction::UNUSED_3F, Instruction::UNUSED_40, Instruction::UNUSED_41, Instruction::UNUSED_42, Instruction::UNUSED_43, Instruction::AGET, Instruction::AGET_WIDE, Instruction::AGET_OBJECT, Instruction::AGET_BOOLEAN, Instruction::AGET_BYTE, Instruction::AGET_CHAR, Instruction::AGET_SHORT, Instruction::APUT, Instruction::APUT_WIDE, Instruction::APUT_OBJECT, Instruction::APUT_BOOLEAN, Instruction::APUT_BYTE, Instruction::APUT_CHAR, Instruction::APUT_SHORT, Instruction::IGET, Instruction::IGET_WIDE, Instruction::IGET_OBJECT, Instruction::IGET_BOOLEAN, Instruction::IGET_BYTE, Instruction::IGET_CHAR, Instruction::IGET_SHORT, Instruction::IPUT, Instruction::IPUT_WIDE, Instruction::IPUT_OBJECT, Instruction::IPUT_BOOLEAN, Instruction::IPUT_BYTE, Instruction::IPUT_CHAR, Instruction::IPUT_SHORT, Instruction::SGET, Instruction::SGET_WIDE, Instruction::SGET_OBJECT, Instruction::SGET_BOOLEAN, Instruction::SGET_BYTE, Instruction::SGET_CHAR, Instruction::SGET_SHORT, Instruction::SPUT, Instruction::SPUT_WIDE, Instruction::SPUT_OBJECT, Instruction::SPUT_BOOLEAN, Instruction::SPUT_BYTE, Instruction::SPUT_CHAR, Instruction::SPUT_SHORT, Instruction::INVOKE_VIRTUAL, Instruction::INVOKE_SUPER, Instruction::INVOKE_DIRECT, Instruction::INVOKE_STATIC, Instruction::INVOKE_INTERFACE, Instruction::RETURN_VOID_BARRIER, Instruction::INVOKE_VIRTUAL_RANGE, Instruction::INVOKE_SUPER_RANGE, Instruction::INVOKE_DIRECT_RANGE, Instruction::INVOKE_STATIC_RANGE, Instruction::INVOKE_INTERFACE_RANGE, Instruction::UNUSED_79, Instruction::UNUSED_7A, Instruction::NEG_INT, Instruction::NOT_INT, Instruction::NEG_LONG, Instruction::NOT_LONG, Instruction::NEG_FLOAT, Instruction::NEG_DOUBLE, Instruction::INT_TO_LONG, Instruction::INT_TO_FLOAT, Instruction::INT_TO_DOUBLE, Instruction::LONG_TO_INT, Instruction::LONG_TO_FLOAT, Instruction::LONG_TO_DOUBLE, Instruction::FLOAT_TO_INT, Instruction::FLOAT_TO_LONG, Instruction::FLOAT_TO_DOUBLE, Instruction::DOUBLE_TO_INT, Instruction::DOUBLE_TO_LONG, Instruction::DOUBLE_TO_FLOAT, Instruction::INT_TO_BYTE, Instruction::INT_TO_CHAR, Instruction::INT_TO_SHORT, Instruction::ADD_INT, Instruction::SUB_INT, Instruction::MUL_INT, Instruction::DIV_INT, Instruction::REM_INT, Instruction::AND_INT, Instruction::OR_INT, Instruction::XOR_INT, Instruction::SHL_INT, Instruction::SHR_INT, Instruction::USHR_INT, Instruction::ADD_LONG, Instruction::SUB_LONG, Instruction::MUL_LONG, Instruction::DIV_LONG, Instruction::REM_LONG, Instruction::AND_LONG, Instruction::OR_LONG, Instruction::XOR_LONG, Instruction::SHL_LONG, Instruction::SHR_LONG, Instruction::USHR_LONG, Instruction::ADD_FLOAT, Instruction::SUB_FLOAT, Instruction::MUL_FLOAT, Instruction::DIV_FLOAT, Instruction::REM_FLOAT, Instruction::ADD_DOUBLE, Instruction::SUB_DOUBLE, Instruction::MUL_DOUBLE, Instruction::DIV_DOUBLE, Instruction::REM_DOUBLE, Instruction::ADD_INT_2ADDR, Instruction::SUB_INT_2ADDR, Instruction::MUL_INT_2ADDR, Instruction::DIV_INT_2ADDR, Instruction::REM_INT_2ADDR, Instruction::AND_INT_2ADDR, Instruction::OR_INT_2ADDR, Instruction::XOR_INT_2ADDR, Instruction::SHL_INT_2ADDR, Instruction::SHR_INT_2ADDR, Instruction::USHR_INT_2ADDR, Instruction::ADD_LONG_2ADDR, Instruction::SUB_LONG_2ADDR, Instruction::MUL_LONG_2ADDR, Instruction::DIV_LONG_2ADDR, Instruction::REM_LONG_2ADDR, Instruction::AND_LONG_2ADDR, Instruction::OR_LONG_2ADDR, Instruction::XOR_LONG_2ADDR, Instruction::SHL_LONG_2ADDR, Instruction::SHR_LONG_2ADDR, Instruction::USHR_LONG_2ADDR, Instruction::ADD_FLOAT_2ADDR, Instruction::SUB_FLOAT_2ADDR, Instruction::MUL_FLOAT_2ADDR, Instruction::DIV_FLOAT_2ADDR, Instruction::REM_FLOAT_2ADDR, Instruction::ADD_DOUBLE_2ADDR, Instruction::SUB_DOUBLE_2ADDR, Instruction::MUL_DOUBLE_2ADDR, Instruction::DIV_DOUBLE_2ADDR, Instruction::REM_DOUBLE_2ADDR, Instruction::ADD_INT_LIT16, Instruction::RSUB_INT, Instruction::MUL_INT_LIT16, Instruction::DIV_INT_LIT16, Instruction::REM_INT_LIT16, Instruction::AND_INT_LIT16, Instruction::OR_INT_LIT16, Instruction::XOR_INT_LIT16, Instruction::ADD_INT_LIT8, Instruction::RSUB_INT_LIT8, Instruction::MUL_INT_LIT8, Instruction::DIV_INT_LIT8, Instruction::REM_INT_LIT8, Instruction::AND_INT_LIT8, Instruction::OR_INT_LIT8, Instruction::XOR_INT_LIT8, Instruction::SHL_INT_LIT8, Instruction::SHR_INT_LIT8, Instruction::USHR_INT_LIT8, Instruction::IGET_QUICK, Instruction::IGET_WIDE_QUICK, Instruction::IGET_OBJECT_QUICK, Instruction::IPUT_QUICK, Instruction::IPUT_WIDE_QUICK, Instruction::IPUT_OBJECT_QUICK, Instruction::INVOKE_VIRTUAL_QUICK, Instruction::INVOKE_VIRTUAL_RANGE_QUICK, Instruction::UNUSED_EB, Instruction::UNUSED_EC, Instruction::UNUSED_ED, Instruction::UNUSED_EE, Instruction::UNUSED_EF, Instruction::UNUSED_F0, Instruction::UNUSED_F1, Instruction::UNUSED_F2, Instruction::UNUSED_F3, Instruction::UNUSED_F4, Instruction::UNUSED_F5, Instruction::UNUSED_F6, Instruction::UNUSED_F7, Instruction::UNUSED_F8, Instruction::UNUSED_F9, Instruction::UNUSED_FA, Instruction::UNUSED_FB, Instruction::UNUSED_FC, Instruction::UNUSED_FD, Instruction::UNUSED_FE, Instruction::UNUSED_FF, // ----- ExtendedMIROpcode ----- kMirOpPhi, kMirOpCopy, kMirOpFusedCmplFloat, kMirOpFusedCmpgFloat, kMirOpFusedCmplDouble, kMirOpFusedCmpgDouble, kMirOpFusedCmpLong, kMirOpNop, kMirOpNullCheck, kMirOpRangeCheck, kMirOpDivZeroCheck, kMirOpCheck, kMirOpCheckPart2, kMirOpSelect, }; // Unsupported opcodes. nullptr can be used when everything is supported. Size of the lists is // recorded below. static const int* kUnsupportedOpcodes[] = { // 0 = kNone. kAllOpcodes, // 1 = kArm, unused (will use kThumb2). kAllOpcodes, // 2 = kArm64. nullptr, // 3 = kThumb2. nullptr, // 4 = kX86. nullptr, // 5 = kX86_64. nullptr, // 6 = kMips. nullptr, // 7 = kMips64. kAllOpcodes }; COMPILE_ASSERT(sizeof(kUnsupportedOpcodes) == 8 * sizeof(int*), kUnsupportedOpcodes_unexp); // Size of the arrays stored above. static const size_t kUnsupportedOpcodesSize[] = { // 0 = kNone. arraysize(kAllOpcodes), // 1 = kArm, unused (will use kThumb2). arraysize(kAllOpcodes), // 2 = kArm64. 0, // 3 = kThumb2. 0, // 4 = kX86. 0, // 5 = kX86_64. 0, // 6 = kMips. 0, // 7 = kMips64. arraysize(kAllOpcodes), }; COMPILE_ASSERT(sizeof(kUnsupportedOpcodesSize) == 8 * sizeof(size_t), kUnsupportedOpcodesSize_unexp); // The maximum amount of Dalvik register in a method for which we will start compiling. Tries to // avoid an abort when we need to manage more SSA registers than we can. static constexpr size_t kMaxAllowedDalvikRegisters = INT16_MAX / 2; CompilationUnit::CompilationUnit(ArenaPool* pool) : compiler_driver(nullptr), class_linker(nullptr), dex_file(nullptr), class_loader(nullptr), class_def_idx(0), method_idx(0), code_item(nullptr), access_flags(0), invoke_type(kDirect), shorty(nullptr), disable_opt(0), enable_debug(0), verbose(false), compiler(nullptr), instruction_set(kNone), target64(false), num_dalvik_registers(0), insns(nullptr), num_ins(0), num_outs(0), num_regs(0), compiler_flip_match(false), arena(pool), arena_stack(pool), mir_graph(nullptr), cg(nullptr), timings("QuickCompiler", true, false), print_pass(false) { } CompilationUnit::~CompilationUnit() { } void CompilationUnit::StartTimingSplit(const char* label) { if (compiler_driver->GetDumpPasses()) { timings.StartTiming(label); } } void CompilationUnit::NewTimingSplit(const char* label) { if (compiler_driver->GetDumpPasses()) { timings.EndTiming(); timings.StartTiming(label); } } void CompilationUnit::EndTiming() { if (compiler_driver->GetDumpPasses()) { timings.EndTiming(); if (enable_debug & (1 << kDebugTimings)) { LOG(INFO) << "TIMINGS " << PrettyMethod(method_idx, *dex_file); LOG(INFO) << Dumpable(timings); } } } static bool CanCompileShorty(const char* shorty, InstructionSet instruction_set) { const char* supported_types = kSupportedTypes[instruction_set]; if (supported_types == nullptr) { // Everything available. return true; } uint32_t shorty_size = strlen(shorty); CHECK_GE(shorty_size, 1u); for (uint32_t i = 0; i < shorty_size; i++) { if (strchr(supported_types, shorty[i]) == nullptr) { return false; } } return true; }; // Skip the method that we do not support currently. static bool CanCompileMethod(uint32_t method_idx, const DexFile& dex_file, CompilationUnit& cu) { // This is a limitation in mir_graph. See MirGraph::SetNumSSARegs. if (cu.num_dalvik_registers > kMaxAllowedDalvikRegisters) { VLOG(compiler) << "Too many dalvik registers : " << cu.num_dalvik_registers; return false; } // Check whether we do have limitations at all. if (kSupportedTypes[cu.instruction_set] == nullptr && kUnsupportedOpcodesSize[cu.instruction_set] == 0U) { return true; } // Check if we can compile the prototype. const char* shorty = dex_file.GetMethodShorty(dex_file.GetMethodId(method_idx)); if (!CanCompileShorty(shorty, cu.instruction_set)) { VLOG(compiler) << "Unsupported shorty : " << shorty; return false; } const int *unsupport_list = kUnsupportedOpcodes[cu.instruction_set]; int unsupport_list_size = kUnsupportedOpcodesSize[cu.instruction_set]; for (unsigned int idx = 0; idx < cu.mir_graph->GetNumBlocks(); idx++) { BasicBlock* bb = cu.mir_graph->GetBasicBlock(idx); if (bb == NULL) continue; if (bb->block_type == kDead) continue; for (MIR* mir = bb->first_mir_insn; mir != nullptr; mir = mir->next) { int opcode = mir->dalvikInsn.opcode; // Check if we support the byte code. if (std::find(unsupport_list, unsupport_list + unsupport_list_size, opcode) != unsupport_list + unsupport_list_size) { if (!MIR::DecodedInstruction::IsPseudoMirOp(opcode)) { VLOG(compiler) << "Unsupported dalvik byte code : " << mir->dalvikInsn.opcode; } else { VLOG(compiler) << "Unsupported extended MIR opcode : " << MIRGraph::extended_mir_op_names_[opcode - kMirOpFirst]; } return false; } // Check if it invokes a prototype that we cannot support. if (Instruction::INVOKE_VIRTUAL == opcode || Instruction::INVOKE_SUPER == opcode || Instruction::INVOKE_DIRECT == opcode || Instruction::INVOKE_STATIC == opcode || Instruction::INVOKE_INTERFACE == opcode) { uint32_t invoke_method_idx = mir->dalvikInsn.vB; const char* invoke_method_shorty = dex_file.GetMethodShorty( dex_file.GetMethodId(invoke_method_idx)); if (!CanCompileShorty(invoke_method_shorty, cu.instruction_set)) { VLOG(compiler) << "Unsupported to invoke '" << PrettyMethod(invoke_method_idx, dex_file) << "' with shorty : " << invoke_method_shorty; return false; } } } } return true; } static CompiledMethod* CompileMethod(CompilerDriver& driver, Compiler* compiler, const DexFile::CodeItem* code_item, uint32_t access_flags, InvokeType invoke_type, uint16_t class_def_idx, uint32_t method_idx, jobject class_loader, const DexFile& dex_file, void* llvm_compilation_unit) { VLOG(compiler) << "Compiling " << PrettyMethod(method_idx, dex_file) << "..."; /* * Skip compilation for pathologically large methods - either by instruction count or num vregs. * Dalvik uses 16-bit uints for instruction and register counts. We'll limit to a quarter * of that, which also guarantees we cannot overflow our 16-bit internal SSA name space. */ if (code_item->insns_size_in_code_units_ >= UINT16_MAX / 4) { LOG(INFO) << "Method exceeds compiler instruction limit: " << code_item->insns_size_in_code_units_ << " in " << PrettyMethod(method_idx, dex_file); return NULL; } if (code_item->registers_size_ >= UINT16_MAX / 4) { LOG(INFO) << "Method exceeds compiler virtual register limit: " << code_item->registers_size_ << " in " << PrettyMethod(method_idx, dex_file); return NULL; } if (!driver.GetCompilerOptions().IsCompilationEnabled()) { return nullptr; } ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); CompilationUnit cu(driver.GetArenaPool()); cu.compiler_driver = &driver; cu.class_linker = class_linker; cu.instruction_set = driver.GetInstructionSet(); if (cu.instruction_set == kArm) { cu.instruction_set = kThumb2; } cu.target64 = Is64BitInstructionSet(cu.instruction_set); cu.compiler = compiler; // TODO: Mips64 is not yet implemented. CHECK((cu.instruction_set == kThumb2) || (cu.instruction_set == kArm64) || (cu.instruction_set == kX86) || (cu.instruction_set == kX86_64) || (cu.instruction_set == kMips)); /* Adjust this value accordingly once inlining is performed */ cu.num_dalvik_registers = code_item->registers_size_; // TODO: set this from command line cu.compiler_flip_match = false; bool use_match = !cu.compiler_method_match.empty(); bool match = use_match && (cu.compiler_flip_match ^ (PrettyMethod(method_idx, dex_file).find(cu.compiler_method_match) != std::string::npos)); if (!use_match || match) { cu.disable_opt = kCompilerOptimizerDisableFlags; cu.enable_debug = kCompilerDebugFlags; cu.verbose = VLOG_IS_ON(compiler) || (cu.enable_debug & (1 << kDebugVerbose)); } if (gVerboseMethods.size() != 0) { cu.verbose = false; for (size_t i = 0; i < gVerboseMethods.size(); ++i) { if (PrettyMethod(method_idx, dex_file).find(gVerboseMethods[i]) != std::string::npos) { cu.verbose = true; break; } } } if (cu.verbose) { cu.enable_debug |= (1 << kDebugCodegenDump); } /* * TODO: rework handling of optimization and debug flags. Should we split out * MIR and backend flags? Need command-line setting as well. */ compiler->InitCompilationUnit(cu); // Disable optimizations according to instruction set. cu.disable_opt |= kDisabledOptimizationsPerISA[cu.instruction_set]; cu.StartTimingSplit("BuildMIRGraph"); cu.mir_graph.reset(new MIRGraph(&cu, &cu.arena)); /* * After creation of the MIR graph, also create the code generator. * The reason we do this is that optimizations on the MIR graph may need to get information * that is only available if a CG exists. */ cu.cg.reset(compiler->GetCodeGenerator(&cu, llvm_compilation_unit)); /* Gathering opcode stats? */ if (kCompilerDebugFlags & (1 << kDebugCountOpcodes)) { cu.mir_graph->EnableOpcodeCounting(); } /* Build the raw MIR graph */ cu.mir_graph->InlineMethod(code_item, access_flags, invoke_type, class_def_idx, method_idx, class_loader, dex_file); if (!CanCompileMethod(method_idx, dex_file, cu)) { VLOG(compiler) << cu.instruction_set << ": Cannot compile method : " << PrettyMethod(method_idx, dex_file); return nullptr; } cu.NewTimingSplit("MIROpt:CheckFilters"); std::string skip_message; if (cu.mir_graph->SkipCompilation(&skip_message)) { VLOG(compiler) << cu.instruction_set << ": Skipping method : " << PrettyMethod(method_idx, dex_file) << " Reason = " << skip_message; return nullptr; } /* Create the pass driver and launch it */ PassDriverMEOpts pass_driver(&cu); pass_driver.Launch(); /* For non-leaf methods check if we should skip compilation when the profiler is enabled. */ if (cu.compiler_driver->ProfilePresent() && !cu.mir_graph->MethodIsLeaf() && cu.mir_graph->SkipCompilationByName(PrettyMethod(method_idx, dex_file))) { return nullptr; } if (cu.enable_debug & (1 << kDebugDumpCheckStats)) { cu.mir_graph->DumpCheckStats(); } if (kCompilerDebugFlags & (1 << kDebugCountOpcodes)) { cu.mir_graph->ShowOpcodeStats(); } /* Reassociate sreg names with original Dalvik vreg names. */ cu.mir_graph->RemapRegLocations(); /* Free Arenas from the cu.arena_stack for reuse by the cu.arena in the codegen. */ if (cu.enable_debug & (1 << kDebugShowMemoryUsage)) { if (cu.arena_stack.PeakBytesAllocated() > 1 * 1024 * 1024) { MemStats stack_stats(cu.arena_stack.GetPeakStats()); LOG(INFO) << PrettyMethod(method_idx, dex_file) << " " << Dumpable(stack_stats); } } cu.arena_stack.Reset(); CompiledMethod* result = NULL; if (cu.mir_graph->PuntToInterpreter()) { VLOG(compiler) << cu.instruction_set << ": Punted method to interpreter: " << PrettyMethod(method_idx, dex_file); return nullptr; } cu.cg->Materialize(); cu.NewTimingSplit("Dedupe"); /* deduping takes up the vast majority of time in GetCompiledMethod(). */ result = cu.cg->GetCompiledMethod(); cu.NewTimingSplit("Cleanup"); if (result) { VLOG(compiler) << cu.instruction_set << ": Compiled " << PrettyMethod(method_idx, dex_file); } else { VLOG(compiler) << cu.instruction_set << ": Deferred " << PrettyMethod(method_idx, dex_file); } if (cu.enable_debug & (1 << kDebugShowMemoryUsage)) { if (cu.arena.BytesAllocated() > (1 * 1024 *1024)) { MemStats mem_stats(cu.arena.GetMemStats()); LOG(INFO) << PrettyMethod(method_idx, dex_file) << " " << Dumpable(mem_stats); } } if (cu.enable_debug & (1 << kDebugShowSummaryMemoryUsage)) { LOG(INFO) << "MEMINFO " << cu.arena.BytesAllocated() << " " << cu.mir_graph->GetNumBlocks() << " " << PrettyMethod(method_idx, dex_file); } cu.EndTiming(); driver.GetTimingsLogger()->AddLogger(cu.timings); return result; } CompiledMethod* CompileOneMethod(CompilerDriver& driver, Compiler* compiler, const DexFile::CodeItem* code_item, uint32_t access_flags, InvokeType invoke_type, uint16_t class_def_idx, uint32_t method_idx, jobject class_loader, const DexFile& dex_file, void* compilation_unit) { return CompileMethod(driver, compiler, code_item, access_flags, invoke_type, class_def_idx, method_idx, class_loader, dex_file, compilation_unit); } } // namespace art extern "C" art::CompiledMethod* ArtQuickCompileMethod(art::CompilerDriver& driver, const art::DexFile::CodeItem* code_item, uint32_t access_flags, art::InvokeType invoke_type, uint16_t class_def_idx, uint32_t method_idx, jobject class_loader, const art::DexFile& dex_file) { // TODO: check method fingerprint here to determine appropriate backend type. Until then, use // build default. art::Compiler* compiler = driver.GetCompiler(); return art::CompileOneMethod(driver, compiler, code_item, access_flags, invoke_type, class_def_idx, method_idx, class_loader, dex_file, NULL /* use thread llvm_info */); }