/art/runtime/ |
D | linear_alloc-inl.h | 33 Arena* arena = allocator_.GetHeadArena(); in SetFirstObject() 58 allocator_.ResetCurrentArena(); in SetupForPostZygoteFork() 80 void* ret = allocator_.Realloc(ptr, old_size, new_size); in Realloc() 85 return allocator_.Realloc(ptr, old_size, new_size); in Realloc() 93 TrackingHeader* storage = new (allocator_.Alloc(size)) TrackingHeader(size, kind); in Alloc() 97 return allocator_.Alloc(size); in Alloc() 107 uint8_t* ptr = allocator_.CurrentPtr() + sizeof(TrackingHeader); in AllocAlign16() 113 if (allocator_.CurrentArenaUnusedBytes() < required_size + mem_tool_bytes) { in AllocAlign16() 123 ptr = static_cast<uint8_t*>(allocator_.Alloc(required_size)); in AllocAlign16() 128 return allocator_.AllocAlign16(size); in AllocAlign16() [all …]
|
D | linear_alloc.h | 79 : lock_("linear alloc"), allocator_(pool), track_allocations_(track_allocs) {} in LinearAlloc() 111 return allocator_.Contains(ptr); in ContainsUnsafe() 120 ArenaAllocator allocator_ GUARDED_BY(lock_);
|
/art/compiler/optimizing/ |
D | code_generation_data.h | 47 return &allocator_; in GetScopedAllocator() 94 : allocator_(std::move(allocator)), in CodeGenerationData() 95 stack_map_stream_(&allocator_, instruction_set), in CodeGenerationData() 96 slow_paths_(allocator_.Adapter(kArenaAllocCodeGenerator)), in CodeGenerationData() 98 allocator_.Adapter(kArenaAllocCodeGenerator)), in CodeGenerationData() 100 allocator_.Adapter(kArenaAllocCodeGenerator)) { in CodeGenerationData() 106 ScopedArenaAllocator allocator_; variable
|
D | instruction_builder.cc | 97 : allocator_(graph->GetAllocator()), in HInstructionBuilder() 148 HPhi* phi = new (allocator_) HPhi( in GetLocalsForWithAllocation() 149 allocator_, in GetLocalsForWithAllocation() 198 HPhi* phi = new (allocator_) HPhi( in InitializeBlockLocals() 199 allocator_, in InitializeBlockLocals() 237 HPhi* phi = new (allocator_) HPhi( in InitializeBlockLocals() 238 allocator_, in InitializeBlockLocals() 299 HEnvironment* environment = new (allocator_) HEnvironment( in InitializeInstruction() 300 allocator_, in InitializeInstruction() 316 HNullCheck* null_check = new (allocator_) HNullCheck(ref, dex_pc); in LoadNullCheckedLocal() [all …]
|
D | gvn.cc | 41 : allocator_(allocator), in ValueSet() 54 : allocator_(allocator), in ValueSet() 92 buckets_[index] = new (allocator_) Node(instruction, hash_code, buckets_[index]); in Add() 177 buckets_[new_index] = node->Dup(allocator_, buckets_[new_index]); in PopulateFromInternal() 233 clone_current = node->Dup(allocator_, nullptr); in CloneBucket() 332 ScopedArenaAllocator* const allocator_; member in art::ValueSet 358 allocator_(graph->GetArenaStack()), in GlobalValueNumberer() 360 sets_(graph->GetBlocks().size(), nullptr, allocator_.Adapter(kArenaAllocGvn)), in GlobalValueNumberer() 362 &allocator_, graph->GetBlocks().size(), /* expandable= */ false, kArenaAllocGvn) {} in GlobalValueNumberer() 372 ScopedArenaAllocator allocator_; member in art::GlobalValueNumberer [all …]
|
D | bounds_check_elimination.cc | 292 : allocator_(allocator), lower_(lower), upper_(upper) {} in ValueRange() 301 ScopedArenaAllocator* GetAllocator() const { return allocator_; } in GetAllocator() 329 return new (allocator_) ValueRange( in Narrow() 330 allocator_, in Narrow() 350 return new (allocator_) ValueRange(allocator_, lower, upper); in Add() 354 ScopedArenaAllocator* const allocator_; member in art::ValueRange 514 allocator_(graph->GetArenaStack()), in BCEVisitor() 518 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() 519 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() 521 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() [all …]
|
D | block_builder.cc | 30 : allocator_(graph->GetAllocator()), in HBasicBlockBuilder() 51 block = new (allocator_) HBasicBlock(graph_, semantic_dex_pc); in MaybeCreateBlockAt() 326 HBasicBlock* new_catch_block = new (allocator_) HBasicBlock(graph_, address); in InsertTryBoundaryBlocks() 327 new_catch_block->AddInstruction(new (allocator_) HGoto(address)); in InsertTryBoundaryBlocks() 335 new (allocator_) TryCatchInformation(iterator.GetHandlerTypeIndex(), *dex_file_)); in InsertTryBoundaryBlocks() 352 HTryBoundary* try_entry = new (allocator_) HTryBoundary( in InsertTryBoundaryBlocks() 381 new (allocator_) HTryBoundary(HTryBoundary::BoundaryKind::kExit, successor->GetDexPc()); in InsertTryBoundaryBlocks() 389 ArenaSet<uint32_t> targets(allocator_->Adapter(kArenaAllocGraphBuilder)); in InsertSynthesizedLoopsForOsr() 417 HBasicBlock* loop_block = new (allocator_) HBasicBlock(graph_, block->GetDexPc()); in InsertSynthesizedLoopsForOsr() 426 loop_block->AddInstruction(new (allocator_) HIf(graph_->GetIntConstant(0), kNoDexPc)); in InsertSynthesizedLoopsForOsr() [all …]
|
D | graph_checker.h | 41 allocator_(graph->GetArenaStack()), in HGraphDelegateVisitor() 42 seen_ids_(&allocator_, graph->GetCurrentInstructionId(), false, kArenaAllocGraphChecker), in HGraphDelegateVisitor() 43 uses_per_instruction_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor() 44 instructions_per_block_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor() 45 phis_per_block_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor() 139 ScopedArenaAllocator allocator_; variable
|
D | ssa_liveness_analysis.h | 303 UsePosition* new_use = new (allocator_) UsePosition(instruction, temp_index, position); in AddTempUse() 353 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position); 364 new (allocator_) EnvUsePosition(environment, input_index, position); 368 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position); 376 new (allocator_) LiveRange(start_block_position, position, nullptr); 392 new (allocator_) LiveRange(start_block_position, position, first_range_); 402 new (allocator_) UsePosition(instruction, input_index, block->GetLifetimeEnd()); in AddPhiUse() 409 new (allocator_) LiveRange(start, end, first_range_); in AddRange() 418 first_range_ = range_search_start_ = new (allocator_) LiveRange(start, end, first_range_); in AddRange() 436 new (allocator_) LiveRange(start, end, nullptr); in AddLoopRange() [all …]
|
D | intrinsics_riscv64.h | 39 : allocator_(allocator), codegen_(codegen) {} in IntrinsicLocationsBuilderRISCV64() 54 ArenaAllocator* const allocator_;
|
D | register_allocation_resolver.cc | 28 : allocator_(codegen->GetGraph()->GetAllocator()), in RegisterAllocationResolver() 537 move = new (allocator_) HParallelMove(allocator_); in AddInputMoveFor() 585 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 600 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 616 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 649 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtExitOf() 671 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtEntryOf() 695 move = new (allocator_) HParallelMove(allocator_); in InsertMoveAfter()
|
D | intrinsics_arm64.cc | 203 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits() 206 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble() 217 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits() 220 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat() 295 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes() 303 CreateIntToIntLocations(allocator_, invoke); in VisitLongReverseBytes() 311 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes() 330 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfLeadingZeros() 338 CreateIntToIntLocations(allocator_, invoke); in VisitLongNumberOfLeadingZeros() 358 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfTrailingZeros() [all …]
|
D | intrinsics_arm64.h | 45 : allocator_(allocator), codegen_(codegen) {} in IntrinsicLocationsBuilderARM64() 60 ArenaAllocator* const allocator_;
|
D | intrinsics_riscv64.cc | 171 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits() 181 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble() 191 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits() 201 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat() 211 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleIsInfinite() 224 CreateFPToIntLocations(allocator_, invoke); in VisitFloatIsInfinite() 250 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekByte() 259 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekIntNative() 268 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekLongNative() 277 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekShortNative() [all …]
|
D | intrinsics_x86_64.cc | 46 : allocator_(codegen->GetGraph()->GetAllocator()), codegen_(codegen) { in IntrinsicLocationsBuilderX86_64() 171 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits() 174 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble() 185 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits() 188 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat() 206 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes() 214 CreateIntToIntLocations(allocator_, invoke); in VisitLongReverseBytes() 222 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes() 275 CreateFPToIntLocations(allocator_, invoke); in VisitFloatIsInfinite() 283 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleIsInfinite() [all …]
|
D | intrinsics_x86.cc | 48 : allocator_(codegen->GetGraph()->GetAllocator()), in IntrinsicLocationsBuilderX86() 203 CreateFPToIntLocations(allocator_, invoke, /* is64bit= */ true); in VisitDoubleDoubleToRawLongBits() 206 CreateIntToFPLocations(allocator_, invoke, /* is64bit= */ true); in VisitDoubleLongBitsToDouble() 217 CreateFPToIntLocations(allocator_, invoke, /* is64bit= */ false); in VisitFloatFloatToRawIntBits() 220 CreateIntToFPLocations(allocator_, invoke, /* is64bit= */ false); in VisitFloatIntBitsToFloat() 272 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes() 280 CreateLongToLongLocations(allocator_, invoke); in VisitLongReverseBytes() 301 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes() 316 CreateFPToFPLocations(allocator_, invoke); in VisitMathSqrt() 347 CreateSSE41FPToFPLocations(allocator_, invoke, codegen_); in VisitMathCeil() [all …]
|
D | scheduler.h | 326 : allocator_(allocator), in SchedulingGraph() 328 nodes_map_(allocator_->Adapter(kArenaAllocScheduler)), in SchedulingGraph() 333 new (allocator_) SchedulingNode(instr, allocator_, is_scheduling_barrier)); 390 ScopedArenaAllocator* const allocator_; variable
|
/art/runtime/verifier/ |
D | reg_type_cache.cc | 57 entries_[i] = new (&allocator_) PreciseConstType(null_handle_, value, i); in FillPrimitiveAndSmallConstantTypes() 61 entries_[id] = new (&allocator_) type( \ in FillPrimitiveAndSmallConstantTypes() 80 new (&allocator_) UndefinedType(null_handle_, "", kUndefinedCacheId); in FillPrimitiveAndSmallConstantTypes() 82 new (&allocator_) ConflictType(null_handle_, "", kConflictCacheId); in FillPrimitiveAndSmallConstantTypes() 84 new (&allocator_) NullType(null_handle_, "", kNullCacheId); in FillPrimitiveAndSmallConstantTypes() 176 char* ptr = allocator_.AllocArray<char>(str.length()); in AddString() 204 entry = new (&allocator_) PreciseReferenceType(handles_.NewHandle(klass), in From() 208 entry = new (&allocator_) ReferenceType(handles_.NewHandle(klass), in From() 223 return AddEntry(new (&allocator_) UnresolvedReferenceType(null_handle_, in From() 236 return AddEntry(new (&allocator_) UnresolvedReferenceType( in MakeUnresolvedReference() [all …]
|
/art/libartbase/base/ |
D | arena_containers.h | 147 allocator_(allocator) { in ArenaAllocatorAdapterKind() 152 allocator_(other.allocator_) { in ArenaAllocatorAdapter() 159 ArenaAllocator* allocator_; 183 allocator_(allocator) { in ArenaAllocatorAdapter() 188 allocator_(other.allocator_) { in ArenaAllocatorAdapter() 204 return allocator_->AllocArray<T>(n, ArenaAllocatorAdapterKind::Kind()); 207 allocator_->MakeInaccessible(p, sizeof(T) * n); in deallocate() 220 ArenaAllocator* allocator_; 233 return lhs.allocator_ == rhs.allocator_;
|
D | arena_bit_vector.cc | 67 return allocator_->Alloc(size, this->Kind()); in Alloc() 74 : ArenaBitVectorAllocatorKind(kind), allocator_(allocator) { } in ArenaBitVectorAllocator() 76 ArenaAlloc* const allocator_; member in art::ArenaBitVectorAllocator
|
D | bit_vector.cc | 33 allocator_(allocator), in BitVector() 61 allocator_->Free(storage_); in ~BitVector() 355 static_cast<uint32_t*>(allocator_->Alloc(new_size * kWordBytes)); in EnsureSize() 362 allocator_->Free(storage_); in EnsureSize() 371 return allocator_; in GetAllocator()
|
D | bit_vector.h | 122 allocator_(other.allocator_), in BitVector() 309 Allocator* const allocator_; // Allocator if expandable. variable
|
/art/compiler/utils/ |
D | assembler.cc | 29 : allocator_(allocator) { in AssemblerBuffer() 31 contents_ = allocator_->AllocArray<uint8_t>(kInitialBufferCapacity, kArenaAllocAssembler); in AssemblerBuffer() 48 if (allocator_->IsRunningOnMemoryTool()) { in ~AssemblerBuffer() 49 allocator_->MakeInaccessible(contents_, Capacity()); in ~AssemblerBuffer() 87 allocator_->Realloc(contents_, old_capacity, new_capacity, kArenaAllocAssembler)); in ExtendCapacity()
|
D | jni_macro_assembler_test.h | 59 allocator_.reset(new ArenaAllocator(&pool_)); in SetUp() 60 assembler_.reset(CreateAssembler(allocator_.get())); in SetUp() 67 allocator_.reset(); in TearDown() 93 std::unique_ptr<ArenaAllocator> allocator_; variable
|
/art/runtime/mirror/ |
D | class_loader.h | 62 GetField64(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_))); in GetAllocator() 66 SetField64<false>(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_), in SetAllocator() 86 uint64_t allocator_; variable
|