/art/runtime/ |
D | linear_alloc.cc | 23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { in LinearAlloc() 28 return allocator_.Realloc(ptr, old_size, new_size); in Realloc() 33 return allocator_.Alloc(size); in Alloc() 38 return allocator_.AllocAlign16(size); in AllocAlign16() 43 return allocator_.BytesUsed(); in GetUsedMemory() 48 return allocator_.GetArenaPool(); in GetArenaPool() 53 return allocator_.Contains(ptr); in Contains() 57 return allocator_.Contains(ptr); in ContainsUnsafe()
|
D | monitor_pool.cc | 60 void* chunk = allocator_.allocate(kChunkSize); in AllocateChunk() 99 allocator_.deallocate(reinterpret_cast<uint8_t*>(monitor_chunks_[i][j]), kChunkSize); in FreeInternal()
|
D | linear_alloc.h | 58 ArenaAllocator allocator_ GUARDED_BY(lock_);
|
/art/libartbase/base/ |
D | arena_containers.h | 143 allocator_(allocator) { in ArenaAllocatorAdapterKind() 148 allocator_(other.allocator_) { in ArenaAllocatorAdapter() 155 ArenaAllocator* allocator_; 179 allocator_(allocator) { in ArenaAllocatorAdapter() 184 allocator_(other.allocator_) { in ArenaAllocatorAdapter() 200 return allocator_->AllocArray<T>(n, ArenaAllocatorAdapterKind::Kind()); 203 allocator_->MakeInaccessible(p, sizeof(T) * n); in deallocate() 216 ArenaAllocator* allocator_; 229 return lhs.allocator_ == rhs.allocator_;
|
D | arena_bit_vector.cc | 66 return allocator_->Alloc(size, this->Kind()); in Alloc() 73 : ArenaBitVectorAllocatorKind(kind), allocator_(allocator) { } in ArenaBitVectorAllocator() 75 ArenaAlloc* const allocator_; member in art::ArenaBitVectorAllocator
|
D | bit_vector.cc | 33 allocator_(allocator), in BitVector() 63 allocator_->Free(storage_); in ~BitVector() 356 static_cast<uint32_t*>(allocator_->Alloc(new_size * kWordBytes)); in EnsureSize() 363 allocator_->Free(storage_); in EnsureSize() 372 return allocator_; in GetAllocator()
|
D | bit_vector.h | 110 allocator_(other.allocator_), in BitVector() 289 Allocator* const allocator_; // Allocator if expandable. variable
|
/art/compiler/utils/ |
D | assembler.cc | 29 : allocator_(allocator) { in AssemblerBuffer() 31 contents_ = allocator_->AllocArray<uint8_t>(kInitialBufferCapacity, kArenaAllocAssembler); in AssemblerBuffer() 48 if (allocator_->IsRunningOnMemoryTool()) { in ~AssemblerBuffer() 49 allocator_->MakeInaccessible(contents_, Capacity()); in ~AssemblerBuffer() 84 allocator_->Realloc(contents_, old_capacity, new_capacity, kArenaAllocAssembler)); in ExtendCapacity()
|
D | jni_macro_assembler_test.h | 62 allocator_.reset(new ArenaAllocator(&pool_)); in SetUp() 63 assembler_.reset(CreateAssembler(allocator_.get())); in SetUp() 80 allocator_.reset(); in TearDown() 144 std::unique_ptr<ArenaAllocator> allocator_; variable
|
/art/compiler/optimizing/ |
D | gvn.cc | 41 : allocator_(allocator), in ValueSet() 54 : allocator_(allocator), in ValueSet() 92 buckets_[index] = new (allocator_) Node(instruction, hash_code, buckets_[index]); in Add() 177 buckets_[new_index] = node->Dup(allocator_, buckets_[new_index]); in PopulateFromInternal() 218 clone_current = node->Dup(allocator_, nullptr); in CloneBucket() 317 ScopedArenaAllocator* const allocator_; member in art::ValueSet 344 allocator_(graph->GetArenaStack()), in GlobalValueNumberer() 346 sets_(graph->GetBlocks().size(), nullptr, allocator_.Adapter(kArenaAllocGvn)), in GlobalValueNumberer() 348 &allocator_, graph->GetBlocks().size(), /* expandable= */ false, kArenaAllocGvn) { in GlobalValueNumberer() 360 ScopedArenaAllocator allocator_; member in art::GlobalValueNumberer [all …]
|
D | bounds_check_elimination.cc | 292 : allocator_(allocator), lower_(lower), upper_(upper) {} in ValueRange() 301 ScopedArenaAllocator* GetAllocator() const { return allocator_; } in GetAllocator() 329 return new (allocator_) ValueRange( in Narrow() 330 allocator_, in Narrow() 350 return new (allocator_) ValueRange(allocator_, lower, upper); in Add() 354 ScopedArenaAllocator* const allocator_; member in art::ValueRange 514 allocator_(graph->GetArenaStack()), in BCEVisitor() 518 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() 519 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() 521 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor() [all …]
|
D | block_builder.cc | 31 : allocator_(graph->GetAllocator()), in HBasicBlockBuilder() 55 block = new (allocator_) HBasicBlock(graph_, semantic_dex_pc); in MaybeCreateBlockAt() 339 HBasicBlock* new_catch_block = new (allocator_) HBasicBlock(graph_, address); in InsertTryBoundaryBlocks() 340 new_catch_block->AddInstruction(new (allocator_) HGoto(address)); in InsertTryBoundaryBlocks() 348 new (allocator_) TryCatchInformation(iterator.GetHandlerTypeIndex(), *dex_file_)); in InsertTryBoundaryBlocks() 365 HTryBoundary* try_entry = new (allocator_) HTryBoundary( in InsertTryBoundaryBlocks() 394 new (allocator_) HTryBoundary(HTryBoundary::BoundaryKind::kExit, successor->GetDexPc()); in InsertTryBoundaryBlocks() 405 graph_->SetEntryBlock(new (allocator_) HBasicBlock(graph_, kNoDexPc)); in Build() 406 graph_->SetExitBlock(new (allocator_) HBasicBlock(graph_, kNoDexPc)); in Build() 424 HBasicBlock* entry_block = new (allocator_) HBasicBlock(graph_, kNoDexPc); in BuildIntrinsic() [all …]
|
D | ssa_liveness_analysis.h | 305 UsePosition* new_use = new (allocator_) UsePosition(instruction, temp_index, position); in AddTempUse() 355 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position); 366 new (allocator_) EnvUsePosition(environment, input_index, position); 370 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position); 378 new (allocator_) LiveRange(start_block_position, position, nullptr); 394 new (allocator_) LiveRange(start_block_position, position, first_range_); 404 new (allocator_) UsePosition(instruction, input_index, block->GetLifetimeEnd()); in AddPhiUse() 411 new (allocator_) LiveRange(start, end, first_range_); in AddRange() 420 first_range_ = range_search_start_ = new (allocator_) LiveRange(start, end, first_range_); in AddRange() 438 new (allocator_) LiveRange(start, end, nullptr); in AddLoopRange() [all …]
|
D | intrinsics_arm64.cc | 268 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits() 271 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble() 282 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits() 285 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat() 324 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes() 332 CreateIntToIntLocations(allocator_, invoke); in VisitLongReverseBytes() 340 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes() 359 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfLeadingZeros() 367 CreateIntToIntLocations(allocator_, invoke); in VisitLongNumberOfLeadingZeros() 387 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfTrailingZeros() [all …]
|
D | graph_checker.h | 36 allocator_(graph->GetArenaStack()), in HGraphDelegateVisitor() 37 seen_ids_(&allocator_, graph->GetCurrentInstructionId(), false, kArenaAllocGraphChecker) { in HGraphDelegateVisitor() 112 ScopedArenaAllocator allocator_; variable
|
D | intrinsics_arm_vixl.cc | 245 : allocator_(codegen->GetGraph()->GetAllocator()), in IntrinsicLocationsBuilderARMVIXL() 294 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits() 297 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble() 308 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits() 311 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat() 370 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfLeadingZeros() 378 CreateLongToLongLocationsWithOverlap(allocator_, invoke); in VisitLongNumberOfLeadingZeros() 416 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfTrailingZeros() 424 CreateLongToLongLocationsWithOverlap(allocator_, invoke); in VisitLongNumberOfTrailingZeros() 432 CreateFPToFPLocations(allocator_, invoke); in VisitMathSqrt() [all …]
|
D | intrinsics_arm64.h | 43 : allocator_(allocator), codegen_(codegen) {} in IntrinsicLocationsBuilderARM64() 60 ArenaAllocator* const allocator_;
|
D | register_allocation_resolver.cc | 28 : allocator_(codegen->GetGraph()->GetAllocator()), in RegisterAllocationResolver() 538 move = new (allocator_) HParallelMove(allocator_); in AddInputMoveFor() 586 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 601 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 617 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt() 649 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtExitOf() 671 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtEntryOf() 695 move = new (allocator_) HParallelMove(allocator_); in InsertMoveAfter()
|
D | scheduler.h | 257 allocator_(allocator), in SchedulingGraph() 259 nodes_map_(allocator_->Adapter(kArenaAllocScheduler)), in SchedulingGraph() 264 new (allocator_) SchedulingNode(instr, allocator_, is_scheduling_barrier)); 318 ScopedArenaAllocator* const allocator_; variable
|
D | stack_map_stream.h | 40 : allocator_(allocator), in StackMapStream() 100 ScopedArenaAllocator* allocator_; variable
|
D | parallel_move_resolver.h | 129 allocator_(allocator) { in ParallelMoveResolverNoSwap() 197 ArenaAllocator* const allocator_; variable
|
D | block_builder.h | 69 ArenaAllocator* const allocator_; variable
|
D | load_store_elimination.cc | 97 allocator_(graph->GetArenaStack()), in LSEVisitor() 102 allocator_.Adapter(kArenaAllocLSE)), in LSEVisitor() 103 allocator_.Adapter(kArenaAllocLSE)), in LSEVisitor() 104 removed_loads_(allocator_.Adapter(kArenaAllocLSE)), in LSEVisitor() 105 substitute_instructions_for_loads_(allocator_.Adapter(kArenaAllocLSE)), in LSEVisitor() 106 possibly_removed_stores_(allocator_.Adapter(kArenaAllocLSE)), in LSEVisitor() 107 singleton_new_instances_(allocator_.Adapter(kArenaAllocLSE)) { in LSEVisitor() 863 ScopedArenaAllocator allocator_; member in art::LSEVisitor
|
/art/runtime/verifier/ |
D | reg_type_cache.cc | 174 char* ptr = allocator_.AllocArray<char>(str.length()); in AddString() 208 new (&allocator_) PreciseReferenceType(klass, AddString(sv_descriptor), entries_.size()); in From() 210 entry = new (&allocator_) ReferenceType(klass, AddString(sv_descriptor), entries_.size()); in From() 224 new (&allocator_) UnresolvedReferenceType(AddString(sv_descriptor), entries_.size())); in From() 235 return AddEntry(new (&allocator_) UnresolvedReferenceType(AddString("a"), entries_.size())); in MakeUnresolvedReference() 264 new (&allocator_) PreciseReferenceType(klass, descriptor, entries_.size())) in InsertClass() 265 : new (&allocator_) ReferenceType(klass, descriptor, entries_.size()); in InsertClass() 284 allocator_(allocator) { in RegTypeCache() 386 ArenaBitVector types(&allocator_, in FromUnresolvedMerge() 466 return AddEntry(new (&allocator_) UnresolvedMergedType(resolved_parts_merged, in FromUnresolvedMerge() [all …]
|
/art/runtime/mirror/ |
D | class_loader.h | 58 GetField64(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_))); in GetAllocator() 62 SetField64<false>(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_), in SetAllocator() 83 uint64_t allocator_; variable
|