Home
last modified time | relevance | path

Searched refs:allocator_ (Results 1 – 25 of 68) sorted by relevance

123

/art/runtime/
Dlinear_alloc-inl.h33 Arena* arena = allocator_.GetHeadArena(); in SetFirstObject()
58 allocator_.ResetCurrentArena(); in SetupForPostZygoteFork()
80 void* ret = allocator_.Realloc(ptr, old_size, new_size); in Realloc()
85 return allocator_.Realloc(ptr, old_size, new_size); in Realloc()
93 TrackingHeader* storage = new (allocator_.Alloc(size)) TrackingHeader(size, kind); in Alloc()
97 return allocator_.Alloc(size); in Alloc()
107 uint8_t* ptr = allocator_.CurrentPtr() + sizeof(TrackingHeader); in AllocAlign16()
113 if (allocator_.CurrentArenaUnusedBytes() < required_size + mem_tool_bytes) { in AllocAlign16()
123 ptr = static_cast<uint8_t*>(allocator_.Alloc(required_size)); in AllocAlign16()
128 return allocator_.AllocAlign16(size); in AllocAlign16()
[all …]
Dlinear_alloc.h79 : lock_("linear alloc"), allocator_(pool), track_allocations_(track_allocs) {} in LinearAlloc()
111 return allocator_.Contains(ptr); in ContainsUnsafe()
120 ArenaAllocator allocator_ GUARDED_BY(lock_);
/art/compiler/optimizing/
Dcode_generation_data.h47 return &allocator_; in GetScopedAllocator()
94 : allocator_(std::move(allocator)), in CodeGenerationData()
95 stack_map_stream_(&allocator_, instruction_set), in CodeGenerationData()
96 slow_paths_(allocator_.Adapter(kArenaAllocCodeGenerator)), in CodeGenerationData()
98 allocator_.Adapter(kArenaAllocCodeGenerator)), in CodeGenerationData()
100 allocator_.Adapter(kArenaAllocCodeGenerator)) { in CodeGenerationData()
106 ScopedArenaAllocator allocator_; variable
Dinstruction_builder.cc97 : allocator_(graph->GetAllocator()), in HInstructionBuilder()
148 HPhi* phi = new (allocator_) HPhi( in GetLocalsForWithAllocation()
149 allocator_, in GetLocalsForWithAllocation()
198 HPhi* phi = new (allocator_) HPhi( in InitializeBlockLocals()
199 allocator_, in InitializeBlockLocals()
237 HPhi* phi = new (allocator_) HPhi( in InitializeBlockLocals()
238 allocator_, in InitializeBlockLocals()
299 HEnvironment* environment = new (allocator_) HEnvironment( in InitializeInstruction()
300 allocator_, in InitializeInstruction()
316 HNullCheck* null_check = new (allocator_) HNullCheck(ref, dex_pc); in LoadNullCheckedLocal()
[all …]
Dgvn.cc41 : allocator_(allocator), in ValueSet()
54 : allocator_(allocator), in ValueSet()
92 buckets_[index] = new (allocator_) Node(instruction, hash_code, buckets_[index]); in Add()
177 buckets_[new_index] = node->Dup(allocator_, buckets_[new_index]); in PopulateFromInternal()
233 clone_current = node->Dup(allocator_, nullptr); in CloneBucket()
332 ScopedArenaAllocator* const allocator_; member in art::ValueSet
358 allocator_(graph->GetArenaStack()), in GlobalValueNumberer()
360 sets_(graph->GetBlocks().size(), nullptr, allocator_.Adapter(kArenaAllocGvn)), in GlobalValueNumberer()
362 &allocator_, graph->GetBlocks().size(), /* expandable= */ false, kArenaAllocGvn) {} in GlobalValueNumberer()
372 ScopedArenaAllocator allocator_; member in art::GlobalValueNumberer
[all …]
Dbounds_check_elimination.cc292 : allocator_(allocator), lower_(lower), upper_(upper) {} in ValueRange()
301 ScopedArenaAllocator* GetAllocator() const { return allocator_; } in GetAllocator()
329 return new (allocator_) ValueRange( in Narrow()
330 allocator_, in Narrow()
350 return new (allocator_) ValueRange(allocator_, lower, upper); in Add()
354 ScopedArenaAllocator* const allocator_; member in art::ValueRange
514 allocator_(graph->GetArenaStack()), in BCEVisitor()
518 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor()
519 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor()
521 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), in BCEVisitor()
[all …]
Dblock_builder.cc30 : allocator_(graph->GetAllocator()), in HBasicBlockBuilder()
51 block = new (allocator_) HBasicBlock(graph_, semantic_dex_pc); in MaybeCreateBlockAt()
326 HBasicBlock* new_catch_block = new (allocator_) HBasicBlock(graph_, address); in InsertTryBoundaryBlocks()
327 new_catch_block->AddInstruction(new (allocator_) HGoto(address)); in InsertTryBoundaryBlocks()
335 new (allocator_) TryCatchInformation(iterator.GetHandlerTypeIndex(), *dex_file_)); in InsertTryBoundaryBlocks()
352 HTryBoundary* try_entry = new (allocator_) HTryBoundary( in InsertTryBoundaryBlocks()
381 new (allocator_) HTryBoundary(HTryBoundary::BoundaryKind::kExit, successor->GetDexPc()); in InsertTryBoundaryBlocks()
389 ArenaSet<uint32_t> targets(allocator_->Adapter(kArenaAllocGraphBuilder)); in InsertSynthesizedLoopsForOsr()
417 HBasicBlock* loop_block = new (allocator_) HBasicBlock(graph_, block->GetDexPc()); in InsertSynthesizedLoopsForOsr()
426 loop_block->AddInstruction(new (allocator_) HIf(graph_->GetIntConstant(0), kNoDexPc)); in InsertSynthesizedLoopsForOsr()
[all …]
Dgraph_checker.h41 allocator_(graph->GetArenaStack()), in HGraphDelegateVisitor()
42 seen_ids_(&allocator_, graph->GetCurrentInstructionId(), false, kArenaAllocGraphChecker), in HGraphDelegateVisitor()
43 uses_per_instruction_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor()
44 instructions_per_block_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor()
45 phis_per_block_(allocator_.Adapter(kArenaAllocGraphChecker)), in HGraphDelegateVisitor()
139 ScopedArenaAllocator allocator_; variable
Dssa_liveness_analysis.h303 UsePosition* new_use = new (allocator_) UsePosition(instruction, temp_index, position); in AddTempUse()
353 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position);
364 new (allocator_) EnvUsePosition(environment, input_index, position);
368 UsePosition* new_use = new (allocator_) UsePosition(instruction, input_index, position);
376 new (allocator_) LiveRange(start_block_position, position, nullptr);
392 new (allocator_) LiveRange(start_block_position, position, first_range_);
402 new (allocator_) UsePosition(instruction, input_index, block->GetLifetimeEnd()); in AddPhiUse()
409 new (allocator_) LiveRange(start, end, first_range_); in AddRange()
418 first_range_ = range_search_start_ = new (allocator_) LiveRange(start, end, first_range_); in AddRange()
436 new (allocator_) LiveRange(start, end, nullptr); in AddLoopRange()
[all …]
Dintrinsics_riscv64.h39 : allocator_(allocator), codegen_(codegen) {} in IntrinsicLocationsBuilderRISCV64()
54 ArenaAllocator* const allocator_;
Dregister_allocation_resolver.cc28 : allocator_(codegen->GetGraph()->GetAllocator()), in RegisterAllocationResolver()
537 move = new (allocator_) HParallelMove(allocator_); in AddInputMoveFor()
585 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt()
600 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt()
616 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAt()
649 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtExitOf()
671 move = new (allocator_) HParallelMove(allocator_); in InsertParallelMoveAtEntryOf()
695 move = new (allocator_) HParallelMove(allocator_); in InsertMoveAfter()
Dintrinsics_arm64.cc203 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits()
206 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble()
217 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits()
220 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat()
295 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes()
303 CreateIntToIntLocations(allocator_, invoke); in VisitLongReverseBytes()
311 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes()
330 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfLeadingZeros()
338 CreateIntToIntLocations(allocator_, invoke); in VisitLongNumberOfLeadingZeros()
358 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerNumberOfTrailingZeros()
[all …]
Dintrinsics_arm64.h45 : allocator_(allocator), codegen_(codegen) {} in IntrinsicLocationsBuilderARM64()
60 ArenaAllocator* const allocator_;
Dintrinsics_riscv64.cc171 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits()
181 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble()
191 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits()
201 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat()
211 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleIsInfinite()
224 CreateFPToIntLocations(allocator_, invoke); in VisitFloatIsInfinite()
250 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekByte()
259 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekIntNative()
268 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekLongNative()
277 CreateIntToIntNoOverlapLocations(allocator_, invoke); in VisitMemoryPeekShortNative()
[all …]
Dintrinsics_x86_64.cc46 : allocator_(codegen->GetGraph()->GetAllocator()), codegen_(codegen) { in IntrinsicLocationsBuilderX86_64()
171 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleDoubleToRawLongBits()
174 CreateIntToFPLocations(allocator_, invoke); in VisitDoubleLongBitsToDouble()
185 CreateFPToIntLocations(allocator_, invoke); in VisitFloatFloatToRawIntBits()
188 CreateIntToFPLocations(allocator_, invoke); in VisitFloatIntBitsToFloat()
206 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes()
214 CreateIntToIntLocations(allocator_, invoke); in VisitLongReverseBytes()
222 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes()
275 CreateFPToIntLocations(allocator_, invoke); in VisitFloatIsInfinite()
283 CreateFPToIntLocations(allocator_, invoke); in VisitDoubleIsInfinite()
[all …]
Dintrinsics_x86.cc48 : allocator_(codegen->GetGraph()->GetAllocator()), in IntrinsicLocationsBuilderX86()
203 CreateFPToIntLocations(allocator_, invoke, /* is64bit= */ true); in VisitDoubleDoubleToRawLongBits()
206 CreateIntToFPLocations(allocator_, invoke, /* is64bit= */ true); in VisitDoubleLongBitsToDouble()
217 CreateFPToIntLocations(allocator_, invoke, /* is64bit= */ false); in VisitFloatFloatToRawIntBits()
220 CreateIntToFPLocations(allocator_, invoke, /* is64bit= */ false); in VisitFloatIntBitsToFloat()
272 CreateIntToIntLocations(allocator_, invoke); in VisitIntegerReverseBytes()
280 CreateLongToLongLocations(allocator_, invoke); in VisitLongReverseBytes()
301 CreateIntToIntLocations(allocator_, invoke); in VisitShortReverseBytes()
316 CreateFPToFPLocations(allocator_, invoke); in VisitMathSqrt()
347 CreateSSE41FPToFPLocations(allocator_, invoke, codegen_); in VisitMathCeil()
[all …]
Dscheduler.h326 : allocator_(allocator), in SchedulingGraph()
328 nodes_map_(allocator_->Adapter(kArenaAllocScheduler)), in SchedulingGraph()
333 new (allocator_) SchedulingNode(instr, allocator_, is_scheduling_barrier));
390 ScopedArenaAllocator* const allocator_; variable
/art/runtime/verifier/
Dreg_type_cache.cc57 entries_[i] = new (&allocator_) PreciseConstType(null_handle_, value, i); in FillPrimitiveAndSmallConstantTypes()
61 entries_[id] = new (&allocator_) type( \ in FillPrimitiveAndSmallConstantTypes()
80 new (&allocator_) UndefinedType(null_handle_, "", kUndefinedCacheId); in FillPrimitiveAndSmallConstantTypes()
82 new (&allocator_) ConflictType(null_handle_, "", kConflictCacheId); in FillPrimitiveAndSmallConstantTypes()
84 new (&allocator_) NullType(null_handle_, "", kNullCacheId); in FillPrimitiveAndSmallConstantTypes()
176 char* ptr = allocator_.AllocArray<char>(str.length()); in AddString()
204 entry = new (&allocator_) PreciseReferenceType(handles_.NewHandle(klass), in From()
208 entry = new (&allocator_) ReferenceType(handles_.NewHandle(klass), in From()
223 return AddEntry(new (&allocator_) UnresolvedReferenceType(null_handle_, in From()
236 return AddEntry(new (&allocator_) UnresolvedReferenceType( in MakeUnresolvedReference()
[all …]
/art/libartbase/base/
Darena_containers.h147 allocator_(allocator) { in ArenaAllocatorAdapterKind()
152 allocator_(other.allocator_) { in ArenaAllocatorAdapter()
159 ArenaAllocator* allocator_;
183 allocator_(allocator) { in ArenaAllocatorAdapter()
188 allocator_(other.allocator_) { in ArenaAllocatorAdapter()
204 return allocator_->AllocArray<T>(n, ArenaAllocatorAdapterKind::Kind());
207 allocator_->MakeInaccessible(p, sizeof(T) * n); in deallocate()
220 ArenaAllocator* allocator_;
233 return lhs.allocator_ == rhs.allocator_;
Darena_bit_vector.cc67 return allocator_->Alloc(size, this->Kind()); in Alloc()
74 : ArenaBitVectorAllocatorKind(kind), allocator_(allocator) { } in ArenaBitVectorAllocator()
76 ArenaAlloc* const allocator_; member in art::ArenaBitVectorAllocator
Dbit_vector.cc33 allocator_(allocator), in BitVector()
61 allocator_->Free(storage_); in ~BitVector()
355 static_cast<uint32_t*>(allocator_->Alloc(new_size * kWordBytes)); in EnsureSize()
362 allocator_->Free(storage_); in EnsureSize()
371 return allocator_; in GetAllocator()
Dbit_vector.h122 allocator_(other.allocator_), in BitVector()
309 Allocator* const allocator_; // Allocator if expandable. variable
/art/compiler/utils/
Dassembler.cc29 : allocator_(allocator) { in AssemblerBuffer()
31 contents_ = allocator_->AllocArray<uint8_t>(kInitialBufferCapacity, kArenaAllocAssembler); in AssemblerBuffer()
48 if (allocator_->IsRunningOnMemoryTool()) { in ~AssemblerBuffer()
49 allocator_->MakeInaccessible(contents_, Capacity()); in ~AssemblerBuffer()
87 allocator_->Realloc(contents_, old_capacity, new_capacity, kArenaAllocAssembler)); in ExtendCapacity()
Djni_macro_assembler_test.h59 allocator_.reset(new ArenaAllocator(&pool_)); in SetUp()
60 assembler_.reset(CreateAssembler(allocator_.get())); in SetUp()
67 allocator_.reset(); in TearDown()
93 std::unique_ptr<ArenaAllocator> allocator_; variable
/art/runtime/mirror/
Dclass_loader.h62 GetField64(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_))); in GetAllocator()
66 SetField64<false>(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_), in SetAllocator()
86 uint64_t allocator_; variable

123