/art/runtime/ |
D | indirect_reference_table_test.cc | 96 EXPECT_EQ(0U, irt.Capacity()); in TEST_F() 118 ASSERT_EQ(0U, irt.Capacity()); in TEST_F() 130 ASSERT_EQ(3U, irt.Capacity()); in TEST_F() 146 ASSERT_EQ(0U, irt.Capacity()); in TEST_F() 167 ASSERT_EQ(4U, irt.Capacity()) << "hole not filled"; in TEST_F() 175 ASSERT_EQ(3U, irt.Capacity()) << "should be 3 after two deletions"; in TEST_F() 182 ASSERT_EQ(0U, irt.Capacity()) << "not empty after split remove"; in TEST_F() 198 ASSERT_EQ(0U, irt.Capacity()) << "switching del not empty"; in TEST_F() 216 ASSERT_EQ(0U, irt.Capacity()) << "temporal del not empty"; in TEST_F() 241 ASSERT_EQ(kTableInitial + 1, irt.Capacity()); in TEST_F() [all …]
|
D | indirect_reference_table.h | 323 size_t Capacity() const { in Capacity() function 329 return IrtIterator(table_, 0, Capacity()); in begin() 333 return IrtIterator(table_, Capacity(), Capacity()); in end()
|
D | jni_env_ext-inl.h | 33 size_t entry_count = locals.Capacity(); in AddLocalReference()
|
D | indirect_reference_table.cc | 156 for (size_t i = 0; i < Capacity(); ++i) { in AssertEmpty() 266 const size_t top_index = Capacity(); in Trim() 285 for (size_t i = 0; i < Capacity(); ++i) { in Dump()
|
D | java_vm_ext.cc | 587 os << "; globals=" << globals_.Capacity(); in DumpForSigQuit() 591 if (weak_globals_.Capacity() > 0) { in DumpForSigQuit() 592 os << " (plus " << weak_globals_.Capacity() << " weak)"; in DumpForSigQuit()
|
D | jni_internal.cc | 2338 const size_t capacity = soa.Env()->locals.Capacity(); in EnsureLocalCapacityInternal()
|
/art/compiler/utils/ |
D | assembler.cc | 61 CHECK_EQ(Capacity(), kInitialBufferCapacity); in AssemblerBuffer() 68 arena_->MakeInaccessible(contents_, Capacity()); in ~AssemblerBuffer() 96 size_t old_capacity = Capacity(); in ExtendCapacity() 110 CHECK_EQ(Capacity(), new_capacity); in ExtendCapacity()
|
D | assembler.h | 115 if (new_size > Capacity()) { in Resize() 211 int ComputeGap() { return buffer_->Capacity() - buffer_->Size(); } in ComputeGap() 238 size_t Capacity() const { in Capacity() function
|
/art/runtime/gc/space/ |
D | malloc_space.cc | 139 CHECK_LE(new_end, Begin() + Capacity()); in MoreCore() 182 << "Capacity " << Capacity(); in CreateZygoteSpace() 226 << ",size=" << PrettySize(Size()) << ",capacity=" << PrettySize(Capacity()) in Dump() 252 size_t new_capacity = Capacity(); in ClampGrowthLimit()
|
D | space.h | 303 virtual size_t Capacity() const { in Capacity() function 377 return Capacity(); in NonGrowthLimitCapacity()
|
D | malloc_space.h | 109 size_t Capacity() const { in Capacity() function
|
D | bump_pointer_space.h | 84 size_t Capacity() const { in Capacity() function
|
D | space_test.h | 334 EXPECT_EQ(space->Capacity(), growth_limit); in SizeFootPrintGrowthLimitAndTrimDriver() 346 EXPECT_EQ(space->Capacity(), capacity); in SizeFootPrintGrowthLimitAndTrimDriver()
|
D | dlmalloc_space.cc | 135 size_t max_allowed = Capacity(); in AllocWithGrowth()
|
D | rosalloc_space.cc | 161 size_t max_allowed = Capacity(); in AllocWithGrowth()
|
/art/runtime/gc/accounting/ |
D | atomic_stack.h | 174 size_t Capacity() const { in Capacity() function
|
/art/runtime/gc/collector/ |
D | mark_compact.cc | 269 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in MarkStackPush() 270 ResizeMarkStack(mark_stack_->Capacity() * 2); in MarkStackPush()
|
D | mark_sweep.cc | 358 ResizeMarkStack(mark_stack_->Capacity() * 2); in ExpandMarkStack() 363 if (UNLIKELY(mark_stack_->Size() < mark_stack_->Capacity())) { in ResizeMarkStack() 384 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in MarkObjectNonNullParallel() 512 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in PushOnMarkStack()
|
D | semi_space.cc | 468 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in MarkStackPush() 469 ResizeMarkStack(mark_stack_->Capacity() * 2); in MarkStackPush()
|
D | concurrent_copying.cc | 146 accounting::ContinuousSpaceBitmap::Create(bitmap_name, space->Begin(), space->Capacity()); in BindBitmaps() 152 space->Begin(), space->Capacity()); in BindBitmaps() 573 const size_t new_size = gc_mark_stack_->Capacity() * 2; in ExpandGcMarkStack()
|
/art/runtime/gc/ |
D | heap.cc | 438 non_moving_space_->SetFootprintLimit(non_moving_space_->Capacity()); in Heap() 679 malloc_space->SetFootprintLimit(malloc_space->Capacity()); in CreateMallocSpaceFromMemMap() 808 mprotect(con_space->Begin(), con_space->Capacity(), PROT_READ | PROT_WRITE); in DumpObject() 2476 madvise(main_space_->Begin(), main_space_->Capacity(), MADV_DONTNEED); in PreZygoteFork() 2531 non_moving_space_->SetFootprintLimit(non_moving_space_->Capacity()); in PreZygoteFork() 3653 malloc_space->SetFootprintLimit(malloc_space->Capacity()); in ClearGrowthLimit() 3659 main_space_backup_->SetFootprintLimit(main_space_backup_->Capacity()); in ClearGrowthLimit()
|
/art/compiler/utils/arm/ |
D | assembler_thumb2.cc | 271 if (required_capacity > buffer_.Capacity()) { in EmitJumpTables()
|