/art/compiler/utils/ |
D | assembler.cc | 42 CHECK_EQ(Capacity(), kInitialBufferCapacity); in AssemblerBuffer() 49 allocator_->MakeInaccessible(contents_, Capacity()); in ~AssemblerBuffer() 77 size_t old_capacity = Capacity(); in ExtendCapacity() 91 CHECK_EQ(Capacity(), new_capacity); in ExtendCapacity()
|
D | assembler.h | 117 if (new_size > Capacity()) { in Resize() 213 int ComputeGap() { return buffer_->Capacity() - buffer_->Size(); } in ComputeGap() 240 size_t Capacity() const { in Capacity() function
|
/art/runtime/ |
D | indirect_reference_table_test.cc | 107 EXPECT_EQ(0U, irt.Capacity()); in TEST_F() 129 ASSERT_EQ(0U, irt.Capacity()); in TEST_F() 141 ASSERT_EQ(3U, irt.Capacity()); in TEST_F() 157 ASSERT_EQ(0U, irt.Capacity()); in TEST_F() 178 ASSERT_EQ(4U, irt.Capacity()) << "hole not filled"; in TEST_F() 186 ASSERT_EQ(3U, irt.Capacity()) << "should be 3 after two deletions"; in TEST_F() 193 ASSERT_EQ(0U, irt.Capacity()) << "not empty after split remove"; in TEST_F() 209 ASSERT_EQ(0U, irt.Capacity()) << "switching del not empty"; in TEST_F() 227 ASSERT_EQ(0U, irt.Capacity()) << "temporal del not empty"; in TEST_F() 253 ASSERT_EQ(kTableInitial + 1, irt.Capacity()); in TEST_F() [all …]
|
D | indirect_reference_table.h | 290 size_t Capacity() const { in Capacity() function 303 return IrtIterator(table_, 0, Capacity()); in begin() 307 return IrtIterator(table_, Capacity(), Capacity()); in end()
|
D | indirect_reference_table.cc | 327 for (size_t i = 0; i < Capacity(); ++i) { in AssertEmpty() 447 const size_t top_index = Capacity(); in Trim() 466 for (size_t i = 0; i < Capacity(); ++i) { in Dump()
|
/art/runtime/jni/ |
D | jni_env_ext-inl.h | 39 size_t entry_count = locals_.Capacity(); in AddLocalReference()
|
D | jni_env_ext.h | 83 return locals_.Capacity(); in GetLocalsCapacity()
|
D | java_vm_ext.cc | 750 os << "; globals=" << globals_.Capacity(); in DumpForSigQuit() 754 if (weak_globals_.Capacity() > 0) { in DumpForSigQuit() 755 os << " (plus " << weak_globals_.Capacity() << " weak)"; in DumpForSigQuit()
|
/art/runtime/gc/space/ |
D | malloc_space.cc | 159 CHECK_LE(new_end, Begin() + Capacity()); in MoreCore() 202 << "Capacity " << Capacity(); in CreateZygoteSpace() 254 << ",size=" << PrettySize(Size()) << ",capacity=" << PrettySize(Capacity()) in Dump() 280 size_t new_capacity = Capacity(); in ClampGrowthLimit()
|
D | space.h | 302 virtual size_t Capacity() const { in Capacity() function 378 return Capacity(); in NonGrowthLimitCapacity()
|
D | malloc_space.h | 101 size_t Capacity() const override { in Capacity() function
|
D | bump_pointer_space.h | 90 size_t Capacity() const override { in Capacity() function
|
D | space_test.h | 332 EXPECT_EQ(space->Capacity(), growth_limit); in SizeFootPrintGrowthLimitAndTrimDriver() 344 EXPECT_EQ(space->Capacity(), capacity); in SizeFootPrintGrowthLimitAndTrimDriver()
|
D | dlmalloc_space.cc | 174 size_t max_allowed = Capacity(); in AllocWithGrowth()
|
D | rosalloc_space.cc | 209 size_t max_allowed = Capacity(); in AllocWithGrowth()
|
D | region_space.cc | 129 accounting::ContinuousSpaceBitmap::Create("region space live bitmap", Begin(), Capacity()); in RegionSpace()
|
/art/runtime/gc/accounting/ |
D | atomic_stack.h | 187 size_t Capacity() const { in Capacity() function
|
/art/runtime/gc/collector/ |
D | semi_space.cc | 370 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in MarkStackPush() 371 ResizeMarkStack(mark_stack_->Capacity() * 2); in MarkStackPush()
|
D | mark_sweep.cc | 384 ResizeMarkStack(mark_stack_->Capacity() * 2); in ExpandMarkStack() 389 if (UNLIKELY(mark_stack_->Size() < mark_stack_->Capacity())) { in ResizeMarkStack() 410 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in MarkObjectNonNullParallel() 531 if (UNLIKELY(mark_stack_->Size() >= mark_stack_->Capacity())) { in PushOnMarkStack()
|