/art/runtime/utils/ |
D | dex_cache_arrays_layout-inl.h | 38 RoundUp(types_offset_ + TypesSize(header.type_ids_size_), MethodsAlignment())), in DexCacheArraysLayout() 40 RoundUp(methods_offset_ + MethodsSize(header.method_ids_size_), StringsAlignment())), in DexCacheArraysLayout() 42 RoundUp(strings_offset_ + StringsSize(header.string_ids_size_), FieldsAlignment())), in DexCacheArraysLayout() 44 RoundUp(fields_offset_ + FieldsSize(header.field_ids_size_), MethodTypesAlignment())), in DexCacheArraysLayout() 46 RoundUp(method_types_offset_ + MethodTypesSize(header.proto_ids_size_), in DexCacheArraysLayout() 48 size_(RoundUp(call_sites_offset_ + CallSitesSize(num_call_sites), Alignment())) { in DexCacheArraysLayout()
|
/art/runtime/ |
D | image.cc | 66 CHECK_EQ(image_begin, RoundUp(image_begin, kPageSize)); in ImageHeader() 67 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kPageSize)); in ImageHeader() 68 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kPageSize)); in ImageHeader() 104 return image_reservation_size_ == RoundUp(image_size_, kPageSize); in IsAppImage() 160 static const size_t kStartPos = RoundUp(sizeof(ImageHeader), kObjectAlignment); in VisitObjects() 164 pos += RoundUp(object->SizeOf(), kObjectAlignment); in VisitObjects()
|
D | nterp_helpers.cc | 107 return RoundUp(frame_size, kStackAlignment); in NterpGetFrameSize()
|
D | backtrace_helper.cc | 108 if (RoundUp(it->pc, align) == reinterpret_cast<uintptr_t>(GetQuickInstrumentationExitPc())) { in CollectImpl()
|
/art/runtime/gc/space/ |
D | malloc_space.cc | 105 *growth_limit = RoundUp(*growth_limit, kPageSize); in CreateMemMap() 106 *capacity = RoundUp(*capacity, kPageSize); in CreateMemMap() 143 growth_limit = RoundUp(growth_limit, kPageSize); in SetGrowthLimit() 186 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), kPageSize))); in CreateZygoteSpace() 191 size_t size = RoundUp(Size(), kPageSize); in CreateZygoteSpace() 203 SetGrowthLimit(RoundUp(size, kPageSize)); in CreateZygoteSpace()
|
D | bump_pointer_space-inl.h | 31 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc() 48 num_bytes = RoundUp(num_bytes, kAlignment); in AllocThreadUnsafe()
|
D | bump_pointer_space.cc | 28 capacity = RoundUp(capacity, kPageSize); in Create() 102 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject() 150 bytes = RoundUp(bytes, kAlignment); in AllocBlock() 234 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
|
D | region_space.cc | 274 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); in ShouldBeEvacuated() 304 size_t obj_alloc_size = RoundUp(obj_size, space::RegionSpace::kRegionSize); in ZeroLiveBytesForLargeObject() 384 num_expected_large_tails = RoundUp(r->BytesAllocated(), kRegionSize) / kRegionSize - 1; in SetFromSpace() 657 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in CheckLiveBytesAgainstRegionBitmap() 954 << (static_cast<float>(live_bytes_) / RoundUp(BytesAllocated(), kRegionSize)); in Dump() 981 prev_object_end = RoundUp(object_end, kAlignment); in GetLongestConsecutiveFreeBytes() 994 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual() 997 *usable_size = RoundUp(num_bytes, kRegionSize); in AllocationSizeNonvirtual()
|
/art/compiler/utils/ |
D | swap_space.cc | 110 size = RoundUp(size, 8U); in Alloc() 149 size_t next_part = std::max(RoundUp(min_size, kPageSize), RoundUp(kMininumMapSize, kPageSize)); in NewFileChunk() 177 size = RoundUp(size, 8U); in Free()
|
/art/runtime/gc/accounting/ |
D | bitmap.cc | 48 const size_t bitmap_size = RoundUp( in AllocateMemMap() 49 RoundUp(num_bits, kBitsPerBitmapWord) / kBitsPerBitmapWord * sizeof(uintptr_t), kPageSize); in AllocateMemMap()
|
/art/dex2oat/linker/arm/ |
D | relative_patcher_thumb2_test.cc | 728 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmAlignment); in TestBakerFieldWide() 794 thunk_offset += RoundUp(expected_thunk.size(), kArmAlignment); in TestBakerFieldWide() 826 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmAlignment); in TestBakerFieldNarrow() 895 thunk_offset += RoundUp(expected_thunk.size(), kArmAlignment); in TestBakerFieldNarrow() 950 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() 971 - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmAlignment) in TEST_F() 972 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmAlignment) in TEST_F() 1018 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() 1025 const uint32_t bne = BneWWithOffset(kLiteralOffset1, RoundUp(raw_code1.size(), kArmAlignment)); in TEST_F() 1048 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() [all …]
|
/art/runtime/arch/x86/ |
D | jni_frame_x86.h | 61 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeOutArgsSize()
|
/art/runtime/arch/arm/ |
D | jni_frame_arm.h | 65 return RoundUp(size, kAapcsStackAlignment); in GetCriticalNativeOutArgsSize()
|
D | quick_entrypoints_cc_arm.cc | 50 fpr_double_index = std::max(fpr_double_index, RoundUp(fpr_index, 2)); in quick_invoke_reg_setup()
|
/art/runtime/mirror/ |
D | array.h | 81 size_t data_offset = RoundUp(OFFSETOF_MEMBER(Array, first_element_), component_size); in DataOffset() 82 DCHECK_EQ(RoundUp(data_offset, component_size), data_offset) in DataOffset() 89 constexpr size_t data_offset = RoundUp(kFirstElementOffset, kComponentSize); in DataOffset() 90 static_assert(RoundUp(data_offset, kComponentSize) == data_offset, "RoundUp fail"); in DataOffset()
|
/art/runtime/arch/arm64/ |
D | jni_frame_arm64.h | 72 return RoundUp(size, kAapcs64StackAlignment); in GetCriticalNativeOutArgsSize()
|
/art/runtime/arch/x86_64/ |
D | jni_frame_x86_64.h | 78 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeOutArgsSize()
|
/art/dex2oat/linker/arm64/ |
D | relative_patcher_arm64_test.cc | 1068 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArm64Alignment); in TestBakerField() 1121 thunk_offset += RoundUp(expected_thunk.size(), kArm64Alignment); in TestBakerField() 1158 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1173 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1174 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1218 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1225 const uint32_t cbnz_offset = RoundUp(raw_code1.size(), kArm64Alignment) - kLiteralOffset1; in TEST_F() 1247 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1262 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1263 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() [all …]
|
/art/libartbase/base/ |
D | scoped_arena_allocator.cc | 96 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); in AllocWithMemoryTool() 150 arena_stack_->top_ptr_ = mark_ptr_ + RoundUp(sizeof(ScopedArenaAllocator), 8); in Reset()
|
D | arena_allocator.h | 271 bytes = RoundUp(bytes, kAlignment); 311 const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment); 317 const size_t aligned_new_size = RoundUp(new_size, kAlignment);
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 550 return RoundUp(size, kThreadLocalBracketQuantumSize); in RoundToBracketSize() 552 return RoundUp(size, kBracketQuantumSize); in RoundToBracketSize() 564 return RoundUp(size, kThreadLocalBracketQuantumSize) / kThreadLocalBracketQuantumSize - 1; in SizeToIndex() 566 return (RoundUp(size, kBracketQuantumSize) - kMaxThreadLocalBracketSize) / kBracketQuantumSize in SizeToIndex() 581 bracket_size = RoundUp(size, kThreadLocalBracketQuantumSize); in SizeToIndexAndBracketSize() 584 bracket_size = RoundUp(size, kBracketQuantumSize); in SizeToIndexAndBracketSize() 875 return RoundUp(bytes, kPageSize); in UsableSize()
|
D | dlmalloc.cc | 72 start = reinterpret_cast<void*>(art::RoundUp(reinterpret_cast<uintptr_t>(start), art::kPageSize)); in DlmallocMadviseCallback()
|
/art/compiler/jni/quick/arm/ |
D | calling_convention_arm.cc | 240 fpr_double_index = (std::max(fpr_double_index * 2, RoundUp(fpr_index, 2))) / 2; in EntrySpills() 427 return RoundUp(total_size, kStackAlignment); in FrameSize() 448 size_t out_args_size = RoundUp(size, kAapcsStackAlignment); in OutArgSize()
|
/art/runtime/gc/ |
D | heap-inl.h | 108 byte_count = RoundUp(byte_count, space::BumpPointerSpace::kAlignment); in AllocObjectWithAllocator() 300 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate() 380 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment); in TryToAllocate()
|
/art/compiler/jni/quick/x86/ |
D | calling_convention_x86.cc | 256 return RoundUp(total_size, kStackAlignment); in FrameSize() 288 size_t out_args_size = RoundUp(size, kNativeStackAlignment); in OutArgSize()
|