Home
last modified time | relevance | path

Searched refs:region (Results 1 – 25 of 26) sorted by relevance

12

/art/libartbase/base/
Dmemory_region_test.cc26 MemoryRegion region(&data, n); in TEST() local
28 ASSERT_EQ(0, region.LoadUnaligned<char>(0)); in TEST()
33 region.LoadUnaligned<uint32_t>(1)); in TEST()
34 ASSERT_EQ(5 + (6 << kBitsPerByte), region.LoadUnaligned<int16_t>(5)); in TEST()
35 ASSERT_EQ(7u, region.LoadUnaligned<unsigned char>(7)); in TEST()
41 MemoryRegion region(&data, n); in TEST() local
43 region.StoreUnaligned<unsigned char>(0u, 7); in TEST()
44 region.StoreUnaligned<int16_t>(1, 6 + (5 << kBitsPerByte)); in TEST()
45 region.StoreUnaligned<uint32_t>(3, in TEST()
50 region.StoreUnaligned<char>(7, 0); in TEST()
Dbit_memory_region.h46 ALWAYS_INLINE explicit BitMemoryRegion(MemoryRegion region) in BitMemoryRegion() argument
47 : BitMemoryRegion(region.begin(), /* bit_start */ 0, region.size_in_bits()) { in BitMemoryRegion()
49 ALWAYS_INLINE BitMemoryRegion(MemoryRegion region, size_t bit_offset, size_t bit_length) in BitMemoryRegion() argument
50 : BitMemoryRegion(region) { in BitMemoryRegion()
307 BitMemoryRegion region(out_->data(), bit_offset_, bit_length); in Allocate()
310 return region; in Allocate()
313 ALWAYS_INLINE void WriteRegion(const BitMemoryRegion& region) { in WriteRegion() argument
314 Allocate(region.size_in_bits()).StoreBits(/* bit_offset */ 0, region, region.size_in_bits()); in WriteRegion()
Dbit_table.h415 MemoryRegion region(const_cast<void*>(bitmap), BitsToBytesRoundUp(num_bits)); in Dedup()
416 DCHECK(num_bits == 0 || BitMemoryRegion(region).LoadBit(num_bits - 1) == 1); in Dedup()
417 DCHECK_EQ(BitMemoryRegion(region).LoadBits(num_bits, region.size_in_bits() - num_bits), 0u); in Dedup()
419 uint32_t hash = hasher(region); in Dedup()
424 if (MemoryRegion::ContentEquals()(region, rows_[it->second])) { in Dedup()
431 void* copy = allocator_->Alloc(region.size(), kArenaAllocBitTableBuilder); in Dedup()
432 memcpy(copy, region.pointer(), region.size()); in Dedup()
433 rows_.push_back(MemoryRegion(copy, region.size())); in Dedup()
Dmemory_region.h129 ALWAYS_INLINE void Extend(const MemoryRegion& region, uintptr_t extra) { in Extend() argument
130 pointer_ = region.pointer(); in Extend()
131 size_ = (region.size() + extra); in Extend()
/art/runtime/jit/
Djit_scoped_code_cache_write.h38 explicit ScopedCodeCacheWrite(const JitMemoryRegion& region) in ScopedCodeCacheWrite() argument
40 region_(region) { in ScopedCodeCacheWrite()
42 const MemMap* const updatable_pages = region.GetUpdatableCodeMapping(); in ScopedCodeCacheWrite()
44 int prot = region.HasDualCodeMapping() ? kProtRW : kProtRWX; in ScopedCodeCacheWrite()
Djit_code_cache.h115 explicit ZygoteMap(JitMemoryRegion* region) in ZygoteMap() argument
116 : map_(), region_(region), compilation_state_(nullptr) {} in ZygoteMap()
201 JitMemoryRegion* region)
244 JitMemoryRegion* region,
262 JitMemoryRegion* region,
276 void Free(Thread* self, JitMemoryRegion* region, const uint8_t* code, const uint8_t* data)
394 bool IsSharedRegion(const JitMemoryRegion& region) const { return &region == &shared_region_; } in IsSharedRegion() argument
399 JitMemoryRegion* region = GetCurrentRegion(); in CanAllocateProfilingInfo() local
400 return region->IsValid() && !IsSharedRegion(*region); in CanAllocateProfilingInfo()
Djit_code_cache.cc232 JitMemoryRegion region; in Create() local
233 if (!region.Initialize(initial_capacity, in Create()
245 jit_code_cache->shared_region_ = std::move(region); in Create()
247 jit_code_cache->private_region_ = std::move(region); in Create()
638 JitMemoryRegion* region, in Commit() argument
653 DCheckRootsAreValid(roots, IsSharedRegion(*region)); in Commit()
664 const uint8_t* code_ptr = region->CommitCode( in Commit()
672 if (!region->CommitData(reserved_data, roots, stack_map)) { in Commit()
719 if (method->IsPreCompiled() && IsSharedRegion(*region)) { in Commit()
734 if (!IsSharedRegion(*region)) { in Commit()
[all …]
Djit.h195 Thread* self, JitMemoryRegion* region, ArtMethod* method, bool baseline, bool osr)
/art/test/1000-non-moving-space-stress/
Dinfo.txt4 reference to an object cleared or moved from a newly allocated region
5 of the region space.
/art/compiler/utils/
Dassembler.cc54 void AssemblerBuffer::ProcessFixups(const MemoryRegion& region) { in ProcessFixups() argument
57 fixup->Process(region, fixup->position()); in ProcessFixups()
Dassembler.h50 virtual void Process(const MemoryRegion& region, int position) = 0;
168 void FinalizeInstructions(const MemoryRegion& region);
272 void ProcessFixups(const MemoryRegion& region);
377 virtual void FinalizeInstructions(const MemoryRegion& region) { in FinalizeInstructions() argument
378 buffer_.FinalizeInstructions(region); in FinalizeInstructions()
Djni_macro_assembler.h61 virtual void FinalizeInstructions(const MemoryRegion& region) = 0;
273 void FinalizeInstructions(const MemoryRegion& region) override { in FinalizeInstructions() argument
274 asm_.FinalizeInstructions(region); in FinalizeInstructions()
/art/test/659-unpadded-array/
Dinfo.txt1 Regression test for the concurrent GC whose region space had
/art/runtime/gc/space/
Dregion_space-inl.h331 mirror::Object* region = nullptr; in AllocLarge() local
336 region = AllocLargeInRange<kForEvac>(cyclic_alloc_region_index_, in AllocLarge()
344 if (region == nullptr) { in AllocLarge()
348 region = AllocLargeInRange<kForEvac>( in AllocLarge()
358 if (region != nullptr) { in AllocLarge()
367 region = AllocLargeInRange<kForEvac>(0, in AllocLarge()
374 if (kForEvac && region != nullptr) { in AllocLarge()
377 return region; in AllocLarge()
Dregion_space.cc312 Region* region = RefToRegionLocked(reinterpret_cast<mirror::Object*>(addr)); in ZeroLiveBytesForLargeObject() local
314 DCHECK(region->IsLarge()); in ZeroLiveBytesForLargeObject()
316 DCHECK(region->IsLargeTail()); in ZeroLiveBytesForLargeObject()
318 region->ZeroLiveBytes(); in ZeroLiveBytesForLargeObject()
/art/test/090-loop-formation/
Dinfo.txt3 in the loop region, and the JIT compiler won't choke on unresolved fields.
/art/compiler/jit/
Djit_compiler.h43 Thread* self, JitMemoryRegion* region, ArtMethod* method, bool baseline, bool osr)
Djit_compiler.cc166 Thread* self, JitMemoryRegion* region, ArtMethod* method, bool baseline, bool osr) { in CompileMethod() argument
186 self, code_cache, region, method, baseline, osr, jit_logger_.get()); in CompileMethod()
/art/runtime/
Dstack_map.cc100 CodeInfo code_info(code_info_data, nullptr, [&](size_t i, auto*, BitMemoryRegion region) { in Dedupe() argument
101 it[i] = dedupe_map_.emplace(region, /*bit_offset=*/0).first; in Dedupe()
102 if (it[i]->second != 0 && region.size_in_bits() > kMinDedupSize) { // Seen before and large? in Dedupe()
232 CodeInfo code_info(code_info_data, &num_bits, [&](size_t i, auto* table, BitMemoryRegion region) { in CollectSizeStats() argument
235 table_stats->AddBits(region.size_in_bits()); in CollectSizeStats()
/art/compiler/
Dcompiler.h76 jit::JitMemoryRegion* region ATTRIBUTE_UNUSED, in JitCompile()
/art/compiler/utils/arm64/
Dassembler_arm64.cc79 void Arm64Assembler::FinalizeInstructions(const MemoryRegion& region) { in FinalizeInstructions() argument
82 region.CopyFrom(0, from); in FinalizeInstructions()
Dassembler_arm64.h94 void FinalizeInstructions(const MemoryRegion& region) override;
/art/compiler/optimizing/
Doptimizing_compiler.cc298 jit::JitMemoryRegion* region,
1216 jit::JitMemoryRegion* region, in JitCompile() argument
1252 region, in JitCompile()
1287 region, in JitCompile()
1297 code_cache->Free(self, region, reserved_code.data(), reserved_data.data()); in JitCompile()
1337 /* is_shared_jit_code= */ code_cache->IsSharedRegion(*region), in JitCompile()
1349 region, in JitCompile()
1395 region, in JitCompile()
1405 code_cache->Free(self, region, reserved_code.data(), reserved_data.data()); in JitCompile()
/art/compiler/utils/arm/
Dassembler_arm_vixl.cc55 void ArmVIXLAssembler::FinalizeInstructions(const MemoryRegion& region) { in FinalizeInstructions() argument
58 region.CopyFrom(0, from); in FinalizeInstructions()
Dassembler_arm_vixl.h180 void FinalizeInstructions(const MemoryRegion& region) override;

12