Home
last modified time | relevance | path

Searched refs:GcRoot (Results 1 – 25 of 67) sorted by relevance

123

/art/runtime/
Dgc_root-inl.h31 inline MirrorType* GcRoot<MirrorType>::Read(GcRootSource* gc_root_source) const { in Read()
37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref) in GcRoot() function
41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref) in GcRoot() function
42 : GcRoot(ref.Ptr()) { } in GcRoot()
Dclass_table-inl.h38 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots()
42 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots()
56 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots()
60 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots()
101 GcRoot<mirror::Class>(before_ptr).Read<kReadBarrierOption>()); in Read()
114 GcRoot<mirror::Class> root(before_ptr); in VisitRoot()
Dgc_root.h187 class GcRoot {
217 ALWAYS_INLINE GcRoot() {} in GcRoot() function
218 explicit ALWAYS_INLINE GcRoot(MirrorType* ref)
220 explicit ALWAYS_INLINE GcRoot(ObjPtr<MirrorType> ref)
247 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) in VisitRootIfNonNull()
263 void VisitRoot(GcRoot<MirrorType>& root) REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
294 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull()
310 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
Dintern_table.h77 std::size_t operator()(const GcRoot<mirror::String>& root) const NO_THREAD_SAFETY_ANALYSIS;
78 bool operator()(const GcRoot<mirror::String>& a, const GcRoot<mirror::String>& b) const
87 bool operator()(const GcRoot<mirror::String>& a, const Utf8String& b) const
93 void MakeEmpty(GcRoot<mirror::String>& item) const { in MakeEmpty()
94 item = GcRoot<mirror::String>(); in MakeEmpty()
96 bool IsEmpty(const GcRoot<mirror::String>& item) const { in IsEmpty()
101 using UnorderedSet = HashSet<GcRoot<mirror::String>,
105 TrackingAllocator<GcRoot<mirror::String>, kAllocatorTagInternTable>>;
333 std::vector<GcRoot<mirror::String>> new_strong_intern_roots_
Dreference_table.cc56 entries_.push_back(GcRoot<mirror::Object>(obj)); in Add()
128 bool operator()(GcRoot<mirror::Object> root1, GcRoot<mirror::Object> root2) const in Dump()
222 GcRoot<mirror::Object>& stack_for_object = it->first; in Dump()
263 for (GcRoot<mirror::Object>& root : entries) { in Dump()
275 GcRoot<mirror::Object> root; in Dump()
288 void Reset(GcRoot<mirror::Object>& _root) { in Dump()
298 for (GcRoot<mirror::Object>& root : sorted_entries) { in Dump()
358 for (GcRoot<mirror::Object>& root : entries_) { in VisitRoots()
Dreference_table.h58 typedef std::vector<GcRoot<mirror::Object>,
59 TrackingAllocator<GcRoot<mirror::Object>, kAllocatorTagReferenceTable>> Table;
Dintern_table.cc133 new_strong_intern_roots_.push_back(GcRoot<mirror::String>(s)); in InsertStrong()
281 for (GcRoot<mirror::String>& entry : weak_interns_.tables_.front().set_) { in PromoteWeakToStrong()
315 std::size_t InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& root) const { in operator ()()
324 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, in operator ()()
325 const GcRoot<mirror::String>& b) const { in operator ()()
332 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, in operator ()()
370 for (GcRoot<mirror::String>& string : table.set_) { in WriteToMemory()
382 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Remove()
394 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Find()
421 tables_.back().set_.insert(GcRoot<mirror::String>(s)); in Insert()
[all …]
Dindirect_reference_table-inl.h114 references_[serial_] = GcRoot<mirror::Object>(obj); in Add()
119 references_[serial_] = GcRoot<mirror::Object>(obj); in SetReference()
Dindirect_reference_table.h161 GcRoot<mirror::Object>* GetReference() { in GetReference()
166 const GcRoot<mirror::Object>* GetReference() const { in GetReference()
179 GcRoot<mirror::Object> references_[kIRTPrevCount];
198 GcRoot<mirror::Object>* operator*() REQUIRES_SHARED(Locks::mutator_lock_) { in REQUIRES_SHARED()
Dart_method-inl.h70 declaring_class_ = GcRoot<mirror::Class>(new_declaring_class); in SetDeclaringClass()
75 GcRoot<mirror::Class> expected_root(expected_class); in CASDeclaringClass()
76 GcRoot<mirror::Class> desired_root(desired_class); in CASDeclaringClass()
77 auto atomic_root_class = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&declaring_class_); in CASDeclaringClass()
Dclass_table_test.cc39 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull()
55 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
Druntime.h788 GcRoot<mirror::Object> GetSentinel() REQUIRES_SHARED(Locks::mutator_lock_);
799 static void ProcessWeakClass(GcRoot<mirror::Class>* root_ptr,
1037 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_exception_;
1038 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_oome_;
1039 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_handling_stack_overflow_;
1040 GcRoot<mirror::Throwable> pre_allocated_NoClassDefFoundError_;
1049 GcRoot<mirror::Object> sentinel_;
/art/runtime/entrypoints/quick/
Dquick_dexcache_entrypoints.cc40 static_assert(sizeof(GcRoot<mirror::Class>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss()
41 static_assert(sizeof(GcRoot<mirror::String>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss()
43 DCHECK_ALIGNED(bss_offset, sizeof(GcRoot<mirror::Object>)); in StoreObjectInBss()
50 GcRoot<mirror::Object>* slot = reinterpret_cast<GcRoot<mirror::Object>*>( in StoreObjectInBss()
56 std::atomic<GcRoot<mirror::Object>>* atomic_slot = in StoreObjectInBss()
57 reinterpret_cast<std::atomic<GcRoot<mirror::Object>>*>(slot); in StoreObjectInBss()
59 atomic_slot->store(GcRoot<mirror::Object>(object), std::memory_order_release); in StoreObjectInBss()
91 sizeof(GcRoot<mirror::Class>)); in StoreTypeInBss()
109 sizeof(GcRoot<mirror::Class>)); in StoreStringInBss()
Dquick_entrypoints.h41 template<class MirrorType> class GcRoot; variable
113 extern "C" mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root)
/art/runtime/gc/
Dsystem_weak_test.cc79 weak_ = GcRoot<mirror::Object>(new_object); in Sweep()
84 GcRoot<mirror::Object> Get() in Get()
94 void Set(GcRoot<mirror::Object> obj) in Set()
107 GcRoot<mirror::Object> weak_ GUARDED_BY(allow_disallow_lock_);
144 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F()
165 cswh.Set(GcRoot<mirror::Object>(mirror::String::AllocFromModifiedUtf8(soa.Self(), "ABC"))); in TEST_F()
189 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F()
Dallocation_record.cc78 GcRoot<mirror::Class>& klass = record->GetClassGcRoot(); in SweepClassObject()
87 klass = GcRoot<mirror::Class>(new_object->AsClass()); in SweepClassObject()
105 it->first = GcRoot<mirror::Object>(nullptr); in SweepAllocationRecords()
114 it->first = GcRoot<mirror::Object>(new_object); in SweepAllocationRecords()
Dallocation_record.h185 GcRoot<mirror::Class>& GetClassGcRoot() REQUIRES_SHARED(Locks::mutator_lock_) { in GetClassGcRoot()
196 GcRoot<mirror::Class> klass_;
213 using EntryPair = std::pair<GcRoot<mirror::Object>, AllocRecord>;
235 entries_.push_back(EntryPair(GcRoot<mirror::Object>(obj), std::move(record))); in Put()
/art/runtime/mirror/
Ddex_cache-inl.h51 first_elem.object = GcRoot<T>(nullptr); in Initialize()
89 GcRoot<mirror::String>* preresolved_strings = GetPreResolvedStrings(); in GetResolvedString()
121 GetPreResolvedStrings()[string_idx.index_] = GcRoot<mirror::String>(resolved); in SetPreResolvedString()
133 GcRoot<mirror::String>*>(PreResolvedStringsOffset(), nullptr); in ClearPreResolvedStrings()
213 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in GetResolvedCallSite()
214 Atomic<GcRoot<mirror::CallSite>>& ref = in GetResolvedCallSite()
215 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target); in GetResolvedCallSite()
224 GcRoot<mirror::CallSite> null_call_site(nullptr); in SetResolvedCallSite()
225 GcRoot<mirror::CallSite> candidate(call_site); in SetResolvedCallSite()
226 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in SetResolvedCallSite()
[all …]
Ddex_cache.h52 GcRoot<T> object;
211 void FixupResolvedCallSites(GcRoot<mirror::CallSite>* dest, const Visitor& visitor)
273 return alignof(GcRoot<mirror::String>); in PreResolvedStringsAlignment()
342 GcRoot<mirror::String>* GetPreResolvedStrings() ALWAYS_INLINE in GetPreResolvedStrings()
344 return GetFieldPtr64<GcRoot<mirror::String>*, kVerifyFlags>(PreResolvedStringsOffset()); in GetPreResolvedStrings()
351 void SetPreResolvedStrings(GcRoot<mirror::String>* strings) in SetPreResolvedStrings()
400 GcRoot<CallSite>* GetResolvedCallSites() in GetResolvedCallSites()
403 return GetFieldPtr<GcRoot<CallSite>*, kVerifyFlags>(ResolvedCallSitesOffset()); in GetResolvedCallSites()
406 void SetResolvedCallSites(GcRoot<CallSite>* resolved_call_sites) in SetResolvedCallSites()
469 return sizeof(GcRoot<mirror::String>) * num_strings; in PreResolvedStringsSize()
[all …]
Ddex_cache.cc108 GcRoot<mirror::CallSite>* call_sites = (dex_file->NumCallSiteIds() == 0) in InitializeDexCache()
110 : reinterpret_cast<GcRoot<CallSite>*>(raw_arrays + layout.CallSitesOffset()); in InitializeDexCache()
224 GcRoot<mirror::String>* strings = in AddPreResolvedStringsArray()
225 linear_alloc->AllocArray<GcRoot<mirror::String>>(self, num_strings); in AddPreResolvedStringsArray()
253 GcRoot<CallSite>* resolved_call_sites, in Init()
/art/runtime/jit/
Dprofiling_info.cc96 GcRoot<mirror::Class> expected_root(existing); in AddInvokeInfo()
97 GcRoot<mirror::Class> desired_root(cls); in AddInvokeInfo()
98 auto atomic_root = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&cache->classes_[i]); in AddInvokeInfo()
Dprofiling_info.h52 GcRoot<mirror::Class> classes_[kIndividualCacheSize];
/art/openjdkjvmti/
Djvmti_weak_table.h147 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in GetTagLocked()
199 size_t operator()(const art::GcRoot<art::mirror::Object>& r) const in operator()
206 bool operator()(const art::GcRoot<art::mirror::Object>& r1, in operator()
207 const art::GcRoot<art::mirror::Object>& r2) const in operator()
213 using TagAllocator = JvmtiAllocator<std::pair<const art::GcRoot<art::mirror::Object>, T>>;
214 std::unordered_map<art::GcRoot<art::mirror::Object>,
Djvmti_weak_table-inl.h70 auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root, in UpdateTableWithReadBarrier()
108 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in RemoveLocked()
153 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in SetLocked()
173 auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag); in SetLocked()
201 auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED, in SweepImpl()
232 tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag); in UpdateTableWith()
/art/runtime/utils/
Ddex_cache_arrays_layout-inl.h74 static_assert(sizeof(GcRoot<T>) == 4U, "Unexpected GcRoot size"); in GcRootAsPointerSize()
92 return alignof(GcRoot<mirror::Class>); in TypesAlignment()
167 return alignof(GcRoot<mirror::CallSite>); in CallSitesAlignment()

123