/art/runtime/ |
D | gc_root-inl.h | 31 inline MirrorType* GcRoot<MirrorType>::Read(GcRootSource* gc_root_source) const { in Read() 37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref) in GcRoot() function 41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref) in GcRoot() function 42 : GcRoot(ref.Ptr()) { } in GcRoot()
|
D | class_table-inl.h | 38 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots() 42 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots() 56 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots() 60 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots() 101 GcRoot<mirror::Class>(before_ptr).Read<kReadBarrierOption>()); in Read() 114 GcRoot<mirror::Class> root(before_ptr); in VisitRoot()
|
D | gc_root.h | 187 class GcRoot { 217 ALWAYS_INLINE GcRoot() {} in GcRoot() function 218 explicit ALWAYS_INLINE GcRoot(MirrorType* ref) 220 explicit ALWAYS_INLINE GcRoot(ObjPtr<MirrorType> ref) 247 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) in VisitRootIfNonNull() 263 void VisitRoot(GcRoot<MirrorType>& root) REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot() 294 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull() 310 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
|
D | intern_table.h | 77 std::size_t operator()(const GcRoot<mirror::String>& root) const NO_THREAD_SAFETY_ANALYSIS; 78 bool operator()(const GcRoot<mirror::String>& a, const GcRoot<mirror::String>& b) const 87 bool operator()(const GcRoot<mirror::String>& a, const Utf8String& b) const 93 void MakeEmpty(GcRoot<mirror::String>& item) const { in MakeEmpty() 94 item = GcRoot<mirror::String>(); in MakeEmpty() 96 bool IsEmpty(const GcRoot<mirror::String>& item) const { in IsEmpty() 101 using UnorderedSet = HashSet<GcRoot<mirror::String>, 105 TrackingAllocator<GcRoot<mirror::String>, kAllocatorTagInternTable>>; 333 std::vector<GcRoot<mirror::String>> new_strong_intern_roots_
|
D | reference_table.cc | 56 entries_.push_back(GcRoot<mirror::Object>(obj)); in Add() 128 bool operator()(GcRoot<mirror::Object> root1, GcRoot<mirror::Object> root2) const in Dump() 222 GcRoot<mirror::Object>& stack_for_object = it->first; in Dump() 263 for (GcRoot<mirror::Object>& root : entries) { in Dump() 275 GcRoot<mirror::Object> root; in Dump() 288 void Reset(GcRoot<mirror::Object>& _root) { in Dump() 298 for (GcRoot<mirror::Object>& root : sorted_entries) { in Dump() 358 for (GcRoot<mirror::Object>& root : entries_) { in VisitRoots()
|
D | reference_table.h | 58 typedef std::vector<GcRoot<mirror::Object>, 59 TrackingAllocator<GcRoot<mirror::Object>, kAllocatorTagReferenceTable>> Table;
|
D | intern_table.cc | 133 new_strong_intern_roots_.push_back(GcRoot<mirror::String>(s)); in InsertStrong() 281 for (GcRoot<mirror::String>& entry : weak_interns_.tables_.front().set_) { in PromoteWeakToStrong() 315 std::size_t InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& root) const { in operator ()() 324 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, in operator ()() 325 const GcRoot<mirror::String>& b) const { in operator ()() 332 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a, in operator ()() 370 for (GcRoot<mirror::String>& string : table.set_) { in WriteToMemory() 382 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Remove() 394 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Find() 421 tables_.back().set_.insert(GcRoot<mirror::String>(s)); in Insert() [all …]
|
D | indirect_reference_table-inl.h | 114 references_[serial_] = GcRoot<mirror::Object>(obj); in Add() 119 references_[serial_] = GcRoot<mirror::Object>(obj); in SetReference()
|
D | indirect_reference_table.h | 161 GcRoot<mirror::Object>* GetReference() { in GetReference() 166 const GcRoot<mirror::Object>* GetReference() const { in GetReference() 179 GcRoot<mirror::Object> references_[kIRTPrevCount]; 198 GcRoot<mirror::Object>* operator*() REQUIRES_SHARED(Locks::mutator_lock_) { in REQUIRES_SHARED()
|
D | art_method-inl.h | 70 declaring_class_ = GcRoot<mirror::Class>(new_declaring_class); in SetDeclaringClass() 75 GcRoot<mirror::Class> expected_root(expected_class); in CASDeclaringClass() 76 GcRoot<mirror::Class> desired_root(desired_class); in CASDeclaringClass() 77 auto atomic_root_class = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&declaring_class_); in CASDeclaringClass()
|
D | class_table_test.cc | 39 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull() 55 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
|
D | runtime.h | 788 GcRoot<mirror::Object> GetSentinel() REQUIRES_SHARED(Locks::mutator_lock_); 799 static void ProcessWeakClass(GcRoot<mirror::Class>* root_ptr, 1037 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_exception_; 1038 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_oome_; 1039 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_handling_stack_overflow_; 1040 GcRoot<mirror::Throwable> pre_allocated_NoClassDefFoundError_; 1049 GcRoot<mirror::Object> sentinel_;
|
/art/runtime/entrypoints/quick/ |
D | quick_dexcache_entrypoints.cc | 40 static_assert(sizeof(GcRoot<mirror::Class>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss() 41 static_assert(sizeof(GcRoot<mirror::String>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss() 43 DCHECK_ALIGNED(bss_offset, sizeof(GcRoot<mirror::Object>)); in StoreObjectInBss() 50 GcRoot<mirror::Object>* slot = reinterpret_cast<GcRoot<mirror::Object>*>( in StoreObjectInBss() 56 std::atomic<GcRoot<mirror::Object>>* atomic_slot = in StoreObjectInBss() 57 reinterpret_cast<std::atomic<GcRoot<mirror::Object>>*>(slot); in StoreObjectInBss() 59 atomic_slot->store(GcRoot<mirror::Object>(object), std::memory_order_release); in StoreObjectInBss() 91 sizeof(GcRoot<mirror::Class>)); in StoreTypeInBss() 109 sizeof(GcRoot<mirror::Class>)); in StoreStringInBss()
|
D | quick_entrypoints.h | 41 template<class MirrorType> class GcRoot; variable 113 extern "C" mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root)
|
/art/runtime/gc/ |
D | system_weak_test.cc | 79 weak_ = GcRoot<mirror::Object>(new_object); in Sweep() 84 GcRoot<mirror::Object> Get() in Get() 94 void Set(GcRoot<mirror::Object> obj) in Set() 107 GcRoot<mirror::Object> weak_ GUARDED_BY(allow_disallow_lock_); 144 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F() 165 cswh.Set(GcRoot<mirror::Object>(mirror::String::AllocFromModifiedUtf8(soa.Self(), "ABC"))); in TEST_F() 189 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F()
|
D | allocation_record.cc | 78 GcRoot<mirror::Class>& klass = record->GetClassGcRoot(); in SweepClassObject() 87 klass = GcRoot<mirror::Class>(new_object->AsClass()); in SweepClassObject() 105 it->first = GcRoot<mirror::Object>(nullptr); in SweepAllocationRecords() 114 it->first = GcRoot<mirror::Object>(new_object); in SweepAllocationRecords()
|
D | allocation_record.h | 185 GcRoot<mirror::Class>& GetClassGcRoot() REQUIRES_SHARED(Locks::mutator_lock_) { in GetClassGcRoot() 196 GcRoot<mirror::Class> klass_; 213 using EntryPair = std::pair<GcRoot<mirror::Object>, AllocRecord>; 235 entries_.push_back(EntryPair(GcRoot<mirror::Object>(obj), std::move(record))); in Put()
|
/art/runtime/mirror/ |
D | dex_cache-inl.h | 51 first_elem.object = GcRoot<T>(nullptr); in Initialize() 89 GcRoot<mirror::String>* preresolved_strings = GetPreResolvedStrings(); in GetResolvedString() 121 GetPreResolvedStrings()[string_idx.index_] = GcRoot<mirror::String>(resolved); in SetPreResolvedString() 133 GcRoot<mirror::String>*>(PreResolvedStringsOffset(), nullptr); in ClearPreResolvedStrings() 213 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in GetResolvedCallSite() 214 Atomic<GcRoot<mirror::CallSite>>& ref = in GetResolvedCallSite() 215 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target); in GetResolvedCallSite() 224 GcRoot<mirror::CallSite> null_call_site(nullptr); in SetResolvedCallSite() 225 GcRoot<mirror::CallSite> candidate(call_site); in SetResolvedCallSite() 226 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in SetResolvedCallSite() [all …]
|
D | dex_cache.h | 52 GcRoot<T> object; 211 void FixupResolvedCallSites(GcRoot<mirror::CallSite>* dest, const Visitor& visitor) 273 return alignof(GcRoot<mirror::String>); in PreResolvedStringsAlignment() 342 GcRoot<mirror::String>* GetPreResolvedStrings() ALWAYS_INLINE in GetPreResolvedStrings() 344 return GetFieldPtr64<GcRoot<mirror::String>*, kVerifyFlags>(PreResolvedStringsOffset()); in GetPreResolvedStrings() 351 void SetPreResolvedStrings(GcRoot<mirror::String>* strings) in SetPreResolvedStrings() 400 GcRoot<CallSite>* GetResolvedCallSites() in GetResolvedCallSites() 403 return GetFieldPtr<GcRoot<CallSite>*, kVerifyFlags>(ResolvedCallSitesOffset()); in GetResolvedCallSites() 406 void SetResolvedCallSites(GcRoot<CallSite>* resolved_call_sites) in SetResolvedCallSites() 469 return sizeof(GcRoot<mirror::String>) * num_strings; in PreResolvedStringsSize() [all …]
|
D | dex_cache.cc | 108 GcRoot<mirror::CallSite>* call_sites = (dex_file->NumCallSiteIds() == 0) in InitializeDexCache() 110 : reinterpret_cast<GcRoot<CallSite>*>(raw_arrays + layout.CallSitesOffset()); in InitializeDexCache() 224 GcRoot<mirror::String>* strings = in AddPreResolvedStringsArray() 225 linear_alloc->AllocArray<GcRoot<mirror::String>>(self, num_strings); in AddPreResolvedStringsArray() 253 GcRoot<CallSite>* resolved_call_sites, in Init()
|
/art/runtime/jit/ |
D | profiling_info.cc | 96 GcRoot<mirror::Class> expected_root(existing); in AddInvokeInfo() 97 GcRoot<mirror::Class> desired_root(cls); in AddInvokeInfo() 98 auto atomic_root = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&cache->classes_[i]); in AddInvokeInfo()
|
D | profiling_info.h | 52 GcRoot<mirror::Class> classes_[kIndividualCacheSize];
|
/art/openjdkjvmti/ |
D | jvmti_weak_table.h | 147 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in GetTagLocked() 199 size_t operator()(const art::GcRoot<art::mirror::Object>& r) const in operator() 206 bool operator()(const art::GcRoot<art::mirror::Object>& r1, in operator() 207 const art::GcRoot<art::mirror::Object>& r2) const in operator() 213 using TagAllocator = JvmtiAllocator<std::pair<const art::GcRoot<art::mirror::Object>, T>>; 214 std::unordered_map<art::GcRoot<art::mirror::Object>,
|
D | jvmti_weak_table-inl.h | 70 auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root, in UpdateTableWithReadBarrier() 108 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in RemoveLocked() 153 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in SetLocked() 173 auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag); in SetLocked() 201 auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED, in SweepImpl() 232 tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag); in UpdateTableWith()
|
/art/runtime/utils/ |
D | dex_cache_arrays_layout-inl.h | 74 static_assert(sizeof(GcRoot<T>) == 4U, "Unexpected GcRoot size"); in GcRootAsPointerSize() 92 return alignof(GcRoot<mirror::Class>); in TypesAlignment() 167 return alignof(GcRoot<mirror::CallSite>); in CallSitesAlignment()
|