| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_MIRROR_DEX_CACHE_H_ |
| #define ART_RUNTIME_MIRROR_DEX_CACHE_H_ |
| |
| #include "array.h" |
| #include "base/array_ref.h" |
| #include "base/atomic_pair.h" |
| #include "base/bit_utils.h" |
| #include "base/locks.h" |
| #include "base/macros.h" |
| #include "dex/dex_file.h" |
| #include "dex/dex_file_types.h" |
| #include "gc_root.h" // Note: must not use -inl here to avoid circular dependency. |
| #include "linear_alloc.h" |
| #include "object.h" |
| #include "object_array.h" |
| |
| namespace art HIDDEN { |
| |
| namespace linker { |
| class ImageWriter; |
| } // namespace linker |
| |
| class ArtField; |
| class ArtMethod; |
| struct DexCacheOffsets; |
| class DexFile; |
| union JValue; |
| class ReflectiveValueVisitor; |
| class Thread; |
| |
| namespace mirror { |
| |
| class CallSite; |
| class Class; |
| class ClassLoader; |
| class DexCache; |
| class MethodType; |
| class String; |
| |
| template <typename T> struct PACKED(8) DexCachePair { |
| GcRoot<T> object; |
| uint32_t index; |
| // The array is initially [ {0,0}, {0,0}, {0,0} ... ] |
| // We maintain the invariant that once a dex cache entry is populated, |
| // the pointer is always non-0 |
| // Any given entry would thus be: |
| // {non-0, non-0} OR {0,0} |
| // |
| // It's generally sufficiently enough then to check if the |
| // lookup index matches the stored index (for a >0 lookup index) |
| // because if it's true the pointer is also non-null. |
| // |
| // For the 0th entry which is a special case, the value is either |
| // {0,0} (initial state) or {non-0, 0} which indicates |
| // that a valid object is stored at that index for a dex section id of 0. |
| // |
| // As an optimization, we want to avoid branching on the object pointer since |
| // it's always non-null if the id branch succeeds (except for the 0th id). |
| // Set the initial state for the 0th entry to be {0,1} which is guaranteed to fail |
| // the lookup id == stored id branch. |
| DexCachePair(ObjPtr<T> object, uint32_t index); |
| DexCachePair() : index(0) {} |
| DexCachePair(const DexCachePair<T>&) = default; |
| DexCachePair& operator=(const DexCachePair<T>&) = default; |
| |
| static void Initialize(std::atomic<DexCachePair<T>>* dex_cache); |
| |
| static uint32_t InvalidIndexForSlot(uint32_t slot) { |
| // Since the cache size is a power of two, 0 will always map to slot 0. |
| // Use 1 for slot 0 and 0 for all other slots. |
| return (slot == 0) ? 1u : 0u; |
| } |
| |
| T* GetObjectForIndex(uint32_t idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| }; |
| |
| template <typename T> struct PACKED(2 * __SIZEOF_POINTER__) NativeDexCachePair { |
| T* object; |
| size_t index; |
| // This is similar to DexCachePair except that we're storing a native pointer |
| // instead of a GC root. See DexCachePair for the details. |
| NativeDexCachePair(T* object, uint32_t index) |
| : object(object), |
| index(index) {} |
| NativeDexCachePair() : object(nullptr), index(0u) { } |
| NativeDexCachePair(const NativeDexCachePair<T>&) = default; |
| NativeDexCachePair& operator=(const NativeDexCachePair<T>&) = default; |
| |
| static void Initialize(std::atomic<NativeDexCachePair<T>>* dex_cache); |
| |
| static uint32_t InvalidIndexForSlot(uint32_t slot) { |
| // Since the cache size is a power of two, 0 will always map to slot 0. |
| // Use 1 for slot 0 and 0 for all other slots. |
| return (slot == 0) ? 1u : 0u; |
| } |
| |
| T* GetObjectForIndex(uint32_t idx) REQUIRES_SHARED(Locks::mutator_lock_) { |
| if (idx != index) { |
| return nullptr; |
| } |
| DCHECK(object != nullptr); |
| return object; |
| } |
| }; |
| |
| template <typename T, size_t size> class NativeDexCachePairArray { |
| public: |
| NativeDexCachePairArray() {} |
| |
| T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { |
| auto pair = GetNativePair(entries_, SlotIndex(index)); |
| return pair.GetObjectForIndex(index); |
| } |
| |
| void Set(uint32_t index, T* value) { |
| NativeDexCachePair<T> pair(value, index); |
| SetNativePair(entries_, SlotIndex(index), pair); |
| } |
| |
| NativeDexCachePair<T> GetNativePair(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { |
| return GetNativePair(entries_, SlotIndex(index)); |
| } |
| |
| void SetNativePair(uint32_t index, NativeDexCachePair<T> value) { |
| SetNativePair(entries_, SlotIndex(index), value); |
| } |
| |
| private: |
| NativeDexCachePair<T> GetNativePair(std::atomic<NativeDexCachePair<T>>* pair_array, size_t idx) { |
| auto* array = reinterpret_cast<AtomicPair<uintptr_t>*>(pair_array); |
| AtomicPair<uintptr_t> value = AtomicPairLoadAcquire(&array[idx]); |
| return NativeDexCachePair<T>(reinterpret_cast<T*>(value.val), value.key); |
| } |
| |
| void SetNativePair(std::atomic<NativeDexCachePair<T>>* pair_array, |
| size_t idx, |
| NativeDexCachePair<T> pair) { |
| auto* array = reinterpret_cast<AtomicPair<uintptr_t>*>(pair_array); |
| AtomicPair<uintptr_t> v(pair.index, reinterpret_cast<size_t>(pair.object)); |
| AtomicPairStoreRelease(&array[idx], v); |
| } |
| |
| uint32_t SlotIndex(uint32_t index) { |
| return index % size; |
| } |
| |
| std::atomic<NativeDexCachePair<T>> entries_[0]; |
| |
| NativeDexCachePairArray(const NativeDexCachePairArray<T, size>&) = delete; |
| NativeDexCachePairArray& operator=(const NativeDexCachePairArray<T, size>&) = delete; |
| }; |
| |
| template <typename T, size_t size> class DexCachePairArray { |
| public: |
| DexCachePairArray() {} |
| |
| T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { |
| return GetPair(index).GetObjectForIndex(index); |
| } |
| |
| void Set(uint32_t index, T* value) REQUIRES_SHARED(Locks::mutator_lock_) { |
| SetPair(index, DexCachePair<T>(value, index)); |
| } |
| |
| DexCachePair<T> GetPair(uint32_t index) { |
| return entries_[SlotIndex(index)].load(std::memory_order_relaxed); |
| } |
| |
| void SetPair(uint32_t index, DexCachePair<T> value) { |
| entries_[SlotIndex(index)].store(value, std::memory_order_relaxed); |
| } |
| |
| void Clear(uint32_t index) { |
| uint32_t slot = SlotIndex(index); |
| // This is racy but should only be called from the transactional interpreter. |
| if (entries_[slot].load(std::memory_order_relaxed).index == index) { |
| DexCachePair<T> cleared(nullptr, DexCachePair<T>::InvalidIndexForSlot(slot)); |
| entries_[slot].store(cleared, std::memory_order_relaxed); |
| } |
| } |
| |
| private: |
| uint32_t SlotIndex(uint32_t index) { |
| return index % size; |
| } |
| |
| std::atomic<DexCachePair<T>> entries_[0]; |
| |
| DexCachePairArray(const DexCachePairArray<T, size>&) = delete; |
| DexCachePairArray& operator=(const DexCachePairArray<T, size>&) = delete; |
| }; |
| |
| template <typename T> class GcRootArray { |
| public: |
| GcRootArray() {} |
| |
| T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| Atomic<GcRoot<T>>* GetGcRoot(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { |
| return &entries_[index]; |
| } |
| |
| // Only to be used in locations that don't need the atomic or will later load |
| // and read atomically. |
| GcRoot<T>* GetGcRootAddress(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { |
| static_assert(sizeof(GcRoot<T>) == sizeof(Atomic<GcRoot<T>>)); |
| return reinterpret_cast<GcRoot<T>*>(&entries_[index]); |
| } |
| |
| void Set(uint32_t index, T* value) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| private: |
| Atomic<GcRoot<T>> entries_[0]; |
| }; |
| |
| template <typename T> class NativeArray { |
| public: |
| NativeArray() {} |
| |
| T* Get(uint32_t index) { |
| return entries_[index].load(std::memory_order_relaxed); |
| } |
| |
| T** GetPtrEntryPtrSize(uint32_t index, PointerSize ptr_size) { |
| if (ptr_size == PointerSize::k64) { |
| return reinterpret_cast<T**>(reinterpret_cast<uint64_t*>(entries_) + index); |
| } else { |
| return reinterpret_cast<T**>(reinterpret_cast<uint32_t*>(entries_) + index); |
| } |
| } |
| |
| void Set(uint32_t index, T* value) { |
| entries_[index].store(value, std::memory_order_relaxed); |
| } |
| |
| private: |
| Atomic<T*> entries_[0]; |
| }; |
| |
| // C++ mirror of java.lang.DexCache. |
| class MANAGED DexCache final : public Object { |
| public: |
| MIRROR_CLASS("Ljava/lang/DexCache;"); |
| |
| // Size of java.lang.DexCache.class. |
| static uint32_t ClassSize(PointerSize pointer_size); |
| |
| // Note: update the image version in image.cc if changing any of these cache sizes. |
| |
| // Size of type dex cache. Needs to be a power of 2 for entrypoint assumptions to hold. |
| static constexpr size_t kDexCacheTypeCacheSize = 1024; |
| static_assert(IsPowerOfTwo(kDexCacheTypeCacheSize), |
| "Type dex cache size is not a power of 2."); |
| |
| // Size of string dex cache. Needs to be a power of 2 for entrypoint assumptions to hold. |
| static constexpr size_t kDexCacheStringCacheSize = 1024; |
| static_assert(IsPowerOfTwo(kDexCacheStringCacheSize), |
| "String dex cache size is not a power of 2."); |
| |
| // Size of field dex cache. Needs to be a power of 2 for entrypoint assumptions to hold. |
| static constexpr size_t kDexCacheFieldCacheSize = 1024; |
| static_assert(IsPowerOfTwo(kDexCacheFieldCacheSize), |
| "Field dex cache size is not a power of 2."); |
| |
| // Size of method dex cache. Needs to be a power of 2 for entrypoint assumptions to hold. |
| static constexpr size_t kDexCacheMethodCacheSize = 1024; |
| static_assert(IsPowerOfTwo(kDexCacheMethodCacheSize), |
| "Method dex cache size is not a power of 2."); |
| |
| // Size of method type dex cache. Needs to be a power of 2 for entrypoint assumptions |
| // to hold. |
| static constexpr size_t kDexCacheMethodTypeCacheSize = 1024; |
| static_assert(IsPowerOfTwo(kDexCacheMethodTypeCacheSize), |
| "MethodType dex cache size is not a power of 2."); |
| |
| // Size of an instance of java.lang.DexCache not including referenced values. |
| static constexpr uint32_t InstanceSize() { |
| return sizeof(DexCache); |
| } |
| |
| // Visit gc-roots in DexCachePair array in [pairs_begin, pairs_end) range. |
| template <typename Visitor> |
| static void VisitDexCachePairRoots(Visitor& visitor, |
| DexCachePair<Object>* pairs_begin, |
| DexCachePair<Object>* pairs_end) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| EXPORT void Initialize(const DexFile* dex_file, ObjPtr<ClassLoader> class_loader) |
| REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::dex_lock_); |
| |
| // Zero all array references. |
| // WARNING: This does not free the memory since it is in LinearAlloc. |
| EXPORT void ResetNativeArrays() REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags, |
| ReadBarrierOption kReadBarrierOption = kWithReadBarrier> |
| ObjPtr<String> GetLocation() REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| String* GetResolvedString(dex::StringIndex string_idx) ALWAYS_INLINE |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| void SetResolvedString(dex::StringIndex string_idx, ObjPtr<mirror::String> resolved) ALWAYS_INLINE |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| // Clear a string for a string_idx, used to undo string intern transactions to make sure |
| // the string isn't kept live. |
| void ClearString(dex::StringIndex string_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| Class* GetResolvedType(dex::TypeIndex type_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| void SetResolvedType(dex::TypeIndex type_idx, ObjPtr<Class> resolved) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| void ClearResolvedType(dex::TypeIndex type_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| ALWAYS_INLINE ArtMethod* GetResolvedMethod(uint32_t method_idx) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| ALWAYS_INLINE void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| ALWAYS_INLINE ArtField* GetResolvedField(uint32_t idx) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| ALWAYS_INLINE void SetResolvedField(uint32_t idx, ArtField* field) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| MethodType* GetResolvedMethodType(dex::ProtoIndex proto_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| void SetResolvedMethodType(dex::ProtoIndex proto_idx, MethodType* resolved) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| // Clear a method type for proto_idx, used to undo method type resolution |
| // in aborted transactions to make sure the method type isn't kept live. |
| void ClearMethodType(dex::ProtoIndex proto_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| CallSite* GetResolvedCallSite(uint32_t call_site_idx) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| // Attempts to bind |call_site_idx| to the call site |resolved|. The |
| // caller must use the return value in place of |resolved|. This is |
| // because multiple threads can invoke the bootstrap method each |
| // producing a call site, but the method handle invocation on the |
| // call site must be on a common agreed value. |
| ObjPtr<CallSite> SetResolvedCallSite(uint32_t call_site_idx, ObjPtr<CallSite> resolved) |
| REQUIRES_SHARED(Locks::mutator_lock_) WARN_UNUSED; |
| |
| const DexFile* GetDexFile() ALWAYS_INLINE REQUIRES_SHARED(Locks::mutator_lock_) { |
| return GetFieldPtr<const DexFile*>(OFFSET_OF_OBJECT_MEMBER(DexCache, dex_file_)); |
| } |
| |
| void SetDexFile(const DexFile* dex_file) REQUIRES_SHARED(Locks::mutator_lock_) { |
| SetFieldPtr<false>(OFFSET_OF_OBJECT_MEMBER(DexCache, dex_file_), dex_file); |
| } |
| |
| EXPORT void SetLocation(ObjPtr<String> location) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| void VisitReflectiveTargets(ReflectiveValueVisitor* visitor) REQUIRES(Locks::mutator_lock_); |
| |
| void SetClassLoader(ObjPtr<ClassLoader> class_loader) REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| ObjPtr<ClassLoader> GetClassLoader() REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| template <VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags, |
| ReadBarrierOption kReadBarrierOption = kWithReadBarrier, |
| typename Visitor> |
| void VisitNativeRoots(const Visitor& visitor) |
| REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::heap_bitmap_lock_); |
| |
| // Sets null to dex cache array fields which were allocated with the startup |
| // allocator. |
| void UnlinkStartupCaches() REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| // Returns whether we should allocate a full array given the number of elements. |
| // Note: update the image version in image.cc if changing this method. |
| static bool ShouldAllocateFullArray(size_t number_of_elements, size_t dex_cache_size) { |
| return number_of_elements <= dex_cache_size; |
| } |
| |
| |
| // NOLINTBEGIN(bugprone-macro-parentheses) |
| #define DEFINE_ARRAY(name, array_kind, getter_setter, type, ids, alloc_kind) \ |
| template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \ |
| array_kind* Get ##getter_setter() \ |
| ALWAYS_INLINE \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return GetFieldPtr<array_kind*, kVerifyFlags>(getter_setter ##Offset()); \ |
| } \ |
| void Set ##getter_setter(array_kind* value) \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| SetFieldPtr<false>(getter_setter ##Offset(), value); \ |
| } \ |
| static constexpr MemberOffset getter_setter ##Offset() { \ |
| return OFFSET_OF_OBJECT_MEMBER(DexCache, name); \ |
| } \ |
| array_kind* Allocate ##getter_setter(bool startup = false) \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return reinterpret_cast<array_kind*>(AllocArray<type>( \ |
| getter_setter ##Offset(), GetDexFile()->ids(), alloc_kind, startup)); \ |
| } \ |
| template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \ |
| size_t Num ##getter_setter() REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return Get ##getter_setter() == nullptr ? 0u : GetDexFile()->ids(); \ |
| } \ |
| |
| #define DEFINE_PAIR_ARRAY(name, pair_kind, getter_setter, type, size, alloc_kind) \ |
| template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \ |
| pair_kind ##Array<type, size>* Get ##getter_setter() \ |
| ALWAYS_INLINE \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return GetFieldPtr<pair_kind ##Array<type, size>*, kVerifyFlags>(getter_setter ##Offset()); \ |
| } \ |
| void Set ##getter_setter(pair_kind ##Array<type, size>* value) \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| SetFieldPtr<false>(getter_setter ##Offset(), value); \ |
| } \ |
| static constexpr MemberOffset getter_setter ##Offset() { \ |
| return OFFSET_OF_OBJECT_MEMBER(DexCache, name); \ |
| } \ |
| pair_kind ##Array<type, size>* Allocate ##getter_setter() \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return reinterpret_cast<pair_kind ##Array<type, size>*>( \ |
| AllocArray<std::atomic<pair_kind<type>>>( \ |
| getter_setter ##Offset(), size, alloc_kind)); \ |
| } \ |
| template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \ |
| size_t Num ##getter_setter() REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| return Get ##getter_setter() == nullptr ? 0u : size; \ |
| } \ |
| |
| #define DEFINE_DUAL_CACHE( \ |
| name, pair_kind, getter_setter, type, pair_size, alloc_pair_kind, \ |
| array_kind, component_type, ids, alloc_array_kind) \ |
| DEFINE_PAIR_ARRAY( \ |
| name, pair_kind, getter_setter, type, pair_size, alloc_pair_kind) \ |
| DEFINE_ARRAY( \ |
| name ##array_, array_kind, getter_setter ##Array, component_type, ids, alloc_array_kind) \ |
| type* Get ##getter_setter ##Entry(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| DCHECK_LT(index, GetDexFile()->ids()); \ |
| auto* array = Get ##getter_setter ##Array(); \ |
| if (array != nullptr) { \ |
| return array->Get(index); \ |
| } \ |
| auto* pairs = Get ##getter_setter(); \ |
| if (pairs != nullptr) { \ |
| return pairs->Get(index); \ |
| } \ |
| return nullptr; \ |
| } \ |
| void Set ##getter_setter ##Entry(uint32_t index, type* resolved) \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| DCHECK_LT(index, GetDexFile()->ids()); \ |
| auto* array = Get ##getter_setter ##Array(); \ |
| if (array != nullptr) { \ |
| array->Set(index, resolved); \ |
| } else { \ |
| auto* pairs = Get ##getter_setter(); \ |
| if (pairs == nullptr) { \ |
| bool should_allocate_full_array = ShouldAllocateFullArray(GetDexFile()->ids(), pair_size); \ |
| if (ShouldAllocateFullArrayAtStartup() || should_allocate_full_array) { \ |
| array = Allocate ##getter_setter ##Array(!should_allocate_full_array); \ |
| array->Set(index, resolved); \ |
| } else { \ |
| pairs = Allocate ##getter_setter(); \ |
| pairs->Set(index, resolved); \ |
| } \ |
| } else { \ |
| pairs->Set(index, resolved); \ |
| } \ |
| } \ |
| } \ |
| void Unlink ##getter_setter ##ArrayIfStartup() \ |
| REQUIRES_SHARED(Locks::mutator_lock_) { \ |
| if (!ShouldAllocateFullArray(GetDexFile()->ids(), pair_size)) { \ |
| Set ##getter_setter ##Array(nullptr) ; \ |
| } \ |
| } |
| |
| DEFINE_ARRAY(resolved_call_sites_, |
| GcRootArray<CallSite>, |
| ResolvedCallSites, |
| GcRoot<CallSite>, |
| NumCallSiteIds, |
| LinearAllocKind::kGCRootArray) |
| |
| DEFINE_DUAL_CACHE(resolved_fields_, |
| NativeDexCachePair, |
| ResolvedFields, |
| ArtField, |
| kDexCacheFieldCacheSize, |
| LinearAllocKind::kNoGCRoots, |
| NativeArray<ArtField>, |
| ArtField, |
| NumFieldIds, |
| LinearAllocKind::kNoGCRoots) |
| |
| DEFINE_DUAL_CACHE(resolved_method_types_, |
| DexCachePair, |
| ResolvedMethodTypes, |
| mirror::MethodType, |
| kDexCacheMethodTypeCacheSize, |
| LinearAllocKind::kDexCacheArray, |
| GcRootArray<mirror::MethodType>, |
| GcRoot<mirror::MethodType>, |
| NumProtoIds, |
| LinearAllocKind::kGCRootArray); |
| |
| DEFINE_DUAL_CACHE(resolved_methods_, |
| NativeDexCachePair, |
| ResolvedMethods, |
| ArtMethod, |
| kDexCacheMethodCacheSize, |
| LinearAllocKind::kNoGCRoots, |
| NativeArray<ArtMethod>, |
| ArtMethod, |
| NumMethodIds, |
| LinearAllocKind::kNoGCRoots) |
| |
| DEFINE_DUAL_CACHE(resolved_types_, |
| DexCachePair, |
| ResolvedTypes, |
| mirror::Class, |
| kDexCacheTypeCacheSize, |
| LinearAllocKind::kDexCacheArray, |
| GcRootArray<mirror::Class>, |
| GcRoot<mirror::Class>, |
| NumTypeIds, |
| LinearAllocKind::kGCRootArray); |
| |
| DEFINE_DUAL_CACHE(strings_, |
| DexCachePair, |
| Strings, |
| mirror::String, |
| kDexCacheStringCacheSize, |
| LinearAllocKind::kDexCacheArray, |
| GcRootArray<mirror::String>, |
| GcRoot<mirror::String>, |
| NumStringIds, |
| LinearAllocKind::kGCRootArray); |
| |
| // NOLINTEND(bugprone-macro-parentheses) |
| |
| private: |
| // Allocate new array in linear alloc and save it in the given fields. |
| template<typename T> |
| T* AllocArray(MemberOffset obj_offset, size_t num, LinearAllocKind kind, bool startup = false) |
| REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| // Visit instance fields of the dex cache as well as its associated arrays. |
| template <bool kVisitNativeRoots, |
| VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags, |
| ReadBarrierOption kReadBarrierOption = kWithReadBarrier, |
| typename Visitor> |
| void VisitReferences(ObjPtr<Class> klass, const Visitor& visitor) |
| REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::heap_bitmap_lock_); |
| |
| // Returns whether we should allocate a full array given the current state of |
| // the runtime and oat files. |
| bool ShouldAllocateFullArrayAtStartup() REQUIRES_SHARED(Locks::mutator_lock_); |
| |
| HeapReference<ClassLoader> class_loader_; |
| HeapReference<String> location_; |
| |
| uint64_t dex_file_; // const DexFile* |
| // |
| uint64_t resolved_call_sites_; // Array of call sites |
| uint64_t resolved_fields_; // NativeDexCacheArray holding ArtField's |
| uint64_t resolved_fields_array_; // Array of ArtField's. |
| uint64_t resolved_method_types_; // DexCacheArray holding mirror::MethodType's |
| uint64_t resolved_method_types_array_; // Array of mirror::MethodType's |
| uint64_t resolved_methods_; // NativeDexCacheArray holding ArtMethod's |
| uint64_t resolved_methods_array_; // Array of ArtMethod's |
| uint64_t resolved_types_; // DexCacheArray holding mirror::Class's |
| uint64_t resolved_types_array_; // Array of resolved types. |
| uint64_t strings_; // DexCacheArray holding mirror::String's |
| uint64_t strings_array_; // Array of String's. |
| |
| friend struct art::DexCacheOffsets; // for verifying offset information |
| friend class linker::ImageWriter; |
| friend class Object; // For VisitReferences |
| DISALLOW_IMPLICIT_CONSTRUCTORS(DexCache); |
| }; |
| |
| } // namespace mirror |
| } // namespace art |
| |
| #endif // ART_RUNTIME_MIRROR_DEX_CACHE_H_ |