blob: 921d8028e72f1a0ba07d3079dcf40d6bb2749f01 [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_MIRROR_DEX_CACHE_H_
#define ART_RUNTIME_MIRROR_DEX_CACHE_H_
#include "array.h"
#include "base/array_ref.h"
#include "base/atomic_pair.h"
#include "base/bit_utils.h"
#include "base/locks.h"
#include "base/macros.h"
#include "dex/dex_file.h"
#include "dex/dex_file_types.h"
#include "gc_root.h" // Note: must not use -inl here to avoid circular dependency.
#include "linear_alloc.h"
#include "object.h"
#include "object_array.h"
namespace art HIDDEN {
namespace linker {
class ImageWriter;
} // namespace linker
class ArtField;
class ArtMethod;
struct DexCacheOffsets;
class DexFile;
union JValue;
class ReflectiveValueVisitor;
class Thread;
namespace mirror {
class CallSite;
class Class;
class ClassLoader;
class DexCache;
class MethodType;
class String;
template <typename T> struct PACKED(8) DexCachePair {
GcRoot<T> object;
uint32_t index;
// The array is initially [ {0,0}, {0,0}, {0,0} ... ]
// We maintain the invariant that once a dex cache entry is populated,
// the pointer is always non-0
// Any given entry would thus be:
// {non-0, non-0} OR {0,0}
//
// It's generally sufficiently enough then to check if the
// lookup index matches the stored index (for a >0 lookup index)
// because if it's true the pointer is also non-null.
//
// For the 0th entry which is a special case, the value is either
// {0,0} (initial state) or {non-0, 0} which indicates
// that a valid object is stored at that index for a dex section id of 0.
//
// As an optimization, we want to avoid branching on the object pointer since
// it's always non-null if the id branch succeeds (except for the 0th id).
// Set the initial state for the 0th entry to be {0,1} which is guaranteed to fail
// the lookup id == stored id branch.
DexCachePair(ObjPtr<T> object, uint32_t index);
DexCachePair() : index(0) {}
DexCachePair(const DexCachePair<T>&) = default;
DexCachePair& operator=(const DexCachePair<T>&) = default;
static void Initialize(std::atomic<DexCachePair<T>>* dex_cache);
static uint32_t InvalidIndexForSlot(uint32_t slot) {
// Since the cache size is a power of two, 0 will always map to slot 0.
// Use 1 for slot 0 and 0 for all other slots.
return (slot == 0) ? 1u : 0u;
}
T* GetObjectForIndex(uint32_t idx) REQUIRES_SHARED(Locks::mutator_lock_);
};
template <typename T> struct PACKED(2 * __SIZEOF_POINTER__) NativeDexCachePair {
T* object;
size_t index;
// This is similar to DexCachePair except that we're storing a native pointer
// instead of a GC root. See DexCachePair for the details.
NativeDexCachePair(T* object, uint32_t index)
: object(object),
index(index) {}
NativeDexCachePair() : object(nullptr), index(0u) { }
NativeDexCachePair(const NativeDexCachePair<T>&) = default;
NativeDexCachePair& operator=(const NativeDexCachePair<T>&) = default;
static void Initialize(std::atomic<NativeDexCachePair<T>>* dex_cache);
static uint32_t InvalidIndexForSlot(uint32_t slot) {
// Since the cache size is a power of two, 0 will always map to slot 0.
// Use 1 for slot 0 and 0 for all other slots.
return (slot == 0) ? 1u : 0u;
}
T* GetObjectForIndex(uint32_t idx) REQUIRES_SHARED(Locks::mutator_lock_) {
if (idx != index) {
return nullptr;
}
DCHECK(object != nullptr);
return object;
}
};
template <typename T, size_t size> class NativeDexCachePairArray {
public:
NativeDexCachePairArray() {}
T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) {
auto pair = GetNativePair(entries_, SlotIndex(index));
return pair.GetObjectForIndex(index);
}
void Set(uint32_t index, T* value) {
NativeDexCachePair<T> pair(value, index);
SetNativePair(entries_, SlotIndex(index), pair);
}
NativeDexCachePair<T> GetNativePair(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) {
return GetNativePair(entries_, SlotIndex(index));
}
void SetNativePair(uint32_t index, NativeDexCachePair<T> value) {
SetNativePair(entries_, SlotIndex(index), value);
}
private:
NativeDexCachePair<T> GetNativePair(std::atomic<NativeDexCachePair<T>>* pair_array, size_t idx) {
auto* array = reinterpret_cast<AtomicPair<uintptr_t>*>(pair_array);
AtomicPair<uintptr_t> value = AtomicPairLoadAcquire(&array[idx]);
return NativeDexCachePair<T>(reinterpret_cast<T*>(value.val), value.key);
}
void SetNativePair(std::atomic<NativeDexCachePair<T>>* pair_array,
size_t idx,
NativeDexCachePair<T> pair) {
auto* array = reinterpret_cast<AtomicPair<uintptr_t>*>(pair_array);
AtomicPair<uintptr_t> v(pair.index, reinterpret_cast<size_t>(pair.object));
AtomicPairStoreRelease(&array[idx], v);
}
uint32_t SlotIndex(uint32_t index) {
return index % size;
}
std::atomic<NativeDexCachePair<T>> entries_[0];
NativeDexCachePairArray(const NativeDexCachePairArray<T, size>&) = delete;
NativeDexCachePairArray& operator=(const NativeDexCachePairArray<T, size>&) = delete;
};
template <typename T, size_t size> class DexCachePairArray {
public:
DexCachePairArray() {}
T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) {
return GetPair(index).GetObjectForIndex(index);
}
void Set(uint32_t index, T* value) REQUIRES_SHARED(Locks::mutator_lock_) {
SetPair(index, DexCachePair<T>(value, index));
}
DexCachePair<T> GetPair(uint32_t index) {
return entries_[SlotIndex(index)].load(std::memory_order_relaxed);
}
void SetPair(uint32_t index, DexCachePair<T> value) {
entries_[SlotIndex(index)].store(value, std::memory_order_relaxed);
}
void Clear(uint32_t index) {
uint32_t slot = SlotIndex(index);
// This is racy but should only be called from the transactional interpreter.
if (entries_[slot].load(std::memory_order_relaxed).index == index) {
DexCachePair<T> cleared(nullptr, DexCachePair<T>::InvalidIndexForSlot(slot));
entries_[slot].store(cleared, std::memory_order_relaxed);
}
}
private:
uint32_t SlotIndex(uint32_t index) {
return index % size;
}
std::atomic<DexCachePair<T>> entries_[0];
DexCachePairArray(const DexCachePairArray<T, size>&) = delete;
DexCachePairArray& operator=(const DexCachePairArray<T, size>&) = delete;
};
template <typename T> class GcRootArray {
public:
GcRootArray() {}
T* Get(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_);
Atomic<GcRoot<T>>* GetGcRoot(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) {
return &entries_[index];
}
// Only to be used in locations that don't need the atomic or will later load
// and read atomically.
GcRoot<T>* GetGcRootAddress(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) {
static_assert(sizeof(GcRoot<T>) == sizeof(Atomic<GcRoot<T>>));
return reinterpret_cast<GcRoot<T>*>(&entries_[index]);
}
void Set(uint32_t index, T* value) REQUIRES_SHARED(Locks::mutator_lock_);
private:
Atomic<GcRoot<T>> entries_[0];
};
template <typename T> class NativeArray {
public:
NativeArray() {}
T* Get(uint32_t index) {
return entries_[index].load(std::memory_order_relaxed);
}
T** GetPtrEntryPtrSize(uint32_t index, PointerSize ptr_size) {
if (ptr_size == PointerSize::k64) {
return reinterpret_cast<T**>(reinterpret_cast<uint64_t*>(entries_) + index);
} else {
return reinterpret_cast<T**>(reinterpret_cast<uint32_t*>(entries_) + index);
}
}
void Set(uint32_t index, T* value) {
entries_[index].store(value, std::memory_order_relaxed);
}
private:
Atomic<T*> entries_[0];
};
// C++ mirror of java.lang.DexCache.
class MANAGED DexCache final : public Object {
public:
MIRROR_CLASS("Ljava/lang/DexCache;");
// Size of java.lang.DexCache.class.
static uint32_t ClassSize(PointerSize pointer_size);
// Note: update the image version in image.cc if changing any of these cache sizes.
// Size of type dex cache. Needs to be a power of 2 for entrypoint assumptions to hold.
static constexpr size_t kDexCacheTypeCacheSize = 1024;
static_assert(IsPowerOfTwo(kDexCacheTypeCacheSize),
"Type dex cache size is not a power of 2.");
// Size of string dex cache. Needs to be a power of 2 for entrypoint assumptions to hold.
static constexpr size_t kDexCacheStringCacheSize = 1024;
static_assert(IsPowerOfTwo(kDexCacheStringCacheSize),
"String dex cache size is not a power of 2.");
// Size of field dex cache. Needs to be a power of 2 for entrypoint assumptions to hold.
static constexpr size_t kDexCacheFieldCacheSize = 1024;
static_assert(IsPowerOfTwo(kDexCacheFieldCacheSize),
"Field dex cache size is not a power of 2.");
// Size of method dex cache. Needs to be a power of 2 for entrypoint assumptions to hold.
static constexpr size_t kDexCacheMethodCacheSize = 1024;
static_assert(IsPowerOfTwo(kDexCacheMethodCacheSize),
"Method dex cache size is not a power of 2.");
// Size of method type dex cache. Needs to be a power of 2 for entrypoint assumptions
// to hold.
static constexpr size_t kDexCacheMethodTypeCacheSize = 1024;
static_assert(IsPowerOfTwo(kDexCacheMethodTypeCacheSize),
"MethodType dex cache size is not a power of 2.");
// Size of an instance of java.lang.DexCache not including referenced values.
static constexpr uint32_t InstanceSize() {
return sizeof(DexCache);
}
// Visit gc-roots in DexCachePair array in [pairs_begin, pairs_end) range.
template <typename Visitor>
static void VisitDexCachePairRoots(Visitor& visitor,
DexCachePair<Object>* pairs_begin,
DexCachePair<Object>* pairs_end)
REQUIRES_SHARED(Locks::mutator_lock_);
EXPORT void Initialize(const DexFile* dex_file, ObjPtr<ClassLoader> class_loader)
REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::dex_lock_);
// Zero all array references.
// WARNING: This does not free the memory since it is in LinearAlloc.
EXPORT void ResetNativeArrays() REQUIRES_SHARED(Locks::mutator_lock_);
template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags,
ReadBarrierOption kReadBarrierOption = kWithReadBarrier>
ObjPtr<String> GetLocation() REQUIRES_SHARED(Locks::mutator_lock_);
String* GetResolvedString(dex::StringIndex string_idx) ALWAYS_INLINE
REQUIRES_SHARED(Locks::mutator_lock_);
void SetResolvedString(dex::StringIndex string_idx, ObjPtr<mirror::String> resolved) ALWAYS_INLINE
REQUIRES_SHARED(Locks::mutator_lock_);
// Clear a string for a string_idx, used to undo string intern transactions to make sure
// the string isn't kept live.
void ClearString(dex::StringIndex string_idx) REQUIRES_SHARED(Locks::mutator_lock_);
Class* GetResolvedType(dex::TypeIndex type_idx) REQUIRES_SHARED(Locks::mutator_lock_);
void SetResolvedType(dex::TypeIndex type_idx, ObjPtr<Class> resolved)
REQUIRES_SHARED(Locks::mutator_lock_);
void ClearResolvedType(dex::TypeIndex type_idx) REQUIRES_SHARED(Locks::mutator_lock_);
ALWAYS_INLINE ArtMethod* GetResolvedMethod(uint32_t method_idx)
REQUIRES_SHARED(Locks::mutator_lock_);
ALWAYS_INLINE void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved)
REQUIRES_SHARED(Locks::mutator_lock_);
ALWAYS_INLINE ArtField* GetResolvedField(uint32_t idx)
REQUIRES_SHARED(Locks::mutator_lock_);
ALWAYS_INLINE void SetResolvedField(uint32_t idx, ArtField* field)
REQUIRES_SHARED(Locks::mutator_lock_);
MethodType* GetResolvedMethodType(dex::ProtoIndex proto_idx) REQUIRES_SHARED(Locks::mutator_lock_);
void SetResolvedMethodType(dex::ProtoIndex proto_idx, MethodType* resolved)
REQUIRES_SHARED(Locks::mutator_lock_);
// Clear a method type for proto_idx, used to undo method type resolution
// in aborted transactions to make sure the method type isn't kept live.
void ClearMethodType(dex::ProtoIndex proto_idx) REQUIRES_SHARED(Locks::mutator_lock_);
CallSite* GetResolvedCallSite(uint32_t call_site_idx) REQUIRES_SHARED(Locks::mutator_lock_);
// Attempts to bind |call_site_idx| to the call site |resolved|. The
// caller must use the return value in place of |resolved|. This is
// because multiple threads can invoke the bootstrap method each
// producing a call site, but the method handle invocation on the
// call site must be on a common agreed value.
ObjPtr<CallSite> SetResolvedCallSite(uint32_t call_site_idx, ObjPtr<CallSite> resolved)
REQUIRES_SHARED(Locks::mutator_lock_) WARN_UNUSED;
const DexFile* GetDexFile() ALWAYS_INLINE REQUIRES_SHARED(Locks::mutator_lock_) {
return GetFieldPtr<const DexFile*>(OFFSET_OF_OBJECT_MEMBER(DexCache, dex_file_));
}
void SetDexFile(const DexFile* dex_file) REQUIRES_SHARED(Locks::mutator_lock_) {
SetFieldPtr<false>(OFFSET_OF_OBJECT_MEMBER(DexCache, dex_file_), dex_file);
}
EXPORT void SetLocation(ObjPtr<String> location) REQUIRES_SHARED(Locks::mutator_lock_);
void VisitReflectiveTargets(ReflectiveValueVisitor* visitor) REQUIRES(Locks::mutator_lock_);
void SetClassLoader(ObjPtr<ClassLoader> class_loader) REQUIRES_SHARED(Locks::mutator_lock_);
ObjPtr<ClassLoader> GetClassLoader() REQUIRES_SHARED(Locks::mutator_lock_);
template <VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags,
ReadBarrierOption kReadBarrierOption = kWithReadBarrier,
typename Visitor>
void VisitNativeRoots(const Visitor& visitor)
REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::heap_bitmap_lock_);
// Sets null to dex cache array fields which were allocated with the startup
// allocator.
void UnlinkStartupCaches() REQUIRES_SHARED(Locks::mutator_lock_);
// Returns whether we should allocate a full array given the number of elements.
// Note: update the image version in image.cc if changing this method.
static bool ShouldAllocateFullArray(size_t number_of_elements, size_t dex_cache_size) {
return number_of_elements <= dex_cache_size;
}
// NOLINTBEGIN(bugprone-macro-parentheses)
#define DEFINE_ARRAY(name, array_kind, getter_setter, type, ids, alloc_kind) \
template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \
array_kind* Get ##getter_setter() \
ALWAYS_INLINE \
REQUIRES_SHARED(Locks::mutator_lock_) { \
return GetFieldPtr<array_kind*, kVerifyFlags>(getter_setter ##Offset()); \
} \
void Set ##getter_setter(array_kind* value) \
REQUIRES_SHARED(Locks::mutator_lock_) { \
SetFieldPtr<false>(getter_setter ##Offset(), value); \
} \
static constexpr MemberOffset getter_setter ##Offset() { \
return OFFSET_OF_OBJECT_MEMBER(DexCache, name); \
} \
array_kind* Allocate ##getter_setter(bool startup = false) \
REQUIRES_SHARED(Locks::mutator_lock_) { \
return reinterpret_cast<array_kind*>(AllocArray<type>( \
getter_setter ##Offset(), GetDexFile()->ids(), alloc_kind, startup)); \
} \
template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \
size_t Num ##getter_setter() REQUIRES_SHARED(Locks::mutator_lock_) { \
return Get ##getter_setter() == nullptr ? 0u : GetDexFile()->ids(); \
} \
#define DEFINE_PAIR_ARRAY(name, pair_kind, getter_setter, type, size, alloc_kind) \
template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \
pair_kind ##Array<type, size>* Get ##getter_setter() \
ALWAYS_INLINE \
REQUIRES_SHARED(Locks::mutator_lock_) { \
return GetFieldPtr<pair_kind ##Array<type, size>*, kVerifyFlags>(getter_setter ##Offset()); \
} \
void Set ##getter_setter(pair_kind ##Array<type, size>* value) \
REQUIRES_SHARED(Locks::mutator_lock_) { \
SetFieldPtr<false>(getter_setter ##Offset(), value); \
} \
static constexpr MemberOffset getter_setter ##Offset() { \
return OFFSET_OF_OBJECT_MEMBER(DexCache, name); \
} \
pair_kind ##Array<type, size>* Allocate ##getter_setter() \
REQUIRES_SHARED(Locks::mutator_lock_) { \
return reinterpret_cast<pair_kind ##Array<type, size>*>( \
AllocArray<std::atomic<pair_kind<type>>>( \
getter_setter ##Offset(), size, alloc_kind)); \
} \
template<VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags> \
size_t Num ##getter_setter() REQUIRES_SHARED(Locks::mutator_lock_) { \
return Get ##getter_setter() == nullptr ? 0u : size; \
} \
#define DEFINE_DUAL_CACHE( \
name, pair_kind, getter_setter, type, pair_size, alloc_pair_kind, \
array_kind, component_type, ids, alloc_array_kind) \
DEFINE_PAIR_ARRAY( \
name, pair_kind, getter_setter, type, pair_size, alloc_pair_kind) \
DEFINE_ARRAY( \
name ##array_, array_kind, getter_setter ##Array, component_type, ids, alloc_array_kind) \
type* Get ##getter_setter ##Entry(uint32_t index) REQUIRES_SHARED(Locks::mutator_lock_) { \
DCHECK_LT(index, GetDexFile()->ids()); \
auto* array = Get ##getter_setter ##Array(); \
if (array != nullptr) { \
return array->Get(index); \
} \
auto* pairs = Get ##getter_setter(); \
if (pairs != nullptr) { \
return pairs->Get(index); \
} \
return nullptr; \
} \
void Set ##getter_setter ##Entry(uint32_t index, type* resolved) \
REQUIRES_SHARED(Locks::mutator_lock_) { \
DCHECK_LT(index, GetDexFile()->ids()); \
auto* array = Get ##getter_setter ##Array(); \
if (array != nullptr) { \
array->Set(index, resolved); \
} else { \
auto* pairs = Get ##getter_setter(); \
if (pairs == nullptr) { \
bool should_allocate_full_array = ShouldAllocateFullArray(GetDexFile()->ids(), pair_size); \
if (ShouldAllocateFullArrayAtStartup() || should_allocate_full_array) { \
array = Allocate ##getter_setter ##Array(!should_allocate_full_array); \
array->Set(index, resolved); \
} else { \
pairs = Allocate ##getter_setter(); \
pairs->Set(index, resolved); \
} \
} else { \
pairs->Set(index, resolved); \
} \
} \
} \
void Unlink ##getter_setter ##ArrayIfStartup() \
REQUIRES_SHARED(Locks::mutator_lock_) { \
if (!ShouldAllocateFullArray(GetDexFile()->ids(), pair_size)) { \
Set ##getter_setter ##Array(nullptr) ; \
} \
}
DEFINE_ARRAY(resolved_call_sites_,
GcRootArray<CallSite>,
ResolvedCallSites,
GcRoot<CallSite>,
NumCallSiteIds,
LinearAllocKind::kGCRootArray)
DEFINE_DUAL_CACHE(resolved_fields_,
NativeDexCachePair,
ResolvedFields,
ArtField,
kDexCacheFieldCacheSize,
LinearAllocKind::kNoGCRoots,
NativeArray<ArtField>,
ArtField,
NumFieldIds,
LinearAllocKind::kNoGCRoots)
DEFINE_DUAL_CACHE(resolved_method_types_,
DexCachePair,
ResolvedMethodTypes,
mirror::MethodType,
kDexCacheMethodTypeCacheSize,
LinearAllocKind::kDexCacheArray,
GcRootArray<mirror::MethodType>,
GcRoot<mirror::MethodType>,
NumProtoIds,
LinearAllocKind::kGCRootArray);
DEFINE_DUAL_CACHE(resolved_methods_,
NativeDexCachePair,
ResolvedMethods,
ArtMethod,
kDexCacheMethodCacheSize,
LinearAllocKind::kNoGCRoots,
NativeArray<ArtMethod>,
ArtMethod,
NumMethodIds,
LinearAllocKind::kNoGCRoots)
DEFINE_DUAL_CACHE(resolved_types_,
DexCachePair,
ResolvedTypes,
mirror::Class,
kDexCacheTypeCacheSize,
LinearAllocKind::kDexCacheArray,
GcRootArray<mirror::Class>,
GcRoot<mirror::Class>,
NumTypeIds,
LinearAllocKind::kGCRootArray);
DEFINE_DUAL_CACHE(strings_,
DexCachePair,
Strings,
mirror::String,
kDexCacheStringCacheSize,
LinearAllocKind::kDexCacheArray,
GcRootArray<mirror::String>,
GcRoot<mirror::String>,
NumStringIds,
LinearAllocKind::kGCRootArray);
// NOLINTEND(bugprone-macro-parentheses)
private:
// Allocate new array in linear alloc and save it in the given fields.
template<typename T>
T* AllocArray(MemberOffset obj_offset, size_t num, LinearAllocKind kind, bool startup = false)
REQUIRES_SHARED(Locks::mutator_lock_);
// Visit instance fields of the dex cache as well as its associated arrays.
template <bool kVisitNativeRoots,
VerifyObjectFlags kVerifyFlags = kDefaultVerifyFlags,
ReadBarrierOption kReadBarrierOption = kWithReadBarrier,
typename Visitor>
void VisitReferences(ObjPtr<Class> klass, const Visitor& visitor)
REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::heap_bitmap_lock_);
// Returns whether we should allocate a full array given the current state of
// the runtime and oat files.
bool ShouldAllocateFullArrayAtStartup() REQUIRES_SHARED(Locks::mutator_lock_);
HeapReference<ClassLoader> class_loader_;
HeapReference<String> location_;
uint64_t dex_file_; // const DexFile*
//
uint64_t resolved_call_sites_; // Array of call sites
uint64_t resolved_fields_; // NativeDexCacheArray holding ArtField's
uint64_t resolved_fields_array_; // Array of ArtField's.
uint64_t resolved_method_types_; // DexCacheArray holding mirror::MethodType's
uint64_t resolved_method_types_array_; // Array of mirror::MethodType's
uint64_t resolved_methods_; // NativeDexCacheArray holding ArtMethod's
uint64_t resolved_methods_array_; // Array of ArtMethod's
uint64_t resolved_types_; // DexCacheArray holding mirror::Class's
uint64_t resolved_types_array_; // Array of resolved types.
uint64_t strings_; // DexCacheArray holding mirror::String's
uint64_t strings_array_; // Array of String's.
friend struct art::DexCacheOffsets; // for verifying offset information
friend class linker::ImageWriter;
friend class Object; // For VisitReferences
DISALLOW_IMPLICIT_CONSTRUCTORS(DexCache);
};
} // namespace mirror
} // namespace art
#endif // ART_RUNTIME_MIRROR_DEX_CACHE_H_