blob: 60a82a29ae77aa356f3f7220bf9ea7202a9fa0b8 [file] [log] [blame]
/*
* Copyright (C) 2014 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_HANDLE_SCOPE_INL_H_
#define ART_RUNTIME_HANDLE_SCOPE_INL_H_
#include "handle_scope.h"
#include "base/mutex.h"
#include "handle.h"
#include "handle_wrapper.h"
#include "mirror/object_reference-inl.h"
#include "obj_ptr-inl.h"
#include "thread-current-inl.h"
#include "verify_object.h"
namespace art {
template<size_t kNumReferences>
inline FixedSizeHandleScope<kNumReferences>::FixedSizeHandleScope(BaseHandleScope* link,
ObjPtr<mirror::Object> fill_value)
: HandleScope(link, kNumReferences) {
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(Thread::Current());
}
static_assert(kNumReferences >= 1, "FixedSizeHandleScope must contain at least 1 reference");
DCHECK_EQ(&storage_[0], GetReferences()); // TODO: Figure out how to use a compile assert.
for (size_t i = 0; i < kNumReferences; ++i) {
SetReference(i, fill_value);
}
}
template<size_t kNumReferences>
inline StackHandleScope<kNumReferences>::StackHandleScope(Thread* self,
ObjPtr<mirror::Object> fill_value)
: FixedSizeHandleScope<kNumReferences>(self->GetTopHandleScope(), fill_value),
self_(self) {
DCHECK_EQ(self, Thread::Current());
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(self_);
}
self_->PushHandleScope(this);
}
template<size_t kNumReferences>
inline StackHandleScope<kNumReferences>::~StackHandleScope() {
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(self_);
}
BaseHandleScope* top_handle_scope = self_->PopHandleScope();
DCHECK_EQ(top_handle_scope, this);
}
inline size_t HandleScope::SizeOf(uint32_t num_references) {
size_t header_size = sizeof(HandleScope);
size_t data_size = sizeof(StackReference<mirror::Object>) * num_references;
return header_size + data_size;
}
inline size_t HandleScope::SizeOf(PointerSize pointer_size, uint32_t num_references) {
// Assume that the layout is packed.
size_t header_size = ReferencesOffset(pointer_size);
size_t data_size = sizeof(StackReference<mirror::Object>) * num_references;
return header_size + data_size;
}
inline ObjPtr<mirror::Object> HandleScope::GetReference(size_t i) const {
DCHECK_LT(i, NumberOfReferences());
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(Thread::Current());
}
return GetReferences()[i].AsMirrorPtr();
}
inline Handle<mirror::Object> HandleScope::GetHandle(size_t i) {
DCHECK_LT(i, NumberOfReferences());
return Handle<mirror::Object>(&GetReferences()[i]);
}
inline MutableHandle<mirror::Object> HandleScope::GetMutableHandle(size_t i) {
DCHECK_LT(i, NumberOfReferences());
return MutableHandle<mirror::Object>(&GetReferences()[i]);
}
inline void HandleScope::SetReference(size_t i, ObjPtr<mirror::Object> object) {
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(Thread::Current());
}
DCHECK_LT(i, NumberOfReferences());
GetReferences()[i].Assign(object);
}
inline bool HandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry) const {
// A HandleScope should always contain something. One created by the
// jni_compiler should have a jobject/jclass as a native method is
// passed in a this pointer or a class
DCHECK_GT(NumberOfReferences(), 0U);
return &GetReferences()[0] <= handle_scope_entry &&
handle_scope_entry <= &GetReferences()[number_of_references_ - 1];
}
template <typename Visitor>
inline void HandleScope::VisitRoots(Visitor& visitor) {
for (size_t i = 0, count = NumberOfReferences(); i < count; ++i) {
// GetReference returns a pointer to the stack reference within the handle scope. If this
// needs to be updated, it will be done by the root visitor.
visitor.VisitRootIfNonNull(GetHandle(i).GetReference());
}
}
template <typename Visitor>
inline void HandleScope::VisitHandles(Visitor& visitor) {
for (size_t i = 0, count = NumberOfReferences(); i < count; ++i) {
if (GetHandle(i) != nullptr) {
visitor.Visit(GetHandle(i));
}
}
}
template<size_t kNumReferences> template<class T>
inline MutableHandle<T> FixedSizeHandleScope<kNumReferences>::NewHandle(T* object) {
return NewHandle(ObjPtr<T>(object));
}
template<size_t kNumReferences> template<class MirrorType>
inline MutableHandle<MirrorType> FixedSizeHandleScope<kNumReferences>::NewHandle(
ObjPtr<MirrorType> object) {
SetReference(pos_, object);
MutableHandle<MirrorType> h(GetHandle<MirrorType>(pos_));
++pos_;
return h;
}
template<size_t kNumReferences> template<class T>
inline HandleWrapper<T> FixedSizeHandleScope<kNumReferences>::NewHandleWrapper(T** object) {
return HandleWrapper<T>(object, NewHandle(*object));
}
template<size_t kNumReferences> template<class T>
inline HandleWrapperObjPtr<T> FixedSizeHandleScope<kNumReferences>::NewHandleWrapper(
ObjPtr<T>* object) {
return HandleWrapperObjPtr<T>(object, NewHandle(*object));
}
template<size_t kNumReferences>
inline void FixedSizeHandleScope<kNumReferences>::SetReference(size_t i,
ObjPtr<mirror::Object> object) {
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(Thread::Current());
}
DCHECK_LT(i, kNumReferences);
VerifyObject(object);
GetReferences()[i].Assign(object);
}
// Number of references contained within this handle scope.
inline uint32_t BaseHandleScope::NumberOfReferences() const {
return LIKELY(!IsVariableSized())
? AsHandleScope()->NumberOfReferences()
: AsVariableSized()->NumberOfReferences();
}
inline bool BaseHandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry) const {
return LIKELY(!IsVariableSized())
? AsHandleScope()->Contains(handle_scope_entry)
: AsVariableSized()->Contains(handle_scope_entry);
}
template <typename Visitor>
inline void BaseHandleScope::VisitRoots(Visitor& visitor) {
if (LIKELY(!IsVariableSized())) {
AsHandleScope()->VisitRoots(visitor);
} else {
AsVariableSized()->VisitRoots(visitor);
}
}
template <typename Visitor>
inline void BaseHandleScope::VisitHandles(Visitor& visitor) {
if (LIKELY(!IsVariableSized())) {
AsHandleScope()->VisitHandles(visitor);
} else {
AsVariableSized()->VisitHandles(visitor);
}
}
inline VariableSizedHandleScope* BaseHandleScope::AsVariableSized() {
DCHECK(IsVariableSized());
return down_cast<VariableSizedHandleScope*>(this);
}
inline HandleScope* BaseHandleScope::AsHandleScope() {
DCHECK(!IsVariableSized());
return down_cast<HandleScope*>(this);
}
inline const VariableSizedHandleScope* BaseHandleScope::AsVariableSized() const {
DCHECK(IsVariableSized());
return down_cast<const VariableSizedHandleScope*>(this);
}
inline const HandleScope* BaseHandleScope::AsHandleScope() const {
DCHECK(!IsVariableSized());
return down_cast<const HandleScope*>(this);
}
template<class T>
inline MutableHandle<T> VariableSizedHandleScope::NewHandle(T* object) {
return NewHandle(ObjPtr<T>(object));
}
template<class MirrorType>
inline MutableHandle<MirrorType> VariableSizedHandleScope::NewHandle(ObjPtr<MirrorType> ptr) {
if (current_scope_->RemainingSlots() == 0) {
current_scope_ = new LocalScopeType(current_scope_);
}
return current_scope_->NewHandle(ptr);
}
inline VariableSizedHandleScope::VariableSizedHandleScope(Thread* const self)
: BaseHandleScope(self->GetTopHandleScope()),
self_(self),
current_scope_(&first_scope_),
first_scope_(/*link=*/ nullptr) {
DCHECK_EQ(self, Thread::Current());
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(self_);
}
self_->PushHandleScope(this);
}
inline VariableSizedHandleScope::~VariableSizedHandleScope() {
if (kDebugLocking) {
Locks::mutator_lock_->AssertSharedHeld(self_);
}
BaseHandleScope* top_handle_scope = self_->PopHandleScope();
DCHECK_EQ(top_handle_scope, this);
// Don't delete first_scope_ since it is not heap allocated.
while (current_scope_ != &first_scope_) {
LocalScopeType* next = down_cast<LocalScopeType*>(current_scope_->GetLink());
delete current_scope_;
current_scope_ = next;
}
}
inline uint32_t VariableSizedHandleScope::NumberOfReferences() const {
uint32_t sum = 0;
const LocalScopeType* cur = current_scope_;
while (cur != nullptr) {
sum += cur->NumberOfReferences();
cur = reinterpret_cast<const LocalScopeType*>(cur->GetLink());
}
return sum;
}
inline bool VariableSizedHandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry)
const {
const LocalScopeType* cur = current_scope_;
while (cur != nullptr) {
if (cur->Contains(handle_scope_entry)) {
return true;
}
cur = reinterpret_cast<const LocalScopeType*>(cur->GetLink());
}
return false;
}
template <typename Visitor>
inline void VariableSizedHandleScope::VisitRoots(Visitor& visitor) {
LocalScopeType* cur = current_scope_;
while (cur != nullptr) {
cur->VisitRoots(visitor);
cur = reinterpret_cast<LocalScopeType*>(cur->GetLink());
}
}
template <typename Visitor>
inline void VariableSizedHandleScope::VisitHandles(Visitor& visitor) {
LocalScopeType* cur = current_scope_;
while (cur != nullptr) {
cur->VisitHandles(visitor);
cur = reinterpret_cast<LocalScopeType*>(cur->GetLink());
}
}
} // namespace art
#endif // ART_RUNTIME_HANDLE_SCOPE_INL_H_