| /* |
| * Copyright (C) 2012 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_SCOPED_THREAD_STATE_CHANGE_H_ |
| #define ART_RUNTIME_SCOPED_THREAD_STATE_CHANGE_H_ |
| |
| #include "base/casts.h" |
| #include "jni_internal.h" |
| #include "thread-inl.h" |
| |
| namespace art { |
| |
| // Scoped change into and out of a particular state. Handles Runnable transitions that require |
| // more complicated suspension checking. The subclasses ScopedObjectAccessUnchecked and |
| // ScopedObjectAccess are used to handle the change into Runnable to get direct access to objects, |
| // the unchecked variant doesn't aid annotalysis. |
| class ScopedThreadStateChange { |
| public: |
| ScopedThreadStateChange(Thread* self, ThreadState new_thread_state) |
| LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) ALWAYS_INLINE |
| : self_(self), thread_state_(new_thread_state), expected_has_no_thread_(false) { |
| if (UNLIKELY(self_ == NULL)) { |
| // Value chosen arbitrarily and won't be used in the destructor since thread_ == NULL. |
| old_thread_state_ = kTerminated; |
| MutexLock mu(NULL, *Locks::runtime_shutdown_lock_); |
| Runtime* runtime = Runtime::Current(); |
| CHECK(runtime == NULL || !runtime->IsStarted() || runtime->IsShuttingDown()); |
| } else { |
| bool runnable_transition; |
| DCHECK_EQ(self, Thread::Current()); |
| // Read state without locks, ok as state is effectively thread local and we're not interested |
| // in the suspend count (this will be handled in the runnable transitions). |
| old_thread_state_ = self->GetState(); |
| runnable_transition = old_thread_state_ == kRunnable || new_thread_state == kRunnable; |
| if (!runnable_transition) { |
| // A suspended transition to another effectively suspended transition, ok to use Unsafe. |
| self_->SetState(new_thread_state); |
| } |
| |
| if (runnable_transition && old_thread_state_ != new_thread_state) { |
| if (new_thread_state == kRunnable) { |
| self_->TransitionFromSuspendedToRunnable(); |
| } else { |
| DCHECK_EQ(old_thread_state_, kRunnable); |
| self_->TransitionFromRunnableToSuspended(new_thread_state); |
| } |
| } |
| } |
| } |
| |
| ~ScopedThreadStateChange() LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) ALWAYS_INLINE { |
| if (UNLIKELY(self_ == NULL)) { |
| if (!expected_has_no_thread_) { |
| MutexLock mu(NULL, *Locks::runtime_shutdown_lock_); |
| Runtime* runtime = Runtime::Current(); |
| bool shutting_down = (runtime == NULL) || runtime->IsShuttingDown(); |
| CHECK(shutting_down); |
| } |
| } else { |
| if (old_thread_state_ != thread_state_) { |
| if (old_thread_state_ == kRunnable) { |
| self_->TransitionFromSuspendedToRunnable(); |
| } else if (thread_state_ == kRunnable) { |
| self_->TransitionFromRunnableToSuspended(old_thread_state_); |
| } else { |
| // A suspended transition to another effectively suspended transition, ok to use Unsafe. |
| self_->SetState(old_thread_state_); |
| } |
| } |
| } |
| } |
| |
| Thread* Self() const { |
| return self_; |
| } |
| |
| protected: |
| // Constructor used by ScopedJniThreadState for an unattached thread that has access to the VM*. |
| ScopedThreadStateChange() |
| : self_(NULL), thread_state_(kTerminated), old_thread_state_(kTerminated), |
| expected_has_no_thread_(true) {} |
| |
| Thread* const self_; |
| const ThreadState thread_state_; |
| |
| private: |
| ThreadState old_thread_state_; |
| const bool expected_has_no_thread_; |
| |
| DISALLOW_COPY_AND_ASSIGN(ScopedThreadStateChange); |
| }; |
| |
| // Entry/exit processing for transitions from Native to Runnable (ie within JNI functions). |
| // |
| // This class performs the necessary thread state switching to and from Runnable and lets us |
| // amortize the cost of working out the current thread. Additionally it lets us check (and repair) |
| // apps that are using a JNIEnv on the wrong thread. The class also decodes and encodes Objects |
| // into jobjects via methods of this class. Performing this here enforces the Runnable thread state |
| // for use of Object, thereby inhibiting the Object being modified by GC whilst native or VM code |
| // is also manipulating the Object. |
| // |
| // The destructor transitions back to the previous thread state, typically Native. In this state |
| // GC and thread suspension may occur. |
| // |
| // For annotalysis the subclass ScopedObjectAccess (below) makes it explicit that a shared of |
| // the mutator_lock_ will be acquired on construction. |
| class ScopedObjectAccessUnchecked : public ScopedThreadStateChange { |
| public: |
| explicit ScopedObjectAccessUnchecked(JNIEnv* env) |
| LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) ALWAYS_INLINE |
| : ScopedThreadStateChange(ThreadForEnv(env), kRunnable), |
| env_(reinterpret_cast<JNIEnvExt*>(env)), vm_(env_->vm) { |
| self_->VerifyStack(); |
| } |
| |
| explicit ScopedObjectAccessUnchecked(Thread* self) |
| LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) |
| : ScopedThreadStateChange(self, kRunnable), |
| env_(reinterpret_cast<JNIEnvExt*>(self->GetJniEnv())), |
| vm_(env_ != NULL ? env_->vm : NULL) { |
| self_->VerifyStack(); |
| } |
| |
| // Used when we want a scoped JNI thread state but have no thread/JNIEnv. Consequently doesn't |
| // change into Runnable or acquire a share on the mutator_lock_. |
| explicit ScopedObjectAccessUnchecked(JavaVM* vm) |
| : ScopedThreadStateChange(), env_(NULL), vm_(reinterpret_cast<JavaVMExt*>(vm)) {} |
| |
| // Here purely to force inlining. |
| ~ScopedObjectAccessUnchecked() ALWAYS_INLINE { |
| } |
| |
| JNIEnvExt* Env() const { |
| return env_; |
| } |
| |
| JavaVMExt* Vm() const { |
| return vm_; |
| } |
| |
| /* |
| * Add a local reference for an object to the indirect reference table associated with the |
| * current stack frame. When the native function returns, the reference will be discarded. |
| * |
| * We need to allow the same reference to be added multiple times, and cope with NULL. |
| * |
| * This will be called on otherwise unreferenced objects. We cannot do GC allocations here, and |
| * it's best if we don't grab a mutex. |
| */ |
| template<typename T> |
| T AddLocalReference(mirror::Object* obj) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| if (obj == NULL) { |
| return NULL; |
| } |
| |
| DCHECK_NE((reinterpret_cast<uintptr_t>(obj) & 0xffff0000), 0xebad0000); |
| |
| IndirectReferenceTable& locals = Env()->locals; |
| |
| uint32_t cookie = Env()->local_ref_cookie; |
| IndirectRef ref = locals.Add(cookie, obj); |
| |
| #if 0 // TODO: fix this to understand PushLocalFrame, so we can turn it on. |
| if (Env()->check_jni) { |
| size_t entry_count = locals.Capacity(); |
| if (entry_count > 16) { |
| LOG(WARNING) << "Warning: more than 16 JNI local references: " |
| << entry_count << " (most recent was a " << PrettyTypeOf(obj) << ")\n" |
| << Dumpable<IndirectReferenceTable>(locals); |
| // TODO: LOG(FATAL) in a later release? |
| } |
| } |
| #endif |
| |
| if (Vm()->work_around_app_jni_bugs) { |
| // Hand out direct pointers to support broken old apps. |
| return reinterpret_cast<T>(obj); |
| } |
| |
| return reinterpret_cast<T>(ref); |
| } |
| |
| template<typename T> |
| T Decode(jobject obj) const |
| LOCKS_EXCLUDED(JavaVMExt::globals_lock, |
| JavaVMExt::weak_globals_lock) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| return down_cast<T>(Self()->DecodeJObject(obj)); |
| } |
| |
| mirror::ArtField* DecodeField(jfieldID fid) const |
| LOCKS_EXCLUDED(JavaVMExt::globals_lock, |
| JavaVMExt::weak_globals_lock) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| #ifdef MOVING_GARBAGE_COLLECTOR |
| // TODO: we should make these unique weak globals if Field instances can ever move. |
| UNIMPLEMENTED(WARNING); |
| #endif |
| return reinterpret_cast<mirror::ArtField*>(fid); |
| } |
| |
| jfieldID EncodeField(mirror::ArtField* field) const |
| LOCKS_EXCLUDED(JavaVMExt::globals_lock, |
| JavaVMExt::weak_globals_lock) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| #ifdef MOVING_GARBAGE_COLLECTOR |
| UNIMPLEMENTED(WARNING); |
| #endif |
| return reinterpret_cast<jfieldID>(field); |
| } |
| |
| mirror::ArtMethod* DecodeMethod(jmethodID mid) const |
| LOCKS_EXCLUDED(JavaVMExt::globals_lock, |
| JavaVMExt::weak_globals_lock) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| #ifdef MOVING_GARBAGE_COLLECTOR |
| // TODO: we should make these unique weak globals if Method instances can ever move. |
| UNIMPLEMENTED(WARNING); |
| #endif |
| return reinterpret_cast<mirror::ArtMethod*>(mid); |
| } |
| |
| jmethodID EncodeMethod(mirror::ArtMethod* method) const |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| DCHECK_EQ(thread_state_, kRunnable); // Don't work with raw objects in non-runnable states. |
| #ifdef MOVING_GARBAGE_COLLECTOR |
| UNIMPLEMENTED(WARNING); |
| #endif |
| return reinterpret_cast<jmethodID>(method); |
| } |
| |
| private: |
| static Thread* ThreadForEnv(JNIEnv* env) { |
| JNIEnvExt* full_env(reinterpret_cast<JNIEnvExt*>(env)); |
| return full_env->self; |
| } |
| |
| // The full JNIEnv. |
| JNIEnvExt* const env_; |
| // The full JavaVM. |
| JavaVMExt* const vm_; |
| |
| DISALLOW_COPY_AND_ASSIGN(ScopedObjectAccessUnchecked); |
| }; |
| |
| // Annotalysis helping variant of the above. |
| class ScopedObjectAccess : public ScopedObjectAccessUnchecked { |
| public: |
| explicit ScopedObjectAccess(JNIEnv* env) |
| LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) |
| SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE |
| : ScopedObjectAccessUnchecked(env) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| } |
| |
| explicit ScopedObjectAccess(Thread* self) |
| LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_) |
| SHARED_LOCK_FUNCTION(Locks::mutator_lock_) |
| : ScopedObjectAccessUnchecked(self) { |
| Locks::mutator_lock_->AssertSharedHeld(Self()); |
| } |
| |
| ~ScopedObjectAccess() UNLOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE { |
| // Base class will release share of lock. Invoked after this destructor. |
| } |
| |
| private: |
| // TODO: remove this constructor. It is used by check JNI's ScopedCheck to make it believe that |
| // routines operating with just a VM are sound, they are not, but when you have just a VM |
| // you cannot call the unsound routines. |
| explicit ScopedObjectAccess(JavaVM* vm) |
| SHARED_LOCK_FUNCTION(Locks::mutator_lock_) |
| : ScopedObjectAccessUnchecked(vm) {} |
| |
| friend class ScopedCheck; |
| DISALLOW_COPY_AND_ASSIGN(ScopedObjectAccess); |
| }; |
| |
| } // namespace art |
| |
| #endif // ART_RUNTIME_SCOPED_THREAD_STATE_CHANGE_H_ |