/* * Copyright (C) 2011 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "trace.h" #include #include "class_linker.h" #include "debugger.h" #include "dex_cache.h" #if !defined(ART_USE_LLVM_COMPILER) #include "oat/runtime/oat_support_entrypoints.h" #endif #include "object_utils.h" #include "os.h" #include "scoped_thread_state_change.h" #include "thread.h" #include "thread_list.h" namespace art { // File format: // header // record 0 // record 1 // ... // // Header format: // u4 magic ('SLOW') // u2 version // u2 offset to data // u8 start date/time in usec // u2 record size in bytes (version >= 2 only) // ... padding to 32 bytes // // Record format v1: // u1 thread ID // u4 method ID | method action // u4 time delta since start, in usec // // Record format v2: // u2 thread ID // u4 method ID | method action // u4 time delta since start, in usec // // Record format v3: // u2 thread ID // u4 method ID | method action // u4 time delta since start, in usec // u4 wall time since start, in usec (when clock == "dual" only) // // 32 bits of microseconds is 70 minutes. // // All values are stored in little-endian order. static const uint32_t kTraceMethodActionMask = 0x03; // two bits static const char kTraceTokenChar = '*'; static const uint16_t kTraceHeaderLength = 32; static const uint32_t kTraceMagicValue = 0x574f4c53; static const uint16_t kTraceVersionSingleClock = 2; static const uint16_t kTraceVersionDualClock = 3; static const uint16_t kTraceRecordSizeSingleClock = 10; // using v2 static const uint16_t kTraceRecordSizeDualClock = 14; // using v3 with two timestamps static ProfilerClockSource gDefaultTraceClockSource = kProfilerClockSourceDual; static inline uint32_t TraceMethodId(uint32_t methodValue) { return (methodValue & ~kTraceMethodActionMask); } static inline uint32_t TraceMethodCombine(uint32_t method, uint8_t traceEvent) { return (method | traceEvent); } void Trace::SetDefaultClockSource(ProfilerClockSource clock_source) { gDefaultTraceClockSource = clock_source; } bool Trace::UseThreadCpuClock() { #if defined(HAVE_POSIX_CLOCKS) return clock_source_ != kProfilerClockSourceWall; #else return false; #endif } bool Trace::UseWallClock() { #if defined(HAVE_POSIX_CLOCKS) return clock_source_ != kProfilerClockSourceThreadCpu; #else return true; #endif } static void MeasureClockOverhead(Trace* trace) { if (trace->UseThreadCpuClock()) { ThreadCpuMicroTime(); } if (trace->UseWallClock()) { MicroTime(); } } static uint32_t GetClockOverhead(Trace* trace) { uint64_t start = ThreadCpuMicroTime(); for (int i = 4000; i > 0; i--) { MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); MeasureClockOverhead(trace); } uint64_t elapsed = ThreadCpuMicroTime() - start; return uint32_t (elapsed / 32); } // TODO: put this somewhere with the big-endian equivalent used by JDWP. static void Append2LE(uint8_t* buf, uint16_t val) { *buf++ = (uint8_t) val; *buf++ = (uint8_t) (val >> 8); } // TODO: put this somewhere with the big-endian equivalent used by JDWP. static void Append4LE(uint8_t* buf, uint32_t val) { *buf++ = (uint8_t) val; *buf++ = (uint8_t) (val >> 8); *buf++ = (uint8_t) (val >> 16); *buf++ = (uint8_t) (val >> 24); } // TODO: put this somewhere with the big-endian equivalent used by JDWP. static void Append8LE(uint8_t* buf, uint64_t val) { *buf++ = (uint8_t) val; *buf++ = (uint8_t) (val >> 8); *buf++ = (uint8_t) (val >> 16); *buf++ = (uint8_t) (val >> 24); *buf++ = (uint8_t) (val >> 32); *buf++ = (uint8_t) (val >> 40); *buf++ = (uint8_t) (val >> 48); *buf++ = (uint8_t) (val >> 56); } static bool InstallStubsClassVisitor(Class* klass, void*) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { Trace* tracer = Runtime::Current()->GetTracer(); for (size_t i = 0; i < klass->NumDirectMethods(); i++) { AbstractMethod* method = klass->GetDirectMethod(i); if (tracer->GetSavedCodeFromMap(method) == NULL) { tracer->SaveAndUpdateCode(method); } } for (size_t i = 0; i < klass->NumVirtualMethods(); i++) { AbstractMethod* method = klass->GetVirtualMethod(i); if (tracer->GetSavedCodeFromMap(method) == NULL) { tracer->SaveAndUpdateCode(method); } } return true; } static bool UninstallStubsClassVisitor(Class* klass, void*) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { Trace* tracer = Runtime::Current()->GetTracer(); for (size_t i = 0; i < klass->NumDirectMethods(); i++) { AbstractMethod* method = klass->GetDirectMethod(i); if (tracer->GetSavedCodeFromMap(method) != NULL) { tracer->ResetSavedCode(method); } } for (size_t i = 0; i < klass->NumVirtualMethods(); i++) { AbstractMethod* method = klass->GetVirtualMethod(i); if (tracer->GetSavedCodeFromMap(method) != NULL) { tracer->ResetSavedCode(method); } } return true; } static void TraceRestoreStack(Thread* self, void*) { struct RestoreStackVisitor : public StackVisitor { RestoreStackVisitor(Thread* self) : StackVisitor(self->GetManagedStack(), self->GetTraceStack(), NULL), self_(self) {} virtual bool VisitFrame() { if (self_->IsTraceStackEmpty()) { return false; // Stop. } uintptr_t pc = GetReturnPc(); if (IsTraceExitPc(pc)) { TraceStackFrame trace_frame = self_->PopTraceStackFrame(); SetReturnPc(trace_frame.return_pc_); CHECK(GetMethod() == trace_frame.method_); } return true; // Continue. } Thread* self_; }; ScopedObjectAccess soa(self); RestoreStackVisitor visitor(self); visitor.WalkStack(); } void Trace::AddSavedCodeToMap(const AbstractMethod* method, const void* code) { saved_code_map_.Put(method, code); } void Trace::RemoveSavedCodeFromMap(const AbstractMethod* method) { saved_code_map_.erase(method); } const void* Trace::GetSavedCodeFromMap(const AbstractMethod* method) { typedef SafeMap::const_iterator It; // TODO: C++0x auto It it = saved_code_map_.find(method); if (it == saved_code_map_.end()) { return NULL; } else { return it->second; } } void Trace::SaveAndUpdateCode(AbstractMethod* method) { #if defined(ART_USE_LLVM_COMPILER) UNUSED(method); UNIMPLEMENTED(FATAL); #else void* trace_stub = GetLogTraceEntryPoint(); CHECK(GetSavedCodeFromMap(method) == NULL); AddSavedCodeToMap(method, method->GetCode()); method->SetCode(trace_stub); #endif } void Trace::ResetSavedCode(AbstractMethod* method) { CHECK(GetSavedCodeFromMap(method) != NULL); method->SetCode(GetSavedCodeFromMap(method)); RemoveSavedCodeFromMap(method); } Trace::Trace(File* trace_file, int buffer_size, int flags) : trace_file_(trace_file), buf_(new uint8_t[buffer_size]()), flags_(flags), clock_source_(gDefaultTraceClockSource), overflow_(false), buffer_size_(buffer_size), start_time_(0), trace_version_(0), record_size_(0), cur_offset_(0) { } void Trace::Start(const char* trace_filename, int trace_fd, int buffer_size, int flags, bool direct_to_ddms) { if (Runtime::Current()->IsMethodTracingActive()) { LOG(INFO) << "Trace already in progress, ignoring this request"; return; } ScopedThreadStateChange tsc(Thread::Current(), kRunnable); Runtime::Current()->GetThreadList()->SuspendAll(); // Open trace file if not going directly to ddms. File* trace_file = NULL; if (!direct_to_ddms) { if (trace_fd < 0) { trace_file = OS::OpenFile(trace_filename, true); } else { trace_file = OS::FileFromFd("tracefile", trace_fd); } if (trace_file == NULL) { PLOG(ERROR) << "Unable to open trace file '" << trace_filename << "'"; Thread::Current()->ThrowNewException("Ljava/lang/RuntimeException;", StringPrintf("Unable to open trace file '%s'", trace_filename).c_str()); Runtime::Current()->GetThreadList()->ResumeAll(); return; } } // Create Trace object. Trace* tracer(new Trace(trace_file, buffer_size, flags)); // Enable count of allocs if specified in the flags. if ((flags && kTraceCountAllocs) != 0) { Runtime::Current()->SetStatsEnabled(true); } Runtime::Current()->EnableMethodTracing(tracer); tracer->BeginTracing(); Runtime::Current()->GetThreadList()->ResumeAll(); } void Trace::Stop() { if (!Runtime::Current()->IsMethodTracingActive()) { LOG(INFO) << "Trace stop requested, but no trace currently running"; return; } ScopedThreadStateChange tsc(Thread::Current(), kRunnable); Runtime::Current()->GetThreadList()->SuspendAll(); Runtime::Current()->GetTracer()->FinishTracing(); Runtime::Current()->DisableMethodTracing(); Runtime::Current()->GetThreadList()->ResumeAll(); } void Trace::Shutdown() { if (!Runtime::Current()->IsMethodTracingActive()) { LOG(INFO) << "Trace shutdown requested, but no trace currently running"; return; } Runtime::Current()->GetTracer()->FinishTracing(); Runtime::Current()->DisableMethodTracing(); } void Trace::BeginTracing() { // Set the start time of tracing. start_time_ = MicroTime(); // Set trace version and record size. if (UseThreadCpuClock() && UseWallClock()) { trace_version_ = kTraceVersionDualClock; record_size_ = kTraceRecordSizeDualClock; } else { trace_version_ = kTraceVersionSingleClock; record_size_ = kTraceRecordSizeSingleClock; } // Set up the beginning of the trace. memset(buf_.get(), 0, kTraceHeaderLength); Append4LE(buf_.get(), kTraceMagicValue); Append2LE(buf_.get() + 4, trace_version_); Append2LE(buf_.get() + 6, kTraceHeaderLength); Append8LE(buf_.get() + 8, start_time_); if (trace_version_ >= kTraceVersionDualClock) { Append2LE(buf_.get() + 16, record_size_); } // Update current offset. cur_offset_ = kTraceHeaderLength; // Install all method tracing stubs. InstallStubs(); } void Trace::FinishTracing() { // Uninstall all method tracing stubs. UninstallStubs(); // Compute elapsed time. uint64_t elapsed = MicroTime() - start_time_; size_t final_offset = cur_offset_; uint32_t clock_overhead = GetClockOverhead(this); if ((flags_ & kTraceCountAllocs) != 0) { Runtime::Current()->SetStatsEnabled(false); } GetVisitedMethods(final_offset); std::ostringstream os; os << StringPrintf("%cversion\n", kTraceTokenChar); os << StringPrintf("%d\n", trace_version_); os << StringPrintf("data-file-overflow=%s\n", overflow_ ? "true" : "false"); if (UseThreadCpuClock()) { if (UseWallClock()) { os << StringPrintf("clock=dual\n"); } else { os << StringPrintf("clock=thread-cpu\n"); } } else { os << StringPrintf("clock=wall\n"); } os << StringPrintf("elapsed-time-usec=%llu\n", elapsed); os << StringPrintf("num-method-calls=%zd\n", (final_offset - kTraceHeaderLength) / record_size_); os << StringPrintf("clock-call-overhead-nsec=%d\n", clock_overhead); os << StringPrintf("vm=art\n"); if ((flags_ & kTraceCountAllocs) != 0) { os << StringPrintf("alloc-count=%d\n", Runtime::Current()->GetStat(KIND_ALLOCATED_OBJECTS)); os << StringPrintf("alloc-size=%d\n", Runtime::Current()->GetStat(KIND_ALLOCATED_BYTES)); os << StringPrintf("gc-count=%d\n", Runtime::Current()->GetStat(KIND_GC_INVOCATIONS)); } os << StringPrintf("%cthreads\n", kTraceTokenChar); DumpThreadList(os); os << StringPrintf("%cmethods\n", kTraceTokenChar); DumpMethodList(os); os << StringPrintf("%cend\n", kTraceTokenChar); std::string header(os.str()); if (trace_file_.get() == NULL) { iovec iov[2]; iov[0].iov_base = reinterpret_cast(const_cast(header.c_str())); iov[0].iov_len = header.length(); iov[1].iov_base = buf_.get(); iov[1].iov_len = final_offset; Dbg::DdmSendChunkV(CHUNK_TYPE("MPSE"), iov, 2); } else { if (!trace_file_->WriteFully(header.c_str(), header.length()) || !trace_file_->WriteFully(buf_.get(), final_offset)) { std::string detail(StringPrintf("Trace data write failed: %s", strerror(errno))); PLOG(ERROR) << detail; Thread::Current()->ThrowNewException("Ljava/lang/RuntimeException;", detail.c_str()); } } } void Trace::LogMethodTraceEvent(Thread* self, const AbstractMethod* method, Trace::TraceEvent event) { if (thread_clock_base_map_.find(self) == thread_clock_base_map_.end()) { uint64_t time = ThreadCpuMicroTime(); thread_clock_base_map_.Put(self, time); } // Advance cur_offset_ atomically. int32_t new_offset; int32_t old_offset; do { old_offset = cur_offset_; new_offset = old_offset + record_size_; if (new_offset > buffer_size_) { overflow_ = true; return; } } while (android_atomic_release_cas(old_offset, new_offset, &cur_offset_) != 0); uint32_t method_value = TraceMethodCombine(reinterpret_cast(method), event); // Write data uint8_t* ptr = buf_.get() + old_offset; Append2LE(ptr, self->GetTid()); Append4LE(ptr + 2, method_value); ptr += 6; if (UseThreadCpuClock()) { uint64_t thread_clock_base = thread_clock_base_map_.find(self)->second; uint32_t thread_clock_diff = ThreadCpuMicroTime() - thread_clock_base; Append4LE(ptr, thread_clock_diff); ptr += 4; } if (UseWallClock()) { uint32_t wall_clock_diff = MicroTime() - start_time_; Append4LE(ptr, wall_clock_diff); } } void Trace::GetVisitedMethods(size_t end_offset) { uint8_t* ptr = buf_.get() + kTraceHeaderLength; uint8_t* end = buf_.get() + end_offset; while (ptr < end) { uint32_t method_value = ptr[2] | (ptr[3] << 8) | (ptr[4] << 16) | (ptr[5] << 24); AbstractMethod* method = reinterpret_cast(TraceMethodId(method_value)); visited_methods_.insert(method); ptr += record_size_; } } void Trace::DumpMethodList(std::ostream& os) { typedef std::set::const_iterator It; // TODO: C++0x auto for (It it = visited_methods_.begin(); it != visited_methods_.end(); ++it) { const AbstractMethod* method = *it; MethodHelper mh(method); os << StringPrintf("%p\t%s\t%s\t%s\t%s\n", method, PrettyDescriptor(mh.GetDeclaringClassDescriptor()).c_str(), mh.GetName(), mh.GetSignature().c_str(), mh.GetDeclaringClassSourceFile()); } } static void DumpThread(Thread* t, void* arg) { std::ostream& os = *reinterpret_cast(arg); std::string name; t->GetThreadName(name); os << t->GetTid() << "\t" << name << "\n"; } void Trace::DumpThreadList(std::ostream& os) { Thread* self = Thread::Current(); Locks::thread_list_lock_->AssertNotHeld(self); MutexLock mu(self, *Locks::thread_list_lock_); Runtime::Current()->GetThreadList()->ForEach(DumpThread, &os); } void Trace::InstallStubs() { Runtime::Current()->GetClassLinker()->VisitClasses(InstallStubsClassVisitor, NULL); } void Trace::UninstallStubs() { Thread* self = Thread::Current(); Locks::thread_list_lock_->AssertNotHeld(self); Runtime::Current()->GetClassLinker()->VisitClasses(UninstallStubsClassVisitor, NULL); MutexLock mu(self, *Locks::thread_list_lock_); Runtime::Current()->GetThreadList()->ForEach(TraceRestoreStack, NULL); } uint32_t TraceMethodUnwindFromCode(Thread* self) { Trace* tracer = Runtime::Current()->GetTracer(); TraceStackFrame trace_frame = self->PopTraceStackFrame(); AbstractMethod* method = trace_frame.method_; uint32_t lr = trace_frame.return_pc_; tracer->LogMethodTraceEvent(self, method, Trace::kMethodTraceUnwind); return lr; } } // namespace art