| /* |
| * Copyright (C) 2013 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define _GNU_SOURCE 1 |
| #include <dirent.h> |
| #include <dlfcn.h> |
| #include <errno.h> |
| #include <fcntl.h> |
| #include <inttypes.h> |
| #include <pthread.h> |
| #include <signal.h> |
| #include <stdint.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <string.h> |
| #include <sys/ptrace.h> |
| #include <sys/stat.h> |
| #include <sys/types.h> |
| #include <sys/wait.h> |
| #include <time.h> |
| #include <unistd.h> |
| |
| #include <algorithm> |
| #include <list> |
| #include <memory> |
| #include <string> |
| #include <vector> |
| |
| #include <backtrace/Backtrace.h> |
| #include <backtrace/BacktraceMap.h> |
| |
| #include <base/stringprintf.h> |
| #include <cutils/atomic.h> |
| #include <cutils/threads.h> |
| |
| #include <gtest/gtest.h> |
| |
| // For the THREAD_SIGNAL definition. |
| #include "BacktraceCurrent.h" |
| #include "thread_utils.h" |
| |
| // Number of microseconds per milliseconds. |
| #define US_PER_MSEC 1000 |
| |
| // Number of nanoseconds in a second. |
| #define NS_PER_SEC 1000000000ULL |
| |
| // Number of simultaneous dumping operations to perform. |
| #define NUM_THREADS 40 |
| |
| // Number of simultaneous threads running in our forked process. |
| #define NUM_PTRACE_THREADS 5 |
| |
| struct thread_t { |
| pid_t tid; |
| int32_t state; |
| pthread_t threadId; |
| void* data; |
| }; |
| |
| struct dump_thread_t { |
| thread_t thread; |
| Backtrace* backtrace; |
| int32_t* now; |
| int32_t done; |
| }; |
| |
| extern "C" { |
| // Prototypes for functions in the test library. |
| int test_level_one(int, int, int, int, void (*)(void*), void*); |
| |
| int test_recursive_call(int, void (*)(void*), void*); |
| } |
| |
| uint64_t NanoTime() { |
| struct timespec t = { 0, 0 }; |
| clock_gettime(CLOCK_MONOTONIC, &t); |
| return static_cast<uint64_t>(t.tv_sec * NS_PER_SEC + t.tv_nsec); |
| } |
| |
| std::string DumpFrames(Backtrace* backtrace) { |
| if (backtrace->NumFrames() == 0) { |
| return " No frames to dump.\n"; |
| } |
| |
| std::string frame; |
| for (size_t i = 0; i < backtrace->NumFrames(); i++) { |
| frame += " " + backtrace->FormatFrameData(i) + '\n'; |
| } |
| return frame; |
| } |
| |
| void WaitForStop(pid_t pid) { |
| uint64_t start = NanoTime(); |
| |
| siginfo_t si; |
| while (ptrace(PTRACE_GETSIGINFO, pid, 0, &si) < 0 && (errno == EINTR || errno == ESRCH)) { |
| if ((NanoTime() - start) > NS_PER_SEC) { |
| printf("The process did not get to a stopping point in 1 second.\n"); |
| break; |
| } |
| usleep(US_PER_MSEC); |
| } |
| } |
| |
| bool ReadyLevelBacktrace(Backtrace* backtrace) { |
| // See if test_level_four is in the backtrace. |
| bool found = false; |
| for (Backtrace::const_iterator it = backtrace->begin(); it != backtrace->end(); ++it) { |
| if (it->func_name == "test_level_four") { |
| found = true; |
| break; |
| } |
| } |
| |
| return found; |
| } |
| |
| void VerifyLevelDump(Backtrace* backtrace) { |
| ASSERT_GT(backtrace->NumFrames(), static_cast<size_t>(0)) |
| << DumpFrames(backtrace); |
| ASSERT_LT(backtrace->NumFrames(), static_cast<size_t>(MAX_BACKTRACE_FRAMES)) |
| << DumpFrames(backtrace); |
| |
| // Look through the frames starting at the highest to find the |
| // frame we want. |
| size_t frame_num = 0; |
| for (size_t i = backtrace->NumFrames()-1; i > 2; i--) { |
| if (backtrace->GetFrame(i)->func_name == "test_level_one") { |
| frame_num = i; |
| break; |
| } |
| } |
| ASSERT_LT(static_cast<size_t>(0), frame_num) << DumpFrames(backtrace); |
| ASSERT_LE(static_cast<size_t>(3), frame_num) << DumpFrames(backtrace); |
| |
| ASSERT_EQ(backtrace->GetFrame(frame_num)->func_name, "test_level_one") |
| << DumpFrames(backtrace); |
| ASSERT_EQ(backtrace->GetFrame(frame_num-1)->func_name, "test_level_two") |
| << DumpFrames(backtrace); |
| ASSERT_EQ(backtrace->GetFrame(frame_num-2)->func_name, "test_level_three") |
| << DumpFrames(backtrace); |
| ASSERT_EQ(backtrace->GetFrame(frame_num-3)->func_name, "test_level_four") |
| << DumpFrames(backtrace); |
| } |
| |
| void VerifyLevelBacktrace(void*) { |
| std::unique_ptr<Backtrace> backtrace( |
| Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyLevelDump(backtrace.get()); |
| } |
| |
| bool ReadyMaxBacktrace(Backtrace* backtrace) { |
| return (backtrace->NumFrames() == MAX_BACKTRACE_FRAMES); |
| } |
| |
| void VerifyMaxDump(Backtrace* backtrace) { |
| ASSERT_EQ(backtrace->NumFrames(), static_cast<size_t>(MAX_BACKTRACE_FRAMES)) |
| << DumpFrames(backtrace); |
| // Verify that the last frame is our recursive call. |
| ASSERT_EQ(backtrace->GetFrame(MAX_BACKTRACE_FRAMES-1)->func_name, "test_recursive_call") |
| << DumpFrames(backtrace); |
| } |
| |
| void VerifyMaxBacktrace(void*) { |
| std::unique_ptr<Backtrace> backtrace( |
| Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyMaxDump(backtrace.get()); |
| } |
| |
| void ThreadSetState(void* data) { |
| thread_t* thread = reinterpret_cast<thread_t*>(data); |
| android_atomic_acquire_store(1, &thread->state); |
| volatile int i = 0; |
| while (thread->state) { |
| i++; |
| } |
| } |
| |
| void VerifyThreadTest(pid_t tid, void (*VerifyFunc)(Backtrace*)) { |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), tid)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyFunc(backtrace.get()); |
| } |
| |
| bool WaitForNonZero(int32_t* value, uint64_t seconds) { |
| uint64_t start = NanoTime(); |
| do { |
| if (android_atomic_acquire_load(value)) { |
| return true; |
| } |
| } while ((NanoTime() - start) < seconds * NS_PER_SEC); |
| return false; |
| } |
| |
| TEST(libbacktrace, local_no_unwind_frames) { |
| // Verify that a local unwind does not include any frames within |
| // libunwind or libbacktrace. |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), getpid())); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| ASSERT_TRUE(backtrace->NumFrames() != 0); |
| for (const auto& frame : *backtrace ) { |
| if (BacktraceMap::IsValid(frame.map)) { |
| const std::string name = basename(frame.map.name.c_str()); |
| ASSERT_TRUE(name != "libunwind.so" && name != "libbacktrace.so") |
| << DumpFrames(backtrace.get()); |
| } |
| break; |
| } |
| } |
| |
| TEST(libbacktrace, local_trace) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelBacktrace, nullptr), 0); |
| } |
| |
| void VerifyIgnoreFrames( |
| Backtrace* bt_all, Backtrace* bt_ign1, |
| Backtrace* bt_ign2, const char* cur_proc) { |
| EXPECT_EQ(bt_all->NumFrames(), bt_ign1->NumFrames() + 1) |
| << "All backtrace:\n" << DumpFrames(bt_all) << "Ignore 1 backtrace:\n" << DumpFrames(bt_ign1); |
| EXPECT_EQ(bt_all->NumFrames(), bt_ign2->NumFrames() + 2) |
| << "All backtrace:\n" << DumpFrames(bt_all) << "Ignore 2 backtrace:\n" << DumpFrames(bt_ign2); |
| |
| // Check all of the frames are the same > the current frame. |
| bool check = (cur_proc == nullptr); |
| for (size_t i = 0; i < bt_ign2->NumFrames(); i++) { |
| if (check) { |
| EXPECT_EQ(bt_ign2->GetFrame(i)->pc, bt_ign1->GetFrame(i+1)->pc); |
| EXPECT_EQ(bt_ign2->GetFrame(i)->sp, bt_ign1->GetFrame(i+1)->sp); |
| EXPECT_EQ(bt_ign2->GetFrame(i)->stack_size, bt_ign1->GetFrame(i+1)->stack_size); |
| |
| EXPECT_EQ(bt_ign2->GetFrame(i)->pc, bt_all->GetFrame(i+2)->pc); |
| EXPECT_EQ(bt_ign2->GetFrame(i)->sp, bt_all->GetFrame(i+2)->sp); |
| EXPECT_EQ(bt_ign2->GetFrame(i)->stack_size, bt_all->GetFrame(i+2)->stack_size); |
| } |
| if (!check && bt_ign2->GetFrame(i)->func_name == cur_proc) { |
| check = true; |
| } |
| } |
| } |
| |
| void VerifyLevelIgnoreFrames(void*) { |
| std::unique_ptr<Backtrace> all( |
| Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(all.get() != nullptr); |
| ASSERT_TRUE(all->Unwind(0)); |
| |
| std::unique_ptr<Backtrace> ign1( |
| Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(ign1.get() != nullptr); |
| ASSERT_TRUE(ign1->Unwind(1)); |
| |
| std::unique_ptr<Backtrace> ign2( |
| Backtrace::Create(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(ign2.get() != nullptr); |
| ASSERT_TRUE(ign2->Unwind(2)); |
| |
| VerifyIgnoreFrames(all.get(), ign1.get(), ign2.get(), "VerifyLevelIgnoreFrames"); |
| } |
| |
| TEST(libbacktrace, local_trace_ignore_frames) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelIgnoreFrames, nullptr), 0); |
| } |
| |
| TEST(libbacktrace, local_max_trace) { |
| ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, VerifyMaxBacktrace, nullptr), 0); |
| } |
| |
| void VerifyProcTest(pid_t pid, pid_t tid, bool share_map, |
| bool (*ReadyFunc)(Backtrace*), |
| void (*VerifyFunc)(Backtrace*)) { |
| pid_t ptrace_tid; |
| if (tid < 0) { |
| ptrace_tid = pid; |
| } else { |
| ptrace_tid = tid; |
| } |
| uint64_t start = NanoTime(); |
| bool verified = false; |
| std::string last_dump; |
| do { |
| usleep(US_PER_MSEC); |
| if (ptrace(PTRACE_ATTACH, ptrace_tid, 0, 0) == 0) { |
| // Wait for the process to get to a stopping point. |
| WaitForStop(ptrace_tid); |
| |
| std::unique_ptr<BacktraceMap> map; |
| if (share_map) { |
| map.reset(BacktraceMap::Create(pid)); |
| } |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(pid, tid, map.get())); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| if (ReadyFunc(backtrace.get())) { |
| VerifyFunc(backtrace.get()); |
| verified = true; |
| } else { |
| last_dump = DumpFrames(backtrace.get()); |
| } |
| |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, ptrace_tid, 0, 0) == 0); |
| } |
| // If 5 seconds have passed, then we are done. |
| } while (!verified && (NanoTime() - start) <= 5 * NS_PER_SEC); |
| ASSERT_TRUE(verified) << "Last backtrace:\n" << last_dump; |
| } |
| |
| TEST(libbacktrace, ptrace_trace) { |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, nullptr, nullptr), 0); |
| _exit(1); |
| } |
| VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyLevelBacktrace, VerifyLevelDump); |
| |
| kill(pid, SIGKILL); |
| int status; |
| ASSERT_EQ(waitpid(pid, &status, 0), pid); |
| } |
| |
| TEST(libbacktrace, ptrace_trace_shared_map) { |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, nullptr, nullptr), 0); |
| _exit(1); |
| } |
| |
| VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, true, ReadyLevelBacktrace, VerifyLevelDump); |
| |
| kill(pid, SIGKILL); |
| int status; |
| ASSERT_EQ(waitpid(pid, &status, 0), pid); |
| } |
| |
| TEST(libbacktrace, ptrace_max_trace) { |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, nullptr, nullptr), 0); |
| _exit(1); |
| } |
| VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyMaxBacktrace, VerifyMaxDump); |
| |
| kill(pid, SIGKILL); |
| int status; |
| ASSERT_EQ(waitpid(pid, &status, 0), pid); |
| } |
| |
| void VerifyProcessIgnoreFrames(Backtrace* bt_all) { |
| std::unique_ptr<Backtrace> ign1(Backtrace::Create(bt_all->Pid(), BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(ign1.get() != nullptr); |
| ASSERT_TRUE(ign1->Unwind(1)); |
| |
| std::unique_ptr<Backtrace> ign2(Backtrace::Create(bt_all->Pid(), BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(ign2.get() != nullptr); |
| ASSERT_TRUE(ign2->Unwind(2)); |
| |
| VerifyIgnoreFrames(bt_all, ign1.get(), ign2.get(), nullptr); |
| } |
| |
| TEST(libbacktrace, ptrace_ignore_frames) { |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, nullptr, nullptr), 0); |
| _exit(1); |
| } |
| VerifyProcTest(pid, BACKTRACE_CURRENT_THREAD, false, ReadyLevelBacktrace, VerifyProcessIgnoreFrames); |
| |
| kill(pid, SIGKILL); |
| int status; |
| ASSERT_EQ(waitpid(pid, &status, 0), pid); |
| } |
| |
| // Create a process with multiple threads and dump all of the threads. |
| void* PtraceThreadLevelRun(void*) { |
| EXPECT_NE(test_level_one(1, 2, 3, 4, nullptr, nullptr), 0); |
| return nullptr; |
| } |
| |
| void GetThreads(pid_t pid, std::vector<pid_t>* threads) { |
| // Get the list of tasks. |
| char task_path[128]; |
| snprintf(task_path, sizeof(task_path), "/proc/%d/task", pid); |
| |
| DIR* tasks_dir = opendir(task_path); |
| ASSERT_TRUE(tasks_dir != nullptr); |
| struct dirent* entry; |
| while ((entry = readdir(tasks_dir)) != nullptr) { |
| char* end; |
| pid_t tid = strtoul(entry->d_name, &end, 10); |
| if (*end == '\0') { |
| threads->push_back(tid); |
| } |
| } |
| closedir(tasks_dir); |
| } |
| |
| TEST(libbacktrace, ptrace_threads) { |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| for (size_t i = 0; i < NUM_PTRACE_THREADS; i++) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| |
| pthread_t thread; |
| ASSERT_TRUE(pthread_create(&thread, &attr, PtraceThreadLevelRun, nullptr) == 0); |
| } |
| ASSERT_NE(test_level_one(1, 2, 3, 4, nullptr, nullptr), 0); |
| _exit(1); |
| } |
| |
| // Check to see that all of the threads are running before unwinding. |
| std::vector<pid_t> threads; |
| uint64_t start = NanoTime(); |
| do { |
| usleep(US_PER_MSEC); |
| threads.clear(); |
| GetThreads(pid, &threads); |
| } while ((threads.size() != NUM_PTRACE_THREADS + 1) && |
| ((NanoTime() - start) <= 5 * NS_PER_SEC)); |
| ASSERT_EQ(threads.size(), static_cast<size_t>(NUM_PTRACE_THREADS + 1)); |
| |
| ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0); |
| WaitForStop(pid); |
| for (std::vector<int>::const_iterator it = threads.begin(); it != threads.end(); ++it) { |
| // Skip the current forked process, we only care about the threads. |
| if (pid == *it) { |
| continue; |
| } |
| VerifyProcTest(pid, *it, false, ReadyLevelBacktrace, VerifyLevelDump); |
| } |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| |
| kill(pid, SIGKILL); |
| int status; |
| ASSERT_EQ(waitpid(pid, &status, 0), pid); |
| } |
| |
| void VerifyLevelThread(void*) { |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), gettid())); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyLevelDump(backtrace.get()); |
| } |
| |
| TEST(libbacktrace, thread_current_level) { |
| ASSERT_NE(test_level_one(1, 2, 3, 4, VerifyLevelThread, nullptr), 0); |
| } |
| |
| void VerifyMaxThread(void*) { |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), gettid())); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyMaxDump(backtrace.get()); |
| } |
| |
| TEST(libbacktrace, thread_current_max) { |
| ASSERT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, VerifyMaxThread, nullptr), 0); |
| } |
| |
| void* ThreadLevelRun(void* data) { |
| thread_t* thread = reinterpret_cast<thread_t*>(data); |
| |
| thread->tid = gettid(); |
| EXPECT_NE(test_level_one(1, 2, 3, 4, ThreadSetState, data), 0); |
| return nullptr; |
| } |
| |
| TEST(libbacktrace, thread_level_trace) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| |
| thread_t thread_data = { 0, 0, 0, nullptr }; |
| pthread_t thread; |
| ASSERT_TRUE(pthread_create(&thread, &attr, ThreadLevelRun, &thread_data) == 0); |
| |
| // Wait up to 2 seconds for the tid to be set. |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2)); |
| |
| // Make sure that the thread signal used is not visible when compiled for |
| // the target. |
| #if !defined(__GLIBC__) |
| ASSERT_LT(THREAD_SIGNAL, SIGRTMIN); |
| #endif |
| |
| // Save the current signal action and make sure it is restored afterwards. |
| struct sigaction cur_action; |
| ASSERT_TRUE(sigaction(THREAD_SIGNAL, nullptr, &cur_action) == 0); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyLevelDump(backtrace.get()); |
| |
| // Tell the thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &thread_data.state); |
| |
| // Verify that the old action was restored. |
| struct sigaction new_action; |
| ASSERT_TRUE(sigaction(THREAD_SIGNAL, nullptr, &new_action) == 0); |
| EXPECT_EQ(cur_action.sa_sigaction, new_action.sa_sigaction); |
| // The SA_RESTORER flag gets set behind our back, so a direct comparison |
| // doesn't work unless we mask the value off. Mips doesn't have this |
| // flag, so skip this on that platform. |
| #if defined(SA_RESTORER) |
| cur_action.sa_flags &= ~SA_RESTORER; |
| new_action.sa_flags &= ~SA_RESTORER; |
| #elif defined(__GLIBC__) |
| // Our host compiler doesn't appear to define this flag for some reason. |
| cur_action.sa_flags &= ~0x04000000; |
| new_action.sa_flags &= ~0x04000000; |
| #endif |
| EXPECT_EQ(cur_action.sa_flags, new_action.sa_flags); |
| } |
| |
| TEST(libbacktrace, thread_ignore_frames) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| |
| thread_t thread_data = { 0, 0, 0, nullptr }; |
| pthread_t thread; |
| ASSERT_TRUE(pthread_create(&thread, &attr, ThreadLevelRun, &thread_data) == 0); |
| |
| // Wait up to 2 seconds for the tid to be set. |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2)); |
| |
| std::unique_ptr<Backtrace> all(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(all.get() != nullptr); |
| ASSERT_TRUE(all->Unwind(0)); |
| |
| std::unique_ptr<Backtrace> ign1(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(ign1.get() != nullptr); |
| ASSERT_TRUE(ign1->Unwind(1)); |
| |
| std::unique_ptr<Backtrace> ign2(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(ign2.get() != nullptr); |
| ASSERT_TRUE(ign2->Unwind(2)); |
| |
| VerifyIgnoreFrames(all.get(), ign1.get(), ign2.get(), nullptr); |
| |
| // Tell the thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &thread_data.state); |
| } |
| |
| void* ThreadMaxRun(void* data) { |
| thread_t* thread = reinterpret_cast<thread_t*>(data); |
| |
| thread->tid = gettid(); |
| EXPECT_NE(test_recursive_call(MAX_BACKTRACE_FRAMES+10, ThreadSetState, data), 0); |
| return nullptr; |
| } |
| |
| TEST(libbacktrace, thread_max_trace) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| |
| thread_t thread_data = { 0, 0, 0, nullptr }; |
| pthread_t thread; |
| ASSERT_TRUE(pthread_create(&thread, &attr, ThreadMaxRun, &thread_data) == 0); |
| |
| // Wait for the tid to be set. |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2)); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| VerifyMaxDump(backtrace.get()); |
| |
| // Tell the thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &thread_data.state); |
| } |
| |
| void* ThreadDump(void* data) { |
| dump_thread_t* dump = reinterpret_cast<dump_thread_t*>(data); |
| while (true) { |
| if (android_atomic_acquire_load(dump->now)) { |
| break; |
| } |
| } |
| |
| // The status of the actual unwind will be checked elsewhere. |
| dump->backtrace = Backtrace::Create(getpid(), dump->thread.tid); |
| dump->backtrace->Unwind(0); |
| |
| android_atomic_acquire_store(1, &dump->done); |
| |
| return nullptr; |
| } |
| |
| TEST(libbacktrace, thread_multiple_dump) { |
| // Dump NUM_THREADS simultaneously. |
| std::vector<thread_t> runners(NUM_THREADS); |
| std::vector<dump_thread_t> dumpers(NUM_THREADS); |
| |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| for (size_t i = 0; i < NUM_THREADS; i++) { |
| // Launch the runners, they will spin in hard loops doing nothing. |
| runners[i].tid = 0; |
| runners[i].state = 0; |
| ASSERT_TRUE(pthread_create(&runners[i].threadId, &attr, ThreadMaxRun, &runners[i]) == 0); |
| } |
| |
| // Wait for tids to be set. |
| for (std::vector<thread_t>::iterator it = runners.begin(); it != runners.end(); ++it) { |
| ASSERT_TRUE(WaitForNonZero(&it->state, 30)); |
| } |
| |
| // Start all of the dumpers at once, they will spin until they are signalled |
| // to begin their dump run. |
| int32_t dump_now = 0; |
| for (size_t i = 0; i < NUM_THREADS; i++) { |
| dumpers[i].thread.tid = runners[i].tid; |
| dumpers[i].thread.state = 0; |
| dumpers[i].done = 0; |
| dumpers[i].now = &dump_now; |
| |
| ASSERT_TRUE(pthread_create(&dumpers[i].thread.threadId, &attr, ThreadDump, &dumpers[i]) == 0); |
| } |
| |
| // Start all of the dumpers going at once. |
| android_atomic_acquire_store(1, &dump_now); |
| |
| for (size_t i = 0; i < NUM_THREADS; i++) { |
| ASSERT_TRUE(WaitForNonZero(&dumpers[i].done, 30)); |
| |
| // Tell the runner thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &runners[i].state); |
| |
| ASSERT_TRUE(dumpers[i].backtrace != nullptr); |
| VerifyMaxDump(dumpers[i].backtrace); |
| |
| delete dumpers[i].backtrace; |
| dumpers[i].backtrace = nullptr; |
| } |
| } |
| |
| TEST(libbacktrace, thread_multiple_dump_same_thread) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| thread_t runner; |
| runner.tid = 0; |
| runner.state = 0; |
| ASSERT_TRUE(pthread_create(&runner.threadId, &attr, ThreadMaxRun, &runner) == 0); |
| |
| // Wait for tids to be set. |
| ASSERT_TRUE(WaitForNonZero(&runner.state, 30)); |
| |
| // Start all of the dumpers at once, they will spin until they are signalled |
| // to begin their dump run. |
| int32_t dump_now = 0; |
| // Dump the same thread NUM_THREADS simultaneously. |
| std::vector<dump_thread_t> dumpers(NUM_THREADS); |
| for (size_t i = 0; i < NUM_THREADS; i++) { |
| dumpers[i].thread.tid = runner.tid; |
| dumpers[i].thread.state = 0; |
| dumpers[i].done = 0; |
| dumpers[i].now = &dump_now; |
| |
| ASSERT_TRUE(pthread_create(&dumpers[i].thread.threadId, &attr, ThreadDump, &dumpers[i]) == 0); |
| } |
| |
| // Start all of the dumpers going at once. |
| android_atomic_acquire_store(1, &dump_now); |
| |
| for (size_t i = 0; i < NUM_THREADS; i++) { |
| ASSERT_TRUE(WaitForNonZero(&dumpers[i].done, 30)); |
| |
| ASSERT_TRUE(dumpers[i].backtrace != nullptr); |
| VerifyMaxDump(dumpers[i].backtrace); |
| |
| delete dumpers[i].backtrace; |
| dumpers[i].backtrace = nullptr; |
| } |
| |
| // Tell the runner thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &runner.state); |
| } |
| |
| // This test is for UnwindMaps that should share the same map cursor when |
| // multiple maps are created for the current process at the same time. |
| TEST(libbacktrace, simultaneous_maps) { |
| BacktraceMap* map1 = BacktraceMap::Create(getpid()); |
| BacktraceMap* map2 = BacktraceMap::Create(getpid()); |
| BacktraceMap* map3 = BacktraceMap::Create(getpid()); |
| |
| Backtrace* back1 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map1); |
| ASSERT_TRUE(back1 != nullptr); |
| EXPECT_TRUE(back1->Unwind(0)); |
| delete back1; |
| delete map1; |
| |
| Backtrace* back2 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map2); |
| ASSERT_TRUE(back2 != nullptr); |
| EXPECT_TRUE(back2->Unwind(0)); |
| delete back2; |
| delete map2; |
| |
| Backtrace* back3 = Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD, map3); |
| ASSERT_TRUE(back3 != nullptr); |
| EXPECT_TRUE(back3->Unwind(0)); |
| delete back3; |
| delete map3; |
| } |
| |
| TEST(libbacktrace, fillin_erases) { |
| BacktraceMap* back_map = BacktraceMap::Create(getpid()); |
| |
| backtrace_map_t map; |
| |
| map.start = 1; |
| map.end = 3; |
| map.flags = 1; |
| map.name = "Initialized"; |
| back_map->FillIn(0, &map); |
| delete back_map; |
| |
| ASSERT_FALSE(BacktraceMap::IsValid(map)); |
| ASSERT_EQ(static_cast<uintptr_t>(0), map.start); |
| ASSERT_EQ(static_cast<uintptr_t>(0), map.end); |
| ASSERT_EQ(0, map.flags); |
| ASSERT_EQ("", map.name); |
| } |
| |
| TEST(libbacktrace, format_test) { |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| |
| backtrace_frame_data_t frame; |
| frame.num = 1; |
| frame.pc = 2; |
| frame.sp = 0; |
| frame.stack_size = 0; |
| frame.func_offset = 0; |
| |
| // Check no map set. |
| frame.num = 1; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 0000000000000002 <unknown>", |
| #else |
| EXPECT_EQ("#01 pc 00000002 <unknown>", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| // Check map name empty, but exists. |
| frame.map.start = 1; |
| frame.map.end = 1; |
| frame.map.load_base = 0; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 0000000000000001 <unknown>", |
| #else |
| EXPECT_EQ("#01 pc 00000001 <unknown>", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| |
| // Check relative pc is set and map name is set. |
| frame.pc = 0x12345679; |
| frame.map.name = "MapFake"; |
| frame.map.start = 1; |
| frame.map.end = 1; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 0000000012345678 MapFake", |
| #else |
| EXPECT_EQ("#01 pc 12345678 MapFake", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| // Check func_name is set, but no func offset. |
| frame.func_name = "ProcFake"; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 0000000012345678 MapFake (ProcFake)", |
| #else |
| EXPECT_EQ("#01 pc 12345678 MapFake (ProcFake)", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| // Check func_name is set, and func offset is non-zero. |
| frame.func_offset = 645; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 0000000012345678 MapFake (ProcFake+645)", |
| #else |
| EXPECT_EQ("#01 pc 12345678 MapFake (ProcFake+645)", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| // Check func_name is set, func offset is non-zero, and load_base is non-zero. |
| frame.func_offset = 645; |
| frame.map.load_base = 100; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 00000000123456dc MapFake (ProcFake+645)", |
| #else |
| EXPECT_EQ("#01 pc 123456dc MapFake (ProcFake+645)", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| |
| // Check a non-zero map offset. |
| frame.map.offset = 0x1000; |
| #if defined(__LP64__) |
| EXPECT_EQ("#01 pc 00000000123456dc MapFake (offset 0x1000) (ProcFake+645)", |
| #else |
| EXPECT_EQ("#01 pc 123456dc MapFake (offset 0x1000) (ProcFake+645)", |
| #endif |
| backtrace->FormatFrameData(&frame)); |
| } |
| |
| struct map_test_t { |
| uintptr_t start; |
| uintptr_t end; |
| }; |
| |
| bool map_sort(map_test_t i, map_test_t j) { |
| return i.start < j.start; |
| } |
| |
| void VerifyMap(pid_t pid) { |
| char buffer[4096]; |
| snprintf(buffer, sizeof(buffer), "/proc/%d/maps", pid); |
| |
| FILE* map_file = fopen(buffer, "r"); |
| ASSERT_TRUE(map_file != nullptr); |
| std::vector<map_test_t> test_maps; |
| while (fgets(buffer, sizeof(buffer), map_file)) { |
| map_test_t map; |
| ASSERT_EQ(2, sscanf(buffer, "%" SCNxPTR "-%" SCNxPTR " ", &map.start, &map.end)); |
| test_maps.push_back(map); |
| } |
| fclose(map_file); |
| std::sort(test_maps.begin(), test_maps.end(), map_sort); |
| |
| std::unique_ptr<BacktraceMap> map(BacktraceMap::Create(pid)); |
| |
| // Basic test that verifies that the map is in the expected order. |
| std::vector<map_test_t>::const_iterator test_it = test_maps.begin(); |
| for (BacktraceMap::const_iterator it = map->begin(); it != map->end(); ++it) { |
| ASSERT_TRUE(test_it != test_maps.end()); |
| ASSERT_EQ(test_it->start, it->start); |
| ASSERT_EQ(test_it->end, it->end); |
| ++test_it; |
| } |
| ASSERT_TRUE(test_it == test_maps.end()); |
| } |
| |
| TEST(libbacktrace, verify_map_remote) { |
| pid_t pid; |
| |
| if ((pid = fork()) == 0) { |
| while (true) { |
| } |
| _exit(0); |
| } |
| ASSERT_LT(0, pid); |
| |
| ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0); |
| |
| // Wait for the process to get to a stopping point. |
| WaitForStop(pid); |
| |
| // The maps should match exactly since the forked process has been paused. |
| VerifyMap(pid); |
| |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| |
| kill(pid, SIGKILL); |
| ASSERT_EQ(waitpid(pid, nullptr, 0), pid); |
| } |
| |
| void InitMemory(uint8_t* memory, size_t bytes) { |
| for (size_t i = 0; i < bytes; i++) { |
| memory[i] = i; |
| if (memory[i] == '\0') { |
| // Don't use '\0' in our data so we can verify that an overread doesn't |
| // occur by using a '\0' as the character after the read data. |
| memory[i] = 23; |
| } |
| } |
| } |
| |
| void* ThreadReadTest(void* data) { |
| thread_t* thread_data = reinterpret_cast<thread_t*>(data); |
| |
| thread_data->tid = gettid(); |
| |
| // Create two map pages. |
| // Mark the second page as not-readable. |
| size_t pagesize = static_cast<size_t>(sysconf(_SC_PAGE_SIZE)); |
| uint8_t* memory; |
| if (posix_memalign(reinterpret_cast<void**>(&memory), pagesize, 2 * pagesize) != 0) { |
| return reinterpret_cast<void*>(-1); |
| } |
| |
| if (mprotect(&memory[pagesize], pagesize, PROT_NONE) != 0) { |
| return reinterpret_cast<void*>(-1); |
| } |
| |
| // Set up a simple pattern in memory. |
| InitMemory(memory, pagesize); |
| |
| thread_data->data = memory; |
| |
| // Tell the caller it's okay to start reading memory. |
| android_atomic_acquire_store(1, &thread_data->state); |
| |
| // Loop waiting for the caller to finish reading the memory. |
| while (thread_data->state) { |
| } |
| |
| // Re-enable read-write on the page so that we don't crash if we try |
| // and access data on this page when freeing the memory. |
| if (mprotect(&memory[pagesize], pagesize, PROT_READ | PROT_WRITE) != 0) { |
| return reinterpret_cast<void*>(-1); |
| } |
| free(memory); |
| |
| android_atomic_acquire_store(1, &thread_data->state); |
| |
| return nullptr; |
| } |
| |
| void RunReadTest(Backtrace* backtrace, uintptr_t read_addr) { |
| size_t pagesize = static_cast<size_t>(sysconf(_SC_PAGE_SIZE)); |
| |
| // Create a page of data to use to do quick compares. |
| uint8_t* expected = new uint8_t[pagesize]; |
| InitMemory(expected, pagesize); |
| |
| uint8_t* data = new uint8_t[2*pagesize]; |
| // Verify that we can only read one page worth of data. |
| size_t bytes_read = backtrace->Read(read_addr, data, 2 * pagesize); |
| ASSERT_EQ(pagesize, bytes_read); |
| ASSERT_TRUE(memcmp(data, expected, pagesize) == 0); |
| |
| // Verify unaligned reads. |
| for (size_t i = 1; i < sizeof(word_t); i++) { |
| bytes_read = backtrace->Read(read_addr + i, data, 2 * sizeof(word_t)); |
| ASSERT_EQ(2 * sizeof(word_t), bytes_read); |
| ASSERT_TRUE(memcmp(data, &expected[i], 2 * sizeof(word_t)) == 0) |
| << "Offset at " << i << " failed"; |
| } |
| |
| // Verify small unaligned reads. |
| for (size_t i = 1; i < sizeof(word_t); i++) { |
| for (size_t j = 1; j < sizeof(word_t); j++) { |
| // Set one byte past what we expect to read, to guarantee we don't overread. |
| data[j] = '\0'; |
| bytes_read = backtrace->Read(read_addr + i, data, j); |
| ASSERT_EQ(j, bytes_read); |
| ASSERT_TRUE(memcmp(data, &expected[i], j) == 0) |
| << "Offset at " << i << " length " << j << " miscompared"; |
| ASSERT_EQ('\0', data[j]) |
| << "Offset at " << i << " length " << j << " wrote too much data"; |
| } |
| } |
| delete[] data; |
| delete[] expected; |
| } |
| |
| TEST(libbacktrace, thread_read) { |
| pthread_attr_t attr; |
| pthread_attr_init(&attr); |
| pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); |
| pthread_t thread; |
| thread_t thread_data = { 0, 0, 0, nullptr }; |
| ASSERT_TRUE(pthread_create(&thread, &attr, ThreadReadTest, &thread_data) == 0); |
| |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 10)); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(getpid(), thread_data.tid)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| |
| RunReadTest(backtrace.get(), reinterpret_cast<uintptr_t>(thread_data.data)); |
| |
| android_atomic_acquire_store(0, &thread_data.state); |
| |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 10)); |
| } |
| |
| volatile uintptr_t g_ready = 0; |
| volatile uintptr_t g_addr = 0; |
| |
| void ForkedReadTest() { |
| // Create two map pages. |
| size_t pagesize = static_cast<size_t>(sysconf(_SC_PAGE_SIZE)); |
| uint8_t* memory; |
| if (posix_memalign(reinterpret_cast<void**>(&memory), pagesize, 2 * pagesize) != 0) { |
| perror("Failed to allocate memory\n"); |
| exit(1); |
| } |
| |
| // Mark the second page as not-readable. |
| if (mprotect(&memory[pagesize], pagesize, PROT_NONE) != 0) { |
| perror("Failed to mprotect memory\n"); |
| exit(1); |
| } |
| |
| // Set up a simple pattern in memory. |
| InitMemory(memory, pagesize); |
| |
| g_addr = reinterpret_cast<uintptr_t>(memory); |
| g_ready = 1; |
| |
| while (1) { |
| usleep(US_PER_MSEC); |
| } |
| } |
| |
| TEST(libbacktrace, process_read) { |
| g_ready = 0; |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| ForkedReadTest(); |
| exit(0); |
| } |
| ASSERT_NE(-1, pid); |
| |
| bool test_executed = false; |
| uint64_t start = NanoTime(); |
| while (1) { |
| if (ptrace(PTRACE_ATTACH, pid, 0, 0) == 0) { |
| WaitForStop(pid); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(pid, pid)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| |
| uintptr_t read_addr; |
| size_t bytes_read = backtrace->Read(reinterpret_cast<uintptr_t>(&g_ready), |
| reinterpret_cast<uint8_t*>(&read_addr), |
| sizeof(uintptr_t)); |
| ASSERT_EQ(sizeof(uintptr_t), bytes_read); |
| if (read_addr) { |
| // The forked process is ready to be read. |
| bytes_read = backtrace->Read(reinterpret_cast<uintptr_t>(&g_addr), |
| reinterpret_cast<uint8_t*>(&read_addr), |
| sizeof(uintptr_t)); |
| ASSERT_EQ(sizeof(uintptr_t), bytes_read); |
| |
| RunReadTest(backtrace.get(), read_addr); |
| |
| test_executed = true; |
| break; |
| } |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| } |
| if ((NanoTime() - start) > 5 * NS_PER_SEC) { |
| break; |
| } |
| usleep(US_PER_MSEC); |
| } |
| kill(pid, SIGKILL); |
| ASSERT_EQ(waitpid(pid, nullptr, 0), pid); |
| |
| ASSERT_TRUE(test_executed); |
| } |
| |
| void VerifyFunctionsFound(const std::vector<std::string>& found_functions) { |
| // We expect to find these functions in libbacktrace_test. If we don't |
| // find them, that's a bug in the memory read handling code in libunwind. |
| std::list<std::string> expected_functions; |
| expected_functions.push_back("test_recursive_call"); |
| expected_functions.push_back("test_level_one"); |
| expected_functions.push_back("test_level_two"); |
| expected_functions.push_back("test_level_three"); |
| expected_functions.push_back("test_level_four"); |
| for (const auto& found_function : found_functions) { |
| for (const auto& expected_function : expected_functions) { |
| if (found_function == expected_function) { |
| expected_functions.remove(found_function); |
| break; |
| } |
| } |
| } |
| ASSERT_TRUE(expected_functions.empty()) << "Not all functions found in shared library."; |
| } |
| |
| const char* CopySharedLibrary() { |
| #if defined(__LP64__) |
| const char* lib_name = "lib64"; |
| #else |
| const char* lib_name = "lib"; |
| #endif |
| |
| #if defined(__BIONIC__) |
| const char* tmp_so_name = "/data/local/tmp/libbacktrace_test.so"; |
| std::string cp_cmd = android::base::StringPrintf("cp /system/%s/libbacktrace_test.so %s", |
| lib_name, tmp_so_name); |
| #else |
| const char* tmp_so_name = "/tmp/libbacktrace_test.so"; |
| if (getenv("ANDROID_HOST_OUT") == NULL) { |
| fprintf(stderr, "ANDROID_HOST_OUT not set, make sure you run lunch."); |
| return nullptr; |
| } |
| std::string cp_cmd = android::base::StringPrintf("cp %s/%s/libbacktrace_test.so %s", |
| getenv("ANDROID_HOST_OUT"), lib_name, |
| tmp_so_name); |
| #endif |
| |
| // Copy the shared so to a tempory directory. |
| system(cp_cmd.c_str()); |
| |
| return tmp_so_name; |
| } |
| |
| TEST(libbacktrace, check_unreadable_elf_local) { |
| const char* tmp_so_name = CopySharedLibrary(); |
| ASSERT_TRUE(tmp_so_name != nullptr); |
| |
| struct stat buf; |
| ASSERT_TRUE(stat(tmp_so_name, &buf) != -1); |
| uintptr_t map_size = buf.st_size; |
| |
| int fd = open(tmp_so_name, O_RDONLY); |
| ASSERT_TRUE(fd != -1); |
| |
| void* map = mmap(NULL, map_size, PROT_READ, MAP_PRIVATE, fd, 0); |
| ASSERT_TRUE(map != MAP_FAILED); |
| close(fd); |
| ASSERT_TRUE(unlink(tmp_so_name) != -1); |
| |
| std::vector<std::string> found_functions; |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(BACKTRACE_CURRENT_PROCESS, |
| BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| |
| // Needed before GetFunctionName will work. |
| backtrace->Unwind(0); |
| |
| // Loop through the entire map, and get every function we can find. |
| map_size += reinterpret_cast<uintptr_t>(map); |
| std::string last_func; |
| for (uintptr_t read_addr = reinterpret_cast<uintptr_t>(map); |
| read_addr < map_size; read_addr += 4) { |
| uintptr_t offset; |
| std::string func_name = backtrace->GetFunctionName(read_addr, &offset); |
| if (!func_name.empty() && last_func != func_name) { |
| found_functions.push_back(func_name); |
| } |
| last_func = func_name; |
| } |
| |
| ASSERT_TRUE(munmap(map, map_size - reinterpret_cast<uintptr_t>(map)) == 0); |
| |
| VerifyFunctionsFound(found_functions); |
| } |
| |
| TEST(libbacktrace, check_unreadable_elf_remote) { |
| const char* tmp_so_name = CopySharedLibrary(); |
| ASSERT_TRUE(tmp_so_name != nullptr); |
| |
| g_ready = 0; |
| |
| struct stat buf; |
| ASSERT_TRUE(stat(tmp_so_name, &buf) != -1); |
| uintptr_t map_size = buf.st_size; |
| |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| int fd = open(tmp_so_name, O_RDONLY); |
| if (fd == -1) { |
| fprintf(stderr, "Failed to open file %s: %s\n", tmp_so_name, strerror(errno)); |
| unlink(tmp_so_name); |
| exit(0); |
| } |
| |
| void* map = mmap(NULL, map_size, PROT_READ, MAP_PRIVATE, fd, 0); |
| if (map == MAP_FAILED) { |
| fprintf(stderr, "Failed to map in memory: %s\n", strerror(errno)); |
| unlink(tmp_so_name); |
| exit(0); |
| } |
| close(fd); |
| if (unlink(tmp_so_name) == -1) { |
| fprintf(stderr, "Failed to unlink: %s\n", strerror(errno)); |
| exit(0); |
| } |
| |
| g_addr = reinterpret_cast<uintptr_t>(map); |
| g_ready = 1; |
| while (true) { |
| usleep(US_PER_MSEC); |
| } |
| exit(0); |
| } |
| ASSERT_TRUE(pid > 0); |
| |
| std::vector<std::string> found_functions; |
| uint64_t start = NanoTime(); |
| while (true) { |
| ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0); |
| |
| // Wait for the process to get to a stopping point. |
| WaitForStop(pid); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(pid, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| |
| uintptr_t read_addr; |
| ASSERT_EQ(sizeof(uintptr_t), backtrace->Read(reinterpret_cast<uintptr_t>(&g_ready), reinterpret_cast<uint8_t*>(&read_addr), sizeof(uintptr_t))); |
| if (read_addr) { |
| ASSERT_EQ(sizeof(uintptr_t), backtrace->Read(reinterpret_cast<uintptr_t>(&g_addr), reinterpret_cast<uint8_t*>(&read_addr), sizeof(uintptr_t))); |
| |
| // Needed before GetFunctionName will work. |
| backtrace->Unwind(0); |
| |
| // Loop through the entire map, and get every function we can find. |
| map_size += read_addr; |
| std::string last_func; |
| for (; read_addr < map_size; read_addr += 4) { |
| uintptr_t offset; |
| std::string func_name = backtrace->GetFunctionName(read_addr, &offset); |
| if (!func_name.empty() && last_func != func_name) { |
| found_functions.push_back(func_name); |
| } |
| last_func = func_name; |
| } |
| break; |
| } |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| |
| if ((NanoTime() - start) > 5 * NS_PER_SEC) { |
| break; |
| } |
| usleep(US_PER_MSEC); |
| } |
| |
| kill(pid, SIGKILL); |
| ASSERT_EQ(waitpid(pid, nullptr, 0), pid); |
| |
| VerifyFunctionsFound(found_functions); |
| } |
| |
| bool FindFuncFrameInBacktrace(Backtrace* backtrace, uintptr_t test_func, size_t* frame_num) { |
| backtrace_map_t map; |
| backtrace->FillInMap(test_func, &map); |
| if (!BacktraceMap::IsValid(map)) { |
| return false; |
| } |
| |
| // Loop through the frames, and find the one that is in the map. |
| *frame_num = 0; |
| for (Backtrace::const_iterator it = backtrace->begin(); it != backtrace->end(); ++it) { |
| if (BacktraceMap::IsValid(it->map) && map.start == it->map.start && |
| it->pc >= test_func) { |
| *frame_num = it->num; |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| void VerifyUnreadableElfFrame(Backtrace* backtrace, uintptr_t test_func, size_t frame_num) { |
| ASSERT_LT(backtrace->NumFrames(), static_cast<size_t>(MAX_BACKTRACE_FRAMES)) |
| << DumpFrames(backtrace); |
| |
| ASSERT_TRUE(frame_num != 0) << DumpFrames(backtrace); |
| // Make sure that there is at least one more frame above the test func call. |
| ASSERT_LT(frame_num, backtrace->NumFrames()) << DumpFrames(backtrace); |
| |
| uintptr_t diff = backtrace->GetFrame(frame_num)->pc - test_func; |
| ASSERT_LT(diff, 200U) << DumpFrames(backtrace); |
| } |
| |
| void VerifyUnreadableElfBacktrace(uintptr_t test_func) { |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(BACKTRACE_CURRENT_PROCESS, |
| BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| size_t frame_num; |
| ASSERT_TRUE(FindFuncFrameInBacktrace(backtrace.get(), test_func, &frame_num)); |
| |
| VerifyUnreadableElfFrame(backtrace.get(), test_func, frame_num); |
| } |
| |
| typedef int (*test_func_t)(int, int, int, int, void (*)(uintptr_t), uintptr_t); |
| |
| TEST(libbacktrace, unwind_through_unreadable_elf_local) { |
| const char* tmp_so_name = CopySharedLibrary(); |
| ASSERT_TRUE(tmp_so_name != nullptr); |
| void* lib_handle = dlopen(tmp_so_name, RTLD_NOW); |
| ASSERT_TRUE(lib_handle != nullptr); |
| ASSERT_TRUE(unlink(tmp_so_name) != -1); |
| |
| test_func_t test_func; |
| test_func = reinterpret_cast<test_func_t>(dlsym(lib_handle, "test_level_one")); |
| ASSERT_TRUE(test_func != nullptr); |
| |
| ASSERT_NE(test_func(1, 2, 3, 4, VerifyUnreadableElfBacktrace, |
| reinterpret_cast<uintptr_t>(test_func)), 0); |
| |
| ASSERT_TRUE(dlclose(lib_handle) == 0); |
| } |
| |
| TEST(libbacktrace, unwind_through_unreadable_elf_remote) { |
| const char* tmp_so_name = CopySharedLibrary(); |
| ASSERT_TRUE(tmp_so_name != nullptr); |
| void* lib_handle = dlopen(tmp_so_name, RTLD_NOW); |
| ASSERT_TRUE(lib_handle != nullptr); |
| ASSERT_TRUE(unlink(tmp_so_name) != -1); |
| |
| test_func_t test_func; |
| test_func = reinterpret_cast<test_func_t>(dlsym(lib_handle, "test_level_one")); |
| ASSERT_TRUE(test_func != nullptr); |
| |
| pid_t pid; |
| if ((pid = fork()) == 0) { |
| test_func(1, 2, 3, 4, 0, 0); |
| exit(0); |
| } |
| ASSERT_TRUE(pid > 0); |
| ASSERT_TRUE(dlclose(lib_handle) == 0); |
| |
| uint64_t start = NanoTime(); |
| bool done = false; |
| while (!done) { |
| ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0); |
| |
| // Wait for the process to get to a stopping point. |
| WaitForStop(pid); |
| |
| std::unique_ptr<Backtrace> backtrace(Backtrace::Create(pid, BACKTRACE_CURRENT_THREAD)); |
| ASSERT_TRUE(backtrace.get() != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| |
| size_t frame_num; |
| if (FindFuncFrameInBacktrace(backtrace.get(), |
| reinterpret_cast<uintptr_t>(test_func), &frame_num)) { |
| |
| VerifyUnreadableElfFrame(backtrace.get(), reinterpret_cast<uintptr_t>(test_func), frame_num); |
| done = true; |
| } |
| |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| |
| if ((NanoTime() - start) > 5 * NS_PER_SEC) { |
| break; |
| } |
| usleep(US_PER_MSEC); |
| } |
| |
| kill(pid, SIGKILL); |
| ASSERT_EQ(waitpid(pid, nullptr, 0), pid); |
| |
| ASSERT_TRUE(done) << "Test function never found in unwind."; |
| } |
| |
| #if defined(ENABLE_PSS_TESTS) |
| #include "GetPss.h" |
| |
| #define MAX_LEAK_BYTES 32*1024UL |
| |
| void CheckForLeak(pid_t pid, pid_t tid) { |
| // Do a few runs to get the PSS stable. |
| for (size_t i = 0; i < 100; i++) { |
| Backtrace* backtrace = Backtrace::Create(pid, tid); |
| ASSERT_TRUE(backtrace != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| delete backtrace; |
| } |
| size_t stable_pss = GetPssBytes(); |
| ASSERT_TRUE(stable_pss != 0); |
| |
| // Loop enough that even a small leak should be detectable. |
| for (size_t i = 0; i < 4096; i++) { |
| Backtrace* backtrace = Backtrace::Create(pid, tid); |
| ASSERT_TRUE(backtrace != nullptr); |
| ASSERT_TRUE(backtrace->Unwind(0)); |
| delete backtrace; |
| } |
| size_t new_pss = GetPssBytes(); |
| ASSERT_TRUE(new_pss != 0); |
| size_t abs_diff = (new_pss > stable_pss) ? new_pss - stable_pss : stable_pss - new_pss; |
| // As long as the new pss is within a certain amount, consider everything okay. |
| ASSERT_LE(abs_diff, MAX_LEAK_BYTES); |
| } |
| |
| TEST(libbacktrace, check_for_leak_local) { |
| CheckForLeak(BACKTRACE_CURRENT_PROCESS, BACKTRACE_CURRENT_THREAD); |
| } |
| |
| TEST(libbacktrace, check_for_leak_local_thread) { |
| thread_t thread_data = { 0, 0, 0, nullptr }; |
| pthread_t thread; |
| ASSERT_TRUE(pthread_create(&thread, nullptr, ThreadLevelRun, &thread_data) == 0); |
| |
| // Wait up to 2 seconds for the tid to be set. |
| ASSERT_TRUE(WaitForNonZero(&thread_data.state, 2)); |
| |
| CheckForLeak(BACKTRACE_CURRENT_PROCESS, thread_data.tid); |
| |
| // Tell the thread to exit its infinite loop. |
| android_atomic_acquire_store(0, &thread_data.state); |
| |
| ASSERT_TRUE(pthread_join(thread, nullptr) == 0); |
| } |
| |
| TEST(libbacktrace, check_for_leak_remote) { |
| pid_t pid; |
| |
| if ((pid = fork()) == 0) { |
| while (true) { |
| } |
| _exit(0); |
| } |
| ASSERT_LT(0, pid); |
| |
| ASSERT_TRUE(ptrace(PTRACE_ATTACH, pid, 0, 0) == 0); |
| |
| // Wait for the process to get to a stopping point. |
| WaitForStop(pid); |
| |
| CheckForLeak(pid, BACKTRACE_CURRENT_THREAD); |
| |
| ASSERT_TRUE(ptrace(PTRACE_DETACH, pid, 0, 0) == 0); |
| |
| kill(pid, SIGKILL); |
| ASSERT_EQ(waitpid(pid, nullptr, 0), pid); |
| } |
| #endif |
| |