blob: e6a91247cbef13a0c16681d770193e5e0d685011 [file] [log] [blame]
/*
* Copyright (C) 2008 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "fault_handler.h"
#include <sys/ucontext.h>
#include "art_method.h"
#include "base/enums.h"
#include "base/hex_dump.h"
#include "base/logging.h" // For VLOG.
#include "base/macros.h"
#include "base/safe_copy.h"
#include "globals.h"
#include "thread-current-inl.h"
#if defined(__APPLE__)
#define ucontext __darwin_ucontext
#if defined(__x86_64__)
// 64 bit mac build.
#define CTX_ESP uc_mcontext->__ss.__rsp
#define CTX_EIP uc_mcontext->__ss.__rip
#define CTX_EAX uc_mcontext->__ss.__rax
#define CTX_METHOD uc_mcontext->__ss.__rdi
#define CTX_RDI uc_mcontext->__ss.__rdi
#define CTX_JMP_BUF uc_mcontext->__ss.__rdi
#else
// 32 bit mac build.
#define CTX_ESP uc_mcontext->__ss.__esp
#define CTX_EIP uc_mcontext->__ss.__eip
#define CTX_EAX uc_mcontext->__ss.__eax
#define CTX_METHOD uc_mcontext->__ss.__eax
#define CTX_JMP_BUF uc_mcontext->__ss.__eax
#endif
#elif defined(__x86_64__)
// 64 bit linux build.
#define CTX_ESP uc_mcontext.gregs[REG_RSP]
#define CTX_EIP uc_mcontext.gregs[REG_RIP]
#define CTX_EAX uc_mcontext.gregs[REG_RAX]
#define CTX_METHOD uc_mcontext.gregs[REG_RDI]
#define CTX_RDI uc_mcontext.gregs[REG_RDI]
#define CTX_JMP_BUF uc_mcontext.gregs[REG_RDI]
#else
// 32 bit linux build.
#define CTX_ESP uc_mcontext.gregs[REG_ESP]
#define CTX_EIP uc_mcontext.gregs[REG_EIP]
#define CTX_EAX uc_mcontext.gregs[REG_EAX]
#define CTX_METHOD uc_mcontext.gregs[REG_EAX]
#define CTX_JMP_BUF uc_mcontext.gregs[REG_EAX]
#endif
//
// X86 (and X86_64) specific fault handler functions.
//
namespace art {
extern "C" void art_quick_throw_null_pointer_exception_from_signal();
extern "C" void art_quick_throw_stack_overflow();
extern "C" void art_quick_test_suspend();
// Get the size of an instruction in bytes.
// Return 0 if the instruction is not handled.
static uint32_t GetInstructionSize(const uint8_t* pc) {
// Don't segfault if pc points to garbage.
char buf[15]; // x86/x86-64 have a maximum instruction length of 15 bytes.
ssize_t bytes = SafeCopy(buf, pc, sizeof(buf));
if (bytes == 0) {
// Nothing was readable.
return 0;
}
if (bytes == -1) {
// SafeCopy not supported, assume that the entire range is readable.
bytes = 16;
} else {
pc = reinterpret_cast<uint8_t*>(buf);
}
#define INCREMENT_PC() \
do { \
pc++; \
if (pc - startpc > bytes) { \
return 0; \
} \
} while (0)
#if defined(__x86_64)
const bool x86_64 = true;
#else
const bool x86_64 = false;
#endif
const uint8_t* startpc = pc;
uint8_t opcode = *pc;
INCREMENT_PC();
uint8_t modrm;
bool has_modrm = false;
bool two_byte = false;
uint32_t displacement_size = 0;
uint32_t immediate_size = 0;
bool operand_size_prefix = false;
// Prefixes.
while (true) {
bool prefix_present = false;
switch (opcode) {
// Group 3
case 0x66:
operand_size_prefix = true;
FALLTHROUGH_INTENDED;
// Group 1
case 0xf0:
case 0xf2:
case 0xf3:
// Group 2
case 0x2e:
case 0x36:
case 0x3e:
case 0x26:
case 0x64:
case 0x65:
// Group 4
case 0x67:
opcode = *pc;
INCREMENT_PC();
prefix_present = true;
break;
}
if (!prefix_present) {
break;
}
}
if (x86_64 && opcode >= 0x40 && opcode <= 0x4f) {
opcode = *pc;
INCREMENT_PC();
}
if (opcode == 0x0f) {
// Two byte opcode
two_byte = true;
opcode = *pc;
INCREMENT_PC();
}
bool unhandled_instruction = false;
if (two_byte) {
switch (opcode) {
case 0x10: // vmovsd/ss
case 0x11: // vmovsd/ss
case 0xb6: // movzx
case 0xb7:
case 0xbe: // movsx
case 0xbf:
modrm = *pc;
INCREMENT_PC();
has_modrm = true;
break;
default:
unhandled_instruction = true;
break;
}
} else {
switch (opcode) {
case 0x88: // mov byte
case 0x89: // mov
case 0x8b:
case 0x38: // cmp with memory.
case 0x39:
case 0x3a:
case 0x3b:
case 0x3c:
case 0x3d:
case 0x85: // test.
modrm = *pc;
INCREMENT_PC();
has_modrm = true;
break;
case 0x80: // group 1, byte immediate.
case 0x83:
case 0xc6:
modrm = *pc;
INCREMENT_PC();
has_modrm = true;
immediate_size = 1;
break;
case 0x81: // group 1, word immediate.
case 0xc7: // mov
modrm = *pc;
INCREMENT_PC();
has_modrm = true;
immediate_size = operand_size_prefix ? 2 : 4;
break;
case 0xf6:
case 0xf7:
modrm = *pc;
INCREMENT_PC();
has_modrm = true;
switch ((modrm >> 3) & 7) { // Extract "reg/opcode" from "modr/m".
case 0: // test
immediate_size = (opcode == 0xf6) ? 1 : (operand_size_prefix ? 2 : 4);
break;
case 2: // not
case 3: // neg
case 4: // mul
case 5: // imul
case 6: // div
case 7: // idiv
break;
default:
unhandled_instruction = true;
break;
}
break;
default:
unhandled_instruction = true;
break;
}
}
if (unhandled_instruction) {
VLOG(signals) << "Unhandled x86 instruction with opcode " << static_cast<int>(opcode);
return 0;
}
if (has_modrm) {
uint8_t mod = (modrm >> 6) & 3U /* 0b11 */;
// Check for SIB.
if (mod != 3U /* 0b11 */ && (modrm & 7U /* 0b111 */) == 4) {
INCREMENT_PC(); // SIB
}
switch (mod) {
case 0U /* 0b00 */: break;
case 1U /* 0b01 */: displacement_size = 1; break;
case 2U /* 0b10 */: displacement_size = 4; break;
case 3U /* 0b11 */:
break;
}
}
// Skip displacement and immediate.
pc += displacement_size + immediate_size;
VLOG(signals) << "x86 instruction length calculated as " << (pc - startpc);
if (pc - startpc > bytes) {
return 0;
}
return pc - startpc;
}
void FaultManager::GetMethodAndReturnPcAndSp(siginfo_t* siginfo, void* context,
ArtMethod** out_method,
uintptr_t* out_return_pc, uintptr_t* out_sp) {
struct ucontext* uc = reinterpret_cast<struct ucontext*>(context);
*out_sp = static_cast<uintptr_t>(uc->CTX_ESP);
VLOG(signals) << "sp: " << std::hex << *out_sp;
if (*out_sp == 0) {
return;
}
// In the case of a stack overflow, the stack is not valid and we can't
// get the method from the top of the stack. However it's in EAX(x86)/RDI(x86_64).
uintptr_t* fault_addr = reinterpret_cast<uintptr_t*>(siginfo->si_addr);
uintptr_t* overflow_addr = reinterpret_cast<uintptr_t*>(
#if defined(__x86_64__)
reinterpret_cast<uint8_t*>(*out_sp) - GetStackOverflowReservedBytes(InstructionSet::kX86_64));
#else
reinterpret_cast<uint8_t*>(*out_sp) - GetStackOverflowReservedBytes(InstructionSet::kX86));
#endif
if (overflow_addr == fault_addr) {
*out_method = reinterpret_cast<ArtMethod*>(uc->CTX_METHOD);
} else {
// The method is at the top of the stack.
*out_method = *reinterpret_cast<ArtMethod**>(*out_sp);
}
uint8_t* pc = reinterpret_cast<uint8_t*>(uc->CTX_EIP);
VLOG(signals) << HexDump(pc, 32, true, "PC ");
if (pc == nullptr) {
// Somebody jumped to 0x0. Definitely not ours, and will definitely segfault below.
*out_method = nullptr;
return;
}
uint32_t instr_size = GetInstructionSize(pc);
if (instr_size == 0) {
// Unknown instruction, tell caller it's not ours.
*out_method = nullptr;
return;
}
*out_return_pc = reinterpret_cast<uintptr_t>(pc + instr_size);
}
bool NullPointerHandler::Action(int, siginfo_t* sig, void* context) {
if (!IsValidImplicitCheck(sig)) {
return false;
}
struct ucontext *uc = reinterpret_cast<struct ucontext*>(context);
uint8_t* pc = reinterpret_cast<uint8_t*>(uc->CTX_EIP);
uint8_t* sp = reinterpret_cast<uint8_t*>(uc->CTX_ESP);
uint32_t instr_size = GetInstructionSize(pc);
if (instr_size == 0) {
// Unknown instruction, can't really happen.
return false;
}
// We need to arrange for the signal handler to return to the null pointer
// exception generator. The return address must be the address of the
// next instruction (this instruction + instruction size). The return address
// is on the stack at the top address of the current frame.
// Push the return address and fault address onto the stack.
uintptr_t retaddr = reinterpret_cast<uintptr_t>(pc + instr_size);
uintptr_t* next_sp = reinterpret_cast<uintptr_t*>(sp - 2 * sizeof(uintptr_t));
next_sp[1] = retaddr;
next_sp[0] = reinterpret_cast<uintptr_t>(sig->si_addr);
uc->CTX_ESP = reinterpret_cast<uintptr_t>(next_sp);
uc->CTX_EIP = reinterpret_cast<uintptr_t>(
art_quick_throw_null_pointer_exception_from_signal);
VLOG(signals) << "Generating null pointer exception";
return true;
}
// A suspend check is done using the following instruction sequence:
// (x86)
// 0xf720f1df: 648B058C000000 mov eax, fs:[0x8c] ; suspend_trigger
// .. some intervening instructions.
// 0xf720f1e6: 8500 test eax, [eax]
// (x86_64)
// 0x7f579de45d9e: 65488B0425A8000000 movq rax, gs:[0xa8] ; suspend_trigger
// .. some intervening instructions.
// 0x7f579de45da7: 8500 test eax, [eax]
// The offset from fs is Thread::ThreadSuspendTriggerOffset().
// To check for a suspend check, we examine the instructions that caused
// the fault.
bool SuspensionHandler::Action(int, siginfo_t*, void* context) {
// These are the instructions to check for. The first one is the mov eax, fs:[xxx]
// where xxx is the offset of the suspend trigger.
uint32_t trigger = Thread::ThreadSuspendTriggerOffset<kRuntimePointerSize>().Int32Value();
VLOG(signals) << "Checking for suspension point";
#if defined(__x86_64__)
uint8_t checkinst1[] = {0x65, 0x48, 0x8b, 0x04, 0x25, static_cast<uint8_t>(trigger & 0xff),
static_cast<uint8_t>((trigger >> 8) & 0xff), 0, 0};
#else
uint8_t checkinst1[] = {0x64, 0x8b, 0x05, static_cast<uint8_t>(trigger & 0xff),
static_cast<uint8_t>((trigger >> 8) & 0xff), 0, 0};
#endif
uint8_t checkinst2[] = {0x85, 0x00};
struct ucontext *uc = reinterpret_cast<struct ucontext*>(context);
uint8_t* pc = reinterpret_cast<uint8_t*>(uc->CTX_EIP);
uint8_t* sp = reinterpret_cast<uint8_t*>(uc->CTX_ESP);
if (pc[0] != checkinst2[0] || pc[1] != checkinst2[1]) {
// Second instruction is not correct (test eax,[eax]).
VLOG(signals) << "Not a suspension point";
return false;
}
// The first instruction can a little bit up the stream due to load hoisting
// in the compiler.
uint8_t* limit = pc - 100; // Compiler will hoist to a max of 20 instructions.
uint8_t* ptr = pc - sizeof(checkinst1);
bool found = false;
while (ptr > limit) {
if (memcmp(ptr, checkinst1, sizeof(checkinst1)) == 0) {
found = true;
break;
}
ptr -= 1;
}
if (found) {
VLOG(signals) << "suspend check match";
// We need to arrange for the signal handler to return to the null pointer
// exception generator. The return address must be the address of the
// next instruction (this instruction + 2). The return address
// is on the stack at the top address of the current frame.
// Push the return address onto the stack.
uintptr_t retaddr = reinterpret_cast<uintptr_t>(pc + 2);
uintptr_t* next_sp = reinterpret_cast<uintptr_t*>(sp - sizeof(uintptr_t));
*next_sp = retaddr;
uc->CTX_ESP = reinterpret_cast<uintptr_t>(next_sp);
uc->CTX_EIP = reinterpret_cast<uintptr_t>(art_quick_test_suspend);
// Now remove the suspend trigger that caused this fault.
Thread::Current()->RemoveSuspendTrigger();
VLOG(signals) << "removed suspend trigger invoking test suspend";
return true;
}
VLOG(signals) << "Not a suspend check match, first instruction mismatch";
return false;
}
// The stack overflow check is done using the following instruction:
// test eax, [esp+ -xxx]
// where 'xxx' is the size of the overflow area.
//
// This is done before any frame is established in the method. The return
// address for the previous method is on the stack at ESP.
bool StackOverflowHandler::Action(int, siginfo_t* info, void* context) {
struct ucontext *uc = reinterpret_cast<struct ucontext*>(context);
uintptr_t sp = static_cast<uintptr_t>(uc->CTX_ESP);
uintptr_t fault_addr = reinterpret_cast<uintptr_t>(info->si_addr);
VLOG(signals) << "fault_addr: " << std::hex << fault_addr;
VLOG(signals) << "checking for stack overflow, sp: " << std::hex << sp <<
", fault_addr: " << fault_addr;
#if defined(__x86_64__)
uintptr_t overflow_addr = sp - GetStackOverflowReservedBytes(InstructionSet::kX86_64);
#else
uintptr_t overflow_addr = sp - GetStackOverflowReservedBytes(InstructionSet::kX86);
#endif
// Check that the fault address is the value expected for a stack overflow.
if (fault_addr != overflow_addr) {
VLOG(signals) << "Not a stack overflow";
return false;
}
VLOG(signals) << "Stack overflow found";
// Since the compiler puts the implicit overflow
// check before the callee save instructions, the SP is already pointing to
// the previous frame.
// Now arrange for the signal handler to return to art_quick_throw_stack_overflow.
uc->CTX_EIP = reinterpret_cast<uintptr_t>(art_quick_throw_stack_overflow);
return true;
}
} // namespace art