blob: 83551a9750731e38ee41fdd5b43cce099ffde175 [file] [log] [blame]
/*
* Copyright (C) 2015 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "stack_map_stream.h"
#include <memory>
#include <vector>
#include "art_method-inl.h"
#include "base/globals.h"
#include "base/stl_util.h"
#include "class_linker.h"
#include "dex/dex_file.h"
#include "dex/dex_file_types.h"
#include "driver/compiler_options.h"
#include "oat/stack_map.h"
#include "optimizing/code_generator.h"
#include "optimizing/nodes.h"
#include "optimizing/optimizing_compiler.h"
#include "runtime.h"
#include "scoped_thread_state_change-inl.h"
namespace art HIDDEN {
constexpr static bool kVerifyStackMaps = kIsDebugBuild;
uint32_t StackMapStream::GetStackMapNativePcOffset(size_t i) {
return StackMap::UnpackNativePc(stack_maps_[i][StackMap::kPackedNativePc], instruction_set_);
}
void StackMapStream::SetStackMapNativePcOffset(size_t i, uint32_t native_pc_offset) {
stack_maps_[i][StackMap::kPackedNativePc] =
StackMap::PackNativePc(native_pc_offset, instruction_set_);
}
void StackMapStream::BeginMethod(size_t frame_size_in_bytes,
size_t core_spill_mask,
size_t fp_spill_mask,
uint32_t num_dex_registers,
bool baseline,
bool debuggable,
bool has_should_deoptimize_flag) {
DCHECK(!in_method_) << "Mismatched Begin/End calls";
in_method_ = true;
DCHECK_EQ(packed_frame_size_, 0u) << "BeginMethod was already called";
DCHECK_ALIGNED(frame_size_in_bytes, kStackAlignment);
packed_frame_size_ = frame_size_in_bytes / kStackAlignment;
core_spill_mask_ = core_spill_mask;
fp_spill_mask_ = fp_spill_mask;
num_dex_registers_ = num_dex_registers;
baseline_ = baseline;
debuggable_ = debuggable;
has_should_deoptimize_flag_ = has_should_deoptimize_flag;
if (kVerifyStackMaps) {
dchecks_.emplace_back([=](const CodeInfo& code_info) {
DCHECK_EQ(code_info.packed_frame_size_, frame_size_in_bytes / kStackAlignment);
DCHECK_EQ(code_info.core_spill_mask_, core_spill_mask);
DCHECK_EQ(code_info.fp_spill_mask_, fp_spill_mask);
DCHECK_EQ(code_info.number_of_dex_registers_, num_dex_registers);
});
}
}
void StackMapStream::EndMethod(size_t code_size) {
DCHECK(in_method_) << "Mismatched Begin/End calls";
in_method_ = false;
code_size_ = code_size;
// Read the stack masks now. The compiler might have updated them.
for (size_t i = 0; i < lazy_stack_masks_.size(); i++) {
BitVector* stack_mask = lazy_stack_masks_[i];
if (stack_mask != nullptr && stack_mask->GetNumberOfBits() != 0) {
stack_maps_[i][StackMap::kStackMaskIndex] =
stack_masks_.Dedup(stack_mask->GetRawStorage(), stack_mask->GetNumberOfBits());
}
}
if (kIsDebugBuild) {
uint32_t packed_code_size = StackMap::PackNativePc(code_size, instruction_set_);
for (size_t i = 0; i < stack_maps_.size(); i++) {
DCHECK_LE(stack_maps_[i][StackMap::kPackedNativePc], packed_code_size);
}
}
if (kVerifyStackMaps) {
dchecks_.emplace_back([=](const CodeInfo& code_info) {
CHECK_EQ(code_info.code_size_, code_size);
});
}
}
void StackMapStream::BeginStackMapEntry(
uint32_t dex_pc,
uint32_t native_pc_offset,
uint32_t register_mask,
BitVector* stack_mask,
StackMap::Kind kind,
bool needs_vreg_info,
const std::vector<uint32_t>& dex_pc_list_for_catch_verification) {
DCHECK(in_method_) << "Call BeginMethod first";
DCHECK(!in_stack_map_) << "Mismatched Begin/End calls";
in_stack_map_ = true;
DCHECK_IMPLIES(!dex_pc_list_for_catch_verification.empty(), kind == StackMap::Kind::Catch);
DCHECK_IMPLIES(!dex_pc_list_for_catch_verification.empty(), kIsDebugBuild);
current_stack_map_ = BitTableBuilder<StackMap>::Entry();
current_stack_map_[StackMap::kKind] = static_cast<uint32_t>(kind);
current_stack_map_[StackMap::kPackedNativePc] =
StackMap::PackNativePc(native_pc_offset, instruction_set_);
current_stack_map_[StackMap::kDexPc] = dex_pc;
if (stack_maps_.size() > 0) {
// Check that non-catch stack maps are sorted by pc.
// Catch stack maps are at the end and may be unordered.
if (stack_maps_.back()[StackMap::kKind] == StackMap::Kind::Catch) {
DCHECK(current_stack_map_[StackMap::kKind] == StackMap::Kind::Catch);
} else if (current_stack_map_[StackMap::kKind] != StackMap::Kind::Catch) {
DCHECK_LE(stack_maps_.back()[StackMap::kPackedNativePc],
current_stack_map_[StackMap::kPackedNativePc]);
}
}
if (register_mask != 0) {
uint32_t shift = LeastSignificantBit(register_mask);
BitTableBuilder<RegisterMask>::Entry entry;
entry[RegisterMask::kValue] = register_mask >> shift;
entry[RegisterMask::kShift] = shift;
current_stack_map_[StackMap::kRegisterMaskIndex] = register_masks_.Dedup(&entry);
}
// The compiler assumes the bit vector will be read during PrepareForFillIn(),
// and it might modify the data before that. Therefore, just store the pointer.
// See ClearSpillSlotsFromLoopPhisInStackMap in code_generator.h.
lazy_stack_masks_.push_back(stack_mask);
current_inline_infos_.clear();
current_dex_registers_.clear();
expected_num_dex_registers_ = needs_vreg_info ? num_dex_registers_ : 0u;
if (kVerifyStackMaps) {
size_t stack_map_index = stack_maps_.size();
// Create lambda method, which will be executed at the very end to verify data.
// Parameters and local variables will be captured(stored) by the lambda "[=]".
dchecks_.emplace_back([=](const CodeInfo& code_info) {
// The `native_pc_offset` may have been overridden using `SetStackMapNativePcOffset(.)`.
uint32_t final_native_pc_offset = GetStackMapNativePcOffset(stack_map_index);
if (kind == StackMap::Kind::Default || kind == StackMap::Kind::OSR) {
StackMap stack_map = code_info.GetStackMapForNativePcOffset(final_native_pc_offset,
instruction_set_);
CHECK_EQ(stack_map.Row(), stack_map_index);
} else if (kind == StackMap::Kind::Catch) {
StackMap stack_map = code_info.GetCatchStackMapForDexPc(
ArrayRef<const uint32_t>(dex_pc_list_for_catch_verification));
CHECK_EQ(stack_map.Row(), stack_map_index);
}
StackMap stack_map = code_info.GetStackMapAt(stack_map_index);
CHECK_EQ(stack_map.GetNativePcOffset(instruction_set_), final_native_pc_offset);
CHECK_EQ(stack_map.GetKind(), static_cast<uint32_t>(kind));
CHECK_EQ(stack_map.GetDexPc(), dex_pc);
CHECK_EQ(code_info.GetRegisterMaskOf(stack_map), register_mask);
BitMemoryRegion seen_stack_mask = code_info.GetStackMaskOf(stack_map);
CHECK_GE(seen_stack_mask.size_in_bits(), stack_mask ? stack_mask->GetNumberOfBits() : 0);
for (size_t b = 0; b < seen_stack_mask.size_in_bits(); b++) {
CHECK_EQ(seen_stack_mask.LoadBit(b), stack_mask != nullptr && stack_mask->IsBitSet(b));
}
});
}
}
void StackMapStream::EndStackMapEntry() {
DCHECK(in_stack_map_) << "Mismatched Begin/End calls";
in_stack_map_ = false;
// Generate index into the InlineInfo table.
size_t inlining_depth = current_inline_infos_.size();
if (!current_inline_infos_.empty()) {
current_inline_infos_.back()[InlineInfo::kIsLast] = InlineInfo::kLast;
current_stack_map_[StackMap::kInlineInfoIndex] =
inline_infos_.Dedup(current_inline_infos_.data(), current_inline_infos_.size());
}
// Generate delta-compressed dex register map.
size_t num_dex_registers = current_dex_registers_.size();
if (!current_dex_registers_.empty()) {
DCHECK_EQ(expected_num_dex_registers_, current_dex_registers_.size());
CreateDexRegisterMap();
}
stack_maps_.Add(current_stack_map_);
if (kVerifyStackMaps) {
size_t stack_map_index = stack_maps_.size() - 1;
dchecks_.emplace_back([=](const CodeInfo& code_info) {
StackMap stack_map = code_info.GetStackMapAt(stack_map_index);
CHECK_EQ(stack_map.HasDexRegisterMap(), (num_dex_registers != 0));
CHECK_EQ(stack_map.HasInlineInfo(), (inlining_depth != 0));
CHECK_EQ(code_info.GetInlineInfosOf(stack_map).size(), inlining_depth);
});
}
}
void StackMapStream::BeginInlineInfoEntry(ArtMethod* method,
uint32_t dex_pc,
uint32_t num_dex_registers,
const DexFile* outer_dex_file,
const CodeGenerator* codegen) {
DCHECK(in_stack_map_) << "Call BeginStackMapEntry first";
DCHECK(!in_inline_info_) << "Mismatched Begin/End calls";
in_inline_info_ = true;
DCHECK_EQ(expected_num_dex_registers_, current_dex_registers_.size());
expected_num_dex_registers_ += num_dex_registers;
BitTableBuilder<InlineInfo>::Entry entry;
entry[InlineInfo::kIsLast] = InlineInfo::kMore;
entry[InlineInfo::kDexPc] = dex_pc;
entry[InlineInfo::kNumberOfDexRegisters] = static_cast<uint32_t>(expected_num_dex_registers_);
if (EncodeArtMethodInInlineInfo(method)) {
entry[InlineInfo::kArtMethodHi] = High32Bits(reinterpret_cast<uintptr_t>(method));
entry[InlineInfo::kArtMethodLo] = Low32Bits(reinterpret_cast<uintptr_t>(method));
} else {
uint32_t is_in_bootclasspath = MethodInfo::kKindNonBCP;
uint32_t dexfile_index = MethodInfo::kSameDexFile;
if (dex_pc != static_cast<uint32_t>(-1)) {
ScopedObjectAccess soa(Thread::Current());
const DexFile* dex_file = method->GetDexFile();
if (!IsSameDexFile(*outer_dex_file, *dex_file)) {
if (method->GetDeclaringClass()->IsBootStrapClassLoaded()) {
ClassLinker* class_linker = Runtime::Current()->GetClassLinker();
const std::vector<const DexFile*>& boot_class_path = class_linker->GetBootClassPath();
auto it = std::find_if(
boot_class_path.begin(), boot_class_path.end(), [dex_file](const DexFile* df) {
return IsSameDexFile(*df, *dex_file);
});
is_in_bootclasspath = MethodInfo::kKindBCP;
dexfile_index = std::distance(boot_class_path.begin(), it);
} else {
const std::vector<const DexFile*>& dex_files =
codegen->GetCompilerOptions().GetDexFilesForOatFile();
auto it = std::find_if(dex_files.begin(), dex_files.end(), [dex_file](const DexFile* df) {
return IsSameDexFile(*df, *dex_file);
});
// No need to set is_in_bootclasspath since the default value works.
dexfile_index = std::distance(dex_files.begin(), it);
}
}
}
uint32_t dex_method_index = method->GetDexMethodIndex();
entry[InlineInfo::kMethodInfoIndex] =
method_infos_.Dedup({dex_method_index, is_in_bootclasspath, dexfile_index});
}
current_inline_infos_.push_back(entry);
if (kVerifyStackMaps) {
size_t stack_map_index = stack_maps_.size();
size_t depth = current_inline_infos_.size() - 1;
dchecks_.emplace_back([=](const CodeInfo& code_info) {
StackMap stack_map = code_info.GetStackMapAt(stack_map_index);
InlineInfo inline_info = code_info.GetInlineInfosOf(stack_map)[depth];
CHECK_EQ(inline_info.GetDexPc(), dex_pc);
bool encode_art_method = EncodeArtMethodInInlineInfo(method);
CHECK_EQ(inline_info.EncodesArtMethod(), encode_art_method);
if (encode_art_method) {
CHECK_EQ(inline_info.GetArtMethod(), method);
} else {
MethodInfo method_info = code_info.GetMethodInfoOf(inline_info);
CHECK_EQ(method_info.GetMethodIndex(), method->GetDexMethodIndex());
CHECK(method_info.GetDexFileIndexKind() == MethodInfo::kKindNonBCP ||
method_info.GetDexFileIndexKind() == MethodInfo::kKindBCP);
ScopedObjectAccess soa(Thread::Current());
if (inline_info.GetDexPc() != static_cast<uint32_t>(-1) &&
!IsSameDexFile(*outer_dex_file, *method->GetDexFile())) {
if (method->GetDeclaringClass()->IsBootStrapClassLoaded()) {
CHECK_EQ(method_info.GetDexFileIndexKind(), MethodInfo::kKindBCP);
ClassLinker* class_linker = Runtime::Current()->GetClassLinker();
const std::vector<const DexFile*>& boot_class_path = class_linker->GetBootClassPath();
DCHECK_LT(method_info.GetDexFileIndex(), boot_class_path.size());
CHECK(IsSameDexFile(*boot_class_path[method_info.GetDexFileIndex()],
*method->GetDexFile()));
} else {
CHECK_EQ(method_info.GetDexFileIndexKind(), MethodInfo::kKindNonBCP);
const std::vector<const DexFile*>& dex_files =
codegen->GetCompilerOptions().GetDexFilesForOatFile();
DCHECK_LT(method_info.GetDexFileIndex(), dex_files.size());
CHECK(IsSameDexFile(*dex_files[method_info.GetDexFileIndex()], *method->GetDexFile()));
}
}
}
});
}
}
void StackMapStream::EndInlineInfoEntry() {
DCHECK(in_inline_info_) << "Mismatched Begin/End calls";
in_inline_info_ = false;
DCHECK_EQ(expected_num_dex_registers_, current_dex_registers_.size());
}
// Create delta-compressed dex register map based on the current list of DexRegisterLocations.
// All dex registers for a stack map are concatenated - inlined registers are just appended.
void StackMapStream::CreateDexRegisterMap() {
// These are fields rather than local variables so that we can reuse the reserved memory.
temp_dex_register_mask_.ClearAllBits();
temp_dex_register_map_.clear();
// Ensure that the arrays that hold previous state are big enough to be safely indexed below.
if (previous_dex_registers_.size() < current_dex_registers_.size()) {
previous_dex_registers_.resize(current_dex_registers_.size(), DexRegisterLocation::None());
dex_register_timestamp_.resize(current_dex_registers_.size(), 0u);
}
// Set bit in the mask for each register that has been changed since the previous stack map.
// Modified registers are stored in the catalogue and the catalogue index added to the list.
for (size_t i = 0; i < current_dex_registers_.size(); i++) {
DexRegisterLocation reg = current_dex_registers_[i];
// Distance is difference between this index and the index of last modification.
uint32_t distance = stack_maps_.size() - dex_register_timestamp_[i];
if (previous_dex_registers_[i] != reg || distance > kMaxDexRegisterMapSearchDistance) {
BitTableBuilder<DexRegisterInfo>::Entry entry;
entry[DexRegisterInfo::kKind] = static_cast<uint32_t>(reg.GetKind());
entry[DexRegisterInfo::kPackedValue] =
DexRegisterInfo::PackValue(reg.GetKind(), reg.GetValue());
uint32_t index = reg.IsLive() ? dex_register_catalog_.Dedup(&entry) : kNoValue;
temp_dex_register_mask_.SetBit(i);
temp_dex_register_map_.push_back({index});
previous_dex_registers_[i] = reg;
dex_register_timestamp_[i] = stack_maps_.size();
}
}
// Set the mask and map for the current StackMap (which includes inlined registers).
if (temp_dex_register_mask_.GetNumberOfBits() != 0) {
current_stack_map_[StackMap::kDexRegisterMaskIndex] =
dex_register_masks_.Dedup(temp_dex_register_mask_.GetRawStorage(),
temp_dex_register_mask_.GetNumberOfBits());
}
if (!current_dex_registers_.empty()) {
current_stack_map_[StackMap::kDexRegisterMapIndex] =
dex_register_maps_.Dedup(temp_dex_register_map_.data(),
temp_dex_register_map_.size());
}
if (kVerifyStackMaps) {
size_t stack_map_index = stack_maps_.size();
// We need to make copy of the current registers for later (when the check is run).
auto expected_dex_registers = std::make_shared<dchecked_vector<DexRegisterLocation>>(
current_dex_registers_.begin(), current_dex_registers_.end());
dchecks_.emplace_back([=](const CodeInfo& code_info) {
StackMap stack_map = code_info.GetStackMapAt(stack_map_index);
uint32_t expected_reg = 0;
for (DexRegisterLocation reg : code_info.GetDexRegisterMapOf(stack_map)) {
CHECK_EQ((*expected_dex_registers)[expected_reg++], reg);
}
for (InlineInfo inline_info : code_info.GetInlineInfosOf(stack_map)) {
DexRegisterMap map = code_info.GetInlineDexRegisterMapOf(stack_map, inline_info);
for (DexRegisterLocation reg : map) {
CHECK_EQ((*expected_dex_registers)[expected_reg++], reg);
}
}
CHECK_EQ(expected_reg, expected_dex_registers->size());
});
}
}
ScopedArenaVector<uint8_t> StackMapStream::Encode() {
DCHECK(in_stack_map_ == false) << "Mismatched Begin/End calls";
DCHECK(in_inline_info_ == false) << "Mismatched Begin/End calls";
uint32_t flags = 0;
flags |= (inline_infos_.size() > 0) ? CodeInfo::kHasInlineInfo : 0;
flags |= baseline_ ? CodeInfo::kIsBaseline : 0;
flags |= debuggable_ ? CodeInfo::kIsDebuggable : 0;
flags |= has_should_deoptimize_flag_ ? CodeInfo::kHasShouldDeoptimizeFlag : 0;
uint32_t bit_table_flags = 0;
ForEachBitTable([&bit_table_flags](size_t i, auto bit_table) {
if (bit_table->size() != 0) { // Record which bit-tables are stored.
bit_table_flags |= 1 << i;
}
});
ScopedArenaVector<uint8_t> buffer(allocator_->Adapter(kArenaAllocStackMapStream));
BitMemoryWriter<ScopedArenaVector<uint8_t>> out(&buffer);
out.WriteInterleavedVarints(std::array<uint32_t, CodeInfo::kNumHeaders>{
flags,
code_size_,
packed_frame_size_,
core_spill_mask_,
fp_spill_mask_,
num_dex_registers_,
bit_table_flags,
});
ForEachBitTable([&out](size_t, auto bit_table) {
if (bit_table->size() != 0) { // Skip empty bit-tables.
bit_table->Encode(out);
}
});
// Verify that we can load the CodeInfo and check some essentials.
size_t number_of_read_bits;
CodeInfo code_info(buffer.data(), &number_of_read_bits);
CHECK_EQ(number_of_read_bits, out.NumberOfWrittenBits());
CHECK_EQ(code_info.GetNumberOfStackMaps(), stack_maps_.size());
CHECK_EQ(CodeInfo::HasInlineInfo(buffer.data()), inline_infos_.size() > 0);
CHECK_EQ(CodeInfo::IsBaseline(buffer.data()), baseline_);
CHECK_EQ(CodeInfo::IsDebuggable(buffer.data()), debuggable_);
CHECK_EQ(CodeInfo::HasShouldDeoptimizeFlag(buffer.data()), has_should_deoptimize_flag_);
// Verify all written data (usually only in debug builds).
if (kVerifyStackMaps) {
for (const auto& dcheck : dchecks_) {
dcheck(code_info);
}
}
return buffer;
}
} // namespace art