/* * Copyright (C) 2011 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "compiled_method.h" namespace art { CompiledMethod::CompiledMethod(InstructionSet instruction_set, const std::vector& code, const size_t frame_size_in_bytes, const uint32_t core_spill_mask, const uint32_t fp_spill_mask, const std::vector& mapping_table, const std::vector& vmap_table) : instruction_set_(instruction_set), frame_size_in_bytes_(frame_size_in_bytes), core_spill_mask_(core_spill_mask), fp_spill_mask_(fp_spill_mask), elf_idx_(-1) { CHECK_NE(code.size(), 0U); DCHECK_EQ(vmap_table.size(), static_cast(__builtin_popcount(core_spill_mask) + __builtin_popcount(fp_spill_mask))); CHECK_LE(vmap_table.size(), (1U << 16) - 1); // length must fit in 2^16-1 size_t code_byte_count = code.size() * sizeof(code[0]); std::vector byte_code(code_byte_count); memcpy(&byte_code[0], &code[0], code_byte_count); std::vector length_prefixed_mapping_table; length_prefixed_mapping_table.push_back(mapping_table.size()); length_prefixed_mapping_table.insert(length_prefixed_mapping_table.end(), mapping_table.begin(), mapping_table.end()); DCHECK_EQ(mapping_table.size() + 1, length_prefixed_mapping_table.size()); std::vector length_prefixed_vmap_table; length_prefixed_vmap_table.push_back(vmap_table.size()); length_prefixed_vmap_table.insert(length_prefixed_vmap_table.end(), vmap_table.begin(), vmap_table.end()); DCHECK_EQ(vmap_table.size() + 1, length_prefixed_vmap_table.size()); DCHECK_EQ(vmap_table.size(), length_prefixed_vmap_table[0]); code_ = byte_code; mapping_table_ = length_prefixed_mapping_table; vmap_table_ = length_prefixed_vmap_table; DCHECK_EQ(vmap_table_[0], static_cast(__builtin_popcount(core_spill_mask) + __builtin_popcount(fp_spill_mask))); } void CompiledMethod::SetGcMap(const std::vector& gc_map) { CHECK_NE(gc_map.size(), 0U); #if !defined(ART_USE_LLVM_COMPILER) && !defined(ART_USE_GREENLAND_COMPILER) // Should only be used with CompiledMethods created with the non-LLVM compilers. CHECK_NE(mapping_table_.size(), 0U); #endif gc_map_ = gc_map; } CompiledMethod::CompiledMethod(InstructionSet instruction_set, const std::vector& code, const size_t frame_size_in_bytes, const uint32_t core_spill_mask, const uint32_t fp_spill_mask) : instruction_set_(instruction_set), code_(code), frame_size_in_bytes_(frame_size_in_bytes), core_spill_mask_(core_spill_mask), fp_spill_mask_(fp_spill_mask), elf_idx_(-1) { CHECK_NE(code.size(), 0U); } CompiledMethod::CompiledMethod(InstructionSet instruction_set, const uint16_t elf_idx, const uint16_t elf_func_idx) : instruction_set_(instruction_set), frame_size_in_bytes_(0), core_spill_mask_(0), fp_spill_mask_(0), elf_idx_(elf_idx), elf_func_idx_(elf_func_idx) { } CompiledMethod::~CompiledMethod() {} InstructionSet CompiledMethod::GetInstructionSet() const { return instruction_set_; } const std::vector& CompiledMethod::GetCode() const { return code_; } size_t CompiledMethod::GetFrameSizeInBytes() const { return frame_size_in_bytes_; } uint32_t CompiledMethod::GetCoreSpillMask() const { return core_spill_mask_; } uint32_t CompiledMethod::GetFpSpillMask() const { return fp_spill_mask_; } const std::vector& CompiledMethod::GetMappingTable() const { return mapping_table_; } const std::vector& CompiledMethod::GetVmapTable() const { return vmap_table_; } const std::vector& CompiledMethod::GetGcMap() const { return gc_map_; } uint32_t CompiledMethod::AlignCode(uint32_t offset) const { return AlignCode(offset, instruction_set_); } uint32_t CompiledMethod::AlignCode(uint32_t offset, InstructionSet instruction_set) { switch (instruction_set) { case kArm: case kThumb2: return RoundUp(offset, kArmAlignment); case kMips: return RoundUp(offset, kMipsAlignment); case kX86: return RoundUp(offset, kX86Alignment); default: LOG(FATAL) << "Unknown InstructionSet: " << instruction_set; return 0; } } size_t CompiledMethod::CodeDelta() const { switch (instruction_set_) { case kArm: case kMips: case kX86: return 0; case kThumb2: { // +1 to set the low-order bit so a BLX will switch to Thumb mode return 1; } default: LOG(FATAL) << "Unknown InstructionSet: " << instruction_set_; return 0; } } const void* CompiledMethod::CodePointer(const void* code_pointer, InstructionSet instruction_set) { switch (instruction_set) { case kArm: case kMips: case kX86: return code_pointer; case kThumb2: { uintptr_t address = reinterpret_cast(code_pointer); // Set the low-order bit so a BLX will switch to Thumb mode address |= 0x1; return reinterpret_cast(address); } default: LOG(FATAL) << "Unknown InstructionSet: " << instruction_set; return NULL; } } #if defined(ART_USE_LLVM_COMPILER) CompiledInvokeStub::CompiledInvokeStub(uint16_t elf_idx) : elf_idx_(elf_idx) { } #endif CompiledInvokeStub::CompiledInvokeStub(std::vector& code) : elf_idx_(-1) { CHECK_NE(code.size(), 0U); code_ = code; } CompiledInvokeStub::~CompiledInvokeStub() {} const std::vector& CompiledInvokeStub::GetCode() const { return code_; } } // namespace art