diff options
Diffstat (limited to 'src/compiler/dex/compiler_ir.h')
-rw-r--r-- | src/compiler/dex/compiler_ir.h | 627 |
1 files changed, 627 insertions, 0 deletions
diff --git a/src/compiler/dex/compiler_ir.h b/src/compiler/dex/compiler_ir.h new file mode 100644 index 0000000000..246d200d16 --- /dev/null +++ b/src/compiler/dex/compiler_ir.h @@ -0,0 +1,627 @@ +/* + * Copyright (C) 2011 The Android Open Source Project + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#ifndef ART_SRC_COMPILER_DEX_COMPILER_IR_H_ +#define ART_SRC_COMPILER_DEX_COMPILER_IR_H_ + +#include <vector> +#include "dex_instruction.h" +#include "compiler.h" +#include "compiler_utility.h" +#include "oat_compilation_unit.h" +#include "safe_map.h" +#include "greenland/ir_builder.h" +#include "llvm/Module.h" +#include "compiler_enums.h" + +namespace art { + +#define SLOW_FIELD_PATH (cu->enable_debug & (1 << kDebugSlowFieldPath)) +#define SLOW_INVOKE_PATH (cu->enable_debug & (1 << kDebugSlowInvokePath)) +#define SLOW_STRING_PATH (cu->enable_debug & (1 << kDebugSlowStringPath)) +#define SLOW_TYPE_PATH (cu->enable_debug & (1 << kDebugSlowTypePath)) +#define EXERCISE_SLOWEST_STRING_PATH (cu->enable_debug & \ + (1 << kDebugSlowestStringPath)) + +// Minimum field size to contain Dalvik v_reg number. +#define VREG_NUM_WIDTH 16 + +struct ArenaBitVector; +struct LIR; +class LLVMInfo; + +struct PromotionMap { + RegLocationType core_location:3; + uint8_t core_reg; + RegLocationType fp_location:3; + uint8_t FpReg; + bool first_in_pair; +}; + +struct RegLocation { + RegLocationType location:3; + unsigned wide:1; + unsigned defined:1; // Do we know the type? + unsigned is_const:1; // Constant, value in cu->constant_values[]. + unsigned fp:1; // Floating point? + unsigned core:1; // Non-floating point? + unsigned ref:1; // Something GC cares about. + unsigned high_word:1; // High word of pair? + unsigned home:1; // Does this represent the home location? + uint8_t low_reg; // First physical register. + uint8_t high_reg; // 2nd physical register (if wide). + int32_t s_reg_low; // SSA name for low Dalvik word. + int32_t orig_sreg; // TODO: remove after Bitcode gen complete + // and consolodate usage w/ s_reg_low. +}; + +struct CompilerTemp { + int s_reg; + ArenaBitVector* bv; +}; + +struct CallInfo { + int num_arg_words; // Note: word count, not arg count. + RegLocation* args; // One for each word of arguments. + RegLocation result; // Eventual target of MOVE_RESULT. + int opt_flags; + InvokeType type; + uint32_t dex_idx; + uint32_t index; // Method idx for invokes, type idx for FilledNewArray. + uintptr_t direct_code; + uintptr_t direct_method; + RegLocation target; // Target of following move_result. + bool skip_this; + bool is_range; + int offset; // Dalvik offset. +}; + + /* + * Data structure tracking the mapping between a Dalvik register (pair) and a + * native register (pair). The idea is to reuse the previously loaded value + * if possible, otherwise to keep the value in a native register as long as + * possible. + */ +struct RegisterInfo { + int reg; // Reg number + bool in_use; // Has it been allocated? + bool is_temp; // Can allocate as temp? + bool pair; // Part of a register pair? + int partner; // If pair, other reg of pair. + bool live; // Is there an associated SSA name? + bool dirty; // If live, is it dirty? + int s_reg; // Name of live value. + LIR *def_start; // Starting inst in last def sequence. + LIR *def_end; // Ending inst in last def sequence. +}; + +struct RegisterPool { + int num_core_regs; + RegisterInfo *core_regs; + int next_core_reg; + int num_fp_regs; + RegisterInfo *FPRegs; + int next_fp_reg; +}; + +#define INVALID_SREG (-1) +#define INVALID_VREG (0xFFFFU) +#define INVALID_REG (0xFF) +#define INVALID_OFFSET (0xDEADF00FU) + +/* SSA encodings for special registers */ +#define SSA_METHOD_BASEREG (-2) +/* First compiler temp basereg, grows smaller */ +#define SSA_CTEMP_BASEREG (SSA_METHOD_BASEREG - 1) + +/* + * Some code patterns cause the generation of excessively large + * methods - in particular initialization sequences. There isn't much + * benefit in optimizing these methods, and the cost can be very high. + * We attempt to identify these cases, and avoid performing most dataflow + * analysis. Two thresholds are used - one for known initializers and one + * for everything else. + */ +#define MANY_BLOCKS_INITIALIZER 1000 /* Threshold for switching dataflow off */ +#define MANY_BLOCKS 4000 /* Non-initializer threshold */ + +// Utility macros to traverse the LIR list. +#define NEXT_LIR(lir) (lir->next) +#define PREV_LIR(lir) (lir->prev) + +// Defines for alias_info (tracks Dalvik register references). +#define DECODE_ALIAS_INFO_REG(X) (X & 0xffff) +#define DECODE_ALIAS_INFO_WIDE_FLAG (0x80000000) +#define DECODE_ALIAS_INFO_WIDE(X) ((X & DECODE_ALIAS_INFO_WIDE_FLAG) ? 1 : 0) +#define ENCODE_ALIAS_INFO(REG, ISWIDE) (REG | (ISWIDE ? DECODE_ALIAS_INFO_WIDE_FLAG : 0)) + +// Common resource macros. +#define ENCODE_CCODE (1ULL << kCCode) +#define ENCODE_FP_STATUS (1ULL << kFPStatus) + +// Abstract memory locations. +#define ENCODE_DALVIK_REG (1ULL << kDalvikReg) +#define ENCODE_LITERAL (1ULL << kLiteral) +#define ENCODE_HEAP_REF (1ULL << kHeapRef) +#define ENCODE_MUST_NOT_ALIAS (1ULL << kMustNotAlias) + +#define ENCODE_ALL (~0ULL) +#define ENCODE_MEM (ENCODE_DALVIK_REG | ENCODE_LITERAL | \ + ENCODE_HEAP_REF | ENCODE_MUST_NOT_ALIAS) + +#define is_pseudo_opcode(opcode) (static_cast<int>(opcode) < 0) + +struct LIR { + int offset; // Offset of this instruction. + int dalvik_offset; // Offset of Dalvik opcode. + LIR* next; + LIR* prev; + LIR* target; + int opcode; + int operands[5]; // [0..4] = [dest, src1, src2, extra, extra2]. + struct { + bool is_nop:1; // LIR is optimized away. + bool pcRelFixup:1; // May need pc-relative fixup. + unsigned int size:5; // Note: size is in bytes. + unsigned int unused:25; + } flags; + int alias_info; // For Dalvik register & litpool disambiguation. + uint64_t use_mask; // Resource mask for use. + uint64_t def_mask; // Resource mask for def. +}; + +extern const char* extended_mir_op_names[kMirOpLast - kMirOpFirst]; + +struct SSARepresentation; + +#define MIR_IGNORE_NULL_CHECK (1 << kMIRIgnoreNullCheck) +#define MIR_NULL_CHECK_ONLY (1 << kMIRNullCheckOnly) +#define MIR_IGNORE_RANGE_CHECK (1 << kMIRIgnoreRangeCheck) +#define MIR_RANGE_CHECK_ONLY (1 << kMIRRangeCheckOnly) +#define MIR_INLINED (1 << kMIRInlined) +#define MIR_INLINED_PRED (1 << kMIRInlinedPred) +#define MIR_CALLEE (1 << kMIRCallee) +#define MIR_IGNORE_SUSPEND_CHECK (1 << kMIRIgnoreSuspendCheck) +#define MIR_DUP (1 << kMIRDup) + +struct Checkstats { + int null_checks; + int null_checks_eliminated; + int range_checks; + int range_checks_eliminated; +}; + +struct MIR { + DecodedInstruction dalvikInsn; + unsigned int width; + unsigned int offset; + MIR* prev; + MIR* next; + SSARepresentation* ssa_rep; + int optimization_flags; + union { + // Establish link between two halves of throwing instructions. + MIR* throw_insn; + // Saved opcode for NOP'd MIRs + Instruction::Code original_opcode; + } meta; +}; + +struct BasicBlockDataFlow; + +struct BasicBlock { + int id; + int dfs_id; + bool visited; + bool hidden; + bool catch_entry; + bool explicit_throw; + bool conditional_branch; + bool terminated_by_return; // Block ends with a Dalvik return opcode. + bool dominates_return; // Is a member of return extended basic block. + uint16_t start_offset; + uint16_t nesting_depth; + BBType block_type; + MIR* first_mir_insn; + MIR* last_mir_insn; + BasicBlock* fall_through; + BasicBlock* taken; + BasicBlock* i_dom; // Immediate dominator. + BasicBlockDataFlow* data_flow_info; + GrowableList* predecessors; + ArenaBitVector* dominators; + ArenaBitVector* i_dominated; // Set nodes being immediately dominated. + ArenaBitVector* dom_frontier; // Dominance frontier. + struct { // For one-to-many successors like. + BlockListType block_list_type; // switch and exception handling. + GrowableList blocks; + } successor_block_list; +}; + +/* + * The "blocks" field in "successor_block_list" points to an array of + * elements with the type "SuccessorBlockInfo". + * For catch blocks, key is type index for the exception. + * For swtich blocks, key is the case value. + */ +struct SuccessorBlockInfo { + BasicBlock* block; + int key; +}; + +struct LoopAnalysis; +struct RegisterPool; +struct ArenaMemBlock; +struct Memstats; +class Codegen; + +#define NOTVISITED (-1) + +struct CompilationUnit { + CompilationUnit() + : num_blocks(0), + compiler(NULL), + class_linker(NULL), + dex_file(NULL), + class_loader(NULL), + class_def_idx(0), + method_idx(0), + code_item(NULL), + access_flags(0), + invoke_type(kDirect), + shorty(NULL), + first_lir_insn(NULL), + last_lir_insn(NULL), + literal_list(NULL), + method_literal_list(NULL), + code_literal_list(NULL), + disable_opt(0), + enable_debug(0), + data_offset(0), + total_size(0), + assembler_status(kSuccess), + assembler_retries(0), + verbose(false), + has_loop(false), + has_invoke(false), + qd_mode(false), + reg_pool(NULL), + instruction_set(kNone), + num_ssa_regs(0), + ssa_base_vregs(NULL), + ssa_subscripts(NULL), + ssa_strings(NULL), + vreg_to_ssa_map(NULL), + ssa_last_defs(NULL), + is_constant_v(NULL), + must_flush_constant_v(NULL), + constant_values(NULL), + reg_location(NULL), + promotion_map(NULL), + method_sreg(0), + num_reachable_blocks(0), + num_dalvik_registers(0), + entry_block(NULL), + exit_block(NULL), + cur_block(NULL), + i_dom_list(NULL), + try_block_addr(NULL), + def_block_matrix(NULL), + temp_block_v(NULL), + temp_dalvik_register_v(NULL), + temp_ssa_register_v(NULL), + temp_ssa_block_id_v(NULL), + block_label_list(NULL), + num_ins(0), + num_outs(0), + num_regs(0), + num_core_spills(0), + num_fp_spills(0), + num_compiler_temps(0), + frame_size(0), + core_spill_mask(0U), + fp_spill_mask(0U), + attrs(0U), + current_dalvik_offset(0), + insns(NULL), + insns_size(0U), + disable_dataflow(false), + def_count(0), + compiler_flip_match(false), + arena_head(NULL), + current_arena(NULL), + num_arena_blocks(0), + mstats(NULL), + checkstats(NULL), + gen_bitcode(false), + llvm_info(NULL), + context(NULL), + module(NULL), + func(NULL), + intrinsic_helper(NULL), + irb(NULL), + placeholder_bb(NULL), + entry_bb(NULL), + entryTarget_bb(NULL), + temp_name(0), +#ifndef NDEBUG + live_sreg(0), +#endif + opcode_count(NULL), + cg(NULL) {} + + int num_blocks; + GrowableList block_list; + Compiler* compiler; // Compiler driving this compiler. + ClassLinker* class_linker; // Linker to resolve fields and methods. + const DexFile* dex_file; // DexFile containing the method being compiled. + jobject class_loader; // compiling method's class loader. + uint32_t class_def_idx; // compiling method's defining class definition index. + uint32_t method_idx; // compiling method's index into method_ids of DexFile. + const DexFile::CodeItem* code_item; // compiling method's DexFile code_item. + uint32_t access_flags; // compiling method's access flags. + InvokeType invoke_type; // compiling method's invocation type. + const char* shorty; // compiling method's shorty. + LIR* first_lir_insn; + LIR* last_lir_insn; + LIR* literal_list; // Constants. + LIR* method_literal_list; // Method literals requiring patching. + LIR* code_literal_list; // Code literals requiring patching. + uint32_t disable_opt; // opt_control_vector flags. + uint32_t enable_debug; // debugControlVector flags. + int data_offset; // starting offset of literal pool. + int total_size; // header + code size. + AssemblerStatus assembler_status; // Success or fix and retry. + int assembler_retries; + std::vector<uint8_t> code_buffer; + /* + * Holds mapping from native PC to dex PC for safepoints where we may deoptimize. + * Native PC is on the return address of the safepointed operation. Dex PC is for + * the instruction being executed at the safepoint. + */ + std::vector<uint32_t> pc2dexMappingTable; + /* + * Holds mapping from Dex PC to native PC for catch entry points. Native PC and Dex PC + * immediately preceed the instruction. + */ + std::vector<uint32_t> dex2pcMappingTable; + std::vector<uint32_t> combined_mapping_table; + std::vector<uint32_t> core_vmap_table; + std::vector<uint32_t> fp_vmap_table; + std::vector<uint8_t> native_gc_map; + std::vector<BasicBlock*> extended_basic_blocks; + bool verbose; + bool has_loop; // Contains a loop. + bool has_invoke; // Contains an invoke instruction. + bool qd_mode; // Compile for code size/compile time. + RegisterPool* reg_pool; + InstructionSet instruction_set; + // Number of total regs used in the whole cu after SSA transformation . + int num_ssa_regs; + // Map SSA reg i to the base virtual register/subscript. + GrowableList* ssa_base_vregs; + GrowableList* ssa_subscripts; + GrowableList* ssa_strings; + + // Map original Dalvik virtual reg i to the current SSA name. + int* vreg_to_ssa_map; // length == method->registers_size + int* ssa_last_defs; // length == method->registers_size + ArenaBitVector* is_constant_v; // length == num_ssa_reg + ArenaBitVector* must_flush_constant_v; // length == num_ssa_reg + int* constant_values; // length == num_ssa_reg + + // Use counts of ssa names. + GrowableList use_counts; // Weighted by nesting depth + GrowableList raw_use_counts; // Not weighted + + // Optimization support. + GrowableList loop_headers; + + // Map SSA names to location. + RegLocation* reg_location; + + // Keep track of Dalvik v_reg to physical register mappings. + PromotionMap* promotion_map; + + // SSA name for Method*. + int method_sreg; + RegLocation method_loc; // Describes location of method*. + + int num_reachable_blocks; + int num_dalvik_registers; // method->registers_size. + BasicBlock* entry_block; + BasicBlock* exit_block; + BasicBlock* cur_block; + GrowableList dfs_order; + GrowableList dfs_post_order; + GrowableList dom_post_order_traversal; + GrowableList throw_launchpads; + GrowableList suspend_launchpads; + GrowableList intrinsic_launchpads; + GrowableList compiler_temps; + int* i_dom_list; + ArenaBitVector* try_block_addr; + ArenaBitVector** def_block_matrix; // num_dalvik_register x num_blocks. + ArenaBitVector* temp_block_v; + ArenaBitVector* temp_dalvik_register_v; + ArenaBitVector* temp_ssa_register_v; // num_ssa_regs. + int* temp_ssa_block_id_v; // working storage for Phi labels. + LIR* block_label_list; + /* + * Frame layout details. + * NOTE: for debug support it will be necessary to add a structure + * to map the Dalvik virtual registers to the promoted registers. + * NOTE: "num" fields are in 4-byte words, "Size" and "Offset" in bytes. + */ + int num_ins; + int num_outs; + int num_regs; // Unlike num_dalvik_registers, does not include ins. + int num_core_spills; + int num_fp_spills; + int num_compiler_temps; + int frame_size; + unsigned int core_spill_mask; + unsigned int fp_spill_mask; + unsigned int attrs; + /* + * TODO: The code generation utilities don't have a built-in + * mechanism to propagate the original Dalvik opcode address to the + * associated generated instructions. For the trace compiler, this wasn't + * necessary because the interpreter handled all throws and debugging + * requests. For now we'll handle this by placing the Dalvik offset + * in the CompilationUnit struct before codegen for each instruction. + * The low-level LIR creation utilites will pull it from here. Rework this. + */ + int current_dalvik_offset; + GrowableList switch_tables; + GrowableList fill_array_data; + const uint16_t* insns; + uint32_t insns_size; + bool disable_dataflow; // Skip dataflow analysis if possible + SafeMap<unsigned int, BasicBlock*> block_map; // FindBlock lookup cache. + SafeMap<unsigned int, unsigned int> block_id_map; // Block collapse lookup cache. + SafeMap<unsigned int, LIR*> boundary_map; // boundary lookup cache. + int def_count; // Used to estimate number of SSA names. + + // If non-empty, apply optimizer/debug flags only to matching methods. + std::string compiler_method_match; + // Flips sense of compiler_method_match - apply flags if doesn't match. + bool compiler_flip_match; + ArenaMemBlock* arena_head; + ArenaMemBlock* current_arena; + int num_arena_blocks; + Memstats* mstats; + Checkstats* checkstats; + bool gen_bitcode; + LLVMInfo* llvm_info; + llvm::LLVMContext* context; + llvm::Module* module; + llvm::Function* func; + greenland::IntrinsicHelper* intrinsic_helper; + greenland::IRBuilder* irb; + llvm::BasicBlock* placeholder_bb; + llvm::BasicBlock* entry_bb; + llvm::BasicBlock* entryTarget_bb; + std::string bitcode_filename; + GrowableList llvm_values; + int32_t temp_name; + SafeMap<llvm::BasicBlock*, LIR*> block_to_label_map; // llvm bb -> LIR label. + SafeMap<int32_t, llvm::BasicBlock*> id_to_block_map; // block id -> llvm bb. + SafeMap<llvm::Value*, RegLocation> loc_map; // llvm Value to loc rec. + std::set<llvm::BasicBlock*> llvm_blocks; +#ifndef NDEBUG + /* + * Sanity checking for the register temp tracking. The same ssa + * name should never be associated with one temp register per + * instruction compilation. + */ + int live_sreg; +#endif + std::set<uint32_t> catches; + int* opcode_count; // Count Dalvik opcodes for tuning. + UniquePtr<Codegen> cg; +}; + +struct SwitchTable { + int offset; + const uint16_t* table; // Original dex table. + int vaddr; // Dalvik offset of switch opcode. + LIR* anchor; // Reference instruction for relative offsets. + LIR** targets; // Array of case targets. +}; + +struct FillArrayData { + int offset; + const uint16_t* table; // Original dex table. + int size; + int vaddr; // Dalvik offset of FILL_ARRAY_DATA opcode. +}; + +#define MAX_PATTERN_LEN 5 + +struct CodePattern { + const Instruction::Code opcodes[MAX_PATTERN_LEN]; + const SpecialCaseHandler handler_code; +}; + +static const CodePattern special_patterns[] = { + {{Instruction::RETURN_VOID}, kNullMethod}, + {{Instruction::CONST, Instruction::RETURN}, kConstFunction}, + {{Instruction::CONST_4, Instruction::RETURN}, kConstFunction}, + {{Instruction::CONST_4, Instruction::RETURN_OBJECT}, kConstFunction}, + {{Instruction::CONST_16, Instruction::RETURN}, kConstFunction}, + {{Instruction::IGET, Instruction:: RETURN}, kIGet}, + {{Instruction::IGET_BOOLEAN, Instruction::RETURN}, kIGetBoolean}, + {{Instruction::IGET_OBJECT, Instruction::RETURN_OBJECT}, kIGetObject}, + {{Instruction::IGET_BYTE, Instruction::RETURN}, kIGetByte}, + {{Instruction::IGET_CHAR, Instruction::RETURN}, kIGetChar}, + {{Instruction::IGET_SHORT, Instruction::RETURN}, kIGetShort}, + {{Instruction::IGET_WIDE, Instruction::RETURN_WIDE}, kIGetWide}, + {{Instruction::IPUT, Instruction::RETURN_VOID}, kIPut}, + {{Instruction::IPUT_BOOLEAN, Instruction::RETURN_VOID}, kIPutBoolean}, + {{Instruction::IPUT_OBJECT, Instruction::RETURN_VOID}, kIPutObject}, + {{Instruction::IPUT_BYTE, Instruction::RETURN_VOID}, kIPutByte}, + {{Instruction::IPUT_CHAR, Instruction::RETURN_VOID}, kIPutChar}, + {{Instruction::IPUT_SHORT, Instruction::RETURN_VOID}, kIPutShort}, + {{Instruction::IPUT_WIDE, Instruction::RETURN_VOID}, kIPutWide}, + {{Instruction::RETURN}, kIdentity}, + {{Instruction::RETURN_OBJECT}, kIdentity}, + {{Instruction::RETURN_WIDE}, kIdentity}, +}; + +static inline bool IsConst(const CompilationUnit* cu, int32_t s_reg) +{ + return (IsBitSet(cu->is_constant_v, s_reg)); +} + +static inline bool IsConst(const CompilationUnit* cu, RegLocation loc) +{ + return (IsConst(cu, loc.orig_sreg)); +} + +static inline int32_t ConstantValue(const CompilationUnit* cu, RegLocation loc) +{ + DCHECK(IsConst(cu, loc)); + return cu->constant_values[loc.orig_sreg]; +} + +static inline int32_t ConstantValue(const CompilationUnit* cu, int32_t s_reg) +{ + DCHECK(IsConst(cu, s_reg)); + return cu->constant_values[s_reg]; +} + +static inline int64_t ConstantValueWide(const CompilationUnit* cu, RegLocation loc) +{ + DCHECK(IsConst(cu, loc)); + return (static_cast<int64_t>(cu->constant_values[loc.orig_sreg + 1]) << 32) | + Low32Bits(static_cast<int64_t>(cu->constant_values[loc.orig_sreg])); +} + +static inline bool IsConstantNullRef(const CompilationUnit* cu, RegLocation loc) +{ + return loc.ref && loc.is_const && (ConstantValue(cu, loc) == 0); +} + +static inline bool MustFlushConstant(const CompilationUnit* cu, RegLocation loc) +{ + DCHECK(IsConst(cu, loc)); + return IsBitSet(cu->must_flush_constant_v, loc.orig_sreg); +} + +} // namespace art + +#endif // ART_SRC_COMPILER_DEX_COMPILER_IR_H_ |