diff options
Diffstat (limited to 'compiler/optimizing')
-rw-r--r-- | compiler/optimizing/code_generator.cc | 19 | ||||
-rw-r--r-- | compiler/optimizing/code_generator_arm_vixl.cc | 2145 | ||||
-rw-r--r-- | compiler/optimizing/code_generator_arm_vixl.h | 405 | ||||
-rw-r--r-- | compiler/optimizing/codegen_test.cc | 32 | ||||
-rw-r--r-- | compiler/optimizing/common_arm.h | 127 |
5 files changed, 5 insertions, 2723 deletions
diff --git a/compiler/optimizing/code_generator.cc b/compiler/optimizing/code_generator.cc index cf633df496..137cd21864 100644 --- a/compiler/optimizing/code_generator.cc +++ b/compiler/optimizing/code_generator.cc @@ -18,7 +18,6 @@ #ifdef ART_ENABLE_CODEGEN_arm #include "code_generator_arm.h" -#include "code_generator_arm_vixl.h" #endif #ifdef ART_ENABLE_CODEGEN_arm64 @@ -576,19 +575,11 @@ std::unique_ptr<CodeGenerator> CodeGenerator::Create(HGraph* graph, #ifdef ART_ENABLE_CODEGEN_arm case kArm: case kThumb2: { - if (kArmUseVIXL32) { - return std::unique_ptr<CodeGenerator>( - new (arena) arm::CodeGeneratorARMVIXL(graph, - *isa_features.AsArmInstructionSetFeatures(), - compiler_options, - stats)); - } else { - return std::unique_ptr<CodeGenerator>( - new (arena) arm::CodeGeneratorARM(graph, - *isa_features.AsArmInstructionSetFeatures(), - compiler_options, - stats)); - } + return std::unique_ptr<CodeGenerator>( + new (arena) arm::CodeGeneratorARM(graph, + *isa_features.AsArmInstructionSetFeatures(), + compiler_options, + stats)); } #endif #ifdef ART_ENABLE_CODEGEN_arm64 diff --git a/compiler/optimizing/code_generator_arm_vixl.cc b/compiler/optimizing/code_generator_arm_vixl.cc deleted file mode 100644 index 4574ad6e29..0000000000 --- a/compiler/optimizing/code_generator_arm_vixl.cc +++ /dev/null @@ -1,2145 +0,0 @@ -/* - * Copyright (C) 2016 The Android Open Source Project - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -#include "code_generator_arm_vixl.h" - -#include "arch/arm/instruction_set_features_arm.h" -#include "art_method.h" -#include "code_generator_utils.h" -#include "common_arm.h" -#include "compiled_method.h" -#include "entrypoints/quick/quick_entrypoints.h" -#include "gc/accounting/card_table.h" -#include "mirror/array-inl.h" -#include "mirror/class-inl.h" -#include "thread.h" -#include "utils/arm/assembler_arm_vixl.h" -#include "utils/arm/managed_register_arm.h" -#include "utils/assembler.h" -#include "utils/stack_checks.h" - -namespace art { -namespace arm { - -namespace vixl32 = vixl::aarch32; -using namespace vixl32; // NOLINT(build/namespaces) - -using helpers::DWARFReg; -using helpers::FromLowSToD; -using helpers::OutputRegister; -using helpers::InputRegisterAt; -using helpers::InputOperandAt; -using helpers::OutputSRegister; -using helpers::InputSRegisterAt; - -using RegisterList = vixl32::RegisterList; - -static bool ExpectedPairLayout(Location location) { - // We expected this for both core and fpu register pairs. - return ((location.low() & 1) == 0) && (location.low() + 1 == location.high()); -} - -static constexpr size_t kArmInstrMaxSizeInBytes = 4u; - -#ifdef __ -#error "ARM Codegen VIXL macro-assembler macro already defined." -#endif - -// TODO: Remove with later pop when codegen complete. -#pragma GCC diagnostic push -#pragma GCC diagnostic ignored "-Wunused-parameter" - -// NOLINT on __ macro to suppress wrong warning/fix (misc-macro-parentheses) from clang-tidy. -#define __ down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()-> // NOLINT -#define QUICK_ENTRY_POINT(x) QUICK_ENTRYPOINT_OFFSET(kArmPointerSize, x).Int32Value() - -// Marker that code is yet to be, and must, be implemented. -#define TODO_VIXL32(level) LOG(level) << __PRETTY_FUNCTION__ << " unimplemented " - -class DivZeroCheckSlowPathARMVIXL : public SlowPathCodeARMVIXL { - public: - explicit DivZeroCheckSlowPathARMVIXL(HDivZeroCheck* instruction) - : SlowPathCodeARMVIXL(instruction) {} - - void EmitNativeCode(CodeGenerator* codegen) OVERRIDE { - CodeGeneratorARMVIXL* armvixl_codegen = down_cast<CodeGeneratorARMVIXL*>(codegen); - __ Bind(GetEntryLabel()); - if (instruction_->CanThrowIntoCatchBlock()) { - // Live registers will be restored in the catch block if caught. - SaveLiveRegisters(codegen, instruction_->GetLocations()); - } - armvixl_codegen->InvokeRuntime(kQuickThrowDivZero, - instruction_, - instruction_->GetDexPc(), - this); - CheckEntrypointTypes<kQuickThrowDivZero, void, void>(); - } - - bool IsFatal() const OVERRIDE { return true; } - - const char* GetDescription() const OVERRIDE { return "DivZeroCheckSlowPathARMVIXL"; } - - private: - DISALLOW_COPY_AND_ASSIGN(DivZeroCheckSlowPathARMVIXL); -}; - -inline vixl32::Condition ARMCondition(IfCondition cond) { - switch (cond) { - case kCondEQ: return eq; - case kCondNE: return ne; - case kCondLT: return lt; - case kCondLE: return le; - case kCondGT: return gt; - case kCondGE: return ge; - case kCondB: return lo; - case kCondBE: return ls; - case kCondA: return hi; - case kCondAE: return hs; - } - LOG(FATAL) << "Unreachable"; - UNREACHABLE(); -} - -// Maps signed condition to unsigned condition. -inline vixl32::Condition ARMUnsignedCondition(IfCondition cond) { - switch (cond) { - case kCondEQ: return eq; - case kCondNE: return ne; - // Signed to unsigned. - case kCondLT: return lo; - case kCondLE: return ls; - case kCondGT: return hi; - case kCondGE: return hs; - // Unsigned remain unchanged. - case kCondB: return lo; - case kCondBE: return ls; - case kCondA: return hi; - case kCondAE: return hs; - } - LOG(FATAL) << "Unreachable"; - UNREACHABLE(); -} - -inline vixl32::Condition ARMFPCondition(IfCondition cond, bool gt_bias) { - // The ARM condition codes can express all the necessary branches, see the - // "Meaning (floating-point)" column in the table A8-1 of the ARMv7 reference manual. - // There is no dex instruction or HIR that would need the missing conditions - // "equal or unordered" or "not equal". - switch (cond) { - case kCondEQ: return eq; - case kCondNE: return ne /* unordered */; - case kCondLT: return gt_bias ? cc : lt /* unordered */; - case kCondLE: return gt_bias ? ls : le /* unordered */; - case kCondGT: return gt_bias ? hi /* unordered */ : gt; - case kCondGE: return gt_bias ? cs /* unordered */ : ge; - default: - LOG(FATAL) << "UNREACHABLE"; - UNREACHABLE(); - } -} - -void SlowPathCodeARMVIXL::SaveLiveRegisters(CodeGenerator* codegen ATTRIBUTE_UNUSED, - LocationSummary* locations ATTRIBUTE_UNUSED) { - TODO_VIXL32(FATAL); -} - -void SlowPathCodeARMVIXL::RestoreLiveRegisters(CodeGenerator* codegen ATTRIBUTE_UNUSED, - LocationSummary* locations ATTRIBUTE_UNUSED) { - TODO_VIXL32(FATAL); -} - -void CodeGeneratorARMVIXL::DumpCoreRegister(std::ostream& stream, int reg) const { - stream << vixl32::Register(reg); -} - -void CodeGeneratorARMVIXL::DumpFloatingPointRegister(std::ostream& stream, int reg) const { - stream << vixl32::SRegister(reg); -} - -static uint32_t ComputeSRegisterMask(const SRegisterList& regs) { - uint32_t mask = 0; - for (uint32_t i = regs.GetFirstSRegister().GetCode(); - i <= regs.GetLastSRegister().GetCode(); - ++i) { - mask |= (1 << i); - } - return mask; -} - -#undef __ - -CodeGeneratorARMVIXL::CodeGeneratorARMVIXL(HGraph* graph, - const ArmInstructionSetFeatures& isa_features, - const CompilerOptions& compiler_options, - OptimizingCompilerStats* stats) - : CodeGenerator(graph, - kNumberOfCoreRegisters, - kNumberOfSRegisters, - kNumberOfRegisterPairs, - kCoreCalleeSaves.GetList(), - ComputeSRegisterMask(kFpuCalleeSaves), - compiler_options, - stats), - block_labels_(graph->GetArena()->Adapter(kArenaAllocCodeGenerator)), - location_builder_(graph, this), - instruction_visitor_(graph, this), - move_resolver_(graph->GetArena(), this), - assembler_(graph->GetArena()), - isa_features_(isa_features) { - // Always save the LR register to mimic Quick. - AddAllocatedRegister(Location::RegisterLocation(LR)); -} - -#define __ reinterpret_cast<ArmVIXLAssembler*>(GetAssembler())->GetVIXLAssembler()-> - -void CodeGeneratorARMVIXL::Finalize(CodeAllocator* allocator) { - GetAssembler()->FinalizeCode(); - CodeGenerator::Finalize(allocator); -} - -void CodeGeneratorARMVIXL::SetupBlockedRegisters() const { - // Don't allocate the dalvik style register pair passing. - blocked_register_pairs_[R1_R2] = true; - - // Stack register, LR and PC are always reserved. - blocked_core_registers_[SP] = true; - blocked_core_registers_[LR] = true; - blocked_core_registers_[PC] = true; - - // Reserve thread register. - blocked_core_registers_[TR] = true; - - // Reserve temp register. - blocked_core_registers_[IP] = true; - - if (GetGraph()->IsDebuggable()) { - // Stubs do not save callee-save floating point registers. If the graph - // is debuggable, we need to deal with these registers differently. For - // now, just block them. - for (uint32_t i = kFpuCalleeSaves.GetFirstSRegister().GetCode(); - i <= kFpuCalleeSaves.GetLastSRegister().GetCode(); - ++i) { - blocked_fpu_registers_[i] = true; - } - } - - UpdateBlockedPairRegisters(); -} - -// Blocks all register pairs containing blocked core registers. -void CodeGeneratorARMVIXL::UpdateBlockedPairRegisters() const { - for (int i = 0; i < kNumberOfRegisterPairs; i++) { - ArmManagedRegister current = - ArmManagedRegister::FromRegisterPair(static_cast<RegisterPair>(i)); - if (blocked_core_registers_[current.AsRegisterPairLow()] - || blocked_core_registers_[current.AsRegisterPairHigh()]) { - blocked_register_pairs_[i] = true; - } - } -} - -void InstructionCodeGeneratorARMVIXL::GenerateSuspendCheck(HSuspendCheck* instruction, - HBasicBlock* successor) { - TODO_VIXL32(FATAL); -} - -InstructionCodeGeneratorARMVIXL::InstructionCodeGeneratorARMVIXL(HGraph* graph, - CodeGeneratorARMVIXL* codegen) - : InstructionCodeGenerator(graph, codegen), - assembler_(codegen->GetAssembler()), - codegen_(codegen) {} - -void CodeGeneratorARMVIXL::ComputeSpillMask() { - core_spill_mask_ = allocated_registers_.GetCoreRegisters() & core_callee_save_mask_; - DCHECK_NE(core_spill_mask_, 0u) << "At least the return address register must be saved"; - // There is no easy instruction to restore just the PC on thumb2. We spill and - // restore another arbitrary register. - core_spill_mask_ |= (1 << kCoreAlwaysSpillRegister.GetCode()); - fpu_spill_mask_ = allocated_registers_.GetFloatingPointRegisters() & fpu_callee_save_mask_; - // We use vpush and vpop for saving and restoring floating point registers, which take - // a SRegister and the number of registers to save/restore after that SRegister. We - // therefore update the `fpu_spill_mask_` to also contain those registers not allocated, - // but in the range. - if (fpu_spill_mask_ != 0) { - uint32_t least_significant_bit = LeastSignificantBit(fpu_spill_mask_); - uint32_t most_significant_bit = MostSignificantBit(fpu_spill_mask_); - for (uint32_t i = least_significant_bit + 1 ; i < most_significant_bit; ++i) { - fpu_spill_mask_ |= (1 << i); - } - } -} - -void CodeGeneratorARMVIXL::GenerateFrameEntry() { - bool skip_overflow_check = - IsLeafMethod() && !FrameNeedsStackCheck(GetFrameSize(), InstructionSet::kArm); - DCHECK(GetCompilerOptions().GetImplicitStackOverflowChecks()); - __ Bind(&frame_entry_label_); - - if (HasEmptyFrame()) { - return; - } - - UseScratchRegisterScope temps(GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - if (!skip_overflow_check) { - __ Sub(temp, sp, static_cast<int32_t>(GetStackOverflowReservedBytes(kArm))); - // The load must immediately precede RecordPcInfo. - { - AssemblerAccurateScope aas(GetVIXLAssembler(), - kArmInstrMaxSizeInBytes, - CodeBufferCheckScope::kMaximumSize); - __ ldr(temp, MemOperand(temp)); - RecordPcInfo(nullptr, 0); - } - } - - __ Push(RegisterList(core_spill_mask_)); - GetAssembler()->cfi().AdjustCFAOffset(kArmWordSize * POPCOUNT(core_spill_mask_)); - GetAssembler()->cfi().RelOffsetForMany(DWARFReg(kMethodRegister), - 0, - core_spill_mask_, - kArmWordSize); - if (fpu_spill_mask_ != 0) { - uint32_t first = LeastSignificantBit(fpu_spill_mask_); - - // Check that list is contiguous. - DCHECK_EQ(fpu_spill_mask_ >> CTZ(fpu_spill_mask_), ~0u >> (32 - POPCOUNT(fpu_spill_mask_))); - - __ Vpush(SRegisterList(vixl32::SRegister(first), POPCOUNT(fpu_spill_mask_))); - GetAssembler()->cfi().AdjustCFAOffset(kArmWordSize * POPCOUNT(fpu_spill_mask_)); - GetAssembler()->cfi().RelOffsetForMany(DWARFReg(s0), - 0, - fpu_spill_mask_, - kArmWordSize); - } - int adjust = GetFrameSize() - FrameEntrySpillSize(); - __ Sub(sp, sp, adjust); - GetAssembler()->cfi().AdjustCFAOffset(adjust); - GetAssembler()->StoreToOffset(kStoreWord, kMethodRegister, sp, 0); -} - -void CodeGeneratorARMVIXL::GenerateFrameExit() { - if (HasEmptyFrame()) { - __ Bx(lr); - return; - } - GetAssembler()->cfi().RememberState(); - int adjust = GetFrameSize() - FrameEntrySpillSize(); - __ Add(sp, sp, adjust); - GetAssembler()->cfi().AdjustCFAOffset(-adjust); - if (fpu_spill_mask_ != 0) { - uint32_t first = LeastSignificantBit(fpu_spill_mask_); - - // Check that list is contiguous. - DCHECK_EQ(fpu_spill_mask_ >> CTZ(fpu_spill_mask_), ~0u >> (32 - POPCOUNT(fpu_spill_mask_))); - - __ Vpop(SRegisterList(vixl32::SRegister(first), POPCOUNT(fpu_spill_mask_))); - GetAssembler()->cfi().AdjustCFAOffset( - -static_cast<int>(kArmWordSize) * POPCOUNT(fpu_spill_mask_)); - GetAssembler()->cfi().RestoreMany(DWARFReg(vixl32::SRegister(0)), - fpu_spill_mask_); - } - // Pop LR into PC to return. - DCHECK_NE(core_spill_mask_ & (1 << kLrCode), 0U); - uint32_t pop_mask = (core_spill_mask_ & (~(1 << kLrCode))) | 1 << kPcCode; - __ Pop(RegisterList(pop_mask)); - GetAssembler()->cfi().RestoreState(); - GetAssembler()->cfi().DefCFAOffset(GetFrameSize()); -} - -void CodeGeneratorARMVIXL::Bind(HBasicBlock* block) { - __ Bind(GetLabelOf(block)); -} - -void CodeGeneratorARMVIXL::MoveConstant(Location destination, int32_t value) { - TODO_VIXL32(FATAL); -} - -void CodeGeneratorARMVIXL::MoveLocation(Location dst, Location src, Primitive::Type dst_type) { - TODO_VIXL32(FATAL); -} - -void CodeGeneratorARMVIXL::AddLocationAsTemp(Location location, LocationSummary* locations) { - TODO_VIXL32(FATAL); -} - -uintptr_t CodeGeneratorARMVIXL::GetAddressOf(HBasicBlock* block) { - TODO_VIXL32(FATAL); - return 0; -} - -void CodeGeneratorARMVIXL::GenerateImplicitNullCheck(HNullCheck* null_check) { - TODO_VIXL32(FATAL); -} - -void CodeGeneratorARMVIXL::GenerateExplicitNullCheck(HNullCheck* null_check) { - TODO_VIXL32(FATAL); -} - -void CodeGeneratorARMVIXL::InvokeRuntime(QuickEntrypointEnum entrypoint, - HInstruction* instruction, - uint32_t dex_pc, - SlowPathCode* slow_path) { - ValidateInvokeRuntime(instruction, slow_path); - GenerateInvokeRuntime(GetThreadOffset<kArmPointerSize>(entrypoint).Int32Value()); - if (EntrypointRequiresStackMap(entrypoint)) { - RecordPcInfo(instruction, dex_pc, slow_path); - } -} - -void CodeGeneratorARMVIXL::InvokeRuntimeWithoutRecordingPcInfo(int32_t entry_point_offset, - HInstruction* instruction, - SlowPathCode* slow_path) { - ValidateInvokeRuntimeWithoutRecordingPcInfo(instruction, slow_path); - GenerateInvokeRuntime(entry_point_offset); -} - -void CodeGeneratorARMVIXL::GenerateInvokeRuntime(int32_t entry_point_offset) { - GetAssembler()->LoadFromOffset(kLoadWord, lr, tr, entry_point_offset); - __ Blx(lr); -} - -// Check if the desired_string_load_kind is supported. If it is, return it, -// otherwise return a fall-back kind that should be used instead. -HLoadString::LoadKind CodeGeneratorARMVIXL::GetSupportedLoadStringKind( - HLoadString::LoadKind desired_string_load_kind) { - TODO_VIXL32(FATAL); - return desired_string_load_kind; -} - -// Check if the desired_class_load_kind is supported. If it is, return it, -// otherwise return a fall-back kind that should be used instead. -HLoadClass::LoadKind CodeGeneratorARMVIXL::GetSupportedLoadClassKind( - HLoadClass::LoadKind desired_class_load_kind) { - TODO_VIXL32(FATAL); - return desired_class_load_kind; -} - -// Check if the desired_dispatch_info is supported. If it is, return it, -// otherwise return a fall-back info that should be used instead. -HInvokeStaticOrDirect::DispatchInfo CodeGeneratorARMVIXL::GetSupportedInvokeStaticOrDirectDispatch( - const HInvokeStaticOrDirect::DispatchInfo& desired_dispatch_info, - MethodReference target_method) { - TODO_VIXL32(FATAL); - return desired_dispatch_info; -} - -// Generate a call to a static or direct method. -void CodeGeneratorARMVIXL::GenerateStaticOrDirectCall(HInvokeStaticOrDirect* invoke, - Location temp) { - TODO_VIXL32(FATAL); -} - -// Generate a call to a virtual method. -void CodeGeneratorARMVIXL::GenerateVirtualCall(HInvokeVirtual* invoke, Location temp) { - TODO_VIXL32(FATAL); -} - -// Copy the result of a call into the given target. -void CodeGeneratorARMVIXL::MoveFromReturnRegister(Location trg, Primitive::Type type) { - TODO_VIXL32(FATAL); -} - -void InstructionCodeGeneratorARMVIXL::HandleGoto(HInstruction* got, HBasicBlock* successor) { - DCHECK(!successor->IsExitBlock()); - HBasicBlock* block = got->GetBlock(); - HInstruction* previous = got->GetPrevious(); - HLoopInformation* info = block->GetLoopInformation(); - - if (info != nullptr && info->IsBackEdge(*block) && info->HasSuspendCheck()) { - codegen_->ClearSpillSlotsFromLoopPhisInStackMap(info->GetSuspendCheck()); - GenerateSuspendCheck(info->GetSuspendCheck(), successor); - return; - } - if (block->IsEntryBlock() && (previous != nullptr) && previous->IsSuspendCheck()) { - GenerateSuspendCheck(previous->AsSuspendCheck(), nullptr); - } - if (!codegen_->GoesToNextBlock(block, successor)) { - __ B(codegen_->GetLabelOf(successor)); - } -} - -void LocationsBuilderARMVIXL::VisitGoto(HGoto* got) { - got->SetLocations(nullptr); -} - -void InstructionCodeGeneratorARMVIXL::VisitGoto(HGoto* got) { - HandleGoto(got, got->GetSuccessor()); -} - -void LocationsBuilderARMVIXL::VisitExit(HExit* exit) { - exit->SetLocations(nullptr); -} - -void InstructionCodeGeneratorARMVIXL::VisitExit(HExit* exit ATTRIBUTE_UNUSED) { -} - -void InstructionCodeGeneratorARMVIXL::GenerateVcmp(HInstruction* instruction) { - Primitive::Type type = instruction->InputAt(0)->GetType(); - Location lhs_loc = instruction->GetLocations()->InAt(0); - Location rhs_loc = instruction->GetLocations()->InAt(1); - if (rhs_loc.IsConstant()) { - // 0.0 is the only immediate that can be encoded directly in - // a VCMP instruction. - // - // Both the JLS (section 15.20.1) and the JVMS (section 6.5) - // specify that in a floating-point comparison, positive zero - // and negative zero are considered equal, so we can use the - // literal 0.0 for both cases here. - // - // Note however that some methods (Float.equal, Float.compare, - // Float.compareTo, Double.equal, Double.compare, - // Double.compareTo, Math.max, Math.min, StrictMath.max, - // StrictMath.min) consider 0.0 to be (strictly) greater than - // -0.0. So if we ever translate calls to these methods into a - // HCompare instruction, we must handle the -0.0 case with - // care here. - DCHECK(rhs_loc.GetConstant()->IsArithmeticZero()); - if (type == Primitive::kPrimFloat) { - __ Vcmp(F32, InputSRegisterAt(instruction, 0), 0.0); - } else { - DCHECK_EQ(type, Primitive::kPrimDouble); - __ Vcmp(F64, FromLowSToD(lhs_loc.AsFpuRegisterPairLow<vixl32::SRegister>()), 0.0); - } - } else { - if (type == Primitive::kPrimFloat) { - __ Vcmp(F32, InputSRegisterAt(instruction, 0), InputSRegisterAt(instruction, 1)); - } else { - DCHECK_EQ(type, Primitive::kPrimDouble); - __ Vcmp(F64, - FromLowSToD(lhs_loc.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(rhs_loc.AsFpuRegisterPairLow<vixl32::SRegister>())); - } - } -} - -void InstructionCodeGeneratorARMVIXL::GenerateFPJumps(HCondition* cond, - vixl32::Label* true_label, - vixl32::Label* false_label ATTRIBUTE_UNUSED) { - // To branch on the result of the FP compare we transfer FPSCR to APSR (encoded as PC in VMRS). - __ Vmrs(RegisterOrAPSR_nzcv(kPcCode), FPSCR); - __ B(ARMFPCondition(cond->GetCondition(), cond->IsGtBias()), true_label); -} - -void InstructionCodeGeneratorARMVIXL::GenerateLongComparesAndJumps(HCondition* cond, - vixl32::Label* true_label, - vixl32::Label* false_label) { - LocationSummary* locations = cond->GetLocations(); - Location left = locations->InAt(0); - Location right = locations->InAt(1); - IfCondition if_cond = cond->GetCondition(); - - vixl32::Register left_high = left.AsRegisterPairHigh<vixl32::Register>(); - vixl32::Register left_low = left.AsRegisterPairLow<vixl32::Register>(); - IfCondition true_high_cond = if_cond; - IfCondition false_high_cond = cond->GetOppositeCondition(); - vixl32::Condition final_condition = ARMUnsignedCondition(if_cond); // unsigned on lower part - - // Set the conditions for the test, remembering that == needs to be - // decided using the low words. - // TODO: consider avoiding jumps with temporary and CMP low+SBC high - switch (if_cond) { - case kCondEQ: - case kCondNE: - // Nothing to do. - break; - case kCondLT: - false_high_cond = kCondGT; - break; - case kCondLE: - true_high_cond = kCondLT; - break; - case kCondGT: - false_high_cond = kCondLT; - break; - case kCondGE: - true_high_cond = kCondGT; - break; - case kCondB: - false_high_cond = kCondA; - break; - case kCondBE: - true_high_cond = kCondB; - break; - case kCondA: - false_high_cond = kCondB; - break; - case kCondAE: - true_high_cond = kCondA; - break; - } - if (right.IsConstant()) { - int64_t value = right.GetConstant()->AsLongConstant()->GetValue(); - int32_t val_low = Low32Bits(value); - int32_t val_high = High32Bits(value); - - __ Cmp(left_high, val_high); - if (if_cond == kCondNE) { - __ B(ARMCondition(true_high_cond), true_label); - } else if (if_cond == kCondEQ) { - __ B(ARMCondition(false_high_cond), false_label); - } else { - __ B(ARMCondition(true_high_cond), true_label); - __ B(ARMCondition(false_high_cond), false_label); - } - // Must be equal high, so compare the lows. - __ Cmp(left_low, val_low); - } else { - vixl32::Register right_high = right.AsRegisterPairHigh<vixl32::Register>(); - vixl32::Register right_low = right.AsRegisterPairLow<vixl32::Register>(); - - __ Cmp(left_high, right_high); - if (if_cond == kCondNE) { - __ B(ARMCondition(true_high_cond), true_label); - } else if (if_cond == kCondEQ) { - __ B(ARMCondition(false_high_cond), false_label); - } else { - __ B(ARMCondition(true_high_cond), true_label); - __ B(ARMCondition(false_high_cond), false_label); - } - // Must be equal high, so compare the lows. - __ Cmp(left_low, right_low); - } - // The last comparison might be unsigned. - // TODO: optimize cases where this is always true/false - __ B(final_condition, true_label); -} - -void InstructionCodeGeneratorARMVIXL::GenerateCompareTestAndBranch(HCondition* condition, - vixl32::Label* true_target_in, - vixl32::Label* false_target_in) { - // Generated branching requires both targets to be explicit. If either of the - // targets is nullptr (fallthrough) use and bind `fallthrough` instead. - vixl32::Label fallthrough; - vixl32::Label* true_target = (true_target_in == nullptr) ? &fallthrough : true_target_in; - vixl32::Label* false_target = (false_target_in == nullptr) ? &fallthrough : false_target_in; - - Primitive::Type type = condition->InputAt(0)->GetType(); - switch (type) { - case Primitive::kPrimLong: - GenerateLongComparesAndJumps(condition, true_target, false_target); - break; - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: - GenerateVcmp(condition); - GenerateFPJumps(condition, true_target, false_target); - break; - default: - LOG(FATAL) << "Unexpected compare type " << type; - } - - if (false_target != &fallthrough) { - __ B(false_target); - } - - if (true_target_in == nullptr || false_target_in == nullptr) { - __ Bind(&fallthrough); - } -} - -void InstructionCodeGeneratorARMVIXL::GenerateTestAndBranch(HInstruction* instruction, - size_t condition_input_index, - vixl32::Label* true_target, - vixl32::Label* false_target) { - HInstruction* cond = instruction->InputAt(condition_input_index); - - if (true_target == nullptr && false_target == nullptr) { - // Nothing to do. The code always falls through. - return; - } else if (cond->IsIntConstant()) { - // Constant condition, statically compared against "true" (integer value 1). - if (cond->AsIntConstant()->IsTrue()) { - if (true_target != nullptr) { - __ B(true_target); - } - } else { - DCHECK(cond->AsIntConstant()->IsFalse()) << cond->AsIntConstant()->GetValue(); - if (false_target != nullptr) { - __ B(false_target); - } - } - return; - } - - // The following code generates these patterns: - // (1) true_target == nullptr && false_target != nullptr - // - opposite condition true => branch to false_target - // (2) true_target != nullptr && false_target == nullptr - // - condition true => branch to true_target - // (3) true_target != nullptr && false_target != nullptr - // - condition true => branch to true_target - // - branch to false_target - if (IsBooleanValueOrMaterializedCondition(cond)) { - // Condition has been materialized, compare the output to 0. - if (kIsDebugBuild) { - Location cond_val = instruction->GetLocations()->InAt(condition_input_index); - DCHECK(cond_val.IsRegister()); - } - if (true_target == nullptr) { - __ Cbz(InputRegisterAt(instruction, condition_input_index), false_target); - } else { - __ Cbnz(InputRegisterAt(instruction, condition_input_index), true_target); - } - } else { - // Condition has not been materialized. Use its inputs as the comparison and - // its condition as the branch condition. - HCondition* condition = cond->AsCondition(); - - // If this is a long or FP comparison that has been folded into - // the HCondition, generate the comparison directly. - Primitive::Type type = condition->InputAt(0)->GetType(); - if (type == Primitive::kPrimLong || Primitive::IsFloatingPointType(type)) { - GenerateCompareTestAndBranch(condition, true_target, false_target); - return; - } - - LocationSummary* locations = cond->GetLocations(); - DCHECK(locations->InAt(0).IsRegister()); - vixl32::Register left = InputRegisterAt(cond, 0); - Location right = locations->InAt(1); - if (right.IsRegister()) { - __ Cmp(left, InputRegisterAt(cond, 1)); - } else { - DCHECK(right.IsConstant()); - __ Cmp(left, CodeGenerator::GetInt32ValueOf(right.GetConstant())); - } - if (true_target == nullptr) { - __ B(ARMCondition(condition->GetOppositeCondition()), false_target); - } else { - __ B(ARMCondition(condition->GetCondition()), true_target); - } - } - - // If neither branch falls through (case 3), the conditional branch to `true_target` - // was already emitted (case 2) and we need to emit a jump to `false_target`. - if (true_target != nullptr && false_target != nullptr) { - __ B(false_target); - } -} - -void LocationsBuilderARMVIXL::VisitIf(HIf* if_instr) { - LocationSummary* locations = new (GetGraph()->GetArena()) LocationSummary(if_instr); - if (IsBooleanValueOrMaterializedCondition(if_instr->InputAt(0))) { - locations->SetInAt(0, Location::RequiresRegister()); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitIf(HIf* if_instr) { - HBasicBlock* true_successor = if_instr->IfTrueSuccessor(); - HBasicBlock* false_successor = if_instr->IfFalseSuccessor(); - vixl32::Label* true_target = - codegen_->GoesToNextBlock(if_instr->GetBlock(), true_successor) ? - nullptr : codegen_->GetLabelOf(true_successor); - vixl32::Label* false_target = - codegen_->GoesToNextBlock(if_instr->GetBlock(), false_successor) ? - nullptr : codegen_->GetLabelOf(false_successor); - GenerateTestAndBranch(if_instr, /* condition_input_index */ 0, true_target, false_target); -} - -void CodeGeneratorARMVIXL::GenerateNop() { - __ Nop(); -} - -void LocationsBuilderARMVIXL::HandleCondition(HCondition* cond) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(cond, LocationSummary::kNoCall); - // Handle the long/FP comparisons made in instruction simplification. - switch (cond->InputAt(0)->GetType()) { - case Primitive::kPrimLong: - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RegisterOrConstant(cond->InputAt(1))); - if (!cond->IsEmittedAtUseSite()) { - locations->SetOut(Location::RequiresRegister(), Location::kOutputOverlap); - } - break; - - // TODO: https://android-review.googlesource.com/#/c/252265/ - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetInAt(1, Location::RequiresFpuRegister()); - if (!cond->IsEmittedAtUseSite()) { - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - } - break; - - default: - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RegisterOrConstant(cond->InputAt(1))); - if (!cond->IsEmittedAtUseSite()) { - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - } - } -} - -void InstructionCodeGeneratorARMVIXL::HandleCondition(HCondition* cond) { - if (cond->IsEmittedAtUseSite()) { - return; - } - - LocationSummary* locations = cond->GetLocations(); - Location right = locations->InAt(1); - vixl32::Register out = OutputRegister(cond); - vixl32::Label true_label, false_label; - - switch (cond->InputAt(0)->GetType()) { - default: { - // Integer case. - if (right.IsRegister()) { - __ Cmp(InputRegisterAt(cond, 0), InputRegisterAt(cond, 1)); - } else { - DCHECK(right.IsConstant()); - __ Cmp(InputRegisterAt(cond, 0), CodeGenerator::GetInt32ValueOf(right.GetConstant())); - } - { - AssemblerAccurateScope aas(GetVIXLAssembler(), - kArmInstrMaxSizeInBytes * 3u, - CodeBufferCheckScope::kMaximumSize); - __ ite(ARMCondition(cond->GetCondition())); - __ mov(ARMCondition(cond->GetCondition()), OutputRegister(cond), 1); - __ mov(ARMCondition(cond->GetOppositeCondition()), OutputRegister(cond), 0); - } - return; - } - case Primitive::kPrimLong: - GenerateLongComparesAndJumps(cond, &true_label, &false_label); - break; - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: - GenerateVcmp(cond); - GenerateFPJumps(cond, &true_label, &false_label); - break; - } - - // Convert the jumps into the result. - vixl32::Label done_label; - - // False case: result = 0. - __ Bind(&false_label); - __ Mov(out, 0); - __ B(&done_label); - - // True case: result = 1. - __ Bind(&true_label); - __ Mov(out, 1); - __ Bind(&done_label); -} - -void LocationsBuilderARMVIXL::VisitEqual(HEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitEqual(HEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitNotEqual(HNotEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitNotEqual(HNotEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitLessThan(HLessThan* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitLessThan(HLessThan* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitLessThanOrEqual(HLessThanOrEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitLessThanOrEqual(HLessThanOrEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitGreaterThan(HGreaterThan* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitGreaterThan(HGreaterThan* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitGreaterThanOrEqual(HGreaterThanOrEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitGreaterThanOrEqual(HGreaterThanOrEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitBelow(HBelow* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitBelow(HBelow* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitBelowOrEqual(HBelowOrEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitBelowOrEqual(HBelowOrEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitAbove(HAbove* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitAbove(HAbove* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitAboveOrEqual(HAboveOrEqual* comp) { - HandleCondition(comp); -} - -void InstructionCodeGeneratorARMVIXL::VisitAboveOrEqual(HAboveOrEqual* comp) { - HandleCondition(comp); -} - -void LocationsBuilderARMVIXL::VisitIntConstant(HIntConstant* constant) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(constant, LocationSummary::kNoCall); - locations->SetOut(Location::ConstantLocation(constant)); -} - -void InstructionCodeGeneratorARMVIXL::VisitIntConstant(HIntConstant* constant ATTRIBUTE_UNUSED) { - // Will be generated at use site. -} - -void LocationsBuilderARMVIXL::VisitLongConstant(HLongConstant* constant) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(constant, LocationSummary::kNoCall); - locations->SetOut(Location::ConstantLocation(constant)); -} - -void InstructionCodeGeneratorARMVIXL::VisitLongConstant(HLongConstant* constant ATTRIBUTE_UNUSED) { - // Will be generated at use site. -} - -void LocationsBuilderARMVIXL::VisitMemoryBarrier(HMemoryBarrier* memory_barrier) { - memory_barrier->SetLocations(nullptr); -} - -void InstructionCodeGeneratorARMVIXL::VisitMemoryBarrier(HMemoryBarrier* memory_barrier) { - codegen_->GenerateMemoryBarrier(memory_barrier->GetBarrierKind()); -} - -void LocationsBuilderARMVIXL::VisitReturnVoid(HReturnVoid* ret) { - ret->SetLocations(nullptr); -} - -void InstructionCodeGeneratorARMVIXL::VisitReturnVoid(HReturnVoid* ret ATTRIBUTE_UNUSED) { - codegen_->GenerateFrameExit(); -} - -void LocationsBuilderARMVIXL::VisitReturn(HReturn* ret) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(ret, LocationSummary::kNoCall); - locations->SetInAt(0, parameter_visitor_.GetReturnLocation(ret->InputAt(0)->GetType())); -} - -void InstructionCodeGeneratorARMVIXL::VisitReturn(HReturn* ret ATTRIBUTE_UNUSED) { - codegen_->GenerateFrameExit(); -} - -void LocationsBuilderARMVIXL::VisitTypeConversion(HTypeConversion* conversion) { - Primitive::Type result_type = conversion->GetResultType(); - Primitive::Type input_type = conversion->GetInputType(); - DCHECK_NE(result_type, input_type); - - // The float-to-long, double-to-long and long-to-float type conversions - // rely on a call to the runtime. - LocationSummary::CallKind call_kind = - (((input_type == Primitive::kPrimFloat || input_type == Primitive::kPrimDouble) - && result_type == Primitive::kPrimLong) - || (input_type == Primitive::kPrimLong && result_type == Primitive::kPrimFloat)) - ? LocationSummary::kCallOnMainOnly - : LocationSummary::kNoCall; - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(conversion, call_kind); - - // The Java language does not allow treating boolean as an integral type but - // our bit representation makes it safe. - - switch (result_type) { - case Primitive::kPrimByte: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to byte is a result of code transformations. - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-byte' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimShort: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to short is a result of code transformations. - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-short' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimInt: - switch (input_type) { - case Primitive::kPrimLong: - // Processing a Dex `long-to-int' instruction. - locations->SetInAt(0, Location::Any()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - - case Primitive::kPrimFloat: - // Processing a Dex `float-to-int' instruction. - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresRegister()); - locations->AddTemp(Location::RequiresFpuRegister()); - break; - - case Primitive::kPrimDouble: - // Processing a Dex `double-to-int' instruction. - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresRegister()); - locations->AddTemp(Location::RequiresFpuRegister()); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimLong: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-long' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - - case Primitive::kPrimFloat: { - // Processing a Dex `float-to-long' instruction. - InvokeRuntimeCallingConvention calling_convention; - locations->SetInAt(0, Location::FpuRegisterLocation( - calling_convention.GetFpuRegisterAt(0))); - locations->SetOut(Location::RegisterPairLocation(R0, R1)); - break; - } - - case Primitive::kPrimDouble: { - // Processing a Dex `double-to-long' instruction. - InvokeRuntimeCallingConvention calling_convention; - locations->SetInAt(0, Location::FpuRegisterPairLocation( - calling_convention.GetFpuRegisterAt(0), - calling_convention.GetFpuRegisterAt(1))); - locations->SetOut(Location::RegisterPairLocation(R0, R1)); - break; - } - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimChar: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to char is a result of code transformations. - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - // Processing a Dex `int-to-char' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimFloat: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-float' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresFpuRegister()); - break; - - case Primitive::kPrimLong: { - // Processing a Dex `long-to-float' instruction. - InvokeRuntimeCallingConvention calling_convention; - locations->SetInAt(0, Location::RegisterPairLocation( - calling_convention.GetRegisterAt(0), calling_convention.GetRegisterAt(1))); - locations->SetOut(Location::FpuRegisterLocation(calling_convention.GetFpuRegisterAt(0))); - break; - } - - case Primitive::kPrimDouble: - // Processing a Dex `double-to-float' instruction. - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - }; - break; - - case Primitive::kPrimDouble: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-double' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresFpuRegister()); - break; - - case Primitive::kPrimLong: - // Processing a Dex `long-to-double' instruction. - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresFpuRegister()); - locations->AddTemp(Location::RequiresFpuRegister()); - locations->AddTemp(Location::RequiresFpuRegister()); - break; - - case Primitive::kPrimFloat: - // Processing a Dex `float-to-double' instruction. - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - }; - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } -} - -void InstructionCodeGeneratorARMVIXL::VisitTypeConversion(HTypeConversion* conversion) { - LocationSummary* locations = conversion->GetLocations(); - Location out = locations->Out(); - Location in = locations->InAt(0); - Primitive::Type result_type = conversion->GetResultType(); - Primitive::Type input_type = conversion->GetInputType(); - DCHECK_NE(result_type, input_type); - switch (result_type) { - case Primitive::kPrimByte: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to byte is a result of code transformations. - __ Sbfx(OutputRegister(conversion), in.AsRegisterPairLow<vixl32::Register>(), 0, 8); - break; - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-byte' instruction. - __ Sbfx(OutputRegister(conversion), InputRegisterAt(conversion, 0), 0, 8); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimShort: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to short is a result of code transformations. - __ Sbfx(OutputRegister(conversion), in.AsRegisterPairLow<vixl32::Register>(), 0, 16); - break; - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-short' instruction. - __ Sbfx(OutputRegister(conversion), InputRegisterAt(conversion, 0), 0, 16); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimInt: - switch (input_type) { - case Primitive::kPrimLong: - // Processing a Dex `long-to-int' instruction. - DCHECK(out.IsRegister()); - if (in.IsRegisterPair()) { - __ Mov(OutputRegister(conversion), in.AsRegisterPairLow<vixl32::Register>()); - } else if (in.IsDoubleStackSlot()) { - GetAssembler()->LoadFromOffset(kLoadWord, - OutputRegister(conversion), - sp, - in.GetStackIndex()); - } else { - DCHECK(in.IsConstant()); - DCHECK(in.GetConstant()->IsLongConstant()); - int64_t value = in.GetConstant()->AsLongConstant()->GetValue(); - __ Mov(OutputRegister(conversion), static_cast<int32_t>(value)); - } - break; - - case Primitive::kPrimFloat: { - // Processing a Dex `float-to-int' instruction. - vixl32::SRegister temp = locations->GetTemp(0).AsFpuRegisterPairLow<vixl32::SRegister>(); - __ Vcvt(I32, F32, temp, InputSRegisterAt(conversion, 0)); - __ Vmov(OutputRegister(conversion), temp); - break; - } - - case Primitive::kPrimDouble: { - // Processing a Dex `double-to-int' instruction. - vixl32::SRegister temp_s = - locations->GetTemp(0).AsFpuRegisterPairLow<vixl32::SRegister>(); - __ Vcvt(I32, F64, temp_s, FromLowSToD(in.AsFpuRegisterPairLow<vixl32::SRegister>())); - __ Vmov(OutputRegister(conversion), temp_s); - break; - } - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimLong: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: - // Processing a Dex `int-to-long' instruction. - DCHECK(out.IsRegisterPair()); - DCHECK(in.IsRegister()); - __ Mov(out.AsRegisterPairLow<vixl32::Register>(), InputRegisterAt(conversion, 0)); - // Sign extension. - __ Asr(out.AsRegisterPairHigh<vixl32::Register>(), - out.AsRegisterPairLow<vixl32::Register>(), - 31); - break; - - case Primitive::kPrimFloat: - // Processing a Dex `float-to-long' instruction. - codegen_->InvokeRuntime(kQuickF2l, conversion, conversion->GetDexPc()); - CheckEntrypointTypes<kQuickF2l, int64_t, float>(); - break; - - case Primitive::kPrimDouble: - // Processing a Dex `double-to-long' instruction. - codegen_->InvokeRuntime(kQuickD2l, conversion, conversion->GetDexPc()); - CheckEntrypointTypes<kQuickD2l, int64_t, double>(); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimChar: - switch (input_type) { - case Primitive::kPrimLong: - // Type conversion from long to char is a result of code transformations. - __ Ubfx(OutputRegister(conversion), in.AsRegisterPairLow<vixl32::Register>(), 0, 16); - break; - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - // Processing a Dex `int-to-char' instruction. - __ Ubfx(OutputRegister(conversion), InputRegisterAt(conversion, 0), 0, 16); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } - break; - - case Primitive::kPrimFloat: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: { - // Processing a Dex `int-to-float' instruction. - __ Vmov(OutputSRegister(conversion), InputRegisterAt(conversion, 0)); - __ Vcvt(F32, I32, OutputSRegister(conversion), OutputSRegister(conversion)); - break; - } - - case Primitive::kPrimLong: - // Processing a Dex `long-to-float' instruction. - codegen_->InvokeRuntime(kQuickL2f, conversion, conversion->GetDexPc()); - CheckEntrypointTypes<kQuickL2f, float, int64_t>(); - break; - - case Primitive::kPrimDouble: - // Processing a Dex `double-to-float' instruction. - __ Vcvt(F32, - F64, - OutputSRegister(conversion), - FromLowSToD(in.AsFpuRegisterPairLow<vixl32::SRegister>())); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - }; - break; - - case Primitive::kPrimDouble: - switch (input_type) { - case Primitive::kPrimBoolean: - // Boolean input is a result of code transformations. - case Primitive::kPrimByte: - case Primitive::kPrimShort: - case Primitive::kPrimInt: - case Primitive::kPrimChar: { - // Processing a Dex `int-to-double' instruction. - __ Vmov(out.AsFpuRegisterPairLow<vixl32::SRegister>(), InputRegisterAt(conversion, 0)); - __ Vcvt(F64, - I32, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - out.AsFpuRegisterPairLow<vixl32::SRegister>()); - break; - } - - case Primitive::kPrimLong: { - // Processing a Dex `long-to-double' instruction. - vixl32::Register low = in.AsRegisterPairLow<vixl32::Register>(); - vixl32::Register high = in.AsRegisterPairHigh<vixl32::Register>(); - - vixl32::SRegister out_s = out.AsFpuRegisterPairLow<vixl32::SRegister>(); - vixl32::DRegister out_d = FromLowSToD(out_s); - - vixl32::SRegister temp_s = - locations->GetTemp(0).AsFpuRegisterPairLow<vixl32::SRegister>(); - vixl32::DRegister temp_d = FromLowSToD(temp_s); - - vixl32::SRegister constant_s = - locations->GetTemp(1).AsFpuRegisterPairLow<vixl32::SRegister>(); - vixl32::DRegister constant_d = FromLowSToD(constant_s); - - // temp_d = int-to-double(high) - __ Vmov(temp_s, high); - __ Vcvt(F64, I32, temp_d, temp_s); - // constant_d = k2Pow32EncodingForDouble - __ Vmov(F64, - constant_d, - vixl32::DOperand(bit_cast<double, int64_t>(k2Pow32EncodingForDouble))); - // out_d = unsigned-to-double(low) - __ Vmov(out_s, low); - __ Vcvt(F64, U32, out_d, out_s); - // out_d += temp_d * constant_d - __ Vmla(F64, out_d, temp_d, constant_d); - break; - } - - case Primitive::kPrimFloat: - // Processing a Dex `float-to-double' instruction. - __ Vcvt(F64, - F32, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - InputSRegisterAt(conversion, 0)); - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - }; - break; - - default: - LOG(FATAL) << "Unexpected type conversion from " << input_type - << " to " << result_type; - } -} - -void LocationsBuilderARMVIXL::VisitAdd(HAdd* add) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(add, LocationSummary::kNoCall); - switch (add->GetResultType()) { - case Primitive::kPrimInt: { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RegisterOrConstant(add->InputAt(1))); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - } - - // TODO: https://android-review.googlesource.com/#/c/254144/ - case Primitive::kPrimLong: { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - } - - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: { - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetInAt(1, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - } - - default: - LOG(FATAL) << "Unexpected add type " << add->GetResultType(); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitAdd(HAdd* add) { - LocationSummary* locations = add->GetLocations(); - Location out = locations->Out(); - Location first = locations->InAt(0); - Location second = locations->InAt(1); - - switch (add->GetResultType()) { - case Primitive::kPrimInt: { - __ Add(OutputRegister(add), InputRegisterAt(add, 0), InputOperandAt(add, 1)); - } - break; - - // TODO: https://android-review.googlesource.com/#/c/254144/ - case Primitive::kPrimLong: { - DCHECK(second.IsRegisterPair()); - __ Adds(out.AsRegisterPairLow<vixl32::Register>(), - first.AsRegisterPairLow<vixl32::Register>(), - Operand(second.AsRegisterPairLow<vixl32::Register>())); - __ Adc(out.AsRegisterPairHigh<vixl32::Register>(), - first.AsRegisterPairHigh<vixl32::Register>(), - second.AsRegisterPairHigh<vixl32::Register>()); - break; - } - - case Primitive::kPrimFloat: { - __ Vadd(F32, OutputSRegister(add), InputSRegisterAt(add, 0), InputSRegisterAt(add, 1)); - } - break; - - case Primitive::kPrimDouble: - __ Vadd(F64, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(first.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(second.AsFpuRegisterPairLow<vixl32::SRegister>())); - break; - - default: - LOG(FATAL) << "Unexpected add type " << add->GetResultType(); - } -} - -void LocationsBuilderARMVIXL::VisitSub(HSub* sub) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(sub, LocationSummary::kNoCall); - switch (sub->GetResultType()) { - case Primitive::kPrimInt: { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RegisterOrConstant(sub->InputAt(1))); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - } - - // TODO: https://android-review.googlesource.com/#/c/254144/ - case Primitive::kPrimLong: { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - } - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: { - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetInAt(1, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - } - default: - LOG(FATAL) << "Unexpected sub type " << sub->GetResultType(); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitSub(HSub* sub) { - LocationSummary* locations = sub->GetLocations(); - Location out = locations->Out(); - Location first = locations->InAt(0); - Location second = locations->InAt(1); - switch (sub->GetResultType()) { - case Primitive::kPrimInt: { - if (second.IsRegister()) { - __ Sub(OutputRegister(sub), InputRegisterAt(sub, 0), InputRegisterAt(sub, 1)); - } else { - __ Sub(OutputRegister(sub), - InputRegisterAt(sub, 0), - second.GetConstant()->AsIntConstant()->GetValue()); - } - break; - } - - // TODO: https://android-review.googlesource.com/#/c/254144/ - case Primitive::kPrimLong: { - DCHECK(second.IsRegisterPair()); - __ Subs(out.AsRegisterPairLow<vixl32::Register>(), - first.AsRegisterPairLow<vixl32::Register>(), - Operand(second.AsRegisterPairLow<vixl32::Register>())); - __ Sbc(out.AsRegisterPairHigh<vixl32::Register>(), - first.AsRegisterPairHigh<vixl32::Register>(), - Operand(second.AsRegisterPairHigh<vixl32::Register>())); - break; - } - - case Primitive::kPrimFloat: { - __ Vsub(F32, OutputSRegister(sub), InputSRegisterAt(sub, 0), InputSRegisterAt(sub, 1)); - break; - } - - case Primitive::kPrimDouble: { - __ Vsub(F64, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(first.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(second.AsFpuRegisterPairLow<vixl32::SRegister>())); - break; - } - - default: - LOG(FATAL) << "Unexpected sub type " << sub->GetResultType(); - } -} - -void LocationsBuilderARMVIXL::VisitMul(HMul* mul) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(mul, LocationSummary::kNoCall); - switch (mul->GetResultType()) { - case Primitive::kPrimInt: - case Primitive::kPrimLong: { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - break; - } - - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: { - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetInAt(1, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - } - - default: - LOG(FATAL) << "Unexpected mul type " << mul->GetResultType(); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitMul(HMul* mul) { - LocationSummary* locations = mul->GetLocations(); - Location out = locations->Out(); - Location first = locations->InAt(0); - Location second = locations->InAt(1); - switch (mul->GetResultType()) { - case Primitive::kPrimInt: { - __ Mul(OutputRegister(mul), InputRegisterAt(mul, 0), InputRegisterAt(mul, 1)); - break; - } - case Primitive::kPrimLong: { - vixl32::Register out_hi = out.AsRegisterPairHigh<vixl32::Register>(); - vixl32::Register out_lo = out.AsRegisterPairLow<vixl32::Register>(); - vixl32::Register in1_hi = first.AsRegisterPairHigh<vixl32::Register>(); - vixl32::Register in1_lo = first.AsRegisterPairLow<vixl32::Register>(); - vixl32::Register in2_hi = second.AsRegisterPairHigh<vixl32::Register>(); - vixl32::Register in2_lo = second.AsRegisterPairLow<vixl32::Register>(); - - // Extra checks to protect caused by the existence of R1_R2. - // The algorithm is wrong if out.hi is either in1.lo or in2.lo: - // (e.g. in1=r0_r1, in2=r2_r3 and out=r1_r2); - DCHECK_NE(out_hi.GetCode(), in1_lo.GetCode()); - DCHECK_NE(out_hi.GetCode(), in2_lo.GetCode()); - - // input: in1 - 64 bits, in2 - 64 bits - // output: out - // formula: out.hi : out.lo = (in1.lo * in2.hi + in1.hi * in2.lo)* 2^32 + in1.lo * in2.lo - // parts: out.hi = in1.lo * in2.hi + in1.hi * in2.lo + (in1.lo * in2.lo)[63:32] - // parts: out.lo = (in1.lo * in2.lo)[31:0] - - UseScratchRegisterScope temps(GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - // temp <- in1.lo * in2.hi - __ Mul(temp, in1_lo, in2_hi); - // out.hi <- in1.lo * in2.hi + in1.hi * in2.lo - __ Mla(out_hi, in1_hi, in2_lo, temp); - // out.lo <- (in1.lo * in2.lo)[31:0]; - __ Umull(out_lo, temp, in1_lo, in2_lo); - // out.hi <- in2.hi * in1.lo + in2.lo * in1.hi + (in1.lo * in2.lo)[63:32] - __ Add(out_hi, out_hi, Operand(temp)); - break; - } - - case Primitive::kPrimFloat: { - __ Vmul(F32, OutputSRegister(mul), InputSRegisterAt(mul, 0), InputSRegisterAt(mul, 1)); - break; - } - - case Primitive::kPrimDouble: { - __ Vmul(F64, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(first.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(second.AsFpuRegisterPairLow<vixl32::SRegister>())); - break; - } - - default: - LOG(FATAL) << "Unexpected mul type " << mul->GetResultType(); - } -} - -void LocationsBuilderARMVIXL::VisitNot(HNot* not_) { - LocationSummary* locations = - new (GetGraph()->GetArena()) LocationSummary(not_, LocationSummary::kNoCall); - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); -} - -void InstructionCodeGeneratorARMVIXL::VisitNot(HNot* not_) { - LocationSummary* locations = not_->GetLocations(); - Location out = locations->Out(); - Location in = locations->InAt(0); - switch (not_->GetResultType()) { - case Primitive::kPrimInt: - __ Mvn(OutputRegister(not_), InputRegisterAt(not_, 0)); - break; - - case Primitive::kPrimLong: - __ Mvn(out.AsRegisterPairLow<vixl32::Register>(), - Operand(in.AsRegisterPairLow<vixl32::Register>())); - __ Mvn(out.AsRegisterPairHigh<vixl32::Register>(), - Operand(in.AsRegisterPairHigh<vixl32::Register>())); - break; - - default: - LOG(FATAL) << "Unimplemented type for not operation " << not_->GetResultType(); - } -} - -void CodeGeneratorARMVIXL::GenerateMemoryBarrier(MemBarrierKind kind) { - // TODO (ported from quick): revisit ARM barrier kinds. - DmbOptions flavor = DmbOptions::ISH; // Quiet C++ warnings. - switch (kind) { - case MemBarrierKind::kAnyStore: - case MemBarrierKind::kLoadAny: - case MemBarrierKind::kAnyAny: { - flavor = DmbOptions::ISH; - break; - } - case MemBarrierKind::kStoreStore: { - flavor = DmbOptions::ISHST; - break; - } - default: - LOG(FATAL) << "Unexpected memory barrier " << kind; - } - __ Dmb(flavor); -} - -void InstructionCodeGeneratorARMVIXL::DivRemOneOrMinusOne(HBinaryOperation* instruction) { - DCHECK(instruction->IsDiv() || instruction->IsRem()); - DCHECK(instruction->GetResultType() == Primitive::kPrimInt); - - LocationSummary* locations = instruction->GetLocations(); - Location second = locations->InAt(1); - DCHECK(second.IsConstant()); - - vixl32::Register out = OutputRegister(instruction); - vixl32::Register dividend = InputRegisterAt(instruction, 0); - int32_t imm = second.GetConstant()->AsIntConstant()->GetValue(); - DCHECK(imm == 1 || imm == -1); - - if (instruction->IsRem()) { - __ Mov(out, 0); - } else { - if (imm == 1) { - __ Mov(out, dividend); - } else { - __ Rsb(out, dividend, 0); - } - } -} - -void InstructionCodeGeneratorARMVIXL::DivRemByPowerOfTwo(HBinaryOperation* instruction) { - DCHECK(instruction->IsDiv() || instruction->IsRem()); - DCHECK(instruction->GetResultType() == Primitive::kPrimInt); - - LocationSummary* locations = instruction->GetLocations(); - Location second = locations->InAt(1); - DCHECK(second.IsConstant()); - - vixl32::Register out = OutputRegister(instruction); - vixl32::Register dividend = InputRegisterAt(instruction, 0); - vixl32::Register temp = locations->GetTemp(0).AsRegister<vixl32::Register>(); - int32_t imm = second.GetConstant()->AsIntConstant()->GetValue(); - uint32_t abs_imm = static_cast<uint32_t>(AbsOrMin(imm)); - int ctz_imm = CTZ(abs_imm); - - if (ctz_imm == 1) { - __ Lsr(temp, dividend, 32 - ctz_imm); - } else { - __ Asr(temp, dividend, 31); - __ Lsr(temp, temp, 32 - ctz_imm); - } - __ Add(out, temp, Operand(dividend)); - - if (instruction->IsDiv()) { - __ Asr(out, out, ctz_imm); - if (imm < 0) { - __ Rsb(out, out, Operand(0)); - } - } else { - __ Ubfx(out, out, 0, ctz_imm); - __ Sub(out, out, Operand(temp)); - } -} - -void InstructionCodeGeneratorARMVIXL::GenerateDivRemWithAnyConstant(HBinaryOperation* instruction) { - DCHECK(instruction->IsDiv() || instruction->IsRem()); - DCHECK(instruction->GetResultType() == Primitive::kPrimInt); - - LocationSummary* locations = instruction->GetLocations(); - Location second = locations->InAt(1); - DCHECK(second.IsConstant()); - - vixl32::Register out = OutputRegister(instruction); - vixl32::Register dividend = InputRegisterAt(instruction, 0); - vixl32::Register temp1 = locations->GetTemp(0).AsRegister<vixl32::Register>(); - vixl32::Register temp2 = locations->GetTemp(1).AsRegister<vixl32::Register>(); - int64_t imm = second.GetConstant()->AsIntConstant()->GetValue(); - - int64_t magic; - int shift; - CalculateMagicAndShiftForDivRem(imm, false /* is_long */, &magic, &shift); - - __ Mov(temp1, magic); - __ Smull(temp2, temp1, dividend, temp1); - - if (imm > 0 && magic < 0) { - __ Add(temp1, temp1, Operand(dividend)); - } else if (imm < 0 && magic > 0) { - __ Sub(temp1, temp1, Operand(dividend)); - } - - if (shift != 0) { - __ Asr(temp1, temp1, shift); - } - - if (instruction->IsDiv()) { - __ Sub(out, temp1, Operand(temp1, vixl32::Shift(ASR), 31)); - } else { - __ Sub(temp1, temp1, Operand(temp1, vixl32::Shift(ASR), 31)); - // TODO: Strength reduction for mls. - __ Mov(temp2, imm); - __ Mls(out, temp1, temp2, dividend); - } -} - -void InstructionCodeGeneratorARMVIXL::GenerateDivRemConstantIntegral( - HBinaryOperation* instruction) { - DCHECK(instruction->IsDiv() || instruction->IsRem()); - DCHECK(instruction->GetResultType() == Primitive::kPrimInt); - - LocationSummary* locations = instruction->GetLocations(); - Location second = locations->InAt(1); - DCHECK(second.IsConstant()); - - int32_t imm = second.GetConstant()->AsIntConstant()->GetValue(); - if (imm == 0) { - // Do not generate anything. DivZeroCheck would prevent any code to be executed. - } else if (imm == 1 || imm == -1) { - DivRemOneOrMinusOne(instruction); - } else if (IsPowerOfTwo(AbsOrMin(imm))) { - DivRemByPowerOfTwo(instruction); - } else { - DCHECK(imm <= -2 || imm >= 2); - GenerateDivRemWithAnyConstant(instruction); - } -} - -void LocationsBuilderARMVIXL::VisitDiv(HDiv* div) { - LocationSummary::CallKind call_kind = LocationSummary::kNoCall; - if (div->GetResultType() == Primitive::kPrimLong) { - // pLdiv runtime call. - call_kind = LocationSummary::kCallOnMainOnly; - } else if (div->GetResultType() == Primitive::kPrimInt && div->InputAt(1)->IsConstant()) { - // sdiv will be replaced by other instruction sequence. - } else if (div->GetResultType() == Primitive::kPrimInt && - !codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { - // pIdivmod runtime call. - call_kind = LocationSummary::kCallOnMainOnly; - } - - LocationSummary* locations = new (GetGraph()->GetArena()) LocationSummary(div, call_kind); - - switch (div->GetResultType()) { - case Primitive::kPrimInt: { - if (div->InputAt(1)->IsConstant()) { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::ConstantLocation(div->InputAt(1)->AsConstant())); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - int32_t value = div->InputAt(1)->AsIntConstant()->GetValue(); - if (value == 1 || value == 0 || value == -1) { - // No temp register required. - } else { - locations->AddTemp(Location::RequiresRegister()); - if (!IsPowerOfTwo(AbsOrMin(value))) { - locations->AddTemp(Location::RequiresRegister()); - } - } - } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { - locations->SetInAt(0, Location::RequiresRegister()); - locations->SetInAt(1, Location::RequiresRegister()); - locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); - } else { - TODO_VIXL32(FATAL); - } - break; - } - case Primitive::kPrimLong: { - TODO_VIXL32(FATAL); - break; - } - case Primitive::kPrimFloat: - case Primitive::kPrimDouble: { - locations->SetInAt(0, Location::RequiresFpuRegister()); - locations->SetInAt(1, Location::RequiresFpuRegister()); - locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); - break; - } - - default: - LOG(FATAL) << "Unexpected div type " << div->GetResultType(); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitDiv(HDiv* div) { - LocationSummary* locations = div->GetLocations(); - Location out = locations->Out(); - Location first = locations->InAt(0); - Location second = locations->InAt(1); - - switch (div->GetResultType()) { - case Primitive::kPrimInt: { - if (second.IsConstant()) { - GenerateDivRemConstantIntegral(div); - } else if (codegen_->GetInstructionSetFeatures().HasDivideInstruction()) { - __ Sdiv(OutputRegister(div), InputRegisterAt(div, 0), InputRegisterAt(div, 1)); - } else { - TODO_VIXL32(FATAL); - } - break; - } - - case Primitive::kPrimLong: { - TODO_VIXL32(FATAL); - break; - } - - case Primitive::kPrimFloat: { - __ Vdiv(F32, OutputSRegister(div), InputSRegisterAt(div, 0), InputSRegisterAt(div, 1)); - break; - } - - case Primitive::kPrimDouble: { - __ Vdiv(F64, - FromLowSToD(out.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(first.AsFpuRegisterPairLow<vixl32::SRegister>()), - FromLowSToD(second.AsFpuRegisterPairLow<vixl32::SRegister>())); - break; - } - - default: - LOG(FATAL) << "Unexpected div type " << div->GetResultType(); - } -} - -void LocationsBuilderARMVIXL::VisitDivZeroCheck(HDivZeroCheck* instruction) { - LocationSummary::CallKind call_kind = instruction->CanThrowIntoCatchBlock() - ? LocationSummary::kCallOnSlowPath - : LocationSummary::kNoCall; - LocationSummary* locations = new (GetGraph()->GetArena()) LocationSummary(instruction, call_kind); - locations->SetInAt(0, Location::RegisterOrConstant(instruction->InputAt(0))); - if (instruction->HasUses()) { - locations->SetOut(Location::SameAsFirstInput()); - } -} - -void InstructionCodeGeneratorARMVIXL::VisitDivZeroCheck(HDivZeroCheck* instruction) { - DivZeroCheckSlowPathARMVIXL* slow_path = - new (GetGraph()->GetArena()) DivZeroCheckSlowPathARMVIXL(instruction); - codegen_->AddSlowPath(slow_path); - - LocationSummary* locations = instruction->GetLocations(); - Location value = locations->InAt(0); - - switch (instruction->GetType()) { - case Primitive::kPrimBoolean: - case Primitive::kPrimByte: - case Primitive::kPrimChar: - case Primitive::kPrimShort: - case Primitive::kPrimInt: { - if (value.IsRegister()) { - __ Cbz(InputRegisterAt(instruction, 0), slow_path->GetEntryLabel()); - } else { - DCHECK(value.IsConstant()) << value; - if (value.GetConstant()->AsIntConstant()->GetValue() == 0) { - __ B(slow_path->GetEntryLabel()); - } - } - break; - } - case Primitive::kPrimLong: { - if (value.IsRegisterPair()) { - UseScratchRegisterScope temps(GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - __ Orrs(temp, - value.AsRegisterPairLow<vixl32::Register>(), - Operand(value.AsRegisterPairHigh<vixl32::Register>())); - __ B(eq, slow_path->GetEntryLabel()); - } else { - DCHECK(value.IsConstant()) << value; - if (value.GetConstant()->AsLongConstant()->GetValue() == 0) { - __ B(slow_path->GetEntryLabel()); - } - } - break; - } - default: - LOG(FATAL) << "Unexpected type for HDivZeroCheck " << instruction->GetType(); - } -} - -void LocationsBuilderARMVIXL::VisitParallelMove(HParallelMove* instruction ATTRIBUTE_UNUSED) { - LOG(FATAL) << "Unreachable"; -} - -void InstructionCodeGeneratorARMVIXL::VisitParallelMove(HParallelMove* instruction) { - codegen_->GetMoveResolver()->EmitNativeCode(instruction); -} - -ArmVIXLAssembler* ParallelMoveResolverARMVIXL::GetAssembler() const { - return codegen_->GetAssembler(); -} - -void ParallelMoveResolverARMVIXL::EmitMove(size_t index) { - MoveOperands* move = moves_[index]; - Location source = move->GetSource(); - Location destination = move->GetDestination(); - - if (source.IsRegister()) { - if (destination.IsRegister()) { - __ Mov(destination.AsRegister<vixl32::Register>(), source.AsRegister<vixl32::Register>()); - } else if (destination.IsFpuRegister()) { - __ Vmov(destination.AsFpuRegister<vixl32::SRegister>(), - source.AsRegister<vixl32::Register>()); - } else { - DCHECK(destination.IsStackSlot()); - GetAssembler()->StoreToOffset(kStoreWord, - source.AsRegister<vixl32::Register>(), - sp, - destination.GetStackIndex()); - } - } else if (source.IsStackSlot()) { - TODO_VIXL32(FATAL); - } else if (source.IsFpuRegister()) { - TODO_VIXL32(FATAL); - } else if (source.IsDoubleStackSlot()) { - TODO_VIXL32(FATAL); - } else if (source.IsRegisterPair()) { - if (destination.IsRegisterPair()) { - __ Mov(destination.AsRegisterPairLow<vixl32::Register>(), - source.AsRegisterPairLow<vixl32::Register>()); - __ Mov(destination.AsRegisterPairHigh<vixl32::Register>(), - source.AsRegisterPairHigh<vixl32::Register>()); - } else if (destination.IsFpuRegisterPair()) { - __ Vmov(FromLowSToD(destination.AsFpuRegisterPairLow<vixl32::SRegister>()), - source.AsRegisterPairLow<vixl32::Register>(), - source.AsRegisterPairHigh<vixl32::Register>()); - } else { - DCHECK(destination.IsDoubleStackSlot()) << destination; - DCHECK(ExpectedPairLayout(source)); - GetAssembler()->StoreToOffset(kStoreWordPair, - source.AsRegisterPairLow<vixl32::Register>(), - sp, - destination.GetStackIndex()); - } - } else if (source.IsFpuRegisterPair()) { - TODO_VIXL32(FATAL); - } else { - DCHECK(source.IsConstant()) << source; - HConstant* constant = source.GetConstant(); - if (constant->IsIntConstant() || constant->IsNullConstant()) { - int32_t value = CodeGenerator::GetInt32ValueOf(constant); - if (destination.IsRegister()) { - __ Mov(destination.AsRegister<vixl32::Register>(), value); - } else { - DCHECK(destination.IsStackSlot()); - UseScratchRegisterScope temps(GetAssembler()->GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - __ Mov(temp, value); - GetAssembler()->StoreToOffset(kStoreWord, temp, sp, destination.GetStackIndex()); - } - } else if (constant->IsLongConstant()) { - int64_t value = constant->AsLongConstant()->GetValue(); - if (destination.IsRegisterPair()) { - __ Mov(destination.AsRegisterPairLow<vixl32::Register>(), Low32Bits(value)); - __ Mov(destination.AsRegisterPairHigh<vixl32::Register>(), High32Bits(value)); - } else { - DCHECK(destination.IsDoubleStackSlot()) << destination; - UseScratchRegisterScope temps(GetAssembler()->GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - __ Mov(temp, Low32Bits(value)); - GetAssembler()->StoreToOffset(kStoreWord, temp, sp, destination.GetStackIndex()); - __ Mov(temp, High32Bits(value)); - GetAssembler()->StoreToOffset(kStoreWord, - temp, - sp, - destination.GetHighStackIndex(kArmWordSize)); - } - } else if (constant->IsDoubleConstant()) { - double value = constant->AsDoubleConstant()->GetValue(); - if (destination.IsFpuRegisterPair()) { - __ Vmov(F64, FromLowSToD(destination.AsFpuRegisterPairLow<vixl32::SRegister>()), value); - } else { - DCHECK(destination.IsDoubleStackSlot()) << destination; - uint64_t int_value = bit_cast<uint64_t, double>(value); - UseScratchRegisterScope temps(GetAssembler()->GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - GetAssembler()->LoadImmediate(temp, Low32Bits(int_value)); - GetAssembler()->StoreToOffset(kStoreWord, temp, sp, destination.GetStackIndex()); - GetAssembler()->LoadImmediate(temp, High32Bits(int_value)); - GetAssembler()->StoreToOffset(kStoreWord, - temp, - sp, - destination.GetHighStackIndex(kArmWordSize)); - } - } else { - DCHECK(constant->IsFloatConstant()) << constant->DebugName(); - float value = constant->AsFloatConstant()->GetValue(); - if (destination.IsFpuRegister()) { - __ Vmov(F32, destination.AsFpuRegister<vixl32::SRegister>(), value); - } else { - DCHECK(destination.IsStackSlot()); - UseScratchRegisterScope temps(GetAssembler()->GetVIXLAssembler()); - vixl32::Register temp = temps.Acquire(); - GetAssembler()->LoadImmediate(temp, bit_cast<int32_t, float>(value)); - GetAssembler()->StoreToOffset(kStoreWord, temp, sp, destination.GetStackIndex()); - } - } - } -} - -void ParallelMoveResolverARMVIXL::Exchange(Register reg, int mem) { - TODO_VIXL32(FATAL); -} - -void ParallelMoveResolverARMVIXL::Exchange(int mem1, int mem2) { - TODO_VIXL32(FATAL); -} - -void ParallelMoveResolverARMVIXL::EmitSwap(size_t index) { - TODO_VIXL32(FATAL); -} - -void ParallelMoveResolverARMVIXL::SpillScratch(int reg ATTRIBUTE_UNUSED) { - TODO_VIXL32(FATAL); -} - -void ParallelMoveResolverARMVIXL::RestoreScratch(int reg ATTRIBUTE_UNUSED) { - TODO_VIXL32(FATAL); -} - - -// TODO: Remove when codegen complete. -#pragma GCC diagnostic pop - -#undef __ -#undef QUICK_ENTRY_POINT -#undef TODO_VIXL32 - -} // namespace arm -} // namespace art diff --git a/compiler/optimizing/code_generator_arm_vixl.h b/compiler/optimizing/code_generator_arm_vixl.h deleted file mode 100644 index d0c2c85b6a..0000000000 --- a/compiler/optimizing/code_generator_arm_vixl.h +++ /dev/null @@ -1,405 +0,0 @@ -/* - * Copyright (C) 2016 The Android Open Source Project - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -#ifndef ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_VIXL_H_ -#define ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_VIXL_H_ - -#include "code_generator_arm.h" -#include "utils/arm/assembler_arm_vixl.h" - -// TODO(VIXL): make vixl clean wrt -Wshadow. -#pragma GCC diagnostic push -#pragma GCC diagnostic ignored "-Wshadow" -#include "aarch32/constants-aarch32.h" -#include "aarch32/instructions-aarch32.h" -#include "aarch32/macro-assembler-aarch32.h" -#pragma GCC diagnostic pop - -// True if VIXL32 should be used for codegen on ARM. -#ifdef USE_VIXL_ARM_BACKEND -static constexpr bool kArmUseVIXL32 = true; -#else -static constexpr bool kArmUseVIXL32 = false; -#endif - -namespace art { -namespace arm { - -static const vixl::aarch32::Register kMethodRegister = vixl::aarch32::r0; -static const vixl::aarch32::Register kCoreAlwaysSpillRegister = vixl::aarch32::r5; -static const vixl::aarch32::RegisterList kCoreCalleeSaves = vixl::aarch32::RegisterList( - (1 << R5) | (1 << R6) | (1 << R7) | (1 << R8) | (1 << R10) | (1 << R11) | (1 << LR)); -// Callee saves s16 to s31 inc. -static const vixl::aarch32::SRegisterList kFpuCalleeSaves = - vixl::aarch32::SRegisterList(vixl::aarch32::s16, 16); - -#define FOR_EACH_IMPLEMENTED_INSTRUCTION(M) \ - M(Above) \ - M(AboveOrEqual) \ - M(Add) \ - M(Below) \ - M(BelowOrEqual) \ - M(Div) \ - M(DivZeroCheck) \ - M(Equal) \ - M(Exit) \ - M(Goto) \ - M(GreaterThan) \ - M(GreaterThanOrEqual) \ - M(If) \ - M(IntConstant) \ - M(LessThan) \ - M(LessThanOrEqual) \ - M(LongConstant) \ - M(MemoryBarrier) \ - M(Mul) \ - M(Not) \ - M(NotEqual) \ - M(ParallelMove) \ - M(Return) \ - M(ReturnVoid) \ - M(Sub) \ - M(TypeConversion) \ - -// TODO: Remove once the VIXL32 backend is implemented completely. -#define FOR_EACH_UNIMPLEMENTED_INSTRUCTION(M) \ - M(And) \ - M(ArrayGet) \ - M(ArrayLength) \ - M(ArraySet) \ - M(BooleanNot) \ - M(BoundsCheck) \ - M(BoundType) \ - M(CheckCast) \ - M(ClassTableGet) \ - M(ClearException) \ - M(ClinitCheck) \ - M(Compare) \ - M(CurrentMethod) \ - M(Deoptimize) \ - M(DoubleConstant) \ - M(FloatConstant) \ - M(InstanceFieldGet) \ - M(InstanceFieldSet) \ - M(InstanceOf) \ - M(InvokeInterface) \ - M(InvokeStaticOrDirect) \ - M(InvokeUnresolved) \ - M(InvokeVirtual) \ - M(LoadClass) \ - M(LoadException) \ - M(LoadString) \ - M(MonitorOperation) \ - M(NativeDebugInfo) \ - M(Neg) \ - M(NewArray) \ - M(NewInstance) \ - M(NullCheck) \ - M(NullConstant) \ - M(Or) \ - M(PackedSwitch) \ - M(ParameterValue) \ - M(Phi) \ - M(Rem) \ - M(Ror) \ - M(Select) \ - M(Shl) \ - M(Shr) \ - M(StaticFieldGet) \ - M(StaticFieldSet) \ - M(SuspendCheck) \ - M(Throw) \ - M(TryBoundary) \ - M(UnresolvedInstanceFieldGet) \ - M(UnresolvedInstanceFieldSet) \ - M(UnresolvedStaticFieldGet) \ - M(UnresolvedStaticFieldSet) \ - M(UShr) \ - M(Xor) \ - -class CodeGeneratorARMVIXL; - -class SlowPathCodeARMVIXL : public SlowPathCode { - public: - explicit SlowPathCodeARMVIXL(HInstruction* instruction) - : SlowPathCode(instruction), entry_label_(), exit_label_() {} - - vixl::aarch32::Label* GetEntryLabel() { return &entry_label_; } - vixl::aarch32::Label* GetExitLabel() { return &exit_label_; } - - void SaveLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE; - void RestoreLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE; - - private: - vixl::aarch32::Label entry_label_; - vixl::aarch32::Label exit_label_; - - DISALLOW_COPY_AND_ASSIGN(SlowPathCodeARMVIXL); -}; - -class ParallelMoveResolverARMVIXL : public ParallelMoveResolverWithSwap { - public: - ParallelMoveResolverARMVIXL(ArenaAllocator* allocator, CodeGeneratorARMVIXL* codegen) - : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} - - void EmitMove(size_t index) OVERRIDE; - void EmitSwap(size_t index) OVERRIDE; - void SpillScratch(int reg) OVERRIDE; - void RestoreScratch(int reg) OVERRIDE; - - ArmVIXLAssembler* GetAssembler() const; - - private: - void Exchange(Register reg, int mem); - void Exchange(int mem1, int mem2); - - CodeGeneratorARMVIXL* const codegen_; - - DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolverARMVIXL); -}; - -#define DEFINE_IMPLEMENTED_INSTRUCTION_VISITOR(Name) \ - void Visit##Name(H##Name*) OVERRIDE; - -#define DEFINE_UNIMPLEMENTED_INSTRUCTION_VISITOR(Name) \ - void Visit##Name(H##Name* instr) OVERRIDE { \ - VisitUnimplemementedInstruction(instr); } - -class LocationsBuilderARMVIXL : public HGraphVisitor { - public: - LocationsBuilderARMVIXL(HGraph* graph, CodeGeneratorARMVIXL* codegen) - : HGraphVisitor(graph), codegen_(codegen) {} - - FOR_EACH_IMPLEMENTED_INSTRUCTION(DEFINE_IMPLEMENTED_INSTRUCTION_VISITOR) - - FOR_EACH_UNIMPLEMENTED_INSTRUCTION(DEFINE_UNIMPLEMENTED_INSTRUCTION_VISITOR) - - private: - void VisitUnimplemementedInstruction(HInstruction* instruction) { - LOG(FATAL) << "Unimplemented Instruction: " << instruction->DebugName(); - } - - void HandleCondition(HCondition* condition); - - CodeGeneratorARMVIXL* const codegen_; - InvokeDexCallingConventionVisitorARM parameter_visitor_; - - DISALLOW_COPY_AND_ASSIGN(LocationsBuilderARMVIXL); -}; - -class InstructionCodeGeneratorARMVIXL : public InstructionCodeGenerator { - public: - InstructionCodeGeneratorARMVIXL(HGraph* graph, CodeGeneratorARMVIXL* codegen); - - FOR_EACH_IMPLEMENTED_INSTRUCTION(DEFINE_IMPLEMENTED_INSTRUCTION_VISITOR) - - FOR_EACH_UNIMPLEMENTED_INSTRUCTION(DEFINE_UNIMPLEMENTED_INSTRUCTION_VISITOR) - - ArmVIXLAssembler* GetAssembler() const { return assembler_; } - vixl::aarch32::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } - - private: - void VisitUnimplemementedInstruction(HInstruction* instruction) { - LOG(FATAL) << "Unimplemented Instruction: " << instruction->DebugName(); - } - - void GenerateSuspendCheck(HSuspendCheck* instruction, HBasicBlock* successor); - void HandleGoto(HInstruction* got, HBasicBlock* successor); - void HandleCondition(HCondition* condition); - void GenerateTestAndBranch(HInstruction* instruction, - size_t condition_input_index, - vixl::aarch32::Label* true_target, - vixl::aarch32::Label* false_target); - void GenerateCompareTestAndBranch(HCondition* condition, - vixl::aarch32::Label* true_target, - vixl::aarch32::Label* false_target); - void GenerateVcmp(HInstruction* instruction); - void GenerateFPJumps(HCondition* cond, - vixl::aarch32::Label* true_label, - vixl::aarch32::Label* false_label); - void GenerateLongComparesAndJumps(HCondition* cond, - vixl::aarch32::Label* true_label, - vixl::aarch32::Label* false_label); - void DivRemOneOrMinusOne(HBinaryOperation* instruction); - void DivRemByPowerOfTwo(HBinaryOperation* instruction); - void GenerateDivRemWithAnyConstant(HBinaryOperation* instruction); - void GenerateDivRemConstantIntegral(HBinaryOperation* instruction); - - ArmVIXLAssembler* const assembler_; - CodeGeneratorARMVIXL* const codegen_; - - DISALLOW_COPY_AND_ASSIGN(InstructionCodeGeneratorARMVIXL); -}; - -class CodeGeneratorARMVIXL : public CodeGenerator { - public: - CodeGeneratorARMVIXL(HGraph* graph, - const ArmInstructionSetFeatures& isa_features, - const CompilerOptions& compiler_options, - OptimizingCompilerStats* stats = nullptr); - - virtual ~CodeGeneratorARMVIXL() {} - - void Initialize() OVERRIDE { - block_labels_.resize(GetGraph()->GetBlocks().size()); - } - - void GenerateFrameEntry() OVERRIDE; - void GenerateFrameExit() OVERRIDE; - void Bind(HBasicBlock* block) OVERRIDE; - void MoveConstant(Location destination, int32_t value) OVERRIDE; - void MoveLocation(Location dst, Location src, Primitive::Type dst_type) OVERRIDE; - void AddLocationAsTemp(Location location, LocationSummary* locations) OVERRIDE; - - ArmVIXLAssembler* GetAssembler() OVERRIDE { return &assembler_; } - - const ArmVIXLAssembler& GetAssembler() const OVERRIDE { return assembler_; } - - vixl::aarch32::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } - - size_t GetWordSize() const OVERRIDE { return kArmWordSize; } - - size_t GetFloatingPointSpillSlotSize() const OVERRIDE { return vixl::aarch32::kRegSizeInBytes; } - - HGraphVisitor* GetLocationBuilder() OVERRIDE { return &location_builder_; } - - HGraphVisitor* GetInstructionVisitor() OVERRIDE { return &instruction_visitor_; } - - uintptr_t GetAddressOf(HBasicBlock* block) OVERRIDE; - - void GenerateMemoryBarrier(MemBarrierKind kind); - void Finalize(CodeAllocator* allocator) OVERRIDE; - void SetupBlockedRegisters() const OVERRIDE; - - // Blocks all register pairs made out of blocked core registers. - void UpdateBlockedPairRegisters() const; - - void DumpCoreRegister(std::ostream& stream, int reg) const OVERRIDE; - void DumpFloatingPointRegister(std::ostream& stream, int reg) const OVERRIDE; - - InstructionSet GetInstructionSet() const OVERRIDE { return InstructionSet::kThumb2; } - - const ArmInstructionSetFeatures& GetInstructionSetFeatures() const { return isa_features_; } - - vixl::aarch32::Label* GetFrameEntryLabel() { return &frame_entry_label_; } - - // Saves the register in the stack. Returns the size taken on stack. - size_t SaveCoreRegister(size_t stack_index ATTRIBUTE_UNUSED, - uint32_t reg_id ATTRIBUTE_UNUSED) OVERRIDE { - UNIMPLEMENTED(INFO) << "TODO: SaveCoreRegister"; - return 0; - } - - // Restores the register from the stack. Returns the size taken on stack. - size_t RestoreCoreRegister(size_t stack_index ATTRIBUTE_UNUSED, - uint32_t reg_id ATTRIBUTE_UNUSED) OVERRIDE { - UNIMPLEMENTED(INFO) << "TODO: RestoreCoreRegister"; - return 0; - } - - size_t SaveFloatingPointRegister(size_t stack_index ATTRIBUTE_UNUSED, - uint32_t reg_id ATTRIBUTE_UNUSED) OVERRIDE { - UNIMPLEMENTED(INFO) << "TODO: SaveFloatingPointRegister"; - return 0; - } - - size_t RestoreFloatingPointRegister(size_t stack_index ATTRIBUTE_UNUSED, - uint32_t reg_id ATTRIBUTE_UNUSED) OVERRIDE { - UNIMPLEMENTED(INFO) << "TODO: RestoreFloatingPointRegister"; - return 0; - } - - bool NeedsTwoRegisters(Primitive::Type type) const OVERRIDE { - return type == Primitive::kPrimDouble || type == Primitive::kPrimLong; - } - - void ComputeSpillMask() OVERRIDE; - - void GenerateImplicitNullCheck(HNullCheck* null_check) OVERRIDE; - void GenerateExplicitNullCheck(HNullCheck* null_check) OVERRIDE; - - ParallelMoveResolver* GetMoveResolver() OVERRIDE { - return &move_resolver_; - } - - // Generate code to invoke a runtime entry point. - void InvokeRuntime(QuickEntrypointEnum entrypoint, - HInstruction* instruction, - uint32_t dex_pc, - SlowPathCode* slow_path = nullptr) OVERRIDE; - - // Generate code to invoke a runtime entry point, but do not record - // PC-related information in a stack map. - void InvokeRuntimeWithoutRecordingPcInfo(int32_t entry_point_offset, - HInstruction* instruction, - SlowPathCode* slow_path); - - void GenerateInvokeRuntime(int32_t entry_point_offset); - - // Check if the desired_string_load_kind is supported. If it is, return it, - // otherwise return a fall-back kind that should be used instead. - HLoadString::LoadKind GetSupportedLoadStringKind( - HLoadString::LoadKind desired_string_load_kind) OVERRIDE; - - // Check if the desired_class_load_kind is supported. If it is, return it, - // otherwise return a fall-back kind that should be used instead. - HLoadClass::LoadKind GetSupportedLoadClassKind( - HLoadClass::LoadKind desired_class_load_kind) OVERRIDE; - - // Check if the desired_dispatch_info is supported. If it is, return it, - // otherwise return a fall-back info that should be used instead. - HInvokeStaticOrDirect::DispatchInfo GetSupportedInvokeStaticOrDirectDispatch( - const HInvokeStaticOrDirect::DispatchInfo& desired_dispatch_info, - MethodReference target_method) OVERRIDE; - - void GenerateStaticOrDirectCall(HInvokeStaticOrDirect* invoke, Location temp) OVERRIDE; - void GenerateVirtualCall(HInvokeVirtual* invoke, Location temp) OVERRIDE; - - void MoveFromReturnRegister(Location trg, Primitive::Type type) OVERRIDE; - - void GenerateNop() OVERRIDE; - - vixl::aarch32::Label* GetLabelOf(HBasicBlock* block) { - block = FirstNonEmptyBlock(block); - return &(block_labels_[block->GetBlockId()]); - } - - private: - // Labels for each block that will be compiled. - // We use a deque so that the `vixl::aarch32::Label` objects do not move in memory. - ArenaDeque<vixl::aarch32::Label> block_labels_; // Indexed by block id. - vixl::aarch32::Label frame_entry_label_; - - LocationsBuilderARMVIXL location_builder_; - InstructionCodeGeneratorARMVIXL instruction_visitor_; - ParallelMoveResolverARMVIXL move_resolver_; - - ArmVIXLAssembler assembler_; - const ArmInstructionSetFeatures& isa_features_; - - DISALLOW_COPY_AND_ASSIGN(CodeGeneratorARMVIXL); -}; - -#undef FOR_EACH_IMPLEMENTED_INSTRUCTION -#undef FOR_EACH_UNIMPLEMENTED_INSTRUCTION -#undef DEFINE_IMPLEMENTED_INSTRUCTION_VISITOR -#undef DEFINE_UNIMPLEMENTED_INSTRUCTION_VISITOR - - -} // namespace arm -} // namespace art - -#endif // ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_VIXL_H_ diff --git a/compiler/optimizing/codegen_test.cc b/compiler/optimizing/codegen_test.cc index f19faa324c..070cbb3894 100644 --- a/compiler/optimizing/codegen_test.cc +++ b/compiler/optimizing/codegen_test.cc @@ -41,7 +41,6 @@ #include "register_allocator_linear_scan.h" #include "ssa_liveness_analysis.h" #include "utils.h" -#include "utils/arm/assembler_arm_vixl.h" #include "utils/arm/managed_register_arm.h" #include "utils/mips/managed_register_mips.h" #include "utils/mips64/managed_register_mips64.h" @@ -49,7 +48,6 @@ #ifdef ART_ENABLE_CODEGEN_arm #include "code_generator_arm.h" -#include "code_generator_arm_vixl.h" #endif #ifdef ART_ENABLE_CODEGEN_arm64 @@ -119,28 +117,6 @@ class TestCodeGeneratorARM : public arm::CodeGeneratorARM { blocked_register_pairs_[arm::R6_R7] = false; } }; - -// A way to test the VIXL32-based code generator on ARM. This will replace -// TestCodeGeneratorARM when the VIXL32-based backend replaces the existing one. -class TestCodeGeneratorARMVIXL : public arm::CodeGeneratorARMVIXL { - public: - TestCodeGeneratorARMVIXL(HGraph* graph, - const ArmInstructionSetFeatures& isa_features, - const CompilerOptions& compiler_options) - : arm::CodeGeneratorARMVIXL(graph, isa_features, compiler_options) { - AddAllocatedRegister(Location::RegisterLocation(arm::R6)); - AddAllocatedRegister(Location::RegisterLocation(arm::R7)); - } - - void SetupBlockedRegisters() const OVERRIDE { - arm::CodeGeneratorARMVIXL::SetupBlockedRegisters(); - blocked_core_registers_[arm::R4] = true; - blocked_core_registers_[arm::R6] = false; - blocked_core_registers_[arm::R7] = false; - // Makes pair R6-R7 available. - blocked_register_pairs_[arm::R6_R7] = false; - } -}; #endif #ifdef ART_ENABLE_CODEGEN_x86 @@ -320,13 +296,6 @@ CodeGenerator* create_codegen_arm(HGraph* graph, const CompilerOptions& compiler *features_arm.get(), compiler_options); } - -CodeGenerator* create_codegen_arm_vixl32(HGraph* graph, const CompilerOptions& compiler_options) { - std::unique_ptr<const ArmInstructionSetFeatures> features_arm( - ArmInstructionSetFeatures::FromCppDefines()); - return new (graph->GetArena()) - TestCodeGeneratorARMVIXL(graph, *features_arm.get(), compiler_options); -} #endif #ifdef ART_ENABLE_CODEGEN_arm64 @@ -382,7 +351,6 @@ static ::std::vector<CodegenTargetConfig> GetTargetConfigs() { #ifdef ART_ENABLE_CODEGEN_arm CodegenTargetConfig(kArm, create_codegen_arm), CodegenTargetConfig(kThumb2, create_codegen_arm), - CodegenTargetConfig(kArm, create_codegen_arm_vixl32), #endif #ifdef ART_ENABLE_CODEGEN_arm64 CodegenTargetConfig(kArm64, create_codegen_arm64), diff --git a/compiler/optimizing/common_arm.h b/compiler/optimizing/common_arm.h deleted file mode 100644 index 853541754d..0000000000 --- a/compiler/optimizing/common_arm.h +++ /dev/null @@ -1,127 +0,0 @@ -/* - * Copyright (C) 2016 The Android Open Source Project - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -#ifndef ART_COMPILER_OPTIMIZING_COMMON_ARM_H_ -#define ART_COMPILER_OPTIMIZING_COMMON_ARM_H_ - -// TODO(VIXL): Make VIXL compile with -Wshadow. -#pragma GCC diagnostic push -#pragma GCC diagnostic ignored "-Wshadow" -#include "aarch32/macro-assembler-aarch32.h" -#pragma GCC diagnostic pop - -namespace art { -namespace arm { -namespace helpers { - -static_assert(vixl::aarch32::kSpCode == SP, "vixl::aarch32::kSpCode must equal ART's SP"); - -inline dwarf::Reg DWARFReg(vixl::aarch32::Register reg) { - return dwarf::Reg::ArmCore(static_cast<int>(reg.GetCode())); -} - -inline dwarf::Reg DWARFReg(vixl::aarch32::SRegister reg) { - return dwarf::Reg::ArmFp(static_cast<int>(reg.GetCode())); -} - -inline vixl::aarch32::DRegister FromLowSToD(vixl::aarch32::SRegister reg) { - DCHECK_EQ(reg.GetCode() % 2, 0u) << reg; - return vixl::aarch32::DRegister(reg.GetCode() / 2); -} - -inline vixl::aarch32::Register RegisterFrom(Location location) { - DCHECK(location.IsRegister()) << location; - return vixl::aarch32::Register(location.reg()); -} - -inline vixl::aarch32::Register RegisterFrom(Location location, Primitive::Type type) { - DCHECK(type != Primitive::kPrimVoid && !Primitive::IsFloatingPointType(type)) << type; - return RegisterFrom(location); -} - -inline vixl::aarch32::DRegister DRegisterFrom(Location location) { - DCHECK(location.IsFpuRegister()) << location; - return vixl::aarch32::DRegister(location.reg()); -} - -inline vixl::aarch32::SRegister SRegisterFrom(Location location) { - DCHECK(location.IsFpuRegister()) << location; - return vixl::aarch32::SRegister(location.reg()); -} - -inline vixl::aarch32::SRegister OutputSRegister(HInstruction* instr) { - Primitive::Type type = instr->GetType(); - DCHECK_EQ(type, Primitive::kPrimFloat) << type; - return SRegisterFrom(instr->GetLocations()->Out()); -} - -inline vixl::aarch32::DRegister OutputDRegister(HInstruction* instr) { - Primitive::Type type = instr->GetType(); - DCHECK_EQ(type, Primitive::kPrimDouble) << type; - return DRegisterFrom(instr->GetLocations()->Out()); -} - -inline vixl::aarch32::SRegister InputSRegisterAt(HInstruction* instr, int input_index) { - Primitive::Type type = instr->InputAt(input_index)->GetType(); - DCHECK_EQ(type, Primitive::kPrimFloat) << type; - return SRegisterFrom(instr->GetLocations()->InAt(input_index)); -} - -inline vixl::aarch32::DRegister InputDRegisterAt(HInstruction* instr, int input_index) { - Primitive::Type type = instr->InputAt(input_index)->GetType(); - DCHECK_EQ(type, Primitive::kPrimDouble) << type; - return DRegisterFrom(instr->GetLocations()->InAt(input_index)); -} - -inline vixl::aarch32::Register OutputRegister(HInstruction* instr) { - return RegisterFrom(instr->GetLocations()->Out(), instr->GetType()); -} - -inline vixl::aarch32::Register InputRegisterAt(HInstruction* instr, int input_index) { - return RegisterFrom(instr->GetLocations()->InAt(input_index), - instr->InputAt(input_index)->GetType()); -} - -inline int64_t Int64ConstantFrom(Location location) { - HConstant* instr = location.GetConstant(); - if (instr->IsIntConstant()) { - return instr->AsIntConstant()->GetValue(); - } else if (instr->IsNullConstant()) { - return 0; - } else { - DCHECK(instr->IsLongConstant()) << instr->DebugName(); - return instr->AsLongConstant()->GetValue(); - } -} - -inline vixl::aarch32::Operand OperandFrom(Location location, Primitive::Type type) { - if (location.IsRegister()) { - return vixl::aarch32::Operand(RegisterFrom(location, type)); - } else { - return vixl::aarch32::Operand(Int64ConstantFrom(location)); - } -} - -inline vixl::aarch32::Operand InputOperandAt(HInstruction* instr, int input_index) { - return OperandFrom(instr->GetLocations()->InAt(input_index), - instr->InputAt(input_index)->GetType()); -} - -} // namespace helpers -} // namespace arm -} // namespace art - -#endif // ART_COMPILER_OPTIMIZING_COMMON_ARM_H_ |