| /* |
| * Copyright (C) 2023 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "code_generator_riscv64.h" |
| |
| #include "android-base/logging.h" |
| #include "android-base/macros.h" |
| #include "arch/riscv64/jni_frame_riscv64.h" |
| #include "arch/riscv64/registers_riscv64.h" |
| #include "base/arena_containers.h" |
| #include "base/macros.h" |
| #include "code_generator_utils.h" |
| #include "dwarf/register.h" |
| #include "heap_poisoning.h" |
| #include "intrinsics_list.h" |
| #include "intrinsics_riscv64.h" |
| #include "jit/profiling_info.h" |
| #include "linker/linker_patch.h" |
| #include "mirror/class-inl.h" |
| #include "optimizing/nodes.h" |
| #include "stack_map_stream.h" |
| #include "utils/label.h" |
| #include "utils/riscv64/assembler_riscv64.h" |
| #include "utils/stack_checks.h" |
| |
| namespace art HIDDEN { |
| namespace riscv64 { |
| |
| // Compare-and-jump packed switch generates approx. 3 + 1.5 * N 32-bit |
| // instructions for N cases. |
| // Table-based packed switch generates approx. 10 32-bit instructions |
| // and N 32-bit data words for N cases. |
| // We switch to the table-based method starting with 6 entries. |
| static constexpr uint32_t kPackedSwitchCompareJumpThreshold = 6; |
| |
| // FCLASS returns a 10-bit classification mask with the two highest bits marking NaNs |
| // (signaling and quiet). To detect a NaN, we can compare (either BGE or BGEU, the sign |
| // bit is always clear) the result with the `kFClassNaNMinValue`. |
| static_assert(kSignalingNaN == 0x100); |
| static_assert(kQuietNaN == 0x200); |
| static constexpr int32_t kFClassNaNMinValue = 0x100; |
| |
| static constexpr XRegister kCoreCalleeSaves[] = { |
| // S1(TR) is excluded as the ART thread register. |
| S0, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, RA |
| }; |
| |
| static constexpr FRegister kFpuCalleeSaves[] = { |
| FS0, FS1, FS2, FS3, FS4, FS5, FS6, FS7, FS8, FS9, FS10, FS11 |
| }; |
| |
| #define QUICK_ENTRY_POINT(x) QUICK_ENTRYPOINT_OFFSET(kRiscv64PointerSize, x).Int32Value() |
| |
| Location RegisterOrZeroBitPatternLocation(HInstruction* instruction) { |
| return IsZeroBitPattern(instruction) |
| ? Location::ConstantLocation(instruction->AsConstant()) |
| : Location::RequiresRegister(); |
| } |
| |
| XRegister InputXRegisterOrZero(Location location) { |
| if (location.IsConstant()) { |
| DCHECK(location.GetConstant()->IsZeroBitPattern()); |
| return Zero; |
| } else { |
| return location.AsRegister<XRegister>(); |
| } |
| } |
| |
| Location Riscv64ReturnLocation(DataType::Type return_type) { |
| switch (return_type) { |
| case DataType::Type::kBool: |
| case DataType::Type::kUint8: |
| case DataType::Type::kInt8: |
| case DataType::Type::kUint16: |
| case DataType::Type::kInt16: |
| case DataType::Type::kUint32: |
| case DataType::Type::kInt32: |
| case DataType::Type::kReference: |
| case DataType::Type::kUint64: |
| case DataType::Type::kInt64: |
| return Location::RegisterLocation(A0); |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| return Location::FpuRegisterLocation(FA0); |
| |
| case DataType::Type::kVoid: |
| return Location::NoLocation(); |
| } |
| UNREACHABLE(); |
| } |
| |
| static RegisterSet OneRegInReferenceOutSaveEverythingCallerSaves() { |
| InvokeRuntimeCallingConvention calling_convention; |
| RegisterSet caller_saves = RegisterSet::Empty(); |
| caller_saves.Add(Location::RegisterLocation(calling_convention.GetRegisterAt(0))); |
| DCHECK_EQ( |
| calling_convention.GetRegisterAt(0), |
| calling_convention.GetReturnLocation(DataType::Type::kReference).AsRegister<XRegister>()); |
| return caller_saves; |
| } |
| |
| template <ClassStatus kStatus> |
| static constexpr int64_t ShiftedSignExtendedClassStatusValue() { |
| // This is used only for status values that have the highest bit set. |
| static_assert(CLZ(enum_cast<uint32_t>(kStatus)) == status_lsb_position); |
| constexpr uint32_t kShiftedStatusValue = enum_cast<uint32_t>(kStatus) << status_lsb_position; |
| static_assert(kShiftedStatusValue >= 0x80000000u); |
| return static_cast<int64_t>(kShiftedStatusValue) - (INT64_C(1) << 32); |
| } |
| |
| Location InvokeRuntimeCallingConvention::GetReturnLocation(DataType::Type return_type) { |
| return Riscv64ReturnLocation(return_type); |
| } |
| |
| Location InvokeDexCallingConventionVisitorRISCV64::GetReturnLocation(DataType::Type type) const { |
| return Riscv64ReturnLocation(type); |
| } |
| |
| Location InvokeDexCallingConventionVisitorRISCV64::GetMethodLocation() const { |
| return Location::RegisterLocation(kArtMethodRegister); |
| } |
| |
| Location InvokeDexCallingConventionVisitorRISCV64::GetNextLocation(DataType::Type type) { |
| Location next_location; |
| if (type == DataType::Type::kVoid) { |
| LOG(FATAL) << "Unexpected parameter type " << type; |
| } |
| |
| // Note: Unlike the RISC-V C/C++ calling convention, managed ABI does not use |
| // GPRs to pass FP args when we run out of FPRs. |
| if (DataType::IsFloatingPointType(type) && |
| float_index_ < calling_convention.GetNumberOfFpuRegisters()) { |
| next_location = |
| Location::FpuRegisterLocation(calling_convention.GetFpuRegisterAt(float_index_++)); |
| } else if (!DataType::IsFloatingPointType(type) && |
| (gp_index_ < calling_convention.GetNumberOfRegisters())) { |
| next_location = Location::RegisterLocation(calling_convention.GetRegisterAt(gp_index_++)); |
| } else { |
| size_t stack_offset = calling_convention.GetStackOffsetOf(stack_index_); |
| next_location = DataType::Is64BitType(type) ? Location::DoubleStackSlot(stack_offset) : |
| Location::StackSlot(stack_offset); |
| } |
| |
| // Space on the stack is reserved for all arguments. |
| stack_index_ += DataType::Is64BitType(type) ? 2 : 1; |
| |
| return next_location; |
| } |
| |
| Location CriticalNativeCallingConventionVisitorRiscv64::GetNextLocation(DataType::Type type) { |
| DCHECK_NE(type, DataType::Type::kReference); |
| |
| Location location = Location::NoLocation(); |
| if (DataType::IsFloatingPointType(type)) { |
| if (fpr_index_ < kParameterFpuRegistersLength) { |
| location = Location::FpuRegisterLocation(kParameterFpuRegisters[fpr_index_]); |
| ++fpr_index_; |
| } |
| // Native ABI allows passing excessive FP args in GPRs. This is facilitated by |
| // inserting fake conversion intrinsic calls (`Double.doubleToRawLongBits()` |
| // or `Float.floatToRawIntBits()`) by `CriticalNativeAbiFixupRiscv64`. |
| // TODO(riscv64): Implement these intrinsics and `CriticalNativeAbiFixupRiscv64`. |
| } else { |
| // Native ABI uses the same core registers as a runtime call. |
| if (gpr_index_ < kRuntimeParameterCoreRegistersLength) { |
| location = Location::RegisterLocation(kRuntimeParameterCoreRegisters[gpr_index_]); |
| ++gpr_index_; |
| } |
| } |
| if (location.IsInvalid()) { |
| if (DataType::Is64BitType(type)) { |
| location = Location::DoubleStackSlot(stack_offset_); |
| } else { |
| location = Location::StackSlot(stack_offset_); |
| } |
| stack_offset_ += kFramePointerSize; |
| |
| if (for_register_allocation_) { |
| location = Location::Any(); |
| } |
| } |
| return location; |
| } |
| |
| Location CriticalNativeCallingConventionVisitorRiscv64::GetReturnLocation( |
| DataType::Type type) const { |
| // The result is returned the same way in native ABI and managed ABI. No result conversion is |
| // needed, see comments in `Riscv64JniCallingConvention::RequiresSmallResultTypeExtension()`. |
| InvokeDexCallingConventionVisitorRISCV64 dex_calling_convention; |
| return dex_calling_convention.GetReturnLocation(type); |
| } |
| |
| Location CriticalNativeCallingConventionVisitorRiscv64::GetMethodLocation() const { |
| // Pass the method in the hidden argument T0. |
| return Location::RegisterLocation(T0); |
| } |
| |
| #define __ down_cast<CodeGeneratorRISCV64*>(codegen)->GetAssembler()-> // NOLINT |
| |
| void LocationsBuilderRISCV64::HandleInvoke(HInvoke* instruction) { |
| InvokeDexCallingConventionVisitorRISCV64 calling_convention_visitor; |
| CodeGenerator::CreateCommonInvokeLocationSummary(instruction, &calling_convention_visitor); |
| } |
| |
| Location LocationsBuilderRISCV64::RegisterOrZeroConstant(HInstruction* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| UNREACHABLE(); |
| } |
| |
| Location LocationsBuilderRISCV64::FpuRegisterOrConstantForStore(HInstruction* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| UNREACHABLE(); |
| } |
| |
| class CompileOptimizedSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| CompileOptimizedSlowPathRISCV64() : SlowPathCodeRISCV64(/*instruction=*/ nullptr) {} |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| uint32_t entrypoint_offset = |
| GetThreadOffset<kRiscv64PointerSize>(kQuickCompileOptimized).Int32Value(); |
| __ Bind(GetEntryLabel()); |
| __ Loadd(RA, TR, entrypoint_offset); |
| // Note: we don't record the call here (and therefore don't generate a stack |
| // map), as the entrypoint should never be suspended. |
| __ Jalr(RA); |
| __ J(GetExitLabel()); |
| } |
| |
| const char* GetDescription() const override { return "CompileOptimizedSlowPath"; } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(CompileOptimizedSlowPathRISCV64); |
| }; |
| |
| class SuspendCheckSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| SuspendCheckSlowPathRISCV64(HSuspendCheck* instruction, HBasicBlock* successor) |
| : SlowPathCodeRISCV64(instruction), successor_(successor) {} |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| LocationSummary* locations = instruction_->GetLocations(); |
| CodeGeneratorRISCV64* riscv64_codegen = down_cast<CodeGeneratorRISCV64*>(codegen); |
| __ Bind(GetEntryLabel()); |
| SaveLiveRegisters(codegen, locations); // Only saves live vector registers for SIMD. |
| riscv64_codegen->InvokeRuntime(kQuickTestSuspend, instruction_, instruction_->GetDexPc(), this); |
| CheckEntrypointTypes<kQuickTestSuspend, void, void>(); |
| RestoreLiveRegisters(codegen, locations); // Only restores live vector registers for SIMD. |
| if (successor_ == nullptr) { |
| __ J(GetReturnLabel()); |
| } else { |
| __ J(riscv64_codegen->GetLabelOf(successor_)); |
| } |
| } |
| |
| Riscv64Label* GetReturnLabel() { |
| DCHECK(successor_ == nullptr); |
| return &return_label_; |
| } |
| |
| const char* GetDescription() const override { return "SuspendCheckSlowPathRISCV64"; } |
| |
| HBasicBlock* GetSuccessor() const { return successor_; } |
| |
| private: |
| // If not null, the block to branch to after the suspend check. |
| HBasicBlock* const successor_; |
| |
| // If `successor_` is null, the label to branch to after the suspend check. |
| Riscv64Label return_label_; |
| |
| DISALLOW_COPY_AND_ASSIGN(SuspendCheckSlowPathRISCV64); |
| }; |
| |
| class NullCheckSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| explicit NullCheckSlowPathRISCV64(HNullCheck* instr) : SlowPathCodeRISCV64(instr) {} |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| CodeGeneratorRISCV64* riscv64_codegen = down_cast<CodeGeneratorRISCV64*>(codegen); |
| __ Bind(GetEntryLabel()); |
| if (instruction_->CanThrowIntoCatchBlock()) { |
| // Live registers will be restored in the catch block if caught. |
| SaveLiveRegisters(codegen, instruction_->GetLocations()); |
| } |
| riscv64_codegen->InvokeRuntime( |
| kQuickThrowNullPointer, instruction_, instruction_->GetDexPc(), this); |
| CheckEntrypointTypes<kQuickThrowNullPointer, void, void>(); |
| } |
| |
| bool IsFatal() const override { return true; } |
| |
| const char* GetDescription() const override { return "NullCheckSlowPathRISCV64"; } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(NullCheckSlowPathRISCV64); |
| }; |
| |
| class BoundsCheckSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| explicit BoundsCheckSlowPathRISCV64(HBoundsCheck* instruction) |
| : SlowPathCodeRISCV64(instruction) {} |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| LocationSummary* locations = instruction_->GetLocations(); |
| CodeGeneratorRISCV64* riscv64_codegen = down_cast<CodeGeneratorRISCV64*>(codegen); |
| __ Bind(GetEntryLabel()); |
| if (instruction_->CanThrowIntoCatchBlock()) { |
| // Live registers will be restored in the catch block if caught. |
| SaveLiveRegisters(codegen, instruction_->GetLocations()); |
| } |
| // We're moving two locations to locations that could overlap, so we need a parallel |
| // move resolver. |
| InvokeRuntimeCallingConvention calling_convention; |
| codegen->EmitParallelMoves(locations->InAt(0), |
| Location::RegisterLocation(calling_convention.GetRegisterAt(0)), |
| DataType::Type::kInt32, |
| locations->InAt(1), |
| Location::RegisterLocation(calling_convention.GetRegisterAt(1)), |
| DataType::Type::kInt32); |
| QuickEntrypointEnum entrypoint = instruction_->AsBoundsCheck()->IsStringCharAt() ? |
| kQuickThrowStringBounds : |
| kQuickThrowArrayBounds; |
| riscv64_codegen->InvokeRuntime(entrypoint, instruction_, instruction_->GetDexPc(), this); |
| CheckEntrypointTypes<kQuickThrowStringBounds, void, int32_t, int32_t>(); |
| CheckEntrypointTypes<kQuickThrowArrayBounds, void, int32_t, int32_t>(); |
| } |
| |
| bool IsFatal() const override { return true; } |
| |
| const char* GetDescription() const override { return "BoundsCheckSlowPathRISCV64"; } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(BoundsCheckSlowPathRISCV64); |
| }; |
| |
| class LoadClassSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| LoadClassSlowPathRISCV64(HLoadClass* cls, HInstruction* at) : SlowPathCodeRISCV64(at), cls_(cls) { |
| DCHECK(at->IsLoadClass() || at->IsClinitCheck()); |
| DCHECK_EQ(instruction_->IsLoadClass(), cls_ == instruction_); |
| } |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| LocationSummary* locations = instruction_->GetLocations(); |
| Location out = locations->Out(); |
| const uint32_t dex_pc = instruction_->GetDexPc(); |
| bool must_resolve_type = instruction_->IsLoadClass() && cls_->MustResolveTypeOnSlowPath(); |
| bool must_do_clinit = instruction_->IsClinitCheck() || cls_->MustGenerateClinitCheck(); |
| |
| CodeGeneratorRISCV64* riscv64_codegen = down_cast<CodeGeneratorRISCV64*>(codegen); |
| __ Bind(GetEntryLabel()); |
| SaveLiveRegisters(codegen, locations); |
| |
| InvokeRuntimeCallingConvention calling_convention; |
| if (must_resolve_type) { |
| DCHECK(IsSameDexFile(cls_->GetDexFile(), riscv64_codegen->GetGraph()->GetDexFile()) || |
| riscv64_codegen->GetCompilerOptions().WithinOatFile(&cls_->GetDexFile()) || |
| ContainsElement(Runtime::Current()->GetClassLinker()->GetBootClassPath(), |
| &cls_->GetDexFile())); |
| dex::TypeIndex type_index = cls_->GetTypeIndex(); |
| __ LoadConst32(calling_convention.GetRegisterAt(0), type_index.index_); |
| if (cls_->NeedsAccessCheck()) { |
| CheckEntrypointTypes<kQuickResolveTypeAndVerifyAccess, void*, uint32_t>(); |
| riscv64_codegen->InvokeRuntime( |
| kQuickResolveTypeAndVerifyAccess, instruction_, dex_pc, this); |
| } else { |
| CheckEntrypointTypes<kQuickResolveType, void*, uint32_t>(); |
| riscv64_codegen->InvokeRuntime(kQuickResolveType, instruction_, dex_pc, this); |
| } |
| // If we also must_do_clinit, the resolved type is now in the correct register. |
| } else { |
| DCHECK(must_do_clinit); |
| Location source = instruction_->IsLoadClass() ? out : locations->InAt(0); |
| riscv64_codegen->MoveLocation( |
| Location::RegisterLocation(calling_convention.GetRegisterAt(0)), source, cls_->GetType()); |
| } |
| if (must_do_clinit) { |
| riscv64_codegen->InvokeRuntime(kQuickInitializeStaticStorage, instruction_, dex_pc, this); |
| CheckEntrypointTypes<kQuickInitializeStaticStorage, void*, mirror::Class*>(); |
| } |
| |
| // Move the class to the desired location. |
| if (out.IsValid()) { |
| DCHECK(out.IsRegister() && !locations->GetLiveRegisters()->ContainsCoreRegister(out.reg())); |
| DataType::Type type = instruction_->GetType(); |
| riscv64_codegen->MoveLocation( |
| out, Location::RegisterLocation(calling_convention.GetRegisterAt(0)), type); |
| } |
| RestoreLiveRegisters(codegen, locations); |
| |
| __ J(GetExitLabel()); |
| } |
| |
| const char* GetDescription() const override { return "LoadClassSlowPathRISCV64"; } |
| |
| private: |
| // The class this slow path will load. |
| HLoadClass* const cls_; |
| |
| DISALLOW_COPY_AND_ASSIGN(LoadClassSlowPathRISCV64); |
| }; |
| |
| class DeoptimizationSlowPathRISCV64 : public SlowPathCodeRISCV64 { |
| public: |
| explicit DeoptimizationSlowPathRISCV64(HDeoptimize* instruction) |
| : SlowPathCodeRISCV64(instruction) {} |
| |
| void EmitNativeCode(CodeGenerator* codegen) override { |
| CodeGeneratorRISCV64* riscv64_codegen = down_cast<CodeGeneratorRISCV64*>(codegen); |
| __ Bind(GetEntryLabel()); |
| LocationSummary* locations = instruction_->GetLocations(); |
| SaveLiveRegisters(codegen, locations); |
| InvokeRuntimeCallingConvention calling_convention; |
| __ LoadConst32(calling_convention.GetRegisterAt(0), |
| static_cast<uint32_t>(instruction_->AsDeoptimize()->GetDeoptimizationKind())); |
| riscv64_codegen->InvokeRuntime(kQuickDeoptimize, instruction_, instruction_->GetDexPc(), this); |
| CheckEntrypointTypes<kQuickDeoptimize, void, DeoptimizationKind>(); |
| } |
| |
| const char* GetDescription() const override { return "DeoptimizationSlowPathRISCV64"; } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(DeoptimizationSlowPathRISCV64); |
| }; |
| |
| #undef __ |
| #define __ down_cast<Riscv64Assembler*>(GetAssembler())-> // NOLINT |
| |
| template <typename Reg, |
| void (Riscv64Assembler::*opS)(Reg, FRegister, FRegister), |
| void (Riscv64Assembler::*opD)(Reg, FRegister, FRegister)> |
| inline void InstructionCodeGeneratorRISCV64::FpBinOp( |
| Reg rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| Riscv64Assembler* assembler = down_cast<CodeGeneratorRISCV64*>(codegen_)->GetAssembler(); |
| if (type == DataType::Type::kFloat32) { |
| (assembler->*opS)(rd, rs1, rs2); |
| } else { |
| DCHECK_EQ(type, DataType::Type::kFloat64); |
| (assembler->*opD)(rd, rs1, rs2); |
| } |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FAdd( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FAddS, &Riscv64Assembler::FAddD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FSub( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FSubS, &Riscv64Assembler::FSubD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FDiv( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FDivS, &Riscv64Assembler::FDivD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FMul( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FMulS, &Riscv64Assembler::FMulD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FMin( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FMinS, &Riscv64Assembler::FMinD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FMax( |
| FRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<FRegister, &Riscv64Assembler::FMaxS, &Riscv64Assembler::FMaxD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FEq( |
| XRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<XRegister, &Riscv64Assembler::FEqS, &Riscv64Assembler::FEqD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FLt( |
| XRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<XRegister, &Riscv64Assembler::FLtS, &Riscv64Assembler::FLtD>(rd, rs1, rs2, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FLe( |
| XRegister rd, FRegister rs1, FRegister rs2, DataType::Type type) { |
| FpBinOp<XRegister, &Riscv64Assembler::FLeS, &Riscv64Assembler::FLeD>(rd, rs1, rs2, type); |
| } |
| |
| template <typename Reg, |
| void (Riscv64Assembler::*opS)(Reg, FRegister), |
| void (Riscv64Assembler::*opD)(Reg, FRegister)> |
| inline void InstructionCodeGeneratorRISCV64::FpUnOp( |
| Reg rd, FRegister rs1, DataType::Type type) { |
| Riscv64Assembler* assembler = down_cast<CodeGeneratorRISCV64*>(codegen_)->GetAssembler(); |
| if (type == DataType::Type::kFloat32) { |
| (assembler->*opS)(rd, rs1); |
| } else { |
| DCHECK_EQ(type, DataType::Type::kFloat64); |
| (assembler->*opD)(rd, rs1); |
| } |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FAbs( |
| FRegister rd, FRegister rs1, DataType::Type type) { |
| FpUnOp<FRegister, &Riscv64Assembler::FAbsS, &Riscv64Assembler::FAbsD>(rd, rs1, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FNeg( |
| FRegister rd, FRegister rs1, DataType::Type type) { |
| FpUnOp<FRegister, &Riscv64Assembler::FNegS, &Riscv64Assembler::FNegD>(rd, rs1, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FMv( |
| FRegister rd, FRegister rs1, DataType::Type type) { |
| FpUnOp<FRegister, &Riscv64Assembler::FMvS, &Riscv64Assembler::FMvD>(rd, rs1, type); |
| } |
| |
| inline void InstructionCodeGeneratorRISCV64::FClass( |
| XRegister rd, FRegister rs1, DataType::Type type) { |
| FpUnOp<XRegister, &Riscv64Assembler::FClassS, &Riscv64Assembler::FClassD>(rd, rs1, type); |
| } |
| |
| Riscv64Assembler* ParallelMoveResolverRISCV64::GetAssembler() const { |
| return codegen_->GetAssembler(); |
| } |
| |
| void ParallelMoveResolverRISCV64::EmitMove(size_t index) { |
| MoveOperands* move = moves_[index]; |
| codegen_->MoveLocation(move->GetDestination(), move->GetSource(), move->GetType()); |
| } |
| |
| void ParallelMoveResolverRISCV64::EmitSwap(size_t index) { |
| MoveOperands* move = moves_[index]; |
| codegen_->SwapLocations(move->GetDestination(), move->GetSource(), move->GetType()); |
| } |
| |
| void ParallelMoveResolverRISCV64::SpillScratch([[maybe_unused]] int reg) { |
| LOG(FATAL) << "Unimplemented"; |
| UNREACHABLE(); |
| } |
| |
| void ParallelMoveResolverRISCV64::RestoreScratch([[maybe_unused]] int reg) { |
| LOG(FATAL) << "Unimplemented"; |
| UNREACHABLE(); |
| } |
| |
| void ParallelMoveResolverRISCV64::Exchange(int index1, int index2, bool double_slot) { |
| // We have 2 scratch X registers and 1 scratch F register that we can use. We prefer |
| // to use X registers for the swap but if both offsets are too big, we need to reserve |
| // one of the X registers for address adjustment and use an F register. |
| bool use_fp_tmp2 = false; |
| if (!IsInt<12>(index2)) { |
| if (!IsInt<12>(index1)) { |
| use_fp_tmp2 = true; |
| } else { |
| std::swap(index1, index2); |
| } |
| } |
| DCHECK_IMPLIES(!IsInt<12>(index2), use_fp_tmp2); |
| |
| Location loc1(double_slot ? Location::DoubleStackSlot(index1) : Location::StackSlot(index1)); |
| Location loc2(double_slot ? Location::DoubleStackSlot(index2) : Location::StackSlot(index2)); |
| riscv64::ScratchRegisterScope srs(GetAssembler()); |
| Location tmp = Location::RegisterLocation(srs.AllocateXRegister()); |
| DataType::Type tmp_type = double_slot ? DataType::Type::kInt64 : DataType::Type::kInt32; |
| Location tmp2 = use_fp_tmp2 |
| ? Location::FpuRegisterLocation(srs.AllocateFRegister()) |
| : Location::RegisterLocation(srs.AllocateXRegister()); |
| DataType::Type tmp2_type = use_fp_tmp2 |
| ? (double_slot ? DataType::Type::kFloat64 : DataType::Type::kFloat32) |
| : tmp_type; |
| |
| codegen_->MoveLocation(tmp, loc1, tmp_type); |
| codegen_->MoveLocation(tmp2, loc2, tmp2_type); |
| if (use_fp_tmp2) { |
| codegen_->MoveLocation(loc2, tmp, tmp_type); |
| } else { |
| // We cannot use `Stored()` or `Storew()` via `MoveLocation()` because we have |
| // no more scratch registers available. Use `Sd()` or `Sw()` explicitly. |
| DCHECK(IsInt<12>(index2)); |
| if (double_slot) { |
| __ Sd(tmp.AsRegister<XRegister>(), SP, index2); |
| } else { |
| __ Sw(tmp.AsRegister<XRegister>(), SP, index2); |
| } |
| srs.FreeXRegister(tmp.AsRegister<XRegister>()); // Free a temporary for `MoveLocation()`. |
| } |
| codegen_->MoveLocation(loc1, tmp2, tmp2_type); |
| } |
| |
| InstructionCodeGeneratorRISCV64::InstructionCodeGeneratorRISCV64(HGraph* graph, |
| CodeGeneratorRISCV64* codegen) |
| : InstructionCodeGenerator(graph, codegen), |
| assembler_(codegen->GetAssembler()), |
| codegen_(codegen) {} |
| |
| void InstructionCodeGeneratorRISCV64::GenerateClassInitializationCheck( |
| SlowPathCodeRISCV64* slow_path, XRegister class_reg) { |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| XRegister tmp2 = srs.AllocateXRegister(); |
| |
| // We shall load the full 32-bit status word with sign-extension and compare as unsigned |
| // to a sign-extended shifted status value. This yields the same comparison as loading and |
| // materializing unsigned but the constant is materialized with a single LUI instruction. |
| __ Loadw(tmp, class_reg, mirror::Class::StatusOffset().SizeValue()); // Sign-extended. |
| __ Li(tmp2, ShiftedSignExtendedClassStatusValue<ClassStatus::kVisiblyInitialized>()); |
| __ Bltu(tmp, tmp2, slow_path->GetEntryLabel()); |
| __ Bind(slow_path->GetExitLabel()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateBitstringTypeCheckCompare( |
| HTypeCheckInstruction* instruction, XRegister temp) { |
| UNUSED(instruction); |
| UNUSED(temp); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateSuspendCheck(HSuspendCheck* instruction, |
| HBasicBlock* successor) { |
| if (instruction->IsNoOp()) { |
| if (successor != nullptr) { |
| __ J(codegen_->GetLabelOf(successor)); |
| } |
| return; |
| } |
| |
| if (codegen_->CanUseImplicitSuspendCheck()) { |
| LOG(FATAL) << "Unimplemented ImplicitSuspendCheck"; |
| return; |
| } |
| |
| SuspendCheckSlowPathRISCV64* slow_path = |
| down_cast<SuspendCheckSlowPathRISCV64*>(instruction->GetSlowPath()); |
| |
| if (slow_path == nullptr) { |
| slow_path = |
| new (codegen_->GetScopedAllocator()) SuspendCheckSlowPathRISCV64(instruction, successor); |
| instruction->SetSlowPath(slow_path); |
| codegen_->AddSlowPath(slow_path); |
| if (successor != nullptr) { |
| DCHECK(successor->IsLoopHeader()); |
| } |
| } else { |
| DCHECK_EQ(slow_path->GetSuccessor(), successor); |
| } |
| |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| __ Loadw(tmp, TR, Thread::ThreadFlagsOffset<kRiscv64PointerSize>().Int32Value()); |
| static_assert(Thread::SuspendOrCheckpointRequestFlags() != std::numeric_limits<uint32_t>::max()); |
| static_assert(IsPowerOfTwo(Thread::SuspendOrCheckpointRequestFlags() + 1u)); |
| // Shift out other bits. Use an instruction that can be 16-bit with the "C" Standard Extension. |
| __ Slli(tmp, tmp, CLZ(static_cast<uint64_t>(Thread::SuspendOrCheckpointRequestFlags()))); |
| if (successor == nullptr) { |
| __ Bnez(tmp, slow_path->GetEntryLabel()); |
| __ Bind(slow_path->GetReturnLabel()); |
| } else { |
| __ Beqz(tmp, codegen_->GetLabelOf(successor)); |
| __ J(slow_path->GetEntryLabel()); |
| // slow_path will return to GetLabelOf(successor). |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateReferenceLoadOneRegister( |
| HInstruction* instruction, |
| Location out, |
| uint32_t offset, |
| Location maybe_temp, |
| ReadBarrierOption read_barrier_option) { |
| UNUSED(instruction); |
| UNUSED(out); |
| UNUSED(offset); |
| UNUSED(maybe_temp); |
| UNUSED(read_barrier_option); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateReferenceLoadTwoRegisters( |
| HInstruction* instruction, |
| Location out, |
| Location obj, |
| uint32_t offset, |
| Location maybe_temp, |
| ReadBarrierOption read_barrier_option) { |
| UNUSED(instruction); |
| UNUSED(out); |
| UNUSED(offset); |
| UNUSED(maybe_temp); |
| UNUSED(read_barrier_option); |
| UNUSED(obj); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateGcRootFieldLoad(HInstruction* instruction, |
| Location root, |
| XRegister obj, |
| uint32_t offset, |
| ReadBarrierOption read_barrier_option, |
| Riscv64Label* label_low) { |
| UNUSED(instruction); |
| UNUSED(root); |
| UNUSED(obj); |
| UNUSED(offset); |
| UNUSED(read_barrier_option); |
| UNUSED(label_low); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateTestAndBranch(HInstruction* instruction, |
| size_t condition_input_index, |
| Riscv64Label* true_target, |
| Riscv64Label* false_target) { |
| HInstruction* cond = instruction->InputAt(condition_input_index); |
| |
| if (true_target == nullptr && false_target == nullptr) { |
| // Nothing to do. The code always falls through. |
| return; |
| } else if (cond->IsIntConstant()) { |
| // Constant condition, statically compared against "true" (integer value 1). |
| if (cond->AsIntConstant()->IsTrue()) { |
| if (true_target != nullptr) { |
| __ J(true_target); |
| } |
| } else { |
| DCHECK(cond->AsIntConstant()->IsFalse()) << cond->AsIntConstant()->GetValue(); |
| if (false_target != nullptr) { |
| __ J(false_target); |
| } |
| } |
| return; |
| } |
| |
| // The following code generates these patterns: |
| // (1) true_target == nullptr && false_target != nullptr |
| // - opposite condition true => branch to false_target |
| // (2) true_target != nullptr && false_target == nullptr |
| // - condition true => branch to true_target |
| // (3) true_target != nullptr && false_target != nullptr |
| // - condition true => branch to true_target |
| // - branch to false_target |
| if (IsBooleanValueOrMaterializedCondition(cond)) { |
| // The condition instruction has been materialized, compare the output to 0. |
| Location cond_val = instruction->GetLocations()->InAt(condition_input_index); |
| DCHECK(cond_val.IsRegister()); |
| if (true_target == nullptr) { |
| __ Beqz(cond_val.AsRegister<XRegister>(), false_target); |
| } else { |
| __ Bnez(cond_val.AsRegister<XRegister>(), true_target); |
| } |
| } else { |
| // The condition instruction has not been materialized, use its inputs as |
| // the comparison and its condition as the branch condition. |
| HCondition* condition = cond->AsCondition(); |
| DataType::Type type = condition->InputAt(0)->GetType(); |
| LocationSummary* locations = condition->GetLocations(); |
| IfCondition if_cond = condition->GetCondition(); |
| Riscv64Label* branch_target = true_target; |
| |
| if (true_target == nullptr) { |
| if_cond = condition->GetOppositeCondition(); |
| branch_target = false_target; |
| } |
| |
| switch (type) { |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| GenerateFpCondition(if_cond, condition->IsGtBias(), type, locations, branch_target); |
| break; |
| default: |
| // Integral types and reference equality. |
| GenerateIntLongCompareAndBranch(if_cond, locations, branch_target); |
| break; |
| } |
| } |
| |
| // If neither branch falls through (case 3), the conditional branch to `true_target` |
| // was already emitted (case 2) and we need to emit a jump to `false_target`. |
| if (true_target != nullptr && false_target != nullptr) { |
| __ J(false_target); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::DivRemOneOrMinusOne(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsDiv() || instruction->IsRem()); |
| DataType::Type type = instruction->GetResultType(); |
| |
| LocationSummary* locations = instruction->GetLocations(); |
| Location second = locations->InAt(1); |
| DCHECK(second.IsConstant()); |
| |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| XRegister dividend = locations->InAt(0).AsRegister<XRegister>(); |
| int64_t imm = Int64FromConstant(second.GetConstant()); |
| DCHECK(imm == 1 || imm == -1); |
| |
| if (instruction->IsRem()) { |
| __ Mv(out, Zero); |
| } else { |
| if (imm == -1) { |
| if (type == DataType::Type::kInt32) { |
| __ Subw(out, Zero, dividend); |
| } else { |
| DCHECK_EQ(type, DataType::Type::kInt64); |
| __ Sub(out, Zero, dividend); |
| } |
| } else if (out != dividend) { |
| __ Mv(out, dividend); |
| } |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::DivRemByPowerOfTwo(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsDiv() || instruction->IsRem()); |
| DataType::Type type = instruction->GetResultType(); |
| DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64) << type; |
| |
| LocationSummary* locations = instruction->GetLocations(); |
| Location second = locations->InAt(1); |
| DCHECK(second.IsConstant()); |
| |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| XRegister dividend = locations->InAt(0).AsRegister<XRegister>(); |
| int64_t imm = Int64FromConstant(second.GetConstant()); |
| int64_t abs_imm = static_cast<uint64_t>(AbsOrMin(imm)); |
| int ctz_imm = CTZ(abs_imm); |
| DCHECK_GE(ctz_imm, 1); // Division by +/-1 is handled by `DivRemOneOrMinusOne()`. |
| |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| // Calculate the negative dividend adjustment `tmp = dividend < 0 ? abs_imm - 1 : 0`. |
| // This adjustment is needed for rounding the division result towards zero. |
| if (type == DataType::Type::kInt32 || ctz_imm == 1) { |
| // A 32-bit dividend is sign-extended to 64-bit, so we can use the upper bits. |
| // And for a 64-bit division by +/-2, we need just the sign bit. |
| DCHECK_IMPLIES(type == DataType::Type::kInt32, ctz_imm < 32); |
| __ Srli(tmp, dividend, 64 - ctz_imm); |
| } else { |
| // For other 64-bit divisions, we need to replicate the sign bit. |
| __ Srai(tmp, dividend, 63); |
| __ Srli(tmp, tmp, 64 - ctz_imm); |
| } |
| // The rest of the calculation can use 64-bit operations even for 32-bit div/rem. |
| __ Add(tmp, tmp, dividend); |
| if (instruction->IsDiv()) { |
| __ Srai(out, tmp, ctz_imm); |
| if (imm < 0) { |
| __ Neg(out, out); |
| } |
| } else { |
| if (ctz_imm <= 11) { |
| __ Andi(tmp, tmp, -abs_imm); |
| } else { |
| ScratchRegisterScope srs2(GetAssembler()); |
| XRegister tmp2 = srs2.AllocateXRegister(); |
| __ Li(tmp2, -abs_imm); |
| __ And(tmp, tmp, tmp2); |
| } |
| __ Sub(out, dividend, tmp); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateDivRemWithAnyConstant(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsDiv() || instruction->IsRem()); |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister dividend = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| Location second = locations->InAt(1); |
| int64_t imm = Int64FromConstant(second.GetConstant()); |
| DataType::Type type = instruction->GetResultType(); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| |
| // TODO: optimize with constant. |
| __ LoadConst64(tmp, imm); |
| if (instruction->IsDiv()) { |
| if (type == DataType::Type::kInt32) { |
| __ Divw(out, dividend, tmp); |
| } else { |
| __ Div(out, dividend, tmp); |
| } |
| } else { |
| if (type == DataType::Type::kInt32) { |
| __ Remw(out, dividend, tmp); |
| } else { |
| __ Rem(out, dividend, tmp); |
| } |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateDivRemIntegral(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsDiv() || instruction->IsRem()); |
| DataType::Type type = instruction->GetResultType(); |
| DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64) << type; |
| |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| Location second = locations->InAt(1); |
| |
| if (second.IsConstant()) { |
| int64_t imm = Int64FromConstant(second.GetConstant()); |
| if (imm == 0) { |
| // Do not generate anything. DivZeroCheck would prevent any code to be executed. |
| } else if (imm == 1 || imm == -1) { |
| DivRemOneOrMinusOne(instruction); |
| } else if (IsPowerOfTwo(AbsOrMin(imm))) { |
| DivRemByPowerOfTwo(instruction); |
| } else { |
| DCHECK(imm <= -2 || imm >= 2); |
| GenerateDivRemWithAnyConstant(instruction); |
| } |
| } else { |
| XRegister dividend = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister divisor = second.AsRegister<XRegister>(); |
| if (instruction->IsDiv()) { |
| if (type == DataType::Type::kInt32) { |
| __ Divw(out, dividend, divisor); |
| } else { |
| __ Div(out, dividend, divisor); |
| } |
| } else { |
| if (type == DataType::Type::kInt32) { |
| __ Remw(out, dividend, divisor); |
| } else { |
| __ Rem(out, dividend, divisor); |
| } |
| } |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateIntLongCondition(IfCondition cond, |
| LocationSummary* locations) { |
| XRegister rd = locations->Out().AsRegister<XRegister>(); |
| XRegister rs1 = locations->InAt(0).AsRegister<XRegister>(); |
| Location rs2_location = locations->InAt(1); |
| bool use_imm = rs2_location.IsConstant(); |
| int64_t imm = use_imm ? CodeGenerator::GetInt64ValueOf(rs2_location.GetConstant()) : 0; |
| XRegister rs2 = use_imm ? kNoXRegister : rs2_location.AsRegister<XRegister>(); |
| switch (cond) { |
| case kCondEQ: |
| case kCondNE: |
| if (!use_imm) { |
| __ Sub(rd, rs1, rs2); // SUB is OK here even for 32-bit comparison. |
| } else if (imm != 0) { |
| DCHECK(IsInt<12>(-imm)); |
| __ Addi(rd, rs1, -imm); // ADDI is OK here even for 32-bit comparison. |
| } // else test `rs1` directly without subtraction for `use_imm && imm == 0`. |
| if (cond == kCondEQ) { |
| __ Seqz(rd, (use_imm && imm == 0) ? rs1 : rd); |
| } else { |
| __ Snez(rd, (use_imm && imm == 0) ? rs1 : rd); |
| } |
| break; |
| |
| case kCondLT: |
| case kCondGE: |
| if (use_imm) { |
| DCHECK(IsInt<12>(imm)); |
| __ Slti(rd, rs1, imm); |
| } else { |
| __ Slt(rd, rs1, rs2); |
| } |
| if (cond == kCondGE) { |
| // Calculate `rs1 >= rhs` as `!(rs1 < rhs)` since there's only the SLT but no SGE. |
| __ Xori(rd, rd, 1); |
| } |
| break; |
| |
| case kCondLE: |
| case kCondGT: |
| if (use_imm) { |
| // Calculate `rs1 <= imm` as `rs1 < imm + 1`. |
| DCHECK(IsInt<12>(imm + 1)); // The value that overflows would fail this check. |
| __ Slti(rd, rs1, imm + 1); |
| } else { |
| __ Slt(rd, rs2, rs1); |
| } |
| if ((cond == kCondGT) == use_imm) { |
| // Calculate `rs1 > imm` as `!(rs1 < imm + 1)` and calculate |
| // `rs1 <= rs2` as `!(rs2 < rs1)` since there's only the SLT but no SGE. |
| __ Xori(rd, rd, 1); |
| } |
| break; |
| |
| case kCondB: |
| case kCondAE: |
| if (use_imm) { |
| // Sltiu sign-extends its 12-bit immediate operand before the comparison |
| // and thus lets us compare directly with unsigned values in the ranges |
| // [0, 0x7ff] and [0x[ffffffff]fffff800, 0x[ffffffff]ffffffff]. |
| DCHECK(IsInt<12>(imm)); |
| __ Sltiu(rd, rs1, imm); |
| } else { |
| __ Sltu(rd, rs1, rs2); |
| } |
| if (cond == kCondAE) { |
| // Calculate `rs1 AE rhs` as `!(rs1 B rhs)` since there's only the SLTU but no SGEU. |
| __ Xori(rd, rd, 1); |
| } |
| break; |
| |
| case kCondBE: |
| case kCondA: |
| if (use_imm) { |
| // Calculate `rs1 BE imm` as `rs1 B imm + 1`. |
| // Sltiu sign-extends its 12-bit immediate operand before the comparison |
| // and thus lets us compare directly with unsigned values in the ranges |
| // [0, 0x7ff] and [0x[ffffffff]fffff800, 0x[ffffffff]ffffffff]. |
| DCHECK(IsInt<12>(imm + 1)); // The value that overflows would fail this check. |
| __ Sltiu(rd, rs1, imm + 1); |
| } else { |
| __ Sltu(rd, rs2, rs1); |
| } |
| if ((cond == kCondA) == use_imm) { |
| // Calculate `rs1 A imm` as `!(rs1 B imm + 1)` and calculate |
| // `rs1 BE rs2` as `!(rs2 B rs1)` since there's only the SLTU but no SGEU. |
| __ Xori(rd, rd, 1); |
| } |
| break; |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateIntLongCompareAndBranch(IfCondition cond, |
| LocationSummary* locations, |
| Riscv64Label* label) { |
| XRegister left = locations->InAt(0).AsRegister<XRegister>(); |
| Location right_location = locations->InAt(1); |
| if (right_location.IsConstant()) { |
| DCHECK_EQ(CodeGenerator::GetInt64ValueOf(right_location.GetConstant()), 0); |
| switch (cond) { |
| case kCondEQ: |
| case kCondBE: // <= 0 if zero |
| __ Beqz(left, label); |
| break; |
| case kCondNE: |
| case kCondA: // > 0 if non-zero |
| __ Bnez(left, label); |
| break; |
| case kCondLT: |
| __ Bltz(left, label); |
| break; |
| case kCondGE: |
| __ Bgez(left, label); |
| break; |
| case kCondLE: |
| __ Blez(left, label); |
| break; |
| case kCondGT: |
| __ Bgtz(left, label); |
| break; |
| case kCondB: // always false |
| break; |
| case kCondAE: // always true |
| __ J(label); |
| break; |
| } |
| } else { |
| XRegister right_reg = right_location.AsRegister<XRegister>(); |
| switch (cond) { |
| case kCondEQ: |
| __ Beq(left, right_reg, label); |
| break; |
| case kCondNE: |
| __ Bne(left, right_reg, label); |
| break; |
| case kCondLT: |
| __ Blt(left, right_reg, label); |
| break; |
| case kCondGE: |
| __ Bge(left, right_reg, label); |
| break; |
| case kCondLE: |
| __ Ble(left, right_reg, label); |
| break; |
| case kCondGT: |
| __ Bgt(left, right_reg, label); |
| break; |
| case kCondB: |
| __ Bltu(left, right_reg, label); |
| break; |
| case kCondAE: |
| __ Bgeu(left, right_reg, label); |
| break; |
| case kCondBE: |
| __ Bleu(left, right_reg, label); |
| break; |
| case kCondA: |
| __ Bgtu(left, right_reg, label); |
| break; |
| } |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenerateFpCondition(IfCondition cond, |
| bool gt_bias, |
| DataType::Type type, |
| LocationSummary* locations, |
| Riscv64Label* label) { |
| // RISCV-V FP compare instructions yield the following values: |
| // l<r l=r l>r Unordered |
| // FEQ l,r 0 1 0 0 |
| // FLT l,r 1 0 0 0 |
| // FLT r,l 0 0 1 0 |
| // FLE l,r 1 1 0 0 |
| // FLE r,l 0 1 1 0 |
| // |
| // We can calculate the `Compare` results using the following formulas: |
| // l<r l=r l>r Unordered |
| // Compare/gt_bias -1 0 1 1 = ((FLE l,r) ^ 1) - (FLT l,r) |
| // Compare/lt_bias -1 0 1 -1 = ((FLE r,l) - 1) + (FLT r,l) |
| // These are emitted in `VisitCompare()`. |
| // |
| // This function emits a fused `Condition(Compare(., .), 0)`. If we compare the |
| // `Compare` results above with 0, we get the following values and formulas: |
| // l<r l=r l>r Unordered |
| // CondEQ/- 0 1 0 0 = (FEQ l, r) |
| // CondNE/- 1 0 1 1 = (FEQ l, r) ^ 1 |
| // CondLT/gt_bias 1 0 0 0 = (FLT l,r) |
| // CondLT/lt_bias 1 0 0 1 = (FLE r,l) ^ 1 |
| // CondLE/gt_bias 1 1 0 0 = (FLE l,r) |
| // CondLE/lt_bias 1 1 0 1 = (FLT r,l) ^ 1 |
| // CondGT/gt_bias 0 0 1 1 = (FLE l,r) ^ 1 |
| // CondGT/lt_bias 0 0 1 0 = (FLT r,l) |
| // CondGE/gt_bias 0 1 1 1 = (FLT l,r) ^ 1 |
| // CondGE/lt_bias 0 1 1 0 = (FLE r,l) |
| // (CondEQ/CondNE comparison with zero yields the same result with gt_bias and lt_bias.) |
| // |
| // If the condition is not materialized, the `^ 1` is not emitted, |
| // instead the condition is reversed by emitting BEQZ instead of BNEZ. |
| |
| FRegister rs1 = locations->InAt(0).AsFpuRegister<FRegister>(); |
| FRegister rs2 = locations->InAt(1).AsFpuRegister<FRegister>(); |
| |
| DCHECK_EQ(label != nullptr, locations->Out().IsInvalid()); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister rd = |
| (label != nullptr) ? srs.AllocateXRegister() : locations->Out().AsRegister<XRegister>(); |
| bool reverse_condition = false; |
| |
| switch (cond) { |
| case kCondEQ: |
| FEq(rd, rs1, rs2, type); |
| break; |
| case kCondNE: |
| FEq(rd, rs1, rs2, type); |
| reverse_condition = true; |
| break; |
| case kCondLT: |
| if (gt_bias) { |
| FLt(rd, rs1, rs2, type); |
| } else { |
| FLe(rd, rs2, rs1, type); |
| reverse_condition = true; |
| } |
| break; |
| case kCondLE: |
| if (gt_bias) { |
| FLe(rd, rs1, rs2, type); |
| } else { |
| FLt(rd, rs2, rs1, type); |
| reverse_condition = true; |
| } |
| break; |
| case kCondGT: |
| if (gt_bias) { |
| FLe(rd, rs1, rs2, type); |
| reverse_condition = true; |
| } else { |
| FLt(rd, rs2, rs1, type); |
| } |
| break; |
| case kCondGE: |
| if (gt_bias) { |
| FLt(rd, rs1, rs2, type); |
| reverse_condition = true; |
| } else { |
| FLe(rd, rs2, rs1, type); |
| } |
| break; |
| default: |
| LOG(FATAL) << "Unexpected floating-point condition " << cond; |
| UNREACHABLE(); |
| } |
| |
| if (label != nullptr) { |
| if (reverse_condition) { |
| __ Beqz(rd, label); |
| } else { |
| __ Bnez(rd, label); |
| } |
| } else { |
| if (reverse_condition) { |
| __ Xori(rd, rd, 1); |
| } |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleGoto(HInstruction* instruction, |
| HBasicBlock* successor) { |
| if (successor->IsExitBlock()) { |
| DCHECK(instruction->GetPrevious()->AlwaysThrows()); |
| return; // no code needed |
| } |
| |
| HBasicBlock* block = instruction->GetBlock(); |
| HInstruction* previous = instruction->GetPrevious(); |
| HLoopInformation* info = block->GetLoopInformation(); |
| |
| if (info != nullptr && info->IsBackEdge(*block) && info->HasSuspendCheck()) { |
| codegen_->MaybeIncrementHotness(/*is_frame_entry=*/ false); |
| GenerateSuspendCheck(info->GetSuspendCheck(), successor); |
| return; // `GenerateSuspendCheck()` emitted the jump. |
| } |
| if (block->IsEntryBlock() && previous != nullptr && previous->IsSuspendCheck()) { |
| GenerateSuspendCheck(previous->AsSuspendCheck(), nullptr); |
| } |
| if (!codegen_->GoesToNextBlock(block, successor)) { |
| __ J(codegen_->GetLabelOf(successor)); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenPackedSwitchWithCompares(XRegister adjusted, |
| XRegister temp, |
| uint32_t num_entries, |
| HBasicBlock* switch_block) { |
| // Note: The `adjusted` register holds `value - lower_bound`. If the `lower_bound` is 0, |
| // `adjusted` is the original `value` register and we must not clobber it. Otherwise, |
| // `adjusted` is the `temp`. The caller already emitted the `adjusted < num_entries` check. |
| |
| // Create a set of compare/jumps. |
| ArrayRef<HBasicBlock* const> successors(switch_block->GetSuccessors()); |
| uint32_t index = 0; |
| for (; num_entries - index >= 2u; index += 2u) { |
| // Jump to `successors[index]` if `value == lower_bound + index`. |
| // Note that `adjusted` holds `value - lower_bound - index`. |
| __ Beqz(adjusted, codegen_->GetLabelOf(successors[index])); |
| if (num_entries - index == 2u) { |
| break; // The last entry shall match, so the branch shall be unconditional. |
| } |
| // Jump to `successors[index + 1]` if `value == lower_bound + index + 1`. |
| // Modify `adjusted` to hold `value - lower_bound - index - 2` for this comparison. |
| __ Addi(temp, adjusted, -2); |
| adjusted = temp; |
| __ Bltz(adjusted, codegen_->GetLabelOf(successors[index + 1])); |
| } |
| // For the last entry, unconditionally jump to `successors[num_entries - 1]`. |
| __ J(codegen_->GetLabelOf(successors[num_entries - 1u])); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenTableBasedPackedSwitch(XRegister adjusted, |
| XRegister temp, |
| uint32_t num_entries, |
| HBasicBlock* switch_block) { |
| // Note: The `adjusted` register holds `value - lower_bound`. If the `lower_bound` is 0, |
| // `adjusted` is the original `value` register and we must not clobber it. Otherwise, |
| // `adjusted` is the `temp`. The caller already emitted the `adjusted < num_entries` check. |
| |
| // Create a jump table. |
| ArenaVector<Riscv64Label*> labels(num_entries, |
| __ GetAllocator()->Adapter(kArenaAllocSwitchTable)); |
| const ArenaVector<HBasicBlock*>& successors = switch_block->GetSuccessors(); |
| for (uint32_t i = 0; i < num_entries; i++) { |
| labels[i] = codegen_->GetLabelOf(successors[i]); |
| } |
| JumpTable* table = __ CreateJumpTable(std::move(labels)); |
| |
| // Load the address of the jump table. |
| // Note: The `LoadLabelAddress()` emits AUIPC+ADD. It is possible to avoid the ADD and |
| // instead embed that offset in the LW below as well as all jump table entries but |
| // that would need some invasive changes in the jump table handling in the assembler. |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister table_base = srs.AllocateXRegister(); |
| __ LoadLabelAddress(table_base, table->GetLabel()); |
| |
| // Load the PC difference from the jump table. |
| // TODO(riscv64): Use SH2ADD from the Zba extension. |
| __ Slli(temp, adjusted, 2); |
| __ Add(temp, temp, table_base); |
| __ Lw(temp, temp, 0); |
| |
| // Compute the absolute target address by adding the table start address |
| // (the table contains offsets to targets relative to its start). |
| __ Add(temp, temp, table_base); |
| // And jump. |
| __ Jr(temp); |
| } |
| |
| int32_t InstructionCodeGeneratorRISCV64::VecAddress(LocationSummary* locations, |
| size_t size, |
| /*out*/ XRegister* adjusted_base) { |
| UNUSED(locations); |
| UNUSED(size); |
| UNUSED(adjusted_base); |
| LOG(FATAL) << "Unimplemented"; |
| UNREACHABLE(); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::GenConditionalMove(HSelect* select) { |
| UNUSED(select); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::HandleBinaryOp(HBinaryOperation* instruction) { |
| DCHECK_EQ(instruction->InputCount(), 2u); |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| DataType::Type type = instruction->GetResultType(); |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: { |
| locations->SetInAt(0, Location::RequiresRegister()); |
| HInstruction* right = instruction->InputAt(1); |
| bool can_use_imm = false; |
| if (instruction->IsMin() || instruction->IsMax()) { |
| can_use_imm = IsZeroBitPattern(instruction); |
| } else if (right->IsConstant()) { |
| int64_t imm = CodeGenerator::GetInt64ValueOf(right->AsConstant()); |
| can_use_imm = IsInt<12>(instruction->IsSub() ? -imm : imm); |
| } |
| if (can_use_imm) { |
| locations->SetInAt(1, Location::ConstantLocation(right->AsConstant())); |
| } else { |
| locations->SetInAt(1, Location::RequiresRegister()); |
| } |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| } |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetInAt(1, Location::RequiresFpuRegister()); |
| if (instruction->IsMin() || instruction->IsMax()) { |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kOutputOverlap); |
| } else { |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); |
| } |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected " << instruction->DebugName() << " type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleBinaryOp(HBinaryOperation* instruction) { |
| DataType::Type type = instruction->GetType(); |
| LocationSummary* locations = instruction->GetLocations(); |
| |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: { |
| XRegister rd = locations->Out().AsRegister<XRegister>(); |
| XRegister rs1 = locations->InAt(0).AsRegister<XRegister>(); |
| Location rs2_location = locations->InAt(1); |
| |
| bool use_imm = rs2_location.IsConstant(); |
| XRegister rs2 = use_imm ? kNoXRegister : rs2_location.AsRegister<XRegister>(); |
| int64_t imm = use_imm ? CodeGenerator::GetInt64ValueOf(rs2_location.GetConstant()) : 0; |
| |
| if (instruction->IsAnd()) { |
| if (use_imm) { |
| __ Andi(rd, rs1, imm); |
| } else { |
| __ And(rd, rs1, rs2); |
| } |
| } else if (instruction->IsOr()) { |
| if (use_imm) { |
| __ Ori(rd, rs1, imm); |
| } else { |
| __ Or(rd, rs1, rs2); |
| } |
| } else if (instruction->IsXor()) { |
| if (use_imm) { |
| __ Xori(rd, rs1, imm); |
| } else { |
| __ Xor(rd, rs1, rs2); |
| } |
| } else if (instruction->IsAdd() || instruction->IsSub()) { |
| if (type == DataType::Type::kInt32) { |
| if (use_imm) { |
| __ Addiw(rd, rs1, instruction->IsSub() ? -imm : imm); |
| } else if (instruction->IsAdd()) { |
| __ Addw(rd, rs1, rs2); |
| } else { |
| DCHECK(instruction->IsSub()); |
| __ Subw(rd, rs1, rs2); |
| } |
| } else { |
| if (use_imm) { |
| __ Addi(rd, rs1, instruction->IsSub() ? -imm : imm); |
| } else if (instruction->IsAdd()) { |
| __ Add(rd, rs1, rs2); |
| } else { |
| DCHECK(instruction->IsSub()); |
| __ Sub(rd, rs1, rs2); |
| } |
| } |
| } else if (instruction->IsMin()) { |
| DCHECK_IMPLIES(use_imm, imm == 0); |
| __ Min(rd, rs1, use_imm ? Zero : rs2); |
| } else { |
| DCHECK(instruction->IsMax()); |
| DCHECK_IMPLIES(use_imm, imm == 0); |
| __ Max(rd, rs1, use_imm ? Zero : rs2); |
| } |
| break; |
| } |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: { |
| FRegister rd = locations->Out().AsFpuRegister<FRegister>(); |
| FRegister rs1 = locations->InAt(0).AsFpuRegister<FRegister>(); |
| FRegister rs2 = locations->InAt(1).AsFpuRegister<FRegister>(); |
| if (instruction->IsAdd()) { |
| FAdd(rd, rs1, rs2, type); |
| } else if (instruction->IsSub()) { |
| FSub(rd, rs1, rs2, type); |
| } else { |
| DCHECK(instruction->IsMin() || instruction->IsMax()); |
| // If one of the operands is NaN and the other is not, riscv64 instructions FMIN/FMAX |
| // return the other operand while we want to return the NaN operand. |
| DCHECK_NE(rd, rs1); // Requested `Location::kOutputOverlap`. |
| DCHECK_NE(rd, rs2); // Requested `Location::kOutputOverlap`. |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| XRegister tmp2 = srs.AllocateXRegister(); |
| Riscv64Label done; |
| // Return `rs1` if it's NaN. |
| FClass(tmp, rs1, type); |
| __ Li(tmp2, kFClassNaNMinValue); |
| FMv(rd, rs1, type); |
| __ Bgeu(tmp, tmp2, &done); |
| // Return `rs2` if it's NaN. |
| FClass(tmp, rs2, type); |
| FMv(rd, rs2, type); |
| __ Bgeu(tmp, tmp2, &done); |
| // Calculate Min/Max for non-NaN arguments. |
| if (instruction->IsMin()) { |
| FMin(rd, rs1, rs2, type); |
| } else { |
| FMax(rd, rs1, rs2, type); |
| } |
| __ Bind(&done); |
| } |
| break; |
| } |
| default: |
| LOG(FATAL) << "Unexpected binary operation type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::HandleCondition(HCondition* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| switch (instruction->InputAt(0)->GetType()) { |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetInAt(1, Location::RequiresFpuRegister()); |
| break; |
| |
| default: { |
| locations->SetInAt(0, Location::RequiresRegister()); |
| HInstruction* rhs = instruction->InputAt(1); |
| bool use_imm = false; |
| if (rhs->IsConstant()) { |
| int64_t imm = CodeGenerator::GetInt64ValueOf(rhs->AsConstant()); |
| if (instruction->IsEmittedAtUseSite()) { |
| // For `HIf`, materialize all non-zero constants with an `HParallelMove`. |
| // Note: For certain constants and conditions, the code could be improved. |
| // For example, 2048 takes two instructions to materialize but the negative |
| // -2048 could be embedded in ADDI for EQ/NE comparison. |
| use_imm = (imm == 0); |
| } else { |
| // Constants that cannot be embedded in an instruction's 12-bit immediate shall be |
| // materialized with an `HParallelMove`. This simplifies the code and avoids cases |
| // with arithmetic overflow. Adjust the `imm` if needed for a particular instruction. |
| switch (instruction->GetCondition()) { |
| case kCondEQ: |
| case kCondNE: |
| imm = -imm; // ADDI with negative immediate (there is no SUBI). |
| break; |
| case kCondLE: |
| case kCondGT: |
| case kCondBE: |
| case kCondA: |
| imm += 1; // SLTI/SLTIU with adjusted immediate (there is no SLEI/SLEIU). |
| break; |
| default: |
| break; |
| } |
| use_imm = IsInt<12>(imm); |
| } |
| } |
| if (use_imm) { |
| locations->SetInAt(1, Location::ConstantLocation(rhs->AsConstant())); |
| } else { |
| locations->SetInAt(1, Location::RequiresRegister()); |
| } |
| break; |
| } |
| } |
| if (!instruction->IsEmittedAtUseSite()) { |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleCondition(HCondition* instruction) { |
| if (instruction->IsEmittedAtUseSite()) { |
| return; |
| } |
| |
| DataType::Type type = instruction->InputAt(0)->GetType(); |
| LocationSummary* locations = instruction->GetLocations(); |
| switch (type) { |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| GenerateFpCondition(instruction->GetCondition(), instruction->IsGtBias(), type, locations); |
| return; |
| default: |
| // Integral types and reference equality. |
| GenerateIntLongCondition(instruction->GetCondition(), locations); |
| return; |
| } |
| } |
| |
| void LocationsBuilderRISCV64::HandleShift(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsShl() || |
| instruction->IsShr() || |
| instruction->IsUShr() || |
| instruction->IsRor()); |
| |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| DataType::Type type = instruction->GetResultType(); |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: { |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetInAt(1, Location::RegisterOrConstant(instruction->InputAt(1))); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| } |
| default: |
| LOG(FATAL) << "Unexpected shift type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleShift(HBinaryOperation* instruction) { |
| DCHECK(instruction->IsShl() || |
| instruction->IsShr() || |
| instruction->IsUShr() || |
| instruction->IsRor()); |
| LocationSummary* locations = instruction->GetLocations(); |
| DataType::Type type = instruction->GetType(); |
| |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: { |
| XRegister rd = locations->Out().AsRegister<XRegister>(); |
| XRegister rs1 = locations->InAt(0).AsRegister<XRegister>(); |
| Location rs2_location = locations->InAt(1); |
| |
| if (rs2_location.IsConstant()) { |
| int64_t imm = CodeGenerator::GetInt64ValueOf(rs2_location.GetConstant()); |
| uint32_t shamt = |
| imm & (type == DataType::Type::kInt32 ? kMaxIntShiftDistance : kMaxLongShiftDistance); |
| |
| if (shamt == 0) { |
| if (rd != rs1) { |
| __ Mv(rd, rs1); |
| } |
| } else if (type == DataType::Type::kInt32) { |
| if (instruction->IsShl()) { |
| __ Slliw(rd, rs1, shamt); |
| } else if (instruction->IsShr()) { |
| __ Sraiw(rd, rs1, shamt); |
| } else if (instruction->IsUShr()) { |
| __ Srliw(rd, rs1, shamt); |
| } else { |
| DCHECK(instruction->IsRor()); |
| __ Roriw(rd, rs1, shamt); |
| } |
| } else { |
| if (instruction->IsShl()) { |
| __ Slli(rd, rs1, shamt); |
| } else if (instruction->IsShr()) { |
| __ Srai(rd, rs1, shamt); |
| } else if (instruction->IsUShr()) { |
| __ Srli(rd, rs1, shamt); |
| } else { |
| DCHECK(instruction->IsRor()); |
| __ Rori(rd, rs1, shamt); |
| } |
| } |
| } else { |
| XRegister rs2 = rs2_location.AsRegister<XRegister>(); |
| if (type == DataType::Type::kInt32) { |
| if (instruction->IsShl()) { |
| __ Sllw(rd, rs1, rs2); |
| } else if (instruction->IsShr()) { |
| __ Sraw(rd, rs1, rs2); |
| } else if (instruction->IsUShr()) { |
| __ Srlw(rd, rs1, rs2); |
| } else { |
| DCHECK(instruction->IsRor()); |
| __ Rorw(rd, rs1, rs2); |
| } |
| } else { |
| if (instruction->IsShl()) { |
| __ Sll(rd, rs1, rs2); |
| } else if (instruction->IsShr()) { |
| __ Sra(rd, rs1, rs2); |
| } else if (instruction->IsUShr()) { |
| __ Srl(rd, rs1, rs2); |
| } else { |
| DCHECK(instruction->IsRor()); |
| __ Ror(rd, rs1, rs2); |
| } |
| } |
| } |
| break; |
| } |
| default: |
| LOG(FATAL) << "Unexpected shift operation type " << type; |
| } |
| } |
| |
| void LocationsBuilderRISCV64::HandleFieldSet(HInstruction* instruction, |
| const FieldInfo& field_info) { |
| UNUSED(instruction); |
| UNUSED(field_info); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleFieldSet(HInstruction* instruction, |
| const FieldInfo& field_info, |
| bool value_can_be_null) { |
| UNUSED(instruction); |
| UNUSED(field_info); |
| UNUSED(value_can_be_null); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::HandleFieldGet(HInstruction* instruction, |
| const FieldInfo& field_info) { |
| UNUSED(instruction); |
| UNUSED(field_info); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::HandleFieldGet(HInstruction* instruction, |
| const FieldInfo& field_info) { |
| UNUSED(instruction); |
| UNUSED(field_info); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitAbove(HAbove* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitAbove(HAbove* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitAboveOrEqual(HAboveOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitAboveOrEqual(HAboveOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitAbs(HAbs* abs) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(abs); |
| switch (abs->GetResultType()) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); |
| break; |
| default: |
| LOG(FATAL) << "Unexpected abs type " << abs->GetResultType(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitAbs(HAbs* abs) { |
| LocationSummary* locations = abs->GetLocations(); |
| switch (abs->GetResultType()) { |
| case DataType::Type::kInt32: { |
| XRegister in = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| __ Sraiw(tmp, in, 31); |
| __ Xor(out, in, tmp); |
| __ Subw(out, out, tmp); |
| break; |
| } |
| case DataType::Type::kInt64: { |
| XRegister in = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| __ Srai(tmp, in, 63); |
| __ Xor(out, in, tmp); |
| __ Sub(out, out, tmp); |
| break; |
| } |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| FAbs(locations->Out().AsFpuRegister<FRegister>(), |
| locations->InAt(0).AsFpuRegister<FRegister>(), |
| abs->GetResultType()); |
| break; |
| default: |
| LOG(FATAL) << "Unexpected abs type " << abs->GetResultType(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitAdd(HAdd* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitAdd(HAdd* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitAnd(HAnd* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitAnd(HAnd* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitArrayGet(HArrayGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitArrayGet(HArrayGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitArrayLength(HArrayLength* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitArrayLength(HArrayLength* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| uint32_t offset = CodeGenerator::GetArrayLengthOffset(instruction); |
| XRegister obj = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| __ Loadwu(out, obj, offset); // Unsigned for string length; does not matter for other arrays. |
| codegen_->MaybeRecordImplicitNullCheck(instruction); |
| // Mask out compression flag from String's array length. |
| if (mirror::kUseStringCompression && instruction->IsStringLength()) { |
| __ Srli(out, out, 1u); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitArraySet(HArraySet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitArraySet(HArraySet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitBelow(HBelow* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitBelow(HBelow* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitBelowOrEqual(HBelowOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitBelowOrEqual(HBelowOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitBooleanNot(HBooleanNot* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitBooleanNot(HBooleanNot* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| __ Xori(locations->Out().AsRegister<XRegister>(), locations->InAt(0).AsRegister<XRegister>(), 1); |
| } |
| |
| void LocationsBuilderRISCV64::VisitBoundsCheck(HBoundsCheck* instruction) { |
| RegisterSet caller_saves = RegisterSet::Empty(); |
| InvokeRuntimeCallingConvention calling_convention; |
| caller_saves.Add(Location::RegisterLocation(calling_convention.GetRegisterAt(0))); |
| caller_saves.Add(Location::RegisterLocation(calling_convention.GetRegisterAt(1))); |
| LocationSummary* locations = codegen_->CreateThrowingSlowPathLocations(instruction, caller_saves); |
| |
| HInstruction* index = instruction->InputAt(0); |
| HInstruction* length = instruction->InputAt(1); |
| |
| bool const_index = false; |
| bool const_length = false; |
| |
| if (length->IsConstant()) { |
| if (index->IsConstant()) { |
| const_index = true; |
| const_length = true; |
| } else { |
| int32_t length_value = length->AsIntConstant()->GetValue(); |
| if (length_value == 0 || length_value == 1) { |
| const_length = true; |
| } |
| } |
| } else if (index->IsConstant()) { |
| int32_t index_value = index->AsIntConstant()->GetValue(); |
| if (index_value <= 0) { |
| const_index = true; |
| } |
| } |
| |
| locations->SetInAt( |
| 0, |
| const_index ? Location::ConstantLocation(index->AsConstant()) : Location::RequiresRegister()); |
| locations->SetInAt(1, |
| const_length ? Location::ConstantLocation(length->AsConstant()) : |
| Location::RequiresRegister()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitBoundsCheck(HBoundsCheck* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| Location index_loc = locations->InAt(0); |
| Location length_loc = locations->InAt(1); |
| |
| if (length_loc.IsConstant()) { |
| int32_t length = length_loc.GetConstant()->AsIntConstant()->GetValue(); |
| if (index_loc.IsConstant()) { |
| int32_t index = index_loc.GetConstant()->AsIntConstant()->GetValue(); |
| if (index < 0 || index >= length) { |
| BoundsCheckSlowPathRISCV64* slow_path = |
| new (codegen_->GetScopedAllocator()) BoundsCheckSlowPathRISCV64(instruction); |
| codegen_->AddSlowPath(slow_path); |
| __ J(slow_path->GetEntryLabel()); |
| } else { |
| // Nothing to be done. |
| } |
| return; |
| } |
| |
| BoundsCheckSlowPathRISCV64* slow_path = |
| new (codegen_->GetScopedAllocator()) BoundsCheckSlowPathRISCV64(instruction); |
| codegen_->AddSlowPath(slow_path); |
| XRegister index = index_loc.AsRegister<XRegister>(); |
| if (length == 0) { |
| __ J(slow_path->GetEntryLabel()); |
| } else { |
| DCHECK_EQ(length, 1); |
| __ Bnez(index, slow_path->GetEntryLabel()); |
| } |
| } else { |
| XRegister length = length_loc.AsRegister<XRegister>(); |
| BoundsCheckSlowPathRISCV64* slow_path = |
| new (codegen_->GetScopedAllocator()) BoundsCheckSlowPathRISCV64(instruction); |
| codegen_->AddSlowPath(slow_path); |
| if (index_loc.IsConstant()) { |
| int32_t index = index_loc.GetConstant()->AsIntConstant()->GetValue(); |
| if (index < 0) { |
| __ J(slow_path->GetEntryLabel()); |
| } else { |
| DCHECK_EQ(index, 0); |
| __ Blez(length, slow_path->GetEntryLabel()); |
| } |
| } else { |
| XRegister index = index_loc.AsRegister<XRegister>(); |
| __ Bgeu(index, length, slow_path->GetEntryLabel()); |
| } |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitBoundType([[maybe_unused]] HBoundType* instruction) { |
| // Nothing to do, this should be removed during prepare for register allocator. |
| LOG(FATAL) << "Unreachable"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitBoundType([[maybe_unused]] HBoundType* instruction) { |
| // Nothing to do, this should be removed during prepare for register allocator. |
| LOG(FATAL) << "Unreachable"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitCheckCast(HCheckCast* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitCheckCast(HCheckCast* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitClassTableGet(HClassTableGet* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitClassTableGet(HClassTableGet* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister in = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister out = locations->Out().AsRegister<XRegister>(); |
| if (instruction->GetTableKind() == HClassTableGet::TableKind::kVTable) { |
| MemberOffset method_offset = |
| mirror::Class::EmbeddedVTableEntryOffset(instruction->GetIndex(), kRiscv64PointerSize); |
| __ Loadd(out, in, method_offset.SizeValue()); |
| } else { |
| uint32_t method_offset = dchecked_integral_cast<uint32_t>( |
| ImTable::OffsetOfElement(instruction->GetIndex(), kRiscv64PointerSize)); |
| __ Loadd(out, in, mirror::Class::ImtPtrOffset(kRiscv64PointerSize).Uint32Value()); |
| __ Loadd(out, out, method_offset); |
| } |
| } |
| |
| static int32_t GetExceptionTlsOffset() { |
| return Thread::ExceptionOffset<kRiscv64PointerSize>().Int32Value(); |
| } |
| |
| void LocationsBuilderRISCV64::VisitClearException(HClearException* instruction) { |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitClearException( |
| [[maybe_unused]] HClearException* instruction) { |
| __ Stored(Zero, TR, GetExceptionTlsOffset()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitClinitCheck(HClinitCheck* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary( |
| instruction, LocationSummary::kCallOnSlowPath); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| if (instruction->HasUses()) { |
| locations->SetOut(Location::SameAsFirstInput()); |
| } |
| // Rely on the type initialization to save everything we need. |
| locations->SetCustomSlowPathCallerSaves(OneRegInReferenceOutSaveEverythingCallerSaves()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitClinitCheck(HClinitCheck* instruction) { |
| // We assume the class is not null. |
| SlowPathCodeRISCV64* slow_path = new (codegen_->GetScopedAllocator()) LoadClassSlowPathRISCV64( |
| instruction->GetLoadClass(), instruction); |
| codegen_->AddSlowPath(slow_path); |
| GenerateClassInitializationCheck(slow_path, |
| instruction->GetLocations()->InAt(0).AsRegister<XRegister>()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitCompare(HCompare* instruction) { |
| DataType::Type in_type = instruction->InputAt(0)->GetType(); |
| |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| |
| switch (in_type) { |
| case DataType::Type::kBool: |
| case DataType::Type::kUint8: |
| case DataType::Type::kInt8: |
| case DataType::Type::kUint16: |
| case DataType::Type::kInt16: |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetInAt(1, RegisterOrZeroBitPatternLocation(instruction->InputAt(1))); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetInAt(1, Location::RequiresFpuRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected type for compare operation " << in_type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitCompare(HCompare* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister result = locations->Out().AsRegister<XRegister>(); |
| DataType::Type in_type = instruction->InputAt(0)->GetType(); |
| |
| // 0 if: left == right |
| // 1 if: left > right |
| // -1 if: left < right |
| switch (in_type) { |
| case DataType::Type::kBool: |
| case DataType::Type::kUint8: |
| case DataType::Type::kInt8: |
| case DataType::Type::kUint16: |
| case DataType::Type::kInt16: |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: { |
| XRegister left = locations->InAt(0).AsRegister<XRegister>(); |
| XRegister right = InputXRegisterOrZero(locations->InAt(1)); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| __ Slt(tmp, left, right); |
| __ Slt(result, right, left); |
| __ Sub(result, result, tmp); |
| break; |
| } |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: { |
| FRegister left = locations->InAt(0).AsFpuRegister<FRegister>(); |
| FRegister right = locations->InAt(1).AsFpuRegister<FRegister>(); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| if (instruction->IsGtBias()) { |
| // ((FLE l,r) ^ 1) - (FLT l,r); see `GenerateFpCondition()`. |
| FLe(tmp, left, right, in_type); |
| FLt(result, left, right, in_type); |
| __ Xori(tmp, tmp, 1); |
| __ Sub(result, tmp, result); |
| } else { |
| // ((FLE r,l) - 1) + (FLT r,l); see `GenerateFpCondition()`. |
| FLe(tmp, right, left, in_type); |
| FLt(result, right, left, in_type); |
| __ Addi(tmp, tmp, -1); |
| __ Add(result, result, tmp); |
| } |
| break; |
| } |
| |
| default: |
| LOG(FATAL) << "Unimplemented compare type " << in_type; |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitConstructorFence(HConstructorFence* instruction) { |
| instruction->SetLocations(nullptr); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitConstructorFence( |
| [[maybe_unused]] HConstructorFence* instruction) { |
| codegen_->GenerateMemoryBarrier(MemBarrierKind::kStoreStore); |
| } |
| |
| void LocationsBuilderRISCV64::VisitCurrentMethod(HCurrentMethod* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetOut(Location::RegisterLocation(kArtMethodRegister)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitCurrentMethod( |
| [[maybe_unused]] HCurrentMethod* instruction) { |
| // Nothing to do, the method is already at its location. |
| } |
| |
| void LocationsBuilderRISCV64::VisitShouldDeoptimizeFlag(HShouldDeoptimizeFlag* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetOut(Location::RequiresRegister()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitShouldDeoptimizeFlag( |
| HShouldDeoptimizeFlag* instruction) { |
| __ Loadw(instruction->GetLocations()->Out().AsRegister<XRegister>(), |
| SP, |
| codegen_->GetStackOffsetOfShouldDeoptimizeFlag()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitDeoptimize(HDeoptimize* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) |
| LocationSummary(instruction, LocationSummary::kCallOnSlowPath); |
| InvokeRuntimeCallingConvention calling_convention; |
| RegisterSet caller_saves = RegisterSet::Empty(); |
| caller_saves.Add(Location::RegisterLocation(calling_convention.GetRegisterAt(0))); |
| locations->SetCustomSlowPathCallerSaves(caller_saves); |
| if (IsBooleanValueOrMaterializedCondition(instruction->InputAt(0))) { |
| locations->SetInAt(0, Location::RequiresRegister()); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitDeoptimize(HDeoptimize* instruction) { |
| SlowPathCodeRISCV64* slow_path = |
| deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathRISCV64>(instruction); |
| GenerateTestAndBranch(instruction, |
| /* condition_input_index= */ 0, |
| slow_path->GetEntryLabel(), |
| /* false_target= */ nullptr); |
| } |
| |
| void LocationsBuilderRISCV64::VisitDiv(HDiv* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetInAt(1, Location::RegisterOrConstant(instruction->InputAt(1))); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetInAt(1, Location::RequiresFpuRegister()); |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected div type " << instruction->GetResultType(); |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitDiv(HDiv* instruction) { |
| DataType::Type type = instruction->GetType(); |
| LocationSummary* locations = instruction->GetLocations(); |
| |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| GenerateDivRemIntegral(instruction); |
| break; |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: { |
| FRegister dst = locations->Out().AsFpuRegister<FRegister>(); |
| FRegister lhs = locations->InAt(0).AsFpuRegister<FRegister>(); |
| FRegister rhs = locations->InAt(1).AsFpuRegister<FRegister>(); |
| FDiv(dst, lhs, rhs, type); |
| break; |
| } |
| default: |
| LOG(FATAL) << "Unexpected div type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitDivZeroCheck(HDivZeroCheck* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitDivZeroCheck(HDivZeroCheck* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitDoubleConstant(HDoubleConstant* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetOut(Location::ConstantLocation(instruction)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitDoubleConstant( |
| [[maybe_unused]] HDoubleConstant* instruction) { |
| // Will be generated at use site. |
| } |
| |
| void LocationsBuilderRISCV64::VisitEqual(HEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitEqual(HEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitExit(HExit* instruction) { |
| instruction->SetLocations(nullptr); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitExit([[maybe_unused]] HExit* instruction) {} |
| |
| void LocationsBuilderRISCV64::VisitFloatConstant(HFloatConstant* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetOut(Location::ConstantLocation(instruction)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitFloatConstant( |
| [[maybe_unused]] HFloatConstant* instruction) { |
| // Will be generated at use site. |
| } |
| |
| void LocationsBuilderRISCV64::VisitGoto(HGoto* instruction) { |
| instruction->SetLocations(nullptr); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitGoto(HGoto* instruction) { |
| HandleGoto(instruction, instruction->GetSuccessor()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitGreaterThan(HGreaterThan* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitGreaterThan(HGreaterThan* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitGreaterThanOrEqual(HGreaterThanOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitGreaterThanOrEqual(HGreaterThanOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitIf(HIf* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| if (IsBooleanValueOrMaterializedCondition(instruction->InputAt(0))) { |
| locations->SetInAt(0, Location::RequiresRegister()); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitIf(HIf* instruction) { |
| HBasicBlock* true_successor = instruction->IfTrueSuccessor(); |
| HBasicBlock* false_successor = instruction->IfFalseSuccessor(); |
| Riscv64Label* true_target = codegen_->GoesToNextBlock(instruction->GetBlock(), true_successor) |
| ? nullptr |
| : codegen_->GetLabelOf(true_successor); |
| Riscv64Label* false_target = codegen_->GoesToNextBlock(instruction->GetBlock(), false_successor) |
| ? nullptr |
| : codegen_->GetLabelOf(false_successor); |
| GenerateTestAndBranch(instruction, /* condition_input_index= */ 0, true_target, false_target); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInstanceFieldGet(HInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInstanceFieldGet(HInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitInstanceFieldSet(HInstanceFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInstanceFieldSet(HInstanceFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitPredicatedInstanceFieldGet( |
| HPredicatedInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitPredicatedInstanceFieldGet( |
| HPredicatedInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitInstanceOf(HInstanceOf* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInstanceOf(HInstanceOf* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitIntConstant(HIntConstant* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| locations->SetOut(Location::ConstantLocation(instruction)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitIntConstant([[maybe_unused]] HIntConstant* instruction) { |
| // Will be generated at use site. |
| } |
| |
| void LocationsBuilderRISCV64::VisitIntermediateAddress(HIntermediateAddress* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitIntermediateAddress(HIntermediateAddress* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokeUnresolved(HInvokeUnresolved* instruction) { |
| // The trampoline uses the same calling convention as dex calling conventions, except |
| // instead of loading arg0/A0 with the target Method*, arg0/A0 will contain the method_idx. |
| HandleInvoke(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokeUnresolved(HInvokeUnresolved* instruction) { |
| codegen_->GenerateInvokeUnresolvedRuntimeCall(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokeInterface(HInvokeInterface* instruction) { |
| HandleInvoke(instruction); |
| // Use T0 as the hidden argument for `art_quick_imt_conflict_trampoline`. |
| if (instruction->GetHiddenArgumentLoadKind() == MethodLoadKind::kRecursive) { |
| instruction->GetLocations()->SetInAt(instruction->GetNumberOfArguments() - 1, |
| Location::RegisterLocation(T0)); |
| } else { |
| instruction->GetLocations()->AddTemp(Location::RegisterLocation(T0)); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokeInterface(HInvokeInterface* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister temp = locations->GetTemp(0).AsRegister<XRegister>(); |
| XRegister receiver = locations->InAt(0).AsRegister<XRegister>(); |
| int32_t class_offset = mirror::Object::ClassOffset().Int32Value(); |
| Offset entry_point = ArtMethod::EntryPointFromQuickCompiledCodeOffset(kRiscv64PointerSize); |
| |
| // /* HeapReference<Class> */ temp = receiver->klass_ |
| __ Loadwu(temp, receiver, class_offset); |
| codegen_->MaybeRecordImplicitNullCheck(instruction); |
| // Instead of simply (possibly) unpoisoning `temp` here, we should |
| // emit a read barrier for the previous class reference load. |
| // However this is not required in practice, as this is an |
| // intermediate/temporary reference and because the current |
| // concurrent copying collector keeps the from-space memory |
| // intact/accessible until the end of the marking phase (the |
| // concurrent copying collector may not in the future). |
| codegen_->MaybeUnpoisonHeapReference(temp); |
| |
| // If we're compiling baseline, update the inline cache. |
| codegen_->MaybeGenerateInlineCacheCheck(instruction, temp); |
| |
| // The register T0 is required to be used for the hidden argument in |
| // `art_quick_imt_conflict_trampoline`. |
| if (instruction->GetHiddenArgumentLoadKind() != MethodLoadKind::kRecursive && |
| instruction->GetHiddenArgumentLoadKind() != MethodLoadKind::kRuntimeCall) { |
| Location hidden_reg = instruction->GetLocations()->GetTemp(1); |
| // Load the resolved interface method in the hidden argument register T0. |
| DCHECK_EQ(T0, hidden_reg.AsRegister<XRegister>()); |
| codegen_->LoadMethod(instruction->GetHiddenArgumentLoadKind(), hidden_reg, instruction); |
| } |
| |
| __ Loadd(temp, temp, mirror::Class::ImtPtrOffset(kRiscv64PointerSize).Uint32Value()); |
| uint32_t method_offset = static_cast<uint32_t>(ImTable::OffsetOfElement( |
| instruction->GetImtIndex(), kRiscv64PointerSize)); |
| // temp = temp->GetImtEntryAt(method_offset); |
| __ Loadd(temp, temp, method_offset); |
| if (instruction->GetHiddenArgumentLoadKind() == MethodLoadKind::kRuntimeCall) { |
| // We pass the method from the IMT in case of a conflict. This will ensure |
| // we go into the runtime to resolve the actual method. |
| Location hidden_reg = instruction->GetLocations()->GetTemp(1); |
| DCHECK_EQ(T0, hidden_reg.AsRegister<XRegister>()); |
| __ Mv(hidden_reg.AsRegister<XRegister>(), temp); |
| } |
| // RA = temp->GetEntryPoint(); |
| __ Loadd(RA, temp, entry_point.Int32Value()); |
| |
| // RA(); |
| __ Jalr(RA); |
| DCHECK(!codegen_->IsLeafMethod()); |
| codegen_->RecordPcInfo(instruction, instruction->GetDexPc()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokeStaticOrDirect(HInvokeStaticOrDirect* instruction) { |
| // Explicit clinit checks triggered by static invokes must have been pruned by |
| // art::PrepareForRegisterAllocation. |
| DCHECK(!instruction->IsStaticWithExplicitClinitCheck()); |
| |
| IntrinsicLocationsBuilderRISCV64 intrinsic(GetGraph()->GetAllocator(), codegen_); |
| if (intrinsic.TryDispatch(instruction)) { |
| return; |
| } |
| |
| if (instruction->GetCodePtrLocation() == CodePtrLocation::kCallCriticalNative) { |
| CriticalNativeCallingConventionVisitorRiscv64 calling_convention_visitor( |
| /*for_register_allocation=*/ true); |
| CodeGenerator::CreateCommonInvokeLocationSummary(instruction, &calling_convention_visitor); |
| } else { |
| HandleInvoke(instruction); |
| } |
| } |
| |
| static bool TryGenerateIntrinsicCode(HInvoke* invoke, CodeGeneratorRISCV64* codegen) { |
| if (invoke->GetLocations()->Intrinsified()) { |
| IntrinsicCodeGeneratorRISCV64 intrinsic(codegen); |
| intrinsic.Dispatch(invoke); |
| return true; |
| } |
| return false; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokeStaticOrDirect( |
| HInvokeStaticOrDirect* instruction) { |
| // Explicit clinit checks triggered by static invokes must have been pruned by |
| // art::PrepareForRegisterAllocation. |
| DCHECK(!instruction->IsStaticWithExplicitClinitCheck()); |
| |
| if (TryGenerateIntrinsicCode(instruction, codegen_)) { |
| return; |
| } |
| |
| LocationSummary* locations = instruction->GetLocations(); |
| codegen_->GenerateStaticOrDirectCall( |
| instruction, locations->HasTemps() ? locations->GetTemp(0) : Location::NoLocation()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokeVirtual(HInvokeVirtual* instruction) { |
| IntrinsicLocationsBuilderRISCV64 intrinsic(GetGraph()->GetAllocator(), codegen_); |
| if (intrinsic.TryDispatch(instruction)) { |
| return; |
| } |
| |
| HandleInvoke(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokeVirtual(HInvokeVirtual* instruction) { |
| if (TryGenerateIntrinsicCode(instruction, codegen_)) { |
| return; |
| } |
| |
| codegen_->GenerateVirtualCall(instruction, instruction->GetLocations()->GetTemp(0)); |
| DCHECK(!codegen_->IsLeafMethod()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokePolymorphic(HInvokePolymorphic* instruction) { |
| HandleInvoke(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokePolymorphic(HInvokePolymorphic* instruction) { |
| codegen_->GenerateInvokePolymorphicCall(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitInvokeCustom(HInvokeCustom* instruction) { |
| HandleInvoke(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitInvokeCustom(HInvokeCustom* instruction) { |
| codegen_->GenerateInvokeCustomCall(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitLessThan(HLessThan* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLessThan(HLessThan* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitLessThanOrEqual(HLessThanOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLessThanOrEqual(HLessThanOrEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitLoadClass(HLoadClass* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLoadClass(HLoadClass* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitLoadException(HLoadException* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetOut(Location::RequiresRegister()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLoadException(HLoadException* instruction) { |
| XRegister out = instruction->GetLocations()->Out().AsRegister<XRegister>(); |
| __ Loadwu(out, TR, GetExceptionTlsOffset()); |
| } |
| |
| void LocationsBuilderRISCV64::VisitLoadMethodHandle(HLoadMethodHandle* instruction) { |
| InvokeRuntimeCallingConvention calling_convention; |
| Location loc = Location::RegisterLocation(calling_convention.GetRegisterAt(0)); |
| CodeGenerator::CreateLoadMethodHandleRuntimeCallLocationSummary(instruction, loc, loc); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLoadMethodHandle(HLoadMethodHandle* instruction) { |
| codegen_->GenerateLoadMethodHandleRuntimeCall(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitLoadMethodType(HLoadMethodType* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLoadMethodType(HLoadMethodType* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitLoadString(HLoadString* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLoadString(HLoadString* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitLongConstant(HLongConstant* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| locations->SetOut(Location::ConstantLocation(instruction)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitLongConstant( |
| [[maybe_unused]] HLongConstant* instruction) { |
| // Will be generated at use site. |
| } |
| |
| void LocationsBuilderRISCV64::VisitMax(HMax* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMax(HMax* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitMemoryBarrier(HMemoryBarrier* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMemoryBarrier(HMemoryBarrier* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitMethodEntryHook(HMethodEntryHook* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMethodEntryHook(HMethodEntryHook* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitMethodExitHook(HMethodExitHook* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMethodExitHook(HMethodExitHook* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitMin(HMin* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMin(HMin* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitMonitorOperation(HMonitorOperation* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary( |
| instruction, LocationSummary::kCallOnMainOnly); |
| InvokeRuntimeCallingConvention calling_convention; |
| locations->SetInAt(0, Location::RegisterLocation(calling_convention.GetRegisterAt(0))); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMonitorOperation(HMonitorOperation* instruction) { |
| codegen_->InvokeRuntime(instruction->IsEnter() ? kQuickLockObject : kQuickUnlockObject, |
| instruction, |
| instruction->GetDexPc()); |
| if (instruction->IsEnter()) { |
| CheckEntrypointTypes<kQuickLockObject, void, mirror::Object*>(); |
| } else { |
| CheckEntrypointTypes<kQuickUnlockObject, void, mirror::Object*>(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitMul(HMul* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetInAt(1, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetInAt(1, Location::RequiresFpuRegister()); |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected mul type " << instruction->GetResultType(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitMul(HMul* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| __ Mulw(locations->Out().AsRegister<XRegister>(), |
| locations->InAt(0).AsRegister<XRegister>(), |
| locations->InAt(1).AsRegister<XRegister>()); |
| break; |
| |
| case DataType::Type::kInt64: |
| __ Mul(locations->Out().AsRegister<XRegister>(), |
| locations->InAt(0).AsRegister<XRegister>(), |
| locations->InAt(1).AsRegister<XRegister>()); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| FMul(locations->Out().AsFpuRegister<FRegister>(), |
| locations->InAt(0).AsFpuRegister<FRegister>(), |
| locations->InAt(1).AsFpuRegister<FRegister>(), |
| instruction->GetResultType()); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected mul type " << instruction->GetResultType(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitNeg(HNeg* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| locations->SetInAt(0, Location::RequiresFpuRegister()); |
| locations->SetOut(Location::RequiresFpuRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected neg type " << instruction->GetResultType(); |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNeg(HNeg* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| __ NegW(locations->Out().AsRegister<XRegister>(), locations->InAt(0).AsRegister<XRegister>()); |
| break; |
| |
| case DataType::Type::kInt64: |
| __ Neg(locations->Out().AsRegister<XRegister>(), locations->InAt(0).AsRegister<XRegister>()); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: |
| FNeg(locations->Out().AsFpuRegister<FRegister>(), |
| locations->InAt(0).AsFpuRegister<FRegister>(), |
| instruction->GetResultType()); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected neg type " << instruction->GetResultType(); |
| UNREACHABLE(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitNewArray(HNewArray* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNewArray(HNewArray* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitNewInstance(HNewInstance* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNewInstance(HNewInstance* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitNop(HNop* instruction) { |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNop([[maybe_unused]] HNop* instruction) { |
| // The environment recording already happened in CodeGenerator::Compile. |
| } |
| |
| void LocationsBuilderRISCV64::VisitNot(HNot* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNot(HNot* instruction) { |
| LocationSummary* locations = instruction->GetLocations(); |
| switch (instruction->GetResultType()) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| __ Not(locations->Out().AsRegister<XRegister>(), locations->InAt(0).AsRegister<XRegister>()); |
| break; |
| |
| default: |
| LOG(FATAL) << "Unexpected type for not operation " << instruction->GetResultType(); |
| UNREACHABLE(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitNotEqual(HNotEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNotEqual(HNotEqual* instruction) { |
| HandleCondition(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitNullConstant(HNullConstant* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNullConstant(HNullConstant* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitNullCheck(HNullCheck* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitNullCheck(HNullCheck* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitOr(HOr* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitOr(HOr* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitPackedSwitch(HPackedSwitch* instruction) { |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, LocationSummary::kNoCall); |
| locations->SetInAt(0, Location::RequiresRegister()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitPackedSwitch(HPackedSwitch* instruction) { |
| int32_t lower_bound = instruction->GetStartValue(); |
| uint32_t num_entries = instruction->GetNumEntries(); |
| LocationSummary* locations = instruction->GetLocations(); |
| XRegister value = locations->InAt(0).AsRegister<XRegister>(); |
| HBasicBlock* switch_block = instruction->GetBlock(); |
| HBasicBlock* default_block = instruction->GetDefaultBlock(); |
| |
| // Prepare a temporary register and an adjusted zero-based value. |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister temp = srs.AllocateXRegister(); |
| XRegister adjusted = value; |
| if (lower_bound != 0) { |
| adjusted = temp; |
| __ AddConst32(temp, value, -lower_bound); |
| } |
| |
| // Jump to the default block if the index is out of the packed switch value range. |
| // Note: We could save one instruction for `num_entries == 1` with BNEZ but the |
| // `HInstructionBuilder` transforms that case to an `HIf`, so let's keep the code simple. |
| CHECK_NE(num_entries, 0u); // `HInstructionBuilder` creates a `HGoto` for empty packed-switch. |
| { |
| ScratchRegisterScope srs2(GetAssembler()); |
| XRegister temp2 = srs2.AllocateXRegister(); |
| __ LoadConst32(temp2, num_entries); |
| __ Bgeu(adjusted, temp2, codegen_->GetLabelOf(default_block)); // Can clobber `TMP` if taken. |
| } |
| |
| if (num_entries >= kPackedSwitchCompareJumpThreshold) { |
| GenTableBasedPackedSwitch(adjusted, temp, num_entries, switch_block); |
| } else { |
| GenPackedSwitchWithCompares(adjusted, temp, num_entries, switch_block); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitParallelMove([[maybe_unused]] HParallelMove* instruction) { |
| LOG(FATAL) << "Unreachable"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitParallelMove(HParallelMove* instruction) { |
| if (instruction->GetNext()->IsSuspendCheck() && |
| instruction->GetBlock()->GetLoopInformation() != nullptr) { |
| HSuspendCheck* suspend_check = instruction->GetNext()->AsSuspendCheck(); |
| // The back edge will generate the suspend check. |
| codegen_->ClearSpillSlotsFromLoopPhisInStackMap(suspend_check, instruction); |
| } |
| |
| codegen_->GetMoveResolver()->EmitNativeCode(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitParameterValue(HParameterValue* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| Location location = parameter_visitor_.GetNextLocation(instruction->GetType()); |
| if (location.IsStackSlot()) { |
| location = Location::StackSlot(location.GetStackIndex() + codegen_->GetFrameSize()); |
| } else if (location.IsDoubleStackSlot()) { |
| location = Location::DoubleStackSlot(location.GetStackIndex() + codegen_->GetFrameSize()); |
| } |
| locations->SetOut(location); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitParameterValue( |
| [[maybe_unused]] HParameterValue* instruction) { |
| // Nothing to do, the parameter is already at its location. |
| } |
| |
| void LocationsBuilderRISCV64::VisitPhi(HPhi* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| for (size_t i = 0, e = locations->GetInputCount(); i < e; ++i) { |
| locations->SetInAt(i, Location::Any()); |
| } |
| locations->SetOut(Location::Any()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitPhi([[maybe_unused]] HPhi* instruction) { |
| LOG(FATAL) << "Unreachable"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitRem(HRem* instruction) { |
| DataType::Type type = instruction->GetResultType(); |
| LocationSummary::CallKind call_kind = |
| DataType::IsFloatingPointType(type) ? LocationSummary::kCallOnMainOnly |
| : LocationSummary::kNoCall; |
| LocationSummary* locations = |
| new (GetGraph()->GetAllocator()) LocationSummary(instruction, call_kind); |
| |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| locations->SetInAt(0, Location::RequiresRegister()); |
| locations->SetInAt(1, Location::RegisterOrConstant(instruction->InputAt(1))); |
| locations->SetOut(Location::RequiresRegister(), Location::kNoOutputOverlap); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: { |
| InvokeRuntimeCallingConvention calling_convention; |
| locations->SetInAt(0, Location::FpuRegisterLocation(calling_convention.GetFpuRegisterAt(0))); |
| locations->SetInAt(1, Location::FpuRegisterLocation(calling_convention.GetFpuRegisterAt(1))); |
| locations->SetOut(calling_convention.GetReturnLocation(type)); |
| break; |
| } |
| |
| default: |
| LOG(FATAL) << "Unexpected rem type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitRem(HRem* instruction) { |
| DataType::Type type = instruction->GetType(); |
| |
| switch (type) { |
| case DataType::Type::kInt32: |
| case DataType::Type::kInt64: |
| GenerateDivRemIntegral(instruction); |
| break; |
| |
| case DataType::Type::kFloat32: |
| case DataType::Type::kFloat64: { |
| QuickEntrypointEnum entrypoint = |
| (type == DataType::Type::kFloat32) ? kQuickFmodf : kQuickFmod; |
| codegen_->InvokeRuntime(entrypoint, instruction, instruction->GetDexPc()); |
| if (type == DataType::Type::kFloat32) { |
| CheckEntrypointTypes<kQuickFmodf, float, float, float>(); |
| } else { |
| CheckEntrypointTypes<kQuickFmod, double, double, double>(); |
| } |
| break; |
| } |
| default: |
| LOG(FATAL) << "Unexpected rem type " << type; |
| UNREACHABLE(); |
| } |
| } |
| |
| void LocationsBuilderRISCV64::VisitReturn(HReturn* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(instruction); |
| DataType::Type return_type = instruction->InputAt(0)->GetType(); |
| DCHECK_NE(return_type, DataType::Type::kVoid); |
| locations->SetInAt(0, Riscv64ReturnLocation(return_type)); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitReturn(HReturn* instruction) { |
| if (GetGraph()->IsCompilingOsr()) { |
| // To simplify callers of an OSR method, we put a floating point return value |
| // in both floating point and core return registers. |
| switch (instruction->InputAt(0)->GetType()) { |
| case DataType::Type::kFloat32: |
| __ FMvXW(A0, FA0); |
| break; |
| case DataType::Type::kFloat64: |
| __ FMvXD(A0, FA0); |
| break; |
| default: |
| break; |
| } |
| } |
| codegen_->GenerateFrameExit(); |
| } |
| |
| void LocationsBuilderRISCV64::VisitReturnVoid(HReturnVoid* instruction) { |
| instruction->SetLocations(nullptr); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitReturnVoid([[maybe_unused]] HReturnVoid* instruction) { |
| codegen_->GenerateFrameExit(); |
| } |
| |
| void LocationsBuilderRISCV64::VisitRor(HRor* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitRor(HRor* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitShl(HShl* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitShl(HShl* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitShr(HShr* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitShr(HShr* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitStaticFieldGet(HStaticFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitStaticFieldGet(HStaticFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitStaticFieldSet(HStaticFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitStaticFieldSet(HStaticFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitStringBuilderAppend(HStringBuilderAppend* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitStringBuilderAppend(HStringBuilderAppend* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitUnresolvedInstanceFieldGet( |
| HUnresolvedInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitUnresolvedInstanceFieldGet( |
| HUnresolvedInstanceFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitUnresolvedInstanceFieldSet( |
| HUnresolvedInstanceFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitUnresolvedInstanceFieldSet( |
| HUnresolvedInstanceFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitUnresolvedStaticFieldGet( |
| HUnresolvedStaticFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitUnresolvedStaticFieldGet( |
| HUnresolvedStaticFieldGet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitUnresolvedStaticFieldSet( |
| HUnresolvedStaticFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitUnresolvedStaticFieldSet( |
| HUnresolvedStaticFieldSet* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitSelect(HSelect* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitSelect(HSelect* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitSub(HSub* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitSub(HSub* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitSuspendCheck(HSuspendCheck* instruction) { |
| LocationSummary* locations = new (GetGraph()->GetAllocator()) |
| LocationSummary(instruction, LocationSummary::kCallOnSlowPath); |
| // In suspend check slow path, usually there are no caller-save registers at all. |
| // If SIMD instructions are present, however, we force spilling all live SIMD |
| // registers in full width (since the runtime only saves/restores lower part). |
| locations->SetCustomSlowPathCallerSaves(GetGraph()->HasSIMD() ? RegisterSet::AllFpu() : |
| RegisterSet::Empty()); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitSuspendCheck(HSuspendCheck* instruction) { |
| HBasicBlock* block = instruction->GetBlock(); |
| if (block->GetLoopInformation() != nullptr) { |
| DCHECK(block->GetLoopInformation()->GetSuspendCheck() == instruction); |
| // The back edge will generate the suspend check. |
| return; |
| } |
| if (block->IsEntryBlock() && instruction->GetNext()->IsGoto()) { |
| // The goto will generate the suspend check. |
| return; |
| } |
| GenerateSuspendCheck(instruction, nullptr); |
| } |
| |
| void LocationsBuilderRISCV64::VisitThrow(HThrow* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitThrow(HThrow* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitTryBoundary(HTryBoundary* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitTryBoundary(HTryBoundary* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitTypeConversion(HTypeConversion* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitTypeConversion(HTypeConversion* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitUShr(HUShr* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitUShr(HUShr* instruction) { |
| HandleShift(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitXor(HXor* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitXor(HXor* instruction) { |
| HandleBinaryOp(instruction); |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecReplicateScalar(HVecReplicateScalar* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecReplicateScalar(HVecReplicateScalar* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecExtractScalar(HVecExtractScalar* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecExtractScalar(HVecExtractScalar* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecReduce(HVecReduce* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecReduce(HVecReduce* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecCnv(HVecCnv* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecCnv(HVecCnv* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecNeg(HVecNeg* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecNeg(HVecNeg* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecAbs(HVecAbs* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecAbs(HVecAbs* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecNot(HVecNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecNot(HVecNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecAdd(HVecAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecAdd(HVecAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecHalvingAdd(HVecHalvingAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecHalvingAdd(HVecHalvingAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecSub(HVecSub* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecSub(HVecSub* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecMul(HVecMul* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecMul(HVecMul* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecDiv(HVecDiv* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecDiv(HVecDiv* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecMin(HVecMin* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecMin(HVecMin* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecMax(HVecMax* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecMax(HVecMax* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecAnd(HVecAnd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecAnd(HVecAnd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecAndNot(HVecAndNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecAndNot(HVecAndNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecOr(HVecOr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecOr(HVecOr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecXor(HVecXor* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecXor(HVecXor* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecSaturationAdd(HVecSaturationAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecSaturationAdd(HVecSaturationAdd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecSaturationSub(HVecSaturationSub* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecSaturationSub(HVecSaturationSub* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecShl(HVecShl* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecShl(HVecShl* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecShr(HVecShr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecShr(HVecShr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecUShr(HVecUShr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecUShr(HVecUShr* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecSetScalars(HVecSetScalars* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecSetScalars(HVecSetScalars* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecMultiplyAccumulate(HVecMultiplyAccumulate* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecMultiplyAccumulate( |
| HVecMultiplyAccumulate* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecSADAccumulate(HVecSADAccumulate* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecSADAccumulate(HVecSADAccumulate* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecDotProd(HVecDotProd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecDotProd(HVecDotProd* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecLoad(HVecLoad* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecLoad(HVecLoad* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecStore(HVecStore* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecStore(HVecStore* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecPredSetAll(HVecPredSetAll* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecPredSetAll(HVecPredSetAll* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecPredWhile(HVecPredWhile* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecPredWhile(HVecPredWhile* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecPredToBoolean(HVecPredToBoolean* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecPredToBoolean(HVecPredToBoolean* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecCondition(HVecCondition* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecCondition(HVecCondition* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void LocationsBuilderRISCV64::VisitVecPredNot(HVecPredNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| void InstructionCodeGeneratorRISCV64::VisitVecPredNot(HVecPredNot* instruction) { |
| UNUSED(instruction); |
| LOG(FATAL) << "Unimplemented"; |
| } |
| |
| namespace detail { |
| |
| // Mark which intrinsics we don't have handcrafted code for. |
| template <Intrinsics T> |
| struct IsUnimplemented { |
| bool is_unimplemented = false; |
| }; |
| |
| #define TRUE_OVERRIDE(Name) \ |
| template <> \ |
| struct IsUnimplemented<Intrinsics::k##Name> { \ |
| bool is_unimplemented = true; \ |
| }; |
| UNIMPLEMENTED_INTRINSIC_LIST_RISCV64(TRUE_OVERRIDE) |
| #undef TRUE_OVERRIDE |
| |
| static constexpr bool kIsIntrinsicUnimplemented[] = { |
| false, // kNone |
| #define IS_UNIMPLEMENTED(Intrinsic, ...) \ |
| IsUnimplemented<Intrinsics::k##Intrinsic>().is_unimplemented, |
| ART_INTRINSICS_LIST(IS_UNIMPLEMENTED) |
| #undef IS_UNIMPLEMENTED |
| }; |
| |
| } // namespace detail |
| |
| CodeGeneratorRISCV64::CodeGeneratorRISCV64(HGraph* graph, |
| const CompilerOptions& compiler_options, |
| OptimizingCompilerStats* stats) |
| : CodeGenerator(graph, |
| kNumberOfXRegisters, |
| kNumberOfFRegisters, |
| /*number_of_register_pairs=*/ 0u, |
| ComputeRegisterMask(kCoreCalleeSaves, arraysize(kCoreCalleeSaves)), |
| ComputeRegisterMask(kFpuCalleeSaves, arraysize(kFpuCalleeSaves)), |
| compiler_options, |
| stats, |
| ArrayRef<const bool>(detail::kIsIntrinsicUnimplemented)), |
| assembler_(graph->GetAllocator(), |
| compiler_options.GetInstructionSetFeatures()->AsRiscv64InstructionSetFeatures()), |
| location_builder_(graph, this), |
| instruction_visitor_(graph, this), |
| block_labels_(nullptr), |
| move_resolver_(graph->GetAllocator(), this), |
| uint32_literals_(std::less<uint32_t>(), |
| graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| uint64_literals_(std::less<uint64_t>(), |
| graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| boot_image_method_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| method_bss_entry_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| boot_image_type_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| type_bss_entry_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| public_type_bss_entry_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| package_type_bss_entry_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| boot_image_string_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| string_bss_entry_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| boot_image_jni_entrypoint_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)), |
| boot_image_other_patches_(graph->GetAllocator()->Adapter(kArenaAllocCodeGenerator)) { |
| // Always mark the RA register to be saved. |
| AddAllocatedRegister(Location::RegisterLocation(RA)); |
| } |
| |
| void CodeGeneratorRISCV64::MaybeIncrementHotness(bool is_frame_entry) { |
| if (GetCompilerOptions().CountHotnessInCompiledCode()) { |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister method = is_frame_entry ? kArtMethodRegister : srs.AllocateXRegister(); |
| if (!is_frame_entry) { |
| __ Loadd(method, SP, 0); |
| } |
| XRegister counter = srs.AllocateXRegister(); |
| __ Loadhu(counter, method, ArtMethod::HotnessCountOffset().Int32Value()); |
| Riscv64Label done; |
| DCHECK_EQ(0u, interpreter::kNterpHotnessValue); |
| __ Beqz(counter, &done); // Can clobber `TMP` if taken. |
| __ Addi(counter, counter, -1); |
| // We may not have another scratch register available for `Storeh`()`, |
| // so we must use the `Sh()` function directly. |
| static_assert(IsInt<12>(ArtMethod::HotnessCountOffset().Int32Value())); |
| __ Sh(counter, method, ArtMethod::HotnessCountOffset().Int32Value()); |
| __ Bind(&done); |
| } |
| |
| if (GetGraph()->IsCompilingBaseline() && !Runtime::Current()->IsAotCompiler()) { |
| SlowPathCodeRISCV64* slow_path = new (GetScopedAllocator()) CompileOptimizedSlowPathRISCV64(); |
| AddSlowPath(slow_path); |
| ProfilingInfo* info = GetGraph()->GetProfilingInfo(); |
| DCHECK(info != nullptr); |
| DCHECK(!HasEmptyFrame()); |
| uint64_t address = reinterpret_cast64<uint64_t>(info); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| __ LoadConst64(tmp, address); |
| XRegister counter = srs.AllocateXRegister(); |
| __ Loadhu(counter, tmp, ProfilingInfo::BaselineHotnessCountOffset().Int32Value()); |
| __ Beqz(counter, slow_path->GetEntryLabel()); // Can clobber `TMP` if taken. |
| __ Addi(counter, counter, -1); |
| // We do not have another scratch register available for `Storeh`()`, |
| // so we must use the `Sh()` function directly. |
| static_assert(IsInt<12>(ProfilingInfo::BaselineHotnessCountOffset().Int32Value())); |
| __ Sh(counter, tmp, ProfilingInfo::BaselineHotnessCountOffset().Int32Value()); |
| __ Bind(slow_path->GetExitLabel()); |
| } |
| } |
| |
| bool CodeGeneratorRISCV64::CanUseImplicitSuspendCheck() const { |
| // TODO(riscv64): Implement implicit suspend checks to reduce code size. |
| return false; |
| } |
| |
| void CodeGeneratorRISCV64::GenerateMemoryBarrier(MemBarrierKind kind) { |
| switch (kind) { |
| case MemBarrierKind::kAnyAny: |
| case MemBarrierKind::kAnyStore: |
| case MemBarrierKind::kLoadAny: |
| case MemBarrierKind::kStoreStore: { |
| // TODO(riscv64): Use more specific fences. |
| __ Fence(); |
| break; |
| } |
| |
| default: |
| LOG(FATAL) << "Unexpected memory barrier " << kind; |
| UNREACHABLE(); |
| } |
| } |
| |
| void CodeGeneratorRISCV64::GenerateFrameEntry() { |
| // Check if we need to generate the clinit check. We will jump to the |
| // resolution stub if the class is not initialized and the executing thread is |
| // not the thread initializing it. |
| // We do this before constructing the frame to get the correct stack trace if |
| // an exception is thrown. |
| if (GetCompilerOptions().ShouldCompileWithClinitCheck(GetGraph()->GetArtMethod())) { |
| Riscv64Label resolution; |
| Riscv64Label memory_barrier; |
| |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| XRegister tmp2 = srs.AllocateXRegister(); |
| |
| // We don't emit a read barrier here to save on code size. We rely on the |
| // resolution trampoline to do a clinit check before re-entering this code. |
| __ Loadwu(tmp2, kArtMethodRegister, ArtMethod::DeclaringClassOffset().Int32Value()); |
| |
| // We shall load the full 32-bit status word with sign-extension and compare as unsigned |
| // to sign-extended shifted status values. This yields the same comparison as loading and |
| // materializing unsigned but the constant is materialized with a single LUI instruction. |
| __ Loadw(tmp, tmp2, mirror::Class::StatusOffset().SizeValue()); // Sign-extended. |
| |
| // Check if we're visibly initialized. |
| __ Li(tmp2, ShiftedSignExtendedClassStatusValue<ClassStatus::kVisiblyInitialized>()); |
| __ Bgeu(tmp, tmp2, &frame_entry_label_); // Can clobber `TMP` if taken. |
| |
| // Check if we're initialized and jump to code that does a memory barrier if so. |
| __ Li(tmp2, ShiftedSignExtendedClassStatusValue<ClassStatus::kInitialized>()); |
| __ Bgeu(tmp, tmp2, &memory_barrier); // Can clobber `TMP` if taken. |
| |
| // Check if we're initializing and the thread initializing is the one |
| // executing the code. |
| __ Li(tmp2, ShiftedSignExtendedClassStatusValue<ClassStatus::kInitializing>()); |
| __ Bltu(tmp, tmp2, &resolution); // Can clobber `TMP` if taken. |
| |
| __ Loadwu(tmp2, kArtMethodRegister, ArtMethod::DeclaringClassOffset().Int32Value()); |
| __ Loadw(tmp, tmp2, mirror::Class::ClinitThreadIdOffset().Int32Value()); |
| __ Loadw(tmp2, TR, Thread::TidOffset<kRiscv64PointerSize>().Int32Value()); |
| __ Beq(tmp, tmp2, &frame_entry_label_); |
| __ Bind(&resolution); |
| |
| // Jump to the resolution stub. |
| ThreadOffset64 entrypoint_offset = |
| GetThreadOffset<kRiscv64PointerSize>(kQuickQuickResolutionTrampoline); |
| __ Loadd(tmp, TR, entrypoint_offset.Int32Value()); |
| __ Jr(tmp); |
| |
| __ Bind(&memory_barrier); |
| GenerateMemoryBarrier(MemBarrierKind::kAnyAny); |
| } |
| __ Bind(&frame_entry_label_); |
| |
| bool do_overflow_check = |
| FrameNeedsStackCheck(GetFrameSize(), InstructionSet::kRiscv64) || !IsLeafMethod(); |
| |
| if (do_overflow_check) { |
| DCHECK(GetCompilerOptions().GetImplicitStackOverflowChecks()); |
| __ Loadw( |
| Zero, SP, -static_cast<int32_t>(GetStackOverflowReservedBytes(InstructionSet::kRiscv64))); |
| RecordPcInfo(nullptr, 0); |
| } |
| |
| if (!HasEmptyFrame()) { |
| // Make sure the frame size isn't unreasonably large. |
| if (GetFrameSize() > GetStackOverflowReservedBytes(InstructionSet::kRiscv64)) { |
| LOG(FATAL) << "Stack frame larger than " |
| << GetStackOverflowReservedBytes(InstructionSet::kRiscv64) << " bytes"; |
| } |
| |
| // Spill callee-saved registers. |
| |
| uint32_t frame_size = GetFrameSize(); |
| |
| IncreaseFrame(frame_size); |
| |
| uint32_t offset = frame_size; |
| for (size_t i = arraysize(kCoreCalleeSaves); i != 0; ) { |
| --i; |
| XRegister reg = kCoreCalleeSaves[i]; |
| if (allocated_registers_.ContainsCoreRegister(reg)) { |
| offset -= kRiscv64DoublewordSize; |
| __ Stored(reg, SP, offset); |
| __ cfi().RelOffset(dwarf::Reg::Riscv64Core(reg), offset); |
| } |
| } |
| |
| for (size_t i = arraysize(kFpuCalleeSaves); i != 0; ) { |
| --i; |
| FRegister reg = kFpuCalleeSaves[i]; |
| if (allocated_registers_.ContainsFloatingPointRegister(reg)) { |
| offset -= kRiscv64DoublewordSize; |
| __ FStored(reg, SP, offset); |
| __ cfi().RelOffset(dwarf::Reg::Riscv64Fp(reg), offset); |
| } |
| } |
| |
| // Save the current method if we need it. Note that we do not |
| // do this in HCurrentMethod, as the instruction might have been removed |
| // in the SSA graph. |
| if (RequiresCurrentMethod()) { |
| __ Stored(kArtMethodRegister, SP, 0); |
| } |
| |
| if (GetGraph()->HasShouldDeoptimizeFlag()) { |
| // Initialize should_deoptimize flag to 0. |
| __ Storew(Zero, SP, GetStackOffsetOfShouldDeoptimizeFlag()); |
| } |
| } |
| MaybeIncrementHotness(/*is_frame_entry=*/ true); |
| } |
| |
| void CodeGeneratorRISCV64::GenerateFrameExit() { |
| __ cfi().RememberState(); |
| |
| if (!HasEmptyFrame()) { |
| // Restore callee-saved registers. |
| |
| // For better instruction scheduling restore RA before other registers. |
| uint32_t offset = GetFrameSize(); |
| for (size_t i = arraysize(kCoreCalleeSaves); i != 0; ) { |
| --i; |
| XRegister reg = kCoreCalleeSaves[i]; |
| if (allocated_registers_.ContainsCoreRegister(reg)) { |
| offset -= kRiscv64DoublewordSize; |
| __ Loadd(reg, SP, offset); |
| __ cfi().Restore(dwarf::Reg::Riscv64Core(reg)); |
| } |
| } |
| |
| for (size_t i = arraysize(kFpuCalleeSaves); i != 0; ) { |
| --i; |
| FRegister reg = kFpuCalleeSaves[i]; |
| if (allocated_registers_.ContainsFloatingPointRegister(reg)) { |
| offset -= kRiscv64DoublewordSize; |
| __ FLoadd(reg, SP, offset); |
| __ cfi().Restore(dwarf::Reg::Riscv64Fp(reg)); |
| } |
| } |
| |
| DecreaseFrame(GetFrameSize()); |
| } |
| |
| __ Jr(RA); |
| |
| __ cfi().RestoreState(); |
| __ cfi().DefCFAOffset(GetFrameSize()); |
| } |
| |
| void CodeGeneratorRISCV64::Bind(HBasicBlock* block) { __ Bind(GetLabelOf(block)); } |
| |
| void CodeGeneratorRISCV64::MoveConstant(Location destination, int32_t value) { |
| DCHECK(destination.IsRegister()); |
| __ LoadConst32(destination.AsRegister<XRegister>(), value); |
| } |
| |
| void CodeGeneratorRISCV64::MoveLocation(Location destination, |
| Location source, |
| DataType::Type dst_type) { |
| if (source.Equals(destination)) { |
| return; |
| } |
| |
| // A valid move type can always be inferred from the destination and source locations. |
| // When moving from and to a register, the `dst_type` can be used to generate 32-bit instead |
| // of 64-bit moves but it's generally OK to use 64-bit moves for 32-bit values in registers. |
| bool unspecified_type = (dst_type == DataType::Type::kVoid); |
| // TODO(riscv64): Is the destination type known in all cases? |
| // TODO(riscv64): Can unspecified `dst_type` move 32-bit GPR to FPR without NaN-boxing? |
| CHECK(!unspecified_type); |
| |
| if (destination.IsRegister() || destination.IsFpuRegister()) { |
| if (unspecified_type) { |
| HConstant* src_cst = source.IsConstant() ? source.GetConstant() : nullptr; |
| if (source.IsStackSlot() || |
| (src_cst != nullptr && |
| (src_cst->IsIntConstant() || src_cst->IsFloatConstant() || src_cst->IsNullConstant()))) { |
| // For stack slots and 32-bit constants, a 32-bit type is appropriate. |
| dst_type = destination.IsRegister() ? DataType::Type::kInt32 : DataType::Type::kFloat32; |
| } else { |
| // If the source is a double stack slot or a 64-bit constant, a 64-bit type |
| // is appropriate. Else the source is a register, and since the type has not |
| // been specified, we chose a 64-bit type to force a 64-bit move. |
| dst_type = destination.IsRegister() ? DataType::Type::kInt64 : DataType::Type::kFloat64; |
| } |
| } |
| DCHECK((destination.IsFpuRegister() && DataType::IsFloatingPointType(dst_type)) || |
| (destination.IsRegister() && !DataType::IsFloatingPointType(dst_type))); |
| |
| if (source.IsStackSlot() || source.IsDoubleStackSlot()) { |
| // Move to GPR/FPR from stack |
| if (DataType::IsFloatingPointType(dst_type)) { |
| if (DataType::Is64BitType(dst_type)) { |
| __ FLoadd(destination.AsFpuRegister<FRegister>(), SP, source.GetStackIndex()); |
| } else { |
| __ FLoadw(destination.AsFpuRegister<FRegister>(), SP, source.GetStackIndex()); |
| } |
| } else { |
| if (DataType::Is64BitType(dst_type)) { |
| __ Loadd(destination.AsRegister<XRegister>(), SP, source.GetStackIndex()); |
| } else if (dst_type == DataType::Type::kReference) { |
| __ Loadwu(destination.AsRegister<XRegister>(), SP, source.GetStackIndex()); |
| } else { |
| __ Loadw(destination.AsRegister<XRegister>(), SP, source.GetStackIndex()); |
| } |
| } |
| } else if (source.IsConstant()) { |
| // Move to GPR/FPR from constant |
| // TODO(riscv64): Consider using literals for difficult-to-materialize 64-bit constants. |
| int64_t value = GetInt64ValueOf(source.GetConstant()->AsConstant()); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister gpr = DataType::IsFloatingPointType(dst_type) |
| ? srs.AllocateXRegister() |
| : destination.AsRegister<XRegister>(); |
| if (DataType::IsFloatingPointType(dst_type) && value == 0) { |
| gpr = Zero; // Note: The scratch register allocated above shall not be used. |
| } else { |
| // Note: For `float` we load the sign-extended value here as it can sometimes yield |
| // a shorter instruction sequence. The higher 32 bits shall be ignored during the |
| // transfer to FP reg and the result shall be correctly NaN-boxed. |
| __ LoadConst64(gpr, value); |
| } |
| if (dst_type == DataType::Type::kFloat32) { |
| __ FMvWX(destination.AsFpuRegister<FRegister>(), gpr); |
| } else if (dst_type == DataType::Type::kFloat64) { |
| __ FMvDX(destination.AsFpuRegister<FRegister>(), gpr); |
| } |
| } else if (source.IsRegister()) { |
| if (destination.IsRegister()) { |
| // Move to GPR from GPR |
| __ Mv(destination.AsRegister<XRegister>(), source.AsRegister<XRegister>()); |
| } else { |
| DCHECK(destination.IsFpuRegister()); |
| if (DataType::Is64BitType(dst_type)) { |
| __ FMvDX(destination.AsFpuRegister<FRegister>(), source.AsRegister<XRegister>()); |
| } else { |
| __ FMvWX(destination.AsFpuRegister<FRegister>(), source.AsRegister<XRegister>()); |
| } |
| } |
| } else if (source.IsFpuRegister()) { |
| if (destination.IsFpuRegister()) { |
| if (GetGraph()->HasSIMD()) { |
| LOG(FATAL) << "Vector extension is unsupported"; |
| UNREACHABLE(); |
| } else { |
| // Move to FPR from FPR |
| if (dst_type == DataType::Type::kFloat32) { |
| __ FMvS(destination.AsFpuRegister<FRegister>(), source.AsFpuRegister<FRegister>()); |
| } else { |
| DCHECK_EQ(dst_type, DataType::Type::kFloat64); |
| __ FMvD(destination.AsFpuRegister<FRegister>(), source.AsFpuRegister<FRegister>()); |
| } |
| } |
| } else { |
| DCHECK(destination.IsRegister()); |
| if (DataType::Is64BitType(dst_type)) { |
| __ FMvXD(destination.AsRegister<XRegister>(), source.AsFpuRegister<FRegister>()); |
| } else { |
| __ FMvXW(destination.AsRegister<XRegister>(), source.AsFpuRegister<FRegister>()); |
| } |
| } |
| } |
| } else if (destination.IsSIMDStackSlot()) { |
| LOG(FATAL) << "SIMD is unsupported"; |
| UNREACHABLE(); |
| } else { // The destination is not a register. It must be a stack slot. |
| DCHECK(destination.IsStackSlot() || destination.IsDoubleStackSlot()); |
| if (source.IsRegister() || source.IsFpuRegister()) { |
| if (unspecified_type) { |
| if (source.IsRegister()) { |
| dst_type = destination.IsStackSlot() ? DataType::Type::kInt32 : DataType::Type::kInt64; |
| } else { |
| dst_type = |
| destination.IsStackSlot() ? DataType::Type::kFloat32 : DataType::Type::kFloat64; |
| } |
| } |
| DCHECK((destination.IsDoubleStackSlot() == DataType::Is64BitType(dst_type)) && |
| (source.IsFpuRegister() == DataType::IsFloatingPointType(dst_type))); |
| // Move to stack from GPR/FPR |
| if (DataType::Is64BitType(dst_type)) { |
| if (source.IsRegister()) { |
| __ Stored(source.AsRegister<XRegister>(), SP, destination.GetStackIndex()); |
| } else { |
| __ FStored(source.AsFpuRegister<FRegister>(), SP, destination.GetStackIndex()); |
| } |
| } else { |
| if (source.IsRegister()) { |
| __ Storew(source.AsRegister<XRegister>(), SP, destination.GetStackIndex()); |
| } else { |
| __ FStorew(source.AsFpuRegister<FRegister>(), SP, destination.GetStackIndex()); |
| } |
| } |
| } else if (source.IsConstant()) { |
| // Move to stack from constant |
| int64_t value = GetInt64ValueOf(source.GetConstant()); |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister gpr = (value != 0) ? srs.AllocateXRegister() : Zero; |
| if (value != 0) { |
| __ LoadConst64(gpr, value); |
| } |
| if (destination.IsStackSlot()) { |
| __ Storew(gpr, SP, destination.GetStackIndex()); |
| } else { |
| DCHECK(destination.IsDoubleStackSlot()); |
| __ Stored(gpr, SP, destination.GetStackIndex()); |
| } |
| } else { |
| DCHECK(source.IsStackSlot() || source.IsDoubleStackSlot()); |
| DCHECK_EQ(source.IsDoubleStackSlot(), destination.IsDoubleStackSlot()); |
| // Move to stack from stack |
| ScratchRegisterScope srs(GetAssembler()); |
| XRegister tmp = srs.AllocateXRegister(); |
| if (destination.IsStackSlot()) { |
| __ Loadw(tmp, SP, source.GetStackIndex()); |
| __ Storew(tmp, SP, destination.GetStackIndex()); |
| } else { |
| __ Loadd(tmp, SP, source.GetStackIndex()); |
| __ Stored(tmp, SP, destination.GetStackIndex()); |
| } |
| } |
| } |
| } |
| |
| void CodeGeneratorRISCV64::AddLocationAsTemp(Location location, LocationSummary* locations) { |
| if (location.IsRegister()) { |
| locations->AddTemp(location); |
| } else { |
| UNIMPLEMENTED(FATAL) << "AddLocationAsTemp not implemented for location " << location; |
| } |
| } |
| |
| void CodeGeneratorRISCV64::SetupBlockedRegisters() const { |
| // ZERO, GP, SP, RA, TP and TR(S1) are reserved and can't be allocated. |
| blocked_core_registers_[Zero] = true; |
| blocked_core_registers_[GP] = true; |
| blocked_core_registers_[SP] = true; |
| blocked_core_registers_[RA] = true; |
| blocked_core_registers_[TP] = true; |
| blocked_core_registers_[TR] = true; // ART Thread register. |
| |
| // TMP(T6), TMP2(T5) and FTMP(FT11) are used as temporary/scratch registers. |
| blocked_core_registers_[TMP] = true; |
| blocked_core_registers_[TMP2] = true; |
| blocked_fpu_registers_[FTMP] = true; |
| |
| if (GetGraph()->IsDebuggable()) { |
| // Stubs do not save callee-save floating point registers. If the graph |
| // is debuggable, we need to deal with these registers differently. For |
| // now, just block them. |
| for (size_t i = 0; i < arraysize(kFpuCalleeSaves); ++i) { |
| blocked_fpu_registers_[kFpuCalleeSaves[i]] = true; |
| } |
| } |
| } |
| |
| size_t CodeGeneratorRISCV64::SaveCoreRegister(size_t stack_index, uint32_t reg_id) { |
| __ Stored(XRegister(reg_id), SP, stack_index); |
| return kRiscv64DoublewordSize; |
| } |
| |
| size_t CodeGeneratorRISCV64::RestoreCoreRegister(size_t stack_index, uint32_t reg_id) { |
| __ Loadd(XRegister(reg_id), SP, stack_index); |
| return kRiscv64DoublewordSize; |
| } |
| |
| size_t CodeGeneratorRISCV64::SaveFloatingPointRegister(size_t stack_index, uint32_t reg_id) { |
| if (GetGraph()->HasSIMD()) { |
| // TODO(riscv64): RISC-V vector extension. |
| UNIMPLEMENTED(FATAL) << "Vector extension is unsupported"; |
| UNREACHABLE(); |
| } |
| __ FStored(FRegister(reg_id), SP, stack_index); |
| return kRiscv64FloatRegSizeInBytes; |
| } |
| |
| size_t CodeGeneratorRISCV64::RestoreFloatingPointRegister(size_t stack_index, uint32_t reg_id) { |
| if (GetGraph()->HasSIMD()) { |
| // TODO(riscv64): RISC-V vector extension. |
| UNIMPLEMENTED(FATAL) << "Vector extension is unsupported"; |
| UNREACHABLE(); |
| } |
| __ FLoadd(FRegister(reg_id), SP, stack_index); |
| return kRiscv64FloatRegSizeInBytes; |
| } |
| |
| void CodeGeneratorRISCV64::DumpCoreRegister(std::ostream& stream, int reg) const { |
| stream << XRegister(reg); |
| } |
| |
| void CodeGeneratorRISCV64::DumpFloatingPointRegister(std::ostream& stream, int reg) const { |
| stream << FRegister(reg); |
| } |
| |
| void CodeGeneratorRISCV64::Finalize() { |
| // Ensure that we fix up branches and literal loads and emit the literal pool. |
| __ FinalizeCode(); |
| |
| // Adjust native pc offsets in stack maps. |
| StackMapStream* stack_map_stream = GetStackMapStream(); |
| for (size_t i = 0, num = stack_map_stream->GetNumberOfStackMaps(); i != num; ++i) { |
| uint32_t old_position = stack_map_stream->GetStackMapNativePcOffset(i); |
| uint32_t new_position = __ GetAdjustedPosition(old_position); |
| DCHECK_GE(new_position, old_position); |
| stack_map_stream->SetStackMapNativePcOffset(i, new_position); |
| } |
| |
| // Adjust pc offsets for the disassembly information. |
| if (disasm_info_ != nullptr) { |
| GeneratedCodeInterval* frame_entry_interval = disasm_info_->GetFrameEntryInterval(); |
| frame_entry_interval->start = __ GetAdjustedPosition(frame_entry_interval->start); |
| frame_entry_interval->end = __ GetAdjustedPosition(frame_entry_interval->end); |
| for (auto& entry : *disasm_info_->GetInstructionIntervals()) { |
| entry.second.start = __ GetAdjustedPosition(entry.second.start); |
| entry.second.end = __ GetAdjustedPosition(entry.second.end); |
| } |
| for (auto& entry : *disasm_info_->GetSlowPathIntervals()) { |
| entry.code_interval.start = __ GetAdjustedPosition(entry.code_interval.start); |
| entry.code_interval.end = __ GetAdjustedPosition(entry.code_interval.end); |
| } |
| } |
| |
| CodeGenerator::Finalize(); |
| } |
| |
| // Generate code to invoke a runtime entry point. |
| void CodeGeneratorRISCV64::InvokeRuntime(QuickEntrypointEnum entrypoint, |
| HInstruction* instruction, |
| uint32_t dex_pc, |
| SlowPathCode* slow_path) { |
| ValidateInvokeRuntime(entrypoint, instruction, slow_path); |
| |
| ThreadOffset64 entrypoint_offset = GetThreadOffset<kRiscv64PointerSize>(entrypoint); |
| |
| // TODO(riscv64): Reduce code size for AOT by using shared trampolines for slow path |
| // runtime calls across the entire oat file. |
| __ Loadd(RA, TR, entrypoint_offset.Int32Value()); |
| __ Jalr(RA); |
| if (EntrypointRequiresStackMap(entrypoint)) { |
| RecordPcInfo(instruction, dex_pc, slow_path); |
| } |
| } |
| |
| // Generate code to invoke a runtime entry point, but do not record |
| // PC-related information in a stack map. |
| void CodeGeneratorRISCV64::InvokeRuntimeWithoutRecordingPcInfo(int32_t entry_point_offset, |
| HInstruction* instruction, |
| SlowPathCode* slow_path) { |
| ValidateInvokeRuntimeWithoutRecordingPcInfo(instruction, slow_path); |
| __ Loadd(RA, TR, entry_point_offset); |
| __ Jalr(RA); |
| } |
| |
| void CodeGeneratorRISCV64::IncreaseFrame(size_t adjustment) { |
| int32_t adjustment32 = dchecked_integral_cast<int32_t>(adjustment); |
| __ AddConst64(SP, SP, -adjustment32); |
| GetAssembler()->cfi().AdjustCFAOffset(adjustment32); |
| } |
| |
| void CodeGeneratorRISCV64::DecreaseFrame(size_t adjustment) { |
| int32_t adjustment32 = dchecked_integral_cast<int32_t>(adjustment); |
| __ AddConst64(SP, SP, adjustment32); |
| GetAssembler()->cfi().AdjustCFAOffset(-adjustment32); |
| } |
| |
| void CodeGeneratorRISCV64::GenerateNop() { |
| __ Nop(); |
| } |
| |
| void CodeGeneratorRISCV64::GenerateImplicitNullCheck(HNullCheck* instruction) { |
| if (CanMoveNullCheckToUser(instruction)) { |
| return; |
| } |
| Location obj = instruction->GetLocations()->InAt(0); |
| |
| __ Lw(Zero, obj.AsRegister<XRegister>(), 0); |
| RecordPcInfo(instruction, instruction->GetDexPc()); |
| } |
| |
| void CodeGeneratorRISCV64::GenerateExplicitNullCheck(HNullCheck* instruction) { |
| SlowPathCodeRISCV64* slow_path = new (GetScopedAllocator()) NullCheckSlowPathRISCV64(instruction); |
| AddSlowPath(slow_path); |
| |
| Location obj = instruction->GetLocations()->InAt(0); |
| |
| __ Beqz(obj.AsRegister<XRegister>(), slow_path->GetEntryLabel()); |
| } |
| |
| HLoadString::LoadKind CodeGeneratorRISCV64::GetSupportedLoadStringKind( |
| HLoadString::LoadKind desired_string_load_kind) { |
| switch (desired_string_load_kind) { |
| case HLoadString::LoadKind::kBootImageLinkTimePcRelative: |
| case HLoadString::LoadKind::kBootImageRelRo: |
| case HLoadString::LoadKind::kBssEntry: |
| DCHECK(!Runtime::Current()->UseJitCompilation()); |
| break; |
| case HLoadString::LoadKind::kJitBootImageAddress: |
| case HLoadString::LoadKind::kJitTableAddress: |
| DCHECK(Runtime::Current()->UseJitCompilation()); |
| break; |
| case HLoadString::LoadKind::kRuntimeCall: |
| break; |
| } |
| return desired_string_load_kind; |
| } |
| |
| HLoadClass::LoadKind CodeGeneratorRISCV64::GetSupportedLoadClassKind( |
| HLoadClass::LoadKind desired_class_load_kind) { |
| switch (desired_class_load_kind) { |
| case HLoadClass::LoadKind::kInvalid: |
| LOG(FATAL) << "UNREACHABLE"; |
| UNREACHABLE(); |
| case HLoadClass::LoadKind::kReferrersClass: |
| break; |
| case HLoadClass::LoadKind::kBootImageLinkTimePcRelative: |
| case HLoadClass::LoadKind::kBootImageRelRo: |
| case HLoadClass::LoadKind::kBssEntry: |
| case HLoadClass::LoadKind::kBssEntryPublic: |
| case HLoadClass::LoadKind::kBssEntryPackage: |
| DCHECK(!Runtime::Current()->UseJitCompilation()); |
| break; |
| case HLoadClass::LoadKind::kJitBootImageAddress: |
| case HLoadClass::LoadKind::kJitTableAddress: |
| DCHECK(Runtime::Current()->UseJitCompilation()); |
| break; |
| case HLoadClass::LoadKind::kRuntimeCall: |
| break; |
| } |
| return desired_class_load_kind; |
| } |
| |
| HInvokeStaticOrDirect::DispatchInfo CodeGeneratorRISCV64::GetSupportedInvokeStaticOrDirectDispatch( |
| const HInvokeStaticOrDirect::DispatchInfo& desired_dispatch_info, ArtMethod* method) { |
| UNUSED(method); |
| // On RISCV64 we support all dispatch types. |
| return desired_dispatch_info; |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageIntrinsicPatch( |
| uint32_t intrinsic_data, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch( |
| /* dex_file= */ nullptr, intrinsic_data, info_high, &boot_image_other_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageRelRoPatch( |
| uint32_t boot_image_offset, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch( |
| /* dex_file= */ nullptr, boot_image_offset, info_high, &boot_image_other_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageMethodPatch( |
| MethodReference target_method, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch( |
| target_method.dex_file, target_method.index, info_high, &boot_image_method_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewMethodBssEntryPatch( |
| MethodReference target_method, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch( |
| target_method.dex_file, target_method.index, info_high, &method_bss_entry_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageTypePatch( |
| const DexFile& dex_file, dex::TypeIndex type_index, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch(&dex_file, type_index.index_, info_high, &boot_image_type_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageJniEntrypointPatch( |
| MethodReference target_method, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch( |
| target_method.dex_file, target_method.index, info_high, &boot_image_jni_entrypoint_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewTypeBssEntryPatch( |
| HLoadClass* load_class, |
| const PcRelativePatchInfo* info_high) { |
| const DexFile& dex_file = load_class->GetDexFile(); |
| dex::TypeIndex type_index = load_class->GetTypeIndex(); |
| ArenaDeque<PcRelativePatchInfo>* patches = nullptr; |
| switch (load_class->GetLoadKind()) { |
| case HLoadClass::LoadKind::kBssEntry: |
| patches = &type_bss_entry_patches_; |
| break; |
| case HLoadClass::LoadKind::kBssEntryPublic: |
| patches = &public_type_bss_entry_patches_; |
| break; |
| case HLoadClass::LoadKind::kBssEntryPackage: |
| patches = &package_type_bss_entry_patches_; |
| break; |
| default: |
| LOG(FATAL) << "Unexpected load kind: " << load_class->GetLoadKind(); |
| UNREACHABLE(); |
| } |
| return NewPcRelativePatch(&dex_file, type_index.index_, info_high, patches); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewBootImageStringPatch( |
| const DexFile& dex_file, dex::StringIndex string_index, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch(&dex_file, string_index.index_, info_high, &boot_image_string_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewStringBssEntryPatch( |
| const DexFile& dex_file, dex::StringIndex string_index, const PcRelativePatchInfo* info_high) { |
| return NewPcRelativePatch(&dex_file, string_index.index_, info_high, &string_bss_entry_patches_); |
| } |
| |
| CodeGeneratorRISCV64::PcRelativePatchInfo* CodeGeneratorRISCV64::NewPcRelativePatch( |
| const DexFile* dex_file, |
| uint32_t offset_or_index, |
| const PcRelativePatchInfo* info_high, |
| ArenaDeque<PcRelativePatchInfo>* patches) { |
| patches->emplace_back(dex_file, offset_or_index, info_high); |
| return &patches->back(); |
| } |
| |
| Literal* CodeGeneratorRISCV64::DeduplicateUint32Literal(uint32_t value) { |
| return uint32_literals_.GetOrCreate(value, |
| [this, value]() { return __ NewLiteral<uint32_t>(value); }); |
| } |
| |
| Literal* CodeGeneratorRISCV64::DeduplicateUint64Literal(uint64_t value) { |
| return uint64_literals_.GetOrCreate(value, |
| [this, value]() { return __ NewLiteral<uint64_t>(value); }); |
| } |
| |
| Literal* CodeGeneratorRISCV64::DeduplicateBootImageAddressLiteral(uint64_t address) { |
| return DeduplicateUint32Literal(dchecked_integral_cast<uint32_t>(address)); |
| } |
| |
| void CodeGeneratorRISCV64::EmitPcRelativeAuipcPlaceholder(PcRelativePatchInfo* info_high, |
| XRegister out) { |
| DCHECK(info_high->pc_insn_label == &info_high->label); |
| __ Bind(&info_high->label); |
| __ Auipc(out, /*imm20=*/ 0x12345); // Placeholder `imm20` patched at link time. |
| } |
| |
| void CodeGeneratorRISCV64::EmitPcRelativeAddiPlaceholder(PcRelativePatchInfo* info_low, |
| XRegister rd, |
| XRegister rs1) { |
| DCHECK(info_low->pc_insn_label != &info_low->label); |
| __ Bind(&info_low->label); |
| __ Addi(rd, rs1, /*imm12=*/ 0x678); // Placeholder `imm12` patched at link time. |
| } |
| |
| void CodeGeneratorRISCV64::EmitPcRelativeLwuPlaceholder(PcRelativePatchInfo* info_low, |
| XRegister rd, |
| XRegister rs1) { |
| DCHECK(info_low->pc_insn_label != &info_low->label); |
| __ Bind(&info_low->label); |
| __ Lwu(rd, rs1, /*offset=*/ 0x678); // Placeholder `offset` patched at link time. |
| } |
| |
| void CodeGeneratorRISCV64::EmitPcRelativeLdPlaceholder(PcRelativePatchInfo* info_low, |
| XRegister rd, |
| XRegister rs1) { |
| DCHECK(info_low->pc_insn_label != &info_low->label); |
| __ Bind(&info_low->label); |
| __ Ld(rd, rs1, /*offset=*/ 0x678); // Placeholder `offset` patched at link time. |
| } |
| |
| template <linker::LinkerPatch (*Factory)(size_t, const DexFile*, uint32_t, uint32_t)> |
| inline void CodeGeneratorRISCV64::EmitPcRelativeLinkerPatches( |
| const ArenaDeque<PcRelativePatchInfo>& infos, |
| ArenaVector<linker::LinkerPatch>* linker_patches) { |
| for (const PcRelativePatchInfo& info : infos) { |
| linker_patches->push_back(Factory(__ GetLabelLocation(&info.label), |
| info.target_dex_file, |
| __ GetLabelLocation(info.pc_insn_label), |
| info.offset_or_index)); |
| } |
| } |
| |
| template <linker::LinkerPatch (*Factory)(size_t, uint32_t, uint32_t)> |
| linker::LinkerPatch NoDexFileAdapter(size_t literal_offset, |
| const DexFile* target_dex_file, |
| uint32_t pc_insn_offset, |
| uint32_t boot_image_offset) { |
| DCHECK(target_dex_file == nullptr); // Unused for these patches, should be null. |
| return Factory(literal_offset, pc_insn_offset, boot_image_offset); |
| } |
| |
| void CodeGeneratorRISCV64::EmitLinkerPatches(ArenaVector<linker::LinkerPatch>* linker_patches) { |
| DCHECK(linker_patches->empty()); |
| size_t size = |
| boot_image_method_patches_.size() + |
| method_bss_entry_patches_.size() + |
| boot_image_type_patches_.size() + |
| type_bss_entry_patches_.size() + |
| public_type_bss_entry_patches_.size() + |
| package_type_bss_entry_patches_.size() + |
| boot_image_string_patches_.size() + |
| string_bss_entry_patches_.size() + |
| boot_image_jni_entrypoint_patches_.size() + |
| boot_image_other_patches_.size(); |
| linker_patches->reserve(size); |
| if (GetCompilerOptions().IsBootImage() || GetCompilerOptions().IsBootImageExtension()) { |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::RelativeMethodPatch>( |
| boot_image_method_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::RelativeTypePatch>( |
| boot_image_type_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::RelativeStringPatch>( |
| boot_image_string_patches_, linker_patches); |
| } else { |
| DCHECK(boot_image_method_patches_.empty()); |
| DCHECK(boot_image_type_patches_.empty()); |
| DCHECK(boot_image_string_patches_.empty()); |
| } |
| if (GetCompilerOptions().IsBootImage()) { |
| EmitPcRelativeLinkerPatches<NoDexFileAdapter<linker::LinkerPatch::IntrinsicReferencePatch>>( |
| boot_image_other_patches_, linker_patches); |
| } else { |
| EmitPcRelativeLinkerPatches<NoDexFileAdapter<linker::LinkerPatch::DataBimgRelRoPatch>>( |
| boot_image_other_patches_, linker_patches); |
| } |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::MethodBssEntryPatch>( |
| method_bss_entry_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::TypeBssEntryPatch>( |
| type_bss_entry_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::PublicTypeBssEntryPatch>( |
| public_type_bss_entry_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::PackageTypeBssEntryPatch>( |
| package_type_bss_entry_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::StringBssEntryPatch>( |
| string_bss_entry_patches_, linker_patches); |
| EmitPcRelativeLinkerPatches<linker::LinkerPatch::RelativeJniEntrypointPatch>( |
| boot_image_jni_entrypoint_patches_, linker_patches); |
| DCHECK_EQ(size, linker_patches->size()); |
| } |
| |
| void CodeGeneratorRISCV64::LoadMethod(MethodLoadKind load_kind, Location temp, HInvoke* invoke) { |
| switch (load_kind) { |
| case MethodLoadKind::kBootImageLinkTimePcRelative: { |
| DCHECK(GetCompilerOptions().IsBootImage() || GetCompilerOptions().IsBootImageExtension()); |
| CodeGeneratorRISCV64::PcRelativePatchInfo* info_high = |
| NewBootImageMethodPatch(invoke->GetResolvedMethodReference()); |
| EmitPcRelativeAuipcPlaceholder(info_high, temp.AsRegister<XRegister>()); |
| CodeGeneratorRISCV64::PcRelativePatchInfo* info_low = |
| NewBootImageMethodPatch(invoke->GetResolvedMethodReference(), info_high); |
| EmitPcRelativeAddiPlaceholder( |
| info_low, temp.AsRegister<XRegister>(), temp.AsRegister<XRegister>()); |
| break; |
| } |
| case MethodLoadKind::kBootImageRelRo: { |
| uint32_t boot_image_offset = GetBootImageOffset(invoke); |
| PcRelativePatchInfo* info_high = NewBootImageRelRoPatch(boot_image_offset); |
| EmitPcRelativeAuipcPlaceholder(info_high, temp.AsRegister<XRegister>()); |
| PcRelativePatchInfo* info_low = NewBootImageRelRoPatch(boot_image_offset, info_high); |
| // Note: Boot image is in the low 4GiB and the entry is 32-bit, so emit a 32-bit load. |
| EmitPcRelativeLwuPlaceholder( |
| info_low, temp.AsRegister<XRegister>(), temp.AsRegister<XRegister>()); |
| break; |
| } |
| case MethodLoadKind::kBssEntry: { |
| PcRelativePatchInfo* info_high = NewMethodBssEntryPatch(invoke->GetMethodReference()); |
| EmitPcRelativeAuipcPlaceholder(info_high, temp.AsRegister<XRegister>()); |
| PcRelativePatchInfo* info_low = |
| NewMethodBssEntryPatch(invoke->GetMethodReference(), info_high); |
| EmitPcRelativeLdPlaceholder( |
| info_low, temp.AsRegister<XRegister>(), temp.AsRegister<XRegister>()); |
| break; |
| } |
| case MethodLoadKind::kJitDirectAddress: { |
| __ LoadConst64(temp.AsRegister<XRegister>(), |
| reinterpret_cast<uint64_t>(invoke->GetResolvedMethod())); |
| break; |
| } |
| case MethodLoadKind::kRuntimeCall: { |
| // Test situation, don't do anything. |
| break; |
| } |
| default: { |
| LOG(FATAL) << "Load kind should have already been handled " << load_kind; |
| UNREACHABLE(); |
| } |
| } |
| } |
| |
| void CodeGeneratorRISCV64::GenerateStaticOrDirectCall(HInvokeStaticOrDirect* invoke, |
| Location temp, |
| SlowPathCode* slow_path) { |
| // All registers are assumed to be correctly set up per the calling convention. |
| Location callee_method = temp; // For all kinds except kRecursive, callee will be in temp. |
| |
| switch (invoke->GetMethodLoadKind()) { |
| case MethodLoadKind::kStringInit: { |
| // temp = thread->string_init_entrypoint |
| uint32_t offset = |
| GetThreadOffset<kRiscv64PointerSize>(invoke->GetStringInitEntryPoint()).Int32Value(); |
| __ Loadd(temp.AsRegister<XRegister>(), TR, offset); |
| break; |
| } |
| case MethodLoadKind::kRecursive: |
| callee_method = invoke->GetLocations()->InAt(invoke->GetCurrentMethodIndex()); |
| break; |
| case MethodLoadKind::kRuntimeCall: |
| GenerateInvokeStaticOrDirectRuntimeCall(invoke, temp, slow_path); |
| return; // No code pointer retrieval; the runtime performs the call directly. |
| case MethodLoadKind::kBootImageLinkTimePcRelative: |
| DCHECK(GetCompilerOptions().IsBootImage() || GetCompilerOptions().IsBootImageExtension()); |
| if (invoke->GetCodePtrLocation() == CodePtrLocation::kCallCriticalNative) { |
| // Do not materialize the method pointer, load directly the entrypoint. |
| CodeGeneratorRISCV64::PcRelativePatchInfo* info_high = |
| NewBootImageJniEntrypointPatch(invoke->GetResolvedMethodReference()); |
| EmitPcRelativeAuipcPlaceholder(info_high, RA); |
| CodeGeneratorRISCV64::PcRelativePatchInfo* info_low = |
| NewBootImageJniEntrypointPatch(invoke->GetResolvedMethodReference(), info_high); |
| EmitPcRelativeLdPlaceholder(info_low, RA, RA); |
| break; |
| } |
| FALLTHROUGH_INTENDED; |
| default: |
| LoadMethod(invoke->GetMethodLoadKind(), temp, invoke); |
| break; |
| } |
| |
| switch (invoke->GetCodePtrLocation()) { |
| case CodePtrLocation::kCallSelf: |
| DCHECK(!GetGraph()->HasShouldDeoptimizeFlag()); |
| __ Jal(&frame_entry_label_); |
| RecordPcInfo(invoke, invoke->GetDexPc(), slow_path); |
| break; |
| case CodePtrLocation::kCallArtMethod: |
| // RA = callee_method->entry_point_from_quick_compiled_code_; |
| __ Loadd(RA, |
| callee_method.AsRegister<XRegister>(), |
| ArtMethod::EntryPointFromQuickCompiledCodeOffset(kRiscv64PointerSize).Int32Value()); |
| // RA() |
| __ Jalr(RA); |
| RecordPcInfo(invoke, invoke->GetDexPc(), slow_path); |
| break; |
| case CodePtrLocation::kCallCriticalNative: { |
| size_t out_frame_size = |
| PrepareCriticalNativeCall<CriticalNativeCallingConventionVisitorRiscv64, |
| kNativeStackAlignment, |
| GetCriticalNativeDirectCallFrameSize>(invoke); |
| if (invoke->GetMethodLoadKind() == MethodLoadKind::kBootImageLinkTimePcRelative) { |
| // Entrypoint is already loaded in RA. |
| } else { |
| // RA = callee_method->ptr_sized_fields_.data_; // EntryPointFromJni |
| MemberOffset offset = ArtMethod::EntryPointFromJniOffset(kRiscv64PointerSize); |
| __ Loadd(RA, callee_method.AsRegister<XRegister>(), offset.Int32Value()); |
| } |
| __ Jalr(RA); |
| RecordPcInfo(invoke, invoke->GetDexPc(), slow_path); |
| // The result is returned the same way in native ABI and managed ABI. No result conversion is |
| // needed, see comments in `Riscv64JniCallingConvention::RequiresSmallResultTypeExtension()`. |
| if (out_frame_size != 0u) { |
| DecreaseFrame(out_frame_size); |
| } |
| break; |
| } |
| } |
| |
| DCHECK(!IsLeafMethod()); |
| } |
| |
| void CodeGeneratorRISCV64::MaybeGenerateInlineCacheCheck(HInstruction* instruction, |
| XRegister klass) { |
| // We know the destination of an intrinsic, so no need to record inline caches. |
| if (!instruction->GetLocations()->Intrinsified() && |
| GetGraph()->IsCompilingBaseline() && |
| !Runtime::Current()->IsAotCompiler()) { |
| DCHECK(!instruction->GetEnvironment()->IsFromInlinedInvoke()); |
| ProfilingInfo* info = GetGraph()->GetProfilingInfo(); |
| DCHECK(info != nullptr); |
| InlineCache* cache = info->GetInlineCache(instruction->GetDexPc()); |
| uint64_t address = reinterpret_cast64<uint64_t>(cache); |
| Riscv64Label done; |
| // The `art_quick_update_inline_cache` expects the inline cache in T5. |
| XRegister ic_reg = T5; |
| ScratchRegisterScope srs(GetAssembler()); |
| DCHECK_EQ(srs.AvailableXRegisters(), 2u); |
| srs.ExcludeXRegister(ic_reg); |
| DCHECK_EQ(srs.AvailableXRegisters(), 1u); |
| __ LoadConst64(ic_reg, address); |
| { |
| ScratchRegisterScope srs2(GetAssembler()); |
| XRegister tmp = srs2.AllocateXRegister(); |
| __ Loadd(tmp, ic_reg, InlineCache::ClassesOffset().Int32Value()); |
| // Fast path for a monomorphic cache. |
| __ Beq(klass, tmp, &done); |
| } |
| InvokeRuntime(kQuickUpdateInlineCache, instruction, instruction->GetDexPc()); |
| __ Bind(&done); |
| } |
| } |
| |
| void CodeGeneratorRISCV64::GenerateVirtualCall(HInvokeVirtual* invoke, |
| Location temp_location, |
| SlowPathCode* slow_path) { |
| // Use the calling convention instead of the location of the receiver, as |
| // intrinsics may have put the receiver in a different register. In the intrinsics |
| // slow path, the arguments have been moved to the right place, so here we are |
| // guaranteed that the receiver is the first register of the calling convention. |
| InvokeDexCallingConvention calling_convention; |
| XRegister receiver = calling_convention.GetRegisterAt(0); |
| XRegister temp = temp_location.AsRegister<XRegister>(); |
| MemberOffset method_offset = |
| mirror::Class::EmbeddedVTableEntryOffset(invoke->GetVTableIndex(), kRiscv64PointerSize); |
| MemberOffset class_offset = mirror::Object::ClassOffset(); |
| Offset entry_point = ArtMethod::EntryPointFromQuickCompiledCodeOffset(kRiscv64PointerSize); |
| |
| // temp = object->GetClass(); |
| __ Loadwu(temp, receiver, class_offset.Int32Value()); |
| MaybeRecordImplicitNullCheck(invoke); |
| // Instead of simply (possibly) unpoisoning `temp` here, we should |
| // emit a read barrier for the previous class reference load. |
| // However this is not required in practice, as this is an |
| // intermediate/temporary reference and because the current |
| // concurrent copying collector keeps the from-space memory |
| // intact/accessible until the end of the marking phase (the |
| // concurrent copying collector may not in the future). |
| MaybeUnpoisonHeapReference(temp); |
| |
| // If we're compiling baseline, update the inline cache. |
| MaybeGenerateInlineCacheCheck(invoke, temp); |
| |
| // temp = temp->GetMethodAt(method_offset); |
| __ Loadd(temp, temp, method_offset.Int32Value()); |
| // RA = temp->GetEntryPoint(); |
| __ Loadd(RA, temp, entry_point.Int32Value()); |
| // RA(); |
| __ Jalr(RA); |
| RecordPcInfo(invoke, invoke->GetDexPc(), slow_path); |
| } |
| |
| void CodeGeneratorRISCV64::MoveFromReturnRegister(Location trg, DataType::Type type) { |
| if (!trg.IsValid()) { |
| DCHECK_EQ(type, DataType::Type::kVoid); |
| return; |
| } |
| |
| DCHECK_NE(type, DataType::Type::kVoid); |
| |
| if (DataType::IsIntegralType(type) || type == DataType::Type::kReference) { |
| XRegister trg_reg = trg.AsRegister<XRegister>(); |
| XRegister res_reg = Riscv64ReturnLocation(type).AsRegister<XRegister>(); |
| if (trg_reg != res_reg) { |
| __ Mv(trg_reg, res_reg); |
| } |
| } else { |
| FRegister trg_reg = trg.AsFpuRegister<FRegister>(); |
| FRegister res_reg = Riscv64ReturnLocation(type).AsFpuRegister<FRegister>(); |
| if (trg_reg != res_reg) { |
| __ FMvD(trg_reg, res_reg); // 64-bit move is OK also for `float`. |
| } |
| } |
| } |
| |
| void CodeGeneratorRISCV64::PoisonHeapReference(XRegister reg) { |
| __ Sub(reg, Zero, reg); // Negate the ref. |
| __ ZextW(reg, reg); // Zero-extend the 32-bit ref. |
| } |
| |
| void CodeGeneratorRISCV64::UnpoisonHeapReference(XRegister reg) { |
| __ Sub(reg, Zero, reg); // Negate the ref. |
| __ ZextW(reg, reg); // Zero-extend the 32-bit ref. |
| } |
| |
| inline void CodeGeneratorRISCV64::MaybePoisonHeapReference(XRegister reg) { |
| if (kPoisonHeapReferences) { |
| PoisonHeapReference(reg); |
| } |
| } |
| |
| inline void CodeGeneratorRISCV64::MaybeUnpoisonHeapReference(XRegister reg) { |
| if (kPoisonHeapReferences) { |
| UnpoisonHeapReference(reg); |
| } |
| } |
| |
| void CodeGeneratorRISCV64::SwapLocations(Location loc1, Location loc2, DataType::Type type) { |
| DCHECK(!loc1.IsConstant()); |
| DCHECK(!loc2.IsConstant()); |
| |
| if (loc1.Equals(loc2)) { |
| return; |
| } |
| |
| bool is_slot1 = loc1.IsStackSlot() || loc1.IsDoubleStackSlot(); |
| bool is_slot2 = loc2.IsStackSlot() || loc2.IsDoubleStackSlot(); |
| bool is_simd1 = loc1.IsSIMDStackSlot(); |
| bool is_simd2 = loc2.IsSIMDStackSlot(); |
| bool is_fp_reg1 = loc1.IsFpuRegister(); |
| bool is_fp_reg2 = loc2.IsFpuRegister(); |
| |
| if ((is_slot1 != is_slot2) || |
| (loc2.IsRegister() && loc1.IsRegister()) || |
| (is_fp_reg2 && is_fp_reg1)) { |
| if ((is_fp_reg2 && is_fp_reg1) && GetGraph()->HasSIMD()) { |
| LOG(FATAL) << "Unsupported"; |
| UNREACHABLE(); |
| } |
| ScratchRegisterScope srs(GetAssembler()); |
| Location tmp = (is_fp_reg2 || is_fp_reg1) |
| ? Location::FpuRegisterLocation(srs.AllocateFRegister()) |
| : Location::RegisterLocation(srs.AllocateXRegister()); |
| MoveLocation(tmp, loc1, type); |
| MoveLocation(loc1, loc2, type); |
| MoveLocation(loc2, tmp, type); |
| } else if (is_slot1 && is_slot2) { |
| move_resolver_.Exchange(loc1.GetStackIndex(), loc2.GetStackIndex(), loc1.IsDoubleStackSlot()); |
| } else if (is_simd1 && is_simd2) { |
| // TODO(riscv64): Add VECTOR/SIMD later. |
| UNIMPLEMENTED(FATAL) << "Vector extension is unsupported"; |
| } else if ((is_fp_reg1 && is_simd2) || (is_fp_reg2 && is_simd1)) { |
| // TODO(riscv64): Add VECTOR/SIMD later. |
| UNIMPLEMENTED(FATAL) << "Vector extension is unsupported"; |
| } else { |
| LOG(FATAL) << "Unimplemented swap between locations " << loc1 << " and " << loc2; |
| } |
| } |
| |
| } // namespace riscv64 |
| } // namespace art |