| /* |
| * Copyright (C) 2017 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_ |
| #define ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_ |
| |
| #include "code_generator_arm_vixl.h" |
| #include "scheduler.h" |
| |
| namespace art { |
| namespace arm { |
| // TODO: Replace CodeGeneratorARMType with CodeGeneratorARMVIXL everywhere? |
| typedef CodeGeneratorARMVIXL CodeGeneratorARMType; |
| |
| // AArch32 instruction latencies. |
| // We currently assume that all ARM CPUs share the same instruction latency list. |
| // The following latencies were tuned based on performance experiments and |
| // automatic tuning using differential evolution approach on various benchmarks. |
| static constexpr uint32_t kArmIntegerOpLatency = 2; |
| static constexpr uint32_t kArmFloatingPointOpLatency = 11; |
| static constexpr uint32_t kArmDataProcWithShifterOpLatency = 4; |
| static constexpr uint32_t kArmMulIntegerLatency = 6; |
| static constexpr uint32_t kArmMulFloatingPointLatency = 11; |
| static constexpr uint32_t kArmDivIntegerLatency = 10; |
| static constexpr uint32_t kArmDivFloatLatency = 20; |
| static constexpr uint32_t kArmDivDoubleLatency = 25; |
| static constexpr uint32_t kArmTypeConversionFloatingPointIntegerLatency = 11; |
| static constexpr uint32_t kArmMemoryLoadLatency = 9; |
| static constexpr uint32_t kArmMemoryStoreLatency = 9; |
| static constexpr uint32_t kArmMemoryBarrierLatency = 6; |
| static constexpr uint32_t kArmBranchLatency = 4; |
| static constexpr uint32_t kArmCallLatency = 5; |
| static constexpr uint32_t kArmCallInternalLatency = 29; |
| static constexpr uint32_t kArmLoadStringInternalLatency = 10; |
| static constexpr uint32_t kArmNopLatency = 2; |
| static constexpr uint32_t kArmLoadWithBakerReadBarrierLatency = 18; |
| static constexpr uint32_t kArmRuntimeTypeCheckLatency = 46; |
| |
| class SchedulingLatencyVisitorARM : public SchedulingLatencyVisitor { |
| public: |
| explicit SchedulingLatencyVisitorARM(CodeGenerator* codegen) |
| : codegen_(down_cast<CodeGeneratorARMType*>(codegen)) {} |
| |
| // Default visitor for instructions not handled specifically below. |
| void VisitInstruction(HInstruction* ATTRIBUTE_UNUSED) { |
| last_visited_latency_ = kArmIntegerOpLatency; |
| } |
| |
| // We add a second unused parameter to be able to use this macro like the others |
| // defined in `nodes.h`. |
| #define FOR_EACH_SCHEDULED_ARM_INSTRUCTION(M) \ |
| M(ArrayGet , unused) \ |
| M(ArrayLength , unused) \ |
| M(ArraySet , unused) \ |
| M(Add , unused) \ |
| M(Sub , unused) \ |
| M(And , unused) \ |
| M(Or , unused) \ |
| M(Ror , unused) \ |
| M(Xor , unused) \ |
| M(Shl , unused) \ |
| M(Shr , unused) \ |
| M(UShr , unused) \ |
| M(Mul , unused) \ |
| M(Div , unused) \ |
| M(Condition , unused) \ |
| M(Compare , unused) \ |
| M(BoundsCheck , unused) \ |
| M(InstanceFieldGet , unused) \ |
| M(InstanceFieldSet , unused) \ |
| M(InstanceOf , unused) \ |
| M(Invoke , unused) \ |
| M(LoadString , unused) \ |
| M(NewArray , unused) \ |
| M(NewInstance , unused) \ |
| M(Rem , unused) \ |
| M(StaticFieldGet , unused) \ |
| M(StaticFieldSet , unused) \ |
| M(SuspendCheck , unused) \ |
| M(TypeConversion , unused) |
| |
| #define FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(M) \ |
| M(BitwiseNegatedRight, unused) \ |
| M(MultiplyAccumulate, unused) \ |
| M(IntermediateAddress, unused) \ |
| M(IntermediateAddressIndex, unused) \ |
| M(DataProcWithShifterOp, unused) |
| |
| #define DECLARE_VISIT_INSTRUCTION(type, unused) \ |
| void Visit##type(H##type* instruction) OVERRIDE; |
| |
| FOR_EACH_SCHEDULED_ARM_INSTRUCTION(DECLARE_VISIT_INSTRUCTION) |
| FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(DECLARE_VISIT_INSTRUCTION) |
| FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION) |
| |
| #undef DECLARE_VISIT_INSTRUCTION |
| |
| private: |
| bool CanGenerateTest(HCondition* cond); |
| void HandleGenerateConditionWithZero(IfCondition cond); |
| void HandleGenerateLongTestConstant(HCondition* cond); |
| void HandleGenerateLongTest(HCondition* cond); |
| void HandleGenerateLongComparesAndJumps(); |
| void HandleGenerateTest(HCondition* cond); |
| void HandleGenerateConditionGeneric(HCondition* cond); |
| void HandleGenerateEqualLong(HCondition* cond); |
| void HandleGenerateConditionLong(HCondition* cond); |
| void HandleGenerateConditionIntegralOrNonPrimitive(HCondition* cond); |
| void HandleCondition(HCondition* instr); |
| void HandleBinaryOperationLantencies(HBinaryOperation* instr); |
| void HandleBitwiseOperationLantencies(HBinaryOperation* instr); |
| void HandleShiftLatencies(HBinaryOperation* instr); |
| void HandleDivRemConstantIntegralLatencies(int32_t imm); |
| void HandleFieldSetLatencies(HInstruction* instruction, const FieldInfo& field_info); |
| void HandleFieldGetLatencies(HInstruction* instruction, const FieldInfo& field_info); |
| void HandleGenerateDataProcInstruction(bool internal_latency = false); |
| void HandleGenerateDataProc(HDataProcWithShifterOp* instruction); |
| void HandleGenerateLongDataProc(HDataProcWithShifterOp* instruction); |
| |
| // The latency setting for each HInstruction depends on how CodeGenerator may generate code, |
| // latency visitors may query CodeGenerator for such information for accurate latency settings. |
| CodeGeneratorARMType* codegen_; |
| }; |
| |
| class HSchedulerARM : public HScheduler { |
| public: |
| HSchedulerARM(ScopedArenaAllocator* allocator, |
| SchedulingNodeSelector* selector, |
| SchedulingLatencyVisitorARM* arm_latency_visitor) |
| : HScheduler(allocator, arm_latency_visitor, selector) {} |
| ~HSchedulerARM() OVERRIDE {} |
| |
| bool IsSchedulable(const HInstruction* instruction) const OVERRIDE { |
| #define CASE_INSTRUCTION_KIND(type, unused) case \ |
| HInstruction::InstructionKind::k##type: |
| switch (instruction->GetKind()) { |
| FOR_EACH_SCHEDULED_SHARED_INSTRUCTION(CASE_INSTRUCTION_KIND) |
| return true; |
| FOR_EACH_CONCRETE_INSTRUCTION_ARM(CASE_INSTRUCTION_KIND) |
| return true; |
| default: |
| return HScheduler::IsSchedulable(instruction); |
| } |
| #undef CASE_INSTRUCTION_KIND |
| } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(HSchedulerARM); |
| }; |
| |
| } // namespace arm |
| } // namespace art |
| |
| #endif // ART_COMPILER_OPTIMIZING_SCHEDULER_ARM_H_ |