From 3177e104e727b802770cebe59dd19d708e0942a7 Mon Sep 17 00:00:00 2001 From: Lena Djokic Date: Wed, 28 Feb 2018 11:32:40 +0100 Subject: MIPS: Use sltiu instead of LoadConst32() + sltu Bltu is sltu + bnez so we can use sltiu + bnez without loading constant (if it is 16-bit constant). Additionally, in VisitInvokeInterface() LoadConst32() is moved to before Jalr(T9) so the load can be taken into the delay slot. Test: ./testrunner.py --target --optimizing in QEMU Change-Id: Ic19f251aeba015be38b7d3690e78b2fe59e7c5ae --- compiler/optimizing/code_generator_mips.cc | 11 +++++------ compiler/optimizing/code_generator_mips64.cc | 4 ++-- 2 files changed, 7 insertions(+), 8 deletions(-) (limited to 'compiler') diff --git a/compiler/optimizing/code_generator_mips.cc b/compiler/optimizing/code_generator_mips.cc index d01b895446..87e6d6834b 100644 --- a/compiler/optimizing/code_generator_mips.cc +++ b/compiler/optimizing/code_generator_mips.cc @@ -1929,8 +1929,8 @@ void InstructionCodeGeneratorMIPS::GenerateClassInitializationCheck(SlowPathCode enum_cast(ClassStatus::kInitialized) << (status_lsb_position % kBitsPerByte); __ LoadFromOffset(kLoadUnsignedByte, TMP, class_reg, status_byte_offset); - __ LoadConst32(AT, shifted_initialized_value); - __ Bltu(TMP, AT, slow_path->GetEntryLabel()); + __ Sltiu(TMP, TMP, shifted_initialized_value); + __ Bnez(TMP, slow_path->GetEntryLabel()); // Even if the initialized flag is set, we need to ensure consistent memory ordering. __ Sync(0); __ Bind(slow_path->GetExitLabel()); @@ -7635,10 +7635,6 @@ void InstructionCodeGeneratorMIPS::VisitInvokeInterface(HInvokeInterface* invoke uint32_t class_offset = mirror::Object::ClassOffset().Int32Value(); Offset entry_point = ArtMethod::EntryPointFromQuickCompiledCodeOffset(kMipsPointerSize); - // Set the hidden argument. - __ LoadConst32(invoke->GetLocations()->GetTemp(1).AsRegister(), - invoke->GetDexMethodIndex()); - // temp = object->GetClass(); if (receiver.IsStackSlot()) { __ LoadFromOffset(kLoadWord, temp, SP, receiver.GetStackIndex()); @@ -7663,6 +7659,9 @@ void InstructionCodeGeneratorMIPS::VisitInvokeInterface(HInvokeInterface* invoke __ LoadFromOffset(kLoadWord, temp, temp, method_offset); // T9 = temp->GetEntryPoint(); __ LoadFromOffset(kLoadWord, T9, temp, entry_point.Int32Value()); + // Set the hidden argument. + __ LoadConst32(invoke->GetLocations()->GetTemp(1).AsRegister(), + invoke->GetDexMethodIndex()); // T9(); __ Jalr(T9); __ NopIfNoReordering(); diff --git a/compiler/optimizing/code_generator_mips64.cc b/compiler/optimizing/code_generator_mips64.cc index e3529f178a..985ac2ca55 100644 --- a/compiler/optimizing/code_generator_mips64.cc +++ b/compiler/optimizing/code_generator_mips64.cc @@ -1773,8 +1773,8 @@ void InstructionCodeGeneratorMIPS64::GenerateClassInitializationCheck(SlowPathCo enum_cast(ClassStatus::kInitialized) << (status_lsb_position % kBitsPerByte); __ LoadFromOffset(kLoadUnsignedByte, TMP, class_reg, status_byte_offset); - __ LoadConst32(AT, shifted_initialized_value); - __ Bltuc(TMP, AT, slow_path->GetEntryLabel()); + __ Sltiu(TMP, TMP, shifted_initialized_value); + __ Bnezc(TMP, slow_path->GetEntryLabel()); // Even if the initialized flag is set, we need to ensure consistent memory ordering. __ Sync(0); __ Bind(slow_path->GetExitLabel()); -- cgit v1.2.3-59-g8ed1b