| /* |
| * Copyright (C) 2014 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include <dirent.h> |
| #include <errno.h> |
| #include <string.h> |
| #include <sys/types.h> |
| |
| #include <fstream> |
| #include <map> |
| #include <regex> |
| |
| #include "gtest/gtest.h" |
| |
| #include "jni/quick/calling_convention.h" |
| #include "utils/arm/jni_macro_assembler_arm_vixl.h" |
| #include "utils/assembler_test_base.h" |
| |
| #include "base/hex_dump.h" |
| #include "base/macros.h" |
| #include "base/malloc_arena_pool.h" |
| #include "common_runtime_test.h" |
| |
| namespace art HIDDEN { |
| namespace arm { |
| |
| // Include results file (generated manually) |
| #include "assembler_thumb_test_expected.cc.inc" |
| |
| class ArmVIXLAssemblerTest : public AssemblerTestBase { |
| public: |
| ArmVIXLAssemblerTest() : pool(), allocator(&pool), assembler(&allocator) { } |
| |
| protected: |
| InstructionSet GetIsa() override { return InstructionSet::kThumb2; } |
| |
| void DumpAndCheck(std::vector<uint8_t>& code, const char* testname, const std::string& expected) { |
| #ifndef ART_TARGET_ANDROID |
| std::string obj_file = scratch_dir_->GetPath() + testname + ".o"; |
| WriteElf</*IsElf64=*/false>(obj_file, InstructionSet::kThumb2, code); |
| std::string disassembly; |
| ASSERT_TRUE(Disassemble(obj_file, &disassembly)); |
| |
| // objdump on buildbot seems to sometimes add annotation like in "bne #226 <.text+0x1e8>". |
| // It is unclear why it does not reproduce locally. As work-around, remove the annotation. |
| std::regex annotation_re(" <\\.text\\+\\w+>"); |
| disassembly = std::regex_replace(disassembly, annotation_re, ""); |
| |
| std::string expected2 = "\n" + |
| obj_file + ": file format elf32-littlearm\n\n" |
| "Disassembly of section .text:\n\n" |
| "00000000 <.text>:\n" + |
| expected; |
| EXPECT_EQ(expected2, disassembly); |
| if (expected2 != disassembly) { |
| std::string out = " \"" + Replace(disassembly, "\n", "\\n\"\n \"") + "\""; |
| printf("C++ formatted disassembler output for %s:\n%s\n", testname, out.c_str()); |
| } |
| #endif // ART_TARGET_ANDROID |
| } |
| |
| #define __ assembler. |
| |
| void EmitAndCheck(const char* testname, const char* expected) { |
| __ FinalizeCode(); |
| size_t cs = __ CodeSize(); |
| std::vector<uint8_t> managed_code(cs); |
| MemoryRegion code(&managed_code[0], managed_code.size()); |
| __ CopyInstructions(code); |
| |
| DumpAndCheck(managed_code, testname, expected); |
| } |
| |
| #undef __ |
| |
| #define __ assembler. |
| |
| MallocArenaPool pool; |
| ArenaAllocator allocator; |
| ArmVIXLJNIMacroAssembler assembler; |
| }; |
| |
| TEST_F(ArmVIXLAssemblerTest, VixlJniHelpers) { |
| // Run the test only with Baker read barriers, as the expected |
| // generated code contains a Marking Register refresh instruction. |
| TEST_DISABLED_WITHOUT_BAKER_READ_BARRIERS(); |
| |
| const bool is_static = true; |
| const bool is_synchronized = false; |
| const bool is_fast_native = false; |
| const bool is_critical_native = false; |
| const char* shorty = "IIFII"; |
| |
| std::unique_ptr<JniCallingConvention> jni_conv( |
| JniCallingConvention::Create(&allocator, |
| is_static, |
| is_synchronized, |
| is_fast_native, |
| is_critical_native, |
| shorty, |
| InstructionSet::kThumb2)); |
| std::unique_ptr<ManagedRuntimeCallingConvention> mr_conv( |
| ManagedRuntimeCallingConvention::Create( |
| &allocator, is_static, is_synchronized, shorty, InstructionSet::kThumb2)); |
| const int frame_size(jni_conv->FrameSize()); |
| ArrayRef<const ManagedRegister> callee_save_regs = jni_conv->CalleeSaveRegisters(); |
| |
| const ManagedRegister method_register = ArmManagedRegister::FromCoreRegister(R0); |
| const ManagedRegister hidden_arg_register = ArmManagedRegister::FromCoreRegister(R4); |
| const ManagedRegister scratch_register = ArmManagedRegister::FromCoreRegister(R12); |
| |
| __ BuildFrame(frame_size, mr_conv->MethodRegister(), callee_save_regs); |
| |
| // Spill arguments. |
| mr_conv->ResetIterator(FrameOffset(frame_size)); |
| for (; mr_conv->HasNext(); mr_conv->Next()) { |
| if (mr_conv->IsCurrentParamInRegister()) { |
| size_t size = mr_conv->IsCurrentParamALongOrDouble() ? 8u : 4u; |
| __ Store(mr_conv->CurrentParamStackOffset(), mr_conv->CurrentParamRegister(), size); |
| } |
| } |
| __ IncreaseFrameSize(32); |
| |
| // Loads |
| __ IncreaseFrameSize(4096); |
| __ Load(method_register, FrameOffset(32), 4); |
| __ Load(method_register, FrameOffset(124), 4); |
| __ Load(method_register, FrameOffset(132), 4); |
| __ Load(method_register, FrameOffset(1020), 4); |
| __ Load(method_register, FrameOffset(1024), 4); |
| __ Load(scratch_register, FrameOffset(4092), 4); |
| __ Load(scratch_register, FrameOffset(4096), 4); |
| __ LoadRawPtrFromThread(scratch_register, ThreadOffset32(512)); |
| |
| // Stores |
| __ Store(FrameOffset(32), method_register, 4); |
| __ Store(FrameOffset(124), method_register, 4); |
| __ Store(FrameOffset(132), method_register, 4); |
| __ Store(FrameOffset(1020), method_register, 4); |
| __ Store(FrameOffset(1024), method_register, 4); |
| __ Store(FrameOffset(4092), scratch_register, 4); |
| __ Store(FrameOffset(4096), scratch_register, 4); |
| __ StoreRawPtr(FrameOffset(48), scratch_register); |
| __ StoreStackPointerToThread(ThreadOffset32(512), false); |
| __ StoreStackPointerToThread(ThreadOffset32(512), true); |
| |
| // MoveArguments |
| static constexpr FrameOffset kInvalidReferenceOffset = |
| JNIMacroAssembler<kArmPointerSize>::kInvalidReferenceOffset; |
| static constexpr size_t kNativePointerSize = static_cast<size_t>(kArmPointerSize); |
| // Normal or @FastNative with parameters (Object, long, long, int, Object). |
| // Note: This shall not spill the reference R1 to [sp, #36]. The JNI compiler spills |
| // references in an separate initial pass before moving arguments and creating `jobject`s. |
| ArgumentLocation move_dests1[] = { |
| ArgumentLocation(ArmManagedRegister::FromCoreRegister(R2), kNativePointerSize), |
| ArgumentLocation(FrameOffset(0), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(8), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(16), kVRegSize), |
| ArgumentLocation(FrameOffset(20), kNativePointerSize), |
| }; |
| ArgumentLocation move_srcs1[] = { |
| ArgumentLocation(ArmManagedRegister::FromCoreRegister(R1), kVRegSize), |
| ArgumentLocation(ArmManagedRegister::FromRegisterPair(R2_R3), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(48), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(56), kVRegSize), |
| ArgumentLocation(FrameOffset(60), kVRegSize), |
| }; |
| FrameOffset move_refs1[] { |
| FrameOffset(36), |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(60), |
| }; |
| __ MoveArguments(ArrayRef<ArgumentLocation>(move_dests1), |
| ArrayRef<ArgumentLocation>(move_srcs1), |
| ArrayRef<FrameOffset>(move_refs1)); |
| // @CriticalNative with parameters (long, long, long, int). |
| ArgumentLocation move_dests2[] = { |
| ArgumentLocation(ArmManagedRegister::FromRegisterPair(R0_R1), 2 * kVRegSize), |
| ArgumentLocation(ArmManagedRegister::FromRegisterPair(R2_R3), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(0), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(8), kVRegSize), |
| }; |
| ArgumentLocation move_srcs2[] = { |
| ArgumentLocation(ArmManagedRegister::FromRegisterPair(R2_R3), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(28), kVRegSize), |
| ArgumentLocation(FrameOffset(32), 2 * kVRegSize), |
| ArgumentLocation(FrameOffset(40), kVRegSize), |
| }; |
| FrameOffset move_refs2[] { |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(kInvalidReferenceOffset), |
| FrameOffset(kInvalidReferenceOffset), |
| }; |
| __ MoveArguments(ArrayRef<ArgumentLocation>(move_dests2), |
| ArrayRef<ArgumentLocation>(move_srcs2), |
| ArrayRef<FrameOffset>(move_refs2)); |
| |
| // Other |
| __ Call(method_register, FrameOffset(48)); |
| __ Copy(FrameOffset(48), FrameOffset(44), 4); |
| __ GetCurrentThread(method_register); |
| __ GetCurrentThread(FrameOffset(48)); |
| __ Move(hidden_arg_register, method_register, 4); |
| __ VerifyObject(scratch_register, false); |
| |
| // Note: `CreateJObject()` may need the scratch register IP. Test with another high register. |
| const ManagedRegister high_register = ArmManagedRegister::FromCoreRegister(R11); |
| __ CreateJObject(high_register, FrameOffset(48), high_register, true); |
| __ CreateJObject(high_register, FrameOffset(48), high_register, false); |
| __ CreateJObject(method_register, FrameOffset(48), high_register, true); |
| __ CreateJObject(method_register, FrameOffset(0), high_register, true); |
| __ CreateJObject(method_register, FrameOffset(1028), high_register, true); |
| __ CreateJObject(high_register, FrameOffset(1028), high_register, true); |
| |
| std::unique_ptr<JNIMacroLabel> exception_slow_path = __ CreateLabel(); |
| __ ExceptionPoll(exception_slow_path.get()); |
| |
| // Push the target out of range of branch emitted by ExceptionPoll. |
| for (int i = 0; i < 64; i++) { |
| __ Store(FrameOffset(2047), scratch_register, 4); |
| } |
| |
| __ DecreaseFrameSize(4096); |
| __ DecreaseFrameSize(32); |
| __ RemoveFrame(frame_size, callee_save_regs, /* may_suspend= */ true); |
| |
| __ Bind(exception_slow_path.get()); |
| __ DeliverPendingException(); |
| |
| EmitAndCheck("VixlJniHelpers", VixlJniHelpersResults); |
| } |
| |
| #undef __ |
| |
| // TODO: Avoid these macros. |
| #define R0 vixl::aarch32::r0 |
| #define R2 vixl::aarch32::r2 |
| #define R4 vixl::aarch32::r4 |
| #define R12 vixl::aarch32::r12 |
| |
| #define __ assembler.asm_. |
| |
| TEST_F(ArmVIXLAssemblerTest, VixlLoadFromOffset) { |
| __ LoadFromOffset(kLoadWord, R2, R4, 12); |
| __ LoadFromOffset(kLoadWord, R2, R4, 0xfff); |
| __ LoadFromOffset(kLoadWord, R2, R4, 0x1000); |
| __ LoadFromOffset(kLoadWord, R2, R4, 0x1000a4); |
| __ LoadFromOffset(kLoadWord, R2, R4, 0x101000); |
| __ LoadFromOffset(kLoadWord, R4, R4, 0x101000); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R2, R4, 12); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R2, R4, 0xfff); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R2, R4, 0x1000); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R2, R4, 0x1000a4); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R2, R4, 0x101000); |
| __ LoadFromOffset(kLoadUnsignedHalfword, R4, R4, 0x101000); |
| __ LoadFromOffset(kLoadWordPair, R2, R4, 12); |
| __ LoadFromOffset(kLoadWordPair, R2, R4, 0x3fc); |
| __ LoadFromOffset(kLoadWordPair, R2, R4, 0x400); |
| __ LoadFromOffset(kLoadWordPair, R2, R4, 0x400a4); |
| __ LoadFromOffset(kLoadWordPair, R2, R4, 0x40400); |
| __ LoadFromOffset(kLoadWordPair, R4, R4, 0x40400); |
| |
| vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); |
| temps.Exclude(R12); |
| __ LoadFromOffset(kLoadWord, R0, R12, 12); // 32-bit because of R12. |
| temps.Include(R12); |
| __ LoadFromOffset(kLoadWord, R2, R4, 0xa4 - 0x100000); |
| |
| __ LoadFromOffset(kLoadSignedByte, R2, R4, 12); |
| __ LoadFromOffset(kLoadUnsignedByte, R2, R4, 12); |
| __ LoadFromOffset(kLoadSignedHalfword, R2, R4, 12); |
| |
| EmitAndCheck("VixlLoadFromOffset", VixlLoadFromOffsetResults); |
| } |
| |
| TEST_F(ArmVIXLAssemblerTest, VixlStoreToOffset) { |
| __ StoreToOffset(kStoreWord, R2, R4, 12); |
| __ StoreToOffset(kStoreWord, R2, R4, 0xfff); |
| __ StoreToOffset(kStoreWord, R2, R4, 0x1000); |
| __ StoreToOffset(kStoreWord, R2, R4, 0x1000a4); |
| __ StoreToOffset(kStoreWord, R2, R4, 0x101000); |
| __ StoreToOffset(kStoreWord, R4, R4, 0x101000); |
| __ StoreToOffset(kStoreHalfword, R2, R4, 12); |
| __ StoreToOffset(kStoreHalfword, R2, R4, 0xfff); |
| __ StoreToOffset(kStoreHalfword, R2, R4, 0x1000); |
| __ StoreToOffset(kStoreHalfword, R2, R4, 0x1000a4); |
| __ StoreToOffset(kStoreHalfword, R2, R4, 0x101000); |
| __ StoreToOffset(kStoreHalfword, R4, R4, 0x101000); |
| __ StoreToOffset(kStoreWordPair, R2, R4, 12); |
| __ StoreToOffset(kStoreWordPair, R2, R4, 0x3fc); |
| __ StoreToOffset(kStoreWordPair, R2, R4, 0x400); |
| __ StoreToOffset(kStoreWordPair, R2, R4, 0x400a4); |
| __ StoreToOffset(kStoreWordPair, R2, R4, 0x40400); |
| __ StoreToOffset(kStoreWordPair, R4, R4, 0x40400); |
| |
| vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); |
| temps.Exclude(R12); |
| __ StoreToOffset(kStoreWord, R0, R12, 12); // 32-bit because of R12. |
| temps.Include(R12); |
| __ StoreToOffset(kStoreWord, R2, R4, 0xa4 - 0x100000); |
| |
| __ StoreToOffset(kStoreByte, R2, R4, 12); |
| |
| EmitAndCheck("VixlStoreToOffset", VixlStoreToOffsetResults); |
| } |
| |
| #undef __ |
| } // namespace arm |
| } // namespace art |