diff options
Diffstat (limited to 'compiler/optimizing')
| -rw-r--r-- | compiler/optimizing/code_generator_mips64.cc | 66 | ||||
| -rw-r--r-- | compiler/optimizing/optimizing_cfi_test_expected.inc | 148 |
2 files changed, 89 insertions, 125 deletions
diff --git a/compiler/optimizing/code_generator_mips64.cc b/compiler/optimizing/code_generator_mips64.cc index 5246dbc5cb..c82533bc7d 100644 --- a/compiler/optimizing/code_generator_mips64.cc +++ b/compiler/optimizing/code_generator_mips64.cc @@ -558,26 +558,21 @@ void CodeGeneratorMIPS64::GenerateFrameEntry() { return; } - // Make sure the frame size isn't unreasonably large. Per the various APIs - // it looks like it should always be less than 2GB in size, which allows - // us using 32-bit signed offsets from the stack pointer. - if (GetFrameSize() > 0x7FFFFFFF) - LOG(FATAL) << "Stack frame larger than 2GB"; + // Make sure the frame size isn't unreasonably large. + if (GetFrameSize() > GetStackOverflowReservedBytes(kMips64)) { + LOG(FATAL) << "Stack frame larger than " << GetStackOverflowReservedBytes(kMips64) << " bytes"; + } // Spill callee-saved registers. - // Note that their cumulative size is small and they can be indexed using - // 16-bit offsets. - - // TODO: increment/decrement SP in one step instead of two or remove this comment. - uint32_t ofs = FrameEntrySpillSize(); + uint32_t ofs = GetFrameSize(); __ IncreaseFrameSize(ofs); for (int i = arraysize(kCoreCalleeSaves) - 1; i >= 0; --i) { GpuRegister reg = kCoreCalleeSaves[i]; if (allocated_registers_.ContainsCoreRegister(reg)) { ofs -= kMips64DoublewordSize; - __ Sd(reg, SP, ofs); + __ StoreToOffset(kStoreDoubleword, reg, SP, ofs); __ cfi().RelOffset(DWARFReg(reg), ofs); } } @@ -586,23 +581,16 @@ void CodeGeneratorMIPS64::GenerateFrameEntry() { FpuRegister reg = kFpuCalleeSaves[i]; if (allocated_registers_.ContainsFloatingPointRegister(reg)) { ofs -= kMips64DoublewordSize; - __ Sdc1(reg, SP, ofs); + __ StoreFpuToOffset(kStoreDoubleword, reg, SP, ofs); __ cfi().RelOffset(DWARFReg(reg), ofs); } } - // Allocate the rest of the frame and store the current method pointer - // at its end. - - __ IncreaseFrameSize(GetFrameSize() - FrameEntrySpillSize()); - // Save the current method if we need it. Note that we do not // do this in HCurrentMethod, as the instruction might have been removed // in the SSA graph. if (RequiresCurrentMethod()) { - static_assert(IsInt<16>(kCurrentMethodStackOffset), - "kCurrentMethodStackOffset must fit into int16_t"); - __ Sd(kMethodRegisterArgument, SP, kCurrentMethodStackOffset); + __ StoreToOffset(kStoreDoubleword, kMethodRegisterArgument, SP, kCurrentMethodStackOffset); } if (GetGraph()->HasShouldDeoptimizeFlag()) { @@ -615,42 +603,32 @@ void CodeGeneratorMIPS64::GenerateFrameExit() { __ cfi().RememberState(); if (!HasEmptyFrame()) { - // Deallocate the rest of the frame. - - __ DecreaseFrameSize(GetFrameSize() - FrameEntrySpillSize()); - // Restore callee-saved registers. - // Note that their cumulative size is small and they can be indexed using - // 16-bit offsets. - - // TODO: increment/decrement SP in one step instead of two or remove this comment. - uint32_t ofs = 0; - - for (size_t i = 0; i < arraysize(kFpuCalleeSaves); ++i) { - FpuRegister reg = kFpuCalleeSaves[i]; - if (allocated_registers_.ContainsFloatingPointRegister(reg)) { - __ Ldc1(reg, SP, ofs); - ofs += kMips64DoublewordSize; + // For better instruction scheduling restore RA before other registers. + uint32_t ofs = GetFrameSize(); + for (int i = arraysize(kCoreCalleeSaves) - 1; i >= 0; --i) { + GpuRegister reg = kCoreCalleeSaves[i]; + if (allocated_registers_.ContainsCoreRegister(reg)) { + ofs -= kMips64DoublewordSize; + __ LoadFromOffset(kLoadDoubleword, reg, SP, ofs); __ cfi().Restore(DWARFReg(reg)); } } - for (size_t i = 0; i < arraysize(kCoreCalleeSaves); ++i) { - GpuRegister reg = kCoreCalleeSaves[i]; - if (allocated_registers_.ContainsCoreRegister(reg)) { - __ Ld(reg, SP, ofs); - ofs += kMips64DoublewordSize; + for (int i = arraysize(kFpuCalleeSaves) - 1; i >= 0; --i) { + FpuRegister reg = kFpuCalleeSaves[i]; + if (allocated_registers_.ContainsFloatingPointRegister(reg)) { + ofs -= kMips64DoublewordSize; + __ LoadFpuFromOffset(kLoadDoubleword, reg, SP, ofs); __ cfi().Restore(DWARFReg(reg)); } } - DCHECK_EQ(ofs, FrameEntrySpillSize()); - __ DecreaseFrameSize(ofs); + __ DecreaseFrameSize(GetFrameSize()); } - __ Jr(RA); - __ Nop(); + __ Jic(RA, 0); __ cfi().RestoreState(); __ cfi().DefCFAOffset(GetFrameSize()); diff --git a/compiler/optimizing/optimizing_cfi_test_expected.inc b/compiler/optimizing/optimizing_cfi_test_expected.inc index d84fe6ccff..60af2b4201 100644 --- a/compiler/optimizing/optimizing_cfi_test_expected.inc +++ b/compiler/optimizing/optimizing_cfi_test_expected.inc @@ -174,53 +174,45 @@ static constexpr uint8_t expected_cfi_kMips[] = { // 0x00000034: .cfi_def_cfa_offset: 64 static constexpr uint8_t expected_asm_kMips64[] = { - 0xD8, 0xFF, 0xBD, 0x67, 0x20, 0x00, 0xBF, 0xFF, 0x18, 0x00, 0xB1, 0xFF, - 0x10, 0x00, 0xB0, 0xFF, 0x08, 0x00, 0xB9, 0xF7, 0x00, 0x00, 0xB8, 0xF7, - 0xE8, 0xFF, 0xBD, 0x67, 0x18, 0x00, 0xBD, 0x67, - 0x00, 0x00, 0xB8, 0xD7, 0x08, 0x00, 0xB9, 0xD7, 0x10, 0x00, 0xB0, 0xDF, - 0x18, 0x00, 0xB1, 0xDF, 0x20, 0x00, 0xBF, 0xDF, 0x28, 0x00, 0xBD, 0x67, - 0x09, 0x00, 0xE0, 0x03, 0x00, 0x00, 0x00, 0x00, + 0xC0, 0xFF, 0xBD, 0x67, 0x38, 0x00, 0xBF, 0xFF, 0x30, 0x00, 0xB1, 0xFF, + 0x28, 0x00, 0xB0, 0xFF, 0x20, 0x00, 0xB9, 0xF7, 0x18, 0x00, 0xB8, 0xF7, + 0x38, 0x00, 0xBF, 0xDF, 0x30, 0x00, 0xB1, 0xDF, 0x28, 0x00, 0xB0, 0xDF, + 0x20, 0x00, 0xB9, 0xD7, 0x18, 0x00, 0xB8, 0xD7, 0x40, 0x00, 0xBD, 0x67, + 0x00, 0x00, 0x1F, 0xD8, }; - static constexpr uint8_t expected_cfi_kMips64[] = { - 0x44, 0x0E, 0x28, 0x44, 0x9F, 0x02, 0x44, 0x91, 0x04, 0x44, 0x90, 0x06, - 0x44, 0xB9, 0x08, 0x44, 0xB8, 0x0A, 0x44, 0x0E, 0x40, 0x0A, 0x44, - 0x0E, 0x28, 0x44, 0xF8, 0x44, 0xF9, 0x44, 0xD0, 0x44, 0xD1, 0x44, 0xDF, - 0x44, 0x0E, 0x00, 0x48, 0x0B, 0x0E, 0x40, + 0x44, 0x0E, 0x40, 0x44, 0x9F, 0x02, 0x44, 0x91, 0x04, 0x44, 0x90, 0x06, + 0x44, 0xB9, 0x08, 0x44, 0xB8, 0x0A, 0x0A, 0x44, 0xDF, 0x44, 0xD1, 0x44, + 0xD0, 0x44, 0xF9, 0x44, 0xF8, 0x44, 0x0E, 0x00, 0x44, 0x0B, 0x0E, 0x40, }; -// 0x00000000: daddiu r29, r29, -40 -// 0x00000004: .cfi_def_cfa_offset: 40 -// 0x00000004: sd r31, +32(r29) +// 0x00000000: daddiu r29, r29, -64 +// 0x00000004: .cfi_def_cfa_offset: 64 +// 0x00000004: sd r31, +56(r29) // 0x00000008: .cfi_offset: r31 at cfa-8 -// 0x00000008: sd r17, +24(r29) +// 0x00000008: sd r17, +48(r29) // 0x0000000c: .cfi_offset: r17 at cfa-16 -// 0x0000000c: sd r16, +16(r29) +// 0x0000000c: sd r16, +40(r29) // 0x00000010: .cfi_offset: r16 at cfa-24 -// 0x00000010: sdc1 f25, +8(r29) +// 0x00000010: sdc1 f25, +32(r29) // 0x00000014: .cfi_offset: r57 at cfa-32 -// 0x00000014: sdc1 f24, +0(r29) +// 0x00000014: sdc1 f24, +24(r29) // 0x00000018: .cfi_offset: r56 at cfa-40 -// 0x00000018: daddiu r29, r29, -24 -// 0x0000001c: .cfi_def_cfa_offset: 64 -// 0x0000001c: .cfi_remember_state -// 0x0000001c: daddiu r29, r29, 24 -// 0x00000020: .cfi_def_cfa_offset: 40 -// 0x00000020: ldc1 f24, +0(r29) -// 0x00000024: .cfi_restore: r56 -// 0x00000024: ldc1 f25, +8(r29) +// 0x00000018: .cfi_remember_state +// 0x00000018: ld r31, +56(r29) +// 0x0000001c: .cfi_restore: r31 +// 0x0000001c: ld r17, +48(r29) +// 0x00000020: .cfi_restore: r17 +// 0x00000020: ld r16, +40(r29) +// 0x00000024: .cfi_restore: r16 +// 0x00000024: ldc1 f25, +32(r29) // 0x00000028: .cfi_restore: r57 -// 0x00000028: ld r16, +16(r29) -// 0x0000002c: .cfi_restore: r16 -// 0x0000002c: ld r17, +24(r29) -// 0x00000030: .cfi_restore: r17 -// 0x00000030: ld r31, +32(r29) -// 0x00000034: .cfi_restore: r31 -// 0x00000034: daddiu r29, r29, 40 -// 0x00000038: .cfi_def_cfa_offset: 0 -// 0x00000038: jr r31 -// 0x0000003c: nop -// 0x00000040: .cfi_restore_state -// 0x00000040: .cfi_def_cfa_offset: 64 +// 0x00000028: ldc1 f24, +24(r29) +// 0x0000002c: .cfi_restore: r56 +// 0x0000002c: daddiu r29, r29, 64 +// 0x00000030: .cfi_def_cfa_offset: 0 +// 0x00000030: jic r31, 0 +// 0x00000034: .cfi_restore_state +// 0x00000034: .cfi_def_cfa_offset: 64 static constexpr uint8_t expected_asm_kThumb2_adjust[] = { #ifdef ART_USE_OLD_ARM_BACKEND @@ -403,58 +395,52 @@ static constexpr uint8_t expected_cfi_kMips_adjust[] = { // 0x00020060: .cfi_def_cfa_offset: 64 static constexpr uint8_t expected_asm_kMips64_adjust_head[] = { - 0xD8, 0xFF, 0xBD, 0x67, 0x20, 0x00, 0xBF, 0xFF, 0x18, 0x00, 0xB1, 0xFF, - 0x10, 0x00, 0xB0, 0xFF, 0x08, 0x00, 0xB9, 0xF7, 0x00, 0x00, 0xB8, 0xF7, - 0xE8, 0xFF, 0xBD, 0x67, 0x02, 0x00, 0xA6, 0x60, - 0x02, 0x00, 0x3E, 0xEC, 0x0C, 0x00, 0x01, 0xD8, + 0xC0, 0xFF, 0xBD, 0x67, 0x38, 0x00, 0xBF, 0xFF, 0x30, 0x00, 0xB1, 0xFF, + 0x28, 0x00, 0xB0, 0xFF, 0x20, 0x00, 0xB9, 0xF7, 0x18, 0x00, 0xB8, 0xF7, + 0x02, 0x00, 0xA6, 0x60, 0x02, 0x00, 0x3E, 0xEC, 0x0C, 0x00, 0x01, 0xD8, }; static constexpr uint8_t expected_asm_kMips64_adjust_tail[] = { - 0x18, 0x00, 0xBD, 0x67, 0x00, 0x00, 0xB8, 0xD7, 0x08, 0x00, 0xB9, 0xD7, - 0x10, 0x00, 0xB0, 0xDF, 0x18, 0x00, 0xB1, 0xDF, 0x20, 0x00, 0xBF, 0xDF, - 0x28, 0x00, 0xBD, 0x67, 0x09, 0x00, 0xE0, 0x03, 0x00, 0x00, 0x00, 0x00, + 0x38, 0x00, 0xBF, 0xDF, 0x30, 0x00, 0xB1, 0xDF, 0x28, 0x00, 0xB0, 0xDF, + 0x20, 0x00, 0xB9, 0xD7, 0x18, 0x00, 0xB8, 0xD7, 0x40, 0x00, 0xBD, 0x67, + 0x00, 0x00, 0x1F, 0xD8, }; static constexpr uint8_t expected_cfi_kMips64_adjust[] = { - 0x44, 0x0E, 0x28, 0x44, 0x9F, 0x02, 0x44, 0x91, 0x04, 0x44, 0x90, 0x06, - 0x44, 0xB9, 0x08, 0x44, 0xB8, 0x0A, 0x44, 0x0E, 0x40, 0x04, 0x10, 0x00, - 0x02, 0x00, 0x0A, 0x44, 0x0E, 0x28, 0x44, 0xF8, 0x44, 0xF9, 0x44, 0xD0, - 0x44, 0xD1, 0x44, 0xDF, 0x44, 0x0E, 0x00, 0x48, 0x0B, 0x0E, 0x40, + 0x44, 0x0E, 0x40, 0x44, 0x9F, 0x02, 0x44, 0x91, 0x04, 0x44, 0x90, 0x06, + 0x44, 0xB9, 0x08, 0x44, 0xB8, 0x0A, 0x04, 0x10, 0x00, 0x02, 0x00, 0x0A, + 0x44, 0xDF, 0x44, 0xD1, 0x44, 0xD0, 0x44, 0xF9, 0x44, 0xF8, 0x44, 0x0E, + 0x00, 0x44, 0x0B, 0x0E, 0x40, }; -// 0x00000000: daddiu r29, r29, -40 -// 0x00000004: .cfi_def_cfa_offset: 40 -// 0x00000004: sd r31, +32(r29) +// 0x00000000: daddiu r29, r29, -64 +// 0x00000004: .cfi_def_cfa_offset: 64 +// 0x00000004: sd r31, +56(r29) // 0x00000008: .cfi_offset: r31 at cfa-8 -// 0x00000008: sd r17, +24(r29) +// 0x00000008: sd r17, +48(r29) // 0x0000000c: .cfi_offset: r17 at cfa-16 -// 0x0000000c: sd r16, +16(r29) +// 0x0000000c: sd r16, +40(r29) // 0x00000010: .cfi_offset: r16 at cfa-24 -// 0x00000010: sdc1 f25, +8(r29) +// 0x00000010: sdc1 f25, +32(r29) // 0x00000014: .cfi_offset: r57 at cfa-32 -// 0x00000014: sdc1 f24, +0(r29) +// 0x00000014: sdc1 f24, +24(r29) // 0x00000018: .cfi_offset: r56 at cfa-40 -// 0x00000018: daddiu r29, r29, -24 -// 0x0000001c: .cfi_def_cfa_offset: 64 -// 0x0000001c: bnec r5, r6, 0x0000002c ; +12 -// 0x00000020: auipc r1, 2 -// 0x00000024: jic r1, 12 ; b 0x00020030 ; +131080 -// 0x00000028: nop +// 0x00000018: bnec r5, r6, 0x00000024 ; +12 +// 0x0000001c: auipc r1, 2 +// 0x00000020: jic r1, 12 ; bc 0x00020028 ; +131080 +// 0x00000024: nop // ... -// 0x00020028: nop -// 0x0002002c: .cfi_remember_state -// 0x0002002c: daddiu r29, r29, 24 -// 0x00020030: .cfi_def_cfa_offset: 40 -// 0x00020030: ldc1 f24, +0(r29) -// 0x00020034: .cfi_restore: r56 -// 0x00020034: ldc1 f25, +8(r29) +// 0x00020024: nop +// 0x00020028: .cfi_remember_state +// 0x00020028: ld r31, +56(r29) +// 0x0002002c: .cfi_restore: r31 +// 0x0002002c: ld r17, +48(r29) +// 0x00020030: .cfi_restore: r17 +// 0x00020030: ld r16, +40(r29) +// 0x00020034: .cfi_restore: r16 +// 0x00020034: ldc1 f25, +32(r29) // 0x00020038: .cfi_restore: r57 -// 0x00020038: ld r16, +16(r29) -// 0x0002003c: .cfi_restore: r16 -// 0x0002003c: ld r17, +24(r29) -// 0x00020040: .cfi_restore: r17 -// 0x00020040: ld r31, +32(r29) -// 0x00020044: .cfi_restore: r31 -// 0x00020044: daddiu r29, r29, 40 -// 0x00020047: .cfi_def_cfa_offset: 0 -// 0x00020048: jr r31 -// 0x0002004c: nop -// 0x00020050: .cfi_restore_state -// 0x00020050: .cfi_def_cfa_offset: 64 +// 0x00020038: ldc1 f24, +24(r29) +// 0x0002003c: .cfi_restore: r56 +// 0x0002003c: daddiu r29, r29, 64 +// 0x00020040: .cfi_def_cfa_offset: 0 +// 0x00020040: jic r31, 0 +// 0x00020044: .cfi_restore_state +// 0x00020044: .cfi_def_cfa_offset: 64 |