| /* |
| * Copyright (C) 2014 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_ARCH_ARM64_ASM_SUPPORT_ARM64_S_ |
| #define ART_RUNTIME_ARCH_ARM64_ASM_SUPPORT_ARM64_S_ |
| |
| #include "asm_support_arm64.h" |
| #include "interpreter/cfi_asm_support.h" |
| |
| // Define special registers. |
| |
| // Register holding Thread::Current(). |
| #define xSELF x19 |
| // Frame Pointer |
| #define xFP x29 |
| // Link Register |
| #define xLR x30 |
| // Define the intraprocedural linkage temporary registers. |
| #define xIP0 x16 |
| #define wIP0 w16 |
| #define xIP1 x17 |
| #define wIP1 w17 |
| |
| #ifdef RESERVE_MARKING_REGISTER |
| // Marking Register, holding Thread::Current()->GetIsGcMarking(). |
| #define wMR w20 |
| #endif |
| |
| // Implicit suspend check register. |
| #define xSUSPEND x21 |
| |
| .macro CFI_EXPRESSION_BREG n, b, offset |
| .if (-0x40 <= (\offset)) && ((\offset) < 0x40) |
| CFI_EXPRESSION_BREG_1(\n, \b, \offset) |
| .elseif (-0x2000 <= (\offset)) && ((\offset) < 0x2000) |
| CFI_EXPRESSION_BREG_2(\n, \b, \offset) |
| .else |
| .error "Unsupported offset" |
| .endif |
| .endm |
| |
| .macro CFI_DEF_CFA_BREG_PLUS_UCONST reg, offset, size |
| .if ((\size) < 0) |
| .error "Size should be positive" |
| .endif |
| .if (((\offset) < -0x40) || ((\offset) >= 0x40)) |
| .error "Unsupported offset" |
| .endif |
| .if ((\size) < 0x80) |
| CFI_DEF_CFA_BREG_PLUS_UCONST_1_1(\reg, \offset, \size) |
| .elseif ((\size) < 0x4000) |
| CFI_DEF_CFA_BREG_PLUS_UCONST_1_2(\reg, \offset, \size) |
| .else |
| .error "Unsupported size" |
| .endif |
| .endm |
| |
| .macro CFI_REMEMBER_STATE |
| .cfi_remember_state |
| .endm |
| |
| // The spec is not clear whether the CFA is part of the saved state and tools |
| // differ in the behaviour, so explicitly set the CFA to avoid any ambiguity. |
| // The restored CFA state should match the CFA state during CFI_REMEMBER_STATE. |
| .macro CFI_RESTORE_STATE_AND_DEF_CFA reg, offset |
| .cfi_restore_state |
| .cfi_def_cfa \reg, \offset |
| .endm |
| |
| .macro ENTRY_ALIGNED name, alignment |
| .type \name, #function |
| .hidden \name // Hide this as a global symbol, so we do not incur plt calls. |
| .global \name |
| // ART-compiled functions have OatQuickMethodHeader but assembly funtions do not. |
| // Prefix the assembly code with 0xFFs, which means there is no method header. |
| .byte 0xFF, 0xFF, 0xFF, 0xFF |
| // Cache alignment for function entry. |
| // Use 0xFF as the last 4 bytes of alignment stand for OatQuickMethodHeader. |
| .balign \alignment, 0xFF |
| \name: |
| .cfi_startproc |
| .endm |
| |
| .macro ENTRY name |
| ENTRY_ALIGNED \name, 16 |
| .endm |
| |
| .macro END name |
| .cfi_endproc |
| .size \name, .-\name |
| .endm |
| |
| .macro UNIMPLEMENTED name |
| ENTRY \name |
| brk 0 |
| END \name |
| .endm |
| |
| // Macro to poison (negate) the reference for heap poisoning. |
| .macro POISON_HEAP_REF rRef |
| #ifdef USE_HEAP_POISONING |
| neg \rRef, \rRef |
| #endif // USE_HEAP_POISONING |
| .endm |
| |
| // Macro to unpoison (negate) the reference for heap poisoning. |
| .macro UNPOISON_HEAP_REF rRef |
| #ifdef USE_HEAP_POISONING |
| neg \rRef, \rRef |
| #endif // USE_HEAP_POISONING |
| .endm |
| |
| .macro INCREASE_FRAME frame_adjustment |
| sub sp, sp, #(\frame_adjustment) |
| .cfi_adjust_cfa_offset (\frame_adjustment) |
| .endm |
| |
| .macro DECREASE_FRAME frame_adjustment |
| add sp, sp, #(\frame_adjustment) |
| .cfi_adjust_cfa_offset -(\frame_adjustment) |
| .endm |
| |
| .macro SAVE_REG reg, offset |
| str \reg, [sp, #(\offset)] |
| .cfi_rel_offset \reg, (\offset) |
| .endm |
| |
| .macro RESTORE_REG_BASE base, reg, offset |
| ldr \reg, [\base, #(\offset)] |
| .cfi_restore \reg |
| .endm |
| |
| .macro RESTORE_REG reg, offset |
| RESTORE_REG_BASE sp, \reg, \offset |
| .endm |
| |
| .macro SAVE_TWO_REGS_BASE base, reg1, reg2, offset |
| stp \reg1, \reg2, [\base, #(\offset)] |
| .cfi_rel_offset \reg1, (\offset) |
| .cfi_rel_offset \reg2, (\offset) + 8 |
| .endm |
| |
| .macro SAVE_TWO_REGS reg1, reg2, offset |
| SAVE_TWO_REGS_BASE sp, \reg1, \reg2, \offset |
| .endm |
| |
| .macro RESTORE_TWO_REGS_BASE base, reg1, reg2, offset |
| ldp \reg1, \reg2, [\base, #(\offset)] |
| .cfi_restore \reg1 |
| .cfi_restore \reg2 |
| .endm |
| |
| .macro RESTORE_TWO_REGS reg1, reg2, offset |
| RESTORE_TWO_REGS_BASE sp, \reg1, \reg2, \offset |
| .endm |
| |
| .macro LOAD_RUNTIME_INSTANCE reg |
| #if __has_feature(hwaddress_sanitizer) |
| adrp \reg, :pg_hi21_nc:_ZN3art7Runtime9instance_E |
| #else |
| adrp \reg, _ZN3art7Runtime9instance_E |
| #endif |
| ldr \reg, [\reg, #:lo12:_ZN3art7Runtime9instance_E] |
| .endm |
| |
| // Macro to refresh the Marking Register (W20). |
| // |
| // This macro must be called at the end of functions implementing |
| // entrypoints that possibly (directly or indirectly) perform a |
| // suspend check (before they return). |
| .macro REFRESH_MARKING_REGISTER |
| #ifdef RESERVE_MARKING_REGISTER |
| ldr wMR, [xSELF, #THREAD_IS_GC_MARKING_OFFSET] |
| #endif |
| .endm |
| |
| // Macro to refresh the suspend check register. |
| // |
| // We do not refresh `xSUSPEND` after every transition to Runnable, so there is |
| // a chance that an implicit suspend check loads null to xSUSPEND but before |
| // causing a SIGSEGV at the next implicit suspend check we make a runtime call |
| // that performs the suspend check explicitly. This can cause a spurious fault |
| // without a pending suspend check request but it should be rare and the fault |
| // overhead was already expected when we triggered the suspend check, we just |
| // pay the price later than expected. |
| .macro REFRESH_SUSPEND_CHECK_REGISTER |
| ldr xSUSPEND, [xSELF, #THREAD_SUSPEND_TRIGGER_OFFSET] |
| .endm |
| |
| /* |
| * Macro that sets up the callee save frame to conform with |
| * Runtime::CreateCalleeSaveMethod(kSaveRefsOnly). |
| */ |
| .macro SETUP_SAVE_REFS_ONLY_FRAME |
| // art::Runtime* xIP0 = art::Runtime::instance_; |
| // Our registers aren't intermixed - just spill in order. |
| LOAD_RUNTIME_INSTANCE xIP0 |
| |
| // ArtMethod* xIP0 = Runtime::instance_->callee_save_methods_[kSaveRefOnly]; |
| ldr xIP0, [xIP0, RUNTIME_SAVE_REFS_ONLY_METHOD_OFFSET] |
| |
| INCREASE_FRAME 96 |
| |
| // Ugly compile-time check, but we only have the preprocessor. |
| #if (FRAME_SIZE_SAVE_REFS_ONLY != 96) |
| #error "FRAME_SIZE_SAVE_REFS_ONLY(ARM64) size not as expected." |
| #endif |
| |
| // GP callee-saves. |
| // x20 paired with ArtMethod* - see below. |
| SAVE_TWO_REGS x21, x22, 16 |
| SAVE_TWO_REGS x23, x24, 32 |
| SAVE_TWO_REGS x25, x26, 48 |
| SAVE_TWO_REGS x27, x28, 64 |
| SAVE_TWO_REGS x29, xLR, 80 |
| |
| // Store ArtMethod* Runtime::callee_save_methods_[kSaveRefsOnly]. |
| // Note: We could avoid saving X20 in the case of Baker read |
| // barriers, as it is overwritten by REFRESH_MARKING_REGISTER |
| // later; but it's not worth handling this special case. |
| stp xIP0, x20, [sp] |
| .cfi_rel_offset x20, 8 |
| |
| // Place sp in Thread::Current()->top_quick_frame. |
| mov xIP0, sp |
| str xIP0, [xSELF, # THREAD_TOP_QUICK_FRAME_OFFSET] |
| .endm |
| |
| // TODO: Probably no need to restore registers preserved by aapcs64. |
| .macro RESTORE_SAVE_REFS_ONLY_FRAME |
| // Callee-saves. |
| // Note: Likewise, we could avoid restoring X20 in the case of Baker |
| // read barriers, as it is overwritten by REFRESH_MARKING_REGISTER |
| // later; but it's not worth handling this special case. |
| RESTORE_REG x20, 8 |
| RESTORE_TWO_REGS x21, x22, 16 |
| RESTORE_TWO_REGS x23, x24, 32 |
| RESTORE_TWO_REGS x25, x26, 48 |
| RESTORE_TWO_REGS x27, x28, 64 |
| RESTORE_TWO_REGS x29, xLR, 80 |
| |
| DECREASE_FRAME 96 |
| .endm |
| |
| .macro SETUP_SAVE_REFS_AND_ARGS_FRAME_INTERNAL base |
| // Ugly compile-time check, but we only have the preprocessor. |
| #if (FRAME_SIZE_SAVE_REFS_AND_ARGS != 224) |
| #error "FRAME_SIZE_SAVE_REFS_AND_ARGS(ARM64) size not as expected." |
| #endif |
| |
| // Stack alignment filler [\base, #8]. |
| // FP args. |
| stp d0, d1, [\base, #16] |
| stp d2, d3, [\base, #32] |
| stp d4, d5, [\base, #48] |
| stp d6, d7, [\base, #64] |
| |
| // Core args. |
| stp x1, x2, [\base, #80] |
| stp x3, x4, [\base, #96] |
| stp x5, x6, [\base, #112] |
| |
| // x7, Callee-saves. |
| // Note: We could avoid saving X20 in the case of Baker read |
| // barriers, as it is overwritten by REFRESH_MARKING_REGISTER |
| // later; but it's not worth handling this special case. |
| stp x7, x20, [\base, #128] |
| .cfi_rel_offset x20, 136 |
| SAVE_TWO_REGS_BASE \base, x21, x22, 144 |
| SAVE_TWO_REGS_BASE \base, x23, x24, 160 |
| SAVE_TWO_REGS_BASE \base, x25, x26, 176 |
| SAVE_TWO_REGS_BASE \base, x27, x28, 192 |
| |
| // x29(callee-save) and LR. |
| SAVE_TWO_REGS_BASE \base, x29, xLR, 208 |
| .endm |
| |
| // TODO: Probably no need to restore registers preserved by aapcs64. (That would require |
| // auditing all users to make sure they restore aapcs64 callee-save registers they clobber.) |
| .macro RESTORE_SAVE_REFS_AND_ARGS_FRAME_INTERNAL base |
| // FP args. |
| ldp d0, d1, [\base, #16] |
| ldp d2, d3, [\base, #32] |
| ldp d4, d5, [\base, #48] |
| ldp d6, d7, [\base, #64] |
| |
| // Core args. |
| ldp x1, x2, [\base, #80] |
| ldp x3, x4, [\base, #96] |
| ldp x5, x6, [\base, #112] |
| |
| // x7, callee-saves and LR. |
| // Note: Likewise, we could avoid restoring X20 in the case of Baker |
| // read barriers, as it is overwritten by REFRESH_MARKING_REGISTER |
| // later; but it's not worth handling this special case. |
| ldp x7, x20, [\base, #128] |
| .cfi_restore x20 |
| RESTORE_TWO_REGS_BASE \base, x21, x22, 144 |
| RESTORE_TWO_REGS_BASE \base, x23, x24, 160 |
| RESTORE_TWO_REGS_BASE \base, x25, x26, 176 |
| RESTORE_TWO_REGS_BASE \base, x27, x28, 192 |
| RESTORE_TWO_REGS_BASE \base, x29, xLR, 208 |
| .endm |
| |
| .macro RESTORE_SAVE_REFS_AND_ARGS_FRAME |
| RESTORE_SAVE_REFS_AND_ARGS_FRAME_INTERNAL sp |
| DECREASE_FRAME FRAME_SIZE_SAVE_REFS_AND_ARGS |
| .endm |
| |
| .macro SAVE_ALL_CALLEE_SAVES offset |
| // FP callee-saves. |
| stp d8, d9, [sp, #(0 + \offset)] |
| stp d10, d11, [sp, #(16 + \offset)] |
| stp d12, d13, [sp, #(32 + \offset)] |
| stp d14, d15, [sp, #(48 + \offset)] |
| |
| // GP callee-saves |
| SAVE_TWO_REGS x19, x20, (64 + \offset) |
| SAVE_TWO_REGS x21, x22, (80 + \offset) |
| SAVE_TWO_REGS x23, x24, (96 + \offset) |
| SAVE_TWO_REGS x25, x26, (112 + \offset) |
| SAVE_TWO_REGS x27, x28, (128 + \offset) |
| SAVE_TWO_REGS x29, xLR, (144 + \offset) |
| .endm |
| |
| /* |
| * Macro that sets up the callee save frame to conform with |
| * Runtime::CreateCalleeSaveMethod(kSaveAllCalleeSaves) |
| */ |
| .macro SETUP_SAVE_ALL_CALLEE_SAVES_FRAME |
| // art::Runtime* xIP0 = art::Runtime::instance_; |
| // Our registers aren't intermixed - just spill in order. |
| LOAD_RUNTIME_INSTANCE xIP0 |
| |
| // ArtMethod* xIP0 = Runtime::instance_->callee_save_methods_[kSaveAllCalleeSaves]; |
| ldr xIP0, [xIP0, RUNTIME_SAVE_ALL_CALLEE_SAVES_METHOD_OFFSET] |
| |
| INCREASE_FRAME 176 |
| |
| // Ugly compile-time check, but we only have the preprocessor. |
| #if (FRAME_SIZE_SAVE_ALL_CALLEE_SAVES != 176) |
| #error "FRAME_SIZE_SAVE_ALL_CALLEE_SAVES(ARM64) size not as expected." |
| #endif |
| |
| // Stack alignment filler [sp, #8]. |
| SAVE_ALL_CALLEE_SAVES 16 |
| |
| // Store ArtMethod* Runtime::callee_save_methods_[kSaveAllCalleeSaves]. |
| str xIP0, [sp] |
| // Place sp in Thread::Current()->top_quick_frame. |
| mov xIP0, sp |
| str xIP0, [xSELF, # THREAD_TOP_QUICK_FRAME_OFFSET] |
| .endm |
| |
| /* |
| * Macro that calls through to artDeliverPendingExceptionFromCode, where the pending |
| * exception is Thread::Current()->exception_ when the runtime method frame is ready. |
| */ |
| .macro DELIVER_PENDING_EXCEPTION_FRAME_READY |
| mov x0, xSELF |
| |
| // Point of no return. |
| bl artDeliverPendingExceptionFromCode // artDeliverPendingExceptionFromCode(Thread*) |
| brk 0 // Unreached |
| .endm |
| |
| /* |
| * Macro that calls through to artDeliverPendingExceptionFromCode, where the pending |
| * exception is Thread::Current()->exception_. |
| */ |
| .macro DELIVER_PENDING_EXCEPTION |
| SETUP_SAVE_ALL_CALLEE_SAVES_FRAME |
| DELIVER_PENDING_EXCEPTION_FRAME_READY |
| .endm |
| |
| .macro RETURN_OR_DELIVER_PENDING_EXCEPTION_REG reg |
| ldr \reg, [xSELF, # THREAD_EXCEPTION_OFFSET] // Get exception field. |
| cbnz \reg, 1f |
| ret |
| 1: |
| DELIVER_PENDING_EXCEPTION |
| .endm |
| |
| .macro RETURN_OR_DELIVER_PENDING_EXCEPTION |
| RETURN_OR_DELIVER_PENDING_EXCEPTION_REG xIP0 |
| .endm |
| |
| // Locking is needed for both managed code and JNI stubs. |
| .macro LOCK_OBJECT_FAST_PATH obj, slow_lock, can_be_null |
| // Use scratch registers x8-x11 as temporaries. |
| ldr w9, [xSELF, #THREAD_ID_OFFSET] |
| .if \can_be_null |
| cbz \obj, \slow_lock |
| .endif |
| // Exclusive load/store has no immediate anymore. |
| add x8, \obj, #MIRROR_OBJECT_LOCK_WORD_OFFSET |
| 1: |
| ldaxr w10, [x8] // Acquire needed only in most common case. |
| eor w11, w10, w9 // Prepare the value to store if unlocked |
| // (thread id, count of 0 and preserved read barrier bits), |
| // or prepare to compare thread id for recursive lock check |
| // (lock_word.ThreadId() ^ self->ThreadId()). |
| tst w10, #LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED // Test the non-gc bits. |
| b.ne 2f // Check if unlocked. |
| // Unlocked case - store w11: original lock word plus thread id, preserved read barrier bits. |
| stxr w10, w11, [x8] |
| cbnz w10, 1b // If the store failed, retry. |
| ret |
| 2: // w10: original lock word, w9: thread id, w11: w10 ^ w11 |
| // Check lock word state and thread id together, |
| tst w11, #(LOCK_WORD_STATE_MASK_SHIFTED | LOCK_WORD_THIN_LOCK_OWNER_MASK_SHIFTED) |
| b.ne \slow_lock |
| add w11, w10, #LOCK_WORD_THIN_LOCK_COUNT_ONE // Increment the recursive lock count. |
| tst w11, #LOCK_WORD_THIN_LOCK_COUNT_MASK_SHIFTED // Test the new thin lock count. |
| b.eq \slow_lock // Zero as the new count indicates overflow, go slow path. |
| stxr w10, w11, [x8] |
| cbnz w10, 1b // If the store failed, retry. |
| ret |
| .endm |
| |
| // Unlocking is needed for both managed code and JNI stubs. |
| .macro UNLOCK_OBJECT_FAST_PATH obj, slow_unlock, can_be_null |
| // Use scratch registers x8-x11 as temporaries. |
| ldr w9, [xSELF, #THREAD_ID_OFFSET] |
| .if \can_be_null |
| cbz \obj, \slow_unlock |
| .endif |
| // Exclusive load/store has no immediate anymore. |
| add x8, \obj, #MIRROR_OBJECT_LOCK_WORD_OFFSET |
| 1: |
| #ifndef USE_READ_BARRIER |
| ldr w10, [x8] |
| #else |
| ldxr w10, [x8] // Need to use atomic instructions for read barrier. |
| #endif |
| eor w11, w10, w9 // Prepare the value to store if simply locked |
| // (mostly 0s, and preserved read barrier bits), |
| // or prepare to compare thread id for recursive lock check |
| // (lock_word.ThreadId() ^ self->ThreadId()). |
| tst w11, #LOCK_WORD_GC_STATE_MASK_SHIFTED_TOGGLED // Test the non-gc bits. |
| b.ne 2f // Locked recursively or by other thread? |
| // Transition to unlocked. |
| #ifndef USE_READ_BARRIER |
| stlr w11, [x8] |
| #else |
| stlxr w10, w11, [x8] // Need to use atomic instructions for read barrier. |
| cbnz w10, 1b // If the store failed, retry. |
| #endif |
| ret |
| 2: |
| // Check lock word state and thread id together. |
| tst w11, #(LOCK_WORD_STATE_MASK_SHIFTED | LOCK_WORD_THIN_LOCK_OWNER_MASK_SHIFTED) |
| b.ne \slow_unlock |
| sub w11, w10, #LOCK_WORD_THIN_LOCK_COUNT_ONE // decrement count |
| #ifndef USE_READ_BARRIER |
| str w11, [x8] |
| #else |
| stxr w10, w11, [x8] // Need to use atomic instructions for read barrier. |
| cbnz w10, 1b // If the store failed, retry. |
| #endif |
| ret |
| .endm |
| |
| #endif // ART_RUNTIME_ARCH_ARM64_ASM_SUPPORT_ARM64_S_ |