blob: 602273f13a9da9b8c8a315ed6674882b3b76c573 [file] [log] [blame]
%def header():
/*
* Copyright (C) 2020 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* This is a #include, not a %include, because we want the C pre-processor
* to expand the macros into assembler assignment statements.
*/
#include "asm_support.h"
#include "arch/arm64/asm_support_arm64.S"
/**
* ARM64 Runtime register usage conventions.
*
* r0 : w0 is 32-bit return register and x0 is 64-bit.
* r0-r7 : Argument registers.
* r8-r15 : Caller save registers (used as temporary registers).
* r16-r17: Also known as ip0-ip1, respectively. Used as scratch registers by
* the linker, by the trampolines and other stubs (the compiler uses
* these as temporary registers).
* r18 : Reserved for platform (SCS, shadow call stack)
* r19 : Pointer to thread-local storage.
* r20-r29: Callee save registers.
* r30 : (lr) is reserved (the link register).
* rsp : (sp) is reserved (the stack pointer).
* rzr : (zr) is reserved (the zero register).
*
* Floating-point registers
* v0-v31
*
* v0 : s0 is return register for singles (32-bit) and d0 for doubles (64-bit).
* This is analogous to the C/C++ (hard-float) calling convention.
* v0-v7 : Floating-point argument registers in both Dalvik and C/C++ conventions.
* Also used as temporary and codegen scratch registers.
*
* v0-v7 and v16-v31 : Caller save registers (used as temporary registers).
* v8-v15 : bottom 64-bits preserved across C calls (d8-d15 are preserved).
*
* v16-v31: Used as codegen temp/scratch.
* v8-v15 : Can be used for promotion.
*
* Must maintain 16-byte stack alignment.
*
* Nterp notes:
*
* The following registers have fixed assignments:
*
* reg nick purpose
* x19 xSELF self (Thread) pointer
* x20 wMR marking register
* x21 xSUSPEND suspend check register
* x29 xFP interpreted frame pointer, used for accessing locals and args
* x22 xPC interpreted program counter, used for fetching instructions
* x23 xINST first 16-bit code unit of current instruction
* x24 xIBASE interpreted instruction base pointer, used for computed goto
* x25 xREFS base of object references of dex registers.
* x16 ip scratch reg
* x17 ip2 scratch reg (used by macros)
*
* Macros are provided for common operations. They MUST NOT alter unspecified registers or
* condition codes.
*/
/* single-purpose registers, given names for clarity */
#define CFI_DEX 22 // DWARF register number of the register holding dex-pc (xPC).
#define CFI_TMP 0 // DWARF register number of the first argument register (r0).
#define xPC x22
#define xINST x23
#define wINST w23
#define xIBASE x24
#define xREFS x25
#define CFI_REFS 25
#define ip x16
#define ip2 x17
#define wip w16
#define wip2 w17
// To avoid putting ifdefs arond the use of wMR, make sure it's defined.
// IsNterpSupported returns false for configurations that don't have wMR (typically CMS).
#ifndef wMR
#define wMR w20
#endif
// Temporary registers while setting up a frame.
#define xNEW_FP x26
#define xNEW_REFS x27
#define CFI_NEW_REFS 27
// +8 for the ArtMethod of the caller.
#define OFFSET_TO_FIRST_ARGUMENT_IN_STACK (CALLEE_SAVES_SIZE + 8)
/*
* Fetch the next instruction from xPC into wINST. Does not advance xPC.
*/
.macro FETCH_INST
ldrh wINST, [xPC]
.endm
/*
* Fetch the next instruction from the specified offset. Advances xPC
* to point to the next instruction. "count" is in 16-bit code units.
*
* Because of the limited size of immediate constants on ARM, this is only
* suitable for small forward movements (i.e. don't try to implement "goto"
* with this).
*
* This must come AFTER anything that can throw an exception, or the
* exception catch may miss. (This also implies that it must come after
* EXPORT_PC.)
*/
.macro FETCH_ADVANCE_INST count
ldrh wINST, [xPC, #((\count)*2)]!
.endm
/*
* Similar to FETCH_ADVANCE_INST, but does not update xPC. Used to load
* xINST ahead of possible exception point. Be sure to manually advance xPC
* later.
*/
.macro PREFETCH_INST count
ldrh wINST, [xPC, #((\count)*2)]
.endm
/* Advance xPC by some number of code units. */
.macro ADVANCE count
add xPC, xPC, #((\count)*2)
.endm
/*
* Fetch a half-word code unit from an offset past the current PC. The
* "count" value is in 16-bit code units. Does not advance xPC.
*
* The "_S" variant works the same but treats the value as signed.
*/
.macro FETCH reg, count
ldrh \reg, [xPC, #((\count)*2)]
.endm
.macro FETCH_S reg, count
ldrsh \reg, [xPC, #((\count)*2)]
.endm
/*
* Fetch one byte from an offset past the current PC. Pass in the same
* "count" as you would for FETCH, and an additional 0/1 indicating which
* byte of the halfword you want (lo/hi).
*/
.macro FETCH_B reg, count, byte
ldrb \reg, [xPC, #((\count)*2+(\byte))]
.endm
/*
* Put the instruction's opcode field into the specified register.
*/
.macro GET_INST_OPCODE reg
and \reg, xINST, #255
.endm
/*
* Begin executing the opcode in _reg. Clobbers reg
*/
.macro GOTO_OPCODE reg
add \reg, xIBASE, \reg, lsl #${handler_size_bits}
br \reg
.endm
/*
* Get/set the 32-bit value from a Dalvik register.
*/
.macro GET_VREG reg, vreg
ldr \reg, [xFP, \vreg, uxtw #2]
.endm
.macro GET_VREG_OBJECT reg, vreg
ldr \reg, [xREFS, \vreg, uxtw #2]
.endm
.macro SET_VREG reg, vreg
str \reg, [xFP, \vreg, uxtw #2]
str wzr, [xREFS, \vreg, uxtw #2]
.endm
.macro SET_VREG_OBJECT reg, vreg
str \reg, [xFP, \vreg, uxtw #2]
str \reg, [xREFS, \vreg, uxtw #2]
.endm
.macro SET_VREG_FLOAT reg, vreg
str \reg, [xFP, \vreg, uxtw #2]
str wzr, [xREFS, \vreg, uxtw #2]
.endm
/*
* Get/set the 64-bit value from a Dalvik register.
*/
.macro LOAD_SCALED_VREG_MASK scaled_mask_reg, unscaled_mask
mov \scaled_mask_reg, \unscaled_mask << 2
.endm
.macro EXTRACT_SCALED_VREG scaled_vreg, scaled_mask_reg, src_reg, lsb
.if \lsb < 2
and \scaled_vreg, \scaled_mask_reg, \src_reg, lsl #(2 - \lsb)
.else
and \scaled_vreg, \scaled_mask_reg, \src_reg, lsr #(\lsb - 2)
.endif
.endm
.macro GET_VREG_WIDE_PRESCALED reg, scaled_vreg
ldr \reg, [xFP, \scaled_vreg, uxtw]
.endm
.macro GET_VREG_WIDE reg, vreg
lsl wip2, \vreg, #2
GET_VREG_WIDE_PRESCALED \reg, wip2
.endm
.macro SET_VREG_WIDE_PRESCALED reg, scaled_vreg
str \reg, [xFP, \scaled_vreg, uxtw]
str xzr, [xREFS, \scaled_vreg, uxtw]
.endm
.macro SET_VREG_WIDE reg, vreg
lsl wip2, \vreg, #2
SET_VREG_WIDE_PRESCALED \reg, wip2
.endm
.macro GET_VREG_DOUBLE_PRESCALED reg, scaled_vreg
GET_VREG_WIDE_PRESCALED \reg, \scaled_vreg
.endm
.macro GET_VREG_DOUBLE reg, vreg
GET_VREG_WIDE \reg, \vreg
.endm
.macro SET_VREG_DOUBLE reg, vreg
SET_VREG_WIDE \reg, \vreg
.endm
/*
* Get the 32-bit value from a Dalvik register and sign-extend to 64-bit.
* Used to avoid an extra instruction in int-to-long.
*/
.macro GET_VREG_S reg, vreg
ldrsw \reg, [xFP, \vreg, uxtw #2]
.endm
// An assembly entry for nterp.
.macro OAT_ENTRY name
.type \name, #function
.hidden \name
.global \name
.balign 16
\name:
.endm
.macro SIZE name
.size \name, .-\name
.endm
.macro NAME_START name
.type \name, #function
.hidden \name // Hide this as a global symbol, so we do not incur plt calls.
.global \name
/* Cache alignment for function entry */
.balign 16
\name:
.endm
.macro NAME_END name
SIZE \name
.endm
// Macro for defining entrypoints into runtime. We don't need to save registers
// (we're not holding references there), but there is no
// kDontSave runtime method. So just use the kSaveRefsOnly runtime method.
.macro NTERP_TRAMPOLINE name, helper
ENTRY \name
SETUP_SAVE_REFS_ONLY_FRAME
bl \helper
RESTORE_SAVE_REFS_ONLY_FRAME
REFRESH_MARKING_REGISTER
ldr xIP0, [xSELF, # THREAD_EXCEPTION_OFFSET] // Get exception field.
cbnz xIP0, nterp_deliver_pending_exception
ret
END \name
.endm
.macro CLEAR_STATIC_VOLATILE_MARKER reg
and \reg, \reg, #-2
.endm
.macro CLEAR_INSTANCE_VOLATILE_MARKER reg
neg \reg, \reg
.endm
.macro EXPORT_PC
str xPC, [xREFS, #-16]
.endm
.macro BRANCH
add xPC, xPC, wINST, sxtw #1 // update xPC
// Update method counter and do a suspend check if the branch is negative or zero.
cmp wINST, #0
b.le 2f
1:
FETCH wINST, 0 // load wINST
GET_INST_OPCODE ip // extract opcode from wINST
GOTO_OPCODE ip // jump to next instruction
2:
ldr x0, [sp]
ldrh w2, [x0, #ART_METHOD_HOTNESS_COUNT_OFFSET]
#if (NTERP_HOTNESS_VALUE != 0)
#error Expected 0 for hotness value
#endif
// If the counter is at zero, handle this in the runtime.
cbz w2, NterpHandleHotnessOverflow
add x2, x2, #-1
strh w2, [x0, #ART_METHOD_HOTNESS_COUNT_OFFSET]
DO_SUSPEND_CHECK continue_label=1b
b 1b
.endm
// Uses x12, x13, and x14 as temporaries.
.macro FETCH_CODE_ITEM_INFO code_item, registers, outs, ins, load_ins
tbz \code_item, #0, 4f
and \code_item, \code_item, #-2 // Remove the extra bit that marks it's a compact dex file
ldrh w13, [\code_item, #COMPACT_CODE_ITEM_FIELDS_OFFSET]
ubfx \registers, w13, #COMPACT_CODE_ITEM_REGISTERS_SIZE_SHIFT, #4
ubfx \outs, w13, #COMPACT_CODE_ITEM_OUTS_SIZE_SHIFT, #4
.if \load_ins
ubfx \ins, w13, #COMPACT_CODE_ITEM_INS_SIZE_SHIFT, #4
.else
ubfx w14, w13, #COMPACT_CODE_ITEM_INS_SIZE_SHIFT, #4
add \registers, \registers, w14
.endif
ldrh w13, [\code_item, #COMPACT_CODE_ITEM_FLAGS_OFFSET]
tst w13, #COMPACT_CODE_ITEM_REGISTERS_INS_OUTS_FLAGS
b.eq 3f
sub x14, \code_item, #4
tst w13, #COMPACT_CODE_ITEM_INSNS_FLAG
csel x14, x14, \code_item, ne
tbz w13, #COMPACT_CODE_ITEM_REGISTERS_BIT, 1f
ldrh w12, [x14, #-2]!
add \registers, \registers, w12
1:
tbz w13, #COMPACT_CODE_ITEM_INS_BIT, 2f
ldrh w12, [x14, #-2]!
.if \load_ins
add \ins, \ins, w12
.else
add \registers, \registers, w12
.endif
2:
tbz w13, #COMPACT_CODE_ITEM_OUTS_BIT, 3f
ldrh w12, [x14, #-2]!
add \outs, \outs, w12
3:
.if \load_ins
add \registers, \registers, \ins
.endif
add \code_item, \code_item, #COMPACT_CODE_ITEM_INSNS_OFFSET
b 5f
4:
// Fetch dex register size.
ldrh \registers, [\code_item, #CODE_ITEM_REGISTERS_SIZE_OFFSET]
// Fetch outs size.
ldrh \outs, [\code_item, #CODE_ITEM_OUTS_SIZE_OFFSET]
.if \load_ins
ldrh \ins, [\code_item, #CODE_ITEM_INS_SIZE_OFFSET]
.endif
add \code_item, \code_item, #CODE_ITEM_INSNS_OFFSET
5:
.endm
.macro TEST_IF_MARKING label
cbnz wMR, \label
.endm
// Setup the stack to start executing the method. Expects:
// - x0 to contain the ArtMethod
//
// Outputs
// - ip contains the dex registers size
// - x28 contains the old stack pointer.
// - \code_item is replaced with a pointer to the instructions
// - if load_ins is 1, w15 contains the ins
//
// Uses ip, ip2, x12, x13, x14 as temporaries.
.macro SETUP_STACK_FRAME code_item, refs, fp, cfi_refs, load_ins
FETCH_CODE_ITEM_INFO \code_item, wip, wip2, w15, \load_ins
// Compute required frame size: ((2 * ip) + ip2) * 4 + 24
// 24 is for saving the previous frame, pc, and method being executed.
add x14, ip, ip
add x14, x14, ip2
lsl x14, x14, #2
add x14, x14, #24
// Compute new stack pointer in x14
sub x14, sp, x14
// Alignment
and x14, x14, #-16
// Set reference and dex registers, align to pointer size for previous frame and dex pc.
add \refs, x14, ip2, lsl #2
add \refs, \refs, 28
and \refs, \refs, #(-__SIZEOF_POINTER__)
add \fp, \refs, ip, lsl #2
// Now setup the stack pointer.
mov x28, sp
.cfi_def_cfa_register x28
mov sp, x14
str x28, [\refs, #-8]
CFI_DEF_CFA_BREG_PLUS_UCONST \cfi_refs, -8, CALLEE_SAVES_SIZE
// Put nulls in reference frame.
cbz ip, 2f
mov ip2, \refs
1:
str xzr, [ip2], #8 // May clear vreg[0].
cmp ip2, \fp
b.lo 1b
2:
// Save the ArtMethod.
str x0, [sp]
.endm
// Increase method hotness and do suspend check before starting executing the method.
.macro START_EXECUTING_INSTRUCTIONS
ldr x0, [sp]
ldrh w2, [x0, #ART_METHOD_HOTNESS_COUNT_OFFSET]
#if (NTERP_HOTNESS_VALUE != 0)
#error Expected 0 for hotness value
#endif
// If the counter is at zero, handle this in the runtime.
cbz w2, 3f
add x2, x2, #-1
strh w2, [x0, #ART_METHOD_HOTNESS_COUNT_OFFSET]
1:
DO_SUSPEND_CHECK continue_label=2f
2:
FETCH_INST
GET_INST_OPCODE ip
GOTO_OPCODE ip
3:
CHECK_AND_UPDATE_SHARED_MEMORY_METHOD if_hot=4f, if_not_hot=1b
4:
mov x1, xzr
mov x2, xFP
bl nterp_hot_method
b 2b
.endm
.macro SPILL_ALL_CALLEE_SAVES
INCREASE_FRAME CALLEE_SAVES_SIZE
// Note: we technically don't need to save x19 and x20,
// but the runtime will expect those values to be there when unwinding
// (see Arm64Context::DoLongJump checking for the thread register).
SAVE_ALL_CALLEE_SAVES 0
.endm
.macro RESTORE_ALL_CALLEE_SAVES
// FP callee-saves
ldp d8, d9, [sp, #0]
ldp d10, d11, [sp, #16]
ldp d12, d13, [sp, #32]
ldp d14, d15, [sp, #48]
// GP callee-saves.
// No need to restore x19 (it's always the thread), and
// don't restore x20 (the marking register) as it may have been updated,
// don't restore x21 (the suspend check register) as it may have been updated.
RESTORE_REG x22, 88
RESTORE_TWO_REGS x23, x24, 96
RESTORE_TWO_REGS x25, x26, 112
RESTORE_TWO_REGS x27, x28, 128
RESTORE_TWO_REGS x29, lr, 144
DECREASE_FRAME CALLEE_SAVES_SIZE
.endm
.macro SPILL_ALL_ARGUMENTS
stp x0, x1, [sp, #-128]!
stp x2, x3, [sp, #16]
stp x4, x5, [sp, #32]
stp x6, x7, [sp, #48]
stp d0, d1, [sp, #64]
stp d2, d3, [sp, #80]
stp d4, d5, [sp, #96]
stp d6, d7, [sp, #112]
.endm
.macro RESTORE_ALL_ARGUMENTS
ldp x2, x3, [sp, #16]
ldp x4, x5, [sp, #32]
ldp x6, x7, [sp, #48]
ldp d0, d1, [sp, #64]
ldp d2, d3, [sp, #80]
ldp d4, d5, [sp, #96]
ldp d6, d7, [sp, #112]
ldp x0, x1, [sp], #128
.endm
// Helper to setup the stack after doing a nterp to nterp call. This will setup:
// - xNEW_FP: the new pointer to dex registers
// - xNEW_REFS: the new pointer to references
// - xPC: the new PC pointer to execute
// - x2: value in instruction to decode the number of arguments.
// - x3: first dex register
// - x4: top of dex register array
//
// The method expects:
// - x0 to contain the ArtMethod
// - x8 to contain the code item
.macro SETUP_STACK_FOR_INVOKE
// We do the same stack overflow check as the compiler. See CanMethodUseNterp
// in how we limit the maximum nterp frame size.
sub x16, sp, #STACK_OVERFLOW_RESERVED_BYTES
ldr wzr, [x16]
// Spill all callee saves to have a consistent stack frame whether we
// are called by compiled code or nterp.
SPILL_ALL_CALLEE_SAVES
// Setup the frame.
SETUP_STACK_FRAME x8, xNEW_REFS, xNEW_FP, CFI_NEW_REFS, load_ins=0
// Make x4 point to the top of the dex register array.
add x4, xNEW_FP, ip, uxtx #2
// Fetch instruction information before replacing xPC.
// TODO: move this down to the method that uses it, fetching it directly from wINST.
FETCH_B w2, 0, 1
// TODO: we could avoid this as instance invokes already fetch it.
FETCH w3, 2
// Set the dex pc pointer.
mov xPC, x8
CFI_DEFINE_DEX_PC_WITH_OFFSET(CFI_TMP, CFI_DEX, 0)
.endm
// Setup arguments based on a non-range nterp to nterp call, and start executing
// the method. We expect:
// - xNEW_FP: the new pointer to dex registers
// - xNEW_REFS: the new pointer to references
// - xPC: the new PC pointer to execute
// - x2: number of arguments (bits 4-7), 5th argument if any (bits 0-3)
// - x3: first dex register
// - x4: top of dex register array
// - x1: receiver if non-static.
.macro SETUP_NON_RANGE_ARGUMENTS_AND_EXECUTE is_static=0, is_string_init=0
// /* op vA, vB, {vC...vG} */
asr ip2, x2, #4
cbz ip2, 6f
mov ip, #-4
cmp ip2, #2
b.lt 1f
b.eq 2f
cmp ip2, #4
b.lt 3f
b.eq 4f
// We use a decrementing ip to store references relative
// to xNEW_FP and dex registers relative to x4
//
// TODO: We could set up ip as the number of registers (this can be an additional output from
// SETUP_STACK_FOR_INVOKE) and then just decrement it by one before copying each arg.
// Maybe even introduce macros NEW_VREG_ADDRESS/NEW_VREG_REF_ADDRESS.
5:
and x2, x2, #15
GET_VREG_OBJECT w5, w2
str w5, [xNEW_FP, ip]
GET_VREG w5, w2
str w5, [x4, ip]
sub ip, ip, #4
4:
asr x2, x3, #12
GET_VREG_OBJECT w5, w2
str w5, [xNEW_FP, ip]
GET_VREG w5, w2
str w5, [x4, ip]
sub ip, ip, #4
3:
ubfx x2, x3, #8, #4
GET_VREG_OBJECT w5, w2
str w5, [xNEW_FP, ip]
GET_VREG w5, w2
str w5, [x4, ip]
sub ip, ip, #4
2:
ubfx x2, x3, #4, #4
GET_VREG_OBJECT w5, w2
str w5, [xNEW_FP, ip]
GET_VREG w5, w2
str w5, [x4, ip]
.if !\is_string_init
sub ip, ip, #4
.endif
1:
.if \is_string_init
// Ignore the first argument
.elseif \is_static
and x2, x3, #0xf
GET_VREG_OBJECT w5, w2
str w5, [xNEW_FP, ip]
GET_VREG w5, w2
str w5, [x4, ip]
.else
str w1, [xNEW_FP, ip]
str w1, [x4, ip]
.endif
6:
// Start executing the method.
mov xFP, xNEW_FP
mov xREFS, xNEW_REFS
CFI_DEF_CFA_BREG_PLUS_UCONST CFI_REFS, -8, CALLEE_SAVES_SIZE
START_EXECUTING_INSTRUCTIONS
.endm
// Setup arguments based on a range nterp to nterp call, and start executing
// the method.
// - xNEW_FP: the new pointer to dex registers
// - xNEW_REFS: the new pointer to references
// - xPC: the new PC pointer to execute
// - x2: number of arguments
// - x3: first dex register
// - x4: top of dex register array
// - x1: receiver if non-static.
//
// Uses ip, ip2, x5, x6 as temporaries.
.macro SETUP_RANGE_ARGUMENTS_AND_EXECUTE is_static=0, is_string_init=0
mov ip, #-4
.if \is_string_init
// Ignore the first argument
sub x2, x2, #1
add x3, x3, #1
.elseif !\is_static
sub x2, x2, #1
add x3, x3, #1
.endif
cbz x2, 2f
add ip2, xREFS, x3, lsl #2 // pointer to first argument in reference array
add ip2, ip2, x2, lsl #2 // pointer to last argument in reference array
add x5, xFP, x3, lsl #2 // pointer to first argument in register array
add x6, x5, x2, lsl #2 // pointer to last argument in register array
1:
ldr w7, [ip2, #-4]!
str w7, [xNEW_FP, ip]
sub x2, x2, 1
ldr w7, [x6, #-4]!
str w7, [x4, ip]
sub ip, ip, 4
cbnz x2, 1b
2:
.if \is_string_init
// Ignore first argument
.elseif !\is_static
str w1, [xNEW_FP, ip]
str w1, [x4, ip]
.endif
mov xFP, xNEW_FP
mov xREFS, xNEW_REFS
CFI_DEF_CFA_BREG_PLUS_UCONST CFI_REFS, -8, CALLEE_SAVES_SIZE
START_EXECUTING_INSTRUCTIONS
.endm
.macro GET_SHORTY dest, is_interface, is_polymorphic, is_custom
stp x0, x1, [sp, #-16]!
.if \is_polymorphic
ldr x0, [sp, #16]
mov x1, xPC
bl NterpGetShortyFromInvokePolymorphic
.elseif \is_custom
ldr x0, [sp, #16]
mov x1, xPC
bl NterpGetShortyFromInvokeCustom
.elseif \is_interface
ldr x0, [sp, #16]
FETCH w1, 1
bl NterpGetShortyFromMethodId
.else
bl NterpGetShorty
.endif
mov \dest, x0
ldp x0, x1, [sp], #16
.endm
.macro GET_SHORTY_SLOW_PATH dest, is_interface
// Save all registers that can hold arguments in the fast path.
stp x0, x1, [sp, #-32]!
str w2, [sp, #16]
str s0, [sp, #20]
.if \is_interface
ldr x0, [sp, #32]
FETCH w1, 1
bl NterpGetShortyFromMethodId
.else
bl NterpGetShorty
.endif
mov \dest, x0
ldr w2, [sp, #16]
ldr s0, [sp, #20]
ldp x0, x1, [sp], #32
.endm
// Input: x0 contains the ArtMethod
// Output: x8 contains the code item
.macro GET_CODE_ITEM
ldr x8, [x0, #ART_METHOD_DATA_OFFSET_64]
.endm
.macro DO_ENTRY_POINT_CHECK call_compiled_code
// On entry, the method is x0, the instance is x1
adr x2, ExecuteNterpImpl
ldr x3, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
cmp x2, x3
b.ne \call_compiled_code
.endm
.macro UPDATE_REGISTERS_FOR_STRING_INIT old_value, new_value
mov wip, wzr
1:
GET_VREG_OBJECT wip2, wip
cmp wip2, \old_value
b.ne 2f
SET_VREG_OBJECT \new_value, wip
2:
add wip, wip, #1
add ip2, xREFS, wip, uxtw #2
cmp ip2, xFP
b.ne 1b
.endm
// Puts the next floating point argument into the expected register,
// fetching values based on a non-range invoke.
// Uses ip and ip2.
.macro LOOP_OVER_SHORTY_LOADING_FPS dreg, sreg, inst, shorty, arg_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #68 // if (wip == 'D') goto FOUND_DOUBLE
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto FOUND_FLOAT
b.eq 3f
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
// Handle extra argument in arg array taken by a long.
cmp wip, #74 // if (wip != 'J') goto LOOP
b.ne 1b
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 1b // goto LOOP
2: // FOUND_DOUBLE
and ip, \inst, #0xf
GET_VREG wip, wip
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
cmp \arg_index, #4
b.eq 5f
and ip2, \inst, #0xf
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 6f
5:
// TODO: Extract from wINST here and below? (Requires using a different register
// in the COMMON_INVOKE_NON_RANGE.)
FETCH_B wip2, 0, 1
and wip2, wip2, #0xf
6:
GET_VREG wip2, wip2
add ip, ip, ip2, lsl #32
fmov \dreg, ip
b 4f
3: // FOUND_FLOAT
cmp \arg_index, #4
b.eq 7f
and ip, \inst, #0xf
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 8f
7:
FETCH_B wip, 0, 1
and wip, wip, #0xf
8:
GET_VREG \sreg, wip
4:
.endm
// Puts the next int/long/object argument in the expected register,
// fetching values based on a non-range invoke.
// Uses ip and ip2.
.macro LOOP_OVER_SHORTY_LOADING_GPRS gpr_reg64, gpr_reg32, inst, shorty, arg_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #74 // if (wip == 'J') goto FOUND_LONG
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto SKIP_FLOAT
b.eq 3f
cmp wip, #68 // if (wip == 'D') goto SKIP_DOUBLE
b.eq 4f
cmp \arg_index, #4
b.eq 7f
and ip, \inst, #0xf
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 8f
7:
FETCH_B wip, 0, 1
and wip, wip, #0xf
8:
GET_VREG \gpr_reg32, wip
b 5f
2: // FOUND_LONG
and ip, \inst, #0xf
GET_VREG wip, wip
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
cmp \arg_index, #4
b.eq 9f
and ip2, \inst, #0xf
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 10f
9:
FETCH_B wip2, 0, 1
and wip2, wip2, #0xf
10:
GET_VREG wip2, wip2
add \gpr_reg64, ip, ip2, lsl #32
b 5f
3: // SKIP_FLOAT
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 1b
4: // SKIP_DOUBLE
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
cmp \arg_index, #4
b.eq 1b
lsr \inst, \inst, #4
add \arg_index, \arg_index, #1
b 1b
5:
.endm
.macro SETUP_RETURN_VALUE shorty
ldrb wip, [\shorty]
cmp ip, #68 // Test if result type char == 'D'.
b.eq 1f
cmp ip, #70 // Test if result type char == 'F'.
b.ne 2f
fmov w0, s0
b 2f
1:
fmov x0, d0
2:
.endm
.macro COMMON_INVOKE_NON_RANGE is_static=0, is_interface=0, suffix="", is_string_init=0, is_polymorphic=0, is_custom=0
.if \is_polymorphic
// We always go to compiled code for polymorphic calls.
.elseif \is_custom
// We always go to compiled code for custom calls.
.else
DO_ENTRY_POINT_CHECK .Lcall_compiled_code_\suffix
GET_CODE_ITEM
.if \is_string_init
bl nterp_to_nterp_string_init_non_range
.elseif \is_static
bl nterp_to_nterp_static_non_range
.else
bl nterp_to_nterp_instance_non_range
.endif
b .Ldone_return_\suffix
.endif
.Lcall_compiled_code_\suffix:
.if \is_polymorphic
// No fast path for polymorphic calls.
.elseif \is_custom
// No fast path for custom calls.
.elseif \is_string_init
// No fast path for string.init.
.else
ldr wip, [x0, #ART_METHOD_ACCESS_FLAGS_OFFSET]
tbz wip, #ART_METHOD_NTERP_INVOKE_FAST_PATH_FLAG_BIT, .Lfast_path_with_few_args_\suffix
FETCH_B wip2, 0, 1
asr ip, ip2, #4
.if \is_static
cbz ip, .Linvoke_fast_path_\suffix
.else
cmp ip, #1
b.eq .Linvoke_fast_path_\suffix
.endif
FETCH w8, 2
cmp ip, #2
.if \is_static
b.lt .Lone_arg_fast_path_\suffix
.endif
b.eq .Ltwo_args_fast_path_\suffix
cmp ip, #4
b.lt .Lthree_args_fast_path_\suffix
b.eq .Lfour_args_fast_path_\suffix
and ip, ip2, #15
GET_VREG w5, wip
.Lfour_args_fast_path_\suffix:
asr ip, x8, #12
GET_VREG w4, wip
.Lthree_args_fast_path_\suffix:
ubfx ip, x8, #8, #4
GET_VREG w3, wip
.Ltwo_args_fast_path_\suffix:
ubfx ip, x8, #4, #4
GET_VREG w2, wip
.Lone_arg_fast_path_\suffix:
.if \is_static
and ip, x8, #0xf
GET_VREG w1, wip
.else
// First argument already in w1.
.endif
.Linvoke_fast_path_\suffix:
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
FETCH_ADVANCE_INST 3
GET_INST_OPCODE ip
GOTO_OPCODE ip
.Lfast_path_with_few_args_\suffix:
// Fast path when we have zero or one argument (modulo 'this'). If there
// is one argument, we can put it in both floating point and core register.
FETCH_B w2, 0, 1
.if \is_static
cmp w2, #(2 << 4)
.else
cmp w2, #(3 << 4)
.endif
b.ge .Lget_shorty_\suffix
.if \is_static
tbz w2, #4, .Linvoke_with_few_args_\suffix
.else
tbnz w2, #4, .Linvoke_with_few_args_\suffix
.endif
FETCH w2, 2
.if \is_static
and w2, w2, #0xf // dex register of first argument
GET_VREG w1, w2
fmov s0, w1
.else
ubfx x2, x2, #4, #4 // dex register of second argument
GET_VREG w2, w2
fmov s0, w2
.endif
.Linvoke_with_few_args_\suffix:
// Check if the next instruction is move-result or move-result-wide.
// If it is, we fetch the shorty and jump to the regular invocation.
FETCH w27, 3
and ip, x27, #0xfe
cmp ip, #0x0a
b.eq .Lget_shorty_and_invoke_\suffix
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
# TODO: Use some other register for shorty and prefetch the instruction directly to wINST.
mov xINST, x27
ADVANCE 3
GET_INST_OPCODE ip
GOTO_OPCODE ip
.Lget_shorty_and_invoke_\suffix:
GET_SHORTY_SLOW_PATH xINST, \is_interface
b .Lgpr_setup_finished_\suffix
.endif
.Lget_shorty_\suffix:
GET_SHORTY xINST, \is_interface, \is_polymorphic, \is_custom
// From this point:
// - xINST contains shorty (in callee-save to switch over return value after call).
// - x0 contains method
// - x1 contains 'this' pointer for instance method.
// - for interface calls, x26 contains the interface method.
add x9, xINST, #1 // shorty + 1 ; ie skip return arg character
FETCH w11, 2 // arguments
.if \is_string_init
lsr x11, x11, #4
mov x10, #1 // ignore first argument
.elseif \is_static
mov x10, xzr // arg_index
.else
lsr x11, x11, #4
mov x10, #1 // ignore first argument
.endif
LOOP_OVER_SHORTY_LOADING_FPS d0, s0, x11, x9, x10, .Lxmm_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_FPS d1, s1, x11, x9, x10, .Lxmm_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_FPS d2, s2, x11, x9, x10, .Lxmm_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_FPS d3, s3, x11, x9, x10, .Lxmm_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_FPS d4, s4, x11, x9, x10, .Lxmm_setup_finished_\suffix
.Lxmm_setup_finished_\suffix:
add x9, xINST, #1 // shorty + 1 ; ie skip return arg character
FETCH w11, 2 // arguments
.if \is_string_init
lsr x11, x11, #4
mov x10, #1 // ignore first argument
LOOP_OVER_SHORTY_LOADING_GPRS x1, w1, x11, x9, x10, .Lgpr_setup_finished_\suffix
.elseif \is_static
mov x10, xzr // arg_index
LOOP_OVER_SHORTY_LOADING_GPRS x1, w1, x11, x9, x10, .Lgpr_setup_finished_\suffix
.else
lsr x11, x11, #4
mov x10, #1 // ignore first argument
.endif
LOOP_OVER_SHORTY_LOADING_GPRS x2, w2, x11, x9, x10, .Lgpr_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_GPRS x3, w3, x11, x9, x10, .Lgpr_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_GPRS x4, w4, x11, x9, x10, .Lgpr_setup_finished_\suffix
LOOP_OVER_SHORTY_LOADING_GPRS x5, w5, x11, x9, x10, .Lgpr_setup_finished_\suffix
.Lgpr_setup_finished_\suffix:
.if \is_polymorphic
bl art_quick_invoke_polymorphic
.elseif \is_custom
bl art_quick_invoke_custom
.else
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
.endif
SETUP_RETURN_VALUE xINST
.Ldone_return_\suffix:
/* resume execution of caller */
.if \is_string_init
FETCH w11, 2 // arguments
and x11, x11, #0xf
GET_VREG w1, w11
UPDATE_REGISTERS_FOR_STRING_INIT w1, w0
.endif
.if \is_polymorphic
FETCH_ADVANCE_INST 4
.else
FETCH_ADVANCE_INST 3
.endif
GET_INST_OPCODE ip
GOTO_OPCODE ip
.endm
// Puts the next floating point argument into the expected register,
// fetching values based on a range invoke.
// Uses ip as temporary.
.macro LOOP_RANGE_OVER_SHORTY_LOADING_FPS dreg, sreg, shorty, arg_index, stack_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #68 // if (wip == 'D') goto FOUND_DOUBLE
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto FOUND_FLOAT
b.eq 3f
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
// Handle extra argument in arg array taken by a long.
cmp wip, #74 // if (wip != 'J') goto LOOP
b.ne 1b
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b // goto LOOP
2: // FOUND_DOUBLE
GET_VREG_DOUBLE \dreg, \arg_index
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 4f
3: // FOUND_FLOAT
GET_VREG \sreg, \arg_index
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
4:
.endm
// Puts the next floating point argument into the expected stack slot,
// fetching values based on a range invoke.
// Uses ip as temporary.
//
// TODO: We could just copy all the vregs to the stack slots in a simple loop
// without looking at the shorty at all. (We could also drop
// the "stack_index" from the macros for loading registers.) We could also do
// that conditionally if argument word count > 6; otherwise we know that all
// args fit into registers.
.macro LOOP_RANGE_OVER_FPs shorty, arg_index, stack_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #68 // if (wip == 'D') goto FOUND_DOUBLE
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto FOUND_FLOAT
b.eq 3f
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
// Handle extra argument in arg array taken by a long.
cmp wip, #74 // if (wip != 'J') goto LOOP
b.ne 1b
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b // goto LOOP
2: // FOUND_DOUBLE
GET_VREG_WIDE ip, \arg_index
add ip2, sp, \stack_index, uxtw #2
str ip, [ip2]
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 1b
3: // FOUND_FLOAT
GET_VREG wip, \arg_index
str wip, [sp, \stack_index, uxtw #2]
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b
.endm
// Puts the next int/long/object argument in the expected register,
// fetching values based on a range invoke.
// Uses ip as temporary.
.macro LOOP_RANGE_OVER_SHORTY_LOADING_GPRS reg64, reg32, shorty, arg_index, stack_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #74 // if (wip == 'J') goto FOUND_LONG
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto SKIP_FLOAT
b.eq 3f
cmp wip, #68 // if (wip == 'D') goto SKIP_DOUBLE
b.eq 4f
GET_VREG \reg32, \arg_index
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 5f
2: // FOUND_LONG
GET_VREG_WIDE \reg64, \arg_index
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 5f
3: // SKIP_FLOAT
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b
4: // SKIP_DOUBLE
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 1b
5:
.endm
// Puts the next int/long/object argument in the expected stack slot,
// fetching values based on a range invoke.
// Uses ip as temporary.
.macro LOOP_RANGE_OVER_INTs shorty, arg_index, stack_index, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #74 // if (wip == 'J') goto FOUND_LONG
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto SKIP_FLOAT
b.eq 3f
cmp wip, #68 // if (wip == 'D') goto SKIP_DOUBLE
b.eq 4f
GET_VREG wip, \arg_index
str wip, [sp, \stack_index, uxtw #2]
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b
2: // FOUND_LONG
GET_VREG_WIDE ip, \arg_index
add ip2, sp, \stack_index, uxtw #2
str ip, [ip2]
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 1b
3: // SKIP_FLOAT
add \arg_index, \arg_index, #1
add \stack_index, \stack_index, #1
b 1b
4: // SKIP_DOUBLE
add \arg_index, \arg_index, #2
add \stack_index, \stack_index, #2
b 1b
.endm
.macro COMMON_INVOKE_RANGE is_static=0, is_interface=0, suffix="", is_string_init=0, is_polymorphic=0, is_custom=0
.if \is_polymorphic
// We always go to compiled code for polymorphic calls.
.elseif \is_custom
// We always go to compiled code for custom calls.
.else
DO_ENTRY_POINT_CHECK .Lcall_compiled_code_range_\suffix
GET_CODE_ITEM
.if \is_string_init
bl nterp_to_nterp_string_init_range
.elseif \is_static
bl nterp_to_nterp_static_range
.else
bl nterp_to_nterp_instance_range
.endif
b .Ldone_return_range_\suffix
.endif
.Lcall_compiled_code_range_\suffix:
.if \is_polymorphic
// No fast path for polymorphic calls.
.elseif \is_custom
// No fast path for custom calls.
.elseif \is_string_init
// No fast path for string.init.
.else
ldr wip, [x0, #ART_METHOD_ACCESS_FLAGS_OFFSET]
tbz wip, #ART_METHOD_NTERP_INVOKE_FAST_PATH_FLAG_BIT, .Lfast_path_with_few_args_range_\suffix
FETCH_B wip2, 0, 1 // Number of arguments
.if \is_static
cbz ip2, .Linvoke_fast_path_range_\suffix
.else
cmp ip2, #1
b.eq .Linvoke_fast_path_range_\suffix
.endif
FETCH wip, 2 // dex register of first argument
add x8, xFP, wip, uxtw #2 // location of first dex register value
cmp ip2, #2
.if \is_static
b.lt .Lone_arg_fast_path_range_\suffix
.endif
b.eq .Ltwo_args_fast_path_range_\suffix
cmp ip2, #4
b.lt .Lthree_args_fast_path_range_\suffix
b.eq .Lfour_args_fast_path_range_\suffix
cmp ip2, #6
b.lt .Lfive_args_fast_path_range_\suffix
b.eq .Lsix_args_fast_path_range_\suffix
cmp ip2, #7
b.eq .Lseven_args_fast_path_range_\suffix
// Setup x8 to point to the stack location of parameters we do not need
// to put parameters in.
add x9, sp, #8 // Add space for the ArtMethod
.Lloop_over_fast_path_range_\suffix:
sub ip2, ip2, #1
ldr wip, [x8, ip2, lsl #2]
str wip, [x9, ip2, lsl #2]
cmp ip2, #7
b.ne .Lloop_over_fast_path_range_\suffix
.Lseven_args_fast_path_range_\suffix:
ldr w7, [x8, #24]
.Lsix_args_fast_path_range_\suffix:
ldr w6, [x8, #20]
.Lfive_args_fast_path_range_\suffix:
ldr w5, [x8, #16]
.Lfour_args_fast_path_range_\suffix:
ldr w4, [x8, #12]
.Lthree_args_fast_path_range_\suffix:
ldr w3, [x8, #8]
.Ltwo_args_fast_path_range_\suffix:
ldr w2, [x8, #4]
.Lone_arg_fast_path_range_\suffix:
.if \is_static
ldr w1, [x8, #0]
.else
// First argument already in w1.
.endif
.Linvoke_fast_path_range_\suffix:
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
FETCH_ADVANCE_INST 3
GET_INST_OPCODE ip
GOTO_OPCODE ip
.Lfast_path_with_few_args_range_\suffix:
// Fast path when we have zero or one argument (modulo 'this'). If there
// is one argument, we can put it in both floating point and core register.
FETCH_B w2, 0, 1 // number of arguments
.if \is_static
cmp w2, #1
.else
cmp w2, #2
.endif
b.lt .Linvoke_with_few_args_range_\suffix
b.ne .Lget_shorty_range_\suffix
FETCH w3, 2 // dex register of first argument
.if \is_static
GET_VREG w1, w3
fmov s0, w1
.else
add w3, w3, #1 // Add 1 for next argument
GET_VREG w2, w3
fmov s0, w2
.endif
.Linvoke_with_few_args_range_\suffix:
// Check if the next instruction is move-result or move-result-wide.
// If it is, we fetch the shorty and jump to the regular invocation.
FETCH w27, 3
and ip, x27, #0xfe
cmp ip, #0x0a
b.eq .Lget_shorty_and_invoke_range_\suffix
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
mov xINST, x27
ADVANCE 3
GET_INST_OPCODE ip
GOTO_OPCODE ip
.Lget_shorty_and_invoke_range_\suffix:
GET_SHORTY_SLOW_PATH xINST, \is_interface
b .Lgpr_setup_finished_range_\suffix
.endif
.Lget_shorty_range_\suffix:
GET_SHORTY xINST, \is_interface, \is_polymorphic, \is_custom
// From this point:
// - xINST contains shorty (in callee-save to switch over return value after call).
// - x0 contains method
// - x1 contains 'this' pointer for instance method.
// - for interface calls, x26 contains the interface method.
add x9, xINST, #1 // shorty + 1 ; ie skip return arg character
FETCH w10, 2 // arguments
.if \is_string_init
add x10, x10, #1 // arg start index
mov x11, #1 // index in stack
.elseif \is_static
mov x11, xzr // index in stack
.else
add x10, x10, #1 // arg start index
mov x11, #1 // index in stack
.endif
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d0, s0, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d1, s1, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d2, s2, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d3, s3, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d4, s4, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d5, s5, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d6, s6, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_FPS d7, s7, x9, w10, w11, .Lxmm_setup_finished_range_\suffix
// Store in the outs array (stored above the ArtMethod in the stack)
add x11, x11, #2 // Add two words for the ArtMethod stored before the outs.
LOOP_RANGE_OVER_FPs x9, w10, w11, .Lxmm_setup_finished_range_\suffix
.Lxmm_setup_finished_range_\suffix:
add x9, xINST, #1 // shorty + 1 ; ie skip return arg character
FETCH w10, 2 // arguments
.if \is_string_init
add x10, x10, #1 // arg start index
mov x11, #1 // index in stack
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x1, w1, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
.elseif \is_static
mov x11, xzr // index in stack
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x1, w1, x9, w10, w11 .Lgpr_setup_finished_range_\suffix
.else
add x10, x10, #1 // arg start index
mov x11, #1 // index in stack
.endif
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x2, w2, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x3, w3, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x4, w4, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x5, w5, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x6, w6, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
LOOP_RANGE_OVER_SHORTY_LOADING_GPRS x7, w7, x9, w10, w11, .Lgpr_setup_finished_range_\suffix
// Store in the outs array (stored above the ArtMethod in the stack)
add x11, x11, #2 // Add two words for the ArtMethod stored before the outs.
LOOP_RANGE_OVER_INTs x9, w10, w11, .Lgpr_setup_finished_range_\suffix
.Lgpr_setup_finished_range_\suffix:
.if \is_polymorphic
bl art_quick_invoke_polymorphic
.elseif \is_custom
bl art_quick_invoke_custom
.else
.if \is_interface
// Setup hidden argument.
mov ip2, x26
.endif
ldr lr, [x0, #ART_METHOD_QUICK_CODE_OFFSET_64]
blr lr
.endif
SETUP_RETURN_VALUE xINST
.Ldone_return_range_\suffix:
/* resume execution of caller */
.if \is_string_init
FETCH w11, 2 // arguments
GET_VREG w1, w11
UPDATE_REGISTERS_FOR_STRING_INIT w1, w0
.endif
.if \is_polymorphic
FETCH_ADVANCE_INST 4
.else
FETCH_ADVANCE_INST 3
.endif
GET_INST_OPCODE ip
GOTO_OPCODE ip
.endm
.macro POISON_HEAP_REF_IF_OBJECT is_object, rRef
.if \is_object
POISON_HEAP_REF \rRef
.endif
.endm
.macro WRITE_BARRIER_IF_OBJECT is_object, value, holder, label
.if \is_object
cbz \value, \label
ldr ip, [xSELF, #THREAD_CARD_TABLE_OFFSET]
lsr wip2, \holder, #CARD_TABLE_CARD_SHIFT
strb wip, [ip, ip2]
\label:
.endif
.endm
// Puts the next int/long/object parameter passed in physical register
// in the expected dex register array entry, and in case of object in the
// expected reference array entry.
.macro LOOP_OVER_SHORTY_STORING_GPRS gpr_64, gpr_32, shorty, arg_offset, regs, refs, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #74 // if (wip == 'J') goto FOUND_LONG
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto SKIP_FLOAT
b.eq 3f
cmp wip, #68 // if (wip == 'D') goto SKIP_DOUBLE
b.eq 4f
str \gpr_32, [\regs, \arg_offset]
cmp wip, #76 // if (wip != 'L') goto NOT_REFERENCE
b.ne 6f
str \gpr_32, [\refs, \arg_offset]
6: // NOT_REFERENCE
add \arg_offset, \arg_offset, #4
b 5f
2: // FOUND_LONG
str \gpr_64, [\regs, \arg_offset]
add \arg_offset, \arg_offset, #8
b 5f
3: // SKIP_FLOAT
add \arg_offset, \arg_offset, #4
b 1b
4: // SKIP_DOUBLE
add \arg_offset, \arg_offset, #8
b 1b
5:
.endm
// Puts the next floating point parameter passed in physical register
// in the expected dex register array entry.
// Uses ip as temporary.
.macro LOOP_OVER_SHORTY_STORING_FPS dreg, sreg, shorty, arg_offset, fp, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #68 // if (wip == 'D') goto FOUND_DOUBLE
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto FOUND_FLOAT
b.eq 3f
add \arg_offset, \arg_offset, #4
// Handle extra argument in arg array taken by a long.
cmp wip, #74 // if (wip != 'J') goto LOOP
b.ne 1b
add \arg_offset, \arg_offset, #4
b 1b // goto LOOP
2: // FOUND_DOUBLE
str \dreg, [\fp, \arg_offset]
add \arg_offset, \arg_offset, #8
b 4f
3: // FOUND_FLOAT
str \sreg, [\fp, \arg_offset]
add \arg_offset, \arg_offset, #4
4:
.endm
// Puts the next floating point parameter passed in stack
// in the expected dex register array entry.
// Uses ip as temporary.
//
// TODO: Or we could just spill regs to the reserved slots in the caller's
// frame and copy all regs in a simple loop. This time, however, we would
// need to look at the shorty anyway to look for the references.
// (The trade-off is different for passing arguments and receiving them.)
.macro LOOP_OVER_FPs shorty, arg_offset, regs, stack_ptr, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #68 // if (wip == 'D') goto FOUND_DOUBLE
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto FOUND_FLOAT
b.eq 3f
add \arg_offset, \arg_offset, #4
// Handle extra argument in arg array taken by a long.
cmp wip, #74 // if (wip != 'J') goto LOOP
b.ne 1b
add \arg_offset, \arg_offset, #4
b 1b // goto LOOP
2: // FOUND_DOUBLE
add ip, \stack_ptr, \arg_offset
ldr ip, [ip, #OFFSET_TO_FIRST_ARGUMENT_IN_STACK]
str ip, [\regs, \arg_offset]
add \arg_offset, \arg_offset, #8
b 1b
3: // FOUND_FLOAT
add ip, \stack_ptr, \arg_offset
ldr wip, [ip, #OFFSET_TO_FIRST_ARGUMENT_IN_STACK]
str wip, [\regs, \arg_offset]
add \arg_offset, \arg_offset, #4
b 1b
.endm
// Puts the next int/long/object parameter passed in stack
// in the expected dex register array entry, and in case of object in the
// expected reference array entry.
// Uses ip and ip2 as temporary.
.macro LOOP_OVER_INTs shorty, arg_offset, regs, refs, stack_ptr, finished
1: // LOOP
ldrb wip, [\shorty], #1 // Load next character in shorty, and increment.
cbz wip, \finished // if (wip == '\0') goto finished
cmp wip, #74 // if (wip == 'J') goto FOUND_LONG
b.eq 2f
cmp wip, #70 // if (wip == 'F') goto SKIP_FLOAT
b.eq 3f
cmp wip, #68 // if (wip == 'D') goto SKIP_DOUBLE
b.eq 4f
add ip2, \stack_ptr, \arg_offset
ldr wip2, [ip2, #OFFSET_TO_FIRST_ARGUMENT_IN_STACK]
str wip2, [\regs, \arg_offset]
cmp wip, #76 // if (wip != 'L') goto loop
b.ne 3f
str wip2, [\refs, \arg_offset]
add \arg_offset, \arg_offset, #4
b 1b
2: // FOUND_LONG
add ip, \stack_ptr, \arg_offset
ldr ip, [ip, #OFFSET_TO_FIRST_ARGUMENT_IN_STACK]
str ip, [\regs, \arg_offset]
add \arg_offset, \arg_offset, #8
b 1b
3: // SKIP_FLOAT
add \arg_offset, \arg_offset, #4
b 1b
4: // SKIP_DOUBLE
add \arg_offset, \arg_offset, #8
b 1b
.endm
.macro SETUP_REFERENCE_PARAMETER_IN_GPR gpr32, regs, refs, ins, arg_offset, finished
str \gpr32, [\regs, \arg_offset]
sub \ins, \ins, #1
str \gpr32, [\refs, \arg_offset]
add \arg_offset, \arg_offset, #4
cbz \ins, \finished
.endm
// Uses ip2 as temporary.
.macro SETUP_REFERENCE_PARAMETERS_IN_STACK regs, refs, ins, stack_ptr, arg_offset
1:
ldr wip2, [\stack_ptr, \arg_offset]
sub \ins, \ins, #1
str wip2, [\regs, \arg_offset]
str wip2, [\refs, \arg_offset]
add \arg_offset, \arg_offset, #4
cbnz \ins, 1b
.endm
.macro CHECK_AND_UPDATE_SHARED_MEMORY_METHOD if_hot, if_not_hot
ldr wip, [x0, #ART_METHOD_ACCESS_FLAGS_OFFSET]
tbz wip, #ART_METHOD_IS_MEMORY_SHARED_FLAG_BIT, \if_hot
ldr wip, [xSELF, #THREAD_SHARED_METHOD_HOTNESS_OFFSET]
cbz wip, \if_hot
add wip, wip, #-1
str wip, [xSELF, #THREAD_SHARED_METHOD_HOTNESS_OFFSET]
b \if_not_hot
.endm
.macro DO_SUSPEND_CHECK continue_label
ldr wip, [xSELF, #THREAD_FLAGS_OFFSET]
tst wip, #THREAD_SUSPEND_OR_CHECKPOINT_REQUEST
b.eq \continue_label
EXPORT_PC
bl art_quick_test_suspend
.endm
%def entry():
/*
* ArtMethod entry point.
*
* On entry:
* x0 ArtMethod* callee
* rest method parameters
*/
OAT_ENTRY ExecuteNterpWithClinitImpl
.cfi_startproc
// For simplicity, we don't do a read barrier here, but instead rely
// on art_quick_resolution_trampoline to always have a suspend point before
// calling back here.
ldr wip, [x0, #ART_METHOD_DECLARING_CLASS_OFFSET]
ldr wip2, [ip, #MIRROR_CLASS_STATUS_OFFSET]
lsr wip2, wip2, #MIRROR_CLASS_STATUS_SHIFT
cmp wip2, #MIRROR_CLASS_STATUS_VISIBLY_INITIALIZED
b.hs ExecuteNterpImpl
cmp wip2, #MIRROR_CLASS_STATUS_INITIALIZED
b.lo .Linitializing_check
dmb ish
b ExecuteNterpImpl
.Linitializing_check:
cmp wip2, #MIRROR_CLASS_STATUS_INITIALIZING
b.lo .Lresolution_trampoline
ldr wip2, [ip, #MIRROR_CLASS_CLINIT_THREAD_ID_OFFSET]
ldr wip, [xSELF, #THREAD_TID_OFFSET]
cmp wip, wip2
b.eq ExecuteNterpImpl
.Lresolution_trampoline:
b art_quick_resolution_trampoline
.cfi_endproc
.type EndExecuteNterpWithClinitImpl, #function
.hidden EndExecuteNterpWithClinitImpl
.global EndExecuteNterpWithClinitImpl
EndExecuteNterpWithClinitImpl:
OAT_ENTRY ExecuteNterpImpl
.cfi_startproc
sub x16, sp, #STACK_OVERFLOW_RESERVED_BYTES
ldr wzr, [x16]
/* Spill callee save regs */
SPILL_ALL_CALLEE_SAVES
ldr xPC, [x0, #ART_METHOD_DATA_OFFSET_64]
// Setup the stack for executing the method.
SETUP_STACK_FRAME xPC, xREFS, xFP, CFI_REFS, load_ins=1
// Setup the parameters
cbz w15, .Lxmm_setup_finished
sub ip2, ip, x15
ldr w26, [x0, #ART_METHOD_ACCESS_FLAGS_OFFSET]
lsl x27, ip2, #2 // x27 is now the offset for inputs into the registers array.
tbz w26, #ART_METHOD_NTERP_ENTRY_POINT_FAST_PATH_FLAG_BIT, .Lsetup_slow_path
// Setup pointer to inputs in FP and pointer to inputs in REFS
add x10, xFP, x27
add x11, xREFS, x27
mov x12, #0
SETUP_REFERENCE_PARAMETER_IN_GPR w1, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w2, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w3, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w4, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w5, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w6, x10, x11, w15, x12, .Lxmm_setup_finished
SETUP_REFERENCE_PARAMETER_IN_GPR w7, x10, x11, w15, x12, .Lxmm_setup_finished
add x28, x28, #OFFSET_TO_FIRST_ARGUMENT_IN_STACK
SETUP_REFERENCE_PARAMETERS_IN_STACK x10, x11, w15, x28, x12
b .Lxmm_setup_finished
.Lsetup_slow_path:
// If the method is not static and there is one argument ('this'), we don't need to fetch the
// shorty.
tbnz w26, #ART_METHOD_IS_STATIC_FLAG_BIT, .Lsetup_with_shorty
str w1, [xFP, x27]
str w1, [xREFS, x27]
cmp w15, #1
b.eq .Lxmm_setup_finished
.Lsetup_with_shorty:
// TODO: Get shorty in a better way and remove below
SPILL_ALL_ARGUMENTS
bl NterpGetShorty
// Save shorty in callee-save xIBASE.
mov xIBASE, x0
RESTORE_ALL_ARGUMENTS
// Setup pointer to inputs in FP and pointer to inputs in REFS
add x10, xFP, x27
add x11, xREFS, x27
mov x12, #0
add x9, xIBASE, #1 // shorty + 1 ; ie skip return arg character
tbnz w26, #ART_METHOD_IS_STATIC_FLAG_BIT, .Lhandle_static_method
add x10, x10, #4
add x11, x11, #4
add x28, x28, #4
b .Lcontinue_setup_gprs
.Lhandle_static_method:
LOOP_OVER_SHORTY_STORING_GPRS x1, w1, x9, x12, x10, x11, .Lgpr_setup_finished
.Lcontinue_setup_gprs:
LOOP_OVER_SHORTY_STORING_GPRS x2, w2, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_SHORTY_STORING_GPRS x3, w3, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_SHORTY_STORING_GPRS x4, w4, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_SHORTY_STORING_GPRS x5, w5, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_SHORTY_STORING_GPRS x6, w6, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_SHORTY_STORING_GPRS x7, w7, x9, x12, x10, x11, .Lgpr_setup_finished
LOOP_OVER_INTs x9, x12, x10, x11, x28, .Lgpr_setup_finished
.Lgpr_setup_finished:
add x9, xIBASE, #1 // shorty + 1 ; ie skip return arg character
mov x12, #0 // reset counter
LOOP_OVER_SHORTY_STORING_FPS d0, s0, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d1, s1, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d2, s2, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d3, s3, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d4, s4, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d5, s5, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d6, s6, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_SHORTY_STORING_FPS d7, s7, x9, x12, x10, .Lxmm_setup_finished
LOOP_OVER_FPs x9, x12, x10, x28, .Lxmm_setup_finished
.Lxmm_setup_finished:
CFI_DEFINE_DEX_PC_WITH_OFFSET(CFI_TMP, CFI_DEX, 0)
// Set rIBASE
adr xIBASE, artNterpAsmInstructionStart
/* start executing the instruction at xPC */
START_EXECUTING_INSTRUCTIONS
/* NOTE: no fallthrough */
// cfi info continues, and covers the whole nterp implementation.
SIZE ExecuteNterpImpl
%def opcode_pre():
%def fetch_from_thread_cache(dest_reg, miss_label):
// Fetch some information from the thread cache.
// Uses ip and ip2 as temporaries.
add ip, xSELF, #THREAD_INTERPRETER_CACHE_OFFSET // cache address
ubfx ip2, xPC, #2, #THREAD_INTERPRETER_CACHE_SIZE_LOG2 // entry index
add ip, ip, ip2, lsl #4 // entry address within the cache
ldp ip, ${dest_reg}, [ip] // entry key (pc) and value (offset)
cmp ip, xPC
b.ne ${miss_label}
%def footer():
/*
* ===========================================================================
* Common subroutines and data
* ===========================================================================
*/
.text
.align 2
// Enclose all code below in a symbol (which gets printed in backtraces).
NAME_START nterp_helper
// Note: mterp also uses the common_* names below for helpers, but that's OK
// as the assembler compiled each interpreter separately.
common_errDivideByZero:
EXPORT_PC
bl art_quick_throw_div_zero
// Expect index in w1, length in w3.
common_errArrayIndex:
EXPORT_PC
mov x0, x1
mov x1, x3
bl art_quick_throw_array_bounds
common_errNullObject:
EXPORT_PC
bl art_quick_throw_null_pointer_exception
NterpCommonInvokeStatic:
COMMON_INVOKE_NON_RANGE is_static=1, suffix="invokeStatic"
NterpCommonInvokeStaticRange:
COMMON_INVOKE_RANGE is_static=1, suffix="invokeStatic"
NterpCommonInvokeInstance:
COMMON_INVOKE_NON_RANGE suffix="invokeInstance"
NterpCommonInvokeInstanceRange:
COMMON_INVOKE_RANGE suffix="invokeInstance"
NterpCommonInvokeInterface:
COMMON_INVOKE_NON_RANGE is_interface=1, suffix="invokeInterface"
NterpCommonInvokeInterfaceRange:
COMMON_INVOKE_RANGE is_interface=1, suffix="invokeInterface"
NterpCommonInvokePolymorphic:
COMMON_INVOKE_NON_RANGE is_polymorphic=1, suffix="invokePolymorphic"
NterpCommonInvokePolymorphicRange:
COMMON_INVOKE_RANGE is_polymorphic=1, suffix="invokePolymorphic"
NterpCommonInvokeCustom:
COMMON_INVOKE_NON_RANGE is_static=1, is_custom=1, suffix="invokeCustom"
NterpCommonInvokeCustomRange:
COMMON_INVOKE_RANGE is_static=1, is_custom=1, suffix="invokeCustom"
NterpHandleStringInit:
COMMON_INVOKE_NON_RANGE is_string_init=1, suffix="stringInit"
NterpHandleStringInitRange:
COMMON_INVOKE_RANGE is_string_init=1, suffix="stringInit"
NterpHandleHotnessOverflow:
CHECK_AND_UPDATE_SHARED_MEMORY_METHOD if_hot=1f, if_not_hot=5f
1:
mov x1, xPC
mov x2, xFP
bl nterp_hot_method
cbnz x0, 3f
2:
FETCH wINST, 0 // load wINST
GET_INST_OPCODE ip // extract opcode from wINST
GOTO_OPCODE ip // jump to next instruction
3:
// Drop the current frame.
ldr ip, [xREFS, #-8]
mov sp, ip
.cfi_def_cfa sp, CALLEE_SAVES_SIZE
// The transition frame of type SaveAllCalleeSaves saves x19 and x20,
// but not managed ABI. So we need to restore callee-saves of the nterp frame,
// and save managed ABI callee saves, which will be restored by the callee upon
// return.
RESTORE_ALL_CALLEE_SAVES
INCREASE_FRAME ((CALLEE_SAVES_SIZE) - 16)
// FP callee-saves
stp d8, d9, [sp, #0]
stp d10, d11, [sp, #16]
stp d12, d13, [sp, #32]
stp d14, d15, [sp, #48]
// GP callee-saves.
SAVE_TWO_REGS x21, x22, 64
SAVE_TWO_REGS x23, x24, 80
SAVE_TWO_REGS x25, x26, 96
SAVE_TWO_REGS x27, x28, 112
SAVE_TWO_REGS x29, lr, 128
// Setup the new frame
ldr x1, [x0, #OSR_DATA_FRAME_SIZE]
// Given stack size contains all callee saved registers, remove them.
sub x1, x1, #(CALLEE_SAVES_SIZE - 16)
// We know x1 cannot be 0, as it at least contains the ArtMethod.
// Remember CFA in a callee-save register.
mov xINST, sp
.cfi_def_cfa_register xINST
sub sp, sp, x1
add x2, x0, #OSR_DATA_MEMORY
4:
sub x1, x1, #8
ldr ip, [x2, x1]
str ip, [sp, x1]
cbnz x1, 4b
// Fetch the native PC to jump to and save it in a callee-save register.
ldr xFP, [x0, #OSR_DATA_NATIVE_PC]
// Free the memory holding OSR Data.
bl free
// Jump to the compiled code.
br xFP
5:
DO_SUSPEND_CHECK continue_label=2b
b 2b
// This is the logical end of ExecuteNterpImpl, where the frame info applies.
// EndExecuteNterpImpl includes the methods below as we want the runtime to
// see them as part of the Nterp PCs.
.cfi_endproc
nterp_to_nterp_static_non_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_NON_RANGE_ARGUMENTS_AND_EXECUTE is_static=1, is_string_init=0
.cfi_endproc
nterp_to_nterp_string_init_non_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_NON_RANGE_ARGUMENTS_AND_EXECUTE is_static=0, is_string_init=1
.cfi_endproc
nterp_to_nterp_instance_non_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_NON_RANGE_ARGUMENTS_AND_EXECUTE is_static=0, is_string_init=0
.cfi_endproc
nterp_to_nterp_static_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_RANGE_ARGUMENTS_AND_EXECUTE is_static=1
.cfi_endproc
nterp_to_nterp_instance_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_RANGE_ARGUMENTS_AND_EXECUTE is_static=0
.cfi_endproc
nterp_to_nterp_string_init_range:
.cfi_startproc
SETUP_STACK_FOR_INVOKE
SETUP_RANGE_ARGUMENTS_AND_EXECUTE is_static=0, is_string_init=1
.cfi_endproc
NAME_END nterp_helper
// This is the end of PCs contained by the OatQuickMethodHeader created for the interpreter
// entry point.
.type EndExecuteNterpImpl, #function
.hidden EndExecuteNterpImpl
.global EndExecuteNterpImpl
EndExecuteNterpImpl:
// Entrypoints into runtime.
NTERP_TRAMPOLINE nterp_get_static_field, NterpGetStaticField
NTERP_TRAMPOLINE nterp_get_instance_field_offset, NterpGetInstanceFieldOffset
NTERP_TRAMPOLINE nterp_filled_new_array, NterpFilledNewArray
NTERP_TRAMPOLINE nterp_filled_new_array_range, NterpFilledNewArrayRange
NTERP_TRAMPOLINE nterp_get_class, NterpGetClass
NTERP_TRAMPOLINE nterp_allocate_object, NterpAllocateObject
NTERP_TRAMPOLINE nterp_get_method, NterpGetMethod
NTERP_TRAMPOLINE nterp_hot_method, NterpHotMethod
NTERP_TRAMPOLINE nterp_load_object, NterpLoadObject
ENTRY nterp_deliver_pending_exception
DELIVER_PENDING_EXCEPTION
END nterp_deliver_pending_exception
// gen_mterp.py will inline the following definitions
// within [ExecuteNterpImpl, EndExecuteNterpImpl).
%def instruction_end():
.type artNterpAsmInstructionEnd, #function
.hidden artNterpAsmInstructionEnd
.global artNterpAsmInstructionEnd
artNterpAsmInstructionEnd:
// artNterpAsmInstructionEnd is used as landing pad for exception handling.
FETCH_INST
GET_INST_OPCODE ip
GOTO_OPCODE ip
%def instruction_start():
.type artNterpAsmInstructionStart, #function
.hidden artNterpAsmInstructionStart
.global artNterpAsmInstructionStart
artNterpAsmInstructionStart = .L_op_nop
.text
%def opcode_name_prefix():
% return "nterp_"
%def opcode_start():
NAME_START nterp_${opcode}
# Explicitly restore CFA, just in case the previous opcode clobbered it (by .cfi_def_*).
CFI_DEF_CFA_BREG_PLUS_UCONST CFI_REFS, -8, CALLEE_SAVES_SIZE
%def opcode_end():
NAME_END nterp_${opcode}
// Advance to the end of this handler. Causes error if we are past that point.
.org nterp_${opcode} + NTERP_HANDLER_SIZE // ${opcode} handler is too big!
%def opcode_slow_path_start(name):
NAME_START ${name}
%def opcode_slow_path_end(name):
NAME_END ${name}