blob: 6985b7384585c8b182329b346977b3369f323c73 [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "arm64_lir.h"
#include "codegen_arm64.h"
#include "dex/quick/mir_to_lir-inl.h"
#include "dex/reg_storage_eq.h"
namespace art {
/* This file contains codegen for the A64 ISA. */
int32_t Arm64Mir2Lir::EncodeImmSingle(uint32_t bits) {
/*
* Valid values will have the form:
*
* aBbb.bbbc.defg.h000.0000.0000.0000.0000
*
* where B = not(b). In other words, if b == 1, then B == 0 and viceversa.
*/
// bits[19..0] are cleared.
if ((bits & 0x0007ffff) != 0)
return -1;
// bits[29..25] are all set or all cleared.
uint32_t b_pattern = (bits >> 16) & 0x3e00;
if (b_pattern != 0 && b_pattern != 0x3e00)
return -1;
// bit[30] and bit[29] are opposite.
if (((bits ^ (bits << 1)) & 0x40000000) == 0)
return -1;
// bits: aBbb.bbbc.defg.h000.0000.0000.0000.0000
// bit7: a000.0000
uint32_t bit7 = ((bits >> 31) & 0x1) << 7;
// bit6: 0b00.0000
uint32_t bit6 = ((bits >> 29) & 0x1) << 6;
// bit5_to_0: 00cd.efgh
uint32_t bit5_to_0 = (bits >> 19) & 0x3f;
return (bit7 | bit6 | bit5_to_0);
}
int32_t Arm64Mir2Lir::EncodeImmDouble(uint64_t bits) {
/*
* Valid values will have the form:
*
* aBbb.bbbb.bbcd.efgh.0000.0000.0000.0000
* 0000.0000.0000.0000.0000.0000.0000.0000
*
* where B = not(b).
*/
// bits[47..0] are cleared.
if ((bits & UINT64_C(0xffffffffffff)) != 0)
return -1;
// bits[61..54] are all set or all cleared.
uint32_t b_pattern = (bits >> 48) & 0x3fc0;
if (b_pattern != 0 && b_pattern != 0x3fc0)
return -1;
// bit[62] and bit[61] are opposite.
if (((bits ^ (bits << 1)) & UINT64_C(0x4000000000000000)) == 0)
return -1;
// bit7: a000.0000
uint32_t bit7 = ((bits >> 63) & 0x1) << 7;
// bit6: 0b00.0000
uint32_t bit6 = ((bits >> 61) & 0x1) << 6;
// bit5_to_0: 00cd.efgh
uint32_t bit5_to_0 = (bits >> 48) & 0x3f;
return (bit7 | bit6 | bit5_to_0);
}
size_t Arm64Mir2Lir::GetLoadStoreSize(LIR* lir) {
bool opcode_is_wide = IS_WIDE(lir->opcode);
A64Opcode opcode = UNWIDE(lir->opcode);
DCHECK(!IsPseudoLirOp(opcode));
const A64EncodingMap *encoder = &EncodingMap[opcode];
uint32_t bits = opcode_is_wide ? encoder->xskeleton : encoder->wskeleton;
return (bits >> 30);
}
size_t Arm64Mir2Lir::GetInstructionOffset(LIR* lir) {
size_t offset = lir->operands[2];
uint64_t check_flags = GetTargetInstFlags(lir->opcode);
DCHECK((check_flags & IS_LOAD) || (check_flags & IS_STORE));
if (check_flags & SCALED_OFFSET_X0) {
DCHECK(check_flags & IS_TERTIARY_OP);
offset = offset * (1 << GetLoadStoreSize(lir));
}
return offset;
}
LIR* Arm64Mir2Lir::LoadFPConstantValue(RegStorage r_dest, int32_t value) {
DCHECK(r_dest.IsSingle());
if (value == 0) {
return NewLIR2(kA64Fmov2sw, r_dest.GetReg(), rwzr);
} else {
int32_t encoded_imm = EncodeImmSingle((uint32_t)value);
if (encoded_imm >= 0) {
return NewLIR2(kA64Fmov2fI, r_dest.GetReg(), encoded_imm);
}
}
LIR* data_target = ScanLiteralPool(literal_list_, value, 0);
if (data_target == NULL) {
// Wide, as we need 8B alignment.
data_target = AddWideData(&literal_list_, value, 0);
}
ScopedMemRefType mem_ref_type(this, ResourceMask::kLiteral);
LIR* load_pc_rel = RawLIR(current_dalvik_offset_, kA64Ldr2fp,
r_dest.GetReg(), 0, 0, 0, 0, data_target);
AppendLIR(load_pc_rel);
return load_pc_rel;
}
LIR* Arm64Mir2Lir::LoadFPConstantValueWide(RegStorage r_dest, int64_t value) {
DCHECK(r_dest.IsDouble());
if (value == 0) {
return NewLIR2(kA64Fmov2Sx, r_dest.GetReg(), rxzr);
} else {
int32_t encoded_imm = EncodeImmDouble(value);
if (encoded_imm >= 0) {
return NewLIR2(WIDE(kA64Fmov2fI), r_dest.GetReg(), encoded_imm);
}
}
// No short form - load from the literal pool.
int32_t val_lo = Low32Bits(value);
int32_t val_hi = High32Bits(value);
LIR* data_target = ScanLiteralPoolWide(literal_list_, val_lo, val_hi);
if (data_target == NULL) {
data_target = AddWideData(&literal_list_, val_lo, val_hi);
}
ScopedMemRefType mem_ref_type(this, ResourceMask::kLiteral);
LIR* load_pc_rel = RawLIR(current_dalvik_offset_, WIDE(kA64Ldr2fp),
r_dest.GetReg(), 0, 0, 0, 0, data_target);
AppendLIR(load_pc_rel);
return load_pc_rel;
}
static int CountLeadingZeros(bool is_wide, uint64_t value) {
return (is_wide) ? __builtin_clzll(value) : __builtin_clz((uint32_t)value);
}
static int CountTrailingZeros(bool is_wide, uint64_t value) {
return (is_wide) ? __builtin_ctzll(value) : __builtin_ctz((uint32_t)value);
}
static int CountSetBits(bool is_wide, uint64_t value) {
return ((is_wide) ?
__builtin_popcountll(value) : __builtin_popcount((uint32_t)value));
}
/**
* @brief Try encoding an immediate in the form required by logical instructions.
*
* @param is_wide Whether @p value is a 64-bit (as opposed to 32-bit) value.
* @param value An integer to be encoded. This is interpreted as 64-bit if @p is_wide is true and as
* 32-bit if @p is_wide is false.
* @return A non-negative integer containing the encoded immediate or -1 if the encoding failed.
* @note This is the inverse of Arm64Mir2Lir::DecodeLogicalImmediate().
*/
int Arm64Mir2Lir::EncodeLogicalImmediate(bool is_wide, uint64_t value) {
unsigned n, imm_s, imm_r;
// Logical immediates are encoded using parameters n, imm_s and imm_r using
// the following table:
//
// N imms immr size S R
// 1 ssssss rrrrrr 64 UInt(ssssss) UInt(rrrrrr)
// 0 0sssss xrrrrr 32 UInt(sssss) UInt(rrrrr)
// 0 10ssss xxrrrr 16 UInt(ssss) UInt(rrrr)
// 0 110sss xxxrrr 8 UInt(sss) UInt(rrr)
// 0 1110ss xxxxrr 4 UInt(ss) UInt(rr)
// 0 11110s xxxxxr 2 UInt(s) UInt(r)
// (s bits must not be all set)
//
// A pattern is constructed of size bits, where the least significant S+1
// bits are set. The pattern is rotated right by R, and repeated across a
// 32 or 64-bit value, depending on destination register width.
//
// To test if an arbitary immediate can be encoded using this scheme, an
// iterative algorithm is used.
//
// 1. If the value has all set or all clear bits, it can't be encoded.
if (value == 0 || value == ~UINT64_C(0) ||
(!is_wide && (uint32_t)value == ~UINT32_C(0))) {
return -1;
}
unsigned lead_zero = CountLeadingZeros(is_wide, value);
unsigned lead_one = CountLeadingZeros(is_wide, ~value);
unsigned trail_zero = CountTrailingZeros(is_wide, value);
unsigned trail_one = CountTrailingZeros(is_wide, ~value);
unsigned set_bits = CountSetBits(is_wide, value);
// The fixed bits in the immediate s field.
// If width == 64 (X reg), start at 0xFFFFFF80.
// If width == 32 (W reg), start at 0xFFFFFFC0, as the iteration for 64-bit
// widths won't be executed.
unsigned width = (is_wide) ? 64 : 32;
int imm_s_fixed = (is_wide) ? -128 : -64;
int imm_s_mask = 0x3f;
for (;;) {
// 2. If the value is two bits wide, it can be encoded.
if (width == 2) {
n = 0;
imm_s = 0x3C;
imm_r = (value & 3) - 1;
break;
}
n = (width == 64) ? 1 : 0;
imm_s = ((imm_s_fixed | (set_bits - 1)) & imm_s_mask);
if ((lead_zero + set_bits) == width) {
imm_r = 0;
} else {
imm_r = (lead_zero > 0) ? (width - trail_zero) : lead_one;
}
// 3. If the sum of leading zeros, trailing zeros and set bits is
// equal to the bit width of the value, it can be encoded.
if (lead_zero + trail_zero + set_bits == width) {
break;
}
// 4. If the sum of leading ones, trailing ones and unset bits in the
// value is equal to the bit width of the value, it can be encoded.
if (lead_one + trail_one + (width - set_bits) == width) {
break;
}
// 5. If the most-significant half of the bitwise value is equal to
// the least-significant half, return to step 2 using the
// least-significant half of the value.
uint64_t mask = (UINT64_C(1) << (width >> 1)) - 1;
if ((value & mask) == ((value >> (width >> 1)) & mask)) {
width >>= 1;
set_bits >>= 1;
imm_s_fixed >>= 1;
continue;
}
// 6. Otherwise, the value can't be encoded.
return -1;
}
return (n << 12 | imm_r << 6 | imm_s);
}
// Maximum number of instructions to use for encoding the immediate.
static const int max_num_ops_per_const_load = 2;
/**
* @brief Return the number of fast halfwords in the given uint64_t integer.
* @details The input integer is split into 4 halfwords (bits 0-15, 16-31, 32-47, 48-63). The
* number of fast halfwords (halfwords that are either 0 or 0xffff) is returned. See below for
* a more accurate description.
* @param value The input 64-bit integer.
* @return Return @c retval such that (retval & 0x7) is the maximum between n and m, where n is
* the number of halfwords with all bits unset (0) and m is the number of halfwords with all bits
* set (0xffff). Additionally (retval & 0x8) is set when m > n.
*/
static int GetNumFastHalfWords(uint64_t value) {
unsigned int num_0000_halfwords = 0;
unsigned int num_ffff_halfwords = 0;
for (int shift = 0; shift < 64; shift += 16) {
uint16_t halfword = static_cast<uint16_t>(value >> shift);
if (halfword == 0)
num_0000_halfwords++;
else if (halfword == UINT16_C(0xffff))
num_ffff_halfwords++;
}
if (num_0000_halfwords >= num_ffff_halfwords) {
DCHECK_LE(num_0000_halfwords, 4U);
return num_0000_halfwords;
} else {
DCHECK_LE(num_ffff_halfwords, 4U);
return num_ffff_halfwords | 0x8;
}
}
// The InexpensiveConstantXXX variants below are used in the promotion algorithm to determine how a
// constant is considered for promotion. If the constant is "inexpensive" then the promotion
// algorithm will give it a low priority for promotion, even when it is referenced many times in
// the code.
bool Arm64Mir2Lir::InexpensiveConstantInt(int32_t value) {
// A 32-bit int can always be loaded with 2 instructions (and without using the literal pool).
// We therefore return true and give it a low priority for promotion.
return true;
}
bool Arm64Mir2Lir::InexpensiveConstantFloat(int32_t value) {
return EncodeImmSingle(value) >= 0;
}
bool Arm64Mir2Lir::InexpensiveConstantLong(int64_t value) {
int num_slow_halfwords = 4 - (GetNumFastHalfWords(value) & 0x7);
if (num_slow_halfwords <= max_num_ops_per_const_load) {
return true;
}
return (EncodeLogicalImmediate(/*is_wide=*/true, value) >= 0);
}
bool Arm64Mir2Lir::InexpensiveConstantDouble(int64_t value) {
return EncodeImmDouble(value) >= 0;
}
// The InexpensiveConstantXXX variants below are used to determine which A64 instructions to use
// when one of the operands is an immediate (e.g. register version or immediate version of add).
bool Arm64Mir2Lir::InexpensiveConstantInt(int32_t value, Instruction::Code opcode) {
switch (opcode) {
case Instruction::IF_EQ:
case Instruction::IF_NE:
case Instruction::IF_LT:
case Instruction::IF_GE:
case Instruction::IF_GT:
case Instruction::IF_LE:
case Instruction::ADD_INT:
case Instruction::ADD_INT_2ADDR:
case Instruction::SUB_INT:
case Instruction::SUB_INT_2ADDR:
// The code below is consistent with the implementation of OpRegRegImm().
{
int32_t abs_value = std::abs(value);
if (abs_value < 0x1000) {
return true;
} else if ((abs_value & UINT64_C(0xfff)) == 0 && ((abs_value >> 12) < 0x1000)) {
return true;
}
return false;
}
case Instruction::SHL_INT:
case Instruction::SHL_INT_2ADDR:
case Instruction::SHR_INT:
case Instruction::SHR_INT_2ADDR:
case Instruction::USHR_INT:
case Instruction::USHR_INT_2ADDR:
return true;
case Instruction::AND_INT:
case Instruction::AND_INT_2ADDR:
case Instruction::AND_INT_LIT16:
case Instruction::AND_INT_LIT8:
case Instruction::OR_INT:
case Instruction::OR_INT_2ADDR:
case Instruction::OR_INT_LIT16:
case Instruction::OR_INT_LIT8:
case Instruction::XOR_INT:
case Instruction::XOR_INT_2ADDR:
case Instruction::XOR_INT_LIT16:
case Instruction::XOR_INT_LIT8:
if (value == 0 || value == INT32_C(-1)) {
return true;
}
return (EncodeLogicalImmediate(/*is_wide=*/false, value) >= 0);
default:
return false;
}
}
/*
* Load a immediate using one single instruction when possible; otherwise
* use a pair of movz and movk instructions.
*
* No additional register clobbering operation performed. Use this version when
* 1) r_dest is freshly returned from AllocTemp or
* 2) The codegen is under fixed register usage
*/
LIR* Arm64Mir2Lir::LoadConstantNoClobber(RegStorage r_dest, int value) {
LIR* res;
if (r_dest.IsFloat()) {
return LoadFPConstantValue(r_dest, value);
}
if (r_dest.Is64Bit()) {
return LoadConstantWide(r_dest, value);
}
// Loading SP/ZR with an immediate is not supported.
DCHECK(!A64_REG_IS_SP(r_dest.GetReg()));
DCHECK(!A64_REG_IS_ZR(r_dest.GetReg()));
// Compute how many movk, movz instructions are needed to load the value.
uint16_t high_bits = High16Bits(value);
uint16_t low_bits = Low16Bits(value);
bool low_fast = ((uint16_t)(low_bits + 1) <= 1);
bool high_fast = ((uint16_t)(high_bits + 1) <= 1);
if (LIKELY(low_fast || high_fast)) {
// 1 instruction is enough to load the immediate.
if (LIKELY(low_bits == high_bits)) {
// Value is either 0 or -1: we can just use wzr.
A64Opcode opcode = LIKELY(low_bits == 0) ? kA64Mov2rr : kA64Mvn2rr;
res = NewLIR2(opcode, r_dest.GetReg(), rwzr);
} else {
uint16_t uniform_bits, useful_bits;
int shift;
if (LIKELY(high_fast)) {
shift = 0;
uniform_bits = high_bits;
useful_bits = low_bits;
} else {
shift = 1;
uniform_bits = low_bits;
useful_bits = high_bits;
}
if (UNLIKELY(uniform_bits != 0)) {
res = NewLIR3(kA64Movn3rdM, r_dest.GetReg(), ~useful_bits, shift);
} else {
res = NewLIR3(kA64Movz3rdM, r_dest.GetReg(), useful_bits, shift);
}
}
} else {
// movk, movz require 2 instructions. Try detecting logical immediates.
int log_imm = EncodeLogicalImmediate(/*is_wide=*/false, value);
if (log_imm >= 0) {
res = NewLIR3(kA64Orr3Rrl, r_dest.GetReg(), rwzr, log_imm);
} else {
// Use 2 instructions.
res = NewLIR3(kA64Movz3rdM, r_dest.GetReg(), low_bits, 0);
NewLIR3(kA64Movk3rdM, r_dest.GetReg(), high_bits, 1);
}
}
return res;
}
// TODO: clean up the names. LoadConstantWide() should really be LoadConstantNoClobberWide().
LIR* Arm64Mir2Lir::LoadConstantWide(RegStorage r_dest, int64_t value) {
if (r_dest.IsFloat()) {
return LoadFPConstantValueWide(r_dest, value);
}
DCHECK(r_dest.Is64Bit());
// Loading SP/ZR with an immediate is not supported.
DCHECK(!A64_REG_IS_SP(r_dest.GetReg()));
DCHECK(!A64_REG_IS_ZR(r_dest.GetReg()));
if (LIKELY(value == INT64_C(0) || value == INT64_C(-1))) {
// value is either 0 or -1: we can just use xzr.
A64Opcode opcode = LIKELY(value == 0) ? WIDE(kA64Mov2rr) : WIDE(kA64Mvn2rr);
return NewLIR2(opcode, r_dest.GetReg(), rxzr);
}
// At least one in value's halfwords is not 0x0, nor 0xffff: find out how many.
uint64_t uvalue = static_cast<uint64_t>(value);
int num_fast_halfwords = GetNumFastHalfWords(uvalue);
int num_slow_halfwords = 4 - (num_fast_halfwords & 0x7);
bool more_ffff_halfwords = (num_fast_halfwords & 0x8) != 0;
if (num_slow_halfwords > 1) {
// A single movz/movn is not enough. Try the logical immediate route.
int log_imm = EncodeLogicalImmediate(/*is_wide=*/true, value);
if (log_imm >= 0) {
return NewLIR3(WIDE(kA64Orr3Rrl), r_dest.GetReg(), rxzr, log_imm);
}
}
if (num_slow_halfwords <= max_num_ops_per_const_load) {
// We can encode the number using a movz/movn followed by one or more movk.
A64Opcode op;
uint16_t background;
LIR* res = nullptr;
// Decide whether to use a movz or a movn.
if (more_ffff_halfwords) {
op = WIDE(kA64Movn3rdM);
background = 0xffff;
} else {
op = WIDE(kA64Movz3rdM);
background = 0;
}
// Emit the first instruction (movz, movn).
int shift;
for (shift = 0; shift < 4; shift++) {
uint16_t halfword = static_cast<uint16_t>(uvalue >> (shift << 4));
if (halfword != background) {
res = NewLIR3(op, r_dest.GetReg(), halfword ^ background, shift);
break;
}
}
// Emit the movk instructions.
for (shift++; shift < 4; shift++) {
uint16_t halfword = static_cast<uint16_t>(uvalue >> (shift << 4));
if (halfword != background) {
NewLIR3(WIDE(kA64Movk3rdM), r_dest.GetReg(), halfword, shift);
}
}
return res;
}
// Use the literal pool.
int32_t val_lo = Low32Bits(value);
int32_t val_hi = High32Bits(value);
LIR* data_target = ScanLiteralPoolWide(literal_list_, val_lo, val_hi);
if (data_target == NULL) {
data_target = AddWideData(&literal_list_, val_lo, val_hi);
}
ScopedMemRefType mem_ref_type(this, ResourceMask::kLiteral);
LIR *res = RawLIR(current_dalvik_offset_, WIDE(kA64Ldr2rp),
r_dest.GetReg(), 0, 0, 0, 0, data_target);
AppendLIR(res);
return res;
}
LIR* Arm64Mir2Lir::OpUnconditionalBranch(LIR* target) {
LIR* res = NewLIR1(kA64B1t, 0 /* offset to be patched during assembly */);
res->target = target;
return res;
}
LIR* Arm64Mir2Lir::OpCondBranch(ConditionCode cc, LIR* target) {
LIR* branch = NewLIR2(kA64B2ct, ArmConditionEncoding(cc),
0 /* offset to be patched */);
branch->target = target;
return branch;
}
LIR* Arm64Mir2Lir::OpReg(OpKind op, RegStorage r_dest_src) {
A64Opcode opcode = kA64Brk1d;
switch (op) {
case kOpBlx:
opcode = kA64Blr1x;
break;
default:
LOG(FATAL) << "Bad opcode " << op;
}
return NewLIR1(opcode, r_dest_src.GetReg());
}
LIR* Arm64Mir2Lir::OpRegRegShift(OpKind op, RegStorage r_dest_src1, RegStorage r_src2, int shift) {
A64Opcode wide = (r_dest_src1.Is64Bit()) ? WIDE(0) : UNWIDE(0);
CHECK_EQ(r_dest_src1.Is64Bit(), r_src2.Is64Bit());
A64Opcode opcode = kA64Brk1d;
switch (op) {
case kOpCmn:
opcode = kA64Cmn3rro;
break;
case kOpCmp:
opcode = kA64Cmp3rro;
break;
case kOpMov:
opcode = kA64Mov2rr;
break;
case kOpMvn:
opcode = kA64Mvn2rr;
break;
case kOpNeg:
opcode = kA64Neg3rro;
break;
case kOpTst:
opcode = kA64Tst3rro;
break;
case kOpRev:
DCHECK_EQ(shift, 0);
// Binary, but rm is encoded twice.
return NewLIR2(kA64Rev2rr | wide, r_dest_src1.GetReg(), r_src2.GetReg());
break;
case kOpRevsh:
// Binary, but rm is encoded twice.
NewLIR2(kA64Rev162rr | wide, r_dest_src1.GetReg(), r_src2.GetReg());
// "sxth r1, r2" is "sbfm r1, r2, #0, #15"
return NewLIR4(kA64Sbfm4rrdd | wide, r_dest_src1.GetReg(), r_dest_src1.GetReg(), 0, 15);
break;
case kOp2Byte:
DCHECK_EQ(shift, ENCODE_NO_SHIFT);
// "sbfx r1, r2, #imm1, #imm2" is "sbfm r1, r2, #imm1, #(imm1 + imm2 - 1)".
// For now we use sbfm directly.
return NewLIR4(kA64Sbfm4rrdd | wide, r_dest_src1.GetReg(), r_src2.GetReg(), 0, 7);
case kOp2Short:
DCHECK_EQ(shift, ENCODE_NO_SHIFT);
// For now we use sbfm rather than its alias, sbfx.
return NewLIR4(kA64Sbfm4rrdd | wide, r_dest_src1.GetReg(), r_src2.GetReg(), 0, 15);
case kOp2Char:
// "ubfx r1, r2, #imm1, #imm2" is "ubfm r1, r2, #imm1, #(imm1 + imm2 - 1)".
// For now we use ubfm directly.
DCHECK_EQ(shift, ENCODE_NO_SHIFT);
return NewLIR4(kA64Ubfm4rrdd | wide, r_dest_src1.GetReg(), r_src2.GetReg(), 0, 15);
default:
return OpRegRegRegShift(op, r_dest_src1, r_dest_src1, r_src2, shift);
}
DCHECK(!IsPseudoLirOp(opcode));
if (EncodingMap[opcode].flags & IS_BINARY_OP) {
DCHECK_EQ(shift, ENCODE_NO_SHIFT);
return NewLIR2(opcode | wide, r_dest_src1.GetReg(), r_src2.GetReg());
} else if (EncodingMap[opcode].flags & IS_TERTIARY_OP) {
A64EncodingKind kind = EncodingMap[opcode].field_loc[2].kind;
if (kind == kFmtShift) {
return NewLIR3(opcode | wide, r_dest_src1.GetReg(), r_src2.GetReg(), shift);
}
}
LOG(FATAL) << "Unexpected encoding operand count";
return NULL;
}
LIR* Arm64Mir2Lir::OpRegRegExtend(OpKind op, RegStorage r_dest_src1, RegStorage r_src2,
A64RegExtEncodings ext, uint8_t amount) {
A64Opcode wide = (r_dest_src1.Is64Bit()) ? WIDE(0) : UNWIDE(0);
A64Opcode opcode = kA64Brk1d;
switch (op) {
case kOpCmn:
opcode = kA64Cmn3Rre;
break;
case kOpCmp:
opcode = kA64Cmp3Rre;
break;
case kOpAdd:
// Note: intentional fallthrough
case kOpSub:
return OpRegRegRegExtend(op, r_dest_src1, r_dest_src1, r_src2, ext, amount);
break;
default:
LOG(FATAL) << "Bad Opcode: " << opcode;
break;
}
DCHECK(!IsPseudoLirOp(opcode));
if (EncodingMap[opcode].flags & IS_TERTIARY_OP) {
A64EncodingKind kind = EncodingMap[opcode].field_loc[2].kind;
if (kind == kFmtExtend) {
return NewLIR3(opcode | wide, r_dest_src1.GetReg(), r_src2.GetReg(),
EncodeExtend(ext, amount));
}
}
LOG(FATAL) << "Unexpected encoding operand count";
return NULL;
}
LIR* Arm64Mir2Lir::OpRegReg(OpKind op, RegStorage r_dest_src1, RegStorage r_src2) {
/* RegReg operations with SP in first parameter need extended register instruction form.
* Only CMN, CMP, ADD & SUB instructions are implemented.
*/
if (r_dest_src1 == rs_sp) {
return OpRegRegExtend(op, r_dest_src1, r_src2, kA64Uxtx, 0);
} else {
return OpRegRegShift(op, r_dest_src1, r_src2, ENCODE_NO_SHIFT);
}
}
LIR* Arm64Mir2Lir::OpMovRegMem(RegStorage r_dest, RegStorage r_base, int offset, MoveType move_type) {
UNIMPLEMENTED(FATAL);
return nullptr;
}
LIR* Arm64Mir2Lir::OpMovMemReg(RegStorage r_base, int offset, RegStorage r_src, MoveType move_type) {
UNIMPLEMENTED(FATAL);
return nullptr;
}
LIR* Arm64Mir2Lir::OpCondRegReg(OpKind op, ConditionCode cc, RegStorage r_dest, RegStorage r_src) {
LOG(FATAL) << "Unexpected use of OpCondRegReg for Arm64";
return NULL;
}
LIR* Arm64Mir2Lir::OpRegRegRegShift(OpKind op, RegStorage r_dest, RegStorage r_src1,
RegStorage r_src2, int shift) {
A64Opcode opcode = kA64Brk1d;
switch (op) {
case kOpAdd:
opcode = kA64Add4rrro;
break;
case kOpSub:
opcode = kA64Sub4rrro;
break;
// case kOpRsub:
// opcode = kA64RsubWWW;
// break;
case kOpAdc:
opcode = kA64Adc3rrr;
break;
case kOpAnd:
opcode = kA64And4rrro;
break;
case kOpXor:
opcode = kA64Eor4rrro;
break;
case kOpMul:
opcode = kA64Mul3rrr;
break;
case kOpDiv:
opcode = kA64Sdiv3rrr;
break;
case kOpOr:
opcode = kA64Orr4rrro;
break;
case kOpSbc:
opcode = kA64Sbc3rrr;
break;
case kOpLsl:
opcode = kA64Lsl3rrr;
break;
case kOpLsr:
opcode = kA64Lsr3rrr;
break;
case kOpAsr:
opcode = kA64Asr3rrr;
break;
case kOpRor:
opcode = kA64Ror3rrr;
break;
default:
LOG(FATAL) << "Bad opcode: " << op;
break;
}
// The instructions above belong to two kinds:
// - 4-operands instructions, where the last operand is a shift/extend immediate,
// - 3-operands instructions with no shift/extend.
A64Opcode widened_opcode = r_dest.Is64Bit() ? WIDE(opcode) : opcode;
CHECK_EQ(r_dest.Is64Bit(), r_src1.Is64Bit());
CHECK_EQ(r_dest.Is64Bit(), r_src2.Is64Bit());
if (EncodingMap[opcode].flags & IS_QUAD_OP) {
DCHECK(!IsExtendEncoding(shift));
return NewLIR4(widened_opcode, r_dest.GetReg(), r_src1.GetReg(), r_src2.GetReg(), shift);
} else {
DCHECK(EncodingMap[opcode].flags & IS_TERTIARY_OP);
DCHECK_EQ(shift, ENCODE_NO_SHIFT);
return NewLIR3(widened_opcode, r_dest.GetReg(), r_src1.GetReg(), r_src2.GetReg());
}
}
LIR* Arm64Mir2Lir::OpRegRegRegExtend(OpKind op, RegStorage r_dest, RegStorage r_src1,
RegStorage r_src2, A64RegExtEncodings ext, uint8_t amount) {
A64Opcode opcode = kA64Brk1d;
switch (op) {
case kOpAdd:
opcode = kA64Add4RRre;
break;
case kOpSub:
opcode = kA64Sub4RRre;
break;
default:
UNIMPLEMENTED(FATAL) << "Unimplemented opcode: " << op;
UNREACHABLE();
}
A64Opcode widened_opcode = r_dest.Is64Bit() ? WIDE(opcode) : opcode;
if (r_dest.Is64Bit()) {
CHECK(r_src1.Is64Bit());
// dest determines whether the op is wide or not. Up-convert src2 when necessary.
// Note: this is not according to aarch64 specifications, but our encoding.
if (!r_src2.Is64Bit()) {
r_src2 = As64BitReg(r_src2);
}
} else {
CHECK(!r_src1.Is64Bit());
CHECK(!r_src2.Is64Bit());
}
// Sanity checks.
// 1) Amount is in the range 0..4
CHECK_LE(amount, 4);
return NewLIR4(widened_opcode, r_dest.GetReg(), r_src1.GetReg(), r_src2.GetReg(),
EncodeExtend(ext, amount));
}
LIR* Arm64Mir2Lir::OpRegRegReg(OpKind op, RegStorage r_dest, RegStorage r_src1, RegStorage r_src2) {
return OpRegRegRegShift(op, r_dest, r_src1, r_src2, ENCODE_NO_SHIFT);
}
LIR* Arm64Mir2Lir::OpRegRegImm(OpKind op, RegStorage r_dest, RegStorage r_src1, int value) {
return OpRegRegImm64(op, r_dest, r_src1, static_cast<int64_t>(value));
}
LIR* Arm64Mir2Lir::OpRegRegImm64(OpKind op, RegStorage r_dest, RegStorage r_src1, int64_t value) {
LIR* res;
bool neg = (value < 0);
int64_t abs_value = (neg) ? -value : value;
A64Opcode opcode = kA64Brk1d;
A64Opcode alt_opcode = kA64Brk1d;
bool is_logical = false;
bool is_wide = r_dest.Is64Bit();
A64Opcode wide = (is_wide) ? WIDE(0) : UNWIDE(0);
int info = 0;
switch (op) {
case kOpLsl: {
// "lsl w1, w2, #imm" is an alias of "ubfm w1, w2, #(-imm MOD 32), #(31-imm)"
// and "lsl x1, x2, #imm" of "ubfm x1, x2, #(-imm MOD 64), #(63-imm)".
// For now, we just use ubfm directly.
int max_value = (is_wide) ? 63 : 31;
return NewLIR4(kA64Ubfm4rrdd | wide, r_dest.GetReg(), r_src1.GetReg(),
(-value) & max_value, max_value - value);
}
case kOpLsr:
return NewLIR3(kA64Lsr3rrd | wide, r_dest.GetReg(), r_src1.GetReg(), value);
case kOpAsr:
return NewLIR3(kA64Asr3rrd | wide, r_dest.GetReg(), r_src1.GetReg(), value);
case kOpRor:
// "ror r1, r2, #imm" is an alias of "extr r1, r2, r2, #imm".
// For now, we just use extr directly.
return NewLIR4(kA64Extr4rrrd | wide, r_dest.GetReg(), r_src1.GetReg(), r_src1.GetReg(),
value);
case kOpAdd:
neg = !neg;
FALLTHROUGH_INTENDED;
case kOpSub:
// Add and sub below read/write sp rather than xzr.
if (abs_value < 0x1000) {
opcode = (neg) ? kA64Add4RRdT : kA64Sub4RRdT;
return NewLIR4(opcode | wide, r_dest.GetReg(), r_src1.GetReg(), abs_value, 0);
} else if ((abs_value & UINT64_C(0xfff)) == 0 && ((abs_value >> 12) < 0x1000)) {
opcode = (neg) ? kA64Add4RRdT : kA64Sub4RRdT;
return NewLIR4(opcode | wide, r_dest.GetReg(), r_src1.GetReg(), abs_value >> 12, 1);
} else {
alt_opcode = (op == kOpAdd) ? kA64Add4RRre : kA64Sub4RRre;
info = EncodeExtend(is_wide ? kA64Uxtx : kA64Uxtw, 0);
}
break;
case kOpAdc:
alt_opcode = kA64Adc3rrr;
break;
case kOpSbc:
alt_opcode = kA64Sbc3rrr;
break;
case kOpOr:
is_logical = true;
opcode = kA64Orr3Rrl;
alt_opcode = kA64Orr4rrro;
break;
case kOpAnd:
is_logical = true;
opcode = kA64And3Rrl;
alt_opcode = kA64And4rrro;
break;
case kOpXor:
is_logical = true;
opcode = kA64Eor3Rrl;
alt_opcode = kA64Eor4rrro;
break;
case kOpMul:
// TUNING: power of 2, shift & add
alt_opcode = kA64Mul3rrr;
break;
default:
LOG(FATAL) << "Bad opcode: " << op;
}
if (is_logical) {
int log_imm = EncodeLogicalImmediate(is_wide, value);
if (log_imm >= 0) {
return NewLIR3(opcode | wide, r_dest.GetReg(), r_src1.GetReg(), log_imm);
} else {
// When the immediate is either 0 or ~0, the logical operation can be trivially reduced
// to a - possibly negated - assignment.
if (value == 0) {
switch (op) {
case kOpOr:
case kOpXor:
// Or/Xor by zero reduces to an assignment.
return NewLIR2(kA64Mov2rr | wide, r_dest.GetReg(), r_src1.GetReg());
default:
// And by zero reduces to a `mov rdest, xzr'.
DCHECK(op == kOpAnd);
return NewLIR2(kA64Mov2rr | wide, r_dest.GetReg(), (is_wide) ? rxzr : rwzr);
}
} else if (value == INT64_C(-1)
|| (!is_wide && static_cast<uint32_t>(value) == ~UINT32_C(0))) {
switch (op) {
case kOpAnd:
// And by -1 reduces to an assignment.
return NewLIR2(kA64Mov2rr | wide, r_dest.GetReg(), r_src1.GetReg());
case kOpXor:
// Xor by -1 reduces to an `mvn rdest, rsrc'.
return NewLIR2(kA64Mvn2rr | wide, r_dest.GetReg(), r_src1.GetReg());
default:
// Or by -1 reduces to a `mvn rdest, xzr'.
DCHECK(op == kOpOr);
return NewLIR2(kA64Mvn2rr | wide, r_dest.GetReg(), (is_wide) ? rxzr : rwzr);
}
}
}
}
RegStorage r_scratch;
if (is_wide) {
r_scratch = AllocTempWide();
LoadConstantWide(r_scratch, value);
} else {
r_scratch = AllocTemp();
LoadConstant(r_scratch, value);
}
if (EncodingMap[alt_opcode].flags & IS_QUAD_OP)
res = NewLIR4(alt_opcode | wide, r_dest.GetReg(), r_src1.GetReg(), r_scratch.GetReg(), info);
else
res = NewLIR3(alt_opcode | wide, r_dest.GetReg(), r_src1.GetReg(), r_scratch.GetReg());
FreeTemp(r_scratch);
return res;
}
LIR* Arm64Mir2Lir::OpRegImm(OpKind op, RegStorage r_dest_src1, int value) {
return OpRegImm64(op, r_dest_src1, static_cast<int64_t>(value));
}
LIR* Arm64Mir2Lir::OpRegImm64(OpKind op, RegStorage r_dest_src1, int64_t value) {
A64Opcode wide = (r_dest_src1.Is64Bit()) ? WIDE(0) : UNWIDE(0);
A64Opcode opcode = kA64Brk1d;
A64Opcode neg_opcode = kA64Brk1d;
bool shift;
bool neg = (value < 0);
uint64_t abs_value = (neg) ? -value : value;
if (LIKELY(abs_value < 0x1000)) {
// abs_value is a 12-bit immediate.
shift = false;
} else if ((abs_value & UINT64_C(0xfff)) == 0 && ((abs_value >> 12) < 0x1000)) {
// abs_value is a shifted 12-bit immediate.
shift = true;
abs_value >>= 12;
} else if (LIKELY(abs_value < 0x1000000 && (op == kOpAdd || op == kOpSub))) {
// Note: It is better to use two ADD/SUB instead of loading a number to a temp register.
// This works for both normal registers and SP.
// For a frame size == 0x2468, it will be encoded as:
// sub sp, #0x2000
// sub sp, #0x468
if (neg) {
op = (op == kOpAdd) ? kOpSub : kOpAdd;
}
OpRegImm64(op, r_dest_src1, abs_value & (~INT64_C(0xfff)));
return OpRegImm64(op, r_dest_src1, abs_value & 0xfff);
} else {
RegStorage r_tmp;
LIR* res;
if (IS_WIDE(wide)) {
r_tmp = AllocTempWide();
res = LoadConstantWide(r_tmp, value);
} else {
r_tmp = AllocTemp();
res = LoadConstant(r_tmp, value);
}
OpRegReg(op, r_dest_src1, r_tmp);
FreeTemp(r_tmp);
return res;
}
switch (op) {
case kOpAdd:
neg_opcode = kA64Sub4RRdT;
opcode = kA64Add4RRdT;
break;
case kOpSub:
neg_opcode = kA64Add4RRdT;
opcode = kA64Sub4RRdT;
break;
case kOpCmp:
neg_opcode = kA64Cmn3RdT;
opcode = kA64Cmp3RdT;
break;
default:
LOG(FATAL) << "Bad op-kind in OpRegImm: " << op;
break;
}
if (UNLIKELY(neg))
opcode = neg_opcode;
if (EncodingMap[opcode].flags & IS_QUAD_OP)
return NewLIR4(opcode | wide, r_dest_src1.GetReg(), r_dest_src1.GetReg(), abs_value,
(shift) ? 1 : 0);
else
return NewLIR3(opcode | wide, r_dest_src1.GetReg(), abs_value, (shift) ? 1 : 0);
}
int Arm64Mir2Lir::EncodeShift(int shift_type, int amount) {
DCHECK_EQ(shift_type & 0x3, shift_type);
DCHECK_EQ(amount & 0x3f, amount);
return ((shift_type & 0x3) << 7) | (amount & 0x3f);
}
int Arm64Mir2Lir::EncodeExtend(int extend_type, int amount) {
DCHECK_EQ(extend_type & 0x7, extend_type);
DCHECK_EQ(amount & 0x7, amount);
return (1 << 6) | ((extend_type & 0x7) << 3) | (amount & 0x7);
}
bool Arm64Mir2Lir::IsExtendEncoding(int encoded_value) {
return ((1 << 6) & encoded_value) != 0;
}
LIR* Arm64Mir2Lir::LoadBaseIndexed(RegStorage r_base, RegStorage r_index, RegStorage r_dest,
int scale, OpSize size) {
LIR* load;
int expected_scale = 0;
A64Opcode opcode = kA64Brk1d;
r_base = Check64BitReg(r_base);
// TODO(Arm64): The sign extension of r_index should be carried out by using an extended
// register offset load (rather than doing the sign extension in a separate instruction).
if (r_index.Is32Bit()) {
// Assemble: ``sxtw xN, wN''.
r_index = As64BitReg(r_index);
NewLIR4(WIDE(kA64Sbfm4rrdd), r_index.GetReg(), r_index.GetReg(), 0, 31);
}
if (r_dest.IsFloat()) {
if (r_dest.IsDouble()) {
DCHECK(size == k64 || size == kDouble);
expected_scale = 3;
opcode = WIDE(kA64Ldr4fXxG);
} else {
DCHECK(r_dest.IsSingle());
DCHECK(size == k32 || size == kSingle);
expected_scale = 2;
opcode = kA64Ldr4fXxG;
}
DCHECK(scale == 0 || scale == expected_scale);
return NewLIR4(opcode, r_dest.GetReg(), r_base.GetReg(), r_index.GetReg(),
(scale != 0) ? 1 : 0);
}
switch (size) {
case kDouble:
case kWord:
case k64:
r_dest = Check64BitReg(r_dest);
opcode = WIDE(kA64Ldr4rXxG);
expected_scale = 3;
break;
case kSingle: // Intentional fall-through.
case k32: // Intentional fall-through.
case kReference:
r_dest = Check32BitReg(r_dest);
opcode = kA64Ldr4rXxG;
expected_scale = 2;
break;
case kUnsignedHalf:
r_dest = Check32BitReg(r_dest);
opcode = kA64Ldrh4wXxd;
expected_scale = 1;
break;
case kSignedHalf:
r_dest = Check32BitReg(r_dest);
opcode = kA64Ldrsh4rXxd;
expected_scale = 1;
break;
case kUnsignedByte:
r_dest = Check32BitReg(r_dest);
opcode = kA64Ldrb3wXx;
break;
case kSignedByte:
r_dest = Check32BitReg(r_dest);
opcode = kA64Ldrsb3rXx;
break;
default:
LOG(FATAL) << "Bad size: " << size;
}
if (UNLIKELY(expected_scale == 0)) {
// This is a tertiary op (e.g. ldrb, ldrsb), it does not not support scale.
DCHECK_NE(EncodingMap[UNWIDE(opcode)].flags & IS_TERTIARY_OP, 0U);
DCHECK_EQ(scale, 0);
load = NewLIR3(opcode, r_dest.GetReg(), r_base.GetReg(), r_index.GetReg());
} else {
DCHECK(scale == 0 || scale == expected_scale);
load = NewLIR4(opcode, r_dest.GetReg(), r_base.GetReg(), r_index.GetReg(),
(scale != 0) ? 1 : 0);
}
return load;
}
LIR* Arm64Mir2Lir::LoadRefIndexed(RegStorage r_base, RegStorage r_index, RegStorage r_dest,
int scale) {
return LoadBaseIndexed(r_base, r_index, As32BitReg(r_dest), scale, kReference);
}
LIR* Arm64Mir2Lir::StoreBaseIndexed(RegStorage r_base, RegStorage r_index, RegStorage r_src,
int scale, OpSize size) {
LIR* store;
int expected_scale = 0;
A64Opcode opcode = kA64Brk1d;
r_base = Check64BitReg(r_base);
// TODO(Arm64): The sign extension of r_index should be carried out by using an extended
// register offset store (rather than doing the sign extension in a separate instruction).
if (r_index.Is32Bit()) {
// Assemble: ``sxtw xN, wN''.
r_index = As64BitReg(r_index);
NewLIR4(WIDE(kA64Sbfm4rrdd), r_index.GetReg(), r_index.GetReg(), 0, 31);
}
if (r_src.IsFloat()) {
if (r_src.IsDouble()) {
DCHECK(size == k64 || size == kDouble);
expected_scale = 3;
opcode = WIDE(kA64Str4fXxG);
} else {
DCHECK(r_src.IsSingle());
DCHECK(size == k32 || size == kSingle);
expected_scale = 2;
opcode = kA64Str4fXxG;
}
DCHECK(scale == 0 || scale == expected_scale);
return NewLIR4(opcode, r_src.GetReg(), r_base.GetReg(), r_index.GetReg(),
(scale != 0) ? 1 : 0);
}
switch (size) {
case kDouble: // Intentional fall-trough.
case kWord: // Intentional fall-trough.
case k64:
r_src = Check64BitReg(r_src);
opcode = WIDE(kA64Str4rXxG);
expected_scale = 3;
break;
case kSingle: // Intentional fall-trough.
case k32: // Intentional fall-trough.
case kReference:
r_src = Check32BitReg(r_src);
opcode = kA64Str4rXxG;
expected_scale = 2;
break;
case kUnsignedHalf:
case kSignedHalf:
r_src = Check32BitReg(r_src);
opcode = kA64Strh4wXxd;
expected_scale = 1;
break;
case kUnsignedByte:
case kSignedByte:
r_src = Check32BitReg(r_src);
opcode = kA64Strb3wXx;
break;
default:
LOG(FATAL) << "Bad size: " << size;
}
if (UNLIKELY(expected_scale == 0)) {
// This is a tertiary op (e.g. strb), it does not not support scale.
DCHECK_NE(EncodingMap[UNWIDE(opcode)].flags & IS_TERTIARY_OP, 0U);
DCHECK_EQ(scale, 0);
store = NewLIR3(opcode, r_src.GetReg(), r_base.GetReg(), r_index.GetReg());
} else {
store = NewLIR4(opcode, r_src.GetReg(), r_base.GetReg(), r_index.GetReg(),
(scale != 0) ? 1 : 0);
}
return store;
}
LIR* Arm64Mir2Lir::StoreRefIndexed(RegStorage r_base, RegStorage r_index, RegStorage r_src,
int scale) {
return StoreBaseIndexed(r_base, r_index, As32BitReg(r_src), scale, kReference);
}
/*
* Load value from base + displacement. Optionally perform null check
* on base (which must have an associated s_reg and MIR). If not
* performing null check, incoming MIR can be null.
*/
LIR* Arm64Mir2Lir::LoadBaseDispBody(RegStorage r_base, int displacement, RegStorage r_dest,
OpSize size) {
LIR* load = NULL;
A64Opcode opcode = kA64Brk1d;
A64Opcode alt_opcode = kA64Brk1d;
int scale = 0;
switch (size) {
case kDouble: // Intentional fall-through.
case kWord: // Intentional fall-through.
case k64:
r_dest = Check64BitReg(r_dest);
scale = 3;
if (r_dest.IsFloat()) {
DCHECK(r_dest.IsDouble());
opcode = WIDE(kA64Ldr3fXD);
alt_opcode = WIDE(kA64Ldur3fXd);
} else {
opcode = WIDE(kA64Ldr3rXD);
alt_opcode = WIDE(kA64Ldur3rXd);
}
break;
case kSingle: // Intentional fall-through.
case k32: // Intentional fall-trough.
case kReference:
r_dest = Check32BitReg(r_dest);
scale = 2;
if (r_dest.IsFloat()) {
DCHECK(r_dest.IsSingle());
opcode = kA64Ldr3fXD;
} else {
opcode = kA64Ldr3rXD;
}
break;
case kUnsignedHalf:
scale = 1;
opcode = kA64Ldrh3wXF;
break;
case kSignedHalf:
scale = 1;
opcode = kA64Ldrsh3rXF;
break;
case kUnsignedByte:
opcode = kA64Ldrb3wXd;
break;
case kSignedByte:
opcode = kA64Ldrsb3rXd;
break;
default:
LOG(FATAL) << "Bad size: " << size;
}
bool displacement_is_aligned = (displacement & ((1 << scale) - 1)) == 0;
int scaled_disp = displacement >> scale;
if (displacement_is_aligned && scaled_disp >= 0 && scaled_disp < 4096) {
// Can use scaled load.
load = NewLIR3(opcode, r_dest.GetReg(), r_base.GetReg(), scaled_disp);
} else if (alt_opcode != kA64Brk1d && IS_SIGNED_IMM9(displacement)) {
// Can use unscaled load.
load = NewLIR3(alt_opcode, r_dest.GetReg(), r_base.GetReg(), displacement);
} else {
// Use long sequence.
// TODO: cleaner support for index/displacement registers? Not a reference, but must match width.
RegStorage r_scratch = AllocTempWide();
LoadConstantWide(r_scratch, displacement);
load = LoadBaseIndexed(r_base, r_scratch, r_dest, 0, size);
FreeTemp(r_scratch);
}
// TODO: in future may need to differentiate Dalvik accesses w/ spills
if (mem_ref_type_ == ResourceMask::kDalvikReg) {
DCHECK(r_base == rs_sp);
AnnotateDalvikRegAccess(load, displacement >> 2, true /* is_load */, r_dest.Is64Bit());
}
return load;
}
LIR* Arm64Mir2Lir::LoadBaseDisp(RegStorage r_base, int displacement, RegStorage r_dest,
OpSize size, VolatileKind is_volatile) {
// LoadBaseDisp() will emit correct insn for atomic load on arm64
// assuming r_dest is correctly prepared using RegClassForFieldLoadStore().
LIR* load = LoadBaseDispBody(r_base, displacement, r_dest, size);
if (UNLIKELY(is_volatile == kVolatile)) {
// TODO: This should generate an acquire load instead of the barrier.
GenMemBarrier(kLoadAny);
}
return load;
}
LIR* Arm64Mir2Lir::LoadRefDisp(RegStorage r_base, int displacement, RegStorage r_dest,
VolatileKind is_volatile) {
return LoadBaseDisp(r_base, displacement, As32BitReg(r_dest), kReference, is_volatile);
}
LIR* Arm64Mir2Lir::StoreBaseDispBody(RegStorage r_base, int displacement, RegStorage r_src,
OpSize size) {
LIR* store = NULL;
A64Opcode opcode = kA64Brk1d;
A64Opcode alt_opcode = kA64Brk1d;
int scale = 0;
switch (size) {
case kDouble: // Intentional fall-through.
case kWord: // Intentional fall-through.
case k64:
r_src = Check64BitReg(r_src);
scale = 3;
if (r_src.IsFloat()) {
DCHECK(r_src.IsDouble());
opcode = WIDE(kA64Str3fXD);
alt_opcode = WIDE(kA64Stur3fXd);
} else {
opcode = WIDE(kA64Str3rXD);
alt_opcode = WIDE(kA64Stur3rXd);
}
break;
case kSingle: // Intentional fall-through.
case k32: // Intentional fall-trough.
case kReference:
r_src = Check32BitReg(r_src);
scale = 2;
if (r_src.IsFloat()) {
DCHECK(r_src.IsSingle());
opcode = kA64Str3fXD;
} else {
opcode = kA64Str3rXD;
}
break;
case kUnsignedHalf:
case kSignedHalf:
scale = 1;
opcode = kA64Strh3wXF;
break;
case kUnsignedByte:
case kSignedByte:
opcode = kA64Strb3wXd;
break;
default:
LOG(FATAL) << "Bad size: " << size;
}
bool displacement_is_aligned = (displacement & ((1 << scale) - 1)) == 0;
int scaled_disp = displacement >> scale;
if (displacement_is_aligned && scaled_disp >= 0 && scaled_disp < 4096) {
// Can use scaled store.
store = NewLIR3(opcode, r_src.GetReg(), r_base.GetReg(), scaled_disp);
} else if (alt_opcode != kA64Brk1d && IS_SIGNED_IMM9(displacement)) {
// Can use unscaled store.
store = NewLIR3(alt_opcode, r_src.GetReg(), r_base.GetReg(), displacement);
} else {
// Use long sequence.
RegStorage r_scratch = AllocTempWide();
LoadConstantWide(r_scratch, displacement);
store = StoreBaseIndexed(r_base, r_scratch, r_src, 0, size);
FreeTemp(r_scratch);
}
// TODO: In future, may need to differentiate Dalvik & spill accesses.
if (mem_ref_type_ == ResourceMask::kDalvikReg) {
DCHECK(r_base == rs_sp);
AnnotateDalvikRegAccess(store, displacement >> 2, false /* is_load */, r_src.Is64Bit());
}
return store;
}
LIR* Arm64Mir2Lir::StoreBaseDisp(RegStorage r_base, int displacement, RegStorage r_src,
OpSize size, VolatileKind is_volatile) {
// TODO: This should generate a release store and no barriers.
if (UNLIKELY(is_volatile == kVolatile)) {
// Ensure that prior accesses become visible to other threads first.
GenMemBarrier(kAnyStore);
}
// StoreBaseDisp() will emit correct insn for atomic store on arm64
// assuming r_dest is correctly prepared using RegClassForFieldLoadStore().
LIR* store = StoreBaseDispBody(r_base, displacement, r_src, size);
if (UNLIKELY(is_volatile == kVolatile)) {
// Preserve order with respect to any subsequent volatile loads.
// We need StoreLoad, but that generally requires the most expensive barrier.
GenMemBarrier(kAnyAny);
}
return store;
}
LIR* Arm64Mir2Lir::StoreRefDisp(RegStorage r_base, int displacement, RegStorage r_src,
VolatileKind is_volatile) {
return StoreBaseDisp(r_base, displacement, As32BitReg(r_src), kReference, is_volatile);
}
LIR* Arm64Mir2Lir::OpFpRegCopy(RegStorage r_dest, RegStorage r_src) {
LOG(FATAL) << "Unexpected use of OpFpRegCopy for Arm64";
return NULL;
}
LIR* Arm64Mir2Lir::OpMem(OpKind op, RegStorage r_base, int disp) {
LOG(FATAL) << "Unexpected use of OpMem for Arm64";
return NULL;
}
LIR* Arm64Mir2Lir::InvokeTrampoline(OpKind op, RegStorage r_tgt, QuickEntrypointEnum trampoline) {
return OpReg(op, r_tgt);
}
} // namespace art