blob: c7ca0035301077c8a866aca802cc1ebafe5948c5 [file] [log] [blame]
/*
* Copyright (C) 2016 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <iostream>
#include <type_traits>
#include "assembler_arm_vixl.h"
#include "base/bit_utils.h"
#include "base/bit_utils_iterator.h"
#include "entrypoints/quick/quick_entrypoints.h"
#include "heap_poisoning.h"
#include "thread.h"
using namespace vixl::aarch32; // NOLINT(build/namespaces)
namespace art HIDDEN {
namespace arm {
#ifdef ___
#error "ARM Assembler macro already defined."
#else
#define ___ vixl_masm_.
#endif
// Thread register definition.
extern const vixl32::Register tr(TR);
// Marking register definition.
extern const vixl32::Register mr(MR);
void ArmVIXLAssembler::FinalizeCode() {
vixl_masm_.FinalizeCode();
}
size_t ArmVIXLAssembler::CodeSize() const {
return vixl_masm_.GetSizeOfCodeGenerated();
}
const uint8_t* ArmVIXLAssembler::CodeBufferBaseAddress() const {
return vixl_masm_.GetBuffer().GetStartAddress<const uint8_t*>();
}
void ArmVIXLAssembler::FinalizeInstructions(const MemoryRegion& region) {
// Copy the instructions from the buffer.
MemoryRegion from(vixl_masm_.GetBuffer()->GetStartAddress<void*>(), CodeSize());
region.CopyFrom(0, from);
}
void ArmVIXLAssembler::PoisonHeapReference(vixl::aarch32::Register reg) {
// reg = -reg.
___ Rsb(reg, reg, 0);
}
void ArmVIXLAssembler::UnpoisonHeapReference(vixl::aarch32::Register reg) {
// reg = -reg.
___ Rsb(reg, reg, 0);
}
void ArmVIXLAssembler::MaybePoisonHeapReference(vixl32::Register reg) {
if (kPoisonHeapReferences) {
PoisonHeapReference(reg);
}
}
void ArmVIXLAssembler::MaybeUnpoisonHeapReference(vixl32::Register reg) {
if (kPoisonHeapReferences) {
UnpoisonHeapReference(reg);
}
}
void ArmVIXLAssembler::GenerateMarkingRegisterCheck(vixl32::Register temp, int code) {
DCHECK(kReserveMarkingRegister);
vixl32::Label mr_is_ok;
// temp = self.tls32_.is.gc_marking
___ Ldr(temp, MemOperand(tr, Thread::IsGcMarkingOffset<kArmPointerSize>().Int32Value()));
// Check that mr == self.tls32_.is.gc_marking.
___ Cmp(mr, temp);
___ B(eq, &mr_is_ok, /* is_far_target= */ false);
___ Bkpt(code);
___ Bind(&mr_is_ok);
}
void ArmVIXLAssembler::LoadImmediate(vixl32::Register rd, int32_t value) {
// TODO(VIXL): Implement this optimization in VIXL.
if (!ShifterOperandCanAlwaysHold(value) && ShifterOperandCanAlwaysHold(~value)) {
___ Mvn(rd, ~value);
} else {
___ Mov(rd, value);
}
}
bool ArmVIXLAssembler::ShifterOperandCanAlwaysHold(uint32_t immediate) {
return vixl_masm_.IsModifiedImmediate(immediate);
}
bool ArmVIXLAssembler::ShifterOperandCanHold(Opcode opcode,
uint32_t immediate,
vixl::aarch32::FlagsUpdate update_flags) {
switch (opcode) {
case ADD:
case SUB:
// Less than (or equal to) 12 bits can be done if we don't need to set condition codes.
if (IsUint<12>(immediate) && update_flags != vixl::aarch32::SetFlags) {
return true;
}
return ShifterOperandCanAlwaysHold(immediate);
case MOV:
// TODO: Support less than or equal to 12bits.
return ShifterOperandCanAlwaysHold(immediate);
case MVN:
default:
return ShifterOperandCanAlwaysHold(immediate);
}
}
bool ArmVIXLAssembler::CanSplitLoadStoreOffset(int32_t allowed_offset_bits,
int32_t offset,
/*out*/ int32_t* add_to_base,
/*out*/ int32_t* offset_for_load_store) {
int32_t other_bits = offset & ~allowed_offset_bits;
if (ShifterOperandCanAlwaysHold(other_bits) || ShifterOperandCanAlwaysHold(-other_bits)) {
*add_to_base = offset & ~allowed_offset_bits;
*offset_for_load_store = offset & allowed_offset_bits;
return true;
}
return false;
}
int32_t ArmVIXLAssembler::AdjustLoadStoreOffset(int32_t allowed_offset_bits,
vixl32::Register temp,
vixl32::Register base,
int32_t offset) {
DCHECK_NE(offset & ~allowed_offset_bits, 0);
int32_t add_to_base, offset_for_load;
if (CanSplitLoadStoreOffset(allowed_offset_bits, offset, &add_to_base, &offset_for_load)) {
___ Add(temp, base, add_to_base);
return offset_for_load;
} else {
___ Mov(temp, offset);
___ Add(temp, temp, base);
return 0;
}
}
// TODO(VIXL): Implement this in VIXL.
int32_t ArmVIXLAssembler::GetAllowedLoadOffsetBits(LoadOperandType type) {
switch (type) {
case kLoadSignedByte:
case kLoadSignedHalfword:
case kLoadUnsignedHalfword:
case kLoadUnsignedByte:
case kLoadWord:
// We can encode imm12 offset.
return 0xfff;
case kLoadSWord:
case kLoadDWord:
case kLoadWordPair:
// We can encode imm8:'00' offset.
return 0xff << 2;
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
}
// TODO(VIXL): Implement this in VIXL.
int32_t ArmVIXLAssembler::GetAllowedStoreOffsetBits(StoreOperandType type) {
switch (type) {
case kStoreHalfword:
case kStoreByte:
case kStoreWord:
// We can encode imm12 offset.
return 0xfff;
case kStoreSWord:
case kStoreDWord:
case kStoreWordPair:
// We can encode imm8:'00' offset.
return 0xff << 2;
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
}
// TODO(VIXL): Implement this in VIXL.
static bool CanHoldLoadOffsetThumb(LoadOperandType type, int offset) {
switch (type) {
case kLoadSignedByte:
case kLoadSignedHalfword:
case kLoadUnsignedHalfword:
case kLoadUnsignedByte:
case kLoadWord:
return IsAbsoluteUint<12>(offset);
case kLoadSWord:
case kLoadDWord:
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode.
case kLoadWordPair:
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset);
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
}
// TODO(VIXL): Implement this in VIXL.
static bool CanHoldStoreOffsetThumb(StoreOperandType type, int offset) {
switch (type) {
case kStoreHalfword:
case kStoreByte:
case kStoreWord:
return IsAbsoluteUint<12>(offset);
case kStoreSWord:
case kStoreDWord:
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode.
case kStoreWordPair:
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset);
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
}
// Implementation note: this method must emit at most one instruction when
// Address::CanHoldStoreOffsetThumb.
// TODO(VIXL): Implement AdjustLoadStoreOffset logic in VIXL.
void ArmVIXLAssembler::StoreToOffset(StoreOperandType type,
vixl32::Register reg,
vixl32::Register base,
int32_t offset) {
vixl32::Register tmp_reg;
UseScratchRegisterScope temps(&vixl_masm_);
if (!CanHoldStoreOffsetThumb(type, offset)) {
CHECK_NE(base.GetCode(), kIpCode);
if ((reg.GetCode() != kIpCode) &&
(!vixl_masm_.GetScratchRegisterList()->IsEmpty()) &&
((type != kStoreWordPair) || (reg.GetCode() + 1 != kIpCode))) {
tmp_reg = temps.Acquire();
} else {
// Be careful not to use ip twice (for `reg` (or `reg` + 1 in
// the case of a word-pair store) and `base`) to build the
// Address object used by the store instruction(s) below.
// Instead, save R5 on the stack (or R6 if R5 is already used by
// `base`), use it as secondary temporary register, and restore
// it after the store instruction has been emitted.
tmp_reg = (base.GetCode() != 5) ? r5 : r6;
___ Push(tmp_reg);
if (base.GetCode() == kSpCode) {
offset += kRegisterSize;
}
}
// TODO: Implement indexed store (not available for STRD), inline AdjustLoadStoreOffset()
// and in the "unsplittable" path get rid of the "add" by using the store indexed instead.
offset = AdjustLoadStoreOffset(GetAllowedStoreOffsetBits(type), tmp_reg, base, offset);
base = tmp_reg;
}
DCHECK(CanHoldStoreOffsetThumb(type, offset));
switch (type) {
case kStoreByte:
___ Strb(reg, MemOperand(base, offset));
break;
case kStoreHalfword:
___ Strh(reg, MemOperand(base, offset));
break;
case kStoreWord:
___ Str(reg, MemOperand(base, offset));
break;
case kStoreWordPair:
___ Strd(reg, vixl32::Register(reg.GetCode() + 1), MemOperand(base, offset));
break;
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
if ((tmp_reg.IsValid()) && (tmp_reg.GetCode() != kIpCode)) {
CHECK(tmp_reg.Is(r5) || tmp_reg.Is(r6)) << tmp_reg;
___ Pop(tmp_reg);
}
}
// Implementation note: this method must emit at most one instruction when
// Address::CanHoldLoadOffsetThumb.
// TODO(VIXL): Implement AdjustLoadStoreOffset logic in VIXL.
void ArmVIXLAssembler::LoadFromOffset(LoadOperandType type,
vixl32::Register dest,
vixl32::Register base,
int32_t offset) {
if (!CanHoldLoadOffsetThumb(type, offset)) {
CHECK(!base.Is(ip));
// Inlined AdjustLoadStoreOffset() allows us to pull a few more tricks.
int32_t allowed_offset_bits = GetAllowedLoadOffsetBits(type);
DCHECK_NE(offset & ~allowed_offset_bits, 0);
int32_t add_to_base, offset_for_load;
if (CanSplitLoadStoreOffset(allowed_offset_bits, offset, &add_to_base, &offset_for_load)) {
// Use reg for the adjusted base. If it's low reg, we may end up using 16-bit load.
AddConstant(dest, base, add_to_base);
base = dest;
offset = offset_for_load;
} else {
UseScratchRegisterScope temps(&vixl_masm_);
vixl32::Register temp = (dest.Is(base)) ? temps.Acquire() : dest;
LoadImmediate(temp, offset);
// TODO: Implement indexed load (not available for LDRD) and use it here to avoid the ADD.
// Use reg for the adjusted base. If it's low reg, we may end up using 16-bit load.
___ Add(dest, dest, (dest.Is(base)) ? temp : base);
base = dest;
offset = 0;
}
}
DCHECK(CanHoldLoadOffsetThumb(type, offset));
switch (type) {
case kLoadSignedByte:
___ Ldrsb(dest, MemOperand(base, offset));
break;
case kLoadUnsignedByte:
___ Ldrb(dest, MemOperand(base, offset));
break;
case kLoadSignedHalfword:
___ Ldrsh(dest, MemOperand(base, offset));
break;
case kLoadUnsignedHalfword:
___ Ldrh(dest, MemOperand(base, offset));
break;
case kLoadWord:
CHECK(!dest.IsSP());
___ Ldr(dest, MemOperand(base, offset));
break;
case kLoadWordPair:
___ Ldrd(dest, vixl32::Register(dest.GetCode() + 1), MemOperand(base, offset));
break;
default:
LOG(FATAL) << "UNREACHABLE";
UNREACHABLE();
}
}
void ArmVIXLAssembler::StoreSToOffset(vixl32::SRegister source,
vixl32::Register base,
int32_t offset) {
___ Vstr(source, MemOperand(base, offset));
}
void ArmVIXLAssembler::StoreDToOffset(vixl32::DRegister source,
vixl32::Register base,
int32_t offset) {
___ Vstr(source, MemOperand(base, offset));
}
void ArmVIXLAssembler::LoadSFromOffset(vixl32::SRegister reg,
vixl32::Register base,
int32_t offset) {
___ Vldr(reg, MemOperand(base, offset));
}
void ArmVIXLAssembler::LoadDFromOffset(vixl32::DRegister reg,
vixl32::Register base,
int32_t offset) {
___ Vldr(reg, MemOperand(base, offset));
}
// Prefer Str to Add/Stm in ArmVIXLAssembler::StoreRegisterList and
// ArmVIXLAssembler::LoadRegisterList where this generates less code (size).
static constexpr int kRegListThreshold = 4;
void ArmVIXLAssembler::StoreRegisterList(RegList regs, size_t stack_offset) {
int number_of_regs = POPCOUNT(static_cast<uint32_t>(regs));
if (number_of_regs != 0) {
if (number_of_regs > kRegListThreshold) {
UseScratchRegisterScope temps(GetVIXLAssembler());
vixl32::Register base = sp;
if (stack_offset != 0) {
base = temps.Acquire();
DCHECK_EQ(regs & (1u << base.GetCode()), 0u);
___ Add(base, sp, Operand::From(stack_offset));
}
___ Stm(base, NO_WRITE_BACK, RegisterList(regs));
} else {
for (uint32_t i : LowToHighBits(static_cast<uint32_t>(regs))) {
___ Str(vixl32::Register(i), MemOperand(sp, stack_offset));
stack_offset += kRegSizeInBytes;
}
}
}
}
void ArmVIXLAssembler::LoadRegisterList(RegList regs, size_t stack_offset) {
int number_of_regs = POPCOUNT(static_cast<uint32_t>(regs));
if (number_of_regs != 0) {
if (number_of_regs > kRegListThreshold) {
UseScratchRegisterScope temps(GetVIXLAssembler());
vixl32::Register base = sp;
if (stack_offset != 0) {
base = temps.Acquire();
___ Add(base, sp, Operand::From(stack_offset));
}
___ Ldm(base, NO_WRITE_BACK, RegisterList(regs));
} else {
for (uint32_t i : LowToHighBits(static_cast<uint32_t>(regs))) {
___ Ldr(vixl32::Register(i), MemOperand(sp, stack_offset));
stack_offset += kRegSizeInBytes;
}
}
}
}
void ArmVIXLAssembler::AddConstant(vixl32::Register rd, int32_t value) {
AddConstant(rd, rd, value);
}
// TODO(VIXL): think about using adds which updates flags where possible.
void ArmVIXLAssembler::AddConstant(vixl32::Register rd,
vixl32::Register rn,
int32_t value) {
DCHECK(vixl_masm_.OutsideITBlock());
// TODO(VIXL): implement this optimization in VIXL.
if (value == 0) {
if (!rd.Is(rn)) {
___ Mov(rd, rn);
}
return;
}
___ Add(rd, rn, value);
}
// Inside IT block we must use assembler, macroassembler instructions are not permitted.
void ArmVIXLAssembler::AddConstantInIt(vixl32::Register rd,
vixl32::Register rn,
int32_t value,
vixl32::Condition cond) {
DCHECK(vixl_masm_.InITBlock());
if (value == 0) {
___ mov(cond, rd, rn);
} else {
___ add(cond, rd, rn, value);
}
}
void ArmVIXLMacroAssembler::CompareAndBranchIfZero(vixl32::Register rn,
vixl32::Label* label,
bool is_far_target) {
if (!is_far_target && rn.IsLow() && !label->IsBound()) {
// In T32, Cbz/Cbnz instructions have following limitations:
// - There are only 7 bits (i:imm5:0) to encode branch target address (cannot be far target).
// - Only low registers (i.e R0 .. R7) can be encoded.
// - Only forward branches (unbound labels) are supported.
Cbz(rn, label);
return;
}
Cmp(rn, 0);
B(eq, label, is_far_target);
}
void ArmVIXLMacroAssembler::CompareAndBranchIfNonZero(vixl32::Register rn,
vixl32::Label* label,
bool is_far_target) {
if (!is_far_target && rn.IsLow() && !label->IsBound()) {
Cbnz(rn, label);
return;
}
Cmp(rn, 0);
B(ne, label, is_far_target);
}
void ArmVIXLMacroAssembler::B(vixl32::Label* label) {
if (!label->IsBound()) {
// Try to use a 16-bit encoding of the B instruction.
DCHECK(OutsideITBlock());
BPreferNear(label);
return;
}
MacroAssembler::B(label);
}
void ArmVIXLMacroAssembler::B(vixl32::Condition cond, vixl32::Label* label, bool is_far_target) {
if (!label->IsBound() && !is_far_target) {
// Try to use a 16-bit encoding of the B instruction.
DCHECK(OutsideITBlock());
BPreferNear(cond, label);
return;
}
MacroAssembler::B(cond, label);
}
} // namespace arm
} // namespace art