blob: 17d5122542367ac829cb5509bfa39411801dea37 [file] [log] [blame]
/*
* Copyright (C) 2014 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_COMPILER_OPTIMIZING_PARALLEL_MOVE_RESOLVER_H_
#define ART_COMPILER_OPTIMIZING_PARALLEL_MOVE_RESOLVER_H_
#include "base/arena_containers.h"
#include "base/macros.h"
#include "base/value_object.h"
#include "data_type.h"
#include "locations.h"
namespace art HIDDEN {
class HParallelMove;
class MoveOperands;
// Helper classes to resolve a set of parallel moves. Architecture dependent code generator must
// have their own subclass that implements corresponding virtual functions.
class ParallelMoveResolver : public ValueObject {
public:
explicit ParallelMoveResolver(ArenaAllocator* allocator)
: moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)) {
moves_.reserve(32);
}
virtual ~ParallelMoveResolver() {}
// Resolve a set of parallel moves, emitting assembler instructions.
virtual void EmitNativeCode(HParallelMove* parallel_move) = 0;
protected:
// Build the initial list of moves.
void BuildInitialMoveList(HParallelMove* parallel_move);
ArenaVector<MoveOperands*> moves_;
private:
DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolver);
};
// This helper class uses swap to resolve dependencies and may emit swap.
class ParallelMoveResolverWithSwap : public ParallelMoveResolver {
public:
explicit ParallelMoveResolverWithSwap(ArenaAllocator* allocator)
: ParallelMoveResolver(allocator) {}
virtual ~ParallelMoveResolverWithSwap() {}
// Resolve a set of parallel moves, emitting assembler instructions.
void EmitNativeCode(HParallelMove* parallel_move) override;
protected:
class ScratchRegisterScope : public ValueObject {
public:
ScratchRegisterScope(ParallelMoveResolverWithSwap* resolver,
int blocked,
int if_scratch,
int number_of_registers);
~ScratchRegisterScope();
int GetRegister() const { return reg_; }
bool IsSpilled() const { return spilled_; }
private:
ParallelMoveResolverWithSwap* resolver_;
int reg_;
bool spilled_;
};
// Return true if the location can be scratched.
bool IsScratchLocation(Location loc);
// Allocate a scratch register for performing a move. The method will try to use
// a register that is the destination of a move, but that move has not been emitted yet.
int AllocateScratchRegister(int blocked, int if_scratch, int register_count, bool* spilled);
// Emit a move.
virtual void EmitMove(size_t index) = 0;
// Execute a move by emitting a swap of two operands.
virtual void EmitSwap(size_t index) = 0;
virtual void SpillScratch(int reg) = 0;
virtual void RestoreScratch(int reg) = 0;
static constexpr int kNoRegister = -1;
private:
// Perform the move at the moves_ index in question (possibly requiring
// other moves to satisfy dependencies).
//
// Return whether another move in the dependency cycle needs to swap. This
// is to handle 64bits swaps:
// 1) In the case of register pairs, where we want the pair to swap first to avoid
// building pairs that are unexpected by the code generator. For example, if
// we were to swap R1 with R2, we would need to update all locations using
// R2 to R1. So a (R2,R3) pair register could become (R1,R3). We could make
// the code generator understand such pairs, but it's easier and cleaner to
// just not create such pairs and exchange pairs in priority.
// 2) Even when the architecture does not have pairs, we must handle 64bits swaps
// first. Consider the case: (R0->R1) (R1->S) (S->R0), where 'S' is a single
// stack slot. If we end up swapping S and R0, S will only contain the low bits
// of R0. If R0->R1 is for a 64bits instruction, R1 will therefore not contain
// the right value.
MoveOperands* PerformMove(size_t index);
DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolverWithSwap);
};
// This helper class uses additional scratch registers to resolve dependencies. It supports all kind
// of dependency cycles and does not care about the register layout.
class ParallelMoveResolverNoSwap : public ParallelMoveResolver {
public:
explicit ParallelMoveResolverNoSwap(ArenaAllocator* allocator)
: ParallelMoveResolver(allocator),
scratches_(allocator->Adapter(kArenaAllocParallelMoveResolver)),
pending_moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)),
allocator_(allocator) {
scratches_.reserve(32);
pending_moves_.reserve(8);
}
virtual ~ParallelMoveResolverNoSwap() {}
// Resolve a set of parallel moves, emitting assembler instructions.
void EmitNativeCode(HParallelMove* parallel_move) override;
protected:
// Called at the beginning of EmitNativeCode(). A subclass may put some architecture dependent
// initialization here.
virtual void PrepareForEmitNativeCode() = 0;
// Called at the end of EmitNativeCode(). A subclass may put some architecture dependent cleanup
// here. All scratch locations will be removed after this call.
virtual void FinishEmitNativeCode() = 0;
// Allocate a scratch location to perform a move from input kind of location. A subclass should
// implement this to get the best fit location. If there is no suitable physical register, it can
// also return a stack slot.
virtual Location AllocateScratchLocationFor(Location::Kind kind) = 0;
// Called after a move which takes a scratch location as source. A subclass can defer the cleanup
// to FinishEmitNativeCode().
virtual void FreeScratchLocation(Location loc) = 0;
// Emit a move.
virtual void EmitMove(size_t index) = 0;
// Return a scratch location from the moves which exactly matches the kind.
// Return Location::NoLocation() if no matching scratch location can be found.
Location GetScratchLocation(Location::Kind kind);
// Add a location to the scratch list which can be returned from GetScratchLocation() to resolve
// dependency cycles.
void AddScratchLocation(Location loc);
// Remove a location from the scratch list.
void RemoveScratchLocation(Location loc);
// List of scratch locations.
ArenaVector<Location> scratches_;
private:
// Perform the move at the given index in `moves_` (possibly requiring other moves to satisfy
// dependencies).
void PerformMove(size_t index);
void UpdateMoveSource(Location from, Location to);
void AddPendingMove(Location source, Location destination, DataType::Type type);
void DeletePendingMove(MoveOperands* move);
// Find a move that may be unblocked after (loc -> XXX) is performed.
MoveOperands* GetUnblockedPendingMove(Location loc);
// Return true if the location is blocked by outstanding moves.
bool IsBlockedByMoves(Location loc);
// Return the number of pending moves.
size_t GetNumberOfPendingMoves();
// Additional pending moves which might be added to resolve dependency cycle.
ArenaVector<MoveOperands*> pending_moves_;
// Used to allocate pending MoveOperands.
ArenaAllocator* const allocator_;
DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolverNoSwap);
};
} // namespace art
#endif // ART_COMPILER_OPTIMIZING_PARALLEL_MOVE_RESOLVER_H_