blob: f541d8fa19326070cab860163d4efbc4ac280bc0 [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_COMPILER_JNI_QUICK_CALLING_CONVENTION_H_
#define ART_COMPILER_JNI_QUICK_CALLING_CONVENTION_H_
#include "base/arena_object.h"
#include "base/array_ref.h"
#include "base/enums.h"
#include "handle_scope.h"
#include "primitive.h"
#include "thread.h"
#include "utils/managed_register.h"
namespace art {
// Top-level abstraction for different calling conventions.
class CallingConvention : public DeletableArenaObject<kArenaAllocCallingConvention> {
public:
bool IsReturnAReference() const { return shorty_[0] == 'L'; }
Primitive::Type GetReturnType() const {
return Primitive::GetType(shorty_[0]);
}
size_t SizeOfReturnValue() const {
size_t result = Primitive::ComponentSize(Primitive::GetType(shorty_[0]));
if (result >= 1 && result < 4) {
result = 4;
}
return result;
}
// Register that holds result of this method invocation.
virtual ManagedRegister ReturnRegister() = 0;
// Register reserved for scratch usage during procedure calls.
virtual ManagedRegister InterproceduralScratchRegister() = 0;
// Offset of Method within the frame.
FrameOffset MethodStackOffset() {
return displacement_;
}
// Iterator interface
// Place iterator at start of arguments. The displacement is applied to
// frame offset methods to account for frames which may be on the stack
// below the one being iterated over.
void ResetIterator(FrameOffset displacement) {
displacement_ = displacement;
itr_slots_ = 0;
itr_args_ = 0;
itr_refs_ = 0;
itr_longs_and_doubles_ = 0;
itr_float_and_doubles_ = 0;
}
virtual ~CallingConvention() {}
protected:
CallingConvention(bool is_static,
bool is_synchronized,
const char* shorty,
PointerSize frame_pointer_size)
: itr_slots_(0), itr_refs_(0), itr_args_(0), itr_longs_and_doubles_(0),
itr_float_and_doubles_(0), displacement_(0),
frame_pointer_size_(frame_pointer_size),
handle_scope_pointer_size_(sizeof(StackReference<mirror::Object>)),
is_static_(is_static), is_synchronized_(is_synchronized),
shorty_(shorty) {
num_args_ = (is_static ? 0 : 1) + strlen(shorty) - 1;
num_ref_args_ = is_static ? 0 : 1; // The implicit this pointer.
num_float_or_double_args_ = 0;
num_long_or_double_args_ = 0;
for (size_t i = 1; i < strlen(shorty); i++) {
char ch = shorty_[i];
switch (ch) {
case 'L':
num_ref_args_++;
break;
case 'J':
num_long_or_double_args_++;
break;
case 'D':
num_long_or_double_args_++;
num_float_or_double_args_++;
break;
case 'F':
num_float_or_double_args_++;
break;
}
}
}
bool IsStatic() const {
return is_static_;
}
bool IsSynchronized() const {
return is_synchronized_;
}
bool IsParamALongOrDouble(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return false; // this argument
}
char ch = shorty_[param];
return (ch == 'J' || ch == 'D');
}
bool IsParamAFloatOrDouble(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return false; // this argument
}
char ch = shorty_[param];
return (ch == 'F' || ch == 'D');
}
bool IsParamADouble(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return false; // this argument
}
return shorty_[param] == 'D';
}
bool IsParamALong(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return false; // this argument
}
return shorty_[param] == 'J';
}
bool IsParamAReference(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return true; // this argument
}
return shorty_[param] == 'L';
}
size_t NumArgs() const {
return num_args_;
}
// Implicit argument count: 1 for instance functions, 0 for static functions.
// (The implicit argument is only relevant to the shorty, i.e.
// the 0th arg is not in the shorty if it's implicit).
size_t NumImplicitArgs() const {
return IsStatic() ? 0 : 1;
}
size_t NumLongOrDoubleArgs() const {
return num_long_or_double_args_;
}
size_t NumFloatOrDoubleArgs() const {
return num_float_or_double_args_;
}
size_t NumReferenceArgs() const {
return num_ref_args_;
}
size_t ParamSize(unsigned int param) const {
DCHECK_LT(param, NumArgs());
if (IsStatic()) {
param++; // 0th argument must skip return value at start of the shorty
} else if (param == 0) {
return sizeof(mirror::HeapReference<mirror::Object>); // this argument
}
size_t result = Primitive::ComponentSize(Primitive::GetType(shorty_[param]));
if (result >= 1 && result < 4) {
result = 4;
}
return result;
}
const char* GetShorty() const {
return shorty_.c_str();
}
// The slot number for current calling_convention argument.
// Note that each slot is 32-bit. When the current argument is bigger
// than 32 bits, return the first slot number for this argument.
unsigned int itr_slots_;
// The number of references iterated past.
unsigned int itr_refs_;
// The argument number along argument list for current argument.
unsigned int itr_args_;
// Number of longs and doubles seen along argument list.
unsigned int itr_longs_and_doubles_;
// Number of float and doubles seen along argument list.
unsigned int itr_float_and_doubles_;
// Space for frames below this on the stack.
FrameOffset displacement_;
// The size of a pointer.
const PointerSize frame_pointer_size_;
// The size of a reference entry within the handle scope.
const size_t handle_scope_pointer_size_;
private:
const bool is_static_;
const bool is_synchronized_;
std::string shorty_;
size_t num_args_;
size_t num_ref_args_;
size_t num_float_or_double_args_;
size_t num_long_or_double_args_;
};
// Abstraction for managed code's calling conventions
// | { Incoming stack args } |
// | { Prior Method* } | <-- Prior SP
// | { Return address } |
// | { Callee saves } |
// | { Spills ... } |
// | { Outgoing stack args } |
// | { Method* } | <-- SP
class ManagedRuntimeCallingConvention : public CallingConvention {
public:
static std::unique_ptr<ManagedRuntimeCallingConvention> Create(ArenaAllocator* arena,
bool is_static,
bool is_synchronized,
const char* shorty,
InstructionSet instruction_set);
// Register that holds the incoming method argument
virtual ManagedRegister MethodRegister() = 0;
// Iterator interface
bool HasNext();
void Next();
bool IsCurrentParamAReference();
bool IsCurrentParamAFloatOrDouble();
bool IsCurrentParamADouble();
bool IsCurrentParamALong();
bool IsCurrentArgExplicit(); // ie a non-implict argument such as this
bool IsCurrentArgPossiblyNull();
size_t CurrentParamSize();
virtual bool IsCurrentParamInRegister() = 0;
virtual bool IsCurrentParamOnStack() = 0;
virtual ManagedRegister CurrentParamRegister() = 0;
virtual FrameOffset CurrentParamStackOffset() = 0;
virtual ~ManagedRuntimeCallingConvention() {}
// Registers to spill to caller's out registers on entry.
virtual const ManagedRegisterEntrySpills& EntrySpills() = 0;
protected:
ManagedRuntimeCallingConvention(bool is_static,
bool is_synchronized,
const char* shorty,
PointerSize frame_pointer_size)
: CallingConvention(is_static, is_synchronized, shorty, frame_pointer_size) {}
};
// Abstraction for JNI calling conventions
// | { Incoming stack args } | <-- Prior SP
// | { Return address } |
// | { Callee saves } | ([1])
// | { Return value spill } | (live on return slow paths)
// | { Local Ref. Table State } |
// | { Stack Indirect Ref. Table |
// | num. refs./link } | (here to prior SP is frame size)
// | { Method* } | <-- Anchor SP written to thread
// | { Outgoing stack args } | <-- SP at point of call
// | Native frame |
//
// [1] We must save all callee saves here to enable any exception throws to restore
// callee saves for frames above this one.
class JniCallingConvention : public CallingConvention {
public:
static std::unique_ptr<JniCallingConvention> Create(ArenaAllocator* arena,
bool is_static,
bool is_synchronized,
bool is_critical_native,
const char* shorty,
InstructionSet instruction_set);
// Size of frame excluding space for outgoing args (its assumed Method* is
// always at the bottom of a frame, but this doesn't work for outgoing
// native args). Includes alignment.
virtual size_t FrameSize() = 0;
// Size of outgoing arguments (stack portion), including alignment.
// -- Arguments that are passed via registers are excluded from this size.
virtual size_t OutArgSize() = 0;
// Number of references in stack indirect reference table
size_t ReferenceCount() const;
// Location where the segment state of the local indirect reference table is saved
FrameOffset SavedLocalReferenceCookieOffset() const;
// Location where the return value of a call can be squirreled if another
// call is made following the native call
FrameOffset ReturnValueSaveLocation() const;
// Register that holds result if it is integer.
virtual ManagedRegister IntReturnRegister() = 0;
// Whether the compiler needs to ensure zero-/sign-extension of a small result type
virtual bool RequiresSmallResultTypeExtension() const = 0;
// Callee save registers to spill prior to native code (which may clobber)
virtual ArrayRef<const ManagedRegister> CalleeSaveRegisters() const = 0;
// Spill mask values
virtual uint32_t CoreSpillMask() const = 0;
virtual uint32_t FpSpillMask() const = 0;
// An extra scratch register live after the call
virtual ManagedRegister ReturnScratchRegister() const = 0;
// Iterator interface
bool HasNext();
virtual void Next();
bool IsCurrentParamAReference();
bool IsCurrentParamAFloatOrDouble();
bool IsCurrentParamADouble();
bool IsCurrentParamALong();
bool IsCurrentParamALongOrDouble() {
return IsCurrentParamALong() || IsCurrentParamADouble();
}
bool IsCurrentParamJniEnv();
size_t CurrentParamSize() const;
virtual bool IsCurrentParamInRegister() = 0;
virtual bool IsCurrentParamOnStack() = 0;
virtual ManagedRegister CurrentParamRegister() = 0;
virtual FrameOffset CurrentParamStackOffset() = 0;
// Iterator interface extension for JNI
FrameOffset CurrentParamHandleScopeEntryOffset();
// Position of handle scope and interior fields
FrameOffset HandleScopeOffset() const {
return FrameOffset(this->displacement_.Int32Value() + static_cast<size_t>(frame_pointer_size_));
// above Method reference
}
FrameOffset HandleScopeLinkOffset() const {
return FrameOffset(HandleScopeOffset().Int32Value() +
HandleScope::LinkOffset(frame_pointer_size_));
}
FrameOffset HandleScopeNumRefsOffset() const {
return FrameOffset(HandleScopeOffset().Int32Value() +
HandleScope::NumberOfReferencesOffset(frame_pointer_size_));
}
FrameOffset HandleReferencesOffset() const {
return FrameOffset(HandleScopeOffset().Int32Value() +
HandleScope::ReferencesOffset(frame_pointer_size_));
}
virtual ~JniCallingConvention() {}
protected:
// Named iterator positions
enum IteratorPos {
kJniEnv = 0,
kObjectOrClass = 1
};
// TODO: remove this constructor once all are changed to the below one.
JniCallingConvention(bool is_static,
bool is_synchronized,
const char* shorty,
PointerSize frame_pointer_size)
: CallingConvention(is_static, is_synchronized, shorty, frame_pointer_size),
is_critical_native_(false) {}
JniCallingConvention(bool is_static,
bool is_synchronized,
bool is_critical_native,
const char* shorty,
PointerSize frame_pointer_size)
: CallingConvention(is_static, is_synchronized, shorty, frame_pointer_size),
is_critical_native_(is_critical_native) {}
// Number of stack slots for outgoing arguments, above which the handle scope is
// located
virtual size_t NumberOfOutgoingStackArgs() = 0;
protected:
size_t NumberOfExtraArgumentsForJni() const;
// Does the transition have a StackHandleScope?
bool HasHandleScope() const;
// Does the transition have a local reference segment state?
bool HasLocalReferenceSegmentState() const;
// Has a JNIEnv* parameter implicitly?
bool HasJniEnv() const;
// Has a 'jclass' parameter implicitly?
bool HasSelfClass() const;
// Are there extra JNI arguments (JNIEnv* and maybe jclass)?
bool HasExtraArgumentsForJni() const;
// Returns the position of itr_args_, fixed up by removing the offset of extra JNI arguments.
unsigned int GetIteratorPositionWithinShorty() const;
// Is the current argument (at the iterator) an extra argument for JNI?
bool IsCurrentArgExtraForJni() const;
const bool is_critical_native_;
private:
// Shorthand for switching on the switch value but only IF there are extra JNI arguments.
//
// Puts the case value into return_value.
// * (switch_value == kJniEnv) => case_jni_env
// * (switch_value == kObjectOrClass) => case_object_or_class
//
// Returns false otherwise (or if there are no extra JNI arguments).
bool SwitchExtraJniArguments(size_t switch_value,
bool case_jni_env,
bool case_object_or_class,
/* out parameters */
bool* return_value) const;
};
} // namespace art
#endif // ART_COMPILER_JNI_QUICK_CALLING_CONVENTION_H_