| /* |
| * Copyright (C) 2013 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_ |
| #define ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_ |
| |
| #include <stddef.h> |
| #include <stdint.h> |
| |
| #include "bit_utils.h" |
| #include "debug_stack.h" |
| #include "dchecked_vector.h" |
| #include "macros.h" |
| #include "memory_tool.h" |
| |
| namespace art { |
| |
| class Arena; |
| class ArenaPool; |
| class ArenaAllocator; |
| class ArenaStack; |
| class ScopedArenaAllocator; |
| class MemStats; |
| |
| template <typename T> |
| class ArenaAllocatorAdapter; |
| |
| static constexpr bool kArenaAllocatorCountAllocations = false; |
| |
| // Type of allocation for memory tuning. |
| enum ArenaAllocKind { |
| kArenaAllocMisc, |
| kArenaAllocSwitchTable, |
| kArenaAllocSlowPaths, |
| kArenaAllocGrowableBitMap, |
| kArenaAllocSTL, |
| kArenaAllocGraphBuilder, |
| kArenaAllocGraph, |
| kArenaAllocBasicBlock, |
| kArenaAllocBlockList, |
| kArenaAllocReversePostOrder, |
| kArenaAllocLinearOrder, |
| kArenaAllocConstantsMap, |
| kArenaAllocPredecessors, |
| kArenaAllocSuccessors, |
| kArenaAllocDominated, |
| kArenaAllocInstruction, |
| kArenaAllocConstructorFenceInputs, |
| kArenaAllocInvokeInputs, |
| kArenaAllocPhiInputs, |
| kArenaAllocTypeCheckInputs, |
| kArenaAllocLoopInfo, |
| kArenaAllocLoopInfoBackEdges, |
| kArenaAllocTryCatchInfo, |
| kArenaAllocUseListNode, |
| kArenaAllocEnvironment, |
| kArenaAllocEnvironmentVRegs, |
| kArenaAllocEnvironmentLocations, |
| kArenaAllocLocationSummary, |
| kArenaAllocSsaBuilder, |
| kArenaAllocMoveOperands, |
| kArenaAllocCodeBuffer, |
| kArenaAllocStackMaps, |
| kArenaAllocOptimization, |
| kArenaAllocGvn, |
| kArenaAllocInductionVarAnalysis, |
| kArenaAllocBoundsCheckElimination, |
| kArenaAllocDCE, |
| kArenaAllocLSA, |
| kArenaAllocLSE, |
| kArenaAllocCFRE, |
| kArenaAllocLICM, |
| kArenaAllocLoopOptimization, |
| kArenaAllocSsaLiveness, |
| kArenaAllocSsaPhiElimination, |
| kArenaAllocReferenceTypePropagation, |
| kArenaAllocSelectGenerator, |
| kArenaAllocSideEffectsAnalysis, |
| kArenaAllocRegisterAllocator, |
| kArenaAllocRegisterAllocatorValidate, |
| kArenaAllocStackMapStream, |
| kArenaAllocVectorNode, |
| kArenaAllocCodeGenerator, |
| kArenaAllocAssembler, |
| kArenaAllocParallelMoveResolver, |
| kArenaAllocGraphChecker, |
| kArenaAllocVerifier, |
| kArenaAllocCallingConvention, |
| kArenaAllocCHA, |
| kArenaAllocScheduler, |
| kArenaAllocProfile, |
| kArenaAllocSuperblockCloner, |
| kNumArenaAllocKinds |
| }; |
| |
| template <bool kCount> |
| class ArenaAllocatorStatsImpl; |
| |
| template <> |
| class ArenaAllocatorStatsImpl<false> { |
| public: |
| ArenaAllocatorStatsImpl() = default; |
| ArenaAllocatorStatsImpl(const ArenaAllocatorStatsImpl& other) = default; |
| ArenaAllocatorStatsImpl& operator = (const ArenaAllocatorStatsImpl& other) = delete; |
| |
| void Copy(const ArenaAllocatorStatsImpl& other ATTRIBUTE_UNUSED) {} |
| void RecordAlloc(size_t bytes ATTRIBUTE_UNUSED, ArenaAllocKind kind ATTRIBUTE_UNUSED) {} |
| size_t NumAllocations() const { return 0u; } |
| size_t BytesAllocated() const { return 0u; } |
| void Dump(std::ostream& os ATTRIBUTE_UNUSED, |
| const Arena* first ATTRIBUTE_UNUSED, |
| ssize_t lost_bytes_adjustment ATTRIBUTE_UNUSED) const {} |
| }; |
| |
| template <bool kCount> |
| class ArenaAllocatorStatsImpl { |
| public: |
| ArenaAllocatorStatsImpl(); |
| ArenaAllocatorStatsImpl(const ArenaAllocatorStatsImpl& other) = default; |
| ArenaAllocatorStatsImpl& operator = (const ArenaAllocatorStatsImpl& other) = delete; |
| |
| void Copy(const ArenaAllocatorStatsImpl& other); |
| void RecordAlloc(size_t bytes, ArenaAllocKind kind); |
| size_t NumAllocations() const; |
| size_t BytesAllocated() const; |
| void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const; |
| |
| private: |
| size_t num_allocations_; |
| dchecked_vector<size_t> alloc_stats_; // Bytes used by various allocation kinds. |
| |
| static const char* const kAllocNames[]; |
| }; |
| |
| typedef ArenaAllocatorStatsImpl<kArenaAllocatorCountAllocations> ArenaAllocatorStats; |
| |
| template <bool kAvailable, bool kValgrind> |
| class ArenaAllocatorMemoryToolCheckImpl { |
| // This is the generic template but since there is a partial specialization |
| // for kValgrind == false, this can be instantiated only for kValgrind == true. |
| static_assert(kValgrind, "This template can be instantiated only for Valgrind."); |
| static_assert(kAvailable, "Valgrind implies memory tool availability."); |
| |
| public: |
| ArenaAllocatorMemoryToolCheckImpl() : is_running_on_valgrind_(RUNNING_ON_MEMORY_TOOL) { } |
| bool IsRunningOnMemoryTool() { return is_running_on_valgrind_; } |
| |
| private: |
| const bool is_running_on_valgrind_; |
| }; |
| |
| template <bool kAvailable> |
| class ArenaAllocatorMemoryToolCheckImpl<kAvailable, false> { |
| public: |
| ArenaAllocatorMemoryToolCheckImpl() { } |
| bool IsRunningOnMemoryTool() { return kAvailable; } |
| }; |
| |
| typedef ArenaAllocatorMemoryToolCheckImpl<kMemoryToolIsAvailable, kMemoryToolIsValgrind> |
| ArenaAllocatorMemoryToolCheck; |
| |
| class ArenaAllocatorMemoryTool : private ArenaAllocatorMemoryToolCheck { |
| public: |
| using ArenaAllocatorMemoryToolCheck::IsRunningOnMemoryTool; |
| |
| void MakeDefined(void* ptr, size_t size) { |
| if (UNLIKELY(IsRunningOnMemoryTool())) { |
| DoMakeDefined(ptr, size); |
| } |
| } |
| void MakeUndefined(void* ptr, size_t size) { |
| if (UNLIKELY(IsRunningOnMemoryTool())) { |
| DoMakeUndefined(ptr, size); |
| } |
| } |
| void MakeInaccessible(void* ptr, size_t size) { |
| if (UNLIKELY(IsRunningOnMemoryTool())) { |
| DoMakeInaccessible(ptr, size); |
| } |
| } |
| |
| private: |
| void DoMakeDefined(void* ptr, size_t size); |
| void DoMakeUndefined(void* ptr, size_t size); |
| void DoMakeInaccessible(void* ptr, size_t size); |
| }; |
| |
| class Arena { |
| public: |
| Arena(); |
| virtual ~Arena() { } |
| // Reset is for pre-use and uses memset for performance. |
| void Reset(); |
| // Release is used inbetween uses and uses madvise for memory usage. |
| virtual void Release() { } |
| uint8_t* Begin() { |
| return memory_; |
| } |
| |
| uint8_t* End() { |
| return memory_ + size_; |
| } |
| |
| size_t Size() const { |
| return size_; |
| } |
| |
| size_t RemainingSpace() const { |
| return Size() - bytes_allocated_; |
| } |
| |
| size_t GetBytesAllocated() const { |
| return bytes_allocated_; |
| } |
| |
| // Return true if ptr is contained in the arena. |
| bool Contains(const void* ptr) const { |
| return memory_ <= ptr && ptr < memory_ + bytes_allocated_; |
| } |
| |
| protected: |
| size_t bytes_allocated_; |
| uint8_t* memory_; |
| size_t size_; |
| Arena* next_; |
| friend class MallocArenaPool; |
| friend class MemMapArenaPool; |
| friend class ArenaAllocator; |
| friend class ArenaStack; |
| friend class ScopedArenaAllocator; |
| template <bool kCount> friend class ArenaAllocatorStatsImpl; |
| |
| friend class ArenaAllocatorTest; |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(Arena); |
| }; |
| |
| class ArenaPool { |
| public: |
| virtual ~ArenaPool() = default; |
| |
| virtual Arena* AllocArena(size_t size) = 0; |
| virtual void FreeArenaChain(Arena* first) = 0; |
| virtual size_t GetBytesAllocated() const = 0; |
| virtual void ReclaimMemory() = 0; |
| virtual void LockReclaimMemory() = 0; |
| // Trim the maps in arenas by madvising, used by JIT to reduce memory usage. |
| virtual void TrimMaps() = 0; |
| |
| protected: |
| ArenaPool() = default; |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(ArenaPool); |
| }; |
| |
| // Fast single-threaded allocator for zero-initialized memory chunks. |
| // |
| // Memory is allocated from ArenaPool in large chunks and then rationed through |
| // the ArenaAllocator. It's returned to the ArenaPool only when the ArenaAllocator |
| // is destroyed. |
| class ArenaAllocator |
| : private DebugStackRefCounter, private ArenaAllocatorStats, private ArenaAllocatorMemoryTool { |
| public: |
| explicit ArenaAllocator(ArenaPool* pool); |
| ~ArenaAllocator(); |
| |
| using ArenaAllocatorMemoryTool::IsRunningOnMemoryTool; |
| using ArenaAllocatorMemoryTool::MakeDefined; |
| using ArenaAllocatorMemoryTool::MakeUndefined; |
| using ArenaAllocatorMemoryTool::MakeInaccessible; |
| |
| // Get adapter for use in STL containers. See arena_containers.h . |
| ArenaAllocatorAdapter<void> Adapter(ArenaAllocKind kind = kArenaAllocSTL); |
| |
| // Returns zeroed memory. |
| void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { |
| if (UNLIKELY(IsRunningOnMemoryTool())) { |
| return AllocWithMemoryTool(bytes, kind); |
| } |
| bytes = RoundUp(bytes, kAlignment); |
| ArenaAllocatorStats::RecordAlloc(bytes, kind); |
| if (UNLIKELY(bytes > static_cast<size_t>(end_ - ptr_))) { |
| return AllocFromNewArena(bytes); |
| } |
| uint8_t* ret = ptr_; |
| DCHECK_ALIGNED(ret, kAlignment); |
| ptr_ += bytes; |
| return ret; |
| } |
| |
| // Returns zeroed memory. |
| void* AllocAlign16(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { |
| // It is an error to request 16-byte aligned allocation of unaligned size. |
| DCHECK_ALIGNED(bytes, 16); |
| if (UNLIKELY(IsRunningOnMemoryTool())) { |
| return AllocWithMemoryToolAlign16(bytes, kind); |
| } |
| uintptr_t padding = |
| ((reinterpret_cast<uintptr_t>(ptr_) + 15u) & 15u) - reinterpret_cast<uintptr_t>(ptr_); |
| ArenaAllocatorStats::RecordAlloc(bytes, kind); |
| if (UNLIKELY(padding + bytes > static_cast<size_t>(end_ - ptr_))) { |
| static_assert(kArenaAlignment >= 16, "Expecting sufficient alignment for new Arena."); |
| return AllocFromNewArena(bytes); |
| } |
| ptr_ += padding; |
| uint8_t* ret = ptr_; |
| DCHECK_ALIGNED(ret, 16); |
| ptr_ += bytes; |
| return ret; |
| } |
| |
| // Realloc never frees the input pointer, it is the caller's job to do this if necessary. |
| void* Realloc(void* ptr, |
| size_t ptr_size, |
| size_t new_size, |
| ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { |
| DCHECK_GE(new_size, ptr_size); |
| DCHECK_EQ(ptr == nullptr, ptr_size == 0u); |
| // We always allocate aligned. |
| const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment); |
| auto* end = reinterpret_cast<uint8_t*>(ptr) + aligned_ptr_size; |
| // If we haven't allocated anything else, we can safely extend. |
| if (end == ptr_) { |
| // Red zone prevents end == ptr_ (unless input = allocator state = null). |
| DCHECK(!IsRunningOnMemoryTool() || ptr_ == nullptr); |
| const size_t aligned_new_size = RoundUp(new_size, kAlignment); |
| const size_t size_delta = aligned_new_size - aligned_ptr_size; |
| // Check remain space. |
| const size_t remain = end_ - ptr_; |
| if (remain >= size_delta) { |
| ptr_ += size_delta; |
| ArenaAllocatorStats::RecordAlloc(size_delta, kind); |
| DCHECK_ALIGNED(ptr_, kAlignment); |
| return ptr; |
| } |
| } |
| auto* new_ptr = Alloc(new_size, kind); // Note: Alloc will take care of aligning new_size. |
| memcpy(new_ptr, ptr, ptr_size); |
| // TODO: Call free on ptr if linear alloc supports free. |
| return new_ptr; |
| } |
| |
| template <typename T> |
| T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { |
| return AllocArray<T>(1, kind); |
| } |
| |
| template <typename T> |
| T* AllocArray(size_t length, ArenaAllocKind kind = kArenaAllocMisc) { |
| return static_cast<T*>(Alloc(length * sizeof(T), kind)); |
| } |
| |
| size_t BytesAllocated() const; |
| |
| MemStats GetMemStats() const; |
| |
| // The BytesUsed method sums up bytes allocated from arenas in arena_head_ and nodes. |
| // TODO: Change BytesAllocated to this behavior? |
| size_t BytesUsed() const; |
| |
| ArenaPool* GetArenaPool() const { |
| return pool_; |
| } |
| |
| bool Contains(const void* ptr) const; |
| |
| // The alignment guaranteed for individual allocations. |
| static constexpr size_t kAlignment = 8u; |
| |
| // The alignment required for the whole Arena rather than individual allocations. |
| static constexpr size_t kArenaAlignment = 16u; |
| |
| private: |
| void* AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind); |
| void* AllocWithMemoryToolAlign16(size_t bytes, ArenaAllocKind kind); |
| uint8_t* AllocFromNewArena(size_t bytes); |
| uint8_t* AllocFromNewArenaWithMemoryTool(size_t bytes); |
| |
| void UpdateBytesAllocated(); |
| |
| ArenaPool* pool_; |
| uint8_t* begin_; |
| uint8_t* end_; |
| uint8_t* ptr_; |
| Arena* arena_head_; |
| |
| template <typename U> |
| friend class ArenaAllocatorAdapter; |
| |
| friend class ArenaAllocatorTest; |
| |
| DISALLOW_COPY_AND_ASSIGN(ArenaAllocator); |
| }; // ArenaAllocator |
| |
| class MemStats { |
| public: |
| MemStats(const char* name, |
| const ArenaAllocatorStats* stats, |
| const Arena* first_arena, |
| ssize_t lost_bytes_adjustment = 0); |
| void Dump(std::ostream& os) const; |
| |
| private: |
| const char* const name_; |
| const ArenaAllocatorStats* const stats_; |
| const Arena* const first_arena_; |
| const ssize_t lost_bytes_adjustment_; |
| }; // MemStats |
| |
| } // namespace art |
| |
| #endif // ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_ |