| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_GC_SPACE_SPACE_H_ |
| #define ART_RUNTIME_GC_SPACE_SPACE_H_ |
| |
| #include <memory> |
| #include <string> |
| |
| #include "base/atomic.h" |
| #include "base/locks.h" |
| #include "base/macros.h" |
| #include "base/mem_map.h" |
| #include "gc/accounting/space_bitmap.h" |
| #include "gc/collector/object_byte_pair.h" |
| #include "runtime_globals.h" |
| |
| namespace art { |
| namespace mirror { |
| class Object; |
| } // namespace mirror |
| |
| namespace gc { |
| |
| class Heap; |
| |
| namespace space { |
| |
| class AllocSpace; |
| class BumpPointerSpace; |
| class ContinuousMemMapAllocSpace; |
| class ContinuousSpace; |
| class DiscontinuousSpace; |
| class MallocSpace; |
| class DlMallocSpace; |
| class RosAllocSpace; |
| class ImageSpace; |
| class LargeObjectSpace; |
| class RegionSpace; |
| class ZygoteSpace; |
| |
| static constexpr bool kDebugSpaces = kIsDebugBuild; |
| |
| // See Space::GetGcRetentionPolicy. |
| enum GcRetentionPolicy { |
| // Objects are retained forever with this policy for a space. |
| kGcRetentionPolicyNeverCollect, |
| // Every GC cycle will attempt to collect objects in this space. |
| kGcRetentionPolicyAlwaysCollect, |
| // Objects will be considered for collection only in "full" GC cycles, ie faster partial |
| // collections won't scan these areas such as the Zygote. |
| kGcRetentionPolicyFullCollect, |
| }; |
| std::ostream& operator<<(std::ostream& os, GcRetentionPolicy policy); |
| |
| enum SpaceType { |
| kSpaceTypeImageSpace, |
| kSpaceTypeMallocSpace, |
| kSpaceTypeZygoteSpace, |
| kSpaceTypeBumpPointerSpace, |
| kSpaceTypeLargeObjectSpace, |
| kSpaceTypeRegionSpace, |
| }; |
| std::ostream& operator<<(std::ostream& os, SpaceType space_type); |
| |
| // A space contains memory allocated for managed objects. |
| class Space { |
| public: |
| // Dump space. Also key method for C++ vtables. |
| virtual void Dump(std::ostream& os) const; |
| |
| // Name of the space. May vary, for example before/after the Zygote fork. |
| const char* GetName() const { |
| return name_.c_str(); |
| } |
| |
| // The policy of when objects are collected associated with this space. |
| GcRetentionPolicy GetGcRetentionPolicy() const { |
| return gc_retention_policy_; |
| } |
| |
| // Is the given object contained within this space? |
| virtual bool Contains(const mirror::Object* obj) const = 0; |
| |
| // The kind of space this: image, alloc, zygote, large object. |
| virtual SpaceType GetType() const = 0; |
| |
| // Is this an image space, ie one backed by a memory mapped image file. |
| bool IsImageSpace() const { |
| return GetType() == kSpaceTypeImageSpace; |
| } |
| ImageSpace* AsImageSpace(); |
| |
| // Is this a dlmalloc backed allocation space? |
| bool IsMallocSpace() const { |
| SpaceType type = GetType(); |
| return type == kSpaceTypeMallocSpace; |
| } |
| MallocSpace* AsMallocSpace(); |
| |
| virtual bool IsDlMallocSpace() const { |
| return false; |
| } |
| virtual DlMallocSpace* AsDlMallocSpace(); |
| |
| virtual bool IsRosAllocSpace() const { |
| return false; |
| } |
| virtual RosAllocSpace* AsRosAllocSpace(); |
| |
| // Is this the space allocated into by the Zygote and no-longer in use for allocation? |
| bool IsZygoteSpace() const { |
| return GetType() == kSpaceTypeZygoteSpace; |
| } |
| virtual ZygoteSpace* AsZygoteSpace(); |
| |
| // Is this space a bump pointer space? |
| bool IsBumpPointerSpace() const { |
| return GetType() == kSpaceTypeBumpPointerSpace; |
| } |
| virtual BumpPointerSpace* AsBumpPointerSpace(); |
| |
| bool IsRegionSpace() const { |
| return GetType() == kSpaceTypeRegionSpace; |
| } |
| virtual RegionSpace* AsRegionSpace(); |
| |
| // Does this space hold large objects and implement the large object space abstraction? |
| bool IsLargeObjectSpace() const { |
| return GetType() == kSpaceTypeLargeObjectSpace; |
| } |
| LargeObjectSpace* AsLargeObjectSpace(); |
| |
| virtual bool IsContinuousSpace() const { |
| return false; |
| } |
| ContinuousSpace* AsContinuousSpace(); |
| |
| virtual bool IsDiscontinuousSpace() const { |
| return false; |
| } |
| DiscontinuousSpace* AsDiscontinuousSpace(); |
| |
| virtual bool IsAllocSpace() const { |
| return false; |
| } |
| virtual AllocSpace* AsAllocSpace(); |
| |
| virtual bool IsContinuousMemMapAllocSpace() const { |
| return false; |
| } |
| virtual ContinuousMemMapAllocSpace* AsContinuousMemMapAllocSpace(); |
| |
| // Returns true if objects in the space are movable. |
| virtual bool CanMoveObjects() const = 0; |
| |
| virtual ~Space() {} |
| |
| protected: |
| Space(const std::string& name, GcRetentionPolicy gc_retention_policy); |
| |
| void SetGcRetentionPolicy(GcRetentionPolicy gc_retention_policy) { |
| gc_retention_policy_ = gc_retention_policy; |
| } |
| |
| // Name of the space that may vary due to the Zygote fork. |
| std::string name_; |
| |
| protected: |
| // When should objects within this space be reclaimed? Not constant as we vary it in the case |
| // of Zygote forking. |
| GcRetentionPolicy gc_retention_policy_; |
| |
| private: |
| friend class art::gc::Heap; |
| DISALLOW_IMPLICIT_CONSTRUCTORS(Space); |
| }; |
| std::ostream& operator<<(std::ostream& os, const Space& space); |
| |
| // AllocSpace interface. |
| class AllocSpace { |
| public: |
| // Number of bytes currently allocated. |
| virtual uint64_t GetBytesAllocated() = 0; |
| // Number of objects currently allocated. |
| virtual uint64_t GetObjectsAllocated() = 0; |
| |
| // Allocate num_bytes without allowing growth. If the allocation |
| // succeeds, the output parameter bytes_allocated will be set to the |
| // actually allocated bytes which is >= num_bytes. |
| // Alloc can be called from multiple threads at the same time and must be thread-safe. |
| // |
| // bytes_tl_bulk_allocated - bytes allocated in bulk ahead of time for a thread local allocation, |
| // if applicable. It is |
| // 1) equal to bytes_allocated if it's not a thread local allocation, |
| // 2) greater than bytes_allocated if it's a thread local |
| // allocation that required a new buffer, or |
| // 3) zero if it's a thread local allocation in an existing |
| // buffer. |
| // This is what is to be added to Heap::num_bytes_allocated_. |
| virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, |
| size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0; |
| |
| // Thread-unsafe allocation for when mutators are suspended, used by the semispace collector. |
| virtual mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, |
| size_t* usable_size, |
| size_t* bytes_tl_bulk_allocated) |
| REQUIRES(Locks::mutator_lock_) { |
| return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); |
| } |
| |
| // Return the storage space required by obj. |
| virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0; |
| |
| // Returns how many bytes were freed. |
| virtual size_t Free(Thread* self, mirror::Object* ptr) = 0; |
| |
| // Free (deallocate) all objects in a list, and return the number of bytes freed. |
| virtual size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) = 0; |
| |
| // Revoke any sort of thread-local buffers that are used to speed up allocations for the given |
| // thread, if the alloc space implementation uses any. |
| // Returns the total free bytes in the revoked thread local runs that's to be subtracted |
| // from Heap::num_bytes_allocated_ or zero if unnecessary. |
| virtual size_t RevokeThreadLocalBuffers(Thread* thread) = 0; |
| |
| // Revoke any sort of thread-local buffers that are used to speed up allocations for all the |
| // threads, if the alloc space implementation uses any. |
| // Returns the total free bytes in the revoked thread local runs that's to be subtracted |
| // from Heap::num_bytes_allocated_ or zero if unnecessary. |
| virtual size_t RevokeAllThreadLocalBuffers() = 0; |
| |
| // Compute largest free contiguous chunk of memory available in the space and |
| // log it if it's smaller than failed_alloc_bytes and return true. |
| // Otherwise leave os untouched and return false. |
| virtual bool LogFragmentationAllocFailure(std::ostream& os, size_t failed_alloc_bytes) = 0; |
| |
| protected: |
| struct SweepCallbackContext { |
| SweepCallbackContext(bool swap_bitmaps, space::Space* space); |
| const bool swap_bitmaps; |
| space::Space* const space; |
| Thread* const self; |
| collector::ObjectBytePair freed; |
| }; |
| |
| AllocSpace() {} |
| virtual ~AllocSpace() {} |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(AllocSpace); |
| }; |
| |
| // Continuous spaces have bitmaps, and an address range. Although not required, objects within |
| // continuous spaces can be marked in the card table. |
| class ContinuousSpace : public Space { |
| public: |
| // Address at which the space begins. |
| uint8_t* Begin() const { |
| return begin_; |
| } |
| |
| // Current address at which the space ends, which may vary as the space is filled. |
| uint8_t* End() const { |
| return end_.load(std::memory_order_relaxed); |
| } |
| |
| // The end of the address range covered by the space. |
| uint8_t* Limit() const { |
| return limit_; |
| } |
| |
| // Change the end of the space. Be careful with use since changing the end of a space to an |
| // invalid value may break the GC. |
| void SetEnd(uint8_t* end) { |
| end_.store(end, std::memory_order_relaxed); |
| } |
| |
| void SetLimit(uint8_t* limit) { |
| limit_ = limit; |
| } |
| |
| // Current size of space |
| size_t Size() const { |
| return End() - Begin(); |
| } |
| |
| virtual accounting::ContinuousSpaceBitmap* GetLiveBitmap() = 0; |
| virtual accounting::ContinuousSpaceBitmap* GetMarkBitmap() = 0; |
| |
| // Maximum which the mapped space can grow to. |
| virtual size_t Capacity() const { |
| return Limit() - Begin(); |
| } |
| |
| // Is object within this space? We check to see if the pointer is beyond the end first as |
| // continuous spaces are iterated over from low to high. |
| bool HasAddress(const mirror::Object* obj) const { |
| const uint8_t* byte_ptr = reinterpret_cast<const uint8_t*>(obj); |
| return byte_ptr >= Begin() && byte_ptr < Limit(); |
| } |
| |
| bool Contains(const mirror::Object* obj) const { |
| return HasAddress(obj); |
| } |
| |
| virtual bool IsContinuousSpace() const { |
| return true; |
| } |
| |
| bool HasBoundBitmaps() REQUIRES(Locks::heap_bitmap_lock_); |
| |
| virtual ~ContinuousSpace() {} |
| |
| protected: |
| ContinuousSpace(const std::string& name, GcRetentionPolicy gc_retention_policy, |
| uint8_t* begin, uint8_t* end, uint8_t* limit) : |
| Space(name, gc_retention_policy), begin_(begin), end_(end), limit_(limit) { |
| } |
| |
| // The beginning of the storage for fast access. |
| uint8_t* begin_; |
| |
| // Current end of the space. |
| Atomic<uint8_t*> end_; |
| |
| // Limit of the space. |
| uint8_t* limit_; |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(ContinuousSpace); |
| }; |
| |
| // A space where objects may be allocated higgledy-piggledy throughout virtual memory. Currently |
| // the card table can't cover these objects and so the write barrier shouldn't be triggered. This |
| // is suitable for use for large primitive arrays. |
| class DiscontinuousSpace : public Space { |
| public: |
| accounting::LargeObjectBitmap* GetLiveBitmap() { |
| return &live_bitmap_; |
| } |
| |
| accounting::LargeObjectBitmap* GetMarkBitmap() { |
| return &mark_bitmap_; |
| } |
| |
| bool IsDiscontinuousSpace() const override { |
| return true; |
| } |
| |
| virtual ~DiscontinuousSpace() {} |
| |
| protected: |
| DiscontinuousSpace(const std::string& name, GcRetentionPolicy gc_retention_policy); |
| |
| accounting::LargeObjectBitmap live_bitmap_; |
| accounting::LargeObjectBitmap mark_bitmap_; |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(DiscontinuousSpace); |
| }; |
| |
| class MemMapSpace : public ContinuousSpace { |
| public: |
| // Size of the space without a limit on its growth. By default this is just the Capacity, but |
| // for the allocation space we support starting with a small heap and then extending it. |
| virtual size_t NonGrowthLimitCapacity() const { |
| return Capacity(); |
| } |
| |
| MemMap* GetMemMap() { |
| return &mem_map_; |
| } |
| |
| const MemMap* GetMemMap() const { |
| return &mem_map_; |
| } |
| |
| MemMap ReleaseMemMap() { |
| return std::move(mem_map_); |
| } |
| |
| protected: |
| MemMapSpace(const std::string& name, |
| MemMap&& mem_map, |
| uint8_t* begin, |
| uint8_t* end, |
| uint8_t* limit, |
| GcRetentionPolicy gc_retention_policy) |
| : ContinuousSpace(name, gc_retention_policy, begin, end, limit), |
| mem_map_(std::move(mem_map)) { |
| } |
| |
| // Underlying storage of the space |
| MemMap mem_map_; |
| |
| private: |
| DISALLOW_IMPLICIT_CONSTRUCTORS(MemMapSpace); |
| }; |
| |
| // Used by the heap compaction interface to enable copying from one type of alloc space to another. |
| class ContinuousMemMapAllocSpace : public MemMapSpace, public AllocSpace { |
| public: |
| bool IsAllocSpace() const override { |
| return true; |
| } |
| AllocSpace* AsAllocSpace() override { |
| return this; |
| } |
| |
| bool IsContinuousMemMapAllocSpace() const override { |
| return true; |
| } |
| ContinuousMemMapAllocSpace* AsContinuousMemMapAllocSpace() override { |
| return this; |
| } |
| |
| // Make the mark bitmap an alias of the live bitmap. Save the current mark bitmap into |
| // `temp_bitmap_`, so that we can restore it later in ContinuousMemMapAllocSpace::UnBindBitmaps. |
| void BindLiveToMarkBitmap() REQUIRES(Locks::heap_bitmap_lock_); |
| // Unalias the mark bitmap from the live bitmap and restore the old mark bitmap. |
| void UnBindBitmaps() REQUIRES(Locks::heap_bitmap_lock_); |
| // Swap the live and mark bitmaps of this space. This is used by the GC for concurrent sweeping. |
| void SwapBitmaps() REQUIRES(Locks::heap_bitmap_lock_); |
| |
| // Clear the space back to an empty space. |
| virtual void Clear() = 0; |
| |
| accounting::ContinuousSpaceBitmap* GetLiveBitmap() override { |
| return &live_bitmap_; |
| } |
| |
| accounting::ContinuousSpaceBitmap* GetMarkBitmap() override { |
| return &mark_bitmap_; |
| } |
| |
| accounting::ContinuousSpaceBitmap* GetTempBitmap() { |
| return &temp_bitmap_; |
| } |
| |
| collector::ObjectBytePair Sweep(bool swap_bitmaps); |
| virtual accounting::ContinuousSpaceBitmap::SweepCallback* GetSweepCallback() = 0; |
| |
| protected: |
| accounting::ContinuousSpaceBitmap live_bitmap_; |
| accounting::ContinuousSpaceBitmap mark_bitmap_; |
| accounting::ContinuousSpaceBitmap temp_bitmap_; |
| |
| ContinuousMemMapAllocSpace(const std::string& name, |
| MemMap&& mem_map, |
| uint8_t* begin, |
| uint8_t* end, |
| uint8_t* limit, |
| GcRetentionPolicy gc_retention_policy) |
| : MemMapSpace(name, std::move(mem_map), begin, end, limit, gc_retention_policy) { |
| } |
| |
| private: |
| friend class gc::Heap; |
| DISALLOW_IMPLICIT_CONSTRUCTORS(ContinuousMemMapAllocSpace); |
| }; |
| |
| } // namespace space |
| } // namespace gc |
| } // namespace art |
| |
| #endif // ART_RUNTIME_GC_SPACE_SPACE_H_ |