| /* |
| * Copyright (C) 2008 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_LIBARTBASE_BASE_MEM_MAP_H_ |
| #define ART_LIBARTBASE_BASE_MEM_MAP_H_ |
| |
| #include <stddef.h> |
| #include <sys/types.h> |
| |
| #include <map> |
| #include <mutex> |
| #include <string> |
| |
| #include "android-base/thread_annotations.h" |
| #include "bit_utils.h" |
| #include "globals.h" |
| #include "macros.h" |
| |
| namespace art { |
| |
| #if defined(__LP64__) && !defined(__Fuchsia__) && \ |
| (defined(__aarch64__) || defined(__riscv) || defined(__APPLE__)) |
| #define USE_ART_LOW_4G_ALLOCATOR 1 |
| #else |
| #if defined(__LP64__) && !defined(__Fuchsia__) && !defined(__x86_64__) |
| #error "Unrecognized 64-bit architecture." |
| #endif |
| #define USE_ART_LOW_4G_ALLOCATOR 0 |
| #endif |
| |
| #ifdef __linux__ |
| static constexpr bool kMadviseZeroes = true; |
| #define HAVE_MREMAP_SYSCALL true |
| #else |
| static constexpr bool kMadviseZeroes = false; |
| // We cannot ever perform MemMap::ReplaceWith on non-linux hosts since the syscall is not |
| // present. |
| #define HAVE_MREMAP_SYSCALL false |
| #endif |
| |
| // Used to keep track of mmap segments. |
| // |
| // On 64b systems not supporting MAP_32BIT, the implementation of MemMap will do a linear scan |
| // for free pages. For security, the start of this scan should be randomized. This requires a |
| // dynamic initializer. |
| // For this to work, it is paramount that there are no other static initializers that access MemMap. |
| // Otherwise, calls might see uninitialized values. |
| class MemMap { |
| public: |
| static constexpr bool kCanReplaceMapping = HAVE_MREMAP_SYSCALL; |
| |
| // Creates an invalid mapping. |
| MemMap() {} |
| |
| // Creates an invalid mapping. Used when we want to be more explicit than MemMap(). |
| static MemMap Invalid() { |
| return MemMap(); |
| } |
| |
| MemMap(MemMap&& other) noexcept REQUIRES(!MemMap::mem_maps_lock_); |
| MemMap& operator=(MemMap&& other) noexcept REQUIRES(!MemMap::mem_maps_lock_) { |
| Reset(); |
| swap(other); |
| return *this; |
| } |
| |
| // Releases the memory mapping. |
| ~MemMap() REQUIRES(!MemMap::mem_maps_lock_); |
| |
| // Swap two MemMaps. |
| void swap(MemMap& other); |
| |
| void Reset() { |
| if (IsValid()) { |
| DoReset(); |
| } |
| } |
| |
| bool IsValid() const { |
| return base_size_ != 0u; |
| } |
| |
| // Replace the data in this memmmap with the data in the memmap pointed to by source. The caller |
| // relinquishes ownership of the source mmap. |
| // |
| // For the call to be successful: |
| // * The range [dest->Begin, dest->Begin() + source->Size()] must not overlap with |
| // [source->Begin(), source->End()]. |
| // * Neither source nor dest may be 'reused' mappings (they must own all the pages associated |
| // with them. |
| // * kCanReplaceMapping must be true. |
| // * Neither source nor dest may use manual redzones. |
| // * Both source and dest must have the same offset from the nearest page boundary. |
| // * mremap must succeed when called on the mappings. |
| // |
| // If this call succeeds it will return true and: |
| // * Invalidate *source |
| // * The protection of this will remain the same. |
| // * The size of this will be the size of the source |
| // * The data in this will be the data from source. |
| // |
| // If this call fails it will return false and make no changes to *source or this. The ownership |
| // of the source mmap is returned to the caller. |
| bool ReplaceWith(/*in-out*/MemMap* source, /*out*/std::string* error); |
| |
| // Set a debug friendly name for a map. It will be prefixed with "dalvik-". |
| static void SetDebugName(void* map_ptr, const char* name, size_t size); |
| |
| // Request an anonymous region of length 'byte_count' and a requested base address. |
| // Use null as the requested base address if you don't care. |
| // |
| // `reuse` allows re-mapping an address range from an existing mapping which retains the |
| // ownership of the memory. Alternatively, `reservation` allows re-mapping the start of an |
| // existing reservation mapping, transferring the ownership of the memory to the new MemMap. |
| // |
| // The word "anonymous" in this context means "not backed by a file". The supplied |
| // 'name' will be used -- on systems that support it -- to give the mapping |
| // a name. |
| // |
| // On success, returns a valid MemMap. On failure, returns an invalid MemMap. |
| static MemMap MapAnonymous(const char* name, |
| uint8_t* addr, |
| size_t byte_count, |
| int prot, |
| bool low_4gb, |
| bool reuse, |
| /*inout*/MemMap* reservation, |
| /*out*/std::string* error_msg, |
| bool use_debug_name = true); |
| |
| // Request an aligned anonymous region, where the alignment must be higher |
| // than the runtime gPageSize. We can't directly ask for a MAP_SHARED |
| // (anonymous or otherwise) mapping to be aligned as in that case file offset |
| // is involved and could make the starting offset to be out of sync with |
| // another mapping of the same file. |
| static MemMap MapAnonymousAligned(const char* name, |
| size_t byte_count, |
| int prot, |
| bool low_4gb, |
| size_t alignment, |
| /*out=*/std::string* error_msg); |
| |
| static MemMap MapAnonymous(const char* name, |
| size_t byte_count, |
| int prot, |
| bool low_4gb, |
| /*out*/std::string* error_msg) { |
| return MapAnonymous(name, |
| /*addr=*/ nullptr, |
| byte_count, |
| prot, |
| low_4gb, |
| /*reuse=*/ false, |
| /*reservation=*/ nullptr, |
| error_msg); |
| } |
| static MemMap MapAnonymous(const char* name, |
| size_t byte_count, |
| int prot, |
| bool low_4gb, |
| MemMap* reservation, |
| /*out*/std::string* error_msg) { |
| return MapAnonymous(name, |
| /*addr=*/ (reservation != nullptr) ? reservation->Begin() : nullptr, |
| byte_count, |
| prot, |
| low_4gb, |
| /*reuse=*/ false, |
| reservation, |
| error_msg); |
| } |
| |
| // Create placeholder for a region allocated by direct call to mmap. |
| // This is useful when we do not have control over the code calling mmap, |
| // but when we still want to keep track of it in the list. |
| // The region is not considered to be owned and will not be unmmaped. |
| static MemMap MapPlaceholder(const char* name, uint8_t* addr, size_t byte_count); |
| |
| // Map part of a file, taking care of non-page aligned offsets. The |
| // "start" offset is absolute, not relative. |
| // |
| // On success, returns a valid MemMap. On failure, returns an invalid MemMap. |
| static MemMap MapFile(size_t byte_count, |
| int prot, |
| int flags, |
| int fd, |
| off_t start, |
| bool low_4gb, |
| const char* filename, |
| std::string* error_msg) { |
| return MapFileAtAddress(nullptr, |
| byte_count, |
| prot, |
| flags, |
| fd, |
| start, |
| /*low_4gb=*/ low_4gb, |
| filename, |
| /*reuse=*/ false, |
| /*reservation=*/ nullptr, |
| error_msg); |
| } |
| |
| // Map part of a file, taking care of non-page aligned offsets. The "start" offset is absolute, |
| // not relative. This version allows requesting a specific address for the base of the mapping. |
| // |
| // `reuse` allows re-mapping an address range from an existing mapping which retains the |
| // ownership of the memory. Alternatively, `reservation` allows re-mapping the start of an |
| // existing reservation mapping, transferring the ownership of the memory to the new MemMap. |
| // |
| // If error_msg is null then we do not print /proc/maps to the log if MapFileAtAddress fails. |
| // This helps improve performance of the fail case since reading and printing /proc/maps takes |
| // several milliseconds in the worst case. |
| // |
| // On success, returns a valid MemMap. On failure, returns an invalid MemMap. |
| static MemMap MapFileAtAddress(uint8_t* addr, |
| size_t byte_count, |
| int prot, |
| int flags, |
| int fd, |
| off_t start, |
| bool low_4gb, |
| const char* filename, |
| bool reuse, |
| /*inout*/MemMap* reservation, |
| /*out*/std::string* error_msg); |
| |
| const std::string& GetName() const { |
| return name_; |
| } |
| |
| bool Sync(); |
| |
| bool Protect(int prot); |
| |
| void FillWithZero(bool release_eagerly); |
| void MadviseDontNeedAndZero() { |
| FillWithZero(/* release_eagerly= */ true); |
| } |
| int MadviseDontFork(); |
| |
| int GetProtect() const { |
| return prot_; |
| } |
| |
| uint8_t* Begin() const { |
| return begin_; |
| } |
| |
| size_t Size() const { |
| return size_; |
| } |
| |
| // Resize the mem-map by unmapping pages at the end. Currently only supports shrinking. |
| void SetSize(size_t new_size); |
| |
| uint8_t* End() const { |
| return Begin() + Size(); |
| } |
| |
| void* BaseBegin() const { |
| return base_begin_; |
| } |
| |
| size_t BaseSize() const { |
| return base_size_; |
| } |
| |
| void* BaseEnd() const { |
| return reinterpret_cast<uint8_t*>(BaseBegin()) + BaseSize(); |
| } |
| |
| bool HasAddress(const void* addr) const { |
| return Begin() <= addr && addr < End(); |
| } |
| |
| // Unmap the pages at end and remap them to create another memory map. |
| MemMap RemapAtEnd(uint8_t* new_end, |
| const char* tail_name, |
| int tail_prot, |
| std::string* error_msg, |
| bool use_debug_name = true); |
| |
| // Unmap the pages of a file at end and remap them to create another memory map. |
| MemMap RemapAtEnd(uint8_t* new_end, |
| const char* tail_name, |
| int tail_prot, |
| int tail_flags, |
| int fd, |
| off_t offset, |
| std::string* error_msg, |
| bool use_debug_name = true); |
| |
| // Take ownership of pages at the beginning of the mapping. The mapping must be an |
| // anonymous reservation mapping, owning entire pages. The `byte_count` must not |
| // exceed the size of this reservation. |
| // |
| // Returns a mapping owning `byte_count` bytes rounded up to entire pages |
| // with size set to the passed `byte_count`. If 'reuse' is true then the caller |
| // is responsible for unmapping the taken pages. |
| MemMap TakeReservedMemory(size_t byte_count, bool reuse = false); |
| |
| static bool CheckNoGaps(MemMap& begin_map, MemMap& end_map) |
| REQUIRES(!MemMap::mem_maps_lock_); |
| static void DumpMaps(std::ostream& os, bool terse = false) |
| REQUIRES(!MemMap::mem_maps_lock_); |
| |
| // Init and Shutdown are NOT thread safe. |
| // Both may be called multiple times and MemMap objects may be created any |
| // time after the first call to Init and before the first call to Shutodwn. |
| static void Init() REQUIRES(!MemMap::mem_maps_lock_); |
| static void Shutdown() REQUIRES(!MemMap::mem_maps_lock_); |
| static bool IsInitialized(); |
| |
| // If the map is PROT_READ, try to read each page of the map to check it is in fact readable (not |
| // faulting). This is used to diagnose a bug b/19894268 where mprotect doesn't seem to be working |
| // intermittently. |
| void TryReadable(); |
| |
| // Align the map by unmapping the unaligned part at the lower end and if 'align_both_ends' is |
| // true, then the higher end as well. |
| void AlignBy(size_t alignment, bool align_both_ends = true); |
| |
| // For annotation reasons. |
| static std::mutex* GetMemMapsLock() RETURN_CAPABILITY(mem_maps_lock_) { |
| return nullptr; |
| } |
| |
| // Reset in a forked process the MemMap whose memory has been madvised MADV_DONTFORK |
| // in the parent process. |
| void ResetInForkedProcess(); |
| |
| // 'redzone_size_ == 0' indicates that we are not using memory-tool on this mapping. |
| size_t GetRedzoneSize() const { return redzone_size_; } |
| |
| #ifdef ART_PAGE_SIZE_AGNOSTIC |
| static inline size_t GetPageSize() { |
| DCHECK_NE(page_size_, 0u); |
| return page_size_; |
| } |
| #else |
| static constexpr size_t GetPageSize() { |
| return GetPageSizeSlow(); |
| } |
| #endif |
| |
| private: |
| MemMap(const std::string& name, |
| uint8_t* begin, |
| size_t size, |
| void* base_begin, |
| size_t base_size, |
| int prot, |
| bool reuse, |
| size_t redzone_size = 0) REQUIRES(!MemMap::mem_maps_lock_); |
| |
| void DoReset(); |
| void Invalidate(); |
| void SwapMembers(MemMap& other); |
| |
| static void DumpMapsLocked(std::ostream& os, bool terse) |
| REQUIRES(MemMap::mem_maps_lock_); |
| static bool HasMemMap(MemMap& map) |
| REQUIRES(MemMap::mem_maps_lock_); |
| static MemMap* GetLargestMemMapAt(void* address) |
| REQUIRES(MemMap::mem_maps_lock_); |
| static bool ContainedWithinExistingMap(uint8_t* ptr, size_t size, std::string* error_msg) |
| REQUIRES(!MemMap::mem_maps_lock_); |
| |
| // Internal version of mmap that supports low 4gb emulation. |
| static void* MapInternal(void* addr, |
| size_t length, |
| int prot, |
| int flags, |
| int fd, |
| off_t offset, |
| bool low_4gb) |
| REQUIRES(!MemMap::mem_maps_lock_); |
| static void* MapInternalArtLow4GBAllocator(size_t length, |
| int prot, |
| int flags, |
| int fd, |
| off_t offset) |
| REQUIRES(!MemMap::mem_maps_lock_); |
| |
| // Release memory owned by a reservation mapping. |
| void ReleaseReservedMemory(size_t byte_count); |
| |
| // member function to access real_munmap |
| static bool CheckMapRequest(uint8_t* expected_ptr, |
| void* actual_ptr, |
| size_t byte_count, |
| std::string* error_msg); |
| |
| static bool CheckReservation(uint8_t* expected_ptr, |
| size_t byte_count, |
| const char* name, |
| const MemMap& reservation, |
| /*out*/std::string* error_msg); |
| |
| std::string name_; |
| uint8_t* begin_ = nullptr; // Start of data. May be changed by AlignBy. |
| size_t size_ = 0u; // Length of data. |
| |
| void* base_begin_ = nullptr; // Page-aligned base address. May be changed by AlignBy. |
| size_t base_size_ = 0u; // Length of mapping. May be changed by RemapAtEnd (ie Zygote). |
| int prot_ = 0; // Protection of the map. |
| |
| // When reuse_ is true, this is just a view of an existing mapping |
| // and we do not take ownership and are not responsible for |
| // unmapping. |
| bool reuse_ = false; |
| |
| // When already_unmapped_ is true the destructor will not call munmap. |
| bool already_unmapped_ = false; |
| |
| size_t redzone_size_ = 0u; |
| |
| #if USE_ART_LOW_4G_ALLOCATOR |
| static uintptr_t next_mem_pos_; // Next memory location to check for low_4g extent. |
| |
| static void* TryMemMapLow4GB(void* ptr, |
| size_t page_aligned_byte_count, |
| int prot, |
| int flags, |
| int fd, |
| off_t offset); |
| #endif |
| |
| static void TargetMMapInit(); |
| static void* TargetMMap(void* start, size_t len, int prot, int flags, int fd, off_t fd_off); |
| static int TargetMUnmap(void* start, size_t len); |
| |
| static std::mutex* mem_maps_lock_; |
| |
| #ifdef ART_PAGE_SIZE_AGNOSTIC |
| static size_t page_size_; |
| #endif |
| |
| friend class MemMapTest; // To allow access to base_begin_ and base_size_. |
| }; |
| |
| inline void swap(MemMap& lhs, MemMap& rhs) { |
| lhs.swap(rhs); |
| } |
| |
| std::ostream& operator<<(std::ostream& os, const MemMap& mem_map); |
| |
| // Zero and maybe release memory if possible, no requirements on alignments. |
| void ZeroMemory(void* address, size_t length, bool release_eagerly); |
| inline void ZeroAndReleaseMemory(void* address, size_t length) { |
| ZeroMemory(address, length, /* release_eagerly= */ true); |
| } |
| |
| } // namespace art |
| |
| #endif // ART_LIBARTBASE_BASE_MEM_MAP_H_ |