summaryrefslogtreecommitdiff
path: root/runtime/linear_alloc-inl.h
diff options
context:
space:
mode:
Diffstat (limited to 'runtime/linear_alloc-inl.h')
-rw-r--r--runtime/linear_alloc-inl.h128
1 files changed, 128 insertions, 0 deletions
diff --git a/runtime/linear_alloc-inl.h b/runtime/linear_alloc-inl.h
new file mode 100644
index 0000000000..a6b3df3ae6
--- /dev/null
+++ b/runtime/linear_alloc-inl.h
@@ -0,0 +1,128 @@
+/*
+ * Copyright (C) 2015 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#ifndef ART_RUNTIME_LINEAR_ALLOC_INL_H_
+#define ART_RUNTIME_LINEAR_ALLOC_INL_H_
+
+#include "linear_alloc.h"
+
+#include "base/gc_visited_arena_pool.h"
+#include "thread-current-inl.h"
+
+namespace art {
+
+inline void LinearAlloc::SetFirstObject(void* begin, size_t bytes) const {
+ DCHECK(track_allocations_);
+ uint8_t* end = static_cast<uint8_t*>(begin) + bytes;
+ Arena* arena = allocator_.GetHeadArena();
+ DCHECK_NE(arena, nullptr);
+ // The object would either be in the head arena or the next one.
+ if (UNLIKELY(begin < arena->Begin() || begin >= arena->End())) {
+ arena = arena->Next();
+ }
+ DCHECK(begin >= arena->Begin() && end <= arena->End());
+ down_cast<TrackedArena*>(arena)->SetFirstObject(static_cast<uint8_t*>(begin), end);
+}
+
+inline void* LinearAlloc::Realloc(Thread* self,
+ void* ptr,
+ size_t old_size,
+ size_t new_size,
+ LinearAllocKind kind) {
+ MutexLock mu(self, lock_);
+ if (track_allocations_) {
+ if (ptr != nullptr) {
+ // Realloc cannot be called on 16-byte aligned as Realloc doesn't guarantee
+ // that. So the header must be immediately prior to ptr.
+ TrackingHeader* header = reinterpret_cast<TrackingHeader*>(ptr) - 1;
+ DCHECK_EQ(header->GetKind(), kind);
+ old_size += sizeof(TrackingHeader);
+ DCHECK_EQ(header->GetSize(), old_size);
+ ptr = header;
+ } else {
+ DCHECK_EQ(old_size, 0u);
+ }
+ new_size += sizeof(TrackingHeader);
+ void* ret = allocator_.Realloc(ptr, old_size, new_size);
+ new (ret) TrackingHeader(new_size, kind);
+ SetFirstObject(ret, new_size);
+ return static_cast<TrackingHeader*>(ret) + 1;
+ } else {
+ return allocator_.Realloc(ptr, old_size, new_size);
+ }
+}
+
+inline void* LinearAlloc::Alloc(Thread* self, size_t size, LinearAllocKind kind) {
+ MutexLock mu(self, lock_);
+ if (track_allocations_) {
+ size += sizeof(TrackingHeader);
+ TrackingHeader* storage = new (allocator_.Alloc(size)) TrackingHeader(size, kind);
+ SetFirstObject(storage, size);
+ return storage + 1;
+ } else {
+ return allocator_.Alloc(size);
+ }
+}
+
+inline void* LinearAlloc::AllocAlign16(Thread* self, size_t size, LinearAllocKind kind) {
+ MutexLock mu(self, lock_);
+ DCHECK_ALIGNED(size, 16);
+ if (track_allocations_) {
+ size_t mem_tool_bytes = ArenaAllocator::IsRunningOnMemoryTool()
+ ? ArenaAllocator::kMemoryToolRedZoneBytes : 0;
+ uint8_t* ptr = allocator_.CurrentPtr() + sizeof(TrackingHeader);
+ uintptr_t padding =
+ RoundUp(reinterpret_cast<uintptr_t>(ptr), 16) - reinterpret_cast<uintptr_t>(ptr);
+ DCHECK_LT(padding, 16u);
+ size_t required_size = size + sizeof(TrackingHeader) + padding;
+
+ if (allocator_.CurrentArenaUnusedBytes() < required_size + mem_tool_bytes) {
+ // The allocator will require a new arena, which is expected to be
+ // 16-byte aligned.
+ static_assert(ArenaAllocator::kArenaAlignment >= 16,
+ "Expecting sufficient alignment for new Arena.");
+ required_size = size + RoundUp(sizeof(TrackingHeader), 16);
+ }
+ // Using ArenaAllocator's AllocAlign16 now would disturb the alignment by
+ // trying to make header 16-byte aligned. The alignment requirements are
+ // already addressed here. Now we want allocator to just bump the pointer.
+ ptr = static_cast<uint8_t*>(allocator_.Alloc(required_size));
+ new (ptr) TrackingHeader(required_size, kind, /*is_16_aligned=*/true);
+ SetFirstObject(ptr, required_size);
+ return AlignUp(ptr + sizeof(TrackingHeader), 16);
+ } else {
+ return allocator_.AllocAlign16(size);
+ }
+}
+
+inline size_t LinearAlloc::GetUsedMemory() const {
+ MutexLock mu(Thread::Current(), lock_);
+ return allocator_.BytesUsed();
+}
+
+inline ArenaPool* LinearAlloc::GetArenaPool() {
+ MutexLock mu(Thread::Current(), lock_);
+ return allocator_.GetArenaPool();
+}
+
+inline bool LinearAlloc::Contains(void* ptr) const {
+ MutexLock mu(Thread::Current(), lock_);
+ return allocator_.Contains(ptr);
+}
+
+} // namespace art
+
+#endif // ART_RUNTIME_LINEAR_ALLOC_INL_H_