/* * Copyright (C) 2018 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "mem_map_arena_pool.h" #include #include #include #include #include #include #include "base/arena_allocator-inl.h" #include "base/mem_map.h" #include "base/systrace.h" namespace art { class MemMapArena final : public Arena { public: MemMapArena(size_t size, bool low_4gb, const char* name); virtual ~MemMapArena(); void Release() override; private: static MemMap Allocate(size_t size, bool low_4gb, const char* name); MemMap map_; }; MemMapArena::MemMapArena(size_t size, bool low_4gb, const char* name) : map_(Allocate(size, low_4gb, name)) { memory_ = map_.Begin(); static_assert(ArenaAllocator::kArenaAlignment <= kPageSize, "Arena should not need stronger alignment than kPageSize."); DCHECK_ALIGNED(memory_, ArenaAllocator::kArenaAlignment); size_ = map_.Size(); } MemMap MemMapArena::Allocate(size_t size, bool low_4gb, const char* name) { // Round up to a full page as that's the smallest unit of allocation for mmap() // and we want to be able to use all memory that we actually allocate. size = RoundUp(size, kPageSize); std::string error_msg; MemMap map = MemMap::MapAnonymous(name, size, PROT_READ | PROT_WRITE, low_4gb, &error_msg); CHECK(map.IsValid()) << error_msg; return map; } MemMapArena::~MemMapArena() { // Destroys MemMap via std::unique_ptr<>. } void MemMapArena::Release() { if (bytes_allocated_ > 0) { map_.MadviseDontNeedAndZero(); bytes_allocated_ = 0; } } MemMapArenaPool::MemMapArenaPool(bool low_4gb, const char* name) : low_4gb_(low_4gb), name_(name), free_arenas_(nullptr) { MemMap::Init(); } MemMapArenaPool::~MemMapArenaPool() { ReclaimMemory(); } void MemMapArenaPool::ReclaimMemory() { while (free_arenas_ != nullptr) { Arena* arena = free_arenas_; free_arenas_ = free_arenas_->next_; delete arena; } } void MemMapArenaPool::LockReclaimMemory() { std::lock_guard lock(lock_); ReclaimMemory(); } Arena* MemMapArenaPool::AllocArena(size_t size) { Arena* ret = nullptr; { std::lock_guard lock(lock_); if (free_arenas_ != nullptr && LIKELY(free_arenas_->Size() >= size)) { ret = free_arenas_; free_arenas_ = free_arenas_->next_; } } if (ret == nullptr) { ret = new MemMapArena(size, low_4gb_, name_); } ret->Reset(); return ret; } void MemMapArenaPool::TrimMaps() { ScopedTrace trace(__PRETTY_FUNCTION__); std::lock_guard lock(lock_); for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { arena->Release(); } } size_t MemMapArenaPool::GetBytesAllocated() const { size_t total = 0; std::lock_guard lock(lock_); for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { total += arena->GetBytesAllocated(); } return total; } void MemMapArenaPool::FreeArenaChain(Arena* first) { if (kRunningOnMemoryTool) { for (Arena* arena = first; arena != nullptr; arena = arena->next_) { MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); } } if (arena_allocator::kArenaAllocatorPreciseTracking) { // Do not reuse arenas when tracking. while (first != nullptr) { Arena* next = first->next_; delete first; first = next; } return; } if (first != nullptr) { Arena* last = first; while (last->next_ != nullptr) { last = last->next_; } std::lock_guard lock(lock_); last->next_ = free_arenas_; free_arenas_ = first; } } } // namespace art