| /* |
| * Copyright (C) 2013 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_GC_SPACE_MALLOC_SPACE_H_ |
| #define ART_RUNTIME_GC_SPACE_MALLOC_SPACE_H_ |
| |
| #include "space.h" |
| |
| #include <iostream> |
| #include <valgrind.h> |
| #include <memcheck/memcheck.h> |
| |
| namespace art { |
| namespace gc { |
| |
| namespace collector { |
| class MarkSweep; |
| } // namespace collector |
| |
| namespace space { |
| |
| class ZygoteSpace; |
| |
| // TODO: Remove define macro |
| #define CHECK_MEMORY_CALL(call, args, what) \ |
| do { \ |
| int rc = call args; \ |
| if (UNLIKELY(rc != 0)) { \ |
| errno = rc; \ |
| PLOG(FATAL) << # call << " failed for " << what; \ |
| } \ |
| } while (false) |
| |
| // A common parent of DlMallocSpace and RosAllocSpace. |
| class MallocSpace : public ContinuousMemMapAllocSpace { |
| public: |
| typedef void(*WalkCallback)(void *start, void *end, size_t num_bytes, void* callback_arg); |
| |
| SpaceType GetType() const { |
| return kSpaceTypeMallocSpace; |
| } |
| |
| // Allocate num_bytes allowing the underlying space to grow. |
| virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, |
| size_t* bytes_allocated, size_t* usable_size) = 0; |
| // Allocate num_bytes without allowing the underlying space to grow. |
| virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, |
| size_t* usable_size) = 0; |
| // Return the storage space required by obj. If usable_size isn't nullptr then it is set to the |
| // amount of the storage space that may be used by obj. |
| virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0; |
| virtual size_t Free(Thread* self, mirror::Object* ptr) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) = 0; |
| virtual size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) = 0; |
| |
| #ifndef NDEBUG |
| virtual void CheckMoreCoreForPrecondition() {} // to be overridden in the debug build. |
| #else |
| void CheckMoreCoreForPrecondition() {} // no-op in the non-debug build. |
| #endif |
| |
| void* MoreCore(intptr_t increment); |
| |
| // Hands unused pages back to the system. |
| virtual size_t Trim() = 0; |
| |
| // Perform a mspace_inspect_all which calls back for each allocation chunk. The chunk may not be |
| // in use, indicated by num_bytes equaling zero. |
| virtual void Walk(WalkCallback callback, void* arg) = 0; |
| |
| // Returns the number of bytes that the space has currently obtained from the system. This is |
| // greater or equal to the amount of live data in the space. |
| virtual size_t GetFootprint() = 0; |
| |
| // Returns the number of bytes that the heap is allowed to obtain from the system via MoreCore. |
| virtual size_t GetFootprintLimit() = 0; |
| |
| // Set the maximum number of bytes that the heap is allowed to obtain from the system via |
| // MoreCore. Note this is used to stop the mspace growing beyond the limit to Capacity. When |
| // allocations fail we GC before increasing the footprint limit and allowing the mspace to grow. |
| virtual void SetFootprintLimit(size_t limit) = 0; |
| |
| // Removes the fork time growth limit on capacity, allowing the application to allocate up to the |
| // maximum reserved size of the heap. |
| void ClearGrowthLimit() { |
| growth_limit_ = NonGrowthLimitCapacity(); |
| } |
| |
| // Override capacity so that we only return the possibly limited capacity |
| size_t Capacity() const { |
| return growth_limit_; |
| } |
| |
| // The total amount of memory reserved for the alloc space. |
| size_t NonGrowthLimitCapacity() const { |
| return GetMemMap()->Size(); |
| } |
| |
| void Dump(std::ostream& os) const; |
| |
| void SetGrowthLimit(size_t growth_limit); |
| |
| virtual MallocSpace* CreateInstance(const std::string& name, MemMap* mem_map, void* allocator, |
| byte* begin, byte* end, byte* limit, size_t growth_limit, |
| bool can_move_objects) = 0; |
| |
| // Splits ourself into a zygote space and new malloc space which has our unused memory. When true, |
| // the low memory mode argument specifies that the heap wishes the created space to be more |
| // aggressive in releasing unused pages. Invalidates the space its called on. |
| ZygoteSpace* CreateZygoteSpace(const char* alloc_space_name, bool low_memory_mode, |
| MallocSpace** out_malloc_space) NO_THREAD_SAFETY_ANALYSIS; |
| virtual uint64_t GetBytesAllocated() = 0; |
| virtual uint64_t GetObjectsAllocated() = 0; |
| |
| // Returns the class of a recently freed object. |
| mirror::Class* FindRecentFreedObject(const mirror::Object* obj); |
| |
| bool CanMoveObjects() const OVERRIDE { |
| return can_move_objects_; |
| } |
| |
| virtual void LogFragmentationAllocFailure(std::ostream& os, size_t failed_alloc_bytes) = 0; |
| |
| protected: |
| MallocSpace(const std::string& name, MemMap* mem_map, byte* begin, byte* end, |
| byte* limit, size_t growth_limit, bool create_bitmaps, bool can_move_objects, |
| size_t starting_size, size_t initial_size); |
| |
| static MemMap* CreateMemMap(const std::string& name, size_t starting_size, size_t* initial_size, |
| size_t* growth_limit, size_t* capacity, byte* requested_begin); |
| |
| // When true the low memory mode argument specifies that the heap wishes the created allocator to |
| // be more aggressive in releasing unused pages. |
| virtual void* CreateAllocator(void* base, size_t morecore_start, size_t initial_size, |
| size_t maximum_size, bool low_memory_mode) = 0; |
| |
| virtual void RegisterRecentFree(mirror::Object* ptr) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) |
| EXCLUSIVE_LOCKS_REQUIRED(lock_); |
| |
| virtual accounting::ContinuousSpaceBitmap::SweepCallback* GetSweepCallback() { |
| return &SweepCallback; |
| } |
| |
| // Recent allocation buffer. |
| static constexpr size_t kRecentFreeCount = kDebugSpaces ? (1 << 16) : 0; |
| static constexpr size_t kRecentFreeMask = kRecentFreeCount - 1; |
| std::pair<const mirror::Object*, mirror::Class*> recent_freed_objects_[kRecentFreeCount]; |
| size_t recent_free_pos_; |
| |
| static size_t bitmap_index_; |
| |
| // Used to ensure mutual exclusion when the allocation spaces data structures are being modified. |
| Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; |
| |
| // The capacity of the alloc space until such time that ClearGrowthLimit is called. |
| // The underlying mem_map_ controls the maximum size we allow the heap to grow to. The growth |
| // limit is a value <= to the mem_map_ capacity used for ergonomic reasons because of the zygote. |
| // Prior to forking the zygote the heap will have a maximally sized mem_map_ but the growth_limit_ |
| // will be set to a lower value. The growth_limit_ is used as the capacity of the alloc_space_, |
| // however, capacity normally can't vary. In the case of the growth_limit_ it can be cleared |
| // one time by a call to ClearGrowthLimit. |
| size_t growth_limit_; |
| |
| // True if objects in the space are movable. |
| const bool can_move_objects_; |
| |
| // Starting and initial sized, used when you reset the space. |
| const size_t starting_size_; |
| const size_t initial_size_; |
| |
| private: |
| static void SweepCallback(size_t num_ptrs, mirror::Object** ptrs, void* arg) |
| SHARED_LOCKS_REQUIRED(Locks::mutator_lock_); |
| |
| DISALLOW_COPY_AND_ASSIGN(MallocSpace); |
| }; |
| |
| } // namespace space |
| } // namespace gc |
| } // namespace art |
| |
| #endif // ART_RUNTIME_GC_SPACE_MALLOC_SPACE_H_ |