| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ART_RUNTIME_GC_SPACE_SPACE_TEST_H_ |
| #define ART_RUNTIME_GC_SPACE_SPACE_TEST_H_ |
| |
| #include <stdint.h> |
| #include <memory> |
| |
| #include "common_runtime_test.h" |
| #include "globals.h" |
| #include "mirror/array-inl.h" |
| #include "mirror/class-inl.h" |
| #include "mirror/class_loader.h" |
| #include "mirror/object-inl.h" |
| #include "scoped_thread_state_change.h" |
| #include "zygote_space.h" |
| |
| namespace art { |
| namespace gc { |
| namespace space { |
| |
| class SpaceTest : public CommonRuntimeTest { |
| public: |
| jobject byte_array_class_; |
| |
| SpaceTest() : byte_array_class_(nullptr) { |
| } |
| |
| void AddSpace(ContinuousSpace* space, bool revoke = true) { |
| Heap* heap = Runtime::Current()->GetHeap(); |
| if (revoke) { |
| heap->RevokeAllThreadLocalBuffers(); |
| } |
| heap->AddSpace(space); |
| heap->SetSpaceAsDefault(space); |
| } |
| |
| mirror::Class* GetByteArrayClass(Thread* self) SHARED_REQUIRES(Locks::mutator_lock_) { |
| StackHandleScope<1> hs(self); |
| auto null_loader(hs.NewHandle<mirror::ClassLoader>(nullptr)); |
| if (byte_array_class_ == nullptr) { |
| mirror::Class* byte_array_class = |
| Runtime::Current()->GetClassLinker()->FindClass(self, "[B", null_loader); |
| EXPECT_TRUE(byte_array_class != nullptr); |
| byte_array_class_ = self->GetJniEnv()->NewLocalRef(byte_array_class); |
| EXPECT_TRUE(byte_array_class_ != nullptr); |
| } |
| return reinterpret_cast<mirror::Class*>(self->DecodeJObject(byte_array_class_)); |
| } |
| |
| mirror::Object* Alloc(space::MallocSpace* alloc_space, Thread* self, size_t bytes, |
| size_t* bytes_allocated, size_t* usable_size, |
| size_t* bytes_tl_bulk_allocated) |
| SHARED_REQUIRES(Locks::mutator_lock_) { |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Class> byte_array_class(hs.NewHandle(GetByteArrayClass(self))); |
| mirror::Object* obj = alloc_space->Alloc(self, bytes, bytes_allocated, usable_size, |
| bytes_tl_bulk_allocated); |
| if (obj != nullptr) { |
| InstallClass(obj, byte_array_class.Get(), bytes); |
| } |
| return obj; |
| } |
| |
| mirror::Object* AllocWithGrowth(space::MallocSpace* alloc_space, Thread* self, size_t bytes, |
| size_t* bytes_allocated, size_t* usable_size, |
| size_t* bytes_tl_bulk_allocated) |
| SHARED_REQUIRES(Locks::mutator_lock_) { |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Class> byte_array_class(hs.NewHandle(GetByteArrayClass(self))); |
| mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size, |
| bytes_tl_bulk_allocated); |
| if (obj != nullptr) { |
| InstallClass(obj, byte_array_class.Get(), bytes); |
| } |
| return obj; |
| } |
| |
| void InstallClass(mirror::Object* o, mirror::Class* byte_array_class, size_t size) |
| SHARED_REQUIRES(Locks::mutator_lock_) { |
| // Note the minimum size, which is the size of a zero-length byte array. |
| EXPECT_GE(size, SizeOfZeroLengthByteArray()); |
| EXPECT_TRUE(byte_array_class != nullptr); |
| o->SetClass(byte_array_class); |
| if (kUseBakerOrBrooksReadBarrier) { |
| // Like the proper heap object allocation, install and verify |
| // the correct read barrier pointer. |
| if (kUseBrooksReadBarrier) { |
| o->SetReadBarrierPointer(o); |
| } |
| o->AssertReadBarrierPointer(); |
| } |
| mirror::Array* arr = o->AsArray<kVerifyNone>(); |
| size_t header_size = SizeOfZeroLengthByteArray(); |
| int32_t length = size - header_size; |
| arr->SetLength(length); |
| EXPECT_EQ(arr->SizeOf<kVerifyNone>(), size); |
| } |
| |
| static size_t SizeOfZeroLengthByteArray() { |
| return mirror::Array::DataOffset(Primitive::ComponentSize(Primitive::kPrimByte)).Uint32Value(); |
| } |
| |
| typedef MallocSpace* (*CreateSpaceFn)(const std::string& name, size_t initial_size, size_t growth_limit, |
| size_t capacity, uint8_t* requested_begin); |
| void InitTestBody(CreateSpaceFn create_space); |
| void ZygoteSpaceTestBody(CreateSpaceFn create_space); |
| void AllocAndFreeTestBody(CreateSpaceFn create_space); |
| void AllocAndFreeListTestBody(CreateSpaceFn create_space); |
| |
| void SizeFootPrintGrowthLimitAndTrimBody(MallocSpace* space, intptr_t object_size, |
| int round, size_t growth_limit); |
| void SizeFootPrintGrowthLimitAndTrimDriver(size_t object_size, CreateSpaceFn create_space); |
| }; |
| |
| static inline size_t test_rand(size_t* seed) { |
| *seed = *seed * 1103515245 + 12345; |
| return *seed; |
| } |
| |
| void SpaceTest::InitTestBody(CreateSpaceFn create_space) { |
| // This will lead to error messages in the log. |
| ScopedLogSeverity sls(LogSeverity::FATAL); |
| |
| { |
| // Init < max == growth |
| std::unique_ptr<Space> space(create_space("test", 16 * MB, 32 * MB, 32 * MB, nullptr)); |
| EXPECT_TRUE(space.get() != nullptr); |
| } |
| { |
| // Init == max == growth |
| std::unique_ptr<Space> space(create_space("test", 16 * MB, 16 * MB, 16 * MB, nullptr)); |
| EXPECT_TRUE(space.get() != nullptr); |
| } |
| { |
| // Init > max == growth |
| std::unique_ptr<Space> space(create_space("test", 32 * MB, 16 * MB, 16 * MB, nullptr)); |
| EXPECT_TRUE(space.get() == nullptr); |
| } |
| { |
| // Growth == init < max |
| std::unique_ptr<Space> space(create_space("test", 16 * MB, 16 * MB, 32 * MB, nullptr)); |
| EXPECT_TRUE(space.get() != nullptr); |
| } |
| { |
| // Growth < init < max |
| std::unique_ptr<Space> space(create_space("test", 16 * MB, 8 * MB, 32 * MB, nullptr)); |
| EXPECT_TRUE(space.get() == nullptr); |
| } |
| { |
| // Init < growth < max |
| std::unique_ptr<Space> space(create_space("test", 8 * MB, 16 * MB, 32 * MB, nullptr)); |
| EXPECT_TRUE(space.get() != nullptr); |
| } |
| { |
| // Init < max < growth |
| std::unique_ptr<Space> space(create_space("test", 8 * MB, 32 * MB, 16 * MB, nullptr)); |
| EXPECT_TRUE(space.get() == nullptr); |
| } |
| } |
| |
| // TODO: This test is not very good, we should improve it. |
| // The test should do more allocations before the creation of the ZygoteSpace, and then do |
| // allocations after the ZygoteSpace is created. The test should also do some GCs to ensure that |
| // the GC works with the ZygoteSpace. |
| void SpaceTest::ZygoteSpaceTestBody(CreateSpaceFn create_space) { |
| size_t dummy; |
| MallocSpace* space(create_space("test", 4 * MB, 16 * MB, 16 * MB, nullptr)); |
| ASSERT_TRUE(space != nullptr); |
| |
| // Make space findable to the heap, will also delete space when runtime is cleaned up |
| AddSpace(space); |
| Thread* self = Thread::Current(); |
| ScopedObjectAccess soa(self); |
| |
| // Succeeds, fits without adjusting the footprint limit. |
| size_t ptr1_bytes_allocated, ptr1_usable_size, ptr1_bytes_tl_bulk_allocated; |
| StackHandleScope<3> hs(soa.Self()); |
| MutableHandle<mirror::Object> ptr1( |
| hs.NewHandle(Alloc(space, self, 1 * MB, &ptr1_bytes_allocated, &ptr1_usable_size, |
| &ptr1_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr1.Get() != nullptr); |
| EXPECT_LE(1U * MB, ptr1_bytes_allocated); |
| EXPECT_LE(1U * MB, ptr1_usable_size); |
| EXPECT_LE(ptr1_usable_size, ptr1_bytes_allocated); |
| EXPECT_EQ(ptr1_bytes_tl_bulk_allocated, ptr1_bytes_allocated); |
| |
| // Fails, requires a higher footprint limit. |
| mirror::Object* ptr2 = Alloc(space, self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr2 == nullptr); |
| |
| // Succeeds, adjusts the footprint. |
| size_t ptr3_bytes_allocated, ptr3_usable_size, ptr3_bytes_tl_bulk_allocated; |
| MutableHandle<mirror::Object> ptr3( |
| hs.NewHandle(AllocWithGrowth(space, self, 8 * MB, &ptr3_bytes_allocated, &ptr3_usable_size, |
| &ptr3_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr3.Get() != nullptr); |
| EXPECT_LE(8U * MB, ptr3_bytes_allocated); |
| EXPECT_LE(8U * MB, ptr3_usable_size); |
| EXPECT_LE(ptr3_usable_size, ptr3_bytes_allocated); |
| EXPECT_EQ(ptr3_bytes_tl_bulk_allocated, ptr3_bytes_allocated); |
| |
| // Fails, requires a higher footprint limit. |
| mirror::Object* ptr4 = space->Alloc(self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr4 == nullptr); |
| |
| // Also fails, requires a higher allowed footprint. |
| mirror::Object* ptr5 = space->AllocWithGrowth(self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr5 == nullptr); |
| |
| // Release some memory. |
| size_t free3 = space->AllocationSize(ptr3.Get(), nullptr); |
| EXPECT_EQ(free3, ptr3_bytes_allocated); |
| EXPECT_EQ(free3, space->Free(self, ptr3.Assign(nullptr))); |
| EXPECT_LE(8U * MB, free3); |
| |
| // Succeeds, now that memory has been freed. |
| size_t ptr6_bytes_allocated, ptr6_usable_size, ptr6_bytes_tl_bulk_allocated; |
| Handle<mirror::Object> ptr6( |
| hs.NewHandle(AllocWithGrowth(space, self, 9 * MB, &ptr6_bytes_allocated, &ptr6_usable_size, |
| &ptr6_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr6.Get() != nullptr); |
| EXPECT_LE(9U * MB, ptr6_bytes_allocated); |
| EXPECT_LE(9U * MB, ptr6_usable_size); |
| EXPECT_LE(ptr6_usable_size, ptr6_bytes_allocated); |
| EXPECT_EQ(ptr6_bytes_tl_bulk_allocated, ptr6_bytes_allocated); |
| |
| // Final clean up. |
| size_t free1 = space->AllocationSize(ptr1.Get(), nullptr); |
| space->Free(self, ptr1.Assign(nullptr)); |
| EXPECT_LE(1U * MB, free1); |
| |
| // Make sure that the zygote space isn't directly at the start of the space. |
| EXPECT_TRUE(space->Alloc(self, 1U * MB, &dummy, nullptr, &dummy) != nullptr); |
| |
| gc::Heap* heap = Runtime::Current()->GetHeap(); |
| space::Space* old_space = space; |
| heap->RemoveSpace(old_space); |
| heap->RevokeAllThreadLocalBuffers(); |
| space::ZygoteSpace* zygote_space = space->CreateZygoteSpace("alloc space", |
| heap->IsLowMemoryMode(), |
| &space); |
| delete old_space; |
| // Add the zygote space. |
| AddSpace(zygote_space, false); |
| |
| // Make space findable to the heap, will also delete space when runtime is cleaned up |
| AddSpace(space, false); |
| |
| // Succeeds, fits without adjusting the footprint limit. |
| ptr1.Assign(Alloc(space, self, 1 * MB, &ptr1_bytes_allocated, &ptr1_usable_size, |
| &ptr1_bytes_tl_bulk_allocated)); |
| EXPECT_TRUE(ptr1.Get() != nullptr); |
| EXPECT_LE(1U * MB, ptr1_bytes_allocated); |
| EXPECT_LE(1U * MB, ptr1_usable_size); |
| EXPECT_LE(ptr1_usable_size, ptr1_bytes_allocated); |
| EXPECT_EQ(ptr1_bytes_tl_bulk_allocated, ptr1_bytes_allocated); |
| |
| // Fails, requires a higher footprint limit. |
| ptr2 = Alloc(space, self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr2 == nullptr); |
| |
| // Succeeds, adjusts the footprint. |
| ptr3.Assign(AllocWithGrowth(space, self, 2 * MB, &ptr3_bytes_allocated, &ptr3_usable_size, |
| &ptr3_bytes_tl_bulk_allocated)); |
| EXPECT_TRUE(ptr3.Get() != nullptr); |
| EXPECT_LE(2U * MB, ptr3_bytes_allocated); |
| EXPECT_LE(2U * MB, ptr3_usable_size); |
| EXPECT_LE(ptr3_usable_size, ptr3_bytes_allocated); |
| EXPECT_EQ(ptr3_bytes_tl_bulk_allocated, ptr3_bytes_allocated); |
| space->Free(self, ptr3.Assign(nullptr)); |
| |
| // Final clean up. |
| free1 = space->AllocationSize(ptr1.Get(), nullptr); |
| space->Free(self, ptr1.Assign(nullptr)); |
| EXPECT_LE(1U * MB, free1); |
| } |
| |
| void SpaceTest::AllocAndFreeTestBody(CreateSpaceFn create_space) { |
| size_t dummy = 0; |
| MallocSpace* space(create_space("test", 4 * MB, 16 * MB, 16 * MB, nullptr)); |
| ASSERT_TRUE(space != nullptr); |
| Thread* self = Thread::Current(); |
| ScopedObjectAccess soa(self); |
| |
| // Make space findable to the heap, will also delete space when runtime is cleaned up |
| AddSpace(space); |
| |
| // Succeeds, fits without adjusting the footprint limit. |
| size_t ptr1_bytes_allocated, ptr1_usable_size, ptr1_bytes_tl_bulk_allocated; |
| StackHandleScope<3> hs(soa.Self()); |
| MutableHandle<mirror::Object> ptr1( |
| hs.NewHandle(Alloc(space, self, 1 * MB, &ptr1_bytes_allocated, &ptr1_usable_size, |
| &ptr1_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr1.Get() != nullptr); |
| EXPECT_LE(1U * MB, ptr1_bytes_allocated); |
| EXPECT_LE(1U * MB, ptr1_usable_size); |
| EXPECT_LE(ptr1_usable_size, ptr1_bytes_allocated); |
| EXPECT_EQ(ptr1_bytes_tl_bulk_allocated, ptr1_bytes_allocated); |
| |
| // Fails, requires a higher footprint limit. |
| mirror::Object* ptr2 = Alloc(space, self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr2 == nullptr); |
| |
| // Succeeds, adjusts the footprint. |
| size_t ptr3_bytes_allocated, ptr3_usable_size, ptr3_bytes_tl_bulk_allocated; |
| MutableHandle<mirror::Object> ptr3( |
| hs.NewHandle(AllocWithGrowth(space, self, 8 * MB, &ptr3_bytes_allocated, &ptr3_usable_size, |
| &ptr3_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr3.Get() != nullptr); |
| EXPECT_LE(8U * MB, ptr3_bytes_allocated); |
| EXPECT_LE(8U * MB, ptr3_usable_size); |
| EXPECT_LE(ptr3_usable_size, ptr3_bytes_allocated); |
| EXPECT_EQ(ptr3_bytes_tl_bulk_allocated, ptr3_bytes_allocated); |
| |
| // Fails, requires a higher footprint limit. |
| mirror::Object* ptr4 = Alloc(space, self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr4 == nullptr); |
| |
| // Also fails, requires a higher allowed footprint. |
| mirror::Object* ptr5 = AllocWithGrowth(space, self, 8 * MB, &dummy, nullptr, &dummy); |
| EXPECT_TRUE(ptr5 == nullptr); |
| |
| // Release some memory. |
| size_t free3 = space->AllocationSize(ptr3.Get(), nullptr); |
| EXPECT_EQ(free3, ptr3_bytes_allocated); |
| space->Free(self, ptr3.Assign(nullptr)); |
| EXPECT_LE(8U * MB, free3); |
| |
| // Succeeds, now that memory has been freed. |
| size_t ptr6_bytes_allocated, ptr6_usable_size, ptr6_bytes_tl_bulk_allocated; |
| Handle<mirror::Object> ptr6( |
| hs.NewHandle(AllocWithGrowth(space, self, 9 * MB, &ptr6_bytes_allocated, &ptr6_usable_size, |
| &ptr6_bytes_tl_bulk_allocated))); |
| EXPECT_TRUE(ptr6.Get() != nullptr); |
| EXPECT_LE(9U * MB, ptr6_bytes_allocated); |
| EXPECT_LE(9U * MB, ptr6_usable_size); |
| EXPECT_LE(ptr6_usable_size, ptr6_bytes_allocated); |
| EXPECT_EQ(ptr6_bytes_tl_bulk_allocated, ptr6_bytes_allocated); |
| |
| // Final clean up. |
| size_t free1 = space->AllocationSize(ptr1.Get(), nullptr); |
| space->Free(self, ptr1.Assign(nullptr)); |
| EXPECT_LE(1U * MB, free1); |
| } |
| |
| void SpaceTest::AllocAndFreeListTestBody(CreateSpaceFn create_space) { |
| MallocSpace* space(create_space("test", 4 * MB, 16 * MB, 16 * MB, nullptr)); |
| ASSERT_TRUE(space != nullptr); |
| |
| // Make space findable to the heap, will also delete space when runtime is cleaned up |
| AddSpace(space); |
| Thread* self = Thread::Current(); |
| ScopedObjectAccess soa(self); |
| |
| // Succeeds, fits without adjusting the max allowed footprint. |
| mirror::Object* lots_of_objects[1024]; |
| for (size_t i = 0; i < arraysize(lots_of_objects); i++) { |
| size_t allocation_size, usable_size, bytes_tl_bulk_allocated; |
| size_t size_of_zero_length_byte_array = SizeOfZeroLengthByteArray(); |
| lots_of_objects[i] = Alloc(space, self, size_of_zero_length_byte_array, &allocation_size, |
| &usable_size, &bytes_tl_bulk_allocated); |
| EXPECT_TRUE(lots_of_objects[i] != nullptr); |
| size_t computed_usable_size; |
| EXPECT_EQ(allocation_size, space->AllocationSize(lots_of_objects[i], &computed_usable_size)); |
| EXPECT_EQ(usable_size, computed_usable_size); |
| EXPECT_TRUE(bytes_tl_bulk_allocated == 0 || |
| bytes_tl_bulk_allocated >= allocation_size); |
| } |
| |
| // Release memory. |
| space->FreeList(self, arraysize(lots_of_objects), lots_of_objects); |
| |
| // Succeeds, fits by adjusting the max allowed footprint. |
| for (size_t i = 0; i < arraysize(lots_of_objects); i++) { |
| size_t allocation_size, usable_size, bytes_tl_bulk_allocated; |
| lots_of_objects[i] = AllocWithGrowth(space, self, 1024, &allocation_size, &usable_size, |
| &bytes_tl_bulk_allocated); |
| EXPECT_TRUE(lots_of_objects[i] != nullptr); |
| size_t computed_usable_size; |
| EXPECT_EQ(allocation_size, space->AllocationSize(lots_of_objects[i], &computed_usable_size)); |
| EXPECT_EQ(usable_size, computed_usable_size); |
| EXPECT_TRUE(bytes_tl_bulk_allocated == 0 || |
| bytes_tl_bulk_allocated >= allocation_size); |
| } |
| |
| // Release memory. |
| space->FreeList(self, arraysize(lots_of_objects), lots_of_objects); |
| } |
| |
| void SpaceTest::SizeFootPrintGrowthLimitAndTrimBody(MallocSpace* space, intptr_t object_size, |
| int round, size_t growth_limit) { |
| if (((object_size > 0 && object_size >= static_cast<intptr_t>(growth_limit))) || |
| ((object_size < 0 && -object_size >= static_cast<intptr_t>(growth_limit)))) { |
| // No allocation can succeed |
| return; |
| } |
| |
| // The space's footprint equals amount of resources requested from system |
| size_t footprint = space->GetFootprint(); |
| |
| // The space must at least have its book keeping allocated |
| EXPECT_GT(footprint, 0u); |
| |
| // But it shouldn't exceed the initial size |
| EXPECT_LE(footprint, growth_limit); |
| |
| // space's size shouldn't exceed the initial size |
| EXPECT_LE(space->Size(), growth_limit); |
| |
| // this invariant should always hold or else the space has grown to be larger than what the |
| // space believes its size is (which will break invariants) |
| EXPECT_GE(space->Size(), footprint); |
| |
| // Fill the space with lots of small objects up to the growth limit |
| size_t max_objects = (growth_limit / (object_size > 0 ? object_size : 8)) + 1; |
| std::unique_ptr<mirror::Object*[]> lots_of_objects(new mirror::Object*[max_objects]); |
| size_t last_object = 0; // last object for which allocation succeeded |
| size_t amount_allocated = 0; // amount of space allocated |
| Thread* self = Thread::Current(); |
| ScopedObjectAccess soa(self); |
| size_t rand_seed = 123456789; |
| for (size_t i = 0; i < max_objects; i++) { |
| size_t alloc_fails = 0; // number of failed allocations |
| size_t max_fails = 30; // number of times we fail allocation before giving up |
| for (; alloc_fails < max_fails; alloc_fails++) { |
| size_t alloc_size; |
| if (object_size > 0) { |
| alloc_size = object_size; |
| } else { |
| alloc_size = test_rand(&rand_seed) % static_cast<size_t>(-object_size); |
| // Note the minimum size, which is the size of a zero-length byte array. |
| size_t size_of_zero_length_byte_array = SizeOfZeroLengthByteArray(); |
| if (alloc_size < size_of_zero_length_byte_array) { |
| alloc_size = size_of_zero_length_byte_array; |
| } |
| } |
| StackHandleScope<1> hs(soa.Self()); |
| auto object(hs.NewHandle<mirror::Object>(nullptr)); |
| size_t bytes_allocated = 0; |
| size_t bytes_tl_bulk_allocated; |
| if (round <= 1) { |
| object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, |
| &bytes_tl_bulk_allocated)); |
| } else { |
| object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr, |
| &bytes_tl_bulk_allocated)); |
| } |
| footprint = space->GetFootprint(); |
| EXPECT_GE(space->Size(), footprint); // invariant |
| if (object.Get() != nullptr) { // allocation succeeded |
| lots_of_objects[i] = object.Get(); |
| size_t allocation_size = space->AllocationSize(object.Get(), nullptr); |
| EXPECT_EQ(bytes_allocated, allocation_size); |
| if (object_size > 0) { |
| EXPECT_GE(allocation_size, static_cast<size_t>(object_size)); |
| } else { |
| EXPECT_GE(allocation_size, 8u); |
| } |
| EXPECT_TRUE(bytes_tl_bulk_allocated == 0 || |
| bytes_tl_bulk_allocated >= allocation_size); |
| amount_allocated += allocation_size; |
| break; |
| } |
| } |
| if (alloc_fails == max_fails) { |
| last_object = i; |
| break; |
| } |
| } |
| CHECK_NE(last_object, 0u); // we should have filled the space |
| EXPECT_GT(amount_allocated, 0u); |
| |
| // We shouldn't have gone past the growth_limit |
| EXPECT_LE(amount_allocated, growth_limit); |
| EXPECT_LE(footprint, growth_limit); |
| EXPECT_LE(space->Size(), growth_limit); |
| |
| // footprint and size should agree with amount allocated |
| EXPECT_GE(footprint, amount_allocated); |
| EXPECT_GE(space->Size(), amount_allocated); |
| |
| // Release storage in a semi-adhoc manner |
| size_t free_increment = 96; |
| while (true) { |
| { |
| ScopedThreadStateChange tsc(self, kNative); |
| // Give the space a haircut. |
| space->Trim(); |
| } |
| |
| // Bounds sanity |
| footprint = space->GetFootprint(); |
| EXPECT_LE(amount_allocated, growth_limit); |
| EXPECT_GE(footprint, amount_allocated); |
| EXPECT_LE(footprint, growth_limit); |
| EXPECT_GE(space->Size(), amount_allocated); |
| EXPECT_LE(space->Size(), growth_limit); |
| |
| if (free_increment == 0) { |
| break; |
| } |
| |
| // Free some objects |
| for (size_t i = 0; i < last_object; i += free_increment) { |
| mirror::Object* object = lots_of_objects.get()[i]; |
| if (object == nullptr) { |
| continue; |
| } |
| size_t allocation_size = space->AllocationSize(object, nullptr); |
| if (object_size > 0) { |
| EXPECT_GE(allocation_size, static_cast<size_t>(object_size)); |
| } else { |
| EXPECT_GE(allocation_size, 8u); |
| } |
| space->Free(self, object); |
| lots_of_objects.get()[i] = nullptr; |
| amount_allocated -= allocation_size; |
| footprint = space->GetFootprint(); |
| EXPECT_GE(space->Size(), footprint); // invariant |
| } |
| |
| free_increment >>= 1; |
| } |
| |
| // The space has become empty here before allocating a large object |
| // below. For RosAlloc, revoke thread-local runs, which are kept |
| // even when empty for a performance reason, so that they won't |
| // cause the following large object allocation to fail due to |
| // potential fragmentation. Note they are normally revoked at each |
| // GC (but no GC here.) |
| space->RevokeAllThreadLocalBuffers(); |
| |
| // All memory was released, try a large allocation to check freed memory is being coalesced |
| StackHandleScope<1> hs(soa.Self()); |
| auto large_object(hs.NewHandle<mirror::Object>(nullptr)); |
| size_t three_quarters_space = (growth_limit / 2) + (growth_limit / 4); |
| size_t bytes_allocated = 0; |
| size_t bytes_tl_bulk_allocated; |
| if (round <= 1) { |
| large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullptr, |
| &bytes_tl_bulk_allocated)); |
| } else { |
| large_object.Assign(AllocWithGrowth(space, self, three_quarters_space, &bytes_allocated, |
| nullptr, &bytes_tl_bulk_allocated)); |
| } |
| EXPECT_TRUE(large_object.Get() != nullptr); |
| |
| // Sanity check footprint |
| footprint = space->GetFootprint(); |
| EXPECT_LE(footprint, growth_limit); |
| EXPECT_GE(space->Size(), footprint); |
| EXPECT_LE(space->Size(), growth_limit); |
| |
| // Clean up |
| space->Free(self, large_object.Assign(nullptr)); |
| |
| // Sanity check footprint |
| footprint = space->GetFootprint(); |
| EXPECT_LE(footprint, growth_limit); |
| EXPECT_GE(space->Size(), footprint); |
| EXPECT_LE(space->Size(), growth_limit); |
| } |
| |
| void SpaceTest::SizeFootPrintGrowthLimitAndTrimDriver(size_t object_size, CreateSpaceFn create_space) { |
| if (object_size < SizeOfZeroLengthByteArray()) { |
| // Too small for the object layout/model. |
| return; |
| } |
| size_t initial_size = 4 * MB; |
| size_t growth_limit = 8 * MB; |
| size_t capacity = 16 * MB; |
| MallocSpace* space(create_space("test", initial_size, growth_limit, capacity, nullptr)); |
| ASSERT_TRUE(space != nullptr); |
| |
| // Basic sanity |
| EXPECT_EQ(space->Capacity(), growth_limit); |
| EXPECT_EQ(space->NonGrowthLimitCapacity(), capacity); |
| |
| // Make space findable to the heap, will also delete space when runtime is cleaned up |
| AddSpace(space); |
| |
| // In this round we don't allocate with growth and therefore can't grow past the initial size. |
| // This effectively makes the growth_limit the initial_size, so assert this. |
| SizeFootPrintGrowthLimitAndTrimBody(space, object_size, 1, initial_size); |
| SizeFootPrintGrowthLimitAndTrimBody(space, object_size, 2, growth_limit); |
| // Remove growth limit |
| space->ClearGrowthLimit(); |
| EXPECT_EQ(space->Capacity(), capacity); |
| SizeFootPrintGrowthLimitAndTrimBody(space, object_size, 3, capacity); |
| } |
| |
| #define TEST_SizeFootPrintGrowthLimitAndTrimStatic(name, spaceName, spaceFn, size) \ |
| TEST_F(spaceName##StaticTest, SizeFootPrintGrowthLimitAndTrim_AllocationsOf_##name) { \ |
| SizeFootPrintGrowthLimitAndTrimDriver(size, spaceFn); \ |
| } |
| |
| #define TEST_SizeFootPrintGrowthLimitAndTrimRandom(name, spaceName, spaceFn, size) \ |
| TEST_F(spaceName##RandomTest, SizeFootPrintGrowthLimitAndTrim_RandomAllocationsWithMax_##name) { \ |
| SizeFootPrintGrowthLimitAndTrimDriver(-size, spaceFn); \ |
| } |
| |
| #define TEST_SPACE_CREATE_FN_BASE(spaceName, spaceFn) \ |
| class spaceName##BaseTest : public SpaceTest { \ |
| }; \ |
| \ |
| TEST_F(spaceName##BaseTest, Init) { \ |
| InitTestBody(spaceFn); \ |
| } \ |
| TEST_F(spaceName##BaseTest, ZygoteSpace) { \ |
| ZygoteSpaceTestBody(spaceFn); \ |
| } \ |
| TEST_F(spaceName##BaseTest, AllocAndFree) { \ |
| AllocAndFreeTestBody(spaceFn); \ |
| } \ |
| TEST_F(spaceName##BaseTest, AllocAndFreeList) { \ |
| AllocAndFreeListTestBody(spaceFn); \ |
| } |
| |
| #define TEST_SPACE_CREATE_FN_STATIC(spaceName, spaceFn) \ |
| class spaceName##StaticTest : public SpaceTest { \ |
| }; \ |
| \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(12B, spaceName, spaceFn, 12) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(16B, spaceName, spaceFn, 16) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(24B, spaceName, spaceFn, 24) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(32B, spaceName, spaceFn, 32) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(64B, spaceName, spaceFn, 64) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(128B, spaceName, spaceFn, 128) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(1KB, spaceName, spaceFn, 1 * KB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(4KB, spaceName, spaceFn, 4 * KB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(1MB, spaceName, spaceFn, 1 * MB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(4MB, spaceName, spaceFn, 4 * MB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimStatic(8MB, spaceName, spaceFn, 8 * MB) |
| |
| #define TEST_SPACE_CREATE_FN_RANDOM(spaceName, spaceFn) \ |
| class spaceName##RandomTest : public SpaceTest { \ |
| }; \ |
| \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(16B, spaceName, spaceFn, 16) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(24B, spaceName, spaceFn, 24) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(32B, spaceName, spaceFn, 32) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(64B, spaceName, spaceFn, 64) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(128B, spaceName, spaceFn, 128) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(1KB, spaceName, spaceFn, 1 * KB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(4KB, spaceName, spaceFn, 4 * KB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(1MB, spaceName, spaceFn, 1 * MB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(4MB, spaceName, spaceFn, 4 * MB) \ |
| TEST_SizeFootPrintGrowthLimitAndTrimRandom(8MB, spaceName, spaceFn, 8 * MB) |
| |
| } // namespace space |
| } // namespace gc |
| } // namespace art |
| |
| #endif // ART_RUNTIME_GC_SPACE_SPACE_TEST_H_ |