blob: 1e5d30d0e7f56b738d0dd6fc1e87ab41d7c63763 [file] [log] [blame]
Ben Murdoch4a90d5f2016-03-22 12:00:34 +00001// Copyright 2012 the V8 project authors. All rights reserved.
2// Redistribution and use in source and binary forms, with or without
3// modification, are permitted provided that the following conditions are
4// met:
5//
6// * Redistributions of source code must retain the above copyright
7// notice, this list of conditions and the following disclaimer.
8// * Redistributions in binary form must reproduce the above
9// copyright notice, this list of conditions and the following
10// disclaimer in the documentation and/or other materials provided
11// with the distribution.
12// * Neither the name of Google Inc. nor the names of its
13// contributors may be used to endorse or promote products derived
14// from this software without specific prior written permission.
15//
16// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
28#include <stdlib.h>
29
30#ifdef __linux__
31#include <errno.h>
32#include <fcntl.h>
33#include <sys/stat.h>
34#include <sys/types.h>
35#include <unistd.h>
36#endif
37
38#include <utility>
39
40#include "src/v8.h"
41
42#include "src/full-codegen/full-codegen.h"
43#include "src/global-handles.h"
44#include "test/cctest/cctest.h"
45#include "test/cctest/heap/heap-tester.h"
Ben Murdoch61f157c2016-09-16 13:49:30 +010046#include "test/cctest/heap/heap-utils.h"
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000047
48using namespace v8::internal;
49using v8::Just;
50
51
52TEST(MarkingDeque) {
53 CcTest::InitializeVM();
54 int mem_size = 20 * kPointerSize;
55 byte* mem = NewArray<byte>(20*kPointerSize);
56 Address low = reinterpret_cast<Address>(mem);
57 Address high = low + mem_size;
58 MarkingDeque s;
59 s.Initialize(low, high);
60
61 Address original_address = reinterpret_cast<Address>(&s);
62 Address current_address = original_address;
63 while (!s.IsFull()) {
64 s.Push(HeapObject::FromAddress(current_address));
65 current_address += kPointerSize;
66 }
67
68 while (!s.IsEmpty()) {
69 Address value = s.Pop()->address();
70 current_address -= kPointerSize;
71 CHECK_EQ(current_address, value);
72 }
73
74 CHECK_EQ(original_address, current_address);
75 DeleteArray(mem);
76}
77
Ben Murdoch61f157c2016-09-16 13:49:30 +010078TEST(Promotion) {
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000079 CcTest::InitializeVM();
Ben Murdoch61f157c2016-09-16 13:49:30 +010080 Isolate* isolate = CcTest::i_isolate();
81 {
82 v8::HandleScope sc(CcTest::isolate());
83 Heap* heap = isolate->heap();
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000084
Ben Murdoch61f157c2016-09-16 13:49:30 +010085 heap::SealCurrentObjects(heap);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000086
Ben Murdoch61f157c2016-09-16 13:49:30 +010087 int array_length =
88 heap::FixedArrayLenFromSize(Page::kMaxRegularHeapObjectSize);
89 Handle<FixedArray> array = isolate->factory()->NewFixedArray(array_length);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000090
Ben Murdoch61f157c2016-09-16 13:49:30 +010091 // Array should be in the new space.
92 CHECK(heap->InSpace(*array, NEW_SPACE));
93 heap->CollectAllGarbage();
94 heap->CollectAllGarbage();
95 CHECK(heap->InSpace(*array, OLD_SPACE));
96 }
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000097}
98
Ben Murdoch4a90d5f2016-03-22 12:00:34 +000099HEAP_TEST(NoPromotion) {
100 CcTest::InitializeVM();
Ben Murdoch61f157c2016-09-16 13:49:30 +0100101 Isolate* isolate = CcTest::i_isolate();
102 {
103 v8::HandleScope sc(CcTest::isolate());
104 Heap* heap = isolate->heap();
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000105
Ben Murdoch61f157c2016-09-16 13:49:30 +0100106 heap::SealCurrentObjects(heap);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000107
Ben Murdoch61f157c2016-09-16 13:49:30 +0100108 int array_length =
109 heap::FixedArrayLenFromSize(Page::kMaxRegularHeapObjectSize);
110 Handle<FixedArray> array = isolate->factory()->NewFixedArray(array_length);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000111
Ben Murdoch61f157c2016-09-16 13:49:30 +0100112 heap->set_force_oom(true);
113 // Array should be in the new space.
114 CHECK(heap->InSpace(*array, NEW_SPACE));
115 heap->CollectAllGarbage();
116 heap->CollectAllGarbage();
117 CHECK(heap->InSpace(*array, NEW_SPACE));
118 }
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000119}
120
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000121HEAP_TEST(MarkCompactCollector) {
122 FLAG_incremental_marking = false;
123 FLAG_retain_maps_for_n_gc = 0;
124 CcTest::InitializeVM();
125 Isolate* isolate = CcTest::i_isolate();
126 Heap* heap = CcTest::heap();
127 Factory* factory = isolate->factory();
128
129 v8::HandleScope sc(CcTest::isolate());
130 Handle<JSGlobalObject> global(isolate->context()->global_object());
131
132 // call mark-compact when heap is empty
133 heap->CollectGarbage(OLD_SPACE, "trigger 1");
134
135 // keep allocating garbage in new space until it fails
136 const int arraysize = 100;
137 AllocationResult allocation;
138 do {
139 allocation = heap->AllocateFixedArray(arraysize);
140 } while (!allocation.IsRetry());
141 heap->CollectGarbage(NEW_SPACE, "trigger 2");
142 heap->AllocateFixedArray(arraysize).ToObjectChecked();
143
144 // keep allocating maps until it fails
145 do {
146 allocation = heap->AllocateMap(JS_OBJECT_TYPE, JSObject::kHeaderSize);
147 } while (!allocation.IsRetry());
148 heap->CollectGarbage(MAP_SPACE, "trigger 3");
149 heap->AllocateMap(JS_OBJECT_TYPE, JSObject::kHeaderSize).ToObjectChecked();
150
151 { HandleScope scope(isolate);
152 // allocate a garbage
153 Handle<String> func_name = factory->InternalizeUtf8String("theFunction");
154 Handle<JSFunction> function = factory->NewFunction(func_name);
155 JSReceiver::SetProperty(global, func_name, function, SLOPPY).Check();
156
157 factory->NewJSObject(function);
158 }
159
160 heap->CollectGarbage(OLD_SPACE, "trigger 4");
161
162 { HandleScope scope(isolate);
163 Handle<String> func_name = factory->InternalizeUtf8String("theFunction");
164 CHECK(Just(true) == JSReceiver::HasOwnProperty(global, func_name));
165 Handle<Object> func_value =
166 Object::GetProperty(global, func_name).ToHandleChecked();
167 CHECK(func_value->IsJSFunction());
168 Handle<JSFunction> function = Handle<JSFunction>::cast(func_value);
169 Handle<JSObject> obj = factory->NewJSObject(function);
170
171 Handle<String> obj_name = factory->InternalizeUtf8String("theObject");
172 JSReceiver::SetProperty(global, obj_name, obj, SLOPPY).Check();
173 Handle<String> prop_name = factory->InternalizeUtf8String("theSlot");
174 Handle<Smi> twenty_three(Smi::FromInt(23), isolate);
175 JSReceiver::SetProperty(obj, prop_name, twenty_three, SLOPPY).Check();
176 }
177
178 heap->CollectGarbage(OLD_SPACE, "trigger 5");
179
180 { HandleScope scope(isolate);
181 Handle<String> obj_name = factory->InternalizeUtf8String("theObject");
182 CHECK(Just(true) == JSReceiver::HasOwnProperty(global, obj_name));
183 Handle<Object> object =
184 Object::GetProperty(global, obj_name).ToHandleChecked();
185 CHECK(object->IsJSObject());
186 Handle<String> prop_name = factory->InternalizeUtf8String("theSlot");
187 CHECK_EQ(*Object::GetProperty(object, prop_name).ToHandleChecked(),
188 Smi::FromInt(23));
189 }
190}
191
192
193// TODO(1600): compaction of map space is temporary removed from GC.
194#if 0
195static Handle<Map> CreateMap(Isolate* isolate) {
196 return isolate->factory()->NewMap(JS_OBJECT_TYPE, JSObject::kHeaderSize);
197}
198
199
200TEST(MapCompact) {
201 FLAG_max_map_space_pages = 16;
202 CcTest::InitializeVM();
203 Isolate* isolate = CcTest::i_isolate();
204 Factory* factory = isolate->factory();
205
206 {
207 v8::HandleScope sc;
208 // keep allocating maps while pointers are still encodable and thus
209 // mark compact is permitted.
210 Handle<JSObject> root = factory->NewJSObjectFromMap(CreateMap());
211 do {
212 Handle<Map> map = CreateMap();
213 map->set_prototype(*root);
214 root = factory->NewJSObjectFromMap(map);
215 } while (CcTest::heap()->map_space()->MapPointersEncodable());
216 }
217 // Now, as we don't have any handles to just allocated maps, we should
218 // be able to trigger map compaction.
219 // To give an additional chance to fail, try to force compaction which
220 // should be impossible right now.
221 CcTest::heap()->CollectAllGarbage(Heap::kForceCompactionMask);
222 // And now map pointers should be encodable again.
223 CHECK(CcTest::heap()->map_space()->MapPointersEncodable());
224}
225#endif
226
227
228static int NumberOfWeakCalls = 0;
Ben Murdochc5610432016-08-08 18:44:38 +0100229static void WeakPointerCallback(const v8::WeakCallbackInfo<void>& data) {
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000230 std::pair<v8::Persistent<v8::Value>*, int>* p =
231 reinterpret_cast<std::pair<v8::Persistent<v8::Value>*, int>*>(
232 data.GetParameter());
233 CHECK_EQ(1234, p->second);
234 NumberOfWeakCalls++;
235 p->first->Reset();
236}
237
238
239HEAP_TEST(ObjectGroups) {
240 FLAG_incremental_marking = false;
241 CcTest::InitializeVM();
242 GlobalHandles* global_handles = CcTest::i_isolate()->global_handles();
243 Heap* heap = CcTest::heap();
244 NumberOfWeakCalls = 0;
245 v8::HandleScope handle_scope(CcTest::isolate());
246
247 Handle<Object> g1s1 =
248 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
249 Handle<Object> g1s2 =
250 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
251 Handle<Object> g1c1 =
252 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
253 std::pair<Handle<Object>*, int> g1s1_and_id(&g1s1, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100254 GlobalHandles::MakeWeak(
255 g1s1.location(), reinterpret_cast<void*>(&g1s1_and_id),
256 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000257 std::pair<Handle<Object>*, int> g1s2_and_id(&g1s2, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100258 GlobalHandles::MakeWeak(
259 g1s2.location(), reinterpret_cast<void*>(&g1s2_and_id),
260 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000261 std::pair<Handle<Object>*, int> g1c1_and_id(&g1c1, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100262 GlobalHandles::MakeWeak(
263 g1c1.location(), reinterpret_cast<void*>(&g1c1_and_id),
264 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000265
266 Handle<Object> g2s1 =
267 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
268 Handle<Object> g2s2 =
269 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
270 Handle<Object> g2c1 =
271 global_handles->Create(heap->AllocateFixedArray(1).ToObjectChecked());
272 std::pair<Handle<Object>*, int> g2s1_and_id(&g2s1, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100273 GlobalHandles::MakeWeak(
274 g2s1.location(), reinterpret_cast<void*>(&g2s1_and_id),
275 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000276 std::pair<Handle<Object>*, int> g2s2_and_id(&g2s2, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100277 GlobalHandles::MakeWeak(
278 g2s2.location(), reinterpret_cast<void*>(&g2s2_and_id),
279 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000280 std::pair<Handle<Object>*, int> g2c1_and_id(&g2c1, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100281 GlobalHandles::MakeWeak(
282 g2c1.location(), reinterpret_cast<void*>(&g2c1_and_id),
283 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000284
285 Handle<Object> root = global_handles->Create(*g1s1); // make a root.
286
287 // Connect group 1 and 2, make a cycle.
288 Handle<FixedArray>::cast(g1s2)->set(0, *g2s2);
289 Handle<FixedArray>::cast(g2s1)->set(0, *g1s1);
290
291 {
292 Object** g1_objects[] = { g1s1.location(), g1s2.location() };
293 Object** g2_objects[] = { g2s1.location(), g2s2.location() };
294 global_handles->AddObjectGroup(g1_objects, 2, NULL);
295 global_handles->SetReference(Handle<HeapObject>::cast(g1s1).location(),
296 g1c1.location());
297 global_handles->AddObjectGroup(g2_objects, 2, NULL);
298 global_handles->SetReference(Handle<HeapObject>::cast(g2s1).location(),
299 g2c1.location());
300 }
301 // Do a full GC
302 heap->CollectGarbage(OLD_SPACE);
303
304 // All object should be alive.
305 CHECK_EQ(0, NumberOfWeakCalls);
306
307 // Weaken the root.
308 std::pair<Handle<Object>*, int> root_and_id(&root, 1234);
Ben Murdochc5610432016-08-08 18:44:38 +0100309 GlobalHandles::MakeWeak(
310 root.location(), reinterpret_cast<void*>(&root_and_id),
311 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000312 // But make children strong roots---all the objects (except for children)
313 // should be collectable now.
314 global_handles->ClearWeakness(g1c1.location());
315 global_handles->ClearWeakness(g2c1.location());
316
317 // Groups are deleted, rebuild groups.
318 {
319 Object** g1_objects[] = { g1s1.location(), g1s2.location() };
320 Object** g2_objects[] = { g2s1.location(), g2s2.location() };
321 global_handles->AddObjectGroup(g1_objects, 2, NULL);
322 global_handles->SetReference(Handle<HeapObject>::cast(g1s1).location(),
323 g1c1.location());
324 global_handles->AddObjectGroup(g2_objects, 2, NULL);
325 global_handles->SetReference(Handle<HeapObject>::cast(g2s1).location(),
326 g2c1.location());
327 }
328
329 heap->CollectGarbage(OLD_SPACE);
330
331 // All objects should be gone. 5 global handles in total.
332 CHECK_EQ(5, NumberOfWeakCalls);
333
334 // And now make children weak again and collect them.
Ben Murdochc5610432016-08-08 18:44:38 +0100335 GlobalHandles::MakeWeak(
336 g1c1.location(), reinterpret_cast<void*>(&g1c1_and_id),
337 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
338 GlobalHandles::MakeWeak(
339 g2c1.location(), reinterpret_cast<void*>(&g2c1_and_id),
340 &WeakPointerCallback, v8::WeakCallbackType::kParameter);
Ben Murdoch4a90d5f2016-03-22 12:00:34 +0000341
342 heap->CollectGarbage(OLD_SPACE);
343 CHECK_EQ(7, NumberOfWeakCalls);
344}
345
346
347class TestRetainedObjectInfo : public v8::RetainedObjectInfo {
348 public:
349 TestRetainedObjectInfo() : has_been_disposed_(false) {}
350
351 bool has_been_disposed() { return has_been_disposed_; }
352
353 virtual void Dispose() {
354 CHECK(!has_been_disposed_);
355 has_been_disposed_ = true;
356 }
357
358 virtual bool IsEquivalent(v8::RetainedObjectInfo* other) {
359 return other == this;
360 }
361
362 virtual intptr_t GetHash() { return 0; }
363
364 virtual const char* GetLabel() { return "whatever"; }
365
366 private:
367 bool has_been_disposed_;
368};
369
370
371TEST(EmptyObjectGroups) {
372 CcTest::InitializeVM();
373 GlobalHandles* global_handles = CcTest::i_isolate()->global_handles();
374
375 v8::HandleScope handle_scope(CcTest::isolate());
376
377 TestRetainedObjectInfo info;
378 global_handles->AddObjectGroup(NULL, 0, &info);
379 CHECK(info.has_been_disposed());
380}
381
382
383#if defined(__has_feature)
384#if __has_feature(address_sanitizer)
385#define V8_WITH_ASAN 1
386#endif
387#endif
388
389
390// Here is a memory use test that uses /proc, and is therefore Linux-only. We
391// do not care how much memory the simulator uses, since it is only there for
392// debugging purposes. Testing with ASAN doesn't make sense, either.
393#if defined(__linux__) && !defined(USE_SIMULATOR) && !defined(V8_WITH_ASAN)
394
395
396static uintptr_t ReadLong(char* buffer, intptr_t* position, int base) {
397 char* end_address = buffer + *position;
398 uintptr_t result = strtoul(buffer + *position, &end_address, base);
399 CHECK(result != ULONG_MAX || errno != ERANGE);
400 CHECK(end_address > buffer + *position);
401 *position = end_address - buffer;
402 return result;
403}
404
405
406// The memory use computed this way is not entirely accurate and depends on
407// the way malloc allocates memory. That's why the memory use may seem to
408// increase even though the sum of the allocated object sizes decreases. It
409// also means that the memory use depends on the kernel and stdlib.
410static intptr_t MemoryInUse() {
411 intptr_t memory_use = 0;
412
413 int fd = open("/proc/self/maps", O_RDONLY);
414 if (fd < 0) return -1;
415
416 const int kBufSize = 10000;
417 char buffer[kBufSize];
418 ssize_t length = read(fd, buffer, kBufSize);
419 intptr_t line_start = 0;
420 CHECK_LT(length, kBufSize); // Make the buffer bigger.
421 CHECK_GT(length, 0); // We have to find some data in the file.
422 while (line_start < length) {
423 if (buffer[line_start] == '\n') {
424 line_start++;
425 continue;
426 }
427 intptr_t position = line_start;
428 uintptr_t start = ReadLong(buffer, &position, 16);
429 CHECK_EQ(buffer[position++], '-');
430 uintptr_t end = ReadLong(buffer, &position, 16);
431 CHECK_EQ(buffer[position++], ' ');
432 CHECK(buffer[position] == '-' || buffer[position] == 'r');
433 bool read_permission = (buffer[position++] == 'r');
434 CHECK(buffer[position] == '-' || buffer[position] == 'w');
435 bool write_permission = (buffer[position++] == 'w');
436 CHECK(buffer[position] == '-' || buffer[position] == 'x');
437 bool execute_permission = (buffer[position++] == 'x');
438 CHECK(buffer[position] == 's' || buffer[position] == 'p');
439 bool private_mapping = (buffer[position++] == 'p');
440 CHECK_EQ(buffer[position++], ' ');
441 uintptr_t offset = ReadLong(buffer, &position, 16);
442 USE(offset);
443 CHECK_EQ(buffer[position++], ' ');
444 uintptr_t major = ReadLong(buffer, &position, 16);
445 USE(major);
446 CHECK_EQ(buffer[position++], ':');
447 uintptr_t minor = ReadLong(buffer, &position, 16);
448 USE(minor);
449 CHECK_EQ(buffer[position++], ' ');
450 uintptr_t inode = ReadLong(buffer, &position, 10);
451 while (position < length && buffer[position] != '\n') position++;
452 if ((read_permission || write_permission || execute_permission) &&
453 private_mapping && inode == 0) {
454 memory_use += (end - start);
455 }
456
457 line_start = position;
458 }
459 close(fd);
460 return memory_use;
461}
462
463
464intptr_t ShortLivingIsolate() {
465 v8::Isolate::CreateParams create_params;
466 create_params.array_buffer_allocator = CcTest::array_buffer_allocator();
467 v8::Isolate* isolate = v8::Isolate::New(create_params);
468 { v8::Isolate::Scope isolate_scope(isolate);
469 v8::Locker lock(isolate);
470 v8::HandleScope handle_scope(isolate);
471 v8::Local<v8::Context> context = v8::Context::New(isolate);
472 CHECK(!context.IsEmpty());
473 }
474 isolate->Dispose();
475 return MemoryInUse();
476}
477
478
479TEST(RegressJoinThreadsOnIsolateDeinit) {
480 intptr_t size_limit = ShortLivingIsolate() * 2;
481 for (int i = 0; i < 10; i++) {
482 CHECK_GT(size_limit, ShortLivingIsolate());
483 }
484}
485
486#endif // __linux__ and !USE_SIMULATOR