blob: d4d43e740859362ab38a8cbf5f3aca694368a417 [file] [log] [blame]
Carl Shapiro1fb86202011-06-27 17:43:13 -07001// Copyright 2011 Google Inc. All Rights Reserved.
Carl Shapiro1fb86202011-06-27 17:43:13 -07002
3#ifndef ART_SRC_HEAP_H_
4#define ART_SRC_HEAP_H_
5
Carl Shapiro58551df2011-07-24 03:09:51 -07006#include <vector>
7
Brian Carlstrom578bbdc2011-07-21 14:07:47 -07008#include "globals.h"
Brian Carlstrom578bbdc2011-07-21 14:07:47 -07009#include "object_bitmap.h"
10#include "thread.h"
Carl Shapiro1fb86202011-06-27 17:43:13 -070011
12namespace art {
13
Brian Carlstroma40f9bc2011-07-26 21:26:07 -070014class Class;
15class Object;
Carl Shapiro69759ea2011-07-21 18:13:35 -070016class Space;
17class HeapBitmap;
18
Carl Shapiro1fb86202011-06-27 17:43:13 -070019class Heap {
20 public:
Brian Carlstrom8a436592011-08-15 21:27:23 -070021 static const size_t kInitialSize = 16 * MB;
Carl Shapiro69759ea2011-07-21 18:13:35 -070022
Brian Carlstrom4a96b602011-07-26 16:40:23 -070023 static const size_t kMaximumSize = 64 * MB;
Carl Shapiro69759ea2011-07-21 18:13:35 -070024
Brian Carlstrom7e93b502011-08-04 14:16:22 -070025 typedef void (RootVistor)(Object* root, void* arg);
26
Brian Carlstrom4a289ed2011-08-16 17:17:49 -070027 // Create a heap with the requested sizes. optional boot image may
28 // be NULL, otherwise it is an image filename created by ImageWriter.
29 static bool Init(size_t starting_size, size_t maximum_size, const char* boot_image_file_name);
Carl Shapiro61e019d2011-07-14 16:53:09 -070030
Carl Shapiro69759ea2011-07-21 18:13:35 -070031 static void Destroy();
Brian Carlstroma7f4f482011-07-17 17:01:34 -070032
Brian Carlstroma40f9bc2011-07-26 21:26:07 -070033 // Allocates and initializes storage for an object instance.
Carl Shapiro58551df2011-07-24 03:09:51 -070034 static Object* AllocObject(Class* klass, size_t num_bytes);
Brian Carlstroma7f4f482011-07-17 17:01:34 -070035
Elliott Hughesa2501992011-08-26 19:39:54 -070036 // Check sanity of given reference. Requires the heap lock.
Ian Rogers408f79a2011-08-23 18:22:33 -070037 static void VerifyObject(Object *obj);
38
Elliott Hughesa2501992011-08-26 19:39:54 -070039 // A weaker test than VerifyObject that doesn't require the heap lock,
40 // and doesn't abort on error, allowing the caller to report more
41 // meaningful diagnostics.
42 static bool IsHeapAddress(Object* obj);
43
Carl Shapiro69759ea2011-07-21 18:13:35 -070044 // Initiates an explicit garbage collection.
45 static void CollectGarbage();
46
47 // Blocks the caller until the garbage collector becomes idle.
48 static void WaitForConcurrentGcToComplete();
49
50 static Mutex* GetLock() {
51 return lock_;
Carl Shapiro5fafe2b2011-07-09 15:34:41 -070052 }
Carl Shapiro61e019d2011-07-14 16:53:09 -070053
Carl Shapiro58551df2011-07-24 03:09:51 -070054 static const std::vector<Space*>& GetSpaces() {
55 return spaces_;
56 }
Carl Shapiro61e019d2011-07-14 16:53:09 -070057
Brian Carlstroma663ea52011-08-19 23:33:41 -070058 static Space* GetBootSpace() {
59 return boot_space_;
60 }
61
Carl Shapiro58551df2011-07-24 03:09:51 -070062 static HeapBitmap* GetLiveBits() {
63 return live_bitmap_;
Carl Shapiro744ad052011-08-06 15:53:36 -070064 }
Carl Shapiro58551df2011-07-24 03:09:51 -070065
66 static HeapBitmap* GetMarkBits() {
67 return mark_bitmap_;
Carl Shapiro744ad052011-08-06 15:53:36 -070068 }
Carl Shapiro58551df2011-07-24 03:09:51 -070069
Brian Carlstrom1f870082011-08-23 16:02:11 -070070 static void SetReferenceOffsets(size_t reference_referent_offset,
71 size_t reference_queue_offset,
72 size_t reference_queueNext_offset,
73 size_t reference_pendingNext_offset,
74 size_t finalizer_reference_zombie_offset) {
75 CHECK_NE(reference_referent_offset, 0U);
76 CHECK_NE(reference_queue_offset, 0U);
77 CHECK_NE(reference_queueNext_offset, 0U);
78 CHECK_NE(reference_pendingNext_offset, 0U);
79 CHECK_NE(finalizer_reference_zombie_offset, 0U);
80 reference_referent_offset_ = reference_referent_offset;
81 reference_queue_offset_ = reference_queue_offset;
82 reference_queueNext_offset_ = reference_queueNext_offset;
83 reference_pendingNext_offset_ = reference_pendingNext_offset;
84 finalizer_reference_zombie_offset_ = finalizer_reference_zombie_offset;
85 }
86
87 static size_t GetReferenceReferentOffset() {
88 DCHECK_NE(reference_referent_offset_, 0U);
89 return reference_referent_offset_;
90 }
91
92 static size_t GetReferenceQueueOffset() {
93 DCHECK_NE(reference_queue_offset_, 0U);
94 return reference_queue_offset_;
95 }
96
97 static size_t GetReferenceQueueNextOffset() {
98 DCHECK_NE(reference_queueNext_offset_, 0U);
99 return reference_queueNext_offset_;
100 }
101
102 static size_t GetReferencePendingNextOffset() {
103 DCHECK_NE(reference_pendingNext_offset_, 0U);
104 return reference_pendingNext_offset_;
105 }
106
107 static size_t GetFinalizerReferenceZombieOffset() {
108 DCHECK_NE(finalizer_reference_zombie_offset_, 0U);
109 return finalizer_reference_zombie_offset_;
110 }
111
Carl Shapiro58551df2011-07-24 03:09:51 -0700112 private:
113 // Allocates uninitialized storage.
Carl Shapiro69759ea2011-07-21 18:13:35 -0700114 static Object* Allocate(size_t num_bytes);
Carl Shapiro58551df2011-07-24 03:09:51 -0700115 static Object* Allocate(Space* space, size_t num_bytes);
116
117 static void RecordAllocation(Space* space, const Object* object);
118 static void RecordFree(Space* space, const Object* object);
Brian Carlstrom9cff8e12011-08-18 16:47:29 -0700119 static void RecordImageAllocations(Space* space);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700120
121 static void CollectGarbageInternal();
122
123 static void GrowForUtilization();
124
125 static Mutex* lock_;
126
Carl Shapiro58551df2011-07-24 03:09:51 -0700127 static std::vector<Space*> spaces_;
Carl Shapiro69759ea2011-07-21 18:13:35 -0700128
Brian Carlstroma663ea52011-08-19 23:33:41 -0700129 // Space loaded from an image
130 static Space* boot_space_;
131
Brian Carlstrom4a289ed2011-08-16 17:17:49 -0700132 // default Space for allocations
133 static Space* alloc_space_;
134
Carl Shapiro69759ea2011-07-21 18:13:35 -0700135 static HeapBitmap* mark_bitmap_;
136
137 static HeapBitmap* live_bitmap_;
138
Carl Shapiro58551df2011-07-24 03:09:51 -0700139 // The maximum size of the heap in bytes.
Carl Shapiro69759ea2011-07-21 18:13:35 -0700140 static size_t maximum_size_;
141
Carl Shapiro58551df2011-07-24 03:09:51 -0700142 // True while the garbage collector is running.
Carl Shapiro69759ea2011-07-21 18:13:35 -0700143 static bool is_gc_running_;
144
Carl Shapiro58551df2011-07-24 03:09:51 -0700145 // Number of bytes allocated. Adjusted after each allocation and
146 // free.
147 static size_t num_bytes_allocated_;
148
149 // Number of objects allocated. Adjusted after each allocation and
150 // free.
151 static size_t num_objects_allocated_;
152
Brian Carlstrom1f870082011-08-23 16:02:11 -0700153 // offset of java.lang.ref.Reference.referent
154 static size_t reference_referent_offset_;
155
156 // offset of java.lang.ref.Reference.queue
157 static size_t reference_queue_offset_;
158
159 // offset of java.lang.ref.Reference.queueNext
160 static size_t reference_queueNext_offset_;
161
162 // offset of java.lang.ref.Reference.pendingNext
163 static size_t reference_pendingNext_offset_;
164
165 // offset of java.lang.ref.FinalizerReference.zombie
166 static size_t finalizer_reference_zombie_offset_;
167
Carl Shapiro69759ea2011-07-21 18:13:35 -0700168 DISALLOW_IMPLICIT_CONSTRUCTORS(Heap);
169};
170
171class HeapLock {
172 public:
173 HeapLock(Heap* heap) : lock_(heap->GetLock()) {
174 lock_->Lock();
175 }
176 ~HeapLock() {
177 lock_->Unlock();
178 }
179 private:
180 Mutex* lock_;
181 DISALLOW_COPY_AND_ASSIGN(HeapLock);
Carl Shapiro1fb86202011-06-27 17:43:13 -0700182};
183
184} // namespace art
185
186#endif // ART_SRC_HEAP_H_