blob: 76c5428299ae48cd6b6960936547674ceff70bfa [file] [log] [blame]
Elliott Hughes2faa5f12012-01-30 14:42:07 -08001/*
2 * Copyright (C) 2011 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
Carl Shapiro69759ea2011-07-21 18:13:35 -070016
17#ifndef ART_SRC_MARK_SWEEP_H_
18#define ART_SRC_MARK_SWEEP_H_
19
Mathieu Chartierd8195f12012-10-05 12:21:28 -070020#include "atomic_stack.h"
Brian Carlstrom578bbdc2011-07-21 14:07:47 -070021#include "macros.h"
Elliott Hughes5e71b522011-10-20 13:12:32 -070022#include "heap_bitmap.h"
Mathieu Chartierb43b7d42012-06-19 13:15:09 -070023#include "object.h"
Ian Rogers0cfe1fb2011-08-26 03:29:44 -070024#include "offsets.h"
Carl Shapiro69759ea2011-07-21 18:13:35 -070025
26namespace art {
27
Mathieu Chartierb43b7d42012-06-19 13:15:09 -070028class CheckObjectVisitor;
Carl Shapiro69759ea2011-07-21 18:13:35 -070029class Class;
Elliott Hughesb3bd5f02012-03-08 21:05:27 -080030class Heap;
Mathieu Chartierb43b7d42012-06-19 13:15:09 -070031class MarkIfReachesAllocspaceVisitor;
32class ModUnionClearCardVisitor;
33class ModUnionVisitor;
34class ModUnionTableBitmap;
Carl Shapiro69759ea2011-07-21 18:13:35 -070035class Object;
Mathieu Chartier357e9be2012-08-01 11:00:14 -070036class TimingLogger;
Carl Shapiro69759ea2011-07-21 18:13:35 -070037
38class MarkSweep {
39 public:
Mathieu Chartierd8195f12012-10-05 12:21:28 -070040 explicit MarkSweep(ObjectStack* mark_stack);
Carl Shapiro58551df2011-07-24 03:09:51 -070041
Carl Shapiro69759ea2011-07-21 18:13:35 -070042 ~MarkSweep();
43
Carl Shapiro58551df2011-07-24 03:09:51 -070044 // Initializes internal structures.
Jesse Wilson078f9b02011-11-18 17:51:47 -050045 void Init();
Carl Shapiro58551df2011-07-24 03:09:51 -070046
Mathieu Chartier7469ebf2012-09-24 16:28:36 -070047 // Find the default mark bitmap.
48 void FindDefaultMarkBitmap();
49
Carl Shapiro69759ea2011-07-21 18:13:35 -070050 // Marks the root set at the start of a garbage collection.
Ian Rogers00f7d0e2012-07-19 15:28:27 -070051 void MarkRoots()
Ian Rogersb726dcb2012-09-05 08:57:23 -070052 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
53 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -070054
Mathieu Chartier262e5ff2012-06-01 17:35:38 -070055 // Verify that image roots point to only marked objects within the alloc space.
Ian Rogersb726dcb2012-09-05 08:57:23 -070056 void VerifyImageRoots() EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Mathieu Chartier262e5ff2012-06-01 17:35:38 -070057
Carl Shapiro58551df2011-07-24 03:09:51 -070058 // Builds a mark stack and recursively mark until it empties.
Mathieu Chartier357e9be2012-08-01 11:00:14 -070059 void RecursiveMark(bool partial, TimingLogger& timings)
Ian Rogersb726dcb2012-09-05 08:57:23 -070060 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
61 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Mathieu Chartiercc236d72012-07-20 10:29:05 -070062
Mathieu Chartier357e9be2012-08-01 11:00:14 -070063 // Copies mark bits from live bitmap of ZygoteSpace to mark bitmap for partial GCs.
Mathieu Chartier2fde5332012-09-14 14:51:54 -070064 void CopyMarkBits(ContinuousSpace* space);
Mathieu Chartier7469ebf2012-09-24 16:28:36 -070065 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
66
67 void BindLiveToMarkBitmap(ContinuousSpace* space)
68 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
69
70 void UnBindBitmaps()
71 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro58551df2011-07-24 03:09:51 -070072
Mathieu Chartier262e5ff2012-06-01 17:35:38 -070073 // Builds a mark stack with objects on dirty cards and recursively mark
74 // until it empties.
Mathieu Chartier357e9be2012-08-01 11:00:14 -070075 void RecursiveMarkDirtyObjects(bool update_finger)
Ian Rogersb726dcb2012-09-05 08:57:23 -070076 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
77 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Mathieu Chartier262e5ff2012-06-01 17:35:38 -070078
Mathieu Chartier357e9be2012-08-01 11:00:14 -070079 // Recursive mark objects on specified cards. Updates finger.
80 void RecursiveMarkCards(CardTable* card_table, const std::vector<byte*>& cards,
81 TimingLogger& timings)
Ian Rogersb726dcb2012-09-05 08:57:23 -070082 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
83 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);;
Mathieu Chartier357e9be2012-08-01 11:00:14 -070084
Carl Shapiro69759ea2011-07-21 18:13:35 -070085 // Remarks the root set after completing the concurrent mark.
Ian Rogers00f7d0e2012-07-19 15:28:27 -070086 void ReMarkRoots()
Ian Rogersb726dcb2012-09-05 08:57:23 -070087 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
88 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -070089
Mathieu Chartiercc236d72012-07-20 10:29:05 -070090 Heap* GetHeap() {
91 return heap_;
92 }
93
Ian Rogers00f7d0e2012-07-19 15:28:27 -070094 void ProcessReferences(bool clear_soft_references)
Ian Rogersb726dcb2012-09-05 08:57:23 -070095 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
96 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
Carl Shapiro58551df2011-07-24 03:09:51 -070097 ProcessReferences(&soft_reference_list_, clear_soft_references,
98 &weak_reference_list_,
99 &finalizer_reference_list_,
100 &phantom_reference_list_);
101 }
102
Carl Shapiro69759ea2011-07-21 18:13:35 -0700103 // Sweeps unmarked objects to complete the garbage collection.
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700104 void Sweep(bool partial, bool swap_bitmaps)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700105 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700106
Mathieu Chartiere0f0cb32012-08-28 11:26:00 -0700107 // Sweeps unmarked objects to complete the garbage collection.
108 void SweepLargeObjects(bool swap_bitmaps)
109 EXCLUSIVE_LOCKS_REQUIRED(GlobalSynchronization::heap_bitmap_lock_);
110
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700111 // Sweep only pointers within an array. WARNING: Trashes objects.
Mathieu Chartierd8195f12012-10-05 12:21:28 -0700112 void SweepArray(TimingLogger& logger, ObjectStack* allocation_stack_, bool swap_bitmaps)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700113 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700114
Elliott Hughesadb460d2011-10-05 17:02:34 -0700115 Object* GetClearedReferences() {
116 return cleared_reference_list_;
117 }
118
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700119 // Proxy for external access to ScanObject.
120 void ScanRoot(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700121 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
122 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700123
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700124 // Blackens an object.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700125 void ScanObject(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700126 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
127 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700128
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700129 void SetFinger(Object* new_finger) {
130 finger_ = new_finger;
131 }
132
133 void DisableFinger() {
134 SetFinger(reinterpret_cast<Object*>(~static_cast<uintptr_t>(0)));
135 }
136
137 size_t GetFreedBytes() const {
138 return freed_bytes_;
139 }
140
141 size_t GetFreedObjects() const {
142 return freed_objects_;
143 }
144
Mathieu Chartiere0f0cb32012-08-28 11:26:00 -0700145 // Everything inside the immune range is marked.
146 void SetImmuneRange(Object* begin, Object* end) {
147 immune_begin_ = begin;
148 immune_end_ = end;
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700149 }
150
Mathieu Chartier7469ebf2012-09-24 16:28:36 -0700151 void SweepSystemWeaks()
152 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
153
154 // Only sweep the weaks which are inside of an allocation stack.
155 void SweepSystemWeaksArray(ObjectStack* allocations)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700156 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700157
Mathieu Chartierc7b83a02012-09-11 18:07:39 -0700158 static bool VerifyIsLiveCallback(const Object* obj, void* arg)
159 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
160
161 void VerifySystemWeaks()
162 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
163
164 // Verify that an object is live, either in a live bitmap or in the allocation stack.
165 void VerifyIsLive(const Object* obj)
166 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
167
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700168 template <typename Visitor>
169 static void VisitObjectReferences(const Object* obj, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700170 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_,
171 Locks::mutator_lock_) {
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700172 DCHECK(obj != NULL);
173 DCHECK(obj->GetClass() != NULL);
174 if (obj->IsClass()) {
175 VisitClassReferences(obj, visitor);
176 } else if (obj->IsArrayInstance()) {
177 VisitArrayReferences(obj, visitor);
178 } else {
179 VisitOtherReferences(obj, visitor);
180 }
181 }
182
Carl Shapiro69759ea2011-07-21 18:13:35 -0700183 private:
184 // Returns true if the object has its bit set in the mark bitmap.
Mathieu Chartier7469ebf2012-09-24 16:28:36 -0700185 bool IsMarked(const Object* object) const;
Carl Shapiro69759ea2011-07-21 18:13:35 -0700186
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700187 static bool IsMarkedCallback(const Object* object, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700188 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Elliott Hughesc33a32b2011-10-11 18:18:07 -0700189
Mathieu Chartier7469ebf2012-09-24 16:28:36 -0700190 static bool IsMarkedArrayCallback(const Object* object, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700191 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Mathieu Chartier46a23632012-08-07 18:44:40 -0700192
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700193 static void MarkObjectVisitor(const Object* root, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700194 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Brian Carlstrom1f870082011-08-23 16:02:11 -0700195
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700196 static void ReMarkObjectVisitor(const Object* root, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700197 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Mathieu Chartier262e5ff2012-06-01 17:35:38 -0700198
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700199 static void VerifyImageRootVisitor(Object* root, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700200 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_,
201 Locks::mutator_lock_);
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700202
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700203 static void ScanDirtyCardCallback(Object* obj, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700204 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
205 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Mathieu Chartier262e5ff2012-06-01 17:35:38 -0700206
Carl Shapiro69759ea2011-07-21 18:13:35 -0700207 // Marks an object.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700208 void MarkObject(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700209 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700210
211 // Yuck.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700212 void MarkObject0(const Object* obj, bool check_finger)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700213 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700214
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700215 static void ScanBitmapCallback(Object* obj, void* finger, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700216 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
217 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro58551df2011-07-24 03:09:51 -0700218
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700219 static void SweepCallback(size_t num_ptrs, Object** ptrs, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700220 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro58551df2011-07-24 03:09:51 -0700221
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700222 // Special sweep for zygote that just marks objects / dirties cards.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700223 static void ZygoteSweepCallback(size_t num_ptrs, Object** ptrs, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700224 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Ian Rogers5d76c432011-10-31 21:42:49 -0700225
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700226 void CheckReference(const Object* obj, const Object* ref, MemberOffset offset, bool is_static)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700227 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700228
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700229 void CheckObject(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700230 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_);
Ian Rogers5d76c432011-10-31 21:42:49 -0700231
Carl Shapiro69759ea2011-07-21 18:13:35 -0700232 // Grays references in instance fields.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700233 void ScanInstanceFields(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700234 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
235 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700236
Mathieu Chartier6f1c9492012-10-15 12:08:41 -0700237 // Verify the roots of the heap and print out information related to any invalid roots.
238 // Called in MarkObject, so may we may not hold the mutator lock.
239 void VerifyRoots()
240 NO_THREAD_SAFETY_ANALYSIS;
241
242 static void VerifyRootCallback(const Object* root, void* arg, size_t vreg,
243 const AbstractMethod* method);
244
245 void VerifyRoot(const Object* root, size_t vreg, const AbstractMethod* method)
246 NO_THREAD_SAFETY_ANALYSIS;
247
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700248 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700249 static void VisitInstanceFieldsReferences(const Object* obj, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700250 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700251 DCHECK(obj != NULL);
252 Class* klass = obj->GetClass();
253 DCHECK(klass != NULL);
254 VisitFieldsReferences(obj, klass->GetReferenceInstanceOffsets(), false, visitor);
255 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700256
Carl Shapiro69759ea2011-07-21 18:13:35 -0700257 // Blackens a class object.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700258 void ScanClass(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700259 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
260 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700261
Carl Shapiro69759ea2011-07-21 18:13:35 -0700262
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700263 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700264 static void VisitClassReferences(const Object* obj, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700265 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700266 VisitInstanceFieldsReferences(obj, visitor);
267 VisitStaticFieldsReferences(obj->AsClass(), visitor);
268 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700269
Carl Shapiro69759ea2011-07-21 18:13:35 -0700270 // Grays references in static fields.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700271 void ScanStaticFields(const Class* klass)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700272 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
273 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700274
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700275 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700276 static void VisitStaticFieldsReferences(const Class* klass, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700277 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700278 DCHECK(klass != NULL);
279 VisitFieldsReferences(klass, klass->GetReferenceStaticOffsets(), true, visitor);
280 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700281
Brian Carlstrom4873d462011-08-21 15:23:39 -0700282 // Used by ScanInstanceFields and ScanStaticFields
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700283 void ScanFields(const Object* obj, uint32_t ref_offsets, bool is_static)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700284 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
285 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Brian Carlstrom4873d462011-08-21 15:23:39 -0700286
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700287 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700288 static void VisitFieldsReferences(const Object* obj, uint32_t ref_offsets, bool is_static,
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700289 const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700290 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_,
291 Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700292 if (ref_offsets != CLASS_WALK_SUPER) {
293 // Found a reference offset bitmap. Mark the specified offsets.
294 while (ref_offsets != 0) {
295 size_t right_shift = CLZ(ref_offsets);
296 MemberOffset field_offset = CLASS_OFFSET_FROM_CLZ(right_shift);
297 const Object* ref = obj->GetFieldObject<const Object*>(field_offset, false);
298 visitor(obj, ref, field_offset, is_static);
299 ref_offsets &= ~(CLASS_HIGH_BIT >> right_shift);
300 }
301 } else {
302 // There is no reference offset bitmap. In the non-static case,
303 // walk up the class inheritance hierarchy and find reference
304 // offsets the hard way. In the static case, just consider this
305 // class.
306 for (const Class* klass = is_static ? obj->AsClass() : obj->GetClass();
307 klass != NULL;
308 klass = is_static ? NULL : klass->GetSuperClass()) {
309 size_t num_reference_fields = (is_static
310 ? klass->NumReferenceStaticFields()
311 : klass->NumReferenceInstanceFields());
312 for (size_t i = 0; i < num_reference_fields; ++i) {
313 Field* field = (is_static
314 ? klass->GetStaticField(i)
315 : klass->GetInstanceField(i));
316 MemberOffset field_offset = field->GetOffset();
317 const Object* ref = obj->GetFieldObject<const Object*>(field_offset, false);
318 visitor(obj, ref, field_offset, is_static);
319 }
320 }
321 }
322 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700323
Carl Shapiro69759ea2011-07-21 18:13:35 -0700324 // Grays references in an array.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700325 void ScanArray(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700326 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
327 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700328
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700329 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700330 static void VisitArrayReferences(const Object* obj, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700331 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700332 visitor(obj, obj->GetClass(), Object::ClassOffset(), false);
333 if (obj->IsObjectArray()) {
334 const ObjectArray<Object>* array = obj->AsObjectArray<Object>();
335 for (int32_t i = 0; i < array->GetLength(); ++i) {
336 const Object* element = array->GetWithoutChecks(i);
337 size_t width = sizeof(Object*);
338 visitor(obj, element, MemberOffset(i * width + Array::DataOffset(width).Int32Value()), false);
339 }
340 }
341 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700342
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700343 void ScanOther(const Object* obj)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700344 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
345 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700346
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700347 template <typename Visitor>
Mathieu Chartierfd678be2012-08-30 14:50:54 -0700348 static void VisitOtherReferences(const Object* obj, const Visitor& visitor)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700349 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_) {
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700350 return VisitInstanceFieldsReferences(obj, visitor);
351 }
Ian Rogers5d76c432011-10-31 21:42:49 -0700352
Carl Shapiro69759ea2011-07-21 18:13:35 -0700353 // Blackens objects grayed during a garbage collection.
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700354 void ScanGrayObjects(bool update_finger)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700355 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700356
357 // Schedules an unmarked object for reference processing.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700358 void DelayReferenceReferent(Object* reference)
Ian Rogers23435d02012-09-24 11:23:12 -0700359 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700360
361 // Recursively blackens objects on the mark stack.
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700362 void ProcessMarkStack()
Ian Rogersb726dcb2012-09-05 08:57:23 -0700363 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
364 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700365
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700366 void EnqueueFinalizerReferences(Object** ref)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700367 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
368 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700369
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700370 void PreserveSomeSoftReferences(Object** ref)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700371 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
372 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700373
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700374 void ClearWhiteReferences(Object** list)
Ian Rogers23435d02012-09-24 11:23:12 -0700375 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_, Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700376
Carl Shapiro58551df2011-07-24 03:09:51 -0700377 void ProcessReferences(Object** soft_references, bool clear_soft_references,
Carl Shapiro69759ea2011-07-21 18:13:35 -0700378 Object** weak_references,
379 Object** finalizer_references,
Ian Rogers00f7d0e2012-07-19 15:28:27 -0700380 Object** phantom_references)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700381 EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_)
382 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
Carl Shapiro69759ea2011-07-21 18:13:35 -0700383
Mathieu Chartier7469ebf2012-09-24 16:28:36 -0700384 void SweepJniWeakGlobals(Heap::IsMarkedTester is_marked, void* arg)
Ian Rogersb726dcb2012-09-05 08:57:23 -0700385 SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_);
Carl Shapiro58551df2011-07-24 03:09:51 -0700386
Mathieu Chartierb062fdd2012-07-03 09:51:48 -0700387 // Current space, we check this space first to avoid searching for the appropriate space for an object.
388 SpaceBitmap* current_mark_bitmap_;
389
Mathieu Chartierd8195f12012-10-05 12:21:28 -0700390 ObjectStack* mark_stack_;
Carl Shapiro69759ea2011-07-21 18:13:35 -0700391
Elliott Hughesb3bd5f02012-03-08 21:05:27 -0800392 Heap* heap_;
Carl Shapiro69759ea2011-07-21 18:13:35 -0700393
394 Object* finger_;
395
Mathieu Chartiere0f0cb32012-08-28 11:26:00 -0700396 // Immune range, every object inside the immune range is assumed to be marked.
397 Object* immune_begin_;
398 Object* immune_end_;
Carl Shapiro69759ea2011-07-21 18:13:35 -0700399
400 Object* soft_reference_list_;
401
402 Object* weak_reference_list_;
403
404 Object* finalizer_reference_list_;
405
406 Object* phantom_reference_list_;
407
408 Object* cleared_reference_list_;
409
Mathieu Chartier357e9be2012-08-01 11:00:14 -0700410 size_t freed_bytes_;
411 size_t freed_objects_;
412
Elliott Hughes352a4242011-10-31 15:15:21 -0700413 size_t class_count_;
414 size_t array_count_;
415 size_t other_count_;
416
Mathieu Chartiere6e06512012-06-26 15:00:26 -0700417 friend class AddIfReachesAllocSpaceVisitor; // Used by mod-union table.
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700418 friend class CheckBitmapVisitor;
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700419 friend class CheckObjectVisitor;
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700420 friend class CheckReferenceVisitor;
Elliott Hughes410c0c82011-09-01 17:58:25 -0700421 friend class InternTableEntryIsUnmarked;
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700422 friend class MarkIfReachesAllocspaceVisitor;
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700423 friend class ModUnionCheckReferences;
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700424 friend class ModUnionClearCardVisitor;
Mathieu Chartiere6e06512012-06-26 15:00:26 -0700425 friend class ModUnionReferenceVisitor;
Mathieu Chartierb43b7d42012-06-19 13:15:09 -0700426 friend class ModUnionVisitor;
427 friend class ModUnionTableBitmap;
Mathieu Chartiere6e06512012-06-26 15:00:26 -0700428 friend class ModUnionTableReferenceCache;
Mathieu Chartiercc236d72012-07-20 10:29:05 -0700429 friend class ModUnionScanImageRootVisitor;
430 friend class ScanBitmapVisitor;
431 friend class ScanImageRootVisitor;
Elliott Hughes410c0c82011-09-01 17:58:25 -0700432
Carl Shapiro69759ea2011-07-21 18:13:35 -0700433 DISALLOW_COPY_AND_ASSIGN(MarkSweep);
434};
435
436} // namespace art
437
438#endif // ART_SRC_MARK_SWEEP_H_