| Ben Murdoch | b8a8cc1 | 2014-11-26 15:28:44 +0000 | [diff] [blame] | 1 | // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 | // Use of this source code is governed by a BSD-style license that can be |
| 3 | // found in the LICENSE file. |
| 4 | |
| 5 | #ifndef V8_OBJECTS_VISITING_H_ |
| 6 | #define V8_OBJECTS_VISITING_H_ |
| 7 | |
| 8 | #include "src/allocation.h" |
| 9 | |
| 10 | // This file provides base classes and auxiliary methods for defining |
| 11 | // static object visitors used during GC. |
| 12 | // Visiting HeapObject body with a normal ObjectVisitor requires performing |
| 13 | // two switches on object's instance type to determine object size and layout |
| 14 | // and one or more virtual method calls on visitor itself. |
| 15 | // Static visitor is different: it provides a dispatch table which contains |
| 16 | // pointers to specialized visit functions. Each map has the visitor_id |
| 17 | // field which contains an index of specialized visitor to use. |
| 18 | |
| 19 | namespace v8 { |
| 20 | namespace internal { |
| 21 | |
| 22 | |
| 23 | // Base class for all static visitors. |
| 24 | class StaticVisitorBase : public AllStatic { |
| 25 | public: |
| 26 | #define VISITOR_ID_LIST(V) \ |
| 27 | V(SeqOneByteString) \ |
| 28 | V(SeqTwoByteString) \ |
| 29 | V(ShortcutCandidate) \ |
| 30 | V(ByteArray) \ |
| 31 | V(FreeSpace) \ |
| 32 | V(FixedArray) \ |
| 33 | V(FixedDoubleArray) \ |
| 34 | V(FixedTypedArray) \ |
| 35 | V(FixedFloat64Array) \ |
| 36 | V(ConstantPoolArray) \ |
| 37 | V(NativeContext) \ |
| 38 | V(AllocationSite) \ |
| 39 | V(DataObject2) \ |
| 40 | V(DataObject3) \ |
| 41 | V(DataObject4) \ |
| 42 | V(DataObject5) \ |
| 43 | V(DataObject6) \ |
| 44 | V(DataObject7) \ |
| 45 | V(DataObject8) \ |
| 46 | V(DataObject9) \ |
| 47 | V(DataObjectGeneric) \ |
| 48 | V(JSObject2) \ |
| 49 | V(JSObject3) \ |
| 50 | V(JSObject4) \ |
| 51 | V(JSObject5) \ |
| 52 | V(JSObject6) \ |
| 53 | V(JSObject7) \ |
| 54 | V(JSObject8) \ |
| 55 | V(JSObject9) \ |
| 56 | V(JSObjectGeneric) \ |
| 57 | V(Struct2) \ |
| 58 | V(Struct3) \ |
| 59 | V(Struct4) \ |
| 60 | V(Struct5) \ |
| 61 | V(Struct6) \ |
| 62 | V(Struct7) \ |
| 63 | V(Struct8) \ |
| 64 | V(Struct9) \ |
| 65 | V(StructGeneric) \ |
| 66 | V(ConsString) \ |
| 67 | V(SlicedString) \ |
| 68 | V(Symbol) \ |
| 69 | V(Oddball) \ |
| 70 | V(Code) \ |
| 71 | V(Map) \ |
| 72 | V(Cell) \ |
| 73 | V(PropertyCell) \ |
| 74 | V(SharedFunctionInfo) \ |
| 75 | V(JSFunction) \ |
| 76 | V(JSWeakCollection) \ |
| 77 | V(JSArrayBuffer) \ |
| 78 | V(JSTypedArray) \ |
| 79 | V(JSDataView) \ |
| 80 | V(JSRegExp) |
| 81 | |
| 82 | // For data objects, JS objects and structs along with generic visitor which |
| 83 | // can visit object of any size we provide visitors specialized by |
| 84 | // object size in words. |
| 85 | // Ids of specialized visitors are declared in a linear order (without |
| 86 | // holes) starting from the id of visitor specialized for 2 words objects |
| 87 | // (base visitor id) and ending with the id of generic visitor. |
| 88 | // Method GetVisitorIdForSize depends on this ordering to calculate visitor |
| 89 | // id of specialized visitor from given instance size, base visitor id and |
| 90 | // generic visitor's id. |
| 91 | enum VisitorId { |
| 92 | #define VISITOR_ID_ENUM_DECL(id) kVisit##id, |
| 93 | VISITOR_ID_LIST(VISITOR_ID_ENUM_DECL) |
| 94 | #undef VISITOR_ID_ENUM_DECL |
| 95 | kVisitorIdCount, |
| 96 | kVisitDataObject = kVisitDataObject2, |
| 97 | kVisitJSObject = kVisitJSObject2, |
| 98 | kVisitStruct = kVisitStruct2, |
| 99 | kMinObjectSizeInWords = 2 |
| 100 | }; |
| 101 | |
| 102 | // Visitor ID should fit in one byte. |
| 103 | STATIC_ASSERT(kVisitorIdCount <= 256); |
| 104 | |
| 105 | // Determine which specialized visitor should be used for given instance type |
| 106 | // and instance type. |
| 107 | static VisitorId GetVisitorId(int instance_type, int instance_size); |
| 108 | |
| 109 | static VisitorId GetVisitorId(Map* map) { |
| 110 | return GetVisitorId(map->instance_type(), map->instance_size()); |
| 111 | } |
| 112 | |
| 113 | // For visitors that allow specialization by size calculate VisitorId based |
| 114 | // on size, base visitor id and generic visitor id. |
| 115 | static VisitorId GetVisitorIdForSize(VisitorId base, VisitorId generic, |
| 116 | int object_size) { |
| 117 | DCHECK((base == kVisitDataObject) || (base == kVisitStruct) || |
| 118 | (base == kVisitJSObject)); |
| 119 | DCHECK(IsAligned(object_size, kPointerSize)); |
| 120 | DCHECK(kMinObjectSizeInWords * kPointerSize <= object_size); |
| 121 | DCHECK(object_size <= Page::kMaxRegularHeapObjectSize); |
| 122 | |
| 123 | const VisitorId specialization = static_cast<VisitorId>( |
| 124 | base + (object_size >> kPointerSizeLog2) - kMinObjectSizeInWords); |
| 125 | |
| 126 | return Min(specialization, generic); |
| 127 | } |
| 128 | }; |
| 129 | |
| 130 | |
| 131 | template <typename Callback> |
| 132 | class VisitorDispatchTable { |
| 133 | public: |
| 134 | void CopyFrom(VisitorDispatchTable* other) { |
| 135 | // We are not using memcpy to guarantee that during update |
| 136 | // every element of callbacks_ array will remain correct |
| 137 | // pointer (memcpy might be implemented as a byte copying loop). |
| 138 | for (int i = 0; i < StaticVisitorBase::kVisitorIdCount; i++) { |
| 139 | base::NoBarrier_Store(&callbacks_[i], other->callbacks_[i]); |
| 140 | } |
| 141 | } |
| 142 | |
| 143 | inline Callback GetVisitorById(StaticVisitorBase::VisitorId id) { |
| 144 | return reinterpret_cast<Callback>(callbacks_[id]); |
| 145 | } |
| 146 | |
| 147 | inline Callback GetVisitor(Map* map) { |
| 148 | return reinterpret_cast<Callback>(callbacks_[map->visitor_id()]); |
| 149 | } |
| 150 | |
| 151 | void Register(StaticVisitorBase::VisitorId id, Callback callback) { |
| 152 | DCHECK(id < StaticVisitorBase::kVisitorIdCount); // id is unsigned. |
| 153 | callbacks_[id] = reinterpret_cast<base::AtomicWord>(callback); |
| 154 | } |
| 155 | |
| 156 | template <typename Visitor, StaticVisitorBase::VisitorId base, |
| 157 | StaticVisitorBase::VisitorId generic, int object_size_in_words> |
| 158 | void RegisterSpecialization() { |
| 159 | static const int size = object_size_in_words * kPointerSize; |
| 160 | Register(StaticVisitorBase::GetVisitorIdForSize(base, generic, size), |
| 161 | &Visitor::template VisitSpecialized<size>); |
| 162 | } |
| 163 | |
| 164 | |
| 165 | template <typename Visitor, StaticVisitorBase::VisitorId base, |
| 166 | StaticVisitorBase::VisitorId generic> |
| 167 | void RegisterSpecializations() { |
| 168 | STATIC_ASSERT((generic - base + StaticVisitorBase::kMinObjectSizeInWords) == |
| 169 | 10); |
| 170 | RegisterSpecialization<Visitor, base, generic, 2>(); |
| 171 | RegisterSpecialization<Visitor, base, generic, 3>(); |
| 172 | RegisterSpecialization<Visitor, base, generic, 4>(); |
| 173 | RegisterSpecialization<Visitor, base, generic, 5>(); |
| 174 | RegisterSpecialization<Visitor, base, generic, 6>(); |
| 175 | RegisterSpecialization<Visitor, base, generic, 7>(); |
| 176 | RegisterSpecialization<Visitor, base, generic, 8>(); |
| 177 | RegisterSpecialization<Visitor, base, generic, 9>(); |
| 178 | Register(generic, &Visitor::Visit); |
| 179 | } |
| 180 | |
| 181 | private: |
| 182 | base::AtomicWord callbacks_[StaticVisitorBase::kVisitorIdCount]; |
| 183 | }; |
| 184 | |
| 185 | |
| 186 | template <typename StaticVisitor> |
| 187 | class BodyVisitorBase : public AllStatic { |
| 188 | public: |
| 189 | INLINE(static void IteratePointers(Heap* heap, HeapObject* object, |
| 190 | int start_offset, int end_offset)) { |
| 191 | Object** start_slot = |
| 192 | reinterpret_cast<Object**>(object->address() + start_offset); |
| 193 | Object** end_slot = |
| 194 | reinterpret_cast<Object**>(object->address() + end_offset); |
| 195 | StaticVisitor::VisitPointers(heap, start_slot, end_slot); |
| 196 | } |
| 197 | }; |
| 198 | |
| 199 | |
| 200 | template <typename StaticVisitor, typename BodyDescriptor, typename ReturnType> |
| 201 | class FlexibleBodyVisitor : public BodyVisitorBase<StaticVisitor> { |
| 202 | public: |
| 203 | INLINE(static ReturnType Visit(Map* map, HeapObject* object)) { |
| 204 | int object_size = BodyDescriptor::SizeOf(map, object); |
| 205 | BodyVisitorBase<StaticVisitor>::IteratePointers( |
| 206 | map->GetHeap(), object, BodyDescriptor::kStartOffset, object_size); |
| 207 | return static_cast<ReturnType>(object_size); |
| 208 | } |
| 209 | |
| 210 | template <int object_size> |
| 211 | static inline ReturnType VisitSpecialized(Map* map, HeapObject* object) { |
| 212 | DCHECK(BodyDescriptor::SizeOf(map, object) == object_size); |
| 213 | BodyVisitorBase<StaticVisitor>::IteratePointers( |
| 214 | map->GetHeap(), object, BodyDescriptor::kStartOffset, object_size); |
| 215 | return static_cast<ReturnType>(object_size); |
| 216 | } |
| 217 | }; |
| 218 | |
| 219 | |
| 220 | template <typename StaticVisitor, typename BodyDescriptor, typename ReturnType> |
| 221 | class FixedBodyVisitor : public BodyVisitorBase<StaticVisitor> { |
| 222 | public: |
| 223 | INLINE(static ReturnType Visit(Map* map, HeapObject* object)) { |
| 224 | BodyVisitorBase<StaticVisitor>::IteratePointers( |
| 225 | map->GetHeap(), object, BodyDescriptor::kStartOffset, |
| 226 | BodyDescriptor::kEndOffset); |
| 227 | return static_cast<ReturnType>(BodyDescriptor::kSize); |
| 228 | } |
| 229 | }; |
| 230 | |
| 231 | |
| 232 | // Base class for visitors used for a linear new space iteration. |
| 233 | // IterateBody returns size of visited object. |
| 234 | // Certain types of objects (i.e. Code objects) are not handled |
| 235 | // by dispatch table of this visitor because they cannot appear |
| 236 | // in the new space. |
| 237 | // |
| 238 | // This class is intended to be used in the following way: |
| 239 | // |
| 240 | // class SomeVisitor : public StaticNewSpaceVisitor<SomeVisitor> { |
| 241 | // ... |
| 242 | // } |
| 243 | // |
| 244 | // This is an example of Curiously recurring template pattern |
| 245 | // (see http://en.wikipedia.org/wiki/Curiously_recurring_template_pattern). |
| 246 | // We use CRTP to guarantee aggressive compile time optimizations (i.e. |
| 247 | // inlining and specialization of StaticVisitor::VisitPointers methods). |
| 248 | template <typename StaticVisitor> |
| 249 | class StaticNewSpaceVisitor : public StaticVisitorBase { |
| 250 | public: |
| 251 | static void Initialize(); |
| 252 | |
| 253 | INLINE(static int IterateBody(Map* map, HeapObject* obj)) { |
| 254 | return table_.GetVisitor(map)(map, obj); |
| 255 | } |
| 256 | |
| 257 | INLINE(static void VisitPointers(Heap* heap, Object** start, Object** end)) { |
| 258 | for (Object** p = start; p < end; p++) StaticVisitor::VisitPointer(heap, p); |
| 259 | } |
| 260 | |
| 261 | private: |
| 262 | INLINE(static int VisitJSFunction(Map* map, HeapObject* object)) { |
| 263 | Heap* heap = map->GetHeap(); |
| 264 | VisitPointers(heap, |
| 265 | HeapObject::RawField(object, JSFunction::kPropertiesOffset), |
| 266 | HeapObject::RawField(object, JSFunction::kCodeEntryOffset)); |
| 267 | |
| 268 | // Don't visit code entry. We are using this visitor only during scavenges. |
| 269 | |
| 270 | VisitPointers( |
| 271 | heap, HeapObject::RawField(object, |
| 272 | JSFunction::kCodeEntryOffset + kPointerSize), |
| 273 | HeapObject::RawField(object, JSFunction::kNonWeakFieldsEndOffset)); |
| 274 | return JSFunction::kSize; |
| 275 | } |
| 276 | |
| 277 | INLINE(static int VisitByteArray(Map* map, HeapObject* object)) { |
| 278 | return reinterpret_cast<ByteArray*>(object)->ByteArraySize(); |
| 279 | } |
| 280 | |
| 281 | INLINE(static int VisitFixedDoubleArray(Map* map, HeapObject* object)) { |
| 282 | int length = reinterpret_cast<FixedDoubleArray*>(object)->length(); |
| 283 | return FixedDoubleArray::SizeFor(length); |
| 284 | } |
| 285 | |
| 286 | INLINE(static int VisitFixedTypedArray(Map* map, HeapObject* object)) { |
| 287 | return reinterpret_cast<FixedTypedArrayBase*>(object)->size(); |
| 288 | } |
| 289 | |
| 290 | INLINE(static int VisitJSObject(Map* map, HeapObject* object)) { |
| 291 | return JSObjectVisitor::Visit(map, object); |
| 292 | } |
| 293 | |
| 294 | INLINE(static int VisitSeqOneByteString(Map* map, HeapObject* object)) { |
| 295 | return SeqOneByteString::cast(object) |
| 296 | ->SeqOneByteStringSize(map->instance_type()); |
| 297 | } |
| 298 | |
| 299 | INLINE(static int VisitSeqTwoByteString(Map* map, HeapObject* object)) { |
| 300 | return SeqTwoByteString::cast(object) |
| 301 | ->SeqTwoByteStringSize(map->instance_type()); |
| 302 | } |
| 303 | |
| 304 | INLINE(static int VisitFreeSpace(Map* map, HeapObject* object)) { |
| 305 | return FreeSpace::cast(object)->Size(); |
| 306 | } |
| 307 | |
| 308 | INLINE(static int VisitJSArrayBuffer(Map* map, HeapObject* object)); |
| 309 | INLINE(static int VisitJSTypedArray(Map* map, HeapObject* object)); |
| 310 | INLINE(static int VisitJSDataView(Map* map, HeapObject* object)); |
| 311 | |
| 312 | class DataObjectVisitor { |
| 313 | public: |
| 314 | template <int object_size> |
| 315 | static inline int VisitSpecialized(Map* map, HeapObject* object) { |
| 316 | return object_size; |
| 317 | } |
| 318 | |
| 319 | INLINE(static int Visit(Map* map, HeapObject* object)) { |
| 320 | return map->instance_size(); |
| 321 | } |
| 322 | }; |
| 323 | |
| 324 | typedef FlexibleBodyVisitor<StaticVisitor, StructBodyDescriptor, int> |
| 325 | StructVisitor; |
| 326 | |
| 327 | typedef FlexibleBodyVisitor<StaticVisitor, JSObject::BodyDescriptor, int> |
| 328 | JSObjectVisitor; |
| 329 | |
| 330 | typedef int (*Callback)(Map* map, HeapObject* object); |
| 331 | |
| 332 | static VisitorDispatchTable<Callback> table_; |
| 333 | }; |
| 334 | |
| 335 | |
| 336 | template <typename StaticVisitor> |
| 337 | VisitorDispatchTable<typename StaticNewSpaceVisitor<StaticVisitor>::Callback> |
| 338 | StaticNewSpaceVisitor<StaticVisitor>::table_; |
| 339 | |
| 340 | |
| 341 | // Base class for visitors used to transitively mark the entire heap. |
| 342 | // IterateBody returns nothing. |
| 343 | // Certain types of objects might not be handled by this base class and |
| 344 | // no visitor function is registered by the generic initialization. A |
| 345 | // specialized visitor function needs to be provided by the inheriting |
| 346 | // class itself for those cases. |
| 347 | // |
| 348 | // This class is intended to be used in the following way: |
| 349 | // |
| 350 | // class SomeVisitor : public StaticMarkingVisitor<SomeVisitor> { |
| 351 | // ... |
| 352 | // } |
| 353 | // |
| 354 | // This is an example of Curiously recurring template pattern. |
| 355 | template <typename StaticVisitor> |
| 356 | class StaticMarkingVisitor : public StaticVisitorBase { |
| 357 | public: |
| 358 | static void Initialize(); |
| 359 | |
| 360 | INLINE(static void IterateBody(Map* map, HeapObject* obj)) { |
| 361 | table_.GetVisitor(map)(map, obj); |
| 362 | } |
| 363 | |
| 364 | INLINE(static void VisitPropertyCell(Map* map, HeapObject* object)); |
| 365 | INLINE(static void VisitCodeEntry(Heap* heap, Address entry_address)); |
| 366 | INLINE(static void VisitEmbeddedPointer(Heap* heap, RelocInfo* rinfo)); |
| 367 | INLINE(static void VisitCell(Heap* heap, RelocInfo* rinfo)); |
| 368 | INLINE(static void VisitDebugTarget(Heap* heap, RelocInfo* rinfo)); |
| 369 | INLINE(static void VisitCodeTarget(Heap* heap, RelocInfo* rinfo)); |
| 370 | INLINE(static void VisitCodeAgeSequence(Heap* heap, RelocInfo* rinfo)); |
| 371 | INLINE(static void VisitExternalReference(RelocInfo* rinfo)) {} |
| 372 | INLINE(static void VisitRuntimeEntry(RelocInfo* rinfo)) {} |
| 373 | // Skip the weak next code link in a code object. |
| 374 | INLINE(static void VisitNextCodeLink(Heap* heap, Object** slot)) {} |
| 375 | |
| 376 | // TODO(mstarzinger): This should be made protected once refactoring is done. |
| 377 | // Mark non-optimize code for functions inlined into the given optimized |
| 378 | // code. This will prevent it from being flushed. |
| 379 | static void MarkInlinedFunctionsCode(Heap* heap, Code* code); |
| 380 | |
| 381 | protected: |
| 382 | INLINE(static void VisitMap(Map* map, HeapObject* object)); |
| 383 | INLINE(static void VisitCode(Map* map, HeapObject* object)); |
| 384 | INLINE(static void VisitSharedFunctionInfo(Map* map, HeapObject* object)); |
| 385 | INLINE(static void VisitConstantPoolArray(Map* map, HeapObject* object)); |
| 386 | INLINE(static void VisitAllocationSite(Map* map, HeapObject* object)); |
| 387 | INLINE(static void VisitWeakCollection(Map* map, HeapObject* object)); |
| 388 | INLINE(static void VisitJSFunction(Map* map, HeapObject* object)); |
| 389 | INLINE(static void VisitJSRegExp(Map* map, HeapObject* object)); |
| 390 | INLINE(static void VisitJSArrayBuffer(Map* map, HeapObject* object)); |
| 391 | INLINE(static void VisitJSTypedArray(Map* map, HeapObject* object)); |
| 392 | INLINE(static void VisitJSDataView(Map* map, HeapObject* object)); |
| 393 | INLINE(static void VisitNativeContext(Map* map, HeapObject* object)); |
| 394 | |
| 395 | // Mark pointers in a Map and its TransitionArray together, possibly |
| 396 | // treating transitions or back pointers weak. |
| 397 | static void MarkMapContents(Heap* heap, Map* map); |
| 398 | static void MarkTransitionArray(Heap* heap, TransitionArray* transitions); |
| 399 | |
| 400 | // Code flushing support. |
| 401 | INLINE(static bool IsFlushable(Heap* heap, JSFunction* function)); |
| 402 | INLINE(static bool IsFlushable(Heap* heap, SharedFunctionInfo* shared_info)); |
| 403 | |
| 404 | // Helpers used by code flushing support that visit pointer fields and treat |
| 405 | // references to code objects either strongly or weakly. |
| 406 | static void VisitSharedFunctionInfoStrongCode(Heap* heap, HeapObject* object); |
| 407 | static void VisitSharedFunctionInfoWeakCode(Heap* heap, HeapObject* object); |
| 408 | static void VisitJSFunctionStrongCode(Heap* heap, HeapObject* object); |
| 409 | static void VisitJSFunctionWeakCode(Heap* heap, HeapObject* object); |
| 410 | |
| 411 | class DataObjectVisitor { |
| 412 | public: |
| 413 | template <int size> |
| 414 | static inline void VisitSpecialized(Map* map, HeapObject* object) {} |
| 415 | |
| 416 | INLINE(static void Visit(Map* map, HeapObject* object)) {} |
| 417 | }; |
| 418 | |
| 419 | typedef FlexibleBodyVisitor<StaticVisitor, FixedArray::BodyDescriptor, void> |
| 420 | FixedArrayVisitor; |
| 421 | |
| 422 | typedef FlexibleBodyVisitor<StaticVisitor, JSObject::BodyDescriptor, void> |
| 423 | JSObjectVisitor; |
| 424 | |
| 425 | typedef FlexibleBodyVisitor<StaticVisitor, StructBodyDescriptor, void> |
| 426 | StructObjectVisitor; |
| 427 | |
| 428 | typedef void (*Callback)(Map* map, HeapObject* object); |
| 429 | |
| 430 | static VisitorDispatchTable<Callback> table_; |
| 431 | }; |
| 432 | |
| 433 | |
| 434 | template <typename StaticVisitor> |
| 435 | VisitorDispatchTable<typename StaticMarkingVisitor<StaticVisitor>::Callback> |
| 436 | StaticMarkingVisitor<StaticVisitor>::table_; |
| 437 | |
| 438 | |
| 439 | class WeakObjectRetainer; |
| 440 | |
| 441 | |
| 442 | // A weak list is single linked list where each element has a weak pointer to |
| 443 | // the next element. Given the head of the list, this function removes dead |
| 444 | // elements from the list and if requested records slots for next-element |
| 445 | // pointers. The template parameter T is a WeakListVisitor that defines how to |
| 446 | // access the next-element pointers. |
| 447 | template <class T> |
| 448 | Object* VisitWeakList(Heap* heap, Object* list, WeakObjectRetainer* retainer); |
| 449 | } |
| 450 | } // namespace v8::internal |
| 451 | |
| 452 | #endif // V8_OBJECTS_VISITING_H_ |