blob: 2cd71a041210017fc46cf0128fa0517791d0c5d7 [file] [log] [blame]
Ian Rogers2dd0e2c2013-01-24 12:42:14 -08001/*
2 * Copyright (C) 2011 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
Mathieu Chartierad2541a2013-10-25 10:05:23 -070017#include <ctime>
18
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080019#include "object.h"
20
Brian Carlstromea46f952013-07-30 01:26:50 -070021#include "art_field.h"
22#include "art_field-inl.h"
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080023#include "array-inl.h"
24#include "class.h"
25#include "class-inl.h"
Ian Rogers04d7aa92013-03-16 14:29:17 -070026#include "class_linker-inl.h"
Ian Rogers1d54e732013-05-02 21:10:01 -070027#include "gc/accounting/card_table-inl.h"
28#include "gc/heap.h"
Ian Rogers04d7aa92013-03-16 14:29:17 -070029#include "iftable-inl.h"
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080030#include "monitor.h"
31#include "object-inl.h"
Ian Rogers04d7aa92013-03-16 14:29:17 -070032#include "object_array-inl.h"
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080033#include "object_utils.h"
34#include "runtime.h"
35#include "sirt_ref.h"
36#include "throwable.h"
37#include "well_known_classes.h"
38
39namespace art {
40namespace mirror {
41
Hiroshi Yamauchi79719282014-04-10 12:46:22 -070042class CopyReferenceFieldsWithReadBarrierVisitor {
43 public:
44 explicit CopyReferenceFieldsWithReadBarrierVisitor(Object* dest_obj)
45 : dest_obj_(dest_obj) {}
46
47 void operator()(Object* obj, MemberOffset offset, bool /* is_static */) const
48 ALWAYS_INLINE SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
49 // GetFieldObject() contains a RB.
Ian Rogersb0fa5dc2014-04-28 16:47:08 -070050 Object* ref = obj->GetFieldObject<Object>(offset);
Hiroshi Yamauchi79719282014-04-10 12:46:22 -070051 // No WB here as a large object space does not have a card table
52 // coverage. Instead, cards will be marked separately.
Ian Rogersb0fa5dc2014-04-28 16:47:08 -070053 dest_obj_->SetFieldObjectWithoutWriteBarrier<false, false>(offset, ref);
Hiroshi Yamauchi79719282014-04-10 12:46:22 -070054 }
55
56 void operator()(mirror::Class* klass, mirror::Reference* ref) const
57 ALWAYS_INLINE SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
58 // Copy java.lang.ref.Reference.referent which isn't visited in
59 // Object::VisitReferences().
60 DCHECK(klass->IsReferenceClass());
61 this->operator()(ref, mirror::Reference::ReferentOffset(), false);
62 }
63
64 private:
65 Object* const dest_obj_;
66};
67
Mathieu Chartier590fee92013-09-13 13:46:47 -070068static Object* CopyObject(Thread* self, mirror::Object* dest, mirror::Object* src, size_t num_bytes)
69 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080070 // Copy instance data. We assume memcpy copies by words.
71 // TODO: expose and use move32.
Mathieu Chartier590fee92013-09-13 13:46:47 -070072 byte* src_bytes = reinterpret_cast<byte*>(src);
73 byte* dst_bytes = reinterpret_cast<byte*>(dest);
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080074 size_t offset = sizeof(Object);
75 memcpy(dst_bytes + offset, src_bytes + offset, num_bytes - offset);
Hiroshi Yamauchi79719282014-04-10 12:46:22 -070076 if (kUseBakerOrBrooksReadBarrier) {
77 // We need a RB here. After the memcpy that covers the whole
78 // object above, copy references fields one by one again with a
79 // RB. TODO: Optimize this later?
80 CopyReferenceFieldsWithReadBarrierVisitor visitor(dest);
81 src->VisitReferences<true>(visitor, visitor);
82 }
Mathieu Chartier590fee92013-09-13 13:46:47 -070083 gc::Heap* heap = Runtime::Current()->GetHeap();
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080084 // Perform write barriers on copied object references.
Mathieu Chartier590fee92013-09-13 13:46:47 -070085 Class* c = src->GetClass();
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080086 if (c->IsArrayClass()) {
87 if (!c->GetComponentType()->IsPrimitive()) {
Ian Rogersef7d42f2014-01-06 12:55:46 -080088 ObjectArray<Object>* array = dest->AsObjectArray<Object>();
Mathieu Chartier590fee92013-09-13 13:46:47 -070089 heap->WriteBarrierArray(dest, 0, array->GetLength());
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080090 }
91 } else {
Mathieu Chartier590fee92013-09-13 13:46:47 -070092 heap->WriteBarrierEveryFieldOf(dest);
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080093 }
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080094 if (c->IsFinalizable()) {
Mathieu Chartier590fee92013-09-13 13:46:47 -070095 SirtRef<Object> sirt_dest(self, dest);
96 heap->AddFinalizerReference(self, dest);
97 return sirt_dest.get();
98 }
99 return dest;
100}
101
Hiroshi Yamauchi4cd662e2014-04-03 16:28:10 -0700102// An allocation pre-fence visitor that copies the object.
103class CopyObjectVisitor {
104 public:
105 explicit CopyObjectVisitor(Thread* self, SirtRef<Object>* orig, size_t num_bytes)
106 : self_(self), orig_(orig), num_bytes_(num_bytes) {
107 }
108
109 void operator()(Object* obj, size_t usable_size) const
110 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
111 UNUSED(usable_size);
112 CopyObject(self_, obj, orig_->get(), num_bytes_);
113 }
114
115 private:
116 Thread* const self_;
117 SirtRef<Object>* const orig_;
118 const size_t num_bytes_;
119 DISALLOW_COPY_AND_ASSIGN(CopyObjectVisitor);
120};
121
Mathieu Chartier590fee92013-09-13 13:46:47 -0700122Object* Object::Clone(Thread* self) {
123 CHECK(!IsClass()) << "Can't clone classes.";
124 // Object::SizeOf gets the right size even if we're an array. Using c->AllocObject() here would
125 // be wrong.
126 gc::Heap* heap = Runtime::Current()->GetHeap();
127 size_t num_bytes = SizeOf();
128 SirtRef<Object> this_object(self, this);
129 Object* copy;
Hiroshi Yamauchi4cd662e2014-04-03 16:28:10 -0700130 CopyObjectVisitor visitor(self, &this_object, num_bytes);
Mathieu Chartier590fee92013-09-13 13:46:47 -0700131 if (heap->IsMovableObject(this)) {
Hiroshi Yamauchi4cd662e2014-04-03 16:28:10 -0700132 copy = heap->AllocObject<true>(self, GetClass(), num_bytes, visitor);
Mathieu Chartier590fee92013-09-13 13:46:47 -0700133 } else {
Hiroshi Yamauchi4cd662e2014-04-03 16:28:10 -0700134 copy = heap->AllocNonMovableObject<true>(self, GetClass(), num_bytes, visitor);
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800135 }
Mathieu Chartier0732d592013-11-06 11:02:50 -0800136 return copy;
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800137}
138
Mathieu Chartier4e6a31e2013-10-31 10:35:05 -0700139int32_t Object::GenerateIdentityHashCode() {
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700140 static AtomicInteger seed(987654321 + std::time(nullptr));
Mathieu Chartier4e6a31e2013-10-31 10:35:05 -0700141 int32_t expected_value, new_value;
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700142 do {
Ian Rogersb122a4b2013-11-19 18:00:50 -0800143 expected_value = static_cast<uint32_t>(seed.Load());
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700144 new_value = expected_value * 1103515245 + 12345;
Mathieu Chartier4e6a31e2013-10-31 10:35:05 -0700145 } while ((expected_value & LockWord::kHashMask) == 0 ||
Ian Rogersb122a4b2013-11-19 18:00:50 -0800146 !seed.CompareAndSwap(expected_value, new_value));
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700147 return expected_value & LockWord::kHashMask;
148}
149
150int32_t Object::IdentityHashCode() const {
Mathieu Chartier590fee92013-09-13 13:46:47 -0700151 mirror::Object* current_this = const_cast<mirror::Object*>(this);
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700152 while (true) {
Mathieu Chartier4d7f61d2014-04-17 14:43:39 -0700153 LockWord lw = current_this->GetLockWord(false);
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700154 switch (lw.GetState()) {
155 case LockWord::kUnlocked: {
156 // Try to compare and swap in a new hash, if we succeed we will return the hash on the next
157 // loop iteration.
158 LockWord hash_word(LockWord::FromHashCode(GenerateIdentityHashCode()));
159 DCHECK_EQ(hash_word.GetState(), LockWord::kHashCode);
160 if (const_cast<Object*>(this)->CasLockWord(lw, hash_word)) {
161 return hash_word.GetHashCode();
162 }
163 break;
164 }
165 case LockWord::kThinLocked: {
166 // Inflate the thin lock to a monitor and stick the hash code inside of the monitor.
167 Thread* self = Thread::Current();
Mathieu Chartier590fee92013-09-13 13:46:47 -0700168 SirtRef<mirror::Object> sirt_this(self, current_this);
169 Monitor::InflateThinLocked(self, sirt_this, lw, GenerateIdentityHashCode());
170 // A GC may have occurred when we switched to kBlocked.
171 current_this = sirt_this.get();
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700172 break;
173 }
174 case LockWord::kFatLocked: {
175 // Already inflated, return the has stored in the monitor.
176 Monitor* monitor = lw.FatLockMonitor();
177 DCHECK(monitor != nullptr);
178 return monitor->GetHashCode();
179 }
180 case LockWord::kHashCode: {
181 return lw.GetHashCode();
182 }
Mathieu Chartier590fee92013-09-13 13:46:47 -0700183 default: {
184 LOG(FATAL) << "Invalid state during hashcode " << lw.GetState();
185 break;
186 }
Mathieu Chartierad2541a2013-10-25 10:05:23 -0700187 }
188 }
189 LOG(FATAL) << "Unreachable";
190 return 0;
191}
192
Ian Rogersef7d42f2014-01-06 12:55:46 -0800193void Object::CheckFieldAssignmentImpl(MemberOffset field_offset, Object* new_value) {
194 Class* c = GetClass();
Mathieu Chartier4e305412014-02-19 10:54:44 -0800195 Runtime* runtime = Runtime::Current();
196 if (runtime->GetClassLinker() == nullptr || !runtime->IsStarted() ||
197 !runtime->GetHeap()->IsObjectValidationEnabled() || !c->IsResolved()) {
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800198 return;
199 }
Ian Rogersef7d42f2014-01-06 12:55:46 -0800200 for (Class* cur = c; cur != NULL; cur = cur->GetSuperClass()) {
Brian Carlstromea46f952013-07-30 01:26:50 -0700201 ObjectArray<ArtField>* fields = cur->GetIFields();
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800202 if (fields != NULL) {
203 size_t num_ref_ifields = cur->NumReferenceInstanceFields();
204 for (size_t i = 0; i < num_ref_ifields; ++i) {
Brian Carlstromea46f952013-07-30 01:26:50 -0700205 ArtField* field = fields->Get(i);
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800206 if (field->GetOffset().Int32Value() == field_offset.Int32Value()) {
207 FieldHelper fh(field);
208 CHECK(fh.GetType()->IsAssignableFrom(new_value->GetClass()));
209 return;
210 }
211 }
212 }
213 }
214 if (c->IsArrayClass()) {
215 // Bounds and assign-ability done in the array setter.
216 return;
217 }
218 if (IsClass()) {
Brian Carlstromea46f952013-07-30 01:26:50 -0700219 ObjectArray<ArtField>* fields = AsClass()->GetSFields();
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800220 if (fields != NULL) {
221 size_t num_ref_sfields = AsClass()->NumReferenceStaticFields();
222 for (size_t i = 0; i < num_ref_sfields; ++i) {
Brian Carlstromea46f952013-07-30 01:26:50 -0700223 ArtField* field = fields->Get(i);
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800224 if (field->GetOffset().Int32Value() == field_offset.Int32Value()) {
225 FieldHelper fh(field);
226 CHECK(fh.GetType()->IsAssignableFrom(new_value->GetClass()));
227 return;
228 }
229 }
230 }
231 }
232 LOG(FATAL) << "Failed to find field for assignment to " << reinterpret_cast<void*>(this)
233 << " of type " << PrettyDescriptor(c) << " at offset " << field_offset;
234}
Ian Rogers2dd0e2c2013-01-24 12:42:14 -0800235
236} // namespace mirror
237} // namespace art