| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "instrumentation.h" |
| |
| #include <functional> |
| #include <optional> |
| #include <sstream> |
| |
| #include <android-base/logging.h> |
| |
| #include "arch/context.h" |
| #include "art_field-inl.h" |
| #include "art_method-inl.h" |
| #include "base/atomic.h" |
| #include "base/callee_save_type.h" |
| #include "class_linker.h" |
| #include "debugger.h" |
| #include "dex/dex_file-inl.h" |
| #include "dex/dex_file_types.h" |
| #include "dex/dex_instruction-inl.h" |
| #include "entrypoints/quick/quick_alloc_entrypoints.h" |
| #include "entrypoints/quick/quick_entrypoints.h" |
| #include "entrypoints/runtime_asm_entrypoints.h" |
| #include "gc_root-inl.h" |
| #include "interpreter/interpreter.h" |
| #include "interpreter/interpreter_common.h" |
| #include "jit/jit.h" |
| #include "jit/jit_code_cache.h" |
| #include "jvalue-inl.h" |
| #include "jvalue.h" |
| #include "mirror/class-inl.h" |
| #include "mirror/dex_cache.h" |
| #include "mirror/object-inl.h" |
| #include "mirror/object_array-inl.h" |
| #include "nth_caller_visitor.h" |
| #include "oat_quick_method_header.h" |
| #include "runtime-inl.h" |
| #include "thread.h" |
| #include "thread_list.h" |
| |
| namespace art { |
| namespace instrumentation { |
| |
| constexpr bool kVerboseInstrumentation = false; |
| |
| void InstrumentationListener::MethodExited( |
| Thread* thread, |
| Handle<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| OptionalFrame frame, |
| MutableHandle<mirror::Object>& return_value) { |
| DCHECK_EQ(method->GetInterfaceMethodIfProxy(kRuntimePointerSize)->GetReturnTypePrimitive(), |
| Primitive::kPrimNot); |
| const void* original_ret = return_value.Get(); |
| JValue v; |
| v.SetL(return_value.Get()); |
| MethodExited(thread, this_object, method, dex_pc, frame, v); |
| DCHECK(original_ret == v.GetL()) << "Return value changed"; |
| } |
| |
| void InstrumentationListener::FieldWritten(Thread* thread, |
| Handle<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| ArtField* field, |
| Handle<mirror::Object> field_value) { |
| DCHECK(!field->IsPrimitiveType()); |
| JValue v; |
| v.SetL(field_value.Get()); |
| FieldWritten(thread, this_object, method, dex_pc, field, v); |
| } |
| |
| // Instrumentation works on non-inlined frames by updating returned PCs |
| // of compiled frames. |
| static constexpr StackVisitor::StackWalkKind kInstrumentationStackWalk = |
| StackVisitor::StackWalkKind::kSkipInlinedFrames; |
| |
| class InstallStubsClassVisitor : public ClassVisitor { |
| public: |
| explicit InstallStubsClassVisitor(Instrumentation* instrumentation) |
| : instrumentation_(instrumentation) {} |
| |
| bool operator()(ObjPtr<mirror::Class> klass) override REQUIRES(Locks::mutator_lock_) { |
| instrumentation_->InstallStubsForClass(klass.Ptr()); |
| return true; // we visit all classes. |
| } |
| |
| private: |
| Instrumentation* const instrumentation_; |
| }; |
| |
| InstrumentationStackPopper::InstrumentationStackPopper(Thread* self) |
| : self_(self), |
| instrumentation_(Runtime::Current()->GetInstrumentation()), |
| pop_until_(0u) {} |
| |
| InstrumentationStackPopper::~InstrumentationStackPopper() { |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| self_->GetInstrumentationStack(); |
| for (auto i = stack->begin(); i != stack->end() && i->first <= pop_until_;) { |
| i = stack->erase(i); |
| } |
| } |
| |
| bool InstrumentationStackPopper::PopFramesTo(uintptr_t stack_pointer, |
| MutableHandle<mirror::Throwable>& exception) { |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| self_->GetInstrumentationStack(); |
| DCHECK(!self_->IsExceptionPending()); |
| if (!instrumentation_->HasMethodUnwindListeners()) { |
| pop_until_ = stack_pointer; |
| return true; |
| } |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Popping frames for exception " << exception->Dump(); |
| } |
| // The instrumentation events expect the exception to be set. |
| self_->SetException(exception.Get()); |
| bool new_exception_thrown = false; |
| auto i = stack->upper_bound(pop_until_); |
| |
| // Now pop all frames until reaching stack_pointer, or a new exception is |
| // thrown. Note that `stack_pointer` doesn't need to be a return PC address |
| // (in fact the exception handling code passes the start of the frame where |
| // the catch handler is). |
| for (; i != stack->end() && i->first <= stack_pointer; i++) { |
| const InstrumentationStackFrame& frame = i->second; |
| ArtMethod* method = frame.method_; |
| // Notify listeners of method unwind. |
| // TODO: improve the dex_pc information here. |
| uint32_t dex_pc = dex::kDexNoIndex; |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Popping for unwind " << method->PrettyMethod(); |
| } |
| if (!method->IsRuntimeMethod() && !frame.interpreter_entry_) { |
| instrumentation_->MethodUnwindEvent(self_, frame.this_object_, method, dex_pc); |
| new_exception_thrown = self_->GetException() != exception.Get(); |
| if (new_exception_thrown) { |
| pop_until_ = i->first; |
| break; |
| } |
| } |
| } |
| if (!new_exception_thrown) { |
| pop_until_ = stack_pointer; |
| } |
| exception.Assign(self_->GetException()); |
| self_->ClearException(); |
| if (kVerboseInstrumentation && new_exception_thrown) { |
| LOG(INFO) << "Did partial pop of frames due to new exception"; |
| } |
| return !new_exception_thrown; |
| } |
| |
| Instrumentation::Instrumentation() |
| : current_force_deopt_id_(0), |
| instrumentation_stubs_installed_(false), |
| entry_exit_stubs_installed_(false), |
| interpreter_stubs_installed_(false), |
| interpret_only_(false), |
| forced_interpret_only_(false), |
| have_method_entry_listeners_(false), |
| have_method_exit_listeners_(false), |
| have_method_unwind_listeners_(false), |
| have_dex_pc_listeners_(false), |
| have_field_read_listeners_(false), |
| have_field_write_listeners_(false), |
| have_exception_thrown_listeners_(false), |
| have_watched_frame_pop_listeners_(false), |
| have_branch_listeners_(false), |
| have_exception_handled_listeners_(false), |
| deoptimized_methods_lock_(new ReaderWriterMutex("deoptimized methods lock", |
| kGenericBottomLock)), |
| deoptimization_enabled_(false), |
| interpreter_handler_table_(kMainHandlerTable), |
| quick_alloc_entry_points_instrumentation_counter_(0), |
| alloc_entrypoints_instrumented_(false), |
| can_use_instrumentation_trampolines_(true) { |
| } |
| |
| void Instrumentation::InstallStubsForClass(ObjPtr<mirror::Class> klass) { |
| if (!klass->IsResolved()) { |
| // We need the class to be resolved to install/uninstall stubs. Otherwise its methods |
| // could not be initialized or linked with regards to class inheritance. |
| } else if (klass->IsErroneousResolved()) { |
| // We can't execute code in a erroneous class: do nothing. |
| } else { |
| for (ArtMethod& method : klass->GetMethods(kRuntimePointerSize)) { |
| InstallStubsForMethod(&method); |
| } |
| } |
| } |
| |
| static void UpdateEntrypoints(ArtMethod* method, const void* quick_code) |
| REQUIRES_SHARED(Locks::mutator_lock_) { |
| if (kIsDebugBuild) { |
| jit::Jit* jit = Runtime::Current()->GetJit(); |
| if (jit != nullptr && jit->GetCodeCache()->ContainsPc(quick_code)) { |
| // Ensure we always have the thumb entrypoint for JIT on arm32. |
| if (kRuntimeISA == InstructionSet::kArm) { |
| CHECK_EQ(reinterpret_cast<uintptr_t>(quick_code) & 1, 1u); |
| } |
| } |
| } |
| method->SetEntryPointFromQuickCompiledCode(quick_code); |
| } |
| |
| bool Instrumentation::NeedDebugVersionFor(ArtMethod* method) const |
| REQUIRES_SHARED(Locks::mutator_lock_) { |
| art::Runtime* runtime = Runtime::Current(); |
| // If anything says we need the debug version or we are debuggable we will need the debug version |
| // of the method. |
| return (runtime->GetRuntimeCallbacks()->MethodNeedsDebugVersion(method) || |
| runtime->IsJavaDebuggable()) && |
| !method->IsNative() && |
| !method->IsProxyMethod(); |
| } |
| |
| void Instrumentation::InstallStubsForMethod(ArtMethod* method) { |
| if (!method->IsInvokable() || method->IsProxyMethod()) { |
| // Do not change stubs for these methods. |
| return; |
| } |
| // Don't stub Proxy.<init>. Note that the Proxy class itself is not a proxy class. |
| // TODO We should remove the need for this since it means we cannot always correctly detect calls |
| // to Proxy.<init> |
| // Annoyingly this can be called before we have actually initialized WellKnownClasses so therefore |
| // we also need to check this based on the declaring-class descriptor. The check is valid because |
| // Proxy only has a single constructor. |
| ArtMethod* well_known_proxy_init = jni::DecodeArtMethod( |
| WellKnownClasses::java_lang_reflect_Proxy_init); |
| if ((LIKELY(well_known_proxy_init != nullptr) && UNLIKELY(method == well_known_proxy_init)) || |
| UNLIKELY(method->IsConstructor() && |
| method->GetDeclaringClass()->DescriptorEquals("Ljava/lang/reflect/Proxy;"))) { |
| return; |
| } |
| const void* new_quick_code; |
| bool uninstall = !entry_exit_stubs_installed_ && !interpreter_stubs_installed_; |
| Runtime* const runtime = Runtime::Current(); |
| ClassLinker* const class_linker = runtime->GetClassLinker(); |
| bool is_class_initialized = method->GetDeclaringClass()->IsInitialized(); |
| if (uninstall) { |
| if ((forced_interpret_only_ || IsDeoptimized(method)) && !method->IsNative()) { |
| new_quick_code = GetQuickToInterpreterBridge(); |
| } else if (is_class_initialized || !method->IsStatic() || method->IsConstructor()) { |
| new_quick_code = GetCodeForInvoke(method); |
| } else { |
| new_quick_code = GetQuickResolutionStub(); |
| } |
| } else { // !uninstall |
| if ((interpreter_stubs_installed_ || forced_interpret_only_ || IsDeoptimized(method)) && |
| !method->IsNative()) { |
| new_quick_code = GetQuickToInterpreterBridge(); |
| } else { |
| // Do not overwrite resolution trampoline. When the trampoline initializes the method's |
| // class, all its static methods code will be set to the instrumentation entry point. |
| // For more details, see ClassLinker::FixupStaticTrampolines. |
| if (is_class_initialized || !method->IsStatic() || method->IsConstructor()) { |
| if (entry_exit_stubs_installed_) { |
| // This needs to be checked first since the instrumentation entrypoint will be able to |
| // find the actual JIT compiled code that corresponds to this method. |
| new_quick_code = GetQuickInstrumentationEntryPoint(); |
| } else if (NeedDebugVersionFor(method)) { |
| // It would be great to search the JIT for its implementation here but we cannot due to |
| // the locks we hold. Instead just set to the interpreter bridge and that code will search |
| // the JIT when it gets called and replace the entrypoint then. |
| new_quick_code = GetQuickToInterpreterBridge(); |
| } else { |
| new_quick_code = class_linker->GetQuickOatCodeFor(method); |
| } |
| } else { |
| new_quick_code = GetQuickResolutionStub(); |
| } |
| } |
| } |
| UpdateEntrypoints(method, new_quick_code); |
| } |
| |
| // Places the instrumentation exit pc as the return PC for every quick frame. This also allows |
| // deoptimization of quick frames to interpreter frames. |
| // Since we may already have done this previously, we need to push new instrumentation frame before |
| // existing instrumentation frames. |
| void InstrumentationInstallStack(Thread* thread, void* arg) |
| REQUIRES(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| struct InstallStackVisitor final : public StackVisitor { |
| InstallStackVisitor(Thread* thread_in, |
| Context* context, |
| uintptr_t instrumentation_exit_pc, |
| uint64_t force_deopt_id) |
| : StackVisitor(thread_in, context, kInstrumentationStackWalk), |
| instrumentation_stack_(thread_in->GetInstrumentationStack()), |
| instrumentation_exit_pc_(instrumentation_exit_pc), |
| reached_existing_instrumentation_frames_(false), |
| last_return_pc_(0), |
| force_deopt_id_(force_deopt_id) {} |
| |
| bool VisitFrame() override REQUIRES_SHARED(Locks::mutator_lock_) { |
| ArtMethod* m = GetMethod(); |
| if (m == nullptr) { |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " Skipping upcall. Frame " << GetFrameId(); |
| } |
| last_return_pc_ = 0; |
| return true; // Ignore upcalls. |
| } |
| if (GetCurrentQuickFrame() == nullptr) { |
| bool interpreter_frame = true; |
| InstrumentationStackFrame instrumentation_frame(GetThisObject().Ptr(), |
| m, |
| /*return_pc=*/ 0, |
| GetFrameId(), |
| interpreter_frame, |
| force_deopt_id_); |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Pushing shadow frame " << instrumentation_frame.Dump(); |
| } |
| shadow_stack_.push_back(instrumentation_frame); |
| return true; // Continue. |
| } |
| uintptr_t return_pc = GetReturnPc(); |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " Installing exit stub in " << DescribeLocation(); |
| } |
| if (return_pc == instrumentation_exit_pc_) { |
| auto it = instrumentation_stack_->find(GetReturnPcAddr()); |
| CHECK(it != instrumentation_stack_->end()); |
| const InstrumentationStackFrame& frame = it->second; |
| if (m->IsRuntimeMethod()) { |
| if (frame.interpreter_entry_) { |
| // This instrumentation frame is for an interpreter bridge and is |
| // pushed when executing the instrumented interpreter bridge. So method |
| // enter event must have been reported. However we need to push a DEX pc |
| // into the dex_pcs_ list to match size of instrumentation stack. |
| uint32_t dex_pc = dex::kDexNoIndex; |
| dex_pcs_.push_back(dex_pc); |
| last_return_pc_ = frame.return_pc_; |
| return true; |
| } |
| } |
| |
| // We've reached a frame which has already been installed with instrumentation exit stub. |
| // We should have already installed instrumentation or be interpreter on previous frames. |
| reached_existing_instrumentation_frames_ = true; |
| |
| CHECK_EQ(m->GetNonObsoleteMethod(), frame.method_->GetNonObsoleteMethod()) |
| << "Expected " << ArtMethod::PrettyMethod(m) |
| << ", Found " << ArtMethod::PrettyMethod(frame.method_); |
| return_pc = frame.return_pc_; |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Ignoring already instrumented " << frame.Dump(); |
| } |
| } else { |
| CHECK_NE(return_pc, 0U); |
| if (UNLIKELY(reached_existing_instrumentation_frames_ && !m->IsRuntimeMethod())) { |
| // We already saw an existing instrumentation frame so this should be a runtime-method |
| // inserted by the interpreter or runtime. |
| std::string thread_name; |
| GetThread()->GetThreadName(thread_name); |
| uint32_t dex_pc = dex::kDexNoIndex; |
| if (last_return_pc_ != 0 && GetCurrentOatQuickMethodHeader() != nullptr) { |
| dex_pc = GetCurrentOatQuickMethodHeader()->ToDexPc( |
| GetCurrentQuickFrame(), last_return_pc_); |
| } |
| LOG(FATAL) << "While walking " << thread_name << " found unexpected non-runtime method" |
| << " without instrumentation exit return or interpreter frame." |
| << " method is " << GetMethod()->PrettyMethod() |
| << " return_pc is " << std::hex << return_pc |
| << " dex pc: " << dex_pc; |
| UNREACHABLE(); |
| } |
| InstrumentationStackFrame instrumentation_frame( |
| m->IsRuntimeMethod() ? nullptr : GetThisObject().Ptr(), |
| m, |
| return_pc, |
| GetFrameId(), // A runtime method still gets a frame id. |
| false, |
| force_deopt_id_); |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Pushing frame " << instrumentation_frame.Dump(); |
| } |
| |
| instrumentation_stack_->insert({GetReturnPcAddr(), instrumentation_frame}); |
| SetReturnPc(instrumentation_exit_pc_); |
| } |
| uint32_t dex_pc = dex::kDexNoIndex; |
| if (last_return_pc_ != 0 && GetCurrentOatQuickMethodHeader() != nullptr) { |
| dex_pc = GetCurrentOatQuickMethodHeader()->ToDexPc(GetCurrentQuickFrame(), last_return_pc_); |
| } |
| dex_pcs_.push_back(dex_pc); |
| last_return_pc_ = return_pc; |
| return true; // Continue. |
| } |
| std::map<uintptr_t, InstrumentationStackFrame>* const instrumentation_stack_; |
| std::vector<InstrumentationStackFrame> shadow_stack_; |
| std::vector<uint32_t> dex_pcs_; |
| const uintptr_t instrumentation_exit_pc_; |
| bool reached_existing_instrumentation_frames_; |
| uintptr_t last_return_pc_; |
| uint64_t force_deopt_id_; |
| }; |
| if (kVerboseInstrumentation) { |
| std::string thread_name; |
| thread->GetThreadName(thread_name); |
| LOG(INFO) << "Installing exit stubs in " << thread_name; |
| } |
| |
| Instrumentation* instrumentation = reinterpret_cast<Instrumentation*>(arg); |
| std::unique_ptr<Context> context(Context::Create()); |
| uintptr_t instrumentation_exit_pc = reinterpret_cast<uintptr_t>(GetQuickInstrumentationExitPc()); |
| InstallStackVisitor visitor( |
| thread, context.get(), instrumentation_exit_pc, instrumentation->current_force_deopt_id_); |
| visitor.WalkStack(true); |
| CHECK_EQ(visitor.dex_pcs_.size(), thread->GetInstrumentationStack()->size()); |
| |
| if (instrumentation->ShouldNotifyMethodEnterExitEvents()) { |
| // Create method enter events for all methods currently on the thread's stack. We only do this |
| // if no debugger is attached to prevent from posting events twice. |
| // TODO: This is the only place we make use of frame_id_. We should create a |
| // std::vector instead and populate it as we walk the stack. |
| auto ssi = visitor.shadow_stack_.rbegin(); |
| for (auto isi = thread->GetInstrumentationStack()->rbegin(), |
| end = thread->GetInstrumentationStack()->rend(); isi != end; ++isi) { |
| while (ssi != visitor.shadow_stack_.rend() && (*ssi).frame_id_ < isi->second.frame_id_) { |
| instrumentation->MethodEnterEvent(thread, (*ssi).this_object_, (*ssi).method_, 0); |
| ++ssi; |
| } |
| uint32_t dex_pc = visitor.dex_pcs_.back(); |
| visitor.dex_pcs_.pop_back(); |
| if (!isi->second.interpreter_entry_ && !isi->second.method_->IsRuntimeMethod()) { |
| instrumentation->MethodEnterEvent( |
| thread, isi->second.this_object_, isi->second.method_, dex_pc); |
| } |
| } |
| } |
| thread->VerifyStack(); |
| } |
| |
| void Instrumentation::InstrumentThreadStack(Thread* thread) { |
| instrumentation_stubs_installed_ = true; |
| InstrumentationInstallStack(thread, this); |
| } |
| |
| // Removes the instrumentation exit pc as the return PC for every quick frame. |
| static void InstrumentationRestoreStack(Thread* thread, void* arg) |
| REQUIRES(Locks::mutator_lock_) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| |
| struct RestoreStackVisitor final : public StackVisitor { |
| RestoreStackVisitor(Thread* thread_in, uintptr_t instrumentation_exit_pc, |
| Instrumentation* instrumentation) |
| : StackVisitor(thread_in, nullptr, kInstrumentationStackWalk), |
| thread_(thread_in), |
| instrumentation_exit_pc_(instrumentation_exit_pc), |
| instrumentation_(instrumentation), |
| instrumentation_stack_(thread_in->GetInstrumentationStack()), |
| frames_removed_(0) {} |
| |
| bool VisitFrame() override REQUIRES_SHARED(Locks::mutator_lock_) { |
| if (instrumentation_stack_->size() == 0) { |
| return false; // Stop. |
| } |
| ArtMethod* m = GetMethod(); |
| if (GetCurrentQuickFrame() == nullptr) { |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " Ignoring a shadow frame. Frame " << GetFrameId() |
| << " Method=" << ArtMethod::PrettyMethod(m); |
| } |
| return true; // Ignore shadow frames. |
| } |
| if (m == nullptr) { |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " Skipping upcall. Frame " << GetFrameId(); |
| } |
| return true; // Ignore upcalls. |
| } |
| auto it = instrumentation_stack_->find(GetReturnPcAddr()); |
| if (it != instrumentation_stack_->end()) { |
| const InstrumentationStackFrame& instrumentation_frame = it->second; |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " Removing exit stub in " << DescribeLocation(); |
| } |
| if (instrumentation_frame.interpreter_entry_) { |
| CHECK(m == Runtime::Current()->GetCalleeSaveMethod(CalleeSaveType::kSaveRefsAndArgs)); |
| } else { |
| CHECK_EQ(m->GetNonObsoleteMethod(), |
| instrumentation_frame.method_->GetNonObsoleteMethod()) |
| << ArtMethod::PrettyMethod(m) |
| << " and " << instrumentation_frame.method_->GetNonObsoleteMethod()->PrettyMethod(); |
| } |
| SetReturnPc(instrumentation_frame.return_pc_); |
| if (instrumentation_->ShouldNotifyMethodEnterExitEvents() && |
| !m->IsRuntimeMethod()) { |
| // Create the method exit events. As the methods didn't really exit the result is 0. |
| // We only do this if no debugger is attached to prevent from posting events twice. |
| JValue val; |
| instrumentation_->MethodExitEvent(thread_, instrumentation_frame.this_object_, m, |
| GetDexPc(), OptionalFrame{}, val); |
| } |
| frames_removed_++; |
| } else { |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << " No exit stub in " << DescribeLocation(); |
| } |
| } |
| return true; // Continue. |
| } |
| Thread* const thread_; |
| const uintptr_t instrumentation_exit_pc_; |
| Instrumentation* const instrumentation_; |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* const instrumentation_stack_; |
| size_t frames_removed_; |
| }; |
| if (kVerboseInstrumentation) { |
| std::string thread_name; |
| thread->GetThreadName(thread_name); |
| LOG(INFO) << "Removing exit stubs in " << thread_name; |
| } |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| thread->GetInstrumentationStack(); |
| if (stack->size() > 0) { |
| Instrumentation* instrumentation = reinterpret_cast<Instrumentation*>(arg); |
| uintptr_t instrumentation_exit_pc = |
| reinterpret_cast<uintptr_t>(GetQuickInstrumentationExitPc()); |
| RestoreStackVisitor visitor(thread, instrumentation_exit_pc, instrumentation); |
| visitor.WalkStack(true); |
| CHECK_EQ(visitor.frames_removed_, stack->size()); |
| stack->clear(); |
| } |
| } |
| |
| void Instrumentation::DeoptimizeAllThreadFrames() { |
| Thread* self = Thread::Current(); |
| MutexLock mu(self, *Locks::thread_list_lock_); |
| ThreadList* tl = Runtime::Current()->GetThreadList(); |
| tl->ForEach([&](Thread* t) { |
| Locks::mutator_lock_->AssertExclusiveHeld(self); |
| InstrumentThreadStack(t); |
| }); |
| current_force_deopt_id_++; |
| } |
| |
| static bool HasEvent(Instrumentation::InstrumentationEvent expected, uint32_t events) { |
| return (events & expected) != 0; |
| } |
| |
| static void PotentiallyAddListenerTo(Instrumentation::InstrumentationEvent event, |
| uint32_t events, |
| std::list<InstrumentationListener*>& list, |
| InstrumentationListener* listener, |
| bool* has_listener) |
| REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !Locks::classlinker_classes_lock_) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| if (!HasEvent(event, events)) { |
| return; |
| } |
| // If there is a free slot in the list, we insert the listener in that slot. |
| // Otherwise we add it to the end of the list. |
| auto it = std::find(list.begin(), list.end(), nullptr); |
| if (it != list.end()) { |
| *it = listener; |
| } else { |
| list.push_back(listener); |
| } |
| Runtime::DoAndMaybeSwitchInterpreter([=](){ *has_listener = true; }); |
| } |
| |
| void Instrumentation::AddListener(InstrumentationListener* listener, uint32_t events) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| PotentiallyAddListenerTo(kMethodEntered, |
| events, |
| method_entry_listeners_, |
| listener, |
| &have_method_entry_listeners_); |
| PotentiallyAddListenerTo(kMethodExited, |
| events, |
| method_exit_listeners_, |
| listener, |
| &have_method_exit_listeners_); |
| PotentiallyAddListenerTo(kMethodUnwind, |
| events, |
| method_unwind_listeners_, |
| listener, |
| &have_method_unwind_listeners_); |
| PotentiallyAddListenerTo(kBranch, |
| events, |
| branch_listeners_, |
| listener, |
| &have_branch_listeners_); |
| PotentiallyAddListenerTo(kDexPcMoved, |
| events, |
| dex_pc_listeners_, |
| listener, |
| &have_dex_pc_listeners_); |
| PotentiallyAddListenerTo(kFieldRead, |
| events, |
| field_read_listeners_, |
| listener, |
| &have_field_read_listeners_); |
| PotentiallyAddListenerTo(kFieldWritten, |
| events, |
| field_write_listeners_, |
| listener, |
| &have_field_write_listeners_); |
| PotentiallyAddListenerTo(kExceptionThrown, |
| events, |
| exception_thrown_listeners_, |
| listener, |
| &have_exception_thrown_listeners_); |
| PotentiallyAddListenerTo(kWatchedFramePop, |
| events, |
| watched_frame_pop_listeners_, |
| listener, |
| &have_watched_frame_pop_listeners_); |
| PotentiallyAddListenerTo(kExceptionHandled, |
| events, |
| exception_handled_listeners_, |
| listener, |
| &have_exception_handled_listeners_); |
| UpdateInterpreterHandlerTable(); |
| } |
| |
| static void PotentiallyRemoveListenerFrom(Instrumentation::InstrumentationEvent event, |
| uint32_t events, |
| std::list<InstrumentationListener*>& list, |
| InstrumentationListener* listener, |
| bool* has_listener) |
| REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !Locks::classlinker_classes_lock_) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| if (!HasEvent(event, events)) { |
| return; |
| } |
| auto it = std::find(list.begin(), list.end(), listener); |
| if (it != list.end()) { |
| // Just update the entry, do not remove from the list. Removing entries in the list |
| // is unsafe when mutators are iterating over it. |
| *it = nullptr; |
| } |
| |
| // Check if the list contains any non-null listener, and update 'has_listener'. |
| for (InstrumentationListener* l : list) { |
| if (l != nullptr) { |
| Runtime::DoAndMaybeSwitchInterpreter([=](){ *has_listener = true; }); |
| return; |
| } |
| } |
| Runtime::DoAndMaybeSwitchInterpreter([=](){ *has_listener = false; }); |
| } |
| |
| void Instrumentation::RemoveListener(InstrumentationListener* listener, uint32_t events) { |
| Locks::mutator_lock_->AssertExclusiveHeld(Thread::Current()); |
| PotentiallyRemoveListenerFrom(kMethodEntered, |
| events, |
| method_entry_listeners_, |
| listener, |
| &have_method_entry_listeners_); |
| PotentiallyRemoveListenerFrom(kMethodExited, |
| events, |
| method_exit_listeners_, |
| listener, |
| &have_method_exit_listeners_); |
| PotentiallyRemoveListenerFrom(kMethodUnwind, |
| events, |
| method_unwind_listeners_, |
| listener, |
| &have_method_unwind_listeners_); |
| PotentiallyRemoveListenerFrom(kBranch, |
| events, |
| branch_listeners_, |
| listener, |
| &have_branch_listeners_); |
| PotentiallyRemoveListenerFrom(kDexPcMoved, |
| events, |
| dex_pc_listeners_, |
| listener, |
| &have_dex_pc_listeners_); |
| PotentiallyRemoveListenerFrom(kFieldRead, |
| events, |
| field_read_listeners_, |
| listener, |
| &have_field_read_listeners_); |
| PotentiallyRemoveListenerFrom(kFieldWritten, |
| events, |
| field_write_listeners_, |
| listener, |
| &have_field_write_listeners_); |
| PotentiallyRemoveListenerFrom(kExceptionThrown, |
| events, |
| exception_thrown_listeners_, |
| listener, |
| &have_exception_thrown_listeners_); |
| PotentiallyRemoveListenerFrom(kWatchedFramePop, |
| events, |
| watched_frame_pop_listeners_, |
| listener, |
| &have_watched_frame_pop_listeners_); |
| PotentiallyRemoveListenerFrom(kExceptionHandled, |
| events, |
| exception_handled_listeners_, |
| listener, |
| &have_exception_handled_listeners_); |
| UpdateInterpreterHandlerTable(); |
| } |
| |
| Instrumentation::InstrumentationLevel Instrumentation::GetCurrentInstrumentationLevel() const { |
| if (interpreter_stubs_installed_) { |
| return InstrumentationLevel::kInstrumentWithInterpreter; |
| } else if (entry_exit_stubs_installed_) { |
| return InstrumentationLevel::kInstrumentWithInstrumentationStubs; |
| } else { |
| return InstrumentationLevel::kInstrumentNothing; |
| } |
| } |
| |
| bool Instrumentation::RequiresInstrumentationInstallation(InstrumentationLevel new_level) const { |
| // We need to reinstall instrumentation if we go to a different level. |
| return GetCurrentInstrumentationLevel() != new_level; |
| } |
| |
| void Instrumentation::UpdateInstrumentationLevels(InstrumentationLevel level) { |
| if (level == InstrumentationLevel::kInstrumentWithInterpreter) { |
| can_use_instrumentation_trampolines_ = false; |
| } |
| if (UNLIKELY(!can_use_instrumentation_trampolines_)) { |
| for (auto& p : requested_instrumentation_levels_) { |
| if (p.second == InstrumentationLevel::kInstrumentWithInstrumentationStubs) { |
| p.second = InstrumentationLevel::kInstrumentWithInterpreter; |
| } |
| } |
| } |
| } |
| |
| void Instrumentation::ConfigureStubs(const char* key, InstrumentationLevel desired_level) { |
| // Store the instrumentation level for this key or remove it. |
| if (desired_level == InstrumentationLevel::kInstrumentNothing) { |
| // The client no longer needs instrumentation. |
| requested_instrumentation_levels_.erase(key); |
| } else { |
| // The client needs instrumentation. |
| requested_instrumentation_levels_.Overwrite(key, desired_level); |
| } |
| |
| UpdateInstrumentationLevels(desired_level); |
| UpdateStubs(); |
| } |
| |
| void Instrumentation::EnableSingleThreadDeopt() { |
| // Single-thread deopt only uses interpreter. |
| can_use_instrumentation_trampolines_ = false; |
| UpdateInstrumentationLevels(InstrumentationLevel::kInstrumentWithInterpreter); |
| UpdateStubs(); |
| } |
| |
| void Instrumentation::UpdateStubs() { |
| // Look for the highest required instrumentation level. |
| InstrumentationLevel requested_level = InstrumentationLevel::kInstrumentNothing; |
| for (const auto& v : requested_instrumentation_levels_) { |
| requested_level = std::max(requested_level, v.second); |
| } |
| |
| DCHECK(can_use_instrumentation_trampolines_ || |
| requested_level != InstrumentationLevel::kInstrumentWithInstrumentationStubs) |
| << "Use trampolines: " << can_use_instrumentation_trampolines_ << " level " |
| << requested_level; |
| |
| interpret_only_ = (requested_level == InstrumentationLevel::kInstrumentWithInterpreter) || |
| forced_interpret_only_; |
| |
| if (!RequiresInstrumentationInstallation(requested_level)) { |
| // We're already set. |
| return; |
| } |
| Thread* const self = Thread::Current(); |
| Runtime* runtime = Runtime::Current(); |
| Locks::mutator_lock_->AssertExclusiveHeld(self); |
| Locks::thread_list_lock_->AssertNotHeld(self); |
| if (requested_level > InstrumentationLevel::kInstrumentNothing) { |
| if (requested_level == InstrumentationLevel::kInstrumentWithInterpreter) { |
| interpreter_stubs_installed_ = true; |
| entry_exit_stubs_installed_ = true; |
| } else { |
| CHECK_EQ(requested_level, InstrumentationLevel::kInstrumentWithInstrumentationStubs); |
| entry_exit_stubs_installed_ = true; |
| interpreter_stubs_installed_ = false; |
| } |
| InstallStubsClassVisitor visitor(this); |
| runtime->GetClassLinker()->VisitClasses(&visitor); |
| instrumentation_stubs_installed_ = true; |
| MutexLock mu(self, *Locks::thread_list_lock_); |
| runtime->GetThreadList()->ForEach(InstrumentationInstallStack, this); |
| } else { |
| interpreter_stubs_installed_ = false; |
| entry_exit_stubs_installed_ = false; |
| InstallStubsClassVisitor visitor(this); |
| runtime->GetClassLinker()->VisitClasses(&visitor); |
| // Restore stack only if there is no method currently deoptimized. |
| bool empty; |
| { |
| ReaderMutexLock mu(self, *GetDeoptimizedMethodsLock()); |
| empty = IsDeoptimizedMethodsEmpty(); // Avoid lock violation. |
| } |
| if (empty) { |
| MutexLock mu(self, *Locks::thread_list_lock_); |
| bool no_remaining_deopts = true; |
| // Check that there are no other forced deoptimizations. Do it here so we only need to lock |
| // thread_list_lock once. |
| // The compiler gets confused on the thread annotations, so use |
| // NO_THREAD_SAFETY_ANALYSIS. Note that we hold the mutator lock |
| // exclusively at this point. |
| Locks::mutator_lock_->AssertExclusiveHeld(self); |
| runtime->GetThreadList()->ForEach([&](Thread* t) NO_THREAD_SAFETY_ANALYSIS { |
| no_remaining_deopts = |
| no_remaining_deopts && !t->IsForceInterpreter() && |
| std::all_of(t->GetInstrumentationStack()->cbegin(), |
| t->GetInstrumentationStack()->cend(), |
| [&](const auto& frame) REQUIRES_SHARED(Locks::mutator_lock_) { |
| return frame.second.force_deopt_id_ == current_force_deopt_id_; |
| }); |
| }); |
| if (no_remaining_deopts) { |
| Runtime::Current()->GetThreadList()->ForEach(InstrumentationRestoreStack, this); |
| // Only do this after restoring, as walking the stack when restoring will see |
| // the instrumentation exit pc. |
| instrumentation_stubs_installed_ = false; |
| } |
| } |
| } |
| } |
| |
| static void ResetQuickAllocEntryPointsForThread(Thread* thread, void* arg ATTRIBUTE_UNUSED) { |
| thread->ResetQuickAllocEntryPointsForThread(kUseReadBarrier && thread->GetIsGcMarking()); |
| } |
| |
| void Instrumentation::SetEntrypointsInstrumented(bool instrumented) { |
| Thread* self = Thread::Current(); |
| Runtime* runtime = Runtime::Current(); |
| Locks::mutator_lock_->AssertNotHeld(self); |
| Locks::instrument_entrypoints_lock_->AssertHeld(self); |
| if (runtime->IsStarted()) { |
| ScopedSuspendAll ssa(__FUNCTION__); |
| MutexLock mu(self, *Locks::runtime_shutdown_lock_); |
| SetQuickAllocEntryPointsInstrumented(instrumented); |
| ResetQuickAllocEntryPoints(); |
| alloc_entrypoints_instrumented_ = instrumented; |
| } else { |
| MutexLock mu(self, *Locks::runtime_shutdown_lock_); |
| SetQuickAllocEntryPointsInstrumented(instrumented); |
| |
| // Note: ResetQuickAllocEntryPoints only works when the runtime is started. Manually run the |
| // update for just this thread. |
| // Note: self may be null. One of those paths is setting instrumentation in the Heap |
| // constructor for gcstress mode. |
| if (self != nullptr) { |
| ResetQuickAllocEntryPointsForThread(self, nullptr); |
| } |
| |
| alloc_entrypoints_instrumented_ = instrumented; |
| } |
| } |
| |
| void Instrumentation::InstrumentQuickAllocEntryPoints() { |
| MutexLock mu(Thread::Current(), *Locks::instrument_entrypoints_lock_); |
| InstrumentQuickAllocEntryPointsLocked(); |
| } |
| |
| void Instrumentation::UninstrumentQuickAllocEntryPoints() { |
| MutexLock mu(Thread::Current(), *Locks::instrument_entrypoints_lock_); |
| UninstrumentQuickAllocEntryPointsLocked(); |
| } |
| |
| void Instrumentation::InstrumentQuickAllocEntryPointsLocked() { |
| Locks::instrument_entrypoints_lock_->AssertHeld(Thread::Current()); |
| if (quick_alloc_entry_points_instrumentation_counter_ == 0) { |
| SetEntrypointsInstrumented(true); |
| } |
| ++quick_alloc_entry_points_instrumentation_counter_; |
| } |
| |
| void Instrumentation::UninstrumentQuickAllocEntryPointsLocked() { |
| Locks::instrument_entrypoints_lock_->AssertHeld(Thread::Current()); |
| CHECK_GT(quick_alloc_entry_points_instrumentation_counter_, 0U); |
| --quick_alloc_entry_points_instrumentation_counter_; |
| if (quick_alloc_entry_points_instrumentation_counter_ == 0) { |
| SetEntrypointsInstrumented(false); |
| } |
| } |
| |
| void Instrumentation::ResetQuickAllocEntryPoints() { |
| Runtime* runtime = Runtime::Current(); |
| if (runtime->IsStarted()) { |
| MutexLock mu(Thread::Current(), *Locks::thread_list_lock_); |
| runtime->GetThreadList()->ForEach(ResetQuickAllocEntryPointsForThread, nullptr); |
| } |
| } |
| |
| void Instrumentation::UpdateMethodsCodeImpl(ArtMethod* method, const void* quick_code) { |
| const void* new_quick_code; |
| if (LIKELY(!instrumentation_stubs_installed_)) { |
| new_quick_code = quick_code; |
| } else { |
| if ((interpreter_stubs_installed_ || IsDeoptimized(method)) && !method->IsNative()) { |
| new_quick_code = GetQuickToInterpreterBridge(); |
| } else { |
| ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); |
| if (class_linker->IsQuickResolutionStub(quick_code) || |
| class_linker->IsQuickToInterpreterBridge(quick_code)) { |
| new_quick_code = quick_code; |
| } else if (entry_exit_stubs_installed_ && |
| // We need to make sure not to replace anything that InstallStubsForMethod |
| // wouldn't. Specifically we cannot stub out Proxy.<init> since subtypes copy the |
| // implementation directly and this will confuse the instrumentation trampolines. |
| // TODO We should remove the need for this since it makes it impossible to profile |
| // Proxy.<init> correctly in all cases. |
| method != jni::DecodeArtMethod(WellKnownClasses::java_lang_reflect_Proxy_init)) { |
| new_quick_code = GetQuickInstrumentationEntryPoint(); |
| if (!method->IsNative() && Runtime::Current()->GetJit() != nullptr) { |
| // Native methods use trampoline entrypoints during interpreter tracing. |
| DCHECK(!Runtime::Current()->GetJit()->GetCodeCache()->GetGarbageCollectCodeUnsafe()); |
| ProfilingInfo* profiling_info = method->GetProfilingInfo(kRuntimePointerSize); |
| // Tracing will look at the saved entry point in the profiling info to know the actual |
| // entrypoint, so we store it here. |
| if (profiling_info != nullptr) { |
| profiling_info->SetSavedEntryPoint(quick_code); |
| } |
| } |
| } else { |
| new_quick_code = quick_code; |
| } |
| } |
| } |
| UpdateEntrypoints(method, new_quick_code); |
| } |
| |
| void Instrumentation::UpdateNativeMethodsCodeToJitCode(ArtMethod* method, const void* quick_code) { |
| // We don't do any read barrier on `method`'s declaring class in this code, as the JIT might |
| // enter here on a soon-to-be deleted ArtMethod. Updating the entrypoint is OK though, as |
| // the ArtMethod is still in memory. |
| const void* new_quick_code = quick_code; |
| if (UNLIKELY(instrumentation_stubs_installed_) && entry_exit_stubs_installed_) { |
| new_quick_code = GetQuickInstrumentationEntryPoint(); |
| } |
| UpdateEntrypoints(method, new_quick_code); |
| } |
| |
| void Instrumentation::UpdateMethodsCode(ArtMethod* method, const void* quick_code) { |
| DCHECK(method->GetDeclaringClass()->IsResolved()); |
| UpdateMethodsCodeImpl(method, quick_code); |
| } |
| |
| void Instrumentation::UpdateMethodsCodeToInterpreterEntryPoint(ArtMethod* method) { |
| UpdateMethodsCodeImpl(method, GetQuickToInterpreterBridge()); |
| } |
| |
| void Instrumentation::UpdateMethodsCodeForJavaDebuggable(ArtMethod* method, |
| const void* quick_code) { |
| // When the runtime is set to Java debuggable, we may update the entry points of |
| // all methods of a class to the interpreter bridge. A method's declaring class |
| // might not be in resolved state yet in that case, so we bypass the DCHECK in |
| // UpdateMethodsCode. |
| UpdateMethodsCodeImpl(method, quick_code); |
| } |
| |
| bool Instrumentation::AddDeoptimizedMethod(ArtMethod* method) { |
| if (IsDeoptimizedMethod(method)) { |
| // Already in the map. Return. |
| return false; |
| } |
| // Not found. Add it. |
| deoptimized_methods_.insert(method); |
| return true; |
| } |
| |
| bool Instrumentation::IsDeoptimizedMethod(ArtMethod* method) { |
| return deoptimized_methods_.find(method) != deoptimized_methods_.end(); |
| } |
| |
| ArtMethod* Instrumentation::BeginDeoptimizedMethod() { |
| if (deoptimized_methods_.empty()) { |
| // Empty. |
| return nullptr; |
| } |
| return *deoptimized_methods_.begin(); |
| } |
| |
| bool Instrumentation::RemoveDeoptimizedMethod(ArtMethod* method) { |
| auto it = deoptimized_methods_.find(method); |
| if (it == deoptimized_methods_.end()) { |
| return false; |
| } |
| deoptimized_methods_.erase(it); |
| return true; |
| } |
| |
| bool Instrumentation::IsDeoptimizedMethodsEmpty() const { |
| return deoptimized_methods_.empty(); |
| } |
| |
| void Instrumentation::Deoptimize(ArtMethod* method) { |
| CHECK(!method->IsNative()); |
| CHECK(!method->IsProxyMethod()); |
| CHECK(method->IsInvokable()); |
| |
| Thread* self = Thread::Current(); |
| { |
| WriterMutexLock mu(self, *GetDeoptimizedMethodsLock()); |
| bool has_not_been_deoptimized = AddDeoptimizedMethod(method); |
| CHECK(has_not_been_deoptimized) << "Method " << ArtMethod::PrettyMethod(method) |
| << " is already deoptimized"; |
| } |
| if (!interpreter_stubs_installed_) { |
| UpdateEntrypoints(method, GetQuickInstrumentationEntryPoint()); |
| |
| // Install instrumentation exit stub and instrumentation frames. We may already have installed |
| // these previously so it will only cover the newly created frames. |
| instrumentation_stubs_installed_ = true; |
| MutexLock mu(self, *Locks::thread_list_lock_); |
| Runtime::Current()->GetThreadList()->ForEach(InstrumentationInstallStack, this); |
| } |
| } |
| |
| void Instrumentation::Undeoptimize(ArtMethod* method) { |
| CHECK(!method->IsNative()); |
| CHECK(!method->IsProxyMethod()); |
| CHECK(method->IsInvokable()); |
| |
| Thread* self = Thread::Current(); |
| bool empty; |
| { |
| WriterMutexLock mu(self, *GetDeoptimizedMethodsLock()); |
| bool found_and_erased = RemoveDeoptimizedMethod(method); |
| CHECK(found_and_erased) << "Method " << ArtMethod::PrettyMethod(method) |
| << " is not deoptimized"; |
| empty = IsDeoptimizedMethodsEmpty(); |
| } |
| |
| // Restore code and possibly stack only if we did not deoptimize everything. |
| if (!interpreter_stubs_installed_) { |
| // Restore its code or resolution trampoline. |
| ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); |
| if (method->IsStatic() && !method->IsConstructor() && |
| !method->GetDeclaringClass()->IsInitialized()) { |
| UpdateEntrypoints(method, GetQuickResolutionStub()); |
| } else { |
| const void* quick_code = NeedDebugVersionFor(method) |
| ? GetQuickToInterpreterBridge() |
| : class_linker->GetQuickOatCodeFor(method); |
| UpdateEntrypoints(method, quick_code); |
| } |
| |
| // If there is no deoptimized method left, we can restore the stack of each thread. |
| if (empty && !entry_exit_stubs_installed_) { |
| MutexLock mu(self, *Locks::thread_list_lock_); |
| Runtime::Current()->GetThreadList()->ForEach(InstrumentationRestoreStack, this); |
| instrumentation_stubs_installed_ = false; |
| } |
| } |
| } |
| |
| bool Instrumentation::IsDeoptimized(ArtMethod* method) { |
| DCHECK(method != nullptr); |
| ReaderMutexLock mu(Thread::Current(), *GetDeoptimizedMethodsLock()); |
| return IsDeoptimizedMethod(method); |
| } |
| |
| void Instrumentation::EnableDeoptimization() { |
| ReaderMutexLock mu(Thread::Current(), *GetDeoptimizedMethodsLock()); |
| CHECK(IsDeoptimizedMethodsEmpty()); |
| CHECK_EQ(deoptimization_enabled_, false); |
| deoptimization_enabled_ = true; |
| } |
| |
| void Instrumentation::DisableDeoptimization(const char* key) { |
| CHECK_EQ(deoptimization_enabled_, true); |
| // If we deoptimized everything, undo it. |
| InstrumentationLevel level = GetCurrentInstrumentationLevel(); |
| if (level == InstrumentationLevel::kInstrumentWithInterpreter) { |
| UndeoptimizeEverything(key); |
| } |
| // Undeoptimized selected methods. |
| while (true) { |
| ArtMethod* method; |
| { |
| ReaderMutexLock mu(Thread::Current(), *GetDeoptimizedMethodsLock()); |
| if (IsDeoptimizedMethodsEmpty()) { |
| break; |
| } |
| method = BeginDeoptimizedMethod(); |
| CHECK(method != nullptr); |
| } |
| Undeoptimize(method); |
| } |
| deoptimization_enabled_ = false; |
| } |
| |
| // Indicates if instrumentation should notify method enter/exit events to the listeners. |
| bool Instrumentation::ShouldNotifyMethodEnterExitEvents() const { |
| if (!HasMethodEntryListeners() && !HasMethodExitListeners()) { |
| return false; |
| } |
| return !deoptimization_enabled_ && !interpreter_stubs_installed_; |
| } |
| |
| void Instrumentation::DeoptimizeEverything(const char* key) { |
| CHECK(deoptimization_enabled_); |
| ConfigureStubs(key, InstrumentationLevel::kInstrumentWithInterpreter); |
| } |
| |
| void Instrumentation::UndeoptimizeEverything(const char* key) { |
| CHECK(interpreter_stubs_installed_); |
| CHECK(deoptimization_enabled_); |
| ConfigureStubs(key, InstrumentationLevel::kInstrumentNothing); |
| } |
| |
| void Instrumentation::EnableMethodTracing(const char* key, bool needs_interpreter) { |
| InstrumentationLevel level; |
| if (needs_interpreter) { |
| level = InstrumentationLevel::kInstrumentWithInterpreter; |
| } else { |
| level = InstrumentationLevel::kInstrumentWithInstrumentationStubs; |
| } |
| ConfigureStubs(key, level); |
| } |
| |
| void Instrumentation::DisableMethodTracing(const char* key) { |
| ConfigureStubs(key, InstrumentationLevel::kInstrumentNothing); |
| } |
| |
| const void* Instrumentation::GetCodeForInvoke(ArtMethod* method) const { |
| // This is called by instrumentation entry only and that should never be getting proxy methods. |
| DCHECK(!method->IsProxyMethod()) << method->PrettyMethod(); |
| ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); |
| if (LIKELY(!instrumentation_stubs_installed_ && !interpreter_stubs_installed_)) { |
| // In general we just return whatever the method thinks its entrypoint is here. The only |
| // exception is if it still has the instrumentation entrypoint. That means we are racing another |
| // thread getting rid of instrumentation which is unexpected but possible. In that case we want |
| // to wait and try to get it from the oat file or jit. |
| const void* code = method->GetEntryPointFromQuickCompiledCodePtrSize(kRuntimePointerSize); |
| DCHECK(code != nullptr); |
| if (code != GetQuickInstrumentationEntryPoint()) { |
| return code; |
| } else if (method->IsNative()) { |
| return class_linker->GetQuickOatCodeFor(method); |
| } |
| // We don't know what it is. Fallthough to try to find the code from the JIT or Oat file. |
| } else if (method->IsNative()) { |
| // TODO We could have JIT compiled native entrypoints. It might be worth it to find these. |
| return class_linker->GetQuickOatCodeFor(method); |
| } else if (UNLIKELY(interpreter_stubs_installed_)) { |
| return GetQuickToInterpreterBridge(); |
| } |
| // Since the method cannot be native due to ifs above we can always fall back to interpreter |
| // bridge. |
| const void* result = GetQuickToInterpreterBridge(); |
| if (!NeedDebugVersionFor(method)) { |
| // If we don't need a debug version we should see what the oat file/class linker has to say. |
| result = class_linker->GetQuickOatCodeFor(method); |
| } |
| // If both those fail try the jit. |
| if (result == GetQuickToInterpreterBridge()) { |
| jit::Jit* jit = Runtime::Current()->GetJit(); |
| if (jit != nullptr) { |
| const void* res = jit->GetCodeCache()->FindCompiledCodeForInstrumentation(method); |
| if (res != nullptr) { |
| result = res; |
| } |
| } |
| } |
| return result; |
| } |
| |
| const void* Instrumentation::GetQuickCodeFor(ArtMethod* method, PointerSize pointer_size) const { |
| ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); |
| if (LIKELY(!instrumentation_stubs_installed_)) { |
| const void* code = method->GetEntryPointFromQuickCompiledCodePtrSize(pointer_size); |
| DCHECK(code != nullptr); |
| if (LIKELY(!class_linker->IsQuickResolutionStub(code) && |
| !class_linker->IsQuickToInterpreterBridge(code)) && |
| !class_linker->IsQuickResolutionStub(code) && |
| !class_linker->IsQuickToInterpreterBridge(code)) { |
| return code; |
| } |
| } |
| return class_linker->GetQuickOatCodeFor(method); |
| } |
| |
| void Instrumentation::MethodEnterEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc) const { |
| DCHECK(!method->IsRuntimeMethod()); |
| if (HasMethodEntryListeners()) { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| for (InstrumentationListener* listener : method_entry_listeners_) { |
| if (listener != nullptr) { |
| listener->MethodEntered(thread, thiz, method, dex_pc); |
| } |
| } |
| } |
| } |
| |
| template <> |
| void Instrumentation::MethodExitEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| OptionalFrame frame, |
| MutableHandle<mirror::Object>& return_value) const { |
| if (HasMethodExitListeners()) { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| for (InstrumentationListener* listener : method_exit_listeners_) { |
| if (listener != nullptr) { |
| listener->MethodExited(thread, thiz, method, dex_pc, frame, return_value); |
| } |
| } |
| } |
| } |
| |
| template<> void Instrumentation::MethodExitEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| OptionalFrame frame, |
| JValue& return_value) const { |
| if (HasMethodExitListeners()) { |
| Thread* self = Thread::Current(); |
| StackHandleScope<2> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| if (method->GetInterfaceMethodIfProxy(kRuntimePointerSize)->GetReturnTypePrimitive() != |
| Primitive::kPrimNot) { |
| for (InstrumentationListener* listener : method_exit_listeners_) { |
| if (listener != nullptr) { |
| listener->MethodExited(thread, thiz, method, dex_pc, frame, return_value); |
| } |
| } |
| } else { |
| MutableHandle<mirror::Object> ret(hs.NewHandle(return_value.GetL())); |
| MethodExitEventImpl(thread, thiz.Get(), method, dex_pc, frame, ret); |
| return_value.SetL(ret.Get()); |
| } |
| } |
| } |
| |
| void Instrumentation::MethodUnwindEvent(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc) const { |
| if (HasMethodUnwindListeners()) { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| for (InstrumentationListener* listener : method_unwind_listeners_) { |
| if (listener != nullptr) { |
| listener->MethodUnwind(thread, thiz, method, dex_pc); |
| } |
| } |
| } |
| } |
| |
| void Instrumentation::DexPcMovedEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc) const { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| for (InstrumentationListener* listener : dex_pc_listeners_) { |
| if (listener != nullptr) { |
| listener->DexPcMoved(thread, thiz, method, dex_pc); |
| } |
| } |
| } |
| |
| void Instrumentation::BranchImpl(Thread* thread, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| int32_t offset) const { |
| for (InstrumentationListener* listener : branch_listeners_) { |
| if (listener != nullptr) { |
| listener->Branch(thread, method, dex_pc, offset); |
| } |
| } |
| } |
| |
| void Instrumentation::WatchedFramePopImpl(Thread* thread, const ShadowFrame& frame) const { |
| for (InstrumentationListener* listener : watched_frame_pop_listeners_) { |
| if (listener != nullptr) { |
| listener->WatchedFramePop(thread, frame); |
| } |
| } |
| } |
| |
| void Instrumentation::FieldReadEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| ArtField* field) const { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| for (InstrumentationListener* listener : field_read_listeners_) { |
| if (listener != nullptr) { |
| listener->FieldRead(thread, thiz, method, dex_pc, field); |
| } |
| } |
| } |
| |
| void Instrumentation::FieldWriteEventImpl(Thread* thread, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uint32_t dex_pc, |
| ArtField* field, |
| const JValue& field_value) const { |
| Thread* self = Thread::Current(); |
| StackHandleScope<2> hs(self); |
| Handle<mirror::Object> thiz(hs.NewHandle(this_object)); |
| if (field->IsPrimitiveType()) { |
| for (InstrumentationListener* listener : field_write_listeners_) { |
| if (listener != nullptr) { |
| listener->FieldWritten(thread, thiz, method, dex_pc, field, field_value); |
| } |
| } |
| } else { |
| Handle<mirror::Object> val(hs.NewHandle(field_value.GetL())); |
| for (InstrumentationListener* listener : field_write_listeners_) { |
| if (listener != nullptr) { |
| listener->FieldWritten(thread, thiz, method, dex_pc, field, val); |
| } |
| } |
| } |
| } |
| |
| void Instrumentation::ExceptionThrownEvent(Thread* thread, |
| ObjPtr<mirror::Throwable> exception_object) const { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Throwable> h_exception(hs.NewHandle(exception_object)); |
| if (HasExceptionThrownListeners()) { |
| DCHECK_EQ(thread->GetException(), h_exception.Get()); |
| thread->ClearException(); |
| for (InstrumentationListener* listener : exception_thrown_listeners_) { |
| if (listener != nullptr) { |
| listener->ExceptionThrown(thread, h_exception); |
| } |
| } |
| // See b/65049545 for discussion about this behavior. |
| thread->AssertNoPendingException(); |
| thread->SetException(h_exception.Get()); |
| } |
| } |
| |
| void Instrumentation::ExceptionHandledEvent(Thread* thread, |
| ObjPtr<mirror::Throwable> exception_object) const { |
| Thread* self = Thread::Current(); |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Throwable> h_exception(hs.NewHandle(exception_object)); |
| if (HasExceptionHandledListeners()) { |
| // We should have cleared the exception so that callers can detect a new one. |
| DCHECK(thread->GetException() == nullptr); |
| for (InstrumentationListener* listener : exception_handled_listeners_) { |
| if (listener != nullptr) { |
| listener->ExceptionHandled(thread, h_exception); |
| } |
| } |
| } |
| } |
| |
| void Instrumentation::PushInstrumentationStackFrame(Thread* self, |
| ObjPtr<mirror::Object> this_object, |
| ArtMethod* method, |
| uintptr_t stack_ptr, |
| uintptr_t lr, |
| bool interpreter_entry) { |
| DCHECK(!self->IsExceptionPending()); |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| self->GetInstrumentationStack(); |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Entering " << ArtMethod::PrettyMethod(method) << " from PC " |
| << reinterpret_cast<void*>(lr); |
| } |
| |
| // We send the enter event before pushing the instrumentation frame to make cleanup easier. If the |
| // event causes an exception we can simply send the unwind event and return. |
| StackHandleScope<1> hs(self); |
| Handle<mirror::Object> h_this(hs.NewHandle(this_object)); |
| if (!interpreter_entry) { |
| MethodEnterEvent(self, h_this.Get(), method, 0); |
| if (self->IsExceptionPending()) { |
| MethodUnwindEvent(self, h_this.Get(), method, 0); |
| return; |
| } |
| } |
| |
| // We have a callee-save frame meaning this value is guaranteed to never be 0. |
| DCHECK(!self->IsExceptionPending()); |
| size_t frame_id = StackVisitor::ComputeNumFrames(self, kInstrumentationStackWalk); |
| |
| instrumentation::InstrumentationStackFrame instrumentation_frame( |
| h_this.Get(), method, lr, frame_id, interpreter_entry, current_force_deopt_id_); |
| stack->insert({stack_ptr, instrumentation_frame}); |
| } |
| |
| DeoptimizationMethodType Instrumentation::GetDeoptimizationMethodType(ArtMethod* method) { |
| if (method->IsRuntimeMethod()) { |
| // Certain methods have strict requirement on whether the dex instruction |
| // should be re-executed upon deoptimization. |
| if (method == Runtime::Current()->GetCalleeSaveMethod( |
| CalleeSaveType::kSaveEverythingForClinit)) { |
| return DeoptimizationMethodType::kKeepDexPc; |
| } |
| if (method == Runtime::Current()->GetCalleeSaveMethod( |
| CalleeSaveType::kSaveEverythingForSuspendCheck)) { |
| return DeoptimizationMethodType::kKeepDexPc; |
| } |
| } |
| return DeoptimizationMethodType::kDefault; |
| } |
| |
| // Try to get the shorty of a runtime method if it's an invocation stub. |
| static char GetRuntimeMethodShorty(Thread* thread) REQUIRES_SHARED(Locks::mutator_lock_) { |
| char shorty = 'V'; |
| StackVisitor::WalkStack( |
| [&shorty](const art::StackVisitor* stack_visitor) REQUIRES_SHARED(Locks::mutator_lock_) { |
| ArtMethod* m = stack_visitor->GetMethod(); |
| if (m == nullptr || m->IsRuntimeMethod()) { |
| return true; |
| } |
| // The first Java method. |
| if (m->IsNative()) { |
| // Use JNI method's shorty for the jni stub. |
| shorty = m->GetShorty()[0]; |
| } else if (m->IsProxyMethod()) { |
| // Proxy method just invokes its proxied method via |
| // art_quick_proxy_invoke_handler. |
| shorty = m->GetInterfaceMethodIfProxy(kRuntimePointerSize)->GetShorty()[0]; |
| } else { |
| const Instruction& instr = m->DexInstructions().InstructionAt(stack_visitor->GetDexPc()); |
| if (instr.IsInvoke()) { |
| auto get_method_index_fn = [](ArtMethod* caller, |
| const Instruction& inst, |
| uint32_t dex_pc) |
| REQUIRES_SHARED(Locks::mutator_lock_) { |
| switch (inst.Opcode()) { |
| case Instruction::INVOKE_VIRTUAL_RANGE_QUICK: |
| case Instruction::INVOKE_VIRTUAL_QUICK: { |
| uint16_t method_idx = caller->GetIndexFromQuickening(dex_pc); |
| CHECK_NE(method_idx, DexFile::kDexNoIndex16); |
| return method_idx; |
| } |
| default: { |
| return static_cast<uint16_t>(inst.VRegB()); |
| } |
| } |
| }; |
| |
| uint16_t method_index = get_method_index_fn(m, instr, stack_visitor->GetDexPc()); |
| const DexFile* dex_file = m->GetDexFile(); |
| if (interpreter::IsStringInit(dex_file, method_index)) { |
| // Invoking string init constructor is turned into invoking |
| // StringFactory.newStringFromChars() which returns a string. |
| shorty = 'L'; |
| } else { |
| shorty = dex_file->GetMethodShorty(method_index)[0]; |
| } |
| |
| } else { |
| // It could be that a non-invoke opcode invokes a stub, which in turn |
| // invokes Java code. In such cases, we should never expect a return |
| // value from the stub. |
| } |
| } |
| // Stop stack walking since we've seen a Java frame. |
| return false; |
| }, |
| thread, |
| /* context= */ nullptr, |
| art::StackVisitor::StackWalkKind::kIncludeInlinedFrames); |
| return shorty; |
| } |
| |
| TwoWordReturn Instrumentation::PopInstrumentationStackFrame(Thread* self, |
| uintptr_t* return_pc_addr, |
| uint64_t* gpr_result, |
| uint64_t* fpr_result) { |
| DCHECK(gpr_result != nullptr); |
| DCHECK(fpr_result != nullptr); |
| // Do the pop. |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| self->GetInstrumentationStack(); |
| CHECK_GT(stack->size(), 0U); |
| auto it = stack->find(reinterpret_cast<uintptr_t>(return_pc_addr)); |
| CHECK(it != stack->end()); |
| InstrumentationStackFrame instrumentation_frame = it->second; |
| stack->erase(it); |
| |
| // Set return PC and check the sanity of the stack. |
| // We don't cache the return pc value in a local as it may change after |
| // sending a method exit event. |
| *return_pc_addr = instrumentation_frame.return_pc_; |
| self->VerifyStack(); |
| |
| ArtMethod* method = instrumentation_frame.method_; |
| uint32_t length; |
| const PointerSize pointer_size = Runtime::Current()->GetClassLinker()->GetImagePointerSize(); |
| char return_shorty; |
| |
| // Runtime method does not call into MethodExitEvent() so there should not be |
| // suspension point below. |
| ScopedAssertNoThreadSuspension ants(__FUNCTION__, method->IsRuntimeMethod()); |
| if (method->IsRuntimeMethod()) { |
| if (method != Runtime::Current()->GetCalleeSaveMethod( |
| CalleeSaveType::kSaveEverythingForClinit)) { |
| // If the caller is at an invocation point and the runtime method is not |
| // for clinit, we need to pass return results to the caller. |
| // We need the correct shorty to decide whether we need to pass the return |
| // result for deoptimization below. |
| return_shorty = GetRuntimeMethodShorty(self); |
| } else { |
| // Some runtime methods such as allocations, unresolved field getters, etc. |
| // have return value. We don't need to set return_value since MethodExitEvent() |
| // below isn't called for runtime methods. Deoptimization doesn't need the |
| // value either since the dex instruction will be re-executed by the |
| // interpreter, except these two cases: |
| // (1) For an invoke, which is handled above to get the correct shorty. |
| // (2) For MONITOR_ENTER/EXIT, which cannot be re-executed since it's not |
| // idempotent. However there is no return value for it anyway. |
| return_shorty = 'V'; |
| } |
| } else { |
| return_shorty = method->GetInterfaceMethodIfProxy(pointer_size)->GetShorty(&length)[0]; |
| } |
| |
| bool is_ref = return_shorty == '[' || return_shorty == 'L'; |
| StackHandleScope<1> hs(self); |
| MutableHandle<mirror::Object> res(hs.NewHandle<mirror::Object>(nullptr)); |
| JValue return_value; |
| if (return_shorty == 'V') { |
| return_value.SetJ(0); |
| } else if (return_shorty == 'F' || return_shorty == 'D') { |
| return_value.SetJ(*fpr_result); |
| } else { |
| return_value.SetJ(*gpr_result); |
| } |
| if (is_ref) { |
| // Take a handle to the return value so we won't lose it if we suspend. |
| res.Assign(return_value.GetL()); |
| } |
| // TODO: improve the dex pc information here, requires knowledge of current PC as opposed to |
| // return_pc. |
| uint32_t dex_pc = dex::kDexNoIndex; |
| if (!method->IsRuntimeMethod() && !instrumentation_frame.interpreter_entry_) { |
| ObjPtr<mirror::Object> this_object = instrumentation_frame.this_object_; |
| // Note that sending the event may change the contents of *return_pc_addr. |
| MethodExitEvent( |
| self, this_object, instrumentation_frame.method_, dex_pc, OptionalFrame{}, return_value); |
| } |
| |
| // Deoptimize if the caller needs to continue execution in the interpreter. Do nothing if we get |
| // back to an upcall. |
| NthCallerVisitor visitor(self, 1, true); |
| visitor.WalkStack(true); |
| bool deoptimize = (visitor.caller != nullptr) && |
| (interpreter_stubs_installed_ || IsDeoptimized(visitor.caller) || |
| self->IsForceInterpreter() || |
| // NB Since structurally obsolete compiled methods might have the offsets of |
| // methods/fields compiled in we need to go back to interpreter whenever we hit |
| // them. |
| visitor.caller->GetDeclaringClass()->IsObsoleteObject() || |
| // Check if we forced all threads to deoptimize in the time between this frame |
| // being created and now. |
| instrumentation_frame.force_deopt_id_ != current_force_deopt_id_ || |
| Dbg::IsForcedInterpreterNeededForUpcall(self, visitor.caller)); |
| if (is_ref) { |
| // Restore the return value if it's a reference since it might have moved. |
| *reinterpret_cast<mirror::Object**>(gpr_result) = res.Get(); |
| } |
| if (deoptimize && Runtime::Current()->IsAsyncDeoptimizeable(*return_pc_addr)) { |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Deoptimizing " |
| << visitor.caller->PrettyMethod() |
| << " by returning from " |
| << method->PrettyMethod() |
| << " with result " |
| << std::hex << return_value.GetJ() << std::dec |
| << " in " |
| << *self; |
| } |
| DeoptimizationMethodType deopt_method_type = GetDeoptimizationMethodType(method); |
| self->PushDeoptimizationContext(return_value, |
| return_shorty == 'L' || return_shorty == '[', |
| /* exception= */ nullptr , |
| /* from_code= */ false, |
| deopt_method_type); |
| return GetTwoWordSuccessValue(*return_pc_addr, |
| reinterpret_cast<uintptr_t>(GetQuickDeoptimizationEntryPoint())); |
| } else { |
| if (deoptimize && !Runtime::Current()->IsAsyncDeoptimizeable(*return_pc_addr)) { |
| VLOG(deopt) << "Got a deoptimization request on un-deoptimizable " << method->PrettyMethod() |
| << " at PC " << reinterpret_cast<void*>(*return_pc_addr); |
| } |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Returning from " << method->PrettyMethod() |
| << " to PC " << reinterpret_cast<void*>(*return_pc_addr); |
| } |
| return GetTwoWordSuccessValue(0, *return_pc_addr); |
| } |
| } |
| |
| uintptr_t Instrumentation::PopFramesForDeoptimization(Thread* self, uintptr_t pop_until) const { |
| std::map<uintptr_t, instrumentation::InstrumentationStackFrame>* stack = |
| self->GetInstrumentationStack(); |
| // Pop all instrumentation frames below `pop_until`. |
| uintptr_t return_pc = 0u; |
| for (auto i = stack->begin(); i != stack->end() && i->first <= pop_until;) { |
| auto e = i; |
| ++i; |
| if (kVerboseInstrumentation) { |
| LOG(INFO) << "Popping for deoptimization " << e->second.method_->PrettyMethod(); |
| } |
| return_pc = e->second.return_pc_; |
| stack->erase(e); |
| } |
| return return_pc; |
| } |
| |
| std::string InstrumentationStackFrame::Dump() const { |
| std::ostringstream os; |
| os << "Frame " << frame_id_ << " " << ArtMethod::PrettyMethod(method_) << ":" |
| << reinterpret_cast<void*>(return_pc_) << " this=" << reinterpret_cast<void*>(this_object_) |
| << " force_deopt_id=" << force_deopt_id_; |
| return os.str(); |
| } |
| |
| } // namespace instrumentation |
| } // namespace art |