blob: 8e7a638d59b332097d8f7cc5871219b136898c6c [file] [log] [blame]
jeffhao725a9572012-11-13 18:20:12 -08001/*
2 * Copyright (C) 2011 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
Brian Carlstromfc0e3212013-07-17 14:40:12 -070017#ifndef ART_RUNTIME_INSTRUMENTATION_H_
18#define ART_RUNTIME_INSTRUMENTATION_H_
jeffhao725a9572012-11-13 18:20:12 -080019
Ian Rogers576ca0c2014-06-06 15:58:22 -070020#include <stdint.h>
Ian Rogers576ca0c2014-06-06 15:58:22 -070021#include <list>
Mathieu Chartiere401d142015-04-22 13:56:20 -070022#include <unordered_set>
Ian Rogers576ca0c2014-06-06 15:58:22 -070023
Ian Rogersd582fa42014-11-05 23:46:43 -080024#include "arch/instruction_set.h"
Andreas Gampe542451c2016-07-26 09:02:02 -070025#include "base/enums.h"
Elliott Hughes76160052012-12-12 16:31:20 -080026#include "base/macros.h"
Ian Rogers719d1a32014-03-06 12:13:39 -080027#include "base/mutex.h"
David Sehr67bf42e2018-02-26 16:43:04 -080028#include "base/safe_map.h"
Hiroshi Yamauchi94f7b492014-07-22 18:08:23 -070029#include "gc_root.h"
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080030
jeffhao725a9572012-11-13 18:20:12 -080031namespace art {
Ian Rogers2dd0e2c2013-01-24 12:42:14 -080032namespace mirror {
Igor Murashkin2ffb7032017-11-08 13:35:21 -080033class Class;
34class Object;
35class Throwable;
Ian Rogers62d6c772013-02-27 08:32:07 -080036} // namespace mirror
Mathieu Chartierc7853442015-03-27 14:35:38 -070037class ArtField;
Mathieu Chartiere401d142015-04-22 13:56:20 -070038class ArtMethod;
Alex Lightd7661582017-05-01 13:48:16 -070039template <typename T> class Handle;
Ian Rogers62d6c772013-02-27 08:32:07 -080040union JValue;
Alex Lighte814f9d2017-07-31 16:14:39 -070041class ShadowFrame;
jeffhao725a9572012-11-13 18:20:12 -080042class Thread;
Mingyao Yang2ee17902017-08-30 11:37:08 -070043enum class DeoptimizationMethodType;
jeffhao725a9572012-11-13 18:20:12 -080044
Ian Rogers62d6c772013-02-27 08:32:07 -080045namespace instrumentation {
jeffhao725a9572012-11-13 18:20:12 -080046
Sebastien Hertzee1997a2013-09-19 14:47:09 +020047// Interpreter handler tables.
48enum InterpreterHandlerTable {
49 kMainHandlerTable = 0, // Main handler table: no suspend check, no instrumentation.
50 kAlternativeHandlerTable = 1, // Alternative handler table: suspend check and/or instrumentation
51 // enabled.
52 kNumHandlerTables
53};
54
Andreas Gampe40da2862015-02-27 12:49:04 -080055// Do we want to deoptimize for method entry and exit listeners or just try to intercept
56// invocations? Deoptimization forces all code to run in the interpreter and considerably hurts the
57// application's performance.
58static constexpr bool kDeoptimizeForAccurateMethodEntryExitListeners = true;
59
Ian Rogers62d6c772013-02-27 08:32:07 -080060// Instrumentation event listener API. Registered listeners will get the appropriate call back for
61// the events they are listening for. The call backs supply the thread, method and dex_pc the event
62// occurred upon. The thread may or may not be Thread::Current().
63struct InstrumentationListener {
64 InstrumentationListener() {}
65 virtual ~InstrumentationListener() {}
66
67 // Call-back for when a method is entered.
Alex Lightd7661582017-05-01 13:48:16 -070068 virtual void MethodEntered(Thread* thread,
69 Handle<mirror::Object> this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -070070 ArtMethod* method,
Andreas Gampebdf7f1c2016-08-30 16:38:47 -070071 uint32_t dex_pc) REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Ian Rogers62d6c772013-02-27 08:32:07 -080072
Alex Lightd7661582017-05-01 13:48:16 -070073 virtual void MethodExited(Thread* thread,
74 Handle<mirror::Object> this_object,
75 ArtMethod* method,
76 uint32_t dex_pc,
77 Handle<mirror::Object> return_value)
78 REQUIRES_SHARED(Locks::mutator_lock_);
79
80 // Call-back for when a method is exited. The implementor should either handler-ize the return
81 // value (if appropriate) or use the alternate MethodExited callback instead if they need to
82 // go through a suspend point.
83 virtual void MethodExited(Thread* thread,
84 Handle<mirror::Object> this_object,
85 ArtMethod* method,
86 uint32_t dex_pc,
Ian Rogers62d6c772013-02-27 08:32:07 -080087 const JValue& return_value)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -070088 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Ian Rogers62d6c772013-02-27 08:32:07 -080089
90 // Call-back for when a method is popped due to an exception throw. A method will either cause a
91 // MethodExited call-back or a MethodUnwind call-back when its activation is removed.
Alex Lightd7661582017-05-01 13:48:16 -070092 virtual void MethodUnwind(Thread* thread,
93 Handle<mirror::Object> this_object,
94 ArtMethod* method,
95 uint32_t dex_pc)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -070096 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Ian Rogers62d6c772013-02-27 08:32:07 -080097
98 // Call-back for when the dex pc moves in a method.
Alex Lightd7661582017-05-01 13:48:16 -070099 virtual void DexPcMoved(Thread* thread,
100 Handle<mirror::Object> this_object,
101 ArtMethod* method,
102 uint32_t new_dex_pc)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700103 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Ian Rogers62d6c772013-02-27 08:32:07 -0800104
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200105 // Call-back for when we read from a field.
Alex Lightd7661582017-05-01 13:48:16 -0700106 virtual void FieldRead(Thread* thread,
107 Handle<mirror::Object> this_object,
108 ArtMethod* method,
109 uint32_t dex_pc,
110 ArtField* field) = 0;
111
112 virtual void FieldWritten(Thread* thread,
113 Handle<mirror::Object> this_object,
114 ArtMethod* method,
115 uint32_t dex_pc,
116 ArtField* field,
117 Handle<mirror::Object> field_value)
118 REQUIRES_SHARED(Locks::mutator_lock_);
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200119
120 // Call-back for when we write into a field.
Alex Lightd7661582017-05-01 13:48:16 -0700121 virtual void FieldWritten(Thread* thread,
122 Handle<mirror::Object> this_object,
123 ArtMethod* method,
124 uint32_t dex_pc,
125 ArtField* field,
126 const JValue& field_value)
127 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200128
Alex Light6e1607e2017-08-23 10:06:18 -0700129 // Call-back when an exception is thrown.
130 virtual void ExceptionThrown(Thread* thread,
Alex Lightd7661582017-05-01 13:48:16 -0700131 Handle<mirror::Throwable> exception_object)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700132 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Mathieu Chartiere5f13e52015-02-24 09:37:21 -0800133
Alex Light9fb1ab12017-09-05 09:32:49 -0700134 // Call-back when an exception is caught/handled by java code.
135 virtual void ExceptionHandled(Thread* thread, Handle<mirror::Throwable> exception_object)
136 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
137
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000138 // Call-back for when we execute a branch.
139 virtual void Branch(Thread* thread,
140 ArtMethod* method,
141 uint32_t dex_pc,
142 int32_t dex_pc_offset)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700143 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100144
145 // Call-back for when we get an invokevirtual or an invokeinterface.
146 virtual void InvokeVirtualOrInterface(Thread* thread,
Alex Lightd7661582017-05-01 13:48:16 -0700147 Handle<mirror::Object> this_object,
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100148 ArtMethod* caller,
149 uint32_t dex_pc,
150 ArtMethod* callee)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700151 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
Alex Lighte814f9d2017-07-31 16:14:39 -0700152
153 // Call-back when a shadow_frame with the needs_notify_pop_ boolean set is popped off the stack by
154 // either return or exceptions. Normally instrumentation listeners should ensure that there are
155 // shadow-frames by deoptimizing stacks.
156 virtual void WatchedFramePop(Thread* thread ATTRIBUTE_UNUSED,
157 const ShadowFrame& frame ATTRIBUTE_UNUSED)
Alex Light05f47742017-09-14 00:34:44 +0000158 REQUIRES_SHARED(Locks::mutator_lock_) = 0;
jeffhao725a9572012-11-13 18:20:12 -0800159};
160
Ian Rogers62d6c772013-02-27 08:32:07 -0800161// Instrumentation is a catch-all for when extra information is required from the runtime. The
162// typical use for instrumentation is for profiling and debugging. Instrumentation may add stubs
163// to method entry and exit, it may also force execution to be switched to the interpreter and
164// trigger deoptimization.
jeffhao725a9572012-11-13 18:20:12 -0800165class Instrumentation {
166 public:
Ian Rogers62d6c772013-02-27 08:32:07 -0800167 enum InstrumentationEvent {
Mathieu Chartiere5f13e52015-02-24 09:37:21 -0800168 kMethodEntered = 0x1,
169 kMethodExited = 0x2,
170 kMethodUnwind = 0x4,
171 kDexPcMoved = 0x8,
172 kFieldRead = 0x10,
173 kFieldWritten = 0x20,
Alex Light6e1607e2017-08-23 10:06:18 -0700174 kExceptionThrown = 0x40,
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000175 kBranch = 0x80,
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100176 kInvokeVirtualOrInterface = 0x100,
Alex Lighte814f9d2017-07-31 16:14:39 -0700177 kWatchedFramePop = 0x200,
Alex Light9fb1ab12017-09-05 09:32:49 -0700178 kExceptionHandled = 0x400,
Ian Rogers62d6c772013-02-27 08:32:07 -0800179 };
jeffhao725a9572012-11-13 18:20:12 -0800180
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200181 enum class InstrumentationLevel {
182 kInstrumentNothing, // execute without instrumentation
183 kInstrumentWithInstrumentationStubs, // execute with instrumentation entry/exit stubs
184 kInstrumentWithInterpreter // execute with interpreter
185 };
186
Mathieu Chartier3b05e9b2014-03-25 09:29:43 -0700187 Instrumentation();
jeffhao725a9572012-11-13 18:20:12 -0800188
Ian Rogers62d6c772013-02-27 08:32:07 -0800189 // Add a listener to be notified of the masked together sent of instrumentation events. This
190 // suspend the runtime to install stubs. You are expected to hold the mutator lock as a proxy
191 // for saying you should have suspended all threads (installing stubs while threads are running
192 // will break).
193 void AddListener(InstrumentationListener* listener, uint32_t events)
Mathieu Chartier90443472015-07-16 20:32:27 -0700194 REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !Locks::classlinker_classes_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800195
Ian Rogers62d6c772013-02-27 08:32:07 -0800196 // Removes a listener possibly removing instrumentation stubs.
197 void RemoveListener(InstrumentationListener* listener, uint32_t events)
Mathieu Chartier90443472015-07-16 20:32:27 -0700198 REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !Locks::classlinker_classes_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800199
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100200 // Deoptimization.
Sebastien Hertza76a6d42014-03-20 16:40:17 +0100201 void EnableDeoptimization()
Mathieu Chartieraa516822015-10-02 15:53:37 -0700202 REQUIRES(Locks::mutator_lock_)
203 REQUIRES(!deoptimized_methods_lock_);
204 // Calls UndeoptimizeEverything which may visit class linker classes through ConfigureStubs.
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200205 void DisableDeoptimization(const char* key)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700206 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
207 REQUIRES(!deoptimized_methods_lock_);
208
Sebastien Hertza76a6d42014-03-20 16:40:17 +0100209 bool AreAllMethodsDeoptimized() const {
210 return interpreter_stubs_installed_;
211 }
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700212 bool ShouldNotifyMethodEnterExitEvents() const REQUIRES_SHARED(Locks::mutator_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100213
Alex Lightbebd7bd2017-07-25 14:05:52 -0700214 bool CanDeoptimize() {
215 return deoptimization_enabled_;
216 }
217
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100218 // Executes everything with interpreter.
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200219 void DeoptimizeEverything(const char* key)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700220 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
221 REQUIRES(!Locks::thread_list_lock_,
222 !Locks::classlinker_classes_lock_,
Mathieu Chartier90443472015-07-16 20:32:27 -0700223 !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100224
Mathieu Chartieraa516822015-10-02 15:53:37 -0700225 // Executes everything with compiled code (or interpreter if there is no code). May visit class
226 // linker classes through ConfigureStubs.
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200227 void UndeoptimizeEverything(const char* key)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700228 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
229 REQUIRES(!Locks::thread_list_lock_,
230 !Locks::classlinker_classes_lock_,
Mathieu Chartier90443472015-07-16 20:32:27 -0700231 !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100232
233 // Deoptimize a method by forcing its execution with the interpreter. Nevertheless, a static
234 // method (except a class initializer) set to the resolution trampoline will be deoptimized only
235 // once its declaring class is initialized.
Mathieu Chartiere401d142015-04-22 13:56:20 -0700236 void Deoptimize(ArtMethod* method)
Mathieu Chartier90443472015-07-16 20:32:27 -0700237 REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100238
239 // Undeoptimze the method by restoring its entrypoints. Nevertheless, a static method
240 // (except a class initializer) set to the resolution trampoline will be updated only once its
241 // declaring class is initialized.
Mathieu Chartiere401d142015-04-22 13:56:20 -0700242 void Undeoptimize(ArtMethod* method)
Mathieu Chartier90443472015-07-16 20:32:27 -0700243 REQUIRES(Locks::mutator_lock_, !Locks::thread_list_lock_, !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100244
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200245 // Indicates whether the method has been deoptimized so it is executed with the interpreter.
Mathieu Chartiere401d142015-04-22 13:56:20 -0700246 bool IsDeoptimized(ArtMethod* method)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700247 REQUIRES(!deoptimized_methods_lock_) REQUIRES_SHARED(Locks::mutator_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100248
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200249 // Enable method tracing by installing instrumentation entry/exit stubs or interpreter.
250 void EnableMethodTracing(const char* key,
251 bool needs_interpreter = kDeoptimizeForAccurateMethodEntryExitListeners)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700252 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
253 REQUIRES(!Locks::thread_list_lock_,
254 !Locks::classlinker_classes_lock_,
Mathieu Chartier90443472015-07-16 20:32:27 -0700255 !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100256
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200257 // Disable method tracing by uninstalling instrumentation entry/exit stubs or interpreter.
258 void DisableMethodTracing(const char* key)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700259 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
260 REQUIRES(!Locks::thread_list_lock_,
261 !Locks::classlinker_classes_lock_,
Mathieu Chartier90443472015-07-16 20:32:27 -0700262 !deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100263
Sebastien Hertzed2be172014-08-19 15:33:43 +0200264 InterpreterHandlerTable GetInterpreterHandlerTable() const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700265 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertzee1997a2013-09-19 14:47:09 +0200266 return interpreter_handler_table_;
267 }
268
Mathieu Chartier90443472015-07-16 20:32:27 -0700269 void InstrumentQuickAllocEntryPoints() REQUIRES(!Locks::instrument_entrypoints_lock_);
270 void UninstrumentQuickAllocEntryPoints() REQUIRES(!Locks::instrument_entrypoints_lock_);
Mathieu Chartier9ef78b52014-09-25 17:03:12 -0700271 void InstrumentQuickAllocEntryPointsLocked()
Mathieu Chartier90443472015-07-16 20:32:27 -0700272 REQUIRES(Locks::instrument_entrypoints_lock_, !Locks::thread_list_lock_,
273 !Locks::runtime_shutdown_lock_);
Mathieu Chartier9ef78b52014-09-25 17:03:12 -0700274 void UninstrumentQuickAllocEntryPointsLocked()
Mathieu Chartier90443472015-07-16 20:32:27 -0700275 REQUIRES(Locks::instrument_entrypoints_lock_, !Locks::thread_list_lock_,
276 !Locks::runtime_shutdown_lock_);
277 void ResetQuickAllocEntryPoints() REQUIRES(Locks::runtime_shutdown_lock_);
Ian Rogersfa824272013-11-05 16:12:57 -0800278
Ian Rogers62d6c772013-02-27 08:32:07 -0800279 // Update the code of a method respecting any installed stubs.
Mathieu Chartiere401d142015-04-22 13:56:20 -0700280 void UpdateMethodsCode(ArtMethod* method, const void* quick_code)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700281 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800282
Nicolas Geoffraya6e0e7d2018-01-26 13:16:50 +0000283 // Update the code of a native method to a JITed stub.
284 void UpdateNativeMethodsCodeToJitCode(ArtMethod* method, const void* quick_code)
285 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
286
Alex Light0a5ec3d2017-07-25 16:50:26 -0700287 // Update the code of a method to the interpreter respecting any installed stubs from debugger.
288 void UpdateMethodsCodeToInterpreterEntryPoint(ArtMethod* method)
289 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
290
Mingyao Yang3fd448a2016-05-10 14:30:41 -0700291 // Update the code of a method respecting any installed stubs from debugger.
Nicolas Geoffraya0619e22016-12-20 13:57:43 +0000292 void UpdateMethodsCodeForJavaDebuggable(ArtMethod* method, const void* quick_code)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700293 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
Mingyao Yang3fd448a2016-05-10 14:30:41 -0700294
Ian Rogers62d6c772013-02-27 08:32:07 -0800295 // Get the quick code for the given method. More efficient than asking the class linker as it
296 // will short-cut to GetCode if instrumentation and static method resolution stubs aren't
297 // installed.
Andreas Gampe542451c2016-07-26 09:02:02 -0700298 const void* GetQuickCodeFor(ArtMethod* method, PointerSize pointer_size) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700299 REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800300
301 void ForceInterpretOnly() {
302 interpret_only_ = true;
303 forced_interpret_only_ = true;
304 }
305
Brian Carlstromea46f952013-07-30 01:26:50 -0700306 // Called by ArtMethod::Invoke to determine dispatch mechanism.
Ian Rogers62d6c772013-02-27 08:32:07 -0800307 bool InterpretOnly() const {
308 return interpret_only_;
309 }
310
Hiroshi Yamauchi563b47c2014-02-28 17:18:37 -0800311 bool IsForcedInterpretOnly() const {
312 return forced_interpret_only_;
313 }
314
Mingyao Yang6ea1a0e2016-01-29 12:12:49 -0800315 // Code is in boot image oat file which isn't compiled as debuggable.
316 // Need debug version (interpreter or jitted) if that's the case.
Nicolas Geoffraya0619e22016-12-20 13:57:43 +0000317 bool NeedDebugVersionFor(ArtMethod* method) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700318 REQUIRES_SHARED(Locks::mutator_lock_);
Mingyao Yang6ea1a0e2016-01-29 12:12:49 -0800319
Ian Rogers62d6c772013-02-27 08:32:07 -0800320 bool AreExitStubsInstalled() const {
321 return instrumentation_stubs_installed_;
322 }
323
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700324 bool HasMethodEntryListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200325 return have_method_entry_listeners_;
326 }
327
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700328 bool HasMethodExitListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200329 return have_method_exit_listeners_;
330 }
331
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700332 bool HasMethodUnwindListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200333 return have_method_unwind_listeners_;
334 }
335
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700336 bool HasDexPcListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200337 return have_dex_pc_listeners_;
338 }
339
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700340 bool HasFieldReadListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200341 return have_field_read_listeners_;
342 }
343
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700344 bool HasFieldWriteListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200345 return have_field_write_listeners_;
346 }
347
Alex Light6e1607e2017-08-23 10:06:18 -0700348 bool HasExceptionThrownListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
349 return have_exception_thrown_listeners_;
Sebastien Hertz9f102032014-05-23 08:59:42 +0200350 }
351
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700352 bool HasBranchListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000353 return have_branch_listeners_;
Mathieu Chartiere5f13e52015-02-24 09:37:21 -0800354 }
355
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700356 bool HasInvokeVirtualOrInterfaceListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100357 return have_invoke_virtual_or_interface_listeners_;
358 }
359
Alex Lighte814f9d2017-07-31 16:14:39 -0700360 bool HasWatchedFramePopListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
361 return have_watched_frame_pop_listeners_;
362 }
363
Alex Light9fb1ab12017-09-05 09:32:49 -0700364 bool HasExceptionHandledListeners() const REQUIRES_SHARED(Locks::mutator_lock_) {
365 return have_exception_handled_listeners_;
366 }
367
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700368 bool IsActive() const REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertzee1997a2013-09-19 14:47:09 +0200369 return have_dex_pc_listeners_ || have_method_entry_listeners_ || have_method_exit_listeners_ ||
Sebastien Hertz42cd43f2014-05-13 14:15:41 +0200370 have_field_read_listeners_ || have_field_write_listeners_ ||
Alex Light6e1607e2017-08-23 10:06:18 -0700371 have_exception_thrown_listeners_ || have_method_unwind_listeners_ ||
Alex Lighte814f9d2017-07-31 16:14:39 -0700372 have_branch_listeners_ || have_invoke_virtual_or_interface_listeners_ ||
Alex Light9fb1ab12017-09-05 09:32:49 -0700373 have_watched_frame_pop_listeners_ || have_exception_handled_listeners_;
Bill Buzbeefd522f92016-02-11 22:37:42 +0000374 }
375
376 // Any instrumentation *other* than what is needed for Jit profiling active?
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700377 bool NonJitProfilingActive() const REQUIRES_SHARED(Locks::mutator_lock_) {
Bill Buzbeefd522f92016-02-11 22:37:42 +0000378 return have_dex_pc_listeners_ || have_method_exit_listeners_ ||
379 have_field_read_listeners_ || have_field_write_listeners_ ||
Alex Light6e1607e2017-08-23 10:06:18 -0700380 have_exception_thrown_listeners_ || have_method_unwind_listeners_ ||
Alex Light9fb1ab12017-09-05 09:32:49 -0700381 have_branch_listeners_ || have_watched_frame_pop_listeners_ ||
382 have_exception_handled_listeners_;
Sebastien Hertzee1997a2013-09-19 14:47:09 +0200383 }
384
Ian Rogers62d6c772013-02-27 08:32:07 -0800385 // Inform listeners that a method has been entered. A dex PC is provided as we may install
386 // listeners into executing code and get method enter events for methods already on the stack.
387 void MethodEnterEvent(Thread* thread, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700388 ArtMethod* method, uint32_t dex_pc) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700389 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200390 if (UNLIKELY(HasMethodEntryListeners())) {
Ian Rogers62d6c772013-02-27 08:32:07 -0800391 MethodEnterEventImpl(thread, this_object, method, dex_pc);
392 }
393 }
394
395 // Inform listeners that a method has been exited.
Alex Lightd7661582017-05-01 13:48:16 -0700396 void MethodExitEvent(Thread* thread,
397 mirror::Object* this_object,
398 ArtMethod* method,
399 uint32_t dex_pc,
Ian Rogers62d6c772013-02-27 08:32:07 -0800400 const JValue& return_value) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700401 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200402 if (UNLIKELY(HasMethodExitListeners())) {
Ian Rogers62d6c772013-02-27 08:32:07 -0800403 MethodExitEventImpl(thread, this_object, method, dex_pc, return_value);
404 }
405 }
406
407 // Inform listeners that a method has been exited due to an exception.
408 void MethodUnwindEvent(Thread* thread, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700409 ArtMethod* method, uint32_t dex_pc) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700410 REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800411
412 // Inform listeners that the dex pc has moved (only supported by the interpreter).
413 void DexPcMovedEvent(Thread* thread, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700414 ArtMethod* method, uint32_t dex_pc) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700415 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz74109f62013-06-07 17:40:09 +0200416 if (UNLIKELY(HasDexPcListeners())) {
Ian Rogers62d6c772013-02-27 08:32:07 -0800417 DexPcMovedEventImpl(thread, this_object, method, dex_pc);
418 }
419 }
420
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000421 // Inform listeners that a branch has been taken (only supported by the interpreter).
422 void Branch(Thread* thread, ArtMethod* method, uint32_t dex_pc, int32_t offset) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700423 REQUIRES_SHARED(Locks::mutator_lock_) {
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000424 if (UNLIKELY(HasBranchListeners())) {
425 BranchImpl(thread, method, dex_pc, offset);
Mathieu Chartiere5f13e52015-02-24 09:37:21 -0800426 }
427 }
428
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200429 // Inform listeners that we read a field (only supported by the interpreter).
430 void FieldReadEvent(Thread* thread, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700431 ArtMethod* method, uint32_t dex_pc,
Mathieu Chartierc7853442015-03-27 14:35:38 -0700432 ArtField* field) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700433 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200434 if (UNLIKELY(HasFieldReadListeners())) {
435 FieldReadEventImpl(thread, this_object, method, dex_pc, field);
436 }
437 }
438
439 // Inform listeners that we write a field (only supported by the interpreter).
440 void FieldWriteEvent(Thread* thread, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700441 ArtMethod* method, uint32_t dex_pc,
Mathieu Chartierc7853442015-03-27 14:35:38 -0700442 ArtField* field, const JValue& field_value) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700443 REQUIRES_SHARED(Locks::mutator_lock_) {
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200444 if (UNLIKELY(HasFieldWriteListeners())) {
445 FieldWriteEventImpl(thread, this_object, method, dex_pc, field, field_value);
446 }
447 }
448
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100449 void InvokeVirtualOrInterface(Thread* thread,
450 mirror::Object* this_object,
451 ArtMethod* caller,
452 uint32_t dex_pc,
453 ArtMethod* callee) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700454 REQUIRES_SHARED(Locks::mutator_lock_) {
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100455 if (UNLIKELY(HasInvokeVirtualOrInterfaceListeners())) {
456 InvokeVirtualOrInterfaceImpl(thread, this_object, caller, dex_pc, callee);
457 }
458 }
459
Alex Lighte814f9d2017-07-31 16:14:39 -0700460 // Inform listeners that a branch has been taken (only supported by the interpreter).
461 void WatchedFramePopped(Thread* thread, const ShadowFrame& frame) const
462 REQUIRES_SHARED(Locks::mutator_lock_) {
463 if (UNLIKELY(HasWatchedFramePopListeners())) {
464 WatchedFramePopImpl(thread, frame);
465 }
466 }
467
Alex Light6e1607e2017-08-23 10:06:18 -0700468 // Inform listeners that an exception was thrown.
469 void ExceptionThrownEvent(Thread* thread, mirror::Throwable* exception_object) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700470 REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800471
Alex Light9fb1ab12017-09-05 09:32:49 -0700472 // Inform listeners that an exception has been handled. This is not sent for native code or for
473 // exceptions which reach the end of the thread's stack.
474 void ExceptionHandledEvent(Thread* thread, mirror::Throwable* exception_object) const
475 REQUIRES_SHARED(Locks::mutator_lock_);
476
Ian Rogers62d6c772013-02-27 08:32:07 -0800477 // Called when an instrumented method is entered. The intended link register (lr) is saved so
478 // that returning causes a branch to the method exit stub. Generates method enter events.
479 void PushInstrumentationStackFrame(Thread* self, mirror::Object* this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700480 ArtMethod* method, uintptr_t lr,
Jeff Hao9a916d32013-06-27 18:45:37 -0700481 bool interpreter_entry)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700482 REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800483
Mingyao Yang2ee17902017-08-30 11:37:08 -0700484 DeoptimizationMethodType GetDeoptimizationMethodType(ArtMethod* method)
485 REQUIRES_SHARED(Locks::mutator_lock_);
486
Ian Rogers62d6c772013-02-27 08:32:07 -0800487 // Called when an instrumented method is exited. Removes the pushed instrumentation frame
Alex Lightb7edcda2017-04-27 13:20:31 -0700488 // returning the intended link register. Generates method exit events. The gpr_result and
489 // fpr_result pointers are pointers to the locations where the integer/pointer and floating point
490 // result values of the function are stored. Both pointers must always be valid but the values
491 // held there will only be meaningful if interpreted as the appropriate type given the function
492 // being returned from.
Andreas Gamped58342c2014-06-05 14:18:08 -0700493 TwoWordReturn PopInstrumentationStackFrame(Thread* self, uintptr_t* return_pc,
Alex Lightb7edcda2017-04-27 13:20:31 -0700494 uint64_t* gpr_result, uint64_t* fpr_result)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700495 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800496
497 // Pops an instrumentation frame from the current thread and generate an unwind event.
Mingyao Yangf711f2c2016-05-23 12:29:39 -0700498 // Returns the return pc for the instrumentation frame that's popped.
499 uintptr_t PopMethodForUnwind(Thread* self, bool is_deoptimization) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700500 REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800501
502 // Call back for configure stubs.
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700503 void InstallStubsForClass(mirror::Class* klass) REQUIRES_SHARED(Locks::mutator_lock_)
Mathieu Chartier90443472015-07-16 20:32:27 -0700504 REQUIRES(!deoptimized_methods_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800505
Mathieu Chartiere401d142015-04-22 13:56:20 -0700506 void InstallStubsForMethod(ArtMethod* method)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700507 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100508
Mingyao Yang99170c62015-07-06 11:10:37 -0700509 // Install instrumentation exit stub on every method of the stack of the given thread.
510 // This is used by the debugger to cause a deoptimization of the thread's stack after updating
511 // local variable(s).
512 void InstrumentThreadStack(Thread* thread)
Alex Light3ae82532017-07-26 13:59:07 -0700513 REQUIRES_SHARED(Locks::mutator_lock_);
Mingyao Yang99170c62015-07-06 11:10:37 -0700514
Sebastien Hertzb2feaaf2015-10-12 13:40:10 +0000515 static size_t ComputeFrameId(Thread* self,
516 size_t frame_depth,
517 size_t inlined_frames_before_frame)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700518 REQUIRES_SHARED(Locks::mutator_lock_);
Sebastien Hertzb2feaaf2015-10-12 13:40:10 +0000519
Mathieu Chartiereebc3af2016-02-29 18:13:38 -0800520 // Does not hold lock, used to check if someone changed from not instrumented to instrumented
521 // during a GC suspend point.
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700522 bool AllocEntrypointsInstrumented() const REQUIRES_SHARED(Locks::mutator_lock_) {
Mathieu Chartier50e93312016-03-16 11:25:29 -0700523 return alloc_entrypoints_instrumented_;
Mathieu Chartiereebc3af2016-02-29 18:13:38 -0800524 }
525
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200526 InstrumentationLevel GetCurrentInstrumentationLevel() const;
527
Alex Lightdba61482016-12-21 08:20:29 -0800528 private:
529 // Returns true if moving to the given instrumentation level requires the installation of stubs.
530 // False otherwise.
531 bool RequiresInstrumentationInstallation(InstrumentationLevel new_level) const;
532
Ian Rogers62d6c772013-02-27 08:32:07 -0800533 // Does the job of installing or removing instrumentation code within methods.
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200534 // In order to support multiple clients using instrumentation at the same time,
535 // the caller must pass a unique key (a string) identifying it so we remind which
536 // instrumentation level it needs. Therefore the current instrumentation level
537 // becomes the highest instrumentation level required by a client.
538 void ConfigureStubs(const char* key, InstrumentationLevel desired_instrumentation_level)
Mathieu Chartieraa516822015-10-02 15:53:37 -0700539 REQUIRES(Locks::mutator_lock_, Roles::uninterruptible_)
540 REQUIRES(!deoptimized_methods_lock_,
541 !Locks::thread_list_lock_,
Mathieu Chartier90443472015-07-16 20:32:27 -0700542 !Locks::classlinker_classes_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800543
Mathieu Chartier90443472015-07-16 20:32:27 -0700544 void UpdateInterpreterHandlerTable() REQUIRES(Locks::mutator_lock_) {
buzbee1452bee2015-03-06 14:43:04 -0800545 /*
546 * TUNING: Dalvik's mterp stashes the actual current handler table base in a
547 * tls field. For Arm, this enables all suspend, debug & tracing checks to be
548 * collapsed into a single conditionally-executed ldw instruction.
549 * Move to Dalvik-style handler-table management for both the goto interpreter and
550 * mterp.
551 */
Sebastien Hertzee1997a2013-09-19 14:47:09 +0200552 interpreter_handler_table_ = IsActive() ? kAlternativeHandlerTable : kMainHandlerTable;
553 }
554
Mathieu Chartier661974a2014-01-09 11:23:53 -0800555 // No thread safety analysis to get around SetQuickAllocEntryPointsInstrumented requiring
556 // exclusive access to mutator lock which you can't get if the runtime isn't started.
Mathieu Chartier9ef78b52014-09-25 17:03:12 -0700557 void SetEntrypointsInstrumented(bool instrumented) NO_THREAD_SAFETY_ANALYSIS;
Mathieu Chartier661974a2014-01-09 11:23:53 -0800558
Alex Lightd7661582017-05-01 13:48:16 -0700559 void MethodEnterEventImpl(Thread* thread,
560 ObjPtr<mirror::Object> this_object,
561 ArtMethod* method,
562 uint32_t dex_pc) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700563 REQUIRES_SHARED(Locks::mutator_lock_);
Alex Lightd7661582017-05-01 13:48:16 -0700564 void MethodExitEventImpl(Thread* thread,
565 ObjPtr<mirror::Object> this_object,
Mathieu Chartiere401d142015-04-22 13:56:20 -0700566 ArtMethod* method,
Alex Lightd7661582017-05-01 13:48:16 -0700567 uint32_t dex_pc,
568 const JValue& return_value) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700569 REQUIRES_SHARED(Locks::mutator_lock_);
Alex Lightd7661582017-05-01 13:48:16 -0700570 void DexPcMovedEventImpl(Thread* thread,
571 ObjPtr<mirror::Object> this_object,
572 ArtMethod* method,
573 uint32_t dex_pc) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700574 REQUIRES_SHARED(Locks::mutator_lock_);
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000575 void BranchImpl(Thread* thread, ArtMethod* method, uint32_t dex_pc, int32_t offset) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700576 REQUIRES_SHARED(Locks::mutator_lock_);
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100577 void InvokeVirtualOrInterfaceImpl(Thread* thread,
Alex Lightd7661582017-05-01 13:48:16 -0700578 ObjPtr<mirror::Object> this_object,
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100579 ArtMethod* caller,
580 uint32_t dex_pc,
581 ArtMethod* callee) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700582 REQUIRES_SHARED(Locks::mutator_lock_);
Alex Lighte814f9d2017-07-31 16:14:39 -0700583 void WatchedFramePopImpl(Thread* thread, const ShadowFrame& frame) const
584 REQUIRES_SHARED(Locks::mutator_lock_);
Alex Lightd7661582017-05-01 13:48:16 -0700585 void FieldReadEventImpl(Thread* thread,
586 ObjPtr<mirror::Object> this_object,
587 ArtMethod* method,
588 uint32_t dex_pc,
589 ArtField* field) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700590 REQUIRES_SHARED(Locks::mutator_lock_);
Alex Lightd7661582017-05-01 13:48:16 -0700591 void FieldWriteEventImpl(Thread* thread,
592 ObjPtr<mirror::Object> this_object,
593 ArtMethod* method,
594 uint32_t dex_pc,
595 ArtField* field,
596 const JValue& field_value) const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700597 REQUIRES_SHARED(Locks::mutator_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800598
Hiroshi Yamauchi799eb3a2014-07-18 15:38:17 -0700599 // Read barrier-aware utility functions for accessing deoptimized_methods_
Mathieu Chartiere401d142015-04-22 13:56:20 -0700600 bool AddDeoptimizedMethod(ArtMethod* method)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700601 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(deoptimized_methods_lock_);
Mathieu Chartiere401d142015-04-22 13:56:20 -0700602 bool IsDeoptimizedMethod(ArtMethod* method)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700603 REQUIRES_SHARED(Locks::mutator_lock_, deoptimized_methods_lock_);
Mathieu Chartiere401d142015-04-22 13:56:20 -0700604 bool RemoveDeoptimizedMethod(ArtMethod* method)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700605 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(deoptimized_methods_lock_);
Mathieu Chartiere401d142015-04-22 13:56:20 -0700606 ArtMethod* BeginDeoptimizedMethod()
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700607 REQUIRES_SHARED(Locks::mutator_lock_, deoptimized_methods_lock_);
Hiroshi Yamauchi799eb3a2014-07-18 15:38:17 -0700608 bool IsDeoptimizedMethodsEmpty() const
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700609 REQUIRES_SHARED(Locks::mutator_lock_, deoptimized_methods_lock_);
Mingyao Yang3fd448a2016-05-10 14:30:41 -0700610 void UpdateMethodsCodeImpl(ArtMethod* method, const void* quick_code)
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700611 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!deoptimized_methods_lock_);
Mingyao Yang3fd448a2016-05-10 14:30:41 -0700612
Hiroshi Yamauchi799eb3a2014-07-18 15:38:17 -0700613
Brian Carlstromea46f952013-07-30 01:26:50 -0700614 // Have we hijacked ArtMethod::code_ so that it calls instrumentation/interpreter code?
Ian Rogers62d6c772013-02-27 08:32:07 -0800615 bool instrumentation_stubs_installed_;
616
Brian Carlstromea46f952013-07-30 01:26:50 -0700617 // Have we hijacked ArtMethod::code_ to reference the enter/exit stubs?
Ian Rogers62d6c772013-02-27 08:32:07 -0800618 bool entry_exit_stubs_installed_;
619
Brian Carlstromea46f952013-07-30 01:26:50 -0700620 // Have we hijacked ArtMethod::code_ to reference the enter interpreter stub?
Ian Rogers62d6c772013-02-27 08:32:07 -0800621 bool interpreter_stubs_installed_;
622
623 // Do we need the fidelity of events that we only get from running within the interpreter?
624 bool interpret_only_;
625
626 // Did the runtime request we only run in the interpreter? ie -Xint mode.
627 bool forced_interpret_only_;
628
629 // Do we have any listeners for method entry events? Short-cut to avoid taking the
630 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200631 bool have_method_entry_listeners_ GUARDED_BY(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800632
633 // Do we have any listeners for method exit events? Short-cut to avoid taking the
634 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200635 bool have_method_exit_listeners_ GUARDED_BY(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800636
637 // Do we have any listeners for method unwind events? Short-cut to avoid taking the
638 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200639 bool have_method_unwind_listeners_ GUARDED_BY(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800640
641 // Do we have any listeners for dex move events? Short-cut to avoid taking the
642 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200643 bool have_dex_pc_listeners_ GUARDED_BY(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800644
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200645 // Do we have any listeners for field read events? Short-cut to avoid taking the
646 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200647 bool have_field_read_listeners_ GUARDED_BY(Locks::mutator_lock_);
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200648
649 // Do we have any listeners for field write events? Short-cut to avoid taking the
650 // instrumentation_lock_.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200651 bool have_field_write_listeners_ GUARDED_BY(Locks::mutator_lock_);
Sebastien Hertz3f52eaf2014-04-04 17:50:18 +0200652
Alex Light6e1607e2017-08-23 10:06:18 -0700653 // Do we have any exception thrown listeners? Short-cut to avoid taking the instrumentation_lock_.
654 bool have_exception_thrown_listeners_ GUARDED_BY(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800655
Alex Lighte814f9d2017-07-31 16:14:39 -0700656 // Do we have any frame pop listeners? Short-cut to avoid taking the instrumentation_lock_.
657 bool have_watched_frame_pop_listeners_ GUARDED_BY(Locks::mutator_lock_);
658
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000659 // Do we have any branch listeners? Short-cut to avoid taking the instrumentation_lock_.
660 bool have_branch_listeners_ GUARDED_BY(Locks::mutator_lock_);
Mathieu Chartiere5f13e52015-02-24 09:37:21 -0800661
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100662 // Do we have any invoke listeners? Short-cut to avoid taking the instrumentation_lock_.
663 bool have_invoke_virtual_or_interface_listeners_ GUARDED_BY(Locks::mutator_lock_);
664
Alex Light9fb1ab12017-09-05 09:32:49 -0700665 // Do we have any exception handled listeners? Short-cut to avoid taking the
666 // instrumentation_lock_.
667 bool have_exception_handled_listeners_ GUARDED_BY(Locks::mutator_lock_);
668
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200669 // Contains the instrumentation level required by each client of the instrumentation identified
670 // by a string key.
671 typedef SafeMap<const char*, InstrumentationLevel> InstrumentationLevelTable;
672 InstrumentationLevelTable requested_instrumentation_levels_ GUARDED_BY(Locks::mutator_lock_);
673
Ian Rogers62d6c772013-02-27 08:32:07 -0800674 // The event listeners, written to with the mutator_lock_ exclusively held.
Nicolas Geoffray514a6162015-11-03 11:44:24 +0000675 // Mutators must be able to iterate over these lists concurrently, that is, with listeners being
676 // added or removed while iterating. The modifying thread holds exclusive lock,
677 // so other threads cannot iterate (i.e. read the data of the list) at the same time but they
678 // do keep iterators that need to remain valid. This is the reason these listeners are std::list
679 // and not for example std::vector: the existing storage for a std::list does not move.
680 // Note that mutators cannot make a copy of these lists before iterating, as the instrumentation
681 // listeners can also be deleted concurrently.
682 // As a result, these lists are never trimmed. That's acceptable given the low number of
683 // listeners we have.
Ian Rogers62d6c772013-02-27 08:32:07 -0800684 std::list<InstrumentationListener*> method_entry_listeners_ GUARDED_BY(Locks::mutator_lock_);
685 std::list<InstrumentationListener*> method_exit_listeners_ GUARDED_BY(Locks::mutator_lock_);
686 std::list<InstrumentationListener*> method_unwind_listeners_ GUARDED_BY(Locks::mutator_lock_);
Nicolas Geoffray81f0f952016-01-20 16:25:19 +0000687 std::list<InstrumentationListener*> branch_listeners_ GUARDED_BY(Locks::mutator_lock_);
Nicolas Geoffray5550ca82015-08-21 18:38:30 +0100688 std::list<InstrumentationListener*> invoke_virtual_or_interface_listeners_
689 GUARDED_BY(Locks::mutator_lock_);
Nicolas Geoffray514a6162015-11-03 11:44:24 +0000690 std::list<InstrumentationListener*> dex_pc_listeners_ GUARDED_BY(Locks::mutator_lock_);
691 std::list<InstrumentationListener*> field_read_listeners_ GUARDED_BY(Locks::mutator_lock_);
692 std::list<InstrumentationListener*> field_write_listeners_ GUARDED_BY(Locks::mutator_lock_);
Alex Light6e1607e2017-08-23 10:06:18 -0700693 std::list<InstrumentationListener*> exception_thrown_listeners_ GUARDED_BY(Locks::mutator_lock_);
Alex Lighte814f9d2017-07-31 16:14:39 -0700694 std::list<InstrumentationListener*> watched_frame_pop_listeners_ GUARDED_BY(Locks::mutator_lock_);
Alex Light9fb1ab12017-09-05 09:32:49 -0700695 std::list<InstrumentationListener*> exception_handled_listeners_ GUARDED_BY(Locks::mutator_lock_);
jeffhao725a9572012-11-13 18:20:12 -0800696
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100697 // The set of methods being deoptimized (by the debugger) which must be executed with interpreter
698 // only.
Mathieu Chartier3b05e9b2014-03-25 09:29:43 -0700699 mutable ReaderWriterMutex deoptimized_methods_lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
Mathieu Chartiere401d142015-04-22 13:56:20 -0700700 std::unordered_set<ArtMethod*> deoptimized_methods_ GUARDED_BY(deoptimized_methods_lock_);
Sebastien Hertz11d40c22014-02-19 18:00:17 +0100701 bool deoptimization_enabled_;
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100702
Ian Rogersfa824272013-11-05 16:12:57 -0800703 // Current interpreter handler table. This is updated each time the thread state flags are
704 // modified.
Sebastien Hertzed2be172014-08-19 15:33:43 +0200705 InterpreterHandlerTable interpreter_handler_table_ GUARDED_BY(Locks::mutator_lock_);
Sebastien Hertzee1997a2013-09-19 14:47:09 +0200706
Ian Rogersfa824272013-11-05 16:12:57 -0800707 // Greater than 0 if quick alloc entry points instrumented.
Mathieu Chartiereebc3af2016-02-29 18:13:38 -0800708 size_t quick_alloc_entry_points_instrumentation_counter_;
Mathieu Chartier50e93312016-03-16 11:25:29 -0700709
710 // alloc_entrypoints_instrumented_ is only updated with all the threads suspended, this is done
711 // to prevent races with the GC where the GC relies on thread suspension only see
712 // alloc_entrypoints_instrumented_ change during suspend points.
713 bool alloc_entrypoints_instrumented_;
714
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200715 friend class InstrumentationTest; // For GetCurrentInstrumentationLevel and ConfigureStubs.
716
jeffhao725a9572012-11-13 18:20:12 -0800717 DISALLOW_COPY_AND_ASSIGN(Instrumentation);
718};
Ian Rogers6a3c1fc2014-10-31 00:33:20 -0700719std::ostream& operator<<(std::ostream& os, const Instrumentation::InstrumentationEvent& rhs);
Sebastien Hertz0462c4c2015-04-01 16:34:17 +0200720std::ostream& operator<<(std::ostream& os, const Instrumentation::InstrumentationLevel& rhs);
jeffhao725a9572012-11-13 18:20:12 -0800721
Ian Rogers62d6c772013-02-27 08:32:07 -0800722// An element in the instrumentation side stack maintained in art::Thread.
723struct InstrumentationStackFrame {
Mingyao Yang2ee17902017-08-30 11:37:08 -0700724 InstrumentationStackFrame(mirror::Object* this_object,
725 ArtMethod* method,
726 uintptr_t return_pc,
727 size_t frame_id,
728 bool interpreter_entry)
729 : this_object_(this_object),
730 method_(method),
731 return_pc_(return_pc),
732 frame_id_(frame_id),
Jeff Hao9a916d32013-06-27 18:45:37 -0700733 interpreter_entry_(interpreter_entry) {
Ian Rogers62d6c772013-02-27 08:32:07 -0800734 }
735
Andreas Gampebdf7f1c2016-08-30 16:38:47 -0700736 std::string Dump() const REQUIRES_SHARED(Locks::mutator_lock_);
Ian Rogers62d6c772013-02-27 08:32:07 -0800737
738 mirror::Object* this_object_;
Mathieu Chartiere401d142015-04-22 13:56:20 -0700739 ArtMethod* method_;
Sebastien Hertz138dbfc2013-12-04 18:15:25 +0100740 uintptr_t return_pc_;
741 size_t frame_id_;
742 bool interpreter_entry_;
Ian Rogers62d6c772013-02-27 08:32:07 -0800743};
744
745} // namespace instrumentation
jeffhao725a9572012-11-13 18:20:12 -0800746} // namespace art
747
Brian Carlstromfc0e3212013-07-17 14:40:12 -0700748#endif // ART_RUNTIME_INSTRUMENTATION_H_