Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 1 | // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 | // Use of this source code is governed by a BSD-style license that can be |
| 3 | // found in the LICENSE file. |
| 4 | |
| 5 | #ifndef V8_FULL_CODEGEN_FULL_CODEGEN_H_ |
| 6 | #define V8_FULL_CODEGEN_FULL_CODEGEN_H_ |
| 7 | |
| 8 | #include "src/allocation.h" |
| 9 | #include "src/assert-scope.h" |
| 10 | #include "src/ast/ast.h" |
| 11 | #include "src/ast/scopes.h" |
| 12 | #include "src/bit-vector.h" |
| 13 | #include "src/code-factory.h" |
| 14 | #include "src/code-stubs.h" |
| 15 | #include "src/codegen.h" |
| 16 | #include "src/compiler.h" |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 17 | #include "src/deoptimizer.h" |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 18 | #include "src/globals.h" |
| 19 | #include "src/objects.h" |
| 20 | |
| 21 | namespace v8 { |
| 22 | namespace internal { |
| 23 | |
| 24 | // Forward declarations. |
| 25 | class JumpPatchSite; |
| 26 | |
| 27 | // ----------------------------------------------------------------------------- |
| 28 | // Full code generator. |
| 29 | |
| 30 | class FullCodeGenerator: public AstVisitor { |
| 31 | public: |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 32 | FullCodeGenerator(MacroAssembler* masm, CompilationInfo* info) |
| 33 | : masm_(masm), |
| 34 | info_(info), |
| 35 | isolate_(info->isolate()), |
| 36 | zone_(info->zone()), |
| 37 | scope_(info->scope()), |
| 38 | nesting_stack_(NULL), |
| 39 | loop_depth_(0), |
| 40 | try_catch_depth_(0), |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 41 | operand_stack_depth_(0), |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 42 | globals_(NULL), |
| 43 | context_(NULL), |
| 44 | bailout_entries_(info->HasDeoptimizationSupport() |
| 45 | ? info->literal()->ast_node_count() |
| 46 | : 0, |
| 47 | info->zone()), |
| 48 | back_edges_(2, info->zone()), |
| 49 | handler_table_(info->zone()), |
| 50 | ic_total_count_(0) { |
| 51 | DCHECK(!info->IsStub()); |
| 52 | Initialize(); |
| 53 | } |
| 54 | |
| 55 | void Initialize(); |
| 56 | |
| 57 | static bool MakeCode(CompilationInfo* info); |
| 58 | |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 59 | // Encode bailout state and pc-offset as a BitField<type, start, size>. |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 60 | // Only use 30 bits because we encode the result as a smi. |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 61 | class BailoutStateField : public BitField<Deoptimizer::BailoutState, 0, 1> {}; |
| 62 | class PcField : public BitField<unsigned, 1, 30 - 1> {}; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 63 | |
| 64 | static const int kMaxBackEdgeWeight = 127; |
| 65 | |
| 66 | // Platform-specific code size multiplier. |
| 67 | #if V8_TARGET_ARCH_IA32 || V8_TARGET_ARCH_X87 |
| 68 | static const int kCodeSizeMultiplier = 105; |
| 69 | #elif V8_TARGET_ARCH_X64 |
| 70 | static const int kCodeSizeMultiplier = 165; |
| 71 | #elif V8_TARGET_ARCH_ARM |
| 72 | static const int kCodeSizeMultiplier = 149; |
| 73 | #elif V8_TARGET_ARCH_ARM64 |
| 74 | static const int kCodeSizeMultiplier = 220; |
| 75 | #elif V8_TARGET_ARCH_PPC64 |
| 76 | static const int kCodeSizeMultiplier = 200; |
| 77 | #elif V8_TARGET_ARCH_PPC |
| 78 | static const int kCodeSizeMultiplier = 200; |
| 79 | #elif V8_TARGET_ARCH_MIPS |
| 80 | static const int kCodeSizeMultiplier = 149; |
| 81 | #elif V8_TARGET_ARCH_MIPS64 |
| 82 | static const int kCodeSizeMultiplier = 149; |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 83 | #elif V8_TARGET_ARCH_S390 |
| 84 | // TODO(joransiu): Copied PPC value. Check this is sensible for S390. |
| 85 | static const int kCodeSizeMultiplier = 200; |
| 86 | #elif V8_TARGET_ARCH_S390X |
| 87 | // TODO(joransiu): Copied PPC value. Check this is sensible for S390X. |
| 88 | static const int kCodeSizeMultiplier = 200; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 89 | #else |
| 90 | #error Unsupported target architecture. |
| 91 | #endif |
| 92 | |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 93 | static Register result_register(); |
| 94 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 95 | private: |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 96 | typedef Deoptimizer::BailoutState BailoutState; |
| 97 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 98 | class Breakable; |
| 99 | class Iteration; |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 100 | class TryFinally; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 101 | |
| 102 | class TestContext; |
| 103 | |
| 104 | class NestedStatement BASE_EMBEDDED { |
| 105 | public: |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 106 | explicit NestedStatement(FullCodeGenerator* codegen) |
| 107 | : codegen_(codegen), |
| 108 | stack_depth_at_target_(codegen->operand_stack_depth_) { |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 109 | // Link into codegen's nesting stack. |
| 110 | previous_ = codegen->nesting_stack_; |
| 111 | codegen->nesting_stack_ = this; |
| 112 | } |
| 113 | virtual ~NestedStatement() { |
| 114 | // Unlink from codegen's nesting stack. |
| 115 | DCHECK_EQ(this, codegen_->nesting_stack_); |
| 116 | codegen_->nesting_stack_ = previous_; |
| 117 | } |
| 118 | |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 119 | virtual Breakable* AsBreakable() { return nullptr; } |
| 120 | virtual Iteration* AsIteration() { return nullptr; } |
| 121 | virtual TryFinally* AsTryFinally() { return nullptr; } |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 122 | |
| 123 | virtual bool IsContinueTarget(Statement* target) { return false; } |
| 124 | virtual bool IsBreakTarget(Statement* target) { return false; } |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 125 | virtual bool IsTryFinally() { return false; } |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 126 | |
| 127 | // Notify the statement that we are exiting it via break, continue, or |
| 128 | // return and give it a chance to generate cleanup code. Return the |
| 129 | // next outer statement in the nesting stack. We accumulate in |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 130 | // {*context_length} the number of context chain links to unwind as we |
| 131 | // traverse the nesting stack from an exit to its target. |
| 132 | virtual NestedStatement* Exit(int* context_length) { return previous_; } |
| 133 | |
| 134 | // Determine the expected operand stack depth when this statement is being |
| 135 | // used as the target of an exit. The caller will drop to this depth. |
| 136 | int GetStackDepthAtTarget() { return stack_depth_at_target_; } |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 137 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 138 | protected: |
| 139 | MacroAssembler* masm() { return codegen_->masm(); } |
| 140 | |
| 141 | FullCodeGenerator* codegen_; |
| 142 | NestedStatement* previous_; |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 143 | int stack_depth_at_target_; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 144 | |
| 145 | private: |
| 146 | DISALLOW_COPY_AND_ASSIGN(NestedStatement); |
| 147 | }; |
| 148 | |
| 149 | // A breakable statement such as a block. |
| 150 | class Breakable : public NestedStatement { |
| 151 | public: |
| 152 | Breakable(FullCodeGenerator* codegen, BreakableStatement* statement) |
| 153 | : NestedStatement(codegen), statement_(statement) { |
| 154 | } |
| 155 | |
| 156 | Breakable* AsBreakable() override { return this; } |
| 157 | bool IsBreakTarget(Statement* target) override { |
| 158 | return statement() == target; |
| 159 | } |
| 160 | |
| 161 | BreakableStatement* statement() { return statement_; } |
| 162 | Label* break_label() { return &break_label_; } |
| 163 | |
| 164 | private: |
| 165 | BreakableStatement* statement_; |
| 166 | Label break_label_; |
| 167 | }; |
| 168 | |
| 169 | // An iteration statement such as a while, for, or do loop. |
| 170 | class Iteration : public Breakable { |
| 171 | public: |
| 172 | Iteration(FullCodeGenerator* codegen, IterationStatement* statement) |
| 173 | : Breakable(codegen, statement) { |
| 174 | } |
| 175 | |
| 176 | Iteration* AsIteration() override { return this; } |
| 177 | bool IsContinueTarget(Statement* target) override { |
| 178 | return statement() == target; |
| 179 | } |
| 180 | |
| 181 | Label* continue_label() { return &continue_label_; } |
| 182 | |
| 183 | private: |
| 184 | Label continue_label_; |
| 185 | }; |
| 186 | |
| 187 | // A nested block statement. |
| 188 | class NestedBlock : public Breakable { |
| 189 | public: |
| 190 | NestedBlock(FullCodeGenerator* codegen, Block* block) |
| 191 | : Breakable(codegen, block) { |
| 192 | } |
| 193 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 194 | NestedStatement* Exit(int* context_length) override { |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 195 | auto block_scope = statement()->AsBlock()->scope(); |
| 196 | if (block_scope != nullptr) { |
| 197 | if (block_scope->ContextLocalCount() > 0) ++(*context_length); |
| 198 | } |
| 199 | return previous_; |
| 200 | } |
| 201 | }; |
| 202 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 203 | // A class literal expression |
| 204 | class NestedClassLiteral : public NestedStatement { |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 205 | public: |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 206 | NestedClassLiteral(FullCodeGenerator* codegen, ClassLiteral* lit) |
| 207 | : NestedStatement(codegen), |
| 208 | needs_context_(lit->scope() != nullptr && |
| 209 | lit->scope()->NeedsContext()) {} |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 210 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 211 | NestedStatement* Exit(int* context_length) override { |
| 212 | if (needs_context_) ++(*context_length); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 213 | return previous_; |
| 214 | } |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 215 | |
| 216 | private: |
| 217 | const bool needs_context_; |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 218 | }; |
| 219 | |
| 220 | class DeferredCommands { |
| 221 | public: |
| 222 | enum Command { kReturn, kThrow, kBreak, kContinue }; |
| 223 | typedef int TokenId; |
| 224 | struct DeferredCommand { |
| 225 | Command command; |
| 226 | TokenId token; |
| 227 | Statement* target; |
| 228 | }; |
| 229 | |
| 230 | DeferredCommands(FullCodeGenerator* codegen, Label* finally_entry) |
| 231 | : codegen_(codegen), |
| 232 | commands_(codegen->zone()), |
| 233 | return_token_(TokenDispenserForFinally::kInvalidToken), |
| 234 | throw_token_(TokenDispenserForFinally::kInvalidToken), |
| 235 | finally_entry_(finally_entry) {} |
| 236 | |
| 237 | void EmitCommands(); |
| 238 | |
| 239 | void RecordBreak(Statement* target); |
| 240 | void RecordContinue(Statement* target); |
| 241 | void RecordReturn(); |
| 242 | void RecordThrow(); |
| 243 | void EmitFallThrough(); |
| 244 | |
| 245 | private: |
| 246 | MacroAssembler* masm() { return codegen_->masm(); } |
| 247 | void EmitJumpToFinally(TokenId token); |
| 248 | |
| 249 | FullCodeGenerator* codegen_; |
| 250 | ZoneVector<DeferredCommand> commands_; |
| 251 | TokenDispenserForFinally dispenser_; |
| 252 | TokenId return_token_; |
| 253 | TokenId throw_token_; |
| 254 | Label* finally_entry_; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 255 | }; |
| 256 | |
| 257 | // The try block of a try/finally statement. |
| 258 | class TryFinally : public NestedStatement { |
| 259 | public: |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 260 | TryFinally(FullCodeGenerator* codegen, DeferredCommands* commands) |
| 261 | : NestedStatement(codegen), deferred_commands_(commands) {} |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 262 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 263 | NestedStatement* Exit(int* context_length) override; |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 264 | |
| 265 | bool IsTryFinally() override { return true; } |
| 266 | TryFinally* AsTryFinally() override { return this; } |
| 267 | |
| 268 | DeferredCommands* deferred_commands() { return deferred_commands_; } |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 269 | |
| 270 | private: |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 271 | DeferredCommands* deferred_commands_; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 272 | }; |
| 273 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 274 | // The body of a with or catch. |
| 275 | class WithOrCatch : public NestedStatement { |
| 276 | public: |
| 277 | explicit WithOrCatch(FullCodeGenerator* codegen) |
| 278 | : NestedStatement(codegen) { |
| 279 | } |
| 280 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 281 | NestedStatement* Exit(int* context_length) override { |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 282 | ++(*context_length); |
| 283 | return previous_; |
| 284 | } |
| 285 | }; |
| 286 | |
| 287 | // A platform-specific utility to overwrite the accumulator register |
| 288 | // with a GC-safe value. |
| 289 | void ClearAccumulator(); |
| 290 | |
| 291 | // Determine whether or not to inline the smi case for the given |
| 292 | // operation. |
| 293 | bool ShouldInlineSmiCase(Token::Value op); |
| 294 | |
| 295 | // Helper function to convert a pure value into a test context. The value |
| 296 | // is expected on the stack or the accumulator, depending on the platform. |
| 297 | // See the platform-specific implementation for details. |
| 298 | void DoTest(Expression* condition, |
| 299 | Label* if_true, |
| 300 | Label* if_false, |
| 301 | Label* fall_through); |
| 302 | void DoTest(const TestContext* context); |
| 303 | |
| 304 | // Helper function to split control flow and avoid a branch to the |
| 305 | // fall-through label if it is set up. |
| 306 | #if V8_TARGET_ARCH_MIPS |
| 307 | void Split(Condition cc, |
| 308 | Register lhs, |
| 309 | const Operand& rhs, |
| 310 | Label* if_true, |
| 311 | Label* if_false, |
| 312 | Label* fall_through); |
| 313 | #elif V8_TARGET_ARCH_MIPS64 |
| 314 | void Split(Condition cc, |
| 315 | Register lhs, |
| 316 | const Operand& rhs, |
| 317 | Label* if_true, |
| 318 | Label* if_false, |
| 319 | Label* fall_through); |
| 320 | #elif V8_TARGET_ARCH_PPC |
| 321 | void Split(Condition cc, Label* if_true, Label* if_false, Label* fall_through, |
| 322 | CRegister cr = cr7); |
| 323 | #else // All other arch. |
| 324 | void Split(Condition cc, |
| 325 | Label* if_true, |
| 326 | Label* if_false, |
| 327 | Label* fall_through); |
| 328 | #endif |
| 329 | |
| 330 | // Load the value of a known (PARAMETER, LOCAL, or CONTEXT) variable into |
| 331 | // a register. Emits a context chain walk if if necessary (so does |
| 332 | // SetVar) so avoid calling both on the same variable. |
| 333 | void GetVar(Register destination, Variable* var); |
| 334 | |
| 335 | // Assign to a known (PARAMETER, LOCAL, or CONTEXT) variable. If it's in |
| 336 | // the context, the write barrier will be emitted and source, scratch0, |
| 337 | // scratch1 will be clobbered. Emits a context chain walk if if necessary |
| 338 | // (so does GetVar) so avoid calling both on the same variable. |
| 339 | void SetVar(Variable* var, |
| 340 | Register source, |
| 341 | Register scratch0, |
| 342 | Register scratch1); |
| 343 | |
| 344 | // An operand used to read/write a stack-allocated (PARAMETER or LOCAL) |
| 345 | // variable. Writing does not need the write barrier. |
| 346 | MemOperand StackOperand(Variable* var); |
| 347 | |
| 348 | // An operand used to read/write a known (PARAMETER, LOCAL, or CONTEXT) |
| 349 | // variable. May emit code to traverse the context chain, loading the |
| 350 | // found context into the scratch register. Writing to this operand will |
| 351 | // need the write barrier if location is CONTEXT. |
| 352 | MemOperand VarOperand(Variable* var, Register scratch); |
| 353 | |
| 354 | void VisitForEffect(Expression* expr) { |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 355 | if (FLAG_verify_operand_stack_depth) EmitOperandStackDepthCheck(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 356 | EffectContext context(this); |
| 357 | Visit(expr); |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 358 | PrepareForBailout(expr, BailoutState::NO_REGISTERS); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 359 | } |
| 360 | |
| 361 | void VisitForAccumulatorValue(Expression* expr) { |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 362 | if (FLAG_verify_operand_stack_depth) EmitOperandStackDepthCheck(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 363 | AccumulatorValueContext context(this); |
| 364 | Visit(expr); |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 365 | PrepareForBailout(expr, BailoutState::TOS_REGISTER); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 366 | } |
| 367 | |
| 368 | void VisitForStackValue(Expression* expr) { |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 369 | if (FLAG_verify_operand_stack_depth) EmitOperandStackDepthCheck(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 370 | StackValueContext context(this); |
| 371 | Visit(expr); |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 372 | PrepareForBailout(expr, BailoutState::NO_REGISTERS); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 373 | } |
| 374 | |
| 375 | void VisitForControl(Expression* expr, |
| 376 | Label* if_true, |
| 377 | Label* if_false, |
| 378 | Label* fall_through) { |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 379 | if (FLAG_verify_operand_stack_depth) EmitOperandStackDepthCheck(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 380 | TestContext context(this, expr, if_true, if_false, fall_through); |
| 381 | Visit(expr); |
| 382 | // For test contexts, we prepare for bailout before branching, not at |
| 383 | // the end of the entire expression. This happens as part of visiting |
| 384 | // the expression. |
| 385 | } |
| 386 | |
| 387 | void VisitInDuplicateContext(Expression* expr); |
| 388 | |
| 389 | void VisitDeclarations(ZoneList<Declaration*>* declarations) override; |
| 390 | void DeclareModules(Handle<FixedArray> descriptions); |
| 391 | void DeclareGlobals(Handle<FixedArray> pairs); |
| 392 | int DeclareGlobalsFlags(); |
| 393 | |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 394 | // Push, pop or drop values onto/from the operand stack. |
| 395 | void PushOperand(Register reg); |
| 396 | void PopOperand(Register reg); |
| 397 | void DropOperands(int count); |
| 398 | |
| 399 | // Convenience helpers for pushing onto the operand stack. |
| 400 | void PushOperand(MemOperand operand); |
| 401 | void PushOperand(Handle<Object> handle); |
| 402 | void PushOperand(Smi* smi); |
| 403 | |
| 404 | // Convenience helpers for pushing/popping multiple operands. |
| 405 | void PushOperands(Register reg1, Register reg2); |
| 406 | void PushOperands(Register reg1, Register reg2, Register reg3); |
| 407 | void PushOperands(Register reg1, Register reg2, Register reg3, Register reg4); |
| 408 | void PopOperands(Register reg1, Register reg2); |
| 409 | |
| 410 | // Convenience helper for calling a runtime function that consumes arguments |
| 411 | // from the operand stack (only usable for functions with known arity). |
| 412 | void CallRuntimeWithOperands(Runtime::FunctionId function_id); |
| 413 | |
| 414 | // Static tracking of the operand stack depth. |
| 415 | void OperandStackDepthDecrement(int count); |
| 416 | void OperandStackDepthIncrement(int count); |
| 417 | |
| 418 | // Generate debug code that verifies that our static tracking of the operand |
| 419 | // stack depth is in sync with the actual operand stack during runtime. |
| 420 | void EmitOperandStackDepthCheck(); |
| 421 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 422 | // Generate code to create an iterator result object. The "value" property is |
| 423 | // set to a value popped from the stack, and "done" is set according to the |
| 424 | // argument. The result object is left in the result register. |
| 425 | void EmitCreateIteratorResult(bool done); |
| 426 | |
| 427 | // Try to perform a comparison as a fast inlined literal compare if |
| 428 | // the operands allow it. Returns true if the compare operations |
| 429 | // has been matched and all code generated; false otherwise. |
| 430 | bool TryLiteralCompare(CompareOperation* compare); |
| 431 | |
| 432 | // Platform-specific code for comparing the type of a value with |
| 433 | // a given literal string. |
| 434 | void EmitLiteralCompareTypeof(Expression* expr, |
| 435 | Expression* sub_expr, |
| 436 | Handle<String> check); |
| 437 | |
| 438 | // Platform-specific code for equality comparison with a nil-like value. |
| 439 | void EmitLiteralCompareNil(CompareOperation* expr, |
| 440 | Expression* sub_expr, |
| 441 | NilValue nil); |
| 442 | |
| 443 | // Bailout support. |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 444 | void PrepareForBailout(Expression* node, Deoptimizer::BailoutState state); |
| 445 | void PrepareForBailoutForId(BailoutId id, Deoptimizer::BailoutState state); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 446 | |
| 447 | // Returns a smi for the index into the FixedArray that backs the feedback |
| 448 | // vector |
| 449 | Smi* SmiFromSlot(FeedbackVectorSlot slot) const { |
| 450 | return Smi::FromInt(TypeFeedbackVector::GetIndexFromSpec( |
| 451 | literal()->feedback_vector_spec(), slot)); |
| 452 | } |
| 453 | |
| 454 | // Record a call's return site offset, used to rebuild the frame if the |
| 455 | // called function was inlined at the site. |
| 456 | void RecordJSReturnSite(Call* call); |
| 457 | |
| 458 | // Prepare for bailout before a test (or compare) and branch. If |
| 459 | // should_normalize, then the following comparison will not handle the |
| 460 | // canonical JS true value so we will insert a (dead) test against true at |
| 461 | // the actual bailout target from the optimized code. If not |
| 462 | // should_normalize, the true and false labels are ignored. |
| 463 | void PrepareForBailoutBeforeSplit(Expression* expr, |
| 464 | bool should_normalize, |
| 465 | Label* if_true, |
| 466 | Label* if_false); |
| 467 | |
| 468 | // If enabled, emit debug code for checking that the current context is |
| 469 | // neither a with nor a catch context. |
| 470 | void EmitDebugCheckDeclarationContext(Variable* variable); |
| 471 | |
| 472 | // This is meant to be called at loop back edges, |back_edge_target| is |
| 473 | // the jump target of the back edge and is used to approximate the amount |
| 474 | // of code inside the loop. |
| 475 | void EmitBackEdgeBookkeeping(IterationStatement* stmt, |
| 476 | Label* back_edge_target); |
| 477 | // Record the OSR AST id corresponding to a back edge in the code. |
| 478 | void RecordBackEdge(BailoutId osr_ast_id); |
| 479 | // Emit a table of back edge ids, pcs and loop depths into the code stream. |
| 480 | // Return the offset of the start of the table. |
| 481 | unsigned EmitBackEdgeTable(); |
| 482 | |
| 483 | void EmitProfilingCounterDecrement(int delta); |
| 484 | void EmitProfilingCounterReset(); |
| 485 | |
| 486 | // Emit code to pop values from the stack associated with nested statements |
| 487 | // like try/catch, try/finally, etc, running the finallies and unwinding the |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 488 | // handlers as needed. Also emits the return sequence if necessary (i.e., |
| 489 | // if the return is not delayed by a finally block). |
| 490 | void EmitUnwindAndReturn(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 491 | |
| 492 | // Platform-specific return sequence |
| 493 | void EmitReturnSequence(); |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 494 | void EmitProfilingCounterHandlingForReturnSequence(bool is_tail_call); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 495 | |
| 496 | // Platform-specific code sequences for calls |
| 497 | void EmitCall(Call* expr, ConvertReceiverMode = ConvertReceiverMode::kAny); |
| 498 | void EmitSuperConstructorCall(Call* expr); |
| 499 | void EmitCallWithLoadIC(Call* expr); |
| 500 | void EmitSuperCallWithLoadIC(Call* expr); |
| 501 | void EmitKeyedCallWithLoadIC(Call* expr, Expression* key); |
| 502 | void EmitKeyedSuperCallWithLoadIC(Call* expr); |
| 503 | void EmitPossiblyEvalCall(Call* expr); |
| 504 | |
| 505 | #define FOR_EACH_FULL_CODE_INTRINSIC(F) \ |
| 506 | F(IsSmi) \ |
| 507 | F(IsArray) \ |
| 508 | F(IsTypedArray) \ |
| 509 | F(IsRegExp) \ |
| 510 | F(IsJSProxy) \ |
| 511 | F(Call) \ |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 512 | F(NewObject) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 513 | F(ValueOf) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 514 | F(StringCharFromCode) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 515 | F(IsJSReceiver) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 516 | F(MathPow) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 517 | F(HasCachedArrayIndex) \ |
| 518 | F(GetCachedArrayIndex) \ |
| 519 | F(GetSuperConstructor) \ |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 520 | F(DebugBreakInOptimizedCode) \ |
| 521 | F(ClassOf) \ |
| 522 | F(StringCharCodeAt) \ |
| 523 | F(SubString) \ |
| 524 | F(RegExpExec) \ |
| 525 | F(RegExpConstructResult) \ |
| 526 | F(ToInteger) \ |
| 527 | F(NumberToString) \ |
| 528 | F(ToString) \ |
| 529 | F(ToLength) \ |
| 530 | F(ToNumber) \ |
| 531 | F(ToName) \ |
| 532 | F(ToObject) \ |
| 533 | F(DebugIsActive) \ |
| 534 | F(CreateIterResultObject) |
| 535 | |
| 536 | #define GENERATOR_DECLARATION(Name) void Emit##Name(CallRuntime* call); |
| 537 | FOR_EACH_FULL_CODE_INTRINSIC(GENERATOR_DECLARATION) |
| 538 | #undef GENERATOR_DECLARATION |
| 539 | |
| 540 | void EmitIntrinsicAsStubCall(CallRuntime* expr, const Callable& callable); |
| 541 | |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 542 | // Emits call to respective code stub. |
| 543 | void EmitHasProperty(); |
| 544 | |
| 545 | // Platform-specific code for restoring context from current JS frame. |
| 546 | void RestoreContext(); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 547 | |
| 548 | // Platform-specific code for loading variables. |
| 549 | void EmitLoadGlobalCheckExtensions(VariableProxy* proxy, |
| 550 | TypeofMode typeof_mode, Label* slow); |
| 551 | MemOperand ContextSlotOperandCheckExtensions(Variable* var, Label* slow); |
| 552 | void EmitDynamicLookupFastCase(VariableProxy* proxy, TypeofMode typeof_mode, |
| 553 | Label* slow, Label* done); |
| 554 | void EmitGlobalVariableLoad(VariableProxy* proxy, TypeofMode typeof_mode); |
| 555 | void EmitVariableLoad(VariableProxy* proxy, |
| 556 | TypeofMode typeof_mode = NOT_INSIDE_TYPEOF); |
| 557 | |
| 558 | void EmitAccessor(ObjectLiteralProperty* property); |
| 559 | |
| 560 | bool NeedsHoleCheckForLoad(VariableProxy* proxy); |
| 561 | |
| 562 | // Expects the arguments and the function already pushed. |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 563 | void EmitResolvePossiblyDirectEval(Call* expr); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 564 | |
| 565 | // Platform-specific support for allocating a new closure based on |
| 566 | // the given function info. |
| 567 | void EmitNewClosure(Handle<SharedFunctionInfo> info, bool pretenure); |
| 568 | |
| 569 | // Re-usable portions of CallRuntime |
| 570 | void EmitLoadJSRuntimeFunction(CallRuntime* expr); |
| 571 | void EmitCallJSRuntimeFunction(CallRuntime* expr); |
| 572 | |
| 573 | // Load a value from a named property. |
| 574 | // The receiver is left on the stack by the IC. |
| 575 | void EmitNamedPropertyLoad(Property* expr); |
| 576 | |
| 577 | // Load a value from super.named property. |
| 578 | // Expect receiver ('this' value) and home_object on the stack. |
| 579 | void EmitNamedSuperPropertyLoad(Property* expr); |
| 580 | |
| 581 | // Load a value from super[keyed] property. |
| 582 | // Expect receiver ('this' value), home_object and key on the stack. |
| 583 | void EmitKeyedSuperPropertyLoad(Property* expr); |
| 584 | |
| 585 | // Load a value from a keyed property. |
| 586 | // The receiver and the key is left on the stack by the IC. |
| 587 | void EmitKeyedPropertyLoad(Property* expr); |
| 588 | |
| 589 | // Adds the properties to the class (function) object and to its prototype. |
| 590 | // Expects the class (function) in the accumulator. The class (function) is |
| 591 | // in the accumulator after installing all the properties. |
| 592 | void EmitClassDefineProperties(ClassLiteral* lit); |
| 593 | |
| 594 | // Pushes the property key as a Name on the stack. |
| 595 | void EmitPropertyKey(ObjectLiteralProperty* property, BailoutId bailout_id); |
| 596 | |
| 597 | // Apply the compound assignment operator. Expects the left operand on top |
| 598 | // of the stack and the right one in the accumulator. |
| 599 | void EmitBinaryOp(BinaryOperation* expr, Token::Value op); |
| 600 | |
| 601 | // Helper functions for generating inlined smi code for certain |
| 602 | // binary operations. |
| 603 | void EmitInlineSmiBinaryOp(BinaryOperation* expr, |
| 604 | Token::Value op, |
| 605 | Expression* left, |
| 606 | Expression* right); |
| 607 | |
| 608 | // Assign to the given expression as if via '='. The right-hand-side value |
| 609 | // is expected in the accumulator. slot is only used if FLAG_vector_stores |
| 610 | // is true. |
| 611 | void EmitAssignment(Expression* expr, FeedbackVectorSlot slot); |
| 612 | |
| 613 | // Complete a variable assignment. The right-hand-side value is expected |
| 614 | // in the accumulator. |
| 615 | void EmitVariableAssignment(Variable* var, Token::Value op, |
| 616 | FeedbackVectorSlot slot); |
| 617 | |
| 618 | // Helper functions to EmitVariableAssignment |
| 619 | void EmitStoreToStackLocalOrContextSlot(Variable* var, |
| 620 | MemOperand location); |
| 621 | |
| 622 | // Complete a named property assignment. The receiver is expected on top |
| 623 | // of the stack and the right-hand-side value in the accumulator. |
| 624 | void EmitNamedPropertyAssignment(Assignment* expr); |
| 625 | |
| 626 | // Complete a super named property assignment. The right-hand-side value |
| 627 | // is expected in accumulator. |
| 628 | void EmitNamedSuperPropertyStore(Property* prop); |
| 629 | |
| 630 | // Complete a super named property assignment. The right-hand-side value |
| 631 | // is expected in accumulator. |
| 632 | void EmitKeyedSuperPropertyStore(Property* prop); |
| 633 | |
| 634 | // Complete a keyed property assignment. The receiver and key are |
| 635 | // expected on top of the stack and the right-hand-side value in the |
| 636 | // accumulator. |
| 637 | void EmitKeyedPropertyAssignment(Assignment* expr); |
| 638 | |
| 639 | static bool NeedsHomeObject(Expression* expr) { |
| 640 | return FunctionLiteral::NeedsHomeObject(expr); |
| 641 | } |
| 642 | |
| 643 | // Adds the [[HomeObject]] to |initializer| if it is a FunctionLiteral. |
| 644 | // The value of the initializer is expected to be at the top of the stack. |
| 645 | // |offset| is the offset in the stack where the home object can be found. |
| 646 | void EmitSetHomeObject(Expression* initializer, int offset, |
| 647 | FeedbackVectorSlot slot); |
| 648 | |
| 649 | void EmitSetHomeObjectAccumulator(Expression* initializer, int offset, |
| 650 | FeedbackVectorSlot slot); |
| 651 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 652 | void EmitLoadStoreICSlot(FeedbackVectorSlot slot); |
| 653 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 654 | void CallIC(Handle<Code> code, |
| 655 | TypeFeedbackId id = TypeFeedbackId::None()); |
| 656 | |
Ben Murdoch | 61f157c | 2016-09-16 13:49:30 +0100 | [diff] [blame] | 657 | void CallLoadIC(TypeFeedbackId id = TypeFeedbackId::None()); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 658 | // Inside typeof reference errors are never thrown. |
Ben Murdoch | 61f157c | 2016-09-16 13:49:30 +0100 | [diff] [blame] | 659 | void CallLoadGlobalIC(TypeofMode typeof_mode, |
| 660 | TypeFeedbackId id = TypeFeedbackId::None()); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 661 | void CallStoreIC(TypeFeedbackId id = TypeFeedbackId::None()); |
| 662 | |
| 663 | void SetFunctionPosition(FunctionLiteral* fun); |
| 664 | void SetReturnPosition(FunctionLiteral* fun); |
| 665 | |
| 666 | enum InsertBreak { INSERT_BREAK, SKIP_BREAK }; |
| 667 | |
| 668 | // During stepping we want to be able to break at each statement, but not at |
| 669 | // every (sub-)expression. That is why by default we insert breaks at every |
| 670 | // statement position, but not at every expression position, unless stated |
| 671 | // otherwise. |
| 672 | void SetStatementPosition(Statement* stmt, |
| 673 | InsertBreak insert_break = INSERT_BREAK); |
Ben Murdoch | c561043 | 2016-08-08 18:44:38 +0100 | [diff] [blame] | 674 | void SetExpressionPosition(Expression* expr); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 675 | |
| 676 | // Consider an expression a statement. As such, we also insert a break. |
| 677 | // This is used in loop headers where we want to break for each iteration. |
| 678 | void SetExpressionAsStatementPosition(Expression* expr); |
| 679 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 680 | void SetCallPosition(Expression* expr, |
| 681 | TailCallMode tail_call_mode = TailCallMode::kDisallow); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 682 | |
| 683 | void SetConstructCallPosition(Expression* expr) { |
| 684 | // Currently call and construct calls are treated the same wrt debugging. |
| 685 | SetCallPosition(expr); |
| 686 | } |
| 687 | |
| 688 | // Non-local control flow support. |
| 689 | void EnterTryBlock(int handler_index, Label* handler); |
| 690 | void ExitTryBlock(int handler_index); |
| 691 | void EnterFinallyBlock(); |
| 692 | void ExitFinallyBlock(); |
| 693 | void ClearPendingMessage(); |
| 694 | |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 695 | void EmitContinue(Statement* target); |
| 696 | void EmitBreak(Statement* target); |
| 697 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 698 | // Loop nesting counter. |
| 699 | int loop_depth() { return loop_depth_; } |
| 700 | void increment_loop_depth() { loop_depth_++; } |
| 701 | void decrement_loop_depth() { |
| 702 | DCHECK(loop_depth_ > 0); |
| 703 | loop_depth_--; |
| 704 | } |
| 705 | |
| 706 | MacroAssembler* masm() const { return masm_; } |
| 707 | |
| 708 | class ExpressionContext; |
| 709 | const ExpressionContext* context() { return context_; } |
| 710 | void set_new_context(const ExpressionContext* context) { context_ = context; } |
| 711 | |
| 712 | Isolate* isolate() const { return isolate_; } |
| 713 | Zone* zone() const { return zone_; } |
| 714 | Handle<Script> script() { return info_->script(); } |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 715 | LanguageMode language_mode() { return scope()->language_mode(); } |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 716 | bool has_simple_parameters() { return info_->has_simple_parameters(); } |
| 717 | FunctionLiteral* literal() const { return info_->literal(); } |
| 718 | Scope* scope() { return scope_; } |
| 719 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 720 | static Register context_register(); |
| 721 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 722 | // Get fields from the stack frame. Offsets are the frame pointer relative |
| 723 | // offsets defined in, e.g., StandardFrameConstants. |
| 724 | void LoadFromFrameField(int frame_offset, Register value); |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 725 | // Set fields in the stack frame. Offsets are the frame pointer relative |
| 726 | // offsets defined in, e.g., StandardFrameConstants. |
| 727 | void StoreToFrameField(int frame_offset, Register value); |
| 728 | |
| 729 | // Load a value from the current context. Indices are defined as an enum |
| 730 | // in v8::internal::Context. |
| 731 | void LoadContextField(Register dst, int context_index); |
| 732 | |
| 733 | // Push the function argument for the runtime functions PushWithContext |
| 734 | // and PushCatchContext. |
| 735 | void PushFunctionArgumentForContextAllocation(); |
| 736 | |
| 737 | void PushCalleeAndWithBaseObject(Call* expr); |
| 738 | |
| 739 | // AST node visit functions. |
| 740 | #define DECLARE_VISIT(type) void Visit##type(type* node) override; |
| 741 | AST_NODE_LIST(DECLARE_VISIT) |
| 742 | #undef DECLARE_VISIT |
| 743 | |
| 744 | void VisitComma(BinaryOperation* expr); |
| 745 | void VisitLogicalExpression(BinaryOperation* expr); |
| 746 | void VisitArithmeticExpression(BinaryOperation* expr); |
| 747 | |
| 748 | void VisitForTypeofValue(Expression* expr); |
| 749 | |
| 750 | void Generate(); |
| 751 | void PopulateDeoptimizationData(Handle<Code> code); |
| 752 | void PopulateTypeFeedbackInfo(Handle<Code> code); |
| 753 | void PopulateHandlerTable(Handle<Code> code); |
| 754 | |
| 755 | bool MustCreateObjectLiteralWithRuntime(ObjectLiteral* expr) const; |
| 756 | bool MustCreateArrayLiteralWithRuntime(ArrayLiteral* expr) const; |
| 757 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 758 | int NewHandlerTableEntry(); |
| 759 | |
| 760 | struct BailoutEntry { |
| 761 | BailoutId id; |
| 762 | unsigned pc_and_state; |
| 763 | }; |
| 764 | |
| 765 | struct BackEdgeEntry { |
| 766 | BailoutId id; |
| 767 | unsigned pc; |
| 768 | uint32_t loop_depth; |
| 769 | }; |
| 770 | |
| 771 | struct HandlerTableEntry { |
| 772 | unsigned range_start; |
| 773 | unsigned range_end; |
| 774 | unsigned handler_offset; |
| 775 | int stack_depth; |
| 776 | int try_catch_depth; |
| 777 | }; |
| 778 | |
| 779 | class ExpressionContext BASE_EMBEDDED { |
| 780 | public: |
| 781 | explicit ExpressionContext(FullCodeGenerator* codegen) |
| 782 | : masm_(codegen->masm()), old_(codegen->context()), codegen_(codegen) { |
| 783 | codegen->set_new_context(this); |
| 784 | } |
| 785 | |
| 786 | virtual ~ExpressionContext() { |
| 787 | codegen_->set_new_context(old_); |
| 788 | } |
| 789 | |
| 790 | Isolate* isolate() const { return codegen_->isolate(); } |
| 791 | |
| 792 | // Convert constant control flow (true or false) to the result expected for |
| 793 | // this expression context. |
| 794 | virtual void Plug(bool flag) const = 0; |
| 795 | |
| 796 | // Emit code to convert a pure value (in a register, known variable |
| 797 | // location, as a literal, or on top of the stack) into the result |
| 798 | // expected according to this expression context. |
| 799 | virtual void Plug(Register reg) const = 0; |
| 800 | virtual void Plug(Variable* var) const = 0; |
| 801 | virtual void Plug(Handle<Object> lit) const = 0; |
| 802 | virtual void Plug(Heap::RootListIndex index) const = 0; |
| 803 | virtual void PlugTOS() const = 0; |
| 804 | |
| 805 | // Emit code to convert pure control flow to a pair of unbound labels into |
| 806 | // the result expected according to this expression context. The |
| 807 | // implementation will bind both labels unless it's a TestContext, which |
| 808 | // won't bind them at this point. |
| 809 | virtual void Plug(Label* materialize_true, |
| 810 | Label* materialize_false) const = 0; |
| 811 | |
| 812 | // Emit code to discard count elements from the top of stack, then convert |
| 813 | // a pure value into the result expected according to this expression |
| 814 | // context. |
| 815 | virtual void DropAndPlug(int count, Register reg) const = 0; |
| 816 | |
| 817 | // Set up branch labels for a test expression. The three Label** parameters |
| 818 | // are output parameters. |
| 819 | virtual void PrepareTest(Label* materialize_true, |
| 820 | Label* materialize_false, |
| 821 | Label** if_true, |
| 822 | Label** if_false, |
| 823 | Label** fall_through) const = 0; |
| 824 | |
| 825 | // Returns true if we are evaluating only for side effects (i.e. if the |
| 826 | // result will be discarded). |
| 827 | virtual bool IsEffect() const { return false; } |
| 828 | |
| 829 | // Returns true if we are evaluating for the value (in accu/on stack). |
| 830 | virtual bool IsAccumulatorValue() const { return false; } |
| 831 | virtual bool IsStackValue() const { return false; } |
| 832 | |
| 833 | // Returns true if we are branching on the value rather than materializing |
| 834 | // it. Only used for asserts. |
| 835 | virtual bool IsTest() const { return false; } |
| 836 | |
| 837 | protected: |
| 838 | FullCodeGenerator* codegen() const { return codegen_; } |
| 839 | MacroAssembler* masm() const { return masm_; } |
| 840 | MacroAssembler* masm_; |
| 841 | |
| 842 | private: |
| 843 | const ExpressionContext* old_; |
| 844 | FullCodeGenerator* codegen_; |
| 845 | }; |
| 846 | |
| 847 | class AccumulatorValueContext : public ExpressionContext { |
| 848 | public: |
| 849 | explicit AccumulatorValueContext(FullCodeGenerator* codegen) |
| 850 | : ExpressionContext(codegen) { } |
| 851 | |
| 852 | void Plug(bool flag) const override; |
| 853 | void Plug(Register reg) const override; |
| 854 | void Plug(Label* materialize_true, Label* materialize_false) const override; |
| 855 | void Plug(Variable* var) const override; |
| 856 | void Plug(Handle<Object> lit) const override; |
| 857 | void Plug(Heap::RootListIndex) const override; |
| 858 | void PlugTOS() const override; |
| 859 | void DropAndPlug(int count, Register reg) const override; |
| 860 | void PrepareTest(Label* materialize_true, Label* materialize_false, |
| 861 | Label** if_true, Label** if_false, |
| 862 | Label** fall_through) const override; |
| 863 | bool IsAccumulatorValue() const override { return true; } |
| 864 | }; |
| 865 | |
| 866 | class StackValueContext : public ExpressionContext { |
| 867 | public: |
| 868 | explicit StackValueContext(FullCodeGenerator* codegen) |
| 869 | : ExpressionContext(codegen) { } |
| 870 | |
| 871 | void Plug(bool flag) const override; |
| 872 | void Plug(Register reg) const override; |
| 873 | void Plug(Label* materialize_true, Label* materialize_false) const override; |
| 874 | void Plug(Variable* var) const override; |
| 875 | void Plug(Handle<Object> lit) const override; |
| 876 | void Plug(Heap::RootListIndex) const override; |
| 877 | void PlugTOS() const override; |
| 878 | void DropAndPlug(int count, Register reg) const override; |
| 879 | void PrepareTest(Label* materialize_true, Label* materialize_false, |
| 880 | Label** if_true, Label** if_false, |
| 881 | Label** fall_through) const override; |
| 882 | bool IsStackValue() const override { return true; } |
| 883 | }; |
| 884 | |
| 885 | class TestContext : public ExpressionContext { |
| 886 | public: |
| 887 | TestContext(FullCodeGenerator* codegen, |
| 888 | Expression* condition, |
| 889 | Label* true_label, |
| 890 | Label* false_label, |
| 891 | Label* fall_through) |
| 892 | : ExpressionContext(codegen), |
| 893 | condition_(condition), |
| 894 | true_label_(true_label), |
| 895 | false_label_(false_label), |
| 896 | fall_through_(fall_through) { } |
| 897 | |
| 898 | static const TestContext* cast(const ExpressionContext* context) { |
| 899 | DCHECK(context->IsTest()); |
| 900 | return reinterpret_cast<const TestContext*>(context); |
| 901 | } |
| 902 | |
| 903 | Expression* condition() const { return condition_; } |
| 904 | Label* true_label() const { return true_label_; } |
| 905 | Label* false_label() const { return false_label_; } |
| 906 | Label* fall_through() const { return fall_through_; } |
| 907 | |
| 908 | void Plug(bool flag) const override; |
| 909 | void Plug(Register reg) const override; |
| 910 | void Plug(Label* materialize_true, Label* materialize_false) const override; |
| 911 | void Plug(Variable* var) const override; |
| 912 | void Plug(Handle<Object> lit) const override; |
| 913 | void Plug(Heap::RootListIndex) const override; |
| 914 | void PlugTOS() const override; |
| 915 | void DropAndPlug(int count, Register reg) const override; |
| 916 | void PrepareTest(Label* materialize_true, Label* materialize_false, |
| 917 | Label** if_true, Label** if_false, |
| 918 | Label** fall_through) const override; |
| 919 | bool IsTest() const override { return true; } |
| 920 | |
| 921 | private: |
| 922 | Expression* condition_; |
| 923 | Label* true_label_; |
| 924 | Label* false_label_; |
| 925 | Label* fall_through_; |
| 926 | }; |
| 927 | |
| 928 | class EffectContext : public ExpressionContext { |
| 929 | public: |
| 930 | explicit EffectContext(FullCodeGenerator* codegen) |
| 931 | : ExpressionContext(codegen) { } |
| 932 | |
| 933 | void Plug(bool flag) const override; |
| 934 | void Plug(Register reg) const override; |
| 935 | void Plug(Label* materialize_true, Label* materialize_false) const override; |
| 936 | void Plug(Variable* var) const override; |
| 937 | void Plug(Handle<Object> lit) const override; |
| 938 | void Plug(Heap::RootListIndex) const override; |
| 939 | void PlugTOS() const override; |
| 940 | void DropAndPlug(int count, Register reg) const override; |
| 941 | void PrepareTest(Label* materialize_true, Label* materialize_false, |
| 942 | Label** if_true, Label** if_false, |
| 943 | Label** fall_through) const override; |
| 944 | bool IsEffect() const override { return true; } |
| 945 | }; |
| 946 | |
| 947 | class EnterBlockScopeIfNeeded { |
| 948 | public: |
| 949 | EnterBlockScopeIfNeeded(FullCodeGenerator* codegen, Scope* scope, |
| 950 | BailoutId entry_id, BailoutId declarations_id, |
| 951 | BailoutId exit_id); |
| 952 | ~EnterBlockScopeIfNeeded(); |
| 953 | |
| 954 | private: |
| 955 | MacroAssembler* masm() const { return codegen_->masm(); } |
| 956 | |
| 957 | FullCodeGenerator* codegen_; |
| 958 | Scope* saved_scope_; |
| 959 | BailoutId exit_id_; |
| 960 | bool needs_block_context_; |
| 961 | }; |
| 962 | |
| 963 | MacroAssembler* masm_; |
| 964 | CompilationInfo* info_; |
| 965 | Isolate* isolate_; |
| 966 | Zone* zone_; |
| 967 | Scope* scope_; |
| 968 | Label return_label_; |
| 969 | NestedStatement* nesting_stack_; |
| 970 | int loop_depth_; |
| 971 | int try_catch_depth_; |
Ben Murdoch | 097c5b2 | 2016-05-18 11:27:45 +0100 | [diff] [blame] | 972 | int operand_stack_depth_; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 973 | ZoneList<Handle<Object> >* globals_; |
| 974 | Handle<FixedArray> modules_; |
| 975 | int module_index_; |
| 976 | const ExpressionContext* context_; |
| 977 | ZoneList<BailoutEntry> bailout_entries_; |
| 978 | ZoneList<BackEdgeEntry> back_edges_; |
| 979 | ZoneVector<HandlerTableEntry> handler_table_; |
| 980 | int ic_total_count_; |
| 981 | Handle<Cell> profiling_counter_; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 982 | |
| 983 | friend class NestedStatement; |
| 984 | |
| 985 | DEFINE_AST_VISITOR_SUBCLASS_MEMBERS(); |
| 986 | DISALLOW_COPY_AND_ASSIGN(FullCodeGenerator); |
| 987 | }; |
| 988 | |
| 989 | |
| 990 | class BackEdgeTable { |
| 991 | public: |
| 992 | BackEdgeTable(Code* code, DisallowHeapAllocation* required) { |
| 993 | DCHECK(code->kind() == Code::FUNCTION); |
| 994 | instruction_start_ = code->instruction_start(); |
| 995 | Address table_address = instruction_start_ + code->back_edge_table_offset(); |
| 996 | length_ = Memory::uint32_at(table_address); |
| 997 | start_ = table_address + kTableLengthSize; |
| 998 | } |
| 999 | |
| 1000 | uint32_t length() { return length_; } |
| 1001 | |
| 1002 | BailoutId ast_id(uint32_t index) { |
| 1003 | return BailoutId(static_cast<int>( |
| 1004 | Memory::uint32_at(entry_at(index) + kAstIdOffset))); |
| 1005 | } |
| 1006 | |
| 1007 | uint32_t loop_depth(uint32_t index) { |
| 1008 | return Memory::uint32_at(entry_at(index) + kLoopDepthOffset); |
| 1009 | } |
| 1010 | |
| 1011 | uint32_t pc_offset(uint32_t index) { |
| 1012 | return Memory::uint32_at(entry_at(index) + kPcOffsetOffset); |
| 1013 | } |
| 1014 | |
| 1015 | Address pc(uint32_t index) { |
| 1016 | return instruction_start_ + pc_offset(index); |
| 1017 | } |
| 1018 | |
Ben Murdoch | da12d29 | 2016-06-02 14:46:10 +0100 | [diff] [blame] | 1019 | enum BackEdgeState { INTERRUPT, ON_STACK_REPLACEMENT }; |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 1020 | |
| 1021 | // Increase allowed loop nesting level by one and patch those matching loops. |
| 1022 | static void Patch(Isolate* isolate, Code* unoptimized_code); |
| 1023 | |
| 1024 | // Patch the back edge to the target state, provided the correct callee. |
| 1025 | static void PatchAt(Code* unoptimized_code, |
| 1026 | Address pc, |
| 1027 | BackEdgeState target_state, |
| 1028 | Code* replacement_code); |
| 1029 | |
| 1030 | // Change all patched back edges back to normal interrupts. |
| 1031 | static void Revert(Isolate* isolate, |
| 1032 | Code* unoptimized_code); |
| 1033 | |
Ben Murdoch | 4a90d5f | 2016-03-22 12:00:34 +0000 | [diff] [blame] | 1034 | // Return the current patch state of the back edge. |
| 1035 | static BackEdgeState GetBackEdgeState(Isolate* isolate, |
| 1036 | Code* unoptimized_code, |
| 1037 | Address pc_after); |
| 1038 | |
| 1039 | #ifdef DEBUG |
| 1040 | // Verify that all back edges of a certain loop depth are patched. |
| 1041 | static bool Verify(Isolate* isolate, Code* unoptimized_code); |
| 1042 | #endif // DEBUG |
| 1043 | |
| 1044 | private: |
| 1045 | Address entry_at(uint32_t index) { |
| 1046 | DCHECK(index < length_); |
| 1047 | return start_ + index * kEntrySize; |
| 1048 | } |
| 1049 | |
| 1050 | static const int kTableLengthSize = kIntSize; |
| 1051 | static const int kAstIdOffset = 0 * kIntSize; |
| 1052 | static const int kPcOffsetOffset = 1 * kIntSize; |
| 1053 | static const int kLoopDepthOffset = 2 * kIntSize; |
| 1054 | static const int kEntrySize = 3 * kIntSize; |
| 1055 | |
| 1056 | Address start_; |
| 1057 | Address instruction_start_; |
| 1058 | uint32_t length_; |
| 1059 | }; |
| 1060 | |
| 1061 | |
| 1062 | } // namespace internal |
| 1063 | } // namespace v8 |
| 1064 | |
| 1065 | #endif // V8_FULL_CODEGEN_FULL_CODEGEN_H_ |