blob: ee4358955f58a6c52feb665431458ee32adfe202 [file] [log] [blame]
Ben Murdochb8a8cc12014-11-26 15:28:44 +00001// Copyright 2013 the V8 project authors. All rights reserved.
2// Use of this source code is governed by a BSD-style license that can be
3// found in the LICENSE file.
4
5#ifndef V8_ARM64_MACRO_ASSEMBLER_ARM64_H_
6#define V8_ARM64_MACRO_ASSEMBLER_ARM64_H_
7
8#include <vector>
9
10#include "src/bailout-reason.h"
11#include "src/globals.h"
12
13#include "src/arm64/assembler-arm64-inl.h"
14#include "src/base/bits.h"
15
16// Simulator specific helpers.
17#if USE_SIMULATOR
18 // TODO(all): If possible automatically prepend an indicator like
19 // UNIMPLEMENTED or LOCATION.
20 #define ASM_UNIMPLEMENTED(message) \
21 __ Debug(message, __LINE__, NO_PARAM)
22 #define ASM_UNIMPLEMENTED_BREAK(message) \
23 __ Debug(message, __LINE__, \
24 FLAG_ignore_asm_unimplemented_break ? NO_PARAM : BREAK)
25 #define ASM_LOCATION(message) \
26 __ Debug("LOCATION: " message, __LINE__, NO_PARAM)
27#else
28 #define ASM_UNIMPLEMENTED(message)
29 #define ASM_UNIMPLEMENTED_BREAK(message)
30 #define ASM_LOCATION(message)
31#endif
32
33
34namespace v8 {
35namespace internal {
36
37#define LS_MACRO_LIST(V) \
38 V(Ldrb, Register&, rt, LDRB_w) \
39 V(Strb, Register&, rt, STRB_w) \
40 V(Ldrsb, Register&, rt, rt.Is64Bits() ? LDRSB_x : LDRSB_w) \
41 V(Ldrh, Register&, rt, LDRH_w) \
42 V(Strh, Register&, rt, STRH_w) \
43 V(Ldrsh, Register&, rt, rt.Is64Bits() ? LDRSH_x : LDRSH_w) \
44 V(Ldr, CPURegister&, rt, LoadOpFor(rt)) \
45 V(Str, CPURegister&, rt, StoreOpFor(rt)) \
46 V(Ldrsw, Register&, rt, LDRSW_x)
47
48#define LSPAIR_MACRO_LIST(V) \
49 V(Ldp, CPURegister&, rt, rt2, LoadPairOpFor(rt, rt2)) \
50 V(Stp, CPURegister&, rt, rt2, StorePairOpFor(rt, rt2)) \
51 V(Ldpsw, CPURegister&, rt, rt2, LDPSW_x)
52
53
54// ----------------------------------------------------------------------------
55// Static helper functions
56
57// Generate a MemOperand for loading a field from an object.
58inline MemOperand FieldMemOperand(Register object, int offset);
59inline MemOperand UntagSmiFieldMemOperand(Register object, int offset);
60
61// Generate a MemOperand for loading a SMI from memory.
62inline MemOperand UntagSmiMemOperand(Register object, int offset);
63
64
65// ----------------------------------------------------------------------------
66// MacroAssembler
67
68enum BranchType {
69 // Copies of architectural conditions.
70 // The associated conditions can be used in place of those, the code will
71 // take care of reinterpreting them with the correct type.
72 integer_eq = eq,
73 integer_ne = ne,
74 integer_hs = hs,
75 integer_lo = lo,
76 integer_mi = mi,
77 integer_pl = pl,
78 integer_vs = vs,
79 integer_vc = vc,
80 integer_hi = hi,
81 integer_ls = ls,
82 integer_ge = ge,
83 integer_lt = lt,
84 integer_gt = gt,
85 integer_le = le,
86 integer_al = al,
87 integer_nv = nv,
88
89 // These two are *different* from the architectural codes al and nv.
90 // 'always' is used to generate unconditional branches.
91 // 'never' is used to not generate a branch (generally as the inverse
92 // branch type of 'always).
93 always, never,
94 // cbz and cbnz
95 reg_zero, reg_not_zero,
96 // tbz and tbnz
97 reg_bit_clear, reg_bit_set,
98
99 // Aliases.
100 kBranchTypeFirstCondition = eq,
101 kBranchTypeLastCondition = nv,
102 kBranchTypeFirstUsingReg = reg_zero,
103 kBranchTypeFirstUsingBit = reg_bit_clear
104};
105
106inline BranchType InvertBranchType(BranchType type) {
107 if (kBranchTypeFirstCondition <= type && type <= kBranchTypeLastCondition) {
108 return static_cast<BranchType>(
109 NegateCondition(static_cast<Condition>(type)));
110 } else {
111 return static_cast<BranchType>(type ^ 1);
112 }
113}
114
115enum RememberedSetAction { EMIT_REMEMBERED_SET, OMIT_REMEMBERED_SET };
116enum SmiCheck { INLINE_SMI_CHECK, OMIT_SMI_CHECK };
117enum PointersToHereCheck {
118 kPointersToHereMaybeInteresting,
119 kPointersToHereAreAlwaysInteresting
120};
121enum LinkRegisterStatus { kLRHasNotBeenSaved, kLRHasBeenSaved };
122enum TargetAddressStorageMode {
123 CAN_INLINE_TARGET_ADDRESS,
124 NEVER_INLINE_TARGET_ADDRESS
125};
126enum UntagMode { kNotSpeculativeUntag, kSpeculativeUntag };
127enum ArrayHasHoles { kArrayCantHaveHoles, kArrayCanHaveHoles };
128enum CopyHint { kCopyUnknown, kCopyShort, kCopyLong };
129enum DiscardMoveMode { kDontDiscardForSameWReg, kDiscardForSameWReg };
130enum SeqStringSetCharCheckIndexType { kIndexIsSmi, kIndexIsInteger32 };
131
132class MacroAssembler : public Assembler {
133 public:
134 MacroAssembler(Isolate* isolate, byte * buffer, unsigned buffer_size);
135
136 inline Handle<Object> CodeObject();
137
138 // Instruction set functions ------------------------------------------------
139 // Logical macros.
140 inline void And(const Register& rd,
141 const Register& rn,
142 const Operand& operand);
143 inline void Ands(const Register& rd,
144 const Register& rn,
145 const Operand& operand);
146 inline void Bic(const Register& rd,
147 const Register& rn,
148 const Operand& operand);
149 inline void Bics(const Register& rd,
150 const Register& rn,
151 const Operand& operand);
152 inline void Orr(const Register& rd,
153 const Register& rn,
154 const Operand& operand);
155 inline void Orn(const Register& rd,
156 const Register& rn,
157 const Operand& operand);
158 inline void Eor(const Register& rd,
159 const Register& rn,
160 const Operand& operand);
161 inline void Eon(const Register& rd,
162 const Register& rn,
163 const Operand& operand);
164 inline void Tst(const Register& rn, const Operand& operand);
165 void LogicalMacro(const Register& rd,
166 const Register& rn,
167 const Operand& operand,
168 LogicalOp op);
169
170 // Add and sub macros.
171 inline void Add(const Register& rd,
172 const Register& rn,
173 const Operand& operand);
174 inline void Adds(const Register& rd,
175 const Register& rn,
176 const Operand& operand);
177 inline void Sub(const Register& rd,
178 const Register& rn,
179 const Operand& operand);
180 inline void Subs(const Register& rd,
181 const Register& rn,
182 const Operand& operand);
183 inline void Cmn(const Register& rn, const Operand& operand);
184 inline void Cmp(const Register& rn, const Operand& operand);
185 inline void Neg(const Register& rd,
186 const Operand& operand);
187 inline void Negs(const Register& rd,
188 const Operand& operand);
189
190 void AddSubMacro(const Register& rd,
191 const Register& rn,
192 const Operand& operand,
193 FlagsUpdate S,
194 AddSubOp op);
195
196 // Add/sub with carry macros.
197 inline void Adc(const Register& rd,
198 const Register& rn,
199 const Operand& operand);
200 inline void Adcs(const Register& rd,
201 const Register& rn,
202 const Operand& operand);
203 inline void Sbc(const Register& rd,
204 const Register& rn,
205 const Operand& operand);
206 inline void Sbcs(const Register& rd,
207 const Register& rn,
208 const Operand& operand);
209 inline void Ngc(const Register& rd,
210 const Operand& operand);
211 inline void Ngcs(const Register& rd,
212 const Operand& operand);
213 void AddSubWithCarryMacro(const Register& rd,
214 const Register& rn,
215 const Operand& operand,
216 FlagsUpdate S,
217 AddSubWithCarryOp op);
218
219 // Move macros.
220 void Mov(const Register& rd,
221 const Operand& operand,
222 DiscardMoveMode discard_mode = kDontDiscardForSameWReg);
223 void Mov(const Register& rd, uint64_t imm);
224 inline void Mvn(const Register& rd, uint64_t imm);
225 void Mvn(const Register& rd, const Operand& operand);
226 static bool IsImmMovn(uint64_t imm, unsigned reg_size);
227 static bool IsImmMovz(uint64_t imm, unsigned reg_size);
228 static unsigned CountClearHalfWords(uint64_t imm, unsigned reg_size);
229
230 // Try to move an immediate into the destination register in a single
231 // instruction. Returns true for success, and updates the contents of dst.
232 // Returns false, otherwise.
233 bool TryOneInstrMoveImmediate(const Register& dst, int64_t imm);
234
235 // Move an immediate into register dst, and return an Operand object for use
236 // with a subsequent instruction that accepts a shift. The value moved into
237 // dst is not necessarily equal to imm; it may have had a shifting operation
238 // applied to it that will be subsequently undone by the shift applied in the
239 // Operand.
240 Operand MoveImmediateForShiftedOp(const Register& dst, int64_t imm);
241
242 // Conditional macros.
243 inline void Ccmp(const Register& rn,
244 const Operand& operand,
245 StatusFlags nzcv,
246 Condition cond);
247 inline void Ccmn(const Register& rn,
248 const Operand& operand,
249 StatusFlags nzcv,
250 Condition cond);
251 void ConditionalCompareMacro(const Register& rn,
252 const Operand& operand,
253 StatusFlags nzcv,
254 Condition cond,
255 ConditionalCompareOp op);
256 void Csel(const Register& rd,
257 const Register& rn,
258 const Operand& operand,
259 Condition cond);
260
261 // Load/store macros.
262#define DECLARE_FUNCTION(FN, REGTYPE, REG, OP) \
263 inline void FN(const REGTYPE REG, const MemOperand& addr);
264 LS_MACRO_LIST(DECLARE_FUNCTION)
265#undef DECLARE_FUNCTION
266
267 void LoadStoreMacro(const CPURegister& rt,
268 const MemOperand& addr,
269 LoadStoreOp op);
270
271#define DECLARE_FUNCTION(FN, REGTYPE, REG, REG2, OP) \
272 inline void FN(const REGTYPE REG, const REGTYPE REG2, const MemOperand& addr);
273 LSPAIR_MACRO_LIST(DECLARE_FUNCTION)
274#undef DECLARE_FUNCTION
275
276 void LoadStorePairMacro(const CPURegister& rt, const CPURegister& rt2,
277 const MemOperand& addr, LoadStorePairOp op);
278
279 // V8-specific load/store helpers.
280 void Load(const Register& rt, const MemOperand& addr, Representation r);
281 void Store(const Register& rt, const MemOperand& addr, Representation r);
282
283 enum AdrHint {
284 // The target must be within the immediate range of adr.
285 kAdrNear,
286 // The target may be outside of the immediate range of adr. Additional
287 // instructions may be emitted.
288 kAdrFar
289 };
290 void Adr(const Register& rd, Label* label, AdrHint = kAdrNear);
291
292 // Remaining instructions are simple pass-through calls to the assembler.
293 inline void Asr(const Register& rd, const Register& rn, unsigned shift);
294 inline void Asr(const Register& rd, const Register& rn, const Register& rm);
295
296 // Branch type inversion relies on these relations.
297 STATIC_ASSERT((reg_zero == (reg_not_zero ^ 1)) &&
298 (reg_bit_clear == (reg_bit_set ^ 1)) &&
299 (always == (never ^ 1)));
300
301 void B(Label* label, BranchType type, Register reg = NoReg, int bit = -1);
302
303 inline void B(Label* label);
304 inline void B(Condition cond, Label* label);
305 void B(Label* label, Condition cond);
306 inline void Bfi(const Register& rd,
307 const Register& rn,
308 unsigned lsb,
309 unsigned width);
310 inline void Bfxil(const Register& rd,
311 const Register& rn,
312 unsigned lsb,
313 unsigned width);
314 inline void Bind(Label* label);
315 inline void Bl(Label* label);
316 inline void Blr(const Register& xn);
317 inline void Br(const Register& xn);
318 inline void Brk(int code);
319 void Cbnz(const Register& rt, Label* label);
320 void Cbz(const Register& rt, Label* label);
321 inline void Cinc(const Register& rd, const Register& rn, Condition cond);
322 inline void Cinv(const Register& rd, const Register& rn, Condition cond);
323 inline void Cls(const Register& rd, const Register& rn);
324 inline void Clz(const Register& rd, const Register& rn);
325 inline void Cneg(const Register& rd, const Register& rn, Condition cond);
326 inline void CzeroX(const Register& rd, Condition cond);
327 inline void CmovX(const Register& rd, const Register& rn, Condition cond);
328 inline void Cset(const Register& rd, Condition cond);
329 inline void Csetm(const Register& rd, Condition cond);
330 inline void Csinc(const Register& rd,
331 const Register& rn,
332 const Register& rm,
333 Condition cond);
334 inline void Csinv(const Register& rd,
335 const Register& rn,
336 const Register& rm,
337 Condition cond);
338 inline void Csneg(const Register& rd,
339 const Register& rn,
340 const Register& rm,
341 Condition cond);
342 inline void Dmb(BarrierDomain domain, BarrierType type);
343 inline void Dsb(BarrierDomain domain, BarrierType type);
344 inline void Debug(const char* message, uint32_t code, Instr params = BREAK);
345 inline void Extr(const Register& rd,
346 const Register& rn,
347 const Register& rm,
348 unsigned lsb);
349 inline void Fabs(const FPRegister& fd, const FPRegister& fn);
350 inline void Fadd(const FPRegister& fd,
351 const FPRegister& fn,
352 const FPRegister& fm);
353 inline void Fccmp(const FPRegister& fn,
354 const FPRegister& fm,
355 StatusFlags nzcv,
356 Condition cond);
357 inline void Fcmp(const FPRegister& fn, const FPRegister& fm);
358 inline void Fcmp(const FPRegister& fn, double value);
359 inline void Fcsel(const FPRegister& fd,
360 const FPRegister& fn,
361 const FPRegister& fm,
362 Condition cond);
363 inline void Fcvt(const FPRegister& fd, const FPRegister& fn);
364 inline void Fcvtas(const Register& rd, const FPRegister& fn);
365 inline void Fcvtau(const Register& rd, const FPRegister& fn);
366 inline void Fcvtms(const Register& rd, const FPRegister& fn);
367 inline void Fcvtmu(const Register& rd, const FPRegister& fn);
368 inline void Fcvtns(const Register& rd, const FPRegister& fn);
369 inline void Fcvtnu(const Register& rd, const FPRegister& fn);
370 inline void Fcvtzs(const Register& rd, const FPRegister& fn);
371 inline void Fcvtzu(const Register& rd, const FPRegister& fn);
372 inline void Fdiv(const FPRegister& fd,
373 const FPRegister& fn,
374 const FPRegister& fm);
375 inline void Fmadd(const FPRegister& fd,
376 const FPRegister& fn,
377 const FPRegister& fm,
378 const FPRegister& fa);
379 inline void Fmax(const FPRegister& fd,
380 const FPRegister& fn,
381 const FPRegister& fm);
382 inline void Fmaxnm(const FPRegister& fd,
383 const FPRegister& fn,
384 const FPRegister& fm);
385 inline void Fmin(const FPRegister& fd,
386 const FPRegister& fn,
387 const FPRegister& fm);
388 inline void Fminnm(const FPRegister& fd,
389 const FPRegister& fn,
390 const FPRegister& fm);
391 inline void Fmov(FPRegister fd, FPRegister fn);
392 inline void Fmov(FPRegister fd, Register rn);
393 // Provide explicit double and float interfaces for FP immediate moves, rather
394 // than relying on implicit C++ casts. This allows signalling NaNs to be
395 // preserved when the immediate matches the format of fd. Most systems convert
396 // signalling NaNs to quiet NaNs when converting between float and double.
397 inline void Fmov(FPRegister fd, double imm);
398 inline void Fmov(FPRegister fd, float imm);
399 // Provide a template to allow other types to be converted automatically.
400 template<typename T>
401 void Fmov(FPRegister fd, T imm) {
402 DCHECK(allow_macro_instructions_);
403 Fmov(fd, static_cast<double>(imm));
404 }
405 inline void Fmov(Register rd, FPRegister fn);
406 inline void Fmsub(const FPRegister& fd,
407 const FPRegister& fn,
408 const FPRegister& fm,
409 const FPRegister& fa);
410 inline void Fmul(const FPRegister& fd,
411 const FPRegister& fn,
412 const FPRegister& fm);
413 inline void Fneg(const FPRegister& fd, const FPRegister& fn);
414 inline void Fnmadd(const FPRegister& fd,
415 const FPRegister& fn,
416 const FPRegister& fm,
417 const FPRegister& fa);
418 inline void Fnmsub(const FPRegister& fd,
419 const FPRegister& fn,
420 const FPRegister& fm,
421 const FPRegister& fa);
422 inline void Frinta(const FPRegister& fd, const FPRegister& fn);
423 inline void Frintm(const FPRegister& fd, const FPRegister& fn);
424 inline void Frintn(const FPRegister& fd, const FPRegister& fn);
Emily Bernierd0a1eb72015-03-24 16:35:39 -0400425 inline void Frintp(const FPRegister& fd, const FPRegister& fn);
Ben Murdochb8a8cc12014-11-26 15:28:44 +0000426 inline void Frintz(const FPRegister& fd, const FPRegister& fn);
427 inline void Fsqrt(const FPRegister& fd, const FPRegister& fn);
428 inline void Fsub(const FPRegister& fd,
429 const FPRegister& fn,
430 const FPRegister& fm);
431 inline void Hint(SystemHint code);
432 inline void Hlt(int code);
433 inline void Isb();
434 inline void Ldnp(const CPURegister& rt,
435 const CPURegister& rt2,
436 const MemOperand& src);
437 // Load a literal from the inline constant pool.
438 inline void Ldr(const CPURegister& rt, const Immediate& imm);
439 // Helper function for double immediate.
440 inline void Ldr(const CPURegister& rt, double imm);
441 inline void Lsl(const Register& rd, const Register& rn, unsigned shift);
442 inline void Lsl(const Register& rd, const Register& rn, const Register& rm);
443 inline void Lsr(const Register& rd, const Register& rn, unsigned shift);
444 inline void Lsr(const Register& rd, const Register& rn, const Register& rm);
445 inline void Madd(const Register& rd,
446 const Register& rn,
447 const Register& rm,
448 const Register& ra);
449 inline void Mneg(const Register& rd, const Register& rn, const Register& rm);
450 inline void Mov(const Register& rd, const Register& rm);
451 inline void Movk(const Register& rd, uint64_t imm, int shift = -1);
452 inline void Mrs(const Register& rt, SystemRegister sysreg);
453 inline void Msr(SystemRegister sysreg, const Register& rt);
454 inline void Msub(const Register& rd,
455 const Register& rn,
456 const Register& rm,
457 const Register& ra);
458 inline void Mul(const Register& rd, const Register& rn, const Register& rm);
459 inline void Nop() { nop(); }
460 inline void Rbit(const Register& rd, const Register& rn);
461 inline void Ret(const Register& xn = lr);
462 inline void Rev(const Register& rd, const Register& rn);
463 inline void Rev16(const Register& rd, const Register& rn);
464 inline void Rev32(const Register& rd, const Register& rn);
465 inline void Ror(const Register& rd, const Register& rs, unsigned shift);
466 inline void Ror(const Register& rd, const Register& rn, const Register& rm);
467 inline void Sbfiz(const Register& rd,
468 const Register& rn,
469 unsigned lsb,
470 unsigned width);
471 inline void Sbfx(const Register& rd,
472 const Register& rn,
473 unsigned lsb,
474 unsigned width);
475 inline void Scvtf(const FPRegister& fd,
476 const Register& rn,
477 unsigned fbits = 0);
478 inline void Sdiv(const Register& rd, const Register& rn, const Register& rm);
479 inline void Smaddl(const Register& rd,
480 const Register& rn,
481 const Register& rm,
482 const Register& ra);
483 inline void Smsubl(const Register& rd,
484 const Register& rn,
485 const Register& rm,
486 const Register& ra);
487 inline void Smull(const Register& rd,
488 const Register& rn,
489 const Register& rm);
490 inline void Smulh(const Register& rd,
491 const Register& rn,
492 const Register& rm);
Emily Bernierd0a1eb72015-03-24 16:35:39 -0400493 inline void Umull(const Register& rd, const Register& rn, const Register& rm);
Ben Murdochb8a8cc12014-11-26 15:28:44 +0000494 inline void Stnp(const CPURegister& rt,
495 const CPURegister& rt2,
496 const MemOperand& dst);
497 inline void Sxtb(const Register& rd, const Register& rn);
498 inline void Sxth(const Register& rd, const Register& rn);
499 inline void Sxtw(const Register& rd, const Register& rn);
500 void Tbnz(const Register& rt, unsigned bit_pos, Label* label);
501 void Tbz(const Register& rt, unsigned bit_pos, Label* label);
502 inline void Ubfiz(const Register& rd,
503 const Register& rn,
504 unsigned lsb,
505 unsigned width);
506 inline void Ubfx(const Register& rd,
507 const Register& rn,
508 unsigned lsb,
509 unsigned width);
510 inline void Ucvtf(const FPRegister& fd,
511 const Register& rn,
512 unsigned fbits = 0);
513 inline void Udiv(const Register& rd, const Register& rn, const Register& rm);
514 inline void Umaddl(const Register& rd,
515 const Register& rn,
516 const Register& rm,
517 const Register& ra);
518 inline void Umsubl(const Register& rd,
519 const Register& rn,
520 const Register& rm,
521 const Register& ra);
522 inline void Uxtb(const Register& rd, const Register& rn);
523 inline void Uxth(const Register& rd, const Register& rn);
524 inline void Uxtw(const Register& rd, const Register& rn);
525
526 // Pseudo-instructions ------------------------------------------------------
527
528 // Compute rd = abs(rm).
529 // This function clobbers the condition flags. On output the overflow flag is
530 // set iff the negation overflowed.
531 //
532 // If rm is the minimum representable value, the result is not representable.
533 // Handlers for each case can be specified using the relevant labels.
534 void Abs(const Register& rd, const Register& rm,
535 Label * is_not_representable = NULL,
536 Label * is_representable = NULL);
537
538 // Push or pop up to 4 registers of the same width to or from the stack,
539 // using the current stack pointer as set by SetStackPointer.
540 //
541 // If an argument register is 'NoReg', all further arguments are also assumed
542 // to be 'NoReg', and are thus not pushed or popped.
543 //
544 // Arguments are ordered such that "Push(a, b);" is functionally equivalent
545 // to "Push(a); Push(b);".
546 //
547 // It is valid to push the same register more than once, and there is no
548 // restriction on the order in which registers are specified.
549 //
550 // It is not valid to pop into the same register more than once in one
551 // operation, not even into the zero register.
552 //
553 // If the current stack pointer (as set by SetStackPointer) is csp, then it
554 // must be aligned to 16 bytes on entry and the total size of the specified
555 // registers must also be a multiple of 16 bytes.
556 //
557 // Even if the current stack pointer is not the system stack pointer (csp),
558 // Push (and derived methods) will still modify the system stack pointer in
559 // order to comply with ABI rules about accessing memory below the system
560 // stack pointer.
561 //
562 // Other than the registers passed into Pop, the stack pointer and (possibly)
563 // the system stack pointer, these methods do not modify any other registers.
564 void Push(const CPURegister& src0, const CPURegister& src1 = NoReg,
565 const CPURegister& src2 = NoReg, const CPURegister& src3 = NoReg);
566 void Push(const CPURegister& src0, const CPURegister& src1,
567 const CPURegister& src2, const CPURegister& src3,
568 const CPURegister& src4, const CPURegister& src5 = NoReg,
569 const CPURegister& src6 = NoReg, const CPURegister& src7 = NoReg);
570 void Pop(const CPURegister& dst0, const CPURegister& dst1 = NoReg,
571 const CPURegister& dst2 = NoReg, const CPURegister& dst3 = NoReg);
572 void Push(const Register& src0, const FPRegister& src1);
573
574 // Alternative forms of Push and Pop, taking a RegList or CPURegList that
575 // specifies the registers that are to be pushed or popped. Higher-numbered
576 // registers are associated with higher memory addresses (as in the A32 push
577 // and pop instructions).
578 //
579 // (Push|Pop)SizeRegList allow you to specify the register size as a
580 // parameter. Only kXRegSizeInBits, kWRegSizeInBits, kDRegSizeInBits and
581 // kSRegSizeInBits are supported.
582 //
583 // Otherwise, (Push|Pop)(CPU|X|W|D|S)RegList is preferred.
584 void PushCPURegList(CPURegList registers);
585 void PopCPURegList(CPURegList registers);
586
587 inline void PushSizeRegList(RegList registers, unsigned reg_size,
588 CPURegister::RegisterType type = CPURegister::kRegister) {
589 PushCPURegList(CPURegList(type, reg_size, registers));
590 }
591 inline void PopSizeRegList(RegList registers, unsigned reg_size,
592 CPURegister::RegisterType type = CPURegister::kRegister) {
593 PopCPURegList(CPURegList(type, reg_size, registers));
594 }
595 inline void PushXRegList(RegList regs) {
596 PushSizeRegList(regs, kXRegSizeInBits);
597 }
598 inline void PopXRegList(RegList regs) {
599 PopSizeRegList(regs, kXRegSizeInBits);
600 }
601 inline void PushWRegList(RegList regs) {
602 PushSizeRegList(regs, kWRegSizeInBits);
603 }
604 inline void PopWRegList(RegList regs) {
605 PopSizeRegList(regs, kWRegSizeInBits);
606 }
607 inline void PushDRegList(RegList regs) {
608 PushSizeRegList(regs, kDRegSizeInBits, CPURegister::kFPRegister);
609 }
610 inline void PopDRegList(RegList regs) {
611 PopSizeRegList(regs, kDRegSizeInBits, CPURegister::kFPRegister);
612 }
613 inline void PushSRegList(RegList regs) {
614 PushSizeRegList(regs, kSRegSizeInBits, CPURegister::kFPRegister);
615 }
616 inline void PopSRegList(RegList regs) {
617 PopSizeRegList(regs, kSRegSizeInBits, CPURegister::kFPRegister);
618 }
619
620 // Push the specified register 'count' times.
621 void PushMultipleTimes(CPURegister src, Register count);
622 void PushMultipleTimes(CPURegister src, int count);
623
624 // This is a convenience method for pushing a single Handle<Object>.
625 inline void Push(Handle<Object> handle);
626 void Push(Smi* smi) { Push(Handle<Smi>(smi, isolate())); }
627
628 // Aliases of Push and Pop, required for V8 compatibility.
629 inline void push(Register src) {
630 Push(src);
631 }
632 inline void pop(Register dst) {
633 Pop(dst);
634 }
635
636 // Sometimes callers need to push or pop multiple registers in a way that is
637 // difficult to structure efficiently for fixed Push or Pop calls. This scope
638 // allows push requests to be queued up, then flushed at once. The
639 // MacroAssembler will try to generate the most efficient sequence required.
640 //
641 // Unlike the other Push and Pop macros, PushPopQueue can handle mixed sets of
642 // register sizes and types.
643 class PushPopQueue {
644 public:
645 explicit PushPopQueue(MacroAssembler* masm) : masm_(masm), size_(0) { }
646
647 ~PushPopQueue() {
648 DCHECK(queued_.empty());
649 }
650
651 void Queue(const CPURegister& rt) {
652 size_ += rt.SizeInBytes();
653 queued_.push_back(rt);
654 }
655
656 enum PreambleDirective {
657 WITH_PREAMBLE,
658 SKIP_PREAMBLE
659 };
660 void PushQueued(PreambleDirective preamble_directive = WITH_PREAMBLE);
661 void PopQueued();
662
663 private:
664 MacroAssembler* masm_;
665 int size_;
666 std::vector<CPURegister> queued_;
667 };
668
669 // Poke 'src' onto the stack. The offset is in bytes.
670 //
671 // If the current stack pointer (according to StackPointer()) is csp, then
672 // csp must be aligned to 16 bytes.
673 void Poke(const CPURegister& src, const Operand& offset);
674
675 // Peek at a value on the stack, and put it in 'dst'. The offset is in bytes.
676 //
677 // If the current stack pointer (according to StackPointer()) is csp, then
678 // csp must be aligned to 16 bytes.
679 void Peek(const CPURegister& dst, const Operand& offset);
680
681 // Poke 'src1' and 'src2' onto the stack. The values written will be adjacent
682 // with 'src2' at a higher address than 'src1'. The offset is in bytes.
683 //
684 // If the current stack pointer (according to StackPointer()) is csp, then
685 // csp must be aligned to 16 bytes.
686 void PokePair(const CPURegister& src1, const CPURegister& src2, int offset);
687
688 // Peek at two values on the stack, and put them in 'dst1' and 'dst2'. The
689 // values peeked will be adjacent, with the value in 'dst2' being from a
690 // higher address than 'dst1'. The offset is in bytes.
691 //
692 // If the current stack pointer (according to StackPointer()) is csp, then
693 // csp must be aligned to 16 bytes.
694 void PeekPair(const CPURegister& dst1, const CPURegister& dst2, int offset);
695
696 // Claim or drop stack space without actually accessing memory.
697 //
698 // In debug mode, both of these will write invalid data into the claimed or
699 // dropped space.
700 //
701 // If the current stack pointer (according to StackPointer()) is csp, then it
702 // must be aligned to 16 bytes and the size claimed or dropped must be a
703 // multiple of 16 bytes.
704 //
705 // Note that unit_size must be specified in bytes. For variants which take a
706 // Register count, the unit size must be a power of two.
707 inline void Claim(uint64_t count, uint64_t unit_size = kXRegSize);
708 inline void Claim(const Register& count,
709 uint64_t unit_size = kXRegSize);
710 inline void Drop(uint64_t count, uint64_t unit_size = kXRegSize);
711 inline void Drop(const Register& count,
712 uint64_t unit_size = kXRegSize);
713
714 // Variants of Claim and Drop, where the 'count' parameter is a SMI held in a
715 // register.
716 inline void ClaimBySMI(const Register& count_smi,
717 uint64_t unit_size = kXRegSize);
718 inline void DropBySMI(const Register& count_smi,
719 uint64_t unit_size = kXRegSize);
720
721 // Compare a register with an operand, and branch to label depending on the
722 // condition. May corrupt the status flags.
723 inline void CompareAndBranch(const Register& lhs,
724 const Operand& rhs,
725 Condition cond,
726 Label* label);
727
728 // Test the bits of register defined by bit_pattern, and branch if ANY of
729 // those bits are set. May corrupt the status flags.
730 inline void TestAndBranchIfAnySet(const Register& reg,
731 const uint64_t bit_pattern,
732 Label* label);
733
734 // Test the bits of register defined by bit_pattern, and branch if ALL of
735 // those bits are clear (ie. not set.) May corrupt the status flags.
736 inline void TestAndBranchIfAllClear(const Register& reg,
737 const uint64_t bit_pattern,
738 Label* label);
739
740 // Insert one or more instructions into the instruction stream that encode
741 // some caller-defined data. The instructions used will be executable with no
742 // side effects.
743 inline void InlineData(uint64_t data);
744
745 // Insert an instrumentation enable marker into the instruction stream.
746 inline void EnableInstrumentation();
747
748 // Insert an instrumentation disable marker into the instruction stream.
749 inline void DisableInstrumentation();
750
751 // Insert an instrumentation event marker into the instruction stream. These
752 // will be picked up by the instrumentation system to annotate an instruction
753 // profile. The argument marker_name must be a printable two character string;
754 // it will be encoded in the event marker.
755 inline void AnnotateInstrumentation(const char* marker_name);
756
757 // If emit_debug_code() is true, emit a run-time check to ensure that
758 // StackPointer() does not point below the system stack pointer.
759 //
760 // Whilst it is architecturally legal for StackPointer() to point below csp,
761 // it can be evidence of a potential bug because the ABI forbids accesses
762 // below csp.
763 //
Emily Bernierd0a1eb72015-03-24 16:35:39 -0400764 // If StackPointer() is the system stack pointer (csp), then csp will be
765 // dereferenced to cause the processor (or simulator) to abort if it is not
766 // properly aligned.
Ben Murdochb8a8cc12014-11-26 15:28:44 +0000767 //
768 // If emit_debug_code() is false, this emits no code.
769 void AssertStackConsistency();
770
771 // Preserve the callee-saved registers (as defined by AAPCS64).
772 //
773 // Higher-numbered registers are pushed before lower-numbered registers, and
774 // thus get higher addresses.
775 // Floating-point registers are pushed before general-purpose registers, and
776 // thus get higher addresses.
777 //
778 // Note that registers are not checked for invalid values. Use this method
779 // only if you know that the GC won't try to examine the values on the stack.
780 //
781 // This method must not be called unless the current stack pointer (as set by
782 // SetStackPointer) is the system stack pointer (csp), and is aligned to
783 // ActivationFrameAlignment().
784 void PushCalleeSavedRegisters();
785
786 // Restore the callee-saved registers (as defined by AAPCS64).
787 //
788 // Higher-numbered registers are popped after lower-numbered registers, and
789 // thus come from higher addresses.
790 // Floating-point registers are popped after general-purpose registers, and
791 // thus come from higher addresses.
792 //
793 // This method must not be called unless the current stack pointer (as set by
794 // SetStackPointer) is the system stack pointer (csp), and is aligned to
795 // ActivationFrameAlignment().
796 void PopCalleeSavedRegisters();
797
798 // Set the current stack pointer, but don't generate any code.
799 inline void SetStackPointer(const Register& stack_pointer) {
800 DCHECK(!TmpList()->IncludesAliasOf(stack_pointer));
801 sp_ = stack_pointer;
802 }
803
804 // Return the current stack pointer, as set by SetStackPointer.
805 inline const Register& StackPointer() const {
806 return sp_;
807 }
808
809 // Align csp for a frame, as per ActivationFrameAlignment, and make it the
810 // current stack pointer.
811 inline void AlignAndSetCSPForFrame() {
812 int sp_alignment = ActivationFrameAlignment();
813 // AAPCS64 mandates at least 16-byte alignment.
814 DCHECK(sp_alignment >= 16);
815 DCHECK(base::bits::IsPowerOfTwo32(sp_alignment));
816 Bic(csp, StackPointer(), sp_alignment - 1);
817 SetStackPointer(csp);
818 }
819
820 // Push the system stack pointer (csp) down to allow the same to be done to
821 // the current stack pointer (according to StackPointer()). This must be
822 // called _before_ accessing the memory.
823 //
824 // This is necessary when pushing or otherwise adding things to the stack, to
825 // satisfy the AAPCS64 constraint that the memory below the system stack
826 // pointer is not accessed. The amount pushed will be increased as necessary
827 // to ensure csp remains aligned to 16 bytes.
828 //
829 // This method asserts that StackPointer() is not csp, since the call does
830 // not make sense in that context.
831 inline void BumpSystemStackPointer(const Operand& space);
832
833 // Re-synchronizes the system stack pointer (csp) with the current stack
Emily Bernierd0a1eb72015-03-24 16:35:39 -0400834 // pointer (according to StackPointer()).
Ben Murdochb8a8cc12014-11-26 15:28:44 +0000835 //
836 // This method asserts that StackPointer() is not csp, since the call does
837 // not make sense in that context.
838 inline void SyncSystemStackPointer();
839
840 // Helpers ------------------------------------------------------------------
841 // Root register.
842 inline void InitializeRootRegister();
843
844 void AssertFPCRState(Register fpcr = NoReg);
845 void ConfigureFPCR();
846 void CanonicalizeNaN(const FPRegister& dst, const FPRegister& src);
847 void CanonicalizeNaN(const FPRegister& reg) {
848 CanonicalizeNaN(reg, reg);
849 }
850
851 // Load an object from the root table.
852 void LoadRoot(CPURegister destination,
853 Heap::RootListIndex index);
854 // Store an object to the root table.
855 void StoreRoot(Register source,
856 Heap::RootListIndex index);
857
858 // Load both TrueValue and FalseValue roots.
859 void LoadTrueFalseRoots(Register true_root, Register false_root);
860
861 void LoadHeapObject(Register dst, Handle<HeapObject> object);
862
863 void LoadObject(Register result, Handle<Object> object) {
864 AllowDeferredHandleDereference heap_object_check;
865 if (object->IsHeapObject()) {
866 LoadHeapObject(result, Handle<HeapObject>::cast(object));
867 } else {
868 DCHECK(object->IsSmi());
869 Mov(result, Operand(object));
870 }
871 }
872
873 static int SafepointRegisterStackIndex(int reg_code);
874
875 // This is required for compatibility with architecture independant code.
876 // Remove if not needed.
877 inline void Move(Register dst, Register src) { Mov(dst, src); }
878
879 void LoadInstanceDescriptors(Register map,
880 Register descriptors);
881 void EnumLengthUntagged(Register dst, Register map);
882 void EnumLengthSmi(Register dst, Register map);
883 void NumberOfOwnDescriptors(Register dst, Register map);
884
885 template<typename Field>
886 void DecodeField(Register dst, Register src) {
887 static const uint64_t shift = Field::kShift;
888 static const uint64_t setbits = CountSetBits(Field::kMask, 32);
889 Ubfx(dst, src, shift, setbits);
890 }
891
892 template<typename Field>
893 void DecodeField(Register reg) {
894 DecodeField<Field>(reg, reg);
895 }
896
897 // ---- SMI and Number Utilities ----
898
899 inline void SmiTag(Register dst, Register src);
900 inline void SmiTag(Register smi);
901 inline void SmiUntag(Register dst, Register src);
902 inline void SmiUntag(Register smi);
903 inline void SmiUntagToDouble(FPRegister dst,
904 Register src,
905 UntagMode mode = kNotSpeculativeUntag);
906 inline void SmiUntagToFloat(FPRegister dst,
907 Register src,
908 UntagMode mode = kNotSpeculativeUntag);
909
910 // Tag and push in one step.
911 inline void SmiTagAndPush(Register src);
912 inline void SmiTagAndPush(Register src1, Register src2);
913
914 inline void JumpIfSmi(Register value,
915 Label* smi_label,
916 Label* not_smi_label = NULL);
917 inline void JumpIfNotSmi(Register value, Label* not_smi_label);
918 inline void JumpIfBothSmi(Register value1,
919 Register value2,
920 Label* both_smi_label,
921 Label* not_smi_label = NULL);
922 inline void JumpIfEitherSmi(Register value1,
923 Register value2,
924 Label* either_smi_label,
925 Label* not_smi_label = NULL);
926 inline void JumpIfEitherNotSmi(Register value1,
927 Register value2,
928 Label* not_smi_label);
929 inline void JumpIfBothNotSmi(Register value1,
930 Register value2,
931 Label* not_smi_label);
932
933 // Abort execution if argument is a smi, enabled via --debug-code.
934 void AssertNotSmi(Register object, BailoutReason reason = kOperandIsASmi);
935 void AssertSmi(Register object, BailoutReason reason = kOperandIsNotASmi);
936
937 inline void ObjectTag(Register tagged_obj, Register obj);
938 inline void ObjectUntag(Register untagged_obj, Register obj);
939
940 // Abort execution if argument is not a name, enabled via --debug-code.
941 void AssertName(Register object);
942
943 // Abort execution if argument is not undefined or an AllocationSite, enabled
944 // via --debug-code.
945 void AssertUndefinedOrAllocationSite(Register object, Register scratch);
946
947 // Abort execution if argument is not a string, enabled via --debug-code.
948 void AssertString(Register object);
949
950 void JumpIfHeapNumber(Register object, Label* on_heap_number,
951 SmiCheckType smi_check_type = DONT_DO_SMI_CHECK);
952 void JumpIfNotHeapNumber(Register object, Label* on_not_heap_number,
953 SmiCheckType smi_check_type = DONT_DO_SMI_CHECK);
954
955 // Sets the vs flag if the input is -0.0.
956 void TestForMinusZero(DoubleRegister input);
957
958 // Jump to label if the input double register contains -0.0.
959 void JumpIfMinusZero(DoubleRegister input, Label* on_negative_zero);
960
961 // Jump to label if the input integer register contains the double precision
962 // floating point representation of -0.0.
963 void JumpIfMinusZero(Register input, Label* on_negative_zero);
964
965 // Generate code to do a lookup in the number string cache. If the number in
966 // the register object is found in the cache the generated code falls through
967 // with the result in the result register. The object and the result register
968 // can be the same. If the number is not found in the cache the code jumps to
969 // the label not_found with only the content of register object unchanged.
970 void LookupNumberStringCache(Register object,
971 Register result,
972 Register scratch1,
973 Register scratch2,
974 Register scratch3,
975 Label* not_found);
976
977 // Saturate a signed 32-bit integer in input to an unsigned 8-bit integer in
978 // output.
979 void ClampInt32ToUint8(Register in_out);
980 void ClampInt32ToUint8(Register output, Register input);
981
982 // Saturate a double in input to an unsigned 8-bit integer in output.
983 void ClampDoubleToUint8(Register output,
984 DoubleRegister input,
985 DoubleRegister dbl_scratch);
986
987 // Try to represent a double as a signed 32-bit int.
988 // This succeeds if the result compares equal to the input, so inputs of -0.0
989 // are represented as 0 and handled as a success.
990 //
991 // On output the Z flag is set if the operation was successful.
992 void TryRepresentDoubleAsInt32(Register as_int,
993 FPRegister value,
994 FPRegister scratch_d,
995 Label* on_successful_conversion = NULL,
996 Label* on_failed_conversion = NULL) {
997 DCHECK(as_int.Is32Bits());
998 TryRepresentDoubleAsInt(as_int, value, scratch_d, on_successful_conversion,
999 on_failed_conversion);
1000 }
1001
1002 // Try to represent a double as a signed 64-bit int.
1003 // This succeeds if the result compares equal to the input, so inputs of -0.0
1004 // are represented as 0 and handled as a success.
1005 //
1006 // On output the Z flag is set if the operation was successful.
1007 void TryRepresentDoubleAsInt64(Register as_int,
1008 FPRegister value,
1009 FPRegister scratch_d,
1010 Label* on_successful_conversion = NULL,
1011 Label* on_failed_conversion = NULL) {
1012 DCHECK(as_int.Is64Bits());
1013 TryRepresentDoubleAsInt(as_int, value, scratch_d, on_successful_conversion,
1014 on_failed_conversion);
1015 }
1016
1017 // ---- Object Utilities ----
1018
1019 // Copy fields from 'src' to 'dst', where both are tagged objects.
1020 // The 'temps' list is a list of X registers which can be used for scratch
1021 // values. The temps list must include at least one register.
1022 //
1023 // Currently, CopyFields cannot make use of more than three registers from
1024 // the 'temps' list.
1025 //
1026 // CopyFields expects to be able to take at least two registers from
1027 // MacroAssembler::TmpList().
1028 void CopyFields(Register dst, Register src, CPURegList temps, unsigned count);
1029
1030 // Starting at address in dst, initialize field_count 64-bit fields with
1031 // 64-bit value in register filler. Register dst is corrupted.
1032 void FillFields(Register dst,
1033 Register field_count,
1034 Register filler);
1035
1036 // Copies a number of bytes from src to dst. All passed registers are
1037 // clobbered. On exit src and dst will point to the place just after where the
1038 // last byte was read or written and length will be zero. Hint may be used to
1039 // determine which is the most efficient algorithm to use for copying.
1040 void CopyBytes(Register dst,
1041 Register src,
1042 Register length,
1043 Register scratch,
1044 CopyHint hint = kCopyUnknown);
1045
1046 // ---- String Utilities ----
1047
1048
1049 // Jump to label if either object is not a sequential one-byte string.
1050 // Optionally perform a smi check on the objects first.
1051 void JumpIfEitherIsNotSequentialOneByteStrings(
1052 Register first, Register second, Register scratch1, Register scratch2,
1053 Label* failure, SmiCheckType smi_check = DO_SMI_CHECK);
1054
1055 // Check if instance type is sequential one-byte string and jump to label if
1056 // it is not.
1057 void JumpIfInstanceTypeIsNotSequentialOneByte(Register type, Register scratch,
1058 Label* failure);
1059
1060 // Checks if both instance types are sequential one-byte strings and jumps to
1061 // label if either is not.
1062 void JumpIfEitherInstanceTypeIsNotSequentialOneByte(
1063 Register first_object_instance_type, Register second_object_instance_type,
1064 Register scratch1, Register scratch2, Label* failure);
1065
1066 // Checks if both instance types are sequential one-byte strings and jumps to
1067 // label if either is not.
1068 void JumpIfBothInstanceTypesAreNotSequentialOneByte(
1069 Register first_object_instance_type, Register second_object_instance_type,
1070 Register scratch1, Register scratch2, Label* failure);
1071
1072 void JumpIfNotUniqueNameInstanceType(Register type, Label* not_unique_name);
1073
1074 // ---- Calling / Jumping helpers ----
1075
1076 // This is required for compatibility in architecture indepenedant code.
1077 inline void jmp(Label* L) { B(L); }
1078
1079 // Passes thrown value to the handler of top of the try handler chain.
1080 // Register value must be x0.
1081 void Throw(Register value,
1082 Register scratch1,
1083 Register scratch2,
1084 Register scratch3,
1085 Register scratch4);
1086
1087 // Propagates an uncatchable exception to the top of the current JS stack's
1088 // handler chain. Register value must be x0.
1089 void ThrowUncatchable(Register value,
1090 Register scratch1,
1091 Register scratch2,
1092 Register scratch3,
1093 Register scratch4);
1094
1095 void CallStub(CodeStub* stub, TypeFeedbackId ast_id = TypeFeedbackId::None());
1096 void TailCallStub(CodeStub* stub);
1097
1098 void CallRuntime(const Runtime::Function* f,
1099 int num_arguments,
1100 SaveFPRegsMode save_doubles = kDontSaveFPRegs);
1101
1102 void CallRuntime(Runtime::FunctionId id,
1103 int num_arguments,
1104 SaveFPRegsMode save_doubles = kDontSaveFPRegs) {
1105 CallRuntime(Runtime::FunctionForId(id), num_arguments, save_doubles);
1106 }
1107
1108 void CallRuntimeSaveDoubles(Runtime::FunctionId id) {
1109 const Runtime::Function* function = Runtime::FunctionForId(id);
1110 CallRuntime(function, function->nargs, kSaveFPRegs);
1111 }
1112
1113 void TailCallRuntime(Runtime::FunctionId fid,
1114 int num_arguments,
1115 int result_size);
1116
1117 int ActivationFrameAlignment();
1118
1119 // Calls a C function.
1120 // The called function is not allowed to trigger a
1121 // garbage collection, since that might move the code and invalidate the
1122 // return address (unless this is somehow accounted for by the called
1123 // function).
1124 void CallCFunction(ExternalReference function,
1125 int num_reg_arguments);
1126 void CallCFunction(ExternalReference function,
1127 int num_reg_arguments,
1128 int num_double_arguments);
1129 void CallCFunction(Register function,
1130 int num_reg_arguments,
1131 int num_double_arguments);
1132
1133 // Calls an API function. Allocates HandleScope, extracts returned value
1134 // from handle and propagates exceptions.
1135 // 'stack_space' is the space to be unwound on exit (includes the call JS
1136 // arguments space and the additional space allocated for the fast call).
1137 // 'spill_offset' is the offset from the stack pointer where
1138 // CallApiFunctionAndReturn can spill registers.
1139 void CallApiFunctionAndReturn(Register function_address,
1140 ExternalReference thunk_ref,
1141 int stack_space,
1142 int spill_offset,
1143 MemOperand return_value_operand,
1144 MemOperand* context_restore_operand);
1145
1146 // The number of register that CallApiFunctionAndReturn will need to save on
1147 // the stack. The space for these registers need to be allocated in the
1148 // ExitFrame before calling CallApiFunctionAndReturn.
1149 static const int kCallApiFunctionSpillSpace = 4;
1150
1151 // Jump to a runtime routine.
1152 void JumpToExternalReference(const ExternalReference& builtin);
1153 // Tail call of a runtime routine (jump).
1154 // Like JumpToExternalReference, but also takes care of passing the number
1155 // of parameters.
1156 void TailCallExternalReference(const ExternalReference& ext,
1157 int num_arguments,
1158 int result_size);
1159 void CallExternalReference(const ExternalReference& ext,
1160 int num_arguments);
1161
1162
1163 // Invoke specified builtin JavaScript function. Adds an entry to
1164 // the unresolved list if the name does not resolve.
1165 void InvokeBuiltin(Builtins::JavaScript id,
1166 InvokeFlag flag,
1167 const CallWrapper& call_wrapper = NullCallWrapper());
1168
1169 // Store the code object for the given builtin in the target register and
1170 // setup the function in the function register.
1171 void GetBuiltinEntry(Register target,
1172 Register function,
1173 Builtins::JavaScript id);
1174
1175 // Store the function for the given builtin in the target register.
1176 void GetBuiltinFunction(Register target, Builtins::JavaScript id);
1177
1178 void Jump(Register target);
1179 void Jump(Address target, RelocInfo::Mode rmode);
1180 void Jump(Handle<Code> code, RelocInfo::Mode rmode);
1181 void Jump(intptr_t target, RelocInfo::Mode rmode);
1182
1183 void Call(Register target);
1184 void Call(Label* target);
1185 void Call(Address target, RelocInfo::Mode rmode);
1186 void Call(Handle<Code> code,
1187 RelocInfo::Mode rmode = RelocInfo::CODE_TARGET,
1188 TypeFeedbackId ast_id = TypeFeedbackId::None());
1189
1190 // For every Call variant, there is a matching CallSize function that returns
1191 // the size (in bytes) of the call sequence.
1192 static int CallSize(Register target);
1193 static int CallSize(Label* target);
1194 static int CallSize(Address target, RelocInfo::Mode rmode);
1195 static int CallSize(Handle<Code> code,
1196 RelocInfo::Mode rmode = RelocInfo::CODE_TARGET,
1197 TypeFeedbackId ast_id = TypeFeedbackId::None());
1198
1199 // Registers used through the invocation chain are hard-coded.
1200 // We force passing the parameters to ensure the contracts are correctly
1201 // honoured by the caller.
1202 // 'function' must be x1.
1203 // 'actual' must use an immediate or x0.
1204 // 'expected' must use an immediate or x2.
1205 // 'call_kind' must be x5.
1206 void InvokePrologue(const ParameterCount& expected,
1207 const ParameterCount& actual,
1208 Handle<Code> code_constant,
1209 Register code_reg,
1210 Label* done,
1211 InvokeFlag flag,
1212 bool* definitely_mismatches,
1213 const CallWrapper& call_wrapper);
1214 void InvokeCode(Register code,
1215 const ParameterCount& expected,
1216 const ParameterCount& actual,
1217 InvokeFlag flag,
1218 const CallWrapper& call_wrapper);
1219 // Invoke the JavaScript function in the given register.
1220 // Changes the current context to the context in the function before invoking.
1221 void InvokeFunction(Register function,
1222 const ParameterCount& actual,
1223 InvokeFlag flag,
1224 const CallWrapper& call_wrapper);
1225 void InvokeFunction(Register function,
1226 const ParameterCount& expected,
1227 const ParameterCount& actual,
1228 InvokeFlag flag,
1229 const CallWrapper& call_wrapper);
1230 void InvokeFunction(Handle<JSFunction> function,
1231 const ParameterCount& expected,
1232 const ParameterCount& actual,
1233 InvokeFlag flag,
1234 const CallWrapper& call_wrapper);
1235
1236
1237 // ---- Floating point helpers ----
1238
1239 // Perform a conversion from a double to a signed int64. If the input fits in
1240 // range of the 64-bit result, execution branches to done. Otherwise,
1241 // execution falls through, and the sign of the result can be used to
1242 // determine if overflow was towards positive or negative infinity.
1243 //
1244 // On successful conversion, the least significant 32 bits of the result are
1245 // equivalent to the ECMA-262 operation "ToInt32".
1246 //
1247 // Only public for the test code in test-code-stubs-arm64.cc.
1248 void TryConvertDoubleToInt64(Register result,
1249 DoubleRegister input,
1250 Label* done);
1251
1252 // Performs a truncating conversion of a floating point number as used by
1253 // the JS bitwise operations. See ECMA-262 9.5: ToInt32.
1254 // Exits with 'result' holding the answer.
1255 void TruncateDoubleToI(Register result, DoubleRegister double_input);
1256
1257 // Performs a truncating conversion of a heap number as used by
1258 // the JS bitwise operations. See ECMA-262 9.5: ToInt32. 'result' and 'input'
1259 // must be different registers. Exits with 'result' holding the answer.
1260 void TruncateHeapNumberToI(Register result, Register object);
1261
1262 // Converts the smi or heap number in object to an int32 using the rules
1263 // for ToInt32 as described in ECMAScript 9.5.: the value is truncated
1264 // and brought into the range -2^31 .. +2^31 - 1. 'result' and 'input' must be
1265 // different registers.
1266 void TruncateNumberToI(Register object,
1267 Register result,
1268 Register heap_number_map,
1269 Label* not_int32);
1270
1271 // ---- Code generation helpers ----
1272
1273 void set_generating_stub(bool value) { generating_stub_ = value; }
1274 bool generating_stub() const { return generating_stub_; }
1275#if DEBUG
1276 void set_allow_macro_instructions(bool value) {
1277 allow_macro_instructions_ = value;
1278 }
1279 bool allow_macro_instructions() const { return allow_macro_instructions_; }
1280#endif
1281 bool use_real_aborts() const { return use_real_aborts_; }
1282 void set_has_frame(bool value) { has_frame_ = value; }
1283 bool has_frame() const { return has_frame_; }
1284 bool AllowThisStubCall(CodeStub* stub);
1285
1286 class NoUseRealAbortsScope {
1287 public:
1288 explicit NoUseRealAbortsScope(MacroAssembler* masm) :
1289 saved_(masm->use_real_aborts_), masm_(masm) {
1290 masm_->use_real_aborts_ = false;
1291 }
1292 ~NoUseRealAbortsScope() {
1293 masm_->use_real_aborts_ = saved_;
1294 }
1295 private:
1296 bool saved_;
1297 MacroAssembler* masm_;
1298 };
1299
1300 // ---------------------------------------------------------------------------
1301 // Debugger Support
1302
1303 void DebugBreak();
1304
1305 // ---------------------------------------------------------------------------
1306 // Exception handling
1307
1308 // Push a new try handler and link into try handler chain.
1309 void PushTryHandler(StackHandler::Kind kind, int handler_index);
1310
1311 // Unlink the stack handler on top of the stack from the try handler chain.
1312 // Must preserve the result register.
1313 void PopTryHandler();
1314
1315
1316 // ---------------------------------------------------------------------------
1317 // Allocation support
1318
1319 // Allocate an object in new space or old pointer space. The object_size is
1320 // specified either in bytes or in words if the allocation flag SIZE_IN_WORDS
1321 // is passed. The allocated object is returned in result.
1322 //
1323 // If the new space is exhausted control continues at the gc_required label.
1324 // In this case, the result and scratch registers may still be clobbered.
1325 // If flags includes TAG_OBJECT, the result is tagged as as a heap object.
1326 void Allocate(Register object_size,
1327 Register result,
1328 Register scratch1,
1329 Register scratch2,
1330 Label* gc_required,
1331 AllocationFlags flags);
1332
1333 void Allocate(int object_size,
1334 Register result,
1335 Register scratch1,
1336 Register scratch2,
1337 Label* gc_required,
1338 AllocationFlags flags);
1339
1340 // Undo allocation in new space. The object passed and objects allocated after
1341 // it will no longer be allocated. The caller must make sure that no pointers
1342 // are left to the object(s) no longer allocated as they would be invalid when
1343 // allocation is undone.
1344 void UndoAllocationInNewSpace(Register object, Register scratch);
1345
1346 void AllocateTwoByteString(Register result,
1347 Register length,
1348 Register scratch1,
1349 Register scratch2,
1350 Register scratch3,
1351 Label* gc_required);
1352 void AllocateOneByteString(Register result, Register length,
1353 Register scratch1, Register scratch2,
1354 Register scratch3, Label* gc_required);
1355 void AllocateTwoByteConsString(Register result,
1356 Register length,
1357 Register scratch1,
1358 Register scratch2,
1359 Label* gc_required);
1360 void AllocateOneByteConsString(Register result, Register length,
1361 Register scratch1, Register scratch2,
1362 Label* gc_required);
1363 void AllocateTwoByteSlicedString(Register result,
1364 Register length,
1365 Register scratch1,
1366 Register scratch2,
1367 Label* gc_required);
1368 void AllocateOneByteSlicedString(Register result, Register length,
1369 Register scratch1, Register scratch2,
1370 Label* gc_required);
1371
1372 // Allocates a heap number or jumps to the gc_required label if the young
1373 // space is full and a scavenge is needed.
1374 // All registers are clobbered.
1375 // If no heap_number_map register is provided, the function will take care of
1376 // loading it.
1377 void AllocateHeapNumber(Register result,
1378 Label* gc_required,
1379 Register scratch1,
1380 Register scratch2,
1381 CPURegister value = NoFPReg,
1382 CPURegister heap_number_map = NoReg,
1383 MutableMode mode = IMMUTABLE);
1384
1385 // ---------------------------------------------------------------------------
1386 // Support functions.
1387
1388 // Try to get function prototype of a function and puts the value in the
1389 // result register. Checks that the function really is a function and jumps
1390 // to the miss label if the fast checks fail. The function register will be
1391 // untouched; the other registers may be clobbered.
1392 enum BoundFunctionAction {
1393 kMissOnBoundFunction,
1394 kDontMissOnBoundFunction
1395 };
1396
1397 void TryGetFunctionPrototype(Register function,
1398 Register result,
1399 Register scratch,
1400 Label* miss,
1401 BoundFunctionAction action =
1402 kDontMissOnBoundFunction);
1403
1404 // Compare object type for heap object. heap_object contains a non-Smi
1405 // whose object type should be compared with the given type. This both
1406 // sets the flags and leaves the object type in the type_reg register.
1407 // It leaves the map in the map register (unless the type_reg and map register
1408 // are the same register). It leaves the heap object in the heap_object
1409 // register unless the heap_object register is the same register as one of the
1410 // other registers.
1411 void CompareObjectType(Register heap_object,
1412 Register map,
1413 Register type_reg,
1414 InstanceType type);
1415
1416
1417 // Compare object type for heap object, and branch if equal (or not.)
1418 // heap_object contains a non-Smi whose object type should be compared with
1419 // the given type. This both sets the flags and leaves the object type in
1420 // the type_reg register. It leaves the map in the map register (unless the
1421 // type_reg and map register are the same register). It leaves the heap
1422 // object in the heap_object register unless the heap_object register is the
1423 // same register as one of the other registers.
1424 void JumpIfObjectType(Register object,
1425 Register map,
1426 Register type_reg,
1427 InstanceType type,
1428 Label* if_cond_pass,
1429 Condition cond = eq);
1430
1431 void JumpIfNotObjectType(Register object,
1432 Register map,
1433 Register type_reg,
1434 InstanceType type,
1435 Label* if_not_object);
1436
1437 // Compare instance type in a map. map contains a valid map object whose
1438 // object type should be compared with the given type. This both
1439 // sets the flags and leaves the object type in the type_reg register.
1440 void CompareInstanceType(Register map,
1441 Register type_reg,
1442 InstanceType type);
1443
1444 // Compare an object's map with the specified map. Condition flags are set
1445 // with result of map compare.
1446 void CompareObjectMap(Register obj, Heap::RootListIndex index);
1447
1448 // Compare an object's map with the specified map. Condition flags are set
1449 // with result of map compare.
1450 void CompareObjectMap(Register obj, Register scratch, Handle<Map> map);
1451
1452 // As above, but the map of the object is already loaded into the register
1453 // which is preserved by the code generated.
1454 void CompareMap(Register obj_map,
1455 Handle<Map> map);
1456
1457 // Check if the map of an object is equal to a specified map and branch to
1458 // label if not. Skip the smi check if not required (object is known to be a
1459 // heap object). If mode is ALLOW_ELEMENT_TRANSITION_MAPS, then also match
1460 // against maps that are ElementsKind transition maps of the specified map.
1461 void CheckMap(Register obj,
1462 Register scratch,
1463 Handle<Map> map,
1464 Label* fail,
1465 SmiCheckType smi_check_type);
1466
1467
1468 void CheckMap(Register obj,
1469 Register scratch,
1470 Heap::RootListIndex index,
1471 Label* fail,
1472 SmiCheckType smi_check_type);
1473
1474 // As above, but the map of the object is already loaded into obj_map, and is
1475 // preserved.
1476 void CheckMap(Register obj_map,
1477 Handle<Map> map,
1478 Label* fail,
1479 SmiCheckType smi_check_type);
1480
Emily Bernierd0a1eb72015-03-24 16:35:39 -04001481 // Check if the map of an object is equal to a specified weak map and branch
1482 // to a specified target if equal. Skip the smi check if not required
1483 // (object is known to be a heap object)
1484 void DispatchWeakMap(Register obj, Register scratch1, Register scratch2,
1485 Handle<WeakCell> cell, Handle<Code> success,
1486 SmiCheckType smi_check_type);
1487
1488 // Compare the given value and the value of weak cell.
1489 void CmpWeakValue(Register value, Handle<WeakCell> cell, Register scratch);
1490
1491 // Load the value of the weak cell in the value register. Branch to the given
1492 // miss label if the weak cell was cleared.
1493 void LoadWeakValue(Register value, Handle<WeakCell> cell, Label* miss);
Ben Murdochb8a8cc12014-11-26 15:28:44 +00001494
1495 // Test the bitfield of the heap object map with mask and set the condition
1496 // flags. The object register is preserved.
1497 void TestMapBitfield(Register object, uint64_t mask);
1498
1499 // Load the elements kind field from a map, and return it in the result
1500 // register.
1501 void LoadElementsKindFromMap(Register result, Register map);
1502
1503 // Compare the object in a register to a value from the root list.
1504 void CompareRoot(const Register& obj, Heap::RootListIndex index);
1505
1506 // Compare the object in a register to a value and jump if they are equal.
1507 void JumpIfRoot(const Register& obj,
1508 Heap::RootListIndex index,
1509 Label* if_equal);
1510
1511 // Compare the object in a register to a value and jump if they are not equal.
1512 void JumpIfNotRoot(const Register& obj,
1513 Heap::RootListIndex index,
1514 Label* if_not_equal);
1515
1516 // Load and check the instance type of an object for being a unique name.
1517 // Loads the type into the second argument register.
1518 // The object and type arguments can be the same register; in that case it
1519 // will be overwritten with the type.
1520 // Fall-through if the object was a string and jump on fail otherwise.
1521 inline void IsObjectNameType(Register object, Register type, Label* fail);
1522
1523 inline void IsObjectJSObjectType(Register heap_object,
1524 Register map,
1525 Register scratch,
1526 Label* fail);
1527
1528 // Check the instance type in the given map to see if it corresponds to a
1529 // JS object type. Jump to the fail label if this is not the case and fall
1530 // through otherwise. However if fail label is NULL, no branch will be
1531 // performed and the flag will be updated. You can test the flag for "le"
1532 // condition to test if it is a valid JS object type.
1533 inline void IsInstanceJSObjectType(Register map,
1534 Register scratch,
1535 Label* fail);
1536
1537 // Load and check the instance type of an object for being a string.
1538 // Loads the type into the second argument register.
1539 // The object and type arguments can be the same register; in that case it
1540 // will be overwritten with the type.
1541 // Jumps to not_string or string appropriate. If the appropriate label is
1542 // NULL, fall through.
1543 inline void IsObjectJSStringType(Register object, Register type,
1544 Label* not_string, Label* string = NULL);
1545
1546 // Compare the contents of a register with an operand, and branch to true,
1547 // false or fall through, depending on condition.
1548 void CompareAndSplit(const Register& lhs,
1549 const Operand& rhs,
1550 Condition cond,
1551 Label* if_true,
1552 Label* if_false,
1553 Label* fall_through);
1554
1555 // Test the bits of register defined by bit_pattern, and branch to
1556 // if_any_set, if_all_clear or fall_through accordingly.
1557 void TestAndSplit(const Register& reg,
1558 uint64_t bit_pattern,
1559 Label* if_all_clear,
1560 Label* if_any_set,
1561 Label* fall_through);
1562
1563 // Check if a map for a JSObject indicates that the object has fast elements.
1564 // Jump to the specified label if it does not.
1565 void CheckFastElements(Register map, Register scratch, Label* fail);
1566
1567 // Check if a map for a JSObject indicates that the object can have both smi
1568 // and HeapObject elements. Jump to the specified label if it does not.
1569 void CheckFastObjectElements(Register map, Register scratch, Label* fail);
1570
1571 // Check to see if number can be stored as a double in FastDoubleElements.
1572 // If it can, store it at the index specified by key_reg in the array,
1573 // otherwise jump to fail.
1574 void StoreNumberToDoubleElements(Register value_reg,
1575 Register key_reg,
1576 Register elements_reg,
1577 Register scratch1,
1578 FPRegister fpscratch1,
1579 Label* fail,
1580 int elements_offset = 0);
1581
1582 // Picks out an array index from the hash field.
1583 // Register use:
1584 // hash - holds the index's hash. Clobbered.
1585 // index - holds the overwritten index on exit.
1586 void IndexFromHash(Register hash, Register index);
1587
1588 // ---------------------------------------------------------------------------
1589 // Inline caching support.
1590
1591 void EmitSeqStringSetCharCheck(Register string,
1592 Register index,
1593 SeqStringSetCharCheckIndexType index_type,
1594 Register scratch,
1595 uint32_t encoding_mask);
1596
1597 // Generate code for checking access rights - used for security checks
1598 // on access to global objects across environments. The holder register
1599 // is left untouched, whereas both scratch registers are clobbered.
1600 void CheckAccessGlobalProxy(Register holder_reg,
1601 Register scratch1,
1602 Register scratch2,
1603 Label* miss);
1604
1605 // Hash the interger value in 'key' register.
1606 // It uses the same algorithm as ComputeIntegerHash in utils.h.
1607 void GetNumberHash(Register key, Register scratch);
1608
1609 // Load value from the dictionary.
1610 //
1611 // elements - holds the slow-case elements of the receiver on entry.
1612 // Unchanged unless 'result' is the same register.
1613 //
1614 // key - holds the smi key on entry.
1615 // Unchanged unless 'result' is the same register.
1616 //
1617 // result - holds the result on exit if the load succeeded.
1618 // Allowed to be the same as 'key' or 'result'.
1619 // Unchanged on bailout so 'key' or 'result' can be used
1620 // in further computation.
1621 void LoadFromNumberDictionary(Label* miss,
1622 Register elements,
1623 Register key,
1624 Register result,
1625 Register scratch0,
1626 Register scratch1,
1627 Register scratch2,
1628 Register scratch3);
1629
1630 // ---------------------------------------------------------------------------
1631 // Frames.
1632
1633 // Activation support.
1634 void EnterFrame(StackFrame::Type type);
Emily Bernierd0a1eb72015-03-24 16:35:39 -04001635 void EnterFrame(StackFrame::Type type, bool load_constant_pool_pointer_reg);
Ben Murdochb8a8cc12014-11-26 15:28:44 +00001636 void LeaveFrame(StackFrame::Type type);
1637
1638 // Returns map with validated enum cache in object register.
1639 void CheckEnumCache(Register object,
1640 Register null_value,
1641 Register scratch0,
1642 Register scratch1,
1643 Register scratch2,
1644 Register scratch3,
1645 Label* call_runtime);
1646
1647 // AllocationMemento support. Arrays may have an associated
1648 // AllocationMemento object that can be checked for in order to pretransition
1649 // to another type.
1650 // On entry, receiver should point to the array object.
1651 // If allocation info is present, the Z flag is set (so that the eq
1652 // condition will pass).
1653 void TestJSArrayForAllocationMemento(Register receiver,
1654 Register scratch1,
1655 Register scratch2,
1656 Label* no_memento_found);
1657
1658 void JumpIfJSArrayHasAllocationMemento(Register receiver,
1659 Register scratch1,
1660 Register scratch2,
1661 Label* memento_found) {
1662 Label no_memento_found;
1663 TestJSArrayForAllocationMemento(receiver, scratch1, scratch2,
1664 &no_memento_found);
1665 B(eq, memento_found);
1666 Bind(&no_memento_found);
1667 }
1668
1669 // The stack pointer has to switch between csp and jssp when setting up and
1670 // destroying the exit frame. Hence preserving/restoring the registers is
1671 // slightly more complicated than simple push/pop operations.
1672 void ExitFramePreserveFPRegs();
1673 void ExitFrameRestoreFPRegs();
1674
1675 // Generates function and stub prologue code.
1676 void StubPrologue();
1677 void Prologue(bool code_pre_aging);
1678
1679 // Enter exit frame. Exit frames are used when calling C code from generated
1680 // (JavaScript) code.
1681 //
1682 // The stack pointer must be jssp on entry, and will be set to csp by this
1683 // function. The frame pointer is also configured, but the only other
1684 // registers modified by this function are the provided scratch register, and
1685 // jssp.
1686 //
1687 // The 'extra_space' argument can be used to allocate some space in the exit
1688 // frame that will be ignored by the GC. This space will be reserved in the
1689 // bottom of the frame immediately above the return address slot.
1690 //
1691 // Set up a stack frame and registers as follows:
1692 // fp[8]: CallerPC (lr)
1693 // fp -> fp[0]: CallerFP (old fp)
1694 // fp[-8]: SPOffset (new csp)
1695 // fp[-16]: CodeObject()
1696 // fp[-16 - fp-size]: Saved doubles, if saved_doubles is true.
1697 // csp[8]: Memory reserved for the caller if extra_space != 0.
1698 // Alignment padding, if necessary.
1699 // csp -> csp[0]: Space reserved for the return address.
1700 //
1701 // This function also stores the new frame information in the top frame, so
1702 // that the new frame becomes the current frame.
1703 void EnterExitFrame(bool save_doubles,
1704 const Register& scratch,
1705 int extra_space = 0);
1706
1707 // Leave the current exit frame, after a C function has returned to generated
1708 // (JavaScript) code.
1709 //
1710 // This effectively unwinds the operation of EnterExitFrame:
1711 // * Preserved doubles are restored (if restore_doubles is true).
1712 // * The frame information is removed from the top frame.
1713 // * The exit frame is dropped.
1714 // * The stack pointer is reset to jssp.
1715 //
1716 // The stack pointer must be csp on entry.
1717 void LeaveExitFrame(bool save_doubles,
1718 const Register& scratch,
1719 bool restore_context);
1720
1721 void LoadContext(Register dst, int context_chain_length);
1722
1723 // Emit code for a truncating division by a constant. The dividend register is
1724 // unchanged. Dividend and result must be different.
1725 void TruncatingDiv(Register result, Register dividend, int32_t divisor);
1726
1727 // ---------------------------------------------------------------------------
1728 // StatsCounter support
1729
1730 void SetCounter(StatsCounter* counter, int value, Register scratch1,
1731 Register scratch2);
1732 void IncrementCounter(StatsCounter* counter, int value, Register scratch1,
1733 Register scratch2);
1734 void DecrementCounter(StatsCounter* counter, int value, Register scratch1,
1735 Register scratch2);
1736
1737 // ---------------------------------------------------------------------------
1738 // Garbage collector support (GC).
1739
1740 enum RememberedSetFinalAction {
1741 kReturnAtEnd,
1742 kFallThroughAtEnd
1743 };
1744
1745 // Record in the remembered set the fact that we have a pointer to new space
1746 // at the address pointed to by the addr register. Only works if addr is not
1747 // in new space.
1748 void RememberedSetHelper(Register object, // Used for debug code.
1749 Register addr,
1750 Register scratch1,
1751 SaveFPRegsMode save_fp,
1752 RememberedSetFinalAction and_then);
1753
1754 // Push and pop the registers that can hold pointers, as defined by the
1755 // RegList constant kSafepointSavedRegisters.
1756 void PushSafepointRegisters();
1757 void PopSafepointRegisters();
1758
1759 void PushSafepointRegistersAndDoubles();
1760 void PopSafepointRegistersAndDoubles();
1761
1762 // Store value in register src in the safepoint stack slot for register dst.
1763 void StoreToSafepointRegisterSlot(Register src, Register dst) {
1764 Poke(src, SafepointRegisterStackIndex(dst.code()) * kPointerSize);
1765 }
1766
1767 // Load the value of the src register from its safepoint stack slot
1768 // into register dst.
1769 void LoadFromSafepointRegisterSlot(Register dst, Register src) {
1770 Peek(src, SafepointRegisterStackIndex(dst.code()) * kPointerSize);
1771 }
1772
1773 void CheckPageFlagSet(const Register& object,
1774 const Register& scratch,
1775 int mask,
1776 Label* if_any_set);
1777
1778 void CheckPageFlagClear(const Register& object,
1779 const Register& scratch,
1780 int mask,
1781 Label* if_all_clear);
1782
Ben Murdochb8a8cc12014-11-26 15:28:44 +00001783 // Check if object is in new space and jump accordingly.
1784 // Register 'object' is preserved.
1785 void JumpIfNotInNewSpace(Register object,
1786 Label* branch) {
1787 InNewSpace(object, ne, branch);
1788 }
1789
1790 void JumpIfInNewSpace(Register object,
1791 Label* branch) {
1792 InNewSpace(object, eq, branch);
1793 }
1794
1795 // Notify the garbage collector that we wrote a pointer into an object.
1796 // |object| is the object being stored into, |value| is the object being
1797 // stored. value and scratch registers are clobbered by the operation.
1798 // The offset is the offset from the start of the object, not the offset from
1799 // the tagged HeapObject pointer. For use with FieldOperand(reg, off).
1800 void RecordWriteField(
1801 Register object,
1802 int offset,
1803 Register value,
1804 Register scratch,
1805 LinkRegisterStatus lr_status,
1806 SaveFPRegsMode save_fp,
1807 RememberedSetAction remembered_set_action = EMIT_REMEMBERED_SET,
1808 SmiCheck smi_check = INLINE_SMI_CHECK,
1809 PointersToHereCheck pointers_to_here_check_for_value =
1810 kPointersToHereMaybeInteresting);
1811
1812 // As above, but the offset has the tag presubtracted. For use with
1813 // MemOperand(reg, off).
1814 inline void RecordWriteContextSlot(
1815 Register context,
1816 int offset,
1817 Register value,
1818 Register scratch,
1819 LinkRegisterStatus lr_status,
1820 SaveFPRegsMode save_fp,
1821 RememberedSetAction remembered_set_action = EMIT_REMEMBERED_SET,
1822 SmiCheck smi_check = INLINE_SMI_CHECK,
1823 PointersToHereCheck pointers_to_here_check_for_value =
1824 kPointersToHereMaybeInteresting) {
1825 RecordWriteField(context,
1826 offset + kHeapObjectTag,
1827 value,
1828 scratch,
1829 lr_status,
1830 save_fp,
1831 remembered_set_action,
1832 smi_check,
1833 pointers_to_here_check_for_value);
1834 }
1835
1836 void RecordWriteForMap(
1837 Register object,
1838 Register map,
1839 Register dst,
1840 LinkRegisterStatus lr_status,
1841 SaveFPRegsMode save_fp);
1842
1843 // For a given |object| notify the garbage collector that the slot |address|
1844 // has been written. |value| is the object being stored. The value and
1845 // address registers are clobbered by the operation.
1846 void RecordWrite(
1847 Register object,
1848 Register address,
1849 Register value,
1850 LinkRegisterStatus lr_status,
1851 SaveFPRegsMode save_fp,
1852 RememberedSetAction remembered_set_action = EMIT_REMEMBERED_SET,
1853 SmiCheck smi_check = INLINE_SMI_CHECK,
1854 PointersToHereCheck pointers_to_here_check_for_value =
1855 kPointersToHereMaybeInteresting);
1856
1857 // Checks the color of an object. If the object is already grey or black
1858 // then we just fall through, since it is already live. If it is white and
1859 // we can determine that it doesn't need to be scanned, then we just mark it
1860 // black and fall through. For the rest we jump to the label so the
1861 // incremental marker can fix its assumptions.
1862 void EnsureNotWhite(Register object,
1863 Register scratch1,
1864 Register scratch2,
1865 Register scratch3,
1866 Register scratch4,
1867 Label* object_is_white_and_not_data);
1868
1869 // Detects conservatively whether an object is data-only, i.e. it does need to
1870 // be scanned by the garbage collector.
1871 void JumpIfDataObject(Register value,
1872 Register scratch,
1873 Label* not_data_object);
1874
1875 // Helper for finding the mark bits for an address.
1876 // Note that the behaviour slightly differs from other architectures.
1877 // On exit:
1878 // - addr_reg is unchanged.
1879 // - The bitmap register points at the word with the mark bits.
1880 // - The shift register contains the index of the first color bit for this
1881 // object in the bitmap.
1882 inline void GetMarkBits(Register addr_reg,
1883 Register bitmap_reg,
1884 Register shift_reg);
1885
1886 // Check if an object has a given incremental marking color.
1887 void HasColor(Register object,
1888 Register scratch0,
1889 Register scratch1,
1890 Label* has_color,
1891 int first_bit,
1892 int second_bit);
1893
1894 void JumpIfBlack(Register object,
1895 Register scratch0,
1896 Register scratch1,
1897 Label* on_black);
1898
1899
1900 // Get the location of a relocated constant (its address in the constant pool)
1901 // from its load site.
1902 void GetRelocatedValueLocation(Register ldr_location,
1903 Register result);
1904
1905
1906 // ---------------------------------------------------------------------------
1907 // Debugging.
1908
1909 // Calls Abort(msg) if the condition cond is not satisfied.
1910 // Use --debug_code to enable.
1911 void Assert(Condition cond, BailoutReason reason);
1912 void AssertRegisterIsClear(Register reg, BailoutReason reason);
1913 void AssertRegisterIsRoot(
1914 Register reg,
1915 Heap::RootListIndex index,
1916 BailoutReason reason = kRegisterDidNotMatchExpectedRoot);
1917 void AssertFastElements(Register elements);
1918
1919 // Abort if the specified register contains the invalid color bit pattern.
1920 // The pattern must be in bits [1:0] of 'reg' register.
1921 //
1922 // If emit_debug_code() is false, this emits no code.
1923 void AssertHasValidColor(const Register& reg);
1924
1925 // Abort if 'object' register doesn't point to a string object.
1926 //
1927 // If emit_debug_code() is false, this emits no code.
1928 void AssertIsString(const Register& object);
1929
1930 // Like Assert(), but always enabled.
1931 void Check(Condition cond, BailoutReason reason);
1932 void CheckRegisterIsClear(Register reg, BailoutReason reason);
1933
1934 // Print a message to stderr and abort execution.
1935 void Abort(BailoutReason reason);
1936
1937 // Conditionally load the cached Array transitioned map of type
1938 // transitioned_kind from the native context if the map in register
1939 // map_in_out is the cached Array map in the native context of
1940 // expected_kind.
1941 void LoadTransitionedArrayMapConditional(
1942 ElementsKind expected_kind,
1943 ElementsKind transitioned_kind,
1944 Register map_in_out,
1945 Register scratch1,
1946 Register scratch2,
1947 Label* no_map_match);
1948
1949 void LoadGlobalFunction(int index, Register function);
1950
1951 // Load the initial map from the global function. The registers function and
1952 // map can be the same, function is then overwritten.
1953 void LoadGlobalFunctionInitialMap(Register function,
1954 Register map,
1955 Register scratch);
1956
1957 CPURegList* TmpList() { return &tmp_list_; }
1958 CPURegList* FPTmpList() { return &fptmp_list_; }
1959
1960 static CPURegList DefaultTmpList();
1961 static CPURegList DefaultFPTmpList();
1962
1963 // Like printf, but print at run-time from generated code.
1964 //
1965 // The caller must ensure that arguments for floating-point placeholders
1966 // (such as %e, %f or %g) are FPRegisters, and that arguments for integer
1967 // placeholders are Registers.
1968 //
1969 // At the moment it is only possible to print the value of csp if it is the
1970 // current stack pointer. Otherwise, the MacroAssembler will automatically
1971 // update csp on every push (using BumpSystemStackPointer), so determining its
1972 // value is difficult.
1973 //
1974 // Format placeholders that refer to more than one argument, or to a specific
1975 // argument, are not supported. This includes formats like "%1$d" or "%.*d".
1976 //
1977 // This function automatically preserves caller-saved registers so that
1978 // calling code can use Printf at any point without having to worry about
1979 // corruption. The preservation mechanism generates a lot of code. If this is
1980 // a problem, preserve the important registers manually and then call
1981 // PrintfNoPreserve. Callee-saved registers are not used by Printf, and are
1982 // implicitly preserved.
1983 void Printf(const char * format,
1984 CPURegister arg0 = NoCPUReg,
1985 CPURegister arg1 = NoCPUReg,
1986 CPURegister arg2 = NoCPUReg,
1987 CPURegister arg3 = NoCPUReg);
1988
1989 // Like Printf, but don't preserve any caller-saved registers, not even 'lr'.
1990 //
1991 // The return code from the system printf call will be returned in x0.
1992 void PrintfNoPreserve(const char * format,
1993 const CPURegister& arg0 = NoCPUReg,
1994 const CPURegister& arg1 = NoCPUReg,
1995 const CPURegister& arg2 = NoCPUReg,
1996 const CPURegister& arg3 = NoCPUReg);
1997
1998 // Code ageing support functions.
1999
2000 // Code ageing on ARM64 works similarly to on ARM. When V8 wants to mark a
2001 // function as old, it replaces some of the function prologue (generated by
2002 // FullCodeGenerator::Generate) with a call to a special stub (ultimately
2003 // generated by GenerateMakeCodeYoungAgainCommon). The stub restores the
2004 // function prologue to its initial young state (indicating that it has been
2005 // recently run) and continues. A young function is therefore one which has a
2006 // normal frame setup sequence, and an old function has a code age sequence
2007 // which calls a code ageing stub.
2008
2009 // Set up a basic stack frame for young code (or code exempt from ageing) with
2010 // type FUNCTION. It may be patched later for code ageing support. This is
2011 // done by to Code::PatchPlatformCodeAge and EmitCodeAgeSequence.
2012 //
2013 // This function takes an Assembler so it can be called from either a
2014 // MacroAssembler or a PatchingAssembler context.
2015 static void EmitFrameSetupForCodeAgePatching(Assembler* assm);
2016
2017 // Call EmitFrameSetupForCodeAgePatching from a MacroAssembler context.
2018 void EmitFrameSetupForCodeAgePatching();
2019
2020 // Emit a code age sequence that calls the relevant code age stub. The code
2021 // generated by this sequence is expected to replace the code generated by
2022 // EmitFrameSetupForCodeAgePatching, and represents an old function.
2023 //
2024 // If stub is NULL, this function generates the code age sequence but omits
2025 // the stub address that is normally embedded in the instruction stream. This
2026 // can be used by debug code to verify code age sequences.
2027 static void EmitCodeAgeSequence(Assembler* assm, Code* stub);
2028
2029 // Call EmitCodeAgeSequence from a MacroAssembler context.
2030 void EmitCodeAgeSequence(Code* stub);
2031
2032 // Return true if the sequence is a young sequence geneated by
2033 // EmitFrameSetupForCodeAgePatching. Otherwise, this method asserts that the
2034 // sequence is a code age sequence (emitted by EmitCodeAgeSequence).
2035 static bool IsYoungSequence(Isolate* isolate, byte* sequence);
2036
2037 // Jumps to found label if a prototype map has dictionary elements.
2038 void JumpIfDictionaryInPrototypeChain(Register object, Register scratch0,
2039 Register scratch1, Label* found);
2040
2041 // Perform necessary maintenance operations before a push or after a pop.
2042 //
2043 // Note that size is specified in bytes.
2044 void PushPreamble(Operand total_size);
2045 void PopPostamble(Operand total_size);
2046
2047 void PushPreamble(int count, int size) { PushPreamble(count * size); }
2048 void PopPostamble(int count, int size) { PopPostamble(count * size); }
2049
2050 private:
2051 // Helpers for CopyFields.
2052 // These each implement CopyFields in a different way.
2053 void CopyFieldsLoopPairsHelper(Register dst, Register src, unsigned count,
2054 Register scratch1, Register scratch2,
2055 Register scratch3, Register scratch4,
2056 Register scratch5);
2057 void CopyFieldsUnrolledPairsHelper(Register dst, Register src, unsigned count,
2058 Register scratch1, Register scratch2,
2059 Register scratch3, Register scratch4);
2060 void CopyFieldsUnrolledHelper(Register dst, Register src, unsigned count,
2061 Register scratch1, Register scratch2,
2062 Register scratch3);
2063
2064 // The actual Push and Pop implementations. These don't generate any code
2065 // other than that required for the push or pop. This allows
2066 // (Push|Pop)CPURegList to bundle together run-time assertions for a large
2067 // block of registers.
2068 //
2069 // Note that size is per register, and is specified in bytes.
2070 void PushHelper(int count, int size,
2071 const CPURegister& src0, const CPURegister& src1,
2072 const CPURegister& src2, const CPURegister& src3);
2073 void PopHelper(int count, int size,
2074 const CPURegister& dst0, const CPURegister& dst1,
2075 const CPURegister& dst2, const CPURegister& dst3);
2076
2077 // Call Printf. On a native build, a simple call will be generated, but if the
2078 // simulator is being used then a suitable pseudo-instruction is used. The
2079 // arguments and stack (csp) must be prepared by the caller as for a normal
2080 // AAPCS64 call to 'printf'.
2081 //
2082 // The 'args' argument should point to an array of variable arguments in their
2083 // proper PCS registers (and in calling order). The argument registers can
2084 // have mixed types. The format string (x0) should not be included.
2085 void CallPrintf(int arg_count = 0, const CPURegister * args = NULL);
2086
2087 // Helper for throwing exceptions. Compute a handler address and jump to
2088 // it. See the implementation for register usage.
2089 void JumpToHandlerEntry(Register exception,
2090 Register object,
2091 Register state,
2092 Register scratch1,
2093 Register scratch2);
2094
2095 // Helper for implementing JumpIfNotInNewSpace and JumpIfInNewSpace.
2096 void InNewSpace(Register object,
2097 Condition cond, // eq for new space, ne otherwise.
2098 Label* branch);
2099
2100 // Try to represent a double as an int so that integer fast-paths may be
2101 // used. Not every valid integer value is guaranteed to be caught.
2102 // It supports both 32-bit and 64-bit integers depending whether 'as_int'
2103 // is a W or X register.
2104 //
2105 // This does not distinguish between +0 and -0, so if this distinction is
2106 // important it must be checked separately.
2107 //
2108 // On output the Z flag is set if the operation was successful.
2109 void TryRepresentDoubleAsInt(Register as_int,
2110 FPRegister value,
2111 FPRegister scratch_d,
2112 Label* on_successful_conversion = NULL,
2113 Label* on_failed_conversion = NULL);
2114
2115 bool generating_stub_;
2116#if DEBUG
2117 // Tell whether any of the macro instruction can be used. When false the
2118 // MacroAssembler will assert if a method which can emit a variable number
2119 // of instructions is called.
2120 bool allow_macro_instructions_;
2121#endif
2122 bool has_frame_;
2123
2124 // The Abort method should call a V8 runtime function, but the CallRuntime
2125 // mechanism depends on CEntryStub. If use_real_aborts is false, Abort will
2126 // use a simpler abort mechanism that doesn't depend on CEntryStub.
2127 //
2128 // The purpose of this is to allow Aborts to be compiled whilst CEntryStub is
2129 // being generated.
2130 bool use_real_aborts_;
2131
2132 // This handle will be patched with the code object on installation.
2133 Handle<Object> code_object_;
2134
2135 // The register to use as a stack pointer for stack operations.
2136 Register sp_;
2137
2138 // Scratch registers available for use by the MacroAssembler.
2139 CPURegList tmp_list_;
2140 CPURegList fptmp_list_;
2141
2142 void InitializeNewString(Register string,
2143 Register length,
2144 Heap::RootListIndex map_index,
2145 Register scratch1,
2146 Register scratch2);
2147
2148 public:
2149 // Far branches resolving.
2150 //
2151 // The various classes of branch instructions with immediate offsets have
2152 // different ranges. While the Assembler will fail to assemble a branch
2153 // exceeding its range, the MacroAssembler offers a mechanism to resolve
2154 // branches to too distant targets, either by tweaking the generated code to
2155 // use branch instructions with wider ranges or generating veneers.
2156 //
2157 // Currently branches to distant targets are resolved using unconditional
2158 // branch isntructions with a range of +-128MB. If that becomes too little
2159 // (!), the mechanism can be extended to generate special veneers for really
2160 // far targets.
2161
2162 // Helps resolve branching to labels potentially out of range.
2163 // If the label is not bound, it registers the information necessary to later
2164 // be able to emit a veneer for this branch if necessary.
2165 // If the label is bound, it returns true if the label (or the previous link
2166 // in the label chain) is out of range. In that case the caller is responsible
2167 // for generating appropriate code.
2168 // Otherwise it returns false.
2169 // This function also checks wether veneers need to be emitted.
2170 bool NeedExtraInstructionsOrRegisterBranch(Label *label,
2171 ImmBranchType branch_type);
2172};
2173
2174
2175// Use this scope when you need a one-to-one mapping bewteen methods and
2176// instructions. This scope prevents the MacroAssembler from being called and
2177// literal pools from being emitted. It also asserts the number of instructions
2178// emitted is what you specified when creating the scope.
2179class InstructionAccurateScope BASE_EMBEDDED {
2180 public:
2181 explicit InstructionAccurateScope(MacroAssembler* masm, size_t count = 0)
2182 : masm_(masm)
2183#ifdef DEBUG
2184 ,
2185 size_(count * kInstructionSize)
2186#endif
2187 {
2188 // Before blocking the const pool, see if it needs to be emitted.
2189 masm_->CheckConstPool(false, true);
2190 masm_->CheckVeneerPool(false, true);
2191
2192 masm_->StartBlockPools();
2193#ifdef DEBUG
2194 if (count != 0) {
2195 masm_->bind(&start_);
2196 }
2197 previous_allow_macro_instructions_ = masm_->allow_macro_instructions();
2198 masm_->set_allow_macro_instructions(false);
2199#endif
2200 }
2201
2202 ~InstructionAccurateScope() {
2203 masm_->EndBlockPools();
2204#ifdef DEBUG
2205 if (start_.is_bound()) {
2206 DCHECK(masm_->SizeOfCodeGeneratedSince(&start_) == size_);
2207 }
2208 masm_->set_allow_macro_instructions(previous_allow_macro_instructions_);
2209#endif
2210 }
2211
2212 private:
2213 MacroAssembler* masm_;
2214#ifdef DEBUG
2215 size_t size_;
2216 Label start_;
2217 bool previous_allow_macro_instructions_;
2218#endif
2219};
2220
2221
2222// This scope utility allows scratch registers to be managed safely. The
2223// MacroAssembler's TmpList() (and FPTmpList()) is used as a pool of scratch
2224// registers. These registers can be allocated on demand, and will be returned
2225// at the end of the scope.
2226//
2227// When the scope ends, the MacroAssembler's lists will be restored to their
2228// original state, even if the lists were modified by some other means.
2229class UseScratchRegisterScope {
2230 public:
2231 explicit UseScratchRegisterScope(MacroAssembler* masm)
2232 : available_(masm->TmpList()),
2233 availablefp_(masm->FPTmpList()),
2234 old_available_(available_->list()),
2235 old_availablefp_(availablefp_->list()) {
2236 DCHECK(available_->type() == CPURegister::kRegister);
2237 DCHECK(availablefp_->type() == CPURegister::kFPRegister);
2238 }
2239
2240 ~UseScratchRegisterScope();
2241
2242 // Take a register from the appropriate temps list. It will be returned
2243 // automatically when the scope ends.
2244 Register AcquireW() { return AcquireNextAvailable(available_).W(); }
2245 Register AcquireX() { return AcquireNextAvailable(available_).X(); }
2246 FPRegister AcquireS() { return AcquireNextAvailable(availablefp_).S(); }
2247 FPRegister AcquireD() { return AcquireNextAvailable(availablefp_).D(); }
2248
2249 Register UnsafeAcquire(const Register& reg) {
2250 return Register(UnsafeAcquire(available_, reg));
2251 }
2252
2253 Register AcquireSameSizeAs(const Register& reg);
2254 FPRegister AcquireSameSizeAs(const FPRegister& reg);
2255
2256 private:
2257 static CPURegister AcquireNextAvailable(CPURegList* available);
2258 static CPURegister UnsafeAcquire(CPURegList* available,
2259 const CPURegister& reg);
2260
2261 // Available scratch registers.
2262 CPURegList* available_; // kRegister
2263 CPURegList* availablefp_; // kFPRegister
2264
2265 // The state of the available lists at the start of this scope.
2266 RegList old_available_; // kRegister
2267 RegList old_availablefp_; // kFPRegister
2268};
2269
2270
2271inline MemOperand ContextMemOperand(Register context, int index) {
2272 return MemOperand(context, Context::SlotOffset(index));
2273}
2274
2275inline MemOperand GlobalObjectMemOperand() {
2276 return ContextMemOperand(cp, Context::GLOBAL_OBJECT_INDEX);
2277}
2278
2279
2280// Encode and decode information about patchable inline SMI checks.
2281class InlineSmiCheckInfo {
2282 public:
2283 explicit InlineSmiCheckInfo(Address info);
2284
2285 bool HasSmiCheck() const {
2286 return smi_check_ != NULL;
2287 }
2288
2289 const Register& SmiRegister() const {
2290 return reg_;
2291 }
2292
2293 Instruction* SmiCheck() const {
2294 return smi_check_;
2295 }
2296
2297 // Use MacroAssembler::InlineData to emit information about patchable inline
2298 // SMI checks. The caller may specify 'reg' as NoReg and an unbound 'site' to
2299 // indicate that there is no inline SMI check. Note that 'reg' cannot be csp.
2300 //
2301 // The generated patch information can be read using the InlineSMICheckInfo
2302 // class.
2303 static void Emit(MacroAssembler* masm, const Register& reg,
2304 const Label* smi_check);
2305
2306 // Emit information to indicate that there is no inline SMI check.
2307 static void EmitNotInlined(MacroAssembler* masm) {
2308 Label unbound;
2309 Emit(masm, NoReg, &unbound);
2310 }
2311
2312 private:
2313 Register reg_;
2314 Instruction* smi_check_;
2315
2316 // Fields in the data encoded by InlineData.
2317
2318 // A width of 5 (Rd_width) for the SMI register preclues the use of csp,
2319 // since kSPRegInternalCode is 63. However, csp should never hold a SMI or be
2320 // used in a patchable check. The Emit() method checks this.
2321 //
2322 // Note that the total size of the fields is restricted by the underlying
2323 // storage size handled by the BitField class, which is a uint32_t.
2324 class RegisterBits : public BitField<unsigned, 0, 5> {};
2325 class DeltaBits : public BitField<uint32_t, 5, 32-5> {};
2326};
2327
2328} } // namespace v8::internal
2329
2330#ifdef GENERATED_CODE_COVERAGE
2331#error "Unsupported option"
2332#define CODE_COVERAGE_STRINGIFY(x) #x
2333#define CODE_COVERAGE_TOSTRING(x) CODE_COVERAGE_STRINGIFY(x)
2334#define __FILE_LINE__ __FILE__ ":" CODE_COVERAGE_TOSTRING(__LINE__)
2335#define ACCESS_MASM(masm) masm->stop(__FILE_LINE__); masm->
2336#else
2337#define ACCESS_MASM(masm) masm->
2338#endif
2339
2340#endif // V8_ARM64_MACRO_ASSEMBLER_ARM64_H_