blob: f22c479780fb1a02ab0eb2108e63ba95ffd1b71f [file] [log] [blame]
Ben Murdochb8a8cc12014-11-26 15:28:44 +00001// Copyright 2013 the V8 project authors. All rights reserved.
2// Use of this source code is governed by a BSD-style license that can be
3// found in the LICENSE file.
4
5#include "src/compiler/code-generator.h"
6
7#include "src/compiler/code-generator-impl.h"
8#include "src/compiler/linkage.h"
9#include "src/compiler/pipeline.h"
10
11namespace v8 {
12namespace internal {
13namespace compiler {
14
15CodeGenerator::CodeGenerator(InstructionSequence* code)
16 : code_(code),
17 current_block_(NULL),
18 current_source_position_(SourcePosition::Invalid()),
19 masm_(code->zone()->isolate(), NULL, 0),
20 resolver_(this),
21 safepoints_(code->zone()),
22 deoptimization_states_(code->zone()),
23 deoptimization_literals_(code->zone()),
24 translations_(code->zone()),
25 last_lazy_deopt_pc_(0) {}
26
27
28Handle<Code> CodeGenerator::GenerateCode() {
29 CompilationInfo* info = linkage()->info();
30
31 // Emit a code line info recording start event.
32 PositionsRecorder* recorder = masm()->positions_recorder();
33 LOG_CODE_EVENT(isolate(), CodeStartLinePosInfoRecordEvent(recorder));
34
35 // Place function entry hook if requested to do so.
36 if (linkage()->GetIncomingDescriptor()->IsJSFunctionCall()) {
37 ProfileEntryHookStub::MaybeCallEntryHook(masm());
38 }
39
40 // Architecture-specific, linkage-specific prologue.
41 info->set_prologue_offset(masm()->pc_offset());
42 AssemblePrologue();
43
44 // Assemble all instructions.
45 for (InstructionSequence::const_iterator i = code()->begin();
46 i != code()->end(); ++i) {
47 AssembleInstruction(*i);
48 }
49
50 FinishCode(masm());
51
52 // Ensure there is space for lazy deopt.
53 if (!info->IsStub()) {
54 int target_offset = masm()->pc_offset() + Deoptimizer::patch_size();
55 while (masm()->pc_offset() < target_offset) {
56 masm()->nop();
57 }
58 }
59
60 safepoints()->Emit(masm(), frame()->GetSpillSlotCount());
61
62 // TODO(titzer): what are the right code flags here?
63 Code::Kind kind = Code::STUB;
64 if (linkage()->GetIncomingDescriptor()->IsJSFunctionCall()) {
65 kind = Code::OPTIMIZED_FUNCTION;
66 }
67 Handle<Code> result = v8::internal::CodeGenerator::MakeCodeEpilogue(
68 masm(), Code::ComputeFlags(kind), info);
69 result->set_is_turbofanned(true);
70 result->set_stack_slots(frame()->GetSpillSlotCount());
71 result->set_safepoint_table_offset(safepoints()->GetCodeOffset());
72
73 PopulateDeoptimizationData(result);
74
75 // Emit a code line info recording stop event.
76 void* line_info = recorder->DetachJITHandlerData();
77 LOG_CODE_EVENT(isolate(), CodeEndLinePosInfoRecordEvent(*result, line_info));
78
79 return result;
80}
81
82
83void CodeGenerator::RecordSafepoint(PointerMap* pointers, Safepoint::Kind kind,
84 int arguments,
85 Safepoint::DeoptMode deopt_mode) {
86 const ZoneList<InstructionOperand*>* operands =
87 pointers->GetNormalizedOperands();
88 Safepoint safepoint =
89 safepoints()->DefineSafepoint(masm(), kind, arguments, deopt_mode);
90 for (int i = 0; i < operands->length(); i++) {
91 InstructionOperand* pointer = operands->at(i);
92 if (pointer->IsStackSlot()) {
93 safepoint.DefinePointerSlot(pointer->index(), zone());
94 } else if (pointer->IsRegister() && (kind & Safepoint::kWithRegisters)) {
95 Register reg = Register::FromAllocationIndex(pointer->index());
96 safepoint.DefinePointerRegister(reg, zone());
97 }
98 }
99}
100
101
102void CodeGenerator::AssembleInstruction(Instruction* instr) {
103 if (instr->IsBlockStart()) {
104 // Bind a label for a block start and handle parallel moves.
105 BlockStartInstruction* block_start = BlockStartInstruction::cast(instr);
106 current_block_ = block_start->block();
107 if (FLAG_code_comments) {
108 // TODO(titzer): these code comments are a giant memory leak.
109 Vector<char> buffer = Vector<char>::New(32);
110 SNPrintF(buffer, "-- B%d start --", block_start->block()->id());
111 masm()->RecordComment(buffer.start());
112 }
113 masm()->bind(block_start->label());
114 }
115 if (instr->IsGapMoves()) {
116 // Handle parallel moves associated with the gap instruction.
117 AssembleGap(GapInstruction::cast(instr));
118 } else if (instr->IsSourcePosition()) {
119 AssembleSourcePosition(SourcePositionInstruction::cast(instr));
120 } else {
121 // Assemble architecture-specific code for the instruction.
122 AssembleArchInstruction(instr);
123
124 // Assemble branches or boolean materializations after this instruction.
125 FlagsMode mode = FlagsModeField::decode(instr->opcode());
126 FlagsCondition condition = FlagsConditionField::decode(instr->opcode());
127 switch (mode) {
128 case kFlags_none:
129 return;
130 case kFlags_set:
131 return AssembleArchBoolean(instr, condition);
132 case kFlags_branch:
133 return AssembleArchBranch(instr, condition);
134 }
135 UNREACHABLE();
136 }
137}
138
139
140void CodeGenerator::AssembleSourcePosition(SourcePositionInstruction* instr) {
141 SourcePosition source_position = instr->source_position();
142 if (source_position == current_source_position_) return;
143 DCHECK(!source_position.IsInvalid());
144 if (!source_position.IsUnknown()) {
145 int code_pos = source_position.raw();
146 masm()->positions_recorder()->RecordPosition(source_position.raw());
147 masm()->positions_recorder()->WriteRecordedPositions();
148 if (FLAG_code_comments) {
149 Vector<char> buffer = Vector<char>::New(256);
150 CompilationInfo* info = linkage()->info();
151 int ln = Script::GetLineNumber(info->script(), code_pos);
152 int cn = Script::GetColumnNumber(info->script(), code_pos);
153 if (info->script()->name()->IsString()) {
154 Handle<String> file(String::cast(info->script()->name()));
155 base::OS::SNPrintF(buffer.start(), buffer.length(), "-- %s:%d:%d --",
156 file->ToCString().get(), ln, cn);
157 } else {
158 base::OS::SNPrintF(buffer.start(), buffer.length(),
159 "-- <unknown>:%d:%d --", ln, cn);
160 }
161 masm()->RecordComment(buffer.start());
162 }
163 }
164 current_source_position_ = source_position;
165}
166
167
168void CodeGenerator::AssembleGap(GapInstruction* instr) {
169 for (int i = GapInstruction::FIRST_INNER_POSITION;
170 i <= GapInstruction::LAST_INNER_POSITION; i++) {
171 GapInstruction::InnerPosition inner_pos =
172 static_cast<GapInstruction::InnerPosition>(i);
173 ParallelMove* move = instr->GetParallelMove(inner_pos);
174 if (move != NULL) resolver()->Resolve(move);
175 }
176}
177
178
179void CodeGenerator::PopulateDeoptimizationData(Handle<Code> code_object) {
180 CompilationInfo* info = linkage()->info();
181 int deopt_count = static_cast<int>(deoptimization_states_.size());
182 if (deopt_count == 0) return;
183 Handle<DeoptimizationInputData> data =
184 DeoptimizationInputData::New(isolate(), deopt_count, TENURED);
185
186 Handle<ByteArray> translation_array =
187 translations_.CreateByteArray(isolate()->factory());
188
189 data->SetTranslationByteArray(*translation_array);
190 data->SetInlinedFunctionCount(Smi::FromInt(0));
191 data->SetOptimizationId(Smi::FromInt(info->optimization_id()));
192 // TODO(jarin) The following code was copied over from Lithium, not sure
193 // whether the scope or the IsOptimizing condition are really needed.
194 if (info->IsOptimizing()) {
195 // Reference to shared function info does not change between phases.
196 AllowDeferredHandleDereference allow_handle_dereference;
197 data->SetSharedFunctionInfo(*info->shared_info());
198 } else {
199 data->SetSharedFunctionInfo(Smi::FromInt(0));
200 }
201
202 Handle<FixedArray> literals = isolate()->factory()->NewFixedArray(
203 static_cast<int>(deoptimization_literals_.size()), TENURED);
204 {
205 AllowDeferredHandleDereference copy_handles;
206 for (unsigned i = 0; i < deoptimization_literals_.size(); i++) {
207 literals->set(i, *deoptimization_literals_[i]);
208 }
209 data->SetLiteralArray(*literals);
210 }
211
212 // No OSR in Turbofan yet...
213 BailoutId osr_ast_id = BailoutId::None();
214 data->SetOsrAstId(Smi::FromInt(osr_ast_id.ToInt()));
215 data->SetOsrPcOffset(Smi::FromInt(-1));
216
217 // Populate deoptimization entries.
218 for (int i = 0; i < deopt_count; i++) {
219 DeoptimizationState* deoptimization_state = deoptimization_states_[i];
220 data->SetAstId(i, deoptimization_state->bailout_id());
221 CHECK_NE(NULL, deoptimization_states_[i]);
222 data->SetTranslationIndex(
223 i, Smi::FromInt(deoptimization_states_[i]->translation_id()));
224 data->SetArgumentsStackHeight(i, Smi::FromInt(0));
225 data->SetPc(i, Smi::FromInt(deoptimization_state->pc_offset()));
226 }
227
228 code_object->set_deoptimization_data(*data);
229}
230
231
232void CodeGenerator::AddSafepointAndDeopt(Instruction* instr) {
233 CallDescriptor::Flags flags(MiscField::decode(instr->opcode()));
234
235 bool needs_frame_state = (flags & CallDescriptor::kNeedsFrameState);
236
237 RecordSafepoint(
238 instr->pointer_map(), Safepoint::kSimple, 0,
239 needs_frame_state ? Safepoint::kLazyDeopt : Safepoint::kNoLazyDeopt);
240
241 if (flags & CallDescriptor::kNeedsNopAfterCall) {
242 AddNopForSmiCodeInlining();
243 }
244
245 if (needs_frame_state) {
246 MarkLazyDeoptSite();
247 // If the frame state is present, it starts at argument 1
248 // (just after the code address).
249 InstructionOperandConverter converter(this, instr);
250 // Deoptimization info starts at argument 1
251 size_t frame_state_offset = 1;
252 FrameStateDescriptor* descriptor =
253 GetFrameStateDescriptor(instr, frame_state_offset);
254 int pc_offset = masm()->pc_offset();
255 int deopt_state_id = BuildTranslation(instr, pc_offset, frame_state_offset,
256 descriptor->state_combine());
257 // If the pre-call frame state differs from the post-call one, produce the
258 // pre-call frame state, too.
259 // TODO(jarin) We might want to avoid building the pre-call frame state
260 // because it is only used to get locals and arguments (by the debugger and
261 // f.arguments), and those are the same in the pre-call and post-call
262 // states.
263 if (descriptor->state_combine() != kIgnoreOutput) {
264 deopt_state_id =
265 BuildTranslation(instr, -1, frame_state_offset, kIgnoreOutput);
266 }
267#if DEBUG
268 // Make sure all the values live in stack slots or they are immediates.
269 // (The values should not live in register because registers are clobbered
270 // by calls.)
271 for (size_t i = 0; i < descriptor->size(); i++) {
272 InstructionOperand* op = instr->InputAt(frame_state_offset + 1 + i);
273 CHECK(op->IsStackSlot() || op->IsImmediate());
274 }
275#endif
276 safepoints()->RecordLazyDeoptimizationIndex(deopt_state_id);
277 }
278}
279
280
281int CodeGenerator::DefineDeoptimizationLiteral(Handle<Object> literal) {
282 int result = static_cast<int>(deoptimization_literals_.size());
283 for (unsigned i = 0; i < deoptimization_literals_.size(); ++i) {
284 if (deoptimization_literals_[i].is_identical_to(literal)) return i;
285 }
286 deoptimization_literals_.push_back(literal);
287 return result;
288}
289
290
291FrameStateDescriptor* CodeGenerator::GetFrameStateDescriptor(
292 Instruction* instr, size_t frame_state_offset) {
293 InstructionOperandConverter i(this, instr);
294 InstructionSequence::StateId state_id = InstructionSequence::StateId::FromInt(
295 i.InputInt32(static_cast<int>(frame_state_offset)));
296 return code()->GetFrameStateDescriptor(state_id);
297}
298
299
300void CodeGenerator::BuildTranslationForFrameStateDescriptor(
301 FrameStateDescriptor* descriptor, Instruction* instr,
302 Translation* translation, size_t frame_state_offset,
303 OutputFrameStateCombine state_combine) {
304 // Outer-most state must be added to translation first.
305 if (descriptor->outer_state() != NULL) {
306 BuildTranslationForFrameStateDescriptor(descriptor->outer_state(), instr,
307 translation, frame_state_offset,
308 kIgnoreOutput);
309 }
310
311 int id = Translation::kSelfLiteralId;
312 if (!descriptor->jsfunction().is_null()) {
313 id = DefineDeoptimizationLiteral(
314 Handle<Object>::cast(descriptor->jsfunction().ToHandleChecked()));
315 }
316
317 switch (descriptor->type()) {
318 case JS_FRAME:
319 translation->BeginJSFrame(
320 descriptor->bailout_id(), id,
321 static_cast<unsigned int>(descriptor->GetHeight(state_combine)));
322 break;
323 case ARGUMENTS_ADAPTOR:
324 translation->BeginArgumentsAdaptorFrame(
325 id, static_cast<unsigned int>(descriptor->parameters_count()));
326 break;
327 }
328
329 frame_state_offset += descriptor->outer_state()->GetTotalSize();
330 for (size_t i = 0; i < descriptor->size(); i++) {
331 AddTranslationForOperand(
332 translation, instr,
333 instr->InputAt(static_cast<int>(frame_state_offset + i)));
334 }
335
336 switch (state_combine) {
337 case kPushOutput:
338 DCHECK(instr->OutputCount() == 1);
339 AddTranslationForOperand(translation, instr, instr->OutputAt(0));
340 break;
341 case kIgnoreOutput:
342 break;
343 }
344}
345
346
347int CodeGenerator::BuildTranslation(Instruction* instr, int pc_offset,
348 size_t frame_state_offset,
349 OutputFrameStateCombine state_combine) {
350 FrameStateDescriptor* descriptor =
351 GetFrameStateDescriptor(instr, frame_state_offset);
352 frame_state_offset++;
353
354 Translation translation(
355 &translations_, static_cast<int>(descriptor->GetFrameCount()),
356 static_cast<int>(descriptor->GetJSFrameCount()), zone());
357 BuildTranslationForFrameStateDescriptor(descriptor, instr, &translation,
358 frame_state_offset, state_combine);
359
360 int deoptimization_id = static_cast<int>(deoptimization_states_.size());
361
362 deoptimization_states_.push_back(new (zone()) DeoptimizationState(
363 descriptor->bailout_id(), translation.index(), pc_offset));
364
365 return deoptimization_id;
366}
367
368
369void CodeGenerator::AddTranslationForOperand(Translation* translation,
370 Instruction* instr,
371 InstructionOperand* op) {
372 if (op->IsStackSlot()) {
373 translation->StoreStackSlot(op->index());
374 } else if (op->IsDoubleStackSlot()) {
375 translation->StoreDoubleStackSlot(op->index());
376 } else if (op->IsRegister()) {
377 InstructionOperandConverter converter(this, instr);
378 translation->StoreRegister(converter.ToRegister(op));
379 } else if (op->IsDoubleRegister()) {
380 InstructionOperandConverter converter(this, instr);
381 translation->StoreDoubleRegister(converter.ToDoubleRegister(op));
382 } else if (op->IsImmediate()) {
383 InstructionOperandConverter converter(this, instr);
384 Constant constant = converter.ToConstant(op);
385 Handle<Object> constant_object;
386 switch (constant.type()) {
387 case Constant::kInt32:
388 constant_object =
389 isolate()->factory()->NewNumberFromInt(constant.ToInt32());
390 break;
391 case Constant::kFloat64:
392 constant_object = isolate()->factory()->NewNumber(constant.ToFloat64());
393 break;
394 case Constant::kHeapObject:
395 constant_object = constant.ToHeapObject();
396 break;
397 default:
398 UNREACHABLE();
399 }
400 int literal_id = DefineDeoptimizationLiteral(constant_object);
401 translation->StoreLiteral(literal_id);
402 } else {
403 UNREACHABLE();
404 }
405}
406
407
408void CodeGenerator::MarkLazyDeoptSite() {
409 last_lazy_deopt_pc_ = masm()->pc_offset();
410}
411
412#if !V8_TURBOFAN_BACKEND
413
414void CodeGenerator::AssembleArchInstruction(Instruction* instr) {
415 UNIMPLEMENTED();
416}
417
418
419void CodeGenerator::AssembleArchBranch(Instruction* instr,
420 FlagsCondition condition) {
421 UNIMPLEMENTED();
422}
423
424
425void CodeGenerator::AssembleArchBoolean(Instruction* instr,
426 FlagsCondition condition) {
427 UNIMPLEMENTED();
428}
429
430
431void CodeGenerator::AssembleDeoptimizerCall(int deoptimization_id) {
432 UNIMPLEMENTED();
433}
434
435
436void CodeGenerator::AssemblePrologue() { UNIMPLEMENTED(); }
437
438
439void CodeGenerator::AssembleReturn() { UNIMPLEMENTED(); }
440
441
442void CodeGenerator::AssembleMove(InstructionOperand* source,
443 InstructionOperand* destination) {
444 UNIMPLEMENTED();
445}
446
447
448void CodeGenerator::AssembleSwap(InstructionOperand* source,
449 InstructionOperand* destination) {
450 UNIMPLEMENTED();
451}
452
453
454void CodeGenerator::AddNopForSmiCodeInlining() { UNIMPLEMENTED(); }
455
456#endif // !V8_TURBOFAN_BACKEND
457
458} // namespace compiler
459} // namespace internal
460} // namespace v8