blob: 1a391ce9bbfe2bd931568a5ecc586a0ff371b998 [file] [log] [blame]
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001/*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
Matthew Gharritye9288852016-07-14 14:08:16 -070017#include "register_allocator_linear_scan.h"
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010018
Nicolas Geoffray234d69d2015-03-09 10:28:50 +000019#include <iostream>
Ian Rogersc7dd2952014-10-21 23:31:19 -070020#include <sstream>
21
Ian Rogerse77493c2014-08-20 15:08:45 -070022#include "base/bit_vector-inl.h"
Andreas Gampe542451c2016-07-26 09:02:02 -070023#include "base/enums.h"
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010024#include "code_generator.h"
Aart Bik96202302016-10-04 17:33:56 -070025#include "linear_order.h"
Matthew Gharrity5d6e27d2016-07-18 13:38:44 -070026#include "register_allocation_resolver.h"
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010027#include "ssa_liveness_analysis.h"
28
29namespace art {
30
31static constexpr size_t kMaxLifetimePosition = -1;
Nicolas Geoffray31d76b42014-06-09 15:02:22 +010032static constexpr size_t kDefaultNumberOfSpillSlots = 4;
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010033
Nicolas Geoffray840e5462015-01-07 16:01:24 +000034// For simplicity, we implement register pairs as (reg, reg + 1).
35// Note that this is a requirement for double registers on ARM, since we
36// allocate SRegister.
37static int GetHighForLowRegister(int reg) { return reg + 1; }
38static bool IsLowRegister(int reg) { return (reg & 1) == 0; }
Nicolas Geoffray234d69d2015-03-09 10:28:50 +000039static bool IsLowOfUnalignedPairInterval(LiveInterval* low) {
40 return GetHighForLowRegister(low->GetRegister()) != low->GetHighInterval()->GetRegister();
41}
Nicolas Geoffray840e5462015-01-07 16:01:24 +000042
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -070043RegisterAllocatorLinearScan::RegisterAllocatorLinearScan(ArenaAllocator* allocator,
44 CodeGenerator* codegen,
45 const SsaLivenessAnalysis& liveness)
46 : RegisterAllocator(allocator, codegen, liveness),
Vladimir Marko2aaa4b52015-09-17 17:03:26 +010047 unhandled_core_intervals_(allocator->Adapter(kArenaAllocRegisterAllocator)),
48 unhandled_fp_intervals_(allocator->Adapter(kArenaAllocRegisterAllocator)),
Nicolas Geoffray39468442014-09-02 15:17:15 +010049 unhandled_(nullptr),
Vladimir Marko2aaa4b52015-09-17 17:03:26 +010050 handled_(allocator->Adapter(kArenaAllocRegisterAllocator)),
51 active_(allocator->Adapter(kArenaAllocRegisterAllocator)),
52 inactive_(allocator->Adapter(kArenaAllocRegisterAllocator)),
53 physical_core_register_intervals_(allocator->Adapter(kArenaAllocRegisterAllocator)),
54 physical_fp_register_intervals_(allocator->Adapter(kArenaAllocRegisterAllocator)),
55 temp_intervals_(allocator->Adapter(kArenaAllocRegisterAllocator)),
56 int_spill_slots_(allocator->Adapter(kArenaAllocRegisterAllocator)),
57 long_spill_slots_(allocator->Adapter(kArenaAllocRegisterAllocator)),
58 float_spill_slots_(allocator->Adapter(kArenaAllocRegisterAllocator)),
59 double_spill_slots_(allocator->Adapter(kArenaAllocRegisterAllocator)),
David Brazdil77a48ae2015-09-15 12:34:04 +000060 catch_phi_spill_slots_(0),
Vladimir Marko2aaa4b52015-09-17 17:03:26 +010061 safepoints_(allocator->Adapter(kArenaAllocRegisterAllocator)),
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010062 processing_core_registers_(false),
63 number_of_registers_(-1),
64 registers_array_(nullptr),
Nicolas Geoffray102cbed2014-10-15 18:31:05 +010065 blocked_core_registers_(codegen->GetBlockedCoreRegisters()),
66 blocked_fp_registers_(codegen->GetBlockedFloatingPointRegisters()),
Vladimir Marko70e97462016-08-09 11:04:26 +010067 reserved_out_slots_(0) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +010068 temp_intervals_.reserve(4);
69 int_spill_slots_.reserve(kDefaultNumberOfSpillSlots);
70 long_spill_slots_.reserve(kDefaultNumberOfSpillSlots);
71 float_spill_slots_.reserve(kDefaultNumberOfSpillSlots);
72 double_spill_slots_.reserve(kDefaultNumberOfSpillSlots);
73
David Brazdil58282f42016-01-14 12:45:10 +000074 codegen->SetupBlockedRegisters();
Vladimir Marko2aaa4b52015-09-17 17:03:26 +010075 physical_core_register_intervals_.resize(codegen->GetNumberOfCoreRegisters(), nullptr);
76 physical_fp_register_intervals_.resize(codegen->GetNumberOfFloatingPointRegisters(), nullptr);
Nicolas Geoffray39468442014-09-02 15:17:15 +010077 // Always reserve for the current method and the graph's max out registers.
78 // TODO: compute it instead.
Mathieu Chartiere401d142015-04-22 13:56:20 -070079 // ArtMethod* takes 2 vregs for 64 bits.
Andreas Gampe542451c2016-07-26 09:02:02 -070080 size_t ptr_size = static_cast<size_t>(InstructionSetPointerSize(codegen->GetInstructionSet()));
81 reserved_out_slots_ = ptr_size / kVRegSize + codegen->GetGraph()->GetMaximumNumberOfOutVRegs();
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010082}
83
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +010084static bool ShouldProcess(bool processing_core_registers, LiveInterval* interval) {
Nicolas Geoffray39468442014-09-02 15:17:15 +010085 if (interval == nullptr) return false;
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +010086 bool is_core_register = (interval->GetType() != Primitive::kPrimDouble)
87 && (interval->GetType() != Primitive::kPrimFloat);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +010088 return processing_core_registers == is_core_register;
89}
90
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -070091void RegisterAllocatorLinearScan::AllocateRegisters() {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +010092 AllocateRegistersInternal();
Matthew Gharrity5d6e27d2016-07-18 13:38:44 -070093 RegisterAllocationResolver(allocator_, codegen_, liveness_)
Vladimir Marko70e97462016-08-09 11:04:26 +010094 .Resolve(ArrayRef<HInstruction* const>(safepoints_),
Matthew Gharrity5d6e27d2016-07-18 13:38:44 -070095 reserved_out_slots_,
96 int_spill_slots_.size(),
97 long_spill_slots_.size(),
98 float_spill_slots_.size(),
99 double_spill_slots_.size(),
100 catch_phi_spill_slots_,
101 temp_intervals_);
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100102
103 if (kIsDebugBuild) {
104 processing_core_registers_ = true;
105 ValidateInternal(true);
106 processing_core_registers_ = false;
107 ValidateInternal(true);
Nicolas Geoffray59768572014-12-01 09:50:04 +0000108 // Check that the linear order is still correct with regards to lifetime positions.
109 // Since only parallel moves have been inserted during the register allocation,
110 // these checks are mostly for making sure these moves have been added correctly.
111 size_t current_liveness = 0;
Aart Bik96202302016-10-04 17:33:56 -0700112 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearOrder()) {
Nicolas Geoffray59768572014-12-01 09:50:04 +0000113 for (HInstructionIterator inst_it(block->GetPhis()); !inst_it.Done(); inst_it.Advance()) {
114 HInstruction* instruction = inst_it.Current();
115 DCHECK_LE(current_liveness, instruction->GetLifetimePosition());
116 current_liveness = instruction->GetLifetimePosition();
117 }
118 for (HInstructionIterator inst_it(block->GetInstructions());
119 !inst_it.Done();
120 inst_it.Advance()) {
121 HInstruction* instruction = inst_it.Current();
122 DCHECK_LE(current_liveness, instruction->GetLifetimePosition()) << instruction->DebugName();
123 current_liveness = instruction->GetLifetimePosition();
124 }
125 }
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100126 }
127}
128
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700129void RegisterAllocatorLinearScan::BlockRegister(Location location, size_t start, size_t end) {
Nicolas Geoffray56b9ee62014-10-09 11:47:51 +0100130 int reg = location.reg();
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100131 DCHECK(location.IsRegister() || location.IsFpuRegister());
132 LiveInterval* interval = location.IsRegister()
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100133 ? physical_core_register_intervals_[reg]
134 : physical_fp_register_intervals_[reg];
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100135 Primitive::Type type = location.IsRegister()
136 ? Primitive::kPrimInt
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000137 : Primitive::kPrimFloat;
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100138 if (interval == nullptr) {
139 interval = LiveInterval::MakeFixedInterval(allocator_, reg, type);
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100140 if (location.IsRegister()) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100141 physical_core_register_intervals_[reg] = interval;
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100142 } else {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100143 physical_fp_register_intervals_[reg] = interval;
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100144 }
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100145 }
146 DCHECK(interval->GetRegister() == reg);
147 interval->AddRange(start, end);
148}
149
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700150void RegisterAllocatorLinearScan::BlockRegisters(size_t start, size_t end, bool caller_save_only) {
David Brazdil77a48ae2015-09-15 12:34:04 +0000151 for (size_t i = 0; i < codegen_->GetNumberOfCoreRegisters(); ++i) {
152 if (!caller_save_only || !codegen_->IsCoreCalleeSaveRegister(i)) {
153 BlockRegister(Location::RegisterLocation(i), start, end);
154 }
155 }
156 for (size_t i = 0; i < codegen_->GetNumberOfFloatingPointRegisters(); ++i) {
157 if (!caller_save_only || !codegen_->IsFloatingPointCalleeSaveRegister(i)) {
158 BlockRegister(Location::FpuRegisterLocation(i), start, end);
159 }
160 }
161}
162
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700163void RegisterAllocatorLinearScan::AllocateRegistersInternal() {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100164 // Iterate post-order, to ensure the list is sorted, and the last added interval
165 // is the one with the lowest start position.
Vladimir Marko2c45bc92016-10-25 16:54:12 +0100166 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearPostOrder()) {
Andreas Gampe277ccbd2014-11-03 21:36:10 -0800167 for (HBackwardInstructionIterator back_it(block->GetInstructions()); !back_it.Done();
168 back_it.Advance()) {
169 ProcessInstruction(back_it.Current());
Nicolas Geoffray39468442014-09-02 15:17:15 +0100170 }
Andreas Gampe277ccbd2014-11-03 21:36:10 -0800171 for (HInstructionIterator inst_it(block->GetPhis()); !inst_it.Done(); inst_it.Advance()) {
172 ProcessInstruction(inst_it.Current());
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100173 }
David Brazdil77a48ae2015-09-15 12:34:04 +0000174
Nicolas Geoffray15bd2282016-01-05 15:55:41 +0000175 if (block->IsCatchBlock() ||
Nicolas Geoffrayad4ed082016-01-27 14:15:23 +0000176 (block->IsLoopHeader() && block->GetLoopInformation()->IsIrreducible())) {
Nicolas Geoffray15bd2282016-01-05 15:55:41 +0000177 // By blocking all registers at the top of each catch block or irreducible loop, we force
178 // intervals belonging to the live-in set of the catch/header block to be spilled.
179 // TODO(ngeoffray): Phis in this block could be allocated in register.
David Brazdil77a48ae2015-09-15 12:34:04 +0000180 size_t position = block->GetLifetimeStart();
181 BlockRegisters(position, position + 1);
182 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100183 }
184
Nicolas Geoffray39468442014-09-02 15:17:15 +0100185 number_of_registers_ = codegen_->GetNumberOfCoreRegisters();
Vladimir Marko5233f932015-09-29 19:01:15 +0100186 registers_array_ = allocator_->AllocArray<size_t>(number_of_registers_,
187 kArenaAllocRegisterAllocator);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100188 processing_core_registers_ = true;
189 unhandled_ = &unhandled_core_intervals_;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100190 for (LiveInterval* fixed : physical_core_register_intervals_) {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100191 if (fixed != nullptr) {
Mingyao Yang296bd602014-10-06 16:47:28 -0700192 // Fixed interval is added to inactive_ instead of unhandled_.
193 // It's also the only type of inactive interval whose start position
194 // can be after the current interval during linear scan.
195 // Fixed interval is never split and never moves to unhandled_.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100196 inactive_.push_back(fixed);
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100197 }
198 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100199 LinearScan();
Nicolas Geoffray39468442014-09-02 15:17:15 +0100200
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100201 inactive_.clear();
202 active_.clear();
203 handled_.clear();
Nicolas Geoffray39468442014-09-02 15:17:15 +0100204
205 number_of_registers_ = codegen_->GetNumberOfFloatingPointRegisters();
Vladimir Marko5233f932015-09-29 19:01:15 +0100206 registers_array_ = allocator_->AllocArray<size_t>(number_of_registers_,
207 kArenaAllocRegisterAllocator);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100208 processing_core_registers_ = false;
209 unhandled_ = &unhandled_fp_intervals_;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100210 for (LiveInterval* fixed : physical_fp_register_intervals_) {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100211 if (fixed != nullptr) {
Mingyao Yang296bd602014-10-06 16:47:28 -0700212 // Fixed interval is added to inactive_ instead of unhandled_.
213 // It's also the only type of inactive interval whose start position
214 // can be after the current interval during linear scan.
215 // Fixed interval is never split and never moves to unhandled_.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100216 inactive_.push_back(fixed);
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100217 }
218 }
Nicolas Geoffray39468442014-09-02 15:17:15 +0100219 LinearScan();
220}
221
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700222void RegisterAllocatorLinearScan::ProcessInstruction(HInstruction* instruction) {
Nicolas Geoffray39468442014-09-02 15:17:15 +0100223 LocationSummary* locations = instruction->GetLocations();
224 size_t position = instruction->GetLifetimePosition();
225
226 if (locations == nullptr) return;
227
228 // Create synthesized intervals for temporaries.
229 for (size_t i = 0; i < locations->GetTempCount(); ++i) {
230 Location temp = locations->GetTemp(i);
Nicolas Geoffray52839d12014-11-07 17:47:25 +0000231 if (temp.IsRegister() || temp.IsFpuRegister()) {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100232 BlockRegister(temp, position, position + 1);
Nicolas Geoffray45b83af2015-07-06 15:12:53 +0000233 // Ensure that an explicit temporary register is marked as being allocated.
234 codegen_->AddAllocatedRegister(temp);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100235 } else {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100236 DCHECK(temp.IsUnallocated());
Roland Levillain5368c212014-11-27 15:03:41 +0000237 switch (temp.GetPolicy()) {
238 case Location::kRequiresRegister: {
239 LiveInterval* interval =
240 LiveInterval::MakeTempInterval(allocator_, Primitive::kPrimInt);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100241 temp_intervals_.push_back(interval);
Nicolas Geoffrayf01d3442015-03-27 17:15:49 +0000242 interval->AddTempUse(instruction, i);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100243 unhandled_core_intervals_.push_back(interval);
Roland Levillain5368c212014-11-27 15:03:41 +0000244 break;
245 }
246
247 case Location::kRequiresFpuRegister: {
248 LiveInterval* interval =
249 LiveInterval::MakeTempInterval(allocator_, Primitive::kPrimDouble);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100250 temp_intervals_.push_back(interval);
Nicolas Geoffrayf01d3442015-03-27 17:15:49 +0000251 interval->AddTempUse(instruction, i);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000252 if (codegen_->NeedsTwoRegisters(Primitive::kPrimDouble)) {
Nicolas Geoffray5588e582015-04-14 14:10:59 +0100253 interval->AddHighInterval(/* is_temp */ true);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000254 LiveInterval* high = interval->GetHighInterval();
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100255 temp_intervals_.push_back(high);
256 unhandled_fp_intervals_.push_back(high);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000257 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100258 unhandled_fp_intervals_.push_back(interval);
Roland Levillain5368c212014-11-27 15:03:41 +0000259 break;
260 }
261
262 default:
263 LOG(FATAL) << "Unexpected policy for temporary location "
264 << temp.GetPolicy();
265 }
Nicolas Geoffray39468442014-09-02 15:17:15 +0100266 }
267 }
268
Nicolas Geoffray3bca0df2014-09-19 11:01:00 +0100269 bool core_register = (instruction->GetType() != Primitive::kPrimDouble)
270 && (instruction->GetType() != Primitive::kPrimFloat);
271
Alexandre Rames8158f282015-08-07 10:26:17 +0100272 if (locations->NeedsSafepoint()) {
Nicolas Geoffrayc0572a42015-02-06 14:35:25 +0000273 if (codegen_->IsLeafMethod()) {
274 // TODO: We do this here because we do not want the suspend check to artificially
275 // create live registers. We should find another place, but this is currently the
276 // simplest.
277 DCHECK(instruction->IsSuspendCheckEntry());
278 instruction->GetBlock()->RemoveInstruction(instruction);
279 return;
Nicolas Geoffray3bca0df2014-09-19 11:01:00 +0100280 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100281 safepoints_.push_back(instruction);
Nicolas Geoffray3bca0df2014-09-19 11:01:00 +0100282 }
283
284 if (locations->WillCall()) {
David Brazdil77a48ae2015-09-15 12:34:04 +0000285 BlockRegisters(position, position + 1, /* caller_save_only */ true);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100286 }
287
Vladimir Marko372f10e2016-05-17 16:30:10 +0100288 for (size_t i = 0; i < locations->GetInputCount(); ++i) {
Nicolas Geoffray39468442014-09-02 15:17:15 +0100289 Location input = locations->InAt(i);
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100290 if (input.IsRegister() || input.IsFpuRegister()) {
291 BlockRegister(input, position, position + 1);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000292 } else if (input.IsPair()) {
293 BlockRegister(input.ToLow(), position, position + 1);
294 BlockRegister(input.ToHigh(), position, position + 1);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100295 }
296 }
297
Nicolas Geoffray39468442014-09-02 15:17:15 +0100298 LiveInterval* current = instruction->GetLiveInterval();
299 if (current == nullptr) return;
300
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100301 ArenaVector<LiveInterval*>& unhandled = core_register
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100302 ? unhandled_core_intervals_
303 : unhandled_fp_intervals_;
304
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100305 DCHECK(unhandled.empty() || current->StartsBeforeOrAt(unhandled.back()));
Nicolas Geoffray87d03762014-11-19 15:17:56 +0000306
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000307 if (codegen_->NeedsTwoRegisters(current->GetType())) {
308 current->AddHighInterval();
309 }
310
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100311 for (size_t safepoint_index = safepoints_.size(); safepoint_index > 0; --safepoint_index) {
312 HInstruction* safepoint = safepoints_[safepoint_index - 1u];
Nicolas Geoffray5588e582015-04-14 14:10:59 +0100313 size_t safepoint_position = safepoint->GetLifetimePosition();
314
315 // Test that safepoints are ordered in the optimal way.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100316 DCHECK(safepoint_index == safepoints_.size() ||
317 safepoints_[safepoint_index]->GetLifetimePosition() < safepoint_position);
Nicolas Geoffray5588e582015-04-14 14:10:59 +0100318
319 if (safepoint_position == current->GetStart()) {
320 // The safepoint is for this instruction, so the location of the instruction
321 // does not need to be saved.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100322 DCHECK_EQ(safepoint_index, safepoints_.size());
Nicolas Geoffray5588e582015-04-14 14:10:59 +0100323 DCHECK_EQ(safepoint, instruction);
324 continue;
325 } else if (current->IsDeadAt(safepoint_position)) {
326 break;
327 } else if (!current->Covers(safepoint_position)) {
328 // Hole in the interval.
329 continue;
330 }
331 current->AddSafepoint(safepoint);
332 }
David Brazdil3fc992f2015-04-16 18:31:55 +0100333 current->ResetSearchCache();
Nicolas Geoffray5588e582015-04-14 14:10:59 +0100334
Nicolas Geoffray39468442014-09-02 15:17:15 +0100335 // Some instructions define their output in fixed register/stack slot. We need
336 // to ensure we know these locations before doing register allocation. For a
337 // given register, we create an interval that covers these locations. The register
338 // will be unavailable at these locations when trying to allocate one for an
339 // interval.
340 //
341 // The backwards walking ensures the ranges are ordered on increasing start positions.
342 Location output = locations->Out();
Calin Juravled0d48522014-11-04 16:40:20 +0000343 if (output.IsUnallocated() && output.GetPolicy() == Location::kSameAsFirstInput) {
344 Location first = locations->InAt(0);
345 if (first.IsRegister() || first.IsFpuRegister()) {
346 current->SetFrom(position + 1);
347 current->SetRegister(first.reg());
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000348 } else if (first.IsPair()) {
349 current->SetFrom(position + 1);
350 current->SetRegister(first.low());
351 LiveInterval* high = current->GetHighInterval();
352 high->SetRegister(first.high());
353 high->SetFrom(position + 1);
Calin Juravled0d48522014-11-04 16:40:20 +0000354 }
355 } else if (output.IsRegister() || output.IsFpuRegister()) {
Nicolas Geoffray39468442014-09-02 15:17:15 +0100356 // Shift the interval's start by one to account for the blocked register.
357 current->SetFrom(position + 1);
Nicolas Geoffray56b9ee62014-10-09 11:47:51 +0100358 current->SetRegister(output.reg());
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100359 BlockRegister(output, position, position + 1);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000360 } else if (output.IsPair()) {
361 current->SetFrom(position + 1);
362 current->SetRegister(output.low());
363 LiveInterval* high = current->GetHighInterval();
364 high->SetRegister(output.high());
365 high->SetFrom(position + 1);
366 BlockRegister(output.ToLow(), position, position + 1);
367 BlockRegister(output.ToHigh(), position, position + 1);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100368 } else if (output.IsStackSlot() || output.IsDoubleStackSlot()) {
369 current->SetSpillSlot(output.GetStackIndex());
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000370 } else {
371 DCHECK(output.IsUnallocated() || output.IsConstant());
Nicolas Geoffray39468442014-09-02 15:17:15 +0100372 }
373
David Brazdil77a48ae2015-09-15 12:34:04 +0000374 if (instruction->IsPhi() && instruction->AsPhi()->IsCatchPhi()) {
375 AllocateSpillSlotForCatchPhi(instruction->AsPhi());
376 }
377
Nicolas Geoffray39468442014-09-02 15:17:15 +0100378 // If needed, add interval to the list of unhandled intervals.
379 if (current->HasSpillSlot() || instruction->IsConstant()) {
Nicolas Geoffrayc8147a72014-10-21 16:06:20 +0100380 // Split just before first register use.
Nicolas Geoffray39468442014-09-02 15:17:15 +0100381 size_t first_register_use = current->FirstRegisterUse();
382 if (first_register_use != kNoLifetime) {
Nicolas Geoffray8cbab3c2015-04-23 15:14:36 +0100383 LiveInterval* split = SplitBetween(current, current->GetStart(), first_register_use - 1);
Nicolas Geoffrayb5f62b32014-10-30 10:58:41 +0000384 // Don't add directly to `unhandled`, it needs to be sorted and the start
Nicolas Geoffray39468442014-09-02 15:17:15 +0100385 // of this new interval might be after intervals already in the list.
386 AddSorted(&unhandled, split);
387 } else {
388 // Nothing to do, we won't allocate a register for this value.
389 }
390 } else {
Nicolas Geoffrayb5f62b32014-10-30 10:58:41 +0000391 // Don't add directly to `unhandled`, temp or safepoint intervals
392 // for this instruction may have been added, and those can be
393 // processed first.
394 AddSorted(&unhandled, current);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100395 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100396}
397
Nicolas Geoffray31d76b42014-06-09 15:02:22 +0100398class AllRangesIterator : public ValueObject {
399 public:
400 explicit AllRangesIterator(LiveInterval* interval)
401 : current_interval_(interval),
402 current_range_(interval->GetFirstRange()) {}
403
404 bool Done() const { return current_interval_ == nullptr; }
405 LiveRange* CurrentRange() const { return current_range_; }
406 LiveInterval* CurrentInterval() const { return current_interval_; }
407
408 void Advance() {
409 current_range_ = current_range_->GetNext();
410 if (current_range_ == nullptr) {
411 current_interval_ = current_interval_->GetNextSibling();
412 if (current_interval_ != nullptr) {
413 current_range_ = current_interval_->GetFirstRange();
414 }
415 }
416 }
417
418 private:
419 LiveInterval* current_interval_;
420 LiveRange* current_range_;
421
422 DISALLOW_COPY_AND_ASSIGN(AllRangesIterator);
423};
424
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700425bool RegisterAllocatorLinearScan::ValidateInternal(bool log_fatal_on_failure) const {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100426 // To simplify unit testing, we eagerly create the array of intervals, and
427 // call the helper method.
Vladimir Markof6a35de2016-03-21 12:01:50 +0000428 ArenaVector<LiveInterval*> intervals(allocator_->Adapter(kArenaAllocRegisterAllocatorValidate));
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100429 for (size_t i = 0; i < liveness_.GetNumberOfSsaValues(); ++i) {
430 HInstruction* instruction = liveness_.GetInstructionFromSsaIndex(i);
431 if (ShouldProcess(processing_core_registers_, instruction->GetLiveInterval())) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100432 intervals.push_back(instruction->GetLiveInterval());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100433 }
434 }
435
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100436 const ArenaVector<LiveInterval*>* physical_register_intervals = processing_core_registers_
437 ? &physical_core_register_intervals_
438 : &physical_fp_register_intervals_;
439 for (LiveInterval* fixed : *physical_register_intervals) {
440 if (fixed != nullptr) {
441 intervals.push_back(fixed);
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100442 }
443 }
444
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100445 for (LiveInterval* temp : temp_intervals_) {
Nicolas Geoffray39468442014-09-02 15:17:15 +0100446 if (ShouldProcess(processing_core_registers_, temp)) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100447 intervals.push_back(temp);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100448 }
449 }
450
Nicolas Geoffray776b3182015-02-23 14:14:57 +0000451 return ValidateIntervals(intervals, GetNumberOfSpillSlots(), reserved_out_slots_, *codegen_,
Nicolas Geoffray39468442014-09-02 15:17:15 +0100452 allocator_, processing_core_registers_, log_fatal_on_failure);
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100453}
454
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700455void RegisterAllocatorLinearScan::DumpInterval(std::ostream& stream, LiveInterval* interval) const {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100456 interval->Dump(stream);
457 stream << ": ";
458 if (interval->HasRegister()) {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100459 if (interval->IsFloatingPoint()) {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100460 codegen_->DumpFloatingPointRegister(stream, interval->GetRegister());
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100461 } else {
462 codegen_->DumpCoreRegister(stream, interval->GetRegister());
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100463 }
464 } else {
465 stream << "spilled";
466 }
467 stream << std::endl;
468}
469
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700470void RegisterAllocatorLinearScan::DumpAllIntervals(std::ostream& stream) const {
Mingyao Yang296bd602014-10-06 16:47:28 -0700471 stream << "inactive: " << std::endl;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100472 for (LiveInterval* inactive_interval : inactive_) {
473 DumpInterval(stream, inactive_interval);
Mingyao Yang296bd602014-10-06 16:47:28 -0700474 }
475 stream << "active: " << std::endl;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100476 for (LiveInterval* active_interval : active_) {
477 DumpInterval(stream, active_interval);
Mingyao Yang296bd602014-10-06 16:47:28 -0700478 }
479 stream << "unhandled: " << std::endl;
480 auto unhandled = (unhandled_ != nullptr) ?
481 unhandled_ : &unhandled_core_intervals_;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100482 for (LiveInterval* unhandled_interval : *unhandled) {
483 DumpInterval(stream, unhandled_interval);
Mingyao Yang296bd602014-10-06 16:47:28 -0700484 }
485 stream << "handled: " << std::endl;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100486 for (LiveInterval* handled_interval : handled_) {
487 DumpInterval(stream, handled_interval);
Mingyao Yang296bd602014-10-06 16:47:28 -0700488 }
489}
490
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100491// By the book implementation of a linear scan register allocator.
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700492void RegisterAllocatorLinearScan::LinearScan() {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100493 while (!unhandled_->empty()) {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100494 // (1) Remove interval with the lowest start position from unhandled.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100495 LiveInterval* current = unhandled_->back();
496 unhandled_->pop_back();
Nicolas Geoffray2e92bc22015-08-20 19:52:26 +0100497
498 // Make sure the interval is an expected state.
Nicolas Geoffray39468442014-09-02 15:17:15 +0100499 DCHECK(!current->IsFixed() && !current->HasSpillSlot());
Nicolas Geoffray2e92bc22015-08-20 19:52:26 +0100500 // Make sure we are going in the right order.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100501 DCHECK(unhandled_->empty() || unhandled_->back()->GetStart() >= current->GetStart());
Nicolas Geoffray2e92bc22015-08-20 19:52:26 +0100502 // Make sure a low interval is always with a high.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100503 DCHECK(!current->IsLowInterval() || unhandled_->back()->IsHighInterval());
Nicolas Geoffray2e92bc22015-08-20 19:52:26 +0100504 // Make sure a high interval is always with a low.
505 DCHECK(current->IsLowInterval() ||
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100506 unhandled_->empty() ||
507 !unhandled_->back()->IsHighInterval());
Nicolas Geoffray87d03762014-11-19 15:17:56 +0000508
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100509 size_t position = current->GetStart();
510
Mingyao Yang296bd602014-10-06 16:47:28 -0700511 // Remember the inactive_ size here since the ones moved to inactive_ from
512 // active_ below shouldn't need to be re-checked.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100513 size_t inactive_intervals_to_handle = inactive_.size();
Mingyao Yang296bd602014-10-06 16:47:28 -0700514
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100515 // (2) Remove currently active intervals that are dead at this position.
516 // Move active intervals that have a lifetime hole at this position
517 // to inactive.
Vladimir Markob95fb772015-09-30 13:32:31 +0100518 auto active_kept_end = std::remove_if(
519 active_.begin(),
520 active_.end(),
521 [this, position](LiveInterval* interval) {
522 if (interval->IsDeadAt(position)) {
523 handled_.push_back(interval);
524 return true;
525 } else if (!interval->Covers(position)) {
526 inactive_.push_back(interval);
527 return true;
528 } else {
529 return false; // Keep this interval.
530 }
531 });
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100532 active_.erase(active_kept_end, active_.end());
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100533
534 // (3) Remove currently inactive intervals that are dead at this position.
535 // Move inactive intervals that cover this position to active.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100536 auto inactive_to_handle_end = inactive_.begin() + inactive_intervals_to_handle;
Vladimir Markob95fb772015-09-30 13:32:31 +0100537 auto inactive_kept_end = std::remove_if(
538 inactive_.begin(),
539 inactive_to_handle_end,
540 [this, position](LiveInterval* interval) {
541 DCHECK(interval->GetStart() < position || interval->IsFixed());
542 if (interval->IsDeadAt(position)) {
543 handled_.push_back(interval);
544 return true;
545 } else if (interval->Covers(position)) {
546 active_.push_back(interval);
547 return true;
548 } else {
549 return false; // Keep this interval.
550 }
551 });
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100552 inactive_.erase(inactive_kept_end, inactive_to_handle_end);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100553
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000554 if (current->IsHighInterval() && !current->GetLowInterval()->HasRegister()) {
555 DCHECK(!current->HasRegister());
556 // Allocating the low part was unsucessful. The splitted interval for the high part
557 // will be handled next (it is in the `unhandled_` list).
558 continue;
559 }
560
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100561 // (4) Try to find an available register.
562 bool success = TryAllocateFreeReg(current);
563
564 // (5) If no register could be found, we need to spill.
565 if (!success) {
566 success = AllocateBlockedReg(current);
567 }
568
569 // (6) If the interval had a register allocated, add it to the list of active
570 // intervals.
571 if (success) {
Nicolas Geoffray98893962015-01-21 12:32:32 +0000572 codegen_->AddAllocatedRegister(processing_core_registers_
573 ? Location::RegisterLocation(current->GetRegister())
574 : Location::FpuRegisterLocation(current->GetRegister()));
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100575 active_.push_back(current);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000576 if (current->HasHighInterval() && !current->GetHighInterval()->HasRegister()) {
577 current->GetHighInterval()->SetRegister(GetHighForLowRegister(current->GetRegister()));
578 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100579 }
580 }
581}
582
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000583static void FreeIfNotCoverAt(LiveInterval* interval, size_t position, size_t* free_until) {
584 DCHECK(!interval->IsHighInterval());
585 // Note that the same instruction may occur multiple times in the input list,
586 // so `free_until` may have changed already.
David Brazdil3fc992f2015-04-16 18:31:55 +0100587 // Since `position` is not the current scan position, we need to use CoversSlow.
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000588 if (interval->IsDeadAt(position)) {
589 // Set the register to be free. Note that inactive intervals might later
590 // update this.
591 free_until[interval->GetRegister()] = kMaxLifetimePosition;
592 if (interval->HasHighInterval()) {
593 DCHECK(interval->GetHighInterval()->IsDeadAt(position));
594 free_until[interval->GetHighInterval()->GetRegister()] = kMaxLifetimePosition;
595 }
David Brazdil3fc992f2015-04-16 18:31:55 +0100596 } else if (!interval->CoversSlow(position)) {
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000597 // The interval becomes inactive at `defined_by`. We make its register
598 // available only until the next use strictly after `defined_by`.
599 free_until[interval->GetRegister()] = interval->FirstUseAfter(position);
600 if (interval->HasHighInterval()) {
David Brazdil3fc992f2015-04-16 18:31:55 +0100601 DCHECK(!interval->GetHighInterval()->CoversSlow(position));
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000602 free_until[interval->GetHighInterval()->GetRegister()] = free_until[interval->GetRegister()];
603 }
604 }
605}
606
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100607// Find a free register. If multiple are found, pick the register that
608// is free the longest.
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700609bool RegisterAllocatorLinearScan::TryAllocateFreeReg(LiveInterval* current) {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100610 size_t* free_until = registers_array_;
611
612 // First set all registers to be free.
613 for (size_t i = 0; i < number_of_registers_; ++i) {
614 free_until[i] = kMaxLifetimePosition;
615 }
616
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100617 // For each active interval, set its register to not free.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100618 for (LiveInterval* interval : active_) {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100619 DCHECK(interval->HasRegister());
620 free_until[interval->GetRegister()] = 0;
621 }
622
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000623 // An interval that starts an instruction (that is, it is not split), may
624 // re-use the registers used by the inputs of that instruciton, based on the
625 // location summary.
626 HInstruction* defined_by = current->GetDefinedBy();
627 if (defined_by != nullptr && !current->IsSplit()) {
628 LocationSummary* locations = defined_by->GetLocations();
629 if (!locations->OutputCanOverlapWithInputs() && locations->Out().IsUnallocated()) {
Vladimir Markoe9004912016-06-16 16:50:52 +0100630 HInputsRef inputs = defined_by->GetInputs();
Vladimir Marko372f10e2016-05-17 16:30:10 +0100631 for (size_t i = 0; i < inputs.size(); ++i) {
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000632 // Take the last interval of the input. It is the location of that interval
633 // that will be used at `defined_by`.
Vladimir Marko372f10e2016-05-17 16:30:10 +0100634 LiveInterval* interval = inputs[i]->GetLiveInterval()->GetLastSibling();
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000635 // Note that interval may have not been processed yet.
636 // TODO: Handle non-split intervals last in the work list.
Nicolas Geoffray94015b92015-06-04 18:21:04 +0100637 if (locations->InAt(i).IsValid()
638 && interval->HasRegister()
639 && interval->SameRegisterKind(*current)) {
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000640 // The input must be live until the end of `defined_by`, to comply to
641 // the linear scan algorithm. So we use `defined_by`'s end lifetime
642 // position to check whether the input is dead or is inactive after
643 // `defined_by`.
David Brazdil3fc992f2015-04-16 18:31:55 +0100644 DCHECK(interval->CoversSlow(defined_by->GetLifetimePosition()));
Nicolas Geoffray829280c2015-01-28 10:20:37 +0000645 size_t position = defined_by->GetLifetimePosition() + 1;
646 FreeIfNotCoverAt(interval, position, free_until);
647 }
648 }
649 }
650 }
651
Mingyao Yang296bd602014-10-06 16:47:28 -0700652 // For each inactive interval, set its register to be free until
653 // the next intersection with `current`.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100654 for (LiveInterval* inactive : inactive_) {
Mingyao Yang296bd602014-10-06 16:47:28 -0700655 // Temp/Slow-path-safepoint interval has no holes.
Vladimir Marko70e97462016-08-09 11:04:26 +0100656 DCHECK(!inactive->IsTemp());
Mingyao Yang296bd602014-10-06 16:47:28 -0700657 if (!current->IsSplit() && !inactive->IsFixed()) {
658 // Neither current nor inactive are fixed.
659 // Thanks to SSA, a non-split interval starting in a hole of an
660 // inactive interval should never intersect with that inactive interval.
661 // Only if it's not fixed though, because fixed intervals don't come from SSA.
662 DCHECK_EQ(inactive->FirstIntersectionWith(current), kNoLifetime);
663 continue;
664 }
665
666 DCHECK(inactive->HasRegister());
667 if (free_until[inactive->GetRegister()] == 0) {
668 // Already used by some active interval. No need to intersect.
669 continue;
670 }
671 size_t next_intersection = inactive->FirstIntersectionWith(current);
672 if (next_intersection != kNoLifetime) {
673 free_until[inactive->GetRegister()] =
674 std::min(free_until[inactive->GetRegister()], next_intersection);
675 }
676 }
677
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000678 int reg = kNoRegister;
Nicolas Geoffray39468442014-09-02 15:17:15 +0100679 if (current->HasRegister()) {
680 // Some instructions have a fixed register output.
681 reg = current->GetRegister();
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000682 if (free_until[reg] == 0) {
683 DCHECK(current->IsHighInterval());
684 // AllocateBlockedReg will spill the holder of the register.
685 return false;
686 }
Nicolas Geoffray39468442014-09-02 15:17:15 +0100687 } else {
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000688 DCHECK(!current->IsHighInterval());
Nicolas Geoffrayfbda5f32015-04-29 14:16:00 +0100689 int hint = current->FindFirstRegisterHint(free_until, liveness_);
Nicolas Geoffrayf2975812015-08-07 18:13:03 -0700690 if ((hint != kNoRegister)
691 // For simplicity, if the hint we are getting for a pair cannot be used,
692 // we are just going to allocate a new pair.
693 && !(current->IsLowInterval() && IsBlocked(GetHighForLowRegister(hint)))) {
Nicolas Geoffray01ef3452014-10-01 11:32:17 +0100694 DCHECK(!IsBlocked(hint));
695 reg = hint;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000696 } else if (current->IsLowInterval()) {
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000697 reg = FindAvailableRegisterPair(free_until, current->GetStart());
Nicolas Geoffray01ef3452014-10-01 11:32:17 +0100698 } else {
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100699 reg = FindAvailableRegister(free_until, current);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100700 }
701 }
702
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000703 DCHECK_NE(reg, kNoRegister);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100704 // If we could not find a register, we need to spill.
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000705 if (free_until[reg] == 0) {
706 return false;
707 }
708
Nicolas Geoffray234d69d2015-03-09 10:28:50 +0000709 if (current->IsLowInterval()) {
710 // If the high register of this interval is not available, we need to spill.
711 int high_reg = current->GetHighInterval()->GetRegister();
712 if (high_reg == kNoRegister) {
713 high_reg = GetHighForLowRegister(reg);
714 }
715 if (free_until[high_reg] == 0) {
716 return false;
717 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100718 }
719
720 current->SetRegister(reg);
721 if (!current->IsDeadAt(free_until[reg])) {
722 // If the register is only available for a subset of live ranges
Nicolas Geoffray82726882015-06-01 13:51:57 +0100723 // covered by `current`, split `current` before the position where
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100724 // the register is not available anymore.
Nicolas Geoffray82726882015-06-01 13:51:57 +0100725 LiveInterval* split = SplitBetween(current, current->GetStart(), free_until[reg]);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100726 DCHECK(split != nullptr);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100727 AddSorted(unhandled_, split);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100728 }
729 return true;
730}
731
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700732bool RegisterAllocatorLinearScan::IsBlocked(int reg) const {
Nicolas Geoffray102cbed2014-10-15 18:31:05 +0100733 return processing_core_registers_
734 ? blocked_core_registers_[reg]
735 : blocked_fp_registers_[reg];
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100736}
737
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700738int RegisterAllocatorLinearScan::FindAvailableRegisterPair(size_t* next_use, size_t starting_at) const {
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000739 int reg = kNoRegister;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000740 // Pick the register pair that is used the last.
741 for (size_t i = 0; i < number_of_registers_; ++i) {
742 if (IsBlocked(i)) continue;
743 if (!IsLowRegister(i)) continue;
744 int high_register = GetHighForLowRegister(i);
745 if (IsBlocked(high_register)) continue;
746 int existing_high_register = GetHighForLowRegister(reg);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000747 if ((reg == kNoRegister) || (next_use[i] >= next_use[reg]
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000748 && next_use[high_register] >= next_use[existing_high_register])) {
749 reg = i;
750 if (next_use[i] == kMaxLifetimePosition
751 && next_use[high_register] == kMaxLifetimePosition) {
752 break;
753 }
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000754 } else if (next_use[reg] <= starting_at || next_use[existing_high_register] <= starting_at) {
755 // If one of the current register is known to be unavailable, just unconditionally
756 // try a new one.
757 reg = i;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000758 }
759 }
760 return reg;
761}
762
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700763bool RegisterAllocatorLinearScan::IsCallerSaveRegister(int reg) const {
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100764 return processing_core_registers_
765 ? !codegen_->IsCoreCalleeSaveRegister(reg)
766 : !codegen_->IsFloatingPointCalleeSaveRegister(reg);
767}
768
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700769int RegisterAllocatorLinearScan::FindAvailableRegister(size_t* next_use, LiveInterval* current) const {
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100770 // We special case intervals that do not span a safepoint to try to find a caller-save
771 // register if one is available. We iterate from 0 to the number of registers,
772 // so if there are caller-save registers available at the end, we continue the iteration.
773 bool prefers_caller_save = !current->HasWillCallSafepoint();
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000774 int reg = kNoRegister;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000775 for (size_t i = 0; i < number_of_registers_; ++i) {
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100776 if (IsBlocked(i)) {
777 // Register cannot be used. Continue.
778 continue;
779 }
780
781 // Best case: we found a register fully available.
782 if (next_use[i] == kMaxLifetimePosition) {
783 if (prefers_caller_save && !IsCallerSaveRegister(i)) {
784 // We can get shorter encodings on some platforms by using
785 // small register numbers. So only update the candidate if the previous
786 // one was not available for the whole method.
787 if (reg == kNoRegister || next_use[reg] != kMaxLifetimePosition) {
788 reg = i;
789 }
790 // Continue the iteration in the hope of finding a caller save register.
791 continue;
792 } else {
793 reg = i;
794 // We know the register is good enough. Return it.
795 break;
796 }
797 }
798
799 // If we had no register before, take this one as a reference.
800 if (reg == kNoRegister) {
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000801 reg = i;
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100802 continue;
803 }
804
805 // Pick the register that is used the last.
806 if (next_use[i] > next_use[reg]) {
807 reg = i;
808 continue;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000809 }
810 }
811 return reg;
812}
813
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100814// Remove interval and its other half if any. Return iterator to the following element.
815static ArenaVector<LiveInterval*>::iterator RemoveIntervalAndPotentialOtherHalf(
816 ArenaVector<LiveInterval*>* intervals, ArenaVector<LiveInterval*>::iterator pos) {
817 DCHECK(intervals->begin() <= pos && pos < intervals->end());
818 LiveInterval* interval = *pos;
819 if (interval->IsLowInterval()) {
820 DCHECK(pos + 1 < intervals->end());
821 DCHECK_EQ(*(pos + 1), interval->GetHighInterval());
822 return intervals->erase(pos, pos + 2);
823 } else if (interval->IsHighInterval()) {
824 DCHECK(intervals->begin() < pos);
825 DCHECK_EQ(*(pos - 1), interval->GetLowInterval());
826 return intervals->erase(pos - 1, pos + 1);
827 } else {
828 return intervals->erase(pos);
829 }
830}
831
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700832bool RegisterAllocatorLinearScan::TrySplitNonPairOrUnalignedPairIntervalAt(size_t position,
833 size_t first_register_use,
834 size_t* next_use) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100835 for (auto it = active_.begin(), end = active_.end(); it != end; ++it) {
836 LiveInterval* active = *it;
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000837 DCHECK(active->HasRegister());
Nicolas Geoffray234d69d2015-03-09 10:28:50 +0000838 if (active->IsFixed()) continue;
839 if (active->IsHighInterval()) continue;
840 if (first_register_use > next_use[active->GetRegister()]) continue;
841
Nicolas Geoffray2e92bc22015-08-20 19:52:26 +0100842 // Split the first interval found that is either:
843 // 1) A non-pair interval.
844 // 2) A pair interval whose high is not low + 1.
845 // 3) A pair interval whose low is not even.
846 if (!active->IsLowInterval() ||
847 IsLowOfUnalignedPairInterval(active) ||
848 !IsLowRegister(active->GetRegister())) {
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000849 LiveInterval* split = Split(active, position);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000850 if (split != active) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100851 handled_.push_back(active);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000852 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100853 RemoveIntervalAndPotentialOtherHalf(&active_, it);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000854 AddSorted(unhandled_, split);
855 return true;
856 }
857 }
858 return false;
859}
860
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100861// Find the register that is used the last, and spill the interval
862// that holds it. If the first use of `current` is after that register
863// we spill `current` instead.
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -0700864bool RegisterAllocatorLinearScan::AllocateBlockedReg(LiveInterval* current) {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100865 size_t first_register_use = current->FirstRegisterUse();
Nicolas Geoffrayda2b2542015-08-06 19:56:45 -0700866 if (current->HasRegister()) {
867 DCHECK(current->IsHighInterval());
868 // The low interval has allocated the register for the high interval. In
869 // case the low interval had to split both intervals, we may end up in a
870 // situation where the high interval does not have a register use anymore.
871 // We must still proceed in order to split currently active and inactive
872 // uses of the high interval's register, and put the high interval in the
873 // active set.
874 DCHECK(first_register_use != kNoLifetime || (current->GetNextSibling() != nullptr));
875 } else if (first_register_use == kNoLifetime) {
Nicolas Geoffray31d76b42014-06-09 15:02:22 +0100876 AllocateSpillSlotFor(current);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100877 return false;
878 }
879
880 // First set all registers as not being used.
881 size_t* next_use = registers_array_;
882 for (size_t i = 0; i < number_of_registers_; ++i) {
883 next_use[i] = kMaxLifetimePosition;
884 }
885
886 // For each active interval, find the next use of its register after the
887 // start of current.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100888 for (LiveInterval* active : active_) {
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100889 DCHECK(active->HasRegister());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100890 if (active->IsFixed()) {
891 next_use[active->GetRegister()] = current->GetStart();
892 } else {
Nicolas Geoffray119a8852016-02-06 17:01:15 +0000893 size_t use = active->FirstRegisterUseAfter(current->GetStart());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100894 if (use != kNoLifetime) {
895 next_use[active->GetRegister()] = use;
896 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100897 }
898 }
899
900 // For each inactive interval, find the next use of its register after the
901 // start of current.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100902 for (LiveInterval* inactive : inactive_) {
Mingyao Yang296bd602014-10-06 16:47:28 -0700903 // Temp/Slow-path-safepoint interval has no holes.
Vladimir Marko70e97462016-08-09 11:04:26 +0100904 DCHECK(!inactive->IsTemp());
Mingyao Yang296bd602014-10-06 16:47:28 -0700905 if (!current->IsSplit() && !inactive->IsFixed()) {
906 // Neither current nor inactive are fixed.
907 // Thanks to SSA, a non-split interval starting in a hole of an
908 // inactive interval should never intersect with that inactive interval.
909 // Only if it's not fixed though, because fixed intervals don't come from SSA.
910 DCHECK_EQ(inactive->FirstIntersectionWith(current), kNoLifetime);
911 continue;
912 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100913 DCHECK(inactive->HasRegister());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100914 size_t next_intersection = inactive->FirstIntersectionWith(current);
915 if (next_intersection != kNoLifetime) {
916 if (inactive->IsFixed()) {
917 next_use[inactive->GetRegister()] =
918 std::min(next_intersection, next_use[inactive->GetRegister()]);
919 } else {
Nicolas Geoffray1ba19812015-04-21 09:12:40 +0100920 size_t use = inactive->FirstUseAfter(current->GetStart());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100921 if (use != kNoLifetime) {
922 next_use[inactive->GetRegister()] = std::min(use, next_use[inactive->GetRegister()]);
923 }
924 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100925 }
926 }
927
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000928 int reg = kNoRegister;
929 bool should_spill = false;
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000930 if (current->HasRegister()) {
931 DCHECK(current->IsHighInterval());
932 reg = current->GetRegister();
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000933 // When allocating the low part, we made sure the high register was available.
Nicolas Geoffray119a8852016-02-06 17:01:15 +0000934 DCHECK_LT(first_register_use, next_use[reg]);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000935 } else if (current->IsLowInterval()) {
Nicolas Geoffray119a8852016-02-06 17:01:15 +0000936 reg = FindAvailableRegisterPair(next_use, first_register_use);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000937 // We should spill if both registers are not available.
Nicolas Geoffray119a8852016-02-06 17:01:15 +0000938 should_spill = (first_register_use >= next_use[reg])
939 || (first_register_use >= next_use[GetHighForLowRegister(reg)]);
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000940 } else {
941 DCHECK(!current->IsHighInterval());
Nicolas Geoffray8826f672015-04-17 09:15:11 +0100942 reg = FindAvailableRegister(next_use, current);
Nicolas Geoffray119a8852016-02-06 17:01:15 +0000943 should_spill = (first_register_use >= next_use[reg]);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100944 }
945
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000946 DCHECK_NE(reg, kNoRegister);
947 if (should_spill) {
Nicolas Geoffray840e5462015-01-07 16:01:24 +0000948 DCHECK(!current->IsHighInterval());
Nicolas Geoffray234d69d2015-03-09 10:28:50 +0000949 bool is_allocation_at_use_site = (current->GetStart() >= (first_register_use - 1));
Nicolas Geoffrayda2b2542015-08-06 19:56:45 -0700950 if (is_allocation_at_use_site) {
951 if (!current->IsLowInterval()) {
952 DumpInterval(std::cerr, current);
953 DumpAllIntervals(std::cerr);
954 // This situation has the potential to infinite loop, so we make it a non-debug CHECK.
955 HInstruction* at = liveness_.GetInstructionFromPosition(first_register_use / 2);
956 CHECK(false) << "There is not enough registers available for "
957 << current->GetParent()->GetDefinedBy()->DebugName() << " "
958 << current->GetParent()->GetDefinedBy()->GetId()
959 << " at " << first_register_use - 1 << " "
960 << (at == nullptr ? "" : at->DebugName());
961 }
962
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000963 // If we're allocating a register for `current` because the instruction at
964 // that position requires it, but we think we should spill, then there are
Nicolas Geoffray234d69d2015-03-09 10:28:50 +0000965 // non-pair intervals or unaligned pair intervals blocking the allocation.
966 // We split the first interval found, and put ourselves first in the
967 // `unhandled_` list.
Nicolas Geoffrayda2b2542015-08-06 19:56:45 -0700968 bool success = TrySplitNonPairOrUnalignedPairIntervalAt(current->GetStart(),
969 first_register_use,
970 next_use);
971 DCHECK(success);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100972 LiveInterval* existing = unhandled_->back();
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000973 DCHECK(existing->IsHighInterval());
974 DCHECK_EQ(existing->GetLowInterval(), current);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100975 unhandled_->push_back(current);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000976 } else {
977 // If the first use of that instruction is after the last use of the found
978 // register, we split this interval just before its first register use.
979 AllocateSpillSlotFor(current);
Nicolas Geoffray8cbab3c2015-04-23 15:14:36 +0100980 LiveInterval* split = SplitBetween(current, current->GetStart(), first_register_use - 1);
Nicolas Geoffrayda2b2542015-08-06 19:56:45 -0700981 DCHECK(current != split);
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +0000982 AddSorted(unhandled_, split);
983 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100984 return false;
985 } else {
986 // Use this register and spill the active and inactives interval that
987 // have that register.
988 current->SetRegister(reg);
989
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100990 for (auto it = active_.begin(), end = active_.end(); it != end; ++it) {
991 LiveInterval* active = *it;
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100992 if (active->GetRegister() == reg) {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +0100993 DCHECK(!active->IsFixed());
Nicolas Geoffraya7062e02014-05-22 12:50:17 +0100994 LiveInterval* split = Split(active, current->GetStart());
Nicolas Geoffraydd8f8872015-01-15 15:37:37 +0000995 if (split != active) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100996 handled_.push_back(active);
Nicolas Geoffraydd8f8872015-01-15 15:37:37 +0000997 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +0100998 RemoveIntervalAndPotentialOtherHalf(&active_, it);
Nicolas Geoffray39468442014-09-02 15:17:15 +0100999 AddSorted(unhandled_, split);
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001000 break;
1001 }
1002 }
1003
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001004 // NOTE: Retrieve end() on each iteration because we're removing elements in the loop body.
1005 for (auto it = inactive_.begin(); it != inactive_.end(); ) {
1006 LiveInterval* inactive = *it;
1007 bool erased = false;
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001008 if (inactive->GetRegister() == reg) {
Mingyao Yang296bd602014-10-06 16:47:28 -07001009 if (!current->IsSplit() && !inactive->IsFixed()) {
1010 // Neither current nor inactive are fixed.
1011 // Thanks to SSA, a non-split interval starting in a hole of an
1012 // inactive interval should never intersect with that inactive interval.
1013 // Only if it's not fixed though, because fixed intervals don't come from SSA.
1014 DCHECK_EQ(inactive->FirstIntersectionWith(current), kNoLifetime);
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001015 } else {
1016 size_t next_intersection = inactive->FirstIntersectionWith(current);
1017 if (next_intersection != kNoLifetime) {
1018 if (inactive->IsFixed()) {
1019 LiveInterval* split = Split(current, next_intersection);
1020 DCHECK_NE(split, current);
1021 AddSorted(unhandled_, split);
1022 } else {
1023 // Split at the start of `current`, which will lead to splitting
1024 // at the end of the lifetime hole of `inactive`.
1025 LiveInterval* split = Split(inactive, current->GetStart());
1026 // If it's inactive, it must start before the current interval.
1027 DCHECK_NE(split, inactive);
1028 it = RemoveIntervalAndPotentialOtherHalf(&inactive_, it);
1029 erased = true;
1030 handled_.push_back(inactive);
1031 AddSorted(unhandled_, split);
Nicolas Geoffray5b168de2015-03-27 10:27:22 +00001032 }
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001033 }
1034 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001035 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001036 // If we have erased the element, `it` already points to the next element.
1037 // Otherwise we need to move to the next element.
1038 if (!erased) {
1039 ++it;
1040 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001041 }
1042
1043 return true;
1044 }
1045}
1046
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -07001047void RegisterAllocatorLinearScan::AddSorted(ArenaVector<LiveInterval*>* array, LiveInterval* interval) {
Nicolas Geoffrayc8147a72014-10-21 16:06:20 +01001048 DCHECK(!interval->IsFixed() && !interval->HasSpillSlot());
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001049 size_t insert_at = 0;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001050 for (size_t i = array->size(); i > 0; --i) {
1051 LiveInterval* current = (*array)[i - 1u];
Nicolas Geoffray6c2dff82015-01-21 14:56:54 +00001052 // High intervals must be processed right after their low equivalent.
1053 if (current->StartsAfter(interval) && !current->IsHighInterval()) {
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001054 insert_at = i;
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001055 break;
1056 }
1057 }
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001058
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001059 // Insert the high interval before the low, to ensure the low is processed before.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001060 auto insert_pos = array->begin() + insert_at;
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001061 if (interval->HasHighInterval()) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001062 array->insert(insert_pos, { interval->GetHighInterval(), interval });
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001063 } else if (interval->HasLowInterval()) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001064 array->insert(insert_pos, { interval, interval->GetLowInterval() });
1065 } else {
1066 array->insert(insert_pos, interval);
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001067 }
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001068}
1069
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -07001070void RegisterAllocatorLinearScan::AllocateSpillSlotFor(LiveInterval* interval) {
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001071 if (interval->IsHighInterval()) {
Nicolas Geoffrayda2b2542015-08-06 19:56:45 -07001072 // The low interval already took care of allocating the spill slot.
1073 DCHECK(!interval->GetLowInterval()->HasRegister());
1074 DCHECK(interval->GetLowInterval()->GetParent()->HasSpillSlot());
Nicolas Geoffray840e5462015-01-07 16:01:24 +00001075 return;
1076 }
1077
Nicolas Geoffray31d76b42014-06-09 15:02:22 +01001078 LiveInterval* parent = interval->GetParent();
1079
1080 // An instruction gets a spill slot for its entire lifetime. If the parent
1081 // of this interval already has a spill slot, there is nothing to do.
1082 if (parent->HasSpillSlot()) {
1083 return;
1084 }
1085
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001086 HInstruction* defined_by = parent->GetDefinedBy();
David Brazdil77a48ae2015-09-15 12:34:04 +00001087 DCHECK(!defined_by->IsPhi() || !defined_by->AsPhi()->IsCatchPhi());
1088
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001089 if (defined_by->IsParameterValue()) {
1090 // Parameters have their own stack slot.
1091 parent->SetSpillSlot(codegen_->GetStackSlotOfParameter(defined_by->AsParameterValue()));
1092 return;
1093 }
1094
Nicolas Geoffray76b1e172015-05-27 17:18:33 +01001095 if (defined_by->IsCurrentMethod()) {
1096 parent->SetSpillSlot(0);
1097 return;
1098 }
1099
Nicolas Geoffray96f89a22014-07-11 10:57:49 +01001100 if (defined_by->IsConstant()) {
1101 // Constants don't need a spill slot.
1102 return;
1103 }
1104
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001105 ArenaVector<size_t>* spill_slots = nullptr;
Nicolas Geoffray776b3182015-02-23 14:14:57 +00001106 switch (interval->GetType()) {
1107 case Primitive::kPrimDouble:
1108 spill_slots = &double_spill_slots_;
1109 break;
1110 case Primitive::kPrimLong:
1111 spill_slots = &long_spill_slots_;
1112 break;
1113 case Primitive::kPrimFloat:
1114 spill_slots = &float_spill_slots_;
1115 break;
1116 case Primitive::kPrimNot:
1117 case Primitive::kPrimInt:
1118 case Primitive::kPrimChar:
1119 case Primitive::kPrimByte:
1120 case Primitive::kPrimBoolean:
1121 case Primitive::kPrimShort:
1122 spill_slots = &int_spill_slots_;
1123 break;
1124 case Primitive::kPrimVoid:
1125 LOG(FATAL) << "Unexpected type for interval " << interval->GetType();
1126 }
1127
Nicolas Geoffray412f10c2014-06-19 10:00:34 +01001128 // Find an available spill slot.
1129 size_t slot = 0;
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001130 for (size_t e = spill_slots->size(); slot < e; ++slot) {
Matthew Gharrityf64a6ab2016-07-11 14:45:01 -07001131 if ((*spill_slots)[slot] <= parent->GetStart()) {
1132 if (!parent->NeedsTwoSpillSlots()) {
1133 // One spill slot is sufficient.
1134 break;
1135 }
1136 if (slot == e - 1 || (*spill_slots)[slot + 1] <= parent->GetStart()) {
1137 // Two spill slots are available.
1138 break;
1139 }
Nicolas Geoffray412f10c2014-06-19 10:00:34 +01001140 }
1141 }
1142
David Brazdil77a48ae2015-09-15 12:34:04 +00001143 size_t end = interval->GetLastSibling()->GetEnd();
Nicolas Geoffray01ef3452014-10-01 11:32:17 +01001144 if (parent->NeedsTwoSpillSlots()) {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001145 if (slot + 2u > spill_slots->size()) {
Nicolas Geoffray3c049742014-09-24 18:10:46 +01001146 // We need a new spill slot.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001147 spill_slots->resize(slot + 2u, end);
Nicolas Geoffray31d76b42014-06-09 15:02:22 +01001148 }
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001149 (*spill_slots)[slot] = end;
1150 (*spill_slots)[slot + 1] = end;
Nicolas Geoffray31d76b42014-06-09 15:02:22 +01001151 } else {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001152 if (slot == spill_slots->size()) {
Nicolas Geoffray3c049742014-09-24 18:10:46 +01001153 // We need a new spill slot.
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001154 spill_slots->push_back(end);
Nicolas Geoffray3c049742014-09-24 18:10:46 +01001155 } else {
Vladimir Marko2aaa4b52015-09-17 17:03:26 +01001156 (*spill_slots)[slot] = end;
Nicolas Geoffray3c049742014-09-24 18:10:46 +01001157 }
Nicolas Geoffray31d76b42014-06-09 15:02:22 +01001158 }
1159
Nicolas Geoffray776b3182015-02-23 14:14:57 +00001160 // Note that the exact spill slot location will be computed when we resolve,
1161 // that is when we know the number of spill slots for each type.
1162 parent->SetSpillSlot(slot);
Nicolas Geoffray86dbb9a2014-06-04 11:12:39 +01001163}
1164
Matthew Gharrity8f49d4b2016-07-14 13:24:00 -07001165void RegisterAllocatorLinearScan::AllocateSpillSlotForCatchPhi(HPhi* phi) {
David Brazdil77a48ae2015-09-15 12:34:04 +00001166 LiveInterval* interval = phi->GetLiveInterval();
1167
1168 HInstruction* previous_phi = phi->GetPrevious();
1169 DCHECK(previous_phi == nullptr ||
1170 previous_phi->AsPhi()->GetRegNumber() <= phi->GetRegNumber())
1171 << "Phis expected to be sorted by vreg number, so that equivalent phis are adjacent.";
1172
1173 if (phi->IsVRegEquivalentOf(previous_phi)) {
1174 // This is an equivalent of the previous phi. We need to assign the same
1175 // catch phi slot.
1176 DCHECK(previous_phi->GetLiveInterval()->HasSpillSlot());
1177 interval->SetSpillSlot(previous_phi->GetLiveInterval()->GetSpillSlot());
1178 } else {
1179 // Allocate a new spill slot for this catch phi.
1180 // TODO: Reuse spill slots when intervals of phis from different catch
1181 // blocks do not overlap.
1182 interval->SetSpillSlot(catch_phi_spill_slots_);
1183 catch_phi_spill_slots_ += interval->NeedsTwoSpillSlots() ? 2 : 1;
1184 }
1185}
1186
Nicolas Geoffraya7062e02014-05-22 12:50:17 +01001187} // namespace art