| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| /* This file contains register alloction support. */ |
| |
| #include "dex/compiler_ir.h" |
| #include "dex/compiler_internals.h" |
| #include "mir_to_lir-inl.h" |
| |
| namespace art { |
| |
| /* |
| * Free all allocated temps in the temp pools. Note that this does |
| * not affect the "liveness" of a temp register, which will stay |
| * live until it is either explicitly killed or reallocated. |
| */ |
| void Mir2Lir::ResetRegPool() { |
| GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_); |
| for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) { |
| info->MarkFree(); |
| } |
| // Reset temp tracking sanity check. |
| if (kIsDebugBuild) { |
| live_sreg_ = INVALID_SREG; |
| } |
| } |
| |
| Mir2Lir::RegisterInfo::RegisterInfo(RegStorage r, const ResourceMask& mask) |
| : reg_(r), is_temp_(false), wide_value_(false), dirty_(false), aliased_(false), partner_(r), |
| s_reg_(INVALID_SREG), def_use_mask_(mask), master_(this), def_start_(nullptr), |
| def_end_(nullptr), alias_chain_(nullptr) { |
| switch (r.StorageSize()) { |
| case 0: storage_mask_ = 0xffffffff; break; |
| case 4: storage_mask_ = 0x00000001; break; |
| case 8: storage_mask_ = 0x00000003; break; |
| case 16: storage_mask_ = 0x0000000f; break; |
| case 32: storage_mask_ = 0x000000ff; break; |
| case 64: storage_mask_ = 0x0000ffff; break; |
| case 128: storage_mask_ = 0xffffffff; break; |
| } |
| used_storage_ = r.Valid() ? ~storage_mask_ : storage_mask_; |
| liveness_ = used_storage_; |
| } |
| |
| Mir2Lir::RegisterPool::RegisterPool(Mir2Lir* m2l, ArenaAllocator* arena, |
| const ArrayRef<const RegStorage>& core_regs, |
| const ArrayRef<const RegStorage>& core64_regs, |
| const ArrayRef<const RegStorage>& sp_regs, |
| const ArrayRef<const RegStorage>& dp_regs, |
| const ArrayRef<const RegStorage>& reserved_regs, |
| const ArrayRef<const RegStorage>& reserved64_regs, |
| const ArrayRef<const RegStorage>& core_temps, |
| const ArrayRef<const RegStorage>& core64_temps, |
| const ArrayRef<const RegStorage>& sp_temps, |
| const ArrayRef<const RegStorage>& dp_temps) : |
| core_regs_(arena, core_regs.size()), next_core_reg_(0), |
| core64_regs_(arena, core64_regs.size()), next_core64_reg_(0), |
| sp_regs_(arena, sp_regs.size()), next_sp_reg_(0), |
| dp_regs_(arena, dp_regs.size()), next_dp_reg_(0), m2l_(m2l) { |
| // Initialize the fast lookup map. |
| m2l_->reginfo_map_.Reset(); |
| if (kIsDebugBuild) { |
| m2l_->reginfo_map_.Resize(RegStorage::kMaxRegs); |
| for (unsigned i = 0; i < RegStorage::kMaxRegs; i++) { |
| m2l_->reginfo_map_.Insert(nullptr); |
| } |
| } else { |
| m2l_->reginfo_map_.SetSize(RegStorage::kMaxRegs); |
| } |
| |
| // Construct the register pool. |
| for (const RegStorage& reg : core_regs) { |
| RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg)); |
| m2l_->reginfo_map_.Put(reg.GetReg(), info); |
| core_regs_.Insert(info); |
| } |
| for (const RegStorage& reg : core64_regs) { |
| RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg)); |
| m2l_->reginfo_map_.Put(reg.GetReg(), info); |
| core64_regs_.Insert(info); |
| } |
| for (const RegStorage& reg : sp_regs) { |
| RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg)); |
| m2l_->reginfo_map_.Put(reg.GetReg(), info); |
| sp_regs_.Insert(info); |
| } |
| for (const RegStorage& reg : dp_regs) { |
| RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg)); |
| m2l_->reginfo_map_.Put(reg.GetReg(), info); |
| dp_regs_.Insert(info); |
| } |
| |
| // Keep special registers from being allocated. |
| for (RegStorage reg : reserved_regs) { |
| m2l_->MarkInUse(reg); |
| } |
| for (RegStorage reg : reserved64_regs) { |
| m2l_->MarkInUse(reg); |
| } |
| |
| // Mark temp regs - all others not in use can be used for promotion |
| for (RegStorage reg : core_temps) { |
| m2l_->MarkTemp(reg); |
| } |
| for (RegStorage reg : core64_temps) { |
| m2l_->MarkTemp(reg); |
| } |
| for (RegStorage reg : sp_temps) { |
| m2l_->MarkTemp(reg); |
| } |
| for (RegStorage reg : dp_temps) { |
| m2l_->MarkTemp(reg); |
| } |
| |
| // Add an entry for InvalidReg with zero'd mask. |
| RegisterInfo* invalid_reg = new (arena) RegisterInfo(RegStorage::InvalidReg(), kEncodeNone); |
| m2l_->reginfo_map_.Put(RegStorage::InvalidReg().GetReg(), invalid_reg); |
| |
| // Existence of core64 registers implies wide references. |
| if (core64_regs_.Size() != 0) { |
| ref_regs_ = &core64_regs_; |
| next_ref_reg_ = &next_core64_reg_; |
| } else { |
| ref_regs_ = &core_regs_; |
| next_ref_reg_ = &next_core_reg_; |
| } |
| } |
| |
| void Mir2Lir::DumpRegPool(GrowableArray<RegisterInfo*>* regs) { |
| LOG(INFO) << "================================================"; |
| GrowableArray<RegisterInfo*>::Iterator it(regs); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| LOG(INFO) << StringPrintf( |
| "R[%d:%d:%c]: T:%d, U:%d, W:%d, p:%d, LV:%d, D:%d, SR:%d, DEF:%d", |
| info->GetReg().GetReg(), info->GetReg().GetRegNum(), info->GetReg().IsFloat() ? 'f' : 'c', |
| info->IsTemp(), info->InUse(), info->IsWide(), info->Partner().GetReg(), info->IsLive(), |
| info->IsDirty(), info->SReg(), info->DefStart() != nullptr); |
| } |
| LOG(INFO) << "================================================"; |
| } |
| |
| void Mir2Lir::DumpCoreRegPool() { |
| DumpRegPool(®_pool_->core_regs_); |
| DumpRegPool(®_pool_->core64_regs_); |
| } |
| |
| void Mir2Lir::DumpFpRegPool() { |
| DumpRegPool(®_pool_->sp_regs_); |
| DumpRegPool(®_pool_->dp_regs_); |
| } |
| |
| void Mir2Lir::DumpRegPools() { |
| LOG(INFO) << "Core registers"; |
| DumpCoreRegPool(); |
| LOG(INFO) << "FP registers"; |
| DumpFpRegPool(); |
| } |
| |
| void Mir2Lir::Clobber(RegStorage reg) { |
| if (UNLIKELY(reg.IsPair())) { |
| DCHECK(!GetRegInfo(reg.GetLow())->IsAliased()); |
| Clobber(reg.GetLow()); |
| DCHECK(!GetRegInfo(reg.GetHigh())->IsAliased()); |
| Clobber(reg.GetHigh()); |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| if (info->IsTemp() && !info->IsDead()) { |
| if (info->GetReg().NotExactlyEquals(info->Partner())) { |
| ClobberBody(GetRegInfo(info->Partner())); |
| } |
| ClobberBody(info); |
| if (info->IsAliased()) { |
| ClobberAliases(info, info->StorageMask()); |
| } else { |
| RegisterInfo* master = info->Master(); |
| if (info != master) { |
| ClobberBody(info->Master()); |
| ClobberAliases(info->Master(), info->StorageMask()); |
| } |
| } |
| } |
| } |
| } |
| |
| void Mir2Lir::ClobberAliases(RegisterInfo* info, uint32_t clobber_mask) { |
| for (RegisterInfo* alias = info->GetAliasChain(); alias != nullptr; |
| alias = alias->GetAliasChain()) { |
| DCHECK(!alias->IsAliased()); // Only the master should be marked as alised. |
| // Only clobber if we have overlap. |
| if ((alias->StorageMask() & clobber_mask) != 0) { |
| ClobberBody(alias); |
| } |
| } |
| } |
| |
| /* |
| * Break the association between a Dalvik vreg and a physical temp register of either register |
| * class. |
| * TODO: Ideally, the public version of this code should not exist. Besides its local usage |
| * in the register utilities, is is also used by code gen routines to work around a deficiency in |
| * local register allocation, which fails to distinguish between the "in" and "out" identities |
| * of Dalvik vregs. This can result in useless register copies when the same Dalvik vreg |
| * is used both as the source and destination register of an operation in which the type |
| * changes (for example: INT_TO_FLOAT v1, v1). Revisit when improved register allocation is |
| * addressed. |
| */ |
| void Mir2Lir::ClobberSReg(int s_reg) { |
| if (s_reg != INVALID_SREG) { |
| if (kIsDebugBuild && s_reg == live_sreg_) { |
| live_sreg_ = INVALID_SREG; |
| } |
| GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_); |
| for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) { |
| if (info->SReg() == s_reg) { |
| if (info->GetReg().NotExactlyEquals(info->Partner())) { |
| // Dealing with a pair - clobber the other half. |
| DCHECK(!info->IsAliased()); |
| ClobberBody(GetRegInfo(info->Partner())); |
| } |
| ClobberBody(info); |
| if (info->IsAliased()) { |
| ClobberAliases(info, info->StorageMask()); |
| } |
| } |
| } |
| } |
| } |
| |
| /* |
| * SSA names associated with the initial definitions of Dalvik |
| * registers are the same as the Dalvik register number (and |
| * thus take the same position in the promotion_map. However, |
| * the special Method* and compiler temp resisters use negative |
| * v_reg numbers to distinguish them and can have an arbitrary |
| * ssa name (above the last original Dalvik register). This function |
| * maps SSA names to positions in the promotion_map array. |
| */ |
| int Mir2Lir::SRegToPMap(int s_reg) { |
| DCHECK_LT(s_reg, mir_graph_->GetNumSSARegs()); |
| DCHECK_GE(s_reg, 0); |
| int v_reg = mir_graph_->SRegToVReg(s_reg); |
| if (v_reg >= 0) { |
| DCHECK_LT(v_reg, cu_->num_dalvik_registers); |
| return v_reg; |
| } else { |
| /* |
| * It must be the case that the v_reg for temporary is less than or equal to the |
| * base reg for temps. For that reason, "position" must be zero or positive. |
| */ |
| unsigned int position = std::abs(v_reg) - std::abs(static_cast<int>(kVRegTempBaseReg)); |
| |
| // The temporaries are placed after dalvik registers in the promotion map |
| DCHECK_LT(position, mir_graph_->GetNumUsedCompilerTemps()); |
| return cu_->num_dalvik_registers + position; |
| } |
| } |
| |
| // TODO: refactor following Alloc/Record routines - much commonality. |
| void Mir2Lir::RecordCorePromotion(RegStorage reg, int s_reg) { |
| int p_map_idx = SRegToPMap(s_reg); |
| int v_reg = mir_graph_->SRegToVReg(s_reg); |
| int reg_num = reg.GetRegNum(); |
| GetRegInfo(reg)->MarkInUse(); |
| core_spill_mask_ |= (1 << reg_num); |
| // Include reg for later sort |
| core_vmap_table_.push_back(reg_num << VREG_NUM_WIDTH | (v_reg & ((1 << VREG_NUM_WIDTH) - 1))); |
| num_core_spills_++; |
| promotion_map_[p_map_idx].core_location = kLocPhysReg; |
| promotion_map_[p_map_idx].core_reg = reg_num; |
| } |
| |
| /* Reserve a callee-save register. Return InvalidReg if none available */ |
| RegStorage Mir2Lir::AllocPreservedCoreReg(int s_reg) { |
| RegStorage res; |
| /* |
| * Note: it really doesn't matter much whether we allocate from the core or core64 |
| * pool for 64-bit targets - but for some targets it does matter whether allocations |
| * happens from the single or double pool. This entire section of code could stand |
| * a good refactoring. |
| */ |
| GrowableArray<RegisterInfo*>::Iterator it(®_pool_->core_regs_); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| if (!info->IsTemp() && !info->InUse()) { |
| res = info->GetReg(); |
| RecordCorePromotion(res, s_reg); |
| break; |
| } |
| } |
| return res; |
| } |
| |
| void Mir2Lir::RecordFpPromotion(RegStorage reg, int s_reg) { |
| DCHECK_NE(cu_->instruction_set, kThumb2); |
| int p_map_idx = SRegToPMap(s_reg); |
| int v_reg = mir_graph_->SRegToVReg(s_reg); |
| int reg_num = reg.GetRegNum(); |
| GetRegInfo(reg)->MarkInUse(); |
| fp_spill_mask_ |= (1 << reg_num); |
| // Include reg for later sort |
| fp_vmap_table_.push_back(reg_num << VREG_NUM_WIDTH | (v_reg & ((1 << VREG_NUM_WIDTH) - 1))); |
| num_fp_spills_++; |
| promotion_map_[p_map_idx].fp_location = kLocPhysReg; |
| promotion_map_[p_map_idx].fp_reg = reg.GetReg(); |
| } |
| |
| // Reserve a callee-save floating point. |
| RegStorage Mir2Lir::AllocPreservedFpReg(int s_reg) { |
| /* |
| * For targets other than Thumb2, it doesn't matter whether we allocate from |
| * the sp_regs_ or dp_regs_ pool. Some refactoring is in order here. |
| */ |
| DCHECK_NE(cu_->instruction_set, kThumb2); |
| RegStorage res; |
| GrowableArray<RegisterInfo*>::Iterator it(®_pool_->sp_regs_); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| if (!info->IsTemp() && !info->InUse()) { |
| res = info->GetReg(); |
| RecordFpPromotion(res, s_reg); |
| break; |
| } |
| } |
| return res; |
| } |
| |
| // TODO: this is Thumb2 only. Remove when DoPromotion refactored. |
| RegStorage Mir2Lir::AllocPreservedDouble(int s_reg) { |
| RegStorage res; |
| UNIMPLEMENTED(FATAL) << "Unexpected use of AllocPreservedDouble"; |
| return res; |
| } |
| |
| // TODO: this is Thumb2 only. Remove when DoPromotion refactored. |
| RegStorage Mir2Lir::AllocPreservedSingle(int s_reg) { |
| RegStorage res; |
| UNIMPLEMENTED(FATAL) << "Unexpected use of AllocPreservedSingle"; |
| return res; |
| } |
| |
| |
| RegStorage Mir2Lir::AllocTempBody(GrowableArray<RegisterInfo*> ®s, int* next_temp, bool required) { |
| int num_regs = regs.Size(); |
| int next = *next_temp; |
| for (int i = 0; i< num_regs; i++) { |
| if (next >= num_regs) |
| next = 0; |
| RegisterInfo* info = regs.Get(next); |
| // Try to allocate a register that doesn't hold a live value. |
| if (info->IsTemp() && !info->InUse() && info->IsDead()) { |
| Clobber(info->GetReg()); |
| info->MarkInUse(); |
| /* |
| * NOTE: "wideness" is an attribute of how the container is used, not its physical size. |
| * The caller will set wideness as appropriate. |
| */ |
| if (info->IsWide()) { |
| RegisterInfo* partner = GetRegInfo(info->Partner()); |
| DCHECK_EQ(info->GetReg().GetRegNum(), partner->Partner().GetRegNum()); |
| DCHECK(partner->IsWide()); |
| info->SetIsWide(false); |
| partner->SetIsWide(false); |
| } |
| *next_temp = next + 1; |
| return info->GetReg(); |
| } |
| next++; |
| } |
| next = *next_temp; |
| // No free non-live regs. Anything we can kill? |
| for (int i = 0; i< num_regs; i++) { |
| if (next >= num_regs) |
| next = 0; |
| RegisterInfo* info = regs.Get(next); |
| if (info->IsTemp() && !info->InUse()) { |
| // Got one. Kill it. |
| ClobberSReg(info->SReg()); |
| Clobber(info->GetReg()); |
| info->MarkInUse(); |
| if (info->IsWide()) { |
| RegisterInfo* partner = GetRegInfo(info->Partner()); |
| DCHECK_EQ(info->GetReg().GetRegNum(), partner->Partner().GetRegNum()); |
| DCHECK(partner->IsWide()); |
| info->SetIsWide(false); |
| partner->SetIsWide(false); |
| } |
| *next_temp = next + 1; |
| return info->GetReg(); |
| } |
| next++; |
| } |
| if (required) { |
| CodegenDump(); |
| DumpRegPools(); |
| LOG(FATAL) << "No free temp registers"; |
| } |
| return RegStorage::InvalidReg(); // No register available |
| } |
| |
| RegStorage Mir2Lir::AllocTemp(bool required) { |
| return AllocTempBody(reg_pool_->core_regs_, ®_pool_->next_core_reg_, required); |
| } |
| |
| RegStorage Mir2Lir::AllocTempWide(bool required) { |
| RegStorage res; |
| if (reg_pool_->core64_regs_.Size() != 0) { |
| res = AllocTempBody(reg_pool_->core64_regs_, ®_pool_->next_core64_reg_, required); |
| } else { |
| RegStorage low_reg = AllocTemp(); |
| RegStorage high_reg = AllocTemp(); |
| res = RegStorage::MakeRegPair(low_reg, high_reg); |
| } |
| if (required) { |
| CheckRegStorage(res, WidenessCheck::kCheckWide, RefCheck::kIgnoreRef, FPCheck::kCheckNotFP); |
| } |
| return res; |
| } |
| |
| RegStorage Mir2Lir::AllocTempRef(bool required) { |
| RegStorage res = AllocTempBody(*reg_pool_->ref_regs_, reg_pool_->next_ref_reg_, required); |
| if (required) { |
| DCHECK(!res.IsPair()); |
| CheckRegStorage(res, WidenessCheck::kCheckNotWide, RefCheck::kCheckRef, FPCheck::kCheckNotFP); |
| } |
| return res; |
| } |
| |
| RegStorage Mir2Lir::AllocTempSingle(bool required) { |
| RegStorage res = AllocTempBody(reg_pool_->sp_regs_, ®_pool_->next_sp_reg_, required); |
| if (required) { |
| DCHECK(res.IsSingle()) << "Reg: 0x" << std::hex << res.GetRawBits(); |
| CheckRegStorage(res, WidenessCheck::kCheckNotWide, RefCheck::kCheckNotRef, FPCheck::kIgnoreFP); |
| } |
| return res; |
| } |
| |
| RegStorage Mir2Lir::AllocTempDouble(bool required) { |
| RegStorage res = AllocTempBody(reg_pool_->dp_regs_, ®_pool_->next_dp_reg_, required); |
| if (required) { |
| DCHECK(res.IsDouble()) << "Reg: 0x" << std::hex << res.GetRawBits(); |
| CheckRegStorage(res, WidenessCheck::kCheckWide, RefCheck::kCheckNotRef, FPCheck::kIgnoreFP); |
| } |
| return res; |
| } |
| |
| RegStorage Mir2Lir::AllocTypedTempWide(bool fp_hint, int reg_class, bool required) { |
| DCHECK_NE(reg_class, kRefReg); // NOTE: the Dalvik width of a reference is always 32 bits. |
| if (((reg_class == kAnyReg) && fp_hint) || (reg_class == kFPReg)) { |
| return AllocTempDouble(required); |
| } |
| return AllocTempWide(required); |
| } |
| |
| RegStorage Mir2Lir::AllocTypedTemp(bool fp_hint, int reg_class, bool required) { |
| if (((reg_class == kAnyReg) && fp_hint) || (reg_class == kFPReg)) { |
| return AllocTempSingle(required); |
| } else if (reg_class == kRefReg) { |
| return AllocTempRef(required); |
| } |
| return AllocTemp(required); |
| } |
| |
| RegStorage Mir2Lir::FindLiveReg(GrowableArray<RegisterInfo*> ®s, int s_reg) { |
| RegStorage res; |
| GrowableArray<RegisterInfo*>::Iterator it(®s); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| if ((info->SReg() == s_reg) && info->IsLive()) { |
| res = info->GetReg(); |
| break; |
| } |
| } |
| return res; |
| } |
| |
| RegStorage Mir2Lir::AllocLiveReg(int s_reg, int reg_class, bool wide) { |
| RegStorage reg; |
| if (reg_class == kRefReg) { |
| reg = FindLiveReg(*reg_pool_->ref_regs_, s_reg); |
| CheckRegStorage(reg, WidenessCheck::kCheckNotWide, RefCheck::kCheckRef, FPCheck::kCheckNotFP); |
| } |
| if (!reg.Valid() && ((reg_class == kAnyReg) || (reg_class == kFPReg))) { |
| reg = FindLiveReg(wide ? reg_pool_->dp_regs_ : reg_pool_->sp_regs_, s_reg); |
| } |
| if (!reg.Valid() && (reg_class != kFPReg)) { |
| if (cu_->target64) { |
| reg = FindLiveReg(wide || reg_class == kRefReg ? reg_pool_->core64_regs_ : |
| reg_pool_->core_regs_, s_reg); |
| } else { |
| reg = FindLiveReg(reg_pool_->core_regs_, s_reg); |
| } |
| } |
| if (reg.Valid()) { |
| if (wide && !reg.IsFloat() && !cu_->target64) { |
| // Only allow reg pairs for core regs on 32-bit targets. |
| RegStorage high_reg = FindLiveReg(reg_pool_->core_regs_, s_reg + 1); |
| if (high_reg.Valid()) { |
| reg = RegStorage::MakeRegPair(reg, high_reg); |
| MarkWide(reg); |
| } else { |
| // Only half available. |
| reg = RegStorage::InvalidReg(); |
| } |
| } |
| if (reg.Valid() && (wide != GetRegInfo(reg)->IsWide())) { |
| // Width mismatch - don't try to reuse. |
| reg = RegStorage::InvalidReg(); |
| } |
| } |
| if (reg.Valid()) { |
| if (reg.IsPair()) { |
| RegisterInfo* info_low = GetRegInfo(reg.GetLow()); |
| RegisterInfo* info_high = GetRegInfo(reg.GetHigh()); |
| if (info_low->IsTemp()) { |
| info_low->MarkInUse(); |
| } |
| if (info_high->IsTemp()) { |
| info_high->MarkInUse(); |
| } |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| if (info->IsTemp()) { |
| info->MarkInUse(); |
| } |
| } |
| } else { |
| // Either not found, or something didn't match up. Clobber to prevent any stale instances. |
| ClobberSReg(s_reg); |
| if (wide) { |
| ClobberSReg(s_reg + 1); |
| } |
| } |
| CheckRegStorage(reg, WidenessCheck::kIgnoreWide, |
| reg_class == kRefReg ? RefCheck::kCheckRef : RefCheck::kIgnoreRef, |
| FPCheck::kIgnoreFP); |
| return reg; |
| } |
| |
| void Mir2Lir::FreeTemp(RegStorage reg) { |
| if (reg.IsPair()) { |
| FreeTemp(reg.GetLow()); |
| FreeTemp(reg.GetHigh()); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| if (p->IsTemp()) { |
| p->MarkFree(); |
| p->SetIsWide(false); |
| p->SetPartner(reg); |
| } |
| } |
| } |
| |
| void Mir2Lir::FreeRegLocTemps(RegLocation rl_keep, RegLocation rl_free) { |
| DCHECK(rl_keep.wide); |
| DCHECK(rl_free.wide); |
| int free_low = rl_free.reg.GetLowReg(); |
| int free_high = rl_free.reg.GetHighReg(); |
| int keep_low = rl_keep.reg.GetLowReg(); |
| int keep_high = rl_keep.reg.GetHighReg(); |
| if ((free_low != keep_low) && (free_low != keep_high) && |
| (free_high != keep_low) && (free_high != keep_high)) { |
| // No overlap, free both |
| FreeTemp(rl_free.reg); |
| } |
| } |
| |
| bool Mir2Lir::IsLive(RegStorage reg) { |
| bool res; |
| if (reg.IsPair()) { |
| RegisterInfo* p_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* p_hi = GetRegInfo(reg.GetHigh()); |
| DCHECK_EQ(p_lo->IsLive(), p_hi->IsLive()); |
| res = p_lo->IsLive() || p_hi->IsLive(); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| res = p->IsLive(); |
| } |
| return res; |
| } |
| |
| bool Mir2Lir::IsTemp(RegStorage reg) { |
| bool res; |
| if (reg.IsPair()) { |
| RegisterInfo* p_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* p_hi = GetRegInfo(reg.GetHigh()); |
| res = p_lo->IsTemp() || p_hi->IsTemp(); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| res = p->IsTemp(); |
| } |
| return res; |
| } |
| |
| bool Mir2Lir::IsPromoted(RegStorage reg) { |
| bool res; |
| if (reg.IsPair()) { |
| RegisterInfo* p_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* p_hi = GetRegInfo(reg.GetHigh()); |
| res = !p_lo->IsTemp() || !p_hi->IsTemp(); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| res = !p->IsTemp(); |
| } |
| return res; |
| } |
| |
| bool Mir2Lir::IsDirty(RegStorage reg) { |
| bool res; |
| if (reg.IsPair()) { |
| RegisterInfo* p_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* p_hi = GetRegInfo(reg.GetHigh()); |
| res = p_lo->IsDirty() || p_hi->IsDirty(); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| res = p->IsDirty(); |
| } |
| return res; |
| } |
| |
| /* |
| * Similar to AllocTemp(), but forces the allocation of a specific |
| * register. No check is made to see if the register was previously |
| * allocated. Use with caution. |
| */ |
| void Mir2Lir::LockTemp(RegStorage reg) { |
| DCHECK(IsTemp(reg)); |
| if (reg.IsPair()) { |
| RegisterInfo* p_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* p_hi = GetRegInfo(reg.GetHigh()); |
| p_lo->MarkInUse(); |
| p_lo->MarkDead(); |
| p_hi->MarkInUse(); |
| p_hi->MarkDead(); |
| } else { |
| RegisterInfo* p = GetRegInfo(reg); |
| p->MarkInUse(); |
| p->MarkDead(); |
| } |
| } |
| |
| void Mir2Lir::ResetDef(RegStorage reg) { |
| if (reg.IsPair()) { |
| GetRegInfo(reg.GetLow())->ResetDefBody(); |
| GetRegInfo(reg.GetHigh())->ResetDefBody(); |
| } else { |
| GetRegInfo(reg)->ResetDefBody(); |
| } |
| } |
| |
| void Mir2Lir::NullifyRange(RegStorage reg, int s_reg) { |
| RegisterInfo* info = nullptr; |
| RegStorage rs = reg.IsPair() ? reg.GetLow() : reg; |
| if (IsTemp(rs)) { |
| info = GetRegInfo(reg); |
| } |
| if ((info != nullptr) && (info->DefStart() != nullptr) && (info->DefEnd() != nullptr)) { |
| DCHECK_EQ(info->SReg(), s_reg); // Make sure we're on the same page. |
| for (LIR* p = info->DefStart();; p = p->next) { |
| NopLIR(p); |
| if (p == info->DefEnd()) { |
| break; |
| } |
| } |
| } |
| } |
| |
| /* |
| * Mark the beginning and end LIR of a def sequence. Note that |
| * on entry start points to the LIR prior to the beginning of the |
| * sequence. |
| */ |
| void Mir2Lir::MarkDef(RegLocation rl, LIR *start, LIR *finish) { |
| DCHECK(!rl.wide); |
| DCHECK(start && start->next); |
| DCHECK(finish); |
| RegisterInfo* p = GetRegInfo(rl.reg); |
| p->SetDefStart(start->next); |
| p->SetDefEnd(finish); |
| } |
| |
| /* |
| * Mark the beginning and end LIR of a def sequence. Note that |
| * on entry start points to the LIR prior to the beginning of the |
| * sequence. |
| */ |
| void Mir2Lir::MarkDefWide(RegLocation rl, LIR *start, LIR *finish) { |
| DCHECK(rl.wide); |
| DCHECK(start && start->next); |
| DCHECK(finish); |
| RegisterInfo* p; |
| if (rl.reg.IsPair()) { |
| p = GetRegInfo(rl.reg.GetLow()); |
| ResetDef(rl.reg.GetHigh()); // Only track low of pair |
| } else { |
| p = GetRegInfo(rl.reg); |
| } |
| p->SetDefStart(start->next); |
| p->SetDefEnd(finish); |
| } |
| |
| void Mir2Lir::ResetDefLoc(RegLocation rl) { |
| DCHECK(!rl.wide); |
| if (IsTemp(rl.reg) && !(cu_->disable_opt & (1 << kSuppressLoads))) { |
| NullifyRange(rl.reg, rl.s_reg_low); |
| } |
| ResetDef(rl.reg); |
| } |
| |
| void Mir2Lir::ResetDefLocWide(RegLocation rl) { |
| DCHECK(rl.wide); |
| // If pair, only track low reg of pair. |
| RegStorage rs = rl.reg.IsPair() ? rl.reg.GetLow() : rl.reg; |
| if (IsTemp(rs) && !(cu_->disable_opt & (1 << kSuppressLoads))) { |
| NullifyRange(rs, rl.s_reg_low); |
| } |
| ResetDef(rs); |
| } |
| |
| void Mir2Lir::ResetDefTracking() { |
| GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_); |
| for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) { |
| info->ResetDefBody(); |
| } |
| } |
| |
| void Mir2Lir::ClobberAllTemps() { |
| GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_); |
| for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) { |
| ClobberBody(info); |
| } |
| } |
| |
| void Mir2Lir::FlushRegWide(RegStorage reg) { |
| if (reg.IsPair()) { |
| RegisterInfo* info1 = GetRegInfo(reg.GetLow()); |
| RegisterInfo* info2 = GetRegInfo(reg.GetHigh()); |
| DCHECK(info1 && info2 && info1->IsWide() && info2->IsWide() && |
| (info1->Partner().ExactlyEquals(info2->GetReg())) && |
| (info2->Partner().ExactlyEquals(info1->GetReg()))); |
| if ((info1->IsLive() && info1->IsDirty()) || (info2->IsLive() && info2->IsDirty())) { |
| if (!(info1->IsTemp() && info2->IsTemp())) { |
| /* Should not happen. If it does, there's a problem in eval_loc */ |
| LOG(FATAL) << "Long half-temp, half-promoted"; |
| } |
| |
| info1->SetIsDirty(false); |
| info2->SetIsDirty(false); |
| if (mir_graph_->SRegToVReg(info2->SReg()) < mir_graph_->SRegToVReg(info1->SReg())) { |
| info1 = info2; |
| } |
| int v_reg = mir_graph_->SRegToVReg(info1->SReg()); |
| ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg); |
| StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile); |
| } |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| if (info->IsLive() && info->IsDirty()) { |
| info->SetIsDirty(false); |
| int v_reg = mir_graph_->SRegToVReg(info->SReg()); |
| ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg); |
| StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile); |
| } |
| } |
| } |
| |
| void Mir2Lir::FlushReg(RegStorage reg) { |
| DCHECK(!reg.IsPair()); |
| RegisterInfo* info = GetRegInfo(reg); |
| if (info->IsLive() && info->IsDirty()) { |
| info->SetIsDirty(false); |
| int v_reg = mir_graph_->SRegToVReg(info->SReg()); |
| ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg); |
| StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, kWord, kNotVolatile); |
| } |
| } |
| |
| void Mir2Lir::FlushSpecificReg(RegisterInfo* info) { |
| if (info->IsWide()) { |
| FlushRegWide(info->GetReg()); |
| } else { |
| FlushReg(info->GetReg()); |
| } |
| } |
| |
| void Mir2Lir::FlushAllRegs() { |
| GrowableArray<RegisterInfo*>::Iterator it(&tempreg_info_); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| if (info->IsDirty() && info->IsLive()) { |
| FlushSpecificReg(info); |
| } |
| info->MarkDead(); |
| info->SetIsWide(false); |
| } |
| } |
| |
| |
| bool Mir2Lir::RegClassMatches(int reg_class, RegStorage reg) { |
| if (reg_class == kAnyReg) { |
| return true; |
| } else if ((reg_class == kCoreReg) || (reg_class == kRefReg)) { |
| /* |
| * For this purpose, consider Core and Ref to be the same class. We aren't dealing |
| * with width here - that should be checked at a higher level (if needed). |
| */ |
| return !reg.IsFloat(); |
| } else { |
| return reg.IsFloat(); |
| } |
| } |
| |
| void Mir2Lir::MarkLive(RegLocation loc) { |
| RegStorage reg = loc.reg; |
| if (!IsTemp(reg)) { |
| return; |
| } |
| int s_reg = loc.s_reg_low; |
| if (s_reg == INVALID_SREG) { |
| // Can't be live if no associated sreg. |
| if (reg.IsPair()) { |
| GetRegInfo(reg.GetLow())->MarkDead(); |
| GetRegInfo(reg.GetHigh())->MarkDead(); |
| } else { |
| GetRegInfo(reg)->MarkDead(); |
| } |
| } else { |
| if (reg.IsPair()) { |
| RegisterInfo* info_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* info_hi = GetRegInfo(reg.GetHigh()); |
| if (info_lo->IsLive() && (info_lo->SReg() == s_reg) && info_hi->IsLive() && |
| (info_hi->SReg() == s_reg)) { |
| return; // Already live. |
| } |
| ClobberSReg(s_reg); |
| ClobberSReg(s_reg + 1); |
| info_lo->MarkLive(s_reg); |
| info_hi->MarkLive(s_reg + 1); |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| if (info->IsLive() && (info->SReg() == s_reg)) { |
| return; // Already live. |
| } |
| ClobberSReg(s_reg); |
| if (loc.wide) { |
| ClobberSReg(s_reg + 1); |
| } |
| info->MarkLive(s_reg); |
| } |
| if (loc.wide) { |
| MarkWide(reg); |
| } else { |
| MarkNarrow(reg); |
| } |
| } |
| } |
| |
| void Mir2Lir::MarkTemp(RegStorage reg) { |
| DCHECK(!reg.IsPair()); |
| RegisterInfo* info = GetRegInfo(reg); |
| tempreg_info_.Insert(info); |
| info->SetIsTemp(true); |
| } |
| |
| void Mir2Lir::UnmarkTemp(RegStorage reg) { |
| DCHECK(!reg.IsPair()); |
| RegisterInfo* info = GetRegInfo(reg); |
| tempreg_info_.Delete(info); |
| info->SetIsTemp(false); |
| } |
| |
| void Mir2Lir::MarkWide(RegStorage reg) { |
| if (reg.IsPair()) { |
| RegisterInfo* info_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* info_hi = GetRegInfo(reg.GetHigh()); |
| // Unpair any old partners. |
| if (info_lo->IsWide() && info_lo->Partner().NotExactlyEquals(info_hi->GetReg())) { |
| GetRegInfo(info_lo->Partner())->SetIsWide(false); |
| } |
| if (info_hi->IsWide() && info_hi->Partner().NotExactlyEquals(info_lo->GetReg())) { |
| GetRegInfo(info_hi->Partner())->SetIsWide(false); |
| } |
| info_lo->SetIsWide(true); |
| info_hi->SetIsWide(true); |
| info_lo->SetPartner(reg.GetHigh()); |
| info_hi->SetPartner(reg.GetLow()); |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| info->SetIsWide(true); |
| info->SetPartner(reg); |
| } |
| } |
| |
| void Mir2Lir::MarkNarrow(RegStorage reg) { |
| DCHECK(!reg.IsPair()); |
| RegisterInfo* info = GetRegInfo(reg); |
| info->SetIsWide(false); |
| info->SetPartner(reg); |
| } |
| |
| void Mir2Lir::MarkClean(RegLocation loc) { |
| if (loc.reg.IsPair()) { |
| RegisterInfo* info = GetRegInfo(loc.reg.GetLow()); |
| info->SetIsDirty(false); |
| info = GetRegInfo(loc.reg.GetHigh()); |
| info->SetIsDirty(false); |
| } else { |
| RegisterInfo* info = GetRegInfo(loc.reg); |
| info->SetIsDirty(false); |
| } |
| } |
| |
| // FIXME: need to verify rules/assumptions about how wide values are treated in 64BitSolos. |
| void Mir2Lir::MarkDirty(RegLocation loc) { |
| if (loc.home) { |
| // If already home, can't be dirty |
| return; |
| } |
| if (loc.reg.IsPair()) { |
| RegisterInfo* info = GetRegInfo(loc.reg.GetLow()); |
| info->SetIsDirty(true); |
| info = GetRegInfo(loc.reg.GetHigh()); |
| info->SetIsDirty(true); |
| } else { |
| RegisterInfo* info = GetRegInfo(loc.reg); |
| info->SetIsDirty(true); |
| } |
| } |
| |
| void Mir2Lir::MarkInUse(RegStorage reg) { |
| if (reg.IsPair()) { |
| GetRegInfo(reg.GetLow())->MarkInUse(); |
| GetRegInfo(reg.GetHigh())->MarkInUse(); |
| } else { |
| GetRegInfo(reg)->MarkInUse(); |
| } |
| } |
| |
| bool Mir2Lir::CheckCorePoolSanity() { |
| GrowableArray<RegisterInfo*>::Iterator it(&tempreg_info_); |
| for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) { |
| if (info->IsTemp() && info->IsLive() && info->IsWide()) { |
| RegStorage my_reg = info->GetReg(); |
| int my_sreg = info->SReg(); |
| RegStorage partner_reg = info->Partner(); |
| RegisterInfo* partner = GetRegInfo(partner_reg); |
| DCHECK(partner != NULL); |
| DCHECK(partner->IsWide()); |
| DCHECK_EQ(my_reg.GetReg(), partner->Partner().GetReg()); |
| DCHECK(partner->IsLive()); |
| int partner_sreg = partner->SReg(); |
| if (my_sreg == INVALID_SREG) { |
| DCHECK_EQ(partner_sreg, INVALID_SREG); |
| } else { |
| int diff = my_sreg - partner_sreg; |
| DCHECK((diff == 0) || (diff == -1) || (diff == 1)); |
| } |
| } |
| if (info->Master() != info) { |
| // Aliased. |
| if (info->IsLive() && (info->SReg() != INVALID_SREG)) { |
| // If I'm live, master should not be live, but should show liveness in alias set. |
| DCHECK_EQ(info->Master()->SReg(), INVALID_SREG); |
| DCHECK(!info->Master()->IsDead()); |
| } |
| // TODO: Add checks in !info->IsDead() case to ensure every live bit is owned by exactly 1 reg. |
| } |
| if (info->IsAliased()) { |
| // Has child aliases. |
| DCHECK_EQ(info->Master(), info); |
| if (info->IsLive() && (info->SReg() != INVALID_SREG)) { |
| // Master live, no child should be dead - all should show liveness in set. |
| for (RegisterInfo* p = info->GetAliasChain(); p != nullptr; p = p->GetAliasChain()) { |
| DCHECK(!p->IsDead()); |
| DCHECK_EQ(p->SReg(), INVALID_SREG); |
| } |
| } else if (!info->IsDead()) { |
| // Master not live, one or more aliases must be. |
| bool live_alias = false; |
| for (RegisterInfo* p = info->GetAliasChain(); p != nullptr; p = p->GetAliasChain()) { |
| live_alias |= p->IsLive(); |
| } |
| DCHECK(live_alias); |
| } |
| } |
| if (info->IsLive() && (info->SReg() == INVALID_SREG)) { |
| // If not fully live, should have INVALID_SREG and def's should be null. |
| DCHECK(info->DefStart() == nullptr); |
| DCHECK(info->DefEnd() == nullptr); |
| } |
| } |
| return true; |
| } |
| |
| /* |
| * Return an updated location record with current in-register status. |
| * If the value lives in live temps, reflect that fact. No code |
| * is generated. If the live value is part of an older pair, |
| * clobber both low and high. |
| * TUNING: clobbering both is a bit heavy-handed, but the alternative |
| * is a bit complex when dealing with FP regs. Examine code to see |
| * if it's worthwhile trying to be more clever here. |
| */ |
| RegLocation Mir2Lir::UpdateLoc(RegLocation loc) { |
| DCHECK(!loc.wide); |
| DCHECK(CheckCorePoolSanity()); |
| if (loc.location != kLocPhysReg) { |
| DCHECK((loc.location == kLocDalvikFrame) || |
| (loc.location == kLocCompilerTemp)); |
| RegStorage reg = AllocLiveReg(loc.s_reg_low, loc.ref ? kRefReg : kAnyReg, false); |
| if (reg.Valid()) { |
| bool match = true; |
| RegisterInfo* info = GetRegInfo(reg); |
| match &= !reg.IsPair(); |
| match &= !info->IsWide(); |
| if (match) { |
| loc.location = kLocPhysReg; |
| loc.reg = reg; |
| } else { |
| Clobber(reg); |
| FreeTemp(reg); |
| } |
| } |
| CheckRegLocation(loc); |
| } |
| return loc; |
| } |
| |
| RegLocation Mir2Lir::UpdateLocWide(RegLocation loc) { |
| DCHECK(loc.wide); |
| DCHECK(CheckCorePoolSanity()); |
| if (loc.location != kLocPhysReg) { |
| DCHECK((loc.location == kLocDalvikFrame) || |
| (loc.location == kLocCompilerTemp)); |
| RegStorage reg = AllocLiveReg(loc.s_reg_low, kAnyReg, true); |
| if (reg.Valid()) { |
| bool match = true; |
| if (reg.IsPair()) { |
| // If we've got a register pair, make sure that it was last used as the same pair. |
| RegisterInfo* info_lo = GetRegInfo(reg.GetLow()); |
| RegisterInfo* info_hi = GetRegInfo(reg.GetHigh()); |
| match &= info_lo->IsWide(); |
| match &= info_hi->IsWide(); |
| match &= (info_lo->Partner().ExactlyEquals(info_hi->GetReg())); |
| match &= (info_hi->Partner().ExactlyEquals(info_lo->GetReg())); |
| } else { |
| RegisterInfo* info = GetRegInfo(reg); |
| match &= info->IsWide(); |
| match &= (info->GetReg().ExactlyEquals(info->Partner())); |
| } |
| if (match) { |
| loc.location = kLocPhysReg; |
| loc.reg = reg; |
| } else { |
| Clobber(reg); |
| FreeTemp(reg); |
| } |
| } |
| CheckRegLocation(loc); |
| } |
| return loc; |
| } |
| |
| /* For use in cases we don't know (or care) width */ |
| RegLocation Mir2Lir::UpdateRawLoc(RegLocation loc) { |
| if (loc.wide) |
| return UpdateLocWide(loc); |
| else |
| return UpdateLoc(loc); |
| } |
| |
| RegLocation Mir2Lir::EvalLocWide(RegLocation loc, int reg_class, bool update) { |
| DCHECK(loc.wide); |
| |
| loc = UpdateLocWide(loc); |
| |
| /* If already in registers, we can assume proper form. Right reg class? */ |
| if (loc.location == kLocPhysReg) { |
| if (!RegClassMatches(reg_class, loc.reg)) { |
| // Wrong register class. Reallocate and transfer ownership. |
| RegStorage new_regs = AllocTypedTempWide(loc.fp, reg_class); |
| // Clobber the old regs. |
| Clobber(loc.reg); |
| // ...and mark the new ones live. |
| loc.reg = new_regs; |
| MarkWide(loc.reg); |
| MarkLive(loc); |
| } |
| CheckRegLocation(loc); |
| return loc; |
| } |
| |
| DCHECK_NE(loc.s_reg_low, INVALID_SREG); |
| DCHECK_NE(GetSRegHi(loc.s_reg_low), INVALID_SREG); |
| |
| loc.reg = AllocTypedTempWide(loc.fp, reg_class); |
| MarkWide(loc.reg); |
| |
| if (update) { |
| loc.location = kLocPhysReg; |
| MarkLive(loc); |
| } |
| CheckRegLocation(loc); |
| return loc; |
| } |
| |
| RegLocation Mir2Lir::EvalLoc(RegLocation loc, int reg_class, bool update) { |
| // Narrow reg_class if the loc is a ref. |
| if (loc.ref && reg_class == kAnyReg) { |
| reg_class = kRefReg; |
| } |
| |
| if (loc.wide) { |
| return EvalLocWide(loc, reg_class, update); |
| } |
| |
| loc = UpdateLoc(loc); |
| |
| if (loc.location == kLocPhysReg) { |
| if (!RegClassMatches(reg_class, loc.reg)) { |
| // Wrong register class. Reallocate and transfer ownership. |
| RegStorage new_reg = AllocTypedTemp(loc.fp, reg_class); |
| // Clobber the old reg. |
| Clobber(loc.reg); |
| // ...and mark the new one live. |
| loc.reg = new_reg; |
| MarkLive(loc); |
| } |
| CheckRegLocation(loc); |
| return loc; |
| } |
| |
| DCHECK_NE(loc.s_reg_low, INVALID_SREG); |
| |
| loc.reg = AllocTypedTemp(loc.fp, reg_class); |
| CheckRegLocation(loc); |
| |
| if (update) { |
| loc.location = kLocPhysReg; |
| MarkLive(loc); |
| } |
| CheckRegLocation(loc); |
| return loc; |
| } |
| |
| /* USE SSA names to count references of base Dalvik v_regs. */ |
| void Mir2Lir::CountRefs(RefCounts* core_counts, RefCounts* fp_counts, size_t num_regs) { |
| for (int i = 0; i < mir_graph_->GetNumSSARegs(); i++) { |
| RegLocation loc = mir_graph_->reg_location_[i]; |
| RefCounts* counts = loc.fp ? fp_counts : core_counts; |
| int p_map_idx = SRegToPMap(loc.s_reg_low); |
| int use_count = mir_graph_->GetUseCount(i); |
| if (loc.fp) { |
| if (loc.wide) { |
| if (WideFPRsAreAliases()) { |
| // Floats and doubles can be counted together. |
| counts[p_map_idx].count += use_count; |
| } else { |
| // Treat doubles as a unit, using upper half of fp_counts array. |
| counts[p_map_idx + num_regs].count += use_count; |
| } |
| i++; |
| } else { |
| counts[p_map_idx].count += use_count; |
| } |
| } else { |
| if (loc.wide && WideGPRsAreAliases()) { |
| i++; |
| } |
| if (!IsInexpensiveConstant(loc)) { |
| counts[p_map_idx].count += use_count; |
| } |
| } |
| } |
| } |
| |
| /* qsort callback function, sort descending */ |
| static int SortCounts(const void *val1, const void *val2) { |
| const Mir2Lir::RefCounts* op1 = reinterpret_cast<const Mir2Lir::RefCounts*>(val1); |
| const Mir2Lir::RefCounts* op2 = reinterpret_cast<const Mir2Lir::RefCounts*>(val2); |
| // Note that we fall back to sorting on reg so we get stable output on differing qsort |
| // implementations (such as on host and target or between local host and build servers). |
| // Note also that if a wide val1 and a non-wide val2 have the same count, then val1 always |
| // ``loses'' (as STARTING_WIDE_SREG is or-ed in val1->s_reg). |
| return (op1->count == op2->count) |
| ? (op1->s_reg - op2->s_reg) |
| : (op1->count < op2->count ? 1 : -1); |
| } |
| |
| void Mir2Lir::DumpCounts(const RefCounts* arr, int size, const char* msg) { |
| LOG(INFO) << msg; |
| for (int i = 0; i < size; i++) { |
| if ((arr[i].s_reg & STARTING_WIDE_SREG) != 0) { |
| LOG(INFO) << "s_reg[64_" << (arr[i].s_reg & ~STARTING_WIDE_SREG) << "]: " << arr[i].count; |
| } else { |
| LOG(INFO) << "s_reg[32_" << arr[i].s_reg << "]: " << arr[i].count; |
| } |
| } |
| } |
| |
| /* |
| * Note: some portions of this code required even if the kPromoteRegs |
| * optimization is disabled. |
| */ |
| void Mir2Lir::DoPromotion() { |
| int dalvik_regs = cu_->num_dalvik_registers; |
| int num_regs = dalvik_regs + mir_graph_->GetNumUsedCompilerTemps(); |
| const int promotion_threshold = 1; |
| // Allocate the promotion map - one entry for each Dalvik vReg or compiler temp |
| promotion_map_ = static_cast<PromotionMap*> |
| (arena_->Alloc(num_regs * sizeof(promotion_map_[0]), kArenaAllocRegAlloc)); |
| |
| // Allow target code to add any special registers |
| AdjustSpillMask(); |
| |
| /* |
| * Simple register promotion. Just do a static count of the uses |
| * of Dalvik registers. Note that we examine the SSA names, but |
| * count based on original Dalvik register name. Count refs |
| * separately based on type in order to give allocation |
| * preference to fp doubles - which must be allocated sequential |
| * physical single fp registers starting with an even-numbered |
| * reg. |
| * TUNING: replace with linear scan once we have the ability |
| * to describe register live ranges for GC. |
| */ |
| size_t core_reg_count_size = WideGPRsAreAliases() ? num_regs : num_regs * 2; |
| size_t fp_reg_count_size = WideFPRsAreAliases() ? num_regs : num_regs * 2; |
| RefCounts *core_regs = |
| static_cast<RefCounts*>(arena_->Alloc(sizeof(RefCounts) * core_reg_count_size, |
| kArenaAllocRegAlloc)); |
| RefCounts *fp_regs = |
| static_cast<RefCounts *>(arena_->Alloc(sizeof(RefCounts) * fp_reg_count_size, |
| kArenaAllocRegAlloc)); |
| // Set ssa names for original Dalvik registers |
| for (int i = 0; i < dalvik_regs; i++) { |
| core_regs[i].s_reg = fp_regs[i].s_reg = i; |
| } |
| |
| // Set ssa names for compiler temporaries |
| for (unsigned int ct_idx = 0; ct_idx < mir_graph_->GetNumUsedCompilerTemps(); ct_idx++) { |
| CompilerTemp* ct = mir_graph_->GetCompilerTemp(ct_idx); |
| core_regs[dalvik_regs + ct_idx].s_reg = ct->s_reg_low; |
| fp_regs[dalvik_regs + ct_idx].s_reg = ct->s_reg_low; |
| } |
| |
| // Duplicate in upper half to represent possible wide starting sregs. |
| for (size_t i = num_regs; i < fp_reg_count_size; i++) { |
| fp_regs[i].s_reg = fp_regs[i - num_regs].s_reg | STARTING_WIDE_SREG; |
| } |
| for (size_t i = num_regs; i < core_reg_count_size; i++) { |
| core_regs[i].s_reg = core_regs[i - num_regs].s_reg | STARTING_WIDE_SREG; |
| } |
| |
| // Sum use counts of SSA regs by original Dalvik vreg. |
| CountRefs(core_regs, fp_regs, num_regs); |
| |
| // Sort the count arrays |
| qsort(core_regs, core_reg_count_size, sizeof(RefCounts), SortCounts); |
| qsort(fp_regs, fp_reg_count_size, sizeof(RefCounts), SortCounts); |
| |
| if (cu_->verbose) { |
| DumpCounts(core_regs, core_reg_count_size, "Core regs after sort"); |
| DumpCounts(fp_regs, fp_reg_count_size, "Fp regs after sort"); |
| } |
| |
| if (!(cu_->disable_opt & (1 << kPromoteRegs))) { |
| // Promote fp regs |
| for (size_t i = 0; (i < fp_reg_count_size) && (fp_regs[i].count >= promotion_threshold); i++) { |
| int low_sreg = fp_regs[i].s_reg & ~STARTING_WIDE_SREG; |
| size_t p_map_idx = SRegToPMap(low_sreg); |
| RegStorage reg = RegStorage::InvalidReg(); |
| if (promotion_map_[p_map_idx].fp_location != kLocPhysReg) { |
| // TODO: break out the Thumb2-specific code. |
| if (cu_->instruction_set == kThumb2) { |
| bool wide = fp_regs[i].s_reg & STARTING_WIDE_SREG; |
| if (wide) { |
| if (promotion_map_[p_map_idx + 1].fp_location != kLocPhysReg) { |
| // Ignore result - if can't alloc double may still be able to alloc singles. |
| AllocPreservedDouble(low_sreg); |
| } |
| // Continue regardless of success - might still be able to grab a single. |
| continue; |
| } else { |
| reg = AllocPreservedSingle(low_sreg); |
| } |
| } else { |
| reg = AllocPreservedFpReg(low_sreg); |
| } |
| if (!reg.Valid()) { |
| break; // No more left |
| } |
| } |
| } |
| |
| // Promote core regs |
| for (size_t i = 0; (i < core_reg_count_size) && |
| (core_regs[i].count >= promotion_threshold); i++) { |
| int low_sreg = core_regs[i].s_reg & ~STARTING_WIDE_SREG; |
| size_t p_map_idx = SRegToPMap(low_sreg); |
| if (promotion_map_[p_map_idx].core_location != kLocPhysReg) { |
| RegStorage reg = AllocPreservedCoreReg(low_sreg); |
| if (!reg.Valid()) { |
| break; // No more left |
| } |
| } |
| } |
| } |
| |
| // Now, update SSA names to new home locations |
| for (int i = 0; i < mir_graph_->GetNumSSARegs(); i++) { |
| RegLocation *curr = &mir_graph_->reg_location_[i]; |
| int p_map_idx = SRegToPMap(curr->s_reg_low); |
| int reg_num = curr->fp ? promotion_map_[p_map_idx].fp_reg : promotion_map_[p_map_idx].core_reg; |
| bool wide = curr->wide || (cu_->target64 && curr->ref); |
| RegStorage reg = RegStorage::InvalidReg(); |
| if (curr->fp && promotion_map_[p_map_idx].fp_location == kLocPhysReg) { |
| if (wide && cu_->instruction_set == kThumb2) { |
| if (promotion_map_[p_map_idx + 1].fp_location == kLocPhysReg) { |
| int high_reg = promotion_map_[p_map_idx+1].fp_reg; |
| // TODO: move target-specific restrictions out of here. |
| if (((reg_num & 0x1) == 0) && ((reg_num + 1) == high_reg)) { |
| reg = RegStorage::FloatSolo64(RegStorage::RegNum(reg_num) >> 1); |
| } |
| } |
| } else { |
| reg = wide ? RegStorage::FloatSolo64(reg_num) : RegStorage::FloatSolo32(reg_num); |
| } |
| } else if (!curr->fp && promotion_map_[p_map_idx].core_location == kLocPhysReg) { |
| if (wide && !cu_->target64) { |
| if (promotion_map_[p_map_idx + 1].core_location == kLocPhysReg) { |
| int high_reg = promotion_map_[p_map_idx+1].core_reg; |
| reg = RegStorage(RegStorage::k64BitPair, reg_num, high_reg); |
| } |
| } else { |
| reg = wide ? RegStorage::Solo64(reg_num) : RegStorage::Solo32(reg_num); |
| } |
| } |
| if (reg.Valid()) { |
| curr->reg = reg; |
| curr->location = kLocPhysReg; |
| curr->home = true; |
| } |
| } |
| if (cu_->verbose) { |
| DumpPromotionMap(); |
| } |
| } |
| |
| /* Returns sp-relative offset in bytes for a VReg */ |
| int Mir2Lir::VRegOffset(int v_reg) { |
| return StackVisitor::GetVRegOffset(cu_->code_item, core_spill_mask_, |
| fp_spill_mask_, frame_size_, v_reg, |
| cu_->instruction_set); |
| } |
| |
| /* Returns sp-relative offset in bytes for a SReg */ |
| int Mir2Lir::SRegOffset(int s_reg) { |
| return VRegOffset(mir_graph_->SRegToVReg(s_reg)); |
| } |
| |
| /* Mark register usage state and return long retloc */ |
| RegLocation Mir2Lir::GetReturnWide(RegisterClass reg_class) { |
| RegLocation res; |
| switch (reg_class) { |
| case kRefReg: LOG(FATAL); break; |
| case kFPReg: res = LocCReturnDouble(); break; |
| default: res = LocCReturnWide(); break; |
| } |
| Clobber(res.reg); |
| LockTemp(res.reg); |
| MarkWide(res.reg); |
| CheckRegLocation(res); |
| return res; |
| } |
| |
| RegLocation Mir2Lir::GetReturn(RegisterClass reg_class) { |
| RegLocation res; |
| switch (reg_class) { |
| case kRefReg: res = LocCReturnRef(); break; |
| case kFPReg: res = LocCReturnFloat(); break; |
| default: res = LocCReturn(); break; |
| } |
| Clobber(res.reg); |
| if (cu_->instruction_set == kMips) { |
| MarkInUse(res.reg); |
| } else { |
| LockTemp(res.reg); |
| } |
| CheckRegLocation(res); |
| return res; |
| } |
| |
| void Mir2Lir::SimpleRegAlloc() { |
| DoPromotion(); |
| |
| if (cu_->verbose && !(cu_->disable_opt & (1 << kPromoteRegs))) { |
| LOG(INFO) << "After Promotion"; |
| mir_graph_->DumpRegLocTable(mir_graph_->reg_location_, mir_graph_->GetNumSSARegs()); |
| } |
| |
| /* Set the frame size */ |
| frame_size_ = ComputeFrameSize(); |
| } |
| |
| /* |
| * Get the "real" sreg number associated with an s_reg slot. In general, |
| * s_reg values passed through codegen are the SSA names created by |
| * dataflow analysis and refer to slot numbers in the mir_graph_->reg_location |
| * array. However, renaming is accomplished by simply replacing RegLocation |
| * entries in the reglocation[] array. Therefore, when location |
| * records for operands are first created, we need to ask the locRecord |
| * identified by the dataflow pass what it's new name is. |
| */ |
| int Mir2Lir::GetSRegHi(int lowSreg) { |
| return (lowSreg == INVALID_SREG) ? INVALID_SREG : lowSreg + 1; |
| } |
| |
| bool Mir2Lir::LiveOut(int s_reg) { |
| // For now. |
| return true; |
| } |
| |
| } // namespace art |