| /* |
| * Copyright (C) 2015 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "load_store_elimination.h" |
| |
| #include "base/array_ref.h" |
| #include "base/scoped_arena_allocator.h" |
| #include "base/scoped_arena_containers.h" |
| #include "escape.h" |
| #include "load_store_analysis.h" |
| #include "side_effects_analysis.h" |
| |
| #include <iostream> |
| |
| /** |
| * The general algorithm of load-store elimination (LSE). |
| * Load-store analysis in the previous pass collects a list of heap locations |
| * and does alias analysis of those heap locations. |
| * LSE keeps track of a list of heap values corresponding to the heap |
| * locations. It visits basic blocks in reverse post order and for |
| * each basic block, visits instructions sequentially, and processes |
| * instructions as follows: |
| * - If the instruction is a load, and the heap location for that load has a |
| * valid heap value, the load can be eliminated. In order to maintain the |
| * validity of all heap locations during the optimization phase, the real |
| * elimination is delayed till the end of LSE. |
| * - If the instruction is a store, it updates the heap value for the heap |
| * location of the store with the store instruction. The real heap value |
| * can be fetched from the store instruction. Heap values are invalidated |
| * for heap locations that may alias with the store instruction's heap |
| * location. The store instruction can be eliminated unless the value stored |
| * is later needed e.g. by a load from the same/aliased heap location or |
| * the heap location persists at method return/deoptimization. |
| * The store instruction is also needed if it's not used to track the heap |
| * value anymore, e.g. when it fails to merge with the heap values from other |
| * predecessors. |
| * - A store that stores the same value as the heap value is eliminated. |
| * - The list of heap values are merged at basic block entry from the basic |
| * block's predecessors. The algorithm is single-pass, so loop side-effects is |
| * used as best effort to decide if a heap location is stored inside the loop. |
| * - A special type of objects called singletons are instantiated in the method |
| * and have a single name, i.e. no aliases. Singletons have exclusive heap |
| * locations since they have no aliases. Singletons are helpful in narrowing |
| * down the life span of a heap location such that they do not always |
| * need to participate in merging heap values. Allocation of a singleton |
| * can be eliminated if that singleton is not used and does not persist |
| * at method return/deoptimization. |
| * - For newly instantiated instances, their heap values are initialized to |
| * language defined default values. |
| * - Some instructions such as invokes are treated as loading and invalidating |
| * all the heap values, depending on the instruction's side effects. |
| * - Finalizable objects are considered as persisting at method |
| * return/deoptimization. |
| * - Currently this LSE algorithm doesn't handle SIMD graph, e.g. with VecLoad |
| * and VecStore instructions. |
| * - Currently this LSE algorithm doesn't handle graph with try-catch, due to |
| * the special block merging structure. |
| */ |
| |
| namespace art { |
| |
| // An unknown heap value. Loads with such a value in the heap location cannot be eliminated. |
| // A heap location can be set to kUnknownHeapValue when: |
| // - initially set a value. |
| // - killed due to aliasing, merging, invocation, or loop side effects. |
| static HInstruction* const kUnknownHeapValue = |
| reinterpret_cast<HInstruction*>(static_cast<uintptr_t>(-1)); |
| |
| // Default heap value after an allocation. |
| // A heap location can be set to that value right after an allocation. |
| static HInstruction* const kDefaultHeapValue = |
| reinterpret_cast<HInstruction*>(static_cast<uintptr_t>(-2)); |
| |
| // Use HGraphDelegateVisitor for which all VisitInvokeXXX() delegate to VisitInvoke(). |
| class LSEVisitor : public HGraphDelegateVisitor { |
| public: |
| LSEVisitor(HGraph* graph, |
| const HeapLocationCollector& heap_locations_collector, |
| const SideEffectsAnalysis& side_effects, |
| OptimizingCompilerStats* stats) |
| : HGraphDelegateVisitor(graph, stats), |
| heap_location_collector_(heap_locations_collector), |
| side_effects_(side_effects), |
| allocator_(graph->GetArenaStack()), |
| heap_values_for_(graph->GetBlocks().size(), |
| ScopedArenaVector<HInstruction*>(heap_locations_collector. |
| GetNumberOfHeapLocations(), |
| kUnknownHeapValue, |
| allocator_.Adapter(kArenaAllocLSE)), |
| allocator_.Adapter(kArenaAllocLSE)), |
| removed_loads_(allocator_.Adapter(kArenaAllocLSE)), |
| substitute_instructions_for_loads_(allocator_.Adapter(kArenaAllocLSE)), |
| possibly_removed_stores_(allocator_.Adapter(kArenaAllocLSE)), |
| singleton_new_instances_(allocator_.Adapter(kArenaAllocLSE)) { |
| } |
| |
| void VisitBasicBlock(HBasicBlock* block) override { |
| // Populate the heap_values array for this block. |
| // TODO: try to reuse the heap_values array from one predecessor if possible. |
| if (block->IsLoopHeader()) { |
| HandleLoopSideEffects(block); |
| } else { |
| MergePredecessorValues(block); |
| } |
| HGraphVisitor::VisitBasicBlock(block); |
| } |
| |
| HTypeConversion* AddTypeConversionIfNecessary(HInstruction* instruction, |
| HInstruction* value, |
| DataType::Type expected_type) { |
| HTypeConversion* type_conversion = nullptr; |
| // Should never add type conversion into boolean value. |
| if (expected_type != DataType::Type::kBool && |
| !DataType::IsTypeConversionImplicit(value->GetType(), expected_type)) { |
| type_conversion = new (GetGraph()->GetAllocator()) HTypeConversion( |
| expected_type, value, instruction->GetDexPc()); |
| instruction->GetBlock()->InsertInstructionBefore(type_conversion, instruction); |
| } |
| return type_conversion; |
| } |
| |
| // Find an instruction's substitute if it's a removed load. |
| // Return the same instruction if it should not be removed. |
| HInstruction* FindSubstitute(HInstruction* instruction) { |
| if (!IsLoad(instruction)) { |
| return instruction; |
| } |
| size_t size = removed_loads_.size(); |
| for (size_t i = 0; i < size; i++) { |
| if (removed_loads_[i] == instruction) { |
| HInstruction* substitute = substitute_instructions_for_loads_[i]; |
| // The substitute list is a flat hierarchy. |
| DCHECK_EQ(FindSubstitute(substitute), substitute); |
| return substitute; |
| } |
| } |
| return instruction; |
| } |
| |
| void AddRemovedLoad(HInstruction* load, HInstruction* heap_value) { |
| DCHECK(IsLoad(load)); |
| DCHECK_EQ(FindSubstitute(heap_value), heap_value) << |
| "Unexpected heap_value that has a substitute " << heap_value->DebugName(); |
| removed_loads_.push_back(load); |
| substitute_instructions_for_loads_.push_back(heap_value); |
| } |
| |
| // Scan the list of removed loads to see if we can reuse `type_conversion`, if |
| // the other removed load has the same substitute and type and is dominated |
| // by `type_conversion`. |
| void TryToReuseTypeConversion(HInstruction* type_conversion, size_t index) { |
| size_t size = removed_loads_.size(); |
| HInstruction* load = removed_loads_[index]; |
| HInstruction* substitute = substitute_instructions_for_loads_[index]; |
| for (size_t j = index + 1; j < size; j++) { |
| HInstruction* load2 = removed_loads_[j]; |
| HInstruction* substitute2 = substitute_instructions_for_loads_[j]; |
| if (load2 == nullptr) { |
| DCHECK(substitute2->IsTypeConversion()); |
| continue; |
| } |
| DCHECK(load2->IsInstanceFieldGet() || |
| load2->IsStaticFieldGet() || |
| load2->IsArrayGet()); |
| DCHECK(substitute2 != nullptr); |
| if (substitute2 == substitute && |
| load2->GetType() == load->GetType() && |
| type_conversion->GetBlock()->Dominates(load2->GetBlock()) && |
| // Don't share across irreducible loop headers. |
| // TODO: can be more fine-grained than this by testing each dominator. |
| (load2->GetBlock() == type_conversion->GetBlock() || |
| !GetGraph()->HasIrreducibleLoops())) { |
| // The removed_loads_ are added in reverse post order. |
| DCHECK(type_conversion->StrictlyDominates(load2)); |
| load2->ReplaceWith(type_conversion); |
| load2->GetBlock()->RemoveInstruction(load2); |
| removed_loads_[j] = nullptr; |
| substitute_instructions_for_loads_[j] = type_conversion; |
| } |
| } |
| } |
| |
| // Remove recorded instructions that should be eliminated. |
| void RemoveInstructions() { |
| size_t size = removed_loads_.size(); |
| DCHECK_EQ(size, substitute_instructions_for_loads_.size()); |
| for (size_t i = 0; i < size; i++) { |
| HInstruction* load = removed_loads_[i]; |
| if (load == nullptr) { |
| // The load has been handled in the scan for type conversion below. |
| DCHECK(substitute_instructions_for_loads_[i]->IsTypeConversion()); |
| continue; |
| } |
| DCHECK(load->IsInstanceFieldGet() || |
| load->IsStaticFieldGet() || |
| load->IsArrayGet()); |
| HInstruction* substitute = substitute_instructions_for_loads_[i]; |
| DCHECK(substitute != nullptr); |
| // We proactively retrieve the substitute for a removed load, so |
| // a load that has a substitute should not be observed as a heap |
| // location value. |
| DCHECK_EQ(FindSubstitute(substitute), substitute); |
| |
| // The load expects to load the heap value as type load->GetType(). |
| // However the tracked heap value may not be of that type. An explicit |
| // type conversion may be needed. |
| // There are actually three types involved here: |
| // (1) tracked heap value's type (type A) |
| // (2) heap location (field or element)'s type (type B) |
| // (3) load's type (type C) |
| // We guarantee that type A stored as type B and then fetched out as |
| // type C is the same as casting from type A to type C directly, since |
| // type B and type C will have the same size which is guarenteed in |
| // HInstanceFieldGet/HStaticFieldGet/HArrayGet's SetType(). |
| // So we only need one type conversion from type A to type C. |
| HTypeConversion* type_conversion = AddTypeConversionIfNecessary( |
| load, substitute, load->GetType()); |
| if (type_conversion != nullptr) { |
| TryToReuseTypeConversion(type_conversion, i); |
| load->ReplaceWith(type_conversion); |
| substitute_instructions_for_loads_[i] = type_conversion; |
| } else { |
| load->ReplaceWith(substitute); |
| } |
| load->GetBlock()->RemoveInstruction(load); |
| } |
| |
| // At this point, stores in possibly_removed_stores_ can be safely removed. |
| for (HInstruction* store : possibly_removed_stores_) { |
| DCHECK(store->IsInstanceFieldSet() || store->IsStaticFieldSet() || store->IsArraySet()); |
| store->GetBlock()->RemoveInstruction(store); |
| } |
| |
| // Eliminate singleton-classified instructions: |
| // * - Constructor fences (they never escape this thread). |
| // * - Allocations (if they are unused). |
| for (HInstruction* new_instance : singleton_new_instances_) { |
| size_t removed = HConstructorFence::RemoveConstructorFences(new_instance); |
| MaybeRecordStat(stats_, |
| MethodCompilationStat::kConstructorFenceRemovedLSE, |
| removed); |
| |
| if (!new_instance->HasNonEnvironmentUses()) { |
| new_instance->RemoveEnvironmentUsers(); |
| new_instance->GetBlock()->RemoveInstruction(new_instance); |
| } |
| } |
| } |
| |
| private: |
| static bool IsLoad(HInstruction* instruction) { |
| if (instruction == kUnknownHeapValue || instruction == kDefaultHeapValue) { |
| return false; |
| } |
| // Unresolved load is not treated as a load. |
| return instruction->IsInstanceFieldGet() || |
| instruction->IsStaticFieldGet() || |
| instruction->IsArrayGet(); |
| } |
| |
| static bool IsStore(HInstruction* instruction) { |
| if (instruction == kUnknownHeapValue || instruction == kDefaultHeapValue) { |
| return false; |
| } |
| // Unresolved store is not treated as a store. |
| return instruction->IsInstanceFieldSet() || |
| instruction->IsArraySet() || |
| instruction->IsStaticFieldSet(); |
| } |
| |
| // Returns the real heap value by finding its substitute or by "peeling" |
| // a store instruction. |
| HInstruction* GetRealHeapValue(HInstruction* heap_value) { |
| if (IsLoad(heap_value)) { |
| return FindSubstitute(heap_value); |
| } |
| if (!IsStore(heap_value)) { |
| return heap_value; |
| } |
| |
| // We keep track of store instructions as the heap values which might be |
| // eliminated if the stores are later found not necessary. The real stored |
| // value needs to be fetched from the store instruction. |
| if (heap_value->IsInstanceFieldSet()) { |
| heap_value = heap_value->AsInstanceFieldSet()->GetValue(); |
| } else if (heap_value->IsStaticFieldSet()) { |
| heap_value = heap_value->AsStaticFieldSet()->GetValue(); |
| } else { |
| DCHECK(heap_value->IsArraySet()); |
| heap_value = heap_value->AsArraySet()->GetValue(); |
| } |
| // heap_value may already be a removed load. |
| return FindSubstitute(heap_value); |
| } |
| |
| // If heap_value is a store, need to keep the store. |
| // This is necessary if a heap value is killed or replaced by another value, |
| // so that the store is no longer used to track heap value. |
| void KeepIfIsStore(HInstruction* heap_value) { |
| if (!IsStore(heap_value)) { |
| return; |
| } |
| auto idx = std::find(possibly_removed_stores_.begin(), |
| possibly_removed_stores_.end(), heap_value); |
| if (idx != possibly_removed_stores_.end()) { |
| // Make sure the store is kept. |
| possibly_removed_stores_.erase(idx); |
| } |
| } |
| |
| // If a heap location X may alias with heap location at `loc_index` |
| // and heap_values of that heap location X holds a store, keep that store. |
| // It's needed for a dependent load that's not eliminated since any store |
| // that may put value into the load's heap location needs to be kept. |
| void KeepStoresIfAliasedToLocation(ScopedArenaVector<HInstruction*>& heap_values, |
| size_t loc_index) { |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| if ((i == loc_index) || heap_location_collector_.MayAlias(i, loc_index)) { |
| KeepIfIsStore(heap_values[i]); |
| } |
| } |
| } |
| |
| void HandleLoopSideEffects(HBasicBlock* block) { |
| DCHECK(block->IsLoopHeader()); |
| int block_id = block->GetBlockId(); |
| ScopedArenaVector<HInstruction*>& heap_values = heap_values_for_[block_id]; |
| HBasicBlock* pre_header = block->GetLoopInformation()->GetPreHeader(); |
| ScopedArenaVector<HInstruction*>& pre_header_heap_values = |
| heap_values_for_[pre_header->GetBlockId()]; |
| |
| // Don't eliminate loads in irreducible loops. |
| // Also keep the stores before the loop. |
| if (block->GetLoopInformation()->IsIrreducible()) { |
| if (kIsDebugBuild) { |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| DCHECK_EQ(heap_values[i], kUnknownHeapValue); |
| } |
| } |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| KeepIfIsStore(pre_header_heap_values[i]); |
| } |
| return; |
| } |
| |
| // Inherit the values from pre-header. |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| heap_values[i] = pre_header_heap_values[i]; |
| } |
| |
| // We do a single pass in reverse post order. For loops, use the side effects as a hint |
| // to see if the heap values should be killed. |
| if (side_effects_.GetLoopEffects(block).DoesAnyWrite()) { |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| HeapLocation* location = heap_location_collector_.GetHeapLocation(i); |
| ReferenceInfo* ref_info = location->GetReferenceInfo(); |
| if (ref_info->IsSingleton() && !location->IsValueKilledByLoopSideEffects()) { |
| // A singleton's field that's not stored into inside a loop is |
| // invariant throughout the loop. Nothing to do. |
| } else { |
| // heap value is killed by loop side effects. |
| KeepIfIsStore(pre_header_heap_values[i]); |
| heap_values[i] = kUnknownHeapValue; |
| } |
| } |
| } else { |
| // The loop doesn't kill any value. |
| } |
| } |
| |
| void MergePredecessorValues(HBasicBlock* block) { |
| ArrayRef<HBasicBlock* const> predecessors(block->GetPredecessors()); |
| if (predecessors.size() == 0) { |
| return; |
| } |
| if (block->IsExitBlock()) { |
| // Exit block doesn't really merge values since the control flow ends in |
| // its predecessors. Each predecessor needs to make sure stores are kept |
| // if necessary. |
| return; |
| } |
| |
| ScopedArenaVector<HInstruction*>& heap_values = heap_values_for_[block->GetBlockId()]; |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| HInstruction* merged_value = nullptr; |
| // If we can merge the store itself from the predecessors, we keep |
| // the store as the heap value as long as possible. In case we cannot |
| // merge the store, we try to merge the values of the stores. |
| HInstruction* merged_store_value = nullptr; |
| // Whether merged_value is a result that's merged from all predecessors. |
| bool from_all_predecessors = true; |
| ReferenceInfo* ref_info = heap_location_collector_.GetHeapLocation(i)->GetReferenceInfo(); |
| HInstruction* ref = ref_info->GetReference(); |
| HInstruction* singleton_ref = nullptr; |
| if (ref_info->IsSingleton()) { |
| // We do more analysis based on singleton's liveness when merging |
| // heap values for such cases. |
| singleton_ref = ref; |
| } |
| |
| for (HBasicBlock* predecessor : predecessors) { |
| HInstruction* pred_value = heap_values_for_[predecessor->GetBlockId()][i]; |
| if (!IsStore(pred_value)) { |
| pred_value = FindSubstitute(pred_value); |
| } |
| DCHECK(pred_value != nullptr); |
| HInstruction* pred_store_value = GetRealHeapValue(pred_value); |
| if ((singleton_ref != nullptr) && |
| !singleton_ref->GetBlock()->Dominates(predecessor)) { |
| // singleton_ref is not live in this predecessor. No need to merge |
| // since singleton_ref is not live at the beginning of this block. |
| DCHECK_EQ(pred_value, kUnknownHeapValue); |
| from_all_predecessors = false; |
| break; |
| } |
| if (merged_value == nullptr) { |
| // First seen heap value. |
| DCHECK(pred_value != nullptr); |
| merged_value = pred_value; |
| } else if (pred_value != merged_value) { |
| // There are conflicting values. |
| merged_value = kUnknownHeapValue; |
| // We may still be able to merge store values. |
| } |
| |
| // Conflicting stores may be storing the same value. We do another merge |
| // of real stored values. |
| if (merged_store_value == nullptr) { |
| // First seen store value. |
| DCHECK(pred_store_value != nullptr); |
| merged_store_value = pred_store_value; |
| } else if (pred_store_value != merged_store_value) { |
| // There are conflicting store values. |
| merged_store_value = kUnknownHeapValue; |
| // There must be conflicting stores also. |
| DCHECK_EQ(merged_value, kUnknownHeapValue); |
| // No need to merge anymore. |
| break; |
| } |
| } |
| |
| if (merged_value == nullptr) { |
| DCHECK(!from_all_predecessors); |
| DCHECK(singleton_ref != nullptr); |
| } |
| if (from_all_predecessors) { |
| if (ref_info->IsSingletonAndRemovable() && |
| (block->IsSingleReturnOrReturnVoidAllowingPhis() || |
| (block->EndsWithReturn() && (merged_value != kUnknownHeapValue || |
| merged_store_value != kUnknownHeapValue)))) { |
| // Values in the singleton are not needed anymore: |
| // (1) if this block consists of a sole return, or |
| // (2) if this block returns and a usable merged value is obtained |
| // (loads prior to the return will always use that value). |
| } else if (!IsStore(merged_value)) { |
| // We don't track merged value as a store anymore. We have to |
| // hold the stores in predecessors live here. |
| for (HBasicBlock* predecessor : predecessors) { |
| ScopedArenaVector<HInstruction*>& pred_values = |
| heap_values_for_[predecessor->GetBlockId()]; |
| KeepIfIsStore(pred_values[i]); |
| } |
| } |
| } else { |
| DCHECK(singleton_ref != nullptr); |
| // singleton_ref is non-existing at the beginning of the block. There is |
| // no need to keep the stores. |
| } |
| |
| if (!from_all_predecessors) { |
| DCHECK(singleton_ref != nullptr); |
| DCHECK((singleton_ref->GetBlock() == block) || |
| !singleton_ref->GetBlock()->Dominates(block)) |
| << "method: " << GetGraph()->GetMethodName(); |
| // singleton_ref is not defined before block or defined only in some of its |
| // predecessors, so block doesn't really have the location at its entry. |
| heap_values[i] = kUnknownHeapValue; |
| } else if (predecessors.size() == 1) { |
| // Inherit heap value from the single predecessor. |
| DCHECK_EQ(heap_values_for_[predecessors[0]->GetBlockId()][i], merged_value); |
| heap_values[i] = merged_value; |
| } else { |
| DCHECK(merged_value == kUnknownHeapValue || |
| merged_value == kDefaultHeapValue || |
| merged_value->GetBlock()->Dominates(block)); |
| if (merged_value != kUnknownHeapValue) { |
| heap_values[i] = merged_value; |
| } else { |
| // Stores in different predecessors may be storing the same value. |
| heap_values[i] = merged_store_value; |
| } |
| } |
| } |
| } |
| |
| // `instruction` is being removed. Try to see if the null check on it |
| // can be removed. This can happen if the same value is set in two branches |
| // but not in dominators. Such as: |
| // int[] a = foo(); |
| // if () { |
| // a[0] = 2; |
| // } else { |
| // a[0] = 2; |
| // } |
| // // a[0] can now be replaced with constant 2, and the null check on it can be removed. |
| void TryRemovingNullCheck(HInstruction* instruction) { |
| HInstruction* prev = instruction->GetPrevious(); |
| if ((prev != nullptr) && prev->IsNullCheck() && (prev == instruction->InputAt(0))) { |
| // Previous instruction is a null check for this instruction. Remove the null check. |
| prev->ReplaceWith(prev->InputAt(0)); |
| prev->GetBlock()->RemoveInstruction(prev); |
| } |
| } |
| |
| HInstruction* GetDefaultValue(DataType::Type type) { |
| switch (type) { |
| case DataType::Type::kReference: |
| return GetGraph()->GetNullConstant(); |
| case DataType::Type::kBool: |
| case DataType::Type::kUint8: |
| case DataType::Type::kInt8: |
| case DataType::Type::kUint16: |
| case DataType::Type::kInt16: |
| case DataType::Type::kInt32: |
| return GetGraph()->GetIntConstant(0); |
| case DataType::Type::kInt64: |
| return GetGraph()->GetLongConstant(0); |
| case DataType::Type::kFloat32: |
| return GetGraph()->GetFloatConstant(0); |
| case DataType::Type::kFloat64: |
| return GetGraph()->GetDoubleConstant(0); |
| default: |
| UNREACHABLE(); |
| } |
| } |
| |
| void VisitGetLocation(HInstruction* instruction, size_t idx) { |
| DCHECK_NE(idx, HeapLocationCollector::kHeapLocationNotFound); |
| ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[instruction->GetBlock()->GetBlockId()]; |
| HInstruction* heap_value = heap_values[idx]; |
| if (heap_value == kDefaultHeapValue) { |
| HInstruction* constant = GetDefaultValue(instruction->GetType()); |
| AddRemovedLoad(instruction, constant); |
| heap_values[idx] = constant; |
| return; |
| } |
| heap_value = GetRealHeapValue(heap_value); |
| if (heap_value == kUnknownHeapValue) { |
| // Load isn't eliminated. Put the load as the value into the HeapLocation. |
| // This acts like GVN but with better aliasing analysis. |
| heap_values[idx] = instruction; |
| KeepStoresIfAliasedToLocation(heap_values, idx); |
| } else { |
| // Load is eliminated. |
| AddRemovedLoad(instruction, heap_value); |
| TryRemovingNullCheck(instruction); |
| } |
| } |
| |
| bool Equal(HInstruction* heap_value, HInstruction* value) { |
| DCHECK(!IsStore(value)) << value->DebugName(); |
| if (heap_value == kUnknownHeapValue) { |
| // Don't compare kUnknownHeapValue with other values. |
| return false; |
| } |
| if (heap_value == value) { |
| return true; |
| } |
| if (heap_value == kDefaultHeapValue && GetDefaultValue(value->GetType()) == value) { |
| return true; |
| } |
| HInstruction* real_heap_value = GetRealHeapValue(heap_value); |
| if (real_heap_value != heap_value) { |
| return Equal(real_heap_value, value); |
| } |
| return false; |
| } |
| |
| void VisitSetLocation(HInstruction* instruction, size_t idx, HInstruction* value) { |
| DCHECK_NE(idx, HeapLocationCollector::kHeapLocationNotFound); |
| DCHECK(!IsStore(value)) << value->DebugName(); |
| // value may already have a substitute. |
| value = FindSubstitute(value); |
| ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[instruction->GetBlock()->GetBlockId()]; |
| HInstruction* heap_value = heap_values[idx]; |
| bool possibly_redundant = false; |
| |
| if (Equal(heap_value, value)) { |
| // Store into the heap location with the same value. |
| // This store can be eliminated right away. |
| instruction->GetBlock()->RemoveInstruction(instruction); |
| return; |
| } else { |
| HLoopInformation* loop_info = instruction->GetBlock()->GetLoopInformation(); |
| if (loop_info == nullptr) { |
| // Store is not in a loop. We try to precisely track the heap value by |
| // the store. |
| possibly_redundant = true; |
| } else if (!loop_info->IsIrreducible()) { |
| // instruction is a store in the loop so the loop must do write. |
| DCHECK(side_effects_.GetLoopEffects(loop_info->GetHeader()).DoesAnyWrite()); |
| ReferenceInfo* ref_info = heap_location_collector_.GetHeapLocation(idx)->GetReferenceInfo(); |
| if (ref_info->IsSingleton() && !loop_info->IsDefinedOutOfTheLoop(ref_info->GetReference())) { |
| // original_ref is created inside the loop. Value stored to it isn't needed at |
| // the loop header. This is true for outer loops also. |
| possibly_redundant = true; |
| } else { |
| // Keep the store since its value may be needed at the loop header. |
| } |
| } else { |
| // Keep the store inside irreducible loops. |
| } |
| } |
| if (possibly_redundant) { |
| possibly_removed_stores_.push_back(instruction); |
| } |
| |
| // Put the store as the heap value. If the value is loaded or needed after |
| // return/deoptimization later, this store isn't really redundant. |
| heap_values[idx] = instruction; |
| |
| // This store may kill values in other heap locations due to aliasing. |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| if (i == idx) { |
| continue; |
| } |
| if (Equal(heap_values[i], value)) { |
| // Same value should be kept even if aliasing happens. |
| continue; |
| } |
| if (heap_values[i] == kUnknownHeapValue) { |
| // Value is already unknown, no need for aliasing check. |
| continue; |
| } |
| if (heap_location_collector_.MayAlias(i, idx)) { |
| // Kill heap locations that may alias and as a result if the heap value |
| // is a store, the store needs to be kept. |
| KeepIfIsStore(heap_values[i]); |
| heap_values[i] = kUnknownHeapValue; |
| } |
| } |
| } |
| |
| void VisitInstanceFieldGet(HInstanceFieldGet* instruction) override { |
| HInstruction* object = instruction->InputAt(0); |
| const FieldInfo& field = instruction->GetFieldInfo(); |
| VisitGetLocation(instruction, heap_location_collector_.GetFieldHeapLocation(object, &field)); |
| } |
| |
| void VisitInstanceFieldSet(HInstanceFieldSet* instruction) override { |
| HInstruction* object = instruction->InputAt(0); |
| const FieldInfo& field = instruction->GetFieldInfo(); |
| HInstruction* value = instruction->InputAt(1); |
| size_t idx = heap_location_collector_.GetFieldHeapLocation(object, &field); |
| VisitSetLocation(instruction, idx, value); |
| } |
| |
| void VisitStaticFieldGet(HStaticFieldGet* instruction) override { |
| HInstruction* cls = instruction->InputAt(0); |
| const FieldInfo& field = instruction->GetFieldInfo(); |
| VisitGetLocation(instruction, heap_location_collector_.GetFieldHeapLocation(cls, &field)); |
| } |
| |
| void VisitStaticFieldSet(HStaticFieldSet* instruction) override { |
| HInstruction* cls = instruction->InputAt(0); |
| const FieldInfo& field = instruction->GetFieldInfo(); |
| size_t idx = heap_location_collector_.GetFieldHeapLocation(cls, &field); |
| VisitSetLocation(instruction, idx, instruction->InputAt(1)); |
| } |
| |
| void VisitArrayGet(HArrayGet* instruction) override { |
| VisitGetLocation(instruction, heap_location_collector_.GetArrayHeapLocation(instruction)); |
| } |
| |
| void VisitArraySet(HArraySet* instruction) override { |
| size_t idx = heap_location_collector_.GetArrayHeapLocation(instruction); |
| VisitSetLocation(instruction, idx, instruction->InputAt(2)); |
| } |
| |
| void VisitDeoptimize(HDeoptimize* instruction) override { |
| const ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[instruction->GetBlock()->GetBlockId()]; |
| for (HInstruction* heap_value : heap_values) { |
| // A store is kept as the heap value for possibly removed stores. |
| // That value stored is generally observeable after deoptimization, except |
| // for singletons that don't escape after deoptimization. |
| if (IsStore(heap_value)) { |
| if (heap_value->IsStaticFieldSet()) { |
| KeepIfIsStore(heap_value); |
| continue; |
| } |
| HInstruction* reference = heap_value->InputAt(0); |
| if (heap_location_collector_.FindReferenceInfoOf(reference)->IsSingleton()) { |
| if (reference->IsNewInstance() && reference->AsNewInstance()->IsFinalizable()) { |
| // Finalizable objects alway escape. |
| KeepIfIsStore(heap_value); |
| continue; |
| } |
| // Check whether the reference for a store is used by an environment local of |
| // HDeoptimize. If not, the singleton is not observed after |
| // deoptimizion. |
| for (const HUseListNode<HEnvironment*>& use : reference->GetEnvUses()) { |
| HEnvironment* user = use.GetUser(); |
| if (user->GetHolder() == instruction) { |
| // The singleton for the store is visible at this deoptimization |
| // point. Need to keep the store so that the heap value is |
| // seen by the interpreter. |
| KeepIfIsStore(heap_value); |
| } |
| } |
| } else { |
| KeepIfIsStore(heap_value); |
| } |
| } |
| } |
| } |
| |
| // Keep necessary stores before exiting a method via return/throw. |
| void HandleExit(HBasicBlock* block) { |
| const ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[block->GetBlockId()]; |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| HInstruction* heap_value = heap_values[i]; |
| ReferenceInfo* ref_info = heap_location_collector_.GetHeapLocation(i)->GetReferenceInfo(); |
| if (!ref_info->IsSingletonAndRemovable()) { |
| KeepIfIsStore(heap_value); |
| } |
| } |
| } |
| |
| void VisitReturn(HReturn* instruction) override { |
| HandleExit(instruction->GetBlock()); |
| } |
| |
| void VisitReturnVoid(HReturnVoid* return_void) override { |
| HandleExit(return_void->GetBlock()); |
| } |
| |
| void VisitThrow(HThrow* throw_instruction) override { |
| HandleExit(throw_instruction->GetBlock()); |
| } |
| |
| void HandleInvoke(HInstruction* instruction) { |
| SideEffects side_effects = instruction->GetSideEffects(); |
| ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[instruction->GetBlock()->GetBlockId()]; |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| ReferenceInfo* ref_info = heap_location_collector_.GetHeapLocation(i)->GetReferenceInfo(); |
| if (ref_info->IsSingleton()) { |
| // Singleton references cannot be seen by the callee. |
| } else { |
| if (side_effects.DoesAnyRead()) { |
| // Invocation may read the heap value. |
| KeepIfIsStore(heap_values[i]); |
| } |
| if (side_effects.DoesAnyWrite()) { |
| // Keep the store since it's not used to track the heap value anymore. |
| KeepIfIsStore(heap_values[i]); |
| heap_values[i] = kUnknownHeapValue; |
| } |
| } |
| } |
| } |
| |
| void VisitInvoke(HInvoke* invoke) override { |
| HandleInvoke(invoke); |
| } |
| |
| void VisitClinitCheck(HClinitCheck* clinit) override { |
| HandleInvoke(clinit); |
| } |
| |
| void VisitUnresolvedInstanceFieldGet(HUnresolvedInstanceFieldGet* instruction) override { |
| // Conservatively treat it as an invocation. |
| HandleInvoke(instruction); |
| } |
| |
| void VisitUnresolvedInstanceFieldSet(HUnresolvedInstanceFieldSet* instruction) override { |
| // Conservatively treat it as an invocation. |
| HandleInvoke(instruction); |
| } |
| |
| void VisitUnresolvedStaticFieldGet(HUnresolvedStaticFieldGet* instruction) override { |
| // Conservatively treat it as an invocation. |
| HandleInvoke(instruction); |
| } |
| |
| void VisitUnresolvedStaticFieldSet(HUnresolvedStaticFieldSet* instruction) override { |
| // Conservatively treat it as an invocation. |
| HandleInvoke(instruction); |
| } |
| |
| void VisitNewInstance(HNewInstance* new_instance) override { |
| ReferenceInfo* ref_info = heap_location_collector_.FindReferenceInfoOf(new_instance); |
| if (ref_info == nullptr) { |
| // new_instance isn't used for field accesses. No need to process it. |
| return; |
| } |
| if (ref_info->IsSingletonAndRemovable() && !new_instance->NeedsChecks()) { |
| DCHECK(!new_instance->IsFinalizable()); |
| // new_instance can potentially be eliminated. |
| singleton_new_instances_.push_back(new_instance); |
| } |
| ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[new_instance->GetBlock()->GetBlockId()]; |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| HInstruction* ref = |
| heap_location_collector_.GetHeapLocation(i)->GetReferenceInfo()->GetReference(); |
| size_t offset = heap_location_collector_.GetHeapLocation(i)->GetOffset(); |
| if (ref == new_instance && offset >= mirror::kObjectHeaderSize) { |
| // Instance fields except the header fields are set to default heap values. |
| heap_values[i] = kDefaultHeapValue; |
| } |
| } |
| } |
| |
| void VisitNewArray(HNewArray* new_array) override { |
| ReferenceInfo* ref_info = heap_location_collector_.FindReferenceInfoOf(new_array); |
| if (ref_info == nullptr) { |
| // new_array isn't used for array accesses. No need to process it. |
| return; |
| } |
| if (ref_info->IsSingletonAndRemovable()) { |
| if (new_array->GetLength()->IsIntConstant() && |
| new_array->GetLength()->AsIntConstant()->GetValue() >= 0) { |
| // new_array can potentially be eliminated. |
| singleton_new_instances_.push_back(new_array); |
| } else { |
| // new_array may throw NegativeArraySizeException. Keep it. |
| } |
| } |
| ScopedArenaVector<HInstruction*>& heap_values = |
| heap_values_for_[new_array->GetBlock()->GetBlockId()]; |
| for (size_t i = 0; i < heap_values.size(); i++) { |
| HeapLocation* location = heap_location_collector_.GetHeapLocation(i); |
| HInstruction* ref = location->GetReferenceInfo()->GetReference(); |
| if (ref == new_array && location->GetIndex() != nullptr) { |
| // Array elements are set to default heap values. |
| heap_values[i] = kDefaultHeapValue; |
| } |
| } |
| } |
| |
| const HeapLocationCollector& heap_location_collector_; |
| const SideEffectsAnalysis& side_effects_; |
| |
| // Use local allocator for allocating memory. |
| ScopedArenaAllocator allocator_; |
| |
| // One array of heap values for each block. |
| ScopedArenaVector<ScopedArenaVector<HInstruction*>> heap_values_for_; |
| |
| // We record the instructions that should be eliminated but may be |
| // used by heap locations. They'll be removed in the end. |
| ScopedArenaVector<HInstruction*> removed_loads_; |
| ScopedArenaVector<HInstruction*> substitute_instructions_for_loads_; |
| |
| // Stores in this list may be removed from the list later when it's |
| // found that the store cannot be eliminated. |
| ScopedArenaVector<HInstruction*> possibly_removed_stores_; |
| |
| ScopedArenaVector<HInstruction*> singleton_new_instances_; |
| |
| DISALLOW_COPY_AND_ASSIGN(LSEVisitor); |
| }; |
| |
| bool LoadStoreElimination::Run() { |
| if (graph_->IsDebuggable() || graph_->HasTryCatch()) { |
| // Debugger may set heap values or trigger deoptimization of callers. |
| // Try/catch support not implemented yet. |
| // Skip this optimization. |
| return false; |
| } |
| const HeapLocationCollector& heap_location_collector = lsa_.GetHeapLocationCollector(); |
| if (heap_location_collector.GetNumberOfHeapLocations() == 0) { |
| // No HeapLocation information from LSA, skip this optimization. |
| return false; |
| } |
| |
| // TODO: analyze VecLoad/VecStore better. |
| if (graph_->HasSIMD()) { |
| return false; |
| } |
| |
| LSEVisitor lse_visitor(graph_, heap_location_collector, side_effects_, stats_); |
| for (HBasicBlock* block : graph_->GetReversePostOrder()) { |
| lse_visitor.VisitBasicBlock(block); |
| } |
| lse_visitor.RemoveInstructions(); |
| |
| return true; |
| } |
| |
| } // namespace art |