blob: db776eaa1acbc33c68727d731f97aa231364aa8b [file] [log] [blame]
/*
* Copyright (C) 2014 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_STACK_MAP_H_
#define ART_RUNTIME_STACK_MAP_H_
#include <limits>
#include "arch/code_offset.h"
#include "base/bit_utils.h"
#include "base/bit_vector.h"
#include "bit_memory_region.h"
#include "leb128.h"
#include "dex_file_types.h"
#include "memory_region.h"
#include "method_info.h"
namespace art {
class VariableIndentationOutputStream;
// Size of a frame slot, in bytes. This constant is a signed value,
// to please the compiler in arithmetic operations involving int32_t
// (signed) values.
static constexpr ssize_t kFrameSlotSize = 4;
// Size of Dex virtual registers.
static constexpr size_t kVRegSize = 4;
class ArtMethod;
class CodeInfo;
class StackMapEncoding;
struct CodeInfoEncoding;
/**
* Classes in the following file are wrapper on stack map information backed
* by a MemoryRegion. As such they read and write to the region, they don't have
* their own fields.
*/
// Dex register location container used by DexRegisterMap and StackMapStream.
class DexRegisterLocation {
public:
/*
* The location kind used to populate the Dex register information in a
* StackMapStream can either be:
* - kStack: vreg stored on the stack, value holds the stack offset;
* - kInRegister: vreg stored in low 32 bits of a core physical register,
* value holds the register number;
* - kInRegisterHigh: vreg stored in high 32 bits of a core physical register,
* value holds the register number;
* - kInFpuRegister: vreg stored in low 32 bits of an FPU register,
* value holds the register number;
* - kInFpuRegisterHigh: vreg stored in high 32 bits of an FPU register,
* value holds the register number;
* - kConstant: value holds the constant;
*
* In addition, DexRegisterMap also uses these values:
* - kInStackLargeOffset: value holds a "large" stack offset (greater than
* or equal to 128 bytes);
* - kConstantLargeValue: value holds a "large" constant (lower than 0, or
* or greater than or equal to 32);
* - kNone: the register has no location, meaning it has not been set.
*/
enum class Kind : uint8_t {
// Short location kinds, for entries fitting on one byte (3 bits
// for the kind, 5 bits for the value) in a DexRegisterMap.
kInStack = 0, // 0b000
kInRegister = 1, // 0b001
kInRegisterHigh = 2, // 0b010
kInFpuRegister = 3, // 0b011
kInFpuRegisterHigh = 4, // 0b100
kConstant = 5, // 0b101
// Large location kinds, requiring a 5-byte encoding (1 byte for the
// kind, 4 bytes for the value).
// Stack location at a large offset, meaning that the offset value
// divided by the stack frame slot size (4 bytes) cannot fit on a
// 5-bit unsigned integer (i.e., this offset value is greater than
// or equal to 2^5 * 4 = 128 bytes).
kInStackLargeOffset = 6, // 0b110
// Large constant, that cannot fit on a 5-bit signed integer (i.e.,
// lower than 0, or greater than or equal to 2^5 = 32).
kConstantLargeValue = 7, // 0b111
// Entries with no location are not stored and do not need own marker.
kNone = static_cast<uint8_t>(-1),
kLastLocationKind = kConstantLargeValue
};
static_assert(
sizeof(Kind) == 1u,
"art::DexRegisterLocation::Kind has a size different from one byte.");
static bool IsShortLocationKind(Kind kind) {
switch (kind) {
case Kind::kInStack:
case Kind::kInRegister:
case Kind::kInRegisterHigh:
case Kind::kInFpuRegister:
case Kind::kInFpuRegisterHigh:
case Kind::kConstant:
return true;
case Kind::kInStackLargeOffset:
case Kind::kConstantLargeValue:
return false;
case Kind::kNone:
LOG(FATAL) << "Unexpected location kind";
}
UNREACHABLE();
}
// Convert `kind` to a "surface" kind, i.e. one that doesn't include
// any value with a "large" qualifier.
// TODO: Introduce another enum type for the surface kind?
static Kind ConvertToSurfaceKind(Kind kind) {
switch (kind) {
case Kind::kInStack:
case Kind::kInRegister:
case Kind::kInRegisterHigh:
case Kind::kInFpuRegister:
case Kind::kInFpuRegisterHigh:
case Kind::kConstant:
return kind;
case Kind::kInStackLargeOffset:
return Kind::kInStack;
case Kind::kConstantLargeValue:
return Kind::kConstant;
case Kind::kNone:
return kind;
}
UNREACHABLE();
}
// Required by art::StackMapStream::LocationCatalogEntriesIndices.
DexRegisterLocation() : kind_(Kind::kNone), value_(0) {}
DexRegisterLocation(Kind kind, int32_t value) : kind_(kind), value_(value) {}
static DexRegisterLocation None() {
return DexRegisterLocation(Kind::kNone, 0);
}
// Get the "surface" kind of the location, i.e., the one that doesn't
// include any value with a "large" qualifier.
Kind GetKind() const {
return ConvertToSurfaceKind(kind_);
}
// Get the value of the location.
int32_t GetValue() const { return value_; }
// Get the actual kind of the location.
Kind GetInternalKind() const { return kind_; }
bool operator==(DexRegisterLocation other) const {
return kind_ == other.kind_ && value_ == other.value_;
}
bool operator!=(DexRegisterLocation other) const {
return !(*this == other);
}
private:
Kind kind_;
int32_t value_;
friend class DexRegisterLocationHashFn;
};
std::ostream& operator<<(std::ostream& stream, const DexRegisterLocation::Kind& kind);
/**
* Store information on unique Dex register locations used in a method.
* The information is of the form:
*
* [DexRegisterLocation+].
*
* DexRegisterLocations are either 1- or 5-byte wide (see art::DexRegisterLocation::Kind).
*/
class DexRegisterLocationCatalog {
public:
explicit DexRegisterLocationCatalog(MemoryRegion region) : region_(region) {}
// Short (compressed) location, fitting on one byte.
typedef uint8_t ShortLocation;
void SetRegisterInfo(size_t offset, const DexRegisterLocation& dex_register_location) {
DexRegisterLocation::Kind kind = ComputeCompressedKind(dex_register_location);
int32_t value = dex_register_location.GetValue();
if (DexRegisterLocation::IsShortLocationKind(kind)) {
// Short location. Compress the kind and the value as a single byte.
if (kind == DexRegisterLocation::Kind::kInStack) {
// Instead of storing stack offsets expressed in bytes for
// short stack locations, store slot offsets. A stack offset
// is a multiple of 4 (kFrameSlotSize). This means that by
// dividing it by 4, we can fit values from the [0, 128)
// interval in a short stack location, and not just values
// from the [0, 32) interval.
DCHECK_EQ(value % kFrameSlotSize, 0);
value /= kFrameSlotSize;
}
DCHECK(IsShortValue(value)) << value;
region_.StoreUnaligned<ShortLocation>(offset, MakeShortLocation(kind, value));
} else {
// Large location. Write the location on one byte and the value
// on 4 bytes.
DCHECK(!IsShortValue(value)) << value;
if (kind == DexRegisterLocation::Kind::kInStackLargeOffset) {
// Also divide large stack offsets by 4 for the sake of consistency.
DCHECK_EQ(value % kFrameSlotSize, 0);
value /= kFrameSlotSize;
}
// Data can be unaligned as the written Dex register locations can
// either be 1-byte or 5-byte wide. Use
// art::MemoryRegion::StoreUnaligned instead of
// art::MemoryRegion::Store to prevent unligned word accesses on ARM.
region_.StoreUnaligned<DexRegisterLocation::Kind>(offset, kind);
region_.StoreUnaligned<int32_t>(offset + sizeof(DexRegisterLocation::Kind), value);
}
}
// Find the offset of the location catalog entry number `location_catalog_entry_index`.
size_t FindLocationOffset(size_t location_catalog_entry_index) const {
size_t offset = kFixedSize;
// Skip the first `location_catalog_entry_index - 1` entries.
for (uint16_t i = 0; i < location_catalog_entry_index; ++i) {
// Read the first next byte and inspect its first 3 bits to decide
// whether it is a short or a large location.
DexRegisterLocation::Kind kind = ExtractKindAtOffset(offset);
if (DexRegisterLocation::IsShortLocationKind(kind)) {
// Short location. Skip the current byte.
offset += SingleShortEntrySize();
} else {
// Large location. Skip the 5 next bytes.
offset += SingleLargeEntrySize();
}
}
return offset;
}
// Get the internal kind of entry at `location_catalog_entry_index`.
DexRegisterLocation::Kind GetLocationInternalKind(size_t location_catalog_entry_index) const {
if (location_catalog_entry_index == kNoLocationEntryIndex) {
return DexRegisterLocation::Kind::kNone;
}
return ExtractKindAtOffset(FindLocationOffset(location_catalog_entry_index));
}
// Get the (surface) kind and value of entry at `location_catalog_entry_index`.
DexRegisterLocation GetDexRegisterLocation(size_t location_catalog_entry_index) const {
if (location_catalog_entry_index == kNoLocationEntryIndex) {
return DexRegisterLocation::None();
}
size_t offset = FindLocationOffset(location_catalog_entry_index);
// Read the first byte and inspect its first 3 bits to get the location.
ShortLocation first_byte = region_.LoadUnaligned<ShortLocation>(offset);
DexRegisterLocation::Kind kind = ExtractKindFromShortLocation(first_byte);
if (DexRegisterLocation::IsShortLocationKind(kind)) {
// Short location. Extract the value from the remaining 5 bits.
int32_t value = ExtractValueFromShortLocation(first_byte);
if (kind == DexRegisterLocation::Kind::kInStack) {
// Convert the stack slot (short) offset to a byte offset value.
value *= kFrameSlotSize;
}
return DexRegisterLocation(kind, value);
} else {
// Large location. Read the four next bytes to get the value.
int32_t value = region_.LoadUnaligned<int32_t>(offset + sizeof(DexRegisterLocation::Kind));
if (kind == DexRegisterLocation::Kind::kInStackLargeOffset) {
// Convert the stack slot (large) offset to a byte offset value.
value *= kFrameSlotSize;
}
return DexRegisterLocation(kind, value);
}
}
// Compute the compressed kind of `location`.
static DexRegisterLocation::Kind ComputeCompressedKind(const DexRegisterLocation& location) {
DexRegisterLocation::Kind kind = location.GetInternalKind();
switch (kind) {
case DexRegisterLocation::Kind::kInStack:
return IsShortStackOffsetValue(location.GetValue())
? DexRegisterLocation::Kind::kInStack
: DexRegisterLocation::Kind::kInStackLargeOffset;
case DexRegisterLocation::Kind::kInRegister:
case DexRegisterLocation::Kind::kInRegisterHigh:
DCHECK_GE(location.GetValue(), 0);
DCHECK_LT(location.GetValue(), 1 << kValueBits);
return kind;
case DexRegisterLocation::Kind::kInFpuRegister:
case DexRegisterLocation::Kind::kInFpuRegisterHigh:
DCHECK_GE(location.GetValue(), 0);
DCHECK_LT(location.GetValue(), 1 << kValueBits);
return kind;
case DexRegisterLocation::Kind::kConstant:
return IsShortConstantValue(location.GetValue())
? DexRegisterLocation::Kind::kConstant
: DexRegisterLocation::Kind::kConstantLargeValue;
case DexRegisterLocation::Kind::kConstantLargeValue:
case DexRegisterLocation::Kind::kInStackLargeOffset:
case DexRegisterLocation::Kind::kNone:
LOG(FATAL) << "Unexpected location kind " << kind;
}
UNREACHABLE();
}
// Can `location` be turned into a short location?
static bool CanBeEncodedAsShortLocation(const DexRegisterLocation& location) {
DexRegisterLocation::Kind kind = location.GetInternalKind();
switch (kind) {
case DexRegisterLocation::Kind::kInStack:
return IsShortStackOffsetValue(location.GetValue());
case DexRegisterLocation::Kind::kInRegister:
case DexRegisterLocation::Kind::kInRegisterHigh:
case DexRegisterLocation::Kind::kInFpuRegister:
case DexRegisterLocation::Kind::kInFpuRegisterHigh:
return true;
case DexRegisterLocation::Kind::kConstant:
return IsShortConstantValue(location.GetValue());
case DexRegisterLocation::Kind::kConstantLargeValue:
case DexRegisterLocation::Kind::kInStackLargeOffset:
case DexRegisterLocation::Kind::kNone:
LOG(FATAL) << "Unexpected location kind " << kind;
}
UNREACHABLE();
}
static size_t EntrySize(const DexRegisterLocation& location) {
return CanBeEncodedAsShortLocation(location) ? SingleShortEntrySize() : SingleLargeEntrySize();
}
static size_t SingleShortEntrySize() {
return sizeof(ShortLocation);
}
static size_t SingleLargeEntrySize() {
return sizeof(DexRegisterLocation::Kind) + sizeof(int32_t);
}
size_t Size() const {
return region_.size();
}
void Dump(VariableIndentationOutputStream* vios,
const CodeInfo& code_info);
// Special (invalid) Dex register location catalog entry index meaning
// that there is no location for a given Dex register (i.e., it is
// mapped to a DexRegisterLocation::Kind::kNone location).
static constexpr size_t kNoLocationEntryIndex = -1;
private:
static constexpr int kFixedSize = 0;
// Width of the kind "field" in a short location, in bits.
static constexpr size_t kKindBits = 3;
// Width of the value "field" in a short location, in bits.
static constexpr size_t kValueBits = 5;
static constexpr uint8_t kKindMask = (1 << kKindBits) - 1;
static constexpr int32_t kValueMask = (1 << kValueBits) - 1;
static constexpr size_t kKindOffset = 0;
static constexpr size_t kValueOffset = kKindBits;
static bool IsShortStackOffsetValue(int32_t value) {
DCHECK_EQ(value % kFrameSlotSize, 0);
return IsShortValue(value / kFrameSlotSize);
}
static bool IsShortConstantValue(int32_t value) {
return IsShortValue(value);
}
static bool IsShortValue(int32_t value) {
return IsUint<kValueBits>(value);
}
static ShortLocation MakeShortLocation(DexRegisterLocation::Kind kind, int32_t value) {
uint8_t kind_integer_value = static_cast<uint8_t>(kind);
DCHECK(IsUint<kKindBits>(kind_integer_value)) << kind_integer_value;
DCHECK(IsShortValue(value)) << value;
return (kind_integer_value & kKindMask) << kKindOffset
| (value & kValueMask) << kValueOffset;
}
static DexRegisterLocation::Kind ExtractKindFromShortLocation(ShortLocation location) {
uint8_t kind = (location >> kKindOffset) & kKindMask;
DCHECK_LE(kind, static_cast<uint8_t>(DexRegisterLocation::Kind::kLastLocationKind));
// We do not encode kNone locations in the stack map.
DCHECK_NE(kind, static_cast<uint8_t>(DexRegisterLocation::Kind::kNone));
return static_cast<DexRegisterLocation::Kind>(kind);
}
static int32_t ExtractValueFromShortLocation(ShortLocation location) {
return (location >> kValueOffset) & kValueMask;
}
// Extract a location kind from the byte at position `offset`.
DexRegisterLocation::Kind ExtractKindAtOffset(size_t offset) const {
ShortLocation first_byte = region_.LoadUnaligned<ShortLocation>(offset);
return ExtractKindFromShortLocation(first_byte);
}
MemoryRegion region_;
friend class CodeInfo;
friend class StackMapStream;
};
/* Information on Dex register locations for a specific PC, mapping a
* stack map's Dex register to a location entry in a DexRegisterLocationCatalog.
* The information is of the form:
*
* [live_bit_mask, entries*]
*
* where entries are concatenated unsigned integer values encoded on a number
* of bits (fixed per DexRegisterMap instances of a CodeInfo object) depending
* on the number of entries in the Dex register location catalog
* (see DexRegisterMap::SingleEntrySizeInBits). The map is 1-byte aligned.
*/
class DexRegisterMap {
public:
explicit DexRegisterMap(MemoryRegion region) : region_(region) {}
DexRegisterMap() {}
bool IsValid() const { return region_.pointer() != nullptr; }
// Get the surface kind of Dex register `dex_register_number`.
DexRegisterLocation::Kind GetLocationKind(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const {
return DexRegisterLocation::ConvertToSurfaceKind(
GetLocationInternalKind(dex_register_number, number_of_dex_registers, code_info, enc));
}
// Get the internal kind of Dex register `dex_register_number`.
DexRegisterLocation::Kind GetLocationInternalKind(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const;
// Get the Dex register location `dex_register_number`.
DexRegisterLocation GetDexRegisterLocation(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const;
int32_t GetStackOffsetInBytes(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const {
DexRegisterLocation location =
GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc);
DCHECK(location.GetKind() == DexRegisterLocation::Kind::kInStack);
// GetDexRegisterLocation returns the offset in bytes.
return location.GetValue();
}
int32_t GetConstant(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const {
DexRegisterLocation location =
GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc);
DCHECK_EQ(location.GetKind(), DexRegisterLocation::Kind::kConstant);
return location.GetValue();
}
int32_t GetMachineRegister(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
const CodeInfo& code_info,
const CodeInfoEncoding& enc) const {
DexRegisterLocation location =
GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc);
DCHECK(location.GetInternalKind() == DexRegisterLocation::Kind::kInRegister ||
location.GetInternalKind() == DexRegisterLocation::Kind::kInRegisterHigh ||
location.GetInternalKind() == DexRegisterLocation::Kind::kInFpuRegister ||
location.GetInternalKind() == DexRegisterLocation::Kind::kInFpuRegisterHigh)
<< location.GetInternalKind();
return location.GetValue();
}
// Get the index of the entry in the Dex register location catalog
// corresponding to `dex_register_number`.
size_t GetLocationCatalogEntryIndex(uint16_t dex_register_number,
uint16_t number_of_dex_registers,
size_t number_of_location_catalog_entries) const {
if (!IsDexRegisterLive(dex_register_number)) {
return DexRegisterLocationCatalog::kNoLocationEntryIndex;
}
if (number_of_location_catalog_entries == 1) {
// We do not allocate space for location maps in the case of a
// single-entry location catalog, as it is useless. The only valid
// entry index is 0;
return 0;
}
// The bit offset of the beginning of the map locations.
size_t map_locations_offset_in_bits =
GetLocationMappingDataOffset(number_of_dex_registers) * kBitsPerByte;
size_t index_in_dex_register_map = GetIndexInDexRegisterMap(dex_register_number);
DCHECK_LT(index_in_dex_register_map, GetNumberOfLiveDexRegisters(number_of_dex_registers));
// The bit size of an entry.
size_t map_entry_size_in_bits = SingleEntrySizeInBits(number_of_location_catalog_entries);
// The bit offset where `index_in_dex_register_map` is located.
size_t entry_offset_in_bits =
map_locations_offset_in_bits + index_in_dex_register_map * map_entry_size_in_bits;
size_t location_catalog_entry_index =
region_.LoadBits(entry_offset_in_bits, map_entry_size_in_bits);
DCHECK_LT(location_catalog_entry_index, number_of_location_catalog_entries);
return location_catalog_entry_index;
}
// Map entry at `index_in_dex_register_map` to `location_catalog_entry_index`.
void SetLocationCatalogEntryIndex(size_t index_in_dex_register_map,
size_t location_catalog_entry_index,
uint16_t number_of_dex_registers,
size_t number_of_location_catalog_entries) {
DCHECK_LT(index_in_dex_register_map, GetNumberOfLiveDexRegisters(number_of_dex_registers));
DCHECK_LT(location_catalog_entry_index, number_of_location_catalog_entries);
if (number_of_location_catalog_entries == 1) {
// We do not allocate space for location maps in the case of a
// single-entry location catalog, as it is useless.
return;
}
// The bit offset of the beginning of the map locations.
size_t map_locations_offset_in_bits =
GetLocationMappingDataOffset(number_of_dex_registers) * kBitsPerByte;
// The bit size of an entry.
size_t map_entry_size_in_bits = SingleEntrySizeInBits(number_of_location_catalog_entries);
// The bit offset where `index_in_dex_register_map` is located.
size_t entry_offset_in_bits =
map_locations_offset_in_bits + index_in_dex_register_map * map_entry_size_in_bits;
region_.StoreBits(entry_offset_in_bits, location_catalog_entry_index, map_entry_size_in_bits);
}
void SetLiveBitMask(uint16_t number_of_dex_registers,
const BitVector& live_dex_registers_mask) {
size_t live_bit_mask_offset_in_bits = GetLiveBitMaskOffset() * kBitsPerByte;
for (uint16_t i = 0; i < number_of_dex_registers; ++i) {
region_.StoreBit(live_bit_mask_offset_in_bits + i, live_dex_registers_mask.IsBitSet(i));
}
}
ALWAYS_INLINE bool IsDexRegisterLive(uint16_t dex_register_number) const {
size_t live_bit_mask_offset_in_bits = GetLiveBitMaskOffset() * kBitsPerByte;
return region_.LoadBit(live_bit_mask_offset_in_bits + dex_register_number);
}
size_t GetNumberOfLiveDexRegisters(uint16_t number_of_dex_registers) const {
size_t number_of_live_dex_registers = 0;
for (size_t i = 0; i < number_of_dex_registers; ++i) {
if (IsDexRegisterLive(i)) {
++number_of_live_dex_registers;
}
}
return number_of_live_dex_registers;
}
static size_t GetLiveBitMaskOffset() {
return kFixedSize;
}
// Compute the size of the live register bit mask (in bytes), for a
// method having `number_of_dex_registers` Dex registers.
static size_t GetLiveBitMaskSize(uint16_t number_of_dex_registers) {
return RoundUp(number_of_dex_registers, kBitsPerByte) / kBitsPerByte;
}
static size_t GetLocationMappingDataOffset(uint16_t number_of_dex_registers) {
return GetLiveBitMaskOffset() + GetLiveBitMaskSize(number_of_dex_registers);
}
size_t GetLocationMappingDataSize(uint16_t number_of_dex_registers,
size_t number_of_location_catalog_entries) const {
size_t location_mapping_data_size_in_bits =
GetNumberOfLiveDexRegisters(number_of_dex_registers)
* SingleEntrySizeInBits(number_of_location_catalog_entries);
return RoundUp(location_mapping_data_size_in_bits, kBitsPerByte) / kBitsPerByte;
}
// Return the size of a map entry in bits. Note that if
// `number_of_location_catalog_entries` equals 1, this function returns 0,
// which is fine, as there is no need to allocate a map for a
// single-entry location catalog; the only valid location catalog entry index
// for a live register in this case is 0 and there is no need to
// store it.
static size_t SingleEntrySizeInBits(size_t number_of_location_catalog_entries) {
// Handle the case of 0, as we cannot pass 0 to art::WhichPowerOf2.
return number_of_location_catalog_entries == 0
? 0u
: WhichPowerOf2(RoundUpToPowerOfTwo(number_of_location_catalog_entries));
}
// Return the size of the DexRegisterMap object, in bytes.
size_t Size() const {
return region_.size();
}
void Dump(VariableIndentationOutputStream* vios,
const CodeInfo& code_info, uint16_t number_of_dex_registers) const;
private:
// Return the index in the Dex register map corresponding to the Dex
// register number `dex_register_number`.
size_t GetIndexInDexRegisterMap(uint16_t dex_register_number) const {
if (!IsDexRegisterLive(dex_register_number)) {
return kInvalidIndexInDexRegisterMap;
}
return GetNumberOfLiveDexRegisters(dex_register_number);
}
// Special (invalid) Dex register map entry index meaning that there
// is no index in the map for a given Dex register (i.e., it must
// have been mapped to a DexRegisterLocation::Kind::kNone location).
static constexpr size_t kInvalidIndexInDexRegisterMap = -1;
static constexpr int kFixedSize = 0;
MemoryRegion region_;
friend class CodeInfo;
friend class StackMapStream;
};
// Represents bit range of bit-packed integer field.
// We reuse the idea from ULEB128p1 to support encoding of -1 (aka 0xFFFFFFFF).
// If min_value is set to -1, we implicitly subtract one from any loaded value,
// and add one to any stored value. This is generalized to any negative values.
// In other words, min_value acts as a base and the stored value is added to it.
struct FieldEncoding {
FieldEncoding(size_t start_offset, size_t end_offset, int32_t min_value = 0)
: start_offset_(start_offset), end_offset_(end_offset), min_value_(min_value) {
DCHECK_LE(start_offset_, end_offset_);
DCHECK_LE(BitSize(), 32u);
}
ALWAYS_INLINE size_t BitSize() const { return end_offset_ - start_offset_; }
template <typename Region>
ALWAYS_INLINE int32_t Load(const Region& region) const {
DCHECK_LE(end_offset_, region.size_in_bits());
return static_cast<int32_t>(region.LoadBits(start_offset_, BitSize())) + min_value_;
}
template <typename Region>
ALWAYS_INLINE void Store(Region region, int32_t value) const {
region.StoreBits(start_offset_, value - min_value_, BitSize());
DCHECK_EQ(Load(region), value);
}
private:
size_t start_offset_;
size_t end_offset_;
int32_t min_value_;
};
class StackMapEncoding {
public:
StackMapEncoding()
: dex_pc_bit_offset_(0),
dex_register_map_bit_offset_(0),
inline_info_bit_offset_(0),
register_mask_index_bit_offset_(0),
stack_mask_index_bit_offset_(0),
total_bit_size_(0) {}
// Set stack map bit layout based on given sizes.
// Returns the size of stack map in bits.
size_t SetFromSizes(size_t native_pc_max,
size_t dex_pc_max,
size_t dex_register_map_size,
size_t number_of_inline_info,
size_t number_of_register_masks,
size_t number_of_stack_masks) {
total_bit_size_ = 0;
DCHECK_EQ(kNativePcBitOffset, total_bit_size_);
total_bit_size_ += MinimumBitsToStore(native_pc_max);
dex_pc_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(1 /* kNoDexPc */ + dex_pc_max);
// We also need +1 for kNoDexRegisterMap, but since the size is strictly
// greater than any offset we might try to encode, we already implicitly have it.
dex_register_map_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(dex_register_map_size);
// We also need +1 for kNoInlineInfo, but since the inline_info_size is strictly
// greater than the offset we might try to encode, we already implicitly have it.
// If inline_info_size is zero, we can encode only kNoInlineInfo (in zero bits).
inline_info_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(number_of_inline_info);
register_mask_index_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(number_of_register_masks);
stack_mask_index_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(number_of_stack_masks);
return total_bit_size_;
}
ALWAYS_INLINE FieldEncoding GetNativePcEncoding() const {
return FieldEncoding(kNativePcBitOffset, dex_pc_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const {
return FieldEncoding(dex_pc_bit_offset_, dex_register_map_bit_offset_, -1 /* min_value */);
}
ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const {
return FieldEncoding(dex_register_map_bit_offset_, inline_info_bit_offset_, -1 /* min_value */);
}
ALWAYS_INLINE FieldEncoding GetInlineInfoEncoding() const {
return FieldEncoding(inline_info_bit_offset_,
register_mask_index_bit_offset_,
-1 /* min_value */);
}
ALWAYS_INLINE FieldEncoding GetRegisterMaskIndexEncoding() const {
return FieldEncoding(register_mask_index_bit_offset_, stack_mask_index_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetStackMaskIndexEncoding() const {
return FieldEncoding(stack_mask_index_bit_offset_, total_bit_size_);
}
ALWAYS_INLINE size_t BitSize() const {
return total_bit_size_;
}
// Encode the encoding into the vector.
template<typename Vector>
void Encode(Vector* dest) const {
static_assert(alignof(StackMapEncoding) == 1, "Should not require alignment");
const uint8_t* ptr = reinterpret_cast<const uint8_t*>(this);
dest->insert(dest->end(), ptr, ptr + sizeof(*this));
}
// Decode the encoding from a pointer, updates the pointer.
void Decode(const uint8_t** ptr) {
*this = *reinterpret_cast<const StackMapEncoding*>(*ptr);
*ptr += sizeof(*this);
}
void Dump(VariableIndentationOutputStream* vios) const;
private:
static constexpr size_t kNativePcBitOffset = 0;
uint8_t dex_pc_bit_offset_;
uint8_t dex_register_map_bit_offset_;
uint8_t inline_info_bit_offset_;
uint8_t register_mask_index_bit_offset_;
uint8_t stack_mask_index_bit_offset_;
uint8_t total_bit_size_;
};
/**
* A Stack Map holds compilation information for a specific PC necessary for:
* - Mapping it to a dex PC,
* - Knowing which stack entries are objects,
* - Knowing which registers hold objects,
* - Knowing the inlining information,
* - Knowing the values of dex registers.
*
* The information is of the form:
*
* [native_pc_offset, dex_pc, dex_register_map_offset, inlining_info_index, register_mask_index,
* stack_mask_index].
*/
class StackMap {
public:
StackMap() {}
explicit StackMap(BitMemoryRegion region) : region_(region) {}
ALWAYS_INLINE bool IsValid() const { return region_.pointer() != nullptr; }
ALWAYS_INLINE uint32_t GetDexPc(const StackMapEncoding& encoding) const {
return encoding.GetDexPcEncoding().Load(region_);
}
ALWAYS_INLINE void SetDexPc(const StackMapEncoding& encoding, uint32_t dex_pc) {
encoding.GetDexPcEncoding().Store(region_, dex_pc);
}
ALWAYS_INLINE uint32_t GetNativePcOffset(const StackMapEncoding& encoding,
InstructionSet instruction_set) const {
CodeOffset offset(
CodeOffset::FromCompressedOffset(encoding.GetNativePcEncoding().Load(region_)));
return offset.Uint32Value(instruction_set);
}
ALWAYS_INLINE void SetNativePcCodeOffset(const StackMapEncoding& encoding,
CodeOffset native_pc_offset) {
encoding.GetNativePcEncoding().Store(region_, native_pc_offset.CompressedValue());
}
ALWAYS_INLINE uint32_t GetDexRegisterMapOffset(const StackMapEncoding& encoding) const {
return encoding.GetDexRegisterMapEncoding().Load(region_);
}
ALWAYS_INLINE void SetDexRegisterMapOffset(const StackMapEncoding& encoding, uint32_t offset) {
encoding.GetDexRegisterMapEncoding().Store(region_, offset);
}
ALWAYS_INLINE uint32_t GetInlineInfoIndex(const StackMapEncoding& encoding) const {
return encoding.GetInlineInfoEncoding().Load(region_);
}
ALWAYS_INLINE void SetInlineInfoIndex(const StackMapEncoding& encoding, uint32_t index) {
encoding.GetInlineInfoEncoding().Store(region_, index);
}
ALWAYS_INLINE uint32_t GetRegisterMaskIndex(const StackMapEncoding& encoding) const {
return encoding.GetRegisterMaskIndexEncoding().Load(region_);
}
ALWAYS_INLINE void SetRegisterMaskIndex(const StackMapEncoding& encoding, uint32_t mask) {
encoding.GetRegisterMaskIndexEncoding().Store(region_, mask);
}
ALWAYS_INLINE uint32_t GetStackMaskIndex(const StackMapEncoding& encoding) const {
return encoding.GetStackMaskIndexEncoding().Load(region_);
}
ALWAYS_INLINE void SetStackMaskIndex(const StackMapEncoding& encoding, uint32_t mask) {
encoding.GetStackMaskIndexEncoding().Store(region_, mask);
}
ALWAYS_INLINE bool HasDexRegisterMap(const StackMapEncoding& encoding) const {
return GetDexRegisterMapOffset(encoding) != kNoDexRegisterMap;
}
ALWAYS_INLINE bool HasInlineInfo(const StackMapEncoding& encoding) const {
return GetInlineInfoIndex(encoding) != kNoInlineInfo;
}
ALWAYS_INLINE bool Equals(const StackMap& other) const {
return region_.pointer() == other.region_.pointer() &&
region_.size() == other.region_.size() &&
region_.BitOffset() == other.region_.BitOffset();
}
void Dump(VariableIndentationOutputStream* vios,
const CodeInfo& code_info,
const CodeInfoEncoding& encoding,
const MethodInfo& method_info,
uint32_t code_offset,
uint16_t number_of_dex_registers,
InstructionSet instruction_set,
const std::string& header_suffix = "") const;
// Special (invalid) offset for the DexRegisterMapOffset field meaning
// that there is no Dex register map for this stack map.
static constexpr uint32_t kNoDexRegisterMap = -1;
// Special (invalid) offset for the InlineDescriptorOffset field meaning
// that there is no inline info for this stack map.
static constexpr uint32_t kNoInlineInfo = -1;
private:
static constexpr int kFixedSize = 0;
BitMemoryRegion region_;
friend class StackMapStream;
};
class InlineInfoEncoding {
public:
void SetFromSizes(size_t method_index_idx_max,
size_t dex_pc_max,
size_t extra_data_max,
size_t dex_register_map_size) {
total_bit_size_ = kMethodIndexBitOffset;
total_bit_size_ += MinimumBitsToStore(method_index_idx_max);
dex_pc_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_);
// Note: We're not encoding the dex pc if there is none. That's the case
// for an intrinsified native method, such as String.charAt().
if (dex_pc_max != dex::kDexNoIndex) {
total_bit_size_ += MinimumBitsToStore(1 /* kNoDexPc */ + dex_pc_max);
}
extra_data_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_);
total_bit_size_ += MinimumBitsToStore(extra_data_max);
// We also need +1 for kNoDexRegisterMap, but since the size is strictly
// greater than any offset we might try to encode, we already implicitly have it.
dex_register_map_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_);
total_bit_size_ += MinimumBitsToStore(dex_register_map_size);
}
ALWAYS_INLINE FieldEncoding GetMethodIndexIdxEncoding() const {
return FieldEncoding(kMethodIndexBitOffset, dex_pc_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const {
return FieldEncoding(dex_pc_bit_offset_, extra_data_bit_offset_, -1 /* min_value */);
}
ALWAYS_INLINE FieldEncoding GetExtraDataEncoding() const {
return FieldEncoding(extra_data_bit_offset_, dex_register_map_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const {
return FieldEncoding(dex_register_map_bit_offset_, total_bit_size_, -1 /* min_value */);
}
ALWAYS_INLINE size_t BitSize() const {
return total_bit_size_;
}
void Dump(VariableIndentationOutputStream* vios) const;
// Encode the encoding into the vector.
template<typename Vector>
void Encode(Vector* dest) const {
static_assert(alignof(InlineInfoEncoding) == 1, "Should not require alignment");
const uint8_t* ptr = reinterpret_cast<const uint8_t*>(this);
dest->insert(dest->end(), ptr, ptr + sizeof(*this));
}
// Decode the encoding from a pointer, updates the pointer.
void Decode(const uint8_t** ptr) {
*this = *reinterpret_cast<const InlineInfoEncoding*>(*ptr);
*ptr += sizeof(*this);
}
private:
static constexpr uint8_t kIsLastBitOffset = 0;
static constexpr uint8_t kMethodIndexBitOffset = 1;
uint8_t dex_pc_bit_offset_;
uint8_t extra_data_bit_offset_;
uint8_t dex_register_map_bit_offset_;
uint8_t total_bit_size_;
};
/**
* Inline information for a specific PC. The information is of the form:
*
* [is_last,
* method_index (or ArtMethod high bits),
* dex_pc,
* extra_data (ArtMethod low bits or 1),
* dex_register_map_offset]+.
*/
class InlineInfo {
public:
explicit InlineInfo(BitMemoryRegion region) : region_(region) {}
ALWAYS_INLINE uint32_t GetDepth(const InlineInfoEncoding& encoding) const {
size_t depth = 0;
while (!GetRegionAtDepth(encoding, depth++).LoadBit(0)) { } // Check is_last bit.
return depth;
}
ALWAYS_INLINE void SetDepth(const InlineInfoEncoding& encoding, uint32_t depth) {
DCHECK_GT(depth, 0u);
for (size_t d = 0; d < depth; ++d) {
GetRegionAtDepth(encoding, d).StoreBit(0, d == depth - 1); // Set is_last bit.
}
}
ALWAYS_INLINE uint32_t GetMethodIndexIdxAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
DCHECK(!EncodesArtMethodAtDepth(encoding, depth));
return encoding.GetMethodIndexIdxEncoding().Load(GetRegionAtDepth(encoding, depth));
}
ALWAYS_INLINE void SetMethodIndexIdxAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth,
uint32_t index) {
encoding.GetMethodIndexIdxEncoding().Store(GetRegionAtDepth(encoding, depth), index);
}
ALWAYS_INLINE uint32_t GetMethodIndexAtDepth(const InlineInfoEncoding& encoding,
const MethodInfo& method_info,
uint32_t depth) const {
return method_info.GetMethodIndex(GetMethodIndexIdxAtDepth(encoding, depth));
}
ALWAYS_INLINE uint32_t GetDexPcAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
return encoding.GetDexPcEncoding().Load(GetRegionAtDepth(encoding, depth));
}
ALWAYS_INLINE void SetDexPcAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth,
uint32_t dex_pc) {
encoding.GetDexPcEncoding().Store(GetRegionAtDepth(encoding, depth), dex_pc);
}
ALWAYS_INLINE bool EncodesArtMethodAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
return (encoding.GetExtraDataEncoding().Load(GetRegionAtDepth(encoding, depth)) & 1) == 0;
}
ALWAYS_INLINE void SetExtraDataAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth,
uint32_t extra_data) {
encoding.GetExtraDataEncoding().Store(GetRegionAtDepth(encoding, depth), extra_data);
}
ALWAYS_INLINE ArtMethod* GetArtMethodAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
uint32_t low_bits = encoding.GetExtraDataEncoding().Load(GetRegionAtDepth(encoding, depth));
uint32_t high_bits = encoding.GetMethodIndexIdxEncoding().Load(
GetRegionAtDepth(encoding, depth));
if (high_bits == 0) {
return reinterpret_cast<ArtMethod*>(low_bits);
} else {
uint64_t address = high_bits;
address = address << 32;
return reinterpret_cast<ArtMethod*>(address | low_bits);
}
}
ALWAYS_INLINE uint32_t GetDexRegisterMapOffsetAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
return encoding.GetDexRegisterMapEncoding().Load(GetRegionAtDepth(encoding, depth));
}
ALWAYS_INLINE void SetDexRegisterMapOffsetAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth,
uint32_t offset) {
encoding.GetDexRegisterMapEncoding().Store(GetRegionAtDepth(encoding, depth), offset);
}
ALWAYS_INLINE bool HasDexRegisterMapAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
return GetDexRegisterMapOffsetAtDepth(encoding, depth) != StackMap::kNoDexRegisterMap;
}
void Dump(VariableIndentationOutputStream* vios,
const CodeInfo& info,
const MethodInfo& method_info,
uint16_t* number_of_dex_registers) const;
private:
ALWAYS_INLINE BitMemoryRegion GetRegionAtDepth(const InlineInfoEncoding& encoding,
uint32_t depth) const {
size_t entry_size = encoding.BitSize();
DCHECK_GT(entry_size, 0u);
return region_.Subregion(depth * entry_size, entry_size);
}
BitMemoryRegion region_;
};
// Bit sized region encoding, may be more than 255 bits.
class BitRegionEncoding {
public:
uint32_t num_bits = 0;
ALWAYS_INLINE size_t BitSize() const {
return num_bits;
}
template<typename Vector>
void Encode(Vector* dest) const {
EncodeUnsignedLeb128(dest, num_bits); // Use leb in case num_bits is greater than 255.
}
void Decode(const uint8_t** ptr) {
num_bits = DecodeUnsignedLeb128(ptr);
}
};
// A table of bit sized encodings.
template <typename Encoding>
struct BitEncodingTable {
static constexpr size_t kInvalidOffset = static_cast<size_t>(-1);
// How the encoding is laid out (serialized).
Encoding encoding;
// Number of entries in the table (serialized).
size_t num_entries;
// Bit offset for the base of the table (computed).
size_t bit_offset = kInvalidOffset;
template<typename Vector>
void Encode(Vector* dest) const {
EncodeUnsignedLeb128(dest, num_entries);
encoding.Encode(dest);
}
ALWAYS_INLINE void Decode(const uint8_t** ptr) {
num_entries = DecodeUnsignedLeb128(ptr);
encoding.Decode(ptr);
}
// Set the bit offset in the table and adds the space used by the table to offset.
void UpdateBitOffset(size_t* offset) {
DCHECK(offset != nullptr);
bit_offset = *offset;
*offset += encoding.BitSize() * num_entries;
}
// Return the bit region for the map at index i.
ALWAYS_INLINE BitMemoryRegion BitRegion(MemoryRegion region, size_t index) const {
DCHECK_NE(bit_offset, kInvalidOffset) << "Invalid table offset";
DCHECK_LT(index, num_entries);
const size_t map_size = encoding.BitSize();
return BitMemoryRegion(region, bit_offset + index * map_size, map_size);
}
};
// A byte sized table of possible variable sized encodings.
struct ByteSizedTable {
static constexpr size_t kInvalidOffset = static_cast<size_t>(-1);
// Number of entries in the table (serialized).
size_t num_entries = 0;
// Number of bytes of the table (serialized).
size_t num_bytes;
// Bit offset for the base of the table (computed).
size_t byte_offset = kInvalidOffset;
template<typename Vector>
void Encode(Vector* dest) const {
EncodeUnsignedLeb128(dest, num_entries);
EncodeUnsignedLeb128(dest, num_bytes);
}
ALWAYS_INLINE void Decode(const uint8_t** ptr) {
num_entries = DecodeUnsignedLeb128(ptr);
num_bytes = DecodeUnsignedLeb128(ptr);
}
// Set the bit offset of the table. Adds the total bit size of the table to offset.
void UpdateBitOffset(size_t* offset) {
DCHECK(offset != nullptr);
DCHECK_ALIGNED(*offset, kBitsPerByte);
byte_offset = *offset / kBitsPerByte;
*offset += num_bytes * kBitsPerByte;
}
};
// Format is [native pc, invoke type, method index].
class InvokeInfoEncoding {
public:
void SetFromSizes(size_t native_pc_max,
size_t invoke_type_max,
size_t method_index_max) {
total_bit_size_ = 0;
DCHECK_EQ(kNativePcBitOffset, total_bit_size_);
total_bit_size_ += MinimumBitsToStore(native_pc_max);
invoke_type_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(invoke_type_max);
method_index_bit_offset_ = total_bit_size_;
total_bit_size_ += MinimumBitsToStore(method_index_max);
}
ALWAYS_INLINE FieldEncoding GetNativePcEncoding() const {
return FieldEncoding(kNativePcBitOffset, invoke_type_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetInvokeTypeEncoding() const {
return FieldEncoding(invoke_type_bit_offset_, method_index_bit_offset_);
}
ALWAYS_INLINE FieldEncoding GetMethodIndexEncoding() const {
return FieldEncoding(method_index_bit_offset_, total_bit_size_);
}
ALWAYS_INLINE size_t BitSize() const {
return total_bit_size_;
}
template<typename Vector>
void Encode(Vector* dest) const {
static_assert(alignof(InvokeInfoEncoding) == 1, "Should not require alignment");
const uint8_t* ptr = reinterpret_cast<const uint8_t*>(this);
dest->insert(dest->end(), ptr, ptr + sizeof(*this));
}
void Decode(const uint8_t** ptr) {
*this = *reinterpret_cast<const InvokeInfoEncoding*>(*ptr);
*ptr += sizeof(*this);
}
private:
static constexpr uint8_t kNativePcBitOffset = 0;
uint8_t invoke_type_bit_offset_;
uint8_t method_index_bit_offset_;
uint8_t total_bit_size_;
};
class InvokeInfo {
public:
explicit InvokeInfo(BitMemoryRegion region) : region_(region) {}
ALWAYS_INLINE uint32_t GetNativePcOffset(const InvokeInfoEncoding& encoding,
InstructionSet instruction_set) const {
CodeOffset offset(
CodeOffset::FromCompressedOffset(encoding.GetNativePcEncoding().Load(region_)));
return offset.Uint32Value(instruction_set);
}
ALWAYS_INLINE void SetNativePcCodeOffset(const InvokeInfoEncoding& encoding,
CodeOffset native_pc_offset) {
encoding.GetNativePcEncoding().Store(region_, native_pc_offset.CompressedValue());
}
ALWAYS_INLINE uint32_t GetInvokeType(const InvokeInfoEncoding& encoding) const {
return encoding.GetInvokeTypeEncoding().Load(region_);
}
ALWAYS_INLINE void SetInvokeType(const InvokeInfoEncoding& encoding, uint32_t invoke_type) {
encoding.GetInvokeTypeEncoding().Store(region_, invoke_type);
}
ALWAYS_INLINE uint32_t GetMethodIndexIdx(const InvokeInfoEncoding& encoding) const {
return encoding.GetMethodIndexEncoding().Load(region_);
}
ALWAYS_INLINE void SetMethodIndexIdx(const InvokeInfoEncoding& encoding,
uint32_t method_index_idx) {
encoding.GetMethodIndexEncoding().Store(region_, method_index_idx);
}
ALWAYS_INLINE uint32_t GetMethodIndex(const InvokeInfoEncoding& encoding,
MethodInfo method_info) const {
return method_info.GetMethodIndex(GetMethodIndexIdx(encoding));
}
bool IsValid() const { return region_.pointer() != nullptr; }
private:
BitMemoryRegion region_;
};
// Most of the fields are encoded as ULEB128 to save space.
struct CodeInfoEncoding {
using SizeType = uint32_t;
static constexpr SizeType kInvalidSize = std::numeric_limits<SizeType>::max();
// Byte sized tables go first to avoid unnecessary alignment bits.
ByteSizedTable dex_register_map;
ByteSizedTable location_catalog;
BitEncodingTable<StackMapEncoding> stack_map;
BitEncodingTable<BitRegionEncoding> register_mask;
BitEncodingTable<BitRegionEncoding> stack_mask;
BitEncodingTable<InvokeInfoEncoding> invoke_info;
BitEncodingTable<InlineInfoEncoding> inline_info;
CodeInfoEncoding() {}
explicit CodeInfoEncoding(const void* data) {
const uint8_t* ptr = reinterpret_cast<const uint8_t*>(data);
dex_register_map.Decode(&ptr);
location_catalog.Decode(&ptr);
stack_map.Decode(&ptr);
register_mask.Decode(&ptr);
stack_mask.Decode(&ptr);
invoke_info.Decode(&ptr);
if (stack_map.encoding.GetInlineInfoEncoding().BitSize() > 0) {
inline_info.Decode(&ptr);
} else {
inline_info = BitEncodingTable<InlineInfoEncoding>();
}
cache_header_size =
dchecked_integral_cast<SizeType>(ptr - reinterpret_cast<const uint8_t*>(data));
ComputeTableOffsets();
}
// Compress is not const since it calculates cache_header_size. This is used by PrepareForFillIn.
template<typename Vector>
void Compress(Vector* dest) {
dex_register_map.Encode(dest);
location_catalog.Encode(dest);
stack_map.Encode(dest);
register_mask.Encode(dest);
stack_mask.Encode(dest);
invoke_info.Encode(dest);
if (stack_map.encoding.GetInlineInfoEncoding().BitSize() > 0) {
inline_info.Encode(dest);
}
cache_header_size = dest->size();
}
ALWAYS_INLINE void ComputeTableOffsets() {
// Skip the header.
size_t bit_offset = HeaderSize() * kBitsPerByte;
// The byte tables must be aligned so they must go first.
dex_register_map.UpdateBitOffset(&bit_offset);
location_catalog.UpdateBitOffset(&bit_offset);
// Other tables don't require alignment.
stack_map.UpdateBitOffset(&bit_offset);
register_mask.UpdateBitOffset(&bit_offset);
stack_mask.UpdateBitOffset(&bit_offset);
invoke_info.UpdateBitOffset(&bit_offset);
inline_info.UpdateBitOffset(&bit_offset);
cache_non_header_size = RoundUp(bit_offset, kBitsPerByte) / kBitsPerByte - HeaderSize();
}
ALWAYS_INLINE size_t HeaderSize() const {
DCHECK_NE(cache_header_size, kInvalidSize) << "Uninitialized";
return cache_header_size;
}
ALWAYS_INLINE size_t NonHeaderSize() const {
DCHECK_NE(cache_non_header_size, kInvalidSize) << "Uninitialized";
return cache_non_header_size;
}
private:
// Computed fields (not serialized).
// Header size in bytes, cached to avoid needing to re-decoding the encoding in HeaderSize.
SizeType cache_header_size = kInvalidSize;
// Non header size in bytes, cached to avoid needing to re-decoding the encoding in NonHeaderSize.
SizeType cache_non_header_size = kInvalidSize;
};
/**
* Wrapper around all compiler information collected for a method.
* The information is of the form:
*
* [CodeInfoEncoding, DexRegisterMap+, DexLocationCatalog+, StackMap+, RegisterMask+, StackMask+,
* InlineInfo*]
*
* where CodeInfoEncoding is of the form:
*
* [ByteSizedTable(dex_register_map), ByteSizedTable(location_catalog),
* BitEncodingTable<StackMapEncoding>, BitEncodingTable<BitRegionEncoding>,
* BitEncodingTable<BitRegionEncoding>, BitEncodingTable<InlineInfoEncoding>]
*/
class CodeInfo {
public:
explicit CodeInfo(MemoryRegion region) : region_(region) {
}
explicit CodeInfo(const void* data) {
CodeInfoEncoding encoding = CodeInfoEncoding(data);
region_ = MemoryRegion(const_cast<void*>(data),
encoding.HeaderSize() + encoding.NonHeaderSize());
}
CodeInfoEncoding ExtractEncoding() const {
CodeInfoEncoding encoding(region_.begin());
AssertValidStackMap(encoding);
return encoding;
}
bool HasInlineInfo(const CodeInfoEncoding& encoding) const {
return encoding.stack_map.encoding.GetInlineInfoEncoding().BitSize() > 0;
}
DexRegisterLocationCatalog GetDexRegisterLocationCatalog(const CodeInfoEncoding& encoding) const {
return DexRegisterLocationCatalog(region_.Subregion(encoding.location_catalog.byte_offset,
encoding.location_catalog.num_bytes));
}
ALWAYS_INLINE size_t GetNumberOfStackMaskBits(const CodeInfoEncoding& encoding) const {
return encoding.stack_mask.encoding.BitSize();
}
ALWAYS_INLINE StackMap GetStackMapAt(size_t index, const CodeInfoEncoding& encoding) const {
return StackMap(encoding.stack_map.BitRegion(region_, index));
}
BitMemoryRegion GetStackMask(size_t index, const CodeInfoEncoding& encoding) const {
return encoding.stack_mask.BitRegion(region_, index);
}
BitMemoryRegion GetStackMaskOf(const CodeInfoEncoding& encoding,
const StackMap& stack_map) const {
return GetStackMask(stack_map.GetStackMaskIndex(encoding.stack_map.encoding), encoding);
}
BitMemoryRegion GetRegisterMask(size_t index, const CodeInfoEncoding& encoding) const {
return encoding.register_mask.BitRegion(region_, index);
}
uint32_t GetRegisterMaskOf(const CodeInfoEncoding& encoding, const StackMap& stack_map) const {
size_t index = stack_map.GetRegisterMaskIndex(encoding.stack_map.encoding);
return GetRegisterMask(index, encoding).LoadBits(0u, encoding.register_mask.encoding.BitSize());
}
uint32_t GetNumberOfLocationCatalogEntries(const CodeInfoEncoding& encoding) const {
return encoding.location_catalog.num_entries;
}
uint32_t GetDexRegisterLocationCatalogSize(const CodeInfoEncoding& encoding) const {
return encoding.location_catalog.num_bytes;
}
uint32_t GetNumberOfStackMaps(const CodeInfoEncoding& encoding) const {
return encoding.stack_map.num_entries;
}
// Get the size of all the stack maps of this CodeInfo object, in bits. Not byte aligned.
ALWAYS_INLINE size_t GetStackMapsSizeInBits(const CodeInfoEncoding& encoding) const {
return encoding.stack_map.encoding.BitSize() * GetNumberOfStackMaps(encoding);
}
InvokeInfo GetInvokeInfo(const CodeInfoEncoding& encoding, size_t index) const {
return InvokeInfo(encoding.invoke_info.BitRegion(region_, index));
}
DexRegisterMap GetDexRegisterMapOf(StackMap stack_map,
const CodeInfoEncoding& encoding,
size_t number_of_dex_registers) const {
if (!stack_map.HasDexRegisterMap(encoding.stack_map.encoding)) {
return DexRegisterMap();
}
const uint32_t offset = encoding.dex_register_map.byte_offset +
stack_map.GetDexRegisterMapOffset(encoding.stack_map.encoding);
size_t size = ComputeDexRegisterMapSizeOf(encoding, offset, number_of_dex_registers);
return DexRegisterMap(region_.Subregion(offset, size));
}
size_t GetDexRegisterMapsSize(const CodeInfoEncoding& encoding,
uint32_t number_of_dex_registers) const {
size_t total = 0;
for (size_t i = 0, e = GetNumberOfStackMaps(encoding); i < e; ++i) {
StackMap stack_map = GetStackMapAt(i, encoding);
DexRegisterMap map(GetDexRegisterMapOf(stack_map, encoding, number_of_dex_registers));
total += map.Size();
}
return total;
}
// Return the `DexRegisterMap` pointed by `inline_info` at depth `depth`.
DexRegisterMap GetDexRegisterMapAtDepth(uint8_t depth,
InlineInfo inline_info,
const CodeInfoEncoding& encoding,
uint32_t number_of_dex_registers) const {
if (!inline_info.HasDexRegisterMapAtDepth(encoding.inline_info.encoding, depth)) {
return DexRegisterMap();
} else {
uint32_t offset = encoding.dex_register_map.byte_offset +
inline_info.GetDexRegisterMapOffsetAtDepth(encoding.inline_info.encoding, depth);
size_t size = ComputeDexRegisterMapSizeOf(encoding, offset, number_of_dex_registers);
return DexRegisterMap(region_.Subregion(offset, size));
}
}
InlineInfo GetInlineInfo(size_t index, const CodeInfoEncoding& encoding) const {
// Since we do not know the depth, we just return the whole remaining map. The caller may
// access the inline info for arbitrary depths. To return the precise inline info we would need
// to count the depth before returning.
// TODO: Clean this up.
const size_t bit_offset = encoding.inline_info.bit_offset +
index * encoding.inline_info.encoding.BitSize();
return InlineInfo(BitMemoryRegion(region_, bit_offset, region_.size_in_bits() - bit_offset));
}
InlineInfo GetInlineInfoOf(StackMap stack_map, const CodeInfoEncoding& encoding) const {
DCHECK(stack_map.HasInlineInfo(encoding.stack_map.encoding));
uint32_t index = stack_map.GetInlineInfoIndex(encoding.stack_map.encoding);
return GetInlineInfo(index, encoding);
}
StackMap GetStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const {
for (size_t i = 0, e = GetNumberOfStackMaps(encoding); i < e; ++i) {
StackMap stack_map = GetStackMapAt(i, encoding);
if (stack_map.GetDexPc(encoding.stack_map.encoding) == dex_pc) {
return stack_map;
}
}
return StackMap();
}
// Searches the stack map list backwards because catch stack maps are stored
// at the end.
StackMap GetCatchStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const {
for (size_t i = GetNumberOfStackMaps(encoding); i > 0; --i) {
StackMap stack_map = GetStackMapAt(i - 1, encoding);
if (stack_map.GetDexPc(encoding.stack_map.encoding) == dex_pc) {
return stack_map;
}
}
return StackMap();
}
StackMap GetOsrStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const {
size_t e = GetNumberOfStackMaps(encoding);
if (e == 0) {
// There cannot be OSR stack map if there is no stack map.
return StackMap();
}
// Walk over all stack maps. If two consecutive stack maps are identical, then we
// have found a stack map suitable for OSR.
const StackMapEncoding& stack_map_encoding = encoding.stack_map.encoding;
for (size_t i = 0; i < e - 1; ++i) {
StackMap stack_map = GetStackMapAt(i, encoding);
if (stack_map.GetDexPc(stack_map_encoding) == dex_pc) {
StackMap other = GetStackMapAt(i + 1, encoding);
if (other.GetDexPc(stack_map_encoding) == dex_pc &&
other.GetNativePcOffset(stack_map_encoding, kRuntimeISA) ==
stack_map.GetNativePcOffset(stack_map_encoding, kRuntimeISA)) {
DCHECK_EQ(other.GetDexRegisterMapOffset(stack_map_encoding),
stack_map.GetDexRegisterMapOffset(stack_map_encoding));
DCHECK(!stack_map.HasInlineInfo(stack_map_encoding));
if (i < e - 2) {
// Make sure there are not three identical stack maps following each other.
DCHECK_NE(
stack_map.GetNativePcOffset(stack_map_encoding, kRuntimeISA),
GetStackMapAt(i + 2, encoding).GetNativePcOffset(stack_map_encoding, kRuntimeISA));
}
return stack_map;
}
}
}
return StackMap();
}
StackMap GetStackMapForNativePcOffset(uint32_t native_pc_offset,
const CodeInfoEncoding& encoding) const {
// TODO: Safepoint stack maps are sorted by native_pc_offset but catch stack
// maps are not. If we knew that the method does not have try/catch,
// we could do binary search.
for (size_t i = 0, e = GetNumberOfStackMaps(encoding); i < e; ++i) {
StackMap stack_map = GetStackMapAt(i, encoding);
if (stack_map.GetNativePcOffset(encoding.stack_map.encoding, kRuntimeISA) ==
native_pc_offset) {
return stack_map;
}
}
return StackMap();
}
InvokeInfo GetInvokeInfoForNativePcOffset(uint32_t native_pc_offset,
const CodeInfoEncoding& encoding) {
for (size_t index = 0; index < encoding.invoke_info.num_entries; index++) {
InvokeInfo item = GetInvokeInfo(encoding, index);
if (item.GetNativePcOffset(encoding.invoke_info.encoding, kRuntimeISA) == native_pc_offset) {
return item;
}
}
return InvokeInfo(BitMemoryRegion());
}
// Dump this CodeInfo object on `os`. `code_offset` is the (absolute)
// native PC of the compiled method and `number_of_dex_registers` the
// number of Dex virtual registers used in this method. If
// `dump_stack_maps` is true, also dump the stack maps and the
// associated Dex register maps.
void Dump(VariableIndentationOutputStream* vios,
uint32_t code_offset,
uint16_t number_of_dex_registers,
bool dump_stack_maps,
InstructionSet instruction_set,
const MethodInfo& method_info) const;
// Check that the code info has valid stack map and abort if it does not.
void AssertValidStackMap(const CodeInfoEncoding& encoding) const {
if (region_.size() != 0 && region_.size_in_bits() < GetStackMapsSizeInBits(encoding)) {
LOG(FATAL) << region_.size() << "\n"
<< encoding.HeaderSize() << "\n"
<< encoding.NonHeaderSize() << "\n"
<< encoding.location_catalog.num_entries << "\n"
<< encoding.stack_map.num_entries << "\n"
<< encoding.stack_map.encoding.BitSize();
}
}
private:
// Compute the size of the Dex register map associated to the stack map at
// `dex_register_map_offset_in_code_info`.
size_t ComputeDexRegisterMapSizeOf(const CodeInfoEncoding& encoding,
uint32_t dex_register_map_offset_in_code_info,
uint16_t number_of_dex_registers) const {
// Offset where the actual mapping data starts within art::DexRegisterMap.
size_t location_mapping_data_offset_in_dex_register_map =
DexRegisterMap::GetLocationMappingDataOffset(number_of_dex_registers);
// Create a temporary art::DexRegisterMap to be able to call
// art::DexRegisterMap::GetNumberOfLiveDexRegisters and
DexRegisterMap dex_register_map_without_locations(
MemoryRegion(region_.Subregion(dex_register_map_offset_in_code_info,
location_mapping_data_offset_in_dex_register_map)));
size_t number_of_live_dex_registers =
dex_register_map_without_locations.GetNumberOfLiveDexRegisters(number_of_dex_registers);
size_t location_mapping_data_size_in_bits =
DexRegisterMap::SingleEntrySizeInBits(GetNumberOfLocationCatalogEntries(encoding))
* number_of_live_dex_registers;
size_t location_mapping_data_size_in_bytes =
RoundUp(location_mapping_data_size_in_bits, kBitsPerByte) / kBitsPerByte;
size_t dex_register_map_size =
location_mapping_data_offset_in_dex_register_map + location_mapping_data_size_in_bytes;
return dex_register_map_size;
}
// Compute the size of a Dex register location catalog starting at offset `origin`
// in `region_` and containing `number_of_dex_locations` entries.
size_t ComputeDexRegisterLocationCatalogSize(uint32_t origin,
uint32_t number_of_dex_locations) const {
// TODO: Ideally, we would like to use art::DexRegisterLocationCatalog::Size or
// art::DexRegisterLocationCatalog::FindLocationOffset, but the
// DexRegisterLocationCatalog is not yet built. Try to factor common code.
size_t offset = origin + DexRegisterLocationCatalog::kFixedSize;
// Skip the first `number_of_dex_locations - 1` entries.
for (uint16_t i = 0; i < number_of_dex_locations; ++i) {
// Read the first next byte and inspect its first 3 bits to decide
// whether it is a short or a large location.
DexRegisterLocationCatalog::ShortLocation first_byte =
region_.LoadUnaligned<DexRegisterLocationCatalog::ShortLocation>(offset);
DexRegisterLocation::Kind kind =
DexRegisterLocationCatalog::ExtractKindFromShortLocation(first_byte);
if (DexRegisterLocation::IsShortLocationKind(kind)) {
// Short location. Skip the current byte.
offset += DexRegisterLocationCatalog::SingleShortEntrySize();
} else {
// Large location. Skip the 5 next bytes.
offset += DexRegisterLocationCatalog::SingleLargeEntrySize();
}
}
size_t size = offset - origin;
return size;
}
MemoryRegion region_;
friend class StackMapStream;
};
#undef ELEMENT_BYTE_OFFSET_AFTER
#undef ELEMENT_BIT_OFFSET_AFTER
} // namespace art
#endif // ART_RUNTIME_STACK_MAP_H_