1409 lines
46 KiB
C++
1409 lines
46 KiB
C++
/*
|
|
* Copyright (C) 2011 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
/* This file contains register alloction support. */
|
|
|
|
#include "dex/compiler_ir.h"
|
|
#include "dex/compiler_internals.h"
|
|
#include "mir_to_lir-inl.h"
|
|
|
|
namespace art {
|
|
|
|
/*
|
|
* Free all allocated temps in the temp pools. Note that this does
|
|
* not affect the "liveness" of a temp register, which will stay
|
|
* live until it is either explicitly killed or reallocated.
|
|
*/
|
|
void Mir2Lir::ResetRegPool() {
|
|
GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_);
|
|
for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) {
|
|
info->MarkFree();
|
|
}
|
|
// Reset temp tracking sanity check.
|
|
if (kIsDebugBuild) {
|
|
live_sreg_ = INVALID_SREG;
|
|
}
|
|
}
|
|
|
|
Mir2Lir::RegisterInfo::RegisterInfo(RegStorage r, const ResourceMask& mask)
|
|
: reg_(r), is_temp_(false), wide_value_(false), dirty_(false), aliased_(false), partner_(r),
|
|
s_reg_(INVALID_SREG), def_use_mask_(mask), master_(this), def_start_(nullptr),
|
|
def_end_(nullptr), alias_chain_(nullptr) {
|
|
switch (r.StorageSize()) {
|
|
case 0: storage_mask_ = 0xffffffff; break;
|
|
case 4: storage_mask_ = 0x00000001; break;
|
|
case 8: storage_mask_ = 0x00000003; break;
|
|
case 16: storage_mask_ = 0x0000000f; break;
|
|
case 32: storage_mask_ = 0x000000ff; break;
|
|
case 64: storage_mask_ = 0x0000ffff; break;
|
|
case 128: storage_mask_ = 0xffffffff; break;
|
|
}
|
|
used_storage_ = r.Valid() ? ~storage_mask_ : storage_mask_;
|
|
liveness_ = used_storage_;
|
|
}
|
|
|
|
Mir2Lir::RegisterPool::RegisterPool(Mir2Lir* m2l, ArenaAllocator* arena,
|
|
const ArrayRef<const RegStorage>& core_regs,
|
|
const ArrayRef<const RegStorage>& core64_regs,
|
|
const ArrayRef<const RegStorage>& sp_regs,
|
|
const ArrayRef<const RegStorage>& dp_regs,
|
|
const ArrayRef<const RegStorage>& reserved_regs,
|
|
const ArrayRef<const RegStorage>& reserved64_regs,
|
|
const ArrayRef<const RegStorage>& core_temps,
|
|
const ArrayRef<const RegStorage>& core64_temps,
|
|
const ArrayRef<const RegStorage>& sp_temps,
|
|
const ArrayRef<const RegStorage>& dp_temps) :
|
|
core_regs_(arena, core_regs.size()), next_core_reg_(0),
|
|
core64_regs_(arena, core64_regs.size()), next_core64_reg_(0),
|
|
sp_regs_(arena, sp_regs.size()), next_sp_reg_(0),
|
|
dp_regs_(arena, dp_regs.size()), next_dp_reg_(0), m2l_(m2l) {
|
|
// Initialize the fast lookup map.
|
|
m2l_->reginfo_map_.Reset();
|
|
if (kIsDebugBuild) {
|
|
m2l_->reginfo_map_.Resize(RegStorage::kMaxRegs);
|
|
for (unsigned i = 0; i < RegStorage::kMaxRegs; i++) {
|
|
m2l_->reginfo_map_.Insert(nullptr);
|
|
}
|
|
} else {
|
|
m2l_->reginfo_map_.SetSize(RegStorage::kMaxRegs);
|
|
}
|
|
|
|
// Construct the register pool.
|
|
for (const RegStorage& reg : core_regs) {
|
|
RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg));
|
|
m2l_->reginfo_map_.Put(reg.GetReg(), info);
|
|
core_regs_.Insert(info);
|
|
}
|
|
for (const RegStorage& reg : core64_regs) {
|
|
RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg));
|
|
m2l_->reginfo_map_.Put(reg.GetReg(), info);
|
|
core64_regs_.Insert(info);
|
|
}
|
|
for (const RegStorage& reg : sp_regs) {
|
|
RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg));
|
|
m2l_->reginfo_map_.Put(reg.GetReg(), info);
|
|
sp_regs_.Insert(info);
|
|
}
|
|
for (const RegStorage& reg : dp_regs) {
|
|
RegisterInfo* info = new (arena) RegisterInfo(reg, m2l_->GetRegMaskCommon(reg));
|
|
m2l_->reginfo_map_.Put(reg.GetReg(), info);
|
|
dp_regs_.Insert(info);
|
|
}
|
|
|
|
// Keep special registers from being allocated.
|
|
for (RegStorage reg : reserved_regs) {
|
|
m2l_->MarkInUse(reg);
|
|
}
|
|
for (RegStorage reg : reserved64_regs) {
|
|
m2l_->MarkInUse(reg);
|
|
}
|
|
|
|
// Mark temp regs - all others not in use can be used for promotion
|
|
for (RegStorage reg : core_temps) {
|
|
m2l_->MarkTemp(reg);
|
|
}
|
|
for (RegStorage reg : core64_temps) {
|
|
m2l_->MarkTemp(reg);
|
|
}
|
|
for (RegStorage reg : sp_temps) {
|
|
m2l_->MarkTemp(reg);
|
|
}
|
|
for (RegStorage reg : dp_temps) {
|
|
m2l_->MarkTemp(reg);
|
|
}
|
|
|
|
// Add an entry for InvalidReg with zero'd mask.
|
|
RegisterInfo* invalid_reg = new (arena) RegisterInfo(RegStorage::InvalidReg(), kEncodeNone);
|
|
m2l_->reginfo_map_.Put(RegStorage::InvalidReg().GetReg(), invalid_reg);
|
|
|
|
// Existence of core64 registers implies wide references.
|
|
if (core64_regs_.Size() != 0) {
|
|
ref_regs_ = &core64_regs_;
|
|
next_ref_reg_ = &next_core64_reg_;
|
|
} else {
|
|
ref_regs_ = &core_regs_;
|
|
next_ref_reg_ = &next_core_reg_;
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::DumpRegPool(GrowableArray<RegisterInfo*>* regs) {
|
|
LOG(INFO) << "================================================";
|
|
GrowableArray<RegisterInfo*>::Iterator it(regs);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
LOG(INFO) << StringPrintf(
|
|
"R[%d:%d:%c]: T:%d, U:%d, W:%d, p:%d, LV:%d, D:%d, SR:%d, DEF:%d",
|
|
info->GetReg().GetReg(), info->GetReg().GetRegNum(), info->GetReg().IsFloat() ? 'f' : 'c',
|
|
info->IsTemp(), info->InUse(), info->IsWide(), info->Partner().GetReg(), info->IsLive(),
|
|
info->IsDirty(), info->SReg(), info->DefStart() != nullptr);
|
|
}
|
|
LOG(INFO) << "================================================";
|
|
}
|
|
|
|
void Mir2Lir::DumpCoreRegPool() {
|
|
DumpRegPool(®_pool_->core_regs_);
|
|
DumpRegPool(®_pool_->core64_regs_);
|
|
}
|
|
|
|
void Mir2Lir::DumpFpRegPool() {
|
|
DumpRegPool(®_pool_->sp_regs_);
|
|
DumpRegPool(®_pool_->dp_regs_);
|
|
}
|
|
|
|
void Mir2Lir::DumpRegPools() {
|
|
LOG(INFO) << "Core registers";
|
|
DumpCoreRegPool();
|
|
LOG(INFO) << "FP registers";
|
|
DumpFpRegPool();
|
|
}
|
|
|
|
void Mir2Lir::Clobber(RegStorage reg) {
|
|
if (UNLIKELY(reg.IsPair())) {
|
|
DCHECK(!GetRegInfo(reg.GetLow())->IsAliased());
|
|
Clobber(reg.GetLow());
|
|
DCHECK(!GetRegInfo(reg.GetHigh())->IsAliased());
|
|
Clobber(reg.GetHigh());
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
if (info->IsTemp() && !info->IsDead()) {
|
|
if (info->GetReg().NotExactlyEquals(info->Partner())) {
|
|
ClobberBody(GetRegInfo(info->Partner()));
|
|
}
|
|
ClobberBody(info);
|
|
if (info->IsAliased()) {
|
|
ClobberAliases(info, info->StorageMask());
|
|
} else {
|
|
RegisterInfo* master = info->Master();
|
|
if (info != master) {
|
|
ClobberBody(info->Master());
|
|
ClobberAliases(info->Master(), info->StorageMask());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::ClobberAliases(RegisterInfo* info, uint32_t clobber_mask) {
|
|
for (RegisterInfo* alias = info->GetAliasChain(); alias != nullptr;
|
|
alias = alias->GetAliasChain()) {
|
|
DCHECK(!alias->IsAliased()); // Only the master should be marked as alised.
|
|
// Only clobber if we have overlap.
|
|
if ((alias->StorageMask() & clobber_mask) != 0) {
|
|
ClobberBody(alias);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Break the association between a Dalvik vreg and a physical temp register of either register
|
|
* class.
|
|
* TODO: Ideally, the public version of this code should not exist. Besides its local usage
|
|
* in the register utilities, is is also used by code gen routines to work around a deficiency in
|
|
* local register allocation, which fails to distinguish between the "in" and "out" identities
|
|
* of Dalvik vregs. This can result in useless register copies when the same Dalvik vreg
|
|
* is used both as the source and destination register of an operation in which the type
|
|
* changes (for example: INT_TO_FLOAT v1, v1). Revisit when improved register allocation is
|
|
* addressed.
|
|
*/
|
|
void Mir2Lir::ClobberSReg(int s_reg) {
|
|
if (s_reg != INVALID_SREG) {
|
|
if (kIsDebugBuild && s_reg == live_sreg_) {
|
|
live_sreg_ = INVALID_SREG;
|
|
}
|
|
GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_);
|
|
for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) {
|
|
if (info->SReg() == s_reg) {
|
|
if (info->GetReg().NotExactlyEquals(info->Partner())) {
|
|
// Dealing with a pair - clobber the other half.
|
|
DCHECK(!info->IsAliased());
|
|
ClobberBody(GetRegInfo(info->Partner()));
|
|
}
|
|
ClobberBody(info);
|
|
if (info->IsAliased()) {
|
|
ClobberAliases(info, info->StorageMask());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* SSA names associated with the initial definitions of Dalvik
|
|
* registers are the same as the Dalvik register number (and
|
|
* thus take the same position in the promotion_map. However,
|
|
* the special Method* and compiler temp resisters use negative
|
|
* v_reg numbers to distinguish them and can have an arbitrary
|
|
* ssa name (above the last original Dalvik register). This function
|
|
* maps SSA names to positions in the promotion_map array.
|
|
*/
|
|
int Mir2Lir::SRegToPMap(int s_reg) {
|
|
DCHECK_LT(s_reg, mir_graph_->GetNumSSARegs());
|
|
DCHECK_GE(s_reg, 0);
|
|
int v_reg = mir_graph_->SRegToVReg(s_reg);
|
|
return v_reg;
|
|
}
|
|
|
|
// TODO: refactor following Alloc/Record routines - much commonality.
|
|
void Mir2Lir::RecordCorePromotion(RegStorage reg, int s_reg) {
|
|
int p_map_idx = SRegToPMap(s_reg);
|
|
int v_reg = mir_graph_->SRegToVReg(s_reg);
|
|
int reg_num = reg.GetRegNum();
|
|
GetRegInfo(reg)->MarkInUse();
|
|
core_spill_mask_ |= (1 << reg_num);
|
|
// Include reg for later sort
|
|
core_vmap_table_.push_back(reg_num << VREG_NUM_WIDTH | (v_reg & ((1 << VREG_NUM_WIDTH) - 1)));
|
|
num_core_spills_++;
|
|
promotion_map_[p_map_idx].core_location = kLocPhysReg;
|
|
promotion_map_[p_map_idx].core_reg = reg_num;
|
|
}
|
|
|
|
/* Reserve a callee-save register. Return InvalidReg if none available */
|
|
RegStorage Mir2Lir::AllocPreservedCoreReg(int s_reg) {
|
|
RegStorage res;
|
|
/*
|
|
* Note: it really doesn't matter much whether we allocate from the core or core64
|
|
* pool for 64-bit targets - but for some targets it does matter whether allocations
|
|
* happens from the single or double pool. This entire section of code could stand
|
|
* a good refactoring.
|
|
*/
|
|
GrowableArray<RegisterInfo*>::Iterator it(®_pool_->core_regs_);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
if (!info->IsTemp() && !info->InUse()) {
|
|
res = info->GetReg();
|
|
RecordCorePromotion(res, s_reg);
|
|
break;
|
|
}
|
|
}
|
|
return res;
|
|
}
|
|
|
|
void Mir2Lir::RecordFpPromotion(RegStorage reg, int s_reg) {
|
|
DCHECK_NE(cu_->instruction_set, kThumb2);
|
|
int p_map_idx = SRegToPMap(s_reg);
|
|
int v_reg = mir_graph_->SRegToVReg(s_reg);
|
|
int reg_num = reg.GetRegNum();
|
|
GetRegInfo(reg)->MarkInUse();
|
|
fp_spill_mask_ |= (1 << reg_num);
|
|
// Include reg for later sort
|
|
fp_vmap_table_.push_back(reg_num << VREG_NUM_WIDTH | (v_reg & ((1 << VREG_NUM_WIDTH) - 1)));
|
|
num_fp_spills_++;
|
|
promotion_map_[p_map_idx].fp_location = kLocPhysReg;
|
|
promotion_map_[p_map_idx].fp_reg = reg.GetReg();
|
|
}
|
|
|
|
// Reserve a callee-save floating point.
|
|
RegStorage Mir2Lir::AllocPreservedFpReg(int s_reg) {
|
|
/*
|
|
* For targets other than Thumb2, it doesn't matter whether we allocate from
|
|
* the sp_regs_ or dp_regs_ pool. Some refactoring is in order here.
|
|
*/
|
|
DCHECK_NE(cu_->instruction_set, kThumb2);
|
|
RegStorage res;
|
|
GrowableArray<RegisterInfo*>::Iterator it(®_pool_->sp_regs_);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
if (!info->IsTemp() && !info->InUse()) {
|
|
res = info->GetReg();
|
|
RecordFpPromotion(res, s_reg);
|
|
break;
|
|
}
|
|
}
|
|
return res;
|
|
}
|
|
|
|
// TODO: this is Thumb2 only. Remove when DoPromotion refactored.
|
|
RegStorage Mir2Lir::AllocPreservedDouble(int s_reg) {
|
|
RegStorage res;
|
|
UNIMPLEMENTED(FATAL) << "Unexpected use of AllocPreservedDouble";
|
|
return res;
|
|
}
|
|
|
|
// TODO: this is Thumb2 only. Remove when DoPromotion refactored.
|
|
RegStorage Mir2Lir::AllocPreservedSingle(int s_reg) {
|
|
RegStorage res;
|
|
UNIMPLEMENTED(FATAL) << "Unexpected use of AllocPreservedSingle";
|
|
return res;
|
|
}
|
|
|
|
|
|
RegStorage Mir2Lir::AllocTempBody(GrowableArray<RegisterInfo*> ®s, int* next_temp, bool required) {
|
|
int num_regs = regs.Size();
|
|
int next = *next_temp;
|
|
for (int i = 0; i< num_regs; i++) {
|
|
if (next >= num_regs)
|
|
next = 0;
|
|
RegisterInfo* info = regs.Get(next);
|
|
// Try to allocate a register that doesn't hold a live value.
|
|
if (info->IsTemp() && !info->InUse() && info->IsDead()) {
|
|
// If it's wide, split it up.
|
|
if (info->IsWide()) {
|
|
// If the pair was associated with a wide value, unmark the partner as well.
|
|
if (info->SReg() != INVALID_SREG) {
|
|
RegisterInfo* partner = GetRegInfo(info->Partner());
|
|
DCHECK_EQ(info->GetReg().GetRegNum(), partner->Partner().GetRegNum());
|
|
DCHECK(partner->IsWide());
|
|
partner->SetIsWide(false);
|
|
}
|
|
info->SetIsWide(false);
|
|
}
|
|
Clobber(info->GetReg());
|
|
info->MarkInUse();
|
|
*next_temp = next + 1;
|
|
return info->GetReg();
|
|
}
|
|
next++;
|
|
}
|
|
next = *next_temp;
|
|
// No free non-live regs. Anything we can kill?
|
|
for (int i = 0; i< num_regs; i++) {
|
|
if (next >= num_regs)
|
|
next = 0;
|
|
RegisterInfo* info = regs.Get(next);
|
|
if (info->IsTemp() && !info->InUse()) {
|
|
// Got one. Kill it.
|
|
ClobberSReg(info->SReg());
|
|
Clobber(info->GetReg());
|
|
info->MarkInUse();
|
|
if (info->IsWide()) {
|
|
RegisterInfo* partner = GetRegInfo(info->Partner());
|
|
DCHECK_EQ(info->GetReg().GetRegNum(), partner->Partner().GetRegNum());
|
|
DCHECK(partner->IsWide());
|
|
info->SetIsWide(false);
|
|
partner->SetIsWide(false);
|
|
}
|
|
*next_temp = next + 1;
|
|
return info->GetReg();
|
|
}
|
|
next++;
|
|
}
|
|
if (required) {
|
|
CodegenDump();
|
|
DumpRegPools();
|
|
LOG(FATAL) << "No free temp registers";
|
|
}
|
|
return RegStorage::InvalidReg(); // No register available
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTemp(bool required) {
|
|
return AllocTempBody(reg_pool_->core_regs_, ®_pool_->next_core_reg_, required);
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTempWide(bool required) {
|
|
RegStorage res;
|
|
if (reg_pool_->core64_regs_.Size() != 0) {
|
|
res = AllocTempBody(reg_pool_->core64_regs_, ®_pool_->next_core64_reg_, required);
|
|
} else {
|
|
RegStorage low_reg = AllocTemp();
|
|
RegStorage high_reg = AllocTemp();
|
|
res = RegStorage::MakeRegPair(low_reg, high_reg);
|
|
}
|
|
if (required) {
|
|
CheckRegStorage(res, WidenessCheck::kCheckWide, RefCheck::kIgnoreRef, FPCheck::kCheckNotFP);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTempRef(bool required) {
|
|
RegStorage res = AllocTempBody(*reg_pool_->ref_regs_, reg_pool_->next_ref_reg_, required);
|
|
if (required) {
|
|
DCHECK(!res.IsPair());
|
|
CheckRegStorage(res, WidenessCheck::kCheckNotWide, RefCheck::kCheckRef, FPCheck::kCheckNotFP);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTempSingle(bool required) {
|
|
RegStorage res = AllocTempBody(reg_pool_->sp_regs_, ®_pool_->next_sp_reg_, required);
|
|
if (required) {
|
|
DCHECK(res.IsSingle()) << "Reg: 0x" << std::hex << res.GetRawBits();
|
|
CheckRegStorage(res, WidenessCheck::kCheckNotWide, RefCheck::kCheckNotRef, FPCheck::kIgnoreFP);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTempDouble(bool required) {
|
|
RegStorage res = AllocTempBody(reg_pool_->dp_regs_, ®_pool_->next_dp_reg_, required);
|
|
if (required) {
|
|
DCHECK(res.IsDouble()) << "Reg: 0x" << std::hex << res.GetRawBits();
|
|
CheckRegStorage(res, WidenessCheck::kCheckWide, RefCheck::kCheckNotRef, FPCheck::kIgnoreFP);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTypedTempWide(bool fp_hint, int reg_class, bool required) {
|
|
DCHECK_NE(reg_class, kRefReg); // NOTE: the Dalvik width of a reference is always 32 bits.
|
|
if (((reg_class == kAnyReg) && fp_hint) || (reg_class == kFPReg)) {
|
|
return AllocTempDouble(required);
|
|
}
|
|
return AllocTempWide(required);
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocTypedTemp(bool fp_hint, int reg_class, bool required) {
|
|
if (((reg_class == kAnyReg) && fp_hint) || (reg_class == kFPReg)) {
|
|
return AllocTempSingle(required);
|
|
} else if (reg_class == kRefReg) {
|
|
return AllocTempRef(required);
|
|
}
|
|
return AllocTemp(required);
|
|
}
|
|
|
|
RegStorage Mir2Lir::FindLiveReg(GrowableArray<RegisterInfo*> ®s, int s_reg) {
|
|
RegStorage res;
|
|
GrowableArray<RegisterInfo*>::Iterator it(®s);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
if ((info->SReg() == s_reg) && info->IsLive()) {
|
|
res = info->GetReg();
|
|
break;
|
|
}
|
|
}
|
|
return res;
|
|
}
|
|
|
|
RegStorage Mir2Lir::AllocLiveReg(int s_reg, int reg_class, bool wide) {
|
|
RegStorage reg;
|
|
if (reg_class == kRefReg) {
|
|
reg = FindLiveReg(*reg_pool_->ref_regs_, s_reg);
|
|
CheckRegStorage(reg, WidenessCheck::kCheckNotWide, RefCheck::kCheckRef, FPCheck::kCheckNotFP);
|
|
}
|
|
if (!reg.Valid() && ((reg_class == kAnyReg) || (reg_class == kFPReg))) {
|
|
reg = FindLiveReg(wide ? reg_pool_->dp_regs_ : reg_pool_->sp_regs_, s_reg);
|
|
}
|
|
if (!reg.Valid() && (reg_class != kFPReg)) {
|
|
if (cu_->target64) {
|
|
reg = FindLiveReg(wide || reg_class == kRefReg ? reg_pool_->core64_regs_ :
|
|
reg_pool_->core_regs_, s_reg);
|
|
} else {
|
|
reg = FindLiveReg(reg_pool_->core_regs_, s_reg);
|
|
}
|
|
}
|
|
if (reg.Valid()) {
|
|
if (wide && !reg.IsFloat() && !cu_->target64) {
|
|
// Only allow reg pairs for core regs on 32-bit targets.
|
|
RegStorage high_reg = FindLiveReg(reg_pool_->core_regs_, s_reg + 1);
|
|
if (high_reg.Valid()) {
|
|
reg = RegStorage::MakeRegPair(reg, high_reg);
|
|
MarkWide(reg);
|
|
} else {
|
|
// Only half available.
|
|
reg = RegStorage::InvalidReg();
|
|
}
|
|
}
|
|
if (reg.Valid() && (wide != GetRegInfo(reg)->IsWide())) {
|
|
// Width mismatch - don't try to reuse.
|
|
reg = RegStorage::InvalidReg();
|
|
}
|
|
}
|
|
if (reg.Valid()) {
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* info_low = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* info_high = GetRegInfo(reg.GetHigh());
|
|
if (info_low->IsTemp()) {
|
|
info_low->MarkInUse();
|
|
}
|
|
if (info_high->IsTemp()) {
|
|
info_high->MarkInUse();
|
|
}
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
if (info->IsTemp()) {
|
|
info->MarkInUse();
|
|
}
|
|
}
|
|
} else {
|
|
// Either not found, or something didn't match up. Clobber to prevent any stale instances.
|
|
ClobberSReg(s_reg);
|
|
if (wide) {
|
|
ClobberSReg(s_reg + 1);
|
|
}
|
|
}
|
|
CheckRegStorage(reg, WidenessCheck::kIgnoreWide,
|
|
reg_class == kRefReg ? RefCheck::kCheckRef : RefCheck::kIgnoreRef,
|
|
FPCheck::kIgnoreFP);
|
|
return reg;
|
|
}
|
|
|
|
void Mir2Lir::FreeTemp(RegStorage reg) {
|
|
if (reg.IsPair()) {
|
|
FreeTemp(reg.GetLow());
|
|
FreeTemp(reg.GetHigh());
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
if (p->IsTemp()) {
|
|
p->MarkFree();
|
|
p->SetIsWide(false);
|
|
p->SetPartner(reg);
|
|
}
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::FreeRegLocTemps(RegLocation rl_keep, RegLocation rl_free) {
|
|
DCHECK(rl_keep.wide);
|
|
DCHECK(rl_free.wide);
|
|
int free_low = rl_free.reg.GetLowReg();
|
|
int free_high = rl_free.reg.GetHighReg();
|
|
int keep_low = rl_keep.reg.GetLowReg();
|
|
int keep_high = rl_keep.reg.GetHighReg();
|
|
if ((free_low != keep_low) && (free_low != keep_high) &&
|
|
(free_high != keep_low) && (free_high != keep_high)) {
|
|
// No overlap, free both
|
|
FreeTemp(rl_free.reg);
|
|
}
|
|
}
|
|
|
|
bool Mir2Lir::IsLive(RegStorage reg) {
|
|
bool res;
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* p_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* p_hi = GetRegInfo(reg.GetHigh());
|
|
DCHECK_EQ(p_lo->IsLive(), p_hi->IsLive());
|
|
res = p_lo->IsLive() || p_hi->IsLive();
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
res = p->IsLive();
|
|
}
|
|
return res;
|
|
}
|
|
|
|
bool Mir2Lir::IsTemp(RegStorage reg) {
|
|
bool res;
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* p_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* p_hi = GetRegInfo(reg.GetHigh());
|
|
res = p_lo->IsTemp() || p_hi->IsTemp();
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
res = p->IsTemp();
|
|
}
|
|
return res;
|
|
}
|
|
|
|
bool Mir2Lir::IsPromoted(RegStorage reg) {
|
|
bool res;
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* p_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* p_hi = GetRegInfo(reg.GetHigh());
|
|
res = !p_lo->IsTemp() || !p_hi->IsTemp();
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
res = !p->IsTemp();
|
|
}
|
|
return res;
|
|
}
|
|
|
|
bool Mir2Lir::IsDirty(RegStorage reg) {
|
|
bool res;
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* p_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* p_hi = GetRegInfo(reg.GetHigh());
|
|
res = p_lo->IsDirty() || p_hi->IsDirty();
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
res = p->IsDirty();
|
|
}
|
|
return res;
|
|
}
|
|
|
|
/*
|
|
* Similar to AllocTemp(), but forces the allocation of a specific
|
|
* register. No check is made to see if the register was previously
|
|
* allocated. Use with caution.
|
|
*/
|
|
void Mir2Lir::LockTemp(RegStorage reg) {
|
|
DCHECK(IsTemp(reg));
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* p_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* p_hi = GetRegInfo(reg.GetHigh());
|
|
p_lo->MarkInUse();
|
|
p_lo->MarkDead();
|
|
p_hi->MarkInUse();
|
|
p_hi->MarkDead();
|
|
} else {
|
|
RegisterInfo* p = GetRegInfo(reg);
|
|
p->MarkInUse();
|
|
p->MarkDead();
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::ResetDef(RegStorage reg) {
|
|
if (reg.IsPair()) {
|
|
GetRegInfo(reg.GetLow())->ResetDefBody();
|
|
GetRegInfo(reg.GetHigh())->ResetDefBody();
|
|
} else {
|
|
GetRegInfo(reg)->ResetDefBody();
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::NullifyRange(RegStorage reg, int s_reg) {
|
|
RegisterInfo* info = nullptr;
|
|
RegStorage rs = reg.IsPair() ? reg.GetLow() : reg;
|
|
if (IsTemp(rs)) {
|
|
info = GetRegInfo(reg);
|
|
}
|
|
if ((info != nullptr) && (info->DefStart() != nullptr) && (info->DefEnd() != nullptr)) {
|
|
DCHECK_EQ(info->SReg(), s_reg); // Make sure we're on the same page.
|
|
for (LIR* p = info->DefStart();; p = p->next) {
|
|
NopLIR(p);
|
|
if (p == info->DefEnd()) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Mark the beginning and end LIR of a def sequence. Note that
|
|
* on entry start points to the LIR prior to the beginning of the
|
|
* sequence.
|
|
*/
|
|
void Mir2Lir::MarkDef(RegLocation rl, LIR *start, LIR *finish) {
|
|
DCHECK(!rl.wide);
|
|
DCHECK(start && start->next);
|
|
DCHECK(finish);
|
|
RegisterInfo* p = GetRegInfo(rl.reg);
|
|
p->SetDefStart(start->next);
|
|
p->SetDefEnd(finish);
|
|
}
|
|
|
|
/*
|
|
* Mark the beginning and end LIR of a def sequence. Note that
|
|
* on entry start points to the LIR prior to the beginning of the
|
|
* sequence.
|
|
*/
|
|
void Mir2Lir::MarkDefWide(RegLocation rl, LIR *start, LIR *finish) {
|
|
DCHECK(rl.wide);
|
|
DCHECK(start && start->next);
|
|
DCHECK(finish);
|
|
RegisterInfo* p;
|
|
if (rl.reg.IsPair()) {
|
|
p = GetRegInfo(rl.reg.GetLow());
|
|
ResetDef(rl.reg.GetHigh()); // Only track low of pair
|
|
} else {
|
|
p = GetRegInfo(rl.reg);
|
|
}
|
|
p->SetDefStart(start->next);
|
|
p->SetDefEnd(finish);
|
|
}
|
|
|
|
void Mir2Lir::ResetDefLoc(RegLocation rl) {
|
|
DCHECK(!rl.wide);
|
|
if (IsTemp(rl.reg) && !(cu_->disable_opt & (1 << kSuppressLoads))) {
|
|
NullifyRange(rl.reg, rl.s_reg_low);
|
|
}
|
|
ResetDef(rl.reg);
|
|
}
|
|
|
|
void Mir2Lir::ResetDefLocWide(RegLocation rl) {
|
|
DCHECK(rl.wide);
|
|
// If pair, only track low reg of pair.
|
|
RegStorage rs = rl.reg.IsPair() ? rl.reg.GetLow() : rl.reg;
|
|
if (IsTemp(rs) && !(cu_->disable_opt & (1 << kSuppressLoads))) {
|
|
NullifyRange(rs, rl.s_reg_low);
|
|
}
|
|
ResetDef(rs);
|
|
}
|
|
|
|
void Mir2Lir::ResetDefTracking() {
|
|
GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_);
|
|
for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) {
|
|
info->ResetDefBody();
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::ClobberAllTemps() {
|
|
GrowableArray<RegisterInfo*>::Iterator iter(&tempreg_info_);
|
|
for (RegisterInfo* info = iter.Next(); info != NULL; info = iter.Next()) {
|
|
ClobberBody(info);
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::FlushRegWide(RegStorage reg) {
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* info1 = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* info2 = GetRegInfo(reg.GetHigh());
|
|
DCHECK(info1 && info2 && info1->IsWide() && info2->IsWide() &&
|
|
(info1->Partner().ExactlyEquals(info2->GetReg())) &&
|
|
(info2->Partner().ExactlyEquals(info1->GetReg())));
|
|
if ((info1->IsLive() && info1->IsDirty()) || (info2->IsLive() && info2->IsDirty())) {
|
|
if (!(info1->IsTemp() && info2->IsTemp())) {
|
|
/* Should not happen. If it does, there's a problem in eval_loc */
|
|
LOG(FATAL) << "Long half-temp, half-promoted";
|
|
}
|
|
|
|
info1->SetIsDirty(false);
|
|
info2->SetIsDirty(false);
|
|
if (mir_graph_->SRegToVReg(info2->SReg()) < mir_graph_->SRegToVReg(info1->SReg())) {
|
|
info1 = info2;
|
|
}
|
|
int v_reg = mir_graph_->SRegToVReg(info1->SReg());
|
|
ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg);
|
|
StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile);
|
|
}
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
if (info->IsLive() && info->IsDirty()) {
|
|
info->SetIsDirty(false);
|
|
int v_reg = mir_graph_->SRegToVReg(info->SReg());
|
|
ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg);
|
|
StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile);
|
|
}
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::FlushReg(RegStorage reg) {
|
|
DCHECK(!reg.IsPair());
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
if (info->IsLive() && info->IsDirty()) {
|
|
info->SetIsDirty(false);
|
|
int v_reg = mir_graph_->SRegToVReg(info->SReg());
|
|
ScopedMemRefType mem_ref_type(this, ResourceMask::kDalvikReg);
|
|
StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, kWord, kNotVolatile);
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::FlushSpecificReg(RegisterInfo* info) {
|
|
if (info->IsWide()) {
|
|
FlushRegWide(info->GetReg());
|
|
} else {
|
|
FlushReg(info->GetReg());
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::FlushAllRegs() {
|
|
GrowableArray<RegisterInfo*>::Iterator it(&tempreg_info_);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
if (info->IsDirty() && info->IsLive()) {
|
|
FlushSpecificReg(info);
|
|
}
|
|
info->MarkDead();
|
|
info->SetIsWide(false);
|
|
}
|
|
}
|
|
|
|
|
|
bool Mir2Lir::RegClassMatches(int reg_class, RegStorage reg) {
|
|
if (reg_class == kAnyReg) {
|
|
return true;
|
|
} else if ((reg_class == kCoreReg) || (reg_class == kRefReg)) {
|
|
/*
|
|
* For this purpose, consider Core and Ref to be the same class. We aren't dealing
|
|
* with width here - that should be checked at a higher level (if needed).
|
|
*/
|
|
return !reg.IsFloat();
|
|
} else {
|
|
return reg.IsFloat();
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::MarkLive(RegLocation loc) {
|
|
RegStorage reg = loc.reg;
|
|
if (!IsTemp(reg)) {
|
|
return;
|
|
}
|
|
int s_reg = loc.s_reg_low;
|
|
if (s_reg == INVALID_SREG) {
|
|
// Can't be live if no associated sreg.
|
|
if (reg.IsPair()) {
|
|
GetRegInfo(reg.GetLow())->MarkDead();
|
|
GetRegInfo(reg.GetHigh())->MarkDead();
|
|
} else {
|
|
GetRegInfo(reg)->MarkDead();
|
|
}
|
|
} else {
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* info_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* info_hi = GetRegInfo(reg.GetHigh());
|
|
if (info_lo->IsLive() && (info_lo->SReg() == s_reg) && info_hi->IsLive() &&
|
|
(info_hi->SReg() == s_reg)) {
|
|
return; // Already live.
|
|
}
|
|
ClobberSReg(s_reg);
|
|
ClobberSReg(s_reg + 1);
|
|
info_lo->MarkLive(s_reg);
|
|
info_hi->MarkLive(s_reg + 1);
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
if (info->IsLive() && (info->SReg() == s_reg)) {
|
|
return; // Already live.
|
|
}
|
|
ClobberSReg(s_reg);
|
|
if (loc.wide) {
|
|
ClobberSReg(s_reg + 1);
|
|
}
|
|
info->MarkLive(s_reg);
|
|
}
|
|
if (loc.wide) {
|
|
MarkWide(reg);
|
|
} else {
|
|
MarkNarrow(reg);
|
|
}
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::MarkTemp(RegStorage reg) {
|
|
DCHECK(!reg.IsPair());
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
tempreg_info_.Insert(info);
|
|
info->SetIsTemp(true);
|
|
}
|
|
|
|
void Mir2Lir::UnmarkTemp(RegStorage reg) {
|
|
DCHECK(!reg.IsPair());
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
tempreg_info_.Delete(info);
|
|
info->SetIsTemp(false);
|
|
}
|
|
|
|
void Mir2Lir::MarkWide(RegStorage reg) {
|
|
if (reg.IsPair()) {
|
|
RegisterInfo* info_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* info_hi = GetRegInfo(reg.GetHigh());
|
|
// Unpair any old partners.
|
|
if (info_lo->IsWide() && info_lo->Partner().NotExactlyEquals(info_hi->GetReg())) {
|
|
GetRegInfo(info_lo->Partner())->SetIsWide(false);
|
|
}
|
|
if (info_hi->IsWide() && info_hi->Partner().NotExactlyEquals(info_lo->GetReg())) {
|
|
GetRegInfo(info_hi->Partner())->SetIsWide(false);
|
|
}
|
|
info_lo->SetIsWide(true);
|
|
info_hi->SetIsWide(true);
|
|
info_lo->SetPartner(reg.GetHigh());
|
|
info_hi->SetPartner(reg.GetLow());
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
info->SetIsWide(true);
|
|
info->SetPartner(reg);
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::MarkNarrow(RegStorage reg) {
|
|
DCHECK(!reg.IsPair());
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
info->SetIsWide(false);
|
|
info->SetPartner(reg);
|
|
}
|
|
|
|
void Mir2Lir::MarkClean(RegLocation loc) {
|
|
if (loc.reg.IsPair()) {
|
|
RegisterInfo* info = GetRegInfo(loc.reg.GetLow());
|
|
info->SetIsDirty(false);
|
|
info = GetRegInfo(loc.reg.GetHigh());
|
|
info->SetIsDirty(false);
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(loc.reg);
|
|
info->SetIsDirty(false);
|
|
}
|
|
}
|
|
|
|
// FIXME: need to verify rules/assumptions about how wide values are treated in 64BitSolos.
|
|
void Mir2Lir::MarkDirty(RegLocation loc) {
|
|
if (loc.home) {
|
|
// If already home, can't be dirty
|
|
return;
|
|
}
|
|
if (loc.reg.IsPair()) {
|
|
RegisterInfo* info = GetRegInfo(loc.reg.GetLow());
|
|
info->SetIsDirty(true);
|
|
info = GetRegInfo(loc.reg.GetHigh());
|
|
info->SetIsDirty(true);
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(loc.reg);
|
|
info->SetIsDirty(true);
|
|
}
|
|
}
|
|
|
|
void Mir2Lir::MarkInUse(RegStorage reg) {
|
|
if (reg.IsPair()) {
|
|
GetRegInfo(reg.GetLow())->MarkInUse();
|
|
GetRegInfo(reg.GetHigh())->MarkInUse();
|
|
} else {
|
|
GetRegInfo(reg)->MarkInUse();
|
|
}
|
|
}
|
|
|
|
bool Mir2Lir::CheckCorePoolSanity() {
|
|
GrowableArray<RegisterInfo*>::Iterator it(&tempreg_info_);
|
|
for (RegisterInfo* info = it.Next(); info != nullptr; info = it.Next()) {
|
|
int my_sreg = info->SReg();
|
|
if (info->IsTemp() && info->IsLive() && info->IsWide() && my_sreg != INVALID_SREG) {
|
|
RegStorage my_reg = info->GetReg();
|
|
RegStorage partner_reg = info->Partner();
|
|
RegisterInfo* partner = GetRegInfo(partner_reg);
|
|
DCHECK(partner != NULL);
|
|
DCHECK(partner->IsWide());
|
|
DCHECK_EQ(my_reg.GetReg(), partner->Partner().GetReg());
|
|
DCHECK(partner->IsLive());
|
|
int partner_sreg = partner->SReg();
|
|
int diff = my_sreg - partner_sreg;
|
|
DCHECK((diff == 0) || (diff == -1) || (diff == 1));
|
|
}
|
|
if (info->Master() != info) {
|
|
// Aliased.
|
|
if (info->IsLive() && (info->SReg() != INVALID_SREG)) {
|
|
// If I'm live, master should not be live, but should show liveness in alias set.
|
|
DCHECK_EQ(info->Master()->SReg(), INVALID_SREG);
|
|
DCHECK(!info->Master()->IsDead());
|
|
}
|
|
// TODO: Add checks in !info->IsDead() case to ensure every live bit is owned by exactly 1 reg.
|
|
}
|
|
if (info->IsAliased()) {
|
|
// Has child aliases.
|
|
DCHECK_EQ(info->Master(), info);
|
|
if (info->IsLive() && (info->SReg() != INVALID_SREG)) {
|
|
// Master live, no child should be dead - all should show liveness in set.
|
|
for (RegisterInfo* p = info->GetAliasChain(); p != nullptr; p = p->GetAliasChain()) {
|
|
DCHECK(!p->IsDead());
|
|
DCHECK_EQ(p->SReg(), INVALID_SREG);
|
|
}
|
|
} else if (!info->IsDead()) {
|
|
// Master not live, one or more aliases must be.
|
|
bool live_alias = false;
|
|
for (RegisterInfo* p = info->GetAliasChain(); p != nullptr; p = p->GetAliasChain()) {
|
|
live_alias |= p->IsLive();
|
|
}
|
|
DCHECK(live_alias);
|
|
}
|
|
}
|
|
if (info->IsLive() && (info->SReg() == INVALID_SREG)) {
|
|
// If not fully live, should have INVALID_SREG and def's should be null.
|
|
DCHECK(info->DefStart() == nullptr);
|
|
DCHECK(info->DefEnd() == nullptr);
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* Return an updated location record with current in-register status.
|
|
* If the value lives in live temps, reflect that fact. No code
|
|
* is generated. If the live value is part of an older pair,
|
|
* clobber both low and high.
|
|
* TUNING: clobbering both is a bit heavy-handed, but the alternative
|
|
* is a bit complex when dealing with FP regs. Examine code to see
|
|
* if it's worthwhile trying to be more clever here.
|
|
*/
|
|
RegLocation Mir2Lir::UpdateLoc(RegLocation loc) {
|
|
DCHECK(!loc.wide);
|
|
DCHECK(CheckCorePoolSanity());
|
|
if (loc.location != kLocPhysReg) {
|
|
DCHECK((loc.location == kLocDalvikFrame) ||
|
|
(loc.location == kLocCompilerTemp));
|
|
RegStorage reg = AllocLiveReg(loc.s_reg_low, loc.ref ? kRefReg : kAnyReg, false);
|
|
if (reg.Valid()) {
|
|
bool match = true;
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
match &= !reg.IsPair();
|
|
match &= !info->IsWide();
|
|
if (match) {
|
|
loc.location = kLocPhysReg;
|
|
loc.reg = reg;
|
|
} else {
|
|
Clobber(reg);
|
|
FreeTemp(reg);
|
|
}
|
|
}
|
|
CheckRegLocation(loc);
|
|
}
|
|
return loc;
|
|
}
|
|
|
|
RegLocation Mir2Lir::UpdateLocWide(RegLocation loc) {
|
|
DCHECK(loc.wide);
|
|
DCHECK(CheckCorePoolSanity());
|
|
if (loc.location != kLocPhysReg) {
|
|
DCHECK((loc.location == kLocDalvikFrame) ||
|
|
(loc.location == kLocCompilerTemp));
|
|
RegStorage reg = AllocLiveReg(loc.s_reg_low, kAnyReg, true);
|
|
if (reg.Valid()) {
|
|
bool match = true;
|
|
if (reg.IsPair()) {
|
|
// If we've got a register pair, make sure that it was last used as the same pair.
|
|
RegisterInfo* info_lo = GetRegInfo(reg.GetLow());
|
|
RegisterInfo* info_hi = GetRegInfo(reg.GetHigh());
|
|
match &= info_lo->IsWide();
|
|
match &= info_hi->IsWide();
|
|
match &= (info_lo->Partner().ExactlyEquals(info_hi->GetReg()));
|
|
match &= (info_hi->Partner().ExactlyEquals(info_lo->GetReg()));
|
|
} else {
|
|
RegisterInfo* info = GetRegInfo(reg);
|
|
match &= info->IsWide();
|
|
match &= (info->GetReg().ExactlyEquals(info->Partner()));
|
|
}
|
|
if (match) {
|
|
loc.location = kLocPhysReg;
|
|
loc.reg = reg;
|
|
} else {
|
|
Clobber(reg);
|
|
FreeTemp(reg);
|
|
}
|
|
}
|
|
CheckRegLocation(loc);
|
|
}
|
|
return loc;
|
|
}
|
|
|
|
/* For use in cases we don't know (or care) width */
|
|
RegLocation Mir2Lir::UpdateRawLoc(RegLocation loc) {
|
|
if (loc.wide)
|
|
return UpdateLocWide(loc);
|
|
else
|
|
return UpdateLoc(loc);
|
|
}
|
|
|
|
RegLocation Mir2Lir::EvalLocWide(RegLocation loc, int reg_class, bool update) {
|
|
DCHECK(loc.wide);
|
|
|
|
loc = UpdateLocWide(loc);
|
|
|
|
/* If already in registers, we can assume proper form. Right reg class? */
|
|
if (loc.location == kLocPhysReg) {
|
|
if (!RegClassMatches(reg_class, loc.reg)) {
|
|
// Wrong register class. Reallocate and transfer ownership.
|
|
RegStorage new_regs = AllocTypedTempWide(loc.fp, reg_class);
|
|
// Clobber the old regs.
|
|
Clobber(loc.reg);
|
|
// ...and mark the new ones live.
|
|
loc.reg = new_regs;
|
|
MarkWide(loc.reg);
|
|
MarkLive(loc);
|
|
}
|
|
CheckRegLocation(loc);
|
|
return loc;
|
|
}
|
|
|
|
DCHECK_NE(loc.s_reg_low, INVALID_SREG);
|
|
DCHECK_NE(GetSRegHi(loc.s_reg_low), INVALID_SREG);
|
|
|
|
loc.reg = AllocTypedTempWide(loc.fp, reg_class);
|
|
MarkWide(loc.reg);
|
|
|
|
if (update) {
|
|
loc.location = kLocPhysReg;
|
|
MarkLive(loc);
|
|
}
|
|
CheckRegLocation(loc);
|
|
return loc;
|
|
}
|
|
|
|
RegLocation Mir2Lir::EvalLoc(RegLocation loc, int reg_class, bool update) {
|
|
// Narrow reg_class if the loc is a ref.
|
|
if (loc.ref && reg_class == kAnyReg) {
|
|
reg_class = kRefReg;
|
|
}
|
|
|
|
if (loc.wide) {
|
|
return EvalLocWide(loc, reg_class, update);
|
|
}
|
|
|
|
loc = UpdateLoc(loc);
|
|
|
|
if (loc.location == kLocPhysReg) {
|
|
if (!RegClassMatches(reg_class, loc.reg)) {
|
|
// Wrong register class. Reallocate and transfer ownership.
|
|
RegStorage new_reg = AllocTypedTemp(loc.fp, reg_class);
|
|
// Clobber the old reg.
|
|
Clobber(loc.reg);
|
|
// ...and mark the new one live.
|
|
loc.reg = new_reg;
|
|
MarkLive(loc);
|
|
}
|
|
CheckRegLocation(loc);
|
|
return loc;
|
|
}
|
|
|
|
DCHECK_NE(loc.s_reg_low, INVALID_SREG);
|
|
|
|
loc.reg = AllocTypedTemp(loc.fp, reg_class);
|
|
CheckRegLocation(loc);
|
|
|
|
if (update) {
|
|
loc.location = kLocPhysReg;
|
|
MarkLive(loc);
|
|
}
|
|
CheckRegLocation(loc);
|
|
return loc;
|
|
}
|
|
|
|
/* USE SSA names to count references of base Dalvik v_regs. */
|
|
void Mir2Lir::CountRefs(RefCounts* core_counts, RefCounts* fp_counts, size_t num_regs) {
|
|
for (int i = 0; i < mir_graph_->GetNumSSARegs(); i++) {
|
|
RegLocation loc = mir_graph_->reg_location_[i];
|
|
RefCounts* counts = loc.fp ? fp_counts : core_counts;
|
|
int p_map_idx = SRegToPMap(loc.s_reg_low);
|
|
int use_count = mir_graph_->GetUseCount(i);
|
|
if (loc.fp) {
|
|
if (loc.wide) {
|
|
if (WideFPRsAreAliases()) {
|
|
// Floats and doubles can be counted together.
|
|
counts[p_map_idx].count += use_count;
|
|
} else {
|
|
// Treat doubles as a unit, using upper half of fp_counts array.
|
|
counts[p_map_idx + num_regs].count += use_count;
|
|
}
|
|
i++;
|
|
} else {
|
|
counts[p_map_idx].count += use_count;
|
|
}
|
|
} else {
|
|
if (loc.wide && WideGPRsAreAliases()) {
|
|
i++;
|
|
}
|
|
if (!IsInexpensiveConstant(loc)) {
|
|
counts[p_map_idx].count += use_count;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* qsort callback function, sort descending */
|
|
static int SortCounts(const void *val1, const void *val2) {
|
|
const Mir2Lir::RefCounts* op1 = reinterpret_cast<const Mir2Lir::RefCounts*>(val1);
|
|
const Mir2Lir::RefCounts* op2 = reinterpret_cast<const Mir2Lir::RefCounts*>(val2);
|
|
// Note that we fall back to sorting on reg so we get stable output on differing qsort
|
|
// implementations (such as on host and target or between local host and build servers).
|
|
// Note also that if a wide val1 and a non-wide val2 have the same count, then val1 always
|
|
// ``loses'' (as STARTING_WIDE_SREG is or-ed in val1->s_reg).
|
|
return (op1->count == op2->count)
|
|
? (op1->s_reg - op2->s_reg)
|
|
: (op1->count < op2->count ? 1 : -1);
|
|
}
|
|
|
|
void Mir2Lir::DumpCounts(const RefCounts* arr, int size, const char* msg) {
|
|
LOG(INFO) << msg;
|
|
for (int i = 0; i < size; i++) {
|
|
if ((arr[i].s_reg & STARTING_WIDE_SREG) != 0) {
|
|
LOG(INFO) << "s_reg[64_" << (arr[i].s_reg & ~STARTING_WIDE_SREG) << "]: " << arr[i].count;
|
|
} else {
|
|
LOG(INFO) << "s_reg[32_" << arr[i].s_reg << "]: " << arr[i].count;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Note: some portions of this code required even if the kPromoteRegs
|
|
* optimization is disabled.
|
|
*/
|
|
void Mir2Lir::DoPromotion() {
|
|
int num_regs = mir_graph_->GetNumOfCodeAndTempVRs();
|
|
const int promotion_threshold = 1;
|
|
// Allocate the promotion map - one entry for each Dalvik vReg or compiler temp
|
|
promotion_map_ = static_cast<PromotionMap*>
|
|
(arena_->Alloc(num_regs * sizeof(promotion_map_[0]), kArenaAllocRegAlloc));
|
|
|
|
// Allow target code to add any special registers
|
|
AdjustSpillMask();
|
|
|
|
/*
|
|
* Simple register promotion. Just do a static count of the uses
|
|
* of Dalvik registers. Note that we examine the SSA names, but
|
|
* count based on original Dalvik register name. Count refs
|
|
* separately based on type in order to give allocation
|
|
* preference to fp doubles - which must be allocated sequential
|
|
* physical single fp registers starting with an even-numbered
|
|
* reg.
|
|
* TUNING: replace with linear scan once we have the ability
|
|
* to describe register live ranges for GC.
|
|
*/
|
|
size_t core_reg_count_size = WideGPRsAreAliases() ? num_regs : num_regs * 2;
|
|
size_t fp_reg_count_size = WideFPRsAreAliases() ? num_regs : num_regs * 2;
|
|
RefCounts *core_regs =
|
|
static_cast<RefCounts*>(arena_->Alloc(sizeof(RefCounts) * core_reg_count_size,
|
|
kArenaAllocRegAlloc));
|
|
RefCounts *fp_regs =
|
|
static_cast<RefCounts *>(arena_->Alloc(sizeof(RefCounts) * fp_reg_count_size,
|
|
kArenaAllocRegAlloc));
|
|
// Set ssa names for original Dalvik registers
|
|
for (int i = 0; i < num_regs; i++) {
|
|
core_regs[i].s_reg = fp_regs[i].s_reg = i;
|
|
}
|
|
|
|
// Duplicate in upper half to represent possible wide starting sregs.
|
|
for (size_t i = num_regs; i < fp_reg_count_size; i++) {
|
|
fp_regs[i].s_reg = fp_regs[i - num_regs].s_reg | STARTING_WIDE_SREG;
|
|
}
|
|
for (size_t i = num_regs; i < core_reg_count_size; i++) {
|
|
core_regs[i].s_reg = core_regs[i - num_regs].s_reg | STARTING_WIDE_SREG;
|
|
}
|
|
|
|
// Sum use counts of SSA regs by original Dalvik vreg.
|
|
CountRefs(core_regs, fp_regs, num_regs);
|
|
|
|
// Sort the count arrays
|
|
qsort(core_regs, core_reg_count_size, sizeof(RefCounts), SortCounts);
|
|
qsort(fp_regs, fp_reg_count_size, sizeof(RefCounts), SortCounts);
|
|
|
|
if (cu_->verbose) {
|
|
DumpCounts(core_regs, core_reg_count_size, "Core regs after sort");
|
|
DumpCounts(fp_regs, fp_reg_count_size, "Fp regs after sort");
|
|
}
|
|
|
|
if (!(cu_->disable_opt & (1 << kPromoteRegs))) {
|
|
// Promote fp regs
|
|
for (size_t i = 0; (i < fp_reg_count_size) && (fp_regs[i].count >= promotion_threshold); i++) {
|
|
int low_sreg = fp_regs[i].s_reg & ~STARTING_WIDE_SREG;
|
|
size_t p_map_idx = SRegToPMap(low_sreg);
|
|
RegStorage reg = RegStorage::InvalidReg();
|
|
if (promotion_map_[p_map_idx].fp_location != kLocPhysReg) {
|
|
// TODO: break out the Thumb2-specific code.
|
|
if (cu_->instruction_set == kThumb2) {
|
|
bool wide = fp_regs[i].s_reg & STARTING_WIDE_SREG;
|
|
if (wide) {
|
|
if (promotion_map_[p_map_idx + 1].fp_location != kLocPhysReg) {
|
|
// Ignore result - if can't alloc double may still be able to alloc singles.
|
|
AllocPreservedDouble(low_sreg);
|
|
}
|
|
// Continue regardless of success - might still be able to grab a single.
|
|
continue;
|
|
} else {
|
|
reg = AllocPreservedSingle(low_sreg);
|
|
}
|
|
} else {
|
|
reg = AllocPreservedFpReg(low_sreg);
|
|
}
|
|
if (!reg.Valid()) {
|
|
break; // No more left
|
|
}
|
|
}
|
|
}
|
|
|
|
// Promote core regs
|
|
for (size_t i = 0; (i < core_reg_count_size) &&
|
|
(core_regs[i].count >= promotion_threshold); i++) {
|
|
int low_sreg = core_regs[i].s_reg & ~STARTING_WIDE_SREG;
|
|
size_t p_map_idx = SRegToPMap(low_sreg);
|
|
if (promotion_map_[p_map_idx].core_location != kLocPhysReg) {
|
|
RegStorage reg = AllocPreservedCoreReg(low_sreg);
|
|
if (!reg.Valid()) {
|
|
break; // No more left
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Now, update SSA names to new home locations
|
|
for (int i = 0; i < mir_graph_->GetNumSSARegs(); i++) {
|
|
RegLocation *curr = &mir_graph_->reg_location_[i];
|
|
int p_map_idx = SRegToPMap(curr->s_reg_low);
|
|
int reg_num = curr->fp ? promotion_map_[p_map_idx].fp_reg : promotion_map_[p_map_idx].core_reg;
|
|
bool wide = curr->wide || (cu_->target64 && curr->ref);
|
|
RegStorage reg = RegStorage::InvalidReg();
|
|
if (curr->fp && promotion_map_[p_map_idx].fp_location == kLocPhysReg) {
|
|
if (wide && cu_->instruction_set == kThumb2) {
|
|
if (promotion_map_[p_map_idx + 1].fp_location == kLocPhysReg) {
|
|
int high_reg = promotion_map_[p_map_idx+1].fp_reg;
|
|
// TODO: move target-specific restrictions out of here.
|
|
if (((reg_num & 0x1) == 0) && ((reg_num + 1) == high_reg)) {
|
|
reg = RegStorage::FloatSolo64(RegStorage::RegNum(reg_num) >> 1);
|
|
}
|
|
}
|
|
} else {
|
|
reg = wide ? RegStorage::FloatSolo64(reg_num) : RegStorage::FloatSolo32(reg_num);
|
|
}
|
|
} else if (!curr->fp && promotion_map_[p_map_idx].core_location == kLocPhysReg) {
|
|
if (wide && !cu_->target64) {
|
|
if (promotion_map_[p_map_idx + 1].core_location == kLocPhysReg) {
|
|
int high_reg = promotion_map_[p_map_idx+1].core_reg;
|
|
reg = RegStorage(RegStorage::k64BitPair, reg_num, high_reg);
|
|
}
|
|
} else {
|
|
reg = wide ? RegStorage::Solo64(reg_num) : RegStorage::Solo32(reg_num);
|
|
}
|
|
}
|
|
if (reg.Valid()) {
|
|
curr->reg = reg;
|
|
curr->location = kLocPhysReg;
|
|
curr->home = true;
|
|
}
|
|
}
|
|
if (cu_->verbose) {
|
|
DumpPromotionMap();
|
|
}
|
|
}
|
|
|
|
/* Returns sp-relative offset in bytes for a VReg */
|
|
int Mir2Lir::VRegOffset(int v_reg) {
|
|
const DexFile::CodeItem* code_item = mir_graph_->GetCurrentDexCompilationUnit()->GetCodeItem();
|
|
return StackVisitor::GetVRegOffset(code_item, core_spill_mask_,
|
|
fp_spill_mask_, frame_size_, v_reg,
|
|
cu_->instruction_set);
|
|
}
|
|
|
|
/* Returns sp-relative offset in bytes for a SReg */
|
|
int Mir2Lir::SRegOffset(int s_reg) {
|
|
return VRegOffset(mir_graph_->SRegToVReg(s_reg));
|
|
}
|
|
|
|
/* Mark register usage state and return long retloc */
|
|
RegLocation Mir2Lir::GetReturnWide(RegisterClass reg_class) {
|
|
RegLocation res;
|
|
switch (reg_class) {
|
|
case kRefReg: LOG(FATAL); break;
|
|
case kFPReg: res = LocCReturnDouble(); break;
|
|
default: res = LocCReturnWide(); break;
|
|
}
|
|
Clobber(res.reg);
|
|
LockTemp(res.reg);
|
|
MarkWide(res.reg);
|
|
CheckRegLocation(res);
|
|
return res;
|
|
}
|
|
|
|
RegLocation Mir2Lir::GetReturn(RegisterClass reg_class) {
|
|
RegLocation res;
|
|
switch (reg_class) {
|
|
case kRefReg: res = LocCReturnRef(); break;
|
|
case kFPReg: res = LocCReturnFloat(); break;
|
|
default: res = LocCReturn(); break;
|
|
}
|
|
Clobber(res.reg);
|
|
if (cu_->instruction_set == kMips) {
|
|
MarkInUse(res.reg);
|
|
} else {
|
|
LockTemp(res.reg);
|
|
}
|
|
CheckRegLocation(res);
|
|
return res;
|
|
}
|
|
|
|
void Mir2Lir::SimpleRegAlloc() {
|
|
DoPromotion();
|
|
|
|
if (cu_->verbose && !(cu_->disable_opt & (1 << kPromoteRegs))) {
|
|
LOG(INFO) << "After Promotion";
|
|
mir_graph_->DumpRegLocTable(mir_graph_->reg_location_, mir_graph_->GetNumSSARegs());
|
|
}
|
|
|
|
/* Set the frame size */
|
|
frame_size_ = ComputeFrameSize();
|
|
}
|
|
|
|
/*
|
|
* Get the "real" sreg number associated with an s_reg slot. In general,
|
|
* s_reg values passed through codegen are the SSA names created by
|
|
* dataflow analysis and refer to slot numbers in the mir_graph_->reg_location
|
|
* array. However, renaming is accomplished by simply replacing RegLocation
|
|
* entries in the reglocation[] array. Therefore, when location
|
|
* records for operands are first created, we need to ask the locRecord
|
|
* identified by the dataflow pass what it's new name is.
|
|
*/
|
|
int Mir2Lir::GetSRegHi(int lowSreg) {
|
|
return (lowSreg == INVALID_SREG) ? INVALID_SREG : lowSreg + 1;
|
|
}
|
|
|
|
bool Mir2Lir::LiveOut(int s_reg) {
|
|
// For now.
|
|
return true;
|
|
}
|
|
|
|
} // namespace art
|