@@ -654,7 +654,7 @@ void SIInstrInfo::copyPhysReg(MachineBasicBlock &MBB,
654
654
// Registers in the sequence are allocated contiguously so we can just
655
655
// use register number to pick one of three round-robin temps.
656
656
unsigned RegNo = DestReg % 3 ;
657
- unsigned Tmp = RS.scavengeRegister (&AMDGPU::VGPR_32RegClass, 0 );
657
+ Register Tmp = RS.scavengeRegister (&AMDGPU::VGPR_32RegClass, 0 );
658
658
if (!Tmp)
659
659
report_fatal_error (" Cannot scavenge VGPR to copy to AGPR" );
660
660
RS.setRegUsed (Tmp);
@@ -938,10 +938,10 @@ void SIInstrInfo::insertVectorSelect(MachineBasicBlock &MBB,
938
938
}
939
939
}
940
940
941
- unsigned SIInstrInfo::insertEQ (MachineBasicBlock *MBB,
941
+ Register SIInstrInfo::insertEQ (MachineBasicBlock *MBB,
942
942
MachineBasicBlock::iterator I,
943
943
const DebugLoc &DL,
944
- unsigned SrcReg, int Value) const {
944
+ Register SrcReg, int Value) const {
945
945
MachineRegisterInfo &MRI = MBB->getParent ()->getRegInfo ();
946
946
Register Reg = MRI.createVirtualRegister (RI.getBoolRC ());
947
947
BuildMI (*MBB, I, DL, get (AMDGPU::V_CMP_EQ_I32_e64), Reg)
@@ -951,10 +951,10 @@ unsigned SIInstrInfo::insertEQ(MachineBasicBlock *MBB,
951
951
return Reg;
952
952
}
953
953
954
- unsigned SIInstrInfo::insertNE (MachineBasicBlock *MBB,
954
+ Register SIInstrInfo::insertNE (MachineBasicBlock *MBB,
955
955
MachineBasicBlock::iterator I,
956
956
const DebugLoc &DL,
957
- unsigned SrcReg, int Value) const {
957
+ Register SrcReg, int Value) const {
958
958
MachineRegisterInfo &MRI = MBB->getParent ()->getRegInfo ();
959
959
Register Reg = MRI.createVirtualRegister (RI.getBoolRC ());
960
960
BuildMI (*MBB, I, DL, get (AMDGPU::V_CMP_NE_I32_e64), Reg)
@@ -1274,7 +1274,7 @@ void SIInstrInfo::loadRegFromStackSlot(MachineBasicBlock &MBB,
1274
1274
// FIXME: Maybe this should not include a memoperand because it will be
1275
1275
// lowered to non-memory instructions.
1276
1276
const MCInstrDesc &OpDesc = get (getSGPRSpillRestoreOpcode (SpillSize));
1277
- if (Register::isVirtualRegister ( DestReg) && SpillSize == 4 ) {
1277
+ if (DestReg. isVirtual ( ) && SpillSize == 4 ) {
1278
1278
MachineRegisterInfo &MRI = MF->getRegInfo ();
1279
1279
MRI.constrainRegClass (DestReg, &AMDGPU::SReg_32_XM0RegClass);
1280
1280
}
@@ -1315,7 +1315,7 @@ unsigned SIInstrInfo::calculateLDSSpillAddress(
1315
1315
unsigned WorkGroupSize = MFI->getMaxFlatWorkGroupSize ();
1316
1316
unsigned WavefrontSize = ST.getWavefrontSize ();
1317
1317
1318
- unsigned TIDReg = MFI->getTIDReg ();
1318
+ Register TIDReg = MFI->getTIDReg ();
1319
1319
if (!MFI->hasCalculatedTID ()) {
1320
1320
MachineBasicBlock &Entry = MBB.getParent ()->front ();
1321
1321
MachineBasicBlock::iterator Insert = Entry.front ();
@@ -1343,8 +1343,8 @@ unsigned SIInstrInfo::calculateLDSSpillAddress(
1343
1343
1344
1344
RS->enterBasicBlock (Entry);
1345
1345
// FIXME: Can we scavenge an SReg_64 and access the subregs?
1346
- unsigned STmp0 = RS->scavengeRegister (&AMDGPU::SGPR_32RegClass, 0 );
1347
- unsigned STmp1 = RS->scavengeRegister (&AMDGPU::SGPR_32RegClass, 0 );
1346
+ Register STmp0 = RS->scavengeRegister (&AMDGPU::SGPR_32RegClass, 0 );
1347
+ Register STmp1 = RS->scavengeRegister (&AMDGPU::SGPR_32RegClass, 0 );
1348
1348
BuildMI (Entry, Insert, DL, get (AMDGPU::S_LOAD_DWORD_IMM), STmp0)
1349
1349
.addReg (InputPtrReg)
1350
1350
.addImm (SI::KernelInputOffsets::NGROUPS_Z);
@@ -2319,7 +2319,7 @@ void SIInstrInfo::insertSelect(MachineBasicBlock &MBB,
2319
2319
2320
2320
I = MIB->getIterator ();
2321
2321
2322
- SmallVector<unsigned , 8 > Regs;
2322
+ SmallVector<Register , 8 > Regs;
2323
2323
for (int Idx = 0 ; Idx != NElts; ++Idx) {
2324
2324
Register DstElt = MRI.createVirtualRegister (EltRC);
2325
2325
Regs.push_back (DstElt);
@@ -3215,7 +3215,7 @@ bool SIInstrInfo::usesConstantBus(const MachineRegisterInfo &MRI,
3215
3215
}
3216
3216
}
3217
3217
3218
- static unsigned findImplicitSGPRRead (const MachineInstr &MI) {
3218
+ static Register findImplicitSGPRRead (const MachineInstr &MI) {
3219
3219
for (const MachineOperand &MO : MI.implicit_operands ()) {
3220
3220
// We only care about reads.
3221
3221
if (MO.isDef ())
@@ -3523,8 +3523,8 @@ bool SIInstrInfo::verifyInstruction(const MachineInstr &MI,
3523
3523
if (AMDGPU::getNamedOperandIdx (Opcode, AMDGPU::OpName::imm) != -1 )
3524
3524
++ConstantBusCount;
3525
3525
3526
- SmallVector<unsigned , 2 > SGPRsUsed;
3527
- unsigned SGPRUsed = findImplicitSGPRRead (MI);
3526
+ SmallVector<Register , 2 > SGPRsUsed;
3527
+ Register SGPRUsed = findImplicitSGPRRead (MI);
3528
3528
if (SGPRUsed != AMDGPU::NoRegister) {
3529
3529
++ConstantBusCount;
3530
3530
SGPRsUsed.push_back (SGPRUsed);
@@ -4316,7 +4316,7 @@ void SIInstrInfo::legalizeOperandsVOP3(MachineRegisterInfo &MRI,
4316
4316
}
4317
4317
}
4318
4318
4319
- unsigned SIInstrInfo::readlaneVGPRToSGPR (unsigned SrcReg, MachineInstr &UseMI,
4319
+ Register SIInstrInfo::readlaneVGPRToSGPR (Register SrcReg, MachineInstr &UseMI,
4320
4320
MachineRegisterInfo &MRI) const {
4321
4321
const TargetRegisterClass *VRC = MRI.getRegClass (SrcReg);
4322
4322
const TargetRegisterClass *SRC = RI.getEquivalentSGPRClass (VRC);
@@ -5722,7 +5722,7 @@ void SIInstrInfo::splitScalar64BitBFE(SetVectorType &Worklist,
5722
5722
}
5723
5723
5724
5724
void SIInstrInfo::addUsersToMoveToVALUWorklist (
5725
- unsigned DstReg,
5725
+ Register DstReg,
5726
5726
MachineRegisterInfo &MRI,
5727
5727
SetVectorType &Worklist) const {
5728
5728
for (MachineRegisterInfo::use_iterator I = MRI.use_begin (DstReg),
@@ -5888,7 +5888,7 @@ const TargetRegisterClass *SIInstrInfo::getDestEquivalentVGPRClass(
5888
5888
}
5889
5889
5890
5890
// Find the one SGPR operand we are allowed to use.
5891
- unsigned SIInstrInfo::findUsedSGPR (const MachineInstr &MI,
5891
+ Register SIInstrInfo::findUsedSGPR (const MachineInstr &MI,
5892
5892
int OpIndices[3 ]) const {
5893
5893
const MCInstrDesc &Desc = MI.getDesc ();
5894
5894
@@ -5901,11 +5901,11 @@ unsigned SIInstrInfo::findUsedSGPR(const MachineInstr &MI,
5901
5901
//
5902
5902
// If the operand's class is an SGPR, we can never move it.
5903
5903
5904
- unsigned SGPRReg = findImplicitSGPRRead (MI);
5904
+ Register SGPRReg = findImplicitSGPRRead (MI);
5905
5905
if (SGPRReg != AMDGPU::NoRegister)
5906
5906
return SGPRReg;
5907
5907
5908
- unsigned UsedSGPRs[3 ] = { AMDGPU::NoRegister };
5908
+ Register UsedSGPRs[3 ] = { AMDGPU::NoRegister };
5909
5909
const MachineRegisterInfo &MRI = MI.getParent ()->getParent ()->getRegInfo ();
5910
5910
5911
5911
for (unsigned i = 0 ; i < 3 ; ++i) {
@@ -6296,7 +6296,7 @@ MachineInstrBuilder
6296
6296
SIInstrInfo::getAddNoCarry (MachineBasicBlock &MBB,
6297
6297
MachineBasicBlock::iterator I,
6298
6298
const DebugLoc &DL,
6299
- unsigned DestReg) const {
6299
+ Register DestReg) const {
6300
6300
if (ST.hasAddNoCarry ())
6301
6301
return BuildMI (MBB, I, DL, get (AMDGPU::V_ADD_U32_e64), DestReg);
6302
6302
0 commit comments