39#include "llvm/Config/llvm-config.h"
62#define DEBUG_TYPE "pre-RA-sched"
64STATISTIC(NumBacktracks,
"Number of times scheduler backtracked");
67STATISTIC(NumPRCopies,
"Number of physical register copies");
71 "Bottom-up register reduction list scheduling",
76 "Similar to list-burr but schedules in source "
77 "order when possible",
82 "Bottom-up register pressure aware list scheduling "
83 "which tries to balance latency and register pressure",
88 "Bottom-up register pressure aware list scheduling "
89 "which tries to balance ILP and register pressure",
94 cl::desc(
"Disable cycle-level precision during preRA scheduling"));
100 cl::desc(
"Disable regpressure priority in sched=list-ilp"));
103 cl::desc(
"Disable live use priority in sched=list-ilp"));
106 cl::desc(
"Disable virtual register cycle interference checks"));
109 cl::desc(
"Disable physreg def-use affinity"));
112 cl::desc(
"Disable no-stall priority in sched=list-ilp"));
115 cl::desc(
"Disable critical path priority in sched=list-ilp"));
118 cl::desc(
"Disable scheduled-height priority in sched=list-ilp"));
121 cl::desc(
"Disable scheduler's two-address hack"));
125 cl::desc(
"Number of instructions to allow ahead of the critical path "
126 "in sched=list-ilp"));
130 cl::desc(
"Average inst/cycle whan no target itinerary exists."));
150 std::vector<SUnit *> PendingQueue;
156 unsigned CurCycle = 0;
159 unsigned MinAvailableCycle = ~0u;
163 unsigned IssueCount = 0u;
168 unsigned NumLiveRegs = 0u;
169 std::unique_ptr<SUnit*[]> LiveRegDefs;
170 std::unique_ptr<SUnit*[]> LiveRegGens;
193 AvailableQueue(availqueue), Topo(SUnits, nullptr) {
201 ~ScheduleDAGRRList()
override {
203 delete AvailableQueue;
211 bool IsReachable(
const SUnit *SU,
const SUnit *TargetSU) {
217 bool WillCreateCycle(
SUnit *SU,
SUnit *TargetSU) {
224 void AddPredQueued(
SUnit *SU,
const SDep &
D) {
246 bool isReady(
SUnit *SU) {
251 void ReleasePred(
SUnit *SU,
const SDep *PredEdge);
252 void ReleasePredecessors(
SUnit *SU);
253 void ReleasePending();
254 void AdvanceToCycle(
unsigned NextCycle);
255 void AdvancePastStalls(
SUnit *SU);
256 void EmitNode(
SUnit *SU);
257 void ScheduleNodeBottomUp(
SUnit*);
258 void CapturePred(
SDep *PredEdge);
259 void UnscheduleNodeBottomUp(
SUnit*);
260 void RestoreHazardCheckerBottomUp();
264 void InsertCopiesAndMoveSuccs(
SUnit*,
unsigned,
270 void releaseInterferences(
unsigned Reg = 0);
272 SUnit *PickNodeToScheduleBottomUp();
273 void ListScheduleBottomUp();
277 unsigned NumSUnits = SUnits.size();
280 if (NewNode->
NodeNum >= NumSUnits)
287 unsigned NumSUnits = SUnits.size();
290 if (NewNode->
NodeNum >= NumSUnits)
314 unsigned &RegClass,
unsigned &
Cost,
320 if (VT == MVT::Untyped) {
325 Register Reg = cast<RegisterSDNode>(
Node->getOperand(1))->getReg();
327 RegClass = RC->
getID();
332 unsigned Opcode =
Node->getMachineOpcode();
333 if (Opcode == TargetOpcode::REG_SEQUENCE) {
334 unsigned DstRCIdx =
Node->getConstantOperandVal(0);
336 RegClass = RC->
getID();
344 assert(RC &&
"Not a valid register class");
345 RegClass = RC->
getID();
356void ScheduleDAGRRList::Schedule() {
358 <<
" '" << BB->getName() <<
"' **********\n");
367 LiveRegDefs.reset(
new SUnit*[
TRI->getNumRegs() + 1]());
368 LiveRegGens.reset(
new SUnit*[
TRI->getNumRegs() + 1]());
369 CallSeqEndForStart.
clear();
370 assert(Interferences.
empty() && LRegsMap.empty() &&
"stale Interferences");
373 BuildSchedGraph(
nullptr);
383 ListScheduleBottomUp();
388 dbgs() <<
"*** Final schedule ***\n";
400void ScheduleDAGRRList::ReleasePred(
SUnit *SU,
const SDep *PredEdge) {
405 dbgs() <<
"*** Scheduling failed! ***\n";
407 dbgs() <<
" has been released too many times!\n";
413 if (!forceUnitLatencies()) {
425 if (Height < MinAvailableCycle)
426 MinAvailableCycle = Height;
428 if (isReady(PredSU)) {
429 AvailableQueue->
push(PredSU);
435 PendingQueue.push_back(PredSU);
459 if (
N->isMachineOpcode()) {
460 if (
N->getMachineOpcode() ==
TII->getCallFrameDestroyOpcode()) {
462 }
else if (
N->getMachineOpcode() ==
TII->getCallFrameSetupOpcode()) {
470 if (
Op.getValueType() == MVT::Other) {
472 goto found_chain_operand;
475 found_chain_operand:;
499 unsigned BestMaxNest = MaxNest;
501 unsigned MyNestLevel = NestLevel;
502 unsigned MyMaxNest = MaxNest;
504 MyNestLevel, MyMaxNest,
TII))
505 if (!Best || (MyMaxNest > BestMaxNest)) {
507 BestMaxNest = MyMaxNest;
511 MaxNest = BestMaxNest;
515 if (
N->isMachineOpcode()) {
516 if (
N->getMachineOpcode() ==
TII->getCallFrameDestroyOpcode()) {
518 MaxNest = std::max(MaxNest, NestLevel);
519 }
else if (
N->getMachineOpcode() ==
TII->getCallFrameSetupOpcode()) {
528 if (
Op.getValueType() == MVT::Other) {
530 goto found_chain_operand;
533 found_chain_operand:;
556void ScheduleDAGRRList::ReleasePredecessors(
SUnit *SU) {
559 ReleasePred(SU, &Pred);
565 SUnit *RegDef = LiveRegDefs[Pred.
getReg()]; (void)RegDef;
567 "interference on register dependence");
569 if (!LiveRegGens[Pred.
getReg()]) {
571 LiveRegGens[Pred.
getReg()] = SU;
579 unsigned CallResource =
TRI->getNumRegs();
580 if (!LiveRegDefs[CallResource])
582 if (
Node->isMachineOpcode() &&
583 Node->getMachineOpcode() ==
TII->getCallFrameDestroyOpcode()) {
584 unsigned NestLevel = 0;
585 unsigned MaxNest = 0;
587 assert(
N &&
"Must find call sequence start");
590 CallSeqEndForStart[
Def] = SU;
593 LiveRegDefs[CallResource] =
Def;
594 LiveRegGens[CallResource] = SU;
601void ScheduleDAGRRList::ReleasePending() {
603 assert(PendingQueue.empty() &&
"pending instrs not allowed in this mode");
608 if (AvailableQueue->
empty())
609 MinAvailableCycle = std::numeric_limits<unsigned>::max();
613 for (
unsigned i = 0, e = PendingQueue.size(); i != e; ++i) {
614 unsigned ReadyCycle = PendingQueue[i]->getHeight();
615 if (ReadyCycle < MinAvailableCycle)
616 MinAvailableCycle = ReadyCycle;
618 if (PendingQueue[i]->isAvailable) {
619 if (!isReady(PendingQueue[i]))
621 AvailableQueue->
push(PendingQueue[i]);
623 PendingQueue[i]->isPending =
false;
624 PendingQueue[i] = PendingQueue.back();
625 PendingQueue.pop_back();
631void ScheduleDAGRRList::AdvanceToCycle(
unsigned NextCycle) {
632 if (NextCycle <= CurCycle)
639 CurCycle = NextCycle;
642 for (; CurCycle != NextCycle; ++CurCycle) {
653void ScheduleDAGRRList::AdvancePastStalls(
SUnit *SU) {
670 AdvanceToCycle(ReadyCycle);
690 AdvanceToCycle(CurCycle + Stalls);
695void ScheduleDAGRRList::EmitNode(
SUnit *SU) {
706 "This target-independent node should not be scheduled.");
738void ScheduleDAGRRList::ScheduleNodeBottomUp(
SUnit *SU) {
743 if (CurCycle < SU->getHeight())
745 <<
"] pipeline stall!\n");
765 AdvanceToCycle(CurCycle + 1);
769 ReleasePredecessors(SU);
775 assert(NumLiveRegs > 0 &&
"NumLiveRegs is already zero!");
777 LiveRegDefs[Succ.
getReg()] =
nullptr;
778 LiveRegGens[Succ.
getReg()] =
nullptr;
779 releaseInterferences(Succ.
getReg());
784 unsigned CallResource =
TRI->getNumRegs();
785 if (LiveRegDefs[CallResource] == SU)
788 if (SUNode->isMachineOpcode() &&
789 SUNode->getMachineOpcode() ==
TII->getCallFrameSetupOpcode()) {
790 assert(NumLiveRegs > 0 &&
"NumLiveRegs is already zero!");
792 LiveRegDefs[CallResource] =
nullptr;
793 LiveRegGens[CallResource] =
nullptr;
794 releaseInterferences(CallResource);
815 AdvanceToCycle(CurCycle + 1);
822void ScheduleDAGRRList::CapturePred(
SDep *PredEdge) {
827 AvailableQueue->
remove(PredSU);
831 "NumSuccsLeft will overflow!");
837void ScheduleDAGRRList::UnscheduleNodeBottomUp(
SUnit *SU) {
844 assert(NumLiveRegs > 0 &&
"NumLiveRegs is already zero!");
846 "Physical register dependency violated?");
848 LiveRegDefs[Pred.
getReg()] =
nullptr;
849 LiveRegGens[Pred.
getReg()] =
nullptr;
850 releaseInterferences(Pred.
getReg());
856 unsigned CallResource =
TRI->getNumRegs();
859 if (SUNode->isMachineOpcode() &&
860 SUNode->getMachineOpcode() ==
TII->getCallFrameSetupOpcode()) {
861 SUnit *SeqEnd = CallSeqEndForStart[SU];
862 assert(SeqEnd &&
"Call sequence start/end must be known");
863 assert(!LiveRegDefs[CallResource]);
864 assert(!LiveRegGens[CallResource]);
866 LiveRegDefs[CallResource] = SU;
867 LiveRegGens[CallResource] = SeqEnd;
873 if (LiveRegGens[CallResource] == SU)
876 if (SUNode->isMachineOpcode() &&
877 SUNode->getMachineOpcode() ==
TII->getCallFrameDestroyOpcode()) {
878 assert(NumLiveRegs > 0 &&
"NumLiveRegs is already zero!");
879 assert(LiveRegDefs[CallResource]);
880 assert(LiveRegGens[CallResource]);
882 LiveRegDefs[CallResource] =
nullptr;
883 LiveRegGens[CallResource] =
nullptr;
884 releaseInterferences(CallResource);
888 for (
auto &Succ : SU->
Succs) {
891 if (!LiveRegDefs[Reg])
895 LiveRegDefs[
Reg] = SU;
899 if (!LiveRegGens[Reg]) {
902 for (
auto &Succ2 : SU->
Succs) {
903 if (Succ2.isAssignedRegDep() && Succ2.getReg() == Reg &&
904 Succ2.getSUnit()->getHeight() < LiveRegGens[Reg]->getHeight())
905 LiveRegGens[
Reg] = Succ2.getSUnit();
919 PendingQueue.push_back(SU);
922 AvailableQueue->
push(SU);
929void ScheduleDAGRRList::RestoreHazardCheckerBottomUp() {
932 unsigned LookAhead = std::min((
unsigned)
Sequence.size(),
937 std::vector<SUnit *>::const_iterator
I = (
Sequence.end() - LookAhead);
938 unsigned HazardCycle = (*I)->getHeight();
941 for (; SU->
getHeight() > HazardCycle; ++HazardCycle) {
950void ScheduleDAGRRList::BacktrackBottomUp(
SUnit *SU,
SUnit *BtSU) {
956 UnscheduleNodeBottomUp(OldSU);
965 RestoreHazardCheckerBottomUp();
975 if (SUNode->isOperandOf(
N))
982SUnit *ScheduleDAGRRList::TryUnfoldSU(
SUnit *SU) {
986 if (!
TII->unfoldMemoryOperand(*DAG,
N, NewNodes))
989 assert(NewNodes.
size() == 2 &&
"Expected a load folding node!");
992 SDNode *LoadNode = NewNodes[0];
993 unsigned NumVals =
N->getNumValues();
999 bool isNewLoad =
true;
1002 LoadSU = &SUnits[LoadNode->
getNodeId()];
1009 LoadSU = CreateNewSUnit(LoadNode);
1012 InitNumRegDefsLeft(LoadSU);
1013 computeLatency(LoadSU);
1019 if (
N->getNodeId() != -1) {
1020 NewSU = &SUnits[
N->getNodeId()];
1028 NewSU = CreateNewSUnit(
N);
1041 InitNumRegDefsLeft(NewSU);
1042 computeLatency(NewSU);
1048 for (
unsigned i = 0; i != NumVals; ++i)
1050 DAG->ReplaceAllUsesOfValueWith(
SDValue(SU->
getNode(), OldNumVals - 1),
1075 for (
const SDep &Pred : ChainPreds) {
1076 RemovePred(SU, Pred);
1078 AddPredQueued(LoadSU, Pred);
1080 for (
const SDep &Pred : LoadPreds) {
1081 RemovePred(SU, Pred);
1083 AddPredQueued(LoadSU, Pred);
1085 for (
const SDep &Pred : NodePreds) {
1086 RemovePred(SU, Pred);
1087 AddPredQueued(NewSU, Pred);
1089 for (
SDep &
D : NodeSuccs) {
1090 SUnit *SuccDep =
D.getSUnit();
1092 RemovePred(SuccDep,
D);
1094 AddPredQueued(SuccDep,
D);
1100 for (
SDep &
D : ChainSuccs) {
1101 SUnit *SuccDep =
D.getSUnit();
1103 RemovePred(SuccDep,
D);
1106 AddPredQueued(SuccDep,
D);
1114 AddPredQueued(NewSU,
D);
1117 AvailableQueue->
addNode(LoadSU);
1119 AvailableQueue->
addNode(NewSU);
1131SUnit *ScheduleDAGRRList::CopyAndMoveSuccessors(
SUnit *SU) {
1139 if (
N->getGluedNode() &&
1140 !
TII->canCopyGluedNodeDuringSchedule(
N)) {
1143 <<
"Giving up because it has incoming glue and the target does not "
1144 "want to copy it\n");
1149 bool TryUnfold =
false;
1150 for (
unsigned i = 0, e =
N->getNumValues(); i != e; ++i) {
1151 MVT VT =
N->getSimpleValueType(i);
1152 if (VT == MVT::Glue) {
1153 LLVM_DEBUG(
dbgs() <<
"Giving up because it has outgoing glue\n");
1155 }
else if (VT == MVT::Other)
1159 MVT VT =
Op.getNode()->getSimpleValueType(
Op.getResNo());
1160 if (VT == MVT::Glue && !
TII->canCopyGluedNodeDuringSchedule(
N)) {
1162 dbgs() <<
"Giving up because it one of the operands is glue and "
1163 "the target does not want to copy it\n");
1170 SUnit *UnfoldSU = TryUnfoldSU(SU);
1181 NewSU = CreateClone(SU);
1186 AddPredQueued(NewSU, Pred);
1202 AddPredQueued(SuccSU,
D);
1207 for (
const auto &[DelSU, DelD] : DelDeps)
1208 RemovePred(DelSU, DelD);
1211 AvailableQueue->
addNode(NewSU);
1219void ScheduleDAGRRList::InsertCopiesAndMoveSuccs(
SUnit *SU,
unsigned Reg,
1223 SUnit *CopyFromSU = CreateNewSUnit(
nullptr);
1227 SUnit *CopyToSU = CreateNewSUnit(
nullptr);
1240 D.setSUnit(CopyToSU);
1241 AddPredQueued(SuccSU,
D);
1251 for (
const auto &[DelSU, DelD] : DelDeps)
1252 RemovePred(DelSU, DelD);
1255 FromDep.setLatency(SU->
Latency);
1256 AddPredQueued(CopyFromSU, FromDep);
1258 ToDep.setLatency(CopyFromSU->
Latency);
1259 AddPredQueued(CopyToSU, ToDep);
1262 AvailableQueue->
addNode(CopyFromSU);
1263 AvailableQueue->
addNode(CopyToSU);
1264 Copies.push_back(CopyFromSU);
1265 Copies.push_back(CopyToSU);
1282 "Physical reg def must be in implicit def list!");
1290 return N->getSimpleValueType(NumRes);
1303 if (!LiveRegDefs[*AliasI])
continue;
1306 if (LiveRegDefs[*AliasI] == SU)
continue;
1313 if (RegAdded.
insert(*AliasI).second) {
1326 for (
unsigned i = 1, e = LiveRegDefs.
size()-1; i != e; ++i) {
1327 if (!LiveRegDefs[i])
continue;
1328 if (LiveRegDefs[i] == SU)
continue;
1330 if (RegAdded.
insert(i).second)
1338 if (
const auto *RegOp = dyn_cast<RegisterMaskSDNode>(
Op.getNode()))
1339 return RegOp->getRegMask();
1347bool ScheduleDAGRRList::
1349 if (NumLiveRegs == 0)
1360 RegAdded, LRegs,
TRI);
1367 unsigned NumOps =
Node->getNumOperands();
1368 if (
Node->getOperand(NumOps-1).getValueType() == MVT::Glue)
1372 unsigned Flags =
Node->getConstantOperandVal(i);
1374 unsigned NumVals =
F.getNumOperandRegisters();
1377 if (
F.isRegDefKind() ||
F.isRegDefEarlyClobberKind() ||
1378 F.isClobberKind()) {
1380 for (; NumVals; --NumVals, ++i) {
1382 if (
Reg.isPhysical())
1393 if (
Reg.isPhysical()) {
1394 SDNode *SrcNode =
Node->getOperand(2).getNode();
1400 if (!
Node->isMachineOpcode())
1405 if (
Node->getMachineOpcode() ==
TII->getCallFrameDestroyOpcode()) {
1407 unsigned CallResource =
TRI->getNumRegs();
1408 if (LiveRegDefs[CallResource]) {
1409 SDNode *Gen = LiveRegGens[CallResource]->getNode();
1413 RegAdded.
insert(CallResource).second)
1428 for (
unsigned i = 0; i < MCID.
getNumDefs(); ++i)
1429 if (MCID.
operands()[i].isOptionalDef()) {
1431 Register Reg = cast<RegisterSDNode>(OptionalDef)->getReg();
1439 return !LRegs.
empty();
1442void ScheduleDAGRRList::releaseInterferences(
unsigned Reg) {
1444 for (
unsigned i = Interferences.
size(); i > 0; --i) {
1445 SUnit *SU = Interferences[i-1];
1446 LRegsMapT::iterator LRegsPos = LRegsMap.find(SU);
1458 AvailableQueue->
push(SU);
1460 if (i < Interferences.
size())
1461 Interferences[i-1] = Interferences.
back();
1463 LRegsMap.erase(LRegsPos);
1471SUnit *ScheduleDAGRRList::PickNodeToScheduleBottomUp() {
1472 SUnit *CurSU = AvailableQueue->
empty() ? nullptr : AvailableQueue->
pop();
1473 auto FindAvailableNode = [&]() {
1476 if (!DelayForLiveRegsBottomUp(CurSU, LRegs))
1479 if (LRegs[0] ==
TRI->getNumRegs())
dbgs() <<
"CallResource";
1482 auto [LRegsIter, LRegsInserted] = LRegsMap.try_emplace(CurSU, LRegs);
1483 if (LRegsInserted) {
1490 LRegsIter->second = LRegs;
1492 CurSU = AvailableQueue->
pop();
1495 FindAvailableNode();
1507 for (
SUnit *TrySU : Interferences) {
1512 SUnit *BtSU =
nullptr;
1513 unsigned LiveCycle = std::numeric_limits<unsigned>::max();
1514 for (
unsigned Reg : LRegs) {
1515 if (LiveRegGens[Reg]->getHeight() < LiveCycle) {
1516 BtSU = LiveRegGens[
Reg];
1520 if (!WillCreateCycle(TrySU, BtSU)) {
1522 BacktrackBottomUp(TrySU, BtSU);
1529 AvailableQueue->
remove(BtSU);
1532 <<
") to SU(" << TrySU->NodeNum <<
")\n");
1537 if (!TrySU->isAvailable || !TrySU->NodeQueueId) {
1538 LLVM_DEBUG(
dbgs() <<
"TrySU not available; choosing node from queue\n");
1539 CurSU = AvailableQueue->
pop();
1543 AvailableQueue->
remove(TrySU);
1546 FindAvailableNode();
1558 SUnit *TrySU = Interferences[0];
1560 assert(LRegs.
size() == 1 &&
"Can't handle this yet!");
1561 unsigned Reg = LRegs[0];
1565 TRI->getMinimalPhysRegClass(Reg, VT);
1575 SUnit *NewDef =
nullptr;
1577 NewDef = CopyAndMoveSuccessors(LRDef);
1578 if (!DestRC && !NewDef)
1584 InsertCopiesAndMoveSuccs(LRDef, Reg, DestRC, RC,
Copies);
1586 <<
" to SU #" <<
Copies.front()->NodeNum <<
"\n");
1592 <<
" to SU #" << TrySU->
NodeNum <<
"\n");
1593 LiveRegDefs[
Reg] = NewDef;
1598 assert(CurSU &&
"Unable to resolve live physical register dependencies!");
1604void ScheduleDAGRRList::ListScheduleBottomUp() {
1606 ReleasePredecessors(&ExitSU);
1609 if (!SUnits.empty()) {
1610 SUnit *RootSU = &SUnits[DAG->getRoot().getNode()->getNodeId()];
1611 assert(RootSU->
Succs.empty() &&
"Graph root shouldn't have successors!");
1613 AvailableQueue->
push(RootSU);
1619 while (!AvailableQueue->
empty() || !Interferences.empty()) {
1621 AvailableQueue->
dump(
this));
1625 SUnit *SU = PickNodeToScheduleBottomUp();
1627 AdvancePastStalls(SU);
1629 ScheduleNodeBottomUp(SU);
1631 while (AvailableQueue->
empty() && !PendingQueue.empty()) {
1633 assert(MinAvailableCycle < std::numeric_limits<unsigned>::max() &&
1634 "MinAvailableCycle uninitialized");
1635 AdvanceToCycle(std::max(CurCycle + 1, MinAvailableCycle));
1643 VerifyScheduledSequence(
true);
1649class RegReductionPQBase;
1652 bool isReady(
SUnit* SU,
unsigned CurCycle)
const {
return true; }
1657struct reverse_sort :
public queue_sort {
1660 reverse_sort(SF &sf) : SortFunc(sf) {}
1662 bool operator()(
SUnit* left,
SUnit* right)
const {
1665 return SortFunc(right, left);
1672struct bu_ls_rr_sort :
public queue_sort {
1675 HasReadyFilter =
false
1678 RegReductionPQBase *SPQ;
1680 bu_ls_rr_sort(RegReductionPQBase *spq) : SPQ(spq) {}
1682 bool operator()(
SUnit* left,
SUnit* right)
const;
1686struct src_ls_rr_sort :
public queue_sort {
1689 HasReadyFilter =
false
1692 RegReductionPQBase *SPQ;
1694 src_ls_rr_sort(RegReductionPQBase *spq) : SPQ(spq) {}
1696 bool operator()(
SUnit* left,
SUnit* right)
const;
1700struct hybrid_ls_rr_sort :
public queue_sort {
1703 HasReadyFilter =
false
1706 RegReductionPQBase *SPQ;
1708 hybrid_ls_rr_sort(RegReductionPQBase *spq) : SPQ(spq) {}
1710 bool isReady(
SUnit *SU,
unsigned CurCycle)
const;
1712 bool operator()(
SUnit* left,
SUnit* right)
const;
1717struct ilp_ls_rr_sort :
public queue_sort {
1720 HasReadyFilter =
false
1723 RegReductionPQBase *SPQ;
1725 ilp_ls_rr_sort(RegReductionPQBase *spq) : SPQ(spq) {}
1727 bool isReady(
SUnit *SU,
unsigned CurCycle)
const;
1729 bool operator()(
SUnit* left,
SUnit* right)
const;
1734 std::vector<SUnit *>
Queue;
1735 unsigned CurQueueId = 0;
1736 bool TracksRegPressure;
1740 std::vector<SUnit> *SUnits =
nullptr;
1746 ScheduleDAGRRList *scheduleDAG =
nullptr;
1749 std::vector<unsigned> SethiUllmanNumbers;
1756 std::vector<unsigned> RegLimit;
1760 bool hasReadyFilter,
1767 SrcOrder(srcorder), MF(mf),
TII(tii),
TRI(tri), TLI(tli) {
1768 if (TracksRegPressure) {
1769 unsigned NumRC =
TRI->getNumRegClasses();
1770 RegLimit.resize(NumRC);
1772 std::fill(RegLimit.begin(), RegLimit.end(), 0);
1779 void setScheduleDAG(ScheduleDAGRRList *scheduleDag) {
1780 scheduleDAG = scheduleDag;
1784 return scheduleDAG->getHazardRec();
1787 void initNodes(std::vector<SUnit> &sunits)
override;
1795 SethiUllmanNumbers.clear();
1799 unsigned getNodePriority(
const SUnit *SU)
const;
1801 unsigned getNodeOrdering(
const SUnit *SU)
const {
1807 bool empty()
const override {
return Queue.empty(); }
1810 assert(!
U->NodeQueueId &&
"Node in the queue already");
1811 U->NodeQueueId = ++CurQueueId;
1818 std::vector<SUnit *>::iterator
I =
llvm::find(Queue, SU);
1819 if (
I != std::prev(
Queue.end()))
1827 void dumpRegPressure()
const;
1829 bool HighRegPressure(
const SUnit *SU)
const;
1831 bool MayReduceRegPressure(
SUnit *SU)
const;
1833 int RegPressureDiff(
SUnit *SU,
unsigned &LiveUses)
const;
1841 void AddPseudoTwoAddrDeps();
1842 void PrescheduleNodesWithMultipleUses();
1843 void CalculateSethiUllmanNumbers();
1847static SUnit *popFromQueueImpl(std::vector<SUnit *> &Q, SF &Picker) {
1848 unsigned BestIdx = 0;
1851 for (
unsigned I = 1, E = std::min(Q.size(), (
decltype(Q.size()))1000);
I != E;
1853 if (Picker(Q[BestIdx], Q[
I]))
1856 if (BestIdx + 1 != Q.size())
1863SUnit *popFromQueue(std::vector<SUnit *> &Q, SF &Picker,
ScheduleDAG *DAG) {
1866 reverse_sort<SF> RPicker(Picker);
1867 return popFromQueueImpl(Q, RPicker);
1871 return popFromQueueImpl(Q, Picker);
1882class RegReductionPriorityQueue :
public RegReductionPQBase {
1892 : RegReductionPQBase(mf, SF::HasReadyFilter, tracksrp, srcorder,
1896 bool isBottomUp()
const override {
return SF::IsBottomUp; }
1898 bool isReady(
SUnit *U)
const override {
1899 return Picker.HasReadyFilter && Picker.isReady(U, getCurCycle());
1902 SUnit *pop()
override {
1903 if (
Queue.empty())
return nullptr;
1905 SUnit *
V = popFromQueue(Queue, Picker, scheduleDAG);
1910#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1913 std::vector<SUnit *> DumpQueue =
Queue;
1914 SF DumpPicker = Picker;
1915 while (!DumpQueue.empty()) {
1916 SUnit *SU = popFromQueue(DumpQueue, DumpPicker, scheduleDAG);
1924using BURegReductionPriorityQueue = RegReductionPriorityQueue<bu_ls_rr_sort>;
1925using SrcRegReductionPriorityQueue = RegReductionPriorityQueue<src_ls_rr_sort>;
1926using HybridBURRPriorityQueue = RegReductionPriorityQueue<hybrid_ls_rr_sort>;
1927using ILPBURRPriorityQueue = RegReductionPriorityQueue<ilp_ls_rr_sort>;
1944 if (LSchedLow != RSchedLow)
1945 return LSchedLow < RSchedLow ? 1 : -1;
1953 if (SUNumbers[SU->
NodeNum] != 0)
1954 return SUNumbers[SU->
NodeNum];
1958 WorkState(
const SUnit *SU) : SU(SU) {}
1960 unsigned PredsProcessed = 0;
1965 while (!WorkList.
empty()) {
1966 auto &Temp = WorkList.
back();
1967 auto *TempSU = Temp.SU;
1968 bool AllPredsKnown =
true;
1970 for (
unsigned P = Temp.PredsProcessed; P < TempSU->Preds.size(); ++
P) {
1971 auto &Pred = TempSU->Preds[
P];
1972 if (Pred.
isCtrl())
continue;
1974 if (SUNumbers[PredSU->
NodeNum] == 0) {
1977 for (
auto It : WorkList)
1978 assert(It.SU != PredSU &&
"Trying to push an element twice?");
1981 Temp.PredsProcessed =
P + 1;
1982 WorkList.push_back(PredSU);
1983 AllPredsKnown =
false;
1992 unsigned SethiUllmanNumber = 0;
1994 for (
const SDep &Pred : TempSU->Preds) {
1995 if (Pred.
isCtrl())
continue;
1997 unsigned PredSethiUllman = SUNumbers[PredSU->
NodeNum];
1998 assert(PredSethiUllman > 0 &&
"We should have evaluated this pred!");
1999 if (PredSethiUllman > SethiUllmanNumber) {
2000 SethiUllmanNumber = PredSethiUllman;
2002 }
else if (PredSethiUllman == SethiUllmanNumber)
2006 SethiUllmanNumber += Extra;
2007 if (SethiUllmanNumber == 0)
2008 SethiUllmanNumber = 1;
2009 SUNumbers[TempSU->NodeNum] = SethiUllmanNumber;
2013 assert(SUNumbers[SU->
NodeNum] > 0 &&
"SethiUllman should never be zero!");
2014 return SUNumbers[SU->
NodeNum];
2019void RegReductionPQBase::CalculateSethiUllmanNumbers() {
2020 SethiUllmanNumbers.assign(SUnits->size(), 0);
2022 for (
const SUnit &SU : *SUnits)
2026void RegReductionPQBase::addNode(
const SUnit *SU) {
2027 unsigned SUSize = SethiUllmanNumbers.size();
2028 if (SUnits->size() > SUSize)
2029 SethiUllmanNumbers.resize(SUSize*2, 0);
2033void RegReductionPQBase::updateNode(
const SUnit *SU) {
2034 SethiUllmanNumbers[SU->
NodeNum] = 0;
2040unsigned RegReductionPQBase::getNodePriority(
const SUnit *SU)
const {
2047 if (Opc == TargetOpcode::EXTRACT_SUBREG ||
2048 Opc == TargetOpcode::SUBREG_TO_REG ||
2049 Opc == TargetOpcode::INSERT_SUBREG)
2065 return SethiUllmanNumbers[SU->
NodeNum];
2067 unsigned Priority = SethiUllmanNumbers[SU->
NodeNum];
2071 return (NP > 0) ? NP : 0;
2081#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
2088 << RegLimit[Id] <<
'\n');
2093bool RegReductionPQBase::HighRegPressure(
const SUnit *SU)
const {
2107 RegDefPos.
IsValid(); RegDefPos.Advance()) {
2108 unsigned RCId,
Cost;
2111 if ((RegPressure[RCId] +
Cost) >= RegLimit[RCId])
2118bool RegReductionPQBase::MayReduceRegPressure(
SUnit *SU)
const {
2121 if (!
N->isMachineOpcode() || !SU->
NumSuccs)
2124 unsigned NumDefs =
TII->get(
N->getMachineOpcode()).getNumDefs();
2125 for (
unsigned i = 0; i != NumDefs; ++i) {
2126 MVT VT =
N->getSimpleValueType(i);
2127 if (!
N->hasAnyUseOfValue(i))
2130 if (RegPressure[RCId] >= RegLimit[RCId])
2143int RegReductionPQBase::RegPressureDiff(
SUnit *SU,
unsigned &LiveUses)
const {
2158 RegDefPos.
IsValid(); RegDefPos.Advance()) {
2159 MVT VT = RegDefPos.GetValue();
2161 if (RegPressure[RCId] >= RegLimit[RCId])
2167 if (!
N || !
N->isMachineOpcode() || !SU->
NumSuccs)
2170 unsigned NumDefs =
TII->get(
N->getMachineOpcode()).getNumDefs();
2171 for (
unsigned i = 0; i != NumDefs; ++i) {
2172 MVT VT =
N->getSimpleValueType(i);
2173 if (!
N->hasAnyUseOfValue(i))
2176 if (RegPressure[RCId] >= RegLimit[RCId])
2182void RegReductionPQBase::scheduledNode(
SUnit *SU) {
2183 if (!TracksRegPressure)
2216 RegDefPos.
IsValid(); RegDefPos.Advance(), --SkipRegDefs) {
2220 unsigned RCId,
Cost;
2232 RegDefPos.
IsValid(); RegDefPos.Advance(), --SkipRegDefs) {
2233 if (SkipRegDefs > 0)
2235 unsigned RCId,
Cost;
2237 if (RegPressure[RCId] <
Cost) {
2241 <<
") has too many regdefs\n");
2251void RegReductionPQBase::unscheduledNode(
SUnit *SU) {
2252 if (!TracksRegPressure)
2258 if (!
N->isMachineOpcode()) {
2262 unsigned Opc =
N->getMachineOpcode();
2263 if (Opc == TargetOpcode::EXTRACT_SUBREG ||
2264 Opc == TargetOpcode::INSERT_SUBREG ||
2265 Opc == TargetOpcode::SUBREG_TO_REG ||
2266 Opc == TargetOpcode::REG_SEQUENCE ||
2267 Opc == TargetOpcode::IMPLICIT_DEF)
2289 if (POpc == TargetOpcode::IMPLICIT_DEF)
2291 if (POpc == TargetOpcode::EXTRACT_SUBREG ||
2292 POpc == TargetOpcode::INSERT_SUBREG ||
2293 POpc == TargetOpcode::SUBREG_TO_REG) {
2299 if (POpc == TargetOpcode::REG_SEQUENCE) {
2302 unsigned RCId = RC->
getID();
2309 for (
unsigned i = 0; i != NumDefs; ++i) {
2324 if (SU->
NumSuccs &&
N->isMachineOpcode()) {
2325 unsigned NumDefs =
TII->get(
N->getMachineOpcode()).getNumDefs();
2326 for (
unsigned i = NumDefs, e =
N->getNumValues(); i != e; ++i) {
2327 MVT VT =
N->getSimpleValueType(i);
2328 if (VT == MVT::Glue || VT == MVT::Other)
2330 if (!
N->hasAnyUseOfValue(i))
2347 unsigned MaxHeight = 0;
2349 if (Succ.
isCtrl())
continue;
2356 if (Height > MaxHeight)
2365 unsigned Scratches = 0;
2367 if (Pred.
isCtrl())
continue;
2376 bool RetVal =
false;
2378 if (Pred.
isCtrl())
continue;
2384 if (Reg.isVirtual()) {
2398 bool RetVal =
false;
2400 if (Succ.
isCtrl())
continue;
2405 if (Reg.isVirtual()) {
2437 if (Pred.
isCtrl())
continue;
2449 if (Pred.
isCtrl())
continue;
2453 "VRegCycle def must be CopyFromReg");
2467 if (Pred.
isCtrl())
continue;
2481 if ((
int)SPQ->getCurCycle() < Height)
return true;
2482 if (SPQ->getHazardRec()->getHazardType(SU, 0)
2491 RegReductionPQBase *SPQ) {
2496 int LHeight = (int)left->
getHeight() + LPenalty;
2497 int RHeight = (int)right->
getHeight() + RPenalty;
2510 if (LHeight != RHeight)
2511 return LHeight > RHeight ? 1 : -1;
2523 if (!SPQ->getHazardRec()->isEnabled()) {
2524 if (LHeight != RHeight)
2525 return LHeight > RHeight ? 1 : -1;
2527 int LDepth = left->
getDepth() - LPenalty;
2528 int RDepth = right->
getDepth() - RPenalty;
2529 if (LDepth != RDepth) {
2531 <<
") depth " << LDepth <<
" vs SU (" << right->
NodeNum
2532 <<
") depth " << RDepth <<
"\n");
2533 return LDepth < RDepth ? 1 : -1;
2549 if (LHasPhysReg != RHasPhysReg) {
2551 static const char *
const PhysRegMsg[] = {
" has no physreg",
2552 " defines a physreg" };
2555 << PhysRegMsg[LHasPhysReg] <<
" SU(" << right->
NodeNum
2556 <<
") " << PhysRegMsg[RHasPhysReg] <<
"\n");
2557 return LHasPhysReg < RHasPhysReg;
2562 unsigned LPriority = SPQ->getNodePriority(left);
2563 unsigned RPriority = SPQ->getNodePriority(right);
2569 RPriority = (RPriority > RNumVals) ? (RPriority - RNumVals) : 0;
2573 LPriority = (LPriority > LNumVals) ? (LPriority - LNumVals) : 0;
2576 if (LPriority != RPriority)
2577 return LPriority > RPriority;
2582 unsigned LOrder = SPQ->getNodeOrdering(left);
2583 unsigned ROrder = SPQ->getNodeOrdering(right);
2587 if ((LOrder || ROrder) && LOrder != ROrder)
2588 return LOrder != 0 && (LOrder < ROrder || ROrder == 0);
2611 return LDist < RDist;
2616 if (LScratch != RScratch)
2617 return LScratch > RScratch;
2621 if ((left->
isCall && RPriority > 0) || (right->
isCall && LPriority > 0))
2640 "NodeQueueId cannot be zero");
2645bool bu_ls_rr_sort::operator()(
SUnit *left,
SUnit *right)
const {
2653bool src_ls_rr_sort::operator()(
SUnit *left,
SUnit *right)
const {
2657 unsigned LOrder = SPQ->getNodeOrdering(left);
2658 unsigned ROrder = SPQ->getNodeOrdering(right);
2662 if ((LOrder || ROrder) && LOrder != ROrder)
2663 return LOrder != 0 && (LOrder < ROrder || ROrder == 0);
2672bool hybrid_ls_rr_sort::isReady(
SUnit *SU,
unsigned CurCycle)
const {
2673 static const unsigned ReadyDelay = 3;
2675 if (SPQ->MayReduceRegPressure(SU))
return true;
2677 if (SU->
getHeight() > (CurCycle + ReadyDelay))
return false;
2679 if (SPQ->getHazardRec()->getHazardType(SU, -ReadyDelay)
2687bool hybrid_ls_rr_sort::operator()(
SUnit *left,
SUnit *right)
const {
2695 bool LHigh = SPQ->HighRegPressure(left);
2696 bool RHigh = SPQ->HighRegPressure(right);
2699 if (LHigh && !RHigh) {
2704 else if (!LHigh && RHigh) {
2709 if (!LHigh && !RHigh) {
2719bool ilp_ls_rr_sort::isReady(
SUnit *SU,
unsigned CurCycle)
const {
2720 if (SU->
getHeight() > CurCycle)
return false;
2722 if (SPQ->getHazardRec()->getHazardType(SU, 0)
2736 if (Opc == TargetOpcode::EXTRACT_SUBREG ||
2737 Opc == TargetOpcode::SUBREG_TO_REG ||
2738 Opc == TargetOpcode::INSERT_SUBREG)
2753bool ilp_ls_rr_sort::operator()(
SUnit *left,
SUnit *right)
const {
2761 unsigned LLiveUses = 0, RLiveUses = 0;
2762 int LPDiff = 0, RPDiff = 0;
2764 LPDiff = SPQ->RegPressureDiff(left, LLiveUses);
2765 RPDiff = SPQ->RegPressureDiff(right, RLiveUses);
2769 <<
"): " << LPDiff <<
" != SU(" << right->
NodeNum
2770 <<
"): " << RPDiff <<
"\n");
2771 return LPDiff > RPDiff;
2777 if (LReduce && !RReduce)
return false;
2778 if (RReduce && !LReduce)
return true;
2783 <<
" != SU(" << right->
NodeNum <<
"): " << RLiveUses
2785 return LLiveUses < RLiveUses;
2791 if (LStall != RStall)
2800 <<
"): " << right->
getDepth() <<
"\n");
2814void RegReductionPQBase::initNodes(std::vector<SUnit> &sunits) {
2818 AddPseudoTwoAddrDeps();
2820 if (!TracksRegPressure && !SrcOrder)
2821 PrescheduleNodesWithMultipleUses();
2823 CalculateSethiUllmanNumbers();
2826 if (scheduleDAG->BB->isSuccessor(scheduleDAG->BB))
2827 for (
SUnit &SU : sunits)
2835bool RegReductionPQBase::canClobber(
const SUnit *SU,
const SUnit *
Op) {
2841 for (
unsigned i = 0; i != NumOps; ++i) {
2857 ScheduleDAGRRList *scheduleDAG,
2863 if (ImpDefs.
empty() && !RegMask)
2868 for (
const SDep &SuccPred : SuccSU->
Preds) {
2874 scheduleDAG->IsReachable(DepSU, SuccPred.
getSUnit()))
2881 if (
TRI->regsOverlap(ImpDef, SuccPred.
getReg()) &&
2882 scheduleDAG->IsReachable(DepSU, SuccPred.
getSUnit()))
2896 unsigned NumDefs =
TII->get(
N->getMachineOpcode()).getNumDefs();
2898 assert(!ImpDefs.
empty() &&
"Caller should check hasPhysRegDefs");
2901 if (!SUNode->isMachineOpcode())
2904 TII->get(SUNode->getMachineOpcode()).implicit_defs();
2906 if (SUImpDefs.
empty() && !SURegMask)
2908 for (
unsigned i = NumDefs, e =
N->getNumValues(); i != e; ++i) {
2909 MVT VT =
N->getSimpleValueType(i);
2910 if (VT == MVT::Glue || VT == MVT::Other)
2912 if (!
N->hasAnyUseOfValue(i))
2918 if (
TRI->regsOverlap(Reg, SUReg))
2956void RegReductionPQBase::PrescheduleNodesWithMultipleUses() {
2958 for (
SUnit &SU : *SUnits) {
2971 cast<RegisterSDNode>(
N->getOperand(1))->getReg().isVirtual())
2974 SDNode *PredFrameSetup =
nullptr;
2989 PredFrameSetup = PredND;
2994 if (PredFrameSetup !=
nullptr)
2998 SUnit *PredSU =
nullptr;
3017 cast<RegisterSDNode>(
N->getOperand(1))->getReg().isVirtual())
3021 for (
const SDep &PredSucc : PredSU->
Succs) {
3023 if (PredSuccSU == &SU)
continue;
3027 goto outer_loop_continue;
3031 goto outer_loop_continue;
3033 if (scheduleDAG->IsReachable(&SU, PredSuccSU))
3034 goto outer_loop_continue;
3040 dbgs() <<
" Prescheduling SU #" << SU.
NodeNum <<
" next to PredSU #"
3042 <<
" to guide scheduling in the presence of multiple uses\n");
3043 for (
unsigned i = 0; i != PredSU->
Succs.size(); ++i) {
3047 if (SuccSU != &SU) {
3049 scheduleDAG->RemovePred(SuccSU, Edge);
3050 scheduleDAG->AddPredQueued(&SU, Edge);
3052 scheduleDAG->AddPredQueued(SuccSU, Edge);
3056 outer_loop_continue:;
3067void RegReductionPQBase::AddPseudoTwoAddrDeps() {
3068 for (
SUnit &SU : *SUnits) {
3077 unsigned Opc =
Node->getMachineOpcode();
3081 for (
unsigned j = 0;
j != NumOps; ++
j) {
3105 while (SuccSU->
Succs.size() == 1 &&
3108 TargetOpcode::COPY_TO_REGCLASS)
3109 SuccSU = SuccSU->
Succs.front().getSUnit();
3122 if (SuccOpc == TargetOpcode::EXTRACT_SUBREG ||
3123 SuccOpc == TargetOpcode::INSERT_SUBREG ||
3124 SuccOpc == TargetOpcode::SUBREG_TO_REG)
3127 (!canClobber(SuccSU, DUSU) ||
3130 !scheduleDAG->IsReachable(SuccSU, &SU)) {
3132 <<
" Adding a pseudo-two-addr edge from SU #"
3151 BURegReductionPriorityQueue *PQ =
3152 new BURegReductionPriorityQueue(*IS->
MF,
false,
false,
TII,
TRI,
nullptr);
3153 ScheduleDAGRRList *SD =
new ScheduleDAGRRList(*IS->
MF,
false, PQ, OptLevel);
3154 PQ->setScheduleDAG(SD);
3165 SrcRegReductionPriorityQueue *PQ =
3166 new SrcRegReductionPriorityQueue(*IS->
MF,
false,
true,
TII,
TRI,
nullptr);
3167 ScheduleDAGRRList *SD =
new ScheduleDAGRRList(*IS->
MF,
false, PQ, OptLevel);
3168 PQ->setScheduleDAG(SD);
3180 HybridBURRPriorityQueue *PQ =
3181 new HybridBURRPriorityQueue(*IS->
MF,
true,
false,
TII,
TRI, TLI);
3183 ScheduleDAGRRList *SD =
new ScheduleDAGRRList(*IS->
MF,
true, PQ, OptLevel);
3184 PQ->setScheduleDAG(SD);
3195 ILPBURRPriorityQueue *PQ =
3196 new ILPBURRPriorityQueue(*IS->
MF,
true,
false,
TII,
TRI, TLI);
3197 ScheduleDAGRRList *SD =
new ScheduleDAGRRList(*IS->
MF,
true, PQ, OptLevel);
3198 PQ->setScheduleDAG(SD);
for(const MachineOperand &MO :llvm::drop_begin(OldMI.operands(), Desc.getNumOperands()))
static msgpack::DocNode getNode(msgpack::DocNode DN, msgpack::Type Type, MCValue Val)
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
#define LLVM_DUMP_METHOD
Mark debug helper function definitions like dump() that should not be stripped from debug builds.
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
This file defines the DenseMap class.
const HexagonInstrInfo * TII
unsigned const TargetRegisterInfo * TRI
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
static bool isLiveOut(const MachineBasicBlock &MBB, unsigned Reg)
static bool canEnableCoalescing(SUnit *SU)
static RegisterScheduler sourceListDAGScheduler("source", "Similar to list-burr but schedules in source " "order when possible", createSourceListDAGScheduler)
static cl::opt< bool > DisableSchedCycles("disable-sched-cycles", cl::Hidden, cl::init(false), cl::desc("Disable cycle-level precision during preRA scheduling"))
static cl::opt< bool > DisableSchedStalls("disable-sched-stalls", cl::Hidden, cl::init(true), cl::desc("Disable no-stall priority in sched=list-ilp"))
static bool hasOnlyLiveInOpers(const SUnit *SU)
hasOnlyLiveInOpers - Return true if SU has only value predecessors that are CopyFromReg from a virtua...
static bool IsChainDependent(SDNode *Outer, SDNode *Inner, unsigned NestLevel, const TargetInstrInfo *TII)
IsChainDependent - Test if Outer is reachable from Inner through chain dependencies.
static bool hasOnlyLiveOutUses(const SUnit *SU)
hasOnlyLiveOutUses - Return true if SU has only value successors that are CopyToReg to a virtual regi...
static MVT getPhysicalRegisterVT(SDNode *N, unsigned Reg, const TargetInstrInfo *TII)
getPhysicalRegisterVT - Returns the ValueType of the physical register definition of the specified no...
static cl::opt< bool > DisableSchedCriticalPath("disable-sched-critical-path", cl::Hidden, cl::init(false), cl::desc("Disable critical path priority in sched=list-ilp"))
static cl::opt< bool > Disable2AddrHack("disable-2addr-hack", cl::Hidden, cl::init(true), cl::desc("Disable scheduler's two-address hack"))
static void CheckForLiveRegDef(SUnit *SU, unsigned Reg, SUnit **LiveRegDefs, SmallSet< unsigned, 4 > &RegAdded, SmallVectorImpl< unsigned > &LRegs, const TargetRegisterInfo *TRI, const SDNode *Node=nullptr)
CheckForLiveRegDef - Return true and update live register vector if the specified register def of the...
static RegisterScheduler ILPListDAGScheduler("list-ilp", "Bottom-up register pressure aware list scheduling " "which tries to balance ILP and register pressure", createILPListDAGScheduler)
static void resetVRegCycle(SUnit *SU)
static RegisterScheduler hybridListDAGScheduler("list-hybrid", "Bottom-up register pressure aware list scheduling " "which tries to balance latency and register pressure", createHybridListDAGScheduler)
static bool canClobberReachingPhysRegUse(const SUnit *DepSU, const SUnit *SU, ScheduleDAGRRList *scheduleDAG, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
canClobberReachingPhysRegUse - True if SU would clobber one of it's successor's explicit physregs who...
static cl::opt< bool > DisableSchedPhysRegJoin("disable-sched-physreg-join", cl::Hidden, cl::init(false), cl::desc("Disable physreg def-use affinity"))
static bool canClobberPhysRegDefs(const SUnit *SuccSU, const SUnit *SU, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
canClobberPhysRegDefs - True if SU would clobber one of SuccSU's physical register defs.
static void GetCostForDef(const ScheduleDAGSDNodes::RegDefIter &RegDefPos, const TargetLowering *TLI, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI, unsigned &RegClass, unsigned &Cost, const MachineFunction &MF)
GetCostForDef - Looks up the register class and cost for a given definition.
static bool BURRSort(SUnit *left, SUnit *right, RegReductionPQBase *SPQ)
static cl::opt< bool > DisableSchedRegPressure("disable-sched-reg-pressure", cl::Hidden, cl::init(false), cl::desc("Disable regpressure priority in sched=list-ilp"))
static bool BUHasStall(SUnit *SU, int Height, RegReductionPQBase *SPQ)
static void initVRegCycle(SUnit *SU)
static constexpr unsigned RegSequenceCost
static cl::opt< int > MaxReorderWindow("max-sched-reorder", cl::Hidden, cl::init(6), cl::desc("Number of instructions to allow ahead of the critical path " "in sched=list-ilp"))
static SDNode * FindCallSeqStart(SDNode *N, unsigned &NestLevel, unsigned &MaxNest, const TargetInstrInfo *TII)
FindCallSeqStart - Starting from the (lowered) CALLSEQ_END node, locate the corresponding (lowered) C...
static bool isOperandOf(const SUnit *SU, SDNode *N)
static cl::opt< bool > DisableSchedVRegCycle("disable-sched-vrcycle", cl::Hidden, cl::init(false), cl::desc("Disable virtual register cycle interference checks"))
static int checkSpecialNodes(const SUnit *left, const SUnit *right)
static cl::opt< bool > DisableSchedLiveUses("disable-sched-live-uses", cl::Hidden, cl::init(true), cl::desc("Disable live use priority in sched=list-ilp"))
static const uint32_t * getNodeRegMask(const SDNode *N)
getNodeRegMask - Returns the register mask attached to an SDNode, if any.
static unsigned closestSucc(const SUnit *SU)
closestSucc - Returns the scheduled cycle of the successor which is closest to the current cycle.
static cl::opt< unsigned > AvgIPC("sched-avg-ipc", cl::Hidden, cl::init(1), cl::desc("Average inst/cycle whan no target itinerary exists."))
static bool hasVRegCycleUse(const SUnit *SU)
static cl::opt< bool > DisableSchedHeight("disable-sched-height", cl::Hidden, cl::init(false), cl::desc("Disable scheduled-height priority in sched=list-ilp"))
static RegisterScheduler burrListDAGScheduler("list-burr", "Bottom-up register reduction list scheduling", createBURRListDAGScheduler)
static unsigned calcMaxScratches(const SUnit *SU)
calcMaxScratches - Returns an cost estimate of the worse case requirement for scratch registers,...
static unsigned CalcNodeSethiUllmanNumber(const SUnit *SU, std::vector< unsigned > &SUNumbers)
CalcNodeSethiUllmanNumber - Compute Sethi Ullman number.
static int BUCompareLatency(SUnit *left, SUnit *right, bool checkPref, RegReductionPQBase *SPQ)
static void CheckForLiveRegDefMasked(SUnit *SU, const uint32_t *RegMask, ArrayRef< SUnit * > LiveRegDefs, SmallSet< unsigned, 4 > &RegAdded, SmallVectorImpl< unsigned > &LRegs)
CheckForLiveRegDefMasked - Check for any live physregs that are clobbered by RegMask,...
This file defines the SmallSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
This file describes how to lower LLVM code to machine code.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
size_t size() const
size - Get the array size.
bool empty() const
empty - Check if the array is empty.
This class represents an Operation in the Expression.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
bool hasOptionalDef() const
Set if this instruction has an optional definition, e.g.
unsigned getNumDefs() const
Return the number of MachineOperands that are register definitions.
int getOperandConstraint(unsigned OpNum, MCOI::OperandConstraint Constraint) const
Returns the value of the specified operand constraint if it is present.
ArrayRef< MCPhysReg > implicit_defs() const
Return a list of registers that are potentially written by any instance of this machine instruction.
bool isCommutable() const
Return true if this may be a 2- or 3-address instruction (of the form "X = op Y, Z,...
MCRegAliasIterator enumerates all registers aliasing Reg.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
static bool clobbersPhysReg(const uint32_t *RegMask, MCRegister PhysReg)
clobbersPhysReg - Returns true if this RegMask clobbers PhysReg.
const TargetRegisterClass * getRegClass(Register Reg) const
Return the register class of the specified virtual register.
Wrapper class representing virtual and physical registers.
Represents one node in the SelectionDAG.
bool isMachineOpcode() const
Test if this node has a post-isel opcode, directly corresponding to a MachineInstr opcode.
int getNodeId() const
Return the unique node id.
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
unsigned getIROrder() const
Return the node ordering.
void setNodeId(int Id)
Set unique node id.
MVT getSimpleValueType(unsigned ResNo) const
Return the type of a specified result as a simple type.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
unsigned getMachineOpcode() const
This may only be called if isMachineOpcode returns true.
const SDValue & getOperand(unsigned Num) const
uint64_t getConstantOperandVal(unsigned Num) const
Helper method returns the integer value of a ConstantSDNode operand.
bool hasAnyUseOfValue(unsigned Value) const
Return true if there are any use of the indicated value.
SDNode * getGluedNode() const
If this node has a glue operand, return the node to which the glue operand points.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
@ Data
Regular data dependence (aka true-dependence).
@ Artificial
Arbitrary strong DAG edge (no real dependence).
unsigned getLatency() const
Returns the latency value for this edge, which roughly means the minimum number of cycles that must e...
bool isAssignedRegDep() const
Tests if this is a Data dependence that is associated with a register.
bool isArtificial() const
Tests if this is an Order dependence that is marked as "artificial", meaning it isn't necessary for c...
bool isCtrl() const
Shorthand for getKind() != SDep::Data.
unsigned getReg() const
Returns the register associated with this edge.
Scheduling unit. This is a node in the scheduling DAG.
bool isCall
Is a function call.
void setHeightToAtLeast(unsigned NewHeight)
If NewHeight is greater than this node's height value, set it to be the new height value.
unsigned NodeQueueId
Queue id of node.
unsigned NodeNum
Entry # of node in the node vector.
bool hasPhysRegClobbers
Has any physreg defs, used or not.
bool isCallOp
Is a function call operand.
const TargetRegisterClass * CopyDstRC
Is a special copy node if != nullptr.
unsigned getHeight() const
Returns the height of this node, which is the length of the maximum path down to any node which has n...
void setHeightDirty()
Sets a flag in this node to indicate that its stored Height value will require recomputation the next...
bool isSucc(const SUnit *N) const
Tests if node N is a successor of this node.
void removePred(const SDep &D)
Removes the specified edge as a pred of the current node if it exists.
unsigned short Latency
Node latency.
unsigned short NumRegDefsLeft
bool isPending
True once pending.
unsigned getDepth() const
Returns the depth of this node, which is the length of the maximum path up to any node which has no p...
bool isScheduled
True once scheduled.
bool isAvailable
True once available.
bool isScheduleLow
True if preferable to schedule low.
bool hasPhysRegDefs
Has physreg defs that are being used.
SmallVector< SDep, 4 > Succs
All sunit successors.
Sched::Preference SchedulingPref
Scheduling preference.
const TargetRegisterClass * CopySrcRC
SDNode * getNode() const
Returns the representative SDNode for this SUnit.
bool isTwoAddress
Is a two-address instruction.
bool isCommutable
Is a commutable instruction.
bool isVRegCycle
May use and def the same vreg.
SmallVector< SDep, 4 > Preds
All sunit predecessors.
bool addPred(const SDep &D, bool Required=true)
Adds the specified edge as a pred of the current node if not already.
RegDefIter - In place iteration over the values defined by an SUnit.
const SDNode * GetNode() const
ScheduleDAGSDNodes - A ScheduleDAG for scheduling SDNode-based DAGs.
SUnit * newSUnit(SDNode *N)
NewSUnit - Creates a new SUnit and return a ptr to it.
virtual void Schedule()=0
Schedule - Order nodes according to selected style, filling in the Sequence member.
virtual bool forceUnitLatencies() const
ForceUnitLatencies - Return true if all scheduling edges should be given a latency value of one.
SUnit * Clone(SUnit *Old)
Clone - Creates a clone of the specified SUnit.
This class can compute a topological ordering for SUnits and provides methods for dynamically updatin...
void RemovePred(SUnit *M, SUnit *N)
Updates the topological ordering to accommodate an edge to be removed from the specified node N from ...
bool WillCreateCycle(SUnit *TargetSU, SUnit *SU)
Returns true if addPred(TargetSU, SU) creates a cycle.
void MarkDirty()
Mark the ordering as temporarily broken, after a new node has been added.
void AddSUnitWithoutPredecessors(const SUnit *SU)
Add a SUnit without predecessors to the end of the topological order.
void AddPred(SUnit *Y, SUnit *X)
Updates the topological ordering to accommodate an edge to be added from SUnit X to SUnit Y.
bool IsReachable(const SUnit *SU, const SUnit *TargetSU)
Checks if SU is reachable from TargetSU.
void AddPredQueued(SUnit *Y, SUnit *X)
Queues an update to the topological ordering to accommodate an edge to be added from SUnit X to SUnit...
virtual void dumpNode(const SUnit &SU) const =0
HazardRecognizer - This determines whether or not an instruction can be issued this cycle,...
unsigned getMaxLookAhead() const
virtual void RecedeCycle()
RecedeCycle - This callback is invoked whenever the next bottom-up instruction to be scheduled cannot...
virtual void Reset()
Reset - This callback is invoked when a new block of instructions is about to be schedule.
virtual void EmitInstruction(SUnit *)
EmitInstruction - This callback is invoked when an instruction is emitted, to advance the hazard stat...
virtual bool atIssueLimit() const
atIssueLimit - Return true if no more instructions may be issued in this cycle.
virtual HazardType getHazardType(SUnit *, int Stalls=0)
getHazardType - Return the hazard type of emitting this node.
This interface is used to plug different priorities computation algorithms into the list scheduler.
void setCurCycle(unsigned Cycle)
virtual void remove(SUnit *SU)=0
virtual void releaseState()=0
virtual void scheduledNode(SUnit *)
As each node is scheduled, this method is invoked.
virtual bool isReady(SUnit *) const
virtual bool tracksRegPressure() const
virtual void dump(ScheduleDAG *) const
bool hasReadyFilter() const
virtual void initNodes(std::vector< SUnit > &SUnits)=0
virtual bool empty() const =0
virtual void unscheduledNode(SUnit *)
virtual void addNode(const SUnit *SU)=0
virtual void updateNode(const SUnit *SU)=0
virtual void push(SUnit *U)=0
SelectionDAGISel - This is the common base class used for SelectionDAG-based pattern-matching instruc...
const TargetLowering * TLI
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
std::pair< const_iterator, bool > insert(const T &V)
insert - Insert an element into the set if it isn't already there.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
TargetInstrInfo - Interface to description of machine instruction set.
virtual ScheduleHazardRecognizer * CreateTargetHazardRecognizer(const TargetSubtargetInfo *STI, const ScheduleDAG *DAG) const
Allocate and return a hazard recognizer to use for this target when scheduling the machine instructio...
virtual uint8_t getRepRegClassCostFor(MVT VT) const
Return the cost of the 'representative' register class for the specified value type.
virtual const TargetRegisterClass * getRepRegClassFor(MVT VT) const
Return the 'representative' register class for the specified value type.
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
unsigned getID() const
Return the register class ID number.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual unsigned getRegPressureLimit(const TargetRegisterClass *RC, MachineFunction &MF) const
Return the register pressure "high water mark" for the specific register class.
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetRegisterInfo * getRegisterInfo() const
getRegisterInfo - If register information is available, return it.
virtual const TargetInstrInfo * getInstrInfo() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
@ EH_LABEL
EH_LABEL - Represents a label in mid basic block used to track locations needed for debug and excepti...
@ CopyFromReg
CopyFromReg - This node indicates that the input value is a virtual or physical register that is defi...
@ EntryToken
EntryToken - This is the marker used to indicate the start of a region.
@ CopyToReg
CopyToReg - This node has three operands: a chain, a register number to set to this value,...
@ LIFETIME_START
This corresponds to the llvm.lifetime.
@ INLINEASM_BR
INLINEASM_BR - Branching version of inline asm. Used by asm-goto.
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
@ INLINEASM
INLINEASM - Represents an inline asm block.
Reg
All possible values of the reg field in the ModR/M byte.
initializer< Ty > init(const Ty &Val)
Sequence
A sequence of states that a pointer may go through in which an objc_retain and objc_release are actua...
NodeAddr< DefNode * > Def
This is an optimization pass for GlobalISel generic memory operations.
void dump(const SparseBitVector< ElementSize > &LHS, raw_ostream &out)
auto find(R &&Range, const T &Val)
Provide wrappers to std::find which take ranges instead of having to pass begin/end explicitly.
ScheduleDAGSDNodes * createBURRListDAGScheduler(SelectionDAGISel *IS, CodeGenOptLevel OptLevel)
createBURRListDAGScheduler - This creates a bottom up register usage reduction list scheduler.
ScheduleDAGSDNodes * createHybridListDAGScheduler(SelectionDAGISel *IS, CodeGenOptLevel)
createHybridListDAGScheduler - This creates a bottom up register pressure aware list scheduler that m...
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
void report_fatal_error(Error Err, bool gen_crash_diag=true)
Report a serious error, calling any installed error handler.
CodeGenOptLevel
Code generation optimization level.
ScheduleDAGSDNodes * createSourceListDAGScheduler(SelectionDAGISel *IS, CodeGenOptLevel OptLevel)
createSourceListDAGScheduler - This creates a bottom up list scheduler that schedules nodes in source...
ScheduleDAGSDNodes * createILPListDAGScheduler(SelectionDAGISel *IS, CodeGenOptLevel)
createILPListDAGScheduler - This creates a bottom up register pressure aware list scheduler that trie...
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
Printable printMBBReference(const MachineBasicBlock &MBB)
Prints a machine basic block reference.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Description of the encoding of one expression Op.