52#include "llvm/Config/llvm-config.h"
76#define DEBUG_TYPE "machine-scheduler"
79 "Number of instructions in source order after pre-RA scheduling");
81 "Number of instructions in source order after post-RA scheduling");
83 "Number of instructions scheduled by pre-RA scheduler");
85 "Number of instructions scheduled by post-RA scheduler");
86STATISTIC(NumClustered,
"Number of load/store pairs clustered");
89 "Number of scheduling units chosen from top queue pre-RA");
91 "Number of scheduling units chosen from bottom queue pre-RA");
93 "Number of scheduling units chosen for NoCand heuristic pre-RA");
95 "Number of scheduling units chosen for Only1 heuristic pre-RA");
97 "Number of scheduling units chosen for PhysReg heuristic pre-RA");
99 "Number of scheduling units chosen for RegExcess heuristic pre-RA");
101 "Number of scheduling units chosen for RegCritical heuristic pre-RA");
103 "Number of scheduling units chosen for Stall heuristic pre-RA");
105 "Number of scheduling units chosen for Cluster heuristic pre-RA");
107 "Number of scheduling units chosen for Weak heuristic pre-RA");
109 "Number of scheduling units chosen for RegMax heuristic pre-RA");
111 NumResourceReducePreRA,
112 "Number of scheduling units chosen for ResourceReduce heuristic pre-RA");
114 NumResourceDemandPreRA,
115 "Number of scheduling units chosen for ResourceDemand heuristic pre-RA");
117 NumTopDepthReducePreRA,
118 "Number of scheduling units chosen for TopDepthReduce heuristic pre-RA");
120 NumTopPathReducePreRA,
121 "Number of scheduling units chosen for TopPathReduce heuristic pre-RA");
123 NumBotHeightReducePreRA,
124 "Number of scheduling units chosen for BotHeightReduce heuristic pre-RA");
126 NumBotPathReducePreRA,
127 "Number of scheduling units chosen for BotPathReduce heuristic pre-RA");
129 "Number of scheduling units chosen for NodeOrder heuristic pre-RA");
131 "Number of scheduling units chosen for FirstValid heuristic pre-RA");
134 "Number of scheduling units chosen from top queue post-RA");
136 "Number of scheduling units chosen from bottom queue post-RA");
138 "Number of scheduling units chosen for NoCand heuristic post-RA");
140 "Number of scheduling units chosen for Only1 heuristic post-RA");
142 "Number of scheduling units chosen for PhysReg heuristic post-RA");
144 "Number of scheduling units chosen for RegExcess heuristic post-RA");
146 NumRegCriticalPostRA,
147 "Number of scheduling units chosen for RegCritical heuristic post-RA");
149 "Number of scheduling units chosen for Stall heuristic post-RA");
151 "Number of scheduling units chosen for Cluster heuristic post-RA");
153 "Number of scheduling units chosen for Weak heuristic post-RA");
155 "Number of scheduling units chosen for RegMax heuristic post-RA");
157 NumResourceReducePostRA,
158 "Number of scheduling units chosen for ResourceReduce heuristic post-RA");
160 NumResourceDemandPostRA,
161 "Number of scheduling units chosen for ResourceDemand heuristic post-RA");
163 NumTopDepthReducePostRA,
164 "Number of scheduling units chosen for TopDepthReduce heuristic post-RA");
166 NumTopPathReducePostRA,
167 "Number of scheduling units chosen for TopPathReduce heuristic post-RA");
169 NumBotHeightReducePostRA,
170 "Number of scheduling units chosen for BotHeightReduce heuristic post-RA");
172 NumBotPathReducePostRA,
173 "Number of scheduling units chosen for BotPathReduce heuristic post-RA");
175 "Number of scheduling units chosen for NodeOrder heuristic post-RA");
177 "Number of scheduling units chosen for FirstValid heuristic post-RA");
181 cl::desc(
"Pre reg-alloc list scheduling direction"),
185 "Force top-down pre reg-alloc list scheduling"),
187 "Force bottom-up pre reg-alloc list scheduling"),
189 "Force bidirectional pre reg-alloc list scheduling")));
193 cl::desc(
"Post reg-alloc list scheduling direction"),
197 "Force top-down post reg-alloc list scheduling"),
199 "Force bottom-up post reg-alloc list scheduling"),
201 "Force bidirectional post reg-alloc list scheduling")));
205 cl::desc(
"Print critical path length to stdout"));
209 cl::desc(
"Verify machine instrs before and after machine scheduling"));
214 cl::desc(
"Pop up a window to show MISched dags after they are processed"));
219 cl::desc(
"Dump resource usage at schedule boundary."));
222 cl::desc(
"Show details of invoking getNextResoufceCycle."));
227#ifdef LLVM_ENABLE_DUMP
236 cl::desc(
"Hide nodes with more predecessor/successor than cutoff"));
242 cl::desc(
"Only schedule this function"));
244 cl::desc(
"Only schedule this MBB#"));
259 cl::desc(
"Enable memop clustering."),
263 cl::desc(
"Switch to fast cluster algorithm with the lost "
264 "of some fusion opportunities"),
268 cl::desc(
"The threshold for fast cluster"),
271#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
274 cl::desc(
"Dump resource usage at schedule boundary."));
277 cl::desc(
"Set width of the columns with "
278 "the resources and schedule units"),
282 cl::desc(
"Set width of the columns showing resource booking."),
286 cl::desc(
"Sort the resources printed in the dump trace"));
297void MachineSchedStrategy::anchor() {}
299void ScheduleDAGMutation::anchor() {}
344 const RequiredAnalyses &Analyses);
368 const RequiredAnalyses &Analyses);
384 MachineSchedulerImpl Impl;
387 MachineSchedulerLegacy();
388 void getAnalysisUsage(AnalysisUsage &AU)
const override;
389 bool runOnMachineFunction(MachineFunction&)
override;
396 PostMachineSchedulerImpl Impl;
399 PostMachineSchedulerLegacy();
400 void getAnalysisUsage(AnalysisUsage &AU)
const override;
401 bool runOnMachineFunction(MachineFunction &)
override;
408char MachineSchedulerLegacy::ID = 0;
413 "Machine Instruction Scheduler",
false,
false)
421 "Machine Instruction Scheduler",
false,
false)
427void MachineSchedulerLegacy::getAnalysisUsage(
AnalysisUsage &AU)
const {
441char PostMachineSchedulerLegacy::ID = 0;
446 "PostRA Machine Instruction Scheduler",
false,
false)
453PostMachineSchedulerLegacy::PostMachineSchedulerLegacy()
458void PostMachineSchedulerLegacy::getAnalysisUsage(
AnalysisUsage &AU)
const {
481 cl::desc(
"Machine instruction scheduler to use"));
489 cl::desc(
"Enable the machine instruction scheduling pass."),
cl::init(
true),
493 "enable-post-misched",
494 cl::desc(
"Enable the post-ra machine instruction scheduling pass."),
501 assert(
I != Beg &&
"reached the top of the region, cannot decrement");
503 if (!
I->isDebugOrPseudoInstr())
522 for(;
I != End; ++
I) {
523 if (!
I->isDebugOrPseudoInstr())
565 const char *MSchedBanner =
"Before machine scheduling.";
567 MF->verify(P, MSchedBanner, &
errs());
569 MF->verify(*MFAM, MSchedBanner, &
errs());
580 const char *MSchedBanner =
"After machine scheduling.";
582 MF->verify(P, MSchedBanner, &
errs());
584 MF->verify(*MFAM, MSchedBanner, &
errs());
611 const char *PostMSchedBanner =
"Before post machine scheduling.";
613 MF->verify(P, PostMSchedBanner, &
errs());
615 MF->verify(*MFAM, PostMSchedBanner, &
errs());
624 const char *PostMSchedBanner =
"After post machine scheduling.";
626 MF->verify(P, PostMSchedBanner, &
errs());
628 MF->verify(*MFAM, PostMSchedBanner, &
errs());
649bool MachineSchedulerLegacy::runOnMachineFunction(
MachineFunction &MF) {
662 auto &MLI = getAnalysis<MachineLoopInfoWrapperPass>().getLI();
663 auto &MDT = getAnalysis<MachineDominatorTreeWrapperPass>().getDomTree();
664 auto &TM = getAnalysis<TargetPassConfig>().getTM<
TargetMachine>();
665 auto &
AA = getAnalysis<AAResultsWrapperPass>().getAAResults();
666 auto &LIS = getAnalysis<LiveIntervalsWrapperPass>().getLIS();
667 auto &MBFI = getAnalysis<MachineBlockFrequencyInfoWrapperPass>().getMBFI();
668 Impl.setLegacyPass(
this);
669 return Impl.run(MF, TM, {MLI, MDT,
AA, LIS, MBFI});
673 : Impl(
std::make_unique<MachineSchedulerImpl>()), TM(TM) {}
679 : Impl(
std::make_unique<PostMachineSchedulerImpl>()), TM(TM) {}
703 Impl->setMFAM(&MFAM);
704 bool Changed = Impl->run(MF, *TM, {MLI, MDT,
AA, LIS, MBFI});
710 .preserve<SlotIndexesAnalysis>()
714bool PostMachineSchedulerLegacy::runOnMachineFunction(
MachineFunction &MF) {
726 auto &MLI = getAnalysis<MachineLoopInfoWrapperPass>().getLI();
727 auto &TM = getAnalysis<TargetPassConfig>().getTM<
TargetMachine>();
728 auto &
AA = getAnalysis<AAResultsWrapperPass>().getAAResults();
729 Impl.setLegacyPass(
this);
730 return Impl.run(MF, TM, {MLI,
AA});
749 Impl->setMFAM(&MFAM);
750 bool Changed = Impl->run(MF, *TM, {MLI,
AA});
773 return MI->isCall() ||
TII->isSchedulingBoundary(*
MI,
MBB, *MF) ||
782 bool RegionsTopDown) {
788 RegionEnd !=
MBB->begin(); RegionEnd =
I) {
791 if (RegionEnd !=
MBB->end() ||
798 unsigned NumRegionInstrs = 0;
800 for (;
I !=
MBB->begin(); --
I) {
804 if (!
MI.isDebugOrPseudoInstr()) {
813 if (NumRegionInstrs != 0)
818 std::reverse(Regions.
begin(), Regions.
end());
857 bool ScheduleSingleMI =
Scheduler.shouldScheduleSingleMIRegions();
861 unsigned NumRegionInstrs = R.NumRegionInstrs;
869 if (
I == RegionEnd || (!ScheduleSingleMI &&
I == std::prev(RegionEnd))) {
877 <<
" " <<
MBB->getName() <<
"\n From: " << *
I
879 if (RegionEnd !=
MBB->end())
dbgs() << *RegionEnd;
880 else dbgs() <<
"End\n";
881 dbgs() <<
" RegionInstrs: " << NumRegionInstrs <<
'\n');
884 errs() <<
":%bb. " <<
MBB->getNumber();
885 errs() <<
" " <<
MBB->getName() <<
" \n";
905#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
907 dbgs() <<
"Queue " << Name <<
": ";
908 for (
const SUnit *SU : Queue)
909 dbgs() << SU->NodeNum <<
" ";
936 dbgs() <<
"*** Scheduling failed! ***\n";
938 dbgs() <<
" has been released too many times!\n";
971 dbgs() <<
"*** Scheduling failed! ***\n";
973 dbgs() <<
" has been released too many times!\n";
1010 unsigned regioninstrs)
1020 else if (
SchedImpl->getPolicy().OnlyBottomUp)
1036 BB->splice(InsertPos,
BB,
MI);
1040 LIS->handleMove(*
MI,
true);
1048#if LLVM_ENABLE_ABI_BREAKING_CHECKS && !defined(NDEBUG)
1053 ++NumInstrsScheduled;
1085 bool IsTopNode =
false;
1090 LLVM_DEBUG(
dbgs() <<
"** ScheduleDAGMI::schedule picking next node\n");
1107 if (&*priorII ==
MI)
1129 dbgs() <<
"*** Final schedule for "
1146 assert(!SU.isBoundaryNode() &&
"Boundary node should not be in SUnits");
1149 SU.biasCriticalPath();
1152 if (!SU.NumPredsLeft)
1155 if (!SU.NumSuccsLeft)
1158 ExitSU.biasCriticalPath();
1168 for (
SUnit *SU : TopRoots)
1207 for (std::vector<std::pair<MachineInstr *, MachineInstr *>>
::iterator
1209 std::pair<MachineInstr *, MachineInstr *>
P = *std::prev(DI);
1220#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1232 dbgs() <<
" * Schedule table (TopDown):\n";
1250 for (
unsigned C = FirstCycle;
C <= LastCycle; ++
C)
1257 dbgs() <<
"Missing SUnit\n";
1260 std::string NodeName(
"SU(");
1261 NodeName += std::to_string(SU->
NodeNum) +
")";
1263 unsigned C = FirstCycle;
1264 for (;
C <= LastCycle; ++
C) {
1282 return std::tie(LHS.AcquireAtCycle, LHS.ReleaseAtCycle) <
1283 std::tie(RHS.AcquireAtCycle, RHS.ReleaseAtCycle);
1287 const std::string ResName =
1288 SchedModel.getResourceName(PI.ProcResourceIdx);
1293 for (
unsigned I = 0, E = PI.ReleaseAtCycle - PI.AcquireAtCycle;
I != E;
1296 while (
C++ <= LastCycle)
1313 dbgs() <<
" * Schedule table (BottomUp):\n";
1326 if ((
int)SU->
BotReadyCycle - PI->ReleaseAtCycle + 1 < LastCycle)
1327 LastCycle = (int)SU->
BotReadyCycle - PI->ReleaseAtCycle + 1;
1332 for (
int C = FirstCycle;
C >= LastCycle; --
C)
1339 dbgs() <<
"Missing SUnit\n";
1342 std::string NodeName(
"SU(");
1343 NodeName += std::to_string(SU->
NodeNum) +
")";
1346 for (;
C >= LastCycle; --
C) {
1363 return std::tie(LHS.AcquireAtCycle, LHS.ReleaseAtCycle) <
1364 std::tie(RHS.AcquireAtCycle, RHS.ReleaseAtCycle);
1368 const std::string ResName =
1369 SchedModel.getResourceName(PI.ProcResourceIdx);
1374 for (
unsigned I = 0, E = PI.ReleaseAtCycle - PI.AcquireAtCycle;
I != E;
1377 while (
C-- >= LastCycle)
1386#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1394 dbgs() <<
"* Schedule table (Bidirectional): not implemented\n";
1396 dbgs() <<
"* Schedule table: DumpDirection not set.\n";
1404 dbgs() <<
"Missing SUnit\n";
1429 if (!Reg.isVirtual())
1434 bool FoundDef =
false;
1436 if (MO2.getReg() == Reg && !MO2.isDead()) {
1447 for (; UI !=
VRegUses.end(); ++UI) {
1463 unsigned regioninstrs)
1477 "ShouldTrackLaneMasks requires ShouldTrackPressure");
1528 dbgs() <<
"Bottom Pressure: ";
1534 "Can't find the region bottom");
1542 unsigned Limit =
RegClassInfo->getRegPressureSetLimit(i);
1551 dbgs() <<
"Excess PSets: ";
1553 dbgs() <<
TRI->getRegPressureSetName(RCPS.getPSet()) <<
" ";
1561 const std::vector<unsigned> &NewMaxPressure) {
1567 unsigned ID = PC.getPSet();
1572 && NewMaxPressure[
ID] <= (
unsigned)std::numeric_limits<int16_t>::max())
1576 if (NewMaxPressure[
ID] >= Limit - 2) {
1578 << NewMaxPressure[
ID]
1579 << ((NewMaxPressure[
ID] > Limit) ?
" > " :
" <= ")
1591 if (!
P.VRegOrUnit.isVirtualReg())
1593 Register Reg =
P.VRegOrUnit.asVirtualReg();
1600 bool Decrement =
P.LaneMask.any();
1604 SUnit &SU = *V2SU.SU;
1614 <<
" UpdateRegPressure: SU(" << SU.
NodeNum <<
") "
1637 assert(VNI &&
"No live value at use.");
1640 SUnit *SU = V2SU.SU;
1664#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
1665 if (
EntrySU.getInstr() !=
nullptr)
1670 dbgs() <<
" Pressure Diff : ";
1673 dbgs() <<
" Single Issue : ";
1674 if (
SchedModel.mustBeginGroup(SU.getInstr()) &&
1681 if (
ExitSU.getInstr() !=
nullptr)
1717 bool IsTopNode =
false;
1722 LLVM_DEBUG(
dbgs() <<
"** ScheduleDAGMILive::schedule picking next node\n");
1731 unsigned SubtreeID =
DFSResult->getSubtreeID(SU);
1749 dbgs() <<
"*** Final schedule for "
1818 if (!
BB->isSuccessor(
BB))
1821 unsigned MaxCyclicLatency = 0;
1824 if (!
P.VRegOrUnit.isVirtualReg())
1826 Register Reg =
P.VRegOrUnit.asVirtualReg();
1837 unsigned LiveOutHeight = DefSU->
getHeight();
1842 SUnit *SU = V2SU.SU;
1854 unsigned CyclicLatency = 0;
1856 CyclicLatency = LiveOutDepth - SU->
getDepth();
1859 if (LiveInHeight > LiveOutHeight) {
1860 if (LiveInHeight - LiveOutHeight < CyclicLatency)
1861 CyclicLatency = LiveInHeight - LiveOutHeight;
1866 << SU->
NodeNum <<
") = " << CyclicLatency <<
"c\n");
1867 if (CyclicLatency > MaxCyclicLatency)
1868 MaxCyclicLatency = CyclicLatency;
1871 LLVM_DEBUG(
dbgs() <<
"Cyclic Critical Path: " << MaxCyclicLatency <<
"c\n");
1872 return MaxCyclicLatency;
1907 SlotIndex SlotIdx =
LIS->getInstructionIndex(*MI).getRegSlot();
1925 if (&*priorII ==
MI)
1942 SlotIndex SlotIdx =
LIS->getInstructionIndex(*MI).getRegSlot();
1977 bool OffsetIsScalable;
1981 : SU(SU), BaseOps(BaseOps),
Offset(
Offset), Width(Width),
1982 OffsetIsScalable(OffsetIsScalable) {}
1986 if (
A->getType() !=
B->getType())
1987 return A->getType() <
B->getType();
1989 return A->getReg() <
B->getReg();
1995 return StackGrowsDown ?
A->getIndex() >
B->getIndex()
1996 :
A->getIndex() <
B->getIndex();
2006 if (std::lexicographical_compare(BaseOps.
begin(), BaseOps.
end(),
2007 RHS.BaseOps.begin(),
RHS.BaseOps.end(),
2010 if (std::lexicographical_compare(
RHS.BaseOps.begin(),
RHS.BaseOps.end(),
2011 BaseOps.
begin(), BaseOps.
end(), Compare))
2019 const TargetInstrInfo *
TII;
2020 const TargetRegisterInfo *
TRI;
2022 bool ReorderWhileClustering;
2025 BaseMemOpClusterMutation(
const TargetInstrInfo *tii,
2026 const TargetRegisterInfo *tri,
bool IsLoad,
2027 bool ReorderWhileClustering)
2028 :
TII(tii),
TRI(tri), IsLoad(IsLoad),
2029 ReorderWhileClustering(ReorderWhileClustering) {}
2031 void apply(ScheduleDAGInstrs *DAGInstrs)
override;
2035 ScheduleDAGInstrs *DAG);
2036 void collectMemOpRecords(std::vector<SUnit> &SUnits,
2037 SmallVectorImpl<MemOpInfo> &MemOpRecords);
2042class StoreClusterMutation :
public BaseMemOpClusterMutation {
2044 StoreClusterMutation(
const TargetInstrInfo *tii,
2045 const TargetRegisterInfo *tri,
2046 bool ReorderWhileClustering)
2047 : BaseMemOpClusterMutation(tii, tri,
false, ReorderWhileClustering) {}
2050class LoadClusterMutation :
public BaseMemOpClusterMutation {
2052 LoadClusterMutation(
const TargetInstrInfo *tii,
const TargetRegisterInfo *tri,
2053 bool ReorderWhileClustering)
2054 : BaseMemOpClusterMutation(tii, tri,
true, ReorderWhileClustering) {}
2059std::unique_ptr<ScheduleDAGMutation>
2062 bool ReorderWhileClustering) {
2064 TII,
TRI, ReorderWhileClustering)
2068std::unique_ptr<ScheduleDAGMutation>
2071 bool ReorderWhileClustering) {
2073 TII,
TRI, ReorderWhileClustering)
2082void BaseMemOpClusterMutation::clusterNeighboringMemOps(
2091 for (
unsigned Idx = 0, End = MemOpRecords.
size(); Idx < (End - 1); ++Idx) {
2093 auto MemOpa = MemOpRecords[Idx];
2096 unsigned NextIdx = Idx + 1;
2097 for (; NextIdx < End; ++NextIdx)
2100 if (!SUnit2ClusterInfo.
count(MemOpRecords[NextIdx].SU->NodeNum) &&
2102 (!DAG->
IsReachable(MemOpRecords[NextIdx].SU, MemOpa.SU) &&
2103 !DAG->
IsReachable(MemOpa.SU, MemOpRecords[NextIdx].SU))))
2108 auto MemOpb = MemOpRecords[NextIdx];
2109 unsigned ClusterLength = 2;
2110 unsigned CurrentClusterBytes = MemOpa.Width.getValue().getKnownMinValue() +
2111 MemOpb.Width.getValue().getKnownMinValue();
2112 auto It = SUnit2ClusterInfo.
find(MemOpa.SU->NodeNum);
2113 if (It != SUnit2ClusterInfo.
end()) {
2114 const auto &[Len, Bytes] = It->second;
2115 ClusterLength = Len + 1;
2116 CurrentClusterBytes = Bytes + MemOpb.Width.getValue().getKnownMinValue();
2119 if (!
TII->shouldClusterMemOps(MemOpa.BaseOps, MemOpa.Offset,
2120 MemOpa.OffsetIsScalable, MemOpb.BaseOps,
2121 MemOpb.Offset, MemOpb.OffsetIsScalable,
2122 ClusterLength, CurrentClusterBytes))
2125 SUnit *SUa = MemOpa.SU;
2126 SUnit *SUb = MemOpb.SU;
2168 SUnit2ClusterInfo[MemOpb.SU->NodeNum] = {ClusterLength,
2169 CurrentClusterBytes};
2172 <<
", Curr cluster bytes: " << CurrentClusterBytes
2183 unsigned ClusterIdx = AllClusters.size();
2185 MemberI->ParentClusterIdx = ClusterIdx;
2188 AllClusters.push_back(Group);
2192void BaseMemOpClusterMutation::collectMemOpRecords(
2194 for (
auto &SU : SUnits) {
2202 bool OffsetIsScalable;
2205 OffsetIsScalable, Width,
TRI)) {
2210 MemOpInfo(&SU, BaseOps,
Offset, OffsetIsScalable, Width));
2213 <<
Offset <<
", OffsetIsScalable: " << OffsetIsScalable
2214 <<
", Width: " << Width <<
"\n");
2217 for (
const auto *
Op : BaseOps)
2223bool BaseMemOpClusterMutation::groupMemOps(
2230 for (
const auto &
MemOp : MemOps) {
2231 unsigned ChainPredID = DAG->
SUnits.size();
2233 for (
const SDep &Pred :
MemOp.SU->Preds) {
2257 collectMemOpRecords(DAG->
SUnits, MemOpRecords);
2259 if (MemOpRecords.
size() < 2)
2266 bool FastCluster = groupMemOps(MemOpRecords, DAG,
Groups);
2268 for (
auto &Group :
Groups) {
2274 clusterNeighboringMemOps(Group.second, FastCluster, DAG);
2289 SlotIndex RegionBeginIdx;
2293 SlotIndex RegionEndIdx;
2296 CopyConstrain(
const TargetInstrInfo *,
const TargetRegisterInfo *) {}
2298 void apply(ScheduleDAGInstrs *DAGInstrs)
override;
2301 void constrainLocalCopy(SUnit *CopySU, ScheduleDAGMILive *DAG);
2306std::unique_ptr<ScheduleDAGMutation>
2309 return std::make_unique<CopyConstrain>(
TII,
TRI);
2352 unsigned LocalReg = SrcReg;
2353 unsigned GlobalReg = DstReg;
2355 if (!LocalLI->
isLocal(RegionBeginIdx, RegionEndIdx)) {
2359 if (!LocalLI->
isLocal(RegionBeginIdx, RegionEndIdx))
2370 if (GlobalSegment == GlobalLI->
end())
2377 if (GlobalSegment->contains(LocalLI->
beginIndex()))
2380 if (GlobalSegment == GlobalLI->
end())
2384 if (GlobalSegment != GlobalLI->
begin()) {
2387 GlobalSegment->start)) {
2398 assert(std::prev(GlobalSegment)->start < LocalLI->beginIndex() &&
2399 "Disconnected LRG within the scheduling region.");
2415 for (
const SDep &Succ : LastLocalSU->
Succs) {
2430 for (
const SDep &Pred : GlobalSU->
Preds) {
2433 if (Pred.
getSUnit() == FirstLocalSU)
2441 for (
SUnit *LU : LocalUses) {
2442 LLVM_DEBUG(
dbgs() <<
" Local use SU(" << LU->NodeNum <<
") -> SU("
2443 << GlobalSU->
NodeNum <<
")\n");
2446 for (
SUnit *GU : GlobalUses) {
2447 LLVM_DEBUG(
dbgs() <<
" Global use SU(" << GU->NodeNum <<
") -> SU("
2448 << FirstLocalSU->
NodeNum <<
")\n");
2460 if (FirstPos == DAG->
end())
2488 unsigned Latency,
bool AfterSchedNode) {
2491 return ResCntFactor >= (int)LFactor;
2493 return ResCntFactor > (int)LFactor;
2506 CheckPending =
false;
2509 MinReadyCycle = std::numeric_limits<unsigned>::max();
2510 ExpectedLatency = 0;
2511 DependentLatency = 0;
2513 MaxExecutedResCount = 0;
2515 IsResourceLimited =
false;
2516 ReservedCycles.clear();
2517 ReservedResourceSegments.clear();
2518 ReservedCyclesIndex.clear();
2519 ResourceGroupSubUnitMasks.clear();
2520#if LLVM_ENABLE_ABI_BREAKING_CHECKS
2524 MaxObservedStall = 0;
2527 ExecutedResCounts.resize(1);
2528 assert(!ExecutedResCounts[0] &&
"nonzero count for bad resource");
2544 unsigned PIdx = PI->ProcResourceIdx;
2546 assert(PI->ReleaseAtCycle >= PI->AcquireAtCycle);
2548 (Factor * (PI->ReleaseAtCycle - PI->AcquireAtCycle));
2560 unsigned ResourceCount =
SchedModel->getNumProcResourceKinds();
2561 ReservedCyclesIndex.resize(ResourceCount);
2562 ExecutedResCounts.resize(ResourceCount);
2563 ResourceGroupSubUnitMasks.resize(ResourceCount,
APInt(ResourceCount, 0));
2564 unsigned NumUnits = 0;
2566 for (
unsigned i = 0; i < ResourceCount; ++i) {
2567 ReservedCyclesIndex[i] = NumUnits;
2568 NumUnits +=
SchedModel->getProcResource(i)->NumUnits;
2570 auto SubUnits =
SchedModel->getProcResource(i)->SubUnitsIdxBegin;
2571 for (
unsigned U = 0, UE =
SchedModel->getProcResource(i)->NumUnits;
2573 ResourceGroupSubUnitMasks[i].setBit(SubUnits[U]);
2593 if (ReadyCycle > CurrCycle)
2594 return ReadyCycle - CurrCycle;
2601 unsigned ReleaseAtCycle,
2602 unsigned AcquireAtCycle) {
2605 return ReservedResourceSegments[InstanceIdx].getFirstAvailableAtFromTop(
2606 CurrCycle, AcquireAtCycle, ReleaseAtCycle);
2608 return ReservedResourceSegments[InstanceIdx].getFirstAvailableAtFromBottom(
2609 CurrCycle, AcquireAtCycle, ReleaseAtCycle);
2612 unsigned NextUnreserved = ReservedCycles[InstanceIdx];
2618 NextUnreserved = std::max(CurrCycle, NextUnreserved + ReleaseAtCycle);
2619 return NextUnreserved;
2625std::pair<unsigned, unsigned>
2627 unsigned ReleaseAtCycle,
2628 unsigned AcquireAtCycle) {
2630 LLVM_DEBUG(
dbgs() <<
" Resource booking (@" << CurrCycle <<
"c): \n");
2632 LLVM_DEBUG(
dbgs() <<
" getNextResourceCycle (@" << CurrCycle <<
"c): \n");
2635 unsigned InstanceIdx = 0;
2636 unsigned StartIndex = ReservedCyclesIndex[PIdx];
2637 unsigned NumberOfInstances =
SchedModel->getProcResource(PIdx)->NumUnits;
2638 assert(NumberOfInstances > 0 &&
2639 "Cannot have zero instances of a ProcResource");
2656 if (ResourceGroupSubUnitMasks[PIdx][PE.ProcResourceIdx])
2658 StartIndex, ReleaseAtCycle, AcquireAtCycle),
2661 auto SubUnits =
SchedModel->getProcResource(PIdx)->SubUnitsIdxBegin;
2662 for (
unsigned I = 0, End = NumberOfInstances;
I < End; ++
I) {
2663 unsigned NextUnreserved, NextInstanceIdx;
2664 std::tie(NextUnreserved, NextInstanceIdx) =
2666 if (MinNextUnreserved > NextUnreserved) {
2667 InstanceIdx = NextInstanceIdx;
2668 MinNextUnreserved = NextUnreserved;
2671 return std::make_pair(MinNextUnreserved, InstanceIdx);
2674 for (
unsigned I = StartIndex, End = StartIndex + NumberOfInstances;
I < End;
2676 unsigned NextUnreserved =
2680 << NextUnreserved <<
"c\n");
2681 if (MinNextUnreserved > NextUnreserved) {
2683 MinNextUnreserved = NextUnreserved;
2688 <<
"[" << InstanceIdx - StartIndex <<
"]"
2689 <<
" available @" << MinNextUnreserved <<
"c"
2691 return std::make_pair(MinNextUnreserved, InstanceIdx);
2711 <<
"hazard: SU(" << SU->
NodeNum <<
") reported by HazardRec\n");
2716 if ((CurrMOps > 0) && (CurrMOps + uops >
SchedModel->getIssueWidth())) {
2718 << uops <<
", CurrMOps = " << CurrMOps <<
", "
2719 <<
"CurrMOps + uops > issue width of "
2728 << (
isTop() ?
"begin" :
"end") <<
" group\n");
2737 unsigned ResIdx = PE.ProcResourceIdx;
2738 unsigned ReleaseAtCycle = PE.ReleaseAtCycle;
2739 unsigned AcquireAtCycle = PE.AcquireAtCycle;
2740 unsigned NRCycle, InstanceIdx;
2741 std::tie(NRCycle, InstanceIdx) =
2743 if (NRCycle > CurrCycle) {
2744#if LLVM_ENABLE_ABI_BREAKING_CHECKS
2745 MaxObservedStall = std::max(ReleaseAtCycle, MaxObservedStall);
2748 <<
"hazard: SU(" << SU->
NodeNum <<
") "
2749 <<
SchedModel->getResourceName(ResIdx) <<
'['
2750 << InstanceIdx - ReservedCyclesIndex[ResIdx] <<
']' <<
"="
2751 << NRCycle <<
"c, is later than "
2752 <<
"CurrCycle = " << CurrCycle <<
"c\n");
2763 SUnit *LateSU =
nullptr;
2764 unsigned RemLatency = 0;
2765 for (
SUnit *SU : ReadySUs) {
2767 if (L > RemLatency) {
2774 << LateSU->
NodeNum <<
") " << RemLatency <<
"c\n");
2788 unsigned OtherCritCount =
Rem->RemIssueCount
2789 + (RetiredMOps *
SchedModel->getMicroOpFactor());
2791 << OtherCritCount /
SchedModel->getMicroOpFactor() <<
'\n');
2792 for (
unsigned PIdx = 1, PEnd =
SchedModel->getNumProcResourceKinds();
2793 PIdx != PEnd; ++PIdx) {
2795 if (OtherCount > OtherCritCount) {
2796 OtherCritCount = OtherCount;
2797 OtherCritIdx = PIdx;
2802 dbgs() <<
" " <<
Available.getName() <<
" + Remain CritRes: "
2803 << OtherCritCount /
SchedModel->getResourceFactor(OtherCritIdx)
2804 <<
" " <<
SchedModel->getResourceName(OtherCritIdx) <<
"\n");
2806 return OtherCritCount;
2813#if LLVM_ENABLE_ABI_BREAKING_CHECKS
2817 if (ReadyCycle > CurrCycle)
2818 MaxObservedStall = std::max(ReadyCycle - CurrCycle, MaxObservedStall);
2821 if (ReadyCycle < MinReadyCycle)
2822 MinReadyCycle = ReadyCycle;
2826 bool IsBuffered =
SchedModel->getMicroOpBufferSize() != 0;
2827 bool HazardDetected = !IsBuffered && ReadyCycle > CurrCycle;
2830 <<
") ReadyCycle = " << ReadyCycle
2831 <<
" is later than CurrCycle = " << CurrCycle
2832 <<
" on an unbuffered resource" <<
"\n");
2837 HazardDetected =
true;
2842 if (!HazardDetected) {
2845 <<
"Move SU(" << SU->
NodeNum <<
") into Available Q\n");
2858 if (
SchedModel->getMicroOpBufferSize() == 0) {
2859 assert(MinReadyCycle < std::numeric_limits<unsigned>::max() &&
2860 "MinReadyCycle uninitialized");
2861 if (MinReadyCycle > NextCycle)
2862 NextCycle = MinReadyCycle;
2865 unsigned DecMOps =
SchedModel->getIssueWidth() * (NextCycle - CurrCycle);
2866 CurrMOps = (CurrMOps <= DecMOps) ? 0 : CurrMOps - DecMOps;
2869 if ((NextCycle - CurrCycle) > DependentLatency)
2870 DependentLatency = 0;
2872 DependentLatency -= (NextCycle - CurrCycle);
2876 CurrCycle = NextCycle;
2879 for (; CurrCycle != NextCycle; ++CurrCycle) {
2886 CheckPending =
true;
2896 ExecutedResCounts[PIdx] +=
Count;
2897 if (ExecutedResCounts[PIdx] > MaxExecutedResCount)
2898 MaxExecutedResCount = ExecutedResCounts[PIdx];
2912 unsigned ReleaseAtCycle,
2914 unsigned AcquireAtCycle) {
2915 unsigned Factor =
SchedModel->getResourceFactor(PIdx);
2916 unsigned Count = Factor * (ReleaseAtCycle- AcquireAtCycle);
2918 << ReleaseAtCycle <<
"x" << Factor <<
"u\n");
2922 assert(
Rem->RemainingCounts[PIdx] >=
Count &&
"resource double counted");
2923 Rem->RemainingCounts[PIdx] -=
Count;
2928 ZoneCritResIdx = PIdx;
2935 unsigned NextAvailable, InstanceIdx;
2936 std::tie(NextAvailable, InstanceIdx) =
2938 if (NextAvailable > CurrCycle) {
2941 <<
'[' << InstanceIdx - ReservedCyclesIndex[PIdx] <<
']'
2942 <<
" reserved until @" << NextAvailable <<
"\n");
2944 return NextAvailable;
2958 CheckPending =
true;
2965 (CurrMOps == 0 || (CurrMOps + IncMOps) <=
SchedModel->getIssueWidth()) &&
2966 "Cannot schedule this instruction's MicroOps in the current cycle.");
2971 unsigned NextCycle = CurrCycle;
2972 switch (
SchedModel->getMicroOpBufferSize()) {
2974 assert(ReadyCycle <= CurrCycle &&
"Broken PendingQueue");
2977 if (ReadyCycle > NextCycle) {
2978 NextCycle = ReadyCycle;
2979 LLVM_DEBUG(
dbgs() <<
" *** Stall until: " << ReadyCycle <<
"\n");
2988 NextCycle = ReadyCycle;
2991 RetiredMOps += IncMOps;
2995 unsigned DecRemIssue = IncMOps *
SchedModel->getMicroOpFactor();
2996 assert(
Rem->RemIssueCount >= DecRemIssue &&
"MOps double counted");
2997 Rem->RemIssueCount -= DecRemIssue;
2998 if (ZoneCritResIdx) {
3000 unsigned ScaledMOps =
3001 RetiredMOps *
SchedModel->getMicroOpFactor();
3009 << ScaledMOps /
SchedModel->getLatencyFactor()
3015 PE =
SchedModel->getWriteProcResEnd(SC); PI != PE; ++PI) {
3017 countResource(SC, PI->ProcResourceIdx, PI->ReleaseAtCycle, NextCycle,
3018 PI->AcquireAtCycle);
3019 if (RCycle > NextCycle)
3029 PE =
SchedModel->getWriteProcResEnd(SC); PI != PE; ++PI) {
3030 unsigned PIdx = PI->ProcResourceIdx;
3031 if (
SchedModel->getProcResource(PIdx)->BufferSize == 0) {
3034 unsigned ReservedUntil, InstanceIdx;
3036 SC, PIdx, PI->ReleaseAtCycle, PI->AcquireAtCycle);
3038 ReservedResourceSegments[InstanceIdx].add(
3040 NextCycle, PI->AcquireAtCycle, PI->ReleaseAtCycle),
3043 ReservedResourceSegments[InstanceIdx].add(
3045 NextCycle, PI->AcquireAtCycle, PI->ReleaseAtCycle),
3050 unsigned ReservedUntil, InstanceIdx;
3052 SC, PIdx, PI->ReleaseAtCycle, PI->AcquireAtCycle);
3054 ReservedCycles[InstanceIdx] =
3055 std::max(ReservedUntil, NextCycle + PI->ReleaseAtCycle);
3057 ReservedCycles[InstanceIdx] = NextCycle;
3064 unsigned &TopLatency =
isTop() ? ExpectedLatency : DependentLatency;
3065 unsigned &BotLatency =
isTop() ? DependentLatency : ExpectedLatency;
3069 << SU->
NodeNum <<
") " << TopLatency <<
"c\n");
3074 << SU->
NodeNum <<
") " << BotLatency <<
"c\n");
3077 if (NextCycle > CurrCycle)
3090 CurrMOps += IncMOps;
3103 while (CurrMOps >=
SchedModel->getIssueWidth()) {
3104 LLVM_DEBUG(
dbgs() <<
" *** Max MOps " << CurrMOps <<
" at cycle "
3105 << CurrCycle <<
'\n');
3116 MinReadyCycle = std::numeric_limits<unsigned>::max();
3120 for (
unsigned I = 0, E =
Pending.size();
I < E; ++
I) {
3126 if (ReadyCycle < MinReadyCycle)
3127 MinReadyCycle = ReadyCycle;
3138 CheckPending =
false;
3167 for (
unsigned i = 0;
Available.empty(); ++i) {
3184#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
3193 unsigned ResourceCount =
SchedModel->getNumProcResourceKinds();
3194 unsigned StartIdx = 0;
3196 for (
unsigned ResIdx = 0; ResIdx < ResourceCount; ++ResIdx) {
3197 const unsigned NumUnits =
SchedModel->getProcResource(ResIdx)->NumUnits;
3198 std::string ResName =
SchedModel->getResourceName(ResIdx);
3199 for (
unsigned UnitIdx = 0; UnitIdx < NumUnits; ++UnitIdx) {
3200 dbgs() << ResName <<
"(" << UnitIdx <<
") = ";
3202 if (ReservedResourceSegments.count(StartIdx + UnitIdx))
3203 dbgs() << ReservedResourceSegments.at(StartIdx + UnitIdx);
3207 dbgs() << ReservedCycles[StartIdx + UnitIdx] <<
"\n";
3209 StartIdx += NumUnits;
3218 if (ZoneCritResIdx) {
3219 ResFactor =
SchedModel->getResourceFactor(ZoneCritResIdx);
3223 ResCount = RetiredMOps * ResFactor;
3225 unsigned LFactor =
SchedModel->getLatencyFactor();
3227 <<
" Retired: " << RetiredMOps;
3229 dbgs() <<
"\n Critical: " << ResCount / LFactor <<
"c, "
3230 << ResCount / ResFactor <<
" "
3231 <<
SchedModel->getResourceName(ZoneCritResIdx)
3232 <<
"\n ExpectedLatency: " << ExpectedLatency <<
"c\n"
3233 << (IsResourceLimited ?
" - Resource" :
" - Latency")
3253 PE =
SchedModel->getWriteProcResEnd(SC); PI != PE; ++PI) {
3254 if (PI->ProcResourceIdx ==
Policy.ReduceResIdx)
3255 ResDelta.CritResources += PI->ReleaseAtCycle;
3256 if (PI->ProcResourceIdx ==
Policy.DemandResIdx)
3257 ResDelta.DemandedResources += PI->ReleaseAtCycle;
3279 RemLatency = std::max(RemLatency,
3281 RemLatency = std::max(RemLatency,
3288bool GenericSchedulerBase::shouldReduceLatency(
const CandPolicy &Policy,
3289 SchedBoundary &CurrZone,
3290 bool ComputeRemLatency,
3291 unsigned &RemLatency)
const {
3301 if (ComputeRemLatency)
3317 unsigned OtherCritIdx = 0;
3318 unsigned OtherCount =
3321 bool OtherResLimited =
false;
3322 unsigned RemLatency = 0;
3323 bool RemLatencyComputed =
false;
3324 if (
SchedModel->hasInstrSchedModel() && OtherCount != 0) {
3326 RemLatencyComputed =
true;
3328 OtherCount, RemLatency,
false);
3334 if (!OtherResLimited &&
3335 (IsPostRA || shouldReduceLatency(Policy, CurrZone, !RemLatencyComputed,
3339 <<
" RemainingLatency " << RemLatency <<
" + "
3341 <<
Rem.CriticalPath <<
"\n");
3348 dbgs() <<
" " << CurrZone.Available.getName() <<
" ResourceLimited: "
3349 << SchedModel->getResourceName(CurrZone.getZoneCritResIdx()) <<
"\n";
3350 }
if (OtherResLimited)
dbgs()
3351 <<
" RemainingLimit: "
3352 <<
SchedModel->getResourceName(OtherCritIdx) <<
"\n";
3354 <<
" Latency limited both directions.\n");
3359 if (OtherResLimited)
3368 case NoCand:
return "NOCAND ";
3369 case Only1:
return "ONLY1 ";
3370 case PhysReg:
return "PHYS-REG ";
3373 case Stall:
return "STALL ";
3374 case Cluster:
return "CLUSTER ";
3375 case Weak:
return "WEAK ";
3376 case RegMax:
return "REG-MAX ";
3392 unsigned ResIdx = 0;
3427 dbgs() <<
" " <<
TRI->getRegPressureSetName(
P.getPSet())
3428 <<
":" <<
P.getUnitInc() <<
" ";
3432 dbgs() <<
" " <<
SchedModel->getProcResource(ResIdx)->Name <<
" ";
3450 if (TryVal < CandVal) {
3454 if (TryVal > CandVal) {
3455 if (Cand.
Reason > Reason)
3466 if (TryVal > CandVal) {
3470 if (TryVal < CandVal) {
3471 if (Cand.
Reason > Reason)
3512 bool IsPostRA =
false) {
3515 << (IsPostRA ?
"post-RA" :
"pre-RA") <<
"]\n");
3534 NumRegExcessPostRA++;
3537 NumRegCriticalPostRA++;
3552 NumResourceReducePostRA++;
3555 NumResourceDemandPostRA++;
3558 NumTopDepthReducePostRA++;
3561 NumTopPathReducePostRA++;
3564 NumBotHeightReducePostRA++;
3567 NumBotPathReducePostRA++;
3570 NumNodeOrderPostRA++;
3573 NumFirstValidPostRA++;
3593 NumRegExcessPreRA++;
3596 NumRegCriticalPreRA++;
3611 NumResourceReducePreRA++;
3614 NumResourceDemandPreRA++;
3617 NumTopDepthReducePreRA++;
3620 NumTopPathReducePreRA++;
3623 NumBotHeightReducePreRA++;
3626 NumBotPathReducePreRA++;
3629 NumNodeOrderPreRA++;
3632 NumFirstValidPreRA++;
3640 bool IsPostRA =
false) {
3646 "(PreRA)GenericScheduler needs vreg liveness");
3652 DAG->computeDFSResult();
3663 if (!
Top.HazardRec) {
3664 Top.HazardRec =
DAG->TII->CreateTargetMIHazardRecognizer(Itin,
DAG);
3666 if (!
Bot.HazardRec) {
3667 Bot.HazardRec =
DAG->TII->CreateTargetMIHazardRecognizer(Itin,
DAG);
3688 for (
unsigned VT = MVT::i64; VT > (
unsigned)MVT::i1; --VT) {
3691 unsigned NIntRegs =
Context->RegClassInfo->getNumAllocatableRegs(
3729#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
3730 dbgs() <<
"GenericScheduler RegionPolicy: "
3731 <<
" ShouldTrackPressure=" <<
RegionPolicy.ShouldTrackPressure
3748 if (
Rem.CyclicCritPath == 0 ||
Rem.CyclicCritPath >=
Rem.CriticalPath)
3752 unsigned IterCount =
3753 std::max(
Rem.CyclicCritPath *
SchedModel->getLatencyFactor(),
3756 unsigned AcyclicCount =
Rem.CriticalPath *
SchedModel->getLatencyFactor();
3758 unsigned InFlightCount =
3759 (AcyclicCount *
Rem.RemIssueCount + IterCount-1) / IterCount;
3760 unsigned BufferLimit =
3763 Rem.IsAcyclicLatencyLimited = InFlightCount > BufferLimit;
3766 dbgs() <<
"IssueCycles="
3767 <<
Rem.RemIssueCount /
SchedModel->getLatencyFactor() <<
"c "
3768 <<
"IterCycles=" << IterCount /
SchedModel->getLatencyFactor()
3769 <<
"c NumIters=" << (AcyclicCount + IterCount - 1) / IterCount
3770 <<
" InFlight=" << InFlightCount /
SchedModel->getMicroOpFactor()
3771 <<
"m BufferLim=" <<
SchedModel->getMicroOpBufferSize() <<
"m\n";
3772 if (
Rem.IsAcyclicLatencyLimited)
dbgs() <<
" ACYCLIC LATENCY LIMIT\n");
3776 Rem.CriticalPath =
DAG->ExitSU.getDepth();
3779 for (
const SUnit *SU :
Bot.Available) {
3785 errs() <<
"Critical Path(GS-RR ): " <<
Rem.CriticalPath <<
" \n";
3789 Rem.CyclicCritPath =
DAG->computeCyclicCriticalPath();
3815 if (TryPSet == CandPSet) {
3820 int TryRank = TryP.
isValid() ?
TRI->getRegPressureSetScore(MF, TryPSet) :
3821 std::numeric_limits<int>::max();
3823 int CandRank = CandP.
isValid() ?
TRI->getRegPressureSetScore(MF, CandPSet) :
3824 std::numeric_limits<int>::max();
3829 return tryGreater(TryRank, CandRank, TryCand, Cand, Reason);
3847 unsigned ScheduledOper = isTop ? 1 : 0;
3848 unsigned UnscheduledOper = isTop ? 0 : 1;
3851 if (
MI->getOperand(ScheduledOper).getReg().isPhysical())
3856 if (
MI->getOperand(UnscheduledOper).getReg().isPhysical())
3857 return AtBoundary ? -1 : 1;
3860 if (
MI->isMoveImmediate()) {
3866 if (
Op.isReg() && !
Op.getReg().isPhysical()) {
3873 return isTop ? -1 : 1;
3885 if (
DAG->isTrackingPressure()) {
3890 DAG->getRegionCriticalPSets(),
3891 DAG->getRegPressure().MaxSetPressure);
3896 &
DAG->getPressureDiff(Cand.
SU),
3898 DAG->getRegionCriticalPSets(),
3899 DAG->getRegPressure().MaxSetPressure);
3903 DAG->getPressureDiff(Cand.
SU),
3905 DAG->getRegionCriticalPSets(),
3906 DAG->getRegPressure().MaxSetPressure);
3911 <<
" Try SU(" << Cand.
SU->
NodeNum <<
") "
3960 bool SameBoundary = Zone !=
nullptr;
3983 bool CandIsClusterSucc =
3985 bool TryCandIsClusterSucc =
3988 if (
tryGreater(TryCandIsClusterSucc, CandIsClusterSucc, TryCand, Cand,
3996 TryCand, Cand,
Weak))
4021 !
Rem.IsAcyclicLatencyLimited &&
tryLatency(TryCand, Cand, *Zone))
4048 for (
SUnit *SU : Q) {
4068 if (
SUnit *SU =
Bot.pickOnlyChoice()) {
4073 if (
SUnit *SU =
Top.pickOnlyChoice()) {
4090 BotCand.Policy != BotPolicy) {
4102 "Last pick result should correspond to re-picking right now");
4110 TopCand.Policy != TopPolicy) {
4122 "Last pick result should correspond to re-picking right now");
4137 IsTopNode = Cand.
AtTop;
4144 if (
DAG->top() ==
DAG->bottom()) {
4146 Bot.Available.empty() &&
Bot.Pending.empty() &&
"ReadyQ garbage");
4151 SU =
Top.pickOnlyChoice();
4162 SU =
Bot.pickOnlyChoice();
4193 Top.removeReady(SU);
4195 Bot.removeReady(SU);
4202 ++NumInstrsInSourceOrderPreRA;
4206 ++NumInstrsInSourceOrderPreRA;
4209 NumInstrsScheduledPreRA += 1;
4222 for (
SDep &Dep : Deps) {
4223 if (Dep.getKind() !=
SDep::Data || !Dep.getReg().isPhysical())
4225 SUnit *DepSU = Dep.getSUnit();
4226 if (isTop ? DepSU->
Succs.size() > 1 : DepSU->
Preds.size() > 1)
4229 if (!Copy->isCopy() && !Copy->isMoveImmediate())
4232 DAG->dumpNode(*Dep.getSUnit()));
4233 DAG->moveInstruction(Copy, InsertPos);
4251 dbgs() <<
" Top Cluster: ";
4252 for (
auto *
N : *TopCluster)
4253 dbgs() <<
N->NodeNum <<
'\t';
4266 dbgs() <<
" Bot Cluster: ";
4267 for (
auto *
N : *BotCluster)
4268 dbgs() <<
N->NodeNum <<
'\t';
4282static MachineSchedRegistry
4302 if (!
Top.HazardRec) {
4303 Top.HazardRec =
DAG->TII->CreateTargetMIHazardRecognizer(Itin,
DAG);
4305 if (!
Bot.HazardRec) {
4306 Bot.HazardRec =
DAG->TII->CreateTargetMIHazardRecognizer(Itin,
DAG);
4343 Rem.CriticalPath =
DAG->ExitSU.getDepth();
4346 for (
const SUnit *SU :
Bot.Available) {
4352 errs() <<
"Critical Path(PGS-RR ): " <<
Rem.CriticalPath <<
" \n";
4371 Top.getLatencyStallCycles(Cand.
SU), TryCand, Cand,
Stall))
4377 bool CandIsClusterSucc =
4379 bool TryCandIsClusterSucc =
4382 if (
tryGreater(TryCandIsClusterSucc, CandIsClusterSucc, TryCand, Cand,
4415 for (
SUnit *SU : Q) {
4434 if (
SUnit *SU =
Bot.pickOnlyChoice()) {
4439 if (
SUnit *SU =
Top.pickOnlyChoice()) {
4456 BotCand.Policy != BotPolicy) {
4468 "Last pick result should correspond to re-picking right now");
4476 TopCand.Policy != TopPolicy) {
4488 "Last pick result should correspond to re-picking right now");
4503 IsTopNode = Cand.
AtTop;
4510 if (
DAG->top() ==
DAG->bottom()) {
4512 Bot.Available.empty() &&
Bot.Pending.empty() &&
"ReadyQ garbage");
4517 SU =
Bot.pickOnlyChoice();
4533 SU =
Top.pickOnlyChoice();
4554 Top.removeReady(SU);
4556 Bot.removeReady(SU);
4563 ++NumInstrsInSourceOrderPostRA;
4567 ++NumInstrsInSourceOrderPostRA;
4570 NumInstrsScheduledPostRA += 1;
4598 const BitVector *ScheduledTrees =
nullptr;
4601 ILPOrder(
bool MaxILP) : MaximizeILP(MaxILP) {}
4606 bool operator()(
const SUnit *
A,
const SUnit *
B)
const {
4609 if (SchedTreeA != SchedTreeB) {
4611 if (ScheduledTrees->
test(SchedTreeA) != ScheduledTrees->
test(SchedTreeB))
4612 return ScheduledTrees->
test(SchedTreeB);
4629class ILPScheduler :
public MachineSchedStrategy {
4630 ScheduleDAGMILive *DAG =
nullptr;
4633 std::vector<SUnit*> ReadyQ;
4636 ILPScheduler(
bool MaximizeILP) :
Cmp(MaximizeILP) {}
4638 void initialize(ScheduleDAGMI *dag)
override {
4640 DAG =
static_cast<ScheduleDAGMILive*
>(dag);
4647 void registerRoots()
override {
4649 std::make_heap(ReadyQ.begin(), ReadyQ.end(), Cmp);
4656 SUnit *pickNode(
bool &IsTopNode)
override {
4657 if (ReadyQ.empty())
return nullptr;
4658 std::pop_heap(ReadyQ.begin(), ReadyQ.end(), Cmp);
4659 SUnit *SU = ReadyQ.back();
4663 <<
"SU(" << SU->
NodeNum <<
") "
4670 <<
"Scheduling " << *SU->
getInstr());
4675 void scheduleTree(
unsigned SubtreeID)
override {
4676 std::make_heap(ReadyQ.begin(), ReadyQ.end(), Cmp);
4681 void schedNode(SUnit *SU,
bool IsTopNode)
override {
4682 assert(!IsTopNode &&
"SchedDFSResult needs bottom-up");
4685 void releaseTopNode(SUnit *)
override { }
4687 void releaseBottomNode(SUnit *SU)
override {
4688 ReadyQ.push_back(SU);
4689 std::push_heap(ReadyQ.begin(), ReadyQ.end(), Cmp);
4716template<
bool IsReverse>
4720 return A->NodeNum >
B->NodeNum;
4722 return A->NodeNum <
B->NodeNum;
4727class InstructionShuffler :
public MachineSchedStrategy {
4734 PriorityQueue<SUnit*, std::vector<SUnit*>, SUnitOrder<false>>
4738 PriorityQueue<SUnit*, std::vector<SUnit*>, SUnitOrder<true>>
4742 InstructionShuffler(
bool alternate,
bool topdown)
4743 : IsAlternating(alternate), IsTopDown(topdown) {}
4753 SUnit *pickNode(
bool &IsTopNode)
override {
4757 if (TopQ.empty())
return nullptr;
4764 if (BottomQ.empty())
return nullptr;
4771 IsTopDown = !IsTopDown;
4775 void schedNode(SUnit *SU,
bool IsTopNode)
override {}
4777 void releaseTopNode(SUnit *SU)
override {
4780 void releaseBottomNode(SUnit *SU)
override {
4792 C, std::make_unique<InstructionShuffler>(Alternate, TopDown));
4796 "shuffle",
"Shuffle machine instructions alternating directions",
4815 return std::string(
G->MF.getName());
4835 return "color=cyan,style=dashed";
4837 return "color=blue,style=dashed";
4854 return G->getGraphNodeLabel(SU);
4858 std::string Str(
"shape=Mrecord");
4863 Str +=
",style=filled,fillcolor=\"#";
4879 errs() <<
"ScheduleDAGMI::viewGraph is only available in debug builds on "
4880 <<
"systems with Graphviz or gv!\n";
4895 return A.first <
B.first;
4898unsigned ResourceSegments::getFirstAvailableAt(
4899 unsigned CurrCycle,
unsigned AcquireAtCycle,
unsigned ReleaseAtCycle,
4901 IntervalBuilder)
const {
4903 "Cannot execute on an un-sorted set of intervals.");
4907 if (AcquireAtCycle == ReleaseAtCycle)
4910 unsigned RetCycle = CurrCycle;
4912 IntervalBuilder(RetCycle, AcquireAtCycle, ReleaseAtCycle);
4913 for (
auto &
Interval : _Intervals) {
4920 "Invalid intervals configuration.");
4921 RetCycle += (unsigned)
Interval.second - (
unsigned)NewInterval.first;
4922 NewInterval = IntervalBuilder(RetCycle, AcquireAtCycle, ReleaseAtCycle);
4928 const unsigned CutOff) {
4929 assert(
A.first <=
A.second &&
"Cannot add negative resource usage");
4930 assert(CutOff > 0 &&
"0-size interval history has no use.");
4936 if (
A.first ==
A.second)
4943 "A resource is being overwritten");
4944 _Intervals.push_back(
A);
4950 while (_Intervals.size() > CutOff)
4951 _Intervals.pop_front();
4956 assert(
A.first <=
A.second &&
"Invalid interval");
4957 assert(
B.first <=
B.second &&
"Invalid interval");
4960 if ((
A.first ==
B.first) || (
A.second ==
B.second))
4965 if ((
A.first >
B.first) && (
A.second <
B.second))
4970 if ((
A.first >
B.first) && (
A.first <
B.second) && (
A.second >
B.second))
4975 if ((
A.first <
B.first) && (
B.first <
A.second) && (
B.second >
B.first))
4981void ResourceSegments::sortAndMerge() {
4982 if (_Intervals.size() <= 1)
4989 auto next = std::next(std::begin(_Intervals));
4990 auto E = std::end(_Intervals);
4991 for (; next != E; ++next) {
4992 if (std::prev(next)->second >= next->first) {
4993 next->first = std::prev(next)->first;
4994 _Intervals.erase(std::prev(next));
MachineInstrBuilder MachineInstrBuilder & DefMI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
Function Alias Analysis false
static const Function * getParent(const Value *V)
This file implements the BitVector class.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
#define clEnumValN(ENUMVAL, FLAGNAME, DESC)
#define LLVM_DUMP_METHOD
Mark debug helper function definitions like dump() that should not be stripped from debug builds.
static std::optional< ArrayRef< InsnRange >::iterator > intersects(const MachineInstr *StartMI, const MachineInstr *EndMI, ArrayRef< InsnRange > Ranges, const InstructionOrdering &Ordering)
Check if the instruction range [StartMI, EndMI] intersects any instruction range in Ranges.
This file defines the DenseMap class.
Generic implementation of equivalence classes through the use Tarjan's efficient union-find algorithm...
const HexagonInstrInfo * TII
A common definition of LaneBitmask for use in TableGen and CodeGen.
static cl::opt< MISched::Direction > PostRADirection("misched-postra-direction", cl::Hidden, cl::desc("Post reg-alloc list scheduling direction"), cl::init(MISched::Unspecified), cl::values(clEnumValN(MISched::TopDown, "topdown", "Force top-down post reg-alloc list scheduling"), clEnumValN(MISched::BottomUp, "bottomup", "Force bottom-up post reg-alloc list scheduling"), clEnumValN(MISched::Bidirectional, "bidirectional", "Force bidirectional post reg-alloc list scheduling")))
static bool isSchedBoundary(MachineBasicBlock::iterator MI, MachineBasicBlock *MBB, MachineFunction *MF, const TargetInstrInfo *TII)
Return true of the given instruction should not be included in a scheduling region.
static MachineSchedRegistry ILPMaxRegistry("ilpmax", "Schedule bottom-up for max ILP", createILPMaxScheduler)
static cl::opt< bool > EnableMemOpCluster("misched-cluster", cl::Hidden, cl::desc("Enable memop clustering."), cl::init(true))
PostRA Machine Instruction Scheduler
static MachineBasicBlock::const_iterator nextIfDebug(MachineBasicBlock::const_iterator I, MachineBasicBlock::const_iterator End)
If this iterator is a debug value, increment until reaching the End or a non-debug instruction.
static const unsigned MinSubtreeSize
static const unsigned InvalidCycle
static cl::opt< bool > MISchedSortResourcesInTrace("misched-sort-resources-in-trace", cl::Hidden, cl::init(true), cl::desc("Sort the resources printed in the dump trace"))
static cl::opt< bool > EnableCyclicPath("misched-cyclicpath", cl::Hidden, cl::desc("Enable cyclic critical path analysis."), cl::init(true))
static MachineBasicBlock::const_iterator priorNonDebug(MachineBasicBlock::const_iterator I, MachineBasicBlock::const_iterator Beg)
Decrement this iterator until reaching the top or a non-debug instr.
static cl::opt< MachineSchedRegistry::ScheduleDAGCtor, false, RegisterPassParser< MachineSchedRegistry > > MachineSchedOpt("misched", cl::init(&useDefaultMachineSched), cl::Hidden, cl::desc("Machine instruction scheduler to use"))
MachineSchedOpt allows command line selection of the scheduler.
static cl::opt< bool > EnableMachineSched("enable-misched", cl::desc("Enable the machine instruction scheduling pass."), cl::init(true), cl::Hidden)
static unsigned computeRemLatency(SchedBoundary &CurrZone)
Compute remaining latency.
static cl::opt< unsigned > MISchedCutoff("misched-cutoff", cl::Hidden, cl::desc("Stop scheduling after N instructions"), cl::init(~0U))
static cl::opt< unsigned > SchedOnlyBlock("misched-only-block", cl::Hidden, cl::desc("Only schedule this MBB#"))
static cl::opt< bool > EnableRegPressure("misched-regpressure", cl::Hidden, cl::desc("Enable register pressure scheduling."), cl::init(true))
static void tracePick(GenericSchedulerBase::CandReason Reason, bool IsTop, bool IsPostRA=false)
static MachineSchedRegistry GenericSchedRegistry("converge", "Standard converging scheduler.", createConvergingSched)
static cl::opt< unsigned > HeaderColWidth("misched-dump-schedule-trace-col-header-width", cl::Hidden, cl::desc("Set width of the columns with " "the resources and schedule units"), cl::init(19))
static cl::opt< bool > ForceFastCluster("force-fast-cluster", cl::Hidden, cl::desc("Switch to fast cluster algorithm with the lost " "of some fusion opportunities"), cl::init(false))
static cl::opt< unsigned > FastClusterThreshold("fast-cluster-threshold", cl::Hidden, cl::desc("The threshold for fast cluster"), cl::init(1000))
static bool checkResourceLimit(unsigned LFactor, unsigned Count, unsigned Latency, bool AfterSchedNode)
Given a Count of resource usage and a Latency value, return true if a SchedBoundary becomes resource ...
static ScheduleDAGInstrs * createInstructionShuffler(MachineSchedContext *C)
static ScheduleDAGInstrs * useDefaultMachineSched(MachineSchedContext *C)
A dummy default scheduler factory indicates whether the scheduler is overridden on the command line.
static bool sortIntervals(const ResourceSegments::IntervalTy &A, const ResourceSegments::IntervalTy &B)
Sort predicate for the intervals stored in an instance of ResourceSegments.
static cl::opt< unsigned > ColWidth("misched-dump-schedule-trace-col-width", cl::Hidden, cl::desc("Set width of the columns showing resource booking."), cl::init(5))
static MachineSchedRegistry DefaultSchedRegistry("default", "Use the target's default scheduler choice.", useDefaultMachineSched)
static cl::opt< std::string > SchedOnlyFunc("misched-only-func", cl::Hidden, cl::desc("Only schedule this function"))
static const char * scheduleTableLegend
static ScheduleDAGInstrs * createConvergingSched(MachineSchedContext *C)
static cl::opt< bool > MischedDetailResourceBooking("misched-detail-resource-booking", cl::Hidden, cl::init(false), cl::desc("Show details of invoking getNextResoufceCycle."))
static cl::opt< unsigned > ViewMISchedCutoff("view-misched-cutoff", cl::Hidden, cl::desc("Hide nodes with more predecessor/successor than cutoff"))
In some situations a few uninteresting nodes depend on nearly all other nodes in the graph,...
static MachineSchedRegistry ShufflerRegistry("shuffle", "Shuffle machine instructions alternating directions", createInstructionShuffler)
static cl::opt< bool > EnablePostRAMachineSched("enable-post-misched", cl::desc("Enable the post-ra machine instruction scheduling pass."), cl::init(true), cl::Hidden)
static void getSchedRegions(MachineBasicBlock *MBB, MBBRegionsVector &Regions, bool RegionsTopDown)
static cl::opt< unsigned > MIResourceCutOff("misched-resource-cutoff", cl::Hidden, cl::desc("Number of intervals to track"), cl::init(10))
static ScheduleDAGInstrs * createILPMaxScheduler(MachineSchedContext *C)
SmallVector< SchedRegion, 16 > MBBRegionsVector
static cl::opt< bool > MISchedDumpReservedCycles("misched-dump-reserved-cycles", cl::Hidden, cl::init(false), cl::desc("Dump resource usage at schedule boundary."))
static cl::opt< unsigned > ReadyListLimit("misched-limit", cl::Hidden, cl::desc("Limit ready list to N instructions"), cl::init(256))
Avoid quadratic complexity in unusually large basic blocks by limiting the size of the ready lists.
static cl::opt< bool > DumpCriticalPathLength("misched-dcpl", cl::Hidden, cl::desc("Print critical path length to stdout"))
static ScheduleDAGInstrs * createILPMinScheduler(MachineSchedContext *C)
static cl::opt< bool > MISchedDumpScheduleTrace("misched-dump-schedule-trace", cl::Hidden, cl::init(false), cl::desc("Dump resource usage at schedule boundary."))
static MachineSchedRegistry ILPMinRegistry("ilpmin", "Schedule bottom-up for min ILP", createILPMinScheduler)
Register const TargetRegisterInfo * TRI
std::pair< uint64_t, uint64_t > Interval
FunctionAnalysisManager FAM
#define INITIALIZE_PASS_DEPENDENCY(depName)
#define INITIALIZE_PASS_END(passName, arg, name, cfg, analysis)
#define INITIALIZE_PASS_BEGIN(passName, arg, name, cfg, analysis)
This file defines the PriorityQueue class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
static void initialize(TargetLibraryInfoImpl &TLI, const Triple &T, const llvm::StringTable &StandardNames, VectorLibrary VecLib)
Initialize the set of available library functions based on the specified target triple.
This file describes how to lower LLVM code to machine code.
Target-Independent Code Generator Pass Configuration Options pass.
static const X86InstrFMA3Group Groups[]
Class recording the (high level) value of a variable.
A manager for alias analyses.
A wrapper pass to provide the legacy pass manager access to a suitably prepared AAResults object.
Class for arbitrary precision integers.
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
Represent the analysis usage information of a pass.
AnalysisUsage & addRequired()
AnalysisUsage & addPreserved()
Add the specified Pass class to the set of analyses preserved by this pass.
LLVM_ABI void setPreservesCFG()
This function should be called by the pass, iff they do not:
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
reverse_iterator rend() const
size_t size() const
size - Get the array size.
reverse_iterator rbegin() const
bool test(unsigned Idx) const
Represents analyses that only rely on functions' control flow.
iterator find(const_arg_type_t< KeyT > Val)
size_type count(const_arg_type_t< KeyT > Val) const
Return 1 if the specified key is in the map, 0 otherwise.
ECValue - The EquivalenceClasses data structure is just a set of these.
EquivalenceClasses - This represents a collection of equivalence classes and supports three efficient...
iterator_range< member_iterator > members(const ECValue &ECV) const
member_iterator unionSets(const ElemTy &V1, const ElemTy &V2)
union - Merge the two equivalence sets for the specified values, inserting them if they do not alread...
void traceCandidate(const SchedCandidate &Cand)
LLVM_ABI void setPolicy(CandPolicy &Policy, bool IsPostRA, SchedBoundary &CurrZone, SchedBoundary *OtherZone)
Set the CandPolicy given a scheduling zone given the current resources and latencies inside and outsi...
MachineSchedPolicy RegionPolicy
const TargetSchedModel * SchedModel
static const char * getReasonStr(GenericSchedulerBase::CandReason Reason)
const MachineSchedContext * Context
CandReason
Represent the type of SchedCandidate found within a single queue.
const TargetRegisterInfo * TRI
void checkAcyclicLatency()
Set IsAcyclicLatencyLimited if the acyclic path is longer than the cyclic critical path by more cycle...
SchedCandidate BotCand
Candidate last picked from Bot boundary.
SchedCandidate TopCand
Candidate last picked from Top boundary.
virtual bool tryCandidate(SchedCandidate &Cand, SchedCandidate &TryCand, SchedBoundary *Zone) const
Apply a set of heuristics to a new candidate.
void dumpPolicy() const override
void initialize(ScheduleDAGMI *dag) override
Initialize the strategy after building the DAG for a new region.
void initCandidate(SchedCandidate &Cand, SUnit *SU, bool AtTop, const RegPressureTracker &RPTracker, RegPressureTracker &TempTracker)
void registerRoots() override
Notify this strategy that all roots have been released (including those that depend on EntrySU or Exi...
void initPolicy(MachineBasicBlock::iterator Begin, MachineBasicBlock::iterator End, unsigned NumRegionInstrs) override
Initialize the per-region scheduling policy.
void reschedulePhysReg(SUnit *SU, bool isTop)
SUnit * pickNode(bool &IsTopNode) override
Pick the best node to balance the schedule. Implements MachineSchedStrategy.
void pickNodeFromQueue(SchedBoundary &Zone, const CandPolicy &ZonePolicy, const RegPressureTracker &RPTracker, SchedCandidate &Candidate)
Pick the best candidate from the queue.
void schedNode(SUnit *SU, bool IsTopNode) override
Update the scheduler's state after scheduling a node.
SUnit * pickNodeBidirectional(bool &IsTopNode)
Pick the best candidate node from either the top or bottom queue.
bool getMemOperandsWithOffsetWidth(const MachineInstr &LdSt, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
Get the base register and byte offset of a load/store instr.
Itinerary data supplied by a subtarget to be used by a target.
LiveInterval - This class represents the liveness of a register, or stack slot.
MachineInstr * getInstructionFromIndex(SlotIndex index) const
Returns the instruction associated with the given index.
SlotIndex getInstructionIndex(const MachineInstr &Instr) const
Returns the base index of the given instruction.
LiveInterval & getInterval(Register Reg)
Result of a LiveRange query.
VNInfo * valueIn() const
Return the value that is live-in to the instruction.
Segments::iterator iterator
LiveQueryResult Query(SlotIndex Idx) const
Query Liveness at Idx.
VNInfo * getVNInfoBefore(SlotIndex Idx) const
getVNInfoBefore - Return the VNInfo that is live up to but not necessarily including Idx,...
SlotIndex beginIndex() const
beginIndex - Return the lowest numbered slot covered.
SlotIndex endIndex() const
endNumber - return the maximum point of the range of the whole, exclusive.
bool isLocal(SlotIndex Start, SlotIndex End) const
True iff this segment is a single segment that lies between the specified boundaries,...
LLVM_ABI iterator find(SlotIndex Pos)
find - Return an iterator pointing to the first segment that ends after Pos, or end().
static LocationSize precise(uint64_t Value)
MachineInstrBundleIterator< const MachineInstr > const_iterator
MachineInstrBundleIterator< MachineInstr > iterator
MachineBlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate machine basic b...
Analysis pass which computes a MachineDominatorTree.
Analysis pass which computes a MachineDominatorTree.
DominatorTree Class - Concrete subclass of DominatorTreeBase that is used to compute a normal dominat...
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
void print(raw_ostream &OS, const SlotIndexes *=nullptr) const
print - Print out the MachineFunction in a format suitable for debugging to the specified stream.
nonconst_iterator getNonConstIterator() const
Representation of each machine instruction.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
bool mayStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly modify memory.
Analysis pass that exposes the MachineLoopInfo for a machine function.
MachineOperand class - Representation of each machine instruction operand.
MachinePassRegistry - Track the registration of machine passes.
MachineSchedRegistry provides a selection of available machine instruction schedulers.
static LLVM_ABI MachinePassRegistry< ScheduleDAGCtor > Registry
ScheduleDAGInstrs *(*)(MachineSchedContext *) ScheduleDAGCtor
LLVM_ABI PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MFAM)
LLVM_ABI MachineSchedulerPass(const TargetMachine *TM)
LLVM_ABI ~MachineSchedulerPass()
static LLVM_ABI PassRegistry * getPassRegistry()
getPassRegistry - Access the global registry object, which is automatically initialized at applicatio...
void initPolicy(MachineBasicBlock::iterator Begin, MachineBasicBlock::iterator End, unsigned NumRegionInstrs) override
Optionally override the per-region scheduling policy.
virtual bool tryCandidate(SchedCandidate &Cand, SchedCandidate &TryCand)
Apply a set of heuristics to a new candidate for PostRA scheduling.
void schedNode(SUnit *SU, bool IsTopNode) override
Called after ScheduleDAGMI has scheduled an instruction and updated scheduled/remaining flags in the ...
SchedCandidate BotCand
Candidate last picked from Bot boundary.
void pickNodeFromQueue(SchedBoundary &Zone, SchedCandidate &Cand)
void initialize(ScheduleDAGMI *Dag) override
Initialize the strategy after building the DAG for a new region.
SchedCandidate TopCand
Candidate last picked from Top boundary.
SUnit * pickNodeBidirectional(bool &IsTopNode)
Pick the best candidate node from either the top or bottom queue.
void registerRoots() override
Notify this strategy that all roots have been released (including those that depend on EntrySU or Exi...
SUnit * pickNode(bool &IsTopNode) override
Pick the next node to schedule.
LLVM_ABI PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MFAM)
LLVM_ABI PostMachineSchedulerPass(const TargetMachine *TM)
LLVM_ABI ~PostMachineSchedulerPass()
A set of analyses that are preserved following a run of a transformation pass.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
PreservedAnalyses & preserveSet()
Mark an analysis set as preserved.
Capture a change in pressure for a single pressure set.
unsigned getPSetOrMax() const
List of PressureChanges in order of increasing, unique PSetID.
LLVM_ABI void dump(const TargetRegisterInfo &TRI) const
LLVM_ABI void addPressureChange(VirtRegOrUnit VRegOrUnit, bool IsDec, const MachineRegisterInfo *MRI)
Add a change in pressure to the pressure diff of a given instruction.
void clear()
clear - Erase all elements from the queue.
Helpers for implementing custom MachineSchedStrategy classes.
ArrayRef< SUnit * > elements()
LLVM_ABI void dump() const
std::vector< SUnit * >::iterator iterator
StringRef getName() const
Track the current register pressure at some position in the instruction stream, and remember the high...
LLVM_ABI void getMaxUpwardPressureDelta(const MachineInstr *MI, PressureDiff *PDiff, RegPressureDelta &Delta, ArrayRef< PressureChange > CriticalPSets, ArrayRef< unsigned > MaxPressureLimit)
Consider the pressure increase caused by traversing this instruction bottom-up.
LLVM_ABI void getMaxDownwardPressureDelta(const MachineInstr *MI, RegPressureDelta &Delta, ArrayRef< PressureChange > CriticalPSets, ArrayRef< unsigned > MaxPressureLimit)
Consider the pressure increase caused by traversing this instruction top-down.
LLVM_ABI void getUpwardPressureDelta(const MachineInstr *MI, PressureDiff &PDiff, RegPressureDelta &Delta, ArrayRef< PressureChange > CriticalPSets, ArrayRef< unsigned > MaxPressureLimit) const
This is the fast version of querying register pressure that does not directly depend on current liven...
List of registers defined and used by a machine instruction.
LLVM_ABI void collect(const MachineInstr &MI, const TargetRegisterInfo &TRI, const MachineRegisterInfo &MRI, bool TrackLaneMasks, bool IgnoreDead)
Analyze the given instruction MI and fill in the Uses, Defs and DeadDefs list based on the MachineOpe...
LLVM_ABI void adjustLaneLiveness(const LiveIntervals &LIS, const MachineRegisterInfo &MRI, SlotIndex Pos, MachineInstr *AddFlagsMI=nullptr)
Use liveness information to find out which uses/defs are partially undefined/dead and adjust the VReg...
LLVM_ABI void detectDeadDefs(const MachineInstr &MI, const LiveIntervals &LIS)
Use liveness information to find dead defs not marked with a dead flag and move them to the DeadDefs ...
RegisterPassParser class - Handle the addition of new machine passes.
Wrapper class representing virtual and physical registers.
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
LLVM_ABI void add(IntervalTy A, const unsigned CutOff=10)
Adds an interval [a, b) to the collection of the instance.
static IntervalTy getResourceIntervalBottom(unsigned C, unsigned AcquireAtCycle, unsigned ReleaseAtCycle)
These function return the interval used by a resource in bottom and top scheduling.
static LLVM_ABI bool intersects(IntervalTy A, IntervalTy B)
Checks whether intervals intersect.
std::pair< int64_t, int64_t > IntervalTy
Represents an interval of discrete integer values closed on the left and open on the right: [a,...
static IntervalTy getResourceIntervalTop(unsigned C, unsigned AcquireAtCycle, unsigned ReleaseAtCycle)
Kind getKind() const
Returns an enum value representing the kind of the dependence.
@ Anti
A register anti-dependence (aka WAR).
@ Data
Regular data dependence (aka true-dependence).
bool isWeak() const
Tests if this a weak dependence.
@ Cluster
Weak DAG edge linking a chain of clustered instrs.
@ Artificial
Arbitrary strong DAG edge (no real dependence).
@ Weak
Arbitrary weak DAG edge.
unsigned getLatency() const
Returns the latency value for this edge, which roughly means the minimum number of cycles that must e...
bool isArtificial() const
Tests if this is an Order dependence that is marked as "artificial", meaning it isn't necessary for c...
bool isCtrl() const
Shorthand for getKind() != SDep::Data.
Register getReg() const
Returns the register associated with this edge.
bool isArtificialDep() const
bool isCtrlDep() const
Tests if this is not an SDep::Data dependence.
Scheduling unit. This is a node in the scheduling DAG.
bool isCall
Is a function call.
unsigned TopReadyCycle
Cycle relative to start when node is ready.
unsigned NodeNum
Entry # of node in the node vector.
bool isUnbuffered
Uses an unbuffered resource.
unsigned getHeight() const
Returns the height of this node, which is the length of the maximum path down to any node which has n...
unsigned short Latency
Node latency.
unsigned getDepth() const
Returns the depth of this node, which is the length of the maximum path up to any node which has no p...
bool isScheduled
True once scheduled.
unsigned ParentClusterIdx
The parent cluster id.
bool hasPhysRegDefs
Has physreg defs that are being used.
unsigned BotReadyCycle
Cycle relative to end when node is ready.
SmallVector< SDep, 4 > Succs
All sunit successors.
bool hasReservedResource
Uses a reserved resource.
bool isBottomReady() const
bool hasPhysRegUses
Has physreg uses.
SmallVector< SDep, 4 > Preds
All sunit predecessors.
MachineInstr * getInstr() const
Returns the representative MachineInstr for this SUnit.
Each Scheduling boundary is associated with ready queues.
LLVM_ABI unsigned getNextResourceCycleByInstance(unsigned InstanceIndex, unsigned ReleaseAtCycle, unsigned AcquireAtCycle)
Compute the next cycle at which the given processor resource unit can be scheduled.
LLVM_ABI void releasePending()
Release pending ready nodes in to the available queue.
unsigned getDependentLatency() const
bool isReservedGroup(unsigned PIdx) const
unsigned getScheduledLatency() const
Get the number of latency cycles "covered" by the scheduled instructions.
LLVM_ABI void incExecutedResources(unsigned PIdx, unsigned Count)
bool isResourceLimited() const
const TargetSchedModel * SchedModel
unsigned getExecutedCount() const
Get a scaled count for the minimum execution time of the scheduled micro-ops that are ready to execut...
LLVM_ABI unsigned getLatencyStallCycles(SUnit *SU)
Get the difference between the given SUnit's ready time and the current cycle.
LLVM_ABI unsigned findMaxLatency(ArrayRef< SUnit * > ReadySUs)
LLVM_ABI void dumpReservedCycles() const
Dump the state of the information that tracks resource usage.
LLVM_ABI unsigned getOtherResourceCount(unsigned &OtherCritIdx)
LLVM_ABI void bumpNode(SUnit *SU)
Move the boundary of scheduled code by one SUnit.
unsigned getCriticalCount() const
Get the scaled count of scheduled micro-ops and resources, including executed resources.
LLVM_ABI SUnit * pickOnlyChoice()
Call this before applying any other heuristics to the Available queue.
LLVM_ABI void releaseNode(SUnit *SU, unsigned ReadyCycle, bool InPQueue, unsigned Idx=0)
Release SU to make it ready.
LLVM_ABI unsigned countResource(const MCSchedClassDesc *SC, unsigned PIdx, unsigned Cycles, unsigned ReadyCycle, unsigned StartAtCycle)
Add the given processor resource to this scheduled zone.
LLVM_ABI ~SchedBoundary()
ScheduleHazardRecognizer * HazardRec
LLVM_ABI void init(ScheduleDAGMI *dag, const TargetSchedModel *smodel, SchedRemainder *rem)
unsigned getResourceCount(unsigned ResIdx) const
LLVM_ABI void bumpCycle(unsigned NextCycle)
Move the boundary of scheduled code by one cycle.
unsigned getCurrMOps() const
Micro-ops issued in the current cycle.
unsigned getCurrCycle() const
Number of cycles to issue the instructions scheduled in this zone.
LLVM_ABI bool checkHazard(SUnit *SU)
Does this SU have a hazard within the current instruction group.
LLVM_ABI std::pair< unsigned, unsigned > getNextResourceCycle(const MCSchedClassDesc *SC, unsigned PIdx, unsigned ReleaseAtCycle, unsigned AcquireAtCycle)
Compute the next cycle at which the given processor resource can be scheduled.
LLVM_ABI void dumpScheduledState() const
LLVM_ABI void removeReady(SUnit *SU)
Remove SU from the ready set for this boundary.
unsigned getZoneCritResIdx() const
unsigned getUnscheduledLatency(SUnit *SU) const
Compute the values of each DAG node for various metrics during DFS.
unsigned getNumInstrs(const SUnit *SU) const
Get the number of instructions in the given subtree and its children.
unsigned getSubtreeID(const SUnit *SU) const
Get the ID of the subtree the given DAG node belongs to.
ILPValue getILP(const SUnit *SU) const
Get the ILP value for a DAG node.
unsigned getSubtreeLevel(unsigned SubtreeID) const
Get the connection level of a subtree.
A ScheduleDAG for scheduling lists of MachineInstr.
SmallVector< ClusterInfo > & getClusters()
Returns the array of the clusters.
virtual void finishBlock()
Cleans up after scheduling in the given block.
MachineBasicBlock::iterator end() const
Returns an iterator to the bottom of the current scheduling region.
std::string getDAGName() const override
Returns a label for the region of code covered by the DAG.
MachineBasicBlock * BB
The block in which to insert instructions.
MachineInstr * FirstDbgValue
virtual void startBlock(MachineBasicBlock *BB)
Prepares to perform scheduling in the given block.
MachineBasicBlock::iterator RegionEnd
The end of the range to be scheduled.
const MCSchedClassDesc * getSchedClass(SUnit *SU) const
Resolves and cache a resolved scheduling class for an SUnit.
DbgValueVector DbgValues
Remember instruction that precedes DBG_VALUE.
bool addEdge(SUnit *SuccSU, const SDep &PredDep)
Add a DAG edge to the given SU with the given predecessor dependence data.
DumpDirection
The direction that should be used to dump the scheduled Sequence.
bool TrackLaneMasks
Whether lane masks should get tracked.
void dumpNode(const SUnit &SU) const override
bool IsReachable(SUnit *SU, SUnit *TargetSU)
IsReachable - Checks if SU is reachable from TargetSU.
MachineBasicBlock::iterator begin() const
Returns an iterator to the top of the current scheduling region.
void buildSchedGraph(AAResults *AA, RegPressureTracker *RPTracker=nullptr, PressureDiffs *PDiffs=nullptr, LiveIntervals *LIS=nullptr, bool TrackLaneMasks=false)
Builds SUnits for the current region.
SUnit * getSUnit(MachineInstr *MI) const
Returns an existing SUnit for this MI, or nullptr.
TargetSchedModel SchedModel
TargetSchedModel provides an interface to the machine model.
bool canAddEdge(SUnit *SuccSU, SUnit *PredSU)
True if an edge can be added from PredSU to SuccSU without creating a cycle.
MachineBasicBlock::iterator RegionBegin
The beginning of the range to be scheduled.
virtual void enterRegion(MachineBasicBlock *bb, MachineBasicBlock::iterator begin, MachineBasicBlock::iterator end, unsigned regioninstrs)
Initialize the DAG and common scheduler state for a new scheduling region.
void dump() const override
void setDumpDirection(DumpDirection D)
ScheduleDAGMILive is an implementation of ScheduleDAGInstrs that schedules machine instructions while...
void scheduleMI(SUnit *SU, bool IsTopNode)
Move an instruction and update register pressure.
void schedule() override
Implement ScheduleDAGInstrs interface for scheduling a sequence of reorderable instructions.
VReg2SUnitMultiMap VRegUses
Maps vregs to the SUnits of their uses in the current scheduling region.
void computeDFSResult()
Compute a DFSResult after DAG building is complete, and before any queue comparisons.
PressureDiff & getPressureDiff(const SUnit *SU)
SchedDFSResult * DFSResult
Information about DAG subtrees.
void enterRegion(MachineBasicBlock *bb, MachineBasicBlock::iterator begin, MachineBasicBlock::iterator end, unsigned regioninstrs) override
Implement the ScheduleDAGInstrs interface for handling the next scheduling region.
void initQueues(ArrayRef< SUnit * > TopRoots, ArrayRef< SUnit * > BotRoots)
Release ExitSU predecessors and setup scheduler queues.
bool ShouldTrackLaneMasks
RegPressureTracker BotRPTracker
void buildDAGWithRegPressure()
Call ScheduleDAGInstrs::buildSchedGraph with register pressure tracking enabled.
std::vector< PressureChange > RegionCriticalPSets
List of pressure sets that exceed the target's pressure limit before scheduling, listed in increasing...
void updateScheduledPressure(const SUnit *SU, const std::vector< unsigned > &NewMaxPressure)
PressureDiffs SUPressureDiffs
unsigned computeCyclicCriticalPath()
Compute the cyclic critical path through the DAG.
void updatePressureDiffs(ArrayRef< VRegMaskOrUnit > LiveUses)
Update the PressureDiff array for liveness after scheduling this instruction.
void collectVRegUses(SUnit &SU)
RegisterClassInfo * RegClassInfo
const SchedDFSResult * getDFSResult() const
Return a non-null DFS result if the scheduling strategy initialized it.
RegPressureTracker RPTracker
bool ShouldTrackPressure
Register pressure in this region computed by initRegPressure.
~ScheduleDAGMILive() override
void dump() const override
BitVector & getScheduledTrees()
MachineBasicBlock::iterator LiveRegionEnd
RegPressureTracker TopRPTracker
ScheduleDAGMI is an implementation of ScheduleDAGInstrs that simply schedules machine instructions ac...
void dumpSchedule() const
dump the scheduled Sequence.
std::unique_ptr< MachineSchedStrategy > SchedImpl
void startBlock(MachineBasicBlock *bb) override
Prepares to perform scheduling in the given block.
void releasePred(SUnit *SU, SDep *PredEdge)
ReleasePred - Decrement the NumSuccsLeft count of a predecessor.
void initQueues(ArrayRef< SUnit * > TopRoots, ArrayRef< SUnit * > BotRoots)
Release ExitSU predecessors and setup scheduler queues.
void moveInstruction(MachineInstr *MI, MachineBasicBlock::iterator InsertPos)
Change the position of an instruction within the basic block and update live ranges and region bounda...
void releasePredecessors(SUnit *SU)
releasePredecessors - Call releasePred on each of SU's predecessors.
void postProcessDAG()
Apply each ScheduleDAGMutation step in order.
void dumpScheduleTraceTopDown() const
Print execution trace of the schedule top-down or bottom-up.
void schedule() override
Implement ScheduleDAGInstrs interface for scheduling a sequence of reorderable instructions.
void findRootsAndBiasEdges(SmallVectorImpl< SUnit * > &TopRoots, SmallVectorImpl< SUnit * > &BotRoots)
MachineBasicBlock::iterator CurrentBottom
The bottom of the unscheduled zone.
virtual bool hasVRegLiveness() const
Return true if this DAG supports VReg liveness and RegPressure.
void enterRegion(MachineBasicBlock *bb, MachineBasicBlock::iterator begin, MachineBasicBlock::iterator end, unsigned regioninstrs) override
Implement the ScheduleDAGInstrs interface for handling the next scheduling region.
LiveIntervals * getLIS() const
void viewGraph(const Twine &Name, const Twine &Title) override
viewGraph - Pop up a ghostview window with the reachable parts of the DAG rendered using 'dot'.
void viewGraph() override
Out-of-line implementation with no arguments is handy for gdb.
void releaseSucc(SUnit *SU, SDep *SuccEdge)
ReleaseSucc - Decrement the NumPredsLeft count of a successor.
void dumpScheduleTraceBottomUp() const
~ScheduleDAGMI() override
void finishBlock() override
Cleans up after scheduling in the given block.
void updateQueues(SUnit *SU, bool IsTopNode)
Update scheduler DAG and queues after scheduling an instruction.
void placeDebugValues()
Reinsert debug_values recorded in ScheduleDAGInstrs::DbgValues.
MachineBasicBlock::iterator CurrentTop
The top of the unscheduled zone.
void releaseSuccessors(SUnit *SU)
releaseSuccessors - Call releaseSucc on each of SU's successors.
std::vector< std::unique_ptr< ScheduleDAGMutation > > Mutations
Ordered list of DAG postprocessing steps.
Mutate the DAG as a postpass after normal DAG building.
MachineRegisterInfo & MRI
Virtual/real register map.
std::vector< SUnit > SUnits
The scheduling units.
const TargetRegisterInfo * TRI
Target processor register info.
SUnit EntrySU
Special node for the region entry.
MachineFunction & MF
Machine function.
void dumpNodeAll(const SUnit &SU) const
SUnit ExitSU
Special node for the region exit.
SlotIndex - An opaque wrapper around machine indexes.
static bool isSameInstr(SlotIndex A, SlotIndex B)
isSameInstr - Return true if A and B refer to the same instruction.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
std::reverse_iterator< const_iterator > const_reverse_iterator
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
iterator_base< SparseMultiSet * > iterator
Information about stack frame layout on the target.
StackDirection getStackGrowthDirection() const
getStackGrowthDirection - Return the direction the stack grows
TargetInstrInfo - Interface to description of machine instruction set.
virtual const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent=false) const
Return the register class that should be used for the specified value type.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
Primary interface to the complete machine description for the target machine.
Target-Independent Code Generator Pass Configuration Options.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
Provide an instruction scheduling machine model to CodeGen passes.
unsigned getMicroOpFactor() const
Multiply number of micro-ops by this factor to normalize it relative to other resources.
ProcResIter getWriteProcResEnd(const MCSchedClassDesc *SC) const
LLVM_ABI bool hasInstrSchedModel() const
Return true if this machine model includes an instruction-level scheduling model.
const MCWriteProcResEntry * ProcResIter
unsigned getResourceFactor(unsigned ResIdx) const
Multiply the number of units consumed for a resource by this factor to normalize it relative to other...
LLVM_ABI unsigned getNumMicroOps(const MachineInstr *MI, const MCSchedClassDesc *SC=nullptr) const
Return the number of issue slots required for this MI.
unsigned getNumProcResourceKinds() const
Get the number of kinds of resources for this target.
ProcResIter getWriteProcResBegin(const MCSchedClassDesc *SC) const
virtual void overridePostRASchedPolicy(MachineSchedPolicy &Policy, const SchedRegion &Region) const
Override generic post-ra scheduling policy within a region.
virtual void overrideSchedPolicy(MachineSchedPolicy &Policy, const SchedRegion &Region) const
Override generic scheduling policy within a region.
virtual bool enableMachineScheduler() const
True if the subtarget should run MachineScheduler after aggressive coalescing.
virtual bool enablePostRAMachineScheduler() const
True if the subtarget should run a machine scheduler after register allocation.
virtual const TargetFrameLowering * getFrameLowering() const
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetLowering * getTargetLowering() const
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
VNInfo - Value Number Information.
SlotIndex def
The index of the defining instruction.
bool isPHIDef() const
Returns true if this value is defined by a PHI instruction (or was, PHI instructions may have been el...
Wrapper class representing a virtual register or register unit.
Base class for the machine scheduler classes.
void scheduleRegions(ScheduleDAGInstrs &Scheduler, bool FixKillFlags)
Main driver for both MachineScheduler and PostMachineScheduler.
Impl class for MachineScheduler.
void setMFAM(MachineFunctionAnalysisManager *MFAM)
void setLegacyPass(MachineFunctionPass *P)
bool run(MachineFunction &MF, const TargetMachine &TM, const RequiredAnalyses &Analyses)
MachineSchedulerImpl()=default
ScheduleDAGInstrs * createMachineScheduler()
Instantiate a ScheduleDAGInstrs that will be owned by the caller.
Impl class for PostMachineScheduler.
bool run(MachineFunction &Func, const TargetMachine &TM, const RequiredAnalyses &Analyses)
void setMFAM(MachineFunctionAnalysisManager *MFAM)
ScheduleDAGInstrs * createPostMachineScheduler()
Instantiate a ScheduleDAGInstrs for PostRA scheduling that will be owned by the caller.
void setLegacyPass(MachineFunctionPass *P)
PostMachineSchedulerImpl()=default
A raw_ostream that writes to an std::string.
This provides a very simple, boring adaptor for a begin and end iterator into a range type.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
Abstract Attribute helper functions.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ C
The default llvm calling convention, compatible with C.
LLVM_ABI StringRef getColorString(unsigned NodeNumber)
Get a color string for this node number.
void apply(Opt *O, const Mod &M, const Mods &... Ms)
ValuesClass values(OptsTy... Options)
Helper to build a ValuesClass by forwarding a variable number of arguments as an initializer list to ...
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
ScheduleDAGMILive * createSchedLive(MachineSchedContext *C)
Create the standard converging machine scheduler.
bool operator<(int64_t V1, const APSInt &V2)
void stable_sort(R &&Range)
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
LLVM_ABI unsigned getWeakLeft(const SUnit *SU, bool isTop)
FormattedString right_justify(StringRef Str, unsigned Width)
right_justify - add spaces before string so total output is Width characters.
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
Printable PrintLaneMask(LaneBitmask LaneMask)
Create Printable object to print LaneBitmasks on a raw_ostream.
AnalysisManager< MachineFunction > MachineFunctionAnalysisManager
LLVM_ABI char & MachineSchedulerID
MachineScheduler - This pass schedules machine instructions.
LLVM_ABI char & PostMachineSchedulerID
PostMachineScheduler - This pass schedules machine instructions postRA.
LLVM_ABI PreservedAnalyses getMachineFunctionPassPreservedAnalyses()
Returns the minimum set of Analyses that all machine function passes must preserve.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
LLVM_ABI bool tryPressure(const PressureChange &TryP, const PressureChange &CandP, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason, const TargetRegisterInfo *TRI, const MachineFunction &MF)
ScheduleDAGMI * createSchedPostRA(MachineSchedContext *C)
Create a generic scheduler with no vreg liveness or DAG mutation passes.
void sort(IteratorTy Start, IteratorTy End)
cl::opt< bool > ViewMISchedDAGs
LLVM_ABI std::unique_ptr< ScheduleDAGMutation > createStoreClusterDAGMutation(const TargetInstrInfo *TII, const TargetRegisterInfo *TRI, bool ReorderWhileClustering=false)
If ReorderWhileClustering is set to true, no attempt will be made to reduce reordering due to store c...
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
FunctionAddr VTableAddr Count
LLVM_ABI cl::opt< bool > VerifyScheduling
bool is_sorted(R &&Range, Compare C)
Wrapper function around std::is_sorted to check if elements in a range R are sorted with respect to a...
LLVM_ABI bool tryLatency(GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, SchedBoundary &Zone)
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
constexpr unsigned InvalidClusterId
FormattedString left_justify(StringRef Str, unsigned Width)
left_justify - append spaces after string so total output is Width characters.
bool isTheSameCluster(unsigned A, unsigned B)
Return whether the input cluster ID's are the same and valid.
LLVM_ABI std::unique_ptr< ScheduleDAGMutation > createLoadClusterDAGMutation(const TargetInstrInfo *TII, const TargetRegisterInfo *TRI, bool ReorderWhileClustering=false)
If ReorderWhileClustering is set to true, no attempt will be made to reduce reordering due to store c...
DWARFExpression::Operation Op
LLVM_ABI bool tryGreater(int TryVal, int CandVal, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason)
SmallPtrSet< SUnit *, 8 > ClusterInfo
Keep record of which SUnit are in the same cluster group.
void ViewGraph(const GraphType &G, const Twine &Name, bool ShortNames=false, const Twine &Title="", GraphProgram::Name Program=GraphProgram::DOT)
ViewGraph - Emit a dot graph, run 'dot', run gv on the postscript file, then cleanup.
ArrayRef(const T &OneElt) -> ArrayRef< T >
LLVM_ABI void initializeMachineSchedulerLegacyPass(PassRegistry &)
LLVM_ABI void dumpRegSetPressure(ArrayRef< unsigned > SetPressure, const TargetRegisterInfo *TRI)
LLVM_ABI bool tryLess(int TryVal, int CandVal, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason)
Return true if this heuristic determines order.
LLVM_ABI std::unique_ptr< ScheduleDAGMutation > createCopyConstrainDAGMutation(const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
LLVM_ABI cl::opt< MISched::Direction > PreRADirection
LLVM_ABI Printable printMBBReference(const MachineBasicBlock &MBB)
Prints a machine basic block reference.
LLVM_ABI void initializePostMachineSchedulerLegacyPass(PassRegistry &)
LLVM_ABI int biasPhysReg(const SUnit *SU, bool isTop)
Minimize physical register live ranges.
cl::opt< bool > PrintDAGs
Implement std::hash so that hash_code can be used in STL containers.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
static std::string getNodeDescription(const SUnit *SU, const ScheduleDAG *G)
static std::string getEdgeAttributes(const SUnit *Node, SUnitIterator EI, const ScheduleDAG *Graph)
If you want to override the dot attributes printed for a particular edge, override this method.
static std::string getGraphName(const ScheduleDAG *G)
static std::string getNodeLabel(const SUnit *SU, const ScheduleDAG *G)
static bool isNodeHidden(const SUnit *Node, const ScheduleDAG *G)
DOTGraphTraits(bool isSimple=false)
static std::string getNodeAttributes(const SUnit *N, const ScheduleDAG *G)
static bool renderGraphFromBottomUp()
DOTGraphTraits - Template class that can be specialized to customize how graphs are converted to 'dot...
DefaultDOTGraphTraits(bool simple=false)
Policy for scheduling the next instruction in the candidate's zone.
Store the state used by GenericScheduler heuristics, required for the lifetime of one invocation of p...
void setBest(SchedCandidate &Best)
void reset(const CandPolicy &NewPolicy)
LLVM_ABI void initResourceDelta(const ScheduleDAGMI *DAG, const TargetSchedModel *SchedModel)
SchedResourceDelta ResDelta
Status of an instruction's critical resource consumption.
unsigned DemandedResources
static constexpr LaneBitmask getNone()
Summarize the scheduling resources required for an instruction of a particular scheduling class.
Identify one of the processor resource kinds consumed by a particular scheduling class for the specif...
MachineSchedContext provides enough context from the MachineScheduler pass for the target to instanti...
const MachineDominatorTree * MDT
RegisterClassInfo * RegClassInfo
MachineBlockFrequencyInfo * MBFI
const MachineLoopInfo * MLI
virtual ~MachineSchedContext()
PressureChange CriticalMax
PressureChange CurrentMax
RegisterPressure computed within a region of instructions delimited by TopPos and BottomPos.
A region of an MBB for scheduling.
Summarize the unscheduled region.
LLVM_ABI void init(ScheduleDAGMI *DAG, const TargetSchedModel *SchedModel)
SmallVector< unsigned, 16 > RemainingCounts
An individual mapping from virtual register number to SUnit.
MachineBlockFrequencyInfo & MBFI
MachineDominatorTree & MDT