LLVM 20.0.0git
ScheduleDAG.cpp
Go to the documentation of this file.
1//===- ScheduleDAG.cpp - Implement the ScheduleDAG class ------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9/// \file Implements the ScheduleDAG class, which is a base class used by
10/// scheduling implementation classes.
11//
12//===----------------------------------------------------------------------===//
13
15#include "llvm/ADT/STLExtras.h"
17#include "llvm/ADT/Statistic.h"
24#include "llvm/Config/llvm-config.h"
27#include "llvm/Support/Debug.h"
29#include <algorithm>
30#include <cassert>
31#include <iterator>
32#include <limits>
33#include <utility>
34#include <vector>
35
36using namespace llvm;
37
38#define DEBUG_TYPE "pre-RA-sched"
39
40STATISTIC(NumNewPredsAdded, "Number of times a single predecessor was added");
41STATISTIC(NumTopoInits,
42 "Number of times the topological order has been recomputed");
43
44#ifndef NDEBUG
46 "stress-sched", cl::Hidden, cl::init(false),
47 cl::desc("Stress test instruction scheduling"));
48#endif
49
50void SchedulingPriorityQueue::anchor() {}
51
53 : TM(mf.getTarget()), TII(mf.getSubtarget().getInstrInfo()),
54 TRI(mf.getSubtarget().getRegisterInfo()), MF(mf),
55 MRI(mf.getRegInfo()) {
56#ifndef NDEBUG
58#endif
59}
60
62
64 SUnits.clear();
65 EntrySU = SUnit();
66 ExitSU = SUnit();
67}
68
69const MCInstrDesc *ScheduleDAG::getNodeDesc(const SDNode *Node) const {
70 if (!Node || !Node->isMachineOpcode()) return nullptr;
71 return &TII->get(Node->getMachineOpcode());
72}
73
75 switch (getKind()) {
76 case Data: dbgs() << "Data"; break;
77 case Anti: dbgs() << "Anti"; break;
78 case Output: dbgs() << "Out "; break;
79 case Order: dbgs() << "Ord "; break;
80 }
81
82 switch (getKind()) {
83 case Data:
84 dbgs() << " Latency=" << getLatency();
85 if (TRI && isAssignedRegDep())
86 dbgs() << " Reg=" << printReg(getReg(), TRI);
87 break;
88 case Anti:
89 case Output:
90 dbgs() << " Latency=" << getLatency();
91 break;
92 case Order:
93 dbgs() << " Latency=" << getLatency();
94 switch(Contents.OrdKind) {
95 case Barrier: dbgs() << " Barrier"; break;
96 case MayAliasMem:
97 case MustAliasMem: dbgs() << " Memory"; break;
98 case Artificial: dbgs() << " Artificial"; break;
99 case Weak: dbgs() << " Weak"; break;
100 case Cluster: dbgs() << " Cluster"; break;
101 }
102 break;
103 }
104}
105
106bool SUnit::addPred(const SDep &D, bool Required) {
107 // If this node already has this dependence, don't add a redundant one.
108 for (SDep &PredDep : Preds) {
109 // Zero-latency weak edges may be added purely for heuristic ordering. Don't
110 // add them if another kind of edge already exists.
111 if (!Required && PredDep.getSUnit() == D.getSUnit())
112 return false;
113 if (PredDep.overlaps(D)) {
114 // Extend the latency if needed. Equivalent to
115 // removePred(PredDep) + addPred(D).
116 if (PredDep.getLatency() < D.getLatency()) {
117 SUnit *PredSU = PredDep.getSUnit();
118 // Find the corresponding successor in N.
119 SDep ForwardD = PredDep;
120 ForwardD.setSUnit(this);
121 for (SDep &SuccDep : PredSU->Succs) {
122 if (SuccDep == ForwardD) {
123 SuccDep.setLatency(D.getLatency());
124 break;
125 }
126 }
127 PredDep.setLatency(D.getLatency());
128 // Changing latency, dirty the involved SUnits.
129 this->setDepthDirty();
131 }
132 return false;
133 }
134 }
135 // Now add a corresponding succ to N.
136 SDep P = D;
137 P.setSUnit(this);
138 SUnit *N = D.getSUnit();
139 // Update the bookkeeping.
140 if (D.getKind() == SDep::Data) {
141 assert(NumPreds < std::numeric_limits<unsigned>::max() &&
142 "NumPreds will overflow!");
143 assert(N->NumSuccs < std::numeric_limits<unsigned>::max() &&
144 "NumSuccs will overflow!");
145 ++NumPreds;
146 ++N->NumSuccs;
147 }
148 if (!N->isScheduled) {
149 if (D.isWeak()) {
151 }
152 else {
153 assert(NumPredsLeft < std::numeric_limits<unsigned>::max() &&
154 "NumPredsLeft will overflow!");
155 ++NumPredsLeft;
156 }
157 }
158 if (!isScheduled) {
159 if (D.isWeak()) {
160 ++N->WeakSuccsLeft;
161 }
162 else {
163 assert(N->NumSuccsLeft < std::numeric_limits<unsigned>::max() &&
164 "NumSuccsLeft will overflow!");
165 ++N->NumSuccsLeft;
166 }
167 }
168 Preds.push_back(D);
169 N->Succs.push_back(P);
170 this->setDepthDirty();
171 N->setHeightDirty();
172 return true;
173}
174
176 // Find the matching predecessor.
178 if (I == Preds.end())
179 return;
180 // Find the corresponding successor in N.
181 SDep P = D;
182 P.setSUnit(this);
183 SUnit *N = D.getSUnit();
185 assert(Succ != N->Succs.end() && "Mismatching preds / succs lists!");
186 // Update the bookkeeping.
187 if (P.getKind() == SDep::Data) {
188 assert(NumPreds > 0 && "NumPreds will underflow!");
189 assert(N->NumSuccs > 0 && "NumSuccs will underflow!");
190 --NumPreds;
191 --N->NumSuccs;
192 }
193 if (!N->isScheduled) {
194 if (D.isWeak()) {
195 assert(WeakPredsLeft > 0 && "WeakPredsLeft will underflow!");
197 } else {
198 assert(NumPredsLeft > 0 && "NumPredsLeft will underflow!");
199 --NumPredsLeft;
200 }
201 }
202 if (!isScheduled) {
203 if (D.isWeak()) {
204 assert(N->WeakSuccsLeft > 0 && "WeakSuccsLeft will underflow!");
205 --N->WeakSuccsLeft;
206 } else {
207 assert(N->NumSuccsLeft > 0 && "NumSuccsLeft will underflow!");
208 --N->NumSuccsLeft;
209 }
210 }
211 N->Succs.erase(Succ);
212 Preds.erase(I);
213 this->setDepthDirty();
214 N->setHeightDirty();
215}
216
218 if (!isDepthCurrent) return;
219 SmallVector<SUnit*, 8> WorkList;
220 WorkList.push_back(this);
221 do {
222 SUnit *SU = WorkList.pop_back_val();
223 SU->isDepthCurrent = false;
224 for (SDep &SuccDep : SU->Succs) {
225 SUnit *SuccSU = SuccDep.getSUnit();
226 if (SuccSU->isDepthCurrent)
227 WorkList.push_back(SuccSU);
228 }
229 } while (!WorkList.empty());
230}
231
233 if (!isHeightCurrent) return;
234 SmallVector<SUnit*, 8> WorkList;
235 WorkList.push_back(this);
236 do {
237 SUnit *SU = WorkList.pop_back_val();
238 SU->isHeightCurrent = false;
239 for (SDep &PredDep : SU->Preds) {
240 SUnit *PredSU = PredDep.getSUnit();
241 if (PredSU->isHeightCurrent)
242 WorkList.push_back(PredSU);
243 }
244 } while (!WorkList.empty());
245}
246
247void SUnit::setDepthToAtLeast(unsigned NewDepth) {
248 if (NewDepth <= getDepth())
249 return;
251 Depth = NewDepth;
252 isDepthCurrent = true;
253}
254
255void SUnit::setHeightToAtLeast(unsigned NewHeight) {
256 if (NewHeight <= getHeight())
257 return;
259 Height = NewHeight;
260 isHeightCurrent = true;
261}
262
263/// Calculates the maximal path from the node to the exit.
264void SUnit::ComputeDepth() {
265 SmallVector<SUnit*, 8> WorkList;
266 WorkList.push_back(this);
267 do {
268 SUnit *Cur = WorkList.back();
269
270 bool Done = true;
271 unsigned MaxPredDepth = 0;
272 for (const SDep &PredDep : Cur->Preds) {
273 SUnit *PredSU = PredDep.getSUnit();
274 if (PredSU->isDepthCurrent)
275 MaxPredDepth = std::max(MaxPredDepth,
276 PredSU->Depth + PredDep.getLatency());
277 else {
278 Done = false;
279 WorkList.push_back(PredSU);
280 }
281 }
282
283 if (Done) {
284 WorkList.pop_back();
285 if (MaxPredDepth != Cur->Depth) {
286 Cur->setDepthDirty();
287 Cur->Depth = MaxPredDepth;
288 }
289 Cur->isDepthCurrent = true;
290 }
291 } while (!WorkList.empty());
292}
293
294/// Calculates the maximal path from the node to the entry.
295void SUnit::ComputeHeight() {
296 SmallVector<SUnit*, 8> WorkList;
297 WorkList.push_back(this);
298 do {
299 SUnit *Cur = WorkList.back();
300
301 bool Done = true;
302 unsigned MaxSuccHeight = 0;
303 for (const SDep &SuccDep : Cur->Succs) {
304 SUnit *SuccSU = SuccDep.getSUnit();
305 if (SuccSU->isHeightCurrent)
306 MaxSuccHeight = std::max(MaxSuccHeight,
307 SuccSU->Height + SuccDep.getLatency());
308 else {
309 Done = false;
310 WorkList.push_back(SuccSU);
311 }
312 }
313
314 if (Done) {
315 WorkList.pop_back();
316 if (MaxSuccHeight != Cur->Height) {
317 Cur->setHeightDirty();
318 Cur->Height = MaxSuccHeight;
319 }
320 Cur->isHeightCurrent = true;
321 }
322 } while (!WorkList.empty());
323}
324
326 if (NumPreds < 2)
327 return;
328
329 SUnit::pred_iterator BestI = Preds.begin();
330 unsigned MaxDepth = BestI->getSUnit()->getDepth();
331 for (SUnit::pred_iterator I = std::next(BestI), E = Preds.end(); I != E;
332 ++I) {
333 if (I->getKind() == SDep::Data && I->getSUnit()->getDepth() > MaxDepth) {
334 MaxDepth = I->getSUnit()->getDepth();
335 BestI = I;
336 }
337 }
338 if (BestI != Preds.begin())
339 std::swap(*Preds.begin(), *BestI);
340}
341
342#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP)
344 dbgs() << " # preds left : " << NumPredsLeft << "\n";
345 dbgs() << " # succs left : " << NumSuccsLeft << "\n";
346 if (WeakPredsLeft)
347 dbgs() << " # weak preds left : " << WeakPredsLeft << "\n";
348 if (WeakSuccsLeft)
349 dbgs() << " # weak succs left : " << WeakSuccsLeft << "\n";
350 dbgs() << " # rdefs left : " << NumRegDefsLeft << "\n";
351 dbgs() << " Latency : " << Latency << "\n";
352 dbgs() << " Depth : " << getDepth() << "\n";
353 dbgs() << " Height : " << getHeight() << "\n";
354}
355
357 if (&SU == &EntrySU)
358 dbgs() << "EntrySU";
359 else if (&SU == &ExitSU)
360 dbgs() << "ExitSU";
361 else
362 dbgs() << "SU(" << SU.NodeNum << ")";
363}
364
366 dumpNode(SU);
367 SU.dumpAttributes();
368 if (SU.Preds.size() > 0) {
369 dbgs() << " Predecessors:\n";
370 for (const SDep &Dep : SU.Preds) {
371 dbgs() << " ";
372 dumpNodeName(*Dep.getSUnit());
373 dbgs() << ": ";
374 Dep.dump(TRI);
375 dbgs() << '\n';
376 }
377 }
378 if (SU.Succs.size() > 0) {
379 dbgs() << " Successors:\n";
380 for (const SDep &Dep : SU.Succs) {
381 dbgs() << " ";
382 dumpNodeName(*Dep.getSUnit());
383 dbgs() << ": ";
384 Dep.dump(TRI);
385 dbgs() << '\n';
386 }
387 }
388}
389#endif
390
391#ifndef NDEBUG
392unsigned ScheduleDAG::VerifyScheduledDAG(bool isBottomUp) {
393 bool AnyNotSched = false;
394 unsigned DeadNodes = 0;
395 for (const SUnit &SUnit : SUnits) {
396 if (!SUnit.isScheduled) {
397 if (SUnit.NumPreds == 0 && SUnit.NumSuccs == 0) {
398 ++DeadNodes;
399 continue;
400 }
401 if (!AnyNotSched)
402 dbgs() << "*** Scheduling failed! ***\n";
404 dbgs() << "has not been scheduled!\n";
405 AnyNotSched = true;
406 }
407 if (SUnit.isScheduled &&
408 (isBottomUp ? SUnit.getHeight() : SUnit.getDepth()) >
409 unsigned(std::numeric_limits<int>::max())) {
410 if (!AnyNotSched)
411 dbgs() << "*** Scheduling failed! ***\n";
413 dbgs() << "has an unexpected "
414 << (isBottomUp ? "Height" : "Depth") << " value!\n";
415 AnyNotSched = true;
416 }
417 if (isBottomUp) {
418 if (SUnit.NumSuccsLeft != 0) {
419 if (!AnyNotSched)
420 dbgs() << "*** Scheduling failed! ***\n";
422 dbgs() << "has successors left!\n";
423 AnyNotSched = true;
424 }
425 } else {
426 if (SUnit.NumPredsLeft != 0) {
427 if (!AnyNotSched)
428 dbgs() << "*** Scheduling failed! ***\n";
430 dbgs() << "has predecessors left!\n";
431 AnyNotSched = true;
432 }
433 }
434 }
435 assert(!AnyNotSched);
436 return SUnits.size() - DeadNodes;
437}
438#endif
439
441 // The idea of the algorithm is taken from
442 // "Online algorithms for managing the topological order of
443 // a directed acyclic graph" by David J. Pearce and Paul H.J. Kelly
444 // This is the MNR algorithm, which was first introduced by
445 // A. Marchetti-Spaccamela, U. Nanni and H. Rohnert in
446 // "Maintaining a topological order under edge insertions".
447 //
448 // Short description of the algorithm:
449 //
450 // Topological ordering, ord, of a DAG maps each node to a topological
451 // index so that for all edges X->Y it is the case that ord(X) < ord(Y).
452 //
453 // This means that if there is a path from the node X to the node Z,
454 // then ord(X) < ord(Z).
455 //
456 // This property can be used to check for reachability of nodes:
457 // if Z is reachable from X, then an insertion of the edge Z->X would
458 // create a cycle.
459 //
460 // The algorithm first computes a topological ordering for the DAG by
461 // initializing the Index2Node and Node2Index arrays and then tries to keep
462 // the ordering up-to-date after edge insertions by reordering the DAG.
463 //
464 // On insertion of the edge X->Y, the algorithm first marks by calling DFS
465 // the nodes reachable from Y, and then shifts them using Shift to lie
466 // immediately after X in Index2Node.
467
468 // Cancel pending updates, mark as valid.
469 Dirty = false;
470 Updates.clear();
471
472 unsigned DAGSize = SUnits.size();
473 std::vector<SUnit*> WorkList;
474 WorkList.reserve(DAGSize);
475
476 Index2Node.resize(DAGSize);
477 Node2Index.resize(DAGSize);
478
479 // Initialize the data structures.
480 if (ExitSU)
481 WorkList.push_back(ExitSU);
482 for (SUnit &SU : SUnits) {
483 int NodeNum = SU.NodeNum;
484 unsigned Degree = SU.Succs.size();
485 // Temporarily use the Node2Index array as scratch space for degree counts.
486 Node2Index[NodeNum] = Degree;
487
488 // Is it a node without dependencies?
489 if (Degree == 0) {
490 assert(SU.Succs.empty() && "SUnit should have no successors");
491 // Collect leaf nodes.
492 WorkList.push_back(&SU);
493 }
494 }
495
496 int Id = DAGSize;
497 while (!WorkList.empty()) {
498 SUnit *SU = WorkList.back();
499 WorkList.pop_back();
500 if (SU->NodeNum < DAGSize)
501 Allocate(SU->NodeNum, --Id);
502 for (const SDep &PredDep : SU->Preds) {
503 SUnit *SU = PredDep.getSUnit();
504 if (SU->NodeNum < DAGSize && !--Node2Index[SU->NodeNum])
505 // If all dependencies of the node are processed already,
506 // then the node can be computed now.
507 WorkList.push_back(SU);
508 }
509 }
510
511 Visited.resize(DAGSize);
512 NumTopoInits++;
513
514#ifndef NDEBUG
515 // Check correctness of the ordering
516 for (SUnit &SU : SUnits) {
517 for (const SDep &PD : SU.Preds) {
518 assert(Node2Index[SU.NodeNum] > Node2Index[PD.getSUnit()->NodeNum] &&
519 "Wrong topological sorting");
520 }
521 }
522#endif
523}
524
525void ScheduleDAGTopologicalSort::FixOrder() {
526 // Recompute from scratch after new nodes have been added.
527 if (Dirty) {
529 return;
530 }
531
532 // Otherwise apply updates one-by-one.
533 for (auto &U : Updates)
534 AddPred(U.first, U.second);
535 Updates.clear();
536}
537
539 // Recomputing the order from scratch is likely more efficient than applying
540 // updates one-by-one for too many updates. The current cut-off is arbitrarily
541 // chosen.
542 Dirty = Dirty || Updates.size() > 10;
543
544 if (Dirty)
545 return;
546
547 Updates.emplace_back(Y, X);
548}
549
551 int UpperBound, LowerBound;
552 LowerBound = Node2Index[Y->NodeNum];
553 UpperBound = Node2Index[X->NodeNum];
554 bool HasLoop = false;
555 // Is Ord(X) < Ord(Y) ?
556 if (LowerBound < UpperBound) {
557 // Update the topological order.
558 Visited.reset();
559 DFS(Y, UpperBound, HasLoop);
560 assert(!HasLoop && "Inserted edge creates a loop!");
561 // Recompute topological indexes.
562 Shift(Visited, LowerBound, UpperBound);
563 }
564
565 NumNewPredsAdded++;
566}
567
569 // InitDAGTopologicalSorting();
570}
571
572void ScheduleDAGTopologicalSort::DFS(const SUnit *SU, int UpperBound,
573 bool &HasLoop) {
574 std::vector<const SUnit*> WorkList;
575 WorkList.reserve(SUnits.size());
576
577 WorkList.push_back(SU);
578 do {
579 SU = WorkList.back();
580 WorkList.pop_back();
581 Visited.set(SU->NodeNum);
582 for (const SDep &SuccDep : llvm::reverse(SU->Succs)) {
583 unsigned s = SuccDep.getSUnit()->NodeNum;
584 // Edges to non-SUnits are allowed but ignored (e.g. ExitSU).
585 if (s >= Node2Index.size())
586 continue;
587 if (Node2Index[s] == UpperBound) {
588 HasLoop = true;
589 return;
590 }
591 // Visit successors if not already and in affected region.
592 if (!Visited.test(s) && Node2Index[s] < UpperBound) {
593 WorkList.push_back(SuccDep.getSUnit());
594 }
595 }
596 } while (!WorkList.empty());
597}
598
599std::vector<int> ScheduleDAGTopologicalSort::GetSubGraph(const SUnit &StartSU,
600 const SUnit &TargetSU,
601 bool &Success) {
602 std::vector<const SUnit*> WorkList;
603 int LowerBound = Node2Index[StartSU.NodeNum];
604 int UpperBound = Node2Index[TargetSU.NodeNum];
605 bool Found = false;
606 BitVector VisitedBack;
607 std::vector<int> Nodes;
608
609 if (LowerBound > UpperBound) {
610 Success = false;
611 return Nodes;
612 }
613
614 WorkList.reserve(SUnits.size());
615 Visited.reset();
616
617 // Starting from StartSU, visit all successors up
618 // to UpperBound.
619 WorkList.push_back(&StartSU);
620 do {
621 const SUnit *SU = WorkList.back();
622 WorkList.pop_back();
623 for (const SDep &SD : llvm::reverse(SU->Succs)) {
624 const SUnit *Succ = SD.getSUnit();
625 unsigned s = Succ->NodeNum;
626 // Edges to non-SUnits are allowed but ignored (e.g. ExitSU).
627 if (Succ->isBoundaryNode())
628 continue;
629 if (Node2Index[s] == UpperBound) {
630 Found = true;
631 continue;
632 }
633 // Visit successors if not already and in affected region.
634 if (!Visited.test(s) && Node2Index[s] < UpperBound) {
635 Visited.set(s);
636 WorkList.push_back(Succ);
637 }
638 }
639 } while (!WorkList.empty());
640
641 if (!Found) {
642 Success = false;
643 return Nodes;
644 }
645
646 WorkList.clear();
647 VisitedBack.resize(SUnits.size());
648 Found = false;
649
650 // Starting from TargetSU, visit all predecessors up
651 // to LowerBound. SUs that are visited by the two
652 // passes are added to Nodes.
653 WorkList.push_back(&TargetSU);
654 do {
655 const SUnit *SU = WorkList.back();
656 WorkList.pop_back();
657 for (const SDep &SD : llvm::reverse(SU->Preds)) {
658 const SUnit *Pred = SD.getSUnit();
659 unsigned s = Pred->NodeNum;
660 // Edges to non-SUnits are allowed but ignored (e.g. EntrySU).
661 if (Pred->isBoundaryNode())
662 continue;
663 if (Node2Index[s] == LowerBound) {
664 Found = true;
665 continue;
666 }
667 if (!VisitedBack.test(s) && Visited.test(s)) {
668 VisitedBack.set(s);
669 WorkList.push_back(Pred);
670 Nodes.push_back(s);
671 }
672 }
673 } while (!WorkList.empty());
674
675 assert(Found && "Error in SUnit Graph!");
676 Success = true;
677 return Nodes;
678}
679
680void ScheduleDAGTopologicalSort::Shift(BitVector& Visited, int LowerBound,
681 int UpperBound) {
682 std::vector<int> L;
683 int shift = 0;
684 int i;
685
686 for (i = LowerBound; i <= UpperBound; ++i) {
687 // w is node at topological index i.
688 int w = Index2Node[i];
689 if (Visited.test(w)) {
690 // Unmark.
691 Visited.reset(w);
692 L.push_back(w);
693 shift = shift + 1;
694 } else {
695 Allocate(w, i - shift);
696 }
697 }
698
699 for (unsigned LI : L) {
700 Allocate(LI, i - shift);
701 i = i + 1;
702 }
703}
704
706 FixOrder();
707 // Is SU reachable from TargetSU via successor edges?
708 if (IsReachable(SU, TargetSU))
709 return true;
710 for (const SDep &PredDep : TargetSU->Preds)
711 if (PredDep.isAssignedRegDep() &&
712 IsReachable(SU, PredDep.getSUnit()))
713 return true;
714 return false;
715}
716
718 assert(SU->NodeNum == Index2Node.size() && "Node cannot be added at the end");
719 assert(SU->NumPreds == 0 && "Can only add SU's with no predecessors");
720 Node2Index.push_back(Index2Node.size());
721 Index2Node.push_back(SU->NodeNum);
722 Visited.resize(Node2Index.size());
723}
724
726 const SUnit *TargetSU) {
727 assert(TargetSU != nullptr && "Invalid target SUnit");
728 assert(SU != nullptr && "Invalid SUnit");
729 FixOrder();
730 // If insertion of the edge SU->TargetSU would create a cycle
731 // then there is a path from TargetSU to SU.
732 int UpperBound, LowerBound;
733 LowerBound = Node2Index[TargetSU->NodeNum];
734 UpperBound = Node2Index[SU->NodeNum];
735 bool HasLoop = false;
736 // Is Ord(TargetSU) < Ord(SU) ?
737 if (LowerBound < UpperBound) {
738 Visited.reset();
739 // There may be a path from TargetSU to SU. Check for it.
740 DFS(TargetSU, UpperBound, HasLoop);
741 }
742 return HasLoop;
743}
744
745void ScheduleDAGTopologicalSort::Allocate(int n, int index) {
746 Node2Index[n] = index;
747 Index2Node[index] = n;
748}
749
751ScheduleDAGTopologicalSort(std::vector<SUnit> &sunits, SUnit *exitsu)
752 : SUnits(sunits), ExitSU(exitsu) {}
753
unsigned const MachineRegisterInfo * MRI
#define Success
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
#define LLVM_DUMP_METHOD
Mark debug helper function definitions like dump() that should not be stripped from debug builds.
Definition: Compiler.h:533
static GCMetadataPrinterRegistry::Add< ErlangGCPrinter > X("erlang", "erlang-compatible garbage collector")
const HexagonInstrInfo * TII
static const unsigned MaxDepth
#define I(x, y, z)
Definition: MD5.cpp:58
unsigned const TargetRegisterInfo * TRI
static GCMetadataPrinterRegistry::Add< OcamlGCMetadataPrinter > Y("ocaml", "ocaml 3.10-compatible collector")
#define P(N)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
This file contains some templates that are useful if you are working with the STL at all.
static cl::opt< bool > StressSchedOpt("stress-sched", cl::Hidden, cl::init(false), cl::desc("Stress test instruction scheduling"))
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition: Statistic.h:166
bool test(unsigned Idx) const
Definition: BitVector.h:461
BitVector & reset()
Definition: BitVector.h:392
void resize(unsigned N, bool t=false)
resize - Grow or shrink the bitvector.
Definition: BitVector.h:341
BitVector & set()
Definition: BitVector.h:351
void push_back(bool Val)
Definition: BitVector.h:466
Describe properties that are true of each instruction in the target description file.
Definition: MCInstrDesc.h:198
const MCInstrDesc & get(unsigned Opcode) const
Return the machine instruction descriptor that corresponds to the specified instruction opcode.
Definition: MCInstrInfo.h:63
Represents one node in the SelectionDAG.
Scheduling dependency.
Definition: ScheduleDAG.h:49
SUnit * getSUnit() const
Definition: ScheduleDAG.h:498
Kind getKind() const
Returns an enum value representing the kind of the dependence.
Definition: ScheduleDAG.h:504
@ Output
A register output-dependence (aka WAW).
Definition: ScheduleDAG.h:55
@ Order
Any other ordering dependency.
Definition: ScheduleDAG.h:56
@ Anti
A register anti-dependence (aka WAR).
Definition: ScheduleDAG.h:54
@ Data
Regular data dependence (aka true-dependence).
Definition: ScheduleDAG.h:53
void setLatency(unsigned Lat)
Sets the latency for this edge.
Definition: ScheduleDAG.h:147
@ Cluster
Weak DAG edge linking a chain of clustered instrs.
Definition: ScheduleDAG.h:74
@ Barrier
An unknown scheduling barrier.
Definition: ScheduleDAG.h:69
@ Artificial
Arbitrary strong DAG edge (no real dependence).
Definition: ScheduleDAG.h:72
@ MayAliasMem
Nonvolatile load/Store instructions that may alias.
Definition: ScheduleDAG.h:70
@ Weak
Arbitrary weak DAG edge.
Definition: ScheduleDAG.h:73
@ MustAliasMem
Nonvolatile load/Store instructions that must alias.
Definition: ScheduleDAG.h:71
unsigned getLatency() const
Returns the latency value for this edge, which roughly means the minimum number of cycles that must e...
Definition: ScheduleDAG.h:142
bool isAssignedRegDep() const
Tests if this is a Data dependence that is associated with a register.
Definition: ScheduleDAG.h:211
unsigned getReg() const
Returns the register associated with this edge.
Definition: ScheduleDAG.h:218
void setSUnit(SUnit *SU)
Definition: ScheduleDAG.h:501
void dump(const TargetRegisterInfo *TRI=nullptr) const
Definition: ScheduleDAG.cpp:74
Scheduling unit. This is a node in the scheduling DAG.
Definition: ScheduleDAG.h:242
void setHeightToAtLeast(unsigned NewHeight)
If NewHeight is greater than this node's height value, set it to be the new height value.
unsigned NumSuccs
Definition: ScheduleDAG.h:273
unsigned NumPreds
Definition: ScheduleDAG.h:272
unsigned NodeNum
Entry # of node in the node vector.
Definition: ScheduleDAG.h:270
unsigned NumSuccsLeft
Definition: ScheduleDAG.h:275
void biasCriticalPath()
Orders this node's predecessor edges such that the critical path edge occurs first.
unsigned getHeight() const
Returns the height of this node, which is the length of the maximum path down to any node which has n...
Definition: ScheduleDAG.h:424
void setHeightDirty()
Sets a flag in this node to indicate that its stored Height value will require recomputation the next...
void removePred(const SDep &D)
Removes the specified edge as a pred of the current node if it exists.
unsigned short Latency
Node latency.
Definition: ScheduleDAG.h:303
SmallVectorImpl< SDep >::iterator pred_iterator
Definition: ScheduleDAG.h:265
bool isBoundaryNode() const
Boundary nodes are placeholders for the boundary of the scheduling region.
Definition: ScheduleDAG.h:358
unsigned short NumRegDefsLeft
Definition: ScheduleDAG.h:302
unsigned getDepth() const
Returns the depth of this node, which is the length of the maximum path up to any node which has no p...
Definition: ScheduleDAG.h:416
bool isScheduled
True once scheduled.
Definition: ScheduleDAG.h:296
unsigned NumPredsLeft
Definition: ScheduleDAG.h:274
void dumpAttributes() const
SmallVector< SDep, 4 > Succs
All sunit successors.
Definition: ScheduleDAG.h:263
unsigned WeakPredsLeft
Definition: ScheduleDAG.h:276
void setDepthDirty()
Sets a flag in this node to indicate that its stored Depth value will require recomputation the next ...
SmallVector< SDep, 4 > Preds
All sunit predecessors.
Definition: ScheduleDAG.h:262
unsigned WeakSuccsLeft
Definition: ScheduleDAG.h:277
void setDepthToAtLeast(unsigned NewDepth)
If NewDepth is greater than this node's depth value, sets it to be the new depth value.
bool addPred(const SDep &D, bool Required=true)
Adds the specified edge as a pred of the current node if not already.
void RemovePred(SUnit *M, SUnit *N)
Updates the topological ordering to accommodate an edge to be removed from the specified node N from ...
bool WillCreateCycle(SUnit *TargetSU, SUnit *SU)
Returns true if addPred(TargetSU, SU) creates a cycle.
void AddSUnitWithoutPredecessors(const SUnit *SU)
Add a SUnit without predecessors to the end of the topological order.
ScheduleDAGTopologicalSort(std::vector< SUnit > &SUnits, SUnit *ExitSU)
std::vector< int > GetSubGraph(const SUnit &StartSU, const SUnit &TargetSU, bool &Success)
Returns an array of SUs that are both in the successor subtree of StartSU and in the predecessor subt...
void InitDAGTopologicalSorting()
Creates the initial topological ordering from the DAG to be scheduled.
void AddPred(SUnit *Y, SUnit *X)
Updates the topological ordering to accommodate an edge to be added from SUnit X to SUnit Y.
bool IsReachable(const SUnit *SU, const SUnit *TargetSU)
Checks if SU is reachable from TargetSU.
void AddPredQueued(SUnit *Y, SUnit *X)
Queues an update to the topological ordering to accommodate an edge to be added from SUnit X to SUnit...
void clearDAG()
Clears the DAG state (between regions).
Definition: ScheduleDAG.cpp:63
const TargetInstrInfo * TII
Target instruction information.
Definition: ScheduleDAG.h:575
std::vector< SUnit > SUnits
The scheduling units.
Definition: ScheduleDAG.h:579
virtual ~ScheduleDAG()
const TargetRegisterInfo * TRI
Target processor register info.
Definition: ScheduleDAG.h:576
SUnit EntrySU
Special node for the region entry.
Definition: ScheduleDAG.h:580
ScheduleDAG(const ScheduleDAG &)=delete
void dumpNodeAll(const SUnit &SU) const
unsigned VerifyScheduledDAG(bool isBottomUp)
Verifies that all SUnits were scheduled and that their state is consistent.
virtual void dumpNode(const SUnit &SU) const =0
void dumpNodeName(const SUnit &SU) const
SUnit ExitSU
Special node for the region exit.
Definition: ScheduleDAG.h:581
bool empty() const
Definition: SmallVector.h:94
void reserve(size_type N)
Definition: SmallVector.h:676
typename SuperClass::iterator iterator
Definition: SmallVector.h:590
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
initializer< Ty > init(const Ty &Val)
Definition: CommandLine.h:443
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
auto find(R &&Range, const T &Val)
Provide wrappers to std::find which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1742
@ Done
Definition: Threading.h:61
auto reverse(ContainerTy &&C)
Definition: STLExtras.h:419
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition: BitVector.h:860
#define N