35#define DEBUG_TYPE "si-memory-legalizer"
36#define PASS_NAME "SI Memory Legalizer"
40 cl::desc(
"Use this to skip inserting cache invalidating instructions."));
62enum class SIAtomicScope {
74enum class SIAtomicAddrSpace {
83 FLAT = GLOBAL |
LDS | SCRATCH,
86 ATOMIC = GLOBAL |
LDS | SCRATCH | GDS,
89 ALL = GLOBAL |
LDS | SCRATCH | GDS | OTHER,
94class SIMemOpInfo final {
97 friend class SIMemOpAccess;
101 SIAtomicScope Scope = SIAtomicScope::SYSTEM;
102 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
103 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
104 bool IsCrossAddressSpaceOrdering =
false;
105 bool IsVolatile =
false;
106 bool IsNonTemporal =
false;
107 bool IsLastUse =
false;
108 bool IsCooperative =
false;
112 const GCNSubtarget &ST,
114 SIAtomicScope Scope = SIAtomicScope::SYSTEM,
115 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC,
116 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL,
117 bool IsCrossAddressSpaceOrdering =
true,
118 AtomicOrdering FailureOrdering = AtomicOrdering::SequentiallyConsistent,
119 bool IsVolatile =
false,
bool IsNonTemporal =
false,
120 bool IsLastUse =
false,
bool IsCooperative =
false)
121 : Ordering(Ordering), FailureOrdering(FailureOrdering), Scope(Scope),
122 OrderingAddrSpace(OrderingAddrSpace), InstrAddrSpace(InstrAddrSpace),
123 IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering),
124 IsVolatile(IsVolatile), IsNonTemporal(IsNonTemporal),
125 IsLastUse(IsLastUse), IsCooperative(IsCooperative) {
127 if (Ordering == AtomicOrdering::NotAtomic) {
128 assert(!IsCooperative &&
"Cannot be cooperative & non-atomic!");
129 assert(Scope == SIAtomicScope::NONE &&
130 OrderingAddrSpace == SIAtomicAddrSpace::NONE &&
131 !IsCrossAddressSpaceOrdering &&
132 FailureOrdering == AtomicOrdering::NotAtomic);
136 assert(Scope != SIAtomicScope::NONE &&
137 (OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
138 SIAtomicAddrSpace::NONE &&
139 (InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
140 SIAtomicAddrSpace::NONE);
145 if ((OrderingAddrSpace == InstrAddrSpace) &&
147 this->IsCrossAddressSpaceOrdering =
false;
151 if ((InstrAddrSpace & ~SIAtomicAddrSpace::SCRATCH) ==
152 SIAtomicAddrSpace::NONE) {
153 this->Scope = std::min(Scope, SIAtomicScope::SINGLETHREAD);
154 }
else if ((InstrAddrSpace &
155 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS)) ==
156 SIAtomicAddrSpace::NONE) {
157 this->Scope = std::min(Scope, SIAtomicScope::WORKGROUP);
158 }
else if ((InstrAddrSpace &
159 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS |
160 SIAtomicAddrSpace::GDS)) == SIAtomicAddrSpace::NONE) {
161 this->Scope = std::min(Scope, SIAtomicScope::AGENT);
166 if (this->Scope == SIAtomicScope::CLUSTER && !
ST.hasClusters())
167 this->Scope = SIAtomicScope::AGENT;
173 SIAtomicScope getScope()
const {
186 return FailureOrdering;
191 SIAtomicAddrSpace getInstrAddrSpace()
const {
192 return InstrAddrSpace;
197 SIAtomicAddrSpace getOrderingAddrSpace()
const {
198 return OrderingAddrSpace;
203 bool getIsCrossAddressSpaceOrdering()
const {
204 return IsCrossAddressSpaceOrdering;
209 bool isVolatile()
const {
215 bool isNonTemporal()
const {
216 return IsNonTemporal;
221 bool isLastUse()
const {
return IsLastUse; }
224 bool isCooperative()
const {
return IsCooperative; }
228 bool isAtomic()
const {
229 return Ordering != AtomicOrdering::NotAtomic;
234class SIMemOpAccess final {
236 const AMDGPUMachineModuleInfo *MMI =
nullptr;
237 const GCNSubtarget &ST;
241 const char *Msg)
const;
247 std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
248 toSIAtomicScope(
SyncScope::ID SSID, SIAtomicAddrSpace InstrAddrSpace)
const;
251 SIAtomicAddrSpace toSIAtomicAddrSpace(
unsigned AS)
const;
255 std::optional<SIMemOpInfo>
261 SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI,
const GCNSubtarget &ST);
264 std::optional<SIMemOpInfo>
269 std::optional<SIMemOpInfo>
274 std::optional<SIMemOpInfo>
279 std::optional<SIMemOpInfo>
285 std::optional<SIMemOpInfo>
289class SICacheControl {
293 const GCNSubtarget &ST;
296 const SIInstrInfo *TII =
nullptr;
303 SICacheControl(
const GCNSubtarget &ST);
308 unsigned Bits)
const;
312 bool canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const;
318 static std::unique_ptr<SICacheControl> create(
const GCNSubtarget &ST);
325 SIAtomicAddrSpace AddrSpace)
const = 0;
332 SIAtomicAddrSpace AddrSpace)
const = 0;
339 SIAtomicAddrSpace AddrSpace)
const = 0;
345 SIAtomicAddrSpace AddrSpace,
346 SIMemOp
Op,
bool IsVolatile,
348 bool IsLastUse =
false)
const = 0;
355 virtual bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
360 virtual bool handleCooperativeAtomic(MachineInstr &
MI)
const {
362 "cooperative atomics are not available on this architecture");
375 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
376 bool IsCrossAddrSpaceOrdering, Position Pos,
386 SIAtomicAddrSpace AddrSpace,
387 Position Pos)
const = 0;
397 SIAtomicAddrSpace AddrSpace,
398 bool IsCrossAddrSpaceOrdering,
399 Position Pos)
const = 0;
402 virtual ~SICacheControl() =
default;
407class SIGfx6CacheControl final :
public SICacheControl {
410 SIGfx6CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
414 SIAtomicAddrSpace AddrSpace)
const override;
418 SIAtomicAddrSpace AddrSpace)
const override;
422 SIAtomicAddrSpace AddrSpace)
const override;
425 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
426 bool IsVolatile,
bool IsNonTemporal,
427 bool IsLastUse)
const override;
430 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
431 bool IsCrossAddrSpaceOrdering, Position Pos,
436 SIAtomicAddrSpace AddrSpace,
437 Position Pos)
const override;
441 SIAtomicAddrSpace AddrSpace,
442 bool IsCrossAddrSpaceOrdering,
443 Position Pos)
const override;
447class SIGfx10CacheControl final :
public SICacheControl {
449 SIGfx10CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
453 SIAtomicAddrSpace AddrSpace)
const override;
457 SIAtomicAddrSpace AddrSpace)
const override {
463 SIAtomicAddrSpace AddrSpace)
const override {
468 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
469 bool IsVolatile,
bool IsNonTemporal,
470 bool IsLastUse)
const override;
473 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
474 bool IsCrossAddrSpaceOrdering, Position Pos,
478 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
481 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
482 Position Pos)
const override {
483 return insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
484 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
489class SIGfx12CacheControl final :
public SICacheControl {
511 SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace)
const;
514 SIGfx12CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {
517 assert(!
ST.hasGFX1250Insts() ||
ST.isCuModeEnabled());
521 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
522 bool IsCrossAddrSpaceOrdering, Position Pos,
526 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
529 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
530 bool IsVolatile,
bool IsNonTemporal,
531 bool IsLastUse)
const override;
533 bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const override;
535 bool handleCooperativeAtomic(MachineInstr &
MI)
const override;
538 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
539 Position Pos)
const override;
543 SIAtomicAddrSpace AddrSpace)
const override {
544 return setAtomicScope(
MI, Scope, AddrSpace);
549 SIAtomicAddrSpace AddrSpace)
const override {
550 return setAtomicScope(
MI, Scope, AddrSpace);
555 SIAtomicAddrSpace AddrSpace)
const override {
556 return setAtomicScope(
MI, Scope, AddrSpace);
560class SIMemoryLegalizer final {
562 const MachineModuleInfo &MMI;
564 std::unique_ptr<SICacheControl> CC =
nullptr;
567 std::list<MachineBasicBlock::iterator> AtomicPseudoMIs;
571 bool isAtomicRet(
const MachineInstr &
MI)
const {
577 bool removeAtomicPseudoMIs();
581 bool expandLoad(
const SIMemOpInfo &MOI,
585 bool expandStore(
const SIMemOpInfo &MOI,
589 bool expandAtomicFence(
const SIMemOpInfo &MOI,
593 bool expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
600 SIMemoryLegalizer(
const MachineModuleInfo &MMI) : MMI(MMI) {};
601 bool run(MachineFunction &MF);
608 SIMemoryLegalizerLegacy() : MachineFunctionPass(ID) {}
610 void getAnalysisUsage(AnalysisUsage &AU)
const override {
615 StringRef getPassName()
const override {
619 bool runOnMachineFunction(MachineFunction &MF)
override;
623 {
"global", SIAtomicAddrSpace::GLOBAL},
624 {
"local", SIAtomicAddrSpace::LDS},
632 OS <<
"unknown address space '" << AS <<
"'; expected one of ";
634 for (
const auto &[Name, Val] : ASNames)
635 OS <<
LS <<
'\'' <<
Name <<
'\'';
643static std::optional<SIAtomicAddrSpace>
645 static constexpr StringLiteral FenceASPrefix =
"amdgpu-synchronize-as";
651 SIAtomicAddrSpace
Result = SIAtomicAddrSpace::NONE;
652 for (
const auto &[Prefix, Suffix] : MMRA) {
653 if (Prefix != FenceASPrefix)
656 if (
auto It = ASNames.find(Suffix); It != ASNames.end())
659 diagnoseUnknownMMRAASName(
MI, Suffix);
662 if (Result == SIAtomicAddrSpace::NONE)
671 const char *Msg)
const {
673 Func.getContext().diagnose(
674 DiagnosticInfoUnsupported(Func, Msg,
MI->getDebugLoc()));
677std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
679 SIAtomicAddrSpace InstrAddrSpace)
const {
681 return std::tuple(SIAtomicScope::SYSTEM, SIAtomicAddrSpace::ATOMIC,
true);
683 return std::tuple(SIAtomicScope::AGENT, SIAtomicAddrSpace::ATOMIC,
true);
685 return std::tuple(SIAtomicScope::CLUSTER, SIAtomicAddrSpace::ATOMIC,
true);
687 return std::tuple(SIAtomicScope::WORKGROUP, SIAtomicAddrSpace::ATOMIC,
690 return std::tuple(SIAtomicScope::WAVEFRONT, SIAtomicAddrSpace::ATOMIC,
693 return std::tuple(SIAtomicScope::SINGLETHREAD, SIAtomicAddrSpace::ATOMIC,
696 return std::tuple(SIAtomicScope::SYSTEM,
697 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
699 return std::tuple(SIAtomicScope::AGENT,
700 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
702 return std::tuple(SIAtomicScope::CLUSTER,
703 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
705 return std::tuple(SIAtomicScope::WORKGROUP,
706 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
708 return std::tuple(SIAtomicScope::WAVEFRONT,
709 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
711 return std::tuple(SIAtomicScope::SINGLETHREAD,
712 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
716SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(
unsigned AS)
const {
718 return SIAtomicAddrSpace::FLAT;
720 return SIAtomicAddrSpace::GLOBAL;
722 return SIAtomicAddrSpace::LDS;
724 return SIAtomicAddrSpace::SCRATCH;
726 return SIAtomicAddrSpace::GDS;
729 return SIAtomicAddrSpace::GLOBAL;
731 return SIAtomicAddrSpace::OTHER;
734SIMemOpAccess::SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI_,
735 const GCNSubtarget &ST)
736 : MMI(&MMI_),
ST(
ST) {}
738std::optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO(
740 assert(
MI->getNumMemOperands() > 0);
745 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
746 bool IsNonTemporal =
true;
748 bool IsLastUse =
false;
749 bool IsCooperative =
false;
753 for (
const auto &MMO :
MI->memoperands()) {
754 IsNonTemporal &= MMO->isNonTemporal();
756 IsLastUse |= MMO->getFlags() &
MOLastUse;
759 toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace());
761 if (OpOrdering != AtomicOrdering::NotAtomic) {
762 const auto &IsSyncScopeInclusion =
764 if (!IsSyncScopeInclusion) {
765 reportUnsupported(
MI,
766 "Unsupported non-inclusive atomic synchronization scope");
770 SSID = *IsSyncScopeInclusion ? SSID : MMO->getSyncScopeID();
772 assert(MMO->getFailureOrdering() != AtomicOrdering::Release &&
773 MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease);
784 Ordering = AtomicOrdering::Monotonic;
786 SIAtomicScope
Scope = SIAtomicScope::NONE;
787 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
788 bool IsCrossAddressSpaceOrdering =
false;
789 if (Ordering != AtomicOrdering::NotAtomic) {
790 auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace);
792 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
795 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
797 if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) ||
798 ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace) ||
799 ((InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) == SIAtomicAddrSpace::NONE)) {
800 reportUnsupported(
MI,
"Unsupported atomic address space");
804 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace, InstrAddrSpace,
805 IsCrossAddressSpaceOrdering, FailureOrdering, IsVolatile,
806 IsNonTemporal, IsLastUse, IsCooperative);
809std::optional<SIMemOpInfo>
813 if (!(
MI->mayLoad() && !
MI->mayStore()))
817 if (
MI->getNumMemOperands() == 0)
818 return SIMemOpInfo(ST);
820 return constructFromMIWithMMO(
MI);
823std::optional<SIMemOpInfo>
827 if (!(!
MI->mayLoad() &&
MI->mayStore()))
831 if (
MI->getNumMemOperands() == 0)
832 return SIMemOpInfo(ST);
834 return constructFromMIWithMMO(
MI);
837std::optional<SIMemOpInfo>
841 if (
MI->getOpcode() != AMDGPU::ATOMIC_FENCE)
848 auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC);
850 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
854 SIAtomicScope
Scope = SIAtomicScope::NONE;
855 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
856 bool IsCrossAddressSpaceOrdering =
false;
857 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
860 if (OrderingAddrSpace != SIAtomicAddrSpace::ATOMIC) {
865 reportUnsupported(
MI,
"Unsupported atomic address space");
869 auto SynchronizeAS = getSynchronizeAddrSpaceMD(*
MI);
871 OrderingAddrSpace = *SynchronizeAS;
873 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace,
874 SIAtomicAddrSpace::ATOMIC, IsCrossAddressSpaceOrdering,
875 AtomicOrdering::NotAtomic);
878std::optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo(
882 if (!(
MI->mayLoad() &&
MI->mayStore()))
886 if (
MI->getNumMemOperands() == 0)
887 return SIMemOpInfo(ST);
889 return constructFromMIWithMMO(
MI);
892std::optional<SIMemOpInfo>
899 return constructFromMIWithMMO(
MI);
902SICacheControl::SICacheControl(
const GCNSubtarget &ST) :
ST(
ST) {
903 TII =
ST.getInstrInfo();
909 unsigned Bits)
const {
910 MachineOperand *CPol =
TII->getNamedOperand(*
MI, AMDGPU::OpName::cpol);
914 CPol->setImm(
CPol->getImm() | Bits);
918bool SICacheControl::canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const {
919 assert((!
ST.hasGloballyAddressableScratch() ||
920 (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE ||
921 (AS & SIAtomicAddrSpace::SCRATCH) == SIAtomicAddrSpace::NONE) &&
922 "scratch instructions should already be replaced by flat "
923 "instructions if GloballyAddressableScratch is enabled");
924 return (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE;
928std::unique_ptr<SICacheControl> SICacheControl::create(
const GCNSubtarget &ST) {
929 GCNSubtarget::Generation Generation =
ST.getGeneration();
930 if (Generation < AMDGPUSubtarget::GFX10)
931 return std::make_unique<SIGfx6CacheControl>(ST);
932 if (Generation < AMDGPUSubtarget::GFX12)
933 return std::make_unique<SIGfx10CacheControl>(ST);
934 return std::make_unique<SIGfx12CacheControl>(ST);
937bool SIGfx6CacheControl::enableLoadCacheBypass(
940 SIAtomicAddrSpace AddrSpace)
const {
943 if (!canAffectGlobalAddrSpace(AddrSpace)) {
955 case SIAtomicScope::SYSTEM:
956 if (
ST.hasGFX940Insts()) {
962 case SIAtomicScope::AGENT:
963 if (
ST.hasGFX940Insts()) {
972 case SIAtomicScope::WORKGROUP:
973 if (
ST.hasGFX940Insts()) {
980 }
else if (
ST.hasGFX90AInsts()) {
985 if (
ST.isTgSplitEnabled())
989 case SIAtomicScope::WAVEFRONT:
990 case SIAtomicScope::SINGLETHREAD:
1000bool SIGfx6CacheControl::enableStoreCacheBypass(
1002 SIAtomicScope Scope,
1003 SIAtomicAddrSpace AddrSpace)
const {
1011 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1013 case SIAtomicScope::SYSTEM:
1017 case SIAtomicScope::AGENT:
1021 case SIAtomicScope::WORKGROUP:
1025 case SIAtomicScope::WAVEFRONT:
1026 case SIAtomicScope::SINGLETHREAD:
1044bool SIGfx6CacheControl::enableRMWCacheBypass(
1046 SIAtomicScope Scope,
1047 SIAtomicAddrSpace AddrSpace)
const {
1057 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1059 case SIAtomicScope::SYSTEM:
1063 case SIAtomicScope::AGENT:
1064 case SIAtomicScope::WORKGROUP:
1065 case SIAtomicScope::WAVEFRONT:
1066 case SIAtomicScope::SINGLETHREAD:
1080bool SIGfx6CacheControl::enableVolatileAndOrNonTemporal(
1082 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1092 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1097 if (
ST.hasGFX940Insts()) {
1100 }
else if (
Op == SIMemOp::LOAD) {
1112 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1113 Position::AFTER, AtomicOrdering::Unordered,
1119 if (IsNonTemporal) {
1120 if (
ST.hasGFX940Insts()) {
1134 SIAtomicScope Scope,
1135 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1136 bool IsCrossAddrSpaceOrdering, Position Pos,
1138 bool AtomicsOnly)
const {
1141 MachineBasicBlock &
MBB = *
MI->getParent();
1144 if (Pos == Position::AFTER)
1148 if (
ST.hasGFX90AInsts() &&
ST.isTgSplitEnabled()) {
1156 if (((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH |
1157 SIAtomicAddrSpace::GDS)) != SIAtomicAddrSpace::NONE) &&
1158 (Scope == SIAtomicScope::WORKGROUP)) {
1160 Scope = SIAtomicScope::AGENT;
1164 AddrSpace &= ~SIAtomicAddrSpace
::LDS;
1168 bool LGKMCnt =
false;
1170 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1171 SIAtomicAddrSpace::NONE) {
1173 case SIAtomicScope::SYSTEM:
1174 case SIAtomicScope::AGENT:
1177 case SIAtomicScope::WORKGROUP:
1178 case SIAtomicScope::WAVEFRONT:
1179 case SIAtomicScope::SINGLETHREAD:
1188 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1190 case SIAtomicScope::SYSTEM:
1191 case SIAtomicScope::AGENT:
1192 case SIAtomicScope::WORKGROUP:
1199 LGKMCnt |= IsCrossAddrSpaceOrdering;
1201 case SIAtomicScope::WAVEFRONT:
1202 case SIAtomicScope::SINGLETHREAD:
1211 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1213 case SIAtomicScope::SYSTEM:
1214 case SIAtomicScope::AGENT:
1221 LGKMCnt |= IsCrossAddrSpaceOrdering;
1223 case SIAtomicScope::WORKGROUP:
1224 case SIAtomicScope::WAVEFRONT:
1225 case SIAtomicScope::SINGLETHREAD:
1234 if (VMCnt || LGKMCnt) {
1235 unsigned WaitCntImmediate =
1241 .
addImm(WaitCntImmediate);
1249 Scope == SIAtomicScope::WORKGROUP &&
1250 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1255 if (Pos == Position::AFTER)
1264 return !ST.isAmdPalOS() && !ST.isMesa3DOS();
1268 SIAtomicScope Scope,
1269 SIAtomicAddrSpace AddrSpace,
1270 Position Pos)
const {
1271 if (!InsertCacheInv)
1276 MachineBasicBlock &
MBB = *
MI->getParent();
1279 if (Pos == Position::AFTER)
1283 ? AMDGPU::BUFFER_WBINVL1_VOL
1284 : AMDGPU::BUFFER_WBINVL1;
1286 if (canAffectGlobalAddrSpace(AddrSpace)) {
1288 case SIAtomicScope::SYSTEM:
1289 if (
ST.hasGFX940Insts()) {
1305 if (
ST.hasGFX90AInsts()) {
1320 case SIAtomicScope::AGENT:
1321 if (
ST.hasGFX940Insts()) {
1336 case SIAtomicScope::WORKGROUP:
1337 if (
ST.isTgSplitEnabled()) {
1338 if (
ST.hasGFX940Insts()) {
1357 }
else if (
ST.hasGFX90AInsts()) {
1363 case SIAtomicScope::WAVEFRONT:
1364 case SIAtomicScope::SINGLETHREAD:
1381 if (Pos == Position::AFTER)
1388 SIAtomicScope Scope,
1389 SIAtomicAddrSpace AddrSpace,
1390 bool IsCrossAddrSpaceOrdering,
1391 Position Pos)
const {
1394 if (
ST.hasGFX90AInsts()) {
1395 MachineBasicBlock &
MBB = *
MI->getParent();
1398 if (Pos == Position::AFTER)
1401 if (canAffectGlobalAddrSpace(AddrSpace)) {
1403 case SIAtomicScope::SYSTEM:
1415 case SIAtomicScope::AGENT:
1416 if (
ST.hasGFX940Insts()) {
1427 case SIAtomicScope::WORKGROUP:
1428 case SIAtomicScope::WAVEFRONT:
1429 case SIAtomicScope::SINGLETHREAD:
1439 if (Pos == Position::AFTER)
1445 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
1446 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
1452bool SIGfx10CacheControl::enableLoadCacheBypass(
1454 SIAtomicAddrSpace AddrSpace)
const {
1458 if (canAffectGlobalAddrSpace(AddrSpace)) {
1460 case SIAtomicScope::SYSTEM:
1461 case SIAtomicScope::AGENT:
1468 case SIAtomicScope::WORKGROUP:
1473 if (!
ST.isCuModeEnabled())
1476 case SIAtomicScope::WAVEFRONT:
1477 case SIAtomicScope::SINGLETHREAD:
1495bool SIGfx10CacheControl::enableVolatileAndOrNonTemporal(
1497 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1508 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1516 if (
Op == SIMemOp::LOAD) {
1529 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1530 Position::AFTER, AtomicOrdering::Unordered,
1535 if (IsNonTemporal) {
1540 if (
Op == SIMemOp::STORE)
1555 SIAtomicScope Scope,
1556 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1557 bool IsCrossAddrSpaceOrdering,
1559 bool AtomicsOnly)
const {
1562 MachineBasicBlock &
MBB = *
MI->getParent();
1565 if (Pos == Position::AFTER)
1570 bool LGKMCnt =
false;
1572 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1573 SIAtomicAddrSpace::NONE) {
1575 case SIAtomicScope::SYSTEM:
1576 case SIAtomicScope::AGENT:
1577 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1579 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1582 case SIAtomicScope::WORKGROUP:
1592 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1594 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1598 case SIAtomicScope::WAVEFRONT:
1599 case SIAtomicScope::SINGLETHREAD:
1608 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1610 case SIAtomicScope::SYSTEM:
1611 case SIAtomicScope::AGENT:
1612 case SIAtomicScope::WORKGROUP:
1619 LGKMCnt |= IsCrossAddrSpaceOrdering;
1621 case SIAtomicScope::WAVEFRONT:
1622 case SIAtomicScope::SINGLETHREAD:
1631 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1633 case SIAtomicScope::SYSTEM:
1634 case SIAtomicScope::AGENT:
1641 LGKMCnt |= IsCrossAddrSpaceOrdering;
1643 case SIAtomicScope::WORKGROUP:
1644 case SIAtomicScope::WAVEFRONT:
1645 case SIAtomicScope::SINGLETHREAD:
1654 if (VMCnt || LGKMCnt) {
1655 unsigned WaitCntImmediate =
1661 .
addImm(WaitCntImmediate);
1669 Scope == SIAtomicScope::WORKGROUP &&
1670 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1682 if (Pos == Position::AFTER)
1689 SIAtomicScope Scope,
1690 SIAtomicAddrSpace AddrSpace,
1691 Position Pos)
const {
1692 if (!InsertCacheInv)
1697 MachineBasicBlock &
MBB = *
MI->getParent();
1700 if (Pos == Position::AFTER)
1703 if (canAffectGlobalAddrSpace(AddrSpace)) {
1705 case SIAtomicScope::SYSTEM:
1706 case SIAtomicScope::AGENT:
1714 case SIAtomicScope::WORKGROUP:
1719 if (!
ST.isCuModeEnabled()) {
1724 case SIAtomicScope::WAVEFRONT:
1725 case SIAtomicScope::SINGLETHREAD:
1740 if (Pos == Position::AFTER)
1748 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1763 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1776bool SIGfx12CacheControl::insertWaitsBeforeSystemScopeStore(
1780 MachineBasicBlock &
MBB = *
MI->getParent();
1784 if (
ST.hasImageInsts()) {
1795 SIAtomicScope Scope,
1796 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1797 bool IsCrossAddrSpaceOrdering,
1799 bool AtomicsOnly)
const {
1802 MachineBasicBlock &
MBB = *
MI->getParent();
1805 bool LOADCnt =
false;
1807 bool STORECnt =
false;
1809 if (Pos == Position::AFTER)
1812 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1813 SIAtomicAddrSpace::NONE) {
1815 case SIAtomicScope::SYSTEM:
1816 case SIAtomicScope::AGENT:
1817 case SIAtomicScope::CLUSTER:
1818 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1820 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1823 case SIAtomicScope::WORKGROUP:
1840 if (!
ST.isCuModeEnabled() ||
ST.hasGFX1250Insts() ||
1842 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1844 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1848 case SIAtomicScope::WAVEFRONT:
1849 case SIAtomicScope::SINGLETHREAD:
1858 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1860 case SIAtomicScope::SYSTEM:
1861 case SIAtomicScope::AGENT:
1862 case SIAtomicScope::CLUSTER:
1863 case SIAtomicScope::WORKGROUP:
1870 DSCnt |= IsCrossAddrSpaceOrdering;
1872 case SIAtomicScope::WAVEFRONT:
1873 case SIAtomicScope::SINGLETHREAD:
1894 if (!AtomicsOnly &&
ST.hasImageInsts()) {
1912 if (Pos == Position::AFTER)
1919 SIAtomicScope Scope,
1920 SIAtomicAddrSpace AddrSpace,
1921 Position Pos)
const {
1922 if (!InsertCacheInv)
1925 MachineBasicBlock &
MBB = *
MI->getParent();
1934 if (!canAffectGlobalAddrSpace(AddrSpace))
1939 case SIAtomicScope::SYSTEM:
1942 case SIAtomicScope::AGENT:
1945 case SIAtomicScope::CLUSTER:
1948 case SIAtomicScope::WORKGROUP:
1956 if (
ST.isCuModeEnabled())
1961 case SIAtomicScope::WAVEFRONT:
1962 case SIAtomicScope::SINGLETHREAD:
1969 if (Pos == Position::AFTER)
1974 if (Pos == Position::AFTER)
1981 SIAtomicScope Scope,
1982 SIAtomicAddrSpace AddrSpace,
1983 bool IsCrossAddrSpaceOrdering,
1984 Position Pos)
const {
1987 MachineBasicBlock &
MBB = *
MI->getParent();
1994 if (canAffectGlobalAddrSpace(AddrSpace)) {
1995 if (Pos == Position::AFTER)
2005 case SIAtomicScope::SYSTEM:
2010 case SIAtomicScope::AGENT:
2012 if (
ST.hasGFX1250Insts()) {
2018 case SIAtomicScope::CLUSTER:
2019 case SIAtomicScope::WORKGROUP:
2021 case SIAtomicScope::WAVEFRONT:
2022 case SIAtomicScope::SINGLETHREAD:
2029 if (Pos == Position::AFTER)
2036 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
2037 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
2043bool SIGfx12CacheControl::enableVolatileAndOrNonTemporal(
2045 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
2054 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
2061 }
else if (IsNonTemporal) {
2069 if (
ST.requiresWaitXCntForSingleAccessInstructions() &&
2071 MachineBasicBlock &
MBB = *
MI->getParent();
2081 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
2082 Position::AFTER, AtomicOrdering::Unordered,
2089bool SIGfx12CacheControl::finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
2090 assert(
MI.mayStore() &&
"Not a Store inst");
2091 const bool IsRMW = (
MI.mayLoad() &&
MI.mayStore());
2094 if (Atomic &&
ST.requiresWaitXCntForSingleAccessInstructions() &&
2096 MachineBasicBlock &
MBB = *
MI.getParent();
2105 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2111 if (
ST.requiresWaitsBeforeSystemScopeStores() && !Atomic &&
2113 Changed |= insertWaitsBeforeSystemScopeStore(
MI.getIterator());
2118bool SIGfx12CacheControl::handleCooperativeAtomic(MachineInstr &
MI)
const {
2119 if (!
ST.hasGFX1250Insts())
2123 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2124 assert(CPol &&
"No CPol operand?");
2132 SIAtomicScope Scope,
2133 SIAtomicAddrSpace AddrSpace)
const {
2136 if (canAffectGlobalAddrSpace(AddrSpace)) {
2138 case SIAtomicScope::SYSTEM:
2141 case SIAtomicScope::AGENT:
2144 case SIAtomicScope::CLUSTER:
2147 case SIAtomicScope::WORKGROUP:
2150 if (!
ST.isCuModeEnabled())
2153 case SIAtomicScope::WAVEFRONT:
2154 case SIAtomicScope::SINGLETHREAD:
2172bool SIMemoryLegalizer::removeAtomicPseudoMIs() {
2173 if (AtomicPseudoMIs.empty())
2176 for (
auto &
MI : AtomicPseudoMIs)
2177 MI->eraseFromParent();
2179 AtomicPseudoMIs.clear();
2183bool SIMemoryLegalizer::expandLoad(
const SIMemOpInfo &MOI,
2189 if (MOI.isAtomic()) {
2191 if (Order == AtomicOrdering::Monotonic ||
2192 Order == AtomicOrdering::Acquire ||
2193 Order == AtomicOrdering::SequentiallyConsistent) {
2194 Changed |= CC->enableLoadCacheBypass(
MI, MOI.getScope(),
2195 MOI.getOrderingAddrSpace());
2200 if (MOI.isCooperative())
2201 Changed |= CC->handleCooperativeAtomic(*
MI);
2203 if (Order == AtomicOrdering::SequentiallyConsistent)
2204 Changed |= CC->insertWait(
MI, MOI.getScope(), MOI.getOrderingAddrSpace(),
2205 SIMemOp::LOAD | SIMemOp::STORE,
2206 MOI.getIsCrossAddressSpaceOrdering(),
2207 Position::BEFORE, Order,
false);
2209 if (Order == AtomicOrdering::Acquire ||
2210 Order == AtomicOrdering::SequentiallyConsistent) {
2213 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2214 SIMemOp::LOAD, MOI.getIsCrossAddressSpaceOrdering(),
2215 Position::AFTER, Order,
true);
2216 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2217 MOI.getOrderingAddrSpace(),
2227 Changed |= CC->enableVolatileAndOrNonTemporal(
2228 MI, MOI.getInstrAddrSpace(), SIMemOp::LOAD, MOI.isVolatile(),
2229 MOI.isNonTemporal(), MOI.isLastUse());
2234bool SIMemoryLegalizer::expandStore(
const SIMemOpInfo &MOI,
2240 MachineInstr &StoreMI = *
MI;
2242 if (MOI.isAtomic()) {
2243 if (MOI.getOrdering() == AtomicOrdering::Monotonic ||
2244 MOI.getOrdering() == AtomicOrdering::Release ||
2245 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) {
2246 Changed |= CC->enableStoreCacheBypass(
MI, MOI.getScope(),
2247 MOI.getOrderingAddrSpace());
2252 if (MOI.isCooperative())
2253 Changed |= CC->handleCooperativeAtomic(*
MI);
2255 if (MOI.getOrdering() == AtomicOrdering::Release ||
2256 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
2257 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2258 MOI.getOrderingAddrSpace(),
2259 MOI.getIsCrossAddressSpaceOrdering(),
2262 Changed |= CC->finalizeStore(StoreMI,
true);
2269 Changed |= CC->enableVolatileAndOrNonTemporal(
2270 MI, MOI.getInstrAddrSpace(), SIMemOp::STORE, MOI.isVolatile(),
2271 MOI.isNonTemporal());
2275 Changed |= CC->finalizeStore(StoreMI,
false);
2279bool SIMemoryLegalizer::expandAtomicFence(
const SIMemOpInfo &MOI,
2281 assert(
MI->getOpcode() == AMDGPU::ATOMIC_FENCE);
2283 AtomicPseudoMIs.push_back(
MI);
2286 const SIAtomicAddrSpace OrderingAddrSpace = MOI.getOrderingAddrSpace();
2288 if (MOI.isAtomic()) {
2290 if (Order == AtomicOrdering::Acquire) {
2292 Changed |= CC->insertWait(
MI, MOI.getScope(), OrderingAddrSpace,
2293 SIMemOp::LOAD | SIMemOp::STORE,
2294 MOI.getIsCrossAddressSpaceOrdering(),
2295 Position::BEFORE, Order,
true);
2298 if (Order == AtomicOrdering::Release ||
2299 Order == AtomicOrdering::AcquireRelease ||
2300 Order == AtomicOrdering::SequentiallyConsistent)
2308 Changed |= CC->insertRelease(
MI, MOI.getScope(), OrderingAddrSpace,
2309 MOI.getIsCrossAddressSpaceOrdering(),
2317 if (Order == AtomicOrdering::Acquire ||
2318 Order == AtomicOrdering::AcquireRelease ||
2319 Order == AtomicOrdering::SequentiallyConsistent)
2320 Changed |= CC->insertAcquire(
MI, MOI.getScope(), OrderingAddrSpace,
2329bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
2334 MachineInstr &RMWMI = *
MI;
2336 if (MOI.isAtomic()) {
2338 if (Order == AtomicOrdering::Monotonic ||
2339 Order == AtomicOrdering::Acquire || Order == AtomicOrdering::Release ||
2340 Order == AtomicOrdering::AcquireRelease ||
2341 Order == AtomicOrdering::SequentiallyConsistent) {
2342 Changed |= CC->enableRMWCacheBypass(
MI, MOI.getScope(),
2343 MOI.getInstrAddrSpace());
2346 if (Order == AtomicOrdering::Release ||
2347 Order == AtomicOrdering::AcquireRelease ||
2348 Order == AtomicOrdering::SequentiallyConsistent ||
2349 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent)
2350 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2351 MOI.getOrderingAddrSpace(),
2352 MOI.getIsCrossAddressSpaceOrdering(),
2355 if (Order == AtomicOrdering::Acquire ||
2356 Order == AtomicOrdering::AcquireRelease ||
2357 Order == AtomicOrdering::SequentiallyConsistent ||
2358 MOI.getFailureOrdering() == AtomicOrdering::Acquire ||
2359 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) {
2362 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2363 isAtomicRet(*
MI) ? SIMemOp::LOAD : SIMemOp::STORE,
2364 MOI.getIsCrossAddressSpaceOrdering(), Position::AFTER,
2366 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2367 MOI.getOrderingAddrSpace(),
2371 Changed |= CC->finalizeStore(RMWMI,
true);
2378bool SIMemoryLegalizer::expandLDSDMA(
const SIMemOpInfo &MOI,
2390 return CC->enableVolatileAndOrNonTemporal(
2391 MI, MOI.getInstrAddrSpace(), OpKind, MOI.isVolatile(),
2392 MOI.isNonTemporal(), MOI.isLastUse());
2395bool SIMemoryLegalizerLegacy::runOnMachineFunction(MachineFunction &MF) {
2396 const MachineModuleInfo &MMI =
2397 getAnalysis<MachineModuleInfoWrapperPass>().getMMI();
2398 return SIMemoryLegalizer(MMI).run(MF);
2405 .getCachedResult<MachineModuleAnalysis>(
2407 assert(MMI &&
"MachineModuleAnalysis must be available");
2408 if (!SIMemoryLegalizer(MMI->getMMI()).run(MF))
2418 CC = SICacheControl::create(ST);
2420 for (
auto &
MBB : MF) {
2424 if (
MI->isBundle() &&
MI->mayLoadOrStore()) {
2427 I != E &&
I->isBundledWithPred(); ++
I) {
2428 I->unbundleFromPred();
2431 MO.setIsInternalRead(
false);
2434 MI->eraseFromParent();
2435 MI =
II->getIterator();
2441 if (
const auto &MOI = MOA.getLoadInfo(
MI)) {
2443 }
else if (
const auto &MOI = MOA.getStoreInfo(
MI)) {
2445 }
else if (
const auto &MOI = MOA.getLDSDMAInfo(
MI)) {
2447 }
else if (
const auto &MOI = MOA.getAtomicFenceInfo(
MI)) {
2449 }
else if (
const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(
MI)) {
2450 Changed |= expandAtomicCmpxchgOrRmw(*MOI,
MI);
2455 Changed |= removeAtomicPseudoMIs();
2461char SIMemoryLegalizerLegacy::
ID = 0;
2465 return new SIMemoryLegalizerLegacy();
static std::optional< LoadInfo > getLoadInfo(const MachineInstr &MI)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU address space definition.
Provides AMDGPU specific target descriptions.
AMDGPU Machine Module Info.
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Atomic ordering constants.
AMD GCN specific subclass of TargetSubtarget.
const HexagonInstrInfo * TII
This header defines various interfaces for pass management in LLVM.
This file provides utility for Memory Model Relaxation Annotations (MMRAs).
uint64_t IntrinsicInst * II
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
static cl::opt< bool > AmdgcnSkipCacheInvalidations("amdgcn-skip-cache-invalidations", cl::init(false), cl::Hidden, cl::desc("Use this to skip inserting cache invalidating instructions."))
static bool canUseBUFFER_WBINVL1_VOL(const GCNSubtarget &ST)
static const uint32_t IV[8]
SyncScope::ID getWorkgroupSSID() const
SyncScope::ID getWavefrontSSID() const
SyncScope::ID getAgentSSID() const
SyncScope::ID getClusterOneAddressSpaceSSID() const
SyncScope::ID getClusterSSID() const
std::optional< bool > isSyncScopeInclusion(SyncScope::ID A, SyncScope::ID B) const
In AMDGPU target synchronization scopes are inclusive, meaning a larger synchronization scope is incl...
SyncScope::ID getAgentOneAddressSpaceSSID() const
SyncScope::ID getSingleThreadOneAddressSpaceSSID() const
SyncScope::ID getWavefrontOneAddressSpaceSSID() const
SyncScope::ID getSystemOneAddressSpaceSSID() const
SyncScope::ID getWorkgroupOneAddressSpaceSSID() const
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
LLVM_ABI void setPreservesCFG()
This function should be called by the pass, iff they do not:
Represents analyses that only rely on functions' control flow.
Diagnostic information for unsupported feature in backend.
FunctionPass class - This class is used to implement most global optimizations.
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
A helper class to return the specified delimiter string after the first invocation of operator String...
Instructions::iterator instr_iterator
MachineInstrBundleIterator< MachineInstr > iterator
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
Representation of each machine instruction.
Ty & getObjFileInfo()
Keep track of various per-module pieces of information for backends that would like to do so.
MachineOperand class - Representation of each machine instruction operand.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
PreservedAnalyses & preserveSet()
Mark an analysis set as preserved.
static bool isVMEM(const MachineInstr &MI)
static bool mayWriteLDSThroughDMA(const MachineInstr &MI)
static bool isBUF(const MachineInstr &MI)
static bool isAtomicRet(const MachineInstr &MI)
static bool isAtomic(const MachineInstr &MI)
static bool isLDSDMA(const MachineInstr &MI)
PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MFAM)
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
A wrapper around a string literal that serves as a proxy for constructing global tables of StringRefs...
StringMap - This is an unconventional map that is specialized for handling keys that are "strings",...
StringRef - Represent a constant reference to a string, i.e.
A raw_ostream that writes to an SmallVector or SmallString.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ BUFFER_STRIDED_POINTER
Address space for 192-bit fat buffer pointers with an additional index.
@ REGION_ADDRESS
Address space for region memory. (GDS)
@ LOCAL_ADDRESS
Address space for local memory.
@ FLAT_ADDRESS
Address space for flat memory.
@ GLOBAL_ADDRESS
Address space for global memory (RAT0, VTX0).
@ BUFFER_FAT_POINTER
Address space for 160-bit buffer fat pointers.
@ PRIVATE_ADDRESS
Address space for private memory.
@ BUFFER_RESOURCE
Address space for 128-bit buffer resources.
constexpr char IsVolatile[]
Key for Kernel::Arg::Metadata::mIsVolatile.
bool isGFX10(const MCSubtargetInfo &STI)
bool isGFX11(const MCSubtargetInfo &STI)
LLVM_ABI IsaVersion getIsaVersion(StringRef GPU)
unsigned encodeWaitcnt(const IsaVersion &Version, unsigned Vmcnt, unsigned Expcnt, unsigned Lgkmcnt)
Encodes Vmcnt, Expcnt and Lgkmcnt into Waitcnt for given isa Version.
unsigned getVmcntBitMask(const IsaVersion &Version)
unsigned getLgkmcntBitMask(const IsaVersion &Version)
unsigned getExpcntBitMask(const IsaVersion &Version)
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Undef
Value of the register doesn't matter.
@ SingleThread
Synchronized with respect to signal handlers executing in the same thread.
@ System
Synchronized with respect to all concurrently executing threads.
initializer< Ty > init(const Ty &Val)
PointerTypeMap run(const Module &M)
Compute the PointerTypeMap for the module M.
Scope
Defines the scope in which this symbol should be visible: Default – Visible in the public interface o...
NodeAddr< FuncNode * > Func
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
OuterAnalysisManagerProxy< ModuleAnalysisManager, MachineFunction > ModuleAnalysisManagerMachineFunctionProxy
Provide the ModuleAnalysisManager to Function proxy.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
char & SIMemoryLegalizerID
LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE()
AnalysisManager< MachineFunction > MachineFunctionAnalysisManager
bool isReleaseOrStronger(AtomicOrdering AO)
LLVM_ABI PreservedAnalyses getMachineFunctionPassPreservedAnalyses()
Returns the minimum set of Analyses that all machine function passes must preserve.
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
AtomicOrdering getMergedAtomicOrdering(AtomicOrdering AO, AtomicOrdering Other)
Return a single atomic ordering that is at least as strong as both the AO and Other orderings for an ...
static const MachineMemOperand::Flags MOCooperative
Mark the MMO of cooperative load/store atomics.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ LLVM_MARK_AS_BITMASK_ENUM
DWARFExpression::Operation Op
static const MachineMemOperand::Flags MOLastUse
Mark the MMO of a load as the last use.
FunctionPass * createSIMemoryLegalizerPass()