35#define DEBUG_TYPE "si-memory-legalizer"
36#define PASS_NAME "SI Memory Legalizer"
40 cl::desc(
"Use this to skip inserting cache invalidating instructions."));
62enum class SIAtomicScope {
74enum class SIAtomicAddrSpace {
83 FLAT = GLOBAL |
LDS | SCRATCH,
86 ATOMIC = GLOBAL |
LDS | SCRATCH | GDS,
89 ALL = GLOBAL |
LDS | SCRATCH | GDS | OTHER,
94class SIMemOpInfo final {
97 friend class SIMemOpAccess;
101 SIAtomicScope Scope = SIAtomicScope::SYSTEM;
102 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
103 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
104 bool IsCrossAddressSpaceOrdering =
false;
105 bool IsVolatile =
false;
106 bool IsNonTemporal =
false;
107 bool IsLastUse =
false;
108 bool IsCooperative =
false;
112 const GCNSubtarget &ST,
114 SIAtomicScope Scope = SIAtomicScope::SYSTEM,
115 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC,
116 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL,
117 bool IsCrossAddressSpaceOrdering =
true,
118 AtomicOrdering FailureOrdering = AtomicOrdering::SequentiallyConsistent,
119 bool IsVolatile =
false,
bool IsNonTemporal =
false,
120 bool IsLastUse =
false,
bool IsCooperative =
false)
121 : Ordering(Ordering), FailureOrdering(FailureOrdering), Scope(Scope),
122 OrderingAddrSpace(OrderingAddrSpace), InstrAddrSpace(InstrAddrSpace),
123 IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering),
124 IsVolatile(IsVolatile), IsNonTemporal(IsNonTemporal),
125 IsLastUse(IsLastUse), IsCooperative(IsCooperative) {
127 if (Ordering == AtomicOrdering::NotAtomic) {
128 assert(!IsCooperative &&
"Cannot be cooperative & non-atomic!");
129 assert(Scope == SIAtomicScope::NONE &&
130 OrderingAddrSpace == SIAtomicAddrSpace::NONE &&
131 !IsCrossAddressSpaceOrdering &&
132 FailureOrdering == AtomicOrdering::NotAtomic);
136 assert(Scope != SIAtomicScope::NONE &&
137 (OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
138 SIAtomicAddrSpace::NONE &&
139 (InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
140 SIAtomicAddrSpace::NONE);
145 if ((OrderingAddrSpace == InstrAddrSpace) &&
147 this->IsCrossAddressSpaceOrdering =
false;
151 if ((InstrAddrSpace & ~SIAtomicAddrSpace::SCRATCH) ==
152 SIAtomicAddrSpace::NONE) {
153 this->Scope = std::min(Scope, SIAtomicScope::SINGLETHREAD);
154 }
else if ((InstrAddrSpace &
155 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS)) ==
156 SIAtomicAddrSpace::NONE) {
157 this->Scope = std::min(Scope, SIAtomicScope::WORKGROUP);
158 }
else if ((InstrAddrSpace &
159 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS |
160 SIAtomicAddrSpace::GDS)) == SIAtomicAddrSpace::NONE) {
161 this->Scope = std::min(Scope, SIAtomicScope::AGENT);
166 if (this->Scope == SIAtomicScope::CLUSTER && !
ST.hasClusters())
167 this->Scope = SIAtomicScope::AGENT;
173 SIAtomicScope getScope()
const {
186 return FailureOrdering;
191 SIAtomicAddrSpace getInstrAddrSpace()
const {
192 return InstrAddrSpace;
197 SIAtomicAddrSpace getOrderingAddrSpace()
const {
198 return OrderingAddrSpace;
203 bool getIsCrossAddressSpaceOrdering()
const {
204 return IsCrossAddressSpaceOrdering;
209 bool isVolatile()
const {
215 bool isNonTemporal()
const {
216 return IsNonTemporal;
221 bool isLastUse()
const {
return IsLastUse; }
224 bool isCooperative()
const {
return IsCooperative; }
228 bool isAtomic()
const {
229 return Ordering != AtomicOrdering::NotAtomic;
234class SIMemOpAccess final {
236 const AMDGPUMachineModuleInfo *MMI =
nullptr;
237 const GCNSubtarget &ST;
241 const char *Msg)
const;
247 std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
248 toSIAtomicScope(
SyncScope::ID SSID, SIAtomicAddrSpace InstrAddrSpace)
const;
251 SIAtomicAddrSpace toSIAtomicAddrSpace(
unsigned AS)
const;
255 std::optional<SIMemOpInfo>
261 SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI,
const GCNSubtarget &ST);
264 std::optional<SIMemOpInfo>
269 std::optional<SIMemOpInfo>
274 std::optional<SIMemOpInfo>
279 std::optional<SIMemOpInfo>
285 std::optional<SIMemOpInfo>
289class SICacheControl {
293 const GCNSubtarget &ST;
296 const SIInstrInfo *TII =
nullptr;
303 SICacheControl(
const GCNSubtarget &ST);
308 unsigned Bits)
const;
312 bool canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const;
318 static std::unique_ptr<SICacheControl> create(
const GCNSubtarget &ST);
325 SIAtomicAddrSpace AddrSpace)
const = 0;
332 SIAtomicAddrSpace AddrSpace)
const = 0;
339 SIAtomicAddrSpace AddrSpace)
const = 0;
345 SIAtomicAddrSpace AddrSpace,
346 SIMemOp
Op,
bool IsVolatile,
348 bool IsLastUse =
false)
const = 0;
355 virtual bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
360 virtual bool handleCooperativeAtomic(MachineInstr &
MI)
const {
362 "cooperative atomics are not available on this architecture");
375 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
376 bool IsCrossAddrSpaceOrdering, Position Pos,
386 SIAtomicAddrSpace AddrSpace,
387 Position Pos)
const = 0;
397 SIAtomicAddrSpace AddrSpace,
398 bool IsCrossAddrSpaceOrdering,
399 Position Pos)
const = 0;
402 virtual ~SICacheControl() =
default;
407class SIGfx6CacheControl final :
public SICacheControl {
410 SIGfx6CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
414 SIAtomicAddrSpace AddrSpace)
const override;
418 SIAtomicAddrSpace AddrSpace)
const override;
422 SIAtomicAddrSpace AddrSpace)
const override;
425 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
426 bool IsVolatile,
bool IsNonTemporal,
427 bool IsLastUse)
const override;
430 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
431 bool IsCrossAddrSpaceOrdering, Position Pos,
436 SIAtomicAddrSpace AddrSpace,
437 Position Pos)
const override;
441 SIAtomicAddrSpace AddrSpace,
442 bool IsCrossAddrSpaceOrdering,
443 Position Pos)
const override;
447class SIGfx10CacheControl final :
public SICacheControl {
449 SIGfx10CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
453 SIAtomicAddrSpace AddrSpace)
const override;
457 SIAtomicAddrSpace AddrSpace)
const override {
463 SIAtomicAddrSpace AddrSpace)
const override {
468 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
469 bool IsVolatile,
bool IsNonTemporal,
470 bool IsLastUse)
const override;
473 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
474 bool IsCrossAddrSpaceOrdering, Position Pos,
478 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
481 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
482 Position Pos)
const override {
483 return insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
484 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
489class SIGfx12CacheControl final :
public SICacheControl {
511 SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace)
const;
514 SIGfx12CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {
517 assert(!
ST.hasGFX1250Insts() ||
ST.isCuModeEnabled());
521 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
522 bool IsCrossAddrSpaceOrdering, Position Pos,
526 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
529 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
530 bool IsVolatile,
bool IsNonTemporal,
531 bool IsLastUse)
const override;
533 bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const override;
535 bool handleCooperativeAtomic(MachineInstr &
MI)
const override;
538 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
539 Position Pos)
const override;
543 SIAtomicAddrSpace AddrSpace)
const override {
544 return setAtomicScope(
MI, Scope, AddrSpace);
549 SIAtomicAddrSpace AddrSpace)
const override {
550 return setAtomicScope(
MI, Scope, AddrSpace);
555 SIAtomicAddrSpace AddrSpace)
const override {
556 return setAtomicScope(
MI, Scope, AddrSpace);
560class SIMemoryLegalizer final {
562 const MachineModuleInfo &MMI;
564 std::unique_ptr<SICacheControl> CC =
nullptr;
567 std::list<MachineBasicBlock::iterator> AtomicPseudoMIs;
571 bool isAtomicRet(
const MachineInstr &
MI)
const {
577 bool removeAtomicPseudoMIs();
581 bool expandLoad(
const SIMemOpInfo &MOI,
585 bool expandStore(
const SIMemOpInfo &MOI,
589 bool expandAtomicFence(
const SIMemOpInfo &MOI,
593 bool expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
600 SIMemoryLegalizer(
const MachineModuleInfo &MMI) : MMI(MMI) {};
601 bool run(MachineFunction &MF);
608 SIMemoryLegalizerLegacy() : MachineFunctionPass(ID) {}
610 void getAnalysisUsage(AnalysisUsage &AU)
const override {
615 StringRef getPassName()
const override {
619 bool runOnMachineFunction(MachineFunction &MF)
override;
623 {
"global", SIAtomicAddrSpace::GLOBAL},
624 {
"local", SIAtomicAddrSpace::LDS},
632 OS <<
"unknown address space '" << AS <<
"'; expected one of ";
634 for (
const auto &[Name, Val] : ASNames)
635 OS <<
LS <<
'\'' <<
Name <<
'\'';
643static std::optional<SIAtomicAddrSpace>
645 static constexpr StringLiteral FenceASPrefix =
"amdgpu-synchronize-as";
651 SIAtomicAddrSpace
Result = SIAtomicAddrSpace::NONE;
652 for (
const auto &[Prefix, Suffix] : MMRA) {
653 if (Prefix != FenceASPrefix)
656 if (
auto It = ASNames.find(Suffix); It != ASNames.end())
659 diagnoseUnknownMMRAASName(
MI, Suffix);
662 if (Result == SIAtomicAddrSpace::NONE)
671 const char *Msg)
const {
673 Func.getContext().diagnose(
674 DiagnosticInfoUnsupported(Func, Msg,
MI->getDebugLoc()));
677std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
679 SIAtomicAddrSpace InstrAddrSpace)
const {
681 return std::tuple(SIAtomicScope::SYSTEM, SIAtomicAddrSpace::ATOMIC,
true);
683 return std::tuple(SIAtomicScope::AGENT, SIAtomicAddrSpace::ATOMIC,
true);
685 return std::tuple(SIAtomicScope::CLUSTER, SIAtomicAddrSpace::ATOMIC,
true);
687 return std::tuple(SIAtomicScope::WORKGROUP, SIAtomicAddrSpace::ATOMIC,
690 return std::tuple(SIAtomicScope::WAVEFRONT, SIAtomicAddrSpace::ATOMIC,
693 return std::tuple(SIAtomicScope::SINGLETHREAD, SIAtomicAddrSpace::ATOMIC,
696 return std::tuple(SIAtomicScope::SYSTEM,
697 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
699 return std::tuple(SIAtomicScope::AGENT,
700 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
702 return std::tuple(SIAtomicScope::CLUSTER,
703 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
705 return std::tuple(SIAtomicScope::WORKGROUP,
706 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
708 return std::tuple(SIAtomicScope::WAVEFRONT,
709 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
711 return std::tuple(SIAtomicScope::SINGLETHREAD,
712 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
716SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(
unsigned AS)
const {
718 return SIAtomicAddrSpace::FLAT;
720 return SIAtomicAddrSpace::GLOBAL;
722 return SIAtomicAddrSpace::LDS;
724 return SIAtomicAddrSpace::SCRATCH;
726 return SIAtomicAddrSpace::GDS;
729 return SIAtomicAddrSpace::GLOBAL;
731 return SIAtomicAddrSpace::OTHER;
734SIMemOpAccess::SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI_,
735 const GCNSubtarget &ST)
736 : MMI(&MMI_),
ST(
ST) {}
738std::optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO(
740 assert(
MI->getNumMemOperands() > 0);
745 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
746 bool IsNonTemporal =
true;
748 bool IsLastUse =
false;
749 bool IsCooperative =
false;
753 for (
const auto &MMO :
MI->memoperands()) {
754 IsNonTemporal &= MMO->isNonTemporal();
756 IsLastUse |= MMO->getFlags() &
MOLastUse;
759 toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace());
761 if (OpOrdering != AtomicOrdering::NotAtomic) {
762 const auto &IsSyncScopeInclusion =
764 if (!IsSyncScopeInclusion) {
765 reportUnsupported(
MI,
766 "Unsupported non-inclusive atomic synchronization scope");
770 SSID = *IsSyncScopeInclusion ? SSID : MMO->getSyncScopeID();
772 assert(MMO->getFailureOrdering() != AtomicOrdering::Release &&
773 MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease);
784 Ordering = AtomicOrdering::Monotonic;
786 SIAtomicScope
Scope = SIAtomicScope::NONE;
787 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
788 bool IsCrossAddressSpaceOrdering =
false;
789 if (Ordering != AtomicOrdering::NotAtomic) {
790 auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace);
792 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
795 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
797 if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) ||
798 ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace) ||
799 ((InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) == SIAtomicAddrSpace::NONE)) {
800 reportUnsupported(
MI,
"Unsupported atomic address space");
804 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace, InstrAddrSpace,
805 IsCrossAddressSpaceOrdering, FailureOrdering, IsVolatile,
806 IsNonTemporal, IsLastUse, IsCooperative);
809std::optional<SIMemOpInfo>
813 if (!(
MI->mayLoad() && !
MI->mayStore()))
817 if (
MI->getNumMemOperands() == 0)
818 return SIMemOpInfo(ST);
820 return constructFromMIWithMMO(
MI);
823std::optional<SIMemOpInfo>
827 if (!(!
MI->mayLoad() &&
MI->mayStore()))
831 if (
MI->getNumMemOperands() == 0)
832 return SIMemOpInfo(ST);
834 return constructFromMIWithMMO(
MI);
837std::optional<SIMemOpInfo>
841 if (
MI->getOpcode() != AMDGPU::ATOMIC_FENCE)
848 auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC);
850 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
854 SIAtomicScope
Scope = SIAtomicScope::NONE;
855 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
856 bool IsCrossAddressSpaceOrdering =
false;
857 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
860 if (OrderingAddrSpace != SIAtomicAddrSpace::ATOMIC) {
865 reportUnsupported(
MI,
"Unsupported atomic address space");
869 auto SynchronizeAS = getSynchronizeAddrSpaceMD(*
MI);
871 OrderingAddrSpace = *SynchronizeAS;
873 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace,
874 SIAtomicAddrSpace::ATOMIC, IsCrossAddressSpaceOrdering,
875 AtomicOrdering::NotAtomic);
878std::optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo(
882 if (!(
MI->mayLoad() &&
MI->mayStore()))
886 if (
MI->getNumMemOperands() == 0)
887 return SIMemOpInfo(ST);
889 return constructFromMIWithMMO(
MI);
892std::optional<SIMemOpInfo>
899 return constructFromMIWithMMO(
MI);
902SICacheControl::SICacheControl(
const GCNSubtarget &ST) :
ST(
ST) {
903 TII =
ST.getInstrInfo();
909 unsigned Bits)
const {
910 MachineOperand *CPol =
TII->getNamedOperand(*
MI, AMDGPU::OpName::cpol);
914 CPol->setImm(
CPol->getImm() | Bits);
918bool SICacheControl::canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const {
919 assert((!
ST.hasGloballyAddressableScratch() ||
920 (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE ||
921 (AS & SIAtomicAddrSpace::SCRATCH) == SIAtomicAddrSpace::NONE) &&
922 "scratch instructions should already be replaced by flat "
923 "instructions if GloballyAddressableScratch is enabled");
924 return (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE;
928std::unique_ptr<SICacheControl> SICacheControl::create(
const GCNSubtarget &ST) {
929 GCNSubtarget::Generation Generation =
ST.getGeneration();
930 if (Generation < AMDGPUSubtarget::GFX10)
931 return std::make_unique<SIGfx6CacheControl>(ST);
932 if (Generation < AMDGPUSubtarget::GFX12)
933 return std::make_unique<SIGfx10CacheControl>(ST);
934 return std::make_unique<SIGfx12CacheControl>(ST);
937bool SIGfx6CacheControl::enableLoadCacheBypass(
940 SIAtomicAddrSpace AddrSpace)
const {
943 if (!canAffectGlobalAddrSpace(AddrSpace)) {
955 case SIAtomicScope::SYSTEM:
956 if (
ST.hasGFX940Insts()) {
962 case SIAtomicScope::AGENT:
963 if (
ST.hasGFX940Insts()) {
972 case SIAtomicScope::WORKGROUP:
973 if (
ST.hasGFX940Insts()) {
980 }
else if (
ST.hasGFX90AInsts()) {
985 if (
ST.isTgSplitEnabled())
989 case SIAtomicScope::WAVEFRONT:
990 case SIAtomicScope::SINGLETHREAD:
1000bool SIGfx6CacheControl::enableStoreCacheBypass(
1002 SIAtomicScope Scope,
1003 SIAtomicAddrSpace AddrSpace)
const {
1011 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1013 case SIAtomicScope::SYSTEM:
1017 case SIAtomicScope::AGENT:
1021 case SIAtomicScope::WORKGROUP:
1025 case SIAtomicScope::WAVEFRONT:
1026 case SIAtomicScope::SINGLETHREAD:
1044bool SIGfx6CacheControl::enableRMWCacheBypass(
1046 SIAtomicScope Scope,
1047 SIAtomicAddrSpace AddrSpace)
const {
1057 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1059 case SIAtomicScope::SYSTEM:
1063 case SIAtomicScope::AGENT:
1064 case SIAtomicScope::WORKGROUP:
1065 case SIAtomicScope::WAVEFRONT:
1066 case SIAtomicScope::SINGLETHREAD:
1080bool SIGfx6CacheControl::enableVolatileAndOrNonTemporal(
1082 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1092 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1097 if (
ST.hasGFX940Insts()) {
1100 }
else if (
Op == SIMemOp::LOAD) {
1112 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1113 Position::AFTER, AtomicOrdering::Unordered,
1119 if (IsNonTemporal) {
1120 if (
ST.hasGFX940Insts()) {
1134 SIAtomicScope Scope,
1135 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1136 bool IsCrossAddrSpaceOrdering, Position Pos,
1138 bool AtomicsOnly)
const {
1141 MachineBasicBlock &
MBB = *
MI->getParent();
1144 if (Pos == Position::AFTER)
1148 if (
ST.hasGFX90AInsts() &&
ST.isTgSplitEnabled()) {
1156 if (((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH |
1157 SIAtomicAddrSpace::GDS)) != SIAtomicAddrSpace::NONE) &&
1158 (Scope == SIAtomicScope::WORKGROUP)) {
1160 Scope = SIAtomicScope::AGENT;
1164 AddrSpace &= ~SIAtomicAddrSpace
::LDS;
1168 bool LGKMCnt =
false;
1170 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1171 SIAtomicAddrSpace::NONE) {
1173 case SIAtomicScope::SYSTEM:
1174 case SIAtomicScope::AGENT:
1177 case SIAtomicScope::WORKGROUP:
1178 case SIAtomicScope::WAVEFRONT:
1179 case SIAtomicScope::SINGLETHREAD:
1188 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1190 case SIAtomicScope::SYSTEM:
1191 case SIAtomicScope::AGENT:
1192 case SIAtomicScope::WORKGROUP:
1199 LGKMCnt |= IsCrossAddrSpaceOrdering;
1201 case SIAtomicScope::WAVEFRONT:
1202 case SIAtomicScope::SINGLETHREAD:
1211 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1213 case SIAtomicScope::SYSTEM:
1214 case SIAtomicScope::AGENT:
1221 LGKMCnt |= IsCrossAddrSpaceOrdering;
1223 case SIAtomicScope::WORKGROUP:
1224 case SIAtomicScope::WAVEFRONT:
1225 case SIAtomicScope::SINGLETHREAD:
1234 if (VMCnt || LGKMCnt) {
1235 unsigned WaitCntImmediate =
1241 .
addImm(WaitCntImmediate);
1249 Scope == SIAtomicScope::WORKGROUP &&
1250 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1255 if (Pos == Position::AFTER)
1264 return !ST.isAmdPalOS() && !ST.isMesa3DOS();
1268 SIAtomicScope Scope,
1269 SIAtomicAddrSpace AddrSpace,
1270 Position Pos)
const {
1271 if (!InsertCacheInv)
1276 MachineBasicBlock &
MBB = *
MI->getParent();
1279 if (Pos == Position::AFTER)
1283 ? AMDGPU::BUFFER_WBINVL1_VOL
1284 : AMDGPU::BUFFER_WBINVL1;
1286 if (canAffectGlobalAddrSpace(AddrSpace)) {
1288 case SIAtomicScope::SYSTEM:
1289 if (
ST.hasGFX940Insts()) {
1305 if (
ST.hasGFX90AInsts()) {
1320 case SIAtomicScope::AGENT:
1321 if (
ST.hasGFX940Insts()) {
1336 case SIAtomicScope::WORKGROUP:
1337 if (
ST.isTgSplitEnabled()) {
1338 if (
ST.hasGFX940Insts()) {
1357 }
else if (
ST.hasGFX90AInsts()) {
1363 case SIAtomicScope::WAVEFRONT:
1364 case SIAtomicScope::SINGLETHREAD:
1381 if (Pos == Position::AFTER)
1388 SIAtomicScope Scope,
1389 SIAtomicAddrSpace AddrSpace,
1390 bool IsCrossAddrSpaceOrdering,
1391 Position Pos)
const {
1394 if (
ST.hasGFX90AInsts()) {
1395 MachineBasicBlock &
MBB = *
MI->getParent();
1398 if (Pos == Position::AFTER)
1401 if (canAffectGlobalAddrSpace(AddrSpace)) {
1403 case SIAtomicScope::SYSTEM:
1415 case SIAtomicScope::AGENT:
1416 if (
ST.hasGFX940Insts()) {
1427 case SIAtomicScope::WORKGROUP:
1428 case SIAtomicScope::WAVEFRONT:
1429 case SIAtomicScope::SINGLETHREAD:
1439 if (Pos == Position::AFTER)
1445 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
1446 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
1452bool SIGfx10CacheControl::enableLoadCacheBypass(
1454 SIAtomicAddrSpace AddrSpace)
const {
1458 if (canAffectGlobalAddrSpace(AddrSpace)) {
1460 case SIAtomicScope::SYSTEM:
1461 case SIAtomicScope::AGENT:
1468 case SIAtomicScope::WORKGROUP:
1473 if (!
ST.isCuModeEnabled())
1476 case SIAtomicScope::WAVEFRONT:
1477 case SIAtomicScope::SINGLETHREAD:
1495bool SIGfx10CacheControl::enableVolatileAndOrNonTemporal(
1497 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1508 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1516 if (
Op == SIMemOp::LOAD) {
1529 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1530 Position::AFTER, AtomicOrdering::Unordered,
1535 if (IsNonTemporal) {
1540 if (
Op == SIMemOp::STORE)
1555 SIAtomicScope Scope,
1556 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1557 bool IsCrossAddrSpaceOrdering,
1559 bool AtomicsOnly)
const {
1562 MachineBasicBlock &
MBB = *
MI->getParent();
1565 if (Pos == Position::AFTER)
1570 bool LGKMCnt =
false;
1572 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1573 SIAtomicAddrSpace::NONE) {
1575 case SIAtomicScope::SYSTEM:
1576 case SIAtomicScope::AGENT:
1577 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1579 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1582 case SIAtomicScope::WORKGROUP:
1592 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1594 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1598 case SIAtomicScope::WAVEFRONT:
1599 case SIAtomicScope::SINGLETHREAD:
1608 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1610 case SIAtomicScope::SYSTEM:
1611 case SIAtomicScope::AGENT:
1612 case SIAtomicScope::WORKGROUP:
1619 LGKMCnt |= IsCrossAddrSpaceOrdering;
1621 case SIAtomicScope::WAVEFRONT:
1622 case SIAtomicScope::SINGLETHREAD:
1631 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1633 case SIAtomicScope::SYSTEM:
1634 case SIAtomicScope::AGENT:
1641 LGKMCnt |= IsCrossAddrSpaceOrdering;
1643 case SIAtomicScope::WORKGROUP:
1644 case SIAtomicScope::WAVEFRONT:
1645 case SIAtomicScope::SINGLETHREAD:
1654 if (VMCnt || LGKMCnt) {
1655 unsigned WaitCntImmediate =
1661 .
addImm(WaitCntImmediate);
1669 Scope == SIAtomicScope::WORKGROUP &&
1670 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1682 if (Pos == Position::AFTER)
1689 SIAtomicScope Scope,
1690 SIAtomicAddrSpace AddrSpace,
1691 Position Pos)
const {
1692 if (!InsertCacheInv)
1697 MachineBasicBlock &
MBB = *
MI->getParent();
1700 if (Pos == Position::AFTER)
1703 if (canAffectGlobalAddrSpace(AddrSpace)) {
1705 case SIAtomicScope::SYSTEM:
1706 case SIAtomicScope::AGENT:
1714 case SIAtomicScope::WORKGROUP:
1719 if (!
ST.isCuModeEnabled()) {
1724 case SIAtomicScope::WAVEFRONT:
1725 case SIAtomicScope::SINGLETHREAD:
1740 if (Pos == Position::AFTER)
1748 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1763 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1776bool SIGfx12CacheControl::insertWaitsBeforeSystemScopeStore(
1780 MachineBasicBlock &
MBB = *
MI->getParent();
1784 if (
ST.hasImageInsts()) {
1795 SIAtomicScope Scope,
1796 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1797 bool IsCrossAddrSpaceOrdering,
1799 bool AtomicsOnly)
const {
1802 MachineBasicBlock &
MBB = *
MI->getParent();
1805 bool LOADCnt =
false;
1807 bool STORECnt =
false;
1809 if (Pos == Position::AFTER)
1812 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1813 SIAtomicAddrSpace::NONE) {
1815 case SIAtomicScope::SYSTEM:
1816 case SIAtomicScope::AGENT:
1817 case SIAtomicScope::CLUSTER:
1818 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1820 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1823 case SIAtomicScope::WORKGROUP:
1840 if (!
ST.isCuModeEnabled() ||
ST.hasGFX1250Insts() ||
1842 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1844 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1848 case SIAtomicScope::WAVEFRONT:
1849 case SIAtomicScope::SINGLETHREAD:
1858 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1860 case SIAtomicScope::SYSTEM:
1861 case SIAtomicScope::AGENT:
1862 case SIAtomicScope::CLUSTER:
1863 case SIAtomicScope::WORKGROUP:
1870 DSCnt |= IsCrossAddrSpaceOrdering;
1872 case SIAtomicScope::WAVEFRONT:
1873 case SIAtomicScope::SINGLETHREAD:
1894 if (!AtomicsOnly &&
ST.hasImageInsts()) {
1912 if (Pos == Position::AFTER)
1919 SIAtomicScope Scope,
1920 SIAtomicAddrSpace AddrSpace,
1921 Position Pos)
const {
1922 if (!InsertCacheInv)
1925 MachineBasicBlock &
MBB = *
MI->getParent();
1934 if (!canAffectGlobalAddrSpace(AddrSpace))
1939 case SIAtomicScope::SYSTEM:
1942 case SIAtomicScope::AGENT:
1945 case SIAtomicScope::CLUSTER:
1948 case SIAtomicScope::WORKGROUP:
1956 if (
ST.isCuModeEnabled())
1961 case SIAtomicScope::WAVEFRONT:
1962 case SIAtomicScope::SINGLETHREAD:
1969 if (Pos == Position::AFTER)
1974 if (Pos == Position::AFTER)
1981 SIAtomicScope Scope,
1982 SIAtomicAddrSpace AddrSpace,
1983 bool IsCrossAddrSpaceOrdering,
1984 Position Pos)
const {
1987 MachineBasicBlock &
MBB = *
MI->getParent();
1994 if (canAffectGlobalAddrSpace(AddrSpace)) {
1995 if (Pos == Position::AFTER)
2005 case SIAtomicScope::SYSTEM:
2010 case SIAtomicScope::AGENT:
2012 if (
ST.hasGFX1250Insts()) {
2018 case SIAtomicScope::CLUSTER:
2019 case SIAtomicScope::WORKGROUP:
2021 case SIAtomicScope::WAVEFRONT:
2022 case SIAtomicScope::SINGLETHREAD:
2029 if (Pos == Position::AFTER)
2036 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
2037 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
2043bool SIGfx12CacheControl::enableVolatileAndOrNonTemporal(
2045 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
2054 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
2061 }
else if (IsNonTemporal) {
2069 if (
ST.requiresWaitXCntForSingleAccessInstructions() &&
2071 MachineBasicBlock &
MBB = *
MI->getParent();
2081 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
2082 Position::AFTER, AtomicOrdering::Unordered,
2089bool SIGfx12CacheControl::finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
2090 assert(
MI.mayStore() &&
"Not a Store inst");
2091 const bool IsRMW = (
MI.mayLoad() &&
MI.mayStore());
2094 if (Atomic &&
ST.requiresWaitXCntForSingleAccessInstructions() &&
2096 MachineBasicBlock &
MBB = *
MI.getParent();
2105 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2111 if (
ST.requiresWaitsBeforeSystemScopeStores() && !Atomic &&
2113 Changed |= insertWaitsBeforeSystemScopeStore(
MI.getIterator());
2118bool SIGfx12CacheControl::handleCooperativeAtomic(MachineInstr &
MI)
const {
2119 if (!
ST.hasGFX1250Insts())
2123 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2124 assert(CPol &&
"No CPol operand?");
2132 SIAtomicScope Scope,
2133 SIAtomicAddrSpace AddrSpace)
const {
2136 if (canAffectGlobalAddrSpace(AddrSpace)) {
2138 case SIAtomicScope::SYSTEM:
2141 case SIAtomicScope::AGENT:
2144 case SIAtomicScope::CLUSTER:
2147 case SIAtomicScope::WORKGROUP:
2150 if (!
ST.isCuModeEnabled())
2153 case SIAtomicScope::WAVEFRONT:
2154 case SIAtomicScope::SINGLETHREAD:
2172bool SIMemoryLegalizer::removeAtomicPseudoMIs() {
2173 if (AtomicPseudoMIs.empty())
2176 for (
auto &
MI : AtomicPseudoMIs)
2177 MI->eraseFromParent();
2179 AtomicPseudoMIs.clear();
2183bool SIMemoryLegalizer::expandLoad(
const SIMemOpInfo &MOI,
2189 if (MOI.isAtomic()) {
2191 if (Order == AtomicOrdering::Monotonic ||
2192 Order == AtomicOrdering::Acquire ||
2193 Order == AtomicOrdering::SequentiallyConsistent) {
2194 Changed |= CC->enableLoadCacheBypass(
MI, MOI.getScope(),
2195 MOI.getOrderingAddrSpace());
2200 if (MOI.isCooperative())
2201 Changed |= CC->handleCooperativeAtomic(*
MI);
2203 if (Order == AtomicOrdering::SequentiallyConsistent)
2204 Changed |= CC->insertWait(
MI, MOI.getScope(), MOI.getOrderingAddrSpace(),
2205 SIMemOp::LOAD | SIMemOp::STORE,
2206 MOI.getIsCrossAddressSpaceOrdering(),
2207 Position::BEFORE, Order,
false);
2209 if (Order == AtomicOrdering::Acquire ||
2210 Order == AtomicOrdering::SequentiallyConsistent) {
2213 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2214 SIMemOp::LOAD, MOI.getIsCrossAddressSpaceOrdering(),
2215 Position::AFTER, Order,
true);
2216 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2217 MOI.getOrderingAddrSpace(),
2227 Changed |= CC->enableVolatileAndOrNonTemporal(
2228 MI, MOI.getInstrAddrSpace(), SIMemOp::LOAD, MOI.isVolatile(),
2229 MOI.isNonTemporal(), MOI.isLastUse());
2234bool SIMemoryLegalizer::expandStore(
const SIMemOpInfo &MOI,
2240 MachineInstr &StoreMI = *
MI;
2242 if (MOI.isAtomic()) {
2243 if (MOI.getOrdering() == AtomicOrdering::Monotonic ||
2244 MOI.getOrdering() == AtomicOrdering::Release ||
2245 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) {
2246 Changed |= CC->enableStoreCacheBypass(
MI, MOI.getScope(),
2247 MOI.getOrderingAddrSpace());
2252 if (MOI.isCooperative())
2253 Changed |= CC->handleCooperativeAtomic(*
MI);
2255 if (MOI.getOrdering() == AtomicOrdering::Release ||
2256 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
2257 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2258 MOI.getOrderingAddrSpace(),
2259 MOI.getIsCrossAddressSpaceOrdering(),
2262 Changed |= CC->finalizeStore(StoreMI,
true);
2269 Changed |= CC->enableVolatileAndOrNonTemporal(
2270 MI, MOI.getInstrAddrSpace(), SIMemOp::STORE, MOI.isVolatile(),
2271 MOI.isNonTemporal());
2275 Changed |= CC->finalizeStore(StoreMI,
false);
2279bool SIMemoryLegalizer::expandAtomicFence(
const SIMemOpInfo &MOI,
2281 assert(
MI->getOpcode() == AMDGPU::ATOMIC_FENCE);
2283 AtomicPseudoMIs.push_back(
MI);
2286 const SIAtomicAddrSpace OrderingAddrSpace = MOI.getOrderingAddrSpace();
2288 if (MOI.isAtomic()) {
2290 if (Order == AtomicOrdering::Acquire) {
2292 Changed |= CC->insertWait(
MI, MOI.getScope(), OrderingAddrSpace,
2293 SIMemOp::LOAD | SIMemOp::STORE,
2294 MOI.getIsCrossAddressSpaceOrdering(),
2295 Position::BEFORE, Order,
true);
2298 if (Order == AtomicOrdering::Release ||
2299 Order == AtomicOrdering::AcquireRelease ||
2300 Order == AtomicOrdering::SequentiallyConsistent)
2308 Changed |= CC->insertRelease(
MI, MOI.getScope(), OrderingAddrSpace,
2309 MOI.getIsCrossAddressSpaceOrdering(),
2317 if (Order == AtomicOrdering::Acquire ||
2318 Order == AtomicOrdering::AcquireRelease ||
2319 Order == AtomicOrdering::SequentiallyConsistent)
2320 Changed |= CC->insertAcquire(
MI, MOI.getScope(), OrderingAddrSpace,
2329bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
2334 MachineInstr &RMWMI = *
MI;
2336 if (MOI.isAtomic()) {
2338 if (Order == AtomicOrdering::Monotonic ||
2339 Order == AtomicOrdering::Acquire || Order == AtomicOrdering::Release ||
2340 Order == AtomicOrdering::AcquireRelease ||
2341 Order == AtomicOrdering::SequentiallyConsistent) {
2342 Changed |= CC->enableRMWCacheBypass(
MI, MOI.getScope(),
2343 MOI.getInstrAddrSpace());
2346 if (Order == AtomicOrdering::Release ||
2347 Order == AtomicOrdering::AcquireRelease ||
2348 Order == AtomicOrdering::SequentiallyConsistent ||
2349 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent)
2350 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2351 MOI.getOrderingAddrSpace(),
2352 MOI.getIsCrossAddressSpaceOrdering(),
2355 if (Order == AtomicOrdering::Acquire ||
2356 Order == AtomicOrdering::AcquireRelease ||
2357 Order == AtomicOrdering::SequentiallyConsistent ||
2358 MOI.getFailureOrdering() == AtomicOrdering::Acquire ||
2359 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) {
2362 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2363 isAtomicRet(*
MI) ? SIMemOp::LOAD : SIMemOp::STORE,
2364 MOI.getIsCrossAddressSpaceOrdering(), Position::AFTER,
2366 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2367 MOI.getOrderingAddrSpace(),
2371 Changed |= CC->finalizeStore(RMWMI,
true);
2378bool SIMemoryLegalizer::expandLDSDMA(
const SIMemOpInfo &MOI,
2390 return CC->enableVolatileAndOrNonTemporal(
2391 MI, MOI.getInstrAddrSpace(), OpKind, MOI.isVolatile(),
2392 MOI.isNonTemporal(), MOI.isLastUse());
2395bool SIMemoryLegalizerLegacy::runOnMachineFunction(MachineFunction &MF) {
2396 const MachineModuleInfo &MMI =
2397 getAnalysis<MachineModuleInfoWrapperPass>().getMMI();
2398 return SIMemoryLegalizer(MMI).run(MF);
2405 .getCachedResult<MachineModuleAnalysis>(
2407 assert(MMI &&
"MachineModuleAnalysis must be available");
2408 if (!SIMemoryLegalizer(MMI->getMMI()).run(MF))
2418 CC = SICacheControl::create(ST);
2420 for (
auto &
MBB : MF) {
2424 if (
MI->isBundle() &&
MI->mayLoadOrStore()) {
2427 I != E &&
I->isBundledWithPred(); ++
I) {
2428 I->unbundleFromPred();
2431 MO.setIsInternalRead(
false);
2434 MI->eraseFromParent();
2435 MI =
II->getIterator();
2441 if (
const auto &MOI = MOA.getLoadInfo(
MI)) {
2443 }
else if (
const auto &MOI = MOA.getStoreInfo(
MI)) {
2445 }
else if (
const auto &MOI = MOA.getLDSDMAInfo(
MI)) {
2447 }
else if (
const auto &MOI = MOA.getAtomicFenceInfo(
MI)) {
2449 }
else if (
const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(
MI)) {
2450 Changed |= expandAtomicCmpxchgOrRmw(*MOI,
MI);
2455 Changed |= removeAtomicPseudoMIs();
2461char SIMemoryLegalizerLegacy::
ID = 0;
2465 return new SIMemoryLegalizerLegacy();
static std::optional< LoadInfo > getLoadInfo(const MachineInstr &MI)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
const TargetInstrInfo & TII
AMDGPU address space definition.
Provides AMDGPU specific target descriptions.
AMDGPU Machine Module Info.
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Atomic ordering constants.
AMD GCN specific subclass of TargetSubtarget.
This header defines various interfaces for pass management in LLVM.
This file provides utility for Memory Model Relaxation Annotations (MMRAs).
uint64_t IntrinsicInst * II
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
static cl::opt< bool > AmdgcnSkipCacheInvalidations("amdgcn-skip-cache-invalidations", cl::init(false), cl::Hidden, cl::desc("Use this to skip inserting cache invalidating instructions."))
static bool canUseBUFFER_WBINVL1_VOL(const GCNSubtarget &ST)
static const uint32_t IV[8]
SyncScope::ID getWorkgroupSSID() const
SyncScope::ID getWavefrontSSID() const
SyncScope::ID getAgentSSID() const
SyncScope::ID getClusterOneAddressSpaceSSID() const
SyncScope::ID getClusterSSID() const
std::optional< bool > isSyncScopeInclusion(SyncScope::ID A, SyncScope::ID B) const
In AMDGPU target synchronization scopes are inclusive, meaning a larger synchronization scope is incl...
SyncScope::ID getAgentOneAddressSpaceSSID() const
SyncScope::ID getSingleThreadOneAddressSpaceSSID() const
SyncScope::ID getWavefrontOneAddressSpaceSSID() const
SyncScope::ID getSystemOneAddressSpaceSSID() const
SyncScope::ID getWorkgroupOneAddressSpaceSSID() const
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
LLVM_ABI void setPreservesCFG()
This function should be called by the pass, iff they do not:
Represents analyses that only rely on functions' control flow.
Diagnostic information for unsupported feature in backend.
FunctionPass class - This class is used to implement most global optimizations.
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
A helper class to return the specified delimiter string after the first invocation of operator String...
const MCInstrDesc & get(unsigned Opcode) const
Return the machine instruction descriptor that corresponds to the specified instruction opcode.
Instructions::iterator instr_iterator
MachineInstrBundleIterator< MachineInstr > iterator
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
Representation of each machine instruction.
Ty & getObjFileInfo()
Keep track of various per-module pieces of information for backends that would like to do so.
MachineOperand class - Representation of each machine instruction operand.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
PreservedAnalyses & preserveSet()
Mark an analysis set as preserved.
static bool isVMEM(const MachineInstr &MI)
static bool mayWriteLDSThroughDMA(const MachineInstr &MI)
static bool isBUF(const MachineInstr &MI)
static bool isAtomicRet(const MachineInstr &MI)
static bool isAtomic(const MachineInstr &MI)
static bool isLDSDMA(const MachineInstr &MI)
PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MFAM)
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
A wrapper around a string literal that serves as a proxy for constructing global tables of StringRefs...
StringMap - This is an unconventional map that is specialized for handling keys that are "strings",...
StringRef - Represent a constant reference to a string, i.e.
A raw_ostream that writes to an SmallVector or SmallString.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ BUFFER_STRIDED_POINTER
Address space for 192-bit fat buffer pointers with an additional index.
@ REGION_ADDRESS
Address space for region memory. (GDS)
@ LOCAL_ADDRESS
Address space for local memory.
@ FLAT_ADDRESS
Address space for flat memory.
@ GLOBAL_ADDRESS
Address space for global memory (RAT0, VTX0).
@ BUFFER_FAT_POINTER
Address space for 160-bit buffer fat pointers.
@ PRIVATE_ADDRESS
Address space for private memory.
@ BUFFER_RESOURCE
Address space for 128-bit buffer resources.
constexpr char IsVolatile[]
Key for Kernel::Arg::Metadata::mIsVolatile.
bool isGFX10(const MCSubtargetInfo &STI)
bool isGFX11(const MCSubtargetInfo &STI)
LLVM_ABI IsaVersion getIsaVersion(StringRef GPU)
unsigned encodeWaitcnt(const IsaVersion &Version, unsigned Vmcnt, unsigned Expcnt, unsigned Lgkmcnt)
Encodes Vmcnt, Expcnt and Lgkmcnt into Waitcnt for given isa Version.
unsigned getVmcntBitMask(const IsaVersion &Version)
unsigned getLgkmcntBitMask(const IsaVersion &Version)
unsigned getExpcntBitMask(const IsaVersion &Version)
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Undef
Value of the register doesn't matter.
@ SingleThread
Synchronized with respect to signal handlers executing in the same thread.
@ System
Synchronized with respect to all concurrently executing threads.
initializer< Ty > init(const Ty &Val)
PointerTypeMap run(const Module &M)
Compute the PointerTypeMap for the module M.
Scope
Defines the scope in which this symbol should be visible: Default – Visible in the public interface o...
NodeAddr< FuncNode * > Func
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
OuterAnalysisManagerProxy< ModuleAnalysisManager, MachineFunction > ModuleAnalysisManagerMachineFunctionProxy
Provide the ModuleAnalysisManager to Function proxy.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
char & SIMemoryLegalizerID
LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE()
AnalysisManager< MachineFunction > MachineFunctionAnalysisManager
bool isReleaseOrStronger(AtomicOrdering AO)
LLVM_ABI PreservedAnalyses getMachineFunctionPassPreservedAnalyses()
Returns the minimum set of Analyses that all machine function passes must preserve.
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
AtomicOrdering getMergedAtomicOrdering(AtomicOrdering AO, AtomicOrdering Other)
Return a single atomic ordering that is at least as strong as both the AO and Other orderings for an ...
static const MachineMemOperand::Flags MOCooperative
Mark the MMO of cooperative load/store atomics.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ LLVM_MARK_AS_BITMASK_ENUM
DWARFExpression::Operation Op
static const MachineMemOperand::Flags MOLastUse
Mark the MMO of a load as the last use.
FunctionPass * createSIMemoryLegalizerPass()