35#define DEBUG_TYPE "si-memory-legalizer"
36#define PASS_NAME "SI Memory Legalizer"
40 cl::desc(
"Use this to skip inserting cache invalidating instructions."));
62enum class SIAtomicScope {
74enum class SIAtomicAddrSpace {
83 FLAT = GLOBAL |
LDS | SCRATCH,
86 ATOMIC = GLOBAL |
LDS | SCRATCH | GDS,
89 ALL = GLOBAL |
LDS | SCRATCH | GDS | OTHER,
94class SIMemOpInfo final {
97 friend class SIMemOpAccess;
101 SIAtomicScope Scope = SIAtomicScope::SYSTEM;
102 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
103 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
104 bool IsCrossAddressSpaceOrdering =
false;
105 bool IsVolatile =
false;
106 bool IsNonTemporal =
false;
107 bool IsLastUse =
false;
108 bool IsCooperative =
false;
112 const GCNSubtarget &ST,
114 SIAtomicScope Scope = SIAtomicScope::SYSTEM,
115 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC,
116 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL,
117 bool IsCrossAddressSpaceOrdering =
true,
118 AtomicOrdering FailureOrdering = AtomicOrdering::SequentiallyConsistent,
119 bool IsVolatile =
false,
bool IsNonTemporal =
false,
120 bool IsLastUse =
false,
bool IsCooperative =
false)
121 : Ordering(Ordering), FailureOrdering(FailureOrdering), Scope(Scope),
122 OrderingAddrSpace(OrderingAddrSpace), InstrAddrSpace(InstrAddrSpace),
123 IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering),
124 IsVolatile(IsVolatile), IsNonTemporal(IsNonTemporal),
125 IsLastUse(IsLastUse), IsCooperative(IsCooperative) {
127 if (Ordering == AtomicOrdering::NotAtomic) {
128 assert(!IsCooperative &&
"Cannot be cooperative & non-atomic!");
129 assert(Scope == SIAtomicScope::NONE &&
130 OrderingAddrSpace == SIAtomicAddrSpace::NONE &&
131 !IsCrossAddressSpaceOrdering &&
132 FailureOrdering == AtomicOrdering::NotAtomic);
136 assert(Scope != SIAtomicScope::NONE &&
137 (OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
138 SIAtomicAddrSpace::NONE &&
139 (InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) !=
140 SIAtomicAddrSpace::NONE);
145 if ((OrderingAddrSpace == InstrAddrSpace) &&
147 this->IsCrossAddressSpaceOrdering =
false;
151 if ((InstrAddrSpace & ~SIAtomicAddrSpace::SCRATCH) ==
152 SIAtomicAddrSpace::NONE) {
153 this->Scope = std::min(Scope, SIAtomicScope::SINGLETHREAD);
154 }
else if ((InstrAddrSpace &
155 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS)) ==
156 SIAtomicAddrSpace::NONE) {
157 this->Scope = std::min(Scope, SIAtomicScope::WORKGROUP);
158 }
else if ((InstrAddrSpace &
159 ~(SIAtomicAddrSpace::SCRATCH | SIAtomicAddrSpace::LDS |
160 SIAtomicAddrSpace::GDS)) == SIAtomicAddrSpace::NONE) {
161 this->Scope = std::min(Scope, SIAtomicScope::AGENT);
166 if (this->Scope == SIAtomicScope::CLUSTER && !
ST.hasClusters())
167 this->Scope = SIAtomicScope::AGENT;
173 SIAtomicScope getScope()
const {
186 return FailureOrdering;
191 SIAtomicAddrSpace getInstrAddrSpace()
const {
192 return InstrAddrSpace;
197 SIAtomicAddrSpace getOrderingAddrSpace()
const {
198 return OrderingAddrSpace;
203 bool getIsCrossAddressSpaceOrdering()
const {
204 return IsCrossAddressSpaceOrdering;
209 bool isVolatile()
const {
215 bool isNonTemporal()
const {
216 return IsNonTemporal;
221 bool isLastUse()
const {
return IsLastUse; }
224 bool isCooperative()
const {
return IsCooperative; }
228 bool isAtomic()
const {
229 return Ordering != AtomicOrdering::NotAtomic;
234class SIMemOpAccess final {
236 const AMDGPUMachineModuleInfo *MMI =
nullptr;
237 const GCNSubtarget &ST;
241 const char *Msg)
const;
247 std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
248 toSIAtomicScope(
SyncScope::ID SSID, SIAtomicAddrSpace InstrAddrSpace)
const;
251 SIAtomicAddrSpace toSIAtomicAddrSpace(
unsigned AS)
const;
255 std::optional<SIMemOpInfo>
261 SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI,
const GCNSubtarget &ST);
264 std::optional<SIMemOpInfo>
269 std::optional<SIMemOpInfo>
274 std::optional<SIMemOpInfo>
279 std::optional<SIMemOpInfo>
285 std::optional<SIMemOpInfo>
289class SICacheControl {
293 const GCNSubtarget &ST;
296 const SIInstrInfo *TII =
nullptr;
303 SICacheControl(
const GCNSubtarget &ST);
308 unsigned Bits)
const;
312 bool canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const;
318 static std::unique_ptr<SICacheControl> create(
const GCNSubtarget &ST);
325 SIAtomicAddrSpace AddrSpace)
const = 0;
332 SIAtomicAddrSpace AddrSpace)
const = 0;
339 SIAtomicAddrSpace AddrSpace)
const = 0;
345 SIAtomicAddrSpace AddrSpace,
346 SIMemOp
Op,
bool IsVolatile,
348 bool IsLastUse =
false)
const = 0;
355 virtual bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
360 virtual bool handleCooperativeAtomic(MachineInstr &
MI)
const {
362 "cooperative atomics are not available on this architecture");
375 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
376 bool IsCrossAddrSpaceOrdering, Position Pos,
386 SIAtomicAddrSpace AddrSpace,
387 Position Pos)
const = 0;
397 SIAtomicAddrSpace AddrSpace,
398 bool IsCrossAddrSpaceOrdering,
399 Position Pos)
const = 0;
402 virtual ~SICacheControl() =
default;
407class SIGfx6CacheControl final :
public SICacheControl {
410 SIGfx6CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
414 SIAtomicAddrSpace AddrSpace)
const override;
418 SIAtomicAddrSpace AddrSpace)
const override;
422 SIAtomicAddrSpace AddrSpace)
const override;
425 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
426 bool IsVolatile,
bool IsNonTemporal,
427 bool IsLastUse)
const override;
430 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
431 bool IsCrossAddrSpaceOrdering, Position Pos,
436 SIAtomicAddrSpace AddrSpace,
437 Position Pos)
const override;
441 SIAtomicAddrSpace AddrSpace,
442 bool IsCrossAddrSpaceOrdering,
443 Position Pos)
const override;
447class SIGfx10CacheControl final :
public SICacheControl {
449 SIGfx10CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {}
453 SIAtomicAddrSpace AddrSpace)
const override;
457 SIAtomicAddrSpace AddrSpace)
const override {
463 SIAtomicAddrSpace AddrSpace)
const override {
468 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
469 bool IsVolatile,
bool IsNonTemporal,
470 bool IsLastUse)
const override;
473 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
474 bool IsCrossAddrSpaceOrdering, Position Pos,
478 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
481 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
482 Position Pos)
const override {
483 return insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
484 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
489class SIGfx12CacheControl final :
public SICacheControl {
511 SIAtomicScope Scope, SIAtomicAddrSpace AddrSpace)
const;
514 SIGfx12CacheControl(
const GCNSubtarget &ST) : SICacheControl(
ST) {
517 assert(!
ST.hasGFX1250Insts() ||
ST.isCuModeEnabled());
521 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
522 bool IsCrossAddrSpaceOrdering, Position Pos,
526 SIAtomicAddrSpace AddrSpace, Position Pos)
const override;
529 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
530 bool IsVolatile,
bool IsNonTemporal,
531 bool IsLastUse)
const override;
533 bool finalizeStore(MachineInstr &
MI,
bool Atomic)
const override;
535 bool handleCooperativeAtomic(MachineInstr &
MI)
const override;
538 SIAtomicAddrSpace AddrSpace,
bool IsCrossAddrSpaceOrdering,
539 Position Pos)
const override;
543 SIAtomicAddrSpace AddrSpace)
const override {
544 return setAtomicScope(
MI, Scope, AddrSpace);
549 SIAtomicAddrSpace AddrSpace)
const override {
550 return setAtomicScope(
MI, Scope, AddrSpace);
555 SIAtomicAddrSpace AddrSpace)
const override {
556 return setAtomicScope(
MI, Scope, AddrSpace);
560class SIMemoryLegalizer final {
562 const MachineModuleInfo &MMI;
564 std::unique_ptr<SICacheControl> CC =
nullptr;
567 std::list<MachineBasicBlock::iterator> AtomicPseudoMIs;
571 bool isAtomicRet(
const MachineInstr &
MI)
const {
577 bool removeAtomicPseudoMIs();
581 bool expandLoad(
const SIMemOpInfo &MOI,
585 bool expandStore(
const SIMemOpInfo &MOI,
589 bool expandAtomicFence(
const SIMemOpInfo &MOI,
593 bool expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
600 SIMemoryLegalizer(
const MachineModuleInfo &MMI) : MMI(MMI) {};
601 bool run(MachineFunction &MF);
608 SIMemoryLegalizerLegacy() : MachineFunctionPass(ID) {}
610 void getAnalysisUsage(AnalysisUsage &AU)
const override {
615 StringRef getPassName()
const override {
619 bool runOnMachineFunction(MachineFunction &MF)
override;
623 {
"global", SIAtomicAddrSpace::GLOBAL},
624 {
"local", SIAtomicAddrSpace::LDS},
632 OS <<
"unknown address space '" << AS <<
"'; expected one of ";
634 for (
const auto &[Name, Val] : ASNames)
635 OS <<
LS <<
'\'' <<
Name <<
'\'';
643static std::optional<SIAtomicAddrSpace>
645 static constexpr StringLiteral FenceASPrefix =
"amdgpu-synchronize-as";
651 SIAtomicAddrSpace
Result = SIAtomicAddrSpace::NONE;
652 for (
const auto &[Prefix, Suffix] : MMRA) {
653 if (Prefix != FenceASPrefix)
656 if (
auto It = ASNames.find(Suffix); It != ASNames.end())
659 diagnoseUnknownMMRAASName(
MI, Suffix);
662 if (Result == SIAtomicAddrSpace::NONE)
671 const char *Msg)
const {
673 Func.getContext().diagnose(
674 DiagnosticInfoUnsupported(Func, Msg,
MI->getDebugLoc()));
677std::optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
679 SIAtomicAddrSpace InstrAddrSpace)
const {
681 return std::tuple(SIAtomicScope::SYSTEM, SIAtomicAddrSpace::ATOMIC,
true);
683 return std::tuple(SIAtomicScope::AGENT, SIAtomicAddrSpace::ATOMIC,
true);
685 return std::tuple(SIAtomicScope::CLUSTER, SIAtomicAddrSpace::ATOMIC,
true);
687 return std::tuple(SIAtomicScope::WORKGROUP, SIAtomicAddrSpace::ATOMIC,
690 return std::tuple(SIAtomicScope::WAVEFRONT, SIAtomicAddrSpace::ATOMIC,
693 return std::tuple(SIAtomicScope::SINGLETHREAD, SIAtomicAddrSpace::ATOMIC,
696 return std::tuple(SIAtomicScope::SYSTEM,
697 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
699 return std::tuple(SIAtomicScope::AGENT,
700 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
702 return std::tuple(SIAtomicScope::CLUSTER,
703 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
705 return std::tuple(SIAtomicScope::WORKGROUP,
706 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
708 return std::tuple(SIAtomicScope::WAVEFRONT,
709 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
711 return std::tuple(SIAtomicScope::SINGLETHREAD,
712 SIAtomicAddrSpace::ATOMIC & InstrAddrSpace,
false);
716SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(
unsigned AS)
const {
718 return SIAtomicAddrSpace::FLAT;
720 return SIAtomicAddrSpace::GLOBAL;
722 return SIAtomicAddrSpace::LDS;
724 return SIAtomicAddrSpace::SCRATCH;
726 return SIAtomicAddrSpace::GDS;
729 return SIAtomicAddrSpace::GLOBAL;
731 return SIAtomicAddrSpace::OTHER;
734SIMemOpAccess::SIMemOpAccess(
const AMDGPUMachineModuleInfo &MMI_,
735 const GCNSubtarget &ST)
736 : MMI(&MMI_),
ST(
ST) {}
738std::optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO(
740 assert(
MI->getNumMemOperands() > 0);
745 SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
746 bool IsNonTemporal =
true;
748 bool IsLastUse =
false;
749 bool IsCooperative =
false;
753 for (
const auto &MMO :
MI->memoperands()) {
754 IsNonTemporal &= MMO->isNonTemporal();
756 IsLastUse |= MMO->getFlags() &
MOLastUse;
759 toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace());
761 if (OpOrdering != AtomicOrdering::NotAtomic) {
762 const auto &IsSyncScopeInclusion =
764 if (!IsSyncScopeInclusion) {
765 reportUnsupported(
MI,
766 "Unsupported non-inclusive atomic synchronization scope");
770 SSID = *IsSyncScopeInclusion ? SSID : MMO->getSyncScopeID();
772 assert(MMO->getFailureOrdering() != AtomicOrdering::Release &&
773 MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease);
779 SIAtomicScope
Scope = SIAtomicScope::NONE;
780 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
781 bool IsCrossAddressSpaceOrdering =
false;
782 if (Ordering != AtomicOrdering::NotAtomic) {
783 auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace);
785 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
788 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
790 if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) ||
791 ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace) ||
792 ((InstrAddrSpace & SIAtomicAddrSpace::ATOMIC) == SIAtomicAddrSpace::NONE)) {
793 reportUnsupported(
MI,
"Unsupported atomic address space");
797 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace, InstrAddrSpace,
798 IsCrossAddressSpaceOrdering, FailureOrdering, IsVolatile,
799 IsNonTemporal, IsLastUse, IsCooperative);
802std::optional<SIMemOpInfo>
806 if (!(
MI->mayLoad() && !
MI->mayStore()))
810 if (
MI->getNumMemOperands() == 0)
811 return SIMemOpInfo(ST);
813 return constructFromMIWithMMO(
MI);
816std::optional<SIMemOpInfo>
820 if (!(!
MI->mayLoad() &&
MI->mayStore()))
824 if (
MI->getNumMemOperands() == 0)
825 return SIMemOpInfo(ST);
827 return constructFromMIWithMMO(
MI);
830std::optional<SIMemOpInfo>
834 if (
MI->getOpcode() != AMDGPU::ATOMIC_FENCE)
841 auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC);
843 reportUnsupported(
MI,
"Unsupported atomic synchronization scope");
847 SIAtomicScope
Scope = SIAtomicScope::NONE;
848 SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
849 bool IsCrossAddressSpaceOrdering =
false;
850 std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
853 if (OrderingAddrSpace != SIAtomicAddrSpace::ATOMIC) {
858 reportUnsupported(
MI,
"Unsupported atomic address space");
862 auto SynchronizeAS = getSynchronizeAddrSpaceMD(*
MI);
864 OrderingAddrSpace = *SynchronizeAS;
866 return SIMemOpInfo(ST, Ordering, Scope, OrderingAddrSpace,
867 SIAtomicAddrSpace::ATOMIC, IsCrossAddressSpaceOrdering,
868 AtomicOrdering::NotAtomic);
871std::optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo(
875 if (!(
MI->mayLoad() &&
MI->mayStore()))
879 if (
MI->getNumMemOperands() == 0)
880 return SIMemOpInfo(ST);
882 return constructFromMIWithMMO(
MI);
885std::optional<SIMemOpInfo>
892 return constructFromMIWithMMO(
MI);
895SICacheControl::SICacheControl(
const GCNSubtarget &ST) :
ST(
ST) {
896 TII =
ST.getInstrInfo();
902 unsigned Bits)
const {
903 MachineOperand *CPol =
TII->getNamedOperand(*
MI, AMDGPU::OpName::cpol);
907 CPol->setImm(
CPol->getImm() | Bits);
911bool SICacheControl::canAffectGlobalAddrSpace(SIAtomicAddrSpace AS)
const {
912 assert((!
ST.hasGloballyAddressableScratch() ||
913 (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE ||
914 (AS & SIAtomicAddrSpace::SCRATCH) == SIAtomicAddrSpace::NONE) &&
915 "scratch instructions should already be replaced by flat "
916 "instructions if GloballyAddressableScratch is enabled");
917 return (AS & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE;
921std::unique_ptr<SICacheControl> SICacheControl::create(
const GCNSubtarget &ST) {
922 GCNSubtarget::Generation Generation =
ST.getGeneration();
923 if (Generation < AMDGPUSubtarget::GFX10)
924 return std::make_unique<SIGfx6CacheControl>(ST);
925 if (Generation < AMDGPUSubtarget::GFX12)
926 return std::make_unique<SIGfx10CacheControl>(ST);
927 return std::make_unique<SIGfx12CacheControl>(ST);
930bool SIGfx6CacheControl::enableLoadCacheBypass(
933 SIAtomicAddrSpace AddrSpace)
const {
936 if (!canAffectGlobalAddrSpace(AddrSpace)) {
948 case SIAtomicScope::SYSTEM:
949 if (
ST.hasGFX940Insts()) {
955 case SIAtomicScope::AGENT:
956 if (
ST.hasGFX940Insts()) {
965 case SIAtomicScope::WORKGROUP:
966 if (
ST.hasGFX940Insts()) {
973 }
else if (
ST.hasGFX90AInsts()) {
978 if (
ST.isTgSplitEnabled())
982 case SIAtomicScope::WAVEFRONT:
983 case SIAtomicScope::SINGLETHREAD:
993bool SIGfx6CacheControl::enableStoreCacheBypass(
996 SIAtomicAddrSpace AddrSpace)
const {
1004 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1006 case SIAtomicScope::SYSTEM:
1010 case SIAtomicScope::AGENT:
1014 case SIAtomicScope::WORKGROUP:
1018 case SIAtomicScope::WAVEFRONT:
1019 case SIAtomicScope::SINGLETHREAD:
1037bool SIGfx6CacheControl::enableRMWCacheBypass(
1039 SIAtomicScope Scope,
1040 SIAtomicAddrSpace AddrSpace)
const {
1050 if (
ST.hasGFX940Insts() && canAffectGlobalAddrSpace(AddrSpace)) {
1052 case SIAtomicScope::SYSTEM:
1056 case SIAtomicScope::AGENT:
1057 case SIAtomicScope::WORKGROUP:
1058 case SIAtomicScope::WAVEFRONT:
1059 case SIAtomicScope::SINGLETHREAD:
1073bool SIGfx6CacheControl::enableVolatileAndOrNonTemporal(
1075 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1085 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1090 if (
ST.hasGFX940Insts()) {
1093 }
else if (
Op == SIMemOp::LOAD) {
1105 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1106 Position::AFTER, AtomicOrdering::Unordered,
1112 if (IsNonTemporal) {
1113 if (
ST.hasGFX940Insts()) {
1127 SIAtomicScope Scope,
1128 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1129 bool IsCrossAddrSpaceOrdering, Position Pos,
1131 bool AtomicsOnly)
const {
1134 MachineBasicBlock &
MBB = *
MI->getParent();
1137 if (Pos == Position::AFTER)
1141 if (
ST.hasGFX90AInsts() &&
ST.isTgSplitEnabled()) {
1149 if (((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH |
1150 SIAtomicAddrSpace::GDS)) != SIAtomicAddrSpace::NONE) &&
1151 (Scope == SIAtomicScope::WORKGROUP)) {
1153 Scope = SIAtomicScope::AGENT;
1157 AddrSpace &= ~SIAtomicAddrSpace
::LDS;
1161 bool LGKMCnt =
false;
1163 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1164 SIAtomicAddrSpace::NONE) {
1166 case SIAtomicScope::SYSTEM:
1167 case SIAtomicScope::AGENT:
1170 case SIAtomicScope::WORKGROUP:
1171 case SIAtomicScope::WAVEFRONT:
1172 case SIAtomicScope::SINGLETHREAD:
1181 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1183 case SIAtomicScope::SYSTEM:
1184 case SIAtomicScope::AGENT:
1185 case SIAtomicScope::WORKGROUP:
1192 LGKMCnt |= IsCrossAddrSpaceOrdering;
1194 case SIAtomicScope::WAVEFRONT:
1195 case SIAtomicScope::SINGLETHREAD:
1204 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1206 case SIAtomicScope::SYSTEM:
1207 case SIAtomicScope::AGENT:
1214 LGKMCnt |= IsCrossAddrSpaceOrdering;
1216 case SIAtomicScope::WORKGROUP:
1217 case SIAtomicScope::WAVEFRONT:
1218 case SIAtomicScope::SINGLETHREAD:
1227 if (VMCnt || LGKMCnt) {
1228 unsigned WaitCntImmediate =
1234 .
addImm(WaitCntImmediate);
1242 Scope == SIAtomicScope::WORKGROUP &&
1243 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1248 if (Pos == Position::AFTER)
1257 return !ST.isAmdPalOS() && !ST.isMesa3DOS();
1261 SIAtomicScope Scope,
1262 SIAtomicAddrSpace AddrSpace,
1263 Position Pos)
const {
1264 if (!InsertCacheInv)
1269 MachineBasicBlock &
MBB = *
MI->getParent();
1272 if (Pos == Position::AFTER)
1276 ? AMDGPU::BUFFER_WBINVL1_VOL
1277 : AMDGPU::BUFFER_WBINVL1;
1279 if (canAffectGlobalAddrSpace(AddrSpace)) {
1281 case SIAtomicScope::SYSTEM:
1282 if (
ST.hasGFX940Insts()) {
1298 if (
ST.hasGFX90AInsts()) {
1313 case SIAtomicScope::AGENT:
1314 if (
ST.hasGFX940Insts()) {
1329 case SIAtomicScope::WORKGROUP:
1330 if (
ST.isTgSplitEnabled()) {
1331 if (
ST.hasGFX940Insts()) {
1350 }
else if (
ST.hasGFX90AInsts()) {
1356 case SIAtomicScope::WAVEFRONT:
1357 case SIAtomicScope::SINGLETHREAD:
1374 if (Pos == Position::AFTER)
1381 SIAtomicScope Scope,
1382 SIAtomicAddrSpace AddrSpace,
1383 bool IsCrossAddrSpaceOrdering,
1384 Position Pos)
const {
1387 if (
ST.hasGFX90AInsts()) {
1388 MachineBasicBlock &
MBB = *
MI->getParent();
1391 if (Pos == Position::AFTER)
1394 if (canAffectGlobalAddrSpace(AddrSpace)) {
1396 case SIAtomicScope::SYSTEM:
1408 case SIAtomicScope::AGENT:
1409 if (
ST.hasGFX940Insts()) {
1420 case SIAtomicScope::WORKGROUP:
1421 case SIAtomicScope::WAVEFRONT:
1422 case SIAtomicScope::SINGLETHREAD:
1432 if (Pos == Position::AFTER)
1438 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
1439 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
1445bool SIGfx10CacheControl::enableLoadCacheBypass(
1447 SIAtomicAddrSpace AddrSpace)
const {
1451 if (canAffectGlobalAddrSpace(AddrSpace)) {
1453 case SIAtomicScope::SYSTEM:
1454 case SIAtomicScope::AGENT:
1461 case SIAtomicScope::WORKGROUP:
1466 if (!
ST.isCuModeEnabled())
1469 case SIAtomicScope::WAVEFRONT:
1470 case SIAtomicScope::SINGLETHREAD:
1488bool SIGfx10CacheControl::enableVolatileAndOrNonTemporal(
1490 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
1501 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
1509 if (
Op == SIMemOp::LOAD) {
1522 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
1523 Position::AFTER, AtomicOrdering::Unordered,
1528 if (IsNonTemporal) {
1533 if (
Op == SIMemOp::STORE)
1548 SIAtomicScope Scope,
1549 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1550 bool IsCrossAddrSpaceOrdering,
1552 bool AtomicsOnly)
const {
1555 MachineBasicBlock &
MBB = *
MI->getParent();
1558 if (Pos == Position::AFTER)
1563 bool LGKMCnt =
false;
1565 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1566 SIAtomicAddrSpace::NONE) {
1568 case SIAtomicScope::SYSTEM:
1569 case SIAtomicScope::AGENT:
1570 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1572 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1575 case SIAtomicScope::WORKGROUP:
1585 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1587 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1591 case SIAtomicScope::WAVEFRONT:
1592 case SIAtomicScope::SINGLETHREAD:
1601 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1603 case SIAtomicScope::SYSTEM:
1604 case SIAtomicScope::AGENT:
1605 case SIAtomicScope::WORKGROUP:
1612 LGKMCnt |= IsCrossAddrSpaceOrdering;
1614 case SIAtomicScope::WAVEFRONT:
1615 case SIAtomicScope::SINGLETHREAD:
1624 if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
1626 case SIAtomicScope::SYSTEM:
1627 case SIAtomicScope::AGENT:
1634 LGKMCnt |= IsCrossAddrSpaceOrdering;
1636 case SIAtomicScope::WORKGROUP:
1637 case SIAtomicScope::WAVEFRONT:
1638 case SIAtomicScope::SINGLETHREAD:
1647 if (VMCnt || LGKMCnt) {
1648 unsigned WaitCntImmediate =
1654 .
addImm(WaitCntImmediate);
1662 Scope == SIAtomicScope::WORKGROUP &&
1663 (AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1675 if (Pos == Position::AFTER)
1682 SIAtomicScope Scope,
1683 SIAtomicAddrSpace AddrSpace,
1684 Position Pos)
const {
1685 if (!InsertCacheInv)
1690 MachineBasicBlock &
MBB = *
MI->getParent();
1693 if (Pos == Position::AFTER)
1696 if (canAffectGlobalAddrSpace(AddrSpace)) {
1698 case SIAtomicScope::SYSTEM:
1699 case SIAtomicScope::AGENT:
1707 case SIAtomicScope::WORKGROUP:
1712 if (!
ST.isCuModeEnabled()) {
1717 case SIAtomicScope::WAVEFRONT:
1718 case SIAtomicScope::SINGLETHREAD:
1733 if (Pos == Position::AFTER)
1741 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1756 MachineOperand *
CPol =
TII->getNamedOperand(*
MI, OpName::cpol);
1769bool SIGfx12CacheControl::insertWaitsBeforeSystemScopeStore(
1773 MachineBasicBlock &
MBB = *
MI->getParent();
1777 if (
ST.hasImageInsts()) {
1788 SIAtomicScope Scope,
1789 SIAtomicAddrSpace AddrSpace, SIMemOp
Op,
1790 bool IsCrossAddrSpaceOrdering,
1792 bool AtomicsOnly)
const {
1795 MachineBasicBlock &
MBB = *
MI->getParent();
1798 bool LOADCnt =
false;
1800 bool STORECnt =
false;
1802 if (Pos == Position::AFTER)
1805 if ((AddrSpace & (SIAtomicAddrSpace::GLOBAL | SIAtomicAddrSpace::SCRATCH)) !=
1806 SIAtomicAddrSpace::NONE) {
1808 case SIAtomicScope::SYSTEM:
1809 case SIAtomicScope::AGENT:
1810 case SIAtomicScope::CLUSTER:
1811 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1813 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1816 case SIAtomicScope::WORKGROUP:
1833 if (!
ST.isCuModeEnabled() ||
ST.hasGFX1250Insts() ||
1835 if ((
Op & SIMemOp::LOAD) != SIMemOp::NONE)
1837 if ((
Op & SIMemOp::STORE) != SIMemOp::NONE)
1841 case SIAtomicScope::WAVEFRONT:
1842 case SIAtomicScope::SINGLETHREAD:
1851 if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
1853 case SIAtomicScope::SYSTEM:
1854 case SIAtomicScope::AGENT:
1855 case SIAtomicScope::CLUSTER:
1856 case SIAtomicScope::WORKGROUP:
1863 DSCnt |= IsCrossAddrSpaceOrdering;
1865 case SIAtomicScope::WAVEFRONT:
1866 case SIAtomicScope::SINGLETHREAD:
1887 if (!AtomicsOnly &&
ST.hasImageInsts()) {
1905 if (Pos == Position::AFTER)
1912 SIAtomicScope Scope,
1913 SIAtomicAddrSpace AddrSpace,
1914 Position Pos)
const {
1915 if (!InsertCacheInv)
1918 MachineBasicBlock &
MBB = *
MI->getParent();
1927 if (!canAffectGlobalAddrSpace(AddrSpace))
1932 case SIAtomicScope::SYSTEM:
1935 case SIAtomicScope::AGENT:
1938 case SIAtomicScope::CLUSTER:
1941 case SIAtomicScope::WORKGROUP:
1949 if (
ST.isCuModeEnabled())
1954 case SIAtomicScope::WAVEFRONT:
1955 case SIAtomicScope::SINGLETHREAD:
1962 if (Pos == Position::AFTER)
1967 if (Pos == Position::AFTER)
1974 SIAtomicScope Scope,
1975 SIAtomicAddrSpace AddrSpace,
1976 bool IsCrossAddrSpaceOrdering,
1977 Position Pos)
const {
1980 MachineBasicBlock &
MBB = *
MI->getParent();
1987 if (canAffectGlobalAddrSpace(AddrSpace)) {
1988 if (Pos == Position::AFTER)
1998 case SIAtomicScope::SYSTEM:
2003 case SIAtomicScope::AGENT:
2005 if (
ST.hasGFX1250Insts()) {
2011 case SIAtomicScope::CLUSTER:
2012 case SIAtomicScope::WORKGROUP:
2014 case SIAtomicScope::WAVEFRONT:
2015 case SIAtomicScope::SINGLETHREAD:
2022 if (Pos == Position::AFTER)
2029 Changed |= insertWait(
MI, Scope, AddrSpace, SIMemOp::LOAD | SIMemOp::STORE,
2030 IsCrossAddrSpaceOrdering, Pos, AtomicOrdering::Release,
2036bool SIGfx12CacheControl::enableVolatileAndOrNonTemporal(
2038 bool IsVolatile,
bool IsNonTemporal,
bool IsLastUse =
false)
const {
2047 assert(
Op == SIMemOp::LOAD ||
Op == SIMemOp::STORE);
2054 }
else if (IsNonTemporal) {
2067 Changed |= insertWait(
MI, SIAtomicScope::SYSTEM, AddrSpace,
Op,
false,
2068 Position::AFTER, AtomicOrdering::Unordered,
2075bool SIGfx12CacheControl::finalizeStore(MachineInstr &
MI,
bool Atomic)
const {
2076 assert(
MI.mayStore() &&
"Not a Store inst");
2077 const bool IsRMW = (
MI.mayLoad() &&
MI.mayStore());
2082 if (Atomic &&
ST.requiresWaitXCntBeforeAtomicStores() &&
TII->isFLAT(
MI)) {
2083 MachineBasicBlock &
MBB = *
MI.getParent();
2092 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2098 if (
ST.requiresWaitsBeforeSystemScopeStores() && !Atomic &&
2100 Changed |= insertWaitsBeforeSystemScopeStore(
MI.getIterator());
2105bool SIGfx12CacheControl::handleCooperativeAtomic(MachineInstr &
MI)
const {
2106 if (!
ST.hasGFX1250Insts())
2110 MachineOperand *
CPol =
TII->getNamedOperand(
MI, OpName::cpol);
2111 assert(CPol &&
"No CPol operand?");
2119 SIAtomicScope Scope,
2120 SIAtomicAddrSpace AddrSpace)
const {
2123 if (canAffectGlobalAddrSpace(AddrSpace)) {
2125 case SIAtomicScope::SYSTEM:
2128 case SIAtomicScope::AGENT:
2131 case SIAtomicScope::CLUSTER:
2134 case SIAtomicScope::WORKGROUP:
2137 if (!
ST.isCuModeEnabled())
2140 case SIAtomicScope::WAVEFRONT:
2141 case SIAtomicScope::SINGLETHREAD:
2159bool SIMemoryLegalizer::removeAtomicPseudoMIs() {
2160 if (AtomicPseudoMIs.empty())
2163 for (
auto &
MI : AtomicPseudoMIs)
2164 MI->eraseFromParent();
2166 AtomicPseudoMIs.clear();
2170bool SIMemoryLegalizer::expandLoad(
const SIMemOpInfo &MOI,
2176 if (MOI.isAtomic()) {
2178 if (Order == AtomicOrdering::Monotonic ||
2179 Order == AtomicOrdering::Acquire ||
2180 Order == AtomicOrdering::SequentiallyConsistent) {
2181 Changed |= CC->enableLoadCacheBypass(
MI, MOI.getScope(),
2182 MOI.getOrderingAddrSpace());
2187 if (MOI.isCooperative())
2188 Changed |= CC->handleCooperativeAtomic(*
MI);
2190 if (Order == AtomicOrdering::SequentiallyConsistent)
2191 Changed |= CC->insertWait(
MI, MOI.getScope(), MOI.getOrderingAddrSpace(),
2192 SIMemOp::LOAD | SIMemOp::STORE,
2193 MOI.getIsCrossAddressSpaceOrdering(),
2194 Position::BEFORE, Order,
false);
2196 if (Order == AtomicOrdering::Acquire ||
2197 Order == AtomicOrdering::SequentiallyConsistent) {
2200 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2201 SIMemOp::LOAD, MOI.getIsCrossAddressSpaceOrdering(),
2202 Position::AFTER, Order,
true);
2203 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2204 MOI.getOrderingAddrSpace(),
2214 Changed |= CC->enableVolatileAndOrNonTemporal(
2215 MI, MOI.getInstrAddrSpace(), SIMemOp::LOAD, MOI.isVolatile(),
2216 MOI.isNonTemporal(), MOI.isLastUse());
2221bool SIMemoryLegalizer::expandStore(
const SIMemOpInfo &MOI,
2227 MachineInstr &StoreMI = *
MI;
2229 if (MOI.isAtomic()) {
2230 if (MOI.getOrdering() == AtomicOrdering::Monotonic ||
2231 MOI.getOrdering() == AtomicOrdering::Release ||
2232 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) {
2233 Changed |= CC->enableStoreCacheBypass(
MI, MOI.getScope(),
2234 MOI.getOrderingAddrSpace());
2239 if (MOI.isCooperative())
2240 Changed |= CC->handleCooperativeAtomic(*
MI);
2242 if (MOI.getOrdering() == AtomicOrdering::Release ||
2243 MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
2244 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2245 MOI.getOrderingAddrSpace(),
2246 MOI.getIsCrossAddressSpaceOrdering(),
2249 Changed |= CC->finalizeStore(StoreMI,
true);
2256 Changed |= CC->enableVolatileAndOrNonTemporal(
2257 MI, MOI.getInstrAddrSpace(), SIMemOp::STORE, MOI.isVolatile(),
2258 MOI.isNonTemporal());
2262 Changed |= CC->finalizeStore(StoreMI,
false);
2266bool SIMemoryLegalizer::expandAtomicFence(
const SIMemOpInfo &MOI,
2268 assert(
MI->getOpcode() == AMDGPU::ATOMIC_FENCE);
2270 AtomicPseudoMIs.push_back(
MI);
2273 const SIAtomicAddrSpace OrderingAddrSpace = MOI.getOrderingAddrSpace();
2275 if (MOI.isAtomic()) {
2277 if (Order == AtomicOrdering::Acquire) {
2279 Changed |= CC->insertWait(
MI, MOI.getScope(), OrderingAddrSpace,
2280 SIMemOp::LOAD | SIMemOp::STORE,
2281 MOI.getIsCrossAddressSpaceOrdering(),
2282 Position::BEFORE, Order,
true);
2285 if (Order == AtomicOrdering::Release ||
2286 Order == AtomicOrdering::AcquireRelease ||
2287 Order == AtomicOrdering::SequentiallyConsistent)
2295 Changed |= CC->insertRelease(
MI, MOI.getScope(), OrderingAddrSpace,
2296 MOI.getIsCrossAddressSpaceOrdering(),
2304 if (Order == AtomicOrdering::Acquire ||
2305 Order == AtomicOrdering::AcquireRelease ||
2306 Order == AtomicOrdering::SequentiallyConsistent)
2307 Changed |= CC->insertAcquire(
MI, MOI.getScope(), OrderingAddrSpace,
2316bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(
const SIMemOpInfo &MOI,
2321 MachineInstr &RMWMI = *
MI;
2323 if (MOI.isAtomic()) {
2325 if (Order == AtomicOrdering::Monotonic ||
2326 Order == AtomicOrdering::Acquire || Order == AtomicOrdering::Release ||
2327 Order == AtomicOrdering::AcquireRelease ||
2328 Order == AtomicOrdering::SequentiallyConsistent) {
2329 Changed |= CC->enableRMWCacheBypass(
MI, MOI.getScope(),
2330 MOI.getInstrAddrSpace());
2333 if (Order == AtomicOrdering::Release ||
2334 Order == AtomicOrdering::AcquireRelease ||
2335 Order == AtomicOrdering::SequentiallyConsistent ||
2336 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent)
2337 Changed |= CC->insertRelease(
MI, MOI.getScope(),
2338 MOI.getOrderingAddrSpace(),
2339 MOI.getIsCrossAddressSpaceOrdering(),
2342 if (Order == AtomicOrdering::Acquire ||
2343 Order == AtomicOrdering::AcquireRelease ||
2344 Order == AtomicOrdering::SequentiallyConsistent ||
2345 MOI.getFailureOrdering() == AtomicOrdering::Acquire ||
2346 MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) {
2349 CC->insertWait(
MI, MOI.getScope(), MOI.getInstrAddrSpace(),
2350 isAtomicRet(*
MI) ? SIMemOp::LOAD : SIMemOp::STORE,
2351 MOI.getIsCrossAddressSpaceOrdering(), Position::AFTER,
2353 Changed |= CC->insertAcquire(
MI, MOI.getScope(),
2354 MOI.getOrderingAddrSpace(),
2358 Changed |= CC->finalizeStore(RMWMI,
true);
2365bool SIMemoryLegalizer::expandLDSDMA(
const SIMemOpInfo &MOI,
2377 return CC->enableVolatileAndOrNonTemporal(
2378 MI, MOI.getInstrAddrSpace(), OpKind, MOI.isVolatile(),
2379 MOI.isNonTemporal(), MOI.isLastUse());
2382bool SIMemoryLegalizerLegacy::runOnMachineFunction(MachineFunction &MF) {
2383 const MachineModuleInfo &MMI =
2384 getAnalysis<MachineModuleInfoWrapperPass>().getMMI();
2385 return SIMemoryLegalizer(MMI).run(MF);
2392 .getCachedResult<MachineModuleAnalysis>(
2394 assert(MMI &&
"MachineModuleAnalysis must be available");
2395 if (!SIMemoryLegalizer(MMI->getMMI()).run(MF))
2405 CC = SICacheControl::create(ST);
2407 for (
auto &
MBB : MF) {
2411 if (
MI->isBundle() &&
MI->mayLoadOrStore()) {
2414 I != E &&
I->isBundledWithPred(); ++
I) {
2415 I->unbundleFromPred();
2418 MO.setIsInternalRead(
false);
2421 MI->eraseFromParent();
2422 MI =
II->getIterator();
2428 if (
const auto &MOI = MOA.getLoadInfo(
MI)) {
2430 }
else if (
const auto &MOI = MOA.getStoreInfo(
MI)) {
2432 }
else if (
const auto &MOI = MOA.getLDSDMAInfo(
MI)) {
2434 }
else if (
const auto &MOI = MOA.getAtomicFenceInfo(
MI)) {
2436 }
else if (
const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(
MI)) {
2437 Changed |= expandAtomicCmpxchgOrRmw(*MOI,
MI);
2442 Changed |= removeAtomicPseudoMIs();
2448char SIMemoryLegalizerLegacy::
ID = 0;
2452 return new SIMemoryLegalizerLegacy();
static std::optional< LoadInfo > getLoadInfo(const MachineInstr &MI)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU address space definition.
Provides AMDGPU specific target descriptions.
AMDGPU Machine Module Info.
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Atomic ordering constants.
AMD GCN specific subclass of TargetSubtarget.
const HexagonInstrInfo * TII
This header defines various interfaces for pass management in LLVM.
This file provides utility for Memory Model Relaxation Annotations (MMRAs).
uint64_t IntrinsicInst * II
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
static cl::opt< bool > AmdgcnSkipCacheInvalidations("amdgcn-skip-cache-invalidations", cl::init(false), cl::Hidden, cl::desc("Use this to skip inserting cache invalidating instructions."))
static bool canUseBUFFER_WBINVL1_VOL(const GCNSubtarget &ST)
static const uint32_t IV[8]
SyncScope::ID getWorkgroupSSID() const
SyncScope::ID getWavefrontSSID() const
SyncScope::ID getAgentSSID() const
SyncScope::ID getClusterOneAddressSpaceSSID() const
SyncScope::ID getClusterSSID() const
std::optional< bool > isSyncScopeInclusion(SyncScope::ID A, SyncScope::ID B) const
In AMDGPU target synchronization scopes are inclusive, meaning a larger synchronization scope is incl...
SyncScope::ID getAgentOneAddressSpaceSSID() const
SyncScope::ID getSingleThreadOneAddressSpaceSSID() const
SyncScope::ID getWavefrontOneAddressSpaceSSID() const
SyncScope::ID getSystemOneAddressSpaceSSID() const
SyncScope::ID getWorkgroupOneAddressSpaceSSID() const
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
LLVM_ABI void setPreservesCFG()
This function should be called by the pass, iff they do not:
Represents analyses that only rely on functions' control flow.
Diagnostic information for unsupported feature in backend.
FunctionPass class - This class is used to implement most global optimizations.
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
A helper class to return the specified delimiter string after the first invocation of operator String...
Instructions::iterator instr_iterator
MachineInstrBundleIterator< MachineInstr > iterator
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
Representation of each machine instruction.
Ty & getObjFileInfo()
Keep track of various per-module pieces of information for backends that would like to do so.
MachineOperand class - Representation of each machine instruction operand.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
PreservedAnalyses & preserveSet()
Mark an analysis set as preserved.
static bool mayWriteLDSThroughDMA(const MachineInstr &MI)
static bool isAtomicRet(const MachineInstr &MI)
static bool isLDSDMA(const MachineInstr &MI)
PreservedAnalyses run(MachineFunction &MF, MachineFunctionAnalysisManager &MFAM)
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
A wrapper around a string literal that serves as a proxy for constructing global tables of StringRefs...
StringMap - This is an unconventional map that is specialized for handling keys that are "strings",...
StringRef - Represent a constant reference to a string, i.e.
A raw_ostream that writes to an SmallVector or SmallString.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ BUFFER_STRIDED_POINTER
Address space for 192-bit fat buffer pointers with an additional index.
@ REGION_ADDRESS
Address space for region memory. (GDS)
@ LOCAL_ADDRESS
Address space for local memory.
@ FLAT_ADDRESS
Address space for flat memory.
@ GLOBAL_ADDRESS
Address space for global memory (RAT0, VTX0).
@ BUFFER_FAT_POINTER
Address space for 160-bit buffer fat pointers.
@ PRIVATE_ADDRESS
Address space for private memory.
@ BUFFER_RESOURCE
Address space for 128-bit buffer resources.
constexpr char IsVolatile[]
Key for Kernel::Arg::Metadata::mIsVolatile.
bool isGFX10(const MCSubtargetInfo &STI)
bool isGFX11(const MCSubtargetInfo &STI)
LLVM_ABI IsaVersion getIsaVersion(StringRef GPU)
unsigned encodeWaitcnt(const IsaVersion &Version, unsigned Vmcnt, unsigned Expcnt, unsigned Lgkmcnt)
Encodes Vmcnt, Expcnt and Lgkmcnt into Waitcnt for given isa Version.
unsigned getVmcntBitMask(const IsaVersion &Version)
unsigned getLgkmcntBitMask(const IsaVersion &Version)
unsigned getExpcntBitMask(const IsaVersion &Version)
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ Undef
Value of the register doesn't matter.
@ SingleThread
Synchronized with respect to signal handlers executing in the same thread.
@ System
Synchronized with respect to all concurrently executing threads.
initializer< Ty > init(const Ty &Val)
PointerTypeMap run(const Module &M)
Compute the PointerTypeMap for the module M.
Scope
Defines the scope in which this symbol should be visible: Default – Visible in the public interface o...
NodeAddr< FuncNode * > Func
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
OuterAnalysisManagerProxy< ModuleAnalysisManager, MachineFunction > ModuleAnalysisManagerMachineFunctionProxy
Provide the ModuleAnalysisManager to Function proxy.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
char & SIMemoryLegalizerID
LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE()
AnalysisManager< MachineFunction > MachineFunctionAnalysisManager
bool isReleaseOrStronger(AtomicOrdering AO)
LLVM_ABI PreservedAnalyses getMachineFunctionPassPreservedAnalyses()
Returns the minimum set of Analyses that all machine function passes must preserve.
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
AtomicOrdering getMergedAtomicOrdering(AtomicOrdering AO, AtomicOrdering Other)
Return a single atomic ordering that is at least as strong as both the AO and Other orderings for an ...
static const MachineMemOperand::Flags MOCooperative
Mark the MMO of cooperative load/store atomics.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ LLVM_MARK_AS_BITMASK_ENUM
DWARFExpression::Operation Op
static const MachineMemOperand::Flags MOLastUse
Mark the MMO of a load as the last use.
FunctionPass * createSIMemoryLegalizerPass()