65#define DEBUG_TYPE "memcpyopt"
68 "enable-memcpyopt-without-libcalls",
cl::Hidden,
69 cl::desc(
"Enable memcpyopt even when libcalls are disabled"));
71STATISTIC(NumMemCpyInstr,
"Number of memcpy instructions deleted");
72STATISTIC(NumMemMoveInstr,
"Number of memmove instructions deleted");
73STATISTIC(NumMemSetInfer,
"Number of memsets inferred");
74STATISTIC(NumMoveToCpy,
"Number of memmoves converted to memcpy");
75STATISTIC(NumCpyToSet,
"Number of memcpys converted to memset");
76STATISTIC(NumCallSlot,
"Number of call slot optimizations performed");
77STATISTIC(NumStackMove,
"Number of stack-move optimizations performed");
106 bool isProfitableToUseMemset(
const DataLayout &
DL)
const;
114bool MemsetRange::isProfitableToUseMemset(
const DataLayout &
DL)
const {
116 if (TheStores.
size() >= 4 || End - Start >= 16)
120 if (TheStores.
size() < 2)
125 for (Instruction *SI : TheStores)
131 if (TheStores.size() == 2)
144 unsigned Bytes = unsigned(End - Start);
145 unsigned MaxIntSize =
DL.getLargestLegalIntTypeSizeInBits() / 8;
148 unsigned NumPointerStores = Bytes / MaxIntSize;
151 unsigned NumByteStores = Bytes % MaxIntSize;
156 return TheStores.size() > NumPointerStores + NumByteStores;
167 const DataLayout &
DL;
170 MemsetRanges(
const DataLayout &
DL) :
DL(
DL) {}
174 const_iterator
begin()
const {
return Ranges.begin(); }
175 const_iterator
end()
const {
return Ranges.end(); }
178 void addInst(int64_t OffsetFromFirst, Instruction *Inst) {
180 addStore(OffsetFromFirst, SI);
185 void addStore(int64_t OffsetFromFirst, StoreInst *SI) {
186 TypeSize StoreSize =
DL.getTypeStoreSize(
SI->getOperand(0)->getType());
189 SI->getPointerOperand(),
SI->getAlign(), SI);
192 void addMemSet(int64_t OffsetFromFirst, MemSetInst *MSI) {
206void MemsetRanges::addRange(int64_t Start, int64_t
Size,
Value *Ptr,
207 MaybeAlign Alignment, Instruction *Inst) {
211 Ranges, [=](
const MemsetRange &O) {
return O.End <
Start; });
216 if (
I ==
Ranges.end() || End < I->Start) {
217 MemsetRange &
R = *
Ranges.insert(
I, MemsetRange());
221 R.Alignment = Alignment;
222 R.TheStores.push_back(Inst);
227 I->TheStores.push_back(Inst);
231 if (
I->Start <= Start &&
I->End >= End)
240 if (Start < I->Start) {
243 I->Alignment = Alignment;
251 range_iterator NextI =
I;
252 while (++NextI !=
Ranges.end() && End >= NextI->Start) {
254 I->TheStores.append(NextI->TheStores.begin(), NextI->TheStores.end());
255 if (NextI->End >
I->End)
271 assert(Start->getParent() == End->
getParent() &&
"Must be in same block");
273 if (Start->getFunction()->doesNotThrow())
278 bool RequiresNoCaptureBeforeUnwind;
280 RequiresNoCaptureBeforeUnwind) &&
281 !RequiresNoCaptureBeforeUnwind)
289void MemCpyOptPass::eraseInstruction(Instruction *
I) {
290 MSSAU->removeMemoryAccess(
I);
291 EEA->removeInstruction(
I);
292 I->eraseFromParent();
303 assert(Start->getBlock() == End->
getBlock() &&
"Only local supported");
309 if (
II &&
II->getIntrinsicID() == Intrinsic::lifetime_start &&
310 SkippedLifetimeStart && !*SkippedLifetimeStart) {
311 *SkippedLifetimeStart =
I;
330 return Start->getBlock() != End->
getBlock() ||
334 if (isa<MemoryUse>(&Acc))
336 Instruction *AccInst =
337 cast<MemoryUseOrDef>(&Acc)->getMemoryInst();
338 return isModSet(AA.getModRefInfo(AccInst, Loc));
352Instruction *MemCpyOptPass::tryMergingIntoMemset(Instruction *StartInst,
359 if (
DL.getTypeStoreSize(
SI->getOperand(0)->getType()).isScalable())
373 MemoryUseOrDef *MemInsertPoint =
nullptr;
374 for (++BI; !BI->isTerminator(); ++BI) {
378 MemInsertPoint = CurrentAcc;
383 if (CB->onlyAccessesInaccessibleMemory())
391 if (BI->mayWriteToMemory() || BI->mayReadFromMemory())
398 if (!NextStore->isSimple())
401 Value *StoredVal = NextStore->getValueOperand();
409 if (
DL.getTypeStoreSize(StoredVal->
getType()).isScalable())
415 ByteVal = StoredByte;
416 if (ByteVal != StoredByte)
420 std::optional<int64_t>
Offset =
429 if (MSI->isVolatile() || ByteVal != MSI->getValue() ||
434 std::optional<int64_t>
Offset =
435 MSI->getDest()->getPointerOffsetFrom(StartPtr,
DL);
451 Ranges.addInst(0, StartInst);
461 for (
const MemsetRange &
Range : Ranges) {
462 if (
Range.TheStores.size() == 1)
466 if (!
Range.isProfitableToUseMemset(
DL))
471 StartPtr =
Range.StartPtr;
473 AMemSet = Builder.CreateMemSet(StartPtr, ByteVal,
Range.End -
Range.Start,
480 dbgs() <<
"With: " << *AMemSet <<
'\n');
481 if (!
Range.TheStores.empty())
486 ? MSSAU->createMemoryAccessBefore(AMemSet,
nullptr, MemInsertPoint)
487 : MSSAU->createMemoryAccessAfter(AMemSet,
nullptr, MemInsertPoint));
488 MSSAU->insertDef(NewDef,
true);
489 MemInsertPoint = NewDef;
492 for (Instruction *SI :
Range.TheStores)
505bool MemCpyOptPass::moveUp(StoreInst *SI, Instruction *
P,
const LoadInst *LI) {
513 DenseSet<Instruction *>
Args;
514 auto AddArg = [&](
Value *Arg) {
516 if (
I &&
I->getParent() ==
SI->getParent()) {
524 if (!AddArg(
SI->getPointerOperand()))
528 SmallVector<Instruction *, 8> ToLift{
SI};
538 for (
auto I = --
SI->getIterator(),
E =
P->getIterator();
I !=
E; --
I) {
546 bool MayAlias =
isModOrRefSet(AA->getModRefInfo(
C, std::nullopt));
548 bool NeedLift =
false;
568 if (
isModSet(AA->getModRefInfo(
C, LoadLoc)))
601 MemoryUseOrDef *MemInsertPoint =
nullptr;
602 if (MemoryUseOrDef *MA = MSSA->getMemoryAccess(
P)) {
608 if (MemoryUseOrDef *MA = MSSA->getMemoryAccess(&
I)) {
618 I->moveBefore(
P->getIterator());
619 assert(MemInsertPoint &&
"Must have found insert point");
620 if (MemoryUseOrDef *MA = MSSA->getMemoryAccess(
I)) {
621 MSSAU->moveAfter(MA, MemInsertPoint);
629bool MemCpyOptPass::processStoreOfLoad(StoreInst *SI, LoadInst *LI,
630 const DataLayout &
DL,
635 BatchAAResults BAA(*AA, EEA);
641 if (
T->isAggregateType() &&
643 (TLI->has(LibFunc_memcpy) && TLI->has(LibFunc_memmove)))) {
653 if (
isModSet(BAA.getModRefInfo(&
I, LoadLoc))) {
663 if (
P == SI || moveUp(SI,
P, LI)) {
668 bool UseMemMove =
false;
669 if (
isModSet(AA->getModRefInfo(SI, LoadLoc)))
674 Builder.CreateTypeSize(Builder.getInt64Ty(),
DL.getTypeStoreSize(
T));
677 M = Builder.CreateMemMove(
SI->getPointerOperand(),
SI->getAlign(),
681 M = Builder.CreateMemCpy(
SI->getPointerOperand(),
SI->getAlign(),
683 M->copyMetadata(*SI, LLVMContext::MD_DIAssignID);
685 LLVM_DEBUG(
dbgs() <<
"Promoting " << *LI <<
" to " << *SI <<
" => " << *M
689 auto *NewAccess = MSSAU->createMemoryAccessAfter(M,
nullptr, LastDef);
697 BBI =
M->getIterator();
705 auto GetCall = [&]() -> CallInst * {
709 MSSA->getWalker()->getClobberingMemoryAccess(LI, BAA)))
714 bool Changed = performCallSlotOptzn(
715 LI, SI,
SI->getPointerOperand()->stripPointerCasts(),
717 DL.getTypeStoreSize(
SI->getOperand(0)->getType()),
718 std::min(
SI->getAlign(), LI->
getAlign()), BAA, GetCall);
729 if (performStackMoveOptzn(LI, SI,
SI->getPointerOperand(),
733 BBI =
SI->getNextNode()->getIterator();
753 if (
SI->getMetadata(LLVMContext::MD_nontemporal))
756 const DataLayout &
DL =
SI->getDataLayout();
758 Value *StoredVal =
SI->getValueOperand();
767 return processStoreOfLoad(SI, LI,
DL, BBI);
788 tryMergingIntoMemset(SI,
SI->getPointerOperand(), ByteVal)) {
789 BBI =
I->getIterator();
796 auto *
T =
V->getType();
797 if (!
T->isAggregateType())
800 TypeSize
Size =
DL.getTypeStoreSize(
T);
801 if (
Size.isScalable())
805 auto *
M = Builder.CreateMemSet(
SI->getPointerOperand(), ByteVal,
Size,
807 M->copyMetadata(*SI, LLVMContext::MD_DIAssignID);
809 LLVM_DEBUG(
dbgs() <<
"Promoting " << *SI <<
" to " << *M <<
"\n");
814 auto *NewAccess = MSSAU->createMemoryAccessBefore(M,
nullptr, StoreDef);
821 BBI =
M->getIterator();
831 BBI =
I->getIterator();
840bool MemCpyOptPass::performCallSlotOptzn(Instruction *cpyLoad,
841 Instruction *cpyStore,
Value *cpyDest,
842 Value *cpySrc, TypeSize cpySize,
845 std::function<CallInst *()> GetC) {
871 std::optional<TypeSize> SrcAllocaSize = srcAlloca->getAllocationSize(
DL);
872 if (!SrcAllocaSize || SrcAllocaSize->isScalable())
874 uint64_t srcSize = SrcAllocaSize->getFixedValue();
876 if (cpySize < srcSize)
879 CallInst *
C = GetC();
884 if (Function *
F =
C->getCalledFunction())
885 if (
F->isIntrinsic() &&
F->getIntrinsicID() == Intrinsic::lifetime_start)
888 if (
C->getParent() != cpyStore->
getParent()) {
893 MemoryLocation DestLoc =
896 : MemoryLocation::getForDest(
cast<MemCpyInst>(cpyStore));
902 MSSA->getMemoryAccess(cpyStore), &SkippedLifetimeStart)) {
903 LLVM_DEBUG(dbgs() <<
"Call Slot: Dest pointer modified after call\n");
910 if (SkippedLifetimeStart) {
913 if (LifetimeArg && LifetimeArg->getParent() ==
C->getParent() &&
914 C->comesBefore(LifetimeArg))
920 bool ExplicitlyDereferenceableOnly;
922 ExplicitlyDereferenceableOnly) ||
925 LLVM_DEBUG(
dbgs() <<
"Call Slot: Dest pointer not dereferenceable\n");
944 LLVM_DEBUG(
dbgs() <<
"Call Slot: Dest may be visible through unwinding\n");
949 Align srcAlign = srcAlloca->getAlign();
950 bool isDestSufficientlyAligned = srcAlign <= cpyDestAlign;
954 LLVM_DEBUG(
dbgs() <<
"Call Slot: Dest not sufficiently aligned\n");
963 while (!srcUseList.empty()) {
964 User *
U = srcUseList.pop_back_val();
973 if (U !=
C && U != cpyLoad) {
974 LLVM_DEBUG(
dbgs() <<
"Call slot: Source accessed by " << *U <<
"\n");
981 bool SrcIsCaptured =
any_of(
C->args(), [&](Use &U) {
982 return U->stripPointerCasts() == cpySrc &&
983 !C->doesNotCapture(C->getArgOperandNo(&U));
1000 MemoryLocation SrcLoc =
1002 for (Instruction &
I :
1003 make_range(++
C->getIterator(),
C->getParent()->end())) {
1006 if (
II->getIntrinsicID() == Intrinsic::lifetime_end &&
1007 II->getArgOperand(0) == srcAlloca)
1030 bool NeedMoveGEP =
false;
1031 if (!DT->dominates(cpyDest,
C)) {
1034 if (
GEP &&
GEP->hasAllConstantIndices() &&
1035 DT->dominates(
GEP->getPointerOperand(),
C))
1057 for (
unsigned ArgI = 0; ArgI <
C->arg_size(); ++ArgI)
1058 if (
C->getArgOperand(ArgI)->stripPointerCasts() == cpySrc &&
1059 cpySrc->
getType() !=
C->getArgOperand(ArgI)->getType())
1063 bool changedArgument =
false;
1064 for (
unsigned ArgI = 0; ArgI <
C->arg_size(); ++ArgI)
1065 if (
C->getArgOperand(ArgI)->stripPointerCasts() == cpySrc) {
1066 changedArgument =
true;
1067 C->setArgOperand(ArgI, cpyDest);
1070 if (!changedArgument)
1074 if (!isDestSufficientlyAligned) {
1081 GEP->moveBefore(
C->getIterator());
1084 if (SkippedLifetimeStart) {
1085 SkippedLifetimeStart->
moveBefore(
C->getIterator());
1086 MSSAU->moveBefore(MSSA->getMemoryAccess(SkippedLifetimeStart),
1087 MSSA->getMemoryAccess(
C));
1091 if (cpyLoad != cpyStore)
1100bool MemCpyOptPass::processMemCpyMemCpyDependence(MemCpyInst *M,
1102 BatchAAResults &BAA) {
1116 int64_t MForwardOffset = 0;
1117 const DataLayout &
DL =
M->getModule()->getDataLayout();
1120 if (
M->getSource() != MDep->
getDest()) {
1121 std::optional<int64_t>
Offset =
1122 M->getSource()->getPointerOffsetFrom(MDep->
getDest(),
DL);
1125 MForwardOffset = *
Offset;
1128 Value *CopyLength =
M->getLength();
1133 if (MForwardOffset != 0 || MDep->
getLength() != CopyLength) {
1139 if (!MDepLen || !MLen)
1141 if (MDepLen->getZExtValue() < MLen->getZExtValue() + MForwardOffset) {
1144 if (MDepLen->getZExtValue() <= (uint64_t)MForwardOffset)
1148 CopyLength = ConstantInt::get(CopyLength->
getType(),
1149 MDepLen->getZExtValue() - MForwardOffset);
1157 if (NewCopySource && NewCopySource->
use_empty())
1170 MCopyLoc = MCopyLoc.getWithNewSize(
1179 if (MForwardOffset > 0) {
1181 std::optional<int64_t> MDestOffset =
1183 if (MDestOffset == MForwardOffset)
1184 CopySource =
M->getDest();
1186 CopySource = Builder.CreateInBoundsPtrAdd(
1187 CopySource, Builder.getInt64(MForwardOffset));
1191 MCopyLoc = MCopyLoc.getWithNewPtr(CopySource);
1192 if (CopySourceAlign)
1205 if (
writtenBetween(MSSA, BAA, MCopyLoc, MSSA->getMemoryAccess(MDep),
1206 MSSA->getMemoryAccess(M)))
1222 bool UseMemMove =
false;
1227 if (
M->isForceInlined())
1233 LLVM_DEBUG(
dbgs() <<
"MemCpyOptPass: Forwarding memcpy->memcpy src:\n"
1241 NewM = Builder.CreateMemMove(
M->getDest(),
M->getDestAlign(), CopySource,
1242 CopySourceAlign, CopyLength,
M->isVolatile());
1243 else if (
M->isForceInlined())
1247 NewM = Builder.CreateMemCpyInline(
M->getDest(),
M->getDestAlign(),
1248 CopySource, CopySourceAlign, CopyLength,
1251 NewM = Builder.CreateMemCpy(
M->getDest(),
M->getDestAlign(), CopySource,
1252 CopySourceAlign, CopyLength,
M->isVolatile());
1258 auto *NewAccess = MSSAU->createMemoryAccessAfter(NewM,
nullptr, LastDef);
1286bool MemCpyOptPass::processMemSetMemCpyDependence(MemCpyInst *MemCpy,
1288 BatchAAResults &BAA) {
1311 MSSA->getMemoryAccess(MemSet),
1312 MSSA->getMemoryAccess(MemCpy)))
1324 if (DestSize == SrcSize) {
1347 "Preserving debug location based on moving memset within BB.");
1348 Builder.SetCurrentDebugLocation(MemSet->
getDebugLoc());
1354 SrcSize = Builder.CreateZExt(SrcSize, DestSize->
getType());
1356 DestSize = Builder.CreateZExt(DestSize, SrcSize->
getType());
1359 Value *Ule = Builder.CreateICmpULE(DestSize, SrcSize);
1360 Value *SizeDiff = Builder.CreateSub(DestSize, SrcSize);
1361 Value *MemsetLen = Builder.CreateSelect(
1362 Ule, ConstantInt::getNullValue(DestSize->
getType()), SizeDiff);
1368 Builder.CreateMemSet(Builder.CreatePtrAdd(Dest, SrcSize),
1369 MemSet->
getOperand(1), MemsetLen, Alignment);
1372 "MemCpy must be a MemoryDef");
1377 MSSAU->createMemoryAccessBefore(NewMemSet,
nullptr, LastDef);
1392 if (
II->getIntrinsicID() == Intrinsic::lifetime_start)
1394 return II->getArgOperand(0) == Alloca;
1428bool MemCpyOptPass::performMemCpyToMemSetOptzn(MemCpyInst *MemCpy,
1430 BatchAAResults &BAA) {
1434 int64_t MOffset = 0;
1439 std::optional<int64_t>
Offset =
1446 if (MOffset != 0 || MemSetSize != CopySize) {
1451 if (!CMemSetSize || !CCopySize ||
1452 CCopySize->getZExtValue() + MOffset > CMemSetSize->getZExtValue()) {
1456 if (CMemSetSize && CCopySize) {
1459 assert(CCopySize->getZExtValue() + MOffset >
1460 CMemSetSize->getZExtValue());
1462 CopySize = MemSetSize;
1465 ConstantInt::get(CopySize->
getType(),
1466 CMemSetSize->getZExtValue() <= (uint64_t)MOffset
1468 : CMemSetSize->getZExtValue() - MOffset);
1478 auto *NewAccess = MSSAU->createMemoryAccessAfter(NewM,
nullptr, LastDef);
1496bool MemCpyOptPass::performStackMoveOptzn(Instruction *Load, Instruction *Store,
1498 TypeSize
Size, BatchAAResults &BAA) {
1513 if (SrcAlloca == DestAlloca)
1533 if (!SrcOffset || *SrcOffset < *DestOffset || *SrcOffset < 0)
1536 if ((*SrcOffset - *DestOffset) % DestAlloca->
getAlign().
value() != 0)
1540 if (!SrcSize || !DestSize)
1542 if (*SrcSize != *DestSize)
1543 if (!SrcSize->isFixed() || !DestSize->isFixed())
1546 if (
Size != *DestSize || *DestOffset != 0) {
1547 LLVM_DEBUG(
dbgs() <<
"Stack Move: Destination alloca size mismatch\n");
1552 bool MoveSrc = !DT->dominates(SrcAlloca, DestAlloca);
1554 if (!DT->dominates(DestAlloca, SrcAlloca))
1563 SmallVector<Instruction *, 4> LifetimeMarkers;
1564 SmallPtrSet<Instruction *, 4> AAMetadataInstrs;
1566 auto CaptureTrackingWithModRef =
1567 [&](
Instruction *AI, function_ref<bool(Instruction *)> ModRefCallback,
1568 bool &AddressCaptured) ->
bool {
1569 SmallVector<Instruction *, 8> Worklist;
1572 Worklist.
reserve(MaxUsesToExplore);
1573 SmallPtrSet<const Use *, 20> Visited;
1574 while (!Worklist.
empty()) {
1576 for (
const Use &U :
I->uses()) {
1579 if (Visited.
size() >= MaxUsesToExplore) {
1582 <<
"Stack Move: Exceeded max uses to see ModRef, bailing\n");
1585 if (!Visited.
insert(&U).second)
1592 if (UI->mayReadOrWriteMemory()) {
1593 if (UI->isLifetimeStartOrEnd()) {
1602 AAMetadataInstrs.
insert(UI);
1604 if (!ModRefCallback(UI))
1619 ModRefInfo DestModRef = ModRefInfo::NoModRef;
1621 SmallVector<BasicBlock *, 8> ReachabilityWorklist;
1622 auto DestModRefCallback = [&](
Instruction *UI) ->
bool {
1632 if (UI->getParent() ==
Store->getParent()) {
1641 if (UI->comesBefore(Store))
1651 ReachabilityWorklist.
push_back(UI->getParent());
1657 bool DestAddressCaptured =
false;
1658 if (!CaptureTrackingWithModRef(DestAlloca, DestModRefCallback,
1659 DestAddressCaptured))
1662 if (!ReachabilityWorklist.
empty() &&
1664 nullptr, DT,
nullptr))
1672 auto SrcModRefCallback = [&](
Instruction *UI) ->
bool {
1675 if (PDT->dominates(Load, UI) || UI == Load || UI == Store)
1685 bool SrcAddressCaptured =
false;
1686 if (!CaptureTrackingWithModRef(SrcAlloca, SrcModRefCallback,
1687 SrcAddressCaptured))
1692 if (DestAddressCaptured && SrcAddressCaptured)
1704 if (*SrcSize != *DestSize) {
1707 if (DestSize->getFixedValue() > SrcSize->getFixedValue()) {
1714 Value *NewDestPtr = SrcAlloca;
1715 if (*SrcOffset != *DestOffset) {
1717 NewDestPtr = Builder.CreateInBoundsPtrAdd(
1718 SrcAlloca, Builder.getInt64(*SrcOffset - *DestOffset));
1729 if (!LifetimeMarkers.
empty()) {
1730 for (Instruction *
I : LifetimeMarkers)
1739 for (Instruction *
I : AAMetadataInstrs) {
1740 I->setMetadata(LLVMContext::MD_alias_scope,
nullptr);
1741 I->setMetadata(LLVMContext::MD_noalias,
nullptr);
1742 I->setMetadata(LLVMContext::MD_tbaa,
nullptr);
1743 I->setMetadata(LLVMContext::MD_tbaa_struct,
nullptr);
1746 LLVM_DEBUG(
dbgs() <<
"Stack Move: Performed stack-move optimization\n");
1768 if (
M->isVolatile())
1772 if (
M->getSource() ==
M->getDest()) {
1785 MemoryUseOrDef *MA = MSSA->getMemoryAccess(M);
1792 if (GV->isConstant() && GV->hasDefinitiveInitializer())
1794 M->getDataLayout())) {
1797 M->getRawDest(), ByteVal,
M->getLength(),
M->getDestAlign(),
false);
1800 MSSAU->createMemoryAccessAfter(NewM,
nullptr, LastDef);
1808 BatchAAResults BAA(*AA, EEA);
1812 const MemoryAccess *DestClobber =
1813 MSSA->getWalker()->getClobberingMemoryAccess(AnyClobber, DestLoc, BAA);
1821 if (DestClobber->
getBlock() ==
M->getParent())
1822 if (processMemSetMemCpyDependence(M, MDep, BAA))
1825 MemoryAccess *SrcClobber = MSSA->getWalker()->getClobberingMemoryAccess(
1837 if (Instruction *
MI = MD->getMemoryInst()) {
1840 if (performCallSlotOptzn(M, M,
M->getDest(),
M->getSource(),
1842 M->getDestAlign().valueOrOne(), BAA,
1843 [
C]() -> CallInst * { return C; })) {
1845 <<
" call: " << *
C <<
"\n"
1846 <<
" memcpy: " << *M <<
"\n");
1854 if (processMemCpyMemCpyDependence(M, MDep, BAA))
1857 if (performMemCpyToMemSetOptzn(M, MDep, BAA)) {
1880 if (performStackMoveOptzn(M, M,
M->getDest(),
M->getSource(),
1883 BBI =
M->getNextNode()->getIterator();
1894bool MemCpyOptPass::isMemMoveMemSetDependency(MemMoveInst *M) {
1895 const auto &
DL =
M->getDataLayout();
1896 MemoryUseOrDef *MemMoveAccess = MSSA->getMemoryAccess(M);
1902 auto *MemMoveSourceOp =
M->getSource();
1908 LocationSize MemMoveLocSize = SourceLoc.
Size;
1909 if (
Source->getPointerOperand() !=
M->getDest() ||
1915 uint64_t MemMoveSize = MemMoveLocSize.
getValue();
1916 LocationSize TotalSize =
1918 MemoryLocation CombinedLoc(
M->getDest(), TotalSize);
1922 BatchAAResults BAA(*AA);
1925 MSSA->getWalker()->getClobberingMemoryAccess(FirstDef, CombinedLoc, BAA));
1935 if (!MemSetLength || MemSetLength->getZExtValue() < MemMoveSize)
1952 if (!
M->isVolatile() && isMemMoveMemSetDependency(M)) {
1962 LLVM_DEBUG(
dbgs() <<
"MemCpyOptPass: Optimizing memmove -> memcpy: " << *M
1966 Type *ArgTys[3] = {
M->getRawDest()->getType(),
M->getRawSource()->getType(),
1967 M->getLength()->getType()};
1969 M->getModule(), Intrinsic::memcpy, ArgTys));
1979bool MemCpyOptPass::processByValArgument(CallBase &CB,
unsigned ArgNo) {
1984 TypeSize ByValSize =
DL.getTypeAllocSize(ByValTy);
1986 MemoryUseOrDef *CallAccess = MSSA->getMemoryAccess(&CB);
1989 MemCpyInst *MDep =
nullptr;
1990 BatchAAResults BAA(*AA, EEA);
1991 MemoryAccess *Clobber = MSSA->getWalker()->getClobberingMemoryAccess(
2005 if (!C1 || !TypeSize::isKnownGE(
2018 if ((!MemDepAlign || *MemDepAlign < *ByValAlign) &&
2034 MSSA->getMemoryAccess(MDep), CallAccess))
2037 LLVM_DEBUG(
dbgs() <<
"MemCpyOptPass: Forwarding memcpy to byval:\n"
2038 <<
" " << *MDep <<
"\n"
2039 <<
" " << CB <<
"\n");
2062bool MemCpyOptPass::processImmutArgument(CallBase &CB,
unsigned ArgNo) {
2063 BatchAAResults BAA(*AA, EEA);
2087 std::optional<TypeSize> AllocaSize = AI->getAllocationSize(
DL);
2090 if (!AllocaSize || AllocaSize->isScalable())
2093 MemoryUseOrDef *CallAccess = MSSA->getMemoryAccess(&CB);
2097 MemCpyInst *MDep =
nullptr;
2098 MemoryAccess *Clobber = MSSA->getWalker()->getClobberingMemoryAccess(
2114 if (!MDepLen || AllocaSize != MDepLen->getValue())
2121 Align AllocaAlign = AI->getAlign();
2122 if (MemDepAlign < AllocaAlign &&
2134 MSSA->getMemoryAccess(MDep), CallAccess))
2141 LLVM_DEBUG(
dbgs() <<
"MemCpyOptPass: Forwarding memcpy to Immut src:\n"
2142 <<
" " << *MDep <<
"\n"
2143 <<
" " << CB <<
"\n");
2153bool MemCpyOptPass::iterateOnFunction(Function &
F) {
2154 bool MadeChange =
false;
2157 for (BasicBlock &BB :
F) {
2162 if (!DT->isReachableFromEntry(&BB))
2169 bool RepeatInstruction =
false;
2172 MadeChange |= processStore(SI, BI);
2174 RepeatInstruction = processMemSet(M, BI);
2176 RepeatInstruction = processMemCpy(M, BI);
2178 RepeatInstruction = processMemMove(M, BI);
2180 for (
unsigned i = 0, e = CB->
arg_size(); i != e; ++i) {
2182 MadeChange |= processByValArgument(*CB, i);
2184 MadeChange |= processImmutArgument(*CB, i);
2189 if (RepeatInstruction) {
2190 if (BI != BB.
begin())
2208 bool MadeChange =
runImpl(
F, &TLI, AA, AC, DT, PDT, &MSSA->getMSSA());
2222 bool MadeChange =
false;
2235 if (!iterateOnFunction(
F))
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file defines the DenseSet and SmallDenseSet classes.
static bool runImpl(Function &F, const TargetLowering &TLI, const LibcallLoweringInfo &Libcalls, AssumptionCache *AC)
This is the interface for a simple mod/ref and alias analysis over globals.
Module.h This file contains the declarations for the Module class.
This header defines various interfaces for pass management in LLVM.
static void eraseInstruction(Instruction &I, ICFLoopSafetyInfo &SafetyInfo, MemorySSAUpdater &MSSAU)
static bool mayBeVisibleThroughUnwinding(Value *V, Instruction *Start, Instruction *End)
static bool isZeroSize(Value *Size)
static bool hasUndefContents(MemorySSA *MSSA, BatchAAResults &AA, Value *V, MemoryDef *Def)
Determine whether the pointer V had only undefined content (due to Def), either because it was freshl...
static bool accessedBetween(BatchAAResults &AA, MemoryLocation Loc, const MemoryUseOrDef *Start, const MemoryUseOrDef *End, Instruction **SkippedLifetimeStart=nullptr)
static bool overreadUndefContents(MemorySSA *MSSA, MemCpyInst *MemCpy, MemIntrinsic *MemSrc, BatchAAResults &BAA)
static cl::opt< bool > EnableMemCpyOptWithoutLibcalls("enable-memcpyopt-without-libcalls", cl::Hidden, cl::desc("Enable memcpyopt even when libcalls are disabled"))
static bool writtenBetween(MemorySSA *MSSA, BatchAAResults &AA, MemoryLocation Loc, const MemoryUseOrDef *Start, const MemoryUseOrDef *End)
This file provides utility analysis objects describing memory locations.
This file exposes an interface to building/using memory SSA to walk memory instructions using a use/d...
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
uint64_t IntrinsicInst * II
This file contains the declarations for profiling metadata utility functions.
This file defines the make_scope_exit function, which executes user-defined cleanup logic at scope ex...
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
A manager for alias analyses.
LLVM_ABI bool isStaticAlloca() const
Return true if this alloca is in the entry block of the function and is a constant size.
Align getAlign() const
Return the alignment of the memory that is being allocated by the instruction.
void setAllocatedType(Type *Ty)
for use only in special circumstances that need to generically transform a whole instruction (eg: IR ...
Type * getAllocatedType() const
Return the type that is being allocated by the instruction.
unsigned getAddressSpace() const
Return the address space for the allocation.
LLVM_ABI std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
void setAlignment(Align Align)
const Value * getArraySize() const
Get the number of elements allocated.
PassT::Result & getResult(IRUnitT &IR, ExtraArgTs... ExtraArgs)
Get the result of an analysis pass for a given IR unit.
A function analysis which provides an AssumptionCache.
A cache of @llvm.assume calls within a function.
iterator begin()
Instruction iterator methods.
LLVM_ABI bool isEntryBlock() const
Return true if this is the entry block of the containing function.
InstListType::iterator iterator
Instruction iterators...
This class is a wrapper over an AAResults, and it is intended to be used only when there are no IR ch...
bool isMustAlias(const MemoryLocation &LocA, const MemoryLocation &LocB)
ModRefInfo getModRefInfo(const Instruction *I, const std::optional< MemoryLocation > &OptLoc)
ModRefInfo callCapturesBefore(const Instruction *I, const MemoryLocation &MemLoc, DominatorTree *DT)
Represents analyses that only rely on functions' control flow.
bool doesNotCapture(unsigned OpNo) const
Determine whether this data operand is not captured.
LLVM_ABI bool paramHasAttr(unsigned ArgNo, Attribute::AttrKind Kind) const
Determine whether the argument or parameter has the given attribute.
bool isByValArgument(unsigned ArgNo) const
Determine whether this argument is passed by value.
MaybeAlign getParamAlign(unsigned ArgNo) const
Extract the alignment for a call or parameter (0=unknown).
bool onlyReadsMemory(unsigned OpNo) const
Type * getParamByValType(unsigned ArgNo) const
Extract the byval type for a call or parameter.
Value * getArgOperand(unsigned i) const
void setArgOperand(unsigned i, Value *v)
unsigned arg_size() const
A parsed version of the target data layout string in and methods for querying it.
Analysis pass which computes a DominatorTree.
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Context-sensitive CaptureAnalysis provider, which computes and caches the earliest common dominator c...
LLVM_ABI void mergeDIAssignID(ArrayRef< const Instruction * > SourceInstructions)
Merge the DIAssignID metadata from this instruction and those attached to instructions in SourceInstr...
const DebugLoc & getDebugLoc() const
Return the debug location for this node as a DebugLoc.
LLVM_ABI const Module * getModule() const
Return the module owning the function this instruction belongs to or nullptr it the function does not...
LLVM_ABI void moveBefore(InstListType::iterator InsertPos)
Unlink this instruction from its current basic block and insert it into the basic block that MovePos ...
LLVM_ABI void dropUnknownNonDebugMetadata(ArrayRef< unsigned > KnownIDs={})
Drop all unknown metadata except for debug locations.
void setDebugLoc(DebugLoc Loc)
Set the debug location information for this instruction.
LLVM_ABI void copyMetadata(const Instruction &SrcInst, ArrayRef< unsigned > WL=ArrayRef< unsigned >())
Copy metadata from SrcInst to this instruction.
LLVM_ABI const DataLayout & getDataLayout() const
Get the data layout of the module this instruction belongs to.
Value * getPointerOperand()
Align getAlign() const
Return the alignment of the access that is being performed.
static LocationSize precise(uint64_t Value)
TypeSize getValue() const
This class wraps the llvm.memcpy intrinsic.
LLVM_ABI PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM)
LLVM_ABI bool runImpl(Function &F, TargetLibraryInfo *TLI, AAResults *AA, AssumptionCache *AC, DominatorTree *DT, PostDominatorTree *PDT, MemorySSA *MSSA)
Value * getLength() const
Value * getRawDest() const
Value * getDest() const
This is just like getRawDest, but it strips off any cast instructions (including addrspacecast) that ...
MaybeAlign getDestAlign() const
This is the common base class for memset/memcpy/memmove.
Value * getRawSource() const
Return the arguments to the instruction.
MaybeAlign getSourceAlign() const
Value * getSource() const
This is just like getRawSource, but it strips off any cast instructions that feed it,...
BasicBlock * getBlock() const
AllAccessType::self_iterator getIterator()
Get the iterators for the all access list and the defs only list We default to the all access list.
Represents a read-write access to memory, whether it is a must-alias, or a may-alias.
Representation for a specific memory location.
static LLVM_ABI MemoryLocation get(const LoadInst *LI)
Return a location with information about the memory reference by the given instruction.
static LLVM_ABI MemoryLocation getForSource(const MemTransferInst *MTI)
Return a location representing the source of a memory transfer.
LocationSize Size
The maximum size of the location, in address-units, or UnknownSize if the size is not known.
static MemoryLocation getBeforeOrAfter(const Value *Ptr, const AAMDNodes &AATags=AAMDNodes())
Return a location that may access any location before or after Ptr, while remaining within the underl...
static LLVM_ABI MemoryLocation getForDest(const MemIntrinsic *MI)
Return a location representing the destination of a memory set or transfer.
An analysis that produces MemorySSA for a function.
MemoryAccess * getClobberingMemoryAccess(const Instruction *I, BatchAAResults &AA)
Given a memory Mod/Ref/ModRef'ing instruction, calling this will give you the nearest dominating Memo...
Encapsulates MemorySSA, including all data associated with memory accesses.
LLVM_ABI bool dominates(const MemoryAccess *A, const MemoryAccess *B) const
Given two memory accesses in potentially different blocks, determine whether MemoryAccess A dominates...
LLVM_ABI void verifyMemorySSA(VerificationLevel=VerificationLevel::Fast) const
Verify that MemorySSA is self consistent (IE definitions dominate all uses, uses appear in the right ...
LLVM_ABI MemorySSAWalker * getWalker()
MemoryUseOrDef * getMemoryAccess(const Instruction *I) const
Given a memory Mod/Ref'ing instruction, get the MemorySSA access associated with it.
bool isLiveOnEntryDef(const MemoryAccess *MA) const
Return true if MA represents the live on entry value.
Class that has the common methods + fields of memory uses/defs.
MemoryAccess * getDefiningAccess() const
Get the access that produces the memory state used by this Use.
Instruction * getMemoryInst() const
Get the instruction that this MemoryUse represents.
const DataLayout & getDataLayout() const
Get the data layout for the module's target platform.
Analysis pass which computes a PostDominatorTree.
PostDominatorTree Class - Concrete subclass of DominatorTree that is used to compute the post-dominat...
A set of analyses that are preserved following a run of a transformation pass.
static PreservedAnalyses all()
Construct a special preserved set that preserves all passes.
PreservedAnalyses & preserveSet()
Mark an analysis set as preserved.
PreservedAnalyses & preserve()
Mark an analysis as preserved.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
void reserve(size_type N)
typename SuperClass::const_iterator const_iterator
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
typename SuperClass::iterator iterator
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Analysis pass providing the TargetLibraryInfo.
Provides information about what library functions are available for the current target.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
LLVM_ABI unsigned getIntegerBitWidth() const
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
void setOperand(unsigned i, Value *Val)
Value * getOperand(unsigned i) const
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
bool hasOneUse() const
Return true if there is exactly one use of this value.
LLVM_ABI void replaceAllUsesWith(Value *V)
Change all uses of this to point to a new Value.
LLVM_ABI const Value * stripPointerCasts() const
Strip off pointer casts, all-zero GEPs and address space casts.
LLVM_ABI std::optional< int64_t > getPointerOffsetFrom(const Value *Other, const DataLayout &DL) const
If this ptr is provably equal to Other plus a constant offset, return that offset in bytes.
constexpr ScalarTy getFixedValue() const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
const ParentTy * getParent() const
reverse_self_iterator getReverseIterator()
self_iterator getIterator()
This provides a very simple, boring adaptor for a begin and end iterator into a range type.
Abstract Attribute helper functions.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
@ C
The default llvm calling convention, compatible with C.
@ BasicBlock
Various leaf nodes.
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > Tys={})
Look up the Function declaration of the intrinsic id in the Module M.
@ User
could "use" a pointer
friend class Instruction
Iterator for Instructions in a `BasicBlock.
LLVM_ABI iterator begin() const
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
bool capturesAddress(CaptureComponents CC)
LLVM_ABI bool isPotentiallyReachableFromMany(SmallVectorImpl< BasicBlock * > &Worklist, const BasicBlock *StopBB, const SmallPtrSetImpl< BasicBlock * > *ExclusionSet, const DominatorTree *DT=nullptr, const LoopInfo *LI=nullptr)
Determine whether there is at least one path from a block in 'Worklist' to 'StopBB' without passing t...
LLVM_ABI void setExplicitlyUnknownBranchWeightsIfProfiled(Instruction &I, StringRef PassName, const Function *F=nullptr)
Like setExplicitlyUnknownBranchWeights(...), but only sets unknown branch weights in the new instruct...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
scope_exit(Callable) -> scope_exit< Callable >
LLVM_ABI bool isDereferenceableAndAlignedPointer(const Value *V, Type *Ty, Align Alignment, const DataLayout &DL, const Instruction *CtxI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr, const TargetLibraryInfo *TLI=nullptr)
Returns true if V is always a dereferenceable pointer with alignment greater or equal than requested.
auto partition_point(R &&Range, Predicate P)
Binary search for the first iterator in a range where a predicate is false.
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
auto cast_or_null(const Y &Val)
LLVM_ABI unsigned getDefaultMaxUsesToExploreForCaptureTracking()
getDefaultMaxUsesToExploreForCaptureTracking - Return default value of the maximal number of uses to ...
LLVM_ABI bool PointerMayBeCapturedBefore(const Value *V, bool ReturnCaptures, const Instruction *I, const DominatorTree *DT, bool IncludeI=false, unsigned MaxUsesToExplore=0, const LoopInfo *LI=nullptr)
PointerMayBeCapturedBefore - Return true if this pointer value may be captured by the enclosing funct...
LLVM_ABI Value * simplifyInstruction(Instruction *I, const SimplifyQuery &Q)
See if we can compute a simplified version of this instruction.
auto dyn_cast_or_null(const Y &Val)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
LLVM_ABI Align getOrEnforceKnownAlignment(Value *V, MaybeAlign PrefAlign, const DataLayout &DL, const Instruction *CxtI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr)
Try to ensure that the alignment of V is at least PrefAlign bytes.
bool isModSet(const ModRefInfo MRI)
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
bool isModOrRefSet(const ModRefInfo MRI)
LLVM_ABI bool isNotVisibleOnUnwind(const Value *Object, bool &RequiresNoCaptureBeforeUnwind)
Return true if Object memory is not visible after an unwind, in the sense that program semantics cann...
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
RNSuccIterator< NodeRef, BlockT, RegionT > succ_begin(NodeRef Node)
ModRefInfo
Flags indicating whether a memory access modifies or references memory.
IRBuilder(LLVMContext &, FolderTy, InserterTy, MDNode *, ArrayRef< OperandBundleDef >) -> IRBuilder< FolderTy, InserterTy >
LLVM_ABI bool VerifyMemorySSA
Enables verification of MemorySSA.
RNSuccIterator< NodeRef, BlockT, RegionT > succ_end(NodeRef Node)
DWARFExpression::Operation Op
LLVM_ABI bool isIdentifiedFunctionLocal(const Value *V)
Return true if V is umabigously identified at the function-level.
LLVM_ABI bool isGuaranteedToTransferExecutionToSuccessor(const Instruction *I)
Return true if this function can prove that the instruction I will always transfer execution to one o...
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI Value * isBytewiseValue(Value *V, const DataLayout &DL)
If the specified value can be set by repeating the same byte in memory, return the i8 value that it i...
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
LLVM_ABI void combineAAMetadata(Instruction *K, const Instruction *J)
Combine metadata of two instructions, where instruction J is a memory access that has been merged int...
bool capturesAnything(CaptureComponents CC)
LLVM_ABI UseCaptureInfo DetermineUseCaptureKind(const Use &U, const Value *Base)
Determine what kind of capture behaviour U may exhibit.
AnalysisManager< Function > FunctionAnalysisManager
Convenience typedef for the Function analysis manager.
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
AAResults AliasAnalysis
Temporary typedef for legacy code that uses a generic AliasAnalysis pointer or reference.
bool capturesAnyProvenance(CaptureComponents CC)
bool isRefSet(const ModRefInfo MRI)
LLVM_ABI bool isWritableObject(const Value *Object, bool &ExplicitlyDereferenceableOnly)
Return true if the Object is writable, in the sense that any location based on this pointer that can ...
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Align valueOrOne() const
For convenience, returns a valid alignment or 1 if undefined.
CaptureComponents UseCC
Components captured by this use.
CaptureComponents ResultCC
Components captured by the return value of the user of this Use.