85#define DEBUG_TYPE "inline-function"
94 cl::desc(
"Convert noalias attributes to metadata during inlining."));
99 cl::desc(
"Use the llvm.experimental.noalias.scope.decl "
100 "intrinsic during inlining."));
108 cl::desc(
"Convert align attributes to assumptions during inlining."));
111 "max-inst-checked-for-throw-during-inlining",
cl::Hidden,
112 cl::desc(
"the maximum number of instructions analyzed for may throw during "
113 "attribute inference in inlined body"),
119 class LandingPadInliningInfo {
130 PHINode *InnerEHValuesPHI =
nullptr;
136 : OuterResumeDest(
II->getUnwindDest()) {
145 UnwindDestPHIValues.
push_back(
PHI->getIncomingValueForBlock(InvokeBB));
154 return OuterResumeDest;
159 LandingPadInst *getLandingPadInst()
const {
return CallerLPad; }
166 void forwardResume(ResumeInst *RI,
167 SmallPtrSetImpl<LandingPadInst*> &InlinedLPads);
171 void addIncomingPHIValuesFor(BasicBlock *BB)
const {
172 addIncomingPHIValuesForInto(BB, OuterResumeDest);
175 void addIncomingPHIValuesForInto(BasicBlock *src, BasicBlock *dest)
const {
177 for (
unsigned i = 0, e = UnwindDestPHIValues.size(); i != e; ++i, ++
I) {
179 phi->addIncoming(UnwindDestPHIValues[i], src);
187 while (It != BB.
end()) {
189 if (IntrinsicCall->isEntry()) {
190 return IntrinsicCall;
199BasicBlock *LandingPadInliningInfo::getInnerResumeDest() {
200 if (InnerResumeDest)
return InnerResumeDest;
206 OuterResumeDest->
getName() +
".body");
209 const unsigned PHICapacity = 2;
214 for (
unsigned i = 0, e = UnwindDestPHIValues.
size(); i != e; ++i, ++
I) {
217 OuterPHI->
getName() +
".lpad-body");
228 InnerEHValuesPHI->
addIncoming(CallerLPad, OuterResumeDest);
231 return InnerResumeDest;
238void LandingPadInliningInfo::forwardResume(
239 ResumeInst *RI, SmallPtrSetImpl<LandingPadInst *> &InlinedLPads) {
248 addIncomingPHIValuesForInto(Src, Dest);
257 return FPI->getParentPad();
269 while (!Worklist.
empty()) {
276 Value *UnwindDestToken =
nullptr;
278 if (CatchSwitch->hasUnwindDest()) {
279 UnwindDestToken = &*CatchSwitch->getUnwindDest()->getFirstNonPHIIt();
287 for (
auto HI = CatchSwitch->handler_begin(),
288 HE = CatchSwitch->handler_end();
289 HI != HE && !UnwindDestToken; ++HI) {
302 auto Memo = MemoMap.
find(ChildPad);
303 if (Memo == MemoMap.
end()) {
310 Value *ChildUnwindDestToken = Memo->second;
311 if (!ChildUnwindDestToken)
318 UnwindDestToken = ChildUnwindDestToken;
329 if (
BasicBlock *RetUnwindDest = CleanupRet->getUnwindDest())
330 UnwindDestToken = &*RetUnwindDest->getFirstNonPHIIt();
335 Value *ChildUnwindDestToken;
337 ChildUnwindDestToken = &*Invoke->getUnwindDest()->getFirstNonPHIIt();
340 auto Memo = MemoMap.
find(ChildPad);
341 if (Memo == MemoMap.
end()) {
348 ChildUnwindDestToken = Memo->second;
349 if (!ChildUnwindDestToken)
361 UnwindDestToken = ChildUnwindDestToken;
367 if (!UnwindDestToken)
378 UnwindParent =
nullptr;
379 bool ExitedOriginalPad =
false;
381 ExitedPad && ExitedPad != UnwindParent;
386 MemoMap[ExitedPad] = UnwindDestToken;
387 ExitedOriginalPad |= (ExitedPad == EHPad);
390 if (ExitedOriginalPad)
391 return UnwindDestToken;
423 EHPad = CPI->getCatchSwitch();
426 auto Memo = MemoMap.
find(EHPad);
427 if (Memo != MemoMap.
end())
432 assert((UnwindDestToken ==
nullptr) != (MemoMap.
count(EHPad) != 0));
434 return UnwindDestToken;
441 MemoMap[EHPad] =
nullptr;
447 Value *AncestorToken;
461 assert(!MemoMap.
count(AncestorPad) || MemoMap[AncestorPad]);
462 auto AncestorMemo = MemoMap.
find(AncestorPad);
463 if (AncestorMemo == MemoMap.
end()) {
466 UnwindDestToken = AncestorMemo->second;
470 LastUselessPad = AncestorPad;
471 MemoMap[LastUselessPad] =
nullptr;
473 TempMemos.
insert(LastUselessPad);
491 while (!Worklist.
empty()) {
493 auto Memo = MemoMap.
find(UselessPad);
494 if (Memo != MemoMap.
end() && Memo->second) {
522 MemoMap[UselessPad] = UnwindDestToken;
524 assert(CatchSwitch->getUnwindDest() ==
nullptr &&
"Expected useless pad");
525 for (
BasicBlock *HandlerBlock : CatchSwitch->handlers()) {
526 auto *CatchPad = &*HandlerBlock->getFirstNonPHIIt();
531 ->getFirstNonPHIIt()) == CatchPad)) &&
532 "Expected useless pad");
546 "Expected useless pad");
553 return UnwindDestToken;
580 if (
F->getIntrinsicID() == Intrinsic::experimental_deoptimize ||
581 F->getIntrinsicID() == Intrinsic::experimental_guard)
593 Value *UnwindDestToken =
600 MemoKey = CatchPad->getCatchSwitch();
602 MemoKey = FuncletPad;
603 assert(FuncletUnwindMap->count(MemoKey) &&
604 (*FuncletUnwindMap)[MemoKey] == UnwindDestToken &&
605 "must get memoized to avoid confusing later searches");
609 bool WasIndirect = OriginallyIndirectCalls.
remove(CI);
633 LandingPadInliningInfo Invoke(
II);
640 InlinedLPads.
insert(
II->getLandingPadInst());
647 InlinedLPad->reserveClauses(OuterNum);
648 for (
unsigned OuterIdx = 0; OuterIdx != OuterNum; ++OuterIdx)
649 InlinedLPad->addClause(OuterLPad->
getClause(OuterIdx));
651 InlinedLPad->setCleanup(
true);
658 &*BB, Invoke.getOuterResumeDest(),
662 Invoke.addIncomingPHIValuesFor(NewBB);
666 Invoke.forwardResume(RI, InlinedLPads);
696 UnwindDestPHIValues.
push_back(
PHI.getIncomingValueForBlock(InvokeBB));
703 for (
Value *V : UnwindDestPHIValues) {
705 PHI->addIncoming(V, Src);
716 if (CRI->unwindsToCaller()) {
717 auto *CleanupPad = CRI->getCleanupPad();
719 CRI->eraseFromParent();
727 FuncletUnwindMap[CleanupPad] =
738 if (CatchSwitch->unwindsToCaller()) {
739 Value *UnwindDestToken;
740 if (
auto *ParentPad =
764 CatchSwitch->getParentPad(), UnwindDest,
765 CatchSwitch->getNumHandlers(), CatchSwitch->
getName(),
766 CatchSwitch->getIterator());
767 for (
BasicBlock *PadBB : CatchSwitch->handlers())
768 NewCatchSwitch->addHandler(PadBB);
773 FuncletUnwindMap[NewCatchSwitch] = UnwindDestToken;
774 Replacement = NewCatchSwitch;
782 I->replaceAllUsesWith(Replacement);
783 I->eraseFromParent();
807 MDNode *CallsiteStackContext) {
813 for (
auto MIBStackIter = MIBStackContext->
op_begin(),
814 CallsiteStackIter = CallsiteStackContext->
op_begin();
815 MIBStackIter != MIBStackContext->
op_end() &&
816 CallsiteStackIter != CallsiteStackContext->
op_end();
817 MIBStackIter++, CallsiteStackIter++) {
821 if (Val1->getZExtValue() != Val2->getZExtValue())
828 Call->setMetadata(LLVMContext::MD_memprof,
nullptr);
832 Call->setMetadata(LLVMContext::MD_callsite,
nullptr);
836 const std::vector<Metadata *> &MIBList,
845 bool MemprofMDAttached =
CallStack.buildAndAttachMIBMetadata(CI);
847 if (!MemprofMDAttached)
857 MDNode *InlinedCallsiteMD,
860 MDNode *ClonedCallsiteMD =
nullptr;
863 if (OrigCallsiteMD) {
868 ClonedCall->
setMetadata(LLVMContext::MD_callsite, ClonedCallsiteMD);
880 std::vector<Metadata *> NewMIBList;
885 for (
auto &MIBOp : OrigMemProfMD->
operands()) {
893 NewMIBList.push_back(MIB);
895 if (NewMIBList.empty()) {
911 bool ContainsMemProfMetadata,
917 if (!CallsiteMD && !ContainsMemProfMetadata)
921 for (
const auto &Entry : VMap) {
926 if (!OrigCall || !ClonedCall)
945 MDNode *MemParallelLoopAccess =
946 CB.
getMetadata(LLVMContext::MD_mem_parallel_loop_access);
950 if (!MemParallelLoopAccess && !AccessGroup && !AliasScope && !NoAlias)
956 if (!
I.mayReadOrWriteMemory())
959 if (MemParallelLoopAccess) {
962 I.getMetadata(LLVMContext::MD_mem_parallel_loop_access),
963 MemParallelLoopAccess);
964 I.setMetadata(LLVMContext::MD_mem_parallel_loop_access,
965 MemParallelLoopAccess);
970 I.getMetadata(LLVMContext::MD_access_group), AccessGroup));
974 I.getMetadata(LLVMContext::MD_alias_scope), AliasScope));
978 I.getMetadata(LLVMContext::MD_noalias), NoAlias));
992 InlineSiteLoc = CI->getZExtValue();
1002 if (!CI || !CI->getMetadata(
"srcloc"))
1004 auto *Callee = CI->getCalledFunction();
1005 if (!Callee || (!Callee->hasFnAttribute(
"dontcall-error") &&
1006 !Callee->hasFnAttribute(
"dontcall-warn")))
1010 if (
MDNode *Existing = CI->getMetadata(
"inlined.from"))
1014 Ops.push_back(MakeMDInt(0));
1017 Ops.push_back(MakeMDInt(InlineSiteLoc));
1037 if (CalledFn && CalledFn->isIntrinsic() &&
I->doesNotThrow() &&
1042 I->getOperandBundlesAsDefs(OpBundles);
1047 I->replaceAllUsesWith(NewInst);
1048 I->eraseFromParent();
1057class ScopedAliasMetadataDeepCloner {
1058 using MetadataMap = DenseMap<const MDNode *, TrackingMDNodeRef>;
1059 SetVector<const MDNode *> MD;
1061 void addRecursiveMetadataUses();
1064 ScopedAliasMetadataDeepCloner(
const Function *
F);
1076ScopedAliasMetadataDeepCloner::ScopedAliasMetadataDeepCloner(
1077 const Function *
F) {
1078 for (
const BasicBlock &BB : *
F) {
1079 for (
const Instruction &
I : BB) {
1080 if (
const MDNode *M =
I.getMetadata(LLVMContext::MD_alias_scope))
1082 if (
const MDNode *M =
I.getMetadata(LLVMContext::MD_noalias))
1087 MD.insert(Decl->getScopeList());
1090 addRecursiveMetadataUses();
1093void ScopedAliasMetadataDeepCloner::addRecursiveMetadataUses() {
1095 while (!
Queue.empty()) {
1099 if (MD.insert(OpMD))
1100 Queue.push_back(OpMD);
1104void ScopedAliasMetadataDeepCloner::clone() {
1105 assert(MDMap.
empty() &&
"clone() already called ?");
1108 for (
const MDNode *
I : MD) {
1110 MDMap[
I].reset(DummyNodes.
back().get());
1117 for (
const MDNode *
I : MD) {
1139 for (BasicBlock &BB :
make_range(FStart, FEnd)) {
1140 for (Instruction &
I : BB) {
1143 if (MDNode *M =
I.getMetadata(LLVMContext::MD_alias_scope))
1144 if (MDNode *MNew = MDMap.
lookup(M))
1145 I.setMetadata(LLVMContext::MD_alias_scope, MNew);
1147 if (MDNode *M =
I.getMetadata(LLVMContext::MD_noalias))
1148 if (MDNode *MNew = MDMap.
lookup(M))
1149 I.setMetadata(LLVMContext::MD_noalias, MNew);
1152 if (MDNode *MNew = MDMap.
lookup(Decl->getScopeList()))
1153 Decl->setScopeList(MNew);
1172 if (CB.
paramHasAttr(Arg.getArgNo(), Attribute::NoAlias) && !Arg.use_empty())
1175 if (NoAliasArgs.
empty())
1195 for (
unsigned i = 0, e = NoAliasArgs.
size(); i != e; ++i) {
1198 std::string Name = std::string(CalledFunc->
getName());
1201 Name +=
A->getName();
1203 Name +=
": argument ";
1211 NewScopes.
insert(std::make_pair(
A, NewScope));
1218 IRBuilder<>(&CB).CreateNoAliasScopeDeclaration(AScopeList);
1228 VMI != VMIE; ++VMI) {
1237 bool IsArgMemOnlyCall =
false, IsFuncCall =
false;
1241 PtrArgs.
push_back(LI->getPointerOperand());
1245 PtrArgs.
push_back(VAAI->getPointerOperand());
1247 PtrArgs.
push_back(CXI->getPointerOperand());
1249 PtrArgs.
push_back(RMWI->getPointerOperand());
1254 if (
Call->doesNotAccessMemory())
1266 IsArgMemOnlyCall =
true;
1273 if (!Arg->getType()->isPointerTy())
1284 if (PtrArgs.
empty() && !IsFuncCall)
1293 for (
const Value *V : PtrArgs) {
1302 bool RequiresNoCaptureBefore =
false, UsesAliasingPtr =
false,
1303 UsesUnknownObject =
false;
1304 for (
const Value *V : ObjSet) {
1319 UsesAliasingPtr =
true;
1321 UsesAliasingPtr =
true;
1327 RequiresNoCaptureBefore =
true;
1333 UsesUnknownObject =
true;
1339 if (UsesUnknownObject)
1344 if (IsFuncCall && !IsArgMemOnlyCall)
1345 RequiresNoCaptureBefore =
true;
1363 if (!RequiresNoCaptureBefore ||
1365 A,
false,
I, &DT,
false,
1386 bool CanAddScopes = !UsesAliasingPtr;
1387 if (CanAddScopes && IsFuncCall)
1388 CanAddScopes = IsArgMemOnlyCall;
1393 Scopes.push_back(NewScopes[
A]);
1396 if (!Scopes.empty())
1398 LLVMContext::MD_alias_scope,
1409 "Expected to be in same basic block!");
1422 auto &Context = CalledFunction->
getContext();
1426 bool HasAttrToPropagate =
false;
1434 Attribute::Dereferenceable, Attribute::DereferenceableOrNull,
1435 Attribute::NonNull, Attribute::NoFPClass,
1436 Attribute::Alignment, Attribute::Range};
1444 ValidObjParamAttrs.
back().addAttribute(Attribute::ReadNone);
1446 ValidObjParamAttrs.
back().addAttribute(Attribute::ReadOnly);
1451 ValidExactParamAttrs.
back().addAttribute(Attr);
1454 HasAttrToPropagate |= ValidObjParamAttrs.
back().hasAttributes();
1455 HasAttrToPropagate |= ValidExactParamAttrs.
back().hasAttributes();
1459 if (!HasAttrToPropagate)
1472 if (InlinedFunctionInfo.
isSimplified(InnerCB, NewInnerCB))
1475 AttributeList AL = NewInnerCB->getAttributes();
1476 for (
unsigned I = 0,
E = InnerCB->arg_size();
I <
E; ++
I) {
1481 if (NewInnerCB->paramHasAttr(
I, Attribute::ByVal))
1485 if (
match(NewInnerCB->getArgOperand(
I),
1501 if (AL.getParamDereferenceableBytes(
I) >
1502 NewAB.getDereferenceableBytes())
1503 NewAB.removeAttribute(Attribute::Dereferenceable);
1504 if (AL.getParamDereferenceableOrNullBytes(
I) >
1505 NewAB.getDereferenceableOrNullBytes())
1506 NewAB.removeAttribute(Attribute::DereferenceableOrNull);
1507 if (AL.getParamAlignment(
I).valueOrOne() >
1508 NewAB.getAlignment().valueOrOne())
1509 NewAB.removeAttribute(Attribute::Alignment);
1510 if (
auto ExistingRange = AL.getParamRange(
I)) {
1511 if (
auto NewRange = NewAB.getRange()) {
1514 NewAB.removeAttribute(Attribute::Range);
1515 NewAB.addRangeAttr(CombinedRange);
1519 if (
FPClassTest ExistingNoFP = AL.getParamNoFPClass(
I))
1520 NewAB.addNoFPClassAttr(ExistingNoFP | NewAB.getNoFPClass());
1522 AL = AL.addParamAttributes(Context,
I, NewAB);
1523 }
else if (NewInnerCB->getArgOperand(
I)->getType()->isPointerTy()) {
1525 const Value *UnderlyingV =
1536 AL = AL.addParamAttributes(Context,
I, ValidObjParamAttrs[ArgNo]);
1543 if (AL.hasParamAttr(
I, Attribute::ReadOnly) &&
1544 AL.hasParamAttr(
I, Attribute::WriteOnly))
1545 AL = AL.addParamAttribute(Context,
I, Attribute::ReadNone);
1548 if (AL.hasParamAttr(
I, Attribute::ReadNone)) {
1549 AL = AL.removeParamAttribute(Context,
I, Attribute::ReadOnly);
1550 AL = AL.removeParamAttribute(Context,
I, Attribute::WriteOnly);
1554 if (AL.hasParamAttr(
I, Attribute::ReadOnly) ||
1555 AL.hasParamAttr(
I, Attribute::ReadNone))
1556 AL = AL.removeParamAttribute(Context,
I, Attribute::Writable);
1558 NewInnerCB->setAttributes(AL);
1572 Valid.addDereferenceableAttr(DerefBytes);
1574 Valid.addDereferenceableOrNullAttr(DerefOrNullBytes);
1576 Valid.addAttribute(Attribute::NoAlias);
1578 Valid.addAttribute(Attribute::NoUndef);
1587 Valid.addAttribute(Attribute::NonNull);
1591 Valid.addRangeAttr(*
Range);
1601 if (!CallSiteValidUB.hasAttributes() && !CallSiteValidPG.hasAttributes())
1604 auto &Context = CalledFunction->
getContext();
1606 for (
auto &BB : *CalledFunction) {
1620 if (InlinedFunctionInfo.
isSimplified(RetVal, NewRetVal))
1640 if (RI->
getParent() != RetVal->getParent() ||
1651 AttributeList AL = NewRetVal->getAttributes();
1652 if (ValidUB.getDereferenceableBytes() < AL.getRetDereferenceableBytes())
1653 ValidUB.removeAttribute(Attribute::Dereferenceable);
1654 if (ValidUB.getDereferenceableOrNullBytes() <
1655 AL.getRetDereferenceableOrNullBytes())
1656 ValidUB.removeAttribute(Attribute::DereferenceableOrNull);
1657 AttributeList NewAL = AL.addRetAttributes(Context, ValidUB);
1690 if (ValidPG.getAlignment().valueOrOne() < AL.getRetAlignment().valueOrOne())
1691 ValidPG.removeAttribute(Attribute::Alignment);
1692 if (ValidPG.hasAttributes()) {
1693 Attribute CBRange = ValidPG.getAttribute(Attribute::Range);
1695 Attribute NewRange = AL.getRetAttr(Attribute::Range);
1697 ValidPG.addRangeAttr(
1702 Attribute CBNoFPClass = ValidPG.getAttribute(Attribute::NoFPClass);
1703 if (CBNoFPClass.
isValid() && AL.hasRetAttr(Attribute::NoFPClass)) {
1704 ValidPG.addNoFPClassAttr(
1706 AL.getRetAttr(Attribute::NoFPClass).getNoFPClass());
1721 (RetVal->hasOneUse() && !RetVal->hasRetAttr(Attribute::NoUndef)))
1722 NewAL = NewAL.addRetAttributes(Context, ValidPG);
1724 NewRetVal->setAttributes(NewAL);
1740 bool DTCalculated =
false;
1744 if (!Arg.getType()->isPointerTy() || Arg.hasPassPointeeByValueCopyAttr() ||
1751 if (!DTCalculated) {
1753 DTCalculated =
true;
1762 DL, ArgVal, Alignment->value());
1775 Builder.getInt64(M->getDataLayout().getTypeStoreSize(ByValType));
1777 Align DstAlign = Dst->getPointerAlignment(M->getDataLayout());
1780 CallInst *CI = Builder.CreateMemCpy(Dst, DstAlign, Src, SrcAlign,
Size);
1824 Align Alignment =
DL.getPrefTypeAlign(ByValType);
1830 Alignment = std::max(Alignment, *ByValAlignment);
1834 nullptr, Alignment, Arg->
getName());
1858 if (Ty == Int8PtrTy)
1863 if (U->getType() != Int8PtrTy)
continue;
1864 if (U->stripPointerCasts() != AI)
continue;
1909 Ctx, InlinedAtNode->getLine(), InlinedAtNode->getColumn(),
1910 InlinedAtNode->getScope(), InlinedAtNode->getInlinedAt());
1919 bool NoInlineLineTables = Fn->
hasFnAttribute(
"no-inline-line-tables");
1925 auto updateLoopInfoLoc = [&Ctx, &InlinedAtNode,
1933 if (!NoInlineLineTables)
1941 if (CalleeHasDebugInfo && !NoInlineLineTables)
1961 I.setDebugLoc(TheCallDL);
1966 assert(DVR->getDebugLoc() &&
"Debug Value must have debug loc");
1967 if (NoInlineLineTables) {
1968 DVR->setDebugLoc(TheCallDL);
1974 DVR->getMarker()->getParent()->
getContext(), IANodes);
1975 DVR->setDebugLoc(IDL);
1979 for (; FI != Fn->
end(); ++FI) {
1982 for (
DbgRecord &DVR :
I.getDbgRecordRange()) {
1988 if (NoInlineLineTables) {
1990 while (BI != FI->end()) {
1991 BI->dropDbgRecords();
1999#define DEBUG_TYPE "assignment-tracking"
2007 errs() <<
"# Finding caller local variables escaped by callee\n");
2010 if (!Arg->getType()->isPointerTy()) {
2022 assert(Arg->getType()->isPtrOrPtrVectorTy());
2023 APInt TmpOffset(
DL.getIndexTypeSizeInBits(Arg->getType()), 0,
false);
2025 Arg->stripAndAccumulateConstantOffsets(
DL, TmpOffset,
true));
2027 LLVM_DEBUG(
errs() <<
" | SKIP: Couldn't walk back to base storage\n");
2040 if (DbgAssign->getDebugLoc().getInlinedAt())
2047 return EscapedLocals;
2053 <<
Start->getParent()->getName() <<
" from "
2066 for (
auto BBI =
Start; BBI != End; ++BBI) {
2072#define DEBUG_TYPE "inline-function"
2086 for (
auto Entry : VMap) {
2092 if (!ClonedBBs.
insert(ClonedBB).second) {
2104 EntryClone, CallerBFI->
getBlockFreq(CallSiteBlock), ClonedBBs);
2114 auto CallSiteCount =
2117 std::min(CallSiteCount.value_or(0), CalleeEntryCount.
getCount());
2122 Function *Callee, int64_t EntryDelta,
2124 auto CalleeCount = Callee->getEntryCount();
2128 const uint64_t PriorEntryCount = CalleeCount->getCount();
2133 (EntryDelta < 0 && static_cast<uint64_t>(-EntryDelta) > PriorEntryCount)
2135 : PriorEntryCount + EntryDelta;
2137 auto updateVTableProfWeight = [](
CallBase *CB,
const uint64_t NewEntryCount,
2146 uint64_t CloneEntryCount = PriorEntryCount - NewEntryCount;
2147 for (
auto Entry : *VMap) {
2150 CI->updateProfWeight(CloneEntryCount, PriorEntryCount);
2151 updateVTableProfWeight(CI, CloneEntryCount, PriorEntryCount);
2156 II->updateProfWeight(CloneEntryCount, PriorEntryCount);
2157 updateVTableProfWeight(
II, CloneEntryCount, PriorEntryCount);
2163 Callee->setEntryCount(NewEntryCount);
2167 if (!VMap || VMap->
count(&BB))
2170 CI->updateProfWeight(NewEntryCount, PriorEntryCount);
2171 updateVTableProfWeight(CI, NewEntryCount, PriorEntryCount);
2174 II->updateProfWeight(NewEntryCount, PriorEntryCount);
2175 updateVTableProfWeight(
II, NewEntryCount, PriorEntryCount);
2203 IsUnsafeClaimRV = !IsRetainRV;
2205 for (
auto *RI : Returns) {
2207 bool InsertRetainCall = IsRetainRV;
2220 if (
II->getIntrinsicID() != Intrinsic::objc_autoreleaseReturnValue ||
2230 if (IsUnsafeClaimRV) {
2231 Builder.SetInsertPoint(
II);
2232 Builder.CreateIntrinsic(Intrinsic::objc_release, RetOpnd);
2234 II->eraseFromParent();
2235 InsertRetainCall =
false;
2254 NewCall->copyMetadata(*CI);
2255 CI->replaceAllUsesWith(NewCall);
2256 CI->eraseFromParent();
2257 InsertRetainCall =
false;
2261 if (InsertRetainCall) {
2265 Builder.SetInsertPoint(RI);
2266 Builder.CreateIntrinsic(Intrinsic::objc_retain, RetOpnd);
2292static std::pair<std::vector<int64_t>, std::vector<int64_t>>
2300 std::vector<int64_t> CalleeCounterMap;
2301 std::vector<int64_t> CalleeCallsiteMap;
2302 CalleeCounterMap.resize(CalleeCounters, -1);
2303 CalleeCallsiteMap.resize(CalleeCallsites, -1);
2306 if (Ins.getNameValue() == &Caller)
2308 const auto OldID =
static_cast<uint32_t>(Ins.getIndex()->getZExtValue());
2309 if (CalleeCounterMap[OldID] == -1)
2311 const auto NewID =
static_cast<uint32_t>(CalleeCounterMap[OldID]);
2313 Ins.setNameValue(&Caller);
2314 Ins.setIndex(NewID);
2319 if (Ins.getNameValue() == &Caller)
2321 const auto OldID =
static_cast<uint32_t>(Ins.getIndex()->getZExtValue());
2322 if (CalleeCallsiteMap[OldID] == -1)
2324 const auto NewID =
static_cast<uint32_t>(CalleeCallsiteMap[OldID]);
2326 Ins.setNameValue(&Caller);
2327 Ins.setIndex(NewID);
2331 std::deque<BasicBlock *> Worklist;
2348 Worklist.push_back(StartBB);
2349 while (!Worklist.empty()) {
2350 auto *BB = Worklist.front();
2351 Worklist.pop_front();
2355 Changed |= RewriteInstrIfNeeded(*BBID);
2359 BBID->moveBefore(BB->getFirstInsertionPt());
2372 Inc->eraseFromParent();
2375 RewriteInstrIfNeeded(*Inc);
2377 }
else if (Inc != BBID) {
2382 Inc->eraseFromParent();
2386 Changed |= RewriteCallsiteInsIfNeeded(*CS);
2391 if (Seen.
insert(Succ).second)
2392 Worklist.push_back(Succ);
2396 "Counter index mapping should be either to -1 or to non-zero index, "
2398 "index corresponds to the entry BB of the caller");
2400 "Callsite index mapping should be either to -1 or to non-zero index, "
2401 "because there should have been at least a callsite - the inlined one "
2402 "- which would have had a 0 index.");
2404 return {std::move(CalleeCounterMap), std::move(CalleeCallsiteMap)};
2424 AAResults *CalleeAAR,
bool InsertLifetime,
2425 bool TrackInlineHistory,
Function *ForwardVarArgsTo,
2428 return InlineFunction(CB, IFI, MergeAttributes, CalleeAAR, InsertLifetime,
2429 TrackInlineHistory, ForwardVarArgsTo, ORE);
2441 static_cast<uint32_t>(CallsiteIDIns->getIndex()->getZExtValue());
2446 auto Ret =
InlineFunction(CB, IFI, MergeAttributes, CalleeAAR, InsertLifetime,
2447 TrackInlineHistory, ForwardVarArgsTo, ORE);
2448 if (!Ret.isSuccess())
2453 CallsiteIDIns->eraseFromParent();
2458 const auto IndicesMaps =
remapIndices(Caller, StartBB, CtxProf,
2459 NumCalleeCounters, NumCalleeCallsites);
2464 const auto &[CalleeCounterMap, CalleeCallsiteMap] = IndicesMaps;
2466 (Ctx.counters().size() +
2467 llvm::count_if(CalleeCounterMap, [](
auto V) { return V != -1; }) ==
2469 "The caller's counters size should have grown by the number of new "
2470 "distinct counters inherited from the inlined callee.");
2471 Ctx.resizeCounters(NewCountersSize);
2475 auto CSIt = Ctx.callsites().find(
CallsiteID);
2476 if (CSIt == Ctx.callsites().end())
2478 auto CalleeCtxIt = CSIt->second.find(CalleeGUID);
2481 if (CalleeCtxIt == CSIt->second.end())
2486 auto &CalleeCtx = CalleeCtxIt->second;
2487 assert(CalleeCtx.guid() == CalleeGUID);
2489 for (
auto I = 0U;
I < CalleeCtx.counters().
size(); ++
I) {
2490 const int64_t NewIndex = CalleeCounterMap[
I];
2491 if (NewIndex >= 0) {
2492 assert(NewIndex != 0 &&
"counter index mapping shouldn't happen to a 0 "
2493 "index, that's the caller's entry BB");
2494 Ctx.counters()[NewIndex] = CalleeCtx.counters()[
I];
2497 for (
auto &[
I, OtherSet] : CalleeCtx.callsites()) {
2498 const int64_t NewCSIdx = CalleeCallsiteMap[
I];
2499 if (NewCSIdx >= 0) {
2501 "callsite index mapping shouldn't happen to a 0 index, the "
2502 "caller must've had at least one callsite (with such an index)");
2503 Ctx.ingestAllContexts(NewCSIdx, std::move(OtherSet));
2513 CtxProf.
update(Updater, Caller);
2536 for (
const auto &
Op : InlineHistory->operands()) {
2538 if (MD->getValue() == CalledFunc) {
2583 "convergent call needs convergencectrl operand");
2594 if (CalledFunc->
hasGC()) {
2595 if (Caller->hasGC() && CalledFunc->
getGC() != Caller->getGC())
2609 Caller->hasPersonalityFn()
2610 ? Caller->getPersonalityFn()->stripPointerCasts()
2612 if (CalledPersonality) {
2617 if (CallerPersonality && CalledPersonality != CallerPersonality)
2623 if (CallerPersonality) {
2626 std::optional<OperandBundleUse> ParentFunclet =
2640 for (
const BasicBlock &CalledBB : *CalledFunc) {
2648 for (
const BasicBlock &CalledBB : *CalledFunc) {
2649 if (CalledBB.isEHPad())
2669 bool MergeAttributes,
AAResults *CalleeAAR,
2670 bool InsertLifetime,
bool TrackInlineHistory,
2677 "CanInlineCallSite should have verified direct call to definition");
2681 bool EHPadForCallUnwindsLocally =
false;
2684 Value *CallSiteUnwindDestToken =
2687 EHPadForCallUnwindsLocally =
2688 CallSiteUnwindDestToken &&
2706 if (CalledFunc->
hasGC()) {
2707 if (!Caller->hasGC())
2708 Caller->setGC(CalledFunc->
getGC());
2711 "CanInlineCallSite should have verified compatible GCs");
2718 if (!Caller->hasPersonalityFn()) {
2719 Caller->setPersonalityFn(CalledPersonality);
2721 assert(Caller->getPersonalityFn()->stripPointerCasts() ==
2722 CalledPersonality &&
2723 "CanInlineCallSite should have verified compatible personality");
2747 auto &
DL = Caller->getDataLayout();
2754 E = CalledFunc->
arg_end();
I != E; ++
I, ++AI, ++ArgNo) {
2755 Value *ActualArg = *AI;
2763 &CB, CalledFunc, IFI,
2765 if (ActualArg != *AI)
2771 VMap[&*
I] = ActualArg;
2791 false, Returns,
".i",
2792 InlinedFunctionInfo);
2794 FirstNewBlock = LastBlock; ++FirstNewBlock;
2809 CalledFunc->
front());
2817 for (ByValInit &
Init : ByValInits)
2819 Caller->getParent(), &*FirstNewBlock, IFI,
2822 std::optional<OperandBundleUse> ParentDeopt =
2849 std::vector<Value *> MergedDeoptArgs;
2850 MergedDeoptArgs.reserve(ParentDeopt->Inputs.size() +
2851 ChildOB.Inputs.size());
2856 OpDefs.
emplace_back(
"deopt", std::move(MergedDeoptArgs));
2886 SAMetadataCloner.clone();
2887 SAMetadataCloner.remap(FirstNewBlock, Caller->end());
2907 if (CalledFunc->
hasMetadata(LLVMContext::MD_implicit_ref)) {
2909 CalledFunc->
getMetadata(LLVMContext::MD_implicit_ref, MDs);
2911 Caller->addMetadata(LLVMContext::MD_implicit_ref, *MD);
2917 FirstNewBlock, Caller->end());
2922 make_range(FirstNewBlock->getIterator(), Caller->end()))
2930 if (IntrinsicCall) {
2943 E = FirstNewBlock->end();
I != E; ) {
2973 Caller->getEntryBlock().splice(
InsertPoint, &*FirstNewBlock,
2990 bool InlinedMustTailCalls =
false, InlinedDeoptimizeCalls =
false;
2994 CallSiteTailKind = CI->getTailCallKind();
3009 if (!VarArgsToForward.
empty() &&
3010 ((ForwardVarArgsTo &&
3016 if (!Attrs.isEmpty() || !VarArgsAttrs.
empty()) {
3017 for (
unsigned ArgNo = 0;
3019 ArgAttrs.
push_back(Attrs.getParamAttrs(ArgNo));
3024 Attrs = AttributeList::get(CI->
getContext(), Attrs.getFnAttrs(),
3025 Attrs.getRetAttrs(), ArgAttrs);
3040 InlinedDeoptimizeCalls |=
3041 F->getIntrinsicID() == Intrinsic::experimental_deoptimize;
3060 ChildTCK = std::min(CallSiteTailKind, ChildTCK);
3079 if ((InsertLifetime || Caller->isPresplitCoroutine()) &&
3081 IRBuilder<> builder(&*FirstNewBlock, FirstNewBlock->begin());
3100 if (InlinedMustTailCalls &&
3101 RI->
getParent()->getTerminatingMustTailCall())
3103 if (InlinedDeoptimizeCalls &&
3104 RI->
getParent()->getTerminatingDeoptimizeCall())
3116 .CreateStackSave(
"savedstack");
3123 if (InlinedMustTailCalls && RI->
getParent()->getTerminatingMustTailCall())
3125 if (InlinedDeoptimizeCalls && RI->
getParent()->getTerminatingDeoptimizeCall())
3160 if (CleanupRet->unwindsToCaller() && EHPadForCallUnwindsLocally)
3178 if (InlinedDeoptimizeCalls) {
3184 if (Caller->getReturnType() == CB.
getType()) {
3186 return RI->
getParent()->getTerminatingDeoptimizeCall() !=
nullptr;
3191 Caller->getParent(), Intrinsic::experimental_deoptimize,
3192 {Caller->getReturnType()});
3218 "Expected at least the deopt operand bundle");
3222 Builder.CreateCall(NewDeoptIntrinsic, CallArgs, OpBundles);
3226 Builder.CreateRetVoid();
3228 Builder.CreateRet(NewDeoptCall);
3243 if (InlinedMustTailCalls) {
3245 Type *NewRetTy = Caller->getReturnType();
3252 RI->
getParent()->getTerminatingMustTailCall();
3253 if (!ReturnedMustTail) {
3265 OldCast->eraseFromParent();
3269 Builder.CreateRet(Builder.CreateBitCast(ReturnedMustTail, NewRetTy));
3285 make_range(FirstNewBlock->getIterator(), Caller->end()))
3297 if (TrackInlineHistory ||
3305 for (
const auto &
Op : CBHistory->operands()) {
3311 ICB->
getMetadata(LLVMContext::MD_inline_history)) {
3312 for (
const auto &
Op : CBHistory->operands()) {
3318 ICB->
setMetadata(LLVMContext::MD_inline_history, NewHistory);
3325 if (Returns.
size() == 1 && std::distance(FirstNewBlock, Caller->end()) == 1) {
3328 FirstNewBlock->end());
3330 Caller->back().eraseFromParent();
3344 if (&CB == R->getReturnValue())
3353 Returns[0]->eraseFromParent();
3355 if (MergeAttributes)
3356 AttributeFuncs::mergeAttributesForInlining(*Caller, *CalledFunc);
3373 CreatedBranchToNormalDest =
3383 CalledFunc->
getName() +
".exit");
3390 CalledFunc->
getName() +
".exit");
3408 Caller->splice(AfterCallBB->
getIterator(), Caller, FirstNewBlock,
3416 if (Returns.
size() > 1) {
3421 PHI->insertBefore(AfterCallBB->
begin());
3432 "Ret value not consistent in function!");
3433 PHI->addIncoming(RI->getReturnValue(), RI->
getParent());
3442 BI->setDebugLoc(
Loc);
3449 if (CreatedBranchToNormalDest)
3450 CreatedBranchToNormalDest->setDebugLoc(
Loc);
3451 }
else if (!Returns.
empty()) {
3455 if (&CB == Returns[0]->getReturnValue())
3462 BasicBlock *ReturnBB = Returns[0]->getParent();
3467 AfterCallBB->
splice(AfterCallBB->
begin(), ReturnBB);
3469 if (CreatedBranchToNormalDest)
3470 CreatedBranchToNormalDest->setDebugLoc(Returns[0]->
getDebugLoc());
3473 Returns[0]->eraseFromParent();
3480 if (CreatedBranchToNormalDest)
3492 if (InlinedMustTailCalls &&
pred_empty(AfterCallBB))
3502 OrigBB->
splice(Br->getIterator(), CalleeEntry);
3505 Br->eraseFromParent();
3516 auto &
DL = Caller->getDataLayout();
3518 PHI->replaceAllUsesWith(V);
3519 PHI->eraseFromParent();
3523 if (MergeAttributes)
3524 AttributeFuncs::mergeAttributesForInlining(*Caller, *CalledFunc);
3529 AAResults *CalleeAAR,
bool InsertLifetime,
bool TrackInlineHistory,
3532 if (Result.isSuccess()) {
3534 TrackInlineHistory, ForwardVarArgsTo, ORE);
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file contains the simple types necessary to represent the attributes associated with functions a...
static void UpdatePHINodes(BasicBlock *OrigBB, BasicBlock *NewBB, ArrayRef< BasicBlock * > Preds, Instruction *BI, bool HasLoopExit)
Update the PHI nodes in OrigBB to include the values coming from NewBB.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static cl::opt< bool > NoAliases("csky-no-aliases", cl::desc("Disable the emission of assembler pseudo instructions"), cl::init(false), cl::Hidden)
This file provides interfaces used to build and manipulate a call graph, which is a very useful tool ...
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file defines the DenseMap class.
This file provides various utilities for inspecting and working with the control flow graph in LLVM I...
Module.h This file contains the declarations for the Module class.
static AttrBuilder IdentifyValidUBGeneratingAttributes(CallBase &CB)
DenseMap< Instruction *, Value * > UnwindDestMemoTy
static BasicBlock * HandleCallsInBlockInlinedThroughInvoke(BasicBlock *BB, BasicBlock *UnwindEdge, SmallSetVector< const Value *, 4 > &OriginallyIndirectCalls, UnwindDestMemoTy *FuncletUnwindMap=nullptr)
When we inline a basic block into an invoke, we have to turn all of the calls that can throw into inv...
static at::StorageToVarsMap collectEscapedLocals(const DataLayout &DL, const CallBase &CB)
Find Alloca and linked DbgAssignIntrinsic for locals escaped by CB.
static void fixupLineNumbers(Function *Fn, Function::iterator FI, Instruction *TheCall, bool CalleeHasDebugInfo)
Update inlined instructions' line numbers to to encode location where these instructions are inlined.
static void removeCallsiteMetadata(CallBase *Call)
static void PropagateInlinedFromMetadata(CallBase &CB, StringRef CalledFuncName, StringRef CallerFuncName, Function::iterator FStart, Function::iterator FEnd)
Track inlining chain via inlined.from metadata for dontcall diagnostics.
static Value * getUnwindDestToken(Instruction *EHPad, UnwindDestMemoTy &MemoMap)
Given an EH pad, find where it unwinds.
static void propagateMemProfMetadata(Function *Callee, CallBase &CB, bool ContainsMemProfMetadata, const ValueMap< const Value *, WeakTrackingVH > &VMap, OptimizationRemarkEmitter *ORE)
static cl::opt< bool > PreserveAlignmentAssumptions("preserve-alignment-assumptions-during-inlining", cl::init(false), cl::Hidden, cl::desc("Convert align attributes to assumptions during inlining."))
static void HandleInlinedLandingPad(InvokeInst *II, BasicBlock *FirstNewBlock, ClonedCodeInfo &InlinedCodeInfo)
If we inlined an invoke site, we need to convert calls in the body of the inlined function into invok...
static Value * getUnwindDestTokenHelper(Instruction *EHPad, UnwindDestMemoTy &MemoMap)
Helper for getUnwindDestToken that does the descendant-ward part of the search.
static DebugLoc inlineDebugLoc(DebugLoc OrigDL, DILocation *InlinedAt, LLVMContext &Ctx, DenseMap< const MDNode *, MDNode * > &IANodes)
Returns a DebugLoc for a new DILocation which is a clone of OrigDL inlined at InlinedAt.
static cl::opt< bool > UseNoAliasIntrinsic("use-noalias-intrinsic-during-inlining", cl::Hidden, cl::init(true), cl::desc("Use the llvm.experimental.noalias.scope.decl " "intrinsic during inlining."))
static void PropagateCallSiteMetadata(CallBase &CB, Function::iterator FStart, Function::iterator FEnd)
When inlining a call site that has !llvm.mem.parallel_loop_access, !llvm.access.group,...
static std::pair< std::vector< int64_t >, std::vector< int64_t > > remapIndices(Function &Caller, BasicBlock *StartBB, PGOContextualProfile &CtxProf, uint32_t CalleeCounters, uint32_t CalleeCallsites)
static AttrBuilder IdentifyValidPoisonGeneratingAttributes(CallBase &CB)
static void updateMemprofMetadata(CallBase *CI, const std::vector< Metadata * > &MIBList, OptimizationRemarkEmitter *ORE)
static void updateCallProfile(Function *Callee, const ValueToValueMapTy &VMap, const ProfileCount &CalleeEntryCount, const CallBase &TheCall, ProfileSummaryInfo *PSI, BlockFrequencyInfo *CallerBFI)
Update the branch metadata for cloned call instructions.
static void updateCallerBFI(BasicBlock *CallSiteBlock, const ValueToValueMapTy &VMap, BlockFrequencyInfo *CallerBFI, BlockFrequencyInfo *CalleeBFI, const BasicBlock &CalleeEntryBlock)
Update the block frequencies of the caller after a callee has been inlined.
static void AddReturnAttributes(CallBase &CB, ValueToValueMapTy &VMap, ClonedCodeInfo &InlinedFunctionInfo)
static void HandleByValArgumentInit(Type *ByValType, Value *Dst, Value *Src, MaybeAlign SrcAlign, Module *M, BasicBlock *InsertBlock, InlineFunctionInfo &IFI, Function *CalledFunc)
static bool MayContainThrowingOrExitingCallAfterCB(CallBase *Begin, ReturnInst *End)
static cl::opt< bool > EnableNoAliasConversion("enable-noalias-to-md-conversion", cl::init(true), cl::Hidden, cl::desc("Convert noalias attributes to metadata during inlining."))
static void AddAliasScopeMetadata(CallBase &CB, ValueToValueMapTy &VMap, const DataLayout &DL, AAResults *CalleeAAR, ClonedCodeInfo &InlinedFunctionInfo)
If the inlined function has noalias arguments, then add new alias scopes for each noalias argument,...
static IntrinsicInst * getConvergenceEntry(BasicBlock &BB)
static void HandleInlinedEHPad(InvokeInst *II, BasicBlock *FirstNewBlock, ClonedCodeInfo &InlinedCodeInfo)
If we inlined an invoke site, we need to convert calls in the body of the inlined function into invok...
static void inlineRetainOrClaimRVCalls(CallBase &CB, objcarc::ARCInstKind RVCallKind, const SmallVectorImpl< ReturnInst * > &Returns)
An operand bundle "clang.arc.attachedcall" on a call indicates the call result is implicitly consumed...
static void fixupAssignments(Function::iterator Start, Function::iterator End)
Update inlined instructions' DIAssignID metadata.
static void propagateMemProfHelper(const CallBase *OrigCall, CallBase *ClonedCall, MDNode *InlinedCallsiteMD, OptimizationRemarkEmitter *ORE)
static bool allocaWouldBeStaticInEntry(const AllocaInst *AI)
Return the result of AI->isStaticAlloca() if AI were moved to the entry block.
static bool isUsedByLifetimeMarker(Value *V)
static void removeMemProfMetadata(CallBase *Call)
static Value * HandleByValArgument(Type *ByValType, Value *Arg, Instruction *TheCall, const Function *CalledFunc, InlineFunctionInfo &IFI, MaybeAlign ByValAlignment)
When inlining a call site that has a byval argument, we have to make the implicit memcpy explicit by ...
static void AddAlignmentAssumptions(CallBase &CB, InlineFunctionInfo &IFI)
If the inlined function has non-byval align arguments, then add @llvm.assume-based alignment assumpti...
static void trackInlinedStores(Function::iterator Start, Function::iterator End, const CallBase &CB)
static cl::opt< unsigned > InlinerAttributeWindow("max-inst-checked-for-throw-during-inlining", cl::Hidden, cl::desc("the maximum number of instructions analyzed for may throw during " "attribute inference in inlined body"), cl::init(4))
static void AddParamAndFnBasicAttributes(const CallBase &CB, ValueToValueMapTy &VMap, ClonedCodeInfo &InlinedFunctionInfo)
static bool haveCommonPrefix(MDNode *MIBStackContext, MDNode *CallsiteStackContext)
static void PropagateOperandBundles(Function::iterator InlinedBB, Instruction *CallSiteEHPad)
Bundle operands of the inlined function must be added to inlined call sites.
static bool hasLifetimeMarkers(AllocaInst *AI)
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static DebugLoc getDebugLoc(MachineBasicBlock::instr_iterator FirstMI, MachineBasicBlock::instr_iterator LastMI)
Return the first DebugLoc that has line number information, given a range of instructions.
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
uint64_t IntrinsicInst * II
This file defines common analysis utilities used by the ObjC ARC Optimizer.
This file defines ARC utility functions which are used by various parts of the compiler.
This file contains the declarations for profiling metadata utility functions.
This file implements a set that has insertion order iteration characteristics.
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
static Value * getParentPad(Value *EHPad)
LLVM_ABI MemoryEffects getMemoryEffects(const CallBase *Call)
Return the behavior of the given call site.
Class for arbitrary precision integers.
an instruction to allocate memory on the stack
bool isSwiftError() const
Return true if this alloca is used as a swifterror argument to a call.
PointerType * getType() const
Overload to return most specific pointer type.
bool isUsedWithInAlloca() const
Return true if this alloca is used as an inalloca argument to a call.
LLVM_ABI std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
const Value * getArraySize() const
Get the number of elements allocated.
This class represents an incoming formal argument to a Function.
unsigned getArgNo() const
Return the index of this formal argument in its containing function.
static LLVM_ABI uint64_t getGUID(const Function &F)
A cache of @llvm.assume calls within a function.
LLVM_ABI void registerAssumption(AssumeInst *CI)
Add an @llvm.assume intrinsic to this function's cache.
An instruction that atomically checks whether a specified value is in a memory location,...
an instruction that atomically reads a memory location, combines it with another value,...
static LLVM_ABI AttributeSet get(LLVMContext &C, const AttrBuilder &B)
Functions, function parameters, and return types can have attributes to indicate how they should be t...
LLVM_ABI const ConstantRange & getRange() const
Returns the value of the range attribute.
LLVM_ABI FPClassTest getNoFPClass() const
Return the FPClassTest for nofpclass.
AttrKind
This enumeration lists the attributes that can be associated with parameters, function results,...
bool isValid() const
Return true if the attribute is any kind of attribute.
LLVM Basic Block Representation.
iterator begin()
Instruction iterator methods.
iterator_range< const_phi_iterator > phis() const
Returns a range that iterates over the phis in the basic block.
LLVM_ABI BasicBlock * splitBasicBlock(iterator I, const Twine &BBName="")
Split the basic block into two basic blocks at the specified instruction.
const Function * getParent() const
Return the enclosing method, or null if none.
LLVM_ABI InstListType::const_iterator getFirstNonPHIIt() const
Returns an iterator to the first instruction in this block that is not a PHINode instruction.
LLVM_ABI SymbolTableList< BasicBlock >::iterator eraseFromParent()
Unlink 'this' from the containing function and delete it.
InstListType::iterator iterator
Instruction iterators...
const Instruction * getTerminator() const LLVM_READONLY
Returns the terminator instruction; assumes that the block is well-formed.
void splice(BasicBlock::iterator ToIt, BasicBlock *FromBB)
Transfer all instructions from FromBB to this basic block at ToIt.
LLVM_ABI void removePredecessor(BasicBlock *Pred, bool KeepOneInputPHIs=false)
Update PHI nodes in this BasicBlock before removal of predecessor Pred.
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
LLVM_ABI void setBlockFreq(const BasicBlock *BB, BlockFrequency Freq)
LLVM_ABI void setBlockFreqAndScale(const BasicBlock *ReferenceBB, BlockFrequency Freq, SmallPtrSetImpl< BasicBlock * > &BlocksToScale)
Set the frequency of ReferenceBB to Freq and scale the frequencies of the blocks in BlocksToScale suc...
LLVM_ABI BlockFrequency getBlockFreq(const BasicBlock *BB) const
getblockFreq - Return block frequency.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
void setCallingConv(CallingConv::ID CC)
LLVM_ABI FPClassTest getRetNoFPClass() const
Extract a test mask for disallowed floating-point value classes for the return value.
MaybeAlign getRetAlign() const
Extract the alignment of the return value.
LLVM_ABI void getOperandBundlesAsDefs(SmallVectorImpl< OperandBundleDef > &Defs) const
Return the list of operand bundles attached to this instruction as a vector of OperandBundleDefs.
OperandBundleUse getOperandBundleAt(unsigned Index) const
Return the operand bundle at a specific index.
std::optional< OperandBundleUse > getOperandBundle(StringRef Name) const
Return an operand bundle by name, if present.
Function * getCalledFunction() const
Returns the function called, or null if this is an indirect function invocation or the function signa...
void removeRetAttrs(const AttributeMask &AttrsToRemove)
Removes the attributes from the return value.
bool hasRetAttr(Attribute::AttrKind Kind) const
Determine whether the return value has the given attribute.
unsigned getNumOperandBundles() const
Return the number of operand bundles associated with this User.
CallingConv::ID getCallingConv() const
LLVM_ABI bool paramHasAttr(unsigned ArgNo, Attribute::AttrKind Kind) const
Determine whether the argument or parameter has the given attribute.
User::op_iterator arg_begin()
Return the iterator pointing to the beginning of the argument list.
Attribute getParamAttr(unsigned ArgNo, Attribute::AttrKind Kind) const
Get the attribute of a given kind from a given arg.
bool isByValArgument(unsigned ArgNo) const
Determine whether this argument is passed by value.
static LLVM_ABI CallBase * addOperandBundle(CallBase *CB, uint32_t ID, OperandBundleDef OB, InsertPosition InsertPt=nullptr)
Create a clone of CB with operand bundle OB added.
MaybeAlign getParamAlign(unsigned ArgNo) const
Extract the alignment for a call or parameter (0=unknown).
AttributeSet getRetAttributes() const
Return the return attributes for this call.
Type * getParamByValType(unsigned ArgNo) const
Extract the byval type for a call or parameter.
Value * getCalledOperand() const
void setAttributes(AttributeList A)
Set the attributes for this call.
LLVM_ABI std::optional< ConstantRange > getRange() const
If this return value has a range attribute, return the value range of the argument.
bool doesNotThrow() const
Determine if the call cannot unwind.
Value * getArgOperand(unsigned i) const
uint64_t getRetDereferenceableBytes() const
Extract the number of dereferenceable bytes for a call or parameter (0=unknown).
bool isConvergent() const
Determine if the invoke is convergent.
FunctionType * getFunctionType() const
static LLVM_ABI CallBase * Create(CallBase *CB, ArrayRef< OperandBundleDef > Bundles, InsertPosition InsertPt=nullptr)
Create a clone of CB with a different set of operand bundles and insert it before InsertPt.
uint64_t getRetDereferenceableOrNullBytes() const
Extract the number of dereferenceable_or_null bytes for a call (0=unknown).
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
unsigned arg_size() const
AttributeList getAttributes() const
Return the attributes for this call.
bool hasOperandBundles() const
Return true if this User has any operand bundles.
LLVM_ABI Function * getCaller()
Helper to get the caller (the parent function).
This class represents a function call, abstracting a target machine's calling convention.
void setTailCallKind(TailCallKind TCK)
TailCallKind getTailCallKind() const
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
bool isMustTailCall() const
static CatchSwitchInst * Create(Value *ParentPad, BasicBlock *UnwindDest, unsigned NumHandlers, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
static CleanupReturnInst * Create(Value *CleanupPad, BasicBlock *UnwindBB=nullptr, InsertPosition InsertBefore=nullptr)
This class represents a range of values.
LLVM_ABI ConstantRange intersectWith(const ConstantRange &CR, PreferredRangeType Type=Smallest) const
Return the range that results from the intersection of this range with another range.
static LLVM_ABI ConstantTokenNone * get(LLVMContext &Context)
Return the ConstantTokenNone.
This is an important base class in LLVM.
const Constant * stripPointerCasts() const
static LLVM_ABI InstrProfIncrementInst * getBBInstrumentation(BasicBlock &BB)
Get the instruction instrumenting a BB, or nullptr if not present.
static LLVM_ABI InstrProfCallsite * getCallsiteInstrumentation(CallBase &CB)
Get the instruction instrumenting a callsite, or nullptr if that cannot be found.
const DILocation * getWithoutAtom() const
uint64_t getAtomGroup() const
uint8_t getAtomRank() const
Subprogram description. Uses SubclassData1.
A parsed version of the target data layout string in and methods for querying it.
Base class for non-instruction debug metadata records that have positions within IR.
Record of a variable value-assignment, aka a non instruction representation of the dbg....
static DebugLoc getCompilerGenerated()
LLVM_ABI unsigned getLine() const
LLVM_ABI DILocation * get() const
Get the underlying DILocation.
LLVM_ABI MDNode * getScope() const
static LLVM_ABI DebugLoc appendInlinedAt(const DebugLoc &DL, DILocation *InlinedAt, LLVMContext &Ctx, DenseMap< const MDNode *, MDNode * > &Cache)
Rebuild the entire inlined-at chain for this instruction so that the top of the chain now is inlined-...
static DebugLoc getTemporary()
LLVM_ABI unsigned getCol() const
LLVM_ABI bool isImplicitCode() const
Check if the DebugLoc corresponds to an implicit code.
static DebugLoc getUnknown()
ValueT lookup(const_arg_type_t< KeyT > Val) const
lookup - Return the entry for the specified key, or a default constructed value if no such entry exis...
iterator find(const_arg_type_t< KeyT > Val)
size_type count(const_arg_type_t< KeyT > Val) const
Return 1 if the specified key is in the map, 0 otherwise.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Implements a dense probed hash-table based set.
void recalculate(ParentType &Func)
recalculate - compute a dominator tree for the given function
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
unsigned getNumParams() const
Return the number of fixed parameters this function type requires.
Class to represent profile counts.
uint64_t getCount() const
const BasicBlock & getEntryBlock() const
BasicBlockListType::iterator iterator
FunctionType * getFunctionType() const
Returns the FunctionType for me.
const BasicBlock & front() const
iterator_range< arg_iterator > args()
DISubprogram * getSubprogram() const
Get the attached subprogram.
bool hasGC() const
hasGC/getGC/setGC/clearGC - The name of the garbage collection algorithm to use during code generatio...
CallingConv::ID getCallingConv() const
getCallingConv()/setCallingConv(CC) - These method get and set the calling convention of this functio...
bool hasPersonalityFn() const
Check whether this function has a personality function.
Constant * getPersonalityFn() const
Get the personality function associated with this function.
bool isIntrinsic() const
isIntrinsic - Returns true if the function's name starts with "llvm.".
MaybeAlign getParamAlign(unsigned ArgNo) const
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
const std::string & getGC() const
std::optional< ProfileCount > getEntryCount(bool AllowSynthetic=false) const
Get the entry count for this function.
Type * getReturnType() const
Returns the type of the ret val.
void setCallingConv(CallingConv::ID CC)
bool onlyReadsMemory() const
Determine if the function does not access or only reads memory.
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
bool hasMetadata() const
Return true if this GlobalObject has any metadata attached to it.
MDNode * getMetadata(unsigned KindID) const
Get the metadata of given kind attached to this GlobalObject.
LLVM_ABI bool isDeclaration() const
Return true if the primary definition of this global value is outside of the current translation unit...
LLVM_ABI CallInst * CreateLifetimeStart(Value *Ptr)
Create a lifetime.start intrinsic.
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
This class captures the data input to the InlineFunction call, and records the auxiliary results prod...
Value * ConvergenceControlToken
bool UpdateProfile
Update profile for callee as well as cloned version.
Instruction * CallSiteEHPad
function_ref< AssumptionCache &(Function &)> GetAssumptionCache
If non-null, InlineFunction will update the callgraph to reflect the changes it makes.
BlockFrequencyInfo * CalleeBFI
SmallVector< AllocaInst *, 4 > StaticAllocas
InlineFunction fills this in with all static allocas that get copied into the caller.
BlockFrequencyInfo * CallerBFI
SmallVector< CallBase *, 8 > InlinedCallSites
All of the new call sites inlined into the caller.
InlineResult is basically true or false.
static InlineResult success()
static InlineResult failure(const char *Reason)
This represents the llvm.instrprof.callsite intrinsic.
This represents the llvm.instrprof.increment intrinsic.
const DebugLoc & getDebugLoc() const
Return the debug location for this node as a DebugLoc.
bool hasMetadata() const
Return true if this instruction has any metadata attached to it.
LLVM_ABI void insertBefore(InstListType::iterator InsertPos)
Insert an unlinked instruction into a basic block immediately before the specified position.
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
LLVM_ABI const Function * getFunction() const
Return the function this instruction belongs to.
MDNode * getMetadata(unsigned KindID) const
Get the metadata of given kind attached to this Instruction.
LLVM_ABI void setMetadata(unsigned KindID, MDNode *Node)
Set the metadata of the specified kind to the specified node.
void setDebugLoc(DebugLoc Loc)
Set the debug location information for this instruction.
LLVM_ABI const DataLayout & getDataLayout() const
Get the data layout of the module this instruction belongs to.
A wrapper class for inspecting calls to intrinsic functions.
static LLVM_ABI bool mayLowerToFunctionCall(Intrinsic::ID IID)
Check if the intrinsic might lower into a regular function call in the course of IR transformations.
This is an important class for using LLVM in a threaded context.
@ OB_clang_arc_attachedcall
The landingpad instruction holds all of the information necessary to generate correct exception handl...
bool isCleanup() const
Return 'true' if this landingpad instruction is a cleanup.
unsigned getNumClauses() const
Get the number of clauses for this landing pad.
Constant * getClause(unsigned Idx) const
Get the value of the clause at index Idx.
An instruction for reading from memory.
MDNode * createAnonymousAliasScope(MDNode *Domain, StringRef Name=StringRef())
Return metadata appropriate for an alias scope root node.
MDNode * createAnonymousAliasScopeDomain(StringRef Name=StringRef())
Return metadata appropriate for an alias scope domain node.
static MDTuple * getDistinct(LLVMContext &Context, ArrayRef< Metadata * > MDs)
void replaceAllUsesWith(Metadata *MD)
RAUW a temporary.
static LLVM_ABI MDNode * concatenate(MDNode *A, MDNode *B)
Methods for metadata merging.
ArrayRef< MDOperand > operands() const
op_iterator op_end() const
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
unsigned getNumOperands() const
Return number of MDNode operands.
op_iterator op_begin() const
LLVMContext & getContext() const
static LLVM_ABI MDString * get(LLVMContext &Context, StringRef Str)
static TempMDTuple getTemporary(LLVMContext &Context, ArrayRef< Metadata * > MDs)
Return a temporary node.
bool onlyAccessesInaccessibleMem() const
Whether this function only (at most) accesses inaccessible memory.
bool onlyAccessesArgPointees() const
Whether this function only (at most) accesses argument memory.
A Module instance is used to store all the information related to an LLVM module.
The instrumented contextual profile, produced by the CtxProfAnalysis.
LLVM_ABI bool isInSpecializedModule() const
LLVM_ABI void update(Visitor, const Function &F)
uint32_t getNumCounters(const Function &F) const
uint32_t allocateNextCounterIndex(const Function &F)
uint32_t getNumCallsites(const Function &F) const
uint32_t allocateNextCallsiteIndex(const Function &F)
A node (context) in the loaded contextual profile, suitable for mutation during IPO passes.
void addIncoming(Value *V, BasicBlock *BB)
Add an incoming value to the end of the PHI list.
static PHINode * Create(Type *Ty, unsigned NumReservedValues, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
Constructors - NumReservedValues is a hint for the number of incoming edges that this phi node will h...
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
static LLVM_ABI PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
Analysis providing profile information.
LLVM_ABI std::optional< uint64_t > getProfileCount(const CallBase &CallInst, BlockFrequencyInfo *BFI, bool AllowSynthetic=false) const
Returns the profile count for CallInst.
Resume the propagation of an exception.
Return a value (possibly void), from a function.
bool remove(const value_type &X)
Remove an item from the set vector.
bool contains(const_arg_type key) const
Check if the SetVector contains the given key.
bool insert(const value_type &X)
Insert a new element into the SetVector.
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
void insert_range(Range &&R)
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
bool contains(ConstPtrType Ptr) const
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
A SetVector that performs no allocations if smaller than a certain size.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void reserve(size_type N)
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
An instruction for storing to memory.
StringRef - Represent a constant reference to a string, i.e.
The instances of the Type class are immutable: once they are created, they are never changed.
static LLVM_ABI IntegerType * getInt64Ty(LLVMContext &C)
LLVM_ABI unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
bool isVoidTy() const
Return true if this is 'void'.
Unconditional Branch instruction.
void setSuccessor(BasicBlock *NewSucc)
static UncondBrInst * Create(BasicBlock *Target, InsertPosition InsertBefore=nullptr)
BasicBlock * getSuccessor(unsigned i=0) const
Value * getOperand(unsigned i) const
This class represents the va_arg llvm instruction, which returns an argument of the specified type gi...
ValueT lookup(const KeyT &Val) const
lookup - Return the entry for the specified key, or a default constructed value if no such entry exis...
size_type count(const KeyT &Val) const
Return 1 if the specified key is in the map, 0 otherwise.
ValueMapIteratorImpl< MapT, const Value *, false > iterator
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
LLVM_ABI void replaceAllUsesWith(Value *V)
Change all uses of this to point to a new Value.
LLVMContext & getContext() const
All values hold a context through their type.
iterator_range< user_iterator > users()
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
std::pair< iterator, bool > insert(const ValueT &V)
const ParentTy * getParent() const
self_iterator getIterator()
Class to build a trie of call stack contexts for a particular profiled allocation call,...
Helper class to iterate through stack ids in both metadata (memprof MIB and callsite) and the corresp...
This provides a very simple, boring adaptor for a begin and end iterator into a range type.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
CallingConv Namespace - This namespace contains an enum with a value for the well-known calling conve...
@ BasicBlock
Various leaf nodes.
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
bool match(Val *V, const Pattern &P)
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
LLVM_ABI void trackAssignments(Function::iterator Start, Function::iterator End, const StorageToVarsMap &Vars, const DataLayout &DL, bool DebugPrints=false)
Track assignments to Vars between Start and End.
LLVM_ABI void remapAssignID(DenseMap< DIAssignID *, DIAssignID * > &Map, Instruction &I)
Replace DIAssignID uses and attachments with IDs from Map.
SmallVector< DbgVariableRecord * > getDVRAssignmentMarkers(const Instruction *Inst)
Return a range of dbg_assign records for which Inst performs the assignment they encode.
DenseMap< const AllocaInst *, SmallSetVector< VarRecord, 2 > > StorageToVarsMap
Map of backing storage to a set of variables that are stored to it.
initializer< Ty > init(const Ty &Val)
std::enable_if_t< detail::IsValidPointer< X, Y >::value, X * > dyn_extract(Y &&MD)
Extract a Value from Metadata, if any.
LLVM_ABI MDNode * getMIBStackNode(const MDNode *MIB)
Returns the stack node from an MIB metadata node.
ARCInstKind getAttachedARCFunctionKind(const CallBase *CB)
This function returns the ARCInstKind of the function attached to operand bundle clang_arc_attachedca...
ARCInstKind
Equivalence classes of instructions in the ARC Model.
@ None
anything that is inert from an ARC perspective.
@ RetainRV
objc_retainAutoreleasedReturnValue
std::optional< Function * > getAttachedARCFunction(const CallBase *CB)
This function returns operand bundle clang_arc_attachedcall's argument, which is the address of the A...
bool isRetainOrClaimRV(ARCInstKind Kind)
Check whether the function is retainRV/unsafeClaimRV.
const Value * GetRCIdentityRoot(const Value *V)
The RCIdentity root of a value V is a dominating value U for which retaining or releasing U is equiva...
bool hasAttachedCallOpBundle(const CallBase *CB)
This is an optimization pass for GlobalISel generic memory operations.
UnaryFunction for_each(R &&Range, UnaryFunction F)
Provide wrappers to std::for_each which take ranges instead of having to pass begin/end explicitly.
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
LLVM_ABI BasicBlock * changeToInvokeAndSplitBasicBlock(CallInst *CI, BasicBlock *UnwindEdge, DomTreeUpdater *DTU=nullptr)
Convert the CallInst to InvokeInst with the specified unwind edge basic block.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
auto successors(const MachineBasicBlock *BB)
LLVM_ABI void CloneAndPruneFunctionInto(Function *NewFunc, const Function *OldFunc, ValueToValueMapTy &VMap, bool ModuleLevelChanges, SmallVectorImpl< ReturnInst * > &Returns, const char *NameSuffix, ClonedCodeInfo &CodeInfo)
This works exactly like CloneFunctionInto, except that it does some simple constant prop and DCE on t...
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
std::string utostr(uint64_t X, bool isNeg=false)
MemoryEffectsBase< IRMemLocation > MemoryEffects
Summary of how a function affects memory in the program.
bool isa_and_nonnull(const Y &Val)
LLVM_ABI InlineResult InlineFunction(CallBase &CB, InlineFunctionInfo &IFI, bool MergeAttributes=false, AAResults *CalleeAAR=nullptr, bool InsertLifetime=true, bool TrackInlineHistory=false, Function *ForwardVarArgsTo=nullptr, OptimizationRemarkEmitter *ORE=nullptr)
This function inlines the called function into the basic block of the caller.
LLVM_ABI bool PointerMayBeCapturedBefore(const Value *V, bool ReturnCaptures, const Instruction *I, const DominatorTree *DT, bool IncludeI=false, unsigned MaxUsesToExplore=0, const LoopInfo *LI=nullptr)
PointerMayBeCapturedBefore - Return true if this pointer value may be captured by the enclosing funct...
LLVM_ABI InlineResult CanInlineCallSite(const CallBase &CB, InlineFunctionInfo &IFI)
Check if it is legal to perform inlining of the function called by CB into the caller at this particu...
bool isScopedEHPersonality(EHPersonality Pers)
Returns true if this personality uses scope-style EH IR instructions: catchswitch,...
LLVM_ABI Value * simplifyInstruction(Instruction *I, const SimplifyQuery &Q)
See if we can compute a simplified version of this instruction.
auto dyn_cast_or_null(const Y &Val)
Align getKnownAlignment(Value *V, const DataLayout &DL, const Instruction *CxtI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr)
Try to infer an alignment for the specified pointer.
LLVM_ABI Align getOrEnforceKnownAlignment(Value *V, MaybeAlign PrefAlign, const DataLayout &DL, const Instruction *CxtI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr)
Try to ensure that the alignment of V is at least PrefAlign bytes.
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
Function::ProfileCount ProfileCount
LLVM_ABI EHPersonality classifyEHPersonality(const Value *Pers)
See if the given exception handling personality function is one that we understand.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
LLVM_ABI unsigned changeToUnreachable(Instruction *I, bool PreserveLCSSA=false, DomTreeUpdater *DTU=nullptr, MemorySSAUpdater *MSSAU=nullptr)
Insert an unreachable instruction before the specified instruction, making it and the rest of the cod...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
IRBuilder(LLVMContext &, FolderTy, InserterTy, MDNode *, ArrayRef< OperandBundleDef >) -> IRBuilder< FolderTy, InserterTy >
LLVM_ABI bool salvageKnowledge(Instruction *I, AssumptionCache *AC=nullptr, DominatorTree *DT=nullptr)
Calls BuildAssumeFromInst and if the resulting llvm.assume is valid insert if before I.
LLVM_ABI void updateProfileCallee(Function *Callee, int64_t EntryDelta, const ValueMap< const Value *, WeakTrackingVH > *VMap=nullptr)
Updates profile information by adjusting the entry count by adding EntryDelta then scaling callsite i...
OperandBundleDefT< Value * > OperandBundleDef
LLVM_ABI bool isAssignmentTrackingEnabled(const Module &M)
Return true if assignment tracking is enabled for module M.
LLVM_ABI void InlineFunctionImpl(CallBase &CB, InlineFunctionInfo &IFI, bool MergeAttributes=false, AAResults *CalleeAAR=nullptr, bool InsertLifetime=true, bool TrackInlineHistory=false, Function *ForwardVarArgsTo=nullptr, OptimizationRemarkEmitter *ORE=nullptr)
This should generally not be used, use InlineFunction instead.
LLVM_ABI MDNode * uniteAccessGroups(MDNode *AccGroups1, MDNode *AccGroups2)
Compute the union of two access-group lists.
DWARFExpression::Operation Op
bool isAsynchronousEHPersonality(EHPersonality Pers)
Returns true if this personality function catches asynchronous exceptions.
ValueMap< const Value *, WeakTrackingVH > ValueToValueMapTy
LLVM_ABI bool isGuaranteedToTransferExecutionToSuccessor(const Instruction *I)
Return true if this function can prove that the instruction I will always transfer execution to one o...
LLVM_ABI bool isEscapeSource(const Value *V)
Returns true if the pointer is one which would have been considered an escape by isNotCapturedBefore.
auto count_if(R &&Range, UnaryPredicate P)
Wrapper function around std::count_if to count the number of times an element satisfying a given pred...
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
bool capturesAnything(CaptureComponents CC)
bool pred_empty(const BasicBlock *BB)
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
LLVM_ABI void getUnderlyingObjects(const Value *V, SmallVectorImpl< const Value * > &Objects, const LoopInfo *LI=nullptr, unsigned MaxLookup=MaxLookupSearchDepth)
This method is similar to getUnderlyingObject except that it can look through phi and select instruct...
LLVM_ABI void updateLoopMetadataDebugLocations(Instruction &I, function_ref< Metadata *(Metadata *)> Updater)
Update the debug locations contained within the MD_loop metadata attached to the instruction I,...
LLVM_ABI bool isIdentifiedObject(const Value *V)
Return true if this pointer refers to a distinct and identifiable object.
LLVM_ABI void scaleProfData(Instruction &I, uint64_t S, uint64_t T)
Scaling the profile data attached to 'I' using the ratio of S/T.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
This struct can be used to capture information about code being cloned, while it is being cloned.
bool ContainsDynamicAllocas
This is set to true if the cloned code contains a 'dynamic' alloca.
bool isSimplified(const Value *From, const Value *To) const
bool ContainsCalls
This is set to true if the cloned code contains a normal call instruction.
bool ContainsMemProfMetadata
This is set to true if there is memprof related metadata (memprof or callsite metadata) in the cloned...
SmallSetVector< const Value *, 4 > OriginallyIndirectCalls
std::vector< WeakTrackingVH > OperandBundleCallSites
All cloned call sites that have operand bundles attached are appended to this vector.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Align valueOrOne() const
For convenience, returns a valid alignment or 1 if undefined.
static Instruction * tryGetVTableInstruction(CallBase *CB)
Helper struct for trackAssignments, below.