LLVM 23.0.0git
AttributorAttributes.cpp
Go to the documentation of this file.
1//===- AttributorAttributes.cpp - Attributes for Attributor deduction -----===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// See the Attributor.h file comment and the class descriptions in that file for
10// more information.
11//
12//===----------------------------------------------------------------------===//
13
15
16#include "llvm/ADT/APInt.h"
17#include "llvm/ADT/ArrayRef.h"
19#include "llvm/ADT/MapVector.h"
21#include "llvm/ADT/STLExtras.h"
23#include "llvm/ADT/SetVector.h"
26#include "llvm/ADT/Statistic.h"
39#include "llvm/IR/Argument.h"
40#include "llvm/IR/Assumptions.h"
41#include "llvm/IR/Attributes.h"
42#include "llvm/IR/BasicBlock.h"
43#include "llvm/IR/Constant.h"
44#include "llvm/IR/Constants.h"
45#include "llvm/IR/DataLayout.h"
47#include "llvm/IR/GlobalValue.h"
48#include "llvm/IR/IRBuilder.h"
49#include "llvm/IR/InlineAsm.h"
50#include "llvm/IR/InstrTypes.h"
51#include "llvm/IR/Instruction.h"
54#include "llvm/IR/IntrinsicsAMDGPU.h"
55#include "llvm/IR/IntrinsicsNVPTX.h"
56#include "llvm/IR/LLVMContext.h"
57#include "llvm/IR/MDBuilder.h"
58#include "llvm/IR/NoFolder.h"
59#include "llvm/IR/Value.h"
60#include "llvm/IR/ValueHandle.h"
75#include <cassert>
76#include <numeric>
77#include <optional>
78#include <string>
79
80using namespace llvm;
81
82#define DEBUG_TYPE "attributor"
83
85 "attributor-manifest-internal", cl::Hidden,
86 cl::desc("Manifest Attributor internal string attributes."),
87 cl::init(false));
88
89static cl::opt<int> MaxHeapToStackSize("max-heap-to-stack-size", cl::init(128),
91
92template <>
94
96
98 "attributor-max-potential-values", cl::Hidden,
99 cl::desc("Maximum number of potential values to be "
100 "tracked for each position."),
102 cl::init(7));
103
105 "attributor-max-potential-values-iterations", cl::Hidden,
106 cl::desc(
107 "Maximum number of iterations we keep dismantling potential values."),
108 cl::init(64));
109
110STATISTIC(NumAAs, "Number of abstract attributes created");
111STATISTIC(NumIndirectCallsPromoted, "Number of indirect calls promoted");
112
113// Some helper macros to deal with statistics tracking.
114//
115// Usage:
116// For simple IR attribute tracking overload trackStatistics in the abstract
117// attribute and choose the right STATS_DECLTRACK_********* macro,
118// e.g.,:
119// void trackStatistics() const override {
120// STATS_DECLTRACK_ARG_ATTR(returned)
121// }
122// If there is a single "increment" side one can use the macro
123// STATS_DECLTRACK with a custom message. If there are multiple increment
124// sides, STATS_DECL and STATS_TRACK can also be used separately.
125//
126#define BUILD_STAT_MSG_IR_ATTR(TYPE, NAME) \
127 ("Number of " #TYPE " marked '" #NAME "'")
128#define BUILD_STAT_NAME(NAME, TYPE) NumIR##TYPE##_##NAME
129#define STATS_DECL_(NAME, MSG) STATISTIC(NAME, MSG);
130#define STATS_DECL(NAME, TYPE, MSG) \
131 STATS_DECL_(BUILD_STAT_NAME(NAME, TYPE), MSG);
132#define STATS_TRACK(NAME, TYPE) ++(BUILD_STAT_NAME(NAME, TYPE));
133#define STATS_DECLTRACK(NAME, TYPE, MSG) \
134 {STATS_DECL(NAME, TYPE, MSG) STATS_TRACK(NAME, TYPE)}
135#define STATS_DECLTRACK_ARG_ATTR(NAME) \
136 STATS_DECLTRACK(NAME, Arguments, BUILD_STAT_MSG_IR_ATTR(arguments, NAME))
137#define STATS_DECLTRACK_CSARG_ATTR(NAME) \
138 STATS_DECLTRACK(NAME, CSArguments, \
139 BUILD_STAT_MSG_IR_ATTR(call site arguments, NAME))
140#define STATS_DECLTRACK_FN_ATTR(NAME) \
141 STATS_DECLTRACK(NAME, Function, BUILD_STAT_MSG_IR_ATTR(functions, NAME))
142#define STATS_DECLTRACK_CS_ATTR(NAME) \
143 STATS_DECLTRACK(NAME, CS, BUILD_STAT_MSG_IR_ATTR(call site, NAME))
144#define STATS_DECLTRACK_FNRET_ATTR(NAME) \
145 STATS_DECLTRACK(NAME, FunctionReturn, \
146 BUILD_STAT_MSG_IR_ATTR(function returns, NAME))
147#define STATS_DECLTRACK_CSRET_ATTR(NAME) \
148 STATS_DECLTRACK(NAME, CSReturn, \
149 BUILD_STAT_MSG_IR_ATTR(call site returns, NAME))
150#define STATS_DECLTRACK_FLOATING_ATTR(NAME) \
151 STATS_DECLTRACK(NAME, Floating, \
152 ("Number of floating values known to be '" #NAME "'"))
153
154// Specialization of the operator<< for abstract attributes subclasses. This
155// disambiguates situations where multiple operators are applicable.
156namespace llvm {
157#define PIPE_OPERATOR(CLASS) \
158 raw_ostream &operator<<(raw_ostream &OS, const CLASS &AA) { \
159 return OS << static_cast<const AbstractAttribute &>(AA); \
160 }
161
201
202#undef PIPE_OPERATOR
203
204template <>
206 const DerefState &R) {
207 ChangeStatus CS0 =
208 clampStateAndIndicateChange(S.DerefBytesState, R.DerefBytesState);
209 ChangeStatus CS1 = clampStateAndIndicateChange(S.GlobalState, R.GlobalState);
210 return CS0 | CS1;
211}
212
213} // namespace llvm
214
215static bool mayBeInCycle(const CycleInfo *CI, const Instruction *I,
216 bool HeaderOnly, Cycle **CPtr = nullptr) {
217 if (!CI)
218 return true;
219 auto *BB = I->getParent();
220 auto *C = CI->getCycle(BB);
221 if (!C)
222 return false;
223 if (CPtr)
224 *CPtr = C;
225 return !HeaderOnly || BB == C->getHeader();
226}
227
228/// Checks if a type could have padding bytes.
229static bool isDenselyPacked(Type *Ty, const DataLayout &DL) {
230 // There is no size information, so be conservative.
231 if (!Ty->isSized())
232 return false;
233
234 // If the alloc size is not equal to the storage size, then there are padding
235 // bytes. For x86_fp80 on x86-64, size: 80 alloc size: 128.
236 if (DL.getTypeSizeInBits(Ty) != DL.getTypeAllocSizeInBits(Ty))
237 return false;
238
239 // FIXME: This isn't the right way to check for padding in vectors with
240 // non-byte-size elements.
241 if (VectorType *SeqTy = dyn_cast<VectorType>(Ty))
242 return isDenselyPacked(SeqTy->getElementType(), DL);
243
244 // For array types, check for padding within members.
245 if (ArrayType *SeqTy = dyn_cast<ArrayType>(Ty))
246 return isDenselyPacked(SeqTy->getElementType(), DL);
247
248 if (!isa<StructType>(Ty))
249 return true;
250
251 // Check for padding within and between elements of a struct.
252 StructType *StructTy = cast<StructType>(Ty);
253 const StructLayout *Layout = DL.getStructLayout(StructTy);
254 uint64_t StartPos = 0;
255 for (unsigned I = 0, E = StructTy->getNumElements(); I < E; ++I) {
256 Type *ElTy = StructTy->getElementType(I);
257 if (!isDenselyPacked(ElTy, DL))
258 return false;
259 if (StartPos != Layout->getElementOffsetInBits(I))
260 return false;
261 StartPos += DL.getTypeAllocSizeInBits(ElTy);
262 }
263
264 return true;
265}
266
267/// Get pointer operand of memory accessing instruction. If \p I is
268/// not a memory accessing instruction, return nullptr. If \p AllowVolatile,
269/// is set to false and the instruction is volatile, return nullptr.
271 bool AllowVolatile) {
272 if (!AllowVolatile && I->isVolatile())
273 return nullptr;
274
275 if (auto *LI = dyn_cast<LoadInst>(I)) {
276 return LI->getPointerOperand();
277 }
278
279 if (auto *SI = dyn_cast<StoreInst>(I)) {
280 return SI->getPointerOperand();
281 }
282
283 if (auto *CXI = dyn_cast<AtomicCmpXchgInst>(I)) {
284 return CXI->getPointerOperand();
285 }
286
287 if (auto *RMWI = dyn_cast<AtomicRMWInst>(I)) {
288 return RMWI->getPointerOperand();
289 }
290
291 return nullptr;
292}
293
294/// Helper function to create a pointer based on \p Ptr, and advanced by \p
295/// Offset bytes.
296static Value *constructPointer(Value *Ptr, int64_t Offset,
297 IRBuilder<NoFolder> &IRB) {
298 LLVM_DEBUG(dbgs() << "Construct pointer: " << *Ptr << " + " << Offset
299 << "-bytes\n");
300
301 if (Offset)
302 Ptr = IRB.CreatePtrAdd(Ptr, IRB.getInt64(Offset),
303 Ptr->getName() + ".b" + Twine(Offset));
304 return Ptr;
305}
306
307static const Value *
309 const Value *Val, const DataLayout &DL, APInt &Offset,
310 bool GetMinOffset, bool AllowNonInbounds,
311 bool UseAssumed = false) {
312
313 auto AttributorAnalysis = [&](Value &V, APInt &ROffset) -> bool {
314 const IRPosition &Pos = IRPosition::value(V);
315 // Only track dependence if we are going to use the assumed info.
316 const AAValueConstantRange *ValueConstantRangeAA =
317 A.getAAFor<AAValueConstantRange>(QueryingAA, Pos,
318 UseAssumed ? DepClassTy::OPTIONAL
320 if (!ValueConstantRangeAA)
321 return false;
322 ConstantRange Range = UseAssumed ? ValueConstantRangeAA->getAssumed()
323 : ValueConstantRangeAA->getKnown();
324 if (Range.isFullSet())
325 return false;
326
327 // We can only use the lower part of the range because the upper part can
328 // be higher than what the value can really be.
329 if (GetMinOffset)
330 ROffset = Range.getSignedMin();
331 else
332 ROffset = Range.getSignedMax();
333 return true;
334 };
335
336 return Val->stripAndAccumulateConstantOffsets(DL, Offset, AllowNonInbounds,
337 /* AllowInvariant */ true,
338 AttributorAnalysis);
339}
340
341static const Value *
343 const Value *Ptr, int64_t &BytesOffset,
344 const DataLayout &DL, bool AllowNonInbounds = false) {
345 APInt OffsetAPInt(DL.getIndexTypeSizeInBits(Ptr->getType()), 0);
346 const Value *Base =
347 stripAndAccumulateOffsets(A, QueryingAA, Ptr, DL, OffsetAPInt,
348 /* GetMinOffset */ true, AllowNonInbounds);
349
350 BytesOffset = OffsetAPInt.getSExtValue();
351 return Base;
352}
353
354/// Clamp the information known for all returned values of a function
355/// (identified by \p QueryingAA) into \p S.
356template <typename AAType, typename StateType = typename AAType::StateType,
357 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind,
358 bool RecurseForSelectAndPHI = true>
360 Attributor &A, const AAType &QueryingAA, StateType &S,
361 const IRPosition::CallBaseContext *CBContext = nullptr) {
362 LLVM_DEBUG(dbgs() << "[Attributor] Clamp return value states for "
363 << QueryingAA << " into " << S << "\n");
364
365 assert((QueryingAA.getIRPosition().getPositionKind() ==
367 QueryingAA.getIRPosition().getPositionKind() ==
369 "Can only clamp returned value states for a function returned or call "
370 "site returned position!");
371
372 // Use an optional state as there might not be any return values and we want
373 // to join (IntegerState::operator&) the state of all there are.
374 std::optional<StateType> T;
375
376 // Callback for each possibly returned value.
377 auto CheckReturnValue = [&](Value &RV) -> bool {
378 const IRPosition &RVPos = IRPosition::value(RV, CBContext);
379 // If possible, use the hasAssumedIRAttr interface.
380 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
381 bool IsKnown;
383 A, &QueryingAA, RVPos, DepClassTy::REQUIRED, IsKnown);
384 }
385
386 const AAType *AA =
387 A.getAAFor<AAType>(QueryingAA, RVPos, DepClassTy::REQUIRED);
388 if (!AA)
389 return false;
390 LLVM_DEBUG(dbgs() << "[Attributor] RV: " << RV
391 << " AA: " << AA->getAsStr(&A) << " @ " << RVPos << "\n");
392 const StateType &AAS = AA->getState();
393 if (!T)
394 T = StateType::getBestState(AAS);
395 *T &= AAS;
396 LLVM_DEBUG(dbgs() << "[Attributor] AA State: " << AAS << " RV State: " << T
397 << "\n");
398 return T->isValidState();
399 };
400
401 if (!A.checkForAllReturnedValues(CheckReturnValue, QueryingAA,
403 RecurseForSelectAndPHI))
404 S.indicatePessimisticFixpoint();
405 else if (T)
406 S ^= *T;
407}
408
409namespace {
410/// Helper class for generic deduction: return value -> returned position.
411template <typename AAType, typename BaseType,
412 typename StateType = typename BaseType::StateType,
413 bool PropagateCallBaseContext = false,
414 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind,
415 bool RecurseForSelectAndPHI = true>
416struct AAReturnedFromReturnedValues : public BaseType {
417 AAReturnedFromReturnedValues(const IRPosition &IRP, Attributor &A)
418 : BaseType(IRP, A) {}
419
420 /// See AbstractAttribute::updateImpl(...).
421 ChangeStatus updateImpl(Attributor &A) override {
422 StateType S(StateType::getBestState(this->getState()));
423 clampReturnedValueStates<AAType, StateType, IRAttributeKind,
424 RecurseForSelectAndPHI>(
425 A, *this, S,
426 PropagateCallBaseContext ? this->getCallBaseContext() : nullptr);
427 // TODO: If we know we visited all returned values, thus no are assumed
428 // dead, we can take the known information from the state T.
429 return clampStateAndIndicateChange<StateType>(this->getState(), S);
430 }
431};
432
433/// Clamp the information known at all call sites for a given argument
434/// (identified by \p QueryingAA) into \p S.
435template <typename AAType, typename StateType = typename AAType::StateType,
436 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
437static void clampCallSiteArgumentStates(Attributor &A, const AAType &QueryingAA,
438 StateType &S) {
439 LLVM_DEBUG(dbgs() << "[Attributor] Clamp call site argument states for "
440 << QueryingAA << " into " << S << "\n");
441
442 assert(QueryingAA.getIRPosition().getPositionKind() ==
444 "Can only clamp call site argument states for an argument position!");
445
446 // Use an optional state as there might not be any return values and we want
447 // to join (IntegerState::operator&) the state of all there are.
448 std::optional<StateType> T;
449
450 // The argument number which is also the call site argument number.
451 unsigned ArgNo = QueryingAA.getIRPosition().getCallSiteArgNo();
452
453 auto CallSiteCheck = [&](AbstractCallSite ACS) {
454 const IRPosition &ACSArgPos = IRPosition::callsite_argument(ACS, ArgNo);
455 // Check if a coresponding argument was found or if it is on not associated
456 // (which can happen for callback calls).
457 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
458 return false;
459
460 // If possible, use the hasAssumedIRAttr interface.
461 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
462 bool IsKnown;
464 A, &QueryingAA, ACSArgPos, DepClassTy::REQUIRED, IsKnown);
465 }
466
467 const AAType *AA =
468 A.getAAFor<AAType>(QueryingAA, ACSArgPos, DepClassTy::REQUIRED);
469 if (!AA)
470 return false;
471 LLVM_DEBUG(dbgs() << "[Attributor] ACS: " << *ACS.getInstruction()
472 << " AA: " << AA->getAsStr(&A) << " @" << ACSArgPos
473 << "\n");
474 const StateType &AAS = AA->getState();
475 if (!T)
476 T = StateType::getBestState(AAS);
477 *T &= AAS;
478 LLVM_DEBUG(dbgs() << "[Attributor] AA State: " << AAS << " CSA State: " << T
479 << "\n");
480 return T->isValidState();
481 };
482
483 bool UsedAssumedInformation = false;
484 if (!A.checkForAllCallSites(CallSiteCheck, QueryingAA, true,
485 UsedAssumedInformation))
486 S.indicatePessimisticFixpoint();
487 else if (T)
488 S ^= *T;
489}
490
491/// This function is the bridge between argument position and the call base
492/// context.
493template <typename AAType, typename BaseType,
494 typename StateType = typename AAType::StateType,
495 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
496bool getArgumentStateFromCallBaseContext(Attributor &A,
497 BaseType &QueryingAttribute,
498 IRPosition &Pos, StateType &State) {
500 "Expected an 'argument' position !");
501 const CallBase *CBContext = Pos.getCallBaseContext();
502 if (!CBContext)
503 return false;
504
505 int ArgNo = Pos.getCallSiteArgNo();
506 assert(ArgNo >= 0 && "Invalid Arg No!");
507 const IRPosition CBArgPos = IRPosition::callsite_argument(*CBContext, ArgNo);
508
509 // If possible, use the hasAssumedIRAttr interface.
510 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
511 bool IsKnown;
513 A, &QueryingAttribute, CBArgPos, DepClassTy::REQUIRED, IsKnown);
514 }
515
516 const auto *AA =
517 A.getAAFor<AAType>(QueryingAttribute, CBArgPos, DepClassTy::REQUIRED);
518 if (!AA)
519 return false;
520 const StateType &CBArgumentState =
521 static_cast<const StateType &>(AA->getState());
522
523 LLVM_DEBUG(dbgs() << "[Attributor] Briding Call site context to argument"
524 << "Position:" << Pos << "CB Arg state:" << CBArgumentState
525 << "\n");
526
527 // NOTE: If we want to do call site grouping it should happen here.
528 State ^= CBArgumentState;
529 return true;
530}
531
532/// Helper class for generic deduction: call site argument -> argument position.
533template <typename AAType, typename BaseType,
534 typename StateType = typename AAType::StateType,
535 bool BridgeCallBaseContext = false,
536 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
537struct AAArgumentFromCallSiteArguments : public BaseType {
538 AAArgumentFromCallSiteArguments(const IRPosition &IRP, Attributor &A)
539 : BaseType(IRP, A) {}
540
541 /// See AbstractAttribute::updateImpl(...).
542 ChangeStatus updateImpl(Attributor &A) override {
543 StateType S = StateType::getBestState(this->getState());
544
545 if (BridgeCallBaseContext) {
546 bool Success =
547 getArgumentStateFromCallBaseContext<AAType, BaseType, StateType,
548 IRAttributeKind>(
549 A, *this, this->getIRPosition(), S);
550 if (Success)
551 return clampStateAndIndicateChange<StateType>(this->getState(), S);
552 }
553 clampCallSiteArgumentStates<AAType, StateType, IRAttributeKind>(A, *this,
554 S);
555
556 // TODO: If we know we visited all incoming values, thus no are assumed
557 // dead, we can take the known information from the state T.
558 return clampStateAndIndicateChange<StateType>(this->getState(), S);
559 }
560};
561
562/// Helper class for generic replication: function returned -> cs returned.
563template <typename AAType, typename BaseType,
564 typename StateType = typename BaseType::StateType,
565 bool IntroduceCallBaseContext = false,
566 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
567struct AACalleeToCallSite : public BaseType {
568 AACalleeToCallSite(const IRPosition &IRP, Attributor &A) : BaseType(IRP, A) {}
569
570 /// See AbstractAttribute::updateImpl(...).
571 ChangeStatus updateImpl(Attributor &A) override {
572 auto IRPKind = this->getIRPosition().getPositionKind();
574 IRPKind == IRPosition::IRP_CALL_SITE) &&
575 "Can only wrap function returned positions for call site "
576 "returned positions!");
577 auto &S = this->getState();
578
579 CallBase &CB = cast<CallBase>(this->getAnchorValue());
580 if (IntroduceCallBaseContext)
581 LLVM_DEBUG(dbgs() << "[Attributor] Introducing call base context:" << CB
582 << "\n");
583
584 ChangeStatus Changed = ChangeStatus::UNCHANGED;
585 auto CalleePred = [&](ArrayRef<const Function *> Callees) {
586 for (const Function *Callee : Callees) {
587 IRPosition FnPos =
589 ? IRPosition::returned(*Callee,
590 IntroduceCallBaseContext ? &CB : nullptr)
591 : IRPosition::function(
592 *Callee, IntroduceCallBaseContext ? &CB : nullptr);
593 // If possible, use the hasAssumedIRAttr interface.
594 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
595 bool IsKnown;
597 A, this, FnPos, DepClassTy::REQUIRED, IsKnown))
598 return false;
599 continue;
600 }
601
602 const AAType *AA =
603 A.getAAFor<AAType>(*this, FnPos, DepClassTy::REQUIRED);
604 if (!AA)
605 return false;
606 Changed |= clampStateAndIndicateChange(S, AA->getState());
607 if (S.isAtFixpoint())
608 return S.isValidState();
609 }
610 return true;
611 };
612 if (!A.checkForAllCallees(CalleePred, *this, CB))
613 return S.indicatePessimisticFixpoint();
614 return Changed;
615 }
616};
617
618/// Helper function to accumulate uses.
619template <class AAType, typename StateType = typename AAType::StateType>
620static void followUsesInContext(AAType &AA, Attributor &A,
622 const Instruction *CtxI,
624 StateType &State) {
625 auto EIt = Explorer.begin(CtxI), EEnd = Explorer.end(CtxI);
626 for (unsigned u = 0; u < Uses.size(); ++u) {
627 const Use *U = Uses[u];
628 if (const Instruction *UserI = dyn_cast<Instruction>(U->getUser())) {
629 bool Found = Explorer.findInContextOf(UserI, EIt, EEnd);
630 if (Found && AA.followUseInMBEC(A, U, UserI, State))
631 Uses.insert_range(llvm::make_pointer_range(UserI->uses()));
632 }
633 }
634}
635
636/// Use the must-be-executed-context around \p I to add information into \p S.
637/// The AAType class is required to have `followUseInMBEC` method with the
638/// following signature and behaviour:
639///
640/// bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I)
641/// U - Underlying use.
642/// I - The user of the \p U.
643/// Returns true if the value should be tracked transitively.
644///
645template <class AAType, typename StateType = typename AAType::StateType>
646static void followUsesInMBEC(AAType &AA, Attributor &A, StateType &S,
647 Instruction &CtxI) {
648 const Value &Val = AA.getIRPosition().getAssociatedValue();
649 if (isa<ConstantData>(Val))
650 return;
651
653 A.getInfoCache().getMustBeExecutedContextExplorer();
654 if (!Explorer)
655 return;
656
657 // Container for (transitive) uses of the associated value.
659 for (const Use &U : Val.uses())
660 Uses.insert(&U);
661
662 followUsesInContext<AAType>(AA, A, *Explorer, &CtxI, Uses, S);
663
664 if (S.isAtFixpoint())
665 return;
666
668 auto Pred = [&](const Instruction *I) {
669 if (const BranchInst *Br = dyn_cast<BranchInst>(I))
670 if (Br->isConditional())
671 BrInsts.push_back(Br);
672 return true;
673 };
674
675 // Here, accumulate conditional branch instructions in the context. We
676 // explore the child paths and collect the known states. The disjunction of
677 // those states can be merged to its own state. Let ParentState_i be a state
678 // to indicate the known information for an i-th branch instruction in the
679 // context. ChildStates are created for its successors respectively.
680 //
681 // ParentS_1 = ChildS_{1, 1} /\ ChildS_{1, 2} /\ ... /\ ChildS_{1, n_1}
682 // ParentS_2 = ChildS_{2, 1} /\ ChildS_{2, 2} /\ ... /\ ChildS_{2, n_2}
683 // ...
684 // ParentS_m = ChildS_{m, 1} /\ ChildS_{m, 2} /\ ... /\ ChildS_{m, n_m}
685 //
686 // Known State |= ParentS_1 \/ ParentS_2 \/... \/ ParentS_m
687 //
688 // FIXME: Currently, recursive branches are not handled. For example, we
689 // can't deduce that ptr must be dereferenced in below function.
690 //
691 // void f(int a, int c, int *ptr) {
692 // if(a)
693 // if (b) {
694 // *ptr = 0;
695 // } else {
696 // *ptr = 1;
697 // }
698 // else {
699 // if (b) {
700 // *ptr = 0;
701 // } else {
702 // *ptr = 1;
703 // }
704 // }
705 // }
706
707 Explorer->checkForAllContext(&CtxI, Pred);
708 for (const BranchInst *Br : BrInsts) {
709 StateType ParentState;
710
711 // The known state of the parent state is a conjunction of children's
712 // known states so it is initialized with a best state.
713 ParentState.indicateOptimisticFixpoint();
714
715 for (const BasicBlock *BB : Br->successors()) {
716 StateType ChildState;
717
718 size_t BeforeSize = Uses.size();
719 followUsesInContext(AA, A, *Explorer, &BB->front(), Uses, ChildState);
720
721 // Erase uses which only appear in the child.
722 for (auto It = Uses.begin() + BeforeSize; It != Uses.end();)
723 It = Uses.erase(It);
724
725 ParentState &= ChildState;
726 }
727
728 // Use only known state.
729 S += ParentState;
730 }
731}
732} // namespace
733
734/// ------------------------ PointerInfo ---------------------------------------
735
736namespace llvm {
737namespace AA {
738namespace PointerInfo {
739
740struct State;
741
742} // namespace PointerInfo
743} // namespace AA
744
745/// Helper for AA::PointerInfo::Access DenseMap/Set usage.
746template <>
749 static inline Access getEmptyKey();
750 static inline Access getTombstoneKey();
751 static unsigned getHashValue(const Access &A);
752 static bool isEqual(const Access &LHS, const Access &RHS);
753};
754
755/// Helper that allows RangeTy as a key in a DenseMap.
756template <> struct DenseMapInfo<AA::RangeTy> {
757 static inline AA::RangeTy getEmptyKey() {
758 auto EmptyKey = DenseMapInfo<int64_t>::getEmptyKey();
759 return AA::RangeTy{EmptyKey, EmptyKey};
760 }
761
762 static inline AA::RangeTy getTombstoneKey() {
763 auto TombstoneKey = DenseMapInfo<int64_t>::getTombstoneKey();
764 return AA::RangeTy{TombstoneKey, TombstoneKey};
765 }
766
772
773 static bool isEqual(const AA::RangeTy &A, const AA::RangeTy B) {
774 return A == B;
775 }
776};
777
778/// Helper for AA::PointerInfo::Access DenseMap/Set usage ignoring everythign
779/// but the instruction
780struct AccessAsInstructionInfo : DenseMapInfo<Instruction *> {
783 static inline Access getEmptyKey();
784 static inline Access getTombstoneKey();
785 static unsigned getHashValue(const Access &A);
786 static bool isEqual(const Access &LHS, const Access &RHS);
787};
788
789} // namespace llvm
790
791/// A type to track pointer/struct usage and accesses for AAPointerInfo.
793 /// Return the best possible representable state.
794 static State getBestState(const State &SIS) { return State(); }
795
796 /// Return the worst possible representable state.
797 static State getWorstState(const State &SIS) {
798 State R;
799 R.indicatePessimisticFixpoint();
800 return R;
801 }
802
803 State() = default;
804 State(State &&SIS) = default;
805
806 const State &getAssumed() const { return *this; }
807
808 /// See AbstractState::isValidState().
809 bool isValidState() const override { return BS.isValidState(); }
810
811 /// See AbstractState::isAtFixpoint().
812 bool isAtFixpoint() const override { return BS.isAtFixpoint(); }
813
814 /// See AbstractState::indicateOptimisticFixpoint().
816 BS.indicateOptimisticFixpoint();
818 }
819
820 /// See AbstractState::indicatePessimisticFixpoint().
822 BS.indicatePessimisticFixpoint();
824 }
825
826 State &operator=(const State &R) {
827 if (this == &R)
828 return *this;
829 BS = R.BS;
830 AccessList = R.AccessList;
831 OffsetBins = R.OffsetBins;
832 RemoteIMap = R.RemoteIMap;
833 ReturnedOffsets = R.ReturnedOffsets;
834 return *this;
835 }
836
838 if (this == &R)
839 return *this;
840 std::swap(BS, R.BS);
841 std::swap(AccessList, R.AccessList);
842 std::swap(OffsetBins, R.OffsetBins);
843 std::swap(RemoteIMap, R.RemoteIMap);
844 std::swap(ReturnedOffsets, R.ReturnedOffsets);
845 return *this;
846 }
847
848 /// Add a new Access to the state at offset \p Offset and with size \p Size.
849 /// The access is associated with \p I, writes \p Content (if anything), and
850 /// is of kind \p Kind. If an Access already exists for the same \p I and same
851 /// \p RemoteI, the two are combined, potentially losing information about
852 /// offset and size. The resulting access must now be moved from its original
853 /// OffsetBin to the bin for its new offset.
854 ///
855 /// \Returns CHANGED, if the state changed, UNCHANGED otherwise.
857 Instruction &I, std::optional<Value *> Content,
859 Instruction *RemoteI = nullptr);
860
863 int64_t numOffsetBins() const { return OffsetBins.size(); }
864
865 const AAPointerInfo::Access &getAccess(unsigned Index) const {
866 return AccessList[Index];
867 }
868
869protected:
870 // Every memory instruction results in an Access object. We maintain a list of
871 // all Access objects that we own, along with the following maps:
872 //
873 // - OffsetBins: RangeTy -> { Access }
874 // - RemoteIMap: RemoteI x LocalI -> Access
875 //
876 // A RemoteI is any instruction that accesses memory. RemoteI is different
877 // from LocalI if and only if LocalI is a call; then RemoteI is some
878 // instruction in the callgraph starting from LocalI. Multiple paths in the
879 // callgraph from LocalI to RemoteI may produce multiple accesses, but these
880 // are all combined into a single Access object. This may result in loss of
881 // information in RangeTy in the Access object.
885
886 /// Flag to determine if the underlying pointer is reaching a return statement
887 /// in the associated function or not. Returns in other functions cause
888 /// invalidation.
890
891 /// See AAPointerInfo::forallInterferingAccesses.
892 template <typename F>
894 if (!isValidState() || !ReturnedOffsets.isUnassigned())
895 return false;
896
897 for (const auto &It : OffsetBins) {
898 AA::RangeTy ItRange = It.getFirst();
899 if (!Range.mayOverlap(ItRange))
900 continue;
901 bool IsExact = Range == ItRange && !Range.offsetOrSizeAreUnknown();
902 for (auto Index : It.getSecond()) {
903 auto &Access = AccessList[Index];
904 if (!CB(Access, IsExact))
905 return false;
906 }
907 }
908 return true;
909 }
910
911 /// See AAPointerInfo::forallInterferingAccesses.
912 template <typename F>
914 AA::RangeTy &Range) const {
915 if (!isValidState() || !ReturnedOffsets.isUnassigned())
916 return false;
917
918 auto LocalList = RemoteIMap.find(&I);
919 if (LocalList == RemoteIMap.end()) {
920 return true;
921 }
922
923 for (unsigned Index : LocalList->getSecond()) {
924 for (auto &R : AccessList[Index]) {
925 Range &= R;
926 if (Range.offsetAndSizeAreUnknown())
927 break;
928 }
929 }
931 }
932
933private:
934 /// State to track fixpoint and validity.
935 BooleanState BS;
936};
937
940 std::optional<Value *> Content, AAPointerInfo::AccessKind Kind, Type *Ty,
941 Instruction *RemoteI) {
942 RemoteI = RemoteI ? RemoteI : &I;
943
944 // Check if we have an access for this instruction, if not, simply add it.
945 auto &LocalList = RemoteIMap[RemoteI];
946 bool AccExists = false;
947 unsigned AccIndex = AccessList.size();
948 for (auto Index : LocalList) {
949 auto &A = AccessList[Index];
950 if (A.getLocalInst() == &I) {
951 AccExists = true;
952 AccIndex = Index;
953 break;
954 }
955 }
956
957 auto AddToBins = [&](const AAPointerInfo::RangeList &ToAdd) {
958 LLVM_DEBUG(if (ToAdd.size()) dbgs()
959 << "[AAPointerInfo] Inserting access in new offset bins\n";);
960
961 for (auto Key : ToAdd) {
962 LLVM_DEBUG(dbgs() << " key " << Key << "\n");
963 OffsetBins[Key].insert(AccIndex);
964 }
965 };
966
967 if (!AccExists) {
968 AccessList.emplace_back(&I, RemoteI, Ranges, Content, Kind, Ty);
969 assert((AccessList.size() == AccIndex + 1) &&
970 "New Access should have been at AccIndex");
971 LocalList.push_back(AccIndex);
972 AddToBins(AccessList[AccIndex].getRanges());
974 }
975
976 // Combine the new Access with the existing Access, and then update the
977 // mapping in the offset bins.
978 AAPointerInfo::Access Acc(&I, RemoteI, Ranges, Content, Kind, Ty);
979 auto &Current = AccessList[AccIndex];
980 auto Before = Current;
981 Current &= Acc;
982 if (Current == Before)
984
985 auto &ExistingRanges = Before.getRanges();
986 auto &NewRanges = Current.getRanges();
987
988 // Ranges that are in the old access but not the new access need to be removed
989 // from the offset bins.
991 AAPointerInfo::RangeList::set_difference(ExistingRanges, NewRanges, ToRemove);
992 LLVM_DEBUG(if (ToRemove.size()) dbgs()
993 << "[AAPointerInfo] Removing access from old offset bins\n";);
994
995 for (auto Key : ToRemove) {
996 LLVM_DEBUG(dbgs() << " key " << Key << "\n");
997 assert(OffsetBins.count(Key) && "Existing Access must be in some bin.");
998 auto &Bin = OffsetBins[Key];
999 assert(Bin.count(AccIndex) &&
1000 "Expected bin to actually contain the Access.");
1001 Bin.erase(AccIndex);
1002 }
1003
1004 // Ranges that are in the new access but not the old access need to be added
1005 // to the offset bins.
1007 AAPointerInfo::RangeList::set_difference(NewRanges, ExistingRanges, ToAdd);
1008 AddToBins(ToAdd);
1009 return ChangeStatus::CHANGED;
1010}
1011
1012namespace {
1013
1014#ifndef NDEBUG
1016 const AAPointerInfo::OffsetInfo &OI) {
1017 OS << llvm::interleaved_array(OI);
1018 return OS;
1019}
1020#endif // NDEBUG
1021
1022struct AAPointerInfoImpl
1023 : public StateWrapper<AA::PointerInfo::State, AAPointerInfo> {
1025 AAPointerInfoImpl(const IRPosition &IRP, Attributor &A) : BaseTy(IRP) {}
1026
1027 /// See AbstractAttribute::getAsStr().
1028 const std::string getAsStr(Attributor *A) const override {
1029 return std::string("PointerInfo ") +
1030 (isValidState() ? (std::string("#") +
1031 std::to_string(OffsetBins.size()) + " bins")
1032 : "<invalid>") +
1033 (reachesReturn()
1034 ? (" (returned:" +
1035 join(map_range(ReturnedOffsets,
1036 [](int64_t O) { return std::to_string(O); }),
1037 ", ") +
1038 ")")
1039 : "");
1040 }
1041
1042 /// See AbstractAttribute::manifest(...).
1043 ChangeStatus manifest(Attributor &A) override {
1044 return AAPointerInfo::manifest(A);
1045 }
1046
1047 const_bin_iterator begin() const override { return State::begin(); }
1048 const_bin_iterator end() const override { return State::end(); }
1049 int64_t numOffsetBins() const override { return State::numOffsetBins(); }
1050 bool reachesReturn() const override {
1051 return !ReturnedOffsets.isUnassigned();
1052 }
1053 void addReturnedOffsetsTo(OffsetInfo &OI) const override {
1054 if (ReturnedOffsets.isUnknown()) {
1055 OI.setUnknown();
1056 return;
1057 }
1058
1059 OffsetInfo MergedOI;
1060 for (auto Offset : ReturnedOffsets) {
1061 OffsetInfo TmpOI = OI;
1062 TmpOI.addToAll(Offset);
1063 MergedOI.merge(TmpOI);
1064 }
1065 OI = std::move(MergedOI);
1066 }
1067
1068 ChangeStatus setReachesReturn(const OffsetInfo &ReachedReturnedOffsets) {
1069 if (ReturnedOffsets.isUnknown())
1070 return ChangeStatus::UNCHANGED;
1071 if (ReachedReturnedOffsets.isUnknown()) {
1072 ReturnedOffsets.setUnknown();
1073 return ChangeStatus::CHANGED;
1074 }
1075 if (ReturnedOffsets.merge(ReachedReturnedOffsets))
1076 return ChangeStatus::CHANGED;
1077 return ChangeStatus::UNCHANGED;
1078 }
1079
1080 bool forallInterferingAccesses(
1081 AA::RangeTy Range,
1082 function_ref<bool(const AAPointerInfo::Access &, bool)> CB)
1083 const override {
1084 return State::forallInterferingAccesses(Range, CB);
1085 }
1086
1087 bool forallInterferingAccesses(
1088 Attributor &A, const AbstractAttribute &QueryingAA, Instruction &I,
1089 bool FindInterferingWrites, bool FindInterferingReads,
1090 function_ref<bool(const Access &, bool)> UserCB, bool &HasBeenWrittenTo,
1091 AA::RangeTy &Range,
1092 function_ref<bool(const Access &)> SkipCB) const override {
1093 HasBeenWrittenTo = false;
1094
1095 SmallPtrSet<const Access *, 8> DominatingWrites;
1096 SmallVector<std::pair<const Access *, bool>, 8> InterferingAccesses;
1097
1098 Function &Scope = *I.getFunction();
1099 bool IsKnownNoSync;
1100 bool IsAssumedNoSync = AA::hasAssumedIRAttr<Attribute::NoSync>(
1101 A, &QueryingAA, IRPosition::function(Scope), DepClassTy::OPTIONAL,
1102 IsKnownNoSync);
1103 const auto *ExecDomainAA = A.lookupAAFor<AAExecutionDomain>(
1104 IRPosition::function(Scope), &QueryingAA, DepClassTy::NONE);
1105 bool AllInSameNoSyncFn = IsAssumedNoSync;
1106 bool InstIsExecutedByInitialThreadOnly =
1107 ExecDomainAA && ExecDomainAA->isExecutedByInitialThreadOnly(I);
1108
1109 // If the function is not ending in aligned barriers, we need the stores to
1110 // be in aligned barriers. The load being in one is not sufficient since the
1111 // store might be executed by a thread that disappears after, causing the
1112 // aligned barrier guarding the load to unblock and the load to read a value
1113 // that has no CFG path to the load.
1114 bool InstIsExecutedInAlignedRegion =
1115 FindInterferingReads && ExecDomainAA &&
1116 ExecDomainAA->isExecutedInAlignedRegion(A, I);
1117
1118 if (InstIsExecutedInAlignedRegion || InstIsExecutedByInitialThreadOnly)
1119 A.recordDependence(*ExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1120
1121 InformationCache &InfoCache = A.getInfoCache();
1122 bool IsThreadLocalObj =
1123 AA::isAssumedThreadLocalObject(A, getAssociatedValue(), *this);
1124
1125 // Helper to determine if we need to consider threading, which we cannot
1126 // right now. However, if the function is (assumed) nosync or the thread
1127 // executing all instructions is the main thread only we can ignore
1128 // threading. Also, thread-local objects do not require threading reasoning.
1129 // Finally, we can ignore threading if either access is executed in an
1130 // aligned region.
1131 auto CanIgnoreThreadingForInst = [&](const Instruction &I) -> bool {
1132 if (IsThreadLocalObj || AllInSameNoSyncFn)
1133 return true;
1134 const auto *FnExecDomainAA =
1135 I.getFunction() == &Scope
1136 ? ExecDomainAA
1137 : A.lookupAAFor<AAExecutionDomain>(
1138 IRPosition::function(*I.getFunction()), &QueryingAA,
1139 DepClassTy::NONE);
1140 if (!FnExecDomainAA)
1141 return false;
1142 if (InstIsExecutedInAlignedRegion ||
1143 (FindInterferingWrites &&
1144 FnExecDomainAA->isExecutedInAlignedRegion(A, I))) {
1145 A.recordDependence(*FnExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1146 return true;
1147 }
1148 if (InstIsExecutedByInitialThreadOnly &&
1149 FnExecDomainAA->isExecutedByInitialThreadOnly(I)) {
1150 A.recordDependence(*FnExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1151 return true;
1152 }
1153 return false;
1154 };
1155
1156 // Helper to determine if the access is executed by the same thread as the
1157 // given instruction, for now it is sufficient to avoid any potential
1158 // threading effects as we cannot deal with them anyway.
1159 auto CanIgnoreThreading = [&](const Access &Acc) -> bool {
1160 return CanIgnoreThreadingForInst(*Acc.getRemoteInst()) ||
1161 (Acc.getRemoteInst() != Acc.getLocalInst() &&
1162 CanIgnoreThreadingForInst(*Acc.getLocalInst()));
1163 };
1164
1165 // TODO: Use inter-procedural reachability and dominance.
1166 bool IsKnownNoRecurse;
1168 A, this, IRPosition::function(Scope), DepClassTy::OPTIONAL,
1169 IsKnownNoRecurse);
1170
1171 // TODO: Use reaching kernels from AAKernelInfo (or move it to
1172 // AAExecutionDomain) such that we allow scopes other than kernels as long
1173 // as the reaching kernels are disjoint.
1174 bool InstInKernel = A.getInfoCache().isKernel(Scope);
1175 bool ObjHasKernelLifetime = false;
1176 const bool UseDominanceReasoning =
1177 FindInterferingWrites && IsKnownNoRecurse;
1178 const DominatorTree *DT =
1179 InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(Scope);
1180
1181 // Helper to check if a value has "kernel lifetime", that is it will not
1182 // outlive a GPU kernel. This is true for shared, constant, and local
1183 // globals on AMD and NVIDIA GPUs.
1184 auto HasKernelLifetime = [&](Value *V, Module &M) {
1185 if (!AA::isGPU(M))
1186 return false;
1187 switch (AA::GPUAddressSpace(V->getType()->getPointerAddressSpace())) {
1188 case AA::GPUAddressSpace::Shared:
1189 case AA::GPUAddressSpace::Constant:
1190 case AA::GPUAddressSpace::Local:
1191 return true;
1192 default:
1193 return false;
1194 };
1195 };
1196
1197 // The IsLiveInCalleeCB will be used by the AA::isPotentiallyReachable query
1198 // to determine if we should look at reachability from the callee. For
1199 // certain pointers we know the lifetime and we do not have to step into the
1200 // callee to determine reachability as the pointer would be dead in the
1201 // callee. See the conditional initialization below.
1202 std::function<bool(const Function &)> IsLiveInCalleeCB;
1203
1204 if (auto *AI = dyn_cast<AllocaInst>(&getAssociatedValue())) {
1205 // If the alloca containing function is not recursive the alloca
1206 // must be dead in the callee.
1207 const Function *AIFn = AI->getFunction();
1208 ObjHasKernelLifetime = A.getInfoCache().isKernel(*AIFn);
1209 bool IsKnownNoRecurse;
1211 A, this, IRPosition::function(*AIFn), DepClassTy::OPTIONAL,
1212 IsKnownNoRecurse)) {
1213 IsLiveInCalleeCB = [AIFn](const Function &Fn) { return AIFn != &Fn; };
1214 }
1215 } else if (auto *GV = dyn_cast<GlobalValue>(&getAssociatedValue())) {
1216 // If the global has kernel lifetime we can stop if we reach a kernel
1217 // as it is "dead" in the (unknown) callees.
1218 ObjHasKernelLifetime = HasKernelLifetime(GV, *GV->getParent());
1219 if (ObjHasKernelLifetime)
1220 IsLiveInCalleeCB = [&A](const Function &Fn) {
1221 return !A.getInfoCache().isKernel(Fn);
1222 };
1223 }
1224
1225 // Set of accesses/instructions that will overwrite the result and are
1226 // therefore blockers in the reachability traversal.
1227 AA::InstExclusionSetTy ExclusionSet;
1228
1229 auto AccessCB = [&](const Access &Acc, bool Exact) {
1230 Function *AccScope = Acc.getRemoteInst()->getFunction();
1231 bool AccInSameScope = AccScope == &Scope;
1232
1233 // If the object has kernel lifetime we can ignore accesses only reachable
1234 // by other kernels. For now we only skip accesses *in* other kernels.
1235 if (InstInKernel && ObjHasKernelLifetime && !AccInSameScope &&
1236 A.getInfoCache().isKernel(*AccScope))
1237 return true;
1238
1239 if (Exact && Acc.isMustAccess() && Acc.getRemoteInst() != &I) {
1240 if (Acc.isWrite() || (isa<LoadInst>(I) && Acc.isWriteOrAssumption()))
1241 ExclusionSet.insert(Acc.getRemoteInst());
1242 }
1243
1244 if ((!FindInterferingWrites || !Acc.isWriteOrAssumption()) &&
1245 (!FindInterferingReads || !Acc.isRead()))
1246 return true;
1247
1248 bool Dominates = FindInterferingWrites && DT && Exact &&
1249 Acc.isMustAccess() && AccInSameScope &&
1250 DT->dominates(Acc.getRemoteInst(), &I);
1251 if (Dominates)
1252 DominatingWrites.insert(&Acc);
1253
1254 // Track if all interesting accesses are in the same `nosync` function as
1255 // the given instruction.
1256 AllInSameNoSyncFn &= Acc.getRemoteInst()->getFunction() == &Scope;
1257
1258 InterferingAccesses.push_back({&Acc, Exact});
1259 return true;
1260 };
1261 if (!State::forallInterferingAccesses(I, AccessCB, Range))
1262 return false;
1263
1264 HasBeenWrittenTo = !DominatingWrites.empty();
1265
1266 // Dominating writes form a chain, find the least/lowest member.
1267 Instruction *LeastDominatingWriteInst = nullptr;
1268 for (const Access *Acc : DominatingWrites) {
1269 if (!LeastDominatingWriteInst) {
1270 LeastDominatingWriteInst = Acc->getRemoteInst();
1271 } else if (DT->dominates(LeastDominatingWriteInst,
1272 Acc->getRemoteInst())) {
1273 LeastDominatingWriteInst = Acc->getRemoteInst();
1274 }
1275 }
1276
1277 // Helper to determine if we can skip a specific write access.
1278 auto CanSkipAccess = [&](const Access &Acc, bool Exact) {
1279 if (SkipCB && SkipCB(Acc))
1280 return true;
1281 if (!CanIgnoreThreading(Acc))
1282 return false;
1283
1284 // Check read (RAW) dependences and write (WAR) dependences as necessary.
1285 // If we successfully excluded all effects we are interested in, the
1286 // access can be skipped.
1287 bool ReadChecked = !FindInterferingReads;
1288 bool WriteChecked = !FindInterferingWrites;
1289
1290 // If the instruction cannot reach the access, the former does not
1291 // interfere with what the access reads.
1292 if (!ReadChecked) {
1293 if (!AA::isPotentiallyReachable(A, I, *Acc.getRemoteInst(), QueryingAA,
1294 &ExclusionSet, IsLiveInCalleeCB))
1295 ReadChecked = true;
1296 }
1297 // If the instruction cannot be reach from the access, the latter does not
1298 // interfere with what the instruction reads.
1299 if (!WriteChecked) {
1300 if (!AA::isPotentiallyReachable(A, *Acc.getRemoteInst(), I, QueryingAA,
1301 &ExclusionSet, IsLiveInCalleeCB))
1302 WriteChecked = true;
1303 }
1304
1305 // If we still might be affected by the write of the access but there are
1306 // dominating writes in the function of the instruction
1307 // (HasBeenWrittenTo), we can try to reason that the access is overwritten
1308 // by them. This would have happend above if they are all in the same
1309 // function, so we only check the inter-procedural case. Effectively, we
1310 // want to show that there is no call after the dominting write that might
1311 // reach the access, and when it returns reach the instruction with the
1312 // updated value. To this end, we iterate all call sites, check if they
1313 // might reach the instruction without going through another access
1314 // (ExclusionSet) and at the same time might reach the access. However,
1315 // that is all part of AAInterFnReachability.
1316 if (!WriteChecked && HasBeenWrittenTo &&
1317 Acc.getRemoteInst()->getFunction() != &Scope) {
1318
1319 const auto *FnReachabilityAA = A.getAAFor<AAInterFnReachability>(
1320 QueryingAA, IRPosition::function(Scope), DepClassTy::OPTIONAL);
1321 if (FnReachabilityAA) {
1322 // Without going backwards in the call tree, can we reach the access
1323 // from the least dominating write. Do not allow to pass the
1324 // instruction itself either.
1325 bool Inserted = ExclusionSet.insert(&I).second;
1326
1327 if (!FnReachabilityAA->instructionCanReach(
1328 A, *LeastDominatingWriteInst,
1329 *Acc.getRemoteInst()->getFunction(), &ExclusionSet))
1330 WriteChecked = true;
1331
1332 if (Inserted)
1333 ExclusionSet.erase(&I);
1334 }
1335 }
1336
1337 if (ReadChecked && WriteChecked)
1338 return true;
1339
1340 if (!DT || !UseDominanceReasoning)
1341 return false;
1342 if (!DominatingWrites.count(&Acc))
1343 return false;
1344 return LeastDominatingWriteInst != Acc.getRemoteInst();
1345 };
1346
1347 // Run the user callback on all accesses we cannot skip and return if
1348 // that succeeded for all or not.
1349 for (auto &It : InterferingAccesses) {
1350 if ((!AllInSameNoSyncFn && !IsThreadLocalObj && !ExecDomainAA) ||
1351 !CanSkipAccess(*It.first, It.second)) {
1352 if (!UserCB(*It.first, It.second))
1353 return false;
1354 }
1355 }
1356 return true;
1357 }
1358
1359 ChangeStatus translateAndAddStateFromCallee(Attributor &A,
1360 const AAPointerInfo &OtherAA,
1361 CallBase &CB) {
1362 using namespace AA::PointerInfo;
1363 if (!OtherAA.getState().isValidState() || !isValidState())
1364 return indicatePessimisticFixpoint();
1365
1366 ChangeStatus Changed = ChangeStatus::UNCHANGED;
1367 const auto &OtherAAImpl = static_cast<const AAPointerInfoImpl &>(OtherAA);
1368 bool IsByval = OtherAAImpl.getAssociatedArgument()->hasByValAttr();
1369 Changed |= setReachesReturn(OtherAAImpl.ReturnedOffsets);
1370
1371 // Combine the accesses bin by bin.
1372 const auto &State = OtherAAImpl.getState();
1373 for (const auto &It : State) {
1374 for (auto Index : It.getSecond()) {
1375 const auto &RAcc = State.getAccess(Index);
1376 if (IsByval && !RAcc.isRead())
1377 continue;
1378 bool UsedAssumedInformation = false;
1379 AccessKind AK = RAcc.getKind();
1380 auto Content = A.translateArgumentToCallSiteContent(
1381 RAcc.getContent(), CB, *this, UsedAssumedInformation);
1382 AK = AccessKind(AK & (IsByval ? AccessKind::AK_R : AccessKind::AK_RW));
1383 AK = AccessKind(AK | (RAcc.isMayAccess() ? AK_MAY : AK_MUST));
1384
1385 Changed |= addAccess(A, RAcc.getRanges(), CB, Content, AK,
1386 RAcc.getType(), RAcc.getRemoteInst());
1387 }
1388 }
1389 return Changed;
1390 }
1391
1392 ChangeStatus translateAndAddState(Attributor &A, const AAPointerInfo &OtherAA,
1393 const OffsetInfo &Offsets, CallBase &CB,
1394 bool IsMustAcc) {
1395 using namespace AA::PointerInfo;
1396 if (!OtherAA.getState().isValidState() || !isValidState())
1397 return indicatePessimisticFixpoint();
1398
1399 const auto &OtherAAImpl = static_cast<const AAPointerInfoImpl &>(OtherAA);
1400
1401 // Combine the accesses bin by bin.
1402 ChangeStatus Changed = ChangeStatus::UNCHANGED;
1403 const auto &State = OtherAAImpl.getState();
1404 for (const auto &It : State) {
1405 for (auto Index : It.getSecond()) {
1406 const auto &RAcc = State.getAccess(Index);
1407 if (!IsMustAcc && RAcc.isAssumption())
1408 continue;
1409 for (auto Offset : Offsets) {
1410 auto NewRanges = Offset == AA::RangeTy::Unknown
1412 : RAcc.getRanges();
1413 if (!NewRanges.isUnknown()) {
1414 NewRanges.addToAllOffsets(Offset);
1415 }
1416 AccessKind AK = RAcc.getKind();
1417 if (!IsMustAcc)
1418 AK = AccessKind((AK & ~AK_MUST) | AK_MAY);
1419 Changed |= addAccess(A, NewRanges, CB, RAcc.getContent(), AK,
1420 RAcc.getType(), RAcc.getRemoteInst());
1421 }
1422 }
1423 }
1424 return Changed;
1425 }
1426
1427 /// Statistic tracking for all AAPointerInfo implementations.
1428 /// See AbstractAttribute::trackStatistics().
1429 void trackPointerInfoStatistics(const IRPosition &IRP) const {}
1430
1431 /// Dump the state into \p O.
1432 void dumpState(raw_ostream &O) {
1433 for (auto &It : OffsetBins) {
1434 O << "[" << It.first.Offset << "-" << It.first.Offset + It.first.Size
1435 << "] : " << It.getSecond().size() << "\n";
1436 for (auto AccIndex : It.getSecond()) {
1437 auto &Acc = AccessList[AccIndex];
1438 O << " - " << Acc.getKind() << " - " << *Acc.getLocalInst() << "\n";
1439 if (Acc.getLocalInst() != Acc.getRemoteInst())
1440 O << " --> " << *Acc.getRemoteInst()
1441 << "\n";
1442 if (!Acc.isWrittenValueYetUndetermined()) {
1443 if (isa_and_nonnull<Function>(Acc.getWrittenValue()))
1444 O << " - c: func " << Acc.getWrittenValue()->getName()
1445 << "\n";
1446 else if (Acc.getWrittenValue())
1447 O << " - c: " << *Acc.getWrittenValue() << "\n";
1448 else
1449 O << " - c: <unknown>\n";
1450 }
1451 }
1452 }
1453 }
1454};
1455
1456struct AAPointerInfoFloating : public AAPointerInfoImpl {
1458 AAPointerInfoFloating(const IRPosition &IRP, Attributor &A)
1459 : AAPointerInfoImpl(IRP, A) {}
1460
1461 /// Deal with an access and signal if it was handled successfully.
1462 bool handleAccess(Attributor &A, Instruction &I,
1463 std::optional<Value *> Content, AccessKind Kind,
1464 OffsetInfo::VecTy &Offsets, ChangeStatus &Changed,
1465 Type &Ty) {
1466 using namespace AA::PointerInfo;
1468 const DataLayout &DL = A.getDataLayout();
1469 TypeSize AccessSize = DL.getTypeStoreSize(&Ty);
1470 if (!AccessSize.isScalable())
1471 Size = AccessSize.getFixedValue();
1472
1473 // Make a strictly ascending list of offsets as required by addAccess()
1474 SmallVector<int64_t> OffsetsSorted(Offsets.begin(), Offsets.end());
1475 llvm::sort(OffsetsSorted);
1476
1478 if (!VT || VT->getElementCount().isScalable() ||
1479 !Content.value_or(nullptr) || !isa<Constant>(*Content) ||
1480 (*Content)->getType() != VT ||
1481 DL.getTypeStoreSize(VT->getElementType()).isScalable()) {
1482 Changed =
1483 Changed | addAccess(A, {OffsetsSorted, Size}, I, Content, Kind, &Ty);
1484 } else {
1485 // Handle vector stores with constant content element-wise.
1486 // TODO: We could look for the elements or create instructions
1487 // representing them.
1488 // TODO: We need to push the Content into the range abstraction
1489 // (AA::RangeTy) to allow different content values for different
1490 // ranges. ranges. Hence, support vectors storing different values.
1491 Type *ElementType = VT->getElementType();
1492 int64_t ElementSize = DL.getTypeStoreSize(ElementType).getFixedValue();
1493 auto *ConstContent = cast<Constant>(*Content);
1494 Type *Int32Ty = Type::getInt32Ty(ElementType->getContext());
1495 SmallVector<int64_t> ElementOffsets(Offsets.begin(), Offsets.end());
1496
1497 for (int i = 0, e = VT->getElementCount().getFixedValue(); i != e; ++i) {
1498 Value *ElementContent = ConstantExpr::getExtractElement(
1499 ConstContent, ConstantInt::get(Int32Ty, i));
1500
1501 // Add the element access.
1502 Changed = Changed | addAccess(A, {ElementOffsets, ElementSize}, I,
1503 ElementContent, Kind, ElementType);
1504
1505 // Advance the offsets for the next element.
1506 for (auto &ElementOffset : ElementOffsets)
1507 ElementOffset += ElementSize;
1508 }
1509 }
1510 return true;
1511 };
1512
1513 /// See AbstractAttribute::updateImpl(...).
1514 ChangeStatus updateImpl(Attributor &A) override;
1515
1516 /// If the indices to \p GEP can be traced to constants, incorporate all
1517 /// of these into \p UsrOI.
1518 ///
1519 /// \return true iff \p UsrOI is updated.
1520 bool collectConstantsForGEP(Attributor &A, const DataLayout &DL,
1521 OffsetInfo &UsrOI, const OffsetInfo &PtrOI,
1522 const GEPOperator *GEP);
1523
1524 /// See AbstractAttribute::trackStatistics()
1525 void trackStatistics() const override {
1526 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1527 }
1528};
1529
1530bool AAPointerInfoFloating::collectConstantsForGEP(Attributor &A,
1531 const DataLayout &DL,
1532 OffsetInfo &UsrOI,
1533 const OffsetInfo &PtrOI,
1534 const GEPOperator *GEP) {
1535 unsigned BitWidth = DL.getIndexTypeSizeInBits(GEP->getType());
1536 SmallMapVector<Value *, APInt, 4> VariableOffsets;
1537 APInt ConstantOffset(BitWidth, 0);
1538
1539 assert(!UsrOI.isUnknown() && !PtrOI.isUnknown() &&
1540 "Don't look for constant values if the offset has already been "
1541 "determined to be unknown.");
1542
1543 if (!GEP->collectOffset(DL, BitWidth, VariableOffsets, ConstantOffset)) {
1544 UsrOI.setUnknown();
1545 return true;
1546 }
1547
1548 LLVM_DEBUG(dbgs() << "[AAPointerInfo] GEP offset is "
1549 << (VariableOffsets.empty() ? "" : "not") << " constant "
1550 << *GEP << "\n");
1551
1552 auto Union = PtrOI;
1553 Union.addToAll(ConstantOffset.getSExtValue());
1554
1555 // Each VI in VariableOffsets has a set of potential constant values. Every
1556 // combination of elements, picked one each from these sets, is separately
1557 // added to the original set of offsets, thus resulting in more offsets.
1558 for (const auto &VI : VariableOffsets) {
1559 auto *PotentialConstantsAA = A.getAAFor<AAPotentialConstantValues>(
1560 *this, IRPosition::value(*VI.first), DepClassTy::OPTIONAL);
1561 if (!PotentialConstantsAA || !PotentialConstantsAA->isValidState()) {
1562 UsrOI.setUnknown();
1563 return true;
1564 }
1565
1566 // UndefValue is treated as a zero, which leaves Union as is.
1567 if (PotentialConstantsAA->undefIsContained())
1568 continue;
1569
1570 // We need at least one constant in every set to compute an actual offset.
1571 // Otherwise, we end up pessimizing AAPointerInfo by respecting offsets that
1572 // don't actually exist. In other words, the absence of constant values
1573 // implies that the operation can be assumed dead for now.
1574 auto &AssumedSet = PotentialConstantsAA->getAssumedSet();
1575 if (AssumedSet.empty())
1576 return false;
1577
1578 OffsetInfo Product;
1579 for (const auto &ConstOffset : AssumedSet) {
1580 auto CopyPerOffset = Union;
1581 CopyPerOffset.addToAll(ConstOffset.getSExtValue() *
1582 VI.second.getZExtValue());
1583 Product.merge(CopyPerOffset);
1584 }
1585 Union = Product;
1586 }
1587
1588 UsrOI = std::move(Union);
1589 return true;
1590}
1591
1592ChangeStatus AAPointerInfoFloating::updateImpl(Attributor &A) {
1593 using namespace AA::PointerInfo;
1595 const DataLayout &DL = A.getDataLayout();
1596 Value &AssociatedValue = getAssociatedValue();
1597
1598 DenseMap<Value *, OffsetInfo> OffsetInfoMap;
1599 OffsetInfoMap[&AssociatedValue].insert(0);
1600
1601 auto HandlePassthroughUser = [&](Value *Usr, Value *CurPtr, bool &Follow) {
1602 // One does not simply walk into a map and assign a reference to a possibly
1603 // new location. That can cause an invalidation before the assignment
1604 // happens, like so:
1605 //
1606 // OffsetInfoMap[Usr] = OffsetInfoMap[CurPtr]; /* bad idea! */
1607 //
1608 // The RHS is a reference that may be invalidated by an insertion caused by
1609 // the LHS. So we ensure that the side-effect of the LHS happens first.
1610
1611 assert(OffsetInfoMap.contains(CurPtr) &&
1612 "CurPtr does not exist in the map!");
1613
1614 auto &UsrOI = OffsetInfoMap[Usr];
1615 auto &PtrOI = OffsetInfoMap[CurPtr];
1616 assert(!PtrOI.isUnassigned() &&
1617 "Cannot pass through if the input Ptr was not visited!");
1618 UsrOI.merge(PtrOI);
1619 Follow = true;
1620 return true;
1621 };
1622
1623 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
1624 Value *CurPtr = U.get();
1625 User *Usr = U.getUser();
1626 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Analyze " << *CurPtr << " in " << *Usr
1627 << "\n");
1628 assert(OffsetInfoMap.count(CurPtr) &&
1629 "The current pointer offset should have been seeded!");
1630 assert(!OffsetInfoMap[CurPtr].isUnassigned() &&
1631 "Current pointer should be assigned");
1632
1633 if (ConstantExpr *CE = dyn_cast<ConstantExpr>(Usr)) {
1634 if (CE->isCast())
1635 return HandlePassthroughUser(Usr, CurPtr, Follow);
1636 if (!isa<GEPOperator>(CE)) {
1637 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Unhandled constant user " << *CE
1638 << "\n");
1639 return false;
1640 }
1641 }
1642 if (auto *GEP = dyn_cast<GEPOperator>(Usr)) {
1643 // Note the order here, the Usr access might change the map, CurPtr is
1644 // already in it though.
1645 auto &UsrOI = OffsetInfoMap[Usr];
1646 auto &PtrOI = OffsetInfoMap[CurPtr];
1647
1648 if (UsrOI.isUnknown())
1649 return true;
1650
1651 if (PtrOI.isUnknown()) {
1652 Follow = true;
1653 UsrOI.setUnknown();
1654 return true;
1655 }
1656
1657 Follow = collectConstantsForGEP(A, DL, UsrOI, PtrOI, GEP);
1658 return true;
1659 }
1660 if (isa<PtrToIntInst>(Usr))
1661 return false;
1662 if (isa<CastInst>(Usr) || isa<SelectInst>(Usr))
1663 return HandlePassthroughUser(Usr, CurPtr, Follow);
1664 // Returns are allowed if they are in the associated functions. Users can
1665 // then check the call site return. Returns from other functions can't be
1666 // tracked and are cause for invalidation.
1667 if (auto *RI = dyn_cast<ReturnInst>(Usr)) {
1668 if (RI->getFunction() == getAssociatedFunction()) {
1669 auto &PtrOI = OffsetInfoMap[CurPtr];
1670 Changed |= setReachesReturn(PtrOI);
1671 return true;
1672 }
1673 return false;
1674 }
1675
1676 // For PHIs we need to take care of the recurrence explicitly as the value
1677 // might change while we iterate through a loop. For now, we give up if
1678 // the PHI is not invariant.
1679 if (auto *PHI = dyn_cast<PHINode>(Usr)) {
1680 // Note the order here, the Usr access might change the map, CurPtr is
1681 // already in it though.
1682 auto [PhiIt, IsFirstPHIUser] = OffsetInfoMap.try_emplace(PHI);
1683 auto &UsrOI = PhiIt->second;
1684 auto &PtrOI = OffsetInfoMap[CurPtr];
1685
1686 // Check if the PHI operand has already an unknown offset as we can't
1687 // improve on that anymore.
1688 if (PtrOI.isUnknown()) {
1689 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand offset unknown "
1690 << *CurPtr << " in " << *PHI << "\n");
1691 Follow = !UsrOI.isUnknown();
1692 UsrOI.setUnknown();
1693 return true;
1694 }
1695
1696 // Check if the PHI is invariant (so far).
1697 if (UsrOI == PtrOI) {
1698 assert(!PtrOI.isUnassigned() &&
1699 "Cannot assign if the current Ptr was not visited!");
1700 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI is invariant (so far)");
1701 return true;
1702 }
1703
1704 // Check if the PHI operand can be traced back to AssociatedValue.
1705 APInt Offset(
1706 DL.getIndexSizeInBits(CurPtr->getType()->getPointerAddressSpace()),
1707 0);
1708 Value *CurPtrBase = CurPtr->stripAndAccumulateConstantOffsets(
1709 DL, Offset, /* AllowNonInbounds */ true);
1710 auto It = OffsetInfoMap.find(CurPtrBase);
1711 if (It == OffsetInfoMap.end()) {
1712 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand is too complex "
1713 << *CurPtr << " in " << *PHI
1714 << " (base: " << *CurPtrBase << ")\n");
1715 UsrOI.setUnknown();
1716 Follow = true;
1717 return true;
1718 }
1719
1720 // Check if the PHI operand is not dependent on the PHI itself. Every
1721 // recurrence is a cyclic net of PHIs in the data flow, and has an
1722 // equivalent Cycle in the control flow. One of those PHIs must be in the
1723 // header of that control flow Cycle. This is independent of the choice of
1724 // Cycles reported by CycleInfo. It is sufficient to check the PHIs in
1725 // every Cycle header; if such a node is marked unknown, this will
1726 // eventually propagate through the whole net of PHIs in the recurrence.
1727 const auto *CI =
1728 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
1729 *PHI->getFunction());
1730 if (mayBeInCycle(CI, cast<Instruction>(Usr), /* HeaderOnly */ true)) {
1731 auto BaseOI = It->getSecond();
1732 BaseOI.addToAll(Offset.getZExtValue());
1733 if (IsFirstPHIUser || BaseOI == UsrOI) {
1734 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI is invariant " << *CurPtr
1735 << " in " << *Usr << "\n");
1736 return HandlePassthroughUser(Usr, CurPtr, Follow);
1737 }
1738
1739 LLVM_DEBUG(
1740 dbgs() << "[AAPointerInfo] PHI operand pointer offset mismatch "
1741 << *CurPtr << " in " << *PHI << "\n");
1742 UsrOI.setUnknown();
1743 Follow = true;
1744 return true;
1745 }
1746
1747 UsrOI.merge(PtrOI);
1748 Follow = true;
1749 return true;
1750 }
1751
1752 if (auto *LoadI = dyn_cast<LoadInst>(Usr)) {
1753 // If the access is to a pointer that may or may not be the associated
1754 // value, e.g. due to a PHI, we cannot assume it will be read.
1755 AccessKind AK = AccessKind::AK_R;
1756 if (getUnderlyingObject(CurPtr) == &AssociatedValue)
1757 AK = AccessKind(AK | AccessKind::AK_MUST);
1758 else
1759 AK = AccessKind(AK | AccessKind::AK_MAY);
1760 if (!handleAccess(A, *LoadI, /* Content */ nullptr, AK,
1761 OffsetInfoMap[CurPtr].Offsets, Changed,
1762 *LoadI->getType()))
1763 return false;
1764
1765 auto IsAssumption = [](Instruction &I) {
1766 if (auto *II = dyn_cast<IntrinsicInst>(&I))
1767 return II->isAssumeLikeIntrinsic();
1768 return false;
1769 };
1770
1771 auto IsImpactedInRange = [&](Instruction *FromI, Instruction *ToI) {
1772 // Check if the assumption and the load are executed together without
1773 // memory modification.
1774 do {
1775 if (FromI->mayWriteToMemory() && !IsAssumption(*FromI))
1776 return true;
1777 FromI = FromI->getNextNode();
1778 } while (FromI && FromI != ToI);
1779 return false;
1780 };
1781
1782 BasicBlock *BB = LoadI->getParent();
1783 auto IsValidAssume = [&](IntrinsicInst &IntrI) {
1784 if (IntrI.getIntrinsicID() != Intrinsic::assume)
1785 return false;
1786 BasicBlock *IntrBB = IntrI.getParent();
1787 if (IntrI.getParent() == BB) {
1788 if (IsImpactedInRange(LoadI->getNextNode(), &IntrI))
1789 return false;
1790 } else {
1791 auto PredIt = pred_begin(IntrBB);
1792 if (PredIt == pred_end(IntrBB))
1793 return false;
1794 if ((*PredIt) != BB)
1795 return false;
1796 if (++PredIt != pred_end(IntrBB))
1797 return false;
1798 for (auto *SuccBB : successors(BB)) {
1799 if (SuccBB == IntrBB)
1800 continue;
1801 if (isa<UnreachableInst>(SuccBB->getTerminator()))
1802 continue;
1803 return false;
1804 }
1805 if (IsImpactedInRange(LoadI->getNextNode(), BB->getTerminator()))
1806 return false;
1807 if (IsImpactedInRange(&IntrBB->front(), &IntrI))
1808 return false;
1809 }
1810 return true;
1811 };
1812
1813 std::pair<Value *, IntrinsicInst *> Assumption;
1814 for (const Use &LoadU : LoadI->uses()) {
1815 if (auto *CmpI = dyn_cast<CmpInst>(LoadU.getUser())) {
1816 if (!CmpI->isEquality() || !CmpI->isTrueWhenEqual())
1817 continue;
1818 for (const Use &CmpU : CmpI->uses()) {
1819 if (auto *IntrI = dyn_cast<IntrinsicInst>(CmpU.getUser())) {
1820 if (!IsValidAssume(*IntrI))
1821 continue;
1822 int Idx = CmpI->getOperandUse(0) == LoadU;
1823 Assumption = {CmpI->getOperand(Idx), IntrI};
1824 break;
1825 }
1826 }
1827 }
1828 if (Assumption.first)
1829 break;
1830 }
1831
1832 // Check if we found an assumption associated with this load.
1833 if (!Assumption.first || !Assumption.second)
1834 return true;
1835
1836 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Assumption found "
1837 << *Assumption.second << ": " << *LoadI
1838 << " == " << *Assumption.first << "\n");
1839 bool UsedAssumedInformation = false;
1840 std::optional<Value *> Content = nullptr;
1841 if (Assumption.first)
1842 Content =
1843 A.getAssumedSimplified(*Assumption.first, *this,
1844 UsedAssumedInformation, AA::Interprocedural);
1845 return handleAccess(
1846 A, *Assumption.second, Content, AccessKind::AK_ASSUMPTION,
1847 OffsetInfoMap[CurPtr].Offsets, Changed, *LoadI->getType());
1848 }
1849
1850 auto HandleStoreLike = [&](Instruction &I, Value *ValueOp, Type &ValueTy,
1851 ArrayRef<Value *> OtherOps, AccessKind AK) {
1852 for (auto *OtherOp : OtherOps) {
1853 if (OtherOp == CurPtr) {
1854 LLVM_DEBUG(
1855 dbgs()
1856 << "[AAPointerInfo] Escaping use in store like instruction " << I
1857 << "\n");
1858 return false;
1859 }
1860 }
1861
1862 // If the access is to a pointer that may or may not be the associated
1863 // value, e.g. due to a PHI, we cannot assume it will be written.
1864 if (getUnderlyingObject(CurPtr) == &AssociatedValue)
1865 AK = AccessKind(AK | AccessKind::AK_MUST);
1866 else
1867 AK = AccessKind(AK | AccessKind::AK_MAY);
1868 bool UsedAssumedInformation = false;
1869 std::optional<Value *> Content = nullptr;
1870 if (ValueOp)
1871 Content = A.getAssumedSimplified(
1872 *ValueOp, *this, UsedAssumedInformation, AA::Interprocedural);
1873 return handleAccess(A, I, Content, AK, OffsetInfoMap[CurPtr].Offsets,
1874 Changed, ValueTy);
1875 };
1876
1877 if (auto *StoreI = dyn_cast<StoreInst>(Usr))
1878 return HandleStoreLike(*StoreI, StoreI->getValueOperand(),
1879 *StoreI->getValueOperand()->getType(),
1880 {StoreI->getValueOperand()}, AccessKind::AK_W);
1881 if (auto *RMWI = dyn_cast<AtomicRMWInst>(Usr))
1882 return HandleStoreLike(*RMWI, nullptr, *RMWI->getValOperand()->getType(),
1883 {RMWI->getValOperand()}, AccessKind::AK_RW);
1884 if (auto *CXI = dyn_cast<AtomicCmpXchgInst>(Usr))
1885 return HandleStoreLike(
1886 *CXI, nullptr, *CXI->getNewValOperand()->getType(),
1887 {CXI->getCompareOperand(), CXI->getNewValOperand()},
1888 AccessKind::AK_RW);
1889
1890 if (auto *CB = dyn_cast<CallBase>(Usr)) {
1891 if (CB->isLifetimeStartOrEnd())
1892 return true;
1893 const auto *TLI =
1894 A.getInfoCache().getTargetLibraryInfoForFunction(*CB->getFunction());
1895 if (getFreedOperand(CB, TLI) == U)
1896 return true;
1897 if (CB->isArgOperand(&U)) {
1898 unsigned ArgNo = CB->getArgOperandNo(&U);
1899 const auto *CSArgPI = A.getAAFor<AAPointerInfo>(
1900 *this, IRPosition::callsite_argument(*CB, ArgNo),
1902 if (!CSArgPI)
1903 return false;
1904 bool IsArgMustAcc = (getUnderlyingObject(CurPtr) == &AssociatedValue);
1905 Changed = translateAndAddState(A, *CSArgPI, OffsetInfoMap[CurPtr], *CB,
1906 IsArgMustAcc) |
1907 Changed;
1908 if (!CSArgPI->reachesReturn())
1909 return isValidState();
1910
1912 if (!Callee || Callee->arg_size() <= ArgNo)
1913 return false;
1914 bool UsedAssumedInformation = false;
1915 auto ReturnedValue = A.getAssumedSimplified(
1916 IRPosition::returned(*Callee), *this, UsedAssumedInformation,
1918 auto *ReturnedArg =
1919 dyn_cast_or_null<Argument>(ReturnedValue.value_or(nullptr));
1920 auto *Arg = Callee->getArg(ArgNo);
1921 if (ReturnedArg && Arg != ReturnedArg)
1922 return true;
1923 bool IsRetMustAcc = IsArgMustAcc && (ReturnedArg == Arg);
1924 const auto *CSRetPI = A.getAAFor<AAPointerInfo>(
1926 if (!CSRetPI)
1927 return false;
1928 OffsetInfo OI = OffsetInfoMap[CurPtr];
1929 CSArgPI->addReturnedOffsetsTo(OI);
1930 Changed =
1931 translateAndAddState(A, *CSRetPI, OI, *CB, IsRetMustAcc) | Changed;
1932 return isValidState();
1933 }
1934 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Call user not handled " << *CB
1935 << "\n");
1936 return false;
1937 }
1938
1939 LLVM_DEBUG(dbgs() << "[AAPointerInfo] User not handled " << *Usr << "\n");
1940 return false;
1941 };
1942 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
1943 assert(OffsetInfoMap.count(OldU) && "Old use should be known already!");
1944 assert(!OffsetInfoMap[OldU].isUnassigned() && "Old use should be assinged");
1945 if (OffsetInfoMap.count(NewU)) {
1946 LLVM_DEBUG({
1947 if (!(OffsetInfoMap[NewU] == OffsetInfoMap[OldU])) {
1948 dbgs() << "[AAPointerInfo] Equivalent use callback failed: "
1949 << OffsetInfoMap[NewU] << " vs " << OffsetInfoMap[OldU]
1950 << "\n";
1951 }
1952 });
1953 return OffsetInfoMap[NewU] == OffsetInfoMap[OldU];
1954 }
1955 bool Unused;
1956 return HandlePassthroughUser(NewU.get(), OldU.get(), Unused);
1957 };
1958 if (!A.checkForAllUses(UsePred, *this, AssociatedValue,
1959 /* CheckBBLivenessOnly */ true, DepClassTy::OPTIONAL,
1960 /* IgnoreDroppableUses */ true, EquivalentUseCB)) {
1961 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Check for all uses failed, abort!\n");
1962 return indicatePessimisticFixpoint();
1963 }
1964
1965 LLVM_DEBUG({
1966 dbgs() << "Accesses by bin after update:\n";
1967 dumpState(dbgs());
1968 });
1969
1970 return Changed;
1971}
1972
1973struct AAPointerInfoReturned final : AAPointerInfoImpl {
1974 AAPointerInfoReturned(const IRPosition &IRP, Attributor &A)
1975 : AAPointerInfoImpl(IRP, A) {}
1976
1977 /// See AbstractAttribute::updateImpl(...).
1978 ChangeStatus updateImpl(Attributor &A) override {
1979 return indicatePessimisticFixpoint();
1980 }
1981
1982 /// See AbstractAttribute::trackStatistics()
1983 void trackStatistics() const override {
1984 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1985 }
1986};
1987
1988struct AAPointerInfoArgument final : AAPointerInfoFloating {
1989 AAPointerInfoArgument(const IRPosition &IRP, Attributor &A)
1990 : AAPointerInfoFloating(IRP, A) {}
1991
1992 /// See AbstractAttribute::trackStatistics()
1993 void trackStatistics() const override {
1994 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1995 }
1996};
1997
1998struct AAPointerInfoCallSiteArgument final : AAPointerInfoFloating {
1999 AAPointerInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
2000 : AAPointerInfoFloating(IRP, A) {}
2001
2002 /// See AbstractAttribute::updateImpl(...).
2003 ChangeStatus updateImpl(Attributor &A) override {
2004 using namespace AA::PointerInfo;
2005 // We handle memory intrinsics explicitly, at least the first (=
2006 // destination) and second (=source) arguments as we know how they are
2007 // accessed.
2008 if (auto *MI = dyn_cast_or_null<MemIntrinsic>(getCtxI())) {
2009 int64_t LengthVal = AA::RangeTy::Unknown;
2010 if (auto Length = MI->getLengthInBytes())
2011 LengthVal = Length->getSExtValue();
2012 unsigned ArgNo = getIRPosition().getCallSiteArgNo();
2013 ChangeStatus Changed = ChangeStatus::UNCHANGED;
2014 if (ArgNo > 1) {
2015 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Unhandled memory intrinsic "
2016 << *MI << "\n");
2017 return indicatePessimisticFixpoint();
2018 } else {
2019 auto Kind =
2020 ArgNo == 0 ? AccessKind::AK_MUST_WRITE : AccessKind::AK_MUST_READ;
2021 Changed =
2022 Changed | addAccess(A, {0, LengthVal}, *MI, nullptr, Kind, nullptr);
2023 }
2024 LLVM_DEBUG({
2025 dbgs() << "Accesses by bin after update:\n";
2026 dumpState(dbgs());
2027 });
2028
2029 return Changed;
2030 }
2031
2032 // TODO: Once we have call site specific value information we can provide
2033 // call site specific liveness information and then it makes
2034 // sense to specialize attributes for call sites arguments instead of
2035 // redirecting requests to the callee argument.
2036 Argument *Arg = getAssociatedArgument();
2037 if (Arg) {
2038 const IRPosition &ArgPos = IRPosition::argument(*Arg);
2039 auto *ArgAA =
2040 A.getAAFor<AAPointerInfo>(*this, ArgPos, DepClassTy::REQUIRED);
2041 if (ArgAA && ArgAA->getState().isValidState())
2042 return translateAndAddStateFromCallee(A, *ArgAA,
2043 *cast<CallBase>(getCtxI()));
2044 if (!Arg->getParent()->isDeclaration())
2045 return indicatePessimisticFixpoint();
2046 }
2047
2048 bool IsKnownNoCapture;
2050 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNoCapture))
2051 return indicatePessimisticFixpoint();
2052
2053 bool IsKnown = false;
2054 if (AA::isAssumedReadNone(A, getIRPosition(), *this, IsKnown))
2055 return ChangeStatus::UNCHANGED;
2056 bool ReadOnly = AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown);
2057 auto Kind =
2058 ReadOnly ? AccessKind::AK_MAY_READ : AccessKind::AK_MAY_READ_WRITE;
2059 return addAccess(A, AA::RangeTy::getUnknown(), *getCtxI(), nullptr, Kind,
2060 nullptr);
2061 }
2062
2063 /// See AbstractAttribute::trackStatistics()
2064 void trackStatistics() const override {
2065 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
2066 }
2067};
2068
2069struct AAPointerInfoCallSiteReturned final : AAPointerInfoFloating {
2070 AAPointerInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
2071 : AAPointerInfoFloating(IRP, A) {}
2072
2073 /// See AbstractAttribute::trackStatistics()
2074 void trackStatistics() const override {
2075 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
2076 }
2077};
2078} // namespace
2079
2080/// -----------------------NoUnwind Function Attribute--------------------------
2081
2082namespace {
2083struct AANoUnwindImpl : AANoUnwind {
2084 AANoUnwindImpl(const IRPosition &IRP, Attributor &A) : AANoUnwind(IRP, A) {}
2085
2086 /// See AbstractAttribute::initialize(...).
2087 void initialize(Attributor &A) override {
2088 bool IsKnown;
2090 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2091 (void)IsKnown;
2092 }
2093
2094 const std::string getAsStr(Attributor *A) const override {
2095 return getAssumed() ? "nounwind" : "may-unwind";
2096 }
2097
2098 /// See AbstractAttribute::updateImpl(...).
2099 ChangeStatus updateImpl(Attributor &A) override {
2100 auto Opcodes = {
2101 (unsigned)Instruction::Invoke, (unsigned)Instruction::CallBr,
2102 (unsigned)Instruction::Call, (unsigned)Instruction::CleanupRet,
2103 (unsigned)Instruction::CatchSwitch, (unsigned)Instruction::Resume};
2104
2105 auto CheckForNoUnwind = [&](Instruction &I) {
2106 if (!I.mayThrow(/* IncludePhaseOneUnwind */ true))
2107 return true;
2108
2109 if (const auto *CB = dyn_cast<CallBase>(&I)) {
2110 bool IsKnownNoUnwind;
2112 A, this, IRPosition::callsite_function(*CB), DepClassTy::REQUIRED,
2113 IsKnownNoUnwind);
2114 }
2115 return false;
2116 };
2117
2118 bool UsedAssumedInformation = false;
2119 if (!A.checkForAllInstructions(CheckForNoUnwind, *this, Opcodes,
2120 UsedAssumedInformation))
2121 return indicatePessimisticFixpoint();
2122
2123 return ChangeStatus::UNCHANGED;
2124 }
2125};
2126
2127struct AANoUnwindFunction final : public AANoUnwindImpl {
2128 AANoUnwindFunction(const IRPosition &IRP, Attributor &A)
2129 : AANoUnwindImpl(IRP, A) {}
2130
2131 /// See AbstractAttribute::trackStatistics()
2132 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nounwind) }
2133};
2134
2135/// NoUnwind attribute deduction for a call sites.
2136struct AANoUnwindCallSite final
2137 : AACalleeToCallSite<AANoUnwind, AANoUnwindImpl> {
2138 AANoUnwindCallSite(const IRPosition &IRP, Attributor &A)
2139 : AACalleeToCallSite<AANoUnwind, AANoUnwindImpl>(IRP, A) {}
2140
2141 /// See AbstractAttribute::trackStatistics()
2142 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nounwind); }
2143};
2144} // namespace
2145
2146/// ------------------------ NoSync Function Attribute -------------------------
2147
2148bool AANoSync::isAlignedBarrier(const CallBase &CB, bool ExecutedAligned) {
2149 switch (CB.getIntrinsicID()) {
2150 case Intrinsic::nvvm_barrier_cta_sync_aligned_all:
2151 case Intrinsic::nvvm_barrier_cta_sync_aligned_count:
2152 case Intrinsic::nvvm_barrier_cta_red_and_aligned_all:
2153 case Intrinsic::nvvm_barrier_cta_red_and_aligned_count:
2154 case Intrinsic::nvvm_barrier_cta_red_or_aligned_all:
2155 case Intrinsic::nvvm_barrier_cta_red_or_aligned_count:
2156 case Intrinsic::nvvm_barrier_cta_red_popc_aligned_all:
2157 case Intrinsic::nvvm_barrier_cta_red_popc_aligned_count:
2158 return true;
2159 case Intrinsic::amdgcn_s_barrier:
2160 if (ExecutedAligned)
2161 return true;
2162 break;
2163 default:
2164 break;
2165 }
2166 return hasAssumption(CB, KnownAssumptionString("ompx_aligned_barrier"));
2167}
2168
2170 if (!I->isAtomic())
2171 return false;
2172
2173 if (auto *FI = dyn_cast<FenceInst>(I))
2174 // All legal orderings for fence are stronger than monotonic.
2175 return FI->getSyncScopeID() != SyncScope::SingleThread;
2176 if (auto *AI = dyn_cast<AtomicCmpXchgInst>(I)) {
2177 // Unordered is not a legal ordering for cmpxchg.
2178 return (AI->getSuccessOrdering() != AtomicOrdering::Monotonic ||
2179 AI->getFailureOrdering() != AtomicOrdering::Monotonic);
2180 }
2181
2182 AtomicOrdering Ordering;
2183 switch (I->getOpcode()) {
2184 case Instruction::AtomicRMW:
2185 Ordering = cast<AtomicRMWInst>(I)->getOrdering();
2186 break;
2187 case Instruction::Store:
2188 Ordering = cast<StoreInst>(I)->getOrdering();
2189 break;
2190 case Instruction::Load:
2191 Ordering = cast<LoadInst>(I)->getOrdering();
2192 break;
2193 default:
2195 "New atomic operations need to be known in the attributor.");
2196 }
2197
2198 return (Ordering != AtomicOrdering::Unordered &&
2199 Ordering != AtomicOrdering::Monotonic);
2200}
2201
2202/// Return true if this intrinsic is nosync. This is only used for intrinsics
2203/// which would be nosync except that they have a volatile flag. All other
2204/// intrinsics are simply annotated with the nosync attribute in Intrinsics.td.
2206 if (auto *MI = dyn_cast<MemIntrinsic>(I))
2207 return !MI->isVolatile();
2208 return false;
2209}
2210
2211namespace {
2212struct AANoSyncImpl : AANoSync {
2213 AANoSyncImpl(const IRPosition &IRP, Attributor &A) : AANoSync(IRP, A) {}
2214
2215 /// See AbstractAttribute::initialize(...).
2216 void initialize(Attributor &A) override {
2217 bool IsKnown;
2218 assert(!AA::hasAssumedIRAttr<Attribute::NoSync>(A, nullptr, getIRPosition(),
2219 DepClassTy::NONE, IsKnown));
2220 (void)IsKnown;
2221 }
2222
2223 const std::string getAsStr(Attributor *A) const override {
2224 return getAssumed() ? "nosync" : "may-sync";
2225 }
2226
2227 /// See AbstractAttribute::updateImpl(...).
2228 ChangeStatus updateImpl(Attributor &A) override;
2229};
2230
2231ChangeStatus AANoSyncImpl::updateImpl(Attributor &A) {
2232
2233 auto CheckRWInstForNoSync = [&](Instruction &I) {
2234 return AA::isNoSyncInst(A, I, *this);
2235 };
2236
2237 auto CheckForNoSync = [&](Instruction &I) {
2238 // At this point we handled all read/write effects and they are all
2239 // nosync, so they can be skipped.
2240 if (I.mayReadOrWriteMemory())
2241 return true;
2242
2243 bool IsKnown;
2244 CallBase &CB = cast<CallBase>(I);
2247 IsKnown))
2248 return true;
2249
2250 // non-convergent and readnone imply nosync.
2251 return !CB.isConvergent();
2252 };
2253
2254 bool UsedAssumedInformation = false;
2255 if (!A.checkForAllReadWriteInstructions(CheckRWInstForNoSync, *this,
2256 UsedAssumedInformation) ||
2257 !A.checkForAllCallLikeInstructions(CheckForNoSync, *this,
2258 UsedAssumedInformation))
2259 return indicatePessimisticFixpoint();
2260
2262}
2263
2264struct AANoSyncFunction final : public AANoSyncImpl {
2265 AANoSyncFunction(const IRPosition &IRP, Attributor &A)
2266 : AANoSyncImpl(IRP, A) {}
2267
2268 /// See AbstractAttribute::trackStatistics()
2269 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nosync) }
2270};
2271
2272/// NoSync attribute deduction for a call sites.
2273struct AANoSyncCallSite final : AACalleeToCallSite<AANoSync, AANoSyncImpl> {
2274 AANoSyncCallSite(const IRPosition &IRP, Attributor &A)
2275 : AACalleeToCallSite<AANoSync, AANoSyncImpl>(IRP, A) {}
2276
2277 /// See AbstractAttribute::trackStatistics()
2278 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nosync); }
2279};
2280} // namespace
2281
2282/// ------------------------ No-Free Attributes ----------------------------
2283
2284namespace {
2285struct AANoFreeImpl : public AANoFree {
2286 AANoFreeImpl(const IRPosition &IRP, Attributor &A) : AANoFree(IRP, A) {}
2287
2288 /// See AbstractAttribute::initialize(...).
2289 void initialize(Attributor &A) override {
2290 bool IsKnown;
2291 assert(!AA::hasAssumedIRAttr<Attribute::NoFree>(A, nullptr, getIRPosition(),
2292 DepClassTy::NONE, IsKnown));
2293 (void)IsKnown;
2294 }
2295
2296 /// See AbstractAttribute::updateImpl(...).
2297 ChangeStatus updateImpl(Attributor &A) override {
2298 auto CheckForNoFree = [&](Instruction &I) {
2299 bool IsKnown;
2302 DepClassTy::REQUIRED, IsKnown);
2303 };
2304
2305 bool UsedAssumedInformation = false;
2306 if (!A.checkForAllCallLikeInstructions(CheckForNoFree, *this,
2307 UsedAssumedInformation))
2308 return indicatePessimisticFixpoint();
2309 return ChangeStatus::UNCHANGED;
2310 }
2311
2312 /// See AbstractAttribute::getAsStr().
2313 const std::string getAsStr(Attributor *A) const override {
2314 return getAssumed() ? "nofree" : "may-free";
2315 }
2316};
2317
2318struct AANoFreeFunction final : public AANoFreeImpl {
2319 AANoFreeFunction(const IRPosition &IRP, Attributor &A)
2320 : AANoFreeImpl(IRP, A) {}
2321
2322 /// See AbstractAttribute::trackStatistics()
2323 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nofree) }
2324};
2325
2326/// NoFree attribute deduction for a call sites.
2327struct AANoFreeCallSite final : AACalleeToCallSite<AANoFree, AANoFreeImpl> {
2328 AANoFreeCallSite(const IRPosition &IRP, Attributor &A)
2329 : AACalleeToCallSite<AANoFree, AANoFreeImpl>(IRP, A) {}
2330
2331 /// See AbstractAttribute::trackStatistics()
2332 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nofree); }
2333};
2334
2335/// NoFree attribute for floating values.
2336struct AANoFreeFloating : AANoFreeImpl {
2337 AANoFreeFloating(const IRPosition &IRP, Attributor &A)
2338 : AANoFreeImpl(IRP, A) {}
2339
2340 /// See AbstractAttribute::trackStatistics()
2341 void trackStatistics() const override{STATS_DECLTRACK_FLOATING_ATTR(nofree)}
2342
2343 /// See Abstract Attribute::updateImpl(...).
2344 ChangeStatus updateImpl(Attributor &A) override {
2345 const IRPosition &IRP = getIRPosition();
2346
2347 bool IsKnown;
2350 DepClassTy::OPTIONAL, IsKnown))
2351 return ChangeStatus::UNCHANGED;
2352
2353 Value &AssociatedValue = getIRPosition().getAssociatedValue();
2354 auto Pred = [&](const Use &U, bool &Follow) -> bool {
2355 Instruction *UserI = cast<Instruction>(U.getUser());
2356 if (auto *CB = dyn_cast<CallBase>(UserI)) {
2357 if (CB->isBundleOperand(&U))
2358 return false;
2359 if (!CB->isArgOperand(&U))
2360 return true;
2361 unsigned ArgNo = CB->getArgOperandNo(&U);
2362
2363 bool IsKnown;
2365 A, this, IRPosition::callsite_argument(*CB, ArgNo),
2366 DepClassTy::REQUIRED, IsKnown);
2367 }
2368
2369 if (isa<GetElementPtrInst>(UserI) || isa<PHINode>(UserI) ||
2370 isa<SelectInst>(UserI)) {
2371 Follow = true;
2372 return true;
2373 }
2374 if (isa<StoreInst>(UserI) || isa<LoadInst>(UserI))
2375 return true;
2376
2377 if (isa<ReturnInst>(UserI) && getIRPosition().isArgumentPosition())
2378 return true;
2379
2380 // Unknown user.
2381 return false;
2382 };
2383 if (!A.checkForAllUses(Pred, *this, AssociatedValue))
2384 return indicatePessimisticFixpoint();
2385
2386 return ChangeStatus::UNCHANGED;
2387 }
2388};
2389
2390/// NoFree attribute for a call site argument.
2391struct AANoFreeArgument final : AANoFreeFloating {
2392 AANoFreeArgument(const IRPosition &IRP, Attributor &A)
2393 : AANoFreeFloating(IRP, A) {}
2394
2395 /// See AbstractAttribute::trackStatistics()
2396 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nofree) }
2397};
2398
2399/// NoFree attribute for call site arguments.
2400struct AANoFreeCallSiteArgument final : AANoFreeFloating {
2401 AANoFreeCallSiteArgument(const IRPosition &IRP, Attributor &A)
2402 : AANoFreeFloating(IRP, A) {}
2403
2404 /// See AbstractAttribute::updateImpl(...).
2405 ChangeStatus updateImpl(Attributor &A) override {
2406 // TODO: Once we have call site specific value information we can provide
2407 // call site specific liveness information and then it makes
2408 // sense to specialize attributes for call sites arguments instead of
2409 // redirecting requests to the callee argument.
2410 Argument *Arg = getAssociatedArgument();
2411 if (!Arg)
2412 return indicatePessimisticFixpoint();
2413 const IRPosition &ArgPos = IRPosition::argument(*Arg);
2414 bool IsKnown;
2416 DepClassTy::REQUIRED, IsKnown))
2417 return ChangeStatus::UNCHANGED;
2418 return indicatePessimisticFixpoint();
2419 }
2420
2421 /// See AbstractAttribute::trackStatistics()
2422 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(nofree) };
2423};
2424
2425/// NoFree attribute for function return value.
2426struct AANoFreeReturned final : AANoFreeFloating {
2427 AANoFreeReturned(const IRPosition &IRP, Attributor &A)
2428 : AANoFreeFloating(IRP, A) {
2429 llvm_unreachable("NoFree is not applicable to function returns!");
2430 }
2431
2432 /// See AbstractAttribute::initialize(...).
2433 void initialize(Attributor &A) override {
2434 llvm_unreachable("NoFree is not applicable to function returns!");
2435 }
2436
2437 /// See AbstractAttribute::updateImpl(...).
2438 ChangeStatus updateImpl(Attributor &A) override {
2439 llvm_unreachable("NoFree is not applicable to function returns!");
2440 }
2441
2442 /// See AbstractAttribute::trackStatistics()
2443 void trackStatistics() const override {}
2444};
2445
2446/// NoFree attribute deduction for a call site return value.
2447struct AANoFreeCallSiteReturned final : AANoFreeFloating {
2448 AANoFreeCallSiteReturned(const IRPosition &IRP, Attributor &A)
2449 : AANoFreeFloating(IRP, A) {}
2450
2451 ChangeStatus manifest(Attributor &A) override {
2452 return ChangeStatus::UNCHANGED;
2453 }
2454 /// See AbstractAttribute::trackStatistics()
2455 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(nofree) }
2456};
2457} // namespace
2458
2459/// ------------------------ NonNull Argument Attribute ------------------------
2460
2462 Attribute::AttrKind ImpliedAttributeKind,
2463 bool IgnoreSubsumingPositions) {
2465 AttrKinds.push_back(Attribute::NonNull);
2468 AttrKinds.push_back(Attribute::Dereferenceable);
2469 if (A.hasAttr(IRP, AttrKinds, IgnoreSubsumingPositions, Attribute::NonNull))
2470 return true;
2471
2472 DominatorTree *DT = nullptr;
2473 AssumptionCache *AC = nullptr;
2474 InformationCache &InfoCache = A.getInfoCache();
2475 if (const Function *Fn = IRP.getAnchorScope()) {
2476 if (!Fn->isDeclaration()) {
2479 }
2480 }
2481
2483 if (IRP.getPositionKind() != IRP_RETURNED) {
2484 Worklist.push_back({IRP.getAssociatedValue(), IRP.getCtxI()});
2485 } else {
2486 bool UsedAssumedInformation = false;
2487 if (!A.checkForAllInstructions(
2488 [&](Instruction &I) {
2489 Worklist.push_back({*cast<ReturnInst>(I).getReturnValue(), &I});
2490 return true;
2491 },
2492 IRP.getAssociatedFunction(), nullptr, {Instruction::Ret},
2493 UsedAssumedInformation, false, /*CheckPotentiallyDead=*/true))
2494 return false;
2495 }
2496
2497 if (llvm::any_of(Worklist, [&](AA::ValueAndContext VAC) {
2498 return !isKnownNonZero(
2499 VAC.getValue(),
2500 SimplifyQuery(A.getDataLayout(), DT, AC, VAC.getCtxI()));
2501 }))
2502 return false;
2503
2504 A.manifestAttrs(IRP, {Attribute::get(IRP.getAnchorValue().getContext(),
2505 Attribute::NonNull)});
2506 return true;
2507}
2508
2509namespace {
2510static int64_t getKnownNonNullAndDerefBytesForUse(
2511 Attributor &A, const AbstractAttribute &QueryingAA, Value &AssociatedValue,
2512 const Use *U, const Instruction *I, bool &IsNonNull, bool &TrackUse) {
2513 TrackUse = false;
2514
2515 const Value *UseV = U->get();
2516 if (!UseV->getType()->isPointerTy())
2517 return 0;
2518
2519 // We need to follow common pointer manipulation uses to the accesses they
2520 // feed into. We can try to be smart to avoid looking through things we do not
2521 // like for now, e.g., non-inbounds GEPs.
2522 if (isa<CastInst>(I)) {
2523 TrackUse = true;
2524 return 0;
2525 }
2526
2528 TrackUse = true;
2529 return 0;
2530 }
2531
2532 Type *PtrTy = UseV->getType();
2533 const Function *F = I->getFunction();
2536 const DataLayout &DL = A.getInfoCache().getDL();
2537 if (const auto *CB = dyn_cast<CallBase>(I)) {
2538 if (CB->isBundleOperand(U)) {
2539 if (RetainedKnowledge RK = getKnowledgeFromUse(
2540 U, {Attribute::NonNull, Attribute::Dereferenceable})) {
2541 IsNonNull |=
2542 (RK.AttrKind == Attribute::NonNull || !NullPointerIsDefined);
2543 return RK.ArgValue;
2544 }
2545 return 0;
2546 }
2547
2548 if (CB->isCallee(U)) {
2549 IsNonNull |= !NullPointerIsDefined;
2550 return 0;
2551 }
2552
2553 unsigned ArgNo = CB->getArgOperandNo(U);
2554 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
2555 // As long as we only use known information there is no need to track
2556 // dependences here.
2557 bool IsKnownNonNull;
2559 DepClassTy::NONE, IsKnownNonNull);
2560 IsNonNull |= IsKnownNonNull;
2561 auto *DerefAA =
2562 A.getAAFor<AADereferenceable>(QueryingAA, IRP, DepClassTy::NONE);
2563 return DerefAA ? DerefAA->getKnownDereferenceableBytes() : 0;
2564 }
2565
2566 std::optional<MemoryLocation> Loc = MemoryLocation::getOrNone(I);
2567 if (!Loc || Loc->Ptr != UseV || !Loc->Size.isPrecise() ||
2568 Loc->Size.isScalable() || I->isVolatile())
2569 return 0;
2570
2571 int64_t Offset;
2572 const Value *Base =
2573 getMinimalBaseOfPointer(A, QueryingAA, Loc->Ptr, Offset, DL);
2574 if (Base && Base == &AssociatedValue) {
2575 int64_t DerefBytes = Loc->Size.getValue() + Offset;
2576 IsNonNull |= !NullPointerIsDefined;
2577 return std::max(int64_t(0), DerefBytes);
2578 }
2579
2580 /// Corner case when an offset is 0.
2582 /*AllowNonInbounds*/ true);
2583 if (Base && Base == &AssociatedValue && Offset == 0) {
2584 int64_t DerefBytes = Loc->Size.getValue();
2585 IsNonNull |= !NullPointerIsDefined;
2586 return std::max(int64_t(0), DerefBytes);
2587 }
2588
2589 return 0;
2590}
2591
2592struct AANonNullImpl : AANonNull {
2593 AANonNullImpl(const IRPosition &IRP, Attributor &A) : AANonNull(IRP, A) {}
2594
2595 /// See AbstractAttribute::initialize(...).
2596 void initialize(Attributor &A) override {
2597 Value &V = *getAssociatedValue().stripPointerCasts();
2598 if (isa<ConstantPointerNull>(V)) {
2599 indicatePessimisticFixpoint();
2600 return;
2601 }
2602
2603 if (Instruction *CtxI = getCtxI())
2604 followUsesInMBEC(*this, A, getState(), *CtxI);
2605 }
2606
2607 /// See followUsesInMBEC
2608 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
2609 AANonNull::StateType &State) {
2610 bool IsNonNull = false;
2611 bool TrackUse = false;
2612 getKnownNonNullAndDerefBytesForUse(A, *this, getAssociatedValue(), U, I,
2613 IsNonNull, TrackUse);
2614 State.setKnown(IsNonNull);
2615 return TrackUse;
2616 }
2617
2618 /// See AbstractAttribute::getAsStr().
2619 const std::string getAsStr(Attributor *A) const override {
2620 return getAssumed() ? "nonnull" : "may-null";
2621 }
2622};
2623
2624/// NonNull attribute for a floating value.
2625struct AANonNullFloating : public AANonNullImpl {
2626 AANonNullFloating(const IRPosition &IRP, Attributor &A)
2627 : AANonNullImpl(IRP, A) {}
2628
2629 /// See AbstractAttribute::updateImpl(...).
2630 ChangeStatus updateImpl(Attributor &A) override {
2631 auto CheckIRP = [&](const IRPosition &IRP) {
2632 bool IsKnownNonNull;
2634 A, *this, IRP, DepClassTy::OPTIONAL, IsKnownNonNull);
2635 };
2636
2637 bool Stripped;
2638 bool UsedAssumedInformation = false;
2639 Value *AssociatedValue = &getAssociatedValue();
2641 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
2642 AA::AnyScope, UsedAssumedInformation))
2643 Stripped = false;
2644 else
2645 Stripped =
2646 Values.size() != 1 || Values.front().getValue() != AssociatedValue;
2647
2648 if (!Stripped) {
2649 bool IsKnown;
2650 if (auto *PHI = dyn_cast<PHINode>(AssociatedValue))
2651 if (llvm::all_of(PHI->incoming_values(), [&](Value *Op) {
2652 return AA::hasAssumedIRAttr<Attribute::NonNull>(
2653 A, this, IRPosition::value(*Op), DepClassTy::OPTIONAL,
2654 IsKnown);
2655 }))
2656 return ChangeStatus::UNCHANGED;
2657 if (auto *Select = dyn_cast<SelectInst>(AssociatedValue))
2659 A, this, IRPosition::value(*Select->getFalseValue()),
2660 DepClassTy::OPTIONAL, IsKnown) &&
2662 A, this, IRPosition::value(*Select->getTrueValue()),
2663 DepClassTy::OPTIONAL, IsKnown))
2664 return ChangeStatus::UNCHANGED;
2665
2666 // If we haven't stripped anything we might still be able to use a
2667 // different AA, but only if the IRP changes. Effectively when we
2668 // interpret this not as a call site value but as a floating/argument
2669 // value.
2670 const IRPosition AVIRP = IRPosition::value(*AssociatedValue);
2671 if (AVIRP == getIRPosition() || !CheckIRP(AVIRP))
2672 return indicatePessimisticFixpoint();
2673 return ChangeStatus::UNCHANGED;
2674 }
2675
2676 for (const auto &VAC : Values)
2677 if (!CheckIRP(IRPosition::value(*VAC.getValue())))
2678 return indicatePessimisticFixpoint();
2679
2680 return ChangeStatus::UNCHANGED;
2681 }
2682
2683 /// See AbstractAttribute::trackStatistics()
2684 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(nonnull) }
2685};
2686
2687/// NonNull attribute for function return value.
2688struct AANonNullReturned final
2689 : AAReturnedFromReturnedValues<AANonNull, AANonNull, AANonNull::StateType,
2690 false, AANonNull::IRAttributeKind, false> {
2691 AANonNullReturned(const IRPosition &IRP, Attributor &A)
2692 : AAReturnedFromReturnedValues<AANonNull, AANonNull, AANonNull::StateType,
2693 false, Attribute::NonNull, false>(IRP, A) {
2694 }
2695
2696 /// See AbstractAttribute::getAsStr().
2697 const std::string getAsStr(Attributor *A) const override {
2698 return getAssumed() ? "nonnull" : "may-null";
2699 }
2700
2701 /// See AbstractAttribute::trackStatistics()
2702 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(nonnull) }
2703};
2704
2705/// NonNull attribute for function argument.
2706struct AANonNullArgument final
2707 : AAArgumentFromCallSiteArguments<AANonNull, AANonNullImpl> {
2708 AANonNullArgument(const IRPosition &IRP, Attributor &A)
2709 : AAArgumentFromCallSiteArguments<AANonNull, AANonNullImpl>(IRP, A) {}
2710
2711 /// See AbstractAttribute::trackStatistics()
2712 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nonnull) }
2713};
2714
2715struct AANonNullCallSiteArgument final : AANonNullFloating {
2716 AANonNullCallSiteArgument(const IRPosition &IRP, Attributor &A)
2717 : AANonNullFloating(IRP, A) {}
2718
2719 /// See AbstractAttribute::trackStatistics()
2720 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(nonnull) }
2721};
2722
2723/// NonNull attribute for a call site return position.
2724struct AANonNullCallSiteReturned final
2725 : AACalleeToCallSite<AANonNull, AANonNullImpl> {
2726 AANonNullCallSiteReturned(const IRPosition &IRP, Attributor &A)
2727 : AACalleeToCallSite<AANonNull, AANonNullImpl>(IRP, A) {}
2728
2729 /// See AbstractAttribute::trackStatistics()
2730 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(nonnull) }
2731};
2732} // namespace
2733
2734/// ------------------------ Must-Progress Attributes --------------------------
2735namespace {
2736struct AAMustProgressImpl : public AAMustProgress {
2737 AAMustProgressImpl(const IRPosition &IRP, Attributor &A)
2738 : AAMustProgress(IRP, A) {}
2739
2740 /// See AbstractAttribute::initialize(...).
2741 void initialize(Attributor &A) override {
2742 bool IsKnown;
2744 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2745 (void)IsKnown;
2746 }
2747
2748 /// See AbstractAttribute::getAsStr()
2749 const std::string getAsStr(Attributor *A) const override {
2750 return getAssumed() ? "mustprogress" : "may-not-progress";
2751 }
2752};
2753
2754struct AAMustProgressFunction final : AAMustProgressImpl {
2755 AAMustProgressFunction(const IRPosition &IRP, Attributor &A)
2756 : AAMustProgressImpl(IRP, A) {}
2757
2758 /// See AbstractAttribute::updateImpl(...).
2759 ChangeStatus updateImpl(Attributor &A) override {
2760 bool IsKnown;
2762 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnown)) {
2763 if (IsKnown)
2764 return indicateOptimisticFixpoint();
2765 return ChangeStatus::UNCHANGED;
2766 }
2767
2768 auto CheckForMustProgress = [&](AbstractCallSite ACS) {
2769 IRPosition IPos = IRPosition::callsite_function(*ACS.getInstruction());
2770 bool IsKnownMustProgress;
2772 A, this, IPos, DepClassTy::REQUIRED, IsKnownMustProgress,
2773 /* IgnoreSubsumingPositions */ true);
2774 };
2775
2776 bool AllCallSitesKnown = true;
2777 if (!A.checkForAllCallSites(CheckForMustProgress, *this,
2778 /* RequireAllCallSites */ true,
2779 AllCallSitesKnown))
2780 return indicatePessimisticFixpoint();
2781
2782 return ChangeStatus::UNCHANGED;
2783 }
2784
2785 /// See AbstractAttribute::trackStatistics()
2786 void trackStatistics() const override {
2787 STATS_DECLTRACK_FN_ATTR(mustprogress)
2788 }
2789};
2790
2791/// MustProgress attribute deduction for a call sites.
2792struct AAMustProgressCallSite final : AAMustProgressImpl {
2793 AAMustProgressCallSite(const IRPosition &IRP, Attributor &A)
2794 : AAMustProgressImpl(IRP, A) {}
2795
2796 /// See AbstractAttribute::updateImpl(...).
2797 ChangeStatus updateImpl(Attributor &A) override {
2798 // TODO: Once we have call site specific value information we can provide
2799 // call site specific liveness information and then it makes
2800 // sense to specialize attributes for call sites arguments instead of
2801 // redirecting requests to the callee argument.
2802 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
2803 bool IsKnownMustProgress;
2805 A, this, FnPos, DepClassTy::REQUIRED, IsKnownMustProgress))
2806 return indicatePessimisticFixpoint();
2807 return ChangeStatus::UNCHANGED;
2808 }
2809
2810 /// See AbstractAttribute::trackStatistics()
2811 void trackStatistics() const override {
2812 STATS_DECLTRACK_CS_ATTR(mustprogress);
2813 }
2814};
2815} // namespace
2816
2817/// ------------------------ No-Recurse Attributes ----------------------------
2818
2819namespace {
2820struct AANoRecurseImpl : public AANoRecurse {
2821 AANoRecurseImpl(const IRPosition &IRP, Attributor &A) : AANoRecurse(IRP, A) {}
2822
2823 /// See AbstractAttribute::initialize(...).
2824 void initialize(Attributor &A) override {
2825 bool IsKnown;
2827 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2828 (void)IsKnown;
2829 }
2830
2831 /// See AbstractAttribute::getAsStr()
2832 const std::string getAsStr(Attributor *A) const override {
2833 return getAssumed() ? "norecurse" : "may-recurse";
2834 }
2835};
2836
2837struct AANoRecurseFunction final : AANoRecurseImpl {
2838 AANoRecurseFunction(const IRPosition &IRP, Attributor &A)
2839 : AANoRecurseImpl(IRP, A) {}
2840
2841 /// See AbstractAttribute::updateImpl(...).
2842 ChangeStatus updateImpl(Attributor &A) override {
2843
2844 // If all live call sites are known to be no-recurse, we are as well.
2845 auto CallSitePred = [&](AbstractCallSite ACS) {
2846 bool IsKnownNoRecurse;
2848 A, this,
2849 IRPosition::function(*ACS.getInstruction()->getFunction()),
2850 DepClassTy::NONE, IsKnownNoRecurse))
2851 return false;
2852 return IsKnownNoRecurse;
2853 };
2854 bool UsedAssumedInformation = false;
2855 if (A.checkForAllCallSites(CallSitePred, *this, true,
2856 UsedAssumedInformation)) {
2857 // If we know all call sites and all are known no-recurse, we are done.
2858 // If all known call sites, which might not be all that exist, are known
2859 // to be no-recurse, we are not done but we can continue to assume
2860 // no-recurse. If one of the call sites we have not visited will become
2861 // live, another update is triggered.
2862 if (!UsedAssumedInformation)
2863 indicateOptimisticFixpoint();
2864 return ChangeStatus::UNCHANGED;
2865 }
2866
2867 const AAInterFnReachability *EdgeReachability =
2868 A.getAAFor<AAInterFnReachability>(*this, getIRPosition(),
2869 DepClassTy::REQUIRED);
2870 if (EdgeReachability && EdgeReachability->canReach(A, *getAnchorScope()))
2871 return indicatePessimisticFixpoint();
2872 return ChangeStatus::UNCHANGED;
2873 }
2874
2875 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(norecurse) }
2876};
2877
2878/// NoRecurse attribute deduction for a call sites.
2879struct AANoRecurseCallSite final
2880 : AACalleeToCallSite<AANoRecurse, AANoRecurseImpl> {
2881 AANoRecurseCallSite(const IRPosition &IRP, Attributor &A)
2882 : AACalleeToCallSite<AANoRecurse, AANoRecurseImpl>(IRP, A) {}
2883
2884 /// See AbstractAttribute::trackStatistics()
2885 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(norecurse); }
2886};
2887} // namespace
2888
2889/// ------------------------ No-Convergent Attribute --------------------------
2890
2891namespace {
2892struct AANonConvergentImpl : public AANonConvergent {
2893 AANonConvergentImpl(const IRPosition &IRP, Attributor &A)
2894 : AANonConvergent(IRP, A) {}
2895
2896 /// See AbstractAttribute::getAsStr()
2897 const std::string getAsStr(Attributor *A) const override {
2898 return getAssumed() ? "non-convergent" : "may-be-convergent";
2899 }
2900};
2901
2902struct AANonConvergentFunction final : AANonConvergentImpl {
2903 AANonConvergentFunction(const IRPosition &IRP, Attributor &A)
2904 : AANonConvergentImpl(IRP, A) {}
2905
2906 /// See AbstractAttribute::updateImpl(...).
2907 ChangeStatus updateImpl(Attributor &A) override {
2908 // If all function calls are known to not be convergent, we are not
2909 // convergent.
2910 auto CalleeIsNotConvergent = [&](Instruction &Inst) {
2911 CallBase &CB = cast<CallBase>(Inst);
2913 if (!Callee || Callee->isIntrinsic()) {
2914 return false;
2915 }
2916 if (Callee->isDeclaration()) {
2917 return !Callee->hasFnAttribute(Attribute::Convergent);
2918 }
2919 const auto *ConvergentAA = A.getAAFor<AANonConvergent>(
2920 *this, IRPosition::function(*Callee), DepClassTy::REQUIRED);
2921 return ConvergentAA && ConvergentAA->isAssumedNotConvergent();
2922 };
2923
2924 bool UsedAssumedInformation = false;
2925 if (!A.checkForAllCallLikeInstructions(CalleeIsNotConvergent, *this,
2926 UsedAssumedInformation)) {
2927 return indicatePessimisticFixpoint();
2928 }
2929 return ChangeStatus::UNCHANGED;
2930 }
2931
2932 ChangeStatus manifest(Attributor &A) override {
2933 if (isKnownNotConvergent() &&
2934 A.hasAttr(getIRPosition(), Attribute::Convergent)) {
2935 A.removeAttrs(getIRPosition(), {Attribute::Convergent});
2936 return ChangeStatus::CHANGED;
2937 }
2938 return ChangeStatus::UNCHANGED;
2939 }
2940
2941 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(convergent) }
2942};
2943} // namespace
2944
2945/// -------------------- Undefined-Behavior Attributes ------------------------
2946
2947namespace {
2948struct AAUndefinedBehaviorImpl : public AAUndefinedBehavior {
2949 AAUndefinedBehaviorImpl(const IRPosition &IRP, Attributor &A)
2950 : AAUndefinedBehavior(IRP, A) {}
2951
2952 /// See AbstractAttribute::updateImpl(...).
2953 // through a pointer (i.e. also branches etc.)
2954 ChangeStatus updateImpl(Attributor &A) override {
2955 const size_t UBPrevSize = KnownUBInsts.size();
2956 const size_t NoUBPrevSize = AssumedNoUBInsts.size();
2957
2958 auto InspectMemAccessInstForUB = [&](Instruction &I) {
2959 // Lang ref now states volatile store is not UB, let's skip them.
2960 if (I.isVolatile() && I.mayWriteToMemory())
2961 return true;
2962
2963 // Skip instructions that are already saved.
2964 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
2965 return true;
2966
2967 // If we reach here, we know we have an instruction
2968 // that accesses memory through a pointer operand,
2969 // for which getPointerOperand() should give it to us.
2970 Value *PtrOp =
2971 const_cast<Value *>(getPointerOperand(&I, /* AllowVolatile */ true));
2972 assert(PtrOp &&
2973 "Expected pointer operand of memory accessing instruction");
2974
2975 // Either we stopped and the appropriate action was taken,
2976 // or we got back a simplified value to continue.
2977 std::optional<Value *> SimplifiedPtrOp =
2978 stopOnUndefOrAssumed(A, PtrOp, &I);
2979 if (!SimplifiedPtrOp || !*SimplifiedPtrOp)
2980 return true;
2981 const Value *PtrOpVal = *SimplifiedPtrOp;
2982
2983 // A memory access through a pointer is considered UB
2984 // only if the pointer has constant null value.
2985 // TODO: Expand it to not only check constant values.
2986 if (!isa<ConstantPointerNull>(PtrOpVal)) {
2987 AssumedNoUBInsts.insert(&I);
2988 return true;
2989 }
2990 const Type *PtrTy = PtrOpVal->getType();
2991
2992 // Because we only consider instructions inside functions,
2993 // assume that a parent function exists.
2994 const Function *F = I.getFunction();
2995
2996 // A memory access using constant null pointer is only considered UB
2997 // if null pointer is _not_ defined for the target platform.
2999 AssumedNoUBInsts.insert(&I);
3000 else
3001 KnownUBInsts.insert(&I);
3002 return true;
3003 };
3004
3005 auto InspectBrInstForUB = [&](Instruction &I) {
3006 // A conditional branch instruction is considered UB if it has `undef`
3007 // condition.
3008
3009 // Skip instructions that are already saved.
3010 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
3011 return true;
3012
3013 // We know we have a branch instruction.
3014 auto *BrInst = cast<BranchInst>(&I);
3015
3016 // Unconditional branches are never considered UB.
3017 if (BrInst->isUnconditional())
3018 return true;
3019
3020 // Either we stopped and the appropriate action was taken,
3021 // or we got back a simplified value to continue.
3022 std::optional<Value *> SimplifiedCond =
3023 stopOnUndefOrAssumed(A, BrInst->getCondition(), BrInst);
3024 if (!SimplifiedCond || !*SimplifiedCond)
3025 return true;
3026 AssumedNoUBInsts.insert(&I);
3027 return true;
3028 };
3029
3030 auto InspectCallSiteForUB = [&](Instruction &I) {
3031 // Check whether a callsite always cause UB or not
3032
3033 // Skip instructions that are already saved.
3034 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
3035 return true;
3036
3037 // Check nonnull and noundef argument attribute violation for each
3038 // callsite.
3039 CallBase &CB = cast<CallBase>(I);
3041 if (!Callee)
3042 return true;
3043 for (unsigned idx = 0; idx < CB.arg_size(); idx++) {
3044 // If current argument is known to be simplified to null pointer and the
3045 // corresponding argument position is known to have nonnull attribute,
3046 // the argument is poison. Furthermore, if the argument is poison and
3047 // the position is known to have noundef attriubte, this callsite is
3048 // considered UB.
3049 if (idx >= Callee->arg_size())
3050 break;
3051 Value *ArgVal = CB.getArgOperand(idx);
3052 if (!ArgVal)
3053 continue;
3054 // Here, we handle three cases.
3055 // (1) Not having a value means it is dead. (we can replace the value
3056 // with undef)
3057 // (2) Simplified to undef. The argument violate noundef attriubte.
3058 // (3) Simplified to null pointer where known to be nonnull.
3059 // The argument is a poison value and violate noundef attribute.
3060 IRPosition CalleeArgumentIRP = IRPosition::callsite_argument(CB, idx);
3061 bool IsKnownNoUndef;
3063 A, this, CalleeArgumentIRP, DepClassTy::NONE, IsKnownNoUndef);
3064 if (!IsKnownNoUndef)
3065 continue;
3066 bool UsedAssumedInformation = false;
3067 std::optional<Value *> SimplifiedVal =
3068 A.getAssumedSimplified(IRPosition::value(*ArgVal), *this,
3069 UsedAssumedInformation, AA::Interprocedural);
3070 if (UsedAssumedInformation)
3071 continue;
3072 if (SimplifiedVal && !*SimplifiedVal)
3073 return true;
3074 if (!SimplifiedVal || isa<UndefValue>(**SimplifiedVal)) {
3075 KnownUBInsts.insert(&I);
3076 continue;
3077 }
3078 if (!ArgVal->getType()->isPointerTy() ||
3079 !isa<ConstantPointerNull>(**SimplifiedVal))
3080 continue;
3081 bool IsKnownNonNull;
3083 A, this, CalleeArgumentIRP, DepClassTy::NONE, IsKnownNonNull);
3084 if (IsKnownNonNull)
3085 KnownUBInsts.insert(&I);
3086 }
3087 return true;
3088 };
3089
3090 auto InspectReturnInstForUB = [&](Instruction &I) {
3091 auto &RI = cast<ReturnInst>(I);
3092 // Either we stopped and the appropriate action was taken,
3093 // or we got back a simplified return value to continue.
3094 std::optional<Value *> SimplifiedRetValue =
3095 stopOnUndefOrAssumed(A, RI.getReturnValue(), &I);
3096 if (!SimplifiedRetValue || !*SimplifiedRetValue)
3097 return true;
3098
3099 // Check if a return instruction always cause UB or not
3100 // Note: It is guaranteed that the returned position of the anchor
3101 // scope has noundef attribute when this is called.
3102 // We also ensure the return position is not "assumed dead"
3103 // because the returned value was then potentially simplified to
3104 // `undef` in AAReturnedValues without removing the `noundef`
3105 // attribute yet.
3106
3107 // When the returned position has noundef attriubte, UB occurs in the
3108 // following cases.
3109 // (1) Returned value is known to be undef.
3110 // (2) The value is known to be a null pointer and the returned
3111 // position has nonnull attribute (because the returned value is
3112 // poison).
3113 if (isa<ConstantPointerNull>(*SimplifiedRetValue)) {
3114 bool IsKnownNonNull;
3116 A, this, IRPosition::returned(*getAnchorScope()), DepClassTy::NONE,
3117 IsKnownNonNull);
3118 if (IsKnownNonNull)
3119 KnownUBInsts.insert(&I);
3120 }
3121
3122 return true;
3123 };
3124
3125 bool UsedAssumedInformation = false;
3126 A.checkForAllInstructions(InspectMemAccessInstForUB, *this,
3127 {Instruction::Load, Instruction::Store,
3128 Instruction::AtomicCmpXchg,
3129 Instruction::AtomicRMW},
3130 UsedAssumedInformation,
3131 /* CheckBBLivenessOnly */ true);
3132 A.checkForAllInstructions(InspectBrInstForUB, *this, {Instruction::Br},
3133 UsedAssumedInformation,
3134 /* CheckBBLivenessOnly */ true);
3135 A.checkForAllCallLikeInstructions(InspectCallSiteForUB, *this,
3136 UsedAssumedInformation);
3137
3138 // If the returned position of the anchor scope has noundef attriubte, check
3139 // all returned instructions.
3140 if (!getAnchorScope()->getReturnType()->isVoidTy()) {
3141 const IRPosition &ReturnIRP = IRPosition::returned(*getAnchorScope());
3142 if (!A.isAssumedDead(ReturnIRP, this, nullptr, UsedAssumedInformation)) {
3143 bool IsKnownNoUndef;
3145 A, this, ReturnIRP, DepClassTy::NONE, IsKnownNoUndef);
3146 if (IsKnownNoUndef)
3147 A.checkForAllInstructions(InspectReturnInstForUB, *this,
3148 {Instruction::Ret}, UsedAssumedInformation,
3149 /* CheckBBLivenessOnly */ true);
3150 }
3151 }
3152
3153 if (NoUBPrevSize != AssumedNoUBInsts.size() ||
3154 UBPrevSize != KnownUBInsts.size())
3155 return ChangeStatus::CHANGED;
3156 return ChangeStatus::UNCHANGED;
3157 }
3158
3159 bool isKnownToCauseUB(Instruction *I) const override {
3160 return KnownUBInsts.count(I);
3161 }
3162
3163 bool isAssumedToCauseUB(Instruction *I) const override {
3164 // In simple words, if an instruction is not in the assumed to _not_
3165 // cause UB, then it is assumed UB (that includes those
3166 // in the KnownUBInsts set). The rest is boilerplate
3167 // is to ensure that it is one of the instructions we test
3168 // for UB.
3169
3170 switch (I->getOpcode()) {
3171 case Instruction::Load:
3172 case Instruction::Store:
3173 case Instruction::AtomicCmpXchg:
3174 case Instruction::AtomicRMW:
3175 return !AssumedNoUBInsts.count(I);
3176 case Instruction::Br: {
3177 auto *BrInst = cast<BranchInst>(I);
3178 if (BrInst->isUnconditional())
3179 return false;
3180 return !AssumedNoUBInsts.count(I);
3181 } break;
3182 default:
3183 return false;
3184 }
3185 return false;
3186 }
3187
3188 ChangeStatus manifest(Attributor &A) override {
3189 if (KnownUBInsts.empty())
3190 return ChangeStatus::UNCHANGED;
3191 for (Instruction *I : KnownUBInsts)
3192 A.changeToUnreachableAfterManifest(I);
3193 return ChangeStatus::CHANGED;
3194 }
3195
3196 /// See AbstractAttribute::getAsStr()
3197 const std::string getAsStr(Attributor *A) const override {
3198 return getAssumed() ? "undefined-behavior" : "no-ub";
3199 }
3200
3201 /// Note: The correctness of this analysis depends on the fact that the
3202 /// following 2 sets will stop changing after some point.
3203 /// "Change" here means that their size changes.
3204 /// The size of each set is monotonically increasing
3205 /// (we only add items to them) and it is upper bounded by the number of
3206 /// instructions in the processed function (we can never save more
3207 /// elements in either set than this number). Hence, at some point,
3208 /// they will stop increasing.
3209 /// Consequently, at some point, both sets will have stopped
3210 /// changing, effectively making the analysis reach a fixpoint.
3211
3212 /// Note: These 2 sets are disjoint and an instruction can be considered
3213 /// one of 3 things:
3214 /// 1) Known to cause UB (AAUndefinedBehavior could prove it) and put it in
3215 /// the KnownUBInsts set.
3216 /// 2) Assumed to cause UB (in every updateImpl, AAUndefinedBehavior
3217 /// has a reason to assume it).
3218 /// 3) Assumed to not cause UB. very other instruction - AAUndefinedBehavior
3219 /// could not find a reason to assume or prove that it can cause UB,
3220 /// hence it assumes it doesn't. We have a set for these instructions
3221 /// so that we don't reprocess them in every update.
3222 /// Note however that instructions in this set may cause UB.
3223
3224protected:
3225 /// A set of all live instructions _known_ to cause UB.
3226 SmallPtrSet<Instruction *, 8> KnownUBInsts;
3227
3228private:
3229 /// A set of all the (live) instructions that are assumed to _not_ cause UB.
3230 SmallPtrSet<Instruction *, 8> AssumedNoUBInsts;
3231
3232 // Should be called on updates in which if we're processing an instruction
3233 // \p I that depends on a value \p V, one of the following has to happen:
3234 // - If the value is assumed, then stop.
3235 // - If the value is known but undef, then consider it UB.
3236 // - Otherwise, do specific processing with the simplified value.
3237 // We return std::nullopt in the first 2 cases to signify that an appropriate
3238 // action was taken and the caller should stop.
3239 // Otherwise, we return the simplified value that the caller should
3240 // use for specific processing.
3241 std::optional<Value *> stopOnUndefOrAssumed(Attributor &A, Value *V,
3242 Instruction *I) {
3243 bool UsedAssumedInformation = false;
3244 std::optional<Value *> SimplifiedV =
3245 A.getAssumedSimplified(IRPosition::value(*V), *this,
3246 UsedAssumedInformation, AA::Interprocedural);
3247 if (!UsedAssumedInformation) {
3248 // Don't depend on assumed values.
3249 if (!SimplifiedV) {
3250 // If it is known (which we tested above) but it doesn't have a value,
3251 // then we can assume `undef` and hence the instruction is UB.
3252 KnownUBInsts.insert(I);
3253 return std::nullopt;
3254 }
3255 if (!*SimplifiedV)
3256 return nullptr;
3257 V = *SimplifiedV;
3258 }
3259 if (isa<UndefValue>(V)) {
3260 KnownUBInsts.insert(I);
3261 return std::nullopt;
3262 }
3263 return V;
3264 }
3265};
3266
3267struct AAUndefinedBehaviorFunction final : AAUndefinedBehaviorImpl {
3268 AAUndefinedBehaviorFunction(const IRPosition &IRP, Attributor &A)
3269 : AAUndefinedBehaviorImpl(IRP, A) {}
3270
3271 /// See AbstractAttribute::trackStatistics()
3272 void trackStatistics() const override {
3273 STATS_DECL(UndefinedBehaviorInstruction, Instruction,
3274 "Number of instructions known to have UB");
3275 BUILD_STAT_NAME(UndefinedBehaviorInstruction, Instruction) +=
3276 KnownUBInsts.size();
3277 }
3278};
3279} // namespace
3280
3281/// ------------------------ Will-Return Attributes ----------------------------
3282
3283namespace {
3284// Helper function that checks whether a function has any cycle which we don't
3285// know if it is bounded or not.
3286// Loops with maximum trip count are considered bounded, any other cycle not.
3287static bool mayContainUnboundedCycle(Function &F, Attributor &A) {
3288 ScalarEvolution *SE =
3289 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(F);
3290 LoopInfo *LI = A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(F);
3291 // If either SCEV or LoopInfo is not available for the function then we assume
3292 // any cycle to be unbounded cycle.
3293 // We use scc_iterator which uses Tarjan algorithm to find all the maximal
3294 // SCCs.To detect if there's a cycle, we only need to find the maximal ones.
3295 if (!SE || !LI) {
3296 for (scc_iterator<Function *> SCCI = scc_begin(&F); !SCCI.isAtEnd(); ++SCCI)
3297 if (SCCI.hasCycle())
3298 return true;
3299 return false;
3300 }
3301
3302 // If there's irreducible control, the function may contain non-loop cycles.
3304 return true;
3305
3306 // Any loop that does not have a max trip count is considered unbounded cycle.
3307 for (auto *L : LI->getLoopsInPreorder()) {
3308 if (!SE->getSmallConstantMaxTripCount(L))
3309 return true;
3310 }
3311 return false;
3312}
3313
3314struct AAWillReturnImpl : public AAWillReturn {
3315 AAWillReturnImpl(const IRPosition &IRP, Attributor &A)
3316 : AAWillReturn(IRP, A) {}
3317
3318 /// See AbstractAttribute::initialize(...).
3319 void initialize(Attributor &A) override {
3320 bool IsKnown;
3322 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
3323 (void)IsKnown;
3324 }
3325
3326 /// Check for `mustprogress` and `readonly` as they imply `willreturn`.
3327 bool isImpliedByMustprogressAndReadonly(Attributor &A, bool KnownOnly) {
3328 if (!A.hasAttr(getIRPosition(), {Attribute::MustProgress}))
3329 return false;
3330
3331 bool IsKnown;
3332 if (AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
3333 return IsKnown || !KnownOnly;
3334 return false;
3335 }
3336
3337 /// See AbstractAttribute::updateImpl(...).
3338 ChangeStatus updateImpl(Attributor &A) override {
3339 if (isImpliedByMustprogressAndReadonly(A, /* KnownOnly */ false))
3340 return ChangeStatus::UNCHANGED;
3341
3342 auto CheckForWillReturn = [&](Instruction &I) {
3344 bool IsKnown;
3346 A, this, IPos, DepClassTy::REQUIRED, IsKnown)) {
3347 if (IsKnown)
3348 return true;
3349 } else {
3350 return false;
3351 }
3352 bool IsKnownNoRecurse;
3354 A, this, IPos, DepClassTy::REQUIRED, IsKnownNoRecurse);
3355 };
3356
3357 bool UsedAssumedInformation = false;
3358 if (!A.checkForAllCallLikeInstructions(CheckForWillReturn, *this,
3359 UsedAssumedInformation))
3360 return indicatePessimisticFixpoint();
3361
3362 return ChangeStatus::UNCHANGED;
3363 }
3364
3365 /// See AbstractAttribute::getAsStr()
3366 const std::string getAsStr(Attributor *A) const override {
3367 return getAssumed() ? "willreturn" : "may-noreturn";
3368 }
3369};
3370
3371struct AAWillReturnFunction final : AAWillReturnImpl {
3372 AAWillReturnFunction(const IRPosition &IRP, Attributor &A)
3373 : AAWillReturnImpl(IRP, A) {}
3374
3375 /// See AbstractAttribute::initialize(...).
3376 void initialize(Attributor &A) override {
3377 AAWillReturnImpl::initialize(A);
3378
3379 Function *F = getAnchorScope();
3380 assert(F && "Did expect an anchor function");
3381 if (F->isDeclaration() || mayContainUnboundedCycle(*F, A))
3382 indicatePessimisticFixpoint();
3383 }
3384
3385 /// See AbstractAttribute::trackStatistics()
3386 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(willreturn) }
3387};
3388
3389/// WillReturn attribute deduction for a call sites.
3390struct AAWillReturnCallSite final
3391 : AACalleeToCallSite<AAWillReturn, AAWillReturnImpl> {
3392 AAWillReturnCallSite(const IRPosition &IRP, Attributor &A)
3393 : AACalleeToCallSite<AAWillReturn, AAWillReturnImpl>(IRP, A) {}
3394
3395 /// See AbstractAttribute::updateImpl(...).
3396 ChangeStatus updateImpl(Attributor &A) override {
3397 if (isImpliedByMustprogressAndReadonly(A, /* KnownOnly */ false))
3398 return ChangeStatus::UNCHANGED;
3399
3400 return AACalleeToCallSite::updateImpl(A);
3401 }
3402
3403 /// See AbstractAttribute::trackStatistics()
3404 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(willreturn); }
3405};
3406} // namespace
3407
3408/// -------------------AAIntraFnReachability Attribute--------------------------
3409
3410/// All information associated with a reachability query. This boilerplate code
3411/// is used by both AAIntraFnReachability and AAInterFnReachability, with
3412/// different \p ToTy values.
3413template <typename ToTy> struct ReachabilityQueryInfo {
3414 enum class Reachable {
3417 };
3418
3419 /// Start here,
3420 const Instruction *From = nullptr;
3421 /// reach this place,
3422 const ToTy *To = nullptr;
3423 /// without going through any of these instructions,
3425 /// and remember if it worked:
3427
3428 /// Precomputed hash for this RQI.
3429 unsigned Hash = 0;
3430
3431 unsigned computeHashValue() const {
3432 assert(Hash == 0 && "Computed hash twice!");
3435 return const_cast<ReachabilityQueryInfo<ToTy> *>(this)->Hash =
3436 detail::combineHashValue(PairDMI ::getHashValue({From, To}),
3437 InstSetDMI::getHashValue(ExclusionSet));
3438 }
3439
3441 : From(From), To(To) {}
3442
3443 /// Constructor replacement to ensure unique and stable sets are used for the
3444 /// cache.
3446 const AA::InstExclusionSetTy *ES, bool MakeUnique)
3447 : From(&From), To(&To), ExclusionSet(ES) {
3448
3449 if (!ES || ES->empty()) {
3450 ExclusionSet = nullptr;
3451 } else if (MakeUnique) {
3452 ExclusionSet = A.getInfoCache().getOrCreateUniqueBlockExecutionSet(ES);
3453 }
3454 }
3455
3458};
3459
3460namespace llvm {
3461template <typename ToTy> struct DenseMapInfo<ReachabilityQueryInfo<ToTy> *> {
3464
3467
3470 return &TombstoneKey;
3471 }
3472 static unsigned getHashValue(const ReachabilityQueryInfo<ToTy> *RQI) {
3473 return RQI->Hash ? RQI->Hash : RQI->computeHashValue();
3474 }
3475 static bool isEqual(const ReachabilityQueryInfo<ToTy> *LHS,
3476 const ReachabilityQueryInfo<ToTy> *RHS) {
3477 if (!PairDMI::isEqual({LHS->From, LHS->To}, {RHS->From, RHS->To}))
3478 return false;
3479 return InstSetDMI::isEqual(LHS->ExclusionSet, RHS->ExclusionSet);
3480 }
3481};
3482
3483#define DefineKeys(ToTy) \
3484 template <> \
3485 ReachabilityQueryInfo<ToTy> \
3486 DenseMapInfo<ReachabilityQueryInfo<ToTy> *>::EmptyKey = \
3487 ReachabilityQueryInfo<ToTy>( \
3488 DenseMapInfo<const Instruction *>::getEmptyKey(), \
3489 DenseMapInfo<const ToTy *>::getEmptyKey()); \
3490 template <> \
3491 ReachabilityQueryInfo<ToTy> \
3492 DenseMapInfo<ReachabilityQueryInfo<ToTy> *>::TombstoneKey = \
3493 ReachabilityQueryInfo<ToTy>( \
3494 DenseMapInfo<const Instruction *>::getTombstoneKey(), \
3495 DenseMapInfo<const ToTy *>::getTombstoneKey());
3496
3498#undef DefineKeys
3499
3500} // namespace llvm
3501
3502namespace {
3503
3504template <typename BaseTy, typename ToTy>
3505struct CachedReachabilityAA : public BaseTy {
3506 using RQITy = ReachabilityQueryInfo<ToTy>;
3507
3508 CachedReachabilityAA(const IRPosition &IRP, Attributor &A) : BaseTy(IRP, A) {}
3509
3510 /// See AbstractAttribute::isQueryAA.
3511 bool isQueryAA() const override { return true; }
3512
3513 /// See AbstractAttribute::updateImpl(...).
3514 ChangeStatus updateImpl(Attributor &A) override {
3515 ChangeStatus Changed = ChangeStatus::UNCHANGED;
3516 for (unsigned u = 0, e = QueryVector.size(); u < e; ++u) {
3517 RQITy *RQI = QueryVector[u];
3518 if (RQI->Result == RQITy::Reachable::No &&
3519 isReachableImpl(A, *RQI, /*IsTemporaryRQI=*/false))
3520 Changed = ChangeStatus::CHANGED;
3521 }
3522 return Changed;
3523 }
3524
3525 virtual bool isReachableImpl(Attributor &A, RQITy &RQI,
3526 bool IsTemporaryRQI) = 0;
3527
3528 bool rememberResult(Attributor &A, typename RQITy::Reachable Result,
3529 RQITy &RQI, bool UsedExclusionSet, bool IsTemporaryRQI) {
3530 RQI.Result = Result;
3531
3532 // Remove the temporary RQI from the cache.
3533 if (IsTemporaryRQI)
3534 QueryCache.erase(&RQI);
3535
3536 // Insert a plain RQI (w/o exclusion set) if that makes sense. Two options:
3537 // 1) If it is reachable, it doesn't matter if we have an exclusion set for
3538 // this query. 2) We did not use the exclusion set, potentially because
3539 // there is none.
3540 if (Result == RQITy::Reachable::Yes || !UsedExclusionSet) {
3541 RQITy PlainRQI(RQI.From, RQI.To);
3542 if (!QueryCache.count(&PlainRQI)) {
3543 RQITy *RQIPtr = new (A.Allocator) RQITy(RQI.From, RQI.To);
3544 RQIPtr->Result = Result;
3545 QueryVector.push_back(RQIPtr);
3546 QueryCache.insert(RQIPtr);
3547 }
3548 }
3549
3550 // Check if we need to insert a new permanent RQI with the exclusion set.
3551 if (IsTemporaryRQI && Result != RQITy::Reachable::Yes && UsedExclusionSet) {
3552 assert((!RQI.ExclusionSet || !RQI.ExclusionSet->empty()) &&
3553 "Did not expect empty set!");
3554 RQITy *RQIPtr = new (A.Allocator)
3555 RQITy(A, *RQI.From, *RQI.To, RQI.ExclusionSet, true);
3556 assert(RQIPtr->Result == RQITy::Reachable::No && "Already reachable?");
3557 RQIPtr->Result = Result;
3558 assert(!QueryCache.count(RQIPtr));
3559 QueryVector.push_back(RQIPtr);
3560 QueryCache.insert(RQIPtr);
3561 }
3562
3563 if (Result == RQITy::Reachable::No && IsTemporaryRQI)
3564 A.registerForUpdate(*this);
3565 return Result == RQITy::Reachable::Yes;
3566 }
3567
3568 const std::string getAsStr(Attributor *A) const override {
3569 // TODO: Return the number of reachable queries.
3570 return "#queries(" + std::to_string(QueryVector.size()) + ")";
3571 }
3572
3573 bool checkQueryCache(Attributor &A, RQITy &StackRQI,
3574 typename RQITy::Reachable &Result) {
3575 if (!this->getState().isValidState()) {
3576 Result = RQITy::Reachable::Yes;
3577 return true;
3578 }
3579
3580 // If we have an exclusion set we might be able to find our answer by
3581 // ignoring it first.
3582 if (StackRQI.ExclusionSet) {
3583 RQITy PlainRQI(StackRQI.From, StackRQI.To);
3584 auto It = QueryCache.find(&PlainRQI);
3585 if (It != QueryCache.end() && (*It)->Result == RQITy::Reachable::No) {
3586 Result = RQITy::Reachable::No;
3587 return true;
3588 }
3589 }
3590
3591 auto It = QueryCache.find(&StackRQI);
3592 if (It != QueryCache.end()) {
3593 Result = (*It)->Result;
3594 return true;
3595 }
3596
3597 // Insert a temporary for recursive queries. We will replace it with a
3598 // permanent entry later.
3599 QueryCache.insert(&StackRQI);
3600 return false;
3601 }
3602
3603private:
3604 SmallVector<RQITy *> QueryVector;
3605 DenseSet<RQITy *> QueryCache;
3606};
3607
3608struct AAIntraFnReachabilityFunction final
3609 : public CachedReachabilityAA<AAIntraFnReachability, Instruction> {
3610 using Base = CachedReachabilityAA<AAIntraFnReachability, Instruction>;
3611 AAIntraFnReachabilityFunction(const IRPosition &IRP, Attributor &A)
3612 : Base(IRP, A) {
3613 DT = A.getInfoCache().getAnalysisResultForFunction<DominatorTreeAnalysis>(
3614 *IRP.getAssociatedFunction());
3615 }
3616
3617 bool isAssumedReachable(
3618 Attributor &A, const Instruction &From, const Instruction &To,
3619 const AA::InstExclusionSetTy *ExclusionSet) const override {
3620 auto *NonConstThis = const_cast<AAIntraFnReachabilityFunction *>(this);
3621 if (&From == &To)
3622 return true;
3623
3624 RQITy StackRQI(A, From, To, ExclusionSet, false);
3625 RQITy::Reachable Result;
3626 if (!NonConstThis->checkQueryCache(A, StackRQI, Result))
3627 return NonConstThis->isReachableImpl(A, StackRQI,
3628 /*IsTemporaryRQI=*/true);
3629 return Result == RQITy::Reachable::Yes;
3630 }
3631
3632 ChangeStatus updateImpl(Attributor &A) override {
3633 // We only depend on liveness. DeadEdges is all we care about, check if any
3634 // of them changed.
3635 auto *LivenessAA =
3636 A.getAAFor<AAIsDead>(*this, getIRPosition(), DepClassTy::OPTIONAL);
3637 if (LivenessAA &&
3638 llvm::all_of(DeadEdges,
3639 [&](const auto &DeadEdge) {
3640 return LivenessAA->isEdgeDead(DeadEdge.first,
3641 DeadEdge.second);
3642 }) &&
3643 llvm::all_of(DeadBlocks, [&](const BasicBlock *BB) {
3644 return LivenessAA->isAssumedDead(BB);
3645 })) {
3646 return ChangeStatus::UNCHANGED;
3647 }
3648 DeadEdges.clear();
3649 DeadBlocks.clear();
3650 return Base::updateImpl(A);
3651 }
3652
3653 bool isReachableImpl(Attributor &A, RQITy &RQI,
3654 bool IsTemporaryRQI) override {
3655 const Instruction *Origin = RQI.From;
3656 bool UsedExclusionSet = false;
3657
3658 auto WillReachInBlock = [&](const Instruction &From, const Instruction &To,
3659 const AA::InstExclusionSetTy *ExclusionSet) {
3660 const Instruction *IP = &From;
3661 while (IP && IP != &To) {
3662 if (ExclusionSet && IP != Origin && ExclusionSet->count(IP)) {
3663 UsedExclusionSet = true;
3664 break;
3665 }
3666 IP = IP->getNextNode();
3667 }
3668 return IP == &To;
3669 };
3670
3671 const BasicBlock *FromBB = RQI.From->getParent();
3672 const BasicBlock *ToBB = RQI.To->getParent();
3673 assert(FromBB->getParent() == ToBB->getParent() &&
3674 "Not an intra-procedural query!");
3675
3676 // Check intra-block reachability, however, other reaching paths are still
3677 // possible.
3678 if (FromBB == ToBB &&
3679 WillReachInBlock(*RQI.From, *RQI.To, RQI.ExclusionSet))
3680 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3681 IsTemporaryRQI);
3682
3683 // Check if reaching the ToBB block is sufficient or if even that would not
3684 // ensure reaching the target. In the latter case we are done.
3685 if (!WillReachInBlock(ToBB->front(), *RQI.To, RQI.ExclusionSet))
3686 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3687 IsTemporaryRQI);
3688
3689 const Function *Fn = FromBB->getParent();
3690 SmallPtrSet<const BasicBlock *, 16> ExclusionBlocks;
3691 if (RQI.ExclusionSet)
3692 for (auto *I : *RQI.ExclusionSet)
3693 if (I->getFunction() == Fn)
3694 ExclusionBlocks.insert(I->getParent());
3695
3696 // Check if we make it out of the FromBB block at all.
3697 if (ExclusionBlocks.count(FromBB) &&
3698 !WillReachInBlock(*RQI.From, *FromBB->getTerminator(),
3699 RQI.ExclusionSet))
3700 return rememberResult(A, RQITy::Reachable::No, RQI, true, IsTemporaryRQI);
3701
3702 auto *LivenessAA =
3703 A.getAAFor<AAIsDead>(*this, getIRPosition(), DepClassTy::OPTIONAL);
3704 if (LivenessAA && LivenessAA->isAssumedDead(ToBB)) {
3705 DeadBlocks.insert(ToBB);
3706 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3707 IsTemporaryRQI);
3708 }
3709
3710 SmallPtrSet<const BasicBlock *, 16> Visited;
3712 Worklist.push_back(FromBB);
3713
3714 DenseSet<std::pair<const BasicBlock *, const BasicBlock *>> LocalDeadEdges;
3715 while (!Worklist.empty()) {
3716 const BasicBlock *BB = Worklist.pop_back_val();
3717 if (!Visited.insert(BB).second)
3718 continue;
3719 for (const BasicBlock *SuccBB : successors(BB)) {
3720 if (LivenessAA && LivenessAA->isEdgeDead(BB, SuccBB)) {
3721 LocalDeadEdges.insert({BB, SuccBB});
3722 continue;
3723 }
3724 // We checked before if we just need to reach the ToBB block.
3725 if (SuccBB == ToBB)
3726 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3727 IsTemporaryRQI);
3728 if (DT && ExclusionBlocks.empty() && DT->dominates(BB, ToBB))
3729 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3730 IsTemporaryRQI);
3731
3732 if (ExclusionBlocks.count(SuccBB)) {
3733 UsedExclusionSet = true;
3734 continue;
3735 }
3736 Worklist.push_back(SuccBB);
3737 }
3738 }
3739
3740 DeadEdges.insert_range(LocalDeadEdges);
3741 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3742 IsTemporaryRQI);
3743 }
3744
3745 /// See AbstractAttribute::trackStatistics()
3746 void trackStatistics() const override {}
3747
3748private:
3749 // Set of assumed dead blocks we used in the last query. If any changes we
3750 // update the state.
3751 DenseSet<const BasicBlock *> DeadBlocks;
3752
3753 // Set of assumed dead edges we used in the last query. If any changes we
3754 // update the state.
3755 DenseSet<std::pair<const BasicBlock *, const BasicBlock *>> DeadEdges;
3756
3757 /// The dominator tree of the function to short-circuit reasoning.
3758 const DominatorTree *DT = nullptr;
3759};
3760} // namespace
3761
3762/// ------------------------ NoAlias Argument Attribute ------------------------
3763
3765 Attribute::AttrKind ImpliedAttributeKind,
3766 bool IgnoreSubsumingPositions) {
3767 assert(ImpliedAttributeKind == Attribute::NoAlias &&
3768 "Unexpected attribute kind");
3769 Value *Val = &IRP.getAssociatedValue();
3771 if (isa<AllocaInst>(Val))
3772 return true;
3773 } else {
3774 IgnoreSubsumingPositions = true;
3775 }
3776
3777 if (isa<UndefValue>(Val))
3778 return true;
3779
3780 if (isa<ConstantPointerNull>(Val) &&
3783 return true;
3784
3785 if (A.hasAttr(IRP, {Attribute::ByVal, Attribute::NoAlias},
3786 IgnoreSubsumingPositions, Attribute::NoAlias))
3787 return true;
3788
3789 return false;
3790}
3791
3792namespace {
3793struct AANoAliasImpl : AANoAlias {
3794 AANoAliasImpl(const IRPosition &IRP, Attributor &A) : AANoAlias(IRP, A) {
3795 assert(getAssociatedType()->isPointerTy() &&
3796 "Noalias is a pointer attribute");
3797 }
3798
3799 const std::string getAsStr(Attributor *A) const override {
3800 return getAssumed() ? "noalias" : "may-alias";
3801 }
3802};
3803
3804/// NoAlias attribute for a floating value.
3805struct AANoAliasFloating final : AANoAliasImpl {
3806 AANoAliasFloating(const IRPosition &IRP, Attributor &A)
3807 : AANoAliasImpl(IRP, A) {}
3808
3809 /// See AbstractAttribute::updateImpl(...).
3810 ChangeStatus updateImpl(Attributor &A) override {
3811 // TODO: Implement this.
3812 return indicatePessimisticFixpoint();
3813 }
3814
3815 /// See AbstractAttribute::trackStatistics()
3816 void trackStatistics() const override {
3818 }
3819};
3820
3821/// NoAlias attribute for an argument.
3822struct AANoAliasArgument final
3823 : AAArgumentFromCallSiteArguments<AANoAlias, AANoAliasImpl> {
3824 using Base = AAArgumentFromCallSiteArguments<AANoAlias, AANoAliasImpl>;
3825 AANoAliasArgument(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
3826
3827 /// See AbstractAttribute::update(...).
3828 ChangeStatus updateImpl(Attributor &A) override {
3829 // We have to make sure no-alias on the argument does not break
3830 // synchronization when this is a callback argument, see also [1] below.
3831 // If synchronization cannot be affected, we delegate to the base updateImpl
3832 // function, otherwise we give up for now.
3833
3834 // If the function is no-sync, no-alias cannot break synchronization.
3835 bool IsKnownNoSycn;
3837 A, this, IRPosition::function_scope(getIRPosition()),
3838 DepClassTy::OPTIONAL, IsKnownNoSycn))
3839 return Base::updateImpl(A);
3840
3841 // If the argument is read-only, no-alias cannot break synchronization.
3842 bool IsKnown;
3843 if (AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
3844 return Base::updateImpl(A);
3845
3846 // If the argument is never passed through callbacks, no-alias cannot break
3847 // synchronization.
3848 bool UsedAssumedInformation = false;
3849 if (A.checkForAllCallSites(
3850 [](AbstractCallSite ACS) { return !ACS.isCallbackCall(); }, *this,
3851 true, UsedAssumedInformation))
3852 return Base::updateImpl(A);
3853
3854 // TODO: add no-alias but make sure it doesn't break synchronization by
3855 // introducing fake uses. See:
3856 // [1] Compiler Optimizations for OpenMP, J. Doerfert and H. Finkel,
3857 // International Workshop on OpenMP 2018,
3858 // http://compilers.cs.uni-saarland.de/people/doerfert/par_opt18.pdf
3859
3860 return indicatePessimisticFixpoint();
3861 }
3862
3863 /// See AbstractAttribute::trackStatistics()
3864 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(noalias) }
3865};
3866
3867struct AANoAliasCallSiteArgument final : AANoAliasImpl {
3868 AANoAliasCallSiteArgument(const IRPosition &IRP, Attributor &A)
3869 : AANoAliasImpl(IRP, A) {}
3870
3871 /// Determine if the underlying value may alias with the call site argument
3872 /// \p OtherArgNo of \p ICS (= the underlying call site).
3873 bool mayAliasWithArgument(Attributor &A, AAResults *&AAR,
3874 const AAMemoryBehavior &MemBehaviorAA,
3875 const CallBase &CB, unsigned OtherArgNo) {
3876 // We do not need to worry about aliasing with the underlying IRP.
3877 if (this->getCalleeArgNo() == (int)OtherArgNo)
3878 return false;
3879
3880 // If it is not a pointer or pointer vector we do not alias.
3881 const Value *ArgOp = CB.getArgOperand(OtherArgNo);
3882 if (!ArgOp->getType()->isPtrOrPtrVectorTy())
3883 return false;
3884
3885 auto *CBArgMemBehaviorAA = A.getAAFor<AAMemoryBehavior>(
3886 *this, IRPosition::callsite_argument(CB, OtherArgNo), DepClassTy::NONE);
3887
3888 // If the argument is readnone, there is no read-write aliasing.
3889 if (CBArgMemBehaviorAA && CBArgMemBehaviorAA->isAssumedReadNone()) {
3890 A.recordDependence(*CBArgMemBehaviorAA, *this, DepClassTy::OPTIONAL);
3891 return false;
3892 }
3893
3894 // If the argument is readonly and the underlying value is readonly, there
3895 // is no read-write aliasing.
3896 bool IsReadOnly = MemBehaviorAA.isAssumedReadOnly();
3897 if (CBArgMemBehaviorAA && CBArgMemBehaviorAA->isAssumedReadOnly() &&
3898 IsReadOnly) {
3899 A.recordDependence(MemBehaviorAA, *this, DepClassTy::OPTIONAL);
3900 A.recordDependence(*CBArgMemBehaviorAA, *this, DepClassTy::OPTIONAL);
3901 return false;
3902 }
3903
3904 // We have to utilize actual alias analysis queries so we need the object.
3905 if (!AAR)
3906 AAR = A.getInfoCache().getAnalysisResultForFunction<AAManager>(
3907 *getAnchorScope());
3908
3909 // Try to rule it out at the call site.
3910 bool IsAliasing = !AAR || !AAR->isNoAlias(&getAssociatedValue(), ArgOp);
3911 LLVM_DEBUG(dbgs() << "[NoAliasCSArg] Check alias between "
3912 "callsite arguments: "
3913 << getAssociatedValue() << " " << *ArgOp << " => "
3914 << (IsAliasing ? "" : "no-") << "alias \n");
3915
3916 return IsAliasing;
3917 }
3918
3919 bool isKnownNoAliasDueToNoAliasPreservation(
3920 Attributor &A, AAResults *&AAR, const AAMemoryBehavior &MemBehaviorAA) {
3921 // We can deduce "noalias" if the following conditions hold.
3922 // (i) Associated value is assumed to be noalias in the definition.
3923 // (ii) Associated value is assumed to be no-capture in all the uses
3924 // possibly executed before this callsite.
3925 // (iii) There is no other pointer argument which could alias with the
3926 // value.
3927
3928 const IRPosition &VIRP = IRPosition::value(getAssociatedValue());
3929 const Function *ScopeFn = VIRP.getAnchorScope();
3930 // Check whether the value is captured in the scope using AANoCapture.
3931 // Look at CFG and check only uses possibly executed before this
3932 // callsite.
3933 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
3934 Instruction *UserI = cast<Instruction>(U.getUser());
3935
3936 // If UserI is the curr instruction and there is a single potential use of
3937 // the value in UserI we allow the use.
3938 // TODO: We should inspect the operands and allow those that cannot alias
3939 // with the value.
3940 if (UserI == getCtxI() && UserI->getNumOperands() == 1)
3941 return true;
3942
3943 if (ScopeFn) {
3944 if (auto *CB = dyn_cast<CallBase>(UserI)) {
3945 if (CB->isArgOperand(&U)) {
3946
3947 unsigned ArgNo = CB->getArgOperandNo(&U);
3948
3949 bool IsKnownNoCapture;
3951 A, this, IRPosition::callsite_argument(*CB, ArgNo),
3952 DepClassTy::OPTIONAL, IsKnownNoCapture))
3953 return true;
3954 }
3955 }
3956
3958 A, *UserI, *getCtxI(), *this, /* ExclusionSet */ nullptr,
3959 [ScopeFn](const Function &Fn) { return &Fn != ScopeFn; }))
3960 return true;
3961 }
3962
3963 // TODO: We should track the capturing uses in AANoCapture but the problem
3964 // is CGSCC runs. For those we would need to "allow" AANoCapture for
3965 // a value in the module slice.
3966 // TODO(captures): Make this more precise.
3967 UseCaptureInfo CI = DetermineUseCaptureKind(U, /*Base=*/nullptr);
3968 if (capturesNothing(CI))
3969 return true;
3970 if (CI.isPassthrough()) {
3971 Follow = true;
3972 return true;
3973 }
3974 LLVM_DEBUG(dbgs() << "[AANoAliasCSArg] Unknown user: " << *UserI << "\n");
3975 return false;
3976 };
3977
3978 bool IsKnownNoCapture;
3979 const AANoCapture *NoCaptureAA = nullptr;
3980 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
3981 A, this, VIRP, DepClassTy::NONE, IsKnownNoCapture, false, &NoCaptureAA);
3982 if (!IsAssumedNoCapture &&
3983 (!NoCaptureAA || !NoCaptureAA->isAssumedNoCaptureMaybeReturned())) {
3984 if (!A.checkForAllUses(UsePred, *this, getAssociatedValue())) {
3985 LLVM_DEBUG(
3986 dbgs() << "[AANoAliasCSArg] " << getAssociatedValue()
3987 << " cannot be noalias as it is potentially captured\n");
3988 return false;
3989 }
3990 }
3991 if (NoCaptureAA)
3992 A.recordDependence(*NoCaptureAA, *this, DepClassTy::OPTIONAL);
3993
3994 // Check there is no other pointer argument which could alias with the
3995 // value passed at this call site.
3996 // TODO: AbstractCallSite
3997 const auto &CB = cast<CallBase>(getAnchorValue());
3998 for (unsigned OtherArgNo = 0; OtherArgNo < CB.arg_size(); OtherArgNo++)
3999 if (mayAliasWithArgument(A, AAR, MemBehaviorAA, CB, OtherArgNo))
4000 return false;
4001
4002 return true;
4003 }
4004
4005 /// See AbstractAttribute::updateImpl(...).
4006 ChangeStatus updateImpl(Attributor &A) override {
4007 // If the argument is readnone we are done as there are no accesses via the
4008 // argument.
4009 auto *MemBehaviorAA =
4010 A.getAAFor<AAMemoryBehavior>(*this, getIRPosition(), DepClassTy::NONE);
4011 if (MemBehaviorAA && MemBehaviorAA->isAssumedReadNone()) {
4012 A.recordDependence(*MemBehaviorAA, *this, DepClassTy::OPTIONAL);
4013 return ChangeStatus::UNCHANGED;
4014 }
4015
4016 bool IsKnownNoAlias;
4017 const IRPosition &VIRP = IRPosition::value(getAssociatedValue());
4019 A, this, VIRP, DepClassTy::REQUIRED, IsKnownNoAlias)) {
4020 LLVM_DEBUG(dbgs() << "[AANoAlias] " << getAssociatedValue()
4021 << " is not no-alias at the definition\n");
4022 return indicatePessimisticFixpoint();
4023 }
4024
4025 AAResults *AAR = nullptr;
4026 if (MemBehaviorAA &&
4027 isKnownNoAliasDueToNoAliasPreservation(A, AAR, *MemBehaviorAA)) {
4028 LLVM_DEBUG(
4029 dbgs() << "[AANoAlias] No-Alias deduced via no-alias preservation\n");
4030 return ChangeStatus::UNCHANGED;
4031 }
4032
4033 return indicatePessimisticFixpoint();
4034 }
4035
4036 /// See AbstractAttribute::trackStatistics()
4037 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(noalias) }
4038};
4039
4040/// NoAlias attribute for function return value.
4041struct AANoAliasReturned final : AANoAliasImpl {
4042 AANoAliasReturned(const IRPosition &IRP, Attributor &A)
4043 : AANoAliasImpl(IRP, A) {}
4044
4045 /// See AbstractAttribute::updateImpl(...).
4046 ChangeStatus updateImpl(Attributor &A) override {
4047
4048 auto CheckReturnValue = [&](Value &RV) -> bool {
4049 if (Constant *C = dyn_cast<Constant>(&RV))
4050 if (C->isNullValue() || isa<UndefValue>(C))
4051 return true;
4052
4053 /// For now, we can only deduce noalias if we have call sites.
4054 /// FIXME: add more support.
4055 if (!isa<CallBase>(&RV))
4056 return false;
4057
4058 const IRPosition &RVPos = IRPosition::value(RV);
4059 bool IsKnownNoAlias;
4061 A, this, RVPos, DepClassTy::REQUIRED, IsKnownNoAlias))
4062 return false;
4063
4064 bool IsKnownNoCapture;
4065 const AANoCapture *NoCaptureAA = nullptr;
4066 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
4067 A, this, RVPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
4068 &NoCaptureAA);
4069 return IsAssumedNoCapture ||
4070 (NoCaptureAA && NoCaptureAA->isAssumedNoCaptureMaybeReturned());
4071 };
4072
4073 if (!A.checkForAllReturnedValues(CheckReturnValue, *this))
4074 return indicatePessimisticFixpoint();
4075
4076 return ChangeStatus::UNCHANGED;
4077 }
4078
4079 /// See AbstractAttribute::trackStatistics()
4080 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noalias) }
4081};
4082
4083/// NoAlias attribute deduction for a call site return value.
4084struct AANoAliasCallSiteReturned final
4085 : AACalleeToCallSite<AANoAlias, AANoAliasImpl> {
4086 AANoAliasCallSiteReturned(const IRPosition &IRP, Attributor &A)
4087 : AACalleeToCallSite<AANoAlias, AANoAliasImpl>(IRP, A) {}
4088
4089 /// See AbstractAttribute::trackStatistics()
4090 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(noalias); }
4091};
4092} // namespace
4093
4094/// -------------------AAIsDead Function Attribute-----------------------
4095
4096namespace {
4097struct AAIsDeadValueImpl : public AAIsDead {
4098 AAIsDeadValueImpl(const IRPosition &IRP, Attributor &A) : AAIsDead(IRP, A) {}
4099
4100 /// See AAIsDead::isAssumedDead().
4101 bool isAssumedDead() const override { return isAssumed(IS_DEAD); }
4102
4103 /// See AAIsDead::isKnownDead().
4104 bool isKnownDead() const override { return isKnown(IS_DEAD); }
4105
4106 /// See AAIsDead::isAssumedDead(BasicBlock *).
4107 bool isAssumedDead(const BasicBlock *BB) const override { return false; }
4108
4109 /// See AAIsDead::isKnownDead(BasicBlock *).
4110 bool isKnownDead(const BasicBlock *BB) const override { return false; }
4111
4112 /// See AAIsDead::isAssumedDead(Instruction *I).
4113 bool isAssumedDead(const Instruction *I) const override {
4114 return I == getCtxI() && isAssumedDead();
4115 }
4116
4117 /// See AAIsDead::isKnownDead(Instruction *I).
4118 bool isKnownDead(const Instruction *I) const override {
4119 return isAssumedDead(I) && isKnownDead();
4120 }
4121
4122 /// See AbstractAttribute::getAsStr().
4123 const std::string getAsStr(Attributor *A) const override {
4124 return isAssumedDead() ? "assumed-dead" : "assumed-live";
4125 }
4126
4127 /// Check if all uses are assumed dead.
4128 bool areAllUsesAssumedDead(Attributor &A, Value &V) {
4129 // Callers might not check the type, void has no uses.
4130 if (V.getType()->isVoidTy() || V.use_empty())
4131 return true;
4132
4133 // If we replace a value with a constant there are no uses left afterwards.
4134 if (!isa<Constant>(V)) {
4135 if (auto *I = dyn_cast<Instruction>(&V))
4136 if (!A.isRunOn(*I->getFunction()))
4137 return false;
4138 bool UsedAssumedInformation = false;
4139 std::optional<Constant *> C =
4140 A.getAssumedConstant(V, *this, UsedAssumedInformation);
4141 if (!C || *C)
4142 return true;
4143 }
4144
4145 auto UsePred = [&](const Use &U, bool &Follow) { return false; };
4146 // Explicitly set the dependence class to required because we want a long
4147 // chain of N dependent instructions to be considered live as soon as one is
4148 // without going through N update cycles. This is not required for
4149 // correctness.
4150 return A.checkForAllUses(UsePred, *this, V, /* CheckBBLivenessOnly */ false,
4151 DepClassTy::REQUIRED,
4152 /* IgnoreDroppableUses */ false);
4153 }
4154
4155 /// Determine if \p I is assumed to be side-effect free.
4156 bool isAssumedSideEffectFree(Attributor &A, Instruction *I) {
4158 return true;
4159
4160 auto *CB = dyn_cast<CallBase>(I);
4161 if (!CB || isa<IntrinsicInst>(CB))
4162 return false;
4163
4164 const IRPosition &CallIRP = IRPosition::callsite_function(*CB);
4165
4166 bool IsKnownNoUnwind;
4168 A, this, CallIRP, DepClassTy::OPTIONAL, IsKnownNoUnwind))
4169 return false;
4170
4171 bool IsKnown;
4172 return AA::isAssumedReadOnly(A, CallIRP, *this, IsKnown);
4173 }
4174};
4175
4176struct AAIsDeadFloating : public AAIsDeadValueImpl {
4177 AAIsDeadFloating(const IRPosition &IRP, Attributor &A)
4178 : AAIsDeadValueImpl(IRP, A) {}
4179
4180 /// See AbstractAttribute::initialize(...).
4181 void initialize(Attributor &A) override {
4182 AAIsDeadValueImpl::initialize(A);
4183
4184 if (isa<UndefValue>(getAssociatedValue())) {
4185 indicatePessimisticFixpoint();
4186 return;
4187 }
4188
4189 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4190 if (!isAssumedSideEffectFree(A, I)) {
4192 indicatePessimisticFixpoint();
4193 else
4194 removeAssumedBits(HAS_NO_EFFECT);
4195 }
4196 }
4197
4198 bool isDeadFence(Attributor &A, FenceInst &FI) {
4199 const auto *ExecDomainAA = A.lookupAAFor<AAExecutionDomain>(
4200 IRPosition::function(*FI.getFunction()), *this, DepClassTy::NONE);
4201 if (!ExecDomainAA || !ExecDomainAA->isNoOpFence(FI))
4202 return false;
4203 A.recordDependence(*ExecDomainAA, *this, DepClassTy::OPTIONAL);
4204 return true;
4205 }
4206
4207 bool isDeadStore(Attributor &A, StoreInst &SI,
4208 SmallSetVector<Instruction *, 8> *AssumeOnlyInst = nullptr) {
4209 // Lang ref now states volatile store is not UB/dead, let's skip them.
4210 if (SI.isVolatile())
4211 return false;
4212
4213 // If we are collecting assumes to be deleted we are in the manifest stage.
4214 // It's problematic to collect the potential copies again now so we use the
4215 // cached ones.
4216 bool UsedAssumedInformation = false;
4217 if (!AssumeOnlyInst) {
4218 PotentialCopies.clear();
4219 if (!AA::getPotentialCopiesOfStoredValue(A, SI, PotentialCopies, *this,
4220 UsedAssumedInformation)) {
4221 LLVM_DEBUG(
4222 dbgs()
4223 << "[AAIsDead] Could not determine potential copies of store!\n");
4224 return false;
4225 }
4226 }
4227 LLVM_DEBUG(dbgs() << "[AAIsDead] Store has " << PotentialCopies.size()
4228 << " potential copies.\n");
4229
4230 InformationCache &InfoCache = A.getInfoCache();
4231 return llvm::all_of(PotentialCopies, [&](Value *V) {
4232 if (A.isAssumedDead(IRPosition::value(*V), this, nullptr,
4233 UsedAssumedInformation))
4234 return true;
4235 if (auto *LI = dyn_cast<LoadInst>(V)) {
4236 if (llvm::all_of(LI->uses(), [&](const Use &U) {
4237 auto &UserI = cast<Instruction>(*U.getUser());
4238 if (InfoCache.isOnlyUsedByAssume(UserI)) {
4239 if (AssumeOnlyInst)
4240 AssumeOnlyInst->insert(&UserI);
4241 return true;
4242 }
4243 return A.isAssumedDead(U, this, nullptr, UsedAssumedInformation);
4244 })) {
4245 return true;
4246 }
4247 }
4248 LLVM_DEBUG(dbgs() << "[AAIsDead] Potential copy " << *V
4249 << " is assumed live!\n");
4250 return false;
4251 });
4252 }
4253
4254 /// See AbstractAttribute::getAsStr().
4255 const std::string getAsStr(Attributor *A) const override {
4256 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4258 if (isValidState())
4259 return "assumed-dead-store";
4261 if (isValidState())
4262 return "assumed-dead-fence";
4263 return AAIsDeadValueImpl::getAsStr(A);
4264 }
4265
4266 /// See AbstractAttribute::updateImpl(...).
4267 ChangeStatus updateImpl(Attributor &A) override {
4268 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4269 if (auto *SI = dyn_cast_or_null<StoreInst>(I)) {
4270 if (!isDeadStore(A, *SI))
4271 return indicatePessimisticFixpoint();
4272 } else if (auto *FI = dyn_cast_or_null<FenceInst>(I)) {
4273 if (!isDeadFence(A, *FI))
4274 return indicatePessimisticFixpoint();
4275 } else {
4276 if (!isAssumedSideEffectFree(A, I))
4277 return indicatePessimisticFixpoint();
4278 if (!areAllUsesAssumedDead(A, getAssociatedValue()))
4279 return indicatePessimisticFixpoint();
4280 }
4282 }
4283
4284 bool isRemovableStore() const override {
4285 return isAssumed(IS_REMOVABLE) && isa<StoreInst>(&getAssociatedValue());
4286 }
4287
4288 /// See AbstractAttribute::manifest(...).
4289 ChangeStatus manifest(Attributor &A) override {
4290 Value &V = getAssociatedValue();
4291 if (auto *I = dyn_cast<Instruction>(&V)) {
4292 // If we get here we basically know the users are all dead. We check if
4293 // isAssumedSideEffectFree returns true here again because it might not be
4294 // the case and only the users are dead but the instruction (=call) is
4295 // still needed.
4296 if (auto *SI = dyn_cast<StoreInst>(I)) {
4297 SmallSetVector<Instruction *, 8> AssumeOnlyInst;
4298 bool IsDead = isDeadStore(A, *SI, &AssumeOnlyInst);
4299 (void)IsDead;
4300 assert(IsDead && "Store was assumed to be dead!");
4301 A.deleteAfterManifest(*I);
4302 for (size_t i = 0; i < AssumeOnlyInst.size(); ++i) {
4303 Instruction *AOI = AssumeOnlyInst[i];
4304 for (auto *Usr : AOI->users())
4305 AssumeOnlyInst.insert(cast<Instruction>(Usr));
4306 A.deleteAfterManifest(*AOI);
4307 }
4308 return ChangeStatus::CHANGED;
4309 }
4310 if (auto *FI = dyn_cast<FenceInst>(I)) {
4311 assert(isDeadFence(A, *FI));
4312 A.deleteAfterManifest(*FI);
4313 return ChangeStatus::CHANGED;
4314 }
4315 if (isAssumedSideEffectFree(A, I) && !isa<InvokeInst>(I)) {
4316 A.deleteAfterManifest(*I);
4317 return ChangeStatus::CHANGED;
4318 }
4319 }
4321 }
4322
4323 /// See AbstractAttribute::trackStatistics()
4324 void trackStatistics() const override {
4326 }
4327
4328private:
4329 // The potential copies of a dead store, used for deletion during manifest.
4330 SmallSetVector<Value *, 4> PotentialCopies;
4331};
4332
4333struct AAIsDeadArgument : public AAIsDeadFloating {
4334 AAIsDeadArgument(const IRPosition &IRP, Attributor &A)
4335 : AAIsDeadFloating(IRP, A) {}
4336
4337 /// See AbstractAttribute::manifest(...).
4338 ChangeStatus manifest(Attributor &A) override {
4339 Argument &Arg = *getAssociatedArgument();
4340 if (A.isValidFunctionSignatureRewrite(Arg, /* ReplacementTypes */ {}))
4341 if (A.registerFunctionSignatureRewrite(
4342 Arg, /* ReplacementTypes */ {},
4345 return ChangeStatus::CHANGED;
4346 }
4347 return ChangeStatus::UNCHANGED;
4348 }
4349
4350 /// See AbstractAttribute::trackStatistics()
4351 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(IsDead) }
4352};
4353
4354struct AAIsDeadCallSiteArgument : public AAIsDeadValueImpl {
4355 AAIsDeadCallSiteArgument(const IRPosition &IRP, Attributor &A)
4356 : AAIsDeadValueImpl(IRP, A) {}
4357
4358 /// See AbstractAttribute::initialize(...).
4359 void initialize(Attributor &A) override {
4360 AAIsDeadValueImpl::initialize(A);
4361 if (isa<UndefValue>(getAssociatedValue()))
4362 indicatePessimisticFixpoint();
4363 }
4364
4365 /// See AbstractAttribute::updateImpl(...).
4366 ChangeStatus updateImpl(Attributor &A) override {
4367 // TODO: Once we have call site specific value information we can provide
4368 // call site specific liveness information and then it makes
4369 // sense to specialize attributes for call sites arguments instead of
4370 // redirecting requests to the callee argument.
4371 Argument *Arg = getAssociatedArgument();
4372 if (!Arg)
4373 return indicatePessimisticFixpoint();
4374 const IRPosition &ArgPos = IRPosition::argument(*Arg);
4375 auto *ArgAA = A.getAAFor<AAIsDead>(*this, ArgPos, DepClassTy::REQUIRED);
4376 if (!ArgAA)
4377 return indicatePessimisticFixpoint();
4378 return clampStateAndIndicateChange(getState(), ArgAA->getState());
4379 }
4380
4381 /// See AbstractAttribute::manifest(...).
4382 ChangeStatus manifest(Attributor &A) override {
4383 CallBase &CB = cast<CallBase>(getAnchorValue());
4384 Use &U = CB.getArgOperandUse(getCallSiteArgNo());
4385 assert(!isa<UndefValue>(U.get()) &&
4386 "Expected undef values to be filtered out!");
4387 UndefValue &UV = *UndefValue::get(U->getType());
4388 if (A.changeUseAfterManifest(U, UV))
4389 return ChangeStatus::CHANGED;
4390 return ChangeStatus::UNCHANGED;
4391 }
4392
4393 /// See AbstractAttribute::trackStatistics()
4394 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(IsDead) }
4395};
4396
4397struct AAIsDeadCallSiteReturned : public AAIsDeadFloating {
4398 AAIsDeadCallSiteReturned(const IRPosition &IRP, Attributor &A)
4399 : AAIsDeadFloating(IRP, A) {}
4400
4401 /// See AAIsDead::isAssumedDead().
4402 bool isAssumedDead() const override {
4403 return AAIsDeadFloating::isAssumedDead() && IsAssumedSideEffectFree;
4404 }
4405
4406 /// See AbstractAttribute::initialize(...).
4407 void initialize(Attributor &A) override {
4408 AAIsDeadFloating::initialize(A);
4409 if (isa<UndefValue>(getAssociatedValue())) {
4410 indicatePessimisticFixpoint();
4411 return;
4412 }
4413
4414 // We track this separately as a secondary state.
4415 IsAssumedSideEffectFree = isAssumedSideEffectFree(A, getCtxI());
4416 }
4417
4418 /// See AbstractAttribute::updateImpl(...).
4419 ChangeStatus updateImpl(Attributor &A) override {
4420 ChangeStatus Changed = ChangeStatus::UNCHANGED;
4421 if (IsAssumedSideEffectFree && !isAssumedSideEffectFree(A, getCtxI())) {
4422 IsAssumedSideEffectFree = false;
4423 Changed = ChangeStatus::CHANGED;
4424 }
4425 if (!areAllUsesAssumedDead(A, getAssociatedValue()))
4426 return indicatePessimisticFixpoint();
4427 return Changed;
4428 }
4429
4430 /// See AbstractAttribute::trackStatistics()
4431 void trackStatistics() const override {
4432 if (IsAssumedSideEffectFree)
4434 else
4435 STATS_DECLTRACK_CSRET_ATTR(UnusedResult)
4436 }
4437
4438 /// See AbstractAttribute::getAsStr().
4439 const std::string getAsStr(Attributor *A) const override {
4440 return isAssumedDead()
4441 ? "assumed-dead"
4442 : (getAssumed() ? "assumed-dead-users" : "assumed-live");
4443 }
4444
4445private:
4446 bool IsAssumedSideEffectFree = true;
4447};
4448
4449struct AAIsDeadReturned : public AAIsDeadValueImpl {
4450 AAIsDeadReturned(const IRPosition &IRP, Attributor &A)
4451 : AAIsDeadValueImpl(IRP, A) {}
4452
4453 /// See AbstractAttribute::updateImpl(...).
4454 ChangeStatus updateImpl(Attributor &A) override {
4455
4456 bool UsedAssumedInformation = false;
4457 A.checkForAllInstructions([](Instruction &) { return true; }, *this,
4458 {Instruction::Ret}, UsedAssumedInformation);
4459
4460 auto PredForCallSite = [&](AbstractCallSite ACS) {
4461 if (ACS.isCallbackCall() || !ACS.getInstruction())
4462 return false;
4463 return areAllUsesAssumedDead(A, *ACS.getInstruction());
4464 };
4465
4466 if (!A.checkForAllCallSites(PredForCallSite, *this, true,
4467 UsedAssumedInformation))
4468 return indicatePessimisticFixpoint();
4469
4470 return ChangeStatus::UNCHANGED;
4471 }
4472
4473 /// See AbstractAttribute::manifest(...).
4474 ChangeStatus manifest(Attributor &A) override {
4475 // TODO: Rewrite the signature to return void?
4476 bool AnyChange = false;
4477 UndefValue &UV = *UndefValue::get(getAssociatedFunction()->getReturnType());
4478 auto RetInstPred = [&](Instruction &I) {
4479 ReturnInst &RI = cast<ReturnInst>(I);
4481 AnyChange |= A.changeUseAfterManifest(RI.getOperandUse(0), UV);
4482 return true;
4483 };
4484 bool UsedAssumedInformation = false;
4485 A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
4486 UsedAssumedInformation);
4487 return AnyChange ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
4488 }
4489
4490 /// See AbstractAttribute::trackStatistics()
4491 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(IsDead) }
4492};
4493
4494struct AAIsDeadFunction : public AAIsDead {
4495 AAIsDeadFunction(const IRPosition &IRP, Attributor &A) : AAIsDead(IRP, A) {}
4496
4497 /// See AbstractAttribute::initialize(...).
4498 void initialize(Attributor &A) override {
4499 Function *F = getAnchorScope();
4500 assert(F && "Did expect an anchor function");
4501 if (!isAssumedDeadInternalFunction(A)) {
4502 ToBeExploredFrom.insert(&F->getEntryBlock().front());
4503 assumeLive(A, F->getEntryBlock());
4504 }
4505 }
4506
4507 bool isAssumedDeadInternalFunction(Attributor &A) {
4508 if (!getAnchorScope()->hasLocalLinkage())
4509 return false;
4510 bool UsedAssumedInformation = false;
4511 return A.checkForAllCallSites([](AbstractCallSite) { return false; }, *this,
4512 true, UsedAssumedInformation);
4513 }
4514
4515 /// See AbstractAttribute::getAsStr().
4516 const std::string getAsStr(Attributor *A) const override {
4517 return "Live[#BB " + std::to_string(AssumedLiveBlocks.size()) + "/" +
4518 std::to_string(getAnchorScope()->size()) + "][#TBEP " +
4519 std::to_string(ToBeExploredFrom.size()) + "][#KDE " +
4520 std::to_string(KnownDeadEnds.size()) + "]";
4521 }
4522
4523 /// See AbstractAttribute::manifest(...).
4524 ChangeStatus manifest(Attributor &A) override {
4525 assert(getState().isValidState() &&
4526 "Attempted to manifest an invalid state!");
4527
4528 ChangeStatus HasChanged = ChangeStatus::UNCHANGED;
4529 Function &F = *getAnchorScope();
4530
4531 if (AssumedLiveBlocks.empty()) {
4532 A.deleteAfterManifest(F);
4533 return ChangeStatus::CHANGED;
4534 }
4535
4536 // Flag to determine if we can change an invoke to a call assuming the
4537 // callee is nounwind. This is not possible if the personality of the
4538 // function allows to catch asynchronous exceptions.
4539 bool Invoke2CallAllowed = !mayCatchAsynchronousExceptions(F);
4540
4541 KnownDeadEnds.set_union(ToBeExploredFrom);
4542 for (const Instruction *DeadEndI : KnownDeadEnds) {
4543 auto *CB = dyn_cast<CallBase>(DeadEndI);
4544 if (!CB)
4545 continue;
4546 bool IsKnownNoReturn;
4548 A, this, IRPosition::callsite_function(*CB), DepClassTy::OPTIONAL,
4549 IsKnownNoReturn);
4550 if (MayReturn && (!Invoke2CallAllowed || !isa<InvokeInst>(CB)))
4551 continue;
4552
4553 if (auto *II = dyn_cast<InvokeInst>(DeadEndI))
4554 A.registerInvokeWithDeadSuccessor(const_cast<InvokeInst &>(*II));
4555 else
4556 A.changeToUnreachableAfterManifest(
4557 const_cast<Instruction *>(DeadEndI->getNextNode()));
4558 HasChanged = ChangeStatus::CHANGED;
4559 }
4560
4561 STATS_DECL(AAIsDead, BasicBlock, "Number of dead basic blocks deleted.");
4562 for (BasicBlock &BB : F)
4563 if (!AssumedLiveBlocks.count(&BB)) {
4564 A.deleteAfterManifest(BB);
4565 ++BUILD_STAT_NAME(AAIsDead, BasicBlock);
4566 HasChanged = ChangeStatus::CHANGED;
4567 }
4568
4569 return HasChanged;
4570 }
4571
4572 /// See AbstractAttribute::updateImpl(...).
4573 ChangeStatus updateImpl(Attributor &A) override;
4574
4575 bool isEdgeDead(const BasicBlock *From, const BasicBlock *To) const override {
4576 assert(From->getParent() == getAnchorScope() &&
4577 To->getParent() == getAnchorScope() &&
4578 "Used AAIsDead of the wrong function");
4579 return isValidState() && !AssumedLiveEdges.count(std::make_pair(From, To));
4580 }
4581
4582 /// See AbstractAttribute::trackStatistics()
4583 void trackStatistics() const override {}
4584
4585 /// Returns true if the function is assumed dead.
4586 bool isAssumedDead() const override { return false; }
4587
4588 /// See AAIsDead::isKnownDead().
4589 bool isKnownDead() const override { return false; }
4590
4591 /// See AAIsDead::isAssumedDead(BasicBlock *).
4592 bool isAssumedDead(const BasicBlock *BB) const override {
4593 assert(BB->getParent() == getAnchorScope() &&
4594 "BB must be in the same anchor scope function.");
4595
4596 if (!getAssumed())
4597 return false;
4598 return !AssumedLiveBlocks.count(BB);
4599 }
4600
4601 /// See AAIsDead::isKnownDead(BasicBlock *).
4602 bool isKnownDead(const BasicBlock *BB) const override {
4603 return getKnown() && isAssumedDead(BB);
4604 }
4605
4606 /// See AAIsDead::isAssumed(Instruction *I).
4607 bool isAssumedDead(const Instruction *I) const override {
4608 assert(I->getParent()->getParent() == getAnchorScope() &&
4609 "Instruction must be in the same anchor scope function.");
4610
4611 if (!getAssumed())
4612 return false;
4613
4614 // If it is not in AssumedLiveBlocks then it for sure dead.
4615 // Otherwise, it can still be after noreturn call in a live block.
4616 if (!AssumedLiveBlocks.count(I->getParent()))
4617 return true;
4618
4619 // If it is not after a liveness barrier it is live.
4620 const Instruction *PrevI = I->getPrevNode();
4621 while (PrevI) {
4622 if (KnownDeadEnds.count(PrevI) || ToBeExploredFrom.count(PrevI))
4623 return true;
4624 PrevI = PrevI->getPrevNode();
4625 }
4626 return false;
4627 }
4628
4629 /// See AAIsDead::isKnownDead(Instruction *I).
4630 bool isKnownDead(const Instruction *I) const override {
4631 return getKnown() && isAssumedDead(I);
4632 }
4633
4634 /// Assume \p BB is (partially) live now and indicate to the Attributor \p A
4635 /// that internal function called from \p BB should now be looked at.
4636 bool assumeLive(Attributor &A, const BasicBlock &BB) {
4637 if (!AssumedLiveBlocks.insert(&BB).second)
4638 return false;
4639
4640 // We assume that all of BB is (probably) live now and if there are calls to
4641 // internal functions we will assume that those are now live as well. This
4642 // is a performance optimization for blocks with calls to a lot of internal
4643 // functions. It can however cause dead functions to be treated as live.
4644 for (const Instruction &I : BB)
4645 if (const auto *CB = dyn_cast<CallBase>(&I))
4647 if (F->hasLocalLinkage())
4648 A.markLiveInternalFunction(*F);
4649 return true;
4650 }
4651
4652 /// Collection of instructions that need to be explored again, e.g., we
4653 /// did assume they do not transfer control to (one of their) successors.
4654 SmallSetVector<const Instruction *, 8> ToBeExploredFrom;
4655
4656 /// Collection of instructions that are known to not transfer control.
4657 SmallSetVector<const Instruction *, 8> KnownDeadEnds;
4658
4659 /// Collection of all assumed live edges
4660 DenseSet<std::pair<const BasicBlock *, const BasicBlock *>> AssumedLiveEdges;
4661
4662 /// Collection of all assumed live BasicBlocks.
4663 DenseSet<const BasicBlock *> AssumedLiveBlocks;
4664};
4665
4666static bool
4667identifyAliveSuccessors(Attributor &A, const CallBase &CB,
4668 AbstractAttribute &AA,
4669 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4670 const IRPosition &IPos = IRPosition::callsite_function(CB);
4671
4672 bool IsKnownNoReturn;
4674 A, &AA, IPos, DepClassTy::OPTIONAL, IsKnownNoReturn))
4675 return !IsKnownNoReturn;
4676 if (CB.isTerminator())
4677 AliveSuccessors.push_back(&CB.getSuccessor(0)->front());
4678 else
4679 AliveSuccessors.push_back(CB.getNextNode());
4680 return false;
4681}
4682
4683static bool
4684identifyAliveSuccessors(Attributor &A, const InvokeInst &II,
4685 AbstractAttribute &AA,
4686 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4687 bool UsedAssumedInformation =
4688 identifyAliveSuccessors(A, cast<CallBase>(II), AA, AliveSuccessors);
4689
4690 // First, determine if we can change an invoke to a call assuming the
4691 // callee is nounwind. This is not possible if the personality of the
4692 // function allows to catch asynchronous exceptions.
4693 if (AAIsDeadFunction::mayCatchAsynchronousExceptions(*II.getFunction())) {
4694 AliveSuccessors.push_back(&II.getUnwindDest()->front());
4695 } else {
4696 const IRPosition &IPos = IRPosition::callsite_function(II);
4697
4698 bool IsKnownNoUnwind;
4700 A, &AA, IPos, DepClassTy::OPTIONAL, IsKnownNoUnwind)) {
4701 UsedAssumedInformation |= !IsKnownNoUnwind;
4702 } else {
4703 AliveSuccessors.push_back(&II.getUnwindDest()->front());
4704 }
4705 }
4706 return UsedAssumedInformation;
4707}
4708
4709static bool
4710identifyAliveSuccessors(Attributor &A, const BranchInst &BI,
4711 AbstractAttribute &AA,
4712 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4713 bool UsedAssumedInformation = false;
4714 if (BI.getNumSuccessors() == 1) {
4715 AliveSuccessors.push_back(&BI.getSuccessor(0)->front());
4716 } else {
4717 std::optional<Constant *> C =
4718 A.getAssumedConstant(*BI.getCondition(), AA, UsedAssumedInformation);
4719 if (!C || isa_and_nonnull<UndefValue>(*C)) {
4720 // No value yet, assume both edges are dead.
4721 } else if (isa_and_nonnull<ConstantInt>(*C)) {
4722 const BasicBlock *SuccBB =
4723 BI.getSuccessor(1 - cast<ConstantInt>(*C)->getValue().getZExtValue());
4724 AliveSuccessors.push_back(&SuccBB->front());
4725 } else {
4726 AliveSuccessors.push_back(&BI.getSuccessor(0)->front());
4727 AliveSuccessors.push_back(&BI.getSuccessor(1)->front());
4728 UsedAssumedInformation = false;
4729 }
4730 }
4731 return UsedAssumedInformation;
4732}
4733
4734static bool
4735identifyAliveSuccessors(Attributor &A, const SwitchInst &SI,
4736 AbstractAttribute &AA,
4737 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4738 bool UsedAssumedInformation = false;
4740 if (!A.getAssumedSimplifiedValues(IRPosition::value(*SI.getCondition()), &AA,
4741 Values, AA::AnyScope,
4742 UsedAssumedInformation)) {
4743 // Something went wrong, assume all successors are live.
4744 for (const BasicBlock *SuccBB : successors(SI.getParent()))
4745 AliveSuccessors.push_back(&SuccBB->front());
4746 return false;
4747 }
4748
4749 if (Values.empty() ||
4750 (Values.size() == 1 &&
4751 isa_and_nonnull<UndefValue>(Values.front().getValue()))) {
4752 // No valid value yet, assume all edges are dead.
4753 return UsedAssumedInformation;
4754 }
4755
4756 Type &Ty = *SI.getCondition()->getType();
4757 SmallPtrSet<ConstantInt *, 8> Constants;
4758 auto CheckForConstantInt = [&](Value *V) {
4759 if (auto *CI = dyn_cast_if_present<ConstantInt>(AA::getWithType(*V, Ty))) {
4760 Constants.insert(CI);
4761 return true;
4762 }
4763 return false;
4764 };
4765
4766 if (!all_of(Values, [&](AA::ValueAndContext &VAC) {
4767 return CheckForConstantInt(VAC.getValue());
4768 })) {
4769 for (const BasicBlock *SuccBB : successors(SI.getParent()))
4770 AliveSuccessors.push_back(&SuccBB->front());
4771 return UsedAssumedInformation;
4772 }
4773
4774 unsigned MatchedCases = 0;
4775 for (const auto &CaseIt : SI.cases()) {
4776 if (Constants.count(CaseIt.getCaseValue())) {
4777 ++MatchedCases;
4778 AliveSuccessors.push_back(&CaseIt.getCaseSuccessor()->front());
4779 }
4780 }
4781
4782 // If all potential values have been matched, we will not visit the default
4783 // case.
4784 if (MatchedCases < Constants.size())
4785 AliveSuccessors.push_back(&SI.getDefaultDest()->front());
4786 return UsedAssumedInformation;
4787}
4788
4789ChangeStatus AAIsDeadFunction::updateImpl(Attributor &A) {
4791
4792 if (AssumedLiveBlocks.empty()) {
4793 if (isAssumedDeadInternalFunction(A))
4795
4796 Function *F = getAnchorScope();
4797 ToBeExploredFrom.insert(&F->getEntryBlock().front());
4798 assumeLive(A, F->getEntryBlock());
4799 Change = ChangeStatus::CHANGED;
4800 }
4801
4802 LLVM_DEBUG(dbgs() << "[AAIsDead] Live [" << AssumedLiveBlocks.size() << "/"
4803 << getAnchorScope()->size() << "] BBs and "
4804 << ToBeExploredFrom.size() << " exploration points and "
4805 << KnownDeadEnds.size() << " known dead ends\n");
4806
4807 // Copy and clear the list of instructions we need to explore from. It is
4808 // refilled with instructions the next update has to look at.
4809 SmallVector<const Instruction *, 8> Worklist(ToBeExploredFrom.begin(),
4810 ToBeExploredFrom.end());
4811 decltype(ToBeExploredFrom) NewToBeExploredFrom;
4812
4814 while (!Worklist.empty()) {
4815 const Instruction *I = Worklist.pop_back_val();
4816 LLVM_DEBUG(dbgs() << "[AAIsDead] Exploration inst: " << *I << "\n");
4817
4818 // Fast forward for uninteresting instructions. We could look for UB here
4819 // though.
4820 while (!I->isTerminator() && !isa<CallBase>(I))
4821 I = I->getNextNode();
4822
4823 AliveSuccessors.clear();
4824
4825 bool UsedAssumedInformation = false;
4826 switch (I->getOpcode()) {
4827 // TODO: look for (assumed) UB to backwards propagate "deadness".
4828 default:
4829 assert(I->isTerminator() &&
4830 "Expected non-terminators to be handled already!");
4831 for (const BasicBlock *SuccBB : successors(I->getParent()))
4832 AliveSuccessors.push_back(&SuccBB->front());
4833 break;
4834 case Instruction::Call:
4835 UsedAssumedInformation = identifyAliveSuccessors(A, cast<CallInst>(*I),
4836 *this, AliveSuccessors);
4837 break;
4838 case Instruction::Invoke:
4839 UsedAssumedInformation = identifyAliveSuccessors(A, cast<InvokeInst>(*I),
4840 *this, AliveSuccessors);
4841 break;
4842 case Instruction::Br:
4843 UsedAssumedInformation = identifyAliveSuccessors(A, cast<BranchInst>(*I),
4844 *this, AliveSuccessors);
4845 break;
4846 case Instruction::Switch:
4847 UsedAssumedInformation = identifyAliveSuccessors(A, cast<SwitchInst>(*I),
4848 *this, AliveSuccessors);
4849 break;
4850 }
4851
4852 if (UsedAssumedInformation) {
4853 NewToBeExploredFrom.insert(I);
4854 } else if (AliveSuccessors.empty() ||
4855 (I->isTerminator() &&
4856 AliveSuccessors.size() < I->getNumSuccessors())) {
4857 if (KnownDeadEnds.insert(I))
4858 Change = ChangeStatus::CHANGED;
4859 }
4860
4861 LLVM_DEBUG(dbgs() << "[AAIsDead] #AliveSuccessors: "
4862 << AliveSuccessors.size() << " UsedAssumedInformation: "
4863 << UsedAssumedInformation << "\n");
4864
4865 for (const Instruction *AliveSuccessor : AliveSuccessors) {
4866 if (!I->isTerminator()) {
4867 assert(AliveSuccessors.size() == 1 &&
4868 "Non-terminator expected to have a single successor!");
4869 Worklist.push_back(AliveSuccessor);
4870 } else {
4871 // record the assumed live edge
4872 auto Edge = std::make_pair(I->getParent(), AliveSuccessor->getParent());
4873 if (AssumedLiveEdges.insert(Edge).second)
4874 Change = ChangeStatus::CHANGED;
4875 if (assumeLive(A, *AliveSuccessor->getParent()))
4876 Worklist.push_back(AliveSuccessor);
4877 }
4878 }
4879 }
4880
4881 // Check if the content of ToBeExploredFrom changed, ignore the order.
4882 if (NewToBeExploredFrom.size() != ToBeExploredFrom.size() ||
4883 llvm::any_of(NewToBeExploredFrom, [&](const Instruction *I) {
4884 return !ToBeExploredFrom.count(I);
4885 })) {
4886 Change = ChangeStatus::CHANGED;
4887 ToBeExploredFrom = std::move(NewToBeExploredFrom);
4888 }
4889
4890 // If we know everything is live there is no need to query for liveness.
4891 // Instead, indicating a pessimistic fixpoint will cause the state to be
4892 // "invalid" and all queries to be answered conservatively without lookups.
4893 // To be in this state we have to (1) finished the exploration and (3) not
4894 // discovered any non-trivial dead end and (2) not ruled unreachable code
4895 // dead.
4896 if (ToBeExploredFrom.empty() &&
4897 getAnchorScope()->size() == AssumedLiveBlocks.size() &&
4898 llvm::all_of(KnownDeadEnds, [](const Instruction *DeadEndI) {
4899 return DeadEndI->isTerminator() && DeadEndI->getNumSuccessors() == 0;
4900 }))
4901 return indicatePessimisticFixpoint();
4902 return Change;
4903}
4904
4905/// Liveness information for a call sites.
4906struct AAIsDeadCallSite final : AAIsDeadFunction {
4907 AAIsDeadCallSite(const IRPosition &IRP, Attributor &A)
4908 : AAIsDeadFunction(IRP, A) {}
4909
4910 /// See AbstractAttribute::initialize(...).
4911 void initialize(Attributor &A) override {
4912 // TODO: Once we have call site specific value information we can provide
4913 // call site specific liveness information and then it makes
4914 // sense to specialize attributes for call sites instead of
4915 // redirecting requests to the callee.
4916 llvm_unreachable("Abstract attributes for liveness are not "
4917 "supported for call sites yet!");
4918 }
4919
4920 /// See AbstractAttribute::updateImpl(...).
4921 ChangeStatus updateImpl(Attributor &A) override {
4922 return indicatePessimisticFixpoint();
4923 }
4924
4925 /// See AbstractAttribute::trackStatistics()
4926 void trackStatistics() const override {}
4927};
4928} // namespace
4929
4930/// -------------------- Dereferenceable Argument Attribute --------------------
4931
4932namespace {
4933struct AADereferenceableImpl : AADereferenceable {
4934 AADereferenceableImpl(const IRPosition &IRP, Attributor &A)
4935 : AADereferenceable(IRP, A) {}
4936 using StateType = DerefState;
4937
4938 /// See AbstractAttribute::initialize(...).
4939 void initialize(Attributor &A) override {
4940 Value &V = *getAssociatedValue().stripPointerCasts();
4942 A.getAttrs(getIRPosition(),
4943 {Attribute::Dereferenceable, Attribute::DereferenceableOrNull},
4944 Attrs, /* IgnoreSubsumingPositions */ false);
4945 for (const Attribute &Attr : Attrs)
4946 takeKnownDerefBytesMaximum(Attr.getValueAsInt());
4947
4948 // Ensure we initialize the non-null AA (if necessary).
4949 bool IsKnownNonNull;
4951 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNonNull);
4952
4953 bool CanBeNull, CanBeFreed;
4954 takeKnownDerefBytesMaximum(V.getPointerDereferenceableBytes(
4955 A.getDataLayout(), CanBeNull, CanBeFreed));
4956
4957 if (Instruction *CtxI = getCtxI())
4958 followUsesInMBEC(*this, A, getState(), *CtxI);
4959 }
4960
4961 /// See AbstractAttribute::getState()
4962 /// {
4963 StateType &getState() override { return *this; }
4964 const StateType &getState() const override { return *this; }
4965 /// }
4966
4967 /// Helper function for collecting accessed bytes in must-be-executed-context
4968 void addAccessedBytesForUse(Attributor &A, const Use *U, const Instruction *I,
4969 DerefState &State) {
4970 const Value *UseV = U->get();
4971 if (!UseV->getType()->isPointerTy())
4972 return;
4973
4974 std::optional<MemoryLocation> Loc = MemoryLocation::getOrNone(I);
4975 if (!Loc || Loc->Ptr != UseV || !Loc->Size.isPrecise() || I->isVolatile())
4976 return;
4977
4978 int64_t Offset;
4980 Loc->Ptr, Offset, A.getDataLayout(), /*AllowNonInbounds*/ true);
4981 if (Base && Base == &getAssociatedValue())
4982 State.addAccessedBytes(Offset, Loc->Size.getValue());
4983 }
4984
4985 /// See followUsesInMBEC
4986 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
4987 AADereferenceable::StateType &State) {
4988 bool IsNonNull = false;
4989 bool TrackUse = false;
4990 int64_t DerefBytes = getKnownNonNullAndDerefBytesForUse(
4991 A, *this, getAssociatedValue(), U, I, IsNonNull, TrackUse);
4992 LLVM_DEBUG(dbgs() << "[AADereferenceable] Deref bytes: " << DerefBytes
4993 << " for instruction " << *I << "\n");
4994
4995 addAccessedBytesForUse(A, U, I, State);
4996 State.takeKnownDerefBytesMaximum(DerefBytes);
4997 return TrackUse;
4998 }
4999
5000 /// See AbstractAttribute::manifest(...).
5001 ChangeStatus manifest(Attributor &A) override {
5002 ChangeStatus Change = AADereferenceable::manifest(A);
5003 bool IsKnownNonNull;
5004 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
5005 A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5006 if (IsAssumedNonNull &&
5007 A.hasAttr(getIRPosition(), Attribute::DereferenceableOrNull)) {
5008 A.removeAttrs(getIRPosition(), {Attribute::DereferenceableOrNull});
5009 return ChangeStatus::CHANGED;
5010 }
5011 return Change;
5012 }
5013
5014 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5015 SmallVectorImpl<Attribute> &Attrs) const override {
5016 // TODO: Add *_globally support
5017 bool IsKnownNonNull;
5018 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
5019 A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5020 if (IsAssumedNonNull)
5021 Attrs.emplace_back(Attribute::getWithDereferenceableBytes(
5022 Ctx, getAssumedDereferenceableBytes()));
5023 else
5024 Attrs.emplace_back(Attribute::getWithDereferenceableOrNullBytes(
5025 Ctx, getAssumedDereferenceableBytes()));
5026 }
5027
5028 /// See AbstractAttribute::getAsStr().
5029 const std::string getAsStr(Attributor *A) const override {
5030 if (!getAssumedDereferenceableBytes())
5031 return "unknown-dereferenceable";
5032 bool IsKnownNonNull;
5033 bool IsAssumedNonNull = false;
5034 if (A)
5036 *A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5037 return std::string("dereferenceable") +
5038 (IsAssumedNonNull ? "" : "_or_null") +
5039 (isAssumedGlobal() ? "_globally" : "") + "<" +
5040 std::to_string(getKnownDereferenceableBytes()) + "-" +
5041 std::to_string(getAssumedDereferenceableBytes()) + ">" +
5042 (!A ? " [non-null is unknown]" : "");
5043 }
5044};
5045
5046/// Dereferenceable attribute for a floating value.
5047struct AADereferenceableFloating : AADereferenceableImpl {
5048 AADereferenceableFloating(const IRPosition &IRP, Attributor &A)
5049 : AADereferenceableImpl(IRP, A) {}
5050
5051 /// See AbstractAttribute::updateImpl(...).
5052 ChangeStatus updateImpl(Attributor &A) override {
5053 bool Stripped;
5054 bool UsedAssumedInformation = false;
5056 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
5057 AA::AnyScope, UsedAssumedInformation)) {
5058 Values.push_back({getAssociatedValue(), getCtxI()});
5059 Stripped = false;
5060 } else {
5061 Stripped = Values.size() != 1 ||
5062 Values.front().getValue() != &getAssociatedValue();
5063 }
5064
5065 const DataLayout &DL = A.getDataLayout();
5066 DerefState T;
5067
5068 auto VisitValueCB = [&](const Value &V) -> bool {
5069 unsigned IdxWidth =
5070 DL.getIndexSizeInBits(V.getType()->getPointerAddressSpace());
5071 APInt Offset(IdxWidth, 0);
5073 A, *this, &V, DL, Offset, /* GetMinOffset */ false,
5074 /* AllowNonInbounds */ true);
5075
5076 const auto *AA = A.getAAFor<AADereferenceable>(
5077 *this, IRPosition::value(*Base), DepClassTy::REQUIRED);
5078 int64_t DerefBytes = 0;
5079 if (!AA || (!Stripped && this == AA)) {
5080 // Use IR information if we did not strip anything.
5081 // TODO: track globally.
5082 bool CanBeNull, CanBeFreed;
5083 DerefBytes =
5084 Base->getPointerDereferenceableBytes(DL, CanBeNull, CanBeFreed);
5085 T.GlobalState.indicatePessimisticFixpoint();
5086 } else {
5087 const DerefState &DS = AA->getState();
5088 DerefBytes = DS.DerefBytesState.getAssumed();
5089 T.GlobalState &= DS.GlobalState;
5090 }
5091
5092 // For now we do not try to "increase" dereferenceability due to negative
5093 // indices as we first have to come up with code to deal with loops and
5094 // for overflows of the dereferenceable bytes.
5095 int64_t OffsetSExt = Offset.getSExtValue();
5096 if (OffsetSExt < 0)
5097 OffsetSExt = 0;
5098
5099 T.takeAssumedDerefBytesMinimum(
5100 std::max(int64_t(0), DerefBytes - OffsetSExt));
5101
5102 if (this == AA) {
5103 if (!Stripped) {
5104 // If nothing was stripped IR information is all we got.
5105 T.takeKnownDerefBytesMaximum(
5106 std::max(int64_t(0), DerefBytes - OffsetSExt));
5107 T.indicatePessimisticFixpoint();
5108 } else if (OffsetSExt > 0) {
5109 // If something was stripped but there is circular reasoning we look
5110 // for the offset. If it is positive we basically decrease the
5111 // dereferenceable bytes in a circular loop now, which will simply
5112 // drive them down to the known value in a very slow way which we
5113 // can accelerate.
5114 T.indicatePessimisticFixpoint();
5115 }
5116 }
5117
5118 return T.isValidState();
5119 };
5120
5121 for (const auto &VAC : Values)
5122 if (!VisitValueCB(*VAC.getValue()))
5123 return indicatePessimisticFixpoint();
5124
5125 return clampStateAndIndicateChange(getState(), T);
5126 }
5127
5128 /// See AbstractAttribute::trackStatistics()
5129 void trackStatistics() const override {
5130 STATS_DECLTRACK_FLOATING_ATTR(dereferenceable)
5131 }
5132};
5133
5134/// Dereferenceable attribute for a return value.
5135struct AADereferenceableReturned final
5136 : AAReturnedFromReturnedValues<AADereferenceable, AADereferenceableImpl> {
5137 using Base =
5138 AAReturnedFromReturnedValues<AADereferenceable, AADereferenceableImpl>;
5139 AADereferenceableReturned(const IRPosition &IRP, Attributor &A)
5140 : Base(IRP, A) {}
5141
5142 /// See AbstractAttribute::trackStatistics()
5143 void trackStatistics() const override {
5144 STATS_DECLTRACK_FNRET_ATTR(dereferenceable)
5145 }
5146};
5147
5148/// Dereferenceable attribute for an argument
5149struct AADereferenceableArgument final
5150 : AAArgumentFromCallSiteArguments<AADereferenceable,
5151 AADereferenceableImpl> {
5152 using Base =
5153 AAArgumentFromCallSiteArguments<AADereferenceable, AADereferenceableImpl>;
5154 AADereferenceableArgument(const IRPosition &IRP, Attributor &A)
5155 : Base(IRP, A) {}
5156
5157 /// See AbstractAttribute::trackStatistics()
5158 void trackStatistics() const override {
5159 STATS_DECLTRACK_ARG_ATTR(dereferenceable)
5160 }
5161};
5162
5163/// Dereferenceable attribute for a call site argument.
5164struct AADereferenceableCallSiteArgument final : AADereferenceableFloating {
5165 AADereferenceableCallSiteArgument(const IRPosition &IRP, Attributor &A)
5166 : AADereferenceableFloating(IRP, A) {}
5167
5168 /// See AbstractAttribute::trackStatistics()
5169 void trackStatistics() const override {
5170 STATS_DECLTRACK_CSARG_ATTR(dereferenceable)
5171 }
5172};
5173
5174/// Dereferenceable attribute deduction for a call site return value.
5175struct AADereferenceableCallSiteReturned final
5176 : AACalleeToCallSite<AADereferenceable, AADereferenceableImpl> {
5177 using Base = AACalleeToCallSite<AADereferenceable, AADereferenceableImpl>;
5178 AADereferenceableCallSiteReturned(const IRPosition &IRP, Attributor &A)
5179 : Base(IRP, A) {}
5180
5181 /// See AbstractAttribute::trackStatistics()
5182 void trackStatistics() const override {
5183 STATS_DECLTRACK_CS_ATTR(dereferenceable);
5184 }
5185};
5186} // namespace
5187
5188// ------------------------ Align Argument Attribute ------------------------
5189
5190namespace {
5191
5192static unsigned getKnownAlignForUse(Attributor &A, AAAlign &QueryingAA,
5193 Value &AssociatedValue, const Use *U,
5194 const Instruction *I, bool &TrackUse) {
5195 // We need to follow common pointer manipulation uses to the accesses they
5196 // feed into.
5197 if (isa<CastInst>(I)) {
5198 // Follow all but ptr2int casts.
5199 TrackUse = !isa<PtrToIntInst>(I);
5200 return 0;
5201 }
5202 if (auto *GEP = dyn_cast<GetElementPtrInst>(I)) {
5203 if (GEP->hasAllConstantIndices())
5204 TrackUse = true;
5205 return 0;
5206 }
5207 if (const IntrinsicInst *II = dyn_cast<IntrinsicInst>(I))
5208 switch (II->getIntrinsicID()) {
5209 case Intrinsic::ptrmask: {
5210 // Is it appropriate to pull attribute in initialization?
5211 const auto *ConstVals = A.getAAFor<AAPotentialConstantValues>(
5212 QueryingAA, IRPosition::value(*II->getOperand(1)), DepClassTy::NONE);
5213 const auto *AlignAA = A.getAAFor<AAAlign>(
5214 QueryingAA, IRPosition::value(*II), DepClassTy::NONE);
5215 if (ConstVals && ConstVals->isValidState() && ConstVals->isAtFixpoint()) {
5216 unsigned ShiftValue = std::min(ConstVals->getAssumedMinTrailingZeros(),
5218 Align ConstAlign(UINT64_C(1) << ShiftValue);
5219 if (ConstAlign >= AlignAA->getKnownAlign())
5220 return Align(1).value();
5221 }
5222 if (AlignAA)
5223 return AlignAA->getKnownAlign().value();
5224 break;
5225 }
5226 case Intrinsic::amdgcn_make_buffer_rsrc: {
5227 const auto *AlignAA = A.getAAFor<AAAlign>(
5228 QueryingAA, IRPosition::value(*II), DepClassTy::NONE);
5229 if (AlignAA)
5230 return AlignAA->getKnownAlign().value();
5231 break;
5232 }
5233 default:
5234 break;
5235 }
5236
5237 MaybeAlign MA;
5238 if (const auto *CB = dyn_cast<CallBase>(I)) {
5239 if (CB->isBundleOperand(U) || CB->isCallee(U))
5240 return 0;
5241
5242 unsigned ArgNo = CB->getArgOperandNo(U);
5243 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
5244 // As long as we only use known information there is no need to track
5245 // dependences here.
5246 auto *AlignAA = A.getAAFor<AAAlign>(QueryingAA, IRP, DepClassTy::NONE);
5247 if (AlignAA)
5248 MA = MaybeAlign(AlignAA->getKnownAlign());
5249 }
5250
5251 const DataLayout &DL = A.getDataLayout();
5252 const Value *UseV = U->get();
5253 if (auto *SI = dyn_cast<StoreInst>(I)) {
5254 if (SI->getPointerOperand() == UseV)
5255 MA = SI->getAlign();
5256 } else if (auto *LI = dyn_cast<LoadInst>(I)) {
5257 if (LI->getPointerOperand() == UseV)
5258 MA = LI->getAlign();
5259 } else if (auto *AI = dyn_cast<AtomicRMWInst>(I)) {
5260 if (AI->getPointerOperand() == UseV)
5261 MA = AI->getAlign();
5262 } else if (auto *AI = dyn_cast<AtomicCmpXchgInst>(I)) {
5263 if (AI->getPointerOperand() == UseV)
5264 MA = AI->getAlign();
5265 }
5266
5267 if (!MA || *MA <= QueryingAA.getKnownAlign())
5268 return 0;
5269
5270 unsigned Alignment = MA->value();
5271 int64_t Offset;
5272
5273 if (const Value *Base = GetPointerBaseWithConstantOffset(UseV, Offset, DL)) {
5274 if (Base == &AssociatedValue) {
5275 // BasePointerAddr + Offset = Alignment * Q for some integer Q.
5276 // So we can say that the maximum power of two which is a divisor of
5277 // gcd(Offset, Alignment) is an alignment.
5278
5279 uint32_t gcd = std::gcd(uint32_t(abs((int32_t)Offset)), Alignment);
5280 Alignment = llvm::bit_floor(gcd);
5281 }
5282 }
5283
5284 return Alignment;
5285}
5286
5287struct AAAlignImpl : AAAlign {
5288 AAAlignImpl(const IRPosition &IRP, Attributor &A) : AAAlign(IRP, A) {}
5289
5290 /// See AbstractAttribute::initialize(...).
5291 void initialize(Attributor &A) override {
5293 A.getAttrs(getIRPosition(), {Attribute::Alignment}, Attrs);
5294 for (const Attribute &Attr : Attrs)
5295 takeKnownMaximum(Attr.getValueAsInt());
5296
5297 Value &V = *getAssociatedValue().stripPointerCasts();
5298 takeKnownMaximum(V.getPointerAlignment(A.getDataLayout()).value());
5299
5300 if (Instruction *CtxI = getCtxI())
5301 followUsesInMBEC(*this, A, getState(), *CtxI);
5302 }
5303
5304 /// See AbstractAttribute::manifest(...).
5305 ChangeStatus manifest(Attributor &A) override {
5306 ChangeStatus InstrChanged = ChangeStatus::UNCHANGED;
5307
5308 // Check for users that allow alignment annotations.
5309 Value &AssociatedValue = getAssociatedValue();
5310 if (isa<ConstantData>(AssociatedValue))
5311 return ChangeStatus::UNCHANGED;
5312
5313 for (const Use &U : AssociatedValue.uses()) {
5314 if (auto *SI = dyn_cast<StoreInst>(U.getUser())) {
5315 if (SI->getPointerOperand() == &AssociatedValue)
5316 if (SI->getAlign() < getAssumedAlign()) {
5317 STATS_DECLTRACK(AAAlign, Store,
5318 "Number of times alignment added to a store");
5319 SI->setAlignment(getAssumedAlign());
5320 InstrChanged = ChangeStatus::CHANGED;
5321 }
5322 } else if (auto *LI = dyn_cast<LoadInst>(U.getUser())) {
5323 if (LI->getPointerOperand() == &AssociatedValue)
5324 if (LI->getAlign() < getAssumedAlign()) {
5325 LI->setAlignment(getAssumedAlign());
5326 STATS_DECLTRACK(AAAlign, Load,
5327 "Number of times alignment added to a load");
5328 InstrChanged = ChangeStatus::CHANGED;
5329 }
5330 } else if (auto *RMW = dyn_cast<AtomicRMWInst>(U.getUser())) {
5331 if (RMW->getPointerOperand() == &AssociatedValue) {
5332 if (RMW->getAlign() < getAssumedAlign()) {
5333 STATS_DECLTRACK(AAAlign, AtomicRMW,
5334 "Number of times alignment added to atomicrmw");
5335
5336 RMW->setAlignment(getAssumedAlign());
5337 InstrChanged = ChangeStatus::CHANGED;
5338 }
5339 }
5340 } else if (auto *CAS = dyn_cast<AtomicCmpXchgInst>(U.getUser())) {
5341 if (CAS->getPointerOperand() == &AssociatedValue) {
5342 if (CAS->getAlign() < getAssumedAlign()) {
5343 STATS_DECLTRACK(AAAlign, AtomicCmpXchg,
5344 "Number of times alignment added to cmpxchg");
5345 CAS->setAlignment(getAssumedAlign());
5346 InstrChanged = ChangeStatus::CHANGED;
5347 }
5348 }
5349 }
5350 }
5351
5352 ChangeStatus Changed = AAAlign::manifest(A);
5353
5354 Align InheritAlign =
5355 getAssociatedValue().getPointerAlignment(A.getDataLayout());
5356 if (InheritAlign >= getAssumedAlign())
5357 return InstrChanged;
5358 return Changed | InstrChanged;
5359 }
5360
5361 // TODO: Provide a helper to determine the implied ABI alignment and check in
5362 // the existing manifest method and a new one for AAAlignImpl that value
5363 // to avoid making the alignment explicit if it did not improve.
5364
5365 /// See AbstractAttribute::getDeducedAttributes
5366 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5367 SmallVectorImpl<Attribute> &Attrs) const override {
5368 if (getAssumedAlign() > 1)
5369 Attrs.emplace_back(
5370 Attribute::getWithAlignment(Ctx, Align(getAssumedAlign())));
5371 }
5372
5373 /// See followUsesInMBEC
5374 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
5375 AAAlign::StateType &State) {
5376 bool TrackUse = false;
5377
5378 unsigned int KnownAlign =
5379 getKnownAlignForUse(A, *this, getAssociatedValue(), U, I, TrackUse);
5380 State.takeKnownMaximum(KnownAlign);
5381
5382 return TrackUse;
5383 }
5384
5385 /// See AbstractAttribute::getAsStr().
5386 const std::string getAsStr(Attributor *A) const override {
5387 return "align<" + std::to_string(getKnownAlign().value()) + "-" +
5388 std::to_string(getAssumedAlign().value()) + ">";
5389 }
5390};
5391
5392/// Align attribute for a floating value.
5393struct AAAlignFloating : AAAlignImpl {
5394 AAAlignFloating(const IRPosition &IRP, Attributor &A) : AAAlignImpl(IRP, A) {}
5395
5396 /// See AbstractAttribute::updateImpl(...).
5397 ChangeStatus updateImpl(Attributor &A) override {
5398 const DataLayout &DL = A.getDataLayout();
5399
5400 bool Stripped;
5401 bool UsedAssumedInformation = false;
5403 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
5404 AA::AnyScope, UsedAssumedInformation)) {
5405 Values.push_back({getAssociatedValue(), getCtxI()});
5406 Stripped = false;
5407 } else {
5408 Stripped = Values.size() != 1 ||
5409 Values.front().getValue() != &getAssociatedValue();
5410 }
5411
5412 StateType T;
5413 auto VisitValueCB = [&](Value &V) -> bool {
5415 return true;
5416 const auto *AA = A.getAAFor<AAAlign>(*this, IRPosition::value(V),
5417 DepClassTy::REQUIRED);
5418 if (!AA || (!Stripped && this == AA)) {
5419 int64_t Offset;
5420 unsigned Alignment = 1;
5421 if (const Value *Base =
5423 // TODO: Use AAAlign for the base too.
5424 Align PA = Base->getPointerAlignment(DL);
5425 // BasePointerAddr + Offset = Alignment * Q for some integer Q.
5426 // So we can say that the maximum power of two which is a divisor of
5427 // gcd(Offset, Alignment) is an alignment.
5428
5429 uint32_t gcd =
5430 std::gcd(uint32_t(abs((int32_t)Offset)), uint32_t(PA.value()));
5431 Alignment = llvm::bit_floor(gcd);
5432 } else {
5433 Alignment = V.getPointerAlignment(DL).value();
5434 }
5435 // Use only IR information if we did not strip anything.
5436 T.takeKnownMaximum(Alignment);
5437 T.indicatePessimisticFixpoint();
5438 } else {
5439 // Use abstract attribute information.
5440 const AAAlign::StateType &DS = AA->getState();
5441 T ^= DS;
5442 }
5443 return T.isValidState();
5444 };
5445
5446 for (const auto &VAC : Values) {
5447 if (!VisitValueCB(*VAC.getValue()))
5448 return indicatePessimisticFixpoint();
5449 }
5450
5451 // TODO: If we know we visited all incoming values, thus no are assumed
5452 // dead, we can take the known information from the state T.
5453 return clampStateAndIndicateChange(getState(), T);
5454 }
5455
5456 /// See AbstractAttribute::trackStatistics()
5457 void trackStatistics() const override { STATS_DECLTRACK_FLOATING_ATTR(align) }
5458};
5459
5460/// Align attribute for function return value.
5461struct AAAlignReturned final
5462 : AAReturnedFromReturnedValues<AAAlign, AAAlignImpl> {
5463 using Base = AAReturnedFromReturnedValues<AAAlign, AAAlignImpl>;
5464 AAAlignReturned(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
5465
5466 /// See AbstractAttribute::trackStatistics()
5467 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(aligned) }
5468};
5469
5470/// Align attribute for function argument.
5471struct AAAlignArgument final
5472 : AAArgumentFromCallSiteArguments<AAAlign, AAAlignImpl> {
5473 using Base = AAArgumentFromCallSiteArguments<AAAlign, AAAlignImpl>;
5474 AAAlignArgument(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
5475
5476 /// See AbstractAttribute::manifest(...).
5477 ChangeStatus manifest(Attributor &A) override {
5478 // If the associated argument is involved in a must-tail call we give up
5479 // because we would need to keep the argument alignments of caller and
5480 // callee in-sync. Just does not seem worth the trouble right now.
5481 if (A.getInfoCache().isInvolvedInMustTailCall(*getAssociatedArgument()))
5482 return ChangeStatus::UNCHANGED;
5483 return Base::manifest(A);
5484 }
5485
5486 /// See AbstractAttribute::trackStatistics()
5487 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(aligned) }
5488};
5489
5490struct AAAlignCallSiteArgument final : AAAlignFloating {
5491 AAAlignCallSiteArgument(const IRPosition &IRP, Attributor &A)
5492 : AAAlignFloating(IRP, A) {}
5493
5494 /// See AbstractAttribute::manifest(...).
5495 ChangeStatus manifest(Attributor &A) override {
5496 // If the associated argument is involved in a must-tail call we give up
5497 // because we would need to keep the argument alignments of caller and
5498 // callee in-sync. Just does not seem worth the trouble right now.
5499 if (Argument *Arg = getAssociatedArgument())
5500 if (A.getInfoCache().isInvolvedInMustTailCall(*Arg))
5501 return ChangeStatus::UNCHANGED;
5502 ChangeStatus Changed = AAAlignImpl::manifest(A);
5503 Align InheritAlign =
5504 getAssociatedValue().getPointerAlignment(A.getDataLayout());
5505 if (InheritAlign >= getAssumedAlign())
5506 Changed = ChangeStatus::UNCHANGED;
5507 return Changed;
5508 }
5509
5510 /// See AbstractAttribute::updateImpl(Attributor &A).
5511 ChangeStatus updateImpl(Attributor &A) override {
5512 ChangeStatus Changed = AAAlignFloating::updateImpl(A);
5513 if (Argument *Arg = getAssociatedArgument()) {
5514 // We only take known information from the argument
5515 // so we do not need to track a dependence.
5516 const auto *ArgAlignAA = A.getAAFor<AAAlign>(
5517 *this, IRPosition::argument(*Arg), DepClassTy::NONE);
5518 if (ArgAlignAA)
5519 takeKnownMaximum(ArgAlignAA->getKnownAlign().value());
5520 }
5521 return Changed;
5522 }
5523
5524 /// See AbstractAttribute::trackStatistics()
5525 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(aligned) }
5526};
5527
5528/// Align attribute deduction for a call site return value.
5529struct AAAlignCallSiteReturned final
5530 : AACalleeToCallSite<AAAlign, AAAlignImpl> {
5531 using Base = AACalleeToCallSite<AAAlign, AAAlignImpl>;
5532 AAAlignCallSiteReturned(const IRPosition &IRP, Attributor &A)
5533 : Base(IRP, A) {}
5534
5535 ChangeStatus updateImpl(Attributor &A) override {
5536 Instruction *I = getIRPosition().getCtxI();
5537 if (const IntrinsicInst *II = dyn_cast<IntrinsicInst>(I)) {
5538 switch (II->getIntrinsicID()) {
5539 case Intrinsic::ptrmask: {
5540 Align Alignment;
5541 bool Valid = false;
5542
5543 const auto *ConstVals = A.getAAFor<AAPotentialConstantValues>(
5544 *this, IRPosition::value(*II->getOperand(1)), DepClassTy::REQUIRED);
5545 if (ConstVals && ConstVals->isValidState()) {
5546 unsigned ShiftValue =
5547 std::min(ConstVals->getAssumedMinTrailingZeros(),
5548 Value::MaxAlignmentExponent);
5549 Alignment = Align(UINT64_C(1) << ShiftValue);
5550 Valid = true;
5551 }
5552
5553 const auto *AlignAA =
5554 A.getAAFor<AAAlign>(*this, IRPosition::value(*(II->getOperand(0))),
5555 DepClassTy::REQUIRED);
5556 if (AlignAA) {
5557 Alignment = std::max(AlignAA->getAssumedAlign(), Alignment);
5558 Valid = true;
5559 }
5560
5561 if (Valid)
5563 this->getState(),
5564 std::min(this->getAssumedAlign(), Alignment).value());
5565 break;
5566 }
5567 // FIXME: Should introduce target specific sub-attributes and letting
5568 // getAAfor<AAAlign> lead to create sub-attribute to handle target
5569 // specific intrinsics.
5570 case Intrinsic::amdgcn_make_buffer_rsrc: {
5571 const auto *AlignAA =
5572 A.getAAFor<AAAlign>(*this, IRPosition::value(*(II->getOperand(0))),
5573 DepClassTy::REQUIRED);
5574 if (AlignAA)
5576 this->getState(), AlignAA->getAssumedAlign().value());
5577 break;
5578 }
5579 default:
5580 break;
5581 }
5582 }
5583 return Base::updateImpl(A);
5584 };
5585 /// See AbstractAttribute::trackStatistics()
5586 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(align); }
5587};
5588} // namespace
5589
5590/// ------------------ Function No-Return Attribute ----------------------------
5591namespace {
5592struct AANoReturnImpl : public AANoReturn {
5593 AANoReturnImpl(const IRPosition &IRP, Attributor &A) : AANoReturn(IRP, A) {}
5594
5595 /// See AbstractAttribute::initialize(...).
5596 void initialize(Attributor &A) override {
5597 bool IsKnown;
5599 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
5600 (void)IsKnown;
5601 }
5602
5603 /// See AbstractAttribute::getAsStr().
5604 const std::string getAsStr(Attributor *A) const override {
5605 return getAssumed() ? "noreturn" : "may-return";
5606 }
5607
5608 /// See AbstractAttribute::updateImpl(Attributor &A).
5609 ChangeStatus updateImpl(Attributor &A) override {
5610 auto CheckForNoReturn = [](Instruction &) { return false; };
5611 bool UsedAssumedInformation = false;
5612 if (!A.checkForAllInstructions(CheckForNoReturn, *this,
5613 {(unsigned)Instruction::Ret},
5614 UsedAssumedInformation))
5615 return indicatePessimisticFixpoint();
5616 return ChangeStatus::UNCHANGED;
5617 }
5618};
5619
5620struct AANoReturnFunction final : AANoReturnImpl {
5621 AANoReturnFunction(const IRPosition &IRP, Attributor &A)
5622 : AANoReturnImpl(IRP, A) {}
5623
5624 /// See AbstractAttribute::trackStatistics()
5625 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(noreturn) }
5626};
5627
5628/// NoReturn attribute deduction for a call sites.
5629struct AANoReturnCallSite final
5630 : AACalleeToCallSite<AANoReturn, AANoReturnImpl> {
5631 AANoReturnCallSite(const IRPosition &IRP, Attributor &A)
5632 : AACalleeToCallSite<AANoReturn, AANoReturnImpl>(IRP, A) {}
5633
5634 /// See AbstractAttribute::trackStatistics()
5635 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(noreturn); }
5636};
5637} // namespace
5638
5639/// ----------------------- Instance Info ---------------------------------
5640
5641namespace {
5642/// A class to hold the state of for no-capture attributes.
5643struct AAInstanceInfoImpl : public AAInstanceInfo {
5644 AAInstanceInfoImpl(const IRPosition &IRP, Attributor &A)
5645 : AAInstanceInfo(IRP, A) {}
5646
5647 /// See AbstractAttribute::initialize(...).
5648 void initialize(Attributor &A) override {
5649 Value &V = getAssociatedValue();
5650 if (auto *C = dyn_cast<Constant>(&V)) {
5651 if (C->isThreadDependent())
5652 indicatePessimisticFixpoint();
5653 else
5654 indicateOptimisticFixpoint();
5655 return;
5656 }
5657 if (auto *CB = dyn_cast<CallBase>(&V))
5658 if (CB->arg_size() == 0 && !CB->mayHaveSideEffects() &&
5659 !CB->mayReadFromMemory()) {
5660 indicateOptimisticFixpoint();
5661 return;
5662 }
5663 if (auto *I = dyn_cast<Instruction>(&V)) {
5664 const auto *CI =
5665 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
5666 *I->getFunction());
5667 if (mayBeInCycle(CI, I, /* HeaderOnly */ false)) {
5668 indicatePessimisticFixpoint();
5669 return;
5670 }
5671 }
5672 }
5673
5674 /// See AbstractAttribute::updateImpl(...).
5675 ChangeStatus updateImpl(Attributor &A) override {
5676 ChangeStatus Changed = ChangeStatus::UNCHANGED;
5677
5678 Value &V = getAssociatedValue();
5679 const Function *Scope = nullptr;
5680 if (auto *I = dyn_cast<Instruction>(&V))
5681 Scope = I->getFunction();
5682 if (auto *A = dyn_cast<Argument>(&V)) {
5683 Scope = A->getParent();
5684 if (!Scope->hasLocalLinkage())
5685 return Changed;
5686 }
5687 if (!Scope)
5688 return indicateOptimisticFixpoint();
5689
5690 bool IsKnownNoRecurse;
5692 A, this, IRPosition::function(*Scope), DepClassTy::OPTIONAL,
5693 IsKnownNoRecurse))
5694 return Changed;
5695
5696 auto UsePred = [&](const Use &U, bool &Follow) {
5697 const Instruction *UserI = dyn_cast<Instruction>(U.getUser());
5698 if (!UserI || isa<GetElementPtrInst>(UserI) || isa<CastInst>(UserI) ||
5699 isa<PHINode>(UserI) || isa<SelectInst>(UserI)) {
5700 Follow = true;
5701 return true;
5702 }
5703 if (isa<LoadInst>(UserI) || isa<CmpInst>(UserI) ||
5704 (isa<StoreInst>(UserI) &&
5705 cast<StoreInst>(UserI)->getValueOperand() != U.get()))
5706 return true;
5707 if (auto *CB = dyn_cast<CallBase>(UserI)) {
5708 // This check is not guaranteeing uniqueness but for now that we cannot
5709 // end up with two versions of \p U thinking it was one.
5711 if (!Callee || !Callee->hasLocalLinkage())
5712 return true;
5713 if (!CB->isArgOperand(&U))
5714 return false;
5715 const auto *ArgInstanceInfoAA = A.getAAFor<AAInstanceInfo>(
5717 DepClassTy::OPTIONAL);
5718 if (!ArgInstanceInfoAA ||
5719 !ArgInstanceInfoAA->isAssumedUniqueForAnalysis())
5720 return false;
5721 // If this call base might reach the scope again we might forward the
5722 // argument back here. This is very conservative.
5724 A, *CB, *Scope, *this, /* ExclusionSet */ nullptr,
5725 [Scope](const Function &Fn) { return &Fn != Scope; }))
5726 return false;
5727 return true;
5728 }
5729 return false;
5730 };
5731
5732 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
5733 if (auto *SI = dyn_cast<StoreInst>(OldU.getUser())) {
5734 auto *Ptr = SI->getPointerOperand()->stripPointerCasts();
5735 if ((isa<AllocaInst>(Ptr) || isNoAliasCall(Ptr)) &&
5736 AA::isDynamicallyUnique(A, *this, *Ptr))
5737 return true;
5738 }
5739 return false;
5740 };
5741
5742 if (!A.checkForAllUses(UsePred, *this, V, /* CheckBBLivenessOnly */ true,
5743 DepClassTy::OPTIONAL,
5744 /* IgnoreDroppableUses */ true, EquivalentUseCB))
5745 return indicatePessimisticFixpoint();
5746
5747 return Changed;
5748 }
5749
5750 /// See AbstractState::getAsStr().
5751 const std::string getAsStr(Attributor *A) const override {
5752 return isAssumedUniqueForAnalysis() ? "<unique [fAa]>" : "<unknown>";
5753 }
5754
5755 /// See AbstractAttribute::trackStatistics()
5756 void trackStatistics() const override {}
5757};
5758
5759/// InstanceInfo attribute for floating values.
5760struct AAInstanceInfoFloating : AAInstanceInfoImpl {
5761 AAInstanceInfoFloating(const IRPosition &IRP, Attributor &A)
5762 : AAInstanceInfoImpl(IRP, A) {}
5763};
5764
5765/// NoCapture attribute for function arguments.
5766struct AAInstanceInfoArgument final : AAInstanceInfoFloating {
5767 AAInstanceInfoArgument(const IRPosition &IRP, Attributor &A)
5768 : AAInstanceInfoFloating(IRP, A) {}
5769};
5770
5771/// InstanceInfo attribute for call site arguments.
5772struct AAInstanceInfoCallSiteArgument final : AAInstanceInfoImpl {
5773 AAInstanceInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
5774 : AAInstanceInfoImpl(IRP, A) {}
5775
5776 /// See AbstractAttribute::updateImpl(...).
5777 ChangeStatus updateImpl(Attributor &A) override {
5778 // TODO: Once we have call site specific value information we can provide
5779 // call site specific liveness information and then it makes
5780 // sense to specialize attributes for call sites arguments instead of
5781 // redirecting requests to the callee argument.
5782 Argument *Arg = getAssociatedArgument();
5783 if (!Arg)
5784 return indicatePessimisticFixpoint();
5785 const IRPosition &ArgPos = IRPosition::argument(*Arg);
5786 auto *ArgAA =
5787 A.getAAFor<AAInstanceInfo>(*this, ArgPos, DepClassTy::REQUIRED);
5788 if (!ArgAA)
5789 return indicatePessimisticFixpoint();
5790 return clampStateAndIndicateChange(getState(), ArgAA->getState());
5791 }
5792};
5793
5794/// InstanceInfo attribute for function return value.
5795struct AAInstanceInfoReturned final : AAInstanceInfoImpl {
5796 AAInstanceInfoReturned(const IRPosition &IRP, Attributor &A)
5797 : AAInstanceInfoImpl(IRP, A) {
5798 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5799 }
5800
5801 /// See AbstractAttribute::initialize(...).
5802 void initialize(Attributor &A) override {
5803 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5804 }
5805
5806 /// See AbstractAttribute::updateImpl(...).
5807 ChangeStatus updateImpl(Attributor &A) override {
5808 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5809 }
5810};
5811
5812/// InstanceInfo attribute deduction for a call site return value.
5813struct AAInstanceInfoCallSiteReturned final : AAInstanceInfoFloating {
5814 AAInstanceInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
5815 : AAInstanceInfoFloating(IRP, A) {}
5816};
5817} // namespace
5818
5819/// ----------------------- Variable Capturing ---------------------------------
5821 Attribute::AttrKind ImpliedAttributeKind,
5822 bool IgnoreSubsumingPositions) {
5823 assert(ImpliedAttributeKind == Attribute::Captures &&
5824 "Unexpected attribute kind");
5825 Value &V = IRP.getAssociatedValue();
5826 if (!isa<Constant>(V) && !IRP.isArgumentPosition())
5827 return V.use_empty();
5828
5829 // You cannot "capture" null in the default address space.
5830 //
5831 // FIXME: This should use NullPointerIsDefined to account for the function
5832 // attribute.
5834 V.getType()->getPointerAddressSpace() == 0)) {
5835 return true;
5836 }
5837
5839 A.getAttrs(IRP, {Attribute::Captures}, Attrs,
5840 /* IgnoreSubsumingPositions */ true);
5841 for (const Attribute &Attr : Attrs)
5842 if (capturesNothing(Attr.getCaptureInfo()))
5843 return true;
5844
5846 if (Argument *Arg = IRP.getAssociatedArgument()) {
5848 A.getAttrs(IRPosition::argument(*Arg),
5849 {Attribute::Captures, Attribute::ByVal}, Attrs,
5850 /* IgnoreSubsumingPositions */ true);
5851 bool ArgNoCapture = any_of(Attrs, [](Attribute Attr) {
5852 return Attr.getKindAsEnum() == Attribute::ByVal ||
5854 });
5855 if (ArgNoCapture) {
5856 A.manifestAttrs(IRP, Attribute::getWithCaptureInfo(
5857 V.getContext(), CaptureInfo::none()));
5858 return true;
5859 }
5860 }
5861
5862 if (const Function *F = IRP.getAssociatedFunction()) {
5863 // Check what state the associated function can actually capture.
5866 if (State.isKnown(NO_CAPTURE)) {
5867 A.manifestAttrs(IRP, Attribute::getWithCaptureInfo(V.getContext(),
5869 return true;
5870 }
5871 }
5872
5873 return false;
5874}
5875
5876/// Set the NOT_CAPTURED_IN_MEM and NOT_CAPTURED_IN_RET bits in \p Known
5877/// depending on the ability of the function associated with \p IRP to capture
5878/// state in memory and through "returning/throwing", respectively.
5880 const Function &F,
5881 BitIntegerState &State) {
5882 // TODO: Once we have memory behavior attributes we should use them here.
5883
5884 // If we know we cannot communicate or write to memory, we do not care about
5885 // ptr2int anymore.
5886 bool ReadOnly = F.onlyReadsMemory();
5887 bool NoThrow = F.doesNotThrow();
5888 bool IsVoidReturn = F.getReturnType()->isVoidTy();
5889 if (ReadOnly && NoThrow && IsVoidReturn) {
5890 State.addKnownBits(NO_CAPTURE);
5891 return;
5892 }
5893
5894 // A function cannot capture state in memory if it only reads memory, it can
5895 // however return/throw state and the state might be influenced by the
5896 // pointer value, e.g., loading from a returned pointer might reveal a bit.
5897 if (ReadOnly)
5898 State.addKnownBits(NOT_CAPTURED_IN_MEM);
5899
5900 // A function cannot communicate state back if it does not through
5901 // exceptions and doesn not return values.
5902 if (NoThrow && IsVoidReturn)
5903 State.addKnownBits(NOT_CAPTURED_IN_RET);
5904
5905 // Check existing "returned" attributes.
5906 int ArgNo = IRP.getCalleeArgNo();
5907 if (!NoThrow || ArgNo < 0 ||
5908 !F.getAttributes().hasAttrSomewhere(Attribute::Returned))
5909 return;
5910
5911 for (unsigned U = 0, E = F.arg_size(); U < E; ++U)
5912 if (F.hasParamAttribute(U, Attribute::Returned)) {
5913 if (U == unsigned(ArgNo))
5914 State.removeAssumedBits(NOT_CAPTURED_IN_RET);
5915 else if (ReadOnly)
5916 State.addKnownBits(NO_CAPTURE);
5917 else
5918 State.addKnownBits(NOT_CAPTURED_IN_RET);
5919 break;
5920 }
5921}
5922
5923namespace {
5924/// A class to hold the state of for no-capture attributes.
5925struct AANoCaptureImpl : public AANoCapture {
5926 AANoCaptureImpl(const IRPosition &IRP, Attributor &A) : AANoCapture(IRP, A) {}
5927
5928 /// See AbstractAttribute::initialize(...).
5929 void initialize(Attributor &A) override {
5930 bool IsKnown;
5932 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
5933 (void)IsKnown;
5934 }
5935
5936 /// See AbstractAttribute::updateImpl(...).
5937 ChangeStatus updateImpl(Attributor &A) override;
5938
5939 /// see AbstractAttribute::isAssumedNoCaptureMaybeReturned(...).
5940 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5941 SmallVectorImpl<Attribute> &Attrs) const override {
5942 if (!isAssumedNoCaptureMaybeReturned())
5943 return;
5944
5945 if (isArgumentPosition()) {
5946 if (isAssumedNoCapture())
5947 Attrs.emplace_back(Attribute::get(Ctx, Attribute::Captures));
5948 else if (ManifestInternal)
5949 Attrs.emplace_back(Attribute::get(Ctx, "no-capture-maybe-returned"));
5950 }
5951 }
5952
5953 /// See AbstractState::getAsStr().
5954 const std::string getAsStr(Attributor *A) const override {
5955 if (isKnownNoCapture())
5956 return "known not-captured";
5957 if (isAssumedNoCapture())
5958 return "assumed not-captured";
5959 if (isKnownNoCaptureMaybeReturned())
5960 return "known not-captured-maybe-returned";
5961 if (isAssumedNoCaptureMaybeReturned())
5962 return "assumed not-captured-maybe-returned";
5963 return "assumed-captured";
5964 }
5965
5966 /// Check the use \p U and update \p State accordingly. Return true if we
5967 /// should continue to update the state.
5968 bool checkUse(Attributor &A, AANoCapture::StateType &State, const Use &U,
5969 bool &Follow) {
5970 Instruction *UInst = cast<Instruction>(U.getUser());
5971 LLVM_DEBUG(dbgs() << "[AANoCapture] Check use: " << *U.get() << " in "
5972 << *UInst << "\n");
5973
5974 // Deal with ptr2int by following uses.
5975 if (isa<PtrToIntInst>(UInst)) {
5976 LLVM_DEBUG(dbgs() << " - ptr2int assume the worst!\n");
5977 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5978 /* Return */ true);
5979 }
5980
5981 // For stores we already checked if we can follow them, if they make it
5982 // here we give up.
5983 if (isa<StoreInst>(UInst))
5984 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5985 /* Return */ true);
5986
5987 // Explicitly catch return instructions.
5988 if (isa<ReturnInst>(UInst)) {
5989 if (UInst->getFunction() == getAnchorScope())
5990 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
5991 /* Return */ true);
5992 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5993 /* Return */ true);
5994 }
5995
5996 // For now we only use special logic for call sites. However, the tracker
5997 // itself knows about a lot of other non-capturing cases already.
5998 auto *CB = dyn_cast<CallBase>(UInst);
5999 if (!CB || !CB->isArgOperand(&U))
6000 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
6001 /* Return */ true);
6002
6003 unsigned ArgNo = CB->getArgOperandNo(&U);
6004 const IRPosition &CSArgPos = IRPosition::callsite_argument(*CB, ArgNo);
6005 // If we have a abstract no-capture attribute for the argument we can use
6006 // it to justify a non-capture attribute here. This allows recursion!
6007 bool IsKnownNoCapture;
6008 const AANoCapture *ArgNoCaptureAA = nullptr;
6009 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
6010 A, this, CSArgPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
6011 &ArgNoCaptureAA);
6012 if (IsAssumedNoCapture)
6013 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
6014 /* Return */ false);
6015 if (ArgNoCaptureAA && ArgNoCaptureAA->isAssumedNoCaptureMaybeReturned()) {
6016 Follow = true;
6017 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
6018 /* Return */ false);
6019 }
6020
6021 // Lastly, we could not find a reason no-capture can be assumed so we don't.
6022 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
6023 /* Return */ true);
6024 }
6025
6026 /// Update \p State according to \p CapturedInMem, \p CapturedInInt, and
6027 /// \p CapturedInRet, then return true if we should continue updating the
6028 /// state.
6029 static bool isCapturedIn(AANoCapture::StateType &State, bool CapturedInMem,
6030 bool CapturedInInt, bool CapturedInRet) {
6031 LLVM_DEBUG(dbgs() << " - captures [Mem " << CapturedInMem << "|Int "
6032 << CapturedInInt << "|Ret " << CapturedInRet << "]\n");
6033 if (CapturedInMem)
6034 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_MEM);
6035 if (CapturedInInt)
6036 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_INT);
6037 if (CapturedInRet)
6038 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_RET);
6039 return State.isAssumed(AANoCapture::NO_CAPTURE_MAYBE_RETURNED);
6040 }
6041};
6042
6043ChangeStatus AANoCaptureImpl::updateImpl(Attributor &A) {
6044 const IRPosition &IRP = getIRPosition();
6045 Value *V = isArgumentPosition() ? IRP.getAssociatedArgument()
6046 : &IRP.getAssociatedValue();
6047 if (!V)
6048 return indicatePessimisticFixpoint();
6049
6050 const Function *F =
6051 isArgumentPosition() ? IRP.getAssociatedFunction() : IRP.getAnchorScope();
6052
6053 // TODO: Is the checkForAllUses below useful for constants?
6054 if (!F)
6055 return indicatePessimisticFixpoint();
6056
6058 const IRPosition &FnPos = IRPosition::function(*F);
6059
6060 // Readonly means we cannot capture through memory.
6061 bool IsKnown;
6062 if (AA::isAssumedReadOnly(A, FnPos, *this, IsKnown)) {
6063 T.addKnownBits(NOT_CAPTURED_IN_MEM);
6064 if (IsKnown)
6065 addKnownBits(NOT_CAPTURED_IN_MEM);
6066 }
6067
6068 // Make sure all returned values are different than the underlying value.
6069 // TODO: we could do this in a more sophisticated way inside
6070 // AAReturnedValues, e.g., track all values that escape through returns
6071 // directly somehow.
6072 auto CheckReturnedArgs = [&](bool &UsedAssumedInformation) {
6074 if (!A.getAssumedSimplifiedValues(IRPosition::returned(*F), this, Values,
6076 UsedAssumedInformation))
6077 return false;
6078 bool SeenConstant = false;
6079 for (const AA::ValueAndContext &VAC : Values) {
6080 if (isa<Constant>(VAC.getValue())) {
6081 if (SeenConstant)
6082 return false;
6083 SeenConstant = true;
6084 } else if (!isa<Argument>(VAC.getValue()) ||
6085 VAC.getValue() == getAssociatedArgument())
6086 return false;
6087 }
6088 return true;
6089 };
6090
6091 bool IsKnownNoUnwind;
6093 A, this, FnPos, DepClassTy::OPTIONAL, IsKnownNoUnwind)) {
6094 bool IsVoidTy = F->getReturnType()->isVoidTy();
6095 bool UsedAssumedInformation = false;
6096 if (IsVoidTy || CheckReturnedArgs(UsedAssumedInformation)) {
6097 T.addKnownBits(NOT_CAPTURED_IN_RET);
6098 if (T.isKnown(NOT_CAPTURED_IN_MEM))
6100 if (IsKnownNoUnwind && (IsVoidTy || !UsedAssumedInformation)) {
6101 addKnownBits(NOT_CAPTURED_IN_RET);
6102 if (isKnown(NOT_CAPTURED_IN_MEM))
6103 return indicateOptimisticFixpoint();
6104 }
6105 }
6106 }
6107
6108 auto UseCheck = [&](const Use &U, bool &Follow) -> bool {
6109 // TODO(captures): Make this more precise.
6110 UseCaptureInfo CI = DetermineUseCaptureKind(U, /*Base=*/nullptr);
6111 if (capturesNothing(CI))
6112 return true;
6113 if (CI.isPassthrough()) {
6114 Follow = true;
6115 return true;
6116 }
6117 return checkUse(A, T, U, Follow);
6118 };
6119
6120 if (!A.checkForAllUses(UseCheck, *this, *V))
6121 return indicatePessimisticFixpoint();
6122
6123 AANoCapture::StateType &S = getState();
6124 auto Assumed = S.getAssumed();
6125 S.intersectAssumedBits(T.getAssumed());
6126 if (!isAssumedNoCaptureMaybeReturned())
6127 return indicatePessimisticFixpoint();
6128 return Assumed == S.getAssumed() ? ChangeStatus::UNCHANGED
6130}
6131
6132/// NoCapture attribute for function arguments.
6133struct AANoCaptureArgument final : AANoCaptureImpl {
6134 AANoCaptureArgument(const IRPosition &IRP, Attributor &A)
6135 : AANoCaptureImpl(IRP, A) {}
6136
6137 /// See AbstractAttribute::trackStatistics()
6138 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nocapture) }
6139};
6140
6141/// NoCapture attribute for call site arguments.
6142struct AANoCaptureCallSiteArgument final : AANoCaptureImpl {
6143 AANoCaptureCallSiteArgument(const IRPosition &IRP, Attributor &A)
6144 : AANoCaptureImpl(IRP, A) {}
6145
6146 /// See AbstractAttribute::updateImpl(...).
6147 ChangeStatus updateImpl(Attributor &A) override {
6148 // TODO: Once we have call site specific value information we can provide
6149 // call site specific liveness information and then it makes
6150 // sense to specialize attributes for call sites arguments instead of
6151 // redirecting requests to the callee argument.
6152 Argument *Arg = getAssociatedArgument();
6153 if (!Arg)
6154 return indicatePessimisticFixpoint();
6155 const IRPosition &ArgPos = IRPosition::argument(*Arg);
6156 bool IsKnownNoCapture;
6157 const AANoCapture *ArgAA = nullptr;
6159 A, this, ArgPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
6160 &ArgAA))
6161 return ChangeStatus::UNCHANGED;
6162 if (!ArgAA || !ArgAA->isAssumedNoCaptureMaybeReturned())
6163 return indicatePessimisticFixpoint();
6164 return clampStateAndIndicateChange(getState(), ArgAA->getState());
6165 }
6166
6167 /// See AbstractAttribute::trackStatistics()
6168 void trackStatistics() const override {
6170 };
6171};
6172
6173/// NoCapture attribute for floating values.
6174struct AANoCaptureFloating final : AANoCaptureImpl {
6175 AANoCaptureFloating(const IRPosition &IRP, Attributor &A)
6176 : AANoCaptureImpl(IRP, A) {}
6177
6178 /// See AbstractAttribute::trackStatistics()
6179 void trackStatistics() const override {
6181 }
6182};
6183
6184/// NoCapture attribute for function return value.
6185struct AANoCaptureReturned final : AANoCaptureImpl {
6186 AANoCaptureReturned(const IRPosition &IRP, Attributor &A)
6187 : AANoCaptureImpl(IRP, A) {
6188 llvm_unreachable("NoCapture is not applicable to function returns!");
6189 }
6190
6191 /// See AbstractAttribute::initialize(...).
6192 void initialize(Attributor &A) override {
6193 llvm_unreachable("NoCapture is not applicable to function returns!");
6194 }
6195
6196 /// See AbstractAttribute::updateImpl(...).
6197 ChangeStatus updateImpl(Attributor &A) override {
6198 llvm_unreachable("NoCapture is not applicable to function returns!");
6199 }
6200
6201 /// See AbstractAttribute::trackStatistics()
6202 void trackStatistics() const override {}
6203};
6204
6205/// NoCapture attribute deduction for a call site return value.
6206struct AANoCaptureCallSiteReturned final : AANoCaptureImpl {
6207 AANoCaptureCallSiteReturned(const IRPosition &IRP, Attributor &A)
6208 : AANoCaptureImpl(IRP, A) {}
6209
6210 /// See AbstractAttribute::initialize(...).
6211 void initialize(Attributor &A) override {
6212 const Function *F = getAnchorScope();
6213 // Check what state the associated function can actually capture.
6214 determineFunctionCaptureCapabilities(getIRPosition(), *F, *this);
6215 }
6216
6217 /// See AbstractAttribute::trackStatistics()
6218 void trackStatistics() const override {
6220 }
6221};
6222} // namespace
6223
6224/// ------------------ Value Simplify Attribute ----------------------------
6225
6226bool ValueSimplifyStateType::unionAssumed(std::optional<Value *> Other) {
6227 // FIXME: Add a typecast support.
6230 if (SimplifiedAssociatedValue == std::optional<Value *>(nullptr))
6231 return false;
6232
6233 LLVM_DEBUG({
6235 dbgs() << "[ValueSimplify] is assumed to be "
6236 << **SimplifiedAssociatedValue << "\n";
6237 else
6238 dbgs() << "[ValueSimplify] is assumed to be <none>\n";
6239 });
6240 return true;
6241}
6242
6243namespace {
6244struct AAValueSimplifyImpl : AAValueSimplify {
6245 AAValueSimplifyImpl(const IRPosition &IRP, Attributor &A)
6246 : AAValueSimplify(IRP, A) {}
6247
6248 /// See AbstractAttribute::initialize(...).
6249 void initialize(Attributor &A) override {
6250 if (getAssociatedValue().getType()->isVoidTy())
6251 indicatePessimisticFixpoint();
6252 if (A.hasSimplificationCallback(getIRPosition()))
6253 indicatePessimisticFixpoint();
6254 }
6255
6256 /// See AbstractAttribute::getAsStr().
6257 const std::string getAsStr(Attributor *A) const override {
6258 LLVM_DEBUG({
6259 dbgs() << "SAV: " << (bool)SimplifiedAssociatedValue << " ";
6260 if (SimplifiedAssociatedValue && *SimplifiedAssociatedValue)
6261 dbgs() << "SAV: " << **SimplifiedAssociatedValue << " ";
6262 });
6263 return isValidState() ? (isAtFixpoint() ? "simplified" : "maybe-simple")
6264 : "not-simple";
6265 }
6266
6267 /// See AbstractAttribute::trackStatistics()
6268 void trackStatistics() const override {}
6269
6270 /// See AAValueSimplify::getAssumedSimplifiedValue()
6271 std::optional<Value *>
6272 getAssumedSimplifiedValue(Attributor &A) const override {
6273 return SimplifiedAssociatedValue;
6274 }
6275
6276 /// Ensure the return value is \p V with type \p Ty, if not possible return
6277 /// nullptr. If \p Check is true we will only verify such an operation would
6278 /// suceed and return a non-nullptr value if that is the case. No IR is
6279 /// generated or modified.
6280 static Value *ensureType(Attributor &A, Value &V, Type &Ty, Instruction *CtxI,
6281 bool Check) {
6282 if (auto *TypedV = AA::getWithType(V, Ty))
6283 return TypedV;
6284 if (CtxI && V.getType()->canLosslesslyBitCastTo(&Ty))
6285 return Check ? &V
6286 : BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
6287 &V, &Ty, "", CtxI->getIterator());
6288 return nullptr;
6289 }
6290
6291 /// Reproduce \p I with type \p Ty or return nullptr if that is not posisble.
6292 /// If \p Check is true we will only verify such an operation would suceed and
6293 /// return a non-nullptr value if that is the case. No IR is generated or
6294 /// modified.
6295 static Value *reproduceInst(Attributor &A,
6296 const AbstractAttribute &QueryingAA,
6297 Instruction &I, Type &Ty, Instruction *CtxI,
6298 bool Check, ValueToValueMapTy &VMap) {
6299 assert(CtxI && "Cannot reproduce an instruction without context!");
6300 if (Check && (I.mayReadFromMemory() ||
6301 !isSafeToSpeculativelyExecute(&I, CtxI, /* DT */ nullptr,
6302 /* TLI */ nullptr)))
6303 return nullptr;
6304 for (Value *Op : I.operands()) {
6305 Value *NewOp = reproduceValue(A, QueryingAA, *Op, Ty, CtxI, Check, VMap);
6306 if (!NewOp) {
6307 assert(Check && "Manifest of new value unexpectedly failed!");
6308 return nullptr;
6309 }
6310 if (!Check)
6311 VMap[Op] = NewOp;
6312 }
6313 if (Check)
6314 return &I;
6315
6316 Instruction *CloneI = I.clone();
6317 // TODO: Try to salvage debug information here.
6318 CloneI->setDebugLoc(DebugLoc());
6319 VMap[&I] = CloneI;
6320 CloneI->insertBefore(CtxI->getIterator());
6321 RemapInstruction(CloneI, VMap);
6322 return CloneI;
6323 }
6324
6325 /// Reproduce \p V with type \p Ty or return nullptr if that is not posisble.
6326 /// If \p Check is true we will only verify such an operation would suceed and
6327 /// return a non-nullptr value if that is the case. No IR is generated or
6328 /// modified.
6329 static Value *reproduceValue(Attributor &A,
6330 const AbstractAttribute &QueryingAA, Value &V,
6331 Type &Ty, Instruction *CtxI, bool Check,
6332 ValueToValueMapTy &VMap) {
6333 if (const auto &NewV = VMap.lookup(&V))
6334 return NewV;
6335 bool UsedAssumedInformation = false;
6336 std::optional<Value *> SimpleV = A.getAssumedSimplified(
6337 V, QueryingAA, UsedAssumedInformation, AA::Interprocedural);
6338 if (!SimpleV.has_value())
6339 return PoisonValue::get(&Ty);
6340 Value *EffectiveV = &V;
6341 if (*SimpleV)
6342 EffectiveV = *SimpleV;
6343 if (auto *C = dyn_cast<Constant>(EffectiveV))
6344 return C;
6345 if (CtxI && AA::isValidAtPosition(AA::ValueAndContext(*EffectiveV, *CtxI),
6346 A.getInfoCache()))
6347 return ensureType(A, *EffectiveV, Ty, CtxI, Check);
6348 if (auto *I = dyn_cast<Instruction>(EffectiveV))
6349 if (Value *NewV = reproduceInst(A, QueryingAA, *I, Ty, CtxI, Check, VMap))
6350 return ensureType(A, *NewV, Ty, CtxI, Check);
6351 return nullptr;
6352 }
6353
6354 /// Return a value we can use as replacement for the associated one, or
6355 /// nullptr if we don't have one that makes sense.
6356 Value *manifestReplacementValue(Attributor &A, Instruction *CtxI) const {
6357 Value *NewV = SimplifiedAssociatedValue
6358 ? *SimplifiedAssociatedValue
6359 : UndefValue::get(getAssociatedType());
6360 if (NewV && NewV != &getAssociatedValue()) {
6361 ValueToValueMapTy VMap;
6362 // First verify we can reprduce the value with the required type at the
6363 // context location before we actually start modifying the IR.
6364 if (reproduceValue(A, *this, *NewV, *getAssociatedType(), CtxI,
6365 /* CheckOnly */ true, VMap))
6366 return reproduceValue(A, *this, *NewV, *getAssociatedType(), CtxI,
6367 /* CheckOnly */ false, VMap);
6368 }
6369 return nullptr;
6370 }
6371
6372 /// Helper function for querying AAValueSimplify and updating candidate.
6373 /// \param IRP The value position we are trying to unify with SimplifiedValue
6374 bool checkAndUpdate(Attributor &A, const AbstractAttribute &QueryingAA,
6375 const IRPosition &IRP, bool Simplify = true) {
6376 bool UsedAssumedInformation = false;
6377 std::optional<Value *> QueryingValueSimplified = &IRP.getAssociatedValue();
6378 if (Simplify)
6379 QueryingValueSimplified = A.getAssumedSimplified(
6380 IRP, QueryingAA, UsedAssumedInformation, AA::Interprocedural);
6381 return unionAssumed(QueryingValueSimplified);
6382 }
6383
6384 /// Returns a candidate is found or not
6385 template <typename AAType> bool askSimplifiedValueFor(Attributor &A) {
6386 if (!getAssociatedValue().getType()->isIntegerTy())
6387 return false;
6388
6389 // This will also pass the call base context.
6390 const auto *AA =
6391 A.getAAFor<AAType>(*this, getIRPosition(), DepClassTy::NONE);
6392 if (!AA)
6393 return false;
6394
6395 std::optional<Constant *> COpt = AA->getAssumedConstant(A);
6396
6397 if (!COpt) {
6398 SimplifiedAssociatedValue = std::nullopt;
6399 A.recordDependence(*AA, *this, DepClassTy::OPTIONAL);
6400 return true;
6401 }
6402 if (auto *C = *COpt) {
6403 SimplifiedAssociatedValue = C;
6404 A.recordDependence(*AA, *this, DepClassTy::OPTIONAL);
6405 return true;
6406 }
6407 return false;
6408 }
6409
6410 bool askSimplifiedValueForOtherAAs(Attributor &A) {
6411 if (askSimplifiedValueFor<AAValueConstantRange>(A))
6412 return true;
6413 if (askSimplifiedValueFor<AAPotentialConstantValues>(A))
6414 return true;
6415 return false;
6416 }
6417
6418 /// See AbstractAttribute::manifest(...).
6419 ChangeStatus manifest(Attributor &A) override {
6420 ChangeStatus Changed = ChangeStatus::UNCHANGED;
6421 for (auto &U : getAssociatedValue().uses()) {
6422 // Check if we need to adjust the insertion point to make sure the IR is
6423 // valid.
6424 Instruction *IP = dyn_cast<Instruction>(U.getUser());
6425 if (auto *PHI = dyn_cast_or_null<PHINode>(IP))
6426 IP = PHI->getIncomingBlock(U)->getTerminator();
6427 if (auto *NewV = manifestReplacementValue(A, IP)) {
6428 LLVM_DEBUG(dbgs() << "[ValueSimplify] " << getAssociatedValue()
6429 << " -> " << *NewV << " :: " << *this << "\n");
6430 if (A.changeUseAfterManifest(U, *NewV))
6431 Changed = ChangeStatus::CHANGED;
6432 }
6433 }
6434
6435 return Changed | AAValueSimplify::manifest(A);
6436 }
6437
6438 /// See AbstractState::indicatePessimisticFixpoint(...).
6439 ChangeStatus indicatePessimisticFixpoint() override {
6440 SimplifiedAssociatedValue = &getAssociatedValue();
6441 return AAValueSimplify::indicatePessimisticFixpoint();
6442 }
6443};
6444
6445struct AAValueSimplifyArgument final : AAValueSimplifyImpl {
6446 AAValueSimplifyArgument(const IRPosition &IRP, Attributor &A)
6447 : AAValueSimplifyImpl(IRP, A) {}
6448
6449 void initialize(Attributor &A) override {
6450 AAValueSimplifyImpl::initialize(A);
6451 if (A.hasAttr(getIRPosition(),
6452 {Attribute::InAlloca, Attribute::Preallocated,
6453 Attribute::StructRet, Attribute::Nest, Attribute::ByVal},
6454 /* IgnoreSubsumingPositions */ true))
6455 indicatePessimisticFixpoint();
6456 }
6457
6458 /// See AbstractAttribute::updateImpl(...).
6459 ChangeStatus updateImpl(Attributor &A) override {
6460 // Byval is only replacable if it is readonly otherwise we would write into
6461 // the replaced value and not the copy that byval creates implicitly.
6462 Argument *Arg = getAssociatedArgument();
6463 if (Arg->hasByValAttr()) {
6464 // TODO: We probably need to verify synchronization is not an issue, e.g.,
6465 // there is no race by not copying a constant byval.
6466 bool IsKnown;
6467 if (!AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
6468 return indicatePessimisticFixpoint();
6469 }
6470
6471 auto Before = SimplifiedAssociatedValue;
6472
6473 auto PredForCallSite = [&](AbstractCallSite ACS) {
6474 const IRPosition &ACSArgPos =
6475 IRPosition::callsite_argument(ACS, getCallSiteArgNo());
6476 // Check if a coresponding argument was found or if it is on not
6477 // associated (which can happen for callback calls).
6478 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
6479 return false;
6480
6481 // Simplify the argument operand explicitly and check if the result is
6482 // valid in the current scope. This avoids refering to simplified values
6483 // in other functions, e.g., we don't want to say a an argument in a
6484 // static function is actually an argument in a different function.
6485 bool UsedAssumedInformation = false;
6486 std::optional<Constant *> SimpleArgOp =
6487 A.getAssumedConstant(ACSArgPos, *this, UsedAssumedInformation);
6488 if (!SimpleArgOp)
6489 return true;
6490 if (!*SimpleArgOp)
6491 return false;
6492 if (!AA::isDynamicallyUnique(A, *this, **SimpleArgOp))
6493 return false;
6494 return unionAssumed(*SimpleArgOp);
6495 };
6496
6497 // Generate a answer specific to a call site context.
6498 bool Success;
6499 bool UsedAssumedInformation = false;
6500 if (hasCallBaseContext() &&
6501 getCallBaseContext()->getCalledOperand() == Arg->getParent())
6502 Success = PredForCallSite(
6503 AbstractCallSite(&getCallBaseContext()->getCalledOperandUse()));
6504 else
6505 Success = A.checkForAllCallSites(PredForCallSite, *this, true,
6506 UsedAssumedInformation);
6507
6508 if (!Success)
6509 if (!askSimplifiedValueForOtherAAs(A))
6510 return indicatePessimisticFixpoint();
6511
6512 // If a candidate was found in this update, return CHANGED.
6513 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6514 : ChangeStatus ::CHANGED;
6515 }
6516
6517 /// See AbstractAttribute::trackStatistics()
6518 void trackStatistics() const override {
6519 STATS_DECLTRACK_ARG_ATTR(value_simplify)
6520 }
6521};
6522
6523struct AAValueSimplifyReturned : AAValueSimplifyImpl {
6524 AAValueSimplifyReturned(const IRPosition &IRP, Attributor &A)
6525 : AAValueSimplifyImpl(IRP, A) {}
6526
6527 /// See AAValueSimplify::getAssumedSimplifiedValue()
6528 std::optional<Value *>
6529 getAssumedSimplifiedValue(Attributor &A) const override {
6530 if (!isValidState())
6531 return nullptr;
6532 return SimplifiedAssociatedValue;
6533 }
6534
6535 /// See AbstractAttribute::updateImpl(...).
6536 ChangeStatus updateImpl(Attributor &A) override {
6537 auto Before = SimplifiedAssociatedValue;
6538
6539 auto ReturnInstCB = [&](Instruction &I) {
6540 auto &RI = cast<ReturnInst>(I);
6541 return checkAndUpdate(
6542 A, *this,
6543 IRPosition::value(*RI.getReturnValue(), getCallBaseContext()));
6544 };
6545
6546 bool UsedAssumedInformation = false;
6547 if (!A.checkForAllInstructions(ReturnInstCB, *this, {Instruction::Ret},
6548 UsedAssumedInformation))
6549 if (!askSimplifiedValueForOtherAAs(A))
6550 return indicatePessimisticFixpoint();
6551
6552 // If a candidate was found in this update, return CHANGED.
6553 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6554 : ChangeStatus ::CHANGED;
6555 }
6556
6557 ChangeStatus manifest(Attributor &A) override {
6558 // We queried AAValueSimplify for the returned values so they will be
6559 // replaced if a simplified form was found. Nothing to do here.
6560 return ChangeStatus::UNCHANGED;
6561 }
6562
6563 /// See AbstractAttribute::trackStatistics()
6564 void trackStatistics() const override {
6565 STATS_DECLTRACK_FNRET_ATTR(value_simplify)
6566 }
6567};
6568
6569struct AAValueSimplifyFloating : AAValueSimplifyImpl {
6570 AAValueSimplifyFloating(const IRPosition &IRP, Attributor &A)
6571 : AAValueSimplifyImpl(IRP, A) {}
6572
6573 /// See AbstractAttribute::initialize(...).
6574 void initialize(Attributor &A) override {
6575 AAValueSimplifyImpl::initialize(A);
6576 Value &V = getAnchorValue();
6577
6578 // TODO: add other stuffs
6579 if (isa<Constant>(V))
6580 indicatePessimisticFixpoint();
6581 }
6582
6583 /// See AbstractAttribute::updateImpl(...).
6584 ChangeStatus updateImpl(Attributor &A) override {
6585 auto Before = SimplifiedAssociatedValue;
6586 if (!askSimplifiedValueForOtherAAs(A))
6587 return indicatePessimisticFixpoint();
6588
6589 // If a candidate was found in this update, return CHANGED.
6590 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6591 : ChangeStatus ::CHANGED;
6592 }
6593
6594 /// See AbstractAttribute::trackStatistics()
6595 void trackStatistics() const override {
6596 STATS_DECLTRACK_FLOATING_ATTR(value_simplify)
6597 }
6598};
6599
6600struct AAValueSimplifyFunction : AAValueSimplifyImpl {
6601 AAValueSimplifyFunction(const IRPosition &IRP, Attributor &A)
6602 : AAValueSimplifyImpl(IRP, A) {}
6603
6604 /// See AbstractAttribute::initialize(...).
6605 void initialize(Attributor &A) override {
6606 SimplifiedAssociatedValue = nullptr;
6607 indicateOptimisticFixpoint();
6608 }
6609 /// See AbstractAttribute::initialize(...).
6610 ChangeStatus updateImpl(Attributor &A) override {
6612 "AAValueSimplify(Function|CallSite)::updateImpl will not be called");
6613 }
6614 /// See AbstractAttribute::trackStatistics()
6615 void trackStatistics() const override {
6616 STATS_DECLTRACK_FN_ATTR(value_simplify)
6617 }
6618};
6619
6620struct AAValueSimplifyCallSite : AAValueSimplifyFunction {
6621 AAValueSimplifyCallSite(const IRPosition &IRP, Attributor &A)
6622 : AAValueSimplifyFunction(IRP, A) {}
6623 /// See AbstractAttribute::trackStatistics()
6624 void trackStatistics() const override {
6625 STATS_DECLTRACK_CS_ATTR(value_simplify)
6626 }
6627};
6628
6629struct AAValueSimplifyCallSiteReturned : AAValueSimplifyImpl {
6630 AAValueSimplifyCallSiteReturned(const IRPosition &IRP, Attributor &A)
6631 : AAValueSimplifyImpl(IRP, A) {}
6632
6633 void initialize(Attributor &A) override {
6634 AAValueSimplifyImpl::initialize(A);
6635 Function *Fn = getAssociatedFunction();
6636 assert(Fn && "Did expect an associted function");
6637 for (Argument &Arg : Fn->args()) {
6638 if (Arg.hasReturnedAttr()) {
6639 auto IRP = IRPosition::callsite_argument(*cast<CallBase>(getCtxI()),
6640 Arg.getArgNo());
6642 checkAndUpdate(A, *this, IRP))
6643 indicateOptimisticFixpoint();
6644 else
6645 indicatePessimisticFixpoint();
6646 return;
6647 }
6648 }
6649 }
6650
6651 /// See AbstractAttribute::updateImpl(...).
6652 ChangeStatus updateImpl(Attributor &A) override {
6653 return indicatePessimisticFixpoint();
6654 }
6655
6656 void trackStatistics() const override {
6657 STATS_DECLTRACK_CSRET_ATTR(value_simplify)
6658 }
6659};
6660
6661struct AAValueSimplifyCallSiteArgument : AAValueSimplifyFloating {
6662 AAValueSimplifyCallSiteArgument(const IRPosition &IRP, Attributor &A)
6663 : AAValueSimplifyFloating(IRP, A) {}
6664
6665 /// See AbstractAttribute::manifest(...).
6666 ChangeStatus manifest(Attributor &A) override {
6667 ChangeStatus Changed = ChangeStatus::UNCHANGED;
6668 // TODO: We should avoid simplification duplication to begin with.
6669 auto *FloatAA = A.lookupAAFor<AAValueSimplify>(
6670 IRPosition::value(getAssociatedValue()), this, DepClassTy::NONE);
6671 if (FloatAA && FloatAA->getState().isValidState())
6672 return Changed;
6673
6674 if (auto *NewV = manifestReplacementValue(A, getCtxI())) {
6675 Use &U = cast<CallBase>(&getAnchorValue())
6676 ->getArgOperandUse(getCallSiteArgNo());
6677 if (A.changeUseAfterManifest(U, *NewV))
6678 Changed = ChangeStatus::CHANGED;
6679 }
6680
6681 return Changed | AAValueSimplify::manifest(A);
6682 }
6683
6684 void trackStatistics() const override {
6685 STATS_DECLTRACK_CSARG_ATTR(value_simplify)
6686 }
6687};
6688} // namespace
6689
6690/// ----------------------- Heap-To-Stack Conversion ---------------------------
6691namespace {
6692struct AAHeapToStackFunction final : public AAHeapToStack {
6693
6694 struct AllocationInfo {
6695 /// The call that allocates the memory.
6696 CallBase *const CB;
6697
6698 /// The library function id for the allocation.
6699 LibFunc LibraryFunctionId = NotLibFunc;
6700
6701 /// The status wrt. a rewrite.
6702 enum {
6703 STACK_DUE_TO_USE,
6704 STACK_DUE_TO_FREE,
6705 INVALID,
6706 } Status = STACK_DUE_TO_USE;
6707
6708 /// Flag to indicate if we encountered a use that might free this allocation
6709 /// but which is not in the deallocation infos.
6710 bool HasPotentiallyFreeingUnknownUses = false;
6711
6712 /// Flag to indicate that we should place the new alloca in the function
6713 /// entry block rather than where the call site (CB) is.
6714 bool MoveAllocaIntoEntry = true;
6715
6716 /// The set of free calls that use this allocation.
6717 SmallSetVector<CallBase *, 1> PotentialFreeCalls{};
6718 };
6719
6720 struct DeallocationInfo {
6721 /// The call that deallocates the memory.
6722 CallBase *const CB;
6723 /// The value freed by the call.
6724 Value *FreedOp;
6725
6726 /// Flag to indicate if we don't know all objects this deallocation might
6727 /// free.
6728 bool MightFreeUnknownObjects = false;
6729
6730 /// The set of allocation calls that are potentially freed.
6731 SmallSetVector<CallBase *, 1> PotentialAllocationCalls{};
6732 };
6733
6734 AAHeapToStackFunction(const IRPosition &IRP, Attributor &A)
6735 : AAHeapToStack(IRP, A) {}
6736
6737 ~AAHeapToStackFunction() override {
6738 // Ensure we call the destructor so we release any memory allocated in the
6739 // sets.
6740 for (auto &It : AllocationInfos)
6741 It.second->~AllocationInfo();
6742 for (auto &It : DeallocationInfos)
6743 It.second->~DeallocationInfo();
6744 }
6745
6746 void initialize(Attributor &A) override {
6747 AAHeapToStack::initialize(A);
6748
6749 const Function *F = getAnchorScope();
6750 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6751
6752 auto AllocationIdentifierCB = [&](Instruction &I) {
6753 CallBase *CB = dyn_cast<CallBase>(&I);
6754 if (!CB)
6755 return true;
6756 if (Value *FreedOp = getFreedOperand(CB, TLI)) {
6757 DeallocationInfos[CB] = new (A.Allocator) DeallocationInfo{CB, FreedOp};
6758 return true;
6759 }
6760 // To do heap to stack, we need to know that the allocation itself is
6761 // removable once uses are rewritten, and that we can initialize the
6762 // alloca to the same pattern as the original allocation result.
6763 if (isRemovableAlloc(CB, TLI)) {
6764 auto *I8Ty = Type::getInt8Ty(CB->getParent()->getContext());
6765 if (nullptr != getInitialValueOfAllocation(CB, TLI, I8Ty)) {
6766 AllocationInfo *AI = new (A.Allocator) AllocationInfo{CB};
6767 AllocationInfos[CB] = AI;
6768 if (TLI)
6769 TLI->getLibFunc(*CB, AI->LibraryFunctionId);
6770 }
6771 }
6772 return true;
6773 };
6774
6775 bool UsedAssumedInformation = false;
6776 bool Success = A.checkForAllCallLikeInstructions(
6777 AllocationIdentifierCB, *this, UsedAssumedInformation,
6778 /* CheckBBLivenessOnly */ false,
6779 /* CheckPotentiallyDead */ true);
6780 (void)Success;
6781 assert(Success && "Did not expect the call base visit callback to fail!");
6782
6784 [](const IRPosition &, const AbstractAttribute *,
6785 bool &) -> std::optional<Value *> { return nullptr; };
6786 for (const auto &It : AllocationInfos)
6787 A.registerSimplificationCallback(IRPosition::callsite_returned(*It.first),
6788 SCB);
6789 for (const auto &It : DeallocationInfos)
6790 A.registerSimplificationCallback(IRPosition::callsite_returned(*It.first),
6791 SCB);
6792 }
6793
6794 const std::string getAsStr(Attributor *A) const override {
6795 unsigned NumH2SMallocs = 0, NumInvalidMallocs = 0;
6796 for (const auto &It : AllocationInfos) {
6797 if (It.second->Status == AllocationInfo::INVALID)
6798 ++NumInvalidMallocs;
6799 else
6800 ++NumH2SMallocs;
6801 }
6802 return "[H2S] Mallocs Good/Bad: " + std::to_string(NumH2SMallocs) + "/" +
6803 std::to_string(NumInvalidMallocs);
6804 }
6805
6806 /// See AbstractAttribute::trackStatistics().
6807 void trackStatistics() const override {
6808 STATS_DECL(
6809 MallocCalls, Function,
6810 "Number of malloc/calloc/aligned_alloc calls converted to allocas");
6811 for (const auto &It : AllocationInfos)
6812 if (It.second->Status != AllocationInfo::INVALID)
6813 ++BUILD_STAT_NAME(MallocCalls, Function);
6814 }
6815
6816 bool isAssumedHeapToStack(const CallBase &CB) const override {
6817 if (isValidState())
6818 if (AllocationInfo *AI =
6819 AllocationInfos.lookup(const_cast<CallBase *>(&CB)))
6820 return AI->Status != AllocationInfo::INVALID;
6821 return false;
6822 }
6823
6824 bool isAssumedHeapToStackRemovedFree(CallBase &CB) const override {
6825 if (!isValidState())
6826 return false;
6827
6828 for (const auto &It : AllocationInfos) {
6829 AllocationInfo &AI = *It.second;
6830 if (AI.Status == AllocationInfo::INVALID)
6831 continue;
6832
6833 if (AI.PotentialFreeCalls.count(&CB))
6834 return true;
6835 }
6836
6837 return false;
6838 }
6839
6840 ChangeStatus manifest(Attributor &A) override {
6841 assert(getState().isValidState() &&
6842 "Attempted to manifest an invalid state!");
6843
6844 ChangeStatus HasChanged = ChangeStatus::UNCHANGED;
6845 Function *F = getAnchorScope();
6846 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6847
6848 for (auto &It : AllocationInfos) {
6849 AllocationInfo &AI = *It.second;
6850 if (AI.Status == AllocationInfo::INVALID)
6851 continue;
6852
6853 for (CallBase *FreeCall : AI.PotentialFreeCalls) {
6854 LLVM_DEBUG(dbgs() << "H2S: Removing free call: " << *FreeCall << "\n");
6855 A.deleteAfterManifest(*FreeCall);
6856 HasChanged = ChangeStatus::CHANGED;
6857 }
6858
6859 LLVM_DEBUG(dbgs() << "H2S: Removing malloc-like call: " << *AI.CB
6860 << "\n");
6861
6862 auto Remark = [&](OptimizationRemark OR) {
6863 LibFunc IsAllocShared;
6864 if (TLI->getLibFunc(*AI.CB, IsAllocShared))
6865 if (IsAllocShared == LibFunc___kmpc_alloc_shared)
6866 return OR << "Moving globalized variable to the stack.";
6867 return OR << "Moving memory allocation from the heap to the stack.";
6868 };
6869 if (AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared)
6870 A.emitRemark<OptimizationRemark>(AI.CB, "OMP110", Remark);
6871 else
6872 A.emitRemark<OptimizationRemark>(AI.CB, "HeapToStack", Remark);
6873
6874 const DataLayout &DL = A.getInfoCache().getDL();
6875 Value *Size;
6876 std::optional<APInt> SizeAPI = getSize(A, *this, AI);
6877 if (SizeAPI) {
6878 Size = ConstantInt::get(AI.CB->getContext(), *SizeAPI);
6879 } else {
6880 LLVMContext &Ctx = AI.CB->getContext();
6881 ObjectSizeOpts Opts;
6882 ObjectSizeOffsetEvaluator Eval(DL, TLI, Ctx, Opts);
6883 SizeOffsetValue SizeOffsetPair = Eval.compute(AI.CB);
6884 assert(SizeOffsetPair != ObjectSizeOffsetEvaluator::unknown() &&
6885 cast<ConstantInt>(SizeOffsetPair.Offset)->isZero());
6886 Size = SizeOffsetPair.Size;
6887 }
6888
6889 BasicBlock::iterator IP = AI.MoveAllocaIntoEntry
6890 ? F->getEntryBlock().begin()
6891 : AI.CB->getIterator();
6892
6893 Align Alignment(1);
6894 if (MaybeAlign RetAlign = AI.CB->getRetAlign())
6895 Alignment = std::max(Alignment, *RetAlign);
6896 if (Value *Align = getAllocAlignment(AI.CB, TLI)) {
6897 std::optional<APInt> AlignmentAPI = getAPInt(A, *this, *Align);
6898 assert(AlignmentAPI && AlignmentAPI->getZExtValue() > 0 &&
6899 "Expected an alignment during manifest!");
6900 Alignment =
6901 std::max(Alignment, assumeAligned(AlignmentAPI->getZExtValue()));
6902 }
6903
6904 // TODO: Hoist the alloca towards the function entry.
6905 unsigned AS = DL.getAllocaAddrSpace();
6906 Instruction *Alloca =
6907 new AllocaInst(Type::getInt8Ty(F->getContext()), AS, Size, Alignment,
6908 AI.CB->getName() + ".h2s", IP);
6909
6910 if (Alloca->getType() != AI.CB->getType())
6911 Alloca = BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
6912 Alloca, AI.CB->getType(), "malloc_cast", AI.CB->getIterator());
6913
6914 auto *I8Ty = Type::getInt8Ty(F->getContext());
6915 auto *InitVal = getInitialValueOfAllocation(AI.CB, TLI, I8Ty);
6916 assert(InitVal &&
6917 "Must be able to materialize initial memory state of allocation");
6918
6919 A.changeAfterManifest(IRPosition::inst(*AI.CB), *Alloca);
6920
6921 if (auto *II = dyn_cast<InvokeInst>(AI.CB)) {
6922 auto *NBB = II->getNormalDest();
6923 BranchInst::Create(NBB, AI.CB->getParent());
6924 A.deleteAfterManifest(*AI.CB);
6925 } else {
6926 A.deleteAfterManifest(*AI.CB);
6927 }
6928
6929 // Initialize the alloca with the same value as used by the allocation
6930 // function. We can skip undef as the initial value of an alloc is
6931 // undef, and the memset would simply end up being DSEd.
6932 if (!isa<UndefValue>(InitVal)) {
6933 IRBuilder<> Builder(Alloca->getNextNode());
6934 // TODO: Use alignment above if align!=1
6935 Builder.CreateMemSet(Alloca, InitVal, Size, std::nullopt);
6936 }
6937 HasChanged = ChangeStatus::CHANGED;
6938 }
6939
6940 return HasChanged;
6941 }
6942
6943 std::optional<APInt> getAPInt(Attributor &A, const AbstractAttribute &AA,
6944 Value &V) {
6945 bool UsedAssumedInformation = false;
6946 std::optional<Constant *> SimpleV =
6947 A.getAssumedConstant(V, AA, UsedAssumedInformation);
6948 if (!SimpleV)
6949 return APInt(64, 0);
6950 if (auto *CI = dyn_cast_or_null<ConstantInt>(*SimpleV))
6951 return CI->getValue();
6952 return std::nullopt;
6953 }
6954
6955 std::optional<APInt> getSize(Attributor &A, const AbstractAttribute &AA,
6956 AllocationInfo &AI) {
6957 auto Mapper = [&](const Value *V) -> const Value * {
6958 bool UsedAssumedInformation = false;
6959 if (std::optional<Constant *> SimpleV =
6960 A.getAssumedConstant(*V, AA, UsedAssumedInformation))
6961 if (*SimpleV)
6962 return *SimpleV;
6963 return V;
6964 };
6965
6966 const Function *F = getAnchorScope();
6967 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6968 return getAllocSize(AI.CB, TLI, Mapper);
6969 }
6970
6971 /// Collection of all malloc-like calls in a function with associated
6972 /// information.
6973 MapVector<CallBase *, AllocationInfo *> AllocationInfos;
6974
6975 /// Collection of all free-like calls in a function with associated
6976 /// information.
6977 MapVector<CallBase *, DeallocationInfo *> DeallocationInfos;
6978
6979 ChangeStatus updateImpl(Attributor &A) override;
6980};
6981
6982ChangeStatus AAHeapToStackFunction::updateImpl(Attributor &A) {
6984 const Function *F = getAnchorScope();
6985 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6986
6987 const auto *LivenessAA =
6988 A.getAAFor<AAIsDead>(*this, IRPosition::function(*F), DepClassTy::NONE);
6989
6990 MustBeExecutedContextExplorer *Explorer =
6991 A.getInfoCache().getMustBeExecutedContextExplorer();
6992
6993 bool StackIsAccessibleByOtherThreads =
6994 A.getInfoCache().stackIsAccessibleByOtherThreads();
6995
6996 LoopInfo *LI =
6997 A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(*F);
6998 std::optional<bool> MayContainIrreducibleControl;
6999 auto IsInLoop = [&](BasicBlock &BB) {
7000 if (&F->getEntryBlock() == &BB)
7001 return false;
7002 if (!MayContainIrreducibleControl.has_value())
7003 MayContainIrreducibleControl = mayContainIrreducibleControl(*F, LI);
7004 if (*MayContainIrreducibleControl)
7005 return true;
7006 if (!LI)
7007 return true;
7008 return LI->getLoopFor(&BB) != nullptr;
7009 };
7010
7011 // Flag to ensure we update our deallocation information at most once per
7012 // updateImpl call and only if we use the free check reasoning.
7013 bool HasUpdatedFrees = false;
7014
7015 auto UpdateFrees = [&]() {
7016 HasUpdatedFrees = true;
7017
7018 for (auto &It : DeallocationInfos) {
7019 DeallocationInfo &DI = *It.second;
7020 // For now we cannot use deallocations that have unknown inputs, skip
7021 // them.
7022 if (DI.MightFreeUnknownObjects)
7023 continue;
7024
7025 // No need to analyze dead calls, ignore them instead.
7026 bool UsedAssumedInformation = false;
7027 if (A.isAssumedDead(*DI.CB, this, LivenessAA, UsedAssumedInformation,
7028 /* CheckBBLivenessOnly */ true))
7029 continue;
7030
7031 // Use the non-optimistic version to get the freed object.
7032 Value *Obj = getUnderlyingObject(DI.FreedOp);
7033 if (!Obj) {
7034 LLVM_DEBUG(dbgs() << "[H2S] Unknown underlying object for free!\n");
7035 DI.MightFreeUnknownObjects = true;
7036 continue;
7037 }
7038
7039 // Free of null and undef can be ignored as no-ops (or UB in the latter
7040 // case).
7042 continue;
7043
7044 CallBase *ObjCB = dyn_cast<CallBase>(Obj);
7045 if (!ObjCB) {
7046 LLVM_DEBUG(dbgs() << "[H2S] Free of a non-call object: " << *Obj
7047 << "\n");
7048 DI.MightFreeUnknownObjects = true;
7049 continue;
7050 }
7051
7052 AllocationInfo *AI = AllocationInfos.lookup(ObjCB);
7053 if (!AI) {
7054 LLVM_DEBUG(dbgs() << "[H2S] Free of a non-allocation object: " << *Obj
7055 << "\n");
7056 DI.MightFreeUnknownObjects = true;
7057 continue;
7058 }
7059
7060 DI.PotentialAllocationCalls.insert(ObjCB);
7061 }
7062 };
7063
7064 auto FreeCheck = [&](AllocationInfo &AI) {
7065 // If the stack is not accessible by other threads, the "must-free" logic
7066 // doesn't apply as the pointer could be shared and needs to be places in
7067 // "shareable" memory.
7068 if (!StackIsAccessibleByOtherThreads) {
7069 bool IsKnownNoSycn;
7071 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNoSycn)) {
7072 LLVM_DEBUG(
7073 dbgs() << "[H2S] found an escaping use, stack is not accessible by "
7074 "other threads and function is not nosync:\n");
7075 return false;
7076 }
7077 }
7078 if (!HasUpdatedFrees)
7079 UpdateFrees();
7080
7081 // TODO: Allow multi exit functions that have different free calls.
7082 if (AI.PotentialFreeCalls.size() != 1) {
7083 LLVM_DEBUG(dbgs() << "[H2S] did not find one free call but "
7084 << AI.PotentialFreeCalls.size() << "\n");
7085 return false;
7086 }
7087 CallBase *UniqueFree = *AI.PotentialFreeCalls.begin();
7088 DeallocationInfo *DI = DeallocationInfos.lookup(UniqueFree);
7089 if (!DI) {
7090 LLVM_DEBUG(
7091 dbgs() << "[H2S] unique free call was not known as deallocation call "
7092 << *UniqueFree << "\n");
7093 return false;
7094 }
7095 if (DI->MightFreeUnknownObjects) {
7096 LLVM_DEBUG(
7097 dbgs() << "[H2S] unique free call might free unknown allocations\n");
7098 return false;
7099 }
7100 if (DI->PotentialAllocationCalls.empty())
7101 return true;
7102 if (DI->PotentialAllocationCalls.size() > 1) {
7103 LLVM_DEBUG(dbgs() << "[H2S] unique free call might free "
7104 << DI->PotentialAllocationCalls.size()
7105 << " different allocations\n");
7106 return false;
7107 }
7108 if (*DI->PotentialAllocationCalls.begin() != AI.CB) {
7109 LLVM_DEBUG(
7110 dbgs()
7111 << "[H2S] unique free call not known to free this allocation but "
7112 << **DI->PotentialAllocationCalls.begin() << "\n");
7113 return false;
7114 }
7115
7116 // __kmpc_alloc_shared and __kmpc_alloc_free are by construction matched.
7117 if (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared) {
7118 Instruction *CtxI = isa<InvokeInst>(AI.CB) ? AI.CB : AI.CB->getNextNode();
7119 if (!Explorer || !Explorer->findInContextOf(UniqueFree, CtxI)) {
7120 LLVM_DEBUG(dbgs() << "[H2S] unique free call might not be executed "
7121 "with the allocation "
7122 << *UniqueFree << "\n");
7123 return false;
7124 }
7125 }
7126 return true;
7127 };
7128
7129 auto UsesCheck = [&](AllocationInfo &AI) {
7130 bool ValidUsesOnly = true;
7131
7132 auto Pred = [&](const Use &U, bool &Follow) -> bool {
7133 Instruction *UserI = cast<Instruction>(U.getUser());
7134 if (isa<LoadInst>(UserI))
7135 return true;
7136 if (auto *SI = dyn_cast<StoreInst>(UserI)) {
7137 if (SI->getValueOperand() == U.get()) {
7139 << "[H2S] escaping store to memory: " << *UserI << "\n");
7140 ValidUsesOnly = false;
7141 } else {
7142 // A store into the malloc'ed memory is fine.
7143 }
7144 return true;
7145 }
7146 if (auto *CB = dyn_cast<CallBase>(UserI)) {
7147 if (!CB->isArgOperand(&U) || CB->isLifetimeStartOrEnd())
7148 return true;
7149 if (DeallocationInfos.count(CB)) {
7150 AI.PotentialFreeCalls.insert(CB);
7151 return true;
7152 }
7153
7154 unsigned ArgNo = CB->getArgOperandNo(&U);
7155 auto CBIRP = IRPosition::callsite_argument(*CB, ArgNo);
7156
7157 bool IsKnownNoCapture;
7158 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
7159 A, this, CBIRP, DepClassTy::OPTIONAL, IsKnownNoCapture);
7160
7161 // If a call site argument use is nofree, we are fine.
7162 bool IsKnownNoFree;
7163 bool IsAssumedNoFree = AA::hasAssumedIRAttr<Attribute::NoFree>(
7164 A, this, CBIRP, DepClassTy::OPTIONAL, IsKnownNoFree);
7165
7166 if (!IsAssumedNoCapture ||
7167 (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared &&
7168 !IsAssumedNoFree)) {
7169 AI.HasPotentiallyFreeingUnknownUses |= !IsAssumedNoFree;
7170
7171 // Emit a missed remark if this is missed OpenMP globalization.
7172 auto Remark = [&](OptimizationRemarkMissed ORM) {
7173 return ORM
7174 << "Could not move globalized variable to the stack. "
7175 "Variable is potentially captured in call. Mark "
7176 "parameter as `__attribute__((noescape))` to override.";
7177 };
7178
7179 if (ValidUsesOnly &&
7180 AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared)
7181 A.emitRemark<OptimizationRemarkMissed>(CB, "OMP113", Remark);
7182
7183 LLVM_DEBUG(dbgs() << "[H2S] Bad user: " << *UserI << "\n");
7184 ValidUsesOnly = false;
7185 }
7186 return true;
7187 }
7188
7189 if (isa<GetElementPtrInst>(UserI) || isa<BitCastInst>(UserI) ||
7190 isa<PHINode>(UserI) || isa<SelectInst>(UserI)) {
7191 Follow = true;
7192 return true;
7193 }
7194 // Unknown user for which we can not track uses further (in a way that
7195 // makes sense).
7196 LLVM_DEBUG(dbgs() << "[H2S] Unknown user: " << *UserI << "\n");
7197 ValidUsesOnly = false;
7198 return true;
7199 };
7200 if (!A.checkForAllUses(Pred, *this, *AI.CB, /* CheckBBLivenessOnly */ false,
7201 DepClassTy::OPTIONAL, /* IgnoreDroppableUses */ true,
7202 [&](const Use &OldU, const Use &NewU) {
7203 auto *SI = dyn_cast<StoreInst>(OldU.getUser());
7204 return !SI || StackIsAccessibleByOtherThreads ||
7205 AA::isAssumedThreadLocalObject(
7206 A, *SI->getPointerOperand(), *this);
7207 }))
7208 return false;
7209 return ValidUsesOnly;
7210 };
7211
7212 // The actual update starts here. We look at all allocations and depending on
7213 // their status perform the appropriate check(s).
7214 for (auto &It : AllocationInfos) {
7215 AllocationInfo &AI = *It.second;
7216 if (AI.Status == AllocationInfo::INVALID)
7217 continue;
7218
7219 if (Value *Align = getAllocAlignment(AI.CB, TLI)) {
7220 std::optional<APInt> APAlign = getAPInt(A, *this, *Align);
7221 if (!APAlign) {
7222 // Can't generate an alloca which respects the required alignment
7223 // on the allocation.
7224 LLVM_DEBUG(dbgs() << "[H2S] Unknown allocation alignment: " << *AI.CB
7225 << "\n");
7226 AI.Status = AllocationInfo::INVALID;
7228 continue;
7229 }
7230 if (APAlign->ugt(llvm::Value::MaximumAlignment) ||
7231 !APAlign->isPowerOf2()) {
7232 LLVM_DEBUG(dbgs() << "[H2S] Invalid allocation alignment: " << APAlign
7233 << "\n");
7234 AI.Status = AllocationInfo::INVALID;
7236 continue;
7237 }
7238 }
7239
7240 std::optional<APInt> Size = getSize(A, *this, AI);
7241 if (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared &&
7242 MaxHeapToStackSize != -1) {
7243 if (!Size || Size->ugt(MaxHeapToStackSize)) {
7244 LLVM_DEBUG({
7245 if (!Size)
7246 dbgs() << "[H2S] Unknown allocation size: " << *AI.CB << "\n";
7247 else
7248 dbgs() << "[H2S] Allocation size too large: " << *AI.CB << " vs. "
7249 << MaxHeapToStackSize << "\n";
7250 });
7251
7252 AI.Status = AllocationInfo::INVALID;
7254 continue;
7255 }
7256 }
7257
7258 switch (AI.Status) {
7259 case AllocationInfo::STACK_DUE_TO_USE:
7260 if (UsesCheck(AI))
7261 break;
7262 AI.Status = AllocationInfo::STACK_DUE_TO_FREE;
7263 [[fallthrough]];
7264 case AllocationInfo::STACK_DUE_TO_FREE:
7265 if (FreeCheck(AI))
7266 break;
7267 AI.Status = AllocationInfo::INVALID;
7269 break;
7270 case AllocationInfo::INVALID:
7271 llvm_unreachable("Invalid allocations should never reach this point!");
7272 };
7273
7274 // Check if we still think we can move it into the entry block. If the
7275 // alloca comes from a converted __kmpc_alloc_shared then we can usually
7276 // ignore the potential compilations associated with loops.
7277 bool IsGlobalizedLocal =
7278 AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared;
7279 if (AI.MoveAllocaIntoEntry &&
7280 (!Size.has_value() ||
7281 (!IsGlobalizedLocal && IsInLoop(*AI.CB->getParent()))))
7282 AI.MoveAllocaIntoEntry = false;
7283 }
7284
7285 return Changed;
7286}
7287} // namespace
7288
7289/// ----------------------- Privatizable Pointers ------------------------------
7290namespace {
7291struct AAPrivatizablePtrImpl : public AAPrivatizablePtr {
7292 AAPrivatizablePtrImpl(const IRPosition &IRP, Attributor &A)
7293 : AAPrivatizablePtr(IRP, A), PrivatizableType(std::nullopt) {}
7294
7295 ChangeStatus indicatePessimisticFixpoint() override {
7296 AAPrivatizablePtr::indicatePessimisticFixpoint();
7297 PrivatizableType = nullptr;
7298 return ChangeStatus::CHANGED;
7299 }
7300
7301 /// Identify the type we can chose for a private copy of the underlying
7302 /// argument. std::nullopt means it is not clear yet, nullptr means there is
7303 /// none.
7304 virtual std::optional<Type *> identifyPrivatizableType(Attributor &A) = 0;
7305
7306 /// Return a privatizable type that encloses both T0 and T1.
7307 /// TODO: This is merely a stub for now as we should manage a mapping as well.
7308 std::optional<Type *> combineTypes(std::optional<Type *> T0,
7309 std::optional<Type *> T1) {
7310 if (!T0)
7311 return T1;
7312 if (!T1)
7313 return T0;
7314 if (T0 == T1)
7315 return T0;
7316 return nullptr;
7317 }
7318
7319 std::optional<Type *> getPrivatizableType() const override {
7320 return PrivatizableType;
7321 }
7322
7323 const std::string getAsStr(Attributor *A) const override {
7324 return isAssumedPrivatizablePtr() ? "[priv]" : "[no-priv]";
7325 }
7326
7327protected:
7328 std::optional<Type *> PrivatizableType;
7329};
7330
7331// TODO: Do this for call site arguments (probably also other values) as well.
7332
7333struct AAPrivatizablePtrArgument final : public AAPrivatizablePtrImpl {
7334 AAPrivatizablePtrArgument(const IRPosition &IRP, Attributor &A)
7335 : AAPrivatizablePtrImpl(IRP, A) {}
7336
7337 /// See AAPrivatizablePtrImpl::identifyPrivatizableType(...)
7338 std::optional<Type *> identifyPrivatizableType(Attributor &A) override {
7339 // If this is a byval argument and we know all the call sites (so we can
7340 // rewrite them), there is no need to check them explicitly.
7341 bool UsedAssumedInformation = false;
7343 A.getAttrs(getIRPosition(), {Attribute::ByVal}, Attrs,
7344 /* IgnoreSubsumingPositions */ true);
7345 if (!Attrs.empty() &&
7346 A.checkForAllCallSites([](AbstractCallSite ACS) { return true; }, *this,
7347 true, UsedAssumedInformation))
7348 return Attrs[0].getValueAsType();
7349
7350 std::optional<Type *> Ty;
7351 unsigned ArgNo = getIRPosition().getCallSiteArgNo();
7352
7353 // Make sure the associated call site argument has the same type at all call
7354 // sites and it is an allocation we know is safe to privatize, for now that
7355 // means we only allow alloca instructions.
7356 // TODO: We can additionally analyze the accesses in the callee to create
7357 // the type from that information instead. That is a little more
7358 // involved and will be done in a follow up patch.
7359 auto CallSiteCheck = [&](AbstractCallSite ACS) {
7360 IRPosition ACSArgPos = IRPosition::callsite_argument(ACS, ArgNo);
7361 // Check if a coresponding argument was found or if it is one not
7362 // associated (which can happen for callback calls).
7363 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
7364 return false;
7365
7366 // Check that all call sites agree on a type.
7367 auto *PrivCSArgAA =
7368 A.getAAFor<AAPrivatizablePtr>(*this, ACSArgPos, DepClassTy::REQUIRED);
7369 if (!PrivCSArgAA)
7370 return false;
7371 std::optional<Type *> CSTy = PrivCSArgAA->getPrivatizableType();
7372
7373 LLVM_DEBUG({
7374 dbgs() << "[AAPrivatizablePtr] ACSPos: " << ACSArgPos << ", CSTy: ";
7375 if (CSTy && *CSTy)
7376 (*CSTy)->print(dbgs());
7377 else if (CSTy)
7378 dbgs() << "<nullptr>";
7379 else
7380 dbgs() << "<none>";
7381 });
7382
7383 Ty = combineTypes(Ty, CSTy);
7384
7385 LLVM_DEBUG({
7386 dbgs() << " : New Type: ";
7387 if (Ty && *Ty)
7388 (*Ty)->print(dbgs());
7389 else if (Ty)
7390 dbgs() << "<nullptr>";
7391 else
7392 dbgs() << "<none>";
7393 dbgs() << "\n";
7394 });
7395
7396 return !Ty || *Ty;
7397 };
7398
7399 if (!A.checkForAllCallSites(CallSiteCheck, *this, true,
7400 UsedAssumedInformation))
7401 return nullptr;
7402 return Ty;
7403 }
7404
7405 /// See AbstractAttribute::updateImpl(...).
7406 ChangeStatus updateImpl(Attributor &A) override {
7407 PrivatizableType = identifyPrivatizableType(A);
7408 if (!PrivatizableType)
7409 return ChangeStatus::UNCHANGED;
7410 if (!*PrivatizableType)
7411 return indicatePessimisticFixpoint();
7412
7413 // The dependence is optional so we don't give up once we give up on the
7414 // alignment.
7415 A.getAAFor<AAAlign>(*this, IRPosition::value(getAssociatedValue()),
7416 DepClassTy::OPTIONAL);
7417
7418 // Avoid arguments with padding for now.
7419 if (!A.hasAttr(getIRPosition(), Attribute::ByVal) &&
7420 !isDenselyPacked(*PrivatizableType, A.getInfoCache().getDL())) {
7421 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Padding detected\n");
7422 return indicatePessimisticFixpoint();
7423 }
7424
7425 // Collect the types that will replace the privatizable type in the function
7426 // signature.
7427 SmallVector<Type *, 16> ReplacementTypes;
7428 identifyReplacementTypes(*PrivatizableType, ReplacementTypes);
7429
7430 // Verify callee and caller agree on how the promoted argument would be
7431 // passed.
7432 Function &Fn = *getIRPosition().getAnchorScope();
7433 const auto *TTI =
7434 A.getInfoCache().getAnalysisResultForFunction<TargetIRAnalysis>(Fn);
7435 if (!TTI) {
7436 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Missing TTI for function "
7437 << Fn.getName() << "\n");
7438 return indicatePessimisticFixpoint();
7439 }
7440
7441 auto CallSiteCheck = [&](AbstractCallSite ACS) {
7442 CallBase *CB = ACS.getInstruction();
7443 return TTI->areTypesABICompatible(
7444 CB->getCaller(),
7446 ReplacementTypes);
7447 };
7448 bool UsedAssumedInformation = false;
7449 if (!A.checkForAllCallSites(CallSiteCheck, *this, true,
7450 UsedAssumedInformation)) {
7451 LLVM_DEBUG(
7452 dbgs() << "[AAPrivatizablePtr] ABI incompatibility detected for "
7453 << Fn.getName() << "\n");
7454 return indicatePessimisticFixpoint();
7455 }
7456
7457 // Register a rewrite of the argument.
7458 Argument *Arg = getAssociatedArgument();
7459 if (!A.isValidFunctionSignatureRewrite(*Arg, ReplacementTypes)) {
7460 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Rewrite not valid\n");
7461 return indicatePessimisticFixpoint();
7462 }
7463
7464 unsigned ArgNo = Arg->getArgNo();
7465
7466 // Helper to check if for the given call site the associated argument is
7467 // passed to a callback where the privatization would be different.
7468 auto IsCompatiblePrivArgOfCallback = [&](CallBase &CB) {
7469 SmallVector<const Use *, 4> CallbackUses;
7470 AbstractCallSite::getCallbackUses(CB, CallbackUses);
7471 for (const Use *U : CallbackUses) {
7472 AbstractCallSite CBACS(U);
7473 assert(CBACS && CBACS.isCallbackCall());
7474 for (Argument &CBArg : CBACS.getCalledFunction()->args()) {
7475 int CBArgNo = CBACS.getCallArgOperandNo(CBArg);
7476
7477 LLVM_DEBUG({
7478 dbgs()
7479 << "[AAPrivatizablePtr] Argument " << *Arg
7480 << "check if can be privatized in the context of its parent ("
7481 << Arg->getParent()->getName()
7482 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7483 "callback ("
7484 << CBArgNo << "@" << CBACS.getCalledFunction()->getName()
7485 << ")\n[AAPrivatizablePtr] " << CBArg << " : "
7486 << CBACS.getCallArgOperand(CBArg) << " vs "
7487 << CB.getArgOperand(ArgNo) << "\n"
7488 << "[AAPrivatizablePtr] " << CBArg << " : "
7489 << CBACS.getCallArgOperandNo(CBArg) << " vs " << ArgNo << "\n";
7490 });
7491
7492 if (CBArgNo != int(ArgNo))
7493 continue;
7494 const auto *CBArgPrivAA = A.getAAFor<AAPrivatizablePtr>(
7495 *this, IRPosition::argument(CBArg), DepClassTy::REQUIRED);
7496 if (CBArgPrivAA && CBArgPrivAA->isValidState()) {
7497 auto CBArgPrivTy = CBArgPrivAA->getPrivatizableType();
7498 if (!CBArgPrivTy)
7499 continue;
7500 if (*CBArgPrivTy == PrivatizableType)
7501 continue;
7502 }
7503
7504 LLVM_DEBUG({
7505 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7506 << " cannot be privatized in the context of its parent ("
7507 << Arg->getParent()->getName()
7508 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7509 "callback ("
7510 << CBArgNo << "@" << CBACS.getCalledFunction()->getName()
7511 << ").\n[AAPrivatizablePtr] for which the argument "
7512 "privatization is not compatible.\n";
7513 });
7514 return false;
7515 }
7516 }
7517 return true;
7518 };
7519
7520 // Helper to check if for the given call site the associated argument is
7521 // passed to a direct call where the privatization would be different.
7522 auto IsCompatiblePrivArgOfDirectCS = [&](AbstractCallSite ACS) {
7523 CallBase *DC = cast<CallBase>(ACS.getInstruction());
7524 int DCArgNo = ACS.getCallArgOperandNo(ArgNo);
7525 assert(DCArgNo >= 0 && unsigned(DCArgNo) < DC->arg_size() &&
7526 "Expected a direct call operand for callback call operand");
7527
7528 Function *DCCallee =
7530 LLVM_DEBUG({
7531 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7532 << " check if be privatized in the context of its parent ("
7533 << Arg->getParent()->getName()
7534 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7535 "direct call of ("
7536 << DCArgNo << "@" << DCCallee->getName() << ").\n";
7537 });
7538
7539 if (unsigned(DCArgNo) < DCCallee->arg_size()) {
7540 const auto *DCArgPrivAA = A.getAAFor<AAPrivatizablePtr>(
7541 *this, IRPosition::argument(*DCCallee->getArg(DCArgNo)),
7542 DepClassTy::REQUIRED);
7543 if (DCArgPrivAA && DCArgPrivAA->isValidState()) {
7544 auto DCArgPrivTy = DCArgPrivAA->getPrivatizableType();
7545 if (!DCArgPrivTy)
7546 return true;
7547 if (*DCArgPrivTy == PrivatizableType)
7548 return true;
7549 }
7550 }
7551
7552 LLVM_DEBUG({
7553 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7554 << " cannot be privatized in the context of its parent ("
7555 << Arg->getParent()->getName()
7556 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7557 "direct call of ("
7559 << ").\n[AAPrivatizablePtr] for which the argument "
7560 "privatization is not compatible.\n";
7561 });
7562 return false;
7563 };
7564
7565 // Helper to check if the associated argument is used at the given abstract
7566 // call site in a way that is incompatible with the privatization assumed
7567 // here.
7568 auto IsCompatiblePrivArgOfOtherCallSite = [&](AbstractCallSite ACS) {
7569 if (ACS.isDirectCall())
7570 return IsCompatiblePrivArgOfCallback(*ACS.getInstruction());
7571 if (ACS.isCallbackCall())
7572 return IsCompatiblePrivArgOfDirectCS(ACS);
7573 return false;
7574 };
7575
7576 if (!A.checkForAllCallSites(IsCompatiblePrivArgOfOtherCallSite, *this, true,
7577 UsedAssumedInformation))
7578 return indicatePessimisticFixpoint();
7579
7580 return ChangeStatus::UNCHANGED;
7581 }
7582
7583 /// Given a type to private \p PrivType, collect the constituates (which are
7584 /// used) in \p ReplacementTypes.
7585 static void
7586 identifyReplacementTypes(Type *PrivType,
7587 SmallVectorImpl<Type *> &ReplacementTypes) {
7588 // TODO: For now we expand the privatization type to the fullest which can
7589 // lead to dead arguments that need to be removed later.
7590 assert(PrivType && "Expected privatizable type!");
7591
7592 // Traverse the type, extract constituate types on the outermost level.
7593 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7594 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++)
7595 ReplacementTypes.push_back(PrivStructType->getElementType(u));
7596 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7597 ReplacementTypes.append(PrivArrayType->getNumElements(),
7598 PrivArrayType->getElementType());
7599 } else {
7600 ReplacementTypes.push_back(PrivType);
7601 }
7602 }
7603
7604 /// Initialize \p Base according to the type \p PrivType at position \p IP.
7605 /// The values needed are taken from the arguments of \p F starting at
7606 /// position \p ArgNo.
7607 static void createInitialization(Type *PrivType, Value &Base, Function &F,
7608 unsigned ArgNo, BasicBlock::iterator IP) {
7609 assert(PrivType && "Expected privatizable type!");
7610
7611 IRBuilder<NoFolder> IRB(IP->getParent(), IP);
7612 const DataLayout &DL = F.getDataLayout();
7613
7614 // Traverse the type, build GEPs and stores.
7615 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7616 const StructLayout *PrivStructLayout = DL.getStructLayout(PrivStructType);
7617 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++) {
7618 Value *Ptr =
7619 constructPointer(&Base, PrivStructLayout->getElementOffset(u), IRB);
7620 new StoreInst(F.getArg(ArgNo + u), Ptr, IP);
7621 }
7622 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7623 Type *PointeeTy = PrivArrayType->getElementType();
7624 uint64_t PointeeTySize = DL.getTypeStoreSize(PointeeTy);
7625 for (unsigned u = 0, e = PrivArrayType->getNumElements(); u < e; u++) {
7626 Value *Ptr = constructPointer(&Base, u * PointeeTySize, IRB);
7627 new StoreInst(F.getArg(ArgNo + u), Ptr, IP);
7628 }
7629 } else {
7630 new StoreInst(F.getArg(ArgNo), &Base, IP);
7631 }
7632 }
7633
7634 /// Extract values from \p Base according to the type \p PrivType at the
7635 /// call position \p ACS. The values are appended to \p ReplacementValues.
7636 void createReplacementValues(Align Alignment, Type *PrivType,
7637 AbstractCallSite ACS, Value *Base,
7638 SmallVectorImpl<Value *> &ReplacementValues) {
7639 assert(Base && "Expected base value!");
7640 assert(PrivType && "Expected privatizable type!");
7641 Instruction *IP = ACS.getInstruction();
7642
7643 IRBuilder<NoFolder> IRB(IP);
7644 const DataLayout &DL = IP->getDataLayout();
7645
7646 // Traverse the type, build GEPs and loads.
7647 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7648 const StructLayout *PrivStructLayout = DL.getStructLayout(PrivStructType);
7649 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++) {
7650 Type *PointeeTy = PrivStructType->getElementType(u);
7651 Value *Ptr =
7652 constructPointer(Base, PrivStructLayout->getElementOffset(u), IRB);
7653 LoadInst *L = new LoadInst(PointeeTy, Ptr, "", IP->getIterator());
7654 L->setAlignment(Alignment);
7655 ReplacementValues.push_back(L);
7656 }
7657 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7658 Type *PointeeTy = PrivArrayType->getElementType();
7659 uint64_t PointeeTySize = DL.getTypeStoreSize(PointeeTy);
7660 for (unsigned u = 0, e = PrivArrayType->getNumElements(); u < e; u++) {
7661 Value *Ptr = constructPointer(Base, u * PointeeTySize, IRB);
7662 LoadInst *L = new LoadInst(PointeeTy, Ptr, "", IP->getIterator());
7663 L->setAlignment(Alignment);
7664 ReplacementValues.push_back(L);
7665 }
7666 } else {
7667 LoadInst *L = new LoadInst(PrivType, Base, "", IP->getIterator());
7668 L->setAlignment(Alignment);
7669 ReplacementValues.push_back(L);
7670 }
7671 }
7672
7673 /// See AbstractAttribute::manifest(...)
7674 ChangeStatus manifest(Attributor &A) override {
7675 if (!PrivatizableType)
7676 return ChangeStatus::UNCHANGED;
7677 assert(*PrivatizableType && "Expected privatizable type!");
7678
7679 // Collect all tail calls in the function as we cannot allow new allocas to
7680 // escape into tail recursion.
7681 // TODO: Be smarter about new allocas escaping into tail calls.
7683 bool UsedAssumedInformation = false;
7684 if (!A.checkForAllInstructions(
7685 [&](Instruction &I) {
7686 CallInst &CI = cast<CallInst>(I);
7687 if (CI.isTailCall())
7688 TailCalls.push_back(&CI);
7689 return true;
7690 },
7691 *this, {Instruction::Call}, UsedAssumedInformation))
7692 return ChangeStatus::UNCHANGED;
7693
7694 Argument *Arg = getAssociatedArgument();
7695 // Query AAAlign attribute for alignment of associated argument to
7696 // determine the best alignment of loads.
7697 const auto *AlignAA =
7698 A.getAAFor<AAAlign>(*this, IRPosition::value(*Arg), DepClassTy::NONE);
7699
7700 // Callback to repair the associated function. A new alloca is placed at the
7701 // beginning and initialized with the values passed through arguments. The
7702 // new alloca replaces the use of the old pointer argument.
7704 [=](const Attributor::ArgumentReplacementInfo &ARI,
7705 Function &ReplacementFn, Function::arg_iterator ArgIt) {
7706 BasicBlock &EntryBB = ReplacementFn.getEntryBlock();
7708 const DataLayout &DL = IP->getDataLayout();
7709 unsigned AS = DL.getAllocaAddrSpace();
7710 Instruction *AI = new AllocaInst(*PrivatizableType, AS,
7711 Arg->getName() + ".priv", IP);
7712 createInitialization(*PrivatizableType, *AI, ReplacementFn,
7713 ArgIt->getArgNo(), IP);
7714
7715 if (AI->getType() != Arg->getType())
7716 AI = BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
7717 AI, Arg->getType(), "", IP);
7718 Arg->replaceAllUsesWith(AI);
7719
7720 for (CallInst *CI : TailCalls)
7721 CI->setTailCall(false);
7722 };
7723
7724 // Callback to repair a call site of the associated function. The elements
7725 // of the privatizable type are loaded prior to the call and passed to the
7726 // new function version.
7728 [=](const Attributor::ArgumentReplacementInfo &ARI,
7729 AbstractCallSite ACS, SmallVectorImpl<Value *> &NewArgOperands) {
7730 // When no alignment is specified for the load instruction,
7731 // natural alignment is assumed.
7732 createReplacementValues(
7733 AlignAA ? AlignAA->getAssumedAlign() : Align(0),
7734 *PrivatizableType, ACS,
7735 ACS.getCallArgOperand(ARI.getReplacedArg().getArgNo()),
7736 NewArgOperands);
7737 };
7738
7739 // Collect the types that will replace the privatizable type in the function
7740 // signature.
7741 SmallVector<Type *, 16> ReplacementTypes;
7742 identifyReplacementTypes(*PrivatizableType, ReplacementTypes);
7743
7744 // Register a rewrite of the argument.
7745 if (A.registerFunctionSignatureRewrite(*Arg, ReplacementTypes,
7746 std::move(FnRepairCB),
7747 std::move(ACSRepairCB)))
7748 return ChangeStatus::CHANGED;
7749 return ChangeStatus::UNCHANGED;
7750 }
7751
7752 /// See AbstractAttribute::trackStatistics()
7753 void trackStatistics() const override {
7754 STATS_DECLTRACK_ARG_ATTR(privatizable_ptr);
7755 }
7756};
7757
7758struct AAPrivatizablePtrFloating : public AAPrivatizablePtrImpl {
7759 AAPrivatizablePtrFloating(const IRPosition &IRP, Attributor &A)
7760 : AAPrivatizablePtrImpl(IRP, A) {}
7761
7762 /// See AbstractAttribute::initialize(...).
7763 void initialize(Attributor &A) override {
7764 // TODO: We can privatize more than arguments.
7765 indicatePessimisticFixpoint();
7766 }
7767
7768 ChangeStatus updateImpl(Attributor &A) override {
7769 llvm_unreachable("AAPrivatizablePtr(Floating|Returned|CallSiteReturned)::"
7770 "updateImpl will not be called");
7771 }
7772
7773 /// See AAPrivatizablePtrImpl::identifyPrivatizableType(...)
7774 std::optional<Type *> identifyPrivatizableType(Attributor &A) override {
7775 Value *Obj = getUnderlyingObject(&getAssociatedValue());
7776 if (!Obj) {
7777 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] No underlying object found!\n");
7778 return nullptr;
7779 }
7780
7781 if (auto *AI = dyn_cast<AllocaInst>(Obj))
7782 if (auto *CI = dyn_cast<ConstantInt>(AI->getArraySize()))
7783 if (CI->isOne())
7784 return AI->getAllocatedType();
7785 if (auto *Arg = dyn_cast<Argument>(Obj)) {
7786 auto *PrivArgAA = A.getAAFor<AAPrivatizablePtr>(
7787 *this, IRPosition::argument(*Arg), DepClassTy::REQUIRED);
7788 if (PrivArgAA && PrivArgAA->isAssumedPrivatizablePtr())
7789 return PrivArgAA->getPrivatizableType();
7790 }
7791
7792 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Underlying object neither valid "
7793 "alloca nor privatizable argument: "
7794 << *Obj << "!\n");
7795 return nullptr;
7796 }
7797
7798 /// See AbstractAttribute::trackStatistics()
7799 void trackStatistics() const override {
7800 STATS_DECLTRACK_FLOATING_ATTR(privatizable_ptr);
7801 }
7802};
7803
7804struct AAPrivatizablePtrCallSiteArgument final
7805 : public AAPrivatizablePtrFloating {
7806 AAPrivatizablePtrCallSiteArgument(const IRPosition &IRP, Attributor &A)
7807 : AAPrivatizablePtrFloating(IRP, A) {}
7808
7809 /// See AbstractAttribute::initialize(...).
7810 void initialize(Attributor &A) override {
7811 if (A.hasAttr(getIRPosition(), Attribute::ByVal))
7812 indicateOptimisticFixpoint();
7813 }
7814
7815 /// See AbstractAttribute::updateImpl(...).
7816 ChangeStatus updateImpl(Attributor &A) override {
7817 PrivatizableType = identifyPrivatizableType(A);
7818 if (!PrivatizableType)
7819 return ChangeStatus::UNCHANGED;
7820 if (!*PrivatizableType)
7821 return indicatePessimisticFixpoint();
7822
7823 const IRPosition &IRP = getIRPosition();
7824 bool IsKnownNoCapture;
7825 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
7826 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoCapture);
7827 if (!IsAssumedNoCapture) {
7828 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer might be captured!\n");
7829 return indicatePessimisticFixpoint();
7830 }
7831
7832 bool IsKnownNoAlias;
7834 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoAlias)) {
7835 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer might alias!\n");
7836 return indicatePessimisticFixpoint();
7837 }
7838
7839 bool IsKnown;
7840 if (!AA::isAssumedReadOnly(A, IRP, *this, IsKnown)) {
7841 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer is written!\n");
7842 return indicatePessimisticFixpoint();
7843 }
7844
7845 return ChangeStatus::UNCHANGED;
7846 }
7847
7848 /// See AbstractAttribute::trackStatistics()
7849 void trackStatistics() const override {
7850 STATS_DECLTRACK_CSARG_ATTR(privatizable_ptr);
7851 }
7852};
7853
7854struct AAPrivatizablePtrCallSiteReturned final
7855 : public AAPrivatizablePtrFloating {
7856 AAPrivatizablePtrCallSiteReturned(const IRPosition &IRP, Attributor &A)
7857 : AAPrivatizablePtrFloating(IRP, A) {}
7858
7859 /// See AbstractAttribute::initialize(...).
7860 void initialize(Attributor &A) override {
7861 // TODO: We can privatize more than arguments.
7862 indicatePessimisticFixpoint();
7863 }
7864
7865 /// See AbstractAttribute::trackStatistics()
7866 void trackStatistics() const override {
7867 STATS_DECLTRACK_CSRET_ATTR(privatizable_ptr);
7868 }
7869};
7870
7871struct AAPrivatizablePtrReturned final : public AAPrivatizablePtrFloating {
7872 AAPrivatizablePtrReturned(const IRPosition &IRP, Attributor &A)
7873 : AAPrivatizablePtrFloating(IRP, A) {}
7874
7875 /// See AbstractAttribute::initialize(...).
7876 void initialize(Attributor &A) override {
7877 // TODO: We can privatize more than arguments.
7878 indicatePessimisticFixpoint();
7879 }
7880
7881 /// See AbstractAttribute::trackStatistics()
7882 void trackStatistics() const override {
7883 STATS_DECLTRACK_FNRET_ATTR(privatizable_ptr);
7884 }
7885};
7886} // namespace
7887
7888/// -------------------- Memory Behavior Attributes ----------------------------
7889/// Includes read-none, read-only, and write-only.
7890/// ----------------------------------------------------------------------------
7891namespace {
7892struct AAMemoryBehaviorImpl : public AAMemoryBehavior {
7893 AAMemoryBehaviorImpl(const IRPosition &IRP, Attributor &A)
7894 : AAMemoryBehavior(IRP, A) {}
7895
7896 /// See AbstractAttribute::initialize(...).
7897 void initialize(Attributor &A) override {
7898 intersectAssumedBits(BEST_STATE);
7899 getKnownStateFromValue(A, getIRPosition(), getState());
7900 AAMemoryBehavior::initialize(A);
7901 }
7902
7903 /// Return the memory behavior information encoded in the IR for \p IRP.
7904 static void getKnownStateFromValue(Attributor &A, const IRPosition &IRP,
7905 BitIntegerState &State,
7906 bool IgnoreSubsumingPositions = false) {
7908 A.getAttrs(IRP, AttrKinds, Attrs, IgnoreSubsumingPositions);
7909 for (const Attribute &Attr : Attrs) {
7910 switch (Attr.getKindAsEnum()) {
7911 case Attribute::ReadNone:
7912 State.addKnownBits(NO_ACCESSES);
7913 break;
7914 case Attribute::ReadOnly:
7915 State.addKnownBits(NO_WRITES);
7916 break;
7917 case Attribute::WriteOnly:
7918 State.addKnownBits(NO_READS);
7919 break;
7920 default:
7921 llvm_unreachable("Unexpected attribute!");
7922 }
7923 }
7924
7925 if (auto *I = dyn_cast<Instruction>(&IRP.getAnchorValue())) {
7926 if (!I->mayReadFromMemory())
7927 State.addKnownBits(NO_READS);
7928 if (!I->mayWriteToMemory())
7929 State.addKnownBits(NO_WRITES);
7930 }
7931 }
7932
7933 /// See AbstractAttribute::getDeducedAttributes(...).
7934 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
7935 SmallVectorImpl<Attribute> &Attrs) const override {
7936 assert(Attrs.size() == 0);
7937 if (isAssumedReadNone())
7938 Attrs.push_back(Attribute::get(Ctx, Attribute::ReadNone));
7939 else if (isAssumedReadOnly())
7940 Attrs.push_back(Attribute::get(Ctx, Attribute::ReadOnly));
7941 else if (isAssumedWriteOnly())
7942 Attrs.push_back(Attribute::get(Ctx, Attribute::WriteOnly));
7943 assert(Attrs.size() <= 1);
7944 }
7945
7946 /// See AbstractAttribute::manifest(...).
7947 ChangeStatus manifest(Attributor &A) override {
7948 const IRPosition &IRP = getIRPosition();
7949
7950 if (A.hasAttr(IRP, Attribute::ReadNone,
7951 /* IgnoreSubsumingPositions */ true))
7952 return ChangeStatus::UNCHANGED;
7953
7954 // Check if we would improve the existing attributes first.
7955 SmallVector<Attribute, 4> DeducedAttrs;
7956 getDeducedAttributes(A, IRP.getAnchorValue().getContext(), DeducedAttrs);
7957 if (llvm::all_of(DeducedAttrs, [&](const Attribute &Attr) {
7958 return A.hasAttr(IRP, Attr.getKindAsEnum(),
7959 /* IgnoreSubsumingPositions */ true);
7960 }))
7961 return ChangeStatus::UNCHANGED;
7962
7963 // Clear existing attributes.
7964 A.removeAttrs(IRP, AttrKinds);
7965 // Clear conflicting writable attribute.
7966 if (isAssumedReadOnly())
7967 A.removeAttrs(IRP, Attribute::Writable);
7968
7969 // Use the generic manifest method.
7970 return IRAttribute::manifest(A);
7971 }
7972
7973 /// See AbstractState::getAsStr().
7974 const std::string getAsStr(Attributor *A) const override {
7975 if (isAssumedReadNone())
7976 return "readnone";
7977 if (isAssumedReadOnly())
7978 return "readonly";
7979 if (isAssumedWriteOnly())
7980 return "writeonly";
7981 return "may-read/write";
7982 }
7983
7984 /// The set of IR attributes AAMemoryBehavior deals with.
7985 static const Attribute::AttrKind AttrKinds[3];
7986};
7987
7988const Attribute::AttrKind AAMemoryBehaviorImpl::AttrKinds[] = {
7989 Attribute::ReadNone, Attribute::ReadOnly, Attribute::WriteOnly};
7990
7991/// Memory behavior attribute for a floating value.
7992struct AAMemoryBehaviorFloating : AAMemoryBehaviorImpl {
7993 AAMemoryBehaviorFloating(const IRPosition &IRP, Attributor &A)
7994 : AAMemoryBehaviorImpl(IRP, A) {}
7995
7996 /// See AbstractAttribute::updateImpl(...).
7997 ChangeStatus updateImpl(Attributor &A) override;
7998
7999 /// See AbstractAttribute::trackStatistics()
8000 void trackStatistics() const override {
8001 if (isAssumedReadNone())
8003 else if (isAssumedReadOnly())
8005 else if (isAssumedWriteOnly())
8007 }
8008
8009private:
8010 /// Return true if users of \p UserI might access the underlying
8011 /// variable/location described by \p U and should therefore be analyzed.
8012 bool followUsersOfUseIn(Attributor &A, const Use &U,
8013 const Instruction *UserI);
8014
8015 /// Update the state according to the effect of use \p U in \p UserI.
8016 void analyzeUseIn(Attributor &A, const Use &U, const Instruction *UserI);
8017};
8018
8019/// Memory behavior attribute for function argument.
8020struct AAMemoryBehaviorArgument : AAMemoryBehaviorFloating {
8021 AAMemoryBehaviorArgument(const IRPosition &IRP, Attributor &A)
8022 : AAMemoryBehaviorFloating(IRP, A) {}
8023
8024 /// See AbstractAttribute::initialize(...).
8025 void initialize(Attributor &A) override {
8026 intersectAssumedBits(BEST_STATE);
8027 const IRPosition &IRP = getIRPosition();
8028 // TODO: Make IgnoreSubsumingPositions a property of an IRAttribute so we
8029 // can query it when we use has/getAttr. That would allow us to reuse the
8030 // initialize of the base class here.
8031 bool HasByVal = A.hasAttr(IRP, {Attribute::ByVal},
8032 /* IgnoreSubsumingPositions */ true);
8033 getKnownStateFromValue(A, IRP, getState(),
8034 /* IgnoreSubsumingPositions */ HasByVal);
8035 }
8036
8037 ChangeStatus manifest(Attributor &A) override {
8038 // TODO: Pointer arguments are not supported on vectors of pointers yet.
8039 if (!getAssociatedValue().getType()->isPointerTy())
8040 return ChangeStatus::UNCHANGED;
8041
8042 // TODO: From readattrs.ll: "inalloca parameters are always
8043 // considered written"
8044 if (A.hasAttr(getIRPosition(),
8045 {Attribute::InAlloca, Attribute::Preallocated})) {
8046 removeKnownBits(NO_WRITES);
8047 removeAssumedBits(NO_WRITES);
8048 }
8049 A.removeAttrs(getIRPosition(), AttrKinds);
8050 return AAMemoryBehaviorFloating::manifest(A);
8051 }
8052
8053 /// See AbstractAttribute::trackStatistics()
8054 void trackStatistics() const override {
8055 if (isAssumedReadNone())
8056 STATS_DECLTRACK_ARG_ATTR(readnone)
8057 else if (isAssumedReadOnly())
8058 STATS_DECLTRACK_ARG_ATTR(readonly)
8059 else if (isAssumedWriteOnly())
8060 STATS_DECLTRACK_ARG_ATTR(writeonly)
8061 }
8062};
8063
8064struct AAMemoryBehaviorCallSiteArgument final : AAMemoryBehaviorArgument {
8065 AAMemoryBehaviorCallSiteArgument(const IRPosition &IRP, Attributor &A)
8066 : AAMemoryBehaviorArgument(IRP, A) {}
8067
8068 /// See AbstractAttribute::initialize(...).
8069 void initialize(Attributor &A) override {
8070 // If we don't have an associated attribute this is either a variadic call
8071 // or an indirect call, either way, nothing to do here.
8072 Argument *Arg = getAssociatedArgument();
8073 if (!Arg) {
8074 indicatePessimisticFixpoint();
8075 return;
8076 }
8077 if (Arg->hasByValAttr()) {
8078 addKnownBits(NO_WRITES);
8079 removeKnownBits(NO_READS);
8080 removeAssumedBits(NO_READS);
8081 }
8082 AAMemoryBehaviorArgument::initialize(A);
8083 if (getAssociatedFunction()->isDeclaration())
8084 indicatePessimisticFixpoint();
8085 }
8086
8087 /// See AbstractAttribute::updateImpl(...).
8088 ChangeStatus updateImpl(Attributor &A) override {
8089 // TODO: Once we have call site specific value information we can provide
8090 // call site specific liveness liveness information and then it makes
8091 // sense to specialize attributes for call sites arguments instead of
8092 // redirecting requests to the callee argument.
8093 Argument *Arg = getAssociatedArgument();
8094 const IRPosition &ArgPos = IRPosition::argument(*Arg);
8095 auto *ArgAA =
8096 A.getAAFor<AAMemoryBehavior>(*this, ArgPos, DepClassTy::REQUIRED);
8097 if (!ArgAA)
8098 return indicatePessimisticFixpoint();
8099 return clampStateAndIndicateChange(getState(), ArgAA->getState());
8100 }
8101
8102 /// See AbstractAttribute::trackStatistics()
8103 void trackStatistics() const override {
8104 if (isAssumedReadNone())
8106 else if (isAssumedReadOnly())
8108 else if (isAssumedWriteOnly())
8110 }
8111};
8112
8113/// Memory behavior attribute for a call site return position.
8114struct AAMemoryBehaviorCallSiteReturned final : AAMemoryBehaviorFloating {
8115 AAMemoryBehaviorCallSiteReturned(const IRPosition &IRP, Attributor &A)
8116 : AAMemoryBehaviorFloating(IRP, A) {}
8117
8118 /// See AbstractAttribute::initialize(...).
8119 void initialize(Attributor &A) override {
8120 AAMemoryBehaviorImpl::initialize(A);
8121 }
8122 /// See AbstractAttribute::manifest(...).
8123 ChangeStatus manifest(Attributor &A) override {
8124 // We do not annotate returned values.
8125 return ChangeStatus::UNCHANGED;
8126 }
8127
8128 /// See AbstractAttribute::trackStatistics()
8129 void trackStatistics() const override {}
8130};
8131
8132/// An AA to represent the memory behavior function attributes.
8133struct AAMemoryBehaviorFunction final : public AAMemoryBehaviorImpl {
8134 AAMemoryBehaviorFunction(const IRPosition &IRP, Attributor &A)
8135 : AAMemoryBehaviorImpl(IRP, A) {}
8136
8137 /// See AbstractAttribute::updateImpl(Attributor &A).
8138 ChangeStatus updateImpl(Attributor &A) override;
8139
8140 /// See AbstractAttribute::manifest(...).
8141 ChangeStatus manifest(Attributor &A) override {
8142 // TODO: It would be better to merge this with AAMemoryLocation, so that
8143 // we could determine read/write per location. This would also have the
8144 // benefit of only one place trying to manifest the memory attribute.
8145 Function &F = cast<Function>(getAnchorValue());
8147 if (isAssumedReadNone())
8148 ME = MemoryEffects::none();
8149 else if (isAssumedReadOnly())
8151 else if (isAssumedWriteOnly())
8153
8154 A.removeAttrs(getIRPosition(), AttrKinds);
8155 // Clear conflicting writable attribute.
8156 if (ME.onlyReadsMemory())
8157 for (Argument &Arg : F.args())
8158 A.removeAttrs(IRPosition::argument(Arg), Attribute::Writable);
8159 return A.manifestAttrs(getIRPosition(),
8160 Attribute::getWithMemoryEffects(F.getContext(), ME));
8161 }
8162
8163 /// See AbstractAttribute::trackStatistics()
8164 void trackStatistics() const override {
8165 if (isAssumedReadNone())
8166 STATS_DECLTRACK_FN_ATTR(readnone)
8167 else if (isAssumedReadOnly())
8168 STATS_DECLTRACK_FN_ATTR(readonly)
8169 else if (isAssumedWriteOnly())
8170 STATS_DECLTRACK_FN_ATTR(writeonly)
8171 }
8172};
8173
8174/// AAMemoryBehavior attribute for call sites.
8175struct AAMemoryBehaviorCallSite final
8176 : AACalleeToCallSite<AAMemoryBehavior, AAMemoryBehaviorImpl> {
8177 AAMemoryBehaviorCallSite(const IRPosition &IRP, Attributor &A)
8178 : AACalleeToCallSite<AAMemoryBehavior, AAMemoryBehaviorImpl>(IRP, A) {}
8179
8180 /// See AbstractAttribute::manifest(...).
8181 ChangeStatus manifest(Attributor &A) override {
8182 // TODO: Deduplicate this with AAMemoryBehaviorFunction.
8183 CallBase &CB = cast<CallBase>(getAnchorValue());
8185 if (isAssumedReadNone())
8186 ME = MemoryEffects::none();
8187 else if (isAssumedReadOnly())
8189 else if (isAssumedWriteOnly())
8191
8192 A.removeAttrs(getIRPosition(), AttrKinds);
8193 // Clear conflicting writable attribute.
8194 if (ME.onlyReadsMemory())
8195 for (Use &U : CB.args())
8196 A.removeAttrs(IRPosition::callsite_argument(CB, U.getOperandNo()),
8197 Attribute::Writable);
8198 return A.manifestAttrs(
8199 getIRPosition(), Attribute::getWithMemoryEffects(CB.getContext(), ME));
8200 }
8201
8202 /// See AbstractAttribute::trackStatistics()
8203 void trackStatistics() const override {
8204 if (isAssumedReadNone())
8205 STATS_DECLTRACK_CS_ATTR(readnone)
8206 else if (isAssumedReadOnly())
8207 STATS_DECLTRACK_CS_ATTR(readonly)
8208 else if (isAssumedWriteOnly())
8209 STATS_DECLTRACK_CS_ATTR(writeonly)
8210 }
8211};
8212
8213ChangeStatus AAMemoryBehaviorFunction::updateImpl(Attributor &A) {
8214
8215 // The current assumed state used to determine a change.
8216 auto AssumedState = getAssumed();
8217
8218 auto CheckRWInst = [&](Instruction &I) {
8219 // If the instruction has an own memory behavior state, use it to restrict
8220 // the local state. No further analysis is required as the other memory
8221 // state is as optimistic as it gets.
8222 if (const auto *CB = dyn_cast<CallBase>(&I)) {
8223 const auto *MemBehaviorAA = A.getAAFor<AAMemoryBehavior>(
8225 if (MemBehaviorAA) {
8226 intersectAssumedBits(MemBehaviorAA->getAssumed());
8227 return !isAtFixpoint();
8228 }
8229 }
8230
8231 // Remove access kind modifiers if necessary.
8232 if (I.mayReadFromMemory())
8233 removeAssumedBits(NO_READS);
8234 if (I.mayWriteToMemory())
8235 removeAssumedBits(NO_WRITES);
8236 return !isAtFixpoint();
8237 };
8238
8239 bool UsedAssumedInformation = false;
8240 if (!A.checkForAllReadWriteInstructions(CheckRWInst, *this,
8241 UsedAssumedInformation))
8242 return indicatePessimisticFixpoint();
8243
8244 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8246}
8247
8248ChangeStatus AAMemoryBehaviorFloating::updateImpl(Attributor &A) {
8249
8250 const IRPosition &IRP = getIRPosition();
8251 const IRPosition &FnPos = IRPosition::function_scope(IRP);
8252 AAMemoryBehavior::StateType &S = getState();
8253
8254 // First, check the function scope. We take the known information and we avoid
8255 // work if the assumed information implies the current assumed information for
8256 // this attribute. This is a valid for all but byval arguments.
8257 Argument *Arg = IRP.getAssociatedArgument();
8258 AAMemoryBehavior::base_t FnMemAssumedState =
8260 if (!Arg || !Arg->hasByValAttr()) {
8261 const auto *FnMemAA =
8262 A.getAAFor<AAMemoryBehavior>(*this, FnPos, DepClassTy::OPTIONAL);
8263 if (FnMemAA) {
8264 FnMemAssumedState = FnMemAA->getAssumed();
8265 S.addKnownBits(FnMemAA->getKnown());
8266 if ((S.getAssumed() & FnMemAA->getAssumed()) == S.getAssumed())
8268 }
8269 }
8270
8271 // The current assumed state used to determine a change.
8272 auto AssumedState = S.getAssumed();
8273
8274 // Make sure the value is not captured (except through "return"), if
8275 // it is, any information derived would be irrelevant anyway as we cannot
8276 // check the potential aliases introduced by the capture. However, no need
8277 // to fall back to anythign less optimistic than the function state.
8278 bool IsKnownNoCapture;
8279 const AANoCapture *ArgNoCaptureAA = nullptr;
8280 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::Captures>(
8281 A, this, IRP, DepClassTy::OPTIONAL, IsKnownNoCapture, false,
8282 &ArgNoCaptureAA);
8283
8284 if (!IsAssumedNoCapture &&
8285 (!ArgNoCaptureAA || !ArgNoCaptureAA->isAssumedNoCaptureMaybeReturned())) {
8286 S.intersectAssumedBits(FnMemAssumedState);
8287 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8289 }
8290
8291 // Visit and expand uses until all are analyzed or a fixpoint is reached.
8292 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
8293 Instruction *UserI = cast<Instruction>(U.getUser());
8294 LLVM_DEBUG(dbgs() << "[AAMemoryBehavior] Use: " << *U << " in " << *UserI
8295 << " \n");
8296
8297 // Droppable users, e.g., llvm::assume does not actually perform any action.
8298 if (UserI->isDroppable())
8299 return true;
8300
8301 // Check if the users of UserI should also be visited.
8302 Follow = followUsersOfUseIn(A, U, UserI);
8303
8304 // If UserI might touch memory we analyze the use in detail.
8305 if (UserI->mayReadOrWriteMemory())
8306 analyzeUseIn(A, U, UserI);
8307
8308 return !isAtFixpoint();
8309 };
8310
8311 if (!A.checkForAllUses(UsePred, *this, getAssociatedValue()))
8312 return indicatePessimisticFixpoint();
8313
8314 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8316}
8317
8318bool AAMemoryBehaviorFloating::followUsersOfUseIn(Attributor &A, const Use &U,
8319 const Instruction *UserI) {
8320 // The loaded value is unrelated to the pointer argument, no need to
8321 // follow the users of the load.
8322 if (isa<LoadInst>(UserI) || isa<ReturnInst>(UserI))
8323 return false;
8324
8325 // By default we follow all uses assuming UserI might leak information on U,
8326 // we have special handling for call sites operands though.
8327 const auto *CB = dyn_cast<CallBase>(UserI);
8328 if (!CB || !CB->isArgOperand(&U))
8329 return true;
8330
8331 // If the use is a call argument known not to be captured, the users of
8332 // the call do not need to be visited because they have to be unrelated to
8333 // the input. Note that this check is not trivial even though we disallow
8334 // general capturing of the underlying argument. The reason is that the
8335 // call might the argument "through return", which we allow and for which we
8336 // need to check call users.
8337 if (U.get()->getType()->isPointerTy()) {
8338 unsigned ArgNo = CB->getArgOperandNo(&U);
8339 bool IsKnownNoCapture;
8341 A, this, IRPosition::callsite_argument(*CB, ArgNo),
8342 DepClassTy::OPTIONAL, IsKnownNoCapture);
8343 }
8344
8345 return true;
8346}
8347
8348void AAMemoryBehaviorFloating::analyzeUseIn(Attributor &A, const Use &U,
8349 const Instruction *UserI) {
8350 assert(UserI->mayReadOrWriteMemory());
8351
8352 switch (UserI->getOpcode()) {
8353 default:
8354 // TODO: Handle all atomics and other side-effect operations we know of.
8355 break;
8356 case Instruction::Load:
8357 // Loads cause the NO_READS property to disappear.
8358 removeAssumedBits(NO_READS);
8359 return;
8360
8361 case Instruction::Store:
8362 // Stores cause the NO_WRITES property to disappear if the use is the
8363 // pointer operand. Note that while capturing was taken care of somewhere
8364 // else we need to deal with stores of the value that is not looked through.
8365 if (cast<StoreInst>(UserI)->getPointerOperand() == U.get())
8366 removeAssumedBits(NO_WRITES);
8367 else
8368 indicatePessimisticFixpoint();
8369 return;
8370
8371 case Instruction::Call:
8372 case Instruction::CallBr:
8373 case Instruction::Invoke: {
8374 // For call sites we look at the argument memory behavior attribute (this
8375 // could be recursive!) in order to restrict our own state.
8376 const auto *CB = cast<CallBase>(UserI);
8377
8378 // Give up on operand bundles.
8379 if (CB->isBundleOperand(&U)) {
8380 indicatePessimisticFixpoint();
8381 return;
8382 }
8383
8384 // Calling a function does read the function pointer, maybe write it if the
8385 // function is self-modifying.
8386 if (CB->isCallee(&U)) {
8387 removeAssumedBits(NO_READS);
8388 break;
8389 }
8390
8391 // Adjust the possible access behavior based on the information on the
8392 // argument.
8393 IRPosition Pos;
8394 if (U.get()->getType()->isPointerTy())
8396 else
8398 const auto *MemBehaviorAA =
8399 A.getAAFor<AAMemoryBehavior>(*this, Pos, DepClassTy::OPTIONAL);
8400 if (!MemBehaviorAA)
8401 break;
8402 // "assumed" has at most the same bits as the MemBehaviorAA assumed
8403 // and at least "known".
8404 intersectAssumedBits(MemBehaviorAA->getAssumed());
8405 return;
8406 }
8407 };
8408
8409 // Generally, look at the "may-properties" and adjust the assumed state if we
8410 // did not trigger special handling before.
8411 if (UserI->mayReadFromMemory())
8412 removeAssumedBits(NO_READS);
8413 if (UserI->mayWriteToMemory())
8414 removeAssumedBits(NO_WRITES);
8415}
8416} // namespace
8417
8418/// -------------------- Memory Locations Attributes ---------------------------
8419/// Includes read-none, argmemonly, inaccessiblememonly,
8420/// inaccessiblememorargmemonly
8421/// ----------------------------------------------------------------------------
8422
8425 if (0 == (MLK & AAMemoryLocation::NO_LOCATIONS))
8426 return "all memory";
8428 return "no memory";
8429 std::string S = "memory:";
8430 if (0 == (MLK & AAMemoryLocation::NO_LOCAL_MEM))
8431 S += "stack,";
8432 if (0 == (MLK & AAMemoryLocation::NO_CONST_MEM))
8433 S += "constant,";
8435 S += "internal global,";
8437 S += "external global,";
8438 if (0 == (MLK & AAMemoryLocation::NO_ARGUMENT_MEM))
8439 S += "argument,";
8441 S += "inaccessible,";
8442 if (0 == (MLK & AAMemoryLocation::NO_MALLOCED_MEM))
8443 S += "malloced,";
8444 if (0 == (MLK & AAMemoryLocation::NO_UNKOWN_MEM))
8445 S += "unknown,";
8446 S.pop_back();
8447 return S;
8448}
8449
8450namespace {
8451struct AAMemoryLocationImpl : public AAMemoryLocation {
8452
8453 AAMemoryLocationImpl(const IRPosition &IRP, Attributor &A)
8454 : AAMemoryLocation(IRP, A), Allocator(A.Allocator) {
8455 AccessKind2Accesses.fill(nullptr);
8456 }
8457
8458 ~AAMemoryLocationImpl() override {
8459 // The AccessSets are allocated via a BumpPtrAllocator, we call
8460 // the destructor manually.
8461 for (AccessSet *AS : AccessKind2Accesses)
8462 if (AS)
8463 AS->~AccessSet();
8464 }
8465
8466 /// See AbstractAttribute::initialize(...).
8467 void initialize(Attributor &A) override {
8468 intersectAssumedBits(BEST_STATE);
8469 getKnownStateFromValue(A, getIRPosition(), getState());
8470 AAMemoryLocation::initialize(A);
8471 }
8472
8473 /// Return the memory behavior information encoded in the IR for \p IRP.
8474 static void getKnownStateFromValue(Attributor &A, const IRPosition &IRP,
8475 BitIntegerState &State,
8476 bool IgnoreSubsumingPositions = false) {
8477 // For internal functions we ignore `argmemonly` and
8478 // `inaccessiblememorargmemonly` as we might break it via interprocedural
8479 // constant propagation. It is unclear if this is the best way but it is
8480 // unlikely this will cause real performance problems. If we are deriving
8481 // attributes for the anchor function we even remove the attribute in
8482 // addition to ignoring it.
8483 // TODO: A better way to handle this would be to add ~NO_GLOBAL_MEM /
8484 // MemoryEffects::Other as a possible location.
8485 bool UseArgMemOnly = true;
8486 Function *AnchorFn = IRP.getAnchorScope();
8487 if (AnchorFn && A.isRunOn(*AnchorFn))
8488 UseArgMemOnly = !AnchorFn->hasLocalLinkage();
8489
8491 A.getAttrs(IRP, {Attribute::Memory}, Attrs, IgnoreSubsumingPositions);
8492 for (const Attribute &Attr : Attrs) {
8493 // TODO: We can map MemoryEffects to Attributor locations more precisely.
8494 MemoryEffects ME = Attr.getMemoryEffects();
8495 if (ME.doesNotAccessMemory()) {
8496 State.addKnownBits(NO_LOCAL_MEM | NO_CONST_MEM);
8497 continue;
8498 }
8499 if (ME.onlyAccessesInaccessibleMem()) {
8500 State.addKnownBits(inverseLocation(NO_INACCESSIBLE_MEM, true, true));
8501 continue;
8502 }
8503 if (ME.onlyAccessesArgPointees()) {
8504 if (UseArgMemOnly)
8505 State.addKnownBits(inverseLocation(NO_ARGUMENT_MEM, true, true));
8506 else {
8507 // Remove location information, only keep read/write info.
8508 ME = MemoryEffects(ME.getModRef());
8509 A.manifestAttrs(IRP,
8510 Attribute::getWithMemoryEffects(
8511 IRP.getAnchorValue().getContext(), ME),
8512 /*ForceReplace*/ true);
8513 }
8514 continue;
8515 }
8517 if (UseArgMemOnly)
8518 State.addKnownBits(inverseLocation(
8519 NO_INACCESSIBLE_MEM | NO_ARGUMENT_MEM, true, true));
8520 else {
8521 // Remove location information, only keep read/write info.
8522 ME = MemoryEffects(ME.getModRef());
8523 A.manifestAttrs(IRP,
8524 Attribute::getWithMemoryEffects(
8525 IRP.getAnchorValue().getContext(), ME),
8526 /*ForceReplace*/ true);
8527 }
8528 continue;
8529 }
8530 }
8531 }
8532
8533 /// See AbstractAttribute::getDeducedAttributes(...).
8534 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
8535 SmallVectorImpl<Attribute> &Attrs) const override {
8536 // TODO: We can map Attributor locations to MemoryEffects more precisely.
8537 assert(Attrs.size() == 0);
8538 if (getIRPosition().getPositionKind() == IRPosition::IRP_FUNCTION) {
8539 if (isAssumedReadNone())
8540 Attrs.push_back(
8541 Attribute::getWithMemoryEffects(Ctx, MemoryEffects::none()));
8542 else if (isAssumedInaccessibleMemOnly())
8543 Attrs.push_back(Attribute::getWithMemoryEffects(
8545 else if (isAssumedArgMemOnly())
8546 Attrs.push_back(
8547 Attribute::getWithMemoryEffects(Ctx, MemoryEffects::argMemOnly()));
8548 else if (isAssumedInaccessibleOrArgMemOnly())
8549 Attrs.push_back(Attribute::getWithMemoryEffects(
8551 }
8552 assert(Attrs.size() <= 1);
8553 }
8554
8555 /// See AbstractAttribute::manifest(...).
8556 ChangeStatus manifest(Attributor &A) override {
8557 // TODO: If AAMemoryLocation and AAMemoryBehavior are merged, we could
8558 // provide per-location modref information here.
8559 const IRPosition &IRP = getIRPosition();
8560
8561 SmallVector<Attribute, 1> DeducedAttrs;
8562 getDeducedAttributes(A, IRP.getAnchorValue().getContext(), DeducedAttrs);
8563 if (DeducedAttrs.size() != 1)
8564 return ChangeStatus::UNCHANGED;
8565 MemoryEffects ME = DeducedAttrs[0].getMemoryEffects();
8566
8567 return A.manifestAttrs(IRP, Attribute::getWithMemoryEffects(
8568 IRP.getAnchorValue().getContext(), ME));
8569 }
8570
8571 /// See AAMemoryLocation::checkForAllAccessesToMemoryKind(...).
8572 bool checkForAllAccessesToMemoryKind(
8573 function_ref<bool(const Instruction *, const Value *, AccessKind,
8574 MemoryLocationsKind)>
8575 Pred,
8576 MemoryLocationsKind RequestedMLK) const override {
8577 if (!isValidState())
8578 return false;
8579
8580 MemoryLocationsKind AssumedMLK = getAssumedNotAccessedLocation();
8581 if (AssumedMLK == NO_LOCATIONS)
8582 return true;
8583
8584 unsigned Idx = 0;
8585 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS;
8586 CurMLK *= 2, ++Idx) {
8587 if (CurMLK & RequestedMLK)
8588 continue;
8589
8590 if (const AccessSet *Accesses = AccessKind2Accesses[Idx])
8591 for (const AccessInfo &AI : *Accesses)
8592 if (!Pred(AI.I, AI.Ptr, AI.Kind, CurMLK))
8593 return false;
8594 }
8595
8596 return true;
8597 }
8598
8599 ChangeStatus indicatePessimisticFixpoint() override {
8600 // If we give up and indicate a pessimistic fixpoint this instruction will
8601 // become an access for all potential access kinds:
8602 // TODO: Add pointers for argmemonly and globals to improve the results of
8603 // checkForAllAccessesToMemoryKind.
8604 bool Changed = false;
8605 MemoryLocationsKind KnownMLK = getKnown();
8606 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
8607 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS; CurMLK *= 2)
8608 if (!(CurMLK & KnownMLK))
8609 updateStateAndAccessesMap(getState(), CurMLK, I, nullptr, Changed,
8610 getAccessKindFromInst(I));
8611 return AAMemoryLocation::indicatePessimisticFixpoint();
8612 }
8613
8614protected:
8615 /// Helper struct to tie together an instruction that has a read or write
8616 /// effect with the pointer it accesses (if any).
8617 struct AccessInfo {
8618
8619 /// The instruction that caused the access.
8620 const Instruction *I;
8621
8622 /// The base pointer that is accessed, or null if unknown.
8623 const Value *Ptr;
8624
8625 /// The kind of access (read/write/read+write).
8627
8628 bool operator==(const AccessInfo &RHS) const {
8629 return I == RHS.I && Ptr == RHS.Ptr && Kind == RHS.Kind;
8630 }
8631 bool operator()(const AccessInfo &LHS, const AccessInfo &RHS) const {
8632 if (LHS.I != RHS.I)
8633 return LHS.I < RHS.I;
8634 if (LHS.Ptr != RHS.Ptr)
8635 return LHS.Ptr < RHS.Ptr;
8636 if (LHS.Kind != RHS.Kind)
8637 return LHS.Kind < RHS.Kind;
8638 return false;
8639 }
8640 };
8641
8642 /// Mapping from *single* memory location kinds, e.g., LOCAL_MEM with the
8643 /// value of NO_LOCAL_MEM, to the accesses encountered for this memory kind.
8644 using AccessSet = SmallSet<AccessInfo, 2, AccessInfo>;
8645 std::array<AccessSet *, llvm::ConstantLog2<VALID_STATE>()>
8646 AccessKind2Accesses;
8647
8648 /// Categorize the pointer arguments of CB that might access memory in
8649 /// AccessedLoc and update the state and access map accordingly.
8650 void
8651 categorizeArgumentPointerLocations(Attributor &A, CallBase &CB,
8652 AAMemoryLocation::StateType &AccessedLocs,
8653 bool &Changed);
8654
8655 /// Return the kind(s) of location that may be accessed by \p V.
8657 categorizeAccessedLocations(Attributor &A, Instruction &I, bool &Changed);
8658
8659 /// Return the access kind as determined by \p I.
8660 AccessKind getAccessKindFromInst(const Instruction *I) {
8661 AccessKind AK = READ_WRITE;
8662 if (I) {
8663 AK = I->mayReadFromMemory() ? READ : NONE;
8664 AK = AccessKind(AK | (I->mayWriteToMemory() ? WRITE : NONE));
8665 }
8666 return AK;
8667 }
8668
8669 /// Update the state \p State and the AccessKind2Accesses given that \p I is
8670 /// an access of kind \p AK to a \p MLK memory location with the access
8671 /// pointer \p Ptr.
8672 void updateStateAndAccessesMap(AAMemoryLocation::StateType &State,
8673 MemoryLocationsKind MLK, const Instruction *I,
8674 const Value *Ptr, bool &Changed,
8675 AccessKind AK = READ_WRITE) {
8676
8677 assert(isPowerOf2_32(MLK) && "Expected a single location set!");
8678 auto *&Accesses = AccessKind2Accesses[llvm::Log2_32(MLK)];
8679 if (!Accesses)
8680 Accesses = new (Allocator) AccessSet();
8681 Changed |= Accesses->insert(AccessInfo{I, Ptr, AK}).second;
8682 if (MLK == NO_UNKOWN_MEM)
8683 MLK = NO_LOCATIONS;
8684 State.removeAssumedBits(MLK);
8685 }
8686
8687 /// Determine the underlying locations kinds for \p Ptr, e.g., globals or
8688 /// arguments, and update the state and access map accordingly.
8689 void categorizePtrValue(Attributor &A, const Instruction &I, const Value &Ptr,
8690 AAMemoryLocation::StateType &State, bool &Changed,
8691 unsigned AccessAS = 0);
8692
8693 /// Used to allocate access sets.
8695};
8696
8697void AAMemoryLocationImpl::categorizePtrValue(
8698 Attributor &A, const Instruction &I, const Value &Ptr,
8699 AAMemoryLocation::StateType &State, bool &Changed, unsigned AccessAS) {
8700 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize pointer locations for "
8701 << Ptr << " ["
8702 << getMemoryLocationsAsStr(State.getAssumed()) << "]\n");
8703
8704 auto Pred = [&](Value &Obj) {
8705 unsigned ObjectAS = Obj.getType()->getPointerAddressSpace();
8706 // TODO: recognize the TBAA used for constant accesses.
8707 MemoryLocationsKind MLK = NO_LOCATIONS;
8708
8709 // Filter accesses to constant (GPU) memory if we have an AS at the access
8710 // site or the object is known to actually have the associated AS.
8711 if ((AccessAS == (unsigned)AA::GPUAddressSpace::Constant ||
8712 (ObjectAS == (unsigned)AA::GPUAddressSpace::Constant &&
8713 isIdentifiedObject(&Obj))) &&
8714 AA::isGPU(*I.getModule()))
8715 return true;
8716
8717 if (isa<UndefValue>(&Obj))
8718 return true;
8719 if (isa<Argument>(&Obj)) {
8720 // TODO: For now we do not treat byval arguments as local copies performed
8721 // on the call edge, though, we should. To make that happen we need to
8722 // teach various passes, e.g., DSE, about the copy effect of a byval. That
8723 // would also allow us to mark functions only accessing byval arguments as
8724 // readnone again, arguably their accesses have no effect outside of the
8725 // function, like accesses to allocas.
8726 MLK = NO_ARGUMENT_MEM;
8727 } else if (auto *GV = dyn_cast<GlobalValue>(&Obj)) {
8728 // Reading constant memory is not treated as a read "effect" by the
8729 // function attr pass so we won't neither. Constants defined by TBAA are
8730 // similar. (We know we do not write it because it is constant.)
8731 if (auto *GVar = dyn_cast<GlobalVariable>(GV))
8732 if (GVar->isConstant())
8733 return true;
8734
8735 if (GV->hasLocalLinkage())
8736 MLK = NO_GLOBAL_INTERNAL_MEM;
8737 else
8738 MLK = NO_GLOBAL_EXTERNAL_MEM;
8739 } else if (isa<ConstantPointerNull>(&Obj) &&
8740 (!NullPointerIsDefined(getAssociatedFunction(), AccessAS) ||
8741 !NullPointerIsDefined(getAssociatedFunction(), ObjectAS))) {
8742 return true;
8743 } else if (isa<AllocaInst>(&Obj)) {
8744 MLK = NO_LOCAL_MEM;
8745 } else if (const auto *CB = dyn_cast<CallBase>(&Obj)) {
8746 bool IsKnownNoAlias;
8749 IsKnownNoAlias))
8750 MLK = NO_MALLOCED_MEM;
8751 else
8752 MLK = NO_UNKOWN_MEM;
8753 } else {
8754 MLK = NO_UNKOWN_MEM;
8755 }
8756
8757 assert(MLK != NO_LOCATIONS && "No location specified!");
8758 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Ptr value can be categorized: "
8759 << Obj << " -> " << getMemoryLocationsAsStr(MLK) << "\n");
8760 updateStateAndAccessesMap(State, MLK, &I, &Obj, Changed,
8761 getAccessKindFromInst(&I));
8762
8763 return true;
8764 };
8765
8766 const auto *AA = A.getAAFor<AAUnderlyingObjects>(
8768 if (!AA || !AA->forallUnderlyingObjects(Pred, AA::Intraprocedural)) {
8769 LLVM_DEBUG(
8770 dbgs() << "[AAMemoryLocation] Pointer locations not categorized\n");
8771 updateStateAndAccessesMap(State, NO_UNKOWN_MEM, &I, nullptr, Changed,
8772 getAccessKindFromInst(&I));
8773 return;
8774 }
8775
8776 LLVM_DEBUG(
8777 dbgs() << "[AAMemoryLocation] Accessed locations with pointer locations: "
8778 << getMemoryLocationsAsStr(State.getAssumed()) << "\n");
8779}
8780
8781void AAMemoryLocationImpl::categorizeArgumentPointerLocations(
8782 Attributor &A, CallBase &CB, AAMemoryLocation::StateType &AccessedLocs,
8783 bool &Changed) {
8784 for (unsigned ArgNo = 0, E = CB.arg_size(); ArgNo < E; ++ArgNo) {
8785
8786 // Skip non-pointer arguments.
8787 const Value *ArgOp = CB.getArgOperand(ArgNo);
8788 if (!ArgOp->getType()->isPtrOrPtrVectorTy())
8789 continue;
8790
8791 // Skip readnone arguments.
8792 const IRPosition &ArgOpIRP = IRPosition::callsite_argument(CB, ArgNo);
8793 const auto *ArgOpMemLocationAA =
8794 A.getAAFor<AAMemoryBehavior>(*this, ArgOpIRP, DepClassTy::OPTIONAL);
8795
8796 if (ArgOpMemLocationAA && ArgOpMemLocationAA->isAssumedReadNone())
8797 continue;
8798
8799 // Categorize potentially accessed pointer arguments as if there was an
8800 // access instruction with them as pointer.
8801 categorizePtrValue(A, CB, *ArgOp, AccessedLocs, Changed);
8802 }
8803}
8804
8806AAMemoryLocationImpl::categorizeAccessedLocations(Attributor &A, Instruction &I,
8807 bool &Changed) {
8808 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize accessed locations for "
8809 << I << "\n");
8810
8811 AAMemoryLocation::StateType AccessedLocs;
8812 AccessedLocs.intersectAssumedBits(NO_LOCATIONS);
8813
8814 if (auto *CB = dyn_cast<CallBase>(&I)) {
8815
8816 // First check if we assume any memory is access is visible.
8817 const auto *CBMemLocationAA = A.getAAFor<AAMemoryLocation>(
8819 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize call site: " << I
8820 << " [" << CBMemLocationAA << "]\n");
8821 if (!CBMemLocationAA) {
8822 updateStateAndAccessesMap(AccessedLocs, NO_UNKOWN_MEM, &I, nullptr,
8823 Changed, getAccessKindFromInst(&I));
8824 return NO_UNKOWN_MEM;
8825 }
8826
8827 if (CBMemLocationAA->isAssumedReadNone())
8828 return NO_LOCATIONS;
8829
8830 if (CBMemLocationAA->isAssumedInaccessibleMemOnly()) {
8831 updateStateAndAccessesMap(AccessedLocs, NO_INACCESSIBLE_MEM, &I, nullptr,
8832 Changed, getAccessKindFromInst(&I));
8833 return AccessedLocs.getAssumed();
8834 }
8835
8836 uint32_t CBAssumedNotAccessedLocs =
8837 CBMemLocationAA->getAssumedNotAccessedLocation();
8838
8839 // Set the argmemonly and global bit as we handle them separately below.
8840 uint32_t CBAssumedNotAccessedLocsNoArgMem =
8841 CBAssumedNotAccessedLocs | NO_ARGUMENT_MEM | NO_GLOBAL_MEM;
8842
8843 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS; CurMLK *= 2) {
8844 if (CBAssumedNotAccessedLocsNoArgMem & CurMLK)
8845 continue;
8846 updateStateAndAccessesMap(AccessedLocs, CurMLK, &I, nullptr, Changed,
8847 getAccessKindFromInst(&I));
8848 }
8849
8850 // Now handle global memory if it might be accessed. This is slightly tricky
8851 // as NO_GLOBAL_MEM has multiple bits set.
8852 bool HasGlobalAccesses = ((~CBAssumedNotAccessedLocs) & NO_GLOBAL_MEM);
8853 if (HasGlobalAccesses) {
8854 auto AccessPred = [&](const Instruction *, const Value *Ptr,
8855 AccessKind Kind, MemoryLocationsKind MLK) {
8856 updateStateAndAccessesMap(AccessedLocs, MLK, &I, Ptr, Changed,
8857 getAccessKindFromInst(&I));
8858 return true;
8859 };
8860 if (!CBMemLocationAA->checkForAllAccessesToMemoryKind(
8861 AccessPred, inverseLocation(NO_GLOBAL_MEM, false, false)))
8862 return AccessedLocs.getWorstState();
8863 }
8864
8865 LLVM_DEBUG(
8866 dbgs() << "[AAMemoryLocation] Accessed state before argument handling: "
8867 << getMemoryLocationsAsStr(AccessedLocs.getAssumed()) << "\n");
8868
8869 // Now handle argument memory if it might be accessed.
8870 bool HasArgAccesses = ((~CBAssumedNotAccessedLocs) & NO_ARGUMENT_MEM);
8871 if (HasArgAccesses)
8872 categorizeArgumentPointerLocations(A, *CB, AccessedLocs, Changed);
8873
8874 LLVM_DEBUG(
8875 dbgs() << "[AAMemoryLocation] Accessed state after argument handling: "
8876 << getMemoryLocationsAsStr(AccessedLocs.getAssumed()) << "\n");
8877
8878 return AccessedLocs.getAssumed();
8879 }
8880
8881 if (const Value *Ptr = getPointerOperand(&I, /* AllowVolatile */ true)) {
8882 LLVM_DEBUG(
8883 dbgs() << "[AAMemoryLocation] Categorize memory access with pointer: "
8884 << I << " [" << *Ptr << "]\n");
8885 categorizePtrValue(A, I, *Ptr, AccessedLocs, Changed,
8886 Ptr->getType()->getPointerAddressSpace());
8887 return AccessedLocs.getAssumed();
8888 }
8889
8890 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Failed to categorize instruction: "
8891 << I << "\n");
8892 updateStateAndAccessesMap(AccessedLocs, NO_UNKOWN_MEM, &I, nullptr, Changed,
8893 getAccessKindFromInst(&I));
8894 return AccessedLocs.getAssumed();
8895}
8896
8897/// An AA to represent the memory behavior function attributes.
8898struct AAMemoryLocationFunction final : public AAMemoryLocationImpl {
8899 AAMemoryLocationFunction(const IRPosition &IRP, Attributor &A)
8900 : AAMemoryLocationImpl(IRP, A) {}
8901
8902 /// See AbstractAttribute::updateImpl(Attributor &A).
8903 ChangeStatus updateImpl(Attributor &A) override {
8904
8905 const auto *MemBehaviorAA =
8906 A.getAAFor<AAMemoryBehavior>(*this, getIRPosition(), DepClassTy::NONE);
8907 if (MemBehaviorAA && MemBehaviorAA->isAssumedReadNone()) {
8908 if (MemBehaviorAA->isKnownReadNone())
8909 return indicateOptimisticFixpoint();
8911 "AAMemoryLocation was not read-none but AAMemoryBehavior was!");
8912 A.recordDependence(*MemBehaviorAA, *this, DepClassTy::OPTIONAL);
8913 return ChangeStatus::UNCHANGED;
8914 }
8915
8916 // The current assumed state used to determine a change.
8917 auto AssumedState = getAssumed();
8918 bool Changed = false;
8919
8920 auto CheckRWInst = [&](Instruction &I) {
8921 MemoryLocationsKind MLK = categorizeAccessedLocations(A, I, Changed);
8922 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Accessed locations for " << I
8923 << ": " << getMemoryLocationsAsStr(MLK) << "\n");
8924 removeAssumedBits(inverseLocation(MLK, false, false));
8925 // Stop once only the valid bit set in the *not assumed location*, thus
8926 // once we don't actually exclude any memory locations in the state.
8927 return getAssumedNotAccessedLocation() != VALID_STATE;
8928 };
8929
8930 bool UsedAssumedInformation = false;
8931 if (!A.checkForAllReadWriteInstructions(CheckRWInst, *this,
8932 UsedAssumedInformation))
8933 return indicatePessimisticFixpoint();
8934
8935 Changed |= AssumedState != getAssumed();
8936 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
8937 }
8938
8939 /// See AbstractAttribute::trackStatistics()
8940 void trackStatistics() const override {
8941 if (isAssumedReadNone())
8942 STATS_DECLTRACK_FN_ATTR(readnone)
8943 else if (isAssumedArgMemOnly())
8944 STATS_DECLTRACK_FN_ATTR(argmemonly)
8945 else if (isAssumedInaccessibleMemOnly())
8946 STATS_DECLTRACK_FN_ATTR(inaccessiblememonly)
8947 else if (isAssumedInaccessibleOrArgMemOnly())
8948 STATS_DECLTRACK_FN_ATTR(inaccessiblememorargmemonly)
8949 }
8950};
8951
8952/// AAMemoryLocation attribute for call sites.
8953struct AAMemoryLocationCallSite final : AAMemoryLocationImpl {
8954 AAMemoryLocationCallSite(const IRPosition &IRP, Attributor &A)
8955 : AAMemoryLocationImpl(IRP, A) {}
8956
8957 /// See AbstractAttribute::updateImpl(...).
8958 ChangeStatus updateImpl(Attributor &A) override {
8959 // TODO: Once we have call site specific value information we can provide
8960 // call site specific liveness liveness information and then it makes
8961 // sense to specialize attributes for call sites arguments instead of
8962 // redirecting requests to the callee argument.
8963 Function *F = getAssociatedFunction();
8964 const IRPosition &FnPos = IRPosition::function(*F);
8965 auto *FnAA =
8966 A.getAAFor<AAMemoryLocation>(*this, FnPos, DepClassTy::REQUIRED);
8967 if (!FnAA)
8968 return indicatePessimisticFixpoint();
8969 bool Changed = false;
8970 auto AccessPred = [&](const Instruction *I, const Value *Ptr,
8971 AccessKind Kind, MemoryLocationsKind MLK) {
8972 updateStateAndAccessesMap(getState(), MLK, I, Ptr, Changed,
8973 getAccessKindFromInst(I));
8974 return true;
8975 };
8976 if (!FnAA->checkForAllAccessesToMemoryKind(AccessPred, ALL_LOCATIONS))
8977 return indicatePessimisticFixpoint();
8978 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
8979 }
8980
8981 /// See AbstractAttribute::trackStatistics()
8982 void trackStatistics() const override {
8983 if (isAssumedReadNone())
8984 STATS_DECLTRACK_CS_ATTR(readnone)
8985 }
8986};
8987} // namespace
8988
8989/// ------------------ denormal-fp-math Attribute -------------------------
8990
8991namespace {
8992struct AADenormalFPMathImpl : public AADenormalFPMath {
8993 AADenormalFPMathImpl(const IRPosition &IRP, Attributor &A)
8994 : AADenormalFPMath(IRP, A) {}
8995
8996 const std::string getAsStr(Attributor *A) const override {
8997 std::string Str("AADenormalFPMath[");
8998 raw_string_ostream OS(Str);
8999
9000 DenormalState Known = getKnown();
9001 if (Known.Mode.isValid())
9002 OS << "denormal-fp-math=" << Known.Mode;
9003 else
9004 OS << "invalid";
9005
9006 if (Known.ModeF32.isValid())
9007 OS << " denormal-fp-math-f32=" << Known.ModeF32;
9008 OS << ']';
9009 return Str;
9010 }
9011};
9012
9013struct AADenormalFPMathFunction final : AADenormalFPMathImpl {
9014 AADenormalFPMathFunction(const IRPosition &IRP, Attributor &A)
9015 : AADenormalFPMathImpl(IRP, A) {}
9016
9017 void initialize(Attributor &A) override {
9018 const Function *F = getAnchorScope();
9019 DenormalFPEnv DenormEnv = F->getDenormalFPEnv();
9020
9021 Known = DenormalState{DenormEnv.DefaultMode, DenormEnv.F32Mode};
9022 if (isModeFixed())
9023 indicateFixpoint();
9024 }
9025
9026 ChangeStatus updateImpl(Attributor &A) override {
9027 ChangeStatus Change = ChangeStatus::UNCHANGED;
9028
9029 auto CheckCallSite = [=, &Change, &A](AbstractCallSite CS) {
9030 Function *Caller = CS.getInstruction()->getFunction();
9031 LLVM_DEBUG(dbgs() << "[AADenormalFPMath] Call " << Caller->getName()
9032 << "->" << getAssociatedFunction()->getName() << '\n');
9033
9034 const auto *CallerInfo = A.getAAFor<AADenormalFPMath>(
9035 *this, IRPosition::function(*Caller), DepClassTy::REQUIRED);
9036 if (!CallerInfo)
9037 return false;
9038
9039 Change = Change | clampStateAndIndicateChange(this->getState(),
9040 CallerInfo->getState());
9041 return true;
9042 };
9043
9044 bool AllCallSitesKnown = true;
9045 if (!A.checkForAllCallSites(CheckCallSite, *this, true, AllCallSitesKnown))
9046 return indicatePessimisticFixpoint();
9047
9048 if (Change == ChangeStatus::CHANGED && isModeFixed())
9049 indicateFixpoint();
9050 return Change;
9051 }
9052
9053 ChangeStatus manifest(Attributor &A) override {
9054 LLVMContext &Ctx = getAssociatedFunction()->getContext();
9055
9056 SmallVector<Attribute, 2> AttrToAdd;
9058
9059 // TODO: Change to use DenormalFPEnv everywhere.
9060 DenormalFPEnv KnownEnv(Known.Mode, Known.ModeF32);
9061
9062 if (KnownEnv == DenormalFPEnv::getDefault()) {
9063 AttrToRemove.push_back(Attribute::DenormalFPEnv);
9064 } else {
9065 AttrToAdd.push_back(Attribute::get(
9066 Ctx, Attribute::DenormalFPEnv,
9067 DenormalFPEnv(Known.Mode, Known.ModeF32).toIntValue()));
9068 }
9069
9070 auto &IRP = getIRPosition();
9071
9072 // TODO: There should be a combined add and remove API.
9073 return A.removeAttrs(IRP, AttrToRemove) |
9074 A.manifestAttrs(IRP, AttrToAdd, /*ForceReplace=*/true);
9075 }
9076
9077 void trackStatistics() const override {
9078 STATS_DECLTRACK_FN_ATTR(denormal_fpenv)
9079 }
9080};
9081} // namespace
9082
9083/// ------------------ Value Constant Range Attribute -------------------------
9084
9085namespace {
9086struct AAValueConstantRangeImpl : AAValueConstantRange {
9087 using StateType = IntegerRangeState;
9088 AAValueConstantRangeImpl(const IRPosition &IRP, Attributor &A)
9089 : AAValueConstantRange(IRP, A) {}
9090
9091 /// See AbstractAttribute::initialize(..).
9092 void initialize(Attributor &A) override {
9093 if (A.hasSimplificationCallback(getIRPosition())) {
9094 indicatePessimisticFixpoint();
9095 return;
9096 }
9097
9098 // Intersect a range given by SCEV.
9099 intersectKnown(getConstantRangeFromSCEV(A, getCtxI()));
9100
9101 // Intersect a range given by LVI.
9102 intersectKnown(getConstantRangeFromLVI(A, getCtxI()));
9103 }
9104
9105 /// See AbstractAttribute::getAsStr().
9106 const std::string getAsStr(Attributor *A) const override {
9107 std::string Str;
9108 llvm::raw_string_ostream OS(Str);
9109 OS << "range(" << getBitWidth() << ")<";
9110 getKnown().print(OS);
9111 OS << " / ";
9112 getAssumed().print(OS);
9113 OS << ">";
9114 return Str;
9115 }
9116
9117 /// Helper function to get a SCEV expr for the associated value at program
9118 /// point \p I.
9119 const SCEV *getSCEV(Attributor &A, const Instruction *I = nullptr) const {
9120 if (!getAnchorScope())
9121 return nullptr;
9122
9123 ScalarEvolution *SE =
9124 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(
9125 *getAnchorScope());
9126
9127 LoopInfo *LI = A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(
9128 *getAnchorScope());
9129
9130 if (!SE || !LI)
9131 return nullptr;
9132
9133 const SCEV *S = SE->getSCEV(&getAssociatedValue());
9134 if (!I)
9135 return S;
9136
9137 return SE->getSCEVAtScope(S, LI->getLoopFor(I->getParent()));
9138 }
9139
9140 /// Helper function to get a range from SCEV for the associated value at
9141 /// program point \p I.
9142 ConstantRange getConstantRangeFromSCEV(Attributor &A,
9143 const Instruction *I = nullptr) const {
9144 if (!getAnchorScope())
9145 return getWorstState(getBitWidth());
9146
9147 ScalarEvolution *SE =
9148 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(
9149 *getAnchorScope());
9150
9151 const SCEV *S = getSCEV(A, I);
9152 if (!SE || !S)
9153 return getWorstState(getBitWidth());
9154
9155 return SE->getUnsignedRange(S);
9156 }
9157
9158 /// Helper function to get a range from LVI for the associated value at
9159 /// program point \p I.
9160 ConstantRange
9161 getConstantRangeFromLVI(Attributor &A,
9162 const Instruction *CtxI = nullptr) const {
9163 if (!getAnchorScope())
9164 return getWorstState(getBitWidth());
9165
9166 LazyValueInfo *LVI =
9167 A.getInfoCache().getAnalysisResultForFunction<LazyValueAnalysis>(
9168 *getAnchorScope());
9169
9170 if (!LVI || !CtxI)
9171 return getWorstState(getBitWidth());
9172 return LVI->getConstantRange(&getAssociatedValue(),
9173 const_cast<Instruction *>(CtxI),
9174 /*UndefAllowed*/ false);
9175 }
9176
9177 /// Return true if \p CtxI is valid for querying outside analyses.
9178 /// This basically makes sure we do not ask intra-procedural analysis
9179 /// about a context in the wrong function or a context that violates
9180 /// dominance assumptions they might have. The \p AllowAACtxI flag indicates
9181 /// if the original context of this AA is OK or should be considered invalid.
9182 bool isValidCtxInstructionForOutsideAnalysis(Attributor &A,
9183 const Instruction *CtxI,
9184 bool AllowAACtxI) const {
9185 if (!CtxI || (!AllowAACtxI && CtxI == getCtxI()))
9186 return false;
9187
9188 // Our context might be in a different function, neither intra-procedural
9189 // analysis (ScalarEvolution nor LazyValueInfo) can handle that.
9190 if (!AA::isValidInScope(getAssociatedValue(), CtxI->getFunction()))
9191 return false;
9192
9193 // If the context is not dominated by the value there are paths to the
9194 // context that do not define the value. This cannot be handled by
9195 // LazyValueInfo so we need to bail.
9196 if (auto *I = dyn_cast<Instruction>(&getAssociatedValue())) {
9197 InformationCache &InfoCache = A.getInfoCache();
9198 const DominatorTree *DT =
9199 InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(
9200 *I->getFunction());
9201 return DT && DT->dominates(I, CtxI);
9202 }
9203
9204 return true;
9205 }
9206
9207 /// See AAValueConstantRange::getKnownConstantRange(..).
9208 ConstantRange
9209 getKnownConstantRange(Attributor &A,
9210 const Instruction *CtxI = nullptr) const override {
9211 if (!isValidCtxInstructionForOutsideAnalysis(A, CtxI,
9212 /* AllowAACtxI */ false))
9213 return getKnown();
9214
9215 ConstantRange LVIR = getConstantRangeFromLVI(A, CtxI);
9216 ConstantRange SCEVR = getConstantRangeFromSCEV(A, CtxI);
9217 return getKnown().intersectWith(SCEVR).intersectWith(LVIR);
9218 }
9219
9220 /// See AAValueConstantRange::getAssumedConstantRange(..).
9221 ConstantRange
9222 getAssumedConstantRange(Attributor &A,
9223 const Instruction *CtxI = nullptr) const override {
9224 // TODO: Make SCEV use Attributor assumption.
9225 // We may be able to bound a variable range via assumptions in
9226 // Attributor. ex.) If x is assumed to be in [1, 3] and y is known to
9227 // evolve to x^2 + x, then we can say that y is in [2, 12].
9228 if (!isValidCtxInstructionForOutsideAnalysis(A, CtxI,
9229 /* AllowAACtxI */ false))
9230 return getAssumed();
9231
9232 ConstantRange LVIR = getConstantRangeFromLVI(A, CtxI);
9233 ConstantRange SCEVR = getConstantRangeFromSCEV(A, CtxI);
9234 return getAssumed().intersectWith(SCEVR).intersectWith(LVIR);
9235 }
9236
9237 /// Helper function to create MDNode for range metadata.
9238 static MDNode *
9239 getMDNodeForConstantRange(Type *Ty, LLVMContext &Ctx,
9240 const ConstantRange &AssumedConstantRange) {
9241 Metadata *LowAndHigh[] = {ConstantAsMetadata::get(ConstantInt::get(
9242 Ty, AssumedConstantRange.getLower())),
9243 ConstantAsMetadata::get(ConstantInt::get(
9244 Ty, AssumedConstantRange.getUpper()))};
9245 return MDNode::get(Ctx, LowAndHigh);
9246 }
9247
9248 /// Return true if \p Assumed is included in ranges from instruction \p I.
9249 static bool isBetterRange(const ConstantRange &Assumed,
9250 const Instruction &I) {
9251 if (Assumed.isFullSet())
9252 return false;
9253
9254 std::optional<ConstantRange> Known;
9255
9256 if (const auto *CB = dyn_cast<CallBase>(&I)) {
9257 Known = CB->getRange();
9258 } else if (MDNode *KnownRanges = I.getMetadata(LLVMContext::MD_range)) {
9259 // If multiple ranges are annotated in IR, we give up to annotate assumed
9260 // range for now.
9261
9262 // TODO: If there exists a known range which containts assumed range, we
9263 // can say assumed range is better.
9264 if (KnownRanges->getNumOperands() > 2)
9265 return false;
9266
9267 ConstantInt *Lower =
9268 mdconst::extract<ConstantInt>(KnownRanges->getOperand(0));
9269 ConstantInt *Upper =
9270 mdconst::extract<ConstantInt>(KnownRanges->getOperand(1));
9271
9272 Known.emplace(Lower->getValue(), Upper->getValue());
9273 }
9274 return !Known || (*Known != Assumed && Known->contains(Assumed));
9275 }
9276
9277 /// Helper function to set range metadata.
9278 static bool
9279 setRangeMetadataIfisBetterRange(Instruction *I,
9280 const ConstantRange &AssumedConstantRange) {
9281 if (isBetterRange(AssumedConstantRange, *I)) {
9282 I->setMetadata(LLVMContext::MD_range,
9283 getMDNodeForConstantRange(I->getType(), I->getContext(),
9284 AssumedConstantRange));
9285 return true;
9286 }
9287 return false;
9288 }
9289 /// Helper function to set range return attribute.
9290 static bool
9291 setRangeRetAttrIfisBetterRange(Attributor &A, const IRPosition &IRP,
9292 Instruction *I,
9293 const ConstantRange &AssumedConstantRange) {
9294 if (isBetterRange(AssumedConstantRange, *I)) {
9295 A.manifestAttrs(IRP,
9296 Attribute::get(I->getContext(), Attribute::Range,
9297 AssumedConstantRange),
9298 /*ForceReplace*/ true);
9299 return true;
9300 }
9301 return false;
9302 }
9303
9304 /// See AbstractAttribute::manifest()
9305 ChangeStatus manifest(Attributor &A) override {
9306 ChangeStatus Changed = ChangeStatus::UNCHANGED;
9307 ConstantRange AssumedConstantRange = getAssumedConstantRange(A);
9308 assert(!AssumedConstantRange.isFullSet() && "Invalid state");
9309
9310 auto &V = getAssociatedValue();
9311 if (!AssumedConstantRange.isEmptySet() &&
9312 !AssumedConstantRange.isSingleElement()) {
9313 if (Instruction *I = dyn_cast<Instruction>(&V)) {
9314 assert(I == getCtxI() && "Should not annotate an instruction which is "
9315 "not the context instruction");
9316 if (isa<LoadInst>(I))
9317 if (setRangeMetadataIfisBetterRange(I, AssumedConstantRange))
9318 Changed = ChangeStatus::CHANGED;
9319 if (isa<CallInst>(I))
9320 if (setRangeRetAttrIfisBetterRange(A, getIRPosition(), I,
9321 AssumedConstantRange))
9322 Changed = ChangeStatus::CHANGED;
9323 }
9324 }
9325
9326 return Changed;
9327 }
9328};
9329
9330struct AAValueConstantRangeArgument final
9331 : AAArgumentFromCallSiteArguments<
9332 AAValueConstantRange, AAValueConstantRangeImpl, IntegerRangeState,
9333 true /* BridgeCallBaseContext */> {
9334 using Base = AAArgumentFromCallSiteArguments<
9335 AAValueConstantRange, AAValueConstantRangeImpl, IntegerRangeState,
9336 true /* BridgeCallBaseContext */>;
9337 AAValueConstantRangeArgument(const IRPosition &IRP, Attributor &A)
9338 : Base(IRP, A) {}
9339
9340 /// See AbstractAttribute::trackStatistics()
9341 void trackStatistics() const override {
9342 STATS_DECLTRACK_ARG_ATTR(value_range)
9343 }
9344};
9345
9346struct AAValueConstantRangeReturned
9347 : AAReturnedFromReturnedValues<AAValueConstantRange,
9348 AAValueConstantRangeImpl,
9349 AAValueConstantRangeImpl::StateType,
9350 /* PropagateCallBaseContext */ true> {
9351 using Base =
9352 AAReturnedFromReturnedValues<AAValueConstantRange,
9353 AAValueConstantRangeImpl,
9354 AAValueConstantRangeImpl::StateType,
9355 /* PropagateCallBaseContext */ true>;
9356 AAValueConstantRangeReturned(const IRPosition &IRP, Attributor &A)
9357 : Base(IRP, A) {}
9358
9359 /// See AbstractAttribute::initialize(...).
9360 void initialize(Attributor &A) override {
9361 if (!A.isFunctionIPOAmendable(*getAssociatedFunction()))
9362 indicatePessimisticFixpoint();
9363 }
9364
9365 /// See AbstractAttribute::trackStatistics()
9366 void trackStatistics() const override {
9367 STATS_DECLTRACK_FNRET_ATTR(value_range)
9368 }
9369};
9370
9371struct AAValueConstantRangeFloating : AAValueConstantRangeImpl {
9372 AAValueConstantRangeFloating(const IRPosition &IRP, Attributor &A)
9373 : AAValueConstantRangeImpl(IRP, A) {}
9374
9375 /// See AbstractAttribute::initialize(...).
9376 void initialize(Attributor &A) override {
9377 AAValueConstantRangeImpl::initialize(A);
9378 if (isAtFixpoint())
9379 return;
9380
9381 Value &V = getAssociatedValue();
9382
9383 if (auto *C = dyn_cast<ConstantInt>(&V)) {
9384 unionAssumed(ConstantRange(C->getValue()));
9385 indicateOptimisticFixpoint();
9386 return;
9387 }
9388
9389 if (isa<UndefValue>(&V)) {
9390 // Collapse the undef state to 0.
9391 unionAssumed(ConstantRange(APInt(getBitWidth(), 0)));
9392 indicateOptimisticFixpoint();
9393 return;
9394 }
9395
9396 if (isa<CallBase>(&V))
9397 return;
9398
9399 if (isa<BinaryOperator>(&V) || isa<CmpInst>(&V) || isa<CastInst>(&V))
9400 return;
9401
9402 // If it is a load instruction with range metadata, use it.
9403 if (LoadInst *LI = dyn_cast<LoadInst>(&V))
9404 if (auto *RangeMD = LI->getMetadata(LLVMContext::MD_range)) {
9405 intersectKnown(getConstantRangeFromMetadata(*RangeMD));
9406 return;
9407 }
9408
9409 // We can work with PHI and select instruction as we traverse their operands
9410 // during update.
9411 if (isa<SelectInst>(V) || isa<PHINode>(V))
9412 return;
9413
9414 // Otherwise we give up.
9415 indicatePessimisticFixpoint();
9416
9417 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] We give up: "
9418 << getAssociatedValue() << "\n");
9419 }
9420
9421 bool calculateBinaryOperator(
9422 Attributor &A, BinaryOperator *BinOp, IntegerRangeState &T,
9423 const Instruction *CtxI,
9424 SmallVectorImpl<const AAValueConstantRange *> &QuerriedAAs) {
9425 Value *LHS = BinOp->getOperand(0);
9426 Value *RHS = BinOp->getOperand(1);
9427
9428 // Simplify the operands first.
9429 bool UsedAssumedInformation = false;
9430 const auto &SimplifiedLHS = A.getAssumedSimplified(
9431 IRPosition::value(*LHS, getCallBaseContext()), *this,
9432 UsedAssumedInformation, AA::Interprocedural);
9433 if (!SimplifiedLHS.has_value())
9434 return true;
9435 if (!*SimplifiedLHS)
9436 return false;
9437 LHS = *SimplifiedLHS;
9438
9439 const auto &SimplifiedRHS = A.getAssumedSimplified(
9440 IRPosition::value(*RHS, getCallBaseContext()), *this,
9441 UsedAssumedInformation, AA::Interprocedural);
9442 if (!SimplifiedRHS.has_value())
9443 return true;
9444 if (!*SimplifiedRHS)
9445 return false;
9446 RHS = *SimplifiedRHS;
9447
9448 // TODO: Allow non integers as well.
9449 if (!LHS->getType()->isIntegerTy() || !RHS->getType()->isIntegerTy())
9450 return false;
9451
9452 auto *LHSAA = A.getAAFor<AAValueConstantRange>(
9453 *this, IRPosition::value(*LHS, getCallBaseContext()),
9454 DepClassTy::REQUIRED);
9455 if (!LHSAA)
9456 return false;
9457 QuerriedAAs.push_back(LHSAA);
9458 auto LHSAARange = LHSAA->getAssumedConstantRange(A, CtxI);
9459
9460 auto *RHSAA = A.getAAFor<AAValueConstantRange>(
9461 *this, IRPosition::value(*RHS, getCallBaseContext()),
9462 DepClassTy::REQUIRED);
9463 if (!RHSAA)
9464 return false;
9465 QuerriedAAs.push_back(RHSAA);
9466 auto RHSAARange = RHSAA->getAssumedConstantRange(A, CtxI);
9467
9468 auto AssumedRange = LHSAARange.binaryOp(BinOp->getOpcode(), RHSAARange);
9469
9470 T.unionAssumed(AssumedRange);
9471
9472 // TODO: Track a known state too.
9473
9474 return T.isValidState();
9475 }
9476
9477 bool calculateCastInst(
9478 Attributor &A, CastInst *CastI, IntegerRangeState &T,
9479 const Instruction *CtxI,
9480 SmallVectorImpl<const AAValueConstantRange *> &QuerriedAAs) {
9481 assert(CastI->getNumOperands() == 1 && "Expected cast to be unary!");
9482 // TODO: Allow non integers as well.
9483 Value *OpV = CastI->getOperand(0);
9484
9485 // Simplify the operand first.
9486 bool UsedAssumedInformation = false;
9487 const auto &SimplifiedOpV = A.getAssumedSimplified(
9488 IRPosition::value(*OpV, getCallBaseContext()), *this,
9489 UsedAssumedInformation, AA::Interprocedural);
9490 if (!SimplifiedOpV.has_value())
9491 return true;
9492 if (!*SimplifiedOpV)
9493 return false;
9494 OpV = *SimplifiedOpV;
9495
9496 if (!OpV->getType()->isIntegerTy())
9497 return false;
9498
9499 auto *OpAA = A.getAAFor<AAValueConstantRange>(
9500 *this, IRPosition::value(*OpV, getCallBaseContext()),
9501 DepClassTy::REQUIRED);
9502 if (!OpAA)
9503 return false;
9504 QuerriedAAs.push_back(OpAA);
9505 T.unionAssumed(OpAA->getAssumed().castOp(CastI->getOpcode(),
9506 getState().getBitWidth()));
9507 return T.isValidState();
9508 }
9509
9510 bool
9511 calculateCmpInst(Attributor &A, CmpInst *CmpI, IntegerRangeState &T,
9512 const Instruction *CtxI,
9513 SmallVectorImpl<const AAValueConstantRange *> &QuerriedAAs) {
9514 Value *LHS = CmpI->getOperand(0);
9515 Value *RHS = CmpI->getOperand(1);
9516
9517 // Simplify the operands first.
9518 bool UsedAssumedInformation = false;
9519 const auto &SimplifiedLHS = A.getAssumedSimplified(
9520 IRPosition::value(*LHS, getCallBaseContext()), *this,
9521 UsedAssumedInformation, AA::Interprocedural);
9522 if (!SimplifiedLHS.has_value())
9523 return true;
9524 if (!*SimplifiedLHS)
9525 return false;
9526 LHS = *SimplifiedLHS;
9527
9528 const auto &SimplifiedRHS = A.getAssumedSimplified(
9529 IRPosition::value(*RHS, getCallBaseContext()), *this,
9530 UsedAssumedInformation, AA::Interprocedural);
9531 if (!SimplifiedRHS.has_value())
9532 return true;
9533 if (!*SimplifiedRHS)
9534 return false;
9535 RHS = *SimplifiedRHS;
9536
9537 // TODO: Allow non integers as well.
9538 if (!LHS->getType()->isIntegerTy() || !RHS->getType()->isIntegerTy())
9539 return false;
9540
9541 auto *LHSAA = A.getAAFor<AAValueConstantRange>(
9542 *this, IRPosition::value(*LHS, getCallBaseContext()),
9543 DepClassTy::REQUIRED);
9544 if (!LHSAA)
9545 return false;
9546 QuerriedAAs.push_back(LHSAA);
9547 auto *RHSAA = A.getAAFor<AAValueConstantRange>(
9548 *this, IRPosition::value(*RHS, getCallBaseContext()),
9549 DepClassTy::REQUIRED);
9550 if (!RHSAA)
9551 return false;
9552 QuerriedAAs.push_back(RHSAA);
9553 auto LHSAARange = LHSAA->getAssumedConstantRange(A, CtxI);
9554 auto RHSAARange = RHSAA->getAssumedConstantRange(A, CtxI);
9555
9556 // If one of them is empty set, we can't decide.
9557 if (LHSAARange.isEmptySet() || RHSAARange.isEmptySet())
9558 return true;
9559
9560 bool MustTrue = false, MustFalse = false;
9561
9562 auto AllowedRegion =
9564
9565 if (AllowedRegion.intersectWith(LHSAARange).isEmptySet())
9566 MustFalse = true;
9567
9568 if (LHSAARange.icmp(CmpI->getPredicate(), RHSAARange))
9569 MustTrue = true;
9570
9571 assert((!MustTrue || !MustFalse) &&
9572 "Either MustTrue or MustFalse should be false!");
9573
9574 if (MustTrue)
9575 T.unionAssumed(ConstantRange(APInt(/* numBits */ 1, /* val */ 1)));
9576 else if (MustFalse)
9577 T.unionAssumed(ConstantRange(APInt(/* numBits */ 1, /* val */ 0)));
9578 else
9579 T.unionAssumed(ConstantRange(/* BitWidth */ 1, /* isFullSet */ true));
9580
9581 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] " << *CmpI << " after "
9582 << (MustTrue ? "true" : (MustFalse ? "false" : "unknown"))
9583 << ": " << T << "\n\t" << *LHSAA << "\t<op>\n\t"
9584 << *RHSAA);
9585
9586 // TODO: Track a known state too.
9587 return T.isValidState();
9588 }
9589
9590 /// See AbstractAttribute::updateImpl(...).
9591 ChangeStatus updateImpl(Attributor &A) override {
9592
9593 IntegerRangeState T(getBitWidth());
9594 auto VisitValueCB = [&](Value &V, const Instruction *CtxI) -> bool {
9596 if (!I || isa<CallBase>(I)) {
9597
9598 // Simplify the operand first.
9599 bool UsedAssumedInformation = false;
9600 const auto &SimplifiedOpV = A.getAssumedSimplified(
9601 IRPosition::value(V, getCallBaseContext()), *this,
9602 UsedAssumedInformation, AA::Interprocedural);
9603 if (!SimplifiedOpV.has_value())
9604 return true;
9605 if (!*SimplifiedOpV)
9606 return false;
9607 Value *VPtr = *SimplifiedOpV;
9608
9609 // If the value is not instruction, we query AA to Attributor.
9610 const auto *AA = A.getAAFor<AAValueConstantRange>(
9611 *this, IRPosition::value(*VPtr, getCallBaseContext()),
9612 DepClassTy::REQUIRED);
9613
9614 // Clamp operator is not used to utilize a program point CtxI.
9615 if (AA)
9616 T.unionAssumed(AA->getAssumedConstantRange(A, CtxI));
9617 else
9618 return false;
9619
9620 return T.isValidState();
9621 }
9622
9624 if (auto *BinOp = dyn_cast<BinaryOperator>(I)) {
9625 if (!calculateBinaryOperator(A, BinOp, T, CtxI, QuerriedAAs))
9626 return false;
9627 } else if (auto *CmpI = dyn_cast<CmpInst>(I)) {
9628 if (!calculateCmpInst(A, CmpI, T, CtxI, QuerriedAAs))
9629 return false;
9630 } else if (auto *CastI = dyn_cast<CastInst>(I)) {
9631 if (!calculateCastInst(A, CastI, T, CtxI, QuerriedAAs))
9632 return false;
9633 } else {
9634 // Give up with other instructions.
9635 // TODO: Add other instructions
9636
9637 T.indicatePessimisticFixpoint();
9638 return false;
9639 }
9640
9641 // Catch circular reasoning in a pessimistic way for now.
9642 // TODO: Check how the range evolves and if we stripped anything, see also
9643 // AADereferenceable or AAAlign for similar situations.
9644 for (const AAValueConstantRange *QueriedAA : QuerriedAAs) {
9645 if (QueriedAA != this)
9646 continue;
9647 // If we are in a stady state we do not need to worry.
9648 if (T.getAssumed() == getState().getAssumed())
9649 continue;
9650 T.indicatePessimisticFixpoint();
9651 }
9652
9653 return T.isValidState();
9654 };
9655
9656 if (!VisitValueCB(getAssociatedValue(), getCtxI()))
9657 return indicatePessimisticFixpoint();
9658
9659 // Ensure that long def-use chains can't cause circular reasoning either by
9660 // introducing a cutoff below.
9661 if (clampStateAndIndicateChange(getState(), T) == ChangeStatus::UNCHANGED)
9662 return ChangeStatus::UNCHANGED;
9663 if (++NumChanges > MaxNumChanges) {
9664 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] performed " << NumChanges
9665 << " but only " << MaxNumChanges
9666 << " are allowed to avoid cyclic reasoning.");
9667 return indicatePessimisticFixpoint();
9668 }
9669 return ChangeStatus::CHANGED;
9670 }
9671
9672 /// See AbstractAttribute::trackStatistics()
9673 void trackStatistics() const override {
9675 }
9676
9677 /// Tracker to bail after too many widening steps of the constant range.
9678 int NumChanges = 0;
9679
9680 /// Upper bound for the number of allowed changes (=widening steps) for the
9681 /// constant range before we give up.
9682 static constexpr int MaxNumChanges = 5;
9683};
9684
9685struct AAValueConstantRangeFunction : AAValueConstantRangeImpl {
9686 AAValueConstantRangeFunction(const IRPosition &IRP, Attributor &A)
9687 : AAValueConstantRangeImpl(IRP, A) {}
9688
9689 /// See AbstractAttribute::initialize(...).
9690 ChangeStatus updateImpl(Attributor &A) override {
9691 llvm_unreachable("AAValueConstantRange(Function|CallSite)::updateImpl will "
9692 "not be called");
9693 }
9694
9695 /// See AbstractAttribute::trackStatistics()
9696 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(value_range) }
9697};
9698
9699struct AAValueConstantRangeCallSite : AAValueConstantRangeFunction {
9700 AAValueConstantRangeCallSite(const IRPosition &IRP, Attributor &A)
9701 : AAValueConstantRangeFunction(IRP, A) {}
9702
9703 /// See AbstractAttribute::trackStatistics()
9704 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(value_range) }
9705};
9706
9707struct AAValueConstantRangeCallSiteReturned
9708 : AACalleeToCallSite<AAValueConstantRange, AAValueConstantRangeImpl,
9709 AAValueConstantRangeImpl::StateType,
9710 /* IntroduceCallBaseContext */ true> {
9711 AAValueConstantRangeCallSiteReturned(const IRPosition &IRP, Attributor &A)
9712 : AACalleeToCallSite<AAValueConstantRange, AAValueConstantRangeImpl,
9713 AAValueConstantRangeImpl::StateType,
9714 /* IntroduceCallBaseContext */ true>(IRP, A) {}
9715
9716 /// See AbstractAttribute::initialize(...).
9717 void initialize(Attributor &A) override {
9718 // If it is a call instruction with range attribute, use the range.
9719 if (CallInst *CI = dyn_cast<CallInst>(&getAssociatedValue())) {
9720 if (std::optional<ConstantRange> Range = CI->getRange())
9721 intersectKnown(*Range);
9722 }
9723
9724 AAValueConstantRangeImpl::initialize(A);
9725 }
9726
9727 /// See AbstractAttribute::trackStatistics()
9728 void trackStatistics() const override {
9729 STATS_DECLTRACK_CSRET_ATTR(value_range)
9730 }
9731};
9732struct AAValueConstantRangeCallSiteArgument : AAValueConstantRangeFloating {
9733 AAValueConstantRangeCallSiteArgument(const IRPosition &IRP, Attributor &A)
9734 : AAValueConstantRangeFloating(IRP, A) {}
9735
9736 /// See AbstractAttribute::manifest()
9737 ChangeStatus manifest(Attributor &A) override {
9738 return ChangeStatus::UNCHANGED;
9739 }
9740
9741 /// See AbstractAttribute::trackStatistics()
9742 void trackStatistics() const override {
9743 STATS_DECLTRACK_CSARG_ATTR(value_range)
9744 }
9745};
9746} // namespace
9747
9748/// ------------------ Potential Values Attribute -------------------------
9749
9750namespace {
9751struct AAPotentialConstantValuesImpl : AAPotentialConstantValues {
9752 using StateType = PotentialConstantIntValuesState;
9753
9754 AAPotentialConstantValuesImpl(const IRPosition &IRP, Attributor &A)
9755 : AAPotentialConstantValues(IRP, A) {}
9756
9757 /// See AbstractAttribute::initialize(..).
9758 void initialize(Attributor &A) override {
9759 if (A.hasSimplificationCallback(getIRPosition()))
9760 indicatePessimisticFixpoint();
9761 else
9762 AAPotentialConstantValues::initialize(A);
9763 }
9764
9765 bool fillSetWithConstantValues(Attributor &A, const IRPosition &IRP, SetTy &S,
9766 bool &ContainsUndef, bool ForSelf) {
9768 bool UsedAssumedInformation = false;
9769 if (!A.getAssumedSimplifiedValues(IRP, *this, Values, AA::Interprocedural,
9770 UsedAssumedInformation)) {
9771 // Avoid recursion when the caller is computing constant values for this
9772 // IRP itself.
9773 if (ForSelf)
9774 return false;
9775 if (!IRP.getAssociatedType()->isIntegerTy())
9776 return false;
9777 auto *PotentialValuesAA = A.getAAFor<AAPotentialConstantValues>(
9778 *this, IRP, DepClassTy::REQUIRED);
9779 if (!PotentialValuesAA || !PotentialValuesAA->getState().isValidState())
9780 return false;
9781 ContainsUndef = PotentialValuesAA->getState().undefIsContained();
9782 S = PotentialValuesAA->getState().getAssumedSet();
9783 return true;
9784 }
9785
9786 // Copy all the constant values, except UndefValue. ContainsUndef is true
9787 // iff Values contains only UndefValue instances. If there are other known
9788 // constants, then UndefValue is dropped.
9789 ContainsUndef = false;
9790 for (auto &It : Values) {
9791 if (isa<UndefValue>(It.getValue())) {
9792 ContainsUndef = true;
9793 continue;
9794 }
9795 auto *CI = dyn_cast<ConstantInt>(It.getValue());
9796 if (!CI)
9797 return false;
9798 S.insert(CI->getValue());
9799 }
9800 ContainsUndef &= S.empty();
9801
9802 return true;
9803 }
9804
9805 /// See AbstractAttribute::getAsStr().
9806 const std::string getAsStr(Attributor *A) const override {
9807 std::string Str;
9808 llvm::raw_string_ostream OS(Str);
9809 OS << getState();
9810 return Str;
9811 }
9812
9813 /// See AbstractAttribute::updateImpl(...).
9814 ChangeStatus updateImpl(Attributor &A) override {
9815 return indicatePessimisticFixpoint();
9816 }
9817};
9818
9819struct AAPotentialConstantValuesArgument final
9820 : AAArgumentFromCallSiteArguments<AAPotentialConstantValues,
9821 AAPotentialConstantValuesImpl,
9822 PotentialConstantIntValuesState> {
9823 using Base = AAArgumentFromCallSiteArguments<AAPotentialConstantValues,
9824 AAPotentialConstantValuesImpl,
9826 AAPotentialConstantValuesArgument(const IRPosition &IRP, Attributor &A)
9827 : Base(IRP, A) {}
9828
9829 /// See AbstractAttribute::trackStatistics()
9830 void trackStatistics() const override {
9831 STATS_DECLTRACK_ARG_ATTR(potential_values)
9832 }
9833};
9834
9835struct AAPotentialConstantValuesReturned
9836 : AAReturnedFromReturnedValues<AAPotentialConstantValues,
9837 AAPotentialConstantValuesImpl> {
9838 using Base = AAReturnedFromReturnedValues<AAPotentialConstantValues,
9839 AAPotentialConstantValuesImpl>;
9840 AAPotentialConstantValuesReturned(const IRPosition &IRP, Attributor &A)
9841 : Base(IRP, A) {}
9842
9843 void initialize(Attributor &A) override {
9844 if (!A.isFunctionIPOAmendable(*getAssociatedFunction()))
9845 indicatePessimisticFixpoint();
9846 Base::initialize(A);
9847 }
9848
9849 /// See AbstractAttribute::trackStatistics()
9850 void trackStatistics() const override {
9851 STATS_DECLTRACK_FNRET_ATTR(potential_values)
9852 }
9853};
9854
9855struct AAPotentialConstantValuesFloating : AAPotentialConstantValuesImpl {
9856 AAPotentialConstantValuesFloating(const IRPosition &IRP, Attributor &A)
9857 : AAPotentialConstantValuesImpl(IRP, A) {}
9858
9859 /// See AbstractAttribute::initialize(..).
9860 void initialize(Attributor &A) override {
9861 AAPotentialConstantValuesImpl::initialize(A);
9862 if (isAtFixpoint())
9863 return;
9864
9865 Value &V = getAssociatedValue();
9866
9867 if (auto *C = dyn_cast<ConstantInt>(&V)) {
9868 unionAssumed(C->getValue());
9869 indicateOptimisticFixpoint();
9870 return;
9871 }
9872
9873 if (isa<UndefValue>(&V)) {
9874 unionAssumedWithUndef();
9875 indicateOptimisticFixpoint();
9876 return;
9877 }
9878
9879 if (isa<BinaryOperator>(&V) || isa<ICmpInst>(&V) || isa<CastInst>(&V))
9880 return;
9881
9882 if (isa<SelectInst>(V) || isa<PHINode>(V) || isa<LoadInst>(V))
9883 return;
9884
9885 indicatePessimisticFixpoint();
9886
9887 LLVM_DEBUG(dbgs() << "[AAPotentialConstantValues] We give up: "
9888 << getAssociatedValue() << "\n");
9889 }
9890
9891 static bool calculateICmpInst(const ICmpInst *ICI, const APInt &LHS,
9892 const APInt &RHS) {
9893 return ICmpInst::compare(LHS, RHS, ICI->getPredicate());
9894 }
9895
9896 static APInt calculateCastInst(const CastInst *CI, const APInt &Src,
9897 uint32_t ResultBitWidth) {
9898 Instruction::CastOps CastOp = CI->getOpcode();
9899 switch (CastOp) {
9900 default:
9901 llvm_unreachable("unsupported or not integer cast");
9902 case Instruction::Trunc:
9903 return Src.trunc(ResultBitWidth);
9904 case Instruction::SExt:
9905 return Src.sext(ResultBitWidth);
9906 case Instruction::ZExt:
9907 return Src.zext(ResultBitWidth);
9908 case Instruction::BitCast:
9909 return Src;
9910 }
9911 }
9912
9913 static APInt calculateBinaryOperator(const BinaryOperator *BinOp,
9914 const APInt &LHS, const APInt &RHS,
9915 bool &SkipOperation, bool &Unsupported) {
9916 Instruction::BinaryOps BinOpcode = BinOp->getOpcode();
9917 // Unsupported is set to true when the binary operator is not supported.
9918 // SkipOperation is set to true when UB occur with the given operand pair
9919 // (LHS, RHS).
9920 // TODO: we should look at nsw and nuw keywords to handle operations
9921 // that create poison or undef value.
9922 switch (BinOpcode) {
9923 default:
9924 Unsupported = true;
9925 return LHS;
9926 case Instruction::Add:
9927 return LHS + RHS;
9928 case Instruction::Sub:
9929 return LHS - RHS;
9930 case Instruction::Mul:
9931 return LHS * RHS;
9932 case Instruction::UDiv:
9933 if (RHS.isZero()) {
9934 SkipOperation = true;
9935 return LHS;
9936 }
9937 return LHS.udiv(RHS);
9938 case Instruction::SDiv:
9939 if (RHS.isZero()) {
9940 SkipOperation = true;
9941 return LHS;
9942 }
9943 return LHS.sdiv(RHS);
9944 case Instruction::URem:
9945 if (RHS.isZero()) {
9946 SkipOperation = true;
9947 return LHS;
9948 }
9949 return LHS.urem(RHS);
9950 case Instruction::SRem:
9951 if (RHS.isZero()) {
9952 SkipOperation = true;
9953 return LHS;
9954 }
9955 return LHS.srem(RHS);
9956 case Instruction::Shl:
9957 return LHS.shl(RHS);
9958 case Instruction::LShr:
9959 return LHS.lshr(RHS);
9960 case Instruction::AShr:
9961 return LHS.ashr(RHS);
9962 case Instruction::And:
9963 return LHS & RHS;
9964 case Instruction::Or:
9965 return LHS | RHS;
9966 case Instruction::Xor:
9967 return LHS ^ RHS;
9968 }
9969 }
9970
9971 bool calculateBinaryOperatorAndTakeUnion(const BinaryOperator *BinOp,
9972 const APInt &LHS, const APInt &RHS) {
9973 bool SkipOperation = false;
9974 bool Unsupported = false;
9975 APInt Result =
9976 calculateBinaryOperator(BinOp, LHS, RHS, SkipOperation, Unsupported);
9977 if (Unsupported)
9978 return false;
9979 // If SkipOperation is true, we can ignore this operand pair (L, R).
9980 if (!SkipOperation)
9981 unionAssumed(Result);
9982 return isValidState();
9983 }
9984
9985 ChangeStatus updateWithICmpInst(Attributor &A, ICmpInst *ICI) {
9986 auto AssumedBefore = getAssumed();
9987 Value *LHS = ICI->getOperand(0);
9988 Value *RHS = ICI->getOperand(1);
9989
9990 bool LHSContainsUndef = false, RHSContainsUndef = false;
9991 SetTy LHSAAPVS, RHSAAPVS;
9992 if (!fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
9993 LHSContainsUndef, /* ForSelf */ false) ||
9994 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
9995 RHSContainsUndef, /* ForSelf */ false))
9996 return indicatePessimisticFixpoint();
9997
9998 // TODO: make use of undef flag to limit potential values aggressively.
9999 bool MaybeTrue = false, MaybeFalse = false;
10000 const APInt Zero(RHS->getType()->getIntegerBitWidth(), 0);
10001 if (LHSContainsUndef && RHSContainsUndef) {
10002 // The result of any comparison between undefs can be soundly replaced
10003 // with undef.
10004 unionAssumedWithUndef();
10005 } else if (LHSContainsUndef) {
10006 for (const APInt &R : RHSAAPVS) {
10007 bool CmpResult = calculateICmpInst(ICI, Zero, R);
10008 MaybeTrue |= CmpResult;
10009 MaybeFalse |= !CmpResult;
10010 if (MaybeTrue & MaybeFalse)
10011 return indicatePessimisticFixpoint();
10012 }
10013 } else if (RHSContainsUndef) {
10014 for (const APInt &L : LHSAAPVS) {
10015 bool CmpResult = calculateICmpInst(ICI, L, Zero);
10016 MaybeTrue |= CmpResult;
10017 MaybeFalse |= !CmpResult;
10018 if (MaybeTrue & MaybeFalse)
10019 return indicatePessimisticFixpoint();
10020 }
10021 } else {
10022 for (const APInt &L : LHSAAPVS) {
10023 for (const APInt &R : RHSAAPVS) {
10024 bool CmpResult = calculateICmpInst(ICI, L, R);
10025 MaybeTrue |= CmpResult;
10026 MaybeFalse |= !CmpResult;
10027 if (MaybeTrue & MaybeFalse)
10028 return indicatePessimisticFixpoint();
10029 }
10030 }
10031 }
10032 if (MaybeTrue)
10033 unionAssumed(APInt(/* numBits */ 1, /* val */ 1));
10034 if (MaybeFalse)
10035 unionAssumed(APInt(/* numBits */ 1, /* val */ 0));
10036 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10037 : ChangeStatus::CHANGED;
10038 }
10039
10040 ChangeStatus updateWithSelectInst(Attributor &A, SelectInst *SI) {
10041 auto AssumedBefore = getAssumed();
10042 Value *LHS = SI->getTrueValue();
10043 Value *RHS = SI->getFalseValue();
10044
10045 bool UsedAssumedInformation = false;
10046 std::optional<Constant *> C = A.getAssumedConstant(
10047 *SI->getCondition(), *this, UsedAssumedInformation);
10048
10049 // Check if we only need one operand.
10050 bool OnlyLeft = false, OnlyRight = false;
10051 if (C && *C && (*C)->isOneValue())
10052 OnlyLeft = true;
10053 else if (C && *C && (*C)->isNullValue())
10054 OnlyRight = true;
10055
10056 bool LHSContainsUndef = false, RHSContainsUndef = false;
10057 SetTy LHSAAPVS, RHSAAPVS;
10058 if (!OnlyRight &&
10059 !fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
10060 LHSContainsUndef, /* ForSelf */ false))
10061 return indicatePessimisticFixpoint();
10062
10063 if (!OnlyLeft &&
10064 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
10065 RHSContainsUndef, /* ForSelf */ false))
10066 return indicatePessimisticFixpoint();
10067
10068 if (OnlyLeft || OnlyRight) {
10069 // select (true/false), lhs, rhs
10070 auto *OpAA = OnlyLeft ? &LHSAAPVS : &RHSAAPVS;
10071 auto Undef = OnlyLeft ? LHSContainsUndef : RHSContainsUndef;
10072
10073 if (Undef)
10074 unionAssumedWithUndef();
10075 else {
10076 for (const auto &It : *OpAA)
10077 unionAssumed(It);
10078 }
10079
10080 } else if (LHSContainsUndef && RHSContainsUndef) {
10081 // select i1 *, undef , undef => undef
10082 unionAssumedWithUndef();
10083 } else {
10084 for (const auto &It : LHSAAPVS)
10085 unionAssumed(It);
10086 for (const auto &It : RHSAAPVS)
10087 unionAssumed(It);
10088 }
10089 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10090 : ChangeStatus::CHANGED;
10091 }
10092
10093 ChangeStatus updateWithCastInst(Attributor &A, CastInst *CI) {
10094 auto AssumedBefore = getAssumed();
10095 if (!CI->isIntegerCast())
10096 return indicatePessimisticFixpoint();
10097 assert(CI->getNumOperands() == 1 && "Expected cast to be unary!");
10098 uint32_t ResultBitWidth = CI->getDestTy()->getIntegerBitWidth();
10099 Value *Src = CI->getOperand(0);
10100
10101 bool SrcContainsUndef = false;
10102 SetTy SrcPVS;
10103 if (!fillSetWithConstantValues(A, IRPosition::value(*Src), SrcPVS,
10104 SrcContainsUndef, /* ForSelf */ false))
10105 return indicatePessimisticFixpoint();
10106
10107 if (SrcContainsUndef)
10108 unionAssumedWithUndef();
10109 else {
10110 for (const APInt &S : SrcPVS) {
10111 APInt T = calculateCastInst(CI, S, ResultBitWidth);
10112 unionAssumed(T);
10113 }
10114 }
10115 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10116 : ChangeStatus::CHANGED;
10117 }
10118
10119 ChangeStatus updateWithBinaryOperator(Attributor &A, BinaryOperator *BinOp) {
10120 auto AssumedBefore = getAssumed();
10121 Value *LHS = BinOp->getOperand(0);
10122 Value *RHS = BinOp->getOperand(1);
10123
10124 bool LHSContainsUndef = false, RHSContainsUndef = false;
10125 SetTy LHSAAPVS, RHSAAPVS;
10126 if (!fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
10127 LHSContainsUndef, /* ForSelf */ false) ||
10128 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
10129 RHSContainsUndef, /* ForSelf */ false))
10130 return indicatePessimisticFixpoint();
10131
10132 const APInt Zero = APInt(LHS->getType()->getIntegerBitWidth(), 0);
10133
10134 // TODO: make use of undef flag to limit potential values aggressively.
10135 if (LHSContainsUndef && RHSContainsUndef) {
10136 if (!calculateBinaryOperatorAndTakeUnion(BinOp, Zero, Zero))
10137 return indicatePessimisticFixpoint();
10138 } else if (LHSContainsUndef) {
10139 for (const APInt &R : RHSAAPVS) {
10140 if (!calculateBinaryOperatorAndTakeUnion(BinOp, Zero, R))
10141 return indicatePessimisticFixpoint();
10142 }
10143 } else if (RHSContainsUndef) {
10144 for (const APInt &L : LHSAAPVS) {
10145 if (!calculateBinaryOperatorAndTakeUnion(BinOp, L, Zero))
10146 return indicatePessimisticFixpoint();
10147 }
10148 } else {
10149 for (const APInt &L : LHSAAPVS) {
10150 for (const APInt &R : RHSAAPVS) {
10151 if (!calculateBinaryOperatorAndTakeUnion(BinOp, L, R))
10152 return indicatePessimisticFixpoint();
10153 }
10154 }
10155 }
10156 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10157 : ChangeStatus::CHANGED;
10158 }
10159
10160 ChangeStatus updateWithInstruction(Attributor &A, Instruction *Inst) {
10161 auto AssumedBefore = getAssumed();
10162 SetTy Incoming;
10163 bool ContainsUndef;
10164 if (!fillSetWithConstantValues(A, IRPosition::value(*Inst), Incoming,
10165 ContainsUndef, /* ForSelf */ true))
10166 return indicatePessimisticFixpoint();
10167 if (ContainsUndef) {
10168 unionAssumedWithUndef();
10169 } else {
10170 for (const auto &It : Incoming)
10171 unionAssumed(It);
10172 }
10173 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10174 : ChangeStatus::CHANGED;
10175 }
10176
10177 /// See AbstractAttribute::updateImpl(...).
10178 ChangeStatus updateImpl(Attributor &A) override {
10179 Value &V = getAssociatedValue();
10181
10182 if (auto *ICI = dyn_cast<ICmpInst>(I))
10183 return updateWithICmpInst(A, ICI);
10184
10185 if (auto *SI = dyn_cast<SelectInst>(I))
10186 return updateWithSelectInst(A, SI);
10187
10188 if (auto *CI = dyn_cast<CastInst>(I))
10189 return updateWithCastInst(A, CI);
10190
10191 if (auto *BinOp = dyn_cast<BinaryOperator>(I))
10192 return updateWithBinaryOperator(A, BinOp);
10193
10194 if (isa<PHINode>(I) || isa<LoadInst>(I))
10195 return updateWithInstruction(A, I);
10196
10197 return indicatePessimisticFixpoint();
10198 }
10199
10200 /// See AbstractAttribute::trackStatistics()
10201 void trackStatistics() const override {
10202 STATS_DECLTRACK_FLOATING_ATTR(potential_values)
10203 }
10204};
10205
10206struct AAPotentialConstantValuesFunction : AAPotentialConstantValuesImpl {
10207 AAPotentialConstantValuesFunction(const IRPosition &IRP, Attributor &A)
10208 : AAPotentialConstantValuesImpl(IRP, A) {}
10209
10210 /// See AbstractAttribute::initialize(...).
10211 ChangeStatus updateImpl(Attributor &A) override {
10213 "AAPotentialConstantValues(Function|CallSite)::updateImpl will "
10214 "not be called");
10215 }
10216
10217 /// See AbstractAttribute::trackStatistics()
10218 void trackStatistics() const override {
10219 STATS_DECLTRACK_FN_ATTR(potential_values)
10220 }
10221};
10222
10223struct AAPotentialConstantValuesCallSite : AAPotentialConstantValuesFunction {
10224 AAPotentialConstantValuesCallSite(const IRPosition &IRP, Attributor &A)
10225 : AAPotentialConstantValuesFunction(IRP, A) {}
10226
10227 /// See AbstractAttribute::trackStatistics()
10228 void trackStatistics() const override {
10229 STATS_DECLTRACK_CS_ATTR(potential_values)
10230 }
10231};
10232
10233struct AAPotentialConstantValuesCallSiteReturned
10234 : AACalleeToCallSite<AAPotentialConstantValues,
10235 AAPotentialConstantValuesImpl> {
10236 AAPotentialConstantValuesCallSiteReturned(const IRPosition &IRP,
10237 Attributor &A)
10238 : AACalleeToCallSite<AAPotentialConstantValues,
10239 AAPotentialConstantValuesImpl>(IRP, A) {}
10240
10241 /// See AbstractAttribute::trackStatistics()
10242 void trackStatistics() const override {
10243 STATS_DECLTRACK_CSRET_ATTR(potential_values)
10244 }
10245};
10246
10247struct AAPotentialConstantValuesCallSiteArgument
10248 : AAPotentialConstantValuesFloating {
10249 AAPotentialConstantValuesCallSiteArgument(const IRPosition &IRP,
10250 Attributor &A)
10251 : AAPotentialConstantValuesFloating(IRP, A) {}
10252
10253 /// See AbstractAttribute::initialize(..).
10254 void initialize(Attributor &A) override {
10255 AAPotentialConstantValuesImpl::initialize(A);
10256 if (isAtFixpoint())
10257 return;
10258
10259 Value &V = getAssociatedValue();
10260
10261 if (auto *C = dyn_cast<ConstantInt>(&V)) {
10262 unionAssumed(C->getValue());
10263 indicateOptimisticFixpoint();
10264 return;
10265 }
10266
10267 if (isa<UndefValue>(&V)) {
10268 unionAssumedWithUndef();
10269 indicateOptimisticFixpoint();
10270 return;
10271 }
10272 }
10273
10274 /// See AbstractAttribute::updateImpl(...).
10275 ChangeStatus updateImpl(Attributor &A) override {
10276 Value &V = getAssociatedValue();
10277 auto AssumedBefore = getAssumed();
10278 auto *AA = A.getAAFor<AAPotentialConstantValues>(
10279 *this, IRPosition::value(V), DepClassTy::REQUIRED);
10280 if (!AA)
10281 return indicatePessimisticFixpoint();
10282 const auto &S = AA->getAssumed();
10283 unionAssumed(S);
10284 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10285 : ChangeStatus::CHANGED;
10286 }
10287
10288 /// See AbstractAttribute::trackStatistics()
10289 void trackStatistics() const override {
10290 STATS_DECLTRACK_CSARG_ATTR(potential_values)
10291 }
10292};
10293} // namespace
10294
10295/// ------------------------ NoUndef Attribute ---------------------------------
10297 Attribute::AttrKind ImpliedAttributeKind,
10298 bool IgnoreSubsumingPositions) {
10299 assert(ImpliedAttributeKind == Attribute::NoUndef &&
10300 "Unexpected attribute kind");
10301 if (A.hasAttr(IRP, {Attribute::NoUndef}, IgnoreSubsumingPositions,
10302 Attribute::NoUndef))
10303 return true;
10304
10305 Value &Val = IRP.getAssociatedValue();
10308 LLVMContext &Ctx = Val.getContext();
10309 A.manifestAttrs(IRP, Attribute::get(Ctx, Attribute::NoUndef));
10310 return true;
10311 }
10312
10313 return false;
10314}
10315
10316namespace {
10317struct AANoUndefImpl : AANoUndef {
10318 AANoUndefImpl(const IRPosition &IRP, Attributor &A) : AANoUndef(IRP, A) {}
10319
10320 /// See AbstractAttribute::initialize(...).
10321 void initialize(Attributor &A) override {
10322 Value &V = getAssociatedValue();
10323 if (isa<UndefValue>(V))
10324 indicatePessimisticFixpoint();
10325 assert(!isImpliedByIR(A, getIRPosition(), Attribute::NoUndef));
10326 }
10327
10328 /// See followUsesInMBEC
10329 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
10330 AANoUndef::StateType &State) {
10331 const Value *UseV = U->get();
10332 const DominatorTree *DT = nullptr;
10333 AssumptionCache *AC = nullptr;
10334 InformationCache &InfoCache = A.getInfoCache();
10335 if (Function *F = getAnchorScope()) {
10336 DT = InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(*F);
10337 AC = InfoCache.getAnalysisResultForFunction<AssumptionAnalysis>(*F);
10338 }
10339 State.setKnown(isGuaranteedNotToBeUndefOrPoison(UseV, AC, I, DT));
10340 bool TrackUse = false;
10341 // Track use for instructions which must produce undef or poison bits when
10342 // at least one operand contains such bits.
10344 TrackUse = true;
10345 return TrackUse;
10346 }
10347
10348 /// See AbstractAttribute::getAsStr().
10349 const std::string getAsStr(Attributor *A) const override {
10350 return getAssumed() ? "noundef" : "may-undef-or-poison";
10351 }
10352
10353 ChangeStatus manifest(Attributor &A) override {
10354 // We don't manifest noundef attribute for dead positions because the
10355 // associated values with dead positions would be replaced with undef
10356 // values.
10357 bool UsedAssumedInformation = false;
10358 if (A.isAssumedDead(getIRPosition(), nullptr, nullptr,
10359 UsedAssumedInformation))
10360 return ChangeStatus::UNCHANGED;
10361 // A position whose simplified value does not have any value is
10362 // considered to be dead. We don't manifest noundef in such positions for
10363 // the same reason above.
10364 if (!A.getAssumedSimplified(getIRPosition(), *this, UsedAssumedInformation,
10366 .has_value())
10367 return ChangeStatus::UNCHANGED;
10368 return AANoUndef::manifest(A);
10369 }
10370};
10371
10372struct AANoUndefFloating : public AANoUndefImpl {
10373 AANoUndefFloating(const IRPosition &IRP, Attributor &A)
10374 : AANoUndefImpl(IRP, A) {}
10375
10376 /// See AbstractAttribute::initialize(...).
10377 void initialize(Attributor &A) override {
10378 AANoUndefImpl::initialize(A);
10379 if (!getState().isAtFixpoint() && getAnchorScope() &&
10380 !getAnchorScope()->isDeclaration())
10381 if (Instruction *CtxI = getCtxI())
10382 followUsesInMBEC(*this, A, getState(), *CtxI);
10383 }
10384
10385 /// See AbstractAttribute::updateImpl(...).
10386 ChangeStatus updateImpl(Attributor &A) override {
10387 auto VisitValueCB = [&](const IRPosition &IRP) -> bool {
10388 bool IsKnownNoUndef;
10390 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoUndef);
10391 };
10392
10393 bool Stripped;
10394 bool UsedAssumedInformation = false;
10395 Value *AssociatedValue = &getAssociatedValue();
10397 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
10398 AA::AnyScope, UsedAssumedInformation))
10399 Stripped = false;
10400 else
10401 Stripped =
10402 Values.size() != 1 || Values.front().getValue() != AssociatedValue;
10403
10404 if (!Stripped) {
10405 // If we haven't stripped anything we might still be able to use a
10406 // different AA, but only if the IRP changes. Effectively when we
10407 // interpret this not as a call site value but as a floating/argument
10408 // value.
10409 const IRPosition AVIRP = IRPosition::value(*AssociatedValue);
10410 if (AVIRP == getIRPosition() || !VisitValueCB(AVIRP))
10411 return indicatePessimisticFixpoint();
10412 return ChangeStatus::UNCHANGED;
10413 }
10414
10415 for (const auto &VAC : Values)
10416 if (!VisitValueCB(IRPosition::value(*VAC.getValue())))
10417 return indicatePessimisticFixpoint();
10418
10419 return ChangeStatus::UNCHANGED;
10420 }
10421
10422 /// See AbstractAttribute::trackStatistics()
10423 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noundef) }
10424};
10425
10426struct AANoUndefReturned final
10427 : AAReturnedFromReturnedValues<AANoUndef, AANoUndefImpl> {
10428 AANoUndefReturned(const IRPosition &IRP, Attributor &A)
10429 : AAReturnedFromReturnedValues<AANoUndef, AANoUndefImpl>(IRP, A) {}
10430
10431 /// See AbstractAttribute::trackStatistics()
10432 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noundef) }
10433};
10434
10435struct AANoUndefArgument final
10436 : AAArgumentFromCallSiteArguments<AANoUndef, AANoUndefImpl> {
10437 AANoUndefArgument(const IRPosition &IRP, Attributor &A)
10438 : AAArgumentFromCallSiteArguments<AANoUndef, AANoUndefImpl>(IRP, A) {}
10439
10440 /// See AbstractAttribute::trackStatistics()
10441 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(noundef) }
10442};
10443
10444struct AANoUndefCallSiteArgument final : AANoUndefFloating {
10445 AANoUndefCallSiteArgument(const IRPosition &IRP, Attributor &A)
10446 : AANoUndefFloating(IRP, A) {}
10447
10448 /// See AbstractAttribute::trackStatistics()
10449 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(noundef) }
10450};
10451
10452struct AANoUndefCallSiteReturned final
10453 : AACalleeToCallSite<AANoUndef, AANoUndefImpl> {
10454 AANoUndefCallSiteReturned(const IRPosition &IRP, Attributor &A)
10455 : AACalleeToCallSite<AANoUndef, AANoUndefImpl>(IRP, A) {}
10456
10457 /// See AbstractAttribute::trackStatistics()
10458 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(noundef) }
10459};
10460
10461/// ------------------------ NoFPClass Attribute -------------------------------
10462
10463struct AANoFPClassImpl : AANoFPClass {
10464 AANoFPClassImpl(const IRPosition &IRP, Attributor &A) : AANoFPClass(IRP, A) {}
10465
10466 void initialize(Attributor &A) override {
10467 const IRPosition &IRP = getIRPosition();
10468
10469 Value &V = IRP.getAssociatedValue();
10470 if (isa<UndefValue>(V)) {
10471 indicateOptimisticFixpoint();
10472 return;
10473 }
10474
10476 A.getAttrs(getIRPosition(), {Attribute::NoFPClass}, Attrs, false);
10477 for (const auto &Attr : Attrs) {
10478 addKnownBits(Attr.getNoFPClass());
10479 }
10480
10481 Instruction *CtxI = getCtxI();
10482
10483 if (getPositionKind() != IRPosition::IRP_RETURNED) {
10484 const DataLayout &DL = A.getDataLayout();
10485 InformationCache &InfoCache = A.getInfoCache();
10486
10487 const DominatorTree *DT = nullptr;
10488 AssumptionCache *AC = nullptr;
10489 const TargetLibraryInfo *TLI = nullptr;
10490 Function *F = getAnchorScope();
10491 if (F) {
10492 TLI = InfoCache.getTargetLibraryInfoForFunction(*F);
10493 if (!F->isDeclaration()) {
10494 DT =
10495 InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(*F);
10496 AC = InfoCache.getAnalysisResultForFunction<AssumptionAnalysis>(*F);
10497 }
10498 }
10499
10500 SimplifyQuery Q(DL, TLI, DT, AC, CtxI);
10501
10502 KnownFPClass KnownFPClass = computeKnownFPClass(&V, fcAllFlags, Q);
10503 addKnownBits(~KnownFPClass.KnownFPClasses);
10504 }
10505
10506 if (CtxI)
10507 followUsesInMBEC(*this, A, getState(), *CtxI);
10508 }
10509
10510 /// See followUsesInMBEC
10511 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
10512 AANoFPClass::StateType &State) {
10513 // TODO: Determine what instructions can be looked through.
10514 auto *CB = dyn_cast<CallBase>(I);
10515 if (!CB)
10516 return false;
10517
10518 if (!CB->isArgOperand(U))
10519 return false;
10520
10521 unsigned ArgNo = CB->getArgOperandNo(U);
10522 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
10523 if (auto *NoFPAA = A.getAAFor<AANoFPClass>(*this, IRP, DepClassTy::NONE))
10524 State.addKnownBits(NoFPAA->getState().getKnown());
10525 return false;
10526 }
10527
10528 const std::string getAsStr(Attributor *A) const override {
10529 std::string Result = "nofpclass";
10530 raw_string_ostream OS(Result);
10531 OS << getKnownNoFPClass() << '/' << getAssumedNoFPClass();
10532 return Result;
10533 }
10534
10535 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
10536 SmallVectorImpl<Attribute> &Attrs) const override {
10537 Attrs.emplace_back(Attribute::getWithNoFPClass(Ctx, getAssumedNoFPClass()));
10538 }
10539};
10540
10541struct AANoFPClassFloating : public AANoFPClassImpl {
10542 AANoFPClassFloating(const IRPosition &IRP, Attributor &A)
10543 : AANoFPClassImpl(IRP, A) {}
10544
10545 /// See AbstractAttribute::updateImpl(...).
10546 ChangeStatus updateImpl(Attributor &A) override {
10548 bool UsedAssumedInformation = false;
10549 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
10550 AA::AnyScope, UsedAssumedInformation)) {
10551 Values.push_back({getAssociatedValue(), getCtxI()});
10552 }
10553
10554 StateType T;
10555 auto VisitValueCB = [&](Value &V, const Instruction *CtxI) -> bool {
10556 const auto *AA = A.getAAFor<AANoFPClass>(*this, IRPosition::value(V),
10557 DepClassTy::REQUIRED);
10558 if (!AA || this == AA) {
10559 T.indicatePessimisticFixpoint();
10560 } else {
10561 const AANoFPClass::StateType &S =
10562 static_cast<const AANoFPClass::StateType &>(AA->getState());
10563 T ^= S;
10564 }
10565 return T.isValidState();
10566 };
10567
10568 for (const auto &VAC : Values)
10569 if (!VisitValueCB(*VAC.getValue(), VAC.getCtxI()))
10570 return indicatePessimisticFixpoint();
10571
10572 return clampStateAndIndicateChange(getState(), T);
10573 }
10574
10575 /// See AbstractAttribute::trackStatistics()
10576 void trackStatistics() const override {
10578 }
10579};
10580
10581struct AANoFPClassReturned final
10582 : AAReturnedFromReturnedValues<AANoFPClass, AANoFPClassImpl,
10583 AANoFPClassImpl::StateType, false,
10584 Attribute::None, false> {
10585 AANoFPClassReturned(const IRPosition &IRP, Attributor &A)
10586 : AAReturnedFromReturnedValues<AANoFPClass, AANoFPClassImpl,
10587 AANoFPClassImpl::StateType, false,
10588 Attribute::None, false>(IRP, A) {}
10589
10590 /// See AbstractAttribute::trackStatistics()
10591 void trackStatistics() const override {
10593 }
10594};
10595
10596struct AANoFPClassArgument final
10597 : AAArgumentFromCallSiteArguments<AANoFPClass, AANoFPClassImpl> {
10598 AANoFPClassArgument(const IRPosition &IRP, Attributor &A)
10599 : AAArgumentFromCallSiteArguments<AANoFPClass, AANoFPClassImpl>(IRP, A) {}
10600
10601 /// See AbstractAttribute::trackStatistics()
10602 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nofpclass) }
10603};
10604
10605struct AANoFPClassCallSiteArgument final : AANoFPClassFloating {
10606 AANoFPClassCallSiteArgument(const IRPosition &IRP, Attributor &A)
10607 : AANoFPClassFloating(IRP, A) {}
10608
10609 /// See AbstractAttribute::trackStatistics()
10610 void trackStatistics() const override {
10612 }
10613};
10614
10615struct AANoFPClassCallSiteReturned final
10616 : AACalleeToCallSite<AANoFPClass, AANoFPClassImpl> {
10617 AANoFPClassCallSiteReturned(const IRPosition &IRP, Attributor &A)
10618 : AACalleeToCallSite<AANoFPClass, AANoFPClassImpl>(IRP, A) {}
10619
10620 /// See AbstractAttribute::trackStatistics()
10621 void trackStatistics() const override {
10623 }
10624};
10625
10626struct AACallEdgesImpl : public AACallEdges {
10627 AACallEdgesImpl(const IRPosition &IRP, Attributor &A) : AACallEdges(IRP, A) {}
10628
10629 const SetVector<Function *> &getOptimisticEdges() const override {
10630 return CalledFunctions;
10631 }
10632
10633 bool hasUnknownCallee() const override { return HasUnknownCallee; }
10634
10635 bool hasNonAsmUnknownCallee() const override {
10636 return HasUnknownCalleeNonAsm;
10637 }
10638
10639 const std::string getAsStr(Attributor *A) const override {
10640 return "CallEdges[" + std::to_string(HasUnknownCallee) + "," +
10641 std::to_string(CalledFunctions.size()) + "]";
10642 }
10643
10644 void trackStatistics() const override {}
10645
10646protected:
10647 void addCalledFunction(Function *Fn, ChangeStatus &Change) {
10648 if (CalledFunctions.insert(Fn)) {
10649 Change = ChangeStatus::CHANGED;
10650 LLVM_DEBUG(dbgs() << "[AACallEdges] New call edge: " << Fn->getName()
10651 << "\n");
10652 }
10653 }
10654
10655 void setHasUnknownCallee(bool NonAsm, ChangeStatus &Change) {
10656 if (!HasUnknownCallee)
10657 Change = ChangeStatus::CHANGED;
10658 if (NonAsm && !HasUnknownCalleeNonAsm)
10659 Change = ChangeStatus::CHANGED;
10660 HasUnknownCalleeNonAsm |= NonAsm;
10661 HasUnknownCallee = true;
10662 }
10663
10664private:
10665 /// Optimistic set of functions that might be called by this position.
10666 SetVector<Function *> CalledFunctions;
10667
10668 /// Is there any call with a unknown callee.
10669 bool HasUnknownCallee = false;
10670
10671 /// Is there any call with a unknown callee, excluding any inline asm.
10672 bool HasUnknownCalleeNonAsm = false;
10673};
10674
10675struct AACallEdgesCallSite : public AACallEdgesImpl {
10676 AACallEdgesCallSite(const IRPosition &IRP, Attributor &A)
10677 : AACallEdgesImpl(IRP, A) {}
10678 /// See AbstractAttribute::updateImpl(...).
10679 ChangeStatus updateImpl(Attributor &A) override {
10680 ChangeStatus Change = ChangeStatus::UNCHANGED;
10681
10682 auto VisitValue = [&](Value &V, const Instruction *CtxI) -> bool {
10683 if (Function *Fn = dyn_cast<Function>(&V)) {
10684 addCalledFunction(Fn, Change);
10685 } else {
10686 LLVM_DEBUG(dbgs() << "[AACallEdges] Unrecognized value: " << V << "\n");
10687 setHasUnknownCallee(true, Change);
10688 }
10689
10690 // Explore all values.
10691 return true;
10692 };
10693
10695 // Process any value that we might call.
10696 auto ProcessCalledOperand = [&](Value *V, Instruction *CtxI) {
10697 if (isa<Constant>(V)) {
10698 VisitValue(*V, CtxI);
10699 return;
10700 }
10701
10702 bool UsedAssumedInformation = false;
10703 Values.clear();
10704 if (!A.getAssumedSimplifiedValues(IRPosition::value(*V), *this, Values,
10705 AA::AnyScope, UsedAssumedInformation)) {
10706 Values.push_back({*V, CtxI});
10707 }
10708 for (auto &VAC : Values)
10709 VisitValue(*VAC.getValue(), VAC.getCtxI());
10710 };
10711
10712 CallBase *CB = cast<CallBase>(getCtxI());
10713
10714 if (auto *IA = dyn_cast<InlineAsm>(CB->getCalledOperand())) {
10715 if (IA->hasSideEffects() &&
10716 !hasAssumption(*CB->getCaller(), "ompx_no_call_asm") &&
10717 !hasAssumption(*CB, "ompx_no_call_asm")) {
10718 setHasUnknownCallee(false, Change);
10719 }
10720 return Change;
10721 }
10722
10723 if (CB->isIndirectCall())
10724 if (auto *IndirectCallAA = A.getAAFor<AAIndirectCallInfo>(
10725 *this, getIRPosition(), DepClassTy::OPTIONAL))
10726 if (IndirectCallAA->foreachCallee(
10727 [&](Function *Fn) { return VisitValue(*Fn, CB); }))
10728 return Change;
10729
10730 // The most simple case.
10731 ProcessCalledOperand(CB->getCalledOperand(), CB);
10732
10733 // Process callback functions.
10734 SmallVector<const Use *, 4u> CallbackUses;
10735 AbstractCallSite::getCallbackUses(*CB, CallbackUses);
10736 for (const Use *U : CallbackUses)
10737 ProcessCalledOperand(U->get(), CB);
10738
10739 return Change;
10740 }
10741};
10742
10743struct AACallEdgesFunction : public AACallEdgesImpl {
10744 AACallEdgesFunction(const IRPosition &IRP, Attributor &A)
10745 : AACallEdgesImpl(IRP, A) {}
10746
10747 /// See AbstractAttribute::updateImpl(...).
10748 ChangeStatus updateImpl(Attributor &A) override {
10749 ChangeStatus Change = ChangeStatus::UNCHANGED;
10750
10751 auto ProcessCallInst = [&](Instruction &Inst) {
10752 CallBase &CB = cast<CallBase>(Inst);
10753
10754 auto *CBEdges = A.getAAFor<AACallEdges>(
10755 *this, IRPosition::callsite_function(CB), DepClassTy::REQUIRED);
10756 if (!CBEdges)
10757 return false;
10758 if (CBEdges->hasNonAsmUnknownCallee())
10759 setHasUnknownCallee(true, Change);
10760 if (CBEdges->hasUnknownCallee())
10761 setHasUnknownCallee(false, Change);
10762
10763 for (Function *F : CBEdges->getOptimisticEdges())
10764 addCalledFunction(F, Change);
10765
10766 return true;
10767 };
10768
10769 // Visit all callable instructions.
10770 bool UsedAssumedInformation = false;
10771 if (!A.checkForAllCallLikeInstructions(ProcessCallInst, *this,
10772 UsedAssumedInformation,
10773 /* CheckBBLivenessOnly */ true)) {
10774 // If we haven't looked at all call like instructions, assume that there
10775 // are unknown callees.
10776 setHasUnknownCallee(true, Change);
10777 }
10778
10779 return Change;
10780 }
10781};
10782
10783/// -------------------AAInterFnReachability Attribute--------------------------
10784
10785struct AAInterFnReachabilityFunction
10786 : public CachedReachabilityAA<AAInterFnReachability, Function> {
10787 using Base = CachedReachabilityAA<AAInterFnReachability, Function>;
10788 AAInterFnReachabilityFunction(const IRPosition &IRP, Attributor &A)
10789 : Base(IRP, A) {}
10790
10791 bool instructionCanReach(
10792 Attributor &A, const Instruction &From, const Function &To,
10793 const AA::InstExclusionSetTy *ExclusionSet) const override {
10794 assert(From.getFunction() == getAnchorScope() && "Queried the wrong AA!");
10795 auto *NonConstThis = const_cast<AAInterFnReachabilityFunction *>(this);
10796
10797 RQITy StackRQI(A, From, To, ExclusionSet, false);
10798 RQITy::Reachable Result;
10799 if (!NonConstThis->checkQueryCache(A, StackRQI, Result))
10800 return NonConstThis->isReachableImpl(A, StackRQI,
10801 /*IsTemporaryRQI=*/true);
10802 return Result == RQITy::Reachable::Yes;
10803 }
10804
10805 bool isReachableImpl(Attributor &A, RQITy &RQI,
10806 bool IsTemporaryRQI) override {
10807 const Instruction *EntryI =
10808 &RQI.From->getFunction()->getEntryBlock().front();
10809 if (EntryI != RQI.From &&
10810 !instructionCanReach(A, *EntryI, *RQI.To, nullptr))
10811 return rememberResult(A, RQITy::Reachable::No, RQI, false,
10812 IsTemporaryRQI);
10813
10814 auto CheckReachableCallBase = [&](CallBase *CB) {
10815 auto *CBEdges = A.getAAFor<AACallEdges>(
10816 *this, IRPosition::callsite_function(*CB), DepClassTy::OPTIONAL);
10817 if (!CBEdges || !CBEdges->getState().isValidState())
10818 return false;
10819 // TODO Check To backwards in this case.
10820 if (CBEdges->hasUnknownCallee())
10821 return false;
10822
10823 for (Function *Fn : CBEdges->getOptimisticEdges()) {
10824 if (Fn == RQI.To)
10825 return false;
10826
10827 if (Fn->isDeclaration()) {
10828 if (Fn->hasFnAttribute(Attribute::NoCallback))
10829 continue;
10830 // TODO Check To backwards in this case.
10831 return false;
10832 }
10833
10834 if (Fn == getAnchorScope()) {
10835 if (EntryI == RQI.From)
10836 continue;
10837 return false;
10838 }
10839
10840 const AAInterFnReachability *InterFnReachability =
10841 A.getAAFor<AAInterFnReachability>(*this, IRPosition::function(*Fn),
10842 DepClassTy::OPTIONAL);
10843
10844 const Instruction &FnFirstInst = Fn->getEntryBlock().front();
10845 if (!InterFnReachability ||
10846 InterFnReachability->instructionCanReach(A, FnFirstInst, *RQI.To,
10847 RQI.ExclusionSet))
10848 return false;
10849 }
10850 return true;
10851 };
10852
10853 const auto *IntraFnReachability = A.getAAFor<AAIntraFnReachability>(
10854 *this, IRPosition::function(*RQI.From->getFunction()),
10855 DepClassTy::OPTIONAL);
10856
10857 // Determine call like instructions that we can reach from the inst.
10858 auto CheckCallBase = [&](Instruction &CBInst) {
10859 // There are usually less nodes in the call graph, check inter function
10860 // reachability first.
10861 if (CheckReachableCallBase(cast<CallBase>(&CBInst)))
10862 return true;
10863 return IntraFnReachability && !IntraFnReachability->isAssumedReachable(
10864 A, *RQI.From, CBInst, RQI.ExclusionSet);
10865 };
10866
10867 bool UsedExclusionSet = /* conservative */ true;
10868 bool UsedAssumedInformation = false;
10869 if (!A.checkForAllCallLikeInstructions(CheckCallBase, *this,
10870 UsedAssumedInformation,
10871 /* CheckBBLivenessOnly */ true))
10872 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
10873 IsTemporaryRQI);
10874
10875 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
10876 IsTemporaryRQI);
10877 }
10878
10879 void trackStatistics() const override {}
10880};
10881} // namespace
10882
10883template <typename AAType>
10884static std::optional<Constant *>
10886 const IRPosition &IRP, Type &Ty) {
10887 if (!Ty.isIntegerTy())
10888 return nullptr;
10889
10890 // This will also pass the call base context.
10891 const auto *AA = A.getAAFor<AAType>(QueryingAA, IRP, DepClassTy::NONE);
10892 if (!AA)
10893 return nullptr;
10894
10895 std::optional<Constant *> COpt = AA->getAssumedConstant(A);
10896
10897 if (!COpt.has_value()) {
10898 A.recordDependence(*AA, QueryingAA, DepClassTy::OPTIONAL);
10899 return std::nullopt;
10900 }
10901 if (auto *C = *COpt) {
10902 A.recordDependence(*AA, QueryingAA, DepClassTy::OPTIONAL);
10903 return C;
10904 }
10905 return nullptr;
10906}
10907
10909 Attributor &A, const AbstractAttribute &AA, const IRPosition &IRP,
10911 Type &Ty = *IRP.getAssociatedType();
10912 std::optional<Value *> V;
10913 for (auto &It : Values) {
10914 V = AA::combineOptionalValuesInAAValueLatice(V, It.getValue(), &Ty);
10915 if (V.has_value() && !*V)
10916 break;
10917 }
10918 if (!V.has_value())
10919 return UndefValue::get(&Ty);
10920 return *V;
10921}
10922
10923namespace {
10924struct AAPotentialValuesImpl : AAPotentialValues {
10925 using StateType = PotentialLLVMValuesState;
10926
10927 AAPotentialValuesImpl(const IRPosition &IRP, Attributor &A)
10928 : AAPotentialValues(IRP, A) {}
10929
10930 /// See AbstractAttribute::initialize(..).
10931 void initialize(Attributor &A) override {
10932 if (A.hasSimplificationCallback(getIRPosition())) {
10933 indicatePessimisticFixpoint();
10934 return;
10935 }
10936 Value *Stripped = getAssociatedValue().stripPointerCasts();
10937 if (isa<Constant>(Stripped) && !isa<ConstantExpr>(Stripped)) {
10938 addValue(A, getState(), *Stripped, getCtxI(), AA::AnyScope,
10939 getAnchorScope());
10940 indicateOptimisticFixpoint();
10941 return;
10942 }
10943 AAPotentialValues::initialize(A);
10944 }
10945
10946 /// See AbstractAttribute::getAsStr().
10947 const std::string getAsStr(Attributor *A) const override {
10948 std::string Str;
10949 llvm::raw_string_ostream OS(Str);
10950 OS << getState();
10951 return Str;
10952 }
10953
10954 template <typename AAType>
10955 static std::optional<Value *> askOtherAA(Attributor &A,
10956 const AbstractAttribute &AA,
10957 const IRPosition &IRP, Type &Ty) {
10959 return &IRP.getAssociatedValue();
10960 std::optional<Constant *> C = askForAssumedConstant<AAType>(A, AA, IRP, Ty);
10961 if (!C)
10962 return std::nullopt;
10963 if (*C)
10964 if (auto *CC = AA::getWithType(**C, Ty))
10965 return CC;
10966 return nullptr;
10967 }
10968
10969 virtual void addValue(Attributor &A, StateType &State, Value &V,
10970 const Instruction *CtxI, AA::ValueScope S,
10971 Function *AnchorScope) const {
10972
10973 IRPosition ValIRP = IRPosition::value(V);
10974 if (auto *CB = dyn_cast_or_null<CallBase>(CtxI)) {
10975 for (const auto &U : CB->args()) {
10976 if (U.get() != &V)
10977 continue;
10978 ValIRP = IRPosition::callsite_argument(*CB, CB->getArgOperandNo(&U));
10979 break;
10980 }
10981 }
10982
10983 Value *VPtr = &V;
10984 if (ValIRP.getAssociatedType()->isIntegerTy()) {
10985 Type &Ty = *getAssociatedType();
10986 std::optional<Value *> SimpleV =
10987 askOtherAA<AAValueConstantRange>(A, *this, ValIRP, Ty);
10988 if (SimpleV.has_value() && !*SimpleV) {
10989 auto *PotentialConstantsAA = A.getAAFor<AAPotentialConstantValues>(
10990 *this, ValIRP, DepClassTy::OPTIONAL);
10991 if (PotentialConstantsAA && PotentialConstantsAA->isValidState()) {
10992 for (const auto &It : PotentialConstantsAA->getAssumedSet())
10993 State.unionAssumed({{*ConstantInt::get(&Ty, It), nullptr}, S});
10994 if (PotentialConstantsAA->undefIsContained())
10995 State.unionAssumed({{*UndefValue::get(&Ty), nullptr}, S});
10996 return;
10997 }
10998 }
10999 if (!SimpleV.has_value())
11000 return;
11001
11002 if (*SimpleV)
11003 VPtr = *SimpleV;
11004 }
11005
11006 if (isa<ConstantInt>(VPtr))
11007 CtxI = nullptr;
11008 if (!AA::isValidInScope(*VPtr, AnchorScope))
11010
11011 State.unionAssumed({{*VPtr, CtxI}, S});
11012 }
11013
11014 /// Helper struct to tie a value+context pair together with the scope for
11015 /// which this is the simplified version.
11016 struct ItemInfo {
11017 AA::ValueAndContext I;
11019
11020 bool operator==(const ItemInfo &II) const {
11021 return II.I == I && II.S == S;
11022 };
11023 bool operator<(const ItemInfo &II) const {
11024 return std::tie(I, S) < std::tie(II.I, II.S);
11025 };
11026 };
11027
11028 bool recurseForValue(Attributor &A, const IRPosition &IRP, AA::ValueScope S) {
11029 SmallMapVector<AA::ValueAndContext, int, 8> ValueScopeMap;
11030 for (auto CS : {AA::Intraprocedural, AA::Interprocedural}) {
11031 if (!(CS & S))
11032 continue;
11033
11034 bool UsedAssumedInformation = false;
11036 if (!A.getAssumedSimplifiedValues(IRP, this, Values, CS,
11037 UsedAssumedInformation))
11038 return false;
11039
11040 for (auto &It : Values)
11041 ValueScopeMap[It] += CS;
11042 }
11043 for (auto &It : ValueScopeMap)
11044 addValue(A, getState(), *It.first.getValue(), It.first.getCtxI(),
11045 AA::ValueScope(It.second), getAnchorScope());
11046
11047 return true;
11048 }
11049
11050 void giveUpOnIntraprocedural(Attributor &A) {
11051 auto NewS = StateType::getBestState(getState());
11052 for (const auto &It : getAssumedSet()) {
11053 if (It.second == AA::Intraprocedural)
11054 continue;
11055 addValue(A, NewS, *It.first.getValue(), It.first.getCtxI(),
11056 AA::Interprocedural, getAnchorScope());
11057 }
11058 assert(!undefIsContained() && "Undef should be an explicit value!");
11059 addValue(A, NewS, getAssociatedValue(), getCtxI(), AA::Intraprocedural,
11060 getAnchorScope());
11061 getState() = NewS;
11062 }
11063
11064 /// See AbstractState::indicatePessimisticFixpoint(...).
11065 ChangeStatus indicatePessimisticFixpoint() override {
11066 getState() = StateType::getBestState(getState());
11067 getState().unionAssumed({{getAssociatedValue(), getCtxI()}, AA::AnyScope});
11068 AAPotentialValues::indicateOptimisticFixpoint();
11069 return ChangeStatus::CHANGED;
11070 }
11071
11072 /// See AbstractAttribute::updateImpl(...).
11073 ChangeStatus updateImpl(Attributor &A) override {
11074 return indicatePessimisticFixpoint();
11075 }
11076
11077 /// See AbstractAttribute::manifest(...).
11078 ChangeStatus manifest(Attributor &A) override {
11081 Values.clear();
11082 if (!getAssumedSimplifiedValues(A, Values, S))
11083 continue;
11084 Value &OldV = getAssociatedValue();
11085 if (isa<UndefValue>(OldV))
11086 continue;
11087 Value *NewV = getSingleValue(A, *this, getIRPosition(), Values);
11088 if (!NewV || NewV == &OldV)
11089 continue;
11090 if (getCtxI() &&
11091 !AA::isValidAtPosition({*NewV, *getCtxI()}, A.getInfoCache()))
11092 continue;
11093 if (A.changeAfterManifest(getIRPosition(), *NewV))
11094 return ChangeStatus::CHANGED;
11095 }
11096 return ChangeStatus::UNCHANGED;
11097 }
11098
11099 bool getAssumedSimplifiedValues(
11100 Attributor &A, SmallVectorImpl<AA::ValueAndContext> &Values,
11101 AA::ValueScope S, bool RecurseForSelectAndPHI = false) const override {
11102 if (!isValidState())
11103 return false;
11104 bool UsedAssumedInformation = false;
11105 for (const auto &It : getAssumedSet())
11106 if (It.second & S) {
11107 if (RecurseForSelectAndPHI && (isa<PHINode>(It.first.getValue()) ||
11108 isa<SelectInst>(It.first.getValue()))) {
11109 if (A.getAssumedSimplifiedValues(
11110 IRPosition::inst(*cast<Instruction>(It.first.getValue())),
11111 this, Values, S, UsedAssumedInformation))
11112 continue;
11113 }
11114 Values.push_back(It.first);
11115 }
11116 assert(!undefIsContained() && "Undef should be an explicit value!");
11117 return true;
11118 }
11119};
11120
11121struct AAPotentialValuesFloating : AAPotentialValuesImpl {
11122 AAPotentialValuesFloating(const IRPosition &IRP, Attributor &A)
11123 : AAPotentialValuesImpl(IRP, A) {}
11124
11125 /// See AbstractAttribute::updateImpl(...).
11126 ChangeStatus updateImpl(Attributor &A) override {
11127 auto AssumedBefore = getAssumed();
11128
11129 genericValueTraversal(A, &getAssociatedValue());
11130
11131 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11132 : ChangeStatus::CHANGED;
11133 }
11134
11135 /// Helper struct to remember which AAIsDead instances we actually used.
11136 struct LivenessInfo {
11137 const AAIsDead *LivenessAA = nullptr;
11138 bool AnyDead = false;
11139 };
11140
11141 /// Check if \p Cmp is a comparison we can simplify.
11142 ///
11143 /// We handle multiple cases, one in which at least one operand is an
11144 /// (assumed) nullptr. If so, try to simplify it using AANonNull on the other
11145 /// operand. Return true if successful, in that case Worklist will be updated.
11146 bool handleCmp(Attributor &A, Value &Cmp, Value *LHS, Value *RHS,
11147 CmpInst::Predicate Pred, ItemInfo II,
11148 SmallVectorImpl<ItemInfo> &Worklist) {
11149
11150 // Simplify the operands first.
11151 bool UsedAssumedInformation = false;
11152 SmallVector<AA::ValueAndContext> LHSValues, RHSValues;
11153 auto GetSimplifiedValues = [&](Value &V,
11155 if (!A.getAssumedSimplifiedValues(
11156 IRPosition::value(V, getCallBaseContext()), this, Values,
11157 AA::Intraprocedural, UsedAssumedInformation)) {
11158 Values.clear();
11159 Values.push_back(AA::ValueAndContext{V, II.I.getCtxI()});
11160 }
11161 return Values.empty();
11162 };
11163 if (GetSimplifiedValues(*LHS, LHSValues))
11164 return true;
11165 if (GetSimplifiedValues(*RHS, RHSValues))
11166 return true;
11167
11168 LLVMContext &Ctx = LHS->getContext();
11169
11170 InformationCache &InfoCache = A.getInfoCache();
11171 Instruction *CmpI = dyn_cast<Instruction>(&Cmp);
11172 Function *F = CmpI ? CmpI->getFunction() : nullptr;
11173 const auto *DT =
11174 F ? InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(*F)
11175 : nullptr;
11176 const auto *TLI =
11177 F ? A.getInfoCache().getTargetLibraryInfoForFunction(*F) : nullptr;
11178 auto *AC =
11179 F ? InfoCache.getAnalysisResultForFunction<AssumptionAnalysis>(*F)
11180 : nullptr;
11181
11182 const DataLayout &DL = A.getDataLayout();
11183 SimplifyQuery Q(DL, TLI, DT, AC, CmpI);
11184
11185 auto CheckPair = [&](Value &LHSV, Value &RHSV) {
11186 if (isa<UndefValue>(LHSV) || isa<UndefValue>(RHSV)) {
11187 addValue(A, getState(), *UndefValue::get(Cmp.getType()),
11188 /* CtxI */ nullptr, II.S, getAnchorScope());
11189 return true;
11190 }
11191
11192 // Handle the trivial case first in which we don't even need to think
11193 // about null or non-null.
11194 if (&LHSV == &RHSV &&
11196 Constant *NewV = ConstantInt::get(Type::getInt1Ty(Ctx),
11198 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11199 getAnchorScope());
11200 return true;
11201 }
11202
11203 auto *TypedLHS = AA::getWithType(LHSV, *LHS->getType());
11204 auto *TypedRHS = AA::getWithType(RHSV, *RHS->getType());
11205 if (TypedLHS && TypedRHS) {
11206 Value *NewV = simplifyCmpInst(Pred, TypedLHS, TypedRHS, Q);
11207 if (NewV && NewV != &Cmp) {
11208 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11209 getAnchorScope());
11210 return true;
11211 }
11212 }
11213
11214 // From now on we only handle equalities (==, !=).
11215 if (!CmpInst::isEquality(Pred))
11216 return false;
11217
11218 bool LHSIsNull = isa<ConstantPointerNull>(LHSV);
11219 bool RHSIsNull = isa<ConstantPointerNull>(RHSV);
11220 if (!LHSIsNull && !RHSIsNull)
11221 return false;
11222
11223 // Left is the nullptr ==/!= non-nullptr case. We'll use AANonNull on the
11224 // non-nullptr operand and if we assume it's non-null we can conclude the
11225 // result of the comparison.
11226 assert((LHSIsNull || RHSIsNull) &&
11227 "Expected nullptr versus non-nullptr comparison at this point");
11228
11229 // The index is the operand that we assume is not null.
11230 unsigned PtrIdx = LHSIsNull;
11231 bool IsKnownNonNull;
11232 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
11233 A, this, IRPosition::value(*(PtrIdx ? &RHSV : &LHSV)),
11234 DepClassTy::REQUIRED, IsKnownNonNull);
11235 if (!IsAssumedNonNull)
11236 return false;
11237
11238 // The new value depends on the predicate, true for != and false for ==.
11239 Constant *NewV =
11240 ConstantInt::get(Type::getInt1Ty(Ctx), Pred == CmpInst::ICMP_NE);
11241 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11242 getAnchorScope());
11243 return true;
11244 };
11245
11246 for (auto &LHSValue : LHSValues)
11247 for (auto &RHSValue : RHSValues)
11248 if (!CheckPair(*LHSValue.getValue(), *RHSValue.getValue()))
11249 return false;
11250 return true;
11251 }
11252
11253 bool handleSelectInst(Attributor &A, SelectInst &SI, ItemInfo II,
11254 SmallVectorImpl<ItemInfo> &Worklist) {
11255 const Instruction *CtxI = II.I.getCtxI();
11256 bool UsedAssumedInformation = false;
11257
11258 std::optional<Constant *> C =
11259 A.getAssumedConstant(*SI.getCondition(), *this, UsedAssumedInformation);
11260 bool NoValueYet = !C.has_value();
11261 if (NoValueYet || isa_and_nonnull<UndefValue>(*C))
11262 return true;
11263 if (auto *CI = dyn_cast_or_null<ConstantInt>(*C)) {
11264 if (CI->isZero())
11265 Worklist.push_back({{*SI.getFalseValue(), CtxI}, II.S});
11266 else
11267 Worklist.push_back({{*SI.getTrueValue(), CtxI}, II.S});
11268 } else if (&SI == &getAssociatedValue()) {
11269 // We could not simplify the condition, assume both values.
11270 Worklist.push_back({{*SI.getTrueValue(), CtxI}, II.S});
11271 Worklist.push_back({{*SI.getFalseValue(), CtxI}, II.S});
11272 } else {
11273 std::optional<Value *> SimpleV = A.getAssumedSimplified(
11274 IRPosition::inst(SI), *this, UsedAssumedInformation, II.S);
11275 if (!SimpleV.has_value())
11276 return true;
11277 if (*SimpleV) {
11278 addValue(A, getState(), **SimpleV, CtxI, II.S, getAnchorScope());
11279 return true;
11280 }
11281 return false;
11282 }
11283 return true;
11284 }
11285
11286 bool handleLoadInst(Attributor &A, LoadInst &LI, ItemInfo II,
11287 SmallVectorImpl<ItemInfo> &Worklist) {
11288 SmallSetVector<Value *, 4> PotentialCopies;
11289 SmallSetVector<Instruction *, 4> PotentialValueOrigins;
11290 bool UsedAssumedInformation = false;
11291 if (!AA::getPotentiallyLoadedValues(A, LI, PotentialCopies,
11292 PotentialValueOrigins, *this,
11293 UsedAssumedInformation,
11294 /* OnlyExact */ true)) {
11295 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Failed to get potentially "
11296 "loaded values for load instruction "
11297 << LI << "\n");
11298 return false;
11299 }
11300
11301 // Do not simplify loads that are only used in llvm.assume if we cannot also
11302 // remove all stores that may feed into the load. The reason is that the
11303 // assume is probably worth something as long as the stores are around.
11304 InformationCache &InfoCache = A.getInfoCache();
11305 if (InfoCache.isOnlyUsedByAssume(LI)) {
11306 if (!llvm::all_of(PotentialValueOrigins, [&](Instruction *I) {
11307 if (!I || isa<AssumeInst>(I))
11308 return true;
11309 if (auto *SI = dyn_cast<StoreInst>(I))
11310 return A.isAssumedDead(SI->getOperandUse(0), this,
11311 /* LivenessAA */ nullptr,
11312 UsedAssumedInformation,
11313 /* CheckBBLivenessOnly */ false);
11314 return A.isAssumedDead(*I, this, /* LivenessAA */ nullptr,
11315 UsedAssumedInformation,
11316 /* CheckBBLivenessOnly */ false);
11317 })) {
11318 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Load is onl used by assumes "
11319 "and we cannot delete all the stores: "
11320 << LI << "\n");
11321 return false;
11322 }
11323 }
11324
11325 // Values have to be dynamically unique or we loose the fact that a
11326 // single llvm::Value might represent two runtime values (e.g.,
11327 // stack locations in different recursive calls).
11328 const Instruction *CtxI = II.I.getCtxI();
11329 bool ScopeIsLocal = (II.S & AA::Intraprocedural);
11330 bool AllLocal = ScopeIsLocal;
11331 bool DynamicallyUnique = llvm::all_of(PotentialCopies, [&](Value *PC) {
11332 AllLocal &= AA::isValidInScope(*PC, getAnchorScope());
11333 return AA::isDynamicallyUnique(A, *this, *PC);
11334 });
11335 if (!DynamicallyUnique) {
11336 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Not all potentially loaded "
11337 "values are dynamically unique: "
11338 << LI << "\n");
11339 return false;
11340 }
11341
11342 for (auto *PotentialCopy : PotentialCopies) {
11343 if (AllLocal) {
11344 Worklist.push_back({{*PotentialCopy, CtxI}, II.S});
11345 } else {
11346 Worklist.push_back({{*PotentialCopy, CtxI}, AA::Interprocedural});
11347 }
11348 }
11349 if (!AllLocal && ScopeIsLocal)
11350 addValue(A, getState(), LI, CtxI, AA::Intraprocedural, getAnchorScope());
11351 return true;
11352 }
11353
11354 bool handlePHINode(
11355 Attributor &A, PHINode &PHI, ItemInfo II,
11356 SmallVectorImpl<ItemInfo> &Worklist,
11357 SmallMapVector<const Function *, LivenessInfo, 4> &LivenessAAs) {
11358 auto GetLivenessInfo = [&](const Function &F) -> LivenessInfo & {
11359 LivenessInfo &LI = LivenessAAs[&F];
11360 if (!LI.LivenessAA)
11361 LI.LivenessAA = A.getAAFor<AAIsDead>(*this, IRPosition::function(F),
11362 DepClassTy::NONE);
11363 return LI;
11364 };
11365
11366 if (&PHI == &getAssociatedValue()) {
11367 LivenessInfo &LI = GetLivenessInfo(*PHI.getFunction());
11368 const auto *CI =
11369 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
11370 *PHI.getFunction());
11371
11372 Cycle *C = nullptr;
11373 bool CyclePHI = mayBeInCycle(CI, &PHI, /* HeaderOnly */ true, &C);
11374 for (unsigned u = 0, e = PHI.getNumIncomingValues(); u < e; u++) {
11375 BasicBlock *IncomingBB = PHI.getIncomingBlock(u);
11376 if (LI.LivenessAA &&
11377 LI.LivenessAA->isEdgeDead(IncomingBB, PHI.getParent())) {
11378 LI.AnyDead = true;
11379 continue;
11380 }
11381 Value *V = PHI.getIncomingValue(u);
11382 if (V == &PHI)
11383 continue;
11384
11385 // If the incoming value is not the PHI but an instruction in the same
11386 // cycle we might have multiple versions of it flying around.
11387 if (CyclePHI && isa<Instruction>(V) &&
11388 (!C || C->contains(cast<Instruction>(V)->getParent())))
11389 return false;
11390
11391 Worklist.push_back({{*V, IncomingBB->getTerminator()}, II.S});
11392 }
11393 return true;
11394 }
11395
11396 bool UsedAssumedInformation = false;
11397 std::optional<Value *> SimpleV = A.getAssumedSimplified(
11398 IRPosition::inst(PHI), *this, UsedAssumedInformation, II.S);
11399 if (!SimpleV.has_value())
11400 return true;
11401 if (!(*SimpleV))
11402 return false;
11403 addValue(A, getState(), **SimpleV, &PHI, II.S, getAnchorScope());
11404 return true;
11405 }
11406
11407 /// Use the generic, non-optimistic InstSimplfy functionality if we managed to
11408 /// simplify any operand of the instruction \p I. Return true if successful,
11409 /// in that case Worklist will be updated.
11410 bool handleGenericInst(Attributor &A, Instruction &I, ItemInfo II,
11411 SmallVectorImpl<ItemInfo> &Worklist) {
11412 bool SomeSimplified = false;
11413 bool UsedAssumedInformation = false;
11414
11415 SmallVector<Value *, 8> NewOps(I.getNumOperands());
11416 int Idx = 0;
11417 for (Value *Op : I.operands()) {
11418 const auto &SimplifiedOp = A.getAssumedSimplified(
11419 IRPosition::value(*Op, getCallBaseContext()), *this,
11420 UsedAssumedInformation, AA::Intraprocedural);
11421 // If we are not sure about any operand we are not sure about the entire
11422 // instruction, we'll wait.
11423 if (!SimplifiedOp.has_value())
11424 return true;
11425
11426 if (*SimplifiedOp)
11427 NewOps[Idx] = *SimplifiedOp;
11428 else
11429 NewOps[Idx] = Op;
11430
11431 SomeSimplified |= (NewOps[Idx] != Op);
11432 ++Idx;
11433 }
11434
11435 // We won't bother with the InstSimplify interface if we didn't simplify any
11436 // operand ourselves.
11437 if (!SomeSimplified)
11438 return false;
11439
11440 InformationCache &InfoCache = A.getInfoCache();
11441 Function *F = I.getFunction();
11442 const auto *DT =
11443 InfoCache.getAnalysisResultForFunction<DominatorTreeAnalysis>(*F);
11444 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
11445 auto *AC = InfoCache.getAnalysisResultForFunction<AssumptionAnalysis>(*F);
11446
11447 const DataLayout &DL = I.getDataLayout();
11448 SimplifyQuery Q(DL, TLI, DT, AC, &I);
11449 Value *NewV = simplifyInstructionWithOperands(&I, NewOps, Q);
11450 if (!NewV || NewV == &I)
11451 return false;
11452
11453 LLVM_DEBUG(dbgs() << "Generic inst " << I << " assumed simplified to "
11454 << *NewV << "\n");
11455 Worklist.push_back({{*NewV, II.I.getCtxI()}, II.S});
11456 return true;
11457 }
11458
11460 Attributor &A, Instruction &I, ItemInfo II,
11461 SmallVectorImpl<ItemInfo> &Worklist,
11462 SmallMapVector<const Function *, LivenessInfo, 4> &LivenessAAs) {
11463 if (auto *CI = dyn_cast<CmpInst>(&I))
11464 return handleCmp(A, *CI, CI->getOperand(0), CI->getOperand(1),
11465 CI->getPredicate(), II, Worklist);
11466
11467 switch (I.getOpcode()) {
11468 case Instruction::Select:
11469 return handleSelectInst(A, cast<SelectInst>(I), II, Worklist);
11470 case Instruction::PHI:
11471 return handlePHINode(A, cast<PHINode>(I), II, Worklist, LivenessAAs);
11472 case Instruction::Load:
11473 return handleLoadInst(A, cast<LoadInst>(I), II, Worklist);
11474 default:
11475 return handleGenericInst(A, I, II, Worklist);
11476 };
11477 return false;
11478 }
11479
11480 void genericValueTraversal(Attributor &A, Value *InitialV) {
11481 SmallMapVector<const Function *, LivenessInfo, 4> LivenessAAs;
11482
11483 SmallSet<ItemInfo, 16> Visited;
11485 Worklist.push_back({{*InitialV, getCtxI()}, AA::AnyScope});
11486
11487 int Iteration = 0;
11488 do {
11489 ItemInfo II = Worklist.pop_back_val();
11490 Value *V = II.I.getValue();
11491 assert(V);
11492 const Instruction *CtxI = II.I.getCtxI();
11493 AA::ValueScope S = II.S;
11494
11495 // Check if we should process the current value. To prevent endless
11496 // recursion keep a record of the values we followed!
11497 if (!Visited.insert(II).second)
11498 continue;
11499
11500 // Make sure we limit the compile time for complex expressions.
11501 if (Iteration++ >= MaxPotentialValuesIterations) {
11502 LLVM_DEBUG(dbgs() << "Generic value traversal reached iteration limit: "
11503 << Iteration << "!\n");
11504 addValue(A, getState(), *V, CtxI, S, getAnchorScope());
11505 continue;
11506 }
11507
11508 // Explicitly look through calls with a "returned" attribute if we do
11509 // not have a pointer as stripPointerCasts only works on them.
11510 Value *NewV = nullptr;
11511 if (V->getType()->isPointerTy()) {
11512 NewV = AA::getWithType(*V->stripPointerCasts(), *V->getType());
11513 } else {
11514 if (auto *CB = dyn_cast<CallBase>(V))
11515 if (auto *Callee =
11517 for (Argument &Arg : Callee->args())
11518 if (Arg.hasReturnedAttr()) {
11519 NewV = CB->getArgOperand(Arg.getArgNo());
11520 break;
11521 }
11522 }
11523 }
11524 if (NewV && NewV != V) {
11525 Worklist.push_back({{*NewV, CtxI}, S});
11526 continue;
11527 }
11528
11529 if (auto *I = dyn_cast<Instruction>(V)) {
11530 if (simplifyInstruction(A, *I, II, Worklist, LivenessAAs))
11531 continue;
11532 }
11533
11534 if (V != InitialV || isa<Argument>(V))
11535 if (recurseForValue(A, IRPosition::value(*V), II.S))
11536 continue;
11537
11538 // If we haven't stripped anything we give up.
11539 if (V == InitialV && CtxI == getCtxI()) {
11540 indicatePessimisticFixpoint();
11541 return;
11542 }
11543
11544 addValue(A, getState(), *V, CtxI, S, getAnchorScope());
11545 } while (!Worklist.empty());
11546
11547 // If we actually used liveness information so we have to record a
11548 // dependence.
11549 for (auto &It : LivenessAAs)
11550 if (It.second.AnyDead)
11551 A.recordDependence(*It.second.LivenessAA, *this, DepClassTy::OPTIONAL);
11552 }
11553
11554 /// See AbstractAttribute::trackStatistics()
11555 void trackStatistics() const override {
11556 STATS_DECLTRACK_FLOATING_ATTR(potential_values)
11557 }
11558};
11559
11560struct AAPotentialValuesArgument final : AAPotentialValuesImpl {
11561 using Base = AAPotentialValuesImpl;
11562 AAPotentialValuesArgument(const IRPosition &IRP, Attributor &A)
11563 : Base(IRP, A) {}
11564
11565 /// See AbstractAttribute::initialize(..).
11566 void initialize(Attributor &A) override {
11567 auto &Arg = cast<Argument>(getAssociatedValue());
11569 indicatePessimisticFixpoint();
11570 }
11571
11572 /// See AbstractAttribute::updateImpl(...).
11573 ChangeStatus updateImpl(Attributor &A) override {
11574 auto AssumedBefore = getAssumed();
11575
11576 unsigned ArgNo = getCalleeArgNo();
11577
11578 bool UsedAssumedInformation = false;
11580 auto CallSitePred = [&](AbstractCallSite ACS) {
11581 const auto CSArgIRP = IRPosition::callsite_argument(ACS, ArgNo);
11582 if (CSArgIRP.getPositionKind() == IRP_INVALID)
11583 return false;
11584
11585 if (!A.getAssumedSimplifiedValues(CSArgIRP, this, Values,
11587 UsedAssumedInformation))
11588 return false;
11589
11590 return isValidState();
11591 };
11592
11593 if (!A.checkForAllCallSites(CallSitePred, *this,
11594 /* RequireAllCallSites */ true,
11595 UsedAssumedInformation))
11596 return indicatePessimisticFixpoint();
11597
11598 Function *Fn = getAssociatedFunction();
11599 bool AnyNonLocal = false;
11600 for (auto &It : Values) {
11601 if (isa<Constant>(It.getValue())) {
11602 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::AnyScope,
11603 getAnchorScope());
11604 continue;
11605 }
11606 if (!AA::isDynamicallyUnique(A, *this, *It.getValue()))
11607 return indicatePessimisticFixpoint();
11608
11609 if (auto *Arg = dyn_cast<Argument>(It.getValue()))
11610 if (Arg->getParent() == Fn) {
11611 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::AnyScope,
11612 getAnchorScope());
11613 continue;
11614 }
11615 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::Interprocedural,
11616 getAnchorScope());
11617 AnyNonLocal = true;
11618 }
11619 assert(!undefIsContained() && "Undef should be an explicit value!");
11620 if (AnyNonLocal)
11621 giveUpOnIntraprocedural(A);
11622
11623 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11624 : ChangeStatus::CHANGED;
11625 }
11626
11627 /// See AbstractAttribute::trackStatistics()
11628 void trackStatistics() const override {
11629 STATS_DECLTRACK_ARG_ATTR(potential_values)
11630 }
11631};
11632
11633struct AAPotentialValuesReturned : public AAPotentialValuesFloating {
11634 using Base = AAPotentialValuesFloating;
11635 AAPotentialValuesReturned(const IRPosition &IRP, Attributor &A)
11636 : Base(IRP, A) {}
11637
11638 /// See AbstractAttribute::initialize(..).
11639 void initialize(Attributor &A) override {
11640 Function *F = getAssociatedFunction();
11641 if (!F || F->isDeclaration() || F->getReturnType()->isVoidTy()) {
11642 indicatePessimisticFixpoint();
11643 return;
11644 }
11645
11646 for (Argument &Arg : F->args())
11647 if (Arg.hasReturnedAttr()) {
11648 addValue(A, getState(), Arg, nullptr, AA::AnyScope, F);
11649 ReturnedArg = &Arg;
11650 break;
11651 }
11652 if (!A.isFunctionIPOAmendable(*F) ||
11653 A.hasSimplificationCallback(getIRPosition())) {
11654 if (!ReturnedArg)
11655 indicatePessimisticFixpoint();
11656 else
11657 indicateOptimisticFixpoint();
11658 }
11659 }
11660
11661 /// See AbstractAttribute::updateImpl(...).
11662 ChangeStatus updateImpl(Attributor &A) override {
11663 auto AssumedBefore = getAssumed();
11664 bool UsedAssumedInformation = false;
11665
11667 Function *AnchorScope = getAnchorScope();
11668 auto HandleReturnedValue = [&](Value &V, Instruction *CtxI,
11669 bool AddValues) {
11671 Values.clear();
11672 if (!A.getAssumedSimplifiedValues(IRPosition::value(V), this, Values, S,
11673 UsedAssumedInformation,
11674 /* RecurseForSelectAndPHI */ true))
11675 return false;
11676 if (!AddValues)
11677 continue;
11678
11679 bool AllInterAreIntra = false;
11680 if (S == AA::Interprocedural)
11681 AllInterAreIntra =
11682 llvm::all_of(Values, [&](const AA::ValueAndContext &VAC) {
11683 return AA::isValidInScope(*VAC.getValue(), AnchorScope);
11684 });
11685
11686 for (const AA::ValueAndContext &VAC : Values) {
11687 addValue(A, getState(), *VAC.getValue(),
11688 VAC.getCtxI() ? VAC.getCtxI() : CtxI,
11689 AllInterAreIntra ? AA::AnyScope : S, AnchorScope);
11690 }
11691 if (AllInterAreIntra)
11692 break;
11693 }
11694 return true;
11695 };
11696
11697 if (ReturnedArg) {
11698 HandleReturnedValue(*ReturnedArg, nullptr, true);
11699 } else {
11700 auto RetInstPred = [&](Instruction &RetI) {
11701 bool AddValues = true;
11702 if (isa<PHINode>(RetI.getOperand(0)) ||
11703 isa<SelectInst>(RetI.getOperand(0))) {
11704 addValue(A, getState(), *RetI.getOperand(0), &RetI, AA::AnyScope,
11705 AnchorScope);
11706 AddValues = false;
11707 }
11708 return HandleReturnedValue(*RetI.getOperand(0), &RetI, AddValues);
11709 };
11710
11711 if (!A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
11712 UsedAssumedInformation,
11713 /* CheckBBLivenessOnly */ true))
11714 return indicatePessimisticFixpoint();
11715 }
11716
11717 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11718 : ChangeStatus::CHANGED;
11719 }
11720
11721 ChangeStatus manifest(Attributor &A) override {
11722 if (ReturnedArg)
11723 return ChangeStatus::UNCHANGED;
11725 if (!getAssumedSimplifiedValues(A, Values, AA::ValueScope::Intraprocedural,
11726 /* RecurseForSelectAndPHI */ true))
11727 return ChangeStatus::UNCHANGED;
11728 Value *NewVal = getSingleValue(A, *this, getIRPosition(), Values);
11729 if (!NewVal)
11730 return ChangeStatus::UNCHANGED;
11731
11732 ChangeStatus Changed = ChangeStatus::UNCHANGED;
11733 if (auto *Arg = dyn_cast<Argument>(NewVal)) {
11734 STATS_DECLTRACK(UniqueReturnValue, FunctionReturn,
11735 "Number of function with unique return");
11736 Changed |= A.manifestAttrs(
11738 {Attribute::get(Arg->getContext(), Attribute::Returned)});
11739 STATS_DECLTRACK_ARG_ATTR(returned);
11740 }
11741
11742 auto RetInstPred = [&](Instruction &RetI) {
11743 Value *RetOp = RetI.getOperand(0);
11744 if (isa<UndefValue>(RetOp) || RetOp == NewVal)
11745 return true;
11746 if (AA::isValidAtPosition({*NewVal, RetI}, A.getInfoCache()))
11747 if (A.changeUseAfterManifest(RetI.getOperandUse(0), *NewVal))
11748 Changed = ChangeStatus::CHANGED;
11749 return true;
11750 };
11751 bool UsedAssumedInformation = false;
11752 (void)A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
11753 UsedAssumedInformation,
11754 /* CheckBBLivenessOnly */ true);
11755 return Changed;
11756 }
11757
11758 ChangeStatus indicatePessimisticFixpoint() override {
11759 return AAPotentialValues::indicatePessimisticFixpoint();
11760 }
11761
11762 /// See AbstractAttribute::trackStatistics()
11763 void trackStatistics() const override{
11764 STATS_DECLTRACK_FNRET_ATTR(potential_values)}
11765
11766 /// The argumented with an existing `returned` attribute.
11767 Argument *ReturnedArg = nullptr;
11768};
11769
11770struct AAPotentialValuesFunction : AAPotentialValuesImpl {
11771 AAPotentialValuesFunction(const IRPosition &IRP, Attributor &A)
11772 : AAPotentialValuesImpl(IRP, A) {}
11773
11774 /// See AbstractAttribute::updateImpl(...).
11775 ChangeStatus updateImpl(Attributor &A) override {
11776 llvm_unreachable("AAPotentialValues(Function|CallSite)::updateImpl will "
11777 "not be called");
11778 }
11779
11780 /// See AbstractAttribute::trackStatistics()
11781 void trackStatistics() const override {
11782 STATS_DECLTRACK_FN_ATTR(potential_values)
11783 }
11784};
11785
11786struct AAPotentialValuesCallSite : AAPotentialValuesFunction {
11787 AAPotentialValuesCallSite(const IRPosition &IRP, Attributor &A)
11788 : AAPotentialValuesFunction(IRP, A) {}
11789
11790 /// See AbstractAttribute::trackStatistics()
11791 void trackStatistics() const override {
11792 STATS_DECLTRACK_CS_ATTR(potential_values)
11793 }
11794};
11795
11796struct AAPotentialValuesCallSiteReturned : AAPotentialValuesImpl {
11797 AAPotentialValuesCallSiteReturned(const IRPosition &IRP, Attributor &A)
11798 : AAPotentialValuesImpl(IRP, A) {}
11799
11800 /// See AbstractAttribute::updateImpl(...).
11801 ChangeStatus updateImpl(Attributor &A) override {
11802 auto AssumedBefore = getAssumed();
11803
11804 Function *Callee = getAssociatedFunction();
11805 if (!Callee)
11806 return indicatePessimisticFixpoint();
11807
11808 bool UsedAssumedInformation = false;
11809 auto *CB = cast<CallBase>(getCtxI());
11810 if (CB->isMustTailCall() &&
11811 !A.isAssumedDead(IRPosition::inst(*CB), this, nullptr,
11812 UsedAssumedInformation))
11813 return indicatePessimisticFixpoint();
11814
11815 Function *Caller = CB->getCaller();
11816
11817 auto AddScope = [&](AA::ValueScope S) {
11819 if (!A.getAssumedSimplifiedValues(IRPosition::returned(*Callee), this,
11820 Values, S, UsedAssumedInformation))
11821 return false;
11822
11823 for (auto &It : Values) {
11824 Value *V = It.getValue();
11825 std::optional<Value *> CallerV = A.translateArgumentToCallSiteContent(
11826 V, *CB, *this, UsedAssumedInformation);
11827 if (!CallerV.has_value()) {
11828 // Nothing to do as long as no value was determined.
11829 continue;
11830 }
11831 V = *CallerV ? *CallerV : V;
11832 if (*CallerV && AA::isDynamicallyUnique(A, *this, *V)) {
11833 if (recurseForValue(A, IRPosition::value(*V), S))
11834 continue;
11835 }
11836 if (S == AA::Intraprocedural && !AA::isValidInScope(*V, Caller)) {
11837 giveUpOnIntraprocedural(A);
11838 return true;
11839 }
11840 addValue(A, getState(), *V, CB, S, getAnchorScope());
11841 }
11842 return true;
11843 };
11844 if (!AddScope(AA::Intraprocedural))
11845 return indicatePessimisticFixpoint();
11846 if (!AddScope(AA::Interprocedural))
11847 return indicatePessimisticFixpoint();
11848 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11849 : ChangeStatus::CHANGED;
11850 }
11851
11852 ChangeStatus indicatePessimisticFixpoint() override {
11853 return AAPotentialValues::indicatePessimisticFixpoint();
11854 }
11855
11856 /// See AbstractAttribute::trackStatistics()
11857 void trackStatistics() const override {
11858 STATS_DECLTRACK_CSRET_ATTR(potential_values)
11859 }
11860};
11861
11862struct AAPotentialValuesCallSiteArgument : AAPotentialValuesFloating {
11863 AAPotentialValuesCallSiteArgument(const IRPosition &IRP, Attributor &A)
11864 : AAPotentialValuesFloating(IRP, A) {}
11865
11866 /// See AbstractAttribute::trackStatistics()
11867 void trackStatistics() const override {
11868 STATS_DECLTRACK_CSARG_ATTR(potential_values)
11869 }
11870};
11871} // namespace
11872
11873/// ---------------------- Assumption Propagation ------------------------------
11874namespace {
11875struct AAAssumptionInfoImpl : public AAAssumptionInfo {
11876 AAAssumptionInfoImpl(const IRPosition &IRP, Attributor &A,
11877 const DenseSet<StringRef> &Known)
11878 : AAAssumptionInfo(IRP, A, Known) {}
11879
11880 /// See AbstractAttribute::manifest(...).
11881 ChangeStatus manifest(Attributor &A) override {
11882 // Don't manifest a universal set if it somehow made it here.
11883 if (getKnown().isUniversal())
11884 return ChangeStatus::UNCHANGED;
11885
11886 const IRPosition &IRP = getIRPosition();
11887 SmallVector<StringRef, 0> Set(getAssumed().getSet().begin(),
11888 getAssumed().getSet().end());
11889 llvm::sort(Set);
11890 return A.manifestAttrs(IRP,
11891 Attribute::get(IRP.getAnchorValue().getContext(),
11893 llvm::join(Set, ",")),
11894 /*ForceReplace=*/true);
11895 }
11896
11897 bool hasAssumption(const StringRef Assumption) const override {
11898 return isValidState() && setContains(Assumption);
11899 }
11900
11901 /// See AbstractAttribute::getAsStr()
11902 const std::string getAsStr(Attributor *A) const override {
11903 const SetContents &Known = getKnown();
11904 const SetContents &Assumed = getAssumed();
11905
11906 SmallVector<StringRef, 0> Set(Known.getSet().begin(), Known.getSet().end());
11907 llvm::sort(Set);
11908 const std::string KnownStr = llvm::join(Set, ",");
11909
11910 std::string AssumedStr = "Universal";
11911 if (!Assumed.isUniversal()) {
11912 Set.assign(Assumed.getSet().begin(), Assumed.getSet().end());
11913 AssumedStr = llvm::join(Set, ",");
11914 }
11915 return "Known [" + KnownStr + "]," + " Assumed [" + AssumedStr + "]";
11916 }
11917};
11918
11919/// Propagates assumption information from parent functions to all of their
11920/// successors. An assumption can be propagated if the containing function
11921/// dominates the called function.
11922///
11923/// We start with a "known" set of assumptions already valid for the associated
11924/// function and an "assumed" set that initially contains all possible
11925/// assumptions. The assumed set is inter-procedurally updated by narrowing its
11926/// contents as concrete values are known. The concrete values are seeded by the
11927/// first nodes that are either entries into the call graph, or contains no
11928/// assumptions. Each node is updated as the intersection of the assumed state
11929/// with all of its predecessors.
11930struct AAAssumptionInfoFunction final : AAAssumptionInfoImpl {
11931 AAAssumptionInfoFunction(const IRPosition &IRP, Attributor &A)
11932 : AAAssumptionInfoImpl(IRP, A,
11933 getAssumptions(*IRP.getAssociatedFunction())) {}
11934
11935 /// See AbstractAttribute::updateImpl(...).
11936 ChangeStatus updateImpl(Attributor &A) override {
11937 bool Changed = false;
11938
11939 auto CallSitePred = [&](AbstractCallSite ACS) {
11940 const auto *AssumptionAA = A.getAAFor<AAAssumptionInfo>(
11941 *this, IRPosition::callsite_function(*ACS.getInstruction()),
11942 DepClassTy::REQUIRED);
11943 if (!AssumptionAA)
11944 return false;
11945 // Get the set of assumptions shared by all of this function's callers.
11946 Changed |= getIntersection(AssumptionAA->getAssumed());
11947 return !getAssumed().empty() || !getKnown().empty();
11948 };
11949
11950 bool UsedAssumedInformation = false;
11951 // Get the intersection of all assumptions held by this node's predecessors.
11952 // If we don't know all the call sites then this is either an entry into the
11953 // call graph or an empty node. This node is known to only contain its own
11954 // assumptions and can be propagated to its successors.
11955 if (!A.checkForAllCallSites(CallSitePred, *this, true,
11956 UsedAssumedInformation))
11957 return indicatePessimisticFixpoint();
11958
11959 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
11960 }
11961
11962 void trackStatistics() const override {}
11963};
11964
11965/// Assumption Info defined for call sites.
11966struct AAAssumptionInfoCallSite final : AAAssumptionInfoImpl {
11967
11968 AAAssumptionInfoCallSite(const IRPosition &IRP, Attributor &A)
11969 : AAAssumptionInfoImpl(IRP, A, getInitialAssumptions(IRP)) {}
11970
11971 /// See AbstractAttribute::initialize(...).
11972 void initialize(Attributor &A) override {
11973 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
11974 A.getAAFor<AAAssumptionInfo>(*this, FnPos, DepClassTy::REQUIRED);
11975 }
11976
11977 /// See AbstractAttribute::updateImpl(...).
11978 ChangeStatus updateImpl(Attributor &A) override {
11979 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
11980 auto *AssumptionAA =
11981 A.getAAFor<AAAssumptionInfo>(*this, FnPos, DepClassTy::REQUIRED);
11982 if (!AssumptionAA)
11983 return indicatePessimisticFixpoint();
11984 bool Changed = getIntersection(AssumptionAA->getAssumed());
11985 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
11986 }
11987
11988 /// See AbstractAttribute::trackStatistics()
11989 void trackStatistics() const override {}
11990
11991private:
11992 /// Helper to initialized the known set as all the assumptions this call and
11993 /// the callee contain.
11994 DenseSet<StringRef> getInitialAssumptions(const IRPosition &IRP) {
11995 const CallBase &CB = cast<CallBase>(IRP.getAssociatedValue());
11996 auto Assumptions = getAssumptions(CB);
11997 if (const Function *F = CB.getCaller())
11998 set_union(Assumptions, getAssumptions(*F));
11999 if (Function *F = IRP.getAssociatedFunction())
12000 set_union(Assumptions, getAssumptions(*F));
12001 return Assumptions;
12002 }
12003};
12004} // namespace
12005
12007 return static_cast<AACallGraphNode *>(const_cast<AACallEdges *>(
12008 A.getOrCreateAAFor<AACallEdges>(IRPosition::function(**I))));
12009}
12010
12012
12013/// ------------------------ UnderlyingObjects ---------------------------------
12014
12015namespace {
12016struct AAUnderlyingObjectsImpl
12017 : StateWrapper<BooleanState, AAUnderlyingObjects> {
12019 AAUnderlyingObjectsImpl(const IRPosition &IRP, Attributor &A) : BaseTy(IRP) {}
12020
12021 /// See AbstractAttribute::getAsStr().
12022 const std::string getAsStr(Attributor *A) const override {
12023 if (!isValidState())
12024 return "<invalid>";
12025 std::string Str;
12027 OS << "underlying objects: inter " << InterAssumedUnderlyingObjects.size()
12028 << " objects, intra " << IntraAssumedUnderlyingObjects.size()
12029 << " objects.\n";
12030 if (!InterAssumedUnderlyingObjects.empty()) {
12031 OS << "inter objects:\n";
12032 for (auto *Obj : InterAssumedUnderlyingObjects)
12033 OS << *Obj << '\n';
12034 }
12035 if (!IntraAssumedUnderlyingObjects.empty()) {
12036 OS << "intra objects:\n";
12037 for (auto *Obj : IntraAssumedUnderlyingObjects)
12038 OS << *Obj << '\n';
12039 }
12040 return Str;
12041 }
12042
12043 /// See AbstractAttribute::trackStatistics()
12044 void trackStatistics() const override {}
12045
12046 /// See AbstractAttribute::updateImpl(...).
12047 ChangeStatus updateImpl(Attributor &A) override {
12048 auto &Ptr = getAssociatedValue();
12049
12050 bool UsedAssumedInformation = false;
12051 auto DoUpdate = [&](SmallSetVector<Value *, 8> &UnderlyingObjects,
12053 SmallPtrSet<Value *, 8> SeenObjects;
12055
12056 if (!A.getAssumedSimplifiedValues(IRPosition::value(Ptr), *this, Values,
12057 Scope, UsedAssumedInformation))
12058 return UnderlyingObjects.insert(&Ptr);
12059
12060 bool Changed = false;
12061
12062 for (unsigned I = 0; I < Values.size(); ++I) {
12063 auto &VAC = Values[I];
12064 auto *Obj = VAC.getValue();
12065 Value *UO = getUnderlyingObject(Obj);
12066 if (!SeenObjects.insert(UO ? UO : Obj).second)
12067 continue;
12068 if (UO && UO != Obj) {
12069 if (isa<AllocaInst>(UO) || isa<GlobalValue>(UO)) {
12070 Changed |= UnderlyingObjects.insert(UO);
12071 continue;
12072 }
12073
12074 const auto *OtherAA = A.getAAFor<AAUnderlyingObjects>(
12075 *this, IRPosition::value(*UO), DepClassTy::OPTIONAL);
12076 auto Pred = [&](Value &V) {
12077 if (&V == UO)
12078 Changed |= UnderlyingObjects.insert(UO);
12079 else
12080 Values.emplace_back(V, nullptr);
12081 return true;
12082 };
12083
12084 if (!OtherAA || !OtherAA->forallUnderlyingObjects(Pred, Scope))
12086 "The forall call should not return false at this position");
12087 UsedAssumedInformation |= !OtherAA->getState().isAtFixpoint();
12088 continue;
12089 }
12090
12091 if (isa<SelectInst>(Obj)) {
12092 Changed |= handleIndirect(A, *Obj, UnderlyingObjects, Scope,
12093 UsedAssumedInformation);
12094 continue;
12095 }
12096 if (auto *PHI = dyn_cast<PHINode>(Obj)) {
12097 // Explicitly look through PHIs as we do not care about dynamically
12098 // uniqueness.
12099 for (unsigned u = 0, e = PHI->getNumIncomingValues(); u < e; u++) {
12100 Changed |=
12101 handleIndirect(A, *PHI->getIncomingValue(u), UnderlyingObjects,
12102 Scope, UsedAssumedInformation);
12103 }
12104 continue;
12105 }
12106
12107 Changed |= UnderlyingObjects.insert(Obj);
12108 }
12109
12110 return Changed;
12111 };
12112
12113 bool Changed = false;
12114 Changed |= DoUpdate(IntraAssumedUnderlyingObjects, AA::Intraprocedural);
12115 Changed |= DoUpdate(InterAssumedUnderlyingObjects, AA::Interprocedural);
12116 if (!UsedAssumedInformation)
12117 indicateOptimisticFixpoint();
12118 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
12119 }
12120
12121 bool forallUnderlyingObjects(
12122 function_ref<bool(Value &)> Pred,
12123 AA::ValueScope Scope = AA::Interprocedural) const override {
12124 if (!isValidState())
12125 return Pred(getAssociatedValue());
12126
12127 auto &AssumedUnderlyingObjects = Scope == AA::Intraprocedural
12128 ? IntraAssumedUnderlyingObjects
12129 : InterAssumedUnderlyingObjects;
12130 for (Value *Obj : AssumedUnderlyingObjects)
12131 if (!Pred(*Obj))
12132 return false;
12133
12134 return true;
12135 }
12136
12137private:
12138 /// Handle the case where the value is not the actual underlying value, such
12139 /// as a phi node or a select instruction.
12140 bool handleIndirect(Attributor &A, Value &V,
12141 SmallSetVector<Value *, 8> &UnderlyingObjects,
12142 AA::ValueScope Scope, bool &UsedAssumedInformation) {
12143 bool Changed = false;
12144 const auto *AA = A.getAAFor<AAUnderlyingObjects>(
12145 *this, IRPosition::value(V), DepClassTy::OPTIONAL);
12146 auto Pred = [&](Value &V) {
12147 Changed |= UnderlyingObjects.insert(&V);
12148 return true;
12149 };
12150 if (!AA || !AA->forallUnderlyingObjects(Pred, Scope))
12152 "The forall call should not return false at this position");
12153 UsedAssumedInformation |= !AA->getState().isAtFixpoint();
12154 return Changed;
12155 }
12156
12157 /// All the underlying objects collected so far via intra procedural scope.
12158 SmallSetVector<Value *, 8> IntraAssumedUnderlyingObjects;
12159 /// All the underlying objects collected so far via inter procedural scope.
12160 SmallSetVector<Value *, 8> InterAssumedUnderlyingObjects;
12161};
12162
12163struct AAUnderlyingObjectsFloating final : AAUnderlyingObjectsImpl {
12164 AAUnderlyingObjectsFloating(const IRPosition &IRP, Attributor &A)
12165 : AAUnderlyingObjectsImpl(IRP, A) {}
12166};
12167
12168struct AAUnderlyingObjectsArgument final : AAUnderlyingObjectsImpl {
12169 AAUnderlyingObjectsArgument(const IRPosition &IRP, Attributor &A)
12170 : AAUnderlyingObjectsImpl(IRP, A) {}
12171};
12172
12173struct AAUnderlyingObjectsCallSite final : AAUnderlyingObjectsImpl {
12174 AAUnderlyingObjectsCallSite(const IRPosition &IRP, Attributor &A)
12175 : AAUnderlyingObjectsImpl(IRP, A) {}
12176};
12177
12178struct AAUnderlyingObjectsCallSiteArgument final : AAUnderlyingObjectsImpl {
12179 AAUnderlyingObjectsCallSiteArgument(const IRPosition &IRP, Attributor &A)
12180 : AAUnderlyingObjectsImpl(IRP, A) {}
12181};
12182
12183struct AAUnderlyingObjectsReturned final : AAUnderlyingObjectsImpl {
12184 AAUnderlyingObjectsReturned(const IRPosition &IRP, Attributor &A)
12185 : AAUnderlyingObjectsImpl(IRP, A) {}
12186};
12187
12188struct AAUnderlyingObjectsCallSiteReturned final : AAUnderlyingObjectsImpl {
12189 AAUnderlyingObjectsCallSiteReturned(const IRPosition &IRP, Attributor &A)
12190 : AAUnderlyingObjectsImpl(IRP, A) {}
12191};
12192
12193struct AAUnderlyingObjectsFunction final : AAUnderlyingObjectsImpl {
12194 AAUnderlyingObjectsFunction(const IRPosition &IRP, Attributor &A)
12195 : AAUnderlyingObjectsImpl(IRP, A) {}
12196};
12197} // namespace
12198
12199/// ------------------------ Global Value Info -------------------------------
12200namespace {
12201struct AAGlobalValueInfoFloating : public AAGlobalValueInfo {
12202 AAGlobalValueInfoFloating(const IRPosition &IRP, Attributor &A)
12203 : AAGlobalValueInfo(IRP, A) {}
12204
12205 /// See AbstractAttribute::initialize(...).
12206 void initialize(Attributor &A) override {}
12207
12208 bool checkUse(Attributor &A, const Use &U, bool &Follow,
12209 SmallVectorImpl<const Value *> &Worklist) {
12210 Instruction *UInst = dyn_cast<Instruction>(U.getUser());
12211 if (!UInst) {
12212 Follow = true;
12213 return true;
12214 }
12215
12216 LLVM_DEBUG(dbgs() << "[AAGlobalValueInfo] Check use: " << *U.get() << " in "
12217 << *UInst << "\n");
12218
12219 if (auto *Cmp = dyn_cast<ICmpInst>(U.getUser())) {
12220 int Idx = &Cmp->getOperandUse(0) == &U;
12221 if (isa<Constant>(Cmp->getOperand(Idx)))
12222 return true;
12223 return U == &getAnchorValue();
12224 }
12225
12226 // Explicitly catch return instructions.
12227 if (isa<ReturnInst>(UInst)) {
12228 auto CallSitePred = [&](AbstractCallSite ACS) {
12229 Worklist.push_back(ACS.getInstruction());
12230 return true;
12231 };
12232 bool UsedAssumedInformation = false;
12233 // TODO: We should traverse the uses or add a "non-call-site" CB.
12234 if (!A.checkForAllCallSites(CallSitePred, *UInst->getFunction(),
12235 /*RequireAllCallSites=*/true, this,
12236 UsedAssumedInformation))
12237 return false;
12238 return true;
12239 }
12240
12241 // For now we only use special logic for call sites. However, the tracker
12242 // itself knows about a lot of other non-capturing cases already.
12243 auto *CB = dyn_cast<CallBase>(UInst);
12244 if (!CB)
12245 return false;
12246 // Direct calls are OK uses.
12247 if (CB->isCallee(&U))
12248 return true;
12249 // Non-argument uses are scary.
12250 if (!CB->isArgOperand(&U))
12251 return false;
12252 // TODO: Iterate callees.
12253 auto *Fn = dyn_cast<Function>(CB->getCalledOperand());
12254 if (!Fn || !A.isFunctionIPOAmendable(*Fn))
12255 return false;
12256
12257 unsigned ArgNo = CB->getArgOperandNo(&U);
12258 Worklist.push_back(Fn->getArg(ArgNo));
12259 return true;
12260 }
12261
12262 ChangeStatus updateImpl(Attributor &A) override {
12263 unsigned NumUsesBefore = Uses.size();
12264
12265 SmallPtrSet<const Value *, 8> Visited;
12267 Worklist.push_back(&getAnchorValue());
12268
12269 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
12270 Uses.insert(&U);
12271 // TODO(captures): Make this more precise.
12272 UseCaptureInfo CI = DetermineUseCaptureKind(U, /*Base=*/nullptr);
12273 if (CI.isPassthrough()) {
12274 Follow = true;
12275 return true;
12276 }
12277 return checkUse(A, U, Follow, Worklist);
12278 };
12279 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
12280 Uses.insert(&OldU);
12281 return true;
12282 };
12283
12284 while (!Worklist.empty()) {
12285 const Value *V = Worklist.pop_back_val();
12286 if (!Visited.insert(V).second)
12287 continue;
12288 if (!A.checkForAllUses(UsePred, *this, *V,
12289 /* CheckBBLivenessOnly */ true,
12290 DepClassTy::OPTIONAL,
12291 /* IgnoreDroppableUses */ true, EquivalentUseCB)) {
12292 return indicatePessimisticFixpoint();
12293 }
12294 }
12295
12296 return Uses.size() == NumUsesBefore ? ChangeStatus::UNCHANGED
12297 : ChangeStatus::CHANGED;
12298 }
12299
12300 bool isPotentialUse(const Use &U) const override {
12301 return !isValidState() || Uses.contains(&U);
12302 }
12303
12304 /// See AbstractAttribute::manifest(...).
12305 ChangeStatus manifest(Attributor &A) override {
12306 return ChangeStatus::UNCHANGED;
12307 }
12308
12309 /// See AbstractAttribute::getAsStr().
12310 const std::string getAsStr(Attributor *A) const override {
12311 return "[" + std::to_string(Uses.size()) + " uses]";
12312 }
12313
12314 void trackStatistics() const override {
12315 STATS_DECLTRACK_FLOATING_ATTR(GlobalValuesTracked);
12316 }
12317
12318private:
12319 /// Set of (transitive) uses of this GlobalValue.
12320 SmallPtrSet<const Use *, 8> Uses;
12321};
12322} // namespace
12323
12324/// ------------------------ Indirect Call Info -------------------------------
12325namespace {
12326struct AAIndirectCallInfoCallSite : public AAIndirectCallInfo {
12327 AAIndirectCallInfoCallSite(const IRPosition &IRP, Attributor &A)
12328 : AAIndirectCallInfo(IRP, A) {}
12329
12330 /// See AbstractAttribute::initialize(...).
12331 void initialize(Attributor &A) override {
12332 auto *MD = getCtxI()->getMetadata(LLVMContext::MD_callees);
12333 if (!MD && !A.isClosedWorldModule())
12334 return;
12335
12336 if (MD) {
12337 for (const auto &Op : MD->operands())
12338 if (Function *Callee = mdconst::dyn_extract_or_null<Function>(Op))
12339 PotentialCallees.insert(Callee);
12340 } else if (A.isClosedWorldModule()) {
12341 ArrayRef<Function *> IndirectlyCallableFunctions =
12342 A.getInfoCache().getIndirectlyCallableFunctions(A);
12343 PotentialCallees.insert_range(IndirectlyCallableFunctions);
12344 }
12345
12346 if (PotentialCallees.empty())
12347 indicateOptimisticFixpoint();
12348 }
12349
12350 ChangeStatus updateImpl(Attributor &A) override {
12351 CallBase *CB = cast<CallBase>(getCtxI());
12352 const Use &CalleeUse = CB->getCalledOperandUse();
12353 Value *FP = CB->getCalledOperand();
12354
12355 SmallSetVector<Function *, 4> AssumedCalleesNow;
12356 bool AllCalleesKnownNow = AllCalleesKnown;
12357
12358 auto CheckPotentialCalleeUse = [&](Function &PotentialCallee,
12359 bool &UsedAssumedInformation) {
12360 const auto *GIAA = A.getAAFor<AAGlobalValueInfo>(
12361 *this, IRPosition::value(PotentialCallee), DepClassTy::OPTIONAL);
12362 if (!GIAA || GIAA->isPotentialUse(CalleeUse))
12363 return true;
12364 UsedAssumedInformation = !GIAA->isAtFixpoint();
12365 return false;
12366 };
12367
12368 auto AddPotentialCallees = [&]() {
12369 for (auto *PotentialCallee : PotentialCallees) {
12370 bool UsedAssumedInformation = false;
12371 if (CheckPotentialCalleeUse(*PotentialCallee, UsedAssumedInformation))
12372 AssumedCalleesNow.insert(PotentialCallee);
12373 }
12374 };
12375
12376 // Use simplification to find potential callees, if !callees was present,
12377 // fallback to that set if necessary.
12378 bool UsedAssumedInformation = false;
12380 if (!A.getAssumedSimplifiedValues(IRPosition::value(*FP), this, Values,
12381 AA::ValueScope::AnyScope,
12382 UsedAssumedInformation)) {
12383 if (PotentialCallees.empty())
12384 return indicatePessimisticFixpoint();
12385 AddPotentialCallees();
12386 }
12387
12388 // Try to find a reason for \p Fn not to be a potential callee. If none was
12389 // found, add it to the assumed callees set.
12390 auto CheckPotentialCallee = [&](Function &Fn) {
12391 if (!PotentialCallees.empty() && !PotentialCallees.count(&Fn))
12392 return false;
12393
12394 auto &CachedResult = FilterResults[&Fn];
12395 if (CachedResult.has_value())
12396 return CachedResult.value();
12397
12398 bool UsedAssumedInformation = false;
12399 if (!CheckPotentialCalleeUse(Fn, UsedAssumedInformation)) {
12400 if (!UsedAssumedInformation)
12401 CachedResult = false;
12402 return false;
12403 }
12404
12405 int NumFnArgs = Fn.arg_size();
12406 int NumCBArgs = CB->arg_size();
12407
12408 // Check if any excess argument (which we fill up with poison) is known to
12409 // be UB on undef.
12410 for (int I = NumCBArgs; I < NumFnArgs; ++I) {
12411 bool IsKnown = false;
12413 A, this, IRPosition::argument(*Fn.getArg(I)),
12414 DepClassTy::OPTIONAL, IsKnown)) {
12415 if (IsKnown)
12416 CachedResult = false;
12417 return false;
12418 }
12419 }
12420
12421 CachedResult = true;
12422 return true;
12423 };
12424
12425 // Check simplification result, prune known UB callees, also restrict it to
12426 // the !callees set, if present.
12427 for (auto &VAC : Values) {
12428 if (isa<UndefValue>(VAC.getValue()))
12429 continue;
12431 VAC.getValue()->getType()->getPointerAddressSpace() == 0)
12432 continue;
12433 // TODO: Check for known UB, e.g., poison + noundef.
12434 if (auto *VACFn = dyn_cast<Function>(VAC.getValue())) {
12435 if (CheckPotentialCallee(*VACFn))
12436 AssumedCalleesNow.insert(VACFn);
12437 continue;
12438 }
12439 if (!PotentialCallees.empty()) {
12440 AddPotentialCallees();
12441 break;
12442 }
12443 AllCalleesKnownNow = false;
12444 }
12445
12446 if (AssumedCalleesNow == AssumedCallees &&
12447 AllCalleesKnown == AllCalleesKnownNow)
12448 return ChangeStatus::UNCHANGED;
12449
12450 std::swap(AssumedCallees, AssumedCalleesNow);
12451 AllCalleesKnown = AllCalleesKnownNow;
12452 return ChangeStatus::CHANGED;
12453 }
12454
12455 /// See AbstractAttribute::manifest(...).
12456 ChangeStatus manifest(Attributor &A) override {
12457 // If we can't specialize at all, give up now.
12458 if (!AllCalleesKnown && AssumedCallees.empty())
12459 return ChangeStatus::UNCHANGED;
12460
12461 CallBase *CB = cast<CallBase>(getCtxI());
12462 bool UsedAssumedInformation = false;
12463 if (A.isAssumedDead(*CB, this, /*LivenessAA=*/nullptr,
12464 UsedAssumedInformation))
12465 return ChangeStatus::UNCHANGED;
12466
12467 ChangeStatus Changed = ChangeStatus::UNCHANGED;
12468 Value *FP = CB->getCalledOperand();
12469 if (FP->getType()->getPointerAddressSpace())
12470 FP = new AddrSpaceCastInst(FP, PointerType::get(FP->getContext(), 0),
12471 FP->getName() + ".as0", CB->getIterator());
12472
12473 bool CBIsVoid = CB->getType()->isVoidTy();
12475 FunctionType *CSFT = CB->getFunctionType();
12476 SmallVector<Value *> CSArgs(CB->args());
12477
12478 // If we know all callees and there are none, the call site is (effectively)
12479 // dead (or UB).
12480 if (AssumedCallees.empty()) {
12481 assert(AllCalleesKnown &&
12482 "Expected all callees to be known if there are none.");
12483 A.changeToUnreachableAfterManifest(CB);
12484 return ChangeStatus::CHANGED;
12485 }
12486
12487 // Special handling for the single callee case.
12488 if (AllCalleesKnown && AssumedCallees.size() == 1) {
12489 auto *NewCallee = AssumedCallees.front();
12490 if (isLegalToPromote(*CB, NewCallee)) {
12491 promoteCall(*CB, NewCallee, nullptr);
12492 NumIndirectCallsPromoted++;
12493 return ChangeStatus::CHANGED;
12494 }
12495 Instruction *NewCall =
12496 CallInst::Create(FunctionCallee(CSFT, NewCallee), CSArgs,
12497 CB->getName(), CB->getIterator());
12498 if (!CBIsVoid)
12499 A.changeAfterManifest(IRPosition::callsite_returned(*CB), *NewCall);
12500 A.deleteAfterManifest(*CB);
12501 return ChangeStatus::CHANGED;
12502 }
12503
12504 // For each potential value we create a conditional
12505 //
12506 // ```
12507 // if (ptr == value) value(args);
12508 // else ...
12509 // ```
12510 //
12511 bool SpecializedForAnyCallees = false;
12512 bool SpecializedForAllCallees = AllCalleesKnown;
12513 ICmpInst *LastCmp = nullptr;
12514 SmallVector<Function *, 8> SkippedAssumedCallees;
12516 for (Function *NewCallee : AssumedCallees) {
12517 if (!A.shouldSpecializeCallSiteForCallee(*this, *CB, *NewCallee,
12518 AssumedCallees.size())) {
12519 SkippedAssumedCallees.push_back(NewCallee);
12520 SpecializedForAllCallees = false;
12521 continue;
12522 }
12523 SpecializedForAnyCallees = true;
12524
12525 LastCmp = new ICmpInst(IP, llvm::CmpInst::ICMP_EQ, FP, NewCallee);
12526 Instruction *ThenTI =
12527 SplitBlockAndInsertIfThen(LastCmp, IP, /* Unreachable */ false);
12528 BasicBlock *CBBB = CB->getParent();
12529 A.registerManifestAddedBasicBlock(*ThenTI->getParent());
12530 A.registerManifestAddedBasicBlock(*IP->getParent());
12531 auto *SplitTI = cast<BranchInst>(LastCmp->getNextNode());
12532 BasicBlock *ElseBB;
12533 if (&*IP == CB) {
12534 ElseBB = BasicBlock::Create(ThenTI->getContext(), "",
12535 ThenTI->getFunction(), CBBB);
12536 A.registerManifestAddedBasicBlock(*ElseBB);
12537 IP = BranchInst::Create(CBBB, ElseBB)->getIterator();
12538 SplitTI->replaceUsesOfWith(CBBB, ElseBB);
12539 } else {
12540 ElseBB = IP->getParent();
12541 ThenTI->replaceUsesOfWith(ElseBB, CBBB);
12542 }
12543 CastInst *RetBC = nullptr;
12544 CallInst *NewCall = nullptr;
12545 if (isLegalToPromote(*CB, NewCallee)) {
12546 auto *CBClone = cast<CallBase>(CB->clone());
12547 CBClone->insertBefore(ThenTI->getIterator());
12548 NewCall = &cast<CallInst>(promoteCall(*CBClone, NewCallee, &RetBC));
12549 NumIndirectCallsPromoted++;
12550 } else {
12551 NewCall = CallInst::Create(FunctionCallee(CSFT, NewCallee), CSArgs,
12552 CB->getName(), ThenTI->getIterator());
12553 }
12554 NewCalls.push_back({NewCall, RetBC});
12555 }
12556
12557 auto AttachCalleeMetadata = [&](CallBase &IndirectCB) {
12558 if (!AllCalleesKnown)
12559 return ChangeStatus::UNCHANGED;
12560 MDBuilder MDB(IndirectCB.getContext());
12561 MDNode *Callees = MDB.createCallees(SkippedAssumedCallees);
12562 IndirectCB.setMetadata(LLVMContext::MD_callees, Callees);
12563 return ChangeStatus::CHANGED;
12564 };
12565
12566 if (!SpecializedForAnyCallees)
12567 return AttachCalleeMetadata(*CB);
12568
12569 // Check if we need the fallback indirect call still.
12570 if (SpecializedForAllCallees) {
12572 LastCmp->eraseFromParent();
12573 new UnreachableInst(IP->getContext(), IP);
12574 IP->eraseFromParent();
12575 } else {
12576 auto *CBClone = cast<CallInst>(CB->clone());
12577 CBClone->setName(CB->getName());
12578 CBClone->insertBefore(*IP->getParent(), IP);
12579 NewCalls.push_back({CBClone, nullptr});
12580 AttachCalleeMetadata(*CBClone);
12581 }
12582
12583 // Check if we need a PHI to merge the results.
12584 if (!CBIsVoid) {
12585 auto *PHI = PHINode::Create(CB->getType(), NewCalls.size(),
12586 CB->getName() + ".phi",
12587 CB->getParent()->getFirstInsertionPt());
12588 for (auto &It : NewCalls) {
12589 CallBase *NewCall = It.first;
12590 Instruction *CallRet = It.second ? It.second : It.first;
12591 if (CallRet->getType() == CB->getType())
12592 PHI->addIncoming(CallRet, CallRet->getParent());
12593 else if (NewCall->getType()->isVoidTy())
12594 PHI->addIncoming(PoisonValue::get(CB->getType()),
12595 NewCall->getParent());
12596 else
12597 llvm_unreachable("Call return should match or be void!");
12598 }
12599 A.changeAfterManifest(IRPosition::callsite_returned(*CB), *PHI);
12600 }
12601
12602 A.deleteAfterManifest(*CB);
12603 Changed = ChangeStatus::CHANGED;
12604
12605 return Changed;
12606 }
12607
12608 /// See AbstractAttribute::getAsStr().
12609 const std::string getAsStr(Attributor *A) const override {
12610 return std::string(AllCalleesKnown ? "eliminate" : "specialize") +
12611 " indirect call site with " + std::to_string(AssumedCallees.size()) +
12612 " functions";
12613 }
12614
12615 void trackStatistics() const override {
12616 if (AllCalleesKnown) {
12618 Eliminated, CallSites,
12619 "Number of indirect call sites eliminated via specialization")
12620 } else {
12621 STATS_DECLTRACK(Specialized, CallSites,
12622 "Number of indirect call sites specialized")
12623 }
12624 }
12625
12626 bool foreachCallee(function_ref<bool(Function *)> CB) const override {
12627 return isValidState() && AllCalleesKnown && all_of(AssumedCallees, CB);
12628 }
12629
12630private:
12631 /// Map to remember filter results.
12632 DenseMap<Function *, std::optional<bool>> FilterResults;
12633
12634 /// If the !callee metadata was present, this set will contain all potential
12635 /// callees (superset).
12636 SmallSetVector<Function *, 4> PotentialCallees;
12637
12638 /// This set contains all currently assumed calllees, which might grow over
12639 /// time.
12640 SmallSetVector<Function *, 4> AssumedCallees;
12641
12642 /// Flag to indicate if all possible callees are in the AssumedCallees set or
12643 /// if there could be others.
12644 bool AllCalleesKnown = true;
12645};
12646} // namespace
12647
12648/// --------------------- Invariant Load Pointer -------------------------------
12649namespace {
12650
12651struct AAInvariantLoadPointerImpl
12652 : public StateWrapper<BitIntegerState<uint8_t, 15>,
12653 AAInvariantLoadPointer> {
12654
12655 enum {
12656 // pointer does not alias within the bounds of the function
12657 IS_NOALIAS = 1 << 0,
12658 // pointer is not involved in any effectful instructions within the bounds
12659 // of the function
12660 IS_NOEFFECT = 1 << 1,
12661 // loads are invariant within the bounds of the function
12662 IS_LOCALLY_INVARIANT = 1 << 2,
12663 // memory lifetime is constrained within the bounds of the function
12664 IS_LOCALLY_CONSTRAINED = 1 << 3,
12665
12666 IS_BEST_STATE = IS_NOALIAS | IS_NOEFFECT | IS_LOCALLY_INVARIANT |
12667 IS_LOCALLY_CONSTRAINED,
12668 };
12669 static_assert(getBestState() == IS_BEST_STATE, "Unexpected best state");
12670
12671 using Base =
12672 StateWrapper<BitIntegerState<uint8_t, 15>, AAInvariantLoadPointer>;
12673
12674 // the BitIntegerState is optimistic about IS_NOALIAS and IS_NOEFFECT, but
12675 // pessimistic about IS_KNOWN_INVARIANT
12676 AAInvariantLoadPointerImpl(const IRPosition &IRP, Attributor &A)
12677 : Base(IRP) {}
12678
12679 bool isKnownInvariant() const final {
12680 return isKnownLocallyInvariant() && isKnown(IS_LOCALLY_CONSTRAINED);
12681 }
12682
12683 bool isKnownLocallyInvariant() const final {
12684 if (isKnown(IS_LOCALLY_INVARIANT))
12685 return true;
12686 return isKnown(IS_NOALIAS | IS_NOEFFECT);
12687 }
12688
12689 bool isAssumedInvariant() const final {
12690 return isAssumedLocallyInvariant() && isAssumed(IS_LOCALLY_CONSTRAINED);
12691 }
12692
12693 bool isAssumedLocallyInvariant() const final {
12694 if (isAssumed(IS_LOCALLY_INVARIANT))
12695 return true;
12696 return isAssumed(IS_NOALIAS | IS_NOEFFECT);
12697 }
12698
12699 ChangeStatus updateImpl(Attributor &A) override {
12700 ChangeStatus Changed = ChangeStatus::UNCHANGED;
12701
12702 Changed |= updateNoAlias(A);
12703 if (requiresNoAlias() && !isAssumed(IS_NOALIAS))
12704 return indicatePessimisticFixpoint();
12705
12706 Changed |= updateNoEffect(A);
12707
12708 Changed |= updateLocalInvariance(A);
12709
12710 return Changed;
12711 }
12712
12713 ChangeStatus manifest(Attributor &A) override {
12714 if (!isKnownInvariant())
12715 return ChangeStatus::UNCHANGED;
12716
12717 ChangeStatus Changed = ChangeStatus::UNCHANGED;
12718 const Value *Ptr = &getAssociatedValue();
12719 const auto TagInvariantLoads = [&](const Use &U, bool &) {
12720 if (U.get() != Ptr)
12721 return true;
12722 auto *I = dyn_cast<Instruction>(U.getUser());
12723 if (!I)
12724 return true;
12725
12726 // Ensure that we are only changing uses from the corresponding callgraph
12727 // SSC in the case that the AA isn't run on the entire module
12728 if (!A.isRunOn(I->getFunction()))
12729 return true;
12730
12731 if (I->hasMetadata(LLVMContext::MD_invariant_load))
12732 return true;
12733
12734 if (auto *LI = dyn_cast<LoadInst>(I)) {
12735 LI->setMetadata(LLVMContext::MD_invariant_load,
12736 MDNode::get(LI->getContext(), {}));
12737 Changed = ChangeStatus::CHANGED;
12738 }
12739 return true;
12740 };
12741
12742 (void)A.checkForAllUses(TagInvariantLoads, *this, *Ptr);
12743 return Changed;
12744 }
12745
12746 /// See AbstractAttribute::getAsStr().
12747 const std::string getAsStr(Attributor *) const override {
12748 if (isKnownInvariant())
12749 return "load-invariant pointer";
12750 return "non-invariant pointer";
12751 }
12752
12753 /// See AbstractAttribute::trackStatistics().
12754 void trackStatistics() const override {}
12755
12756private:
12757 /// Indicate that noalias is required for the pointer to be invariant.
12758 bool requiresNoAlias() const {
12759 switch (getPositionKind()) {
12760 default:
12761 // Conservatively default to require noalias.
12762 return true;
12763 case IRP_FLOAT:
12764 case IRP_RETURNED:
12765 case IRP_CALL_SITE:
12766 return false;
12767 case IRP_CALL_SITE_RETURNED: {
12768 const auto &CB = cast<CallBase>(getAnchorValue());
12770 &CB, /*MustPreserveNullness=*/false);
12771 }
12772 case IRP_ARGUMENT: {
12773 const Function *F = getAssociatedFunction();
12774 assert(F && "no associated function for argument");
12775 return !isCallableCC(F->getCallingConv());
12776 }
12777 }
12778 }
12779
12780 bool isExternal() const {
12781 const Function *F = getAssociatedFunction();
12782 if (!F)
12783 return true;
12784 return isCallableCC(F->getCallingConv()) &&
12785 getPositionKind() != IRP_CALL_SITE_RETURNED;
12786 }
12787
12788 ChangeStatus updateNoAlias(Attributor &A) {
12789 if (isKnown(IS_NOALIAS) || !isAssumed(IS_NOALIAS))
12790 return ChangeStatus::UNCHANGED;
12791
12792 // Try to use AANoAlias.
12793 if (const auto *ANoAlias = A.getOrCreateAAFor<AANoAlias>(
12794 getIRPosition(), this, DepClassTy::REQUIRED)) {
12795 if (ANoAlias->isKnownNoAlias()) {
12796 addKnownBits(IS_NOALIAS);
12797 return ChangeStatus::CHANGED;
12798 }
12799
12800 if (!ANoAlias->isAssumedNoAlias()) {
12801 removeAssumedBits(IS_NOALIAS);
12802 return ChangeStatus::CHANGED;
12803 }
12804
12805 return ChangeStatus::UNCHANGED;
12806 }
12807
12808 // Try to infer noalias from argument attribute, since it is applicable for
12809 // the duration of the function.
12810 if (const Argument *Arg = getAssociatedArgument()) {
12811 if (Arg->hasNoAliasAttr()) {
12812 addKnownBits(IS_NOALIAS);
12813 return ChangeStatus::UNCHANGED;
12814 }
12815
12816 // Noalias information is not provided, and cannot be inferred,
12817 // so we conservatively assume the pointer aliases.
12818 removeAssumedBits(IS_NOALIAS);
12819 return ChangeStatus::CHANGED;
12820 }
12821
12822 return ChangeStatus::UNCHANGED;
12823 }
12824
12825 ChangeStatus updateNoEffect(Attributor &A) {
12826 if (isKnown(IS_NOEFFECT) || !isAssumed(IS_NOEFFECT))
12827 return ChangeStatus::UNCHANGED;
12828
12829 if (!getAssociatedFunction())
12830 return indicatePessimisticFixpoint();
12831
12832 if (isa<AllocaInst>(&getAssociatedValue()))
12833 return indicatePessimisticFixpoint();
12834
12835 const auto HasNoEffectLoads = [&](const Use &U, bool &) {
12836 const auto *LI = dyn_cast<LoadInst>(U.getUser());
12837 return !LI || !LI->mayHaveSideEffects();
12838 };
12839 if (!A.checkForAllUses(HasNoEffectLoads, *this, getAssociatedValue()))
12840 return indicatePessimisticFixpoint();
12841
12842 if (const auto *AMemoryBehavior = A.getOrCreateAAFor<AAMemoryBehavior>(
12843 getIRPosition(), this, DepClassTy::REQUIRED)) {
12844 // For non-instructions, try to use AAMemoryBehavior to infer the readonly
12845 // attribute
12846 if (!AMemoryBehavior->isAssumedReadOnly())
12847 return indicatePessimisticFixpoint();
12848
12849 if (AMemoryBehavior->isKnownReadOnly()) {
12850 addKnownBits(IS_NOEFFECT);
12851 return ChangeStatus::UNCHANGED;
12852 }
12853
12854 return ChangeStatus::UNCHANGED;
12855 }
12856
12857 if (const Argument *Arg = getAssociatedArgument()) {
12858 if (Arg->onlyReadsMemory()) {
12859 addKnownBits(IS_NOEFFECT);
12860 return ChangeStatus::UNCHANGED;
12861 }
12862
12863 // Readonly information is not provided, and cannot be inferred from
12864 // AAMemoryBehavior.
12865 return indicatePessimisticFixpoint();
12866 }
12867
12868 return ChangeStatus::UNCHANGED;
12869 }
12870
12871 ChangeStatus updateLocalInvariance(Attributor &A) {
12872 if (isKnown(IS_LOCALLY_INVARIANT) || !isAssumed(IS_LOCALLY_INVARIANT))
12873 return ChangeStatus::UNCHANGED;
12874
12875 // try to infer invariance from underlying objects
12876 const auto *AUO = A.getOrCreateAAFor<AAUnderlyingObjects>(
12877 getIRPosition(), this, DepClassTy::REQUIRED);
12878 if (!AUO)
12879 return ChangeStatus::UNCHANGED;
12880
12881 bool UsedAssumedInformation = false;
12882 const auto IsLocallyInvariantLoadIfPointer = [&](const Value &V) {
12883 if (!V.getType()->isPointerTy())
12884 return true;
12885 const auto *IsInvariantLoadPointer =
12886 A.getOrCreateAAFor<AAInvariantLoadPointer>(IRPosition::value(V), this,
12887 DepClassTy::REQUIRED);
12888 // Conservatively fail if invariance cannot be inferred.
12889 if (!IsInvariantLoadPointer)
12890 return false;
12891
12892 if (IsInvariantLoadPointer->isKnownLocallyInvariant())
12893 return true;
12894 if (!IsInvariantLoadPointer->isAssumedLocallyInvariant())
12895 return false;
12896
12897 UsedAssumedInformation = true;
12898 return true;
12899 };
12900 if (!AUO->forallUnderlyingObjects(IsLocallyInvariantLoadIfPointer))
12901 return indicatePessimisticFixpoint();
12902
12903 if (const auto *CB = dyn_cast<CallBase>(&getAnchorValue())) {
12905 CB, /*MustPreserveNullness=*/false)) {
12906 for (const Value *Arg : CB->args()) {
12907 if (!IsLocallyInvariantLoadIfPointer(*Arg))
12908 return indicatePessimisticFixpoint();
12909 }
12910 }
12911 }
12912
12913 if (!UsedAssumedInformation) {
12914 // Pointer is known and not just assumed to be locally invariant.
12915 addKnownBits(IS_LOCALLY_INVARIANT);
12916 return ChangeStatus::CHANGED;
12917 }
12918
12919 return ChangeStatus::UNCHANGED;
12920 }
12921};
12922
12923struct AAInvariantLoadPointerFloating final : AAInvariantLoadPointerImpl {
12924 AAInvariantLoadPointerFloating(const IRPosition &IRP, Attributor &A)
12925 : AAInvariantLoadPointerImpl(IRP, A) {}
12926};
12927
12928struct AAInvariantLoadPointerReturned final : AAInvariantLoadPointerImpl {
12929 AAInvariantLoadPointerReturned(const IRPosition &IRP, Attributor &A)
12930 : AAInvariantLoadPointerImpl(IRP, A) {}
12931
12932 void initialize(Attributor &) override {
12933 removeAssumedBits(IS_LOCALLY_CONSTRAINED);
12934 }
12935};
12936
12937struct AAInvariantLoadPointerCallSiteReturned final
12938 : AAInvariantLoadPointerImpl {
12939 AAInvariantLoadPointerCallSiteReturned(const IRPosition &IRP, Attributor &A)
12940 : AAInvariantLoadPointerImpl(IRP, A) {}
12941
12942 void initialize(Attributor &A) override {
12943 const Function *F = getAssociatedFunction();
12944 assert(F && "no associated function for return from call");
12945
12946 if (!F->isDeclaration() && !F->isIntrinsic())
12947 return AAInvariantLoadPointerImpl::initialize(A);
12948
12949 const auto &CB = cast<CallBase>(getAnchorValue());
12951 &CB, /*MustPreserveNullness=*/false))
12952 return AAInvariantLoadPointerImpl::initialize(A);
12953
12954 if (F->onlyReadsMemory() && F->hasNoSync())
12955 return AAInvariantLoadPointerImpl::initialize(A);
12956
12957 // At this point, the function is opaque, so we conservatively assume
12958 // non-invariance.
12959 indicatePessimisticFixpoint();
12960 }
12961};
12962
12963struct AAInvariantLoadPointerArgument final : AAInvariantLoadPointerImpl {
12964 AAInvariantLoadPointerArgument(const IRPosition &IRP, Attributor &A)
12965 : AAInvariantLoadPointerImpl(IRP, A) {}
12966
12967 void initialize(Attributor &) override {
12968 const Function *F = getAssociatedFunction();
12969 assert(F && "no associated function for argument");
12970
12971 if (!isCallableCC(F->getCallingConv())) {
12972 addKnownBits(IS_LOCALLY_CONSTRAINED);
12973 return;
12974 }
12975
12976 if (!F->hasLocalLinkage())
12977 removeAssumedBits(IS_LOCALLY_CONSTRAINED);
12978 }
12979};
12980
12981struct AAInvariantLoadPointerCallSiteArgument final
12982 : AAInvariantLoadPointerImpl {
12983 AAInvariantLoadPointerCallSiteArgument(const IRPosition &IRP, Attributor &A)
12984 : AAInvariantLoadPointerImpl(IRP, A) {}
12985};
12986} // namespace
12987
12988/// ------------------------ Address Space ------------------------------------
12989namespace {
12990
12991template <typename InstType>
12992static bool makeChange(Attributor &A, InstType *MemInst, const Use &U,
12993 Value *OriginalValue, PointerType *NewPtrTy,
12994 bool UseOriginalValue) {
12995 if (U.getOperandNo() != InstType::getPointerOperandIndex())
12996 return false;
12997
12998 if (MemInst->isVolatile()) {
12999 auto *TTI = A.getInfoCache().getAnalysisResultForFunction<TargetIRAnalysis>(
13000 *MemInst->getFunction());
13001 unsigned NewAS = NewPtrTy->getPointerAddressSpace();
13002 if (!TTI || !TTI->hasVolatileVariant(MemInst, NewAS))
13003 return false;
13004 }
13005
13006 if (UseOriginalValue) {
13007 A.changeUseAfterManifest(const_cast<Use &>(U), *OriginalValue);
13008 return true;
13009 }
13010
13011 Instruction *CastInst = new AddrSpaceCastInst(OriginalValue, NewPtrTy);
13012 CastInst->insertBefore(MemInst->getIterator());
13013 A.changeUseAfterManifest(const_cast<Use &>(U), *CastInst);
13014 return true;
13015}
13016
13017struct AAAddressSpaceImpl : public AAAddressSpace {
13018 AAAddressSpaceImpl(const IRPosition &IRP, Attributor &A)
13019 : AAAddressSpace(IRP, A) {}
13020
13021 uint32_t getAddressSpace() const override {
13022 assert(isValidState() && "the AA is invalid");
13023 return AssumedAddressSpace;
13024 }
13025
13026 /// See AbstractAttribute::initialize(...).
13027 void initialize(Attributor &A) override {
13028 assert(getAssociatedType()->isPtrOrPtrVectorTy() &&
13029 "Associated value is not a pointer");
13030
13031 if (!A.getInfoCache().getFlatAddressSpace().has_value()) {
13032 indicatePessimisticFixpoint();
13033 return;
13034 }
13035
13036 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
13037 unsigned AS = getAssociatedType()->getPointerAddressSpace();
13038 if (AS != FlatAS) {
13039 [[maybe_unused]] bool R = takeAddressSpace(AS);
13040 assert(R && "The take should happen");
13041 indicateOptimisticFixpoint();
13042 }
13043 }
13044
13045 ChangeStatus updateImpl(Attributor &A) override {
13046 uint32_t OldAddressSpace = AssumedAddressSpace;
13047 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
13048
13049 auto CheckAddressSpace = [&](Value &Obj) {
13050 // Ignore undef.
13051 if (isa<UndefValue>(&Obj))
13052 return true;
13053
13054 // If the object already has a non-flat address space, we simply take it.
13055 unsigned ObjAS = Obj.getType()->getPointerAddressSpace();
13056 if (ObjAS != FlatAS)
13057 return takeAddressSpace(ObjAS);
13058
13059 // At this point, we know Obj is in the flat address space. For a final
13060 // attempt, we want to use getAssumedAddrSpace, but first we must get the
13061 // associated function, if possible.
13062 Function *F = nullptr;
13063 if (auto *Arg = dyn_cast<Argument>(&Obj))
13064 F = Arg->getParent();
13065 else if (auto *I = dyn_cast<Instruction>(&Obj))
13066 F = I->getFunction();
13067
13068 // Use getAssumedAddrSpace if the associated function exists.
13069 if (F) {
13070 auto *TTI =
13071 A.getInfoCache().getAnalysisResultForFunction<TargetIRAnalysis>(*F);
13072 unsigned AssumedAS = TTI->getAssumedAddrSpace(&Obj);
13073 if (AssumedAS != ~0U)
13074 return takeAddressSpace(AssumedAS);
13075 }
13076
13077 // Now we can't do anything else but to take the flat AS.
13078 return takeAddressSpace(FlatAS);
13079 };
13080
13081 auto *AUO = A.getOrCreateAAFor<AAUnderlyingObjects>(getIRPosition(), this,
13082 DepClassTy::REQUIRED);
13083 if (!AUO->forallUnderlyingObjects(CheckAddressSpace))
13084 return indicatePessimisticFixpoint();
13085
13086 return OldAddressSpace == AssumedAddressSpace ? ChangeStatus::UNCHANGED
13087 : ChangeStatus::CHANGED;
13088 }
13089
13090 /// See AbstractAttribute::manifest(...).
13091 ChangeStatus manifest(Attributor &A) override {
13092 unsigned NewAS = getAddressSpace();
13093
13094 if (NewAS == InvalidAddressSpace ||
13095 NewAS == getAssociatedType()->getPointerAddressSpace())
13096 return ChangeStatus::UNCHANGED;
13097
13098 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
13099
13100 Value *AssociatedValue = &getAssociatedValue();
13101 Value *OriginalValue = peelAddrspacecast(AssociatedValue, FlatAS);
13102
13103 PointerType *NewPtrTy =
13104 PointerType::get(getAssociatedType()->getContext(), NewAS);
13105 bool UseOriginalValue =
13106 OriginalValue->getType()->getPointerAddressSpace() == NewAS;
13107
13108 bool Changed = false;
13109
13110 auto Pred = [&](const Use &U, bool &) {
13111 if (U.get() != AssociatedValue)
13112 return true;
13113 auto *Inst = dyn_cast<Instruction>(U.getUser());
13114 if (!Inst)
13115 return true;
13116 // This is a WA to make sure we only change uses from the corresponding
13117 // CGSCC if the AA is run on CGSCC instead of the entire module.
13118 if (!A.isRunOn(Inst->getFunction()))
13119 return true;
13120 if (auto *LI = dyn_cast<LoadInst>(Inst)) {
13121 Changed |=
13122 makeChange(A, LI, U, OriginalValue, NewPtrTy, UseOriginalValue);
13123 } else if (auto *SI = dyn_cast<StoreInst>(Inst)) {
13124 Changed |=
13125 makeChange(A, SI, U, OriginalValue, NewPtrTy, UseOriginalValue);
13126 } else if (auto *RMW = dyn_cast<AtomicRMWInst>(Inst)) {
13127 Changed |=
13128 makeChange(A, RMW, U, OriginalValue, NewPtrTy, UseOriginalValue);
13129 } else if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(Inst)) {
13130 Changed |=
13131 makeChange(A, CmpX, U, OriginalValue, NewPtrTy, UseOriginalValue);
13132 }
13133 return true;
13134 };
13135
13136 // It doesn't matter if we can't check all uses as we can simply
13137 // conservatively ignore those that can not be visited.
13138 (void)A.checkForAllUses(Pred, *this, getAssociatedValue(),
13139 /* CheckBBLivenessOnly */ true);
13140
13141 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
13142 }
13143
13144 /// See AbstractAttribute::getAsStr().
13145 const std::string getAsStr(Attributor *A) const override {
13146 if (!isValidState())
13147 return "addrspace(<invalid>)";
13148 return "addrspace(" +
13149 (AssumedAddressSpace == InvalidAddressSpace
13150 ? "none"
13151 : std::to_string(AssumedAddressSpace)) +
13152 ")";
13153 }
13154
13155private:
13156 uint32_t AssumedAddressSpace = InvalidAddressSpace;
13157
13158 bool takeAddressSpace(uint32_t AS) {
13159 if (AssumedAddressSpace == InvalidAddressSpace) {
13160 AssumedAddressSpace = AS;
13161 return true;
13162 }
13163 return AssumedAddressSpace == AS;
13164 }
13165
13166 static Value *peelAddrspacecast(Value *V, unsigned FlatAS) {
13167 if (auto *I = dyn_cast<AddrSpaceCastInst>(V)) {
13168 assert(I->getSrcAddressSpace() != FlatAS &&
13169 "there should not be flat AS -> non-flat AS");
13170 return I->getPointerOperand();
13171 }
13172 if (auto *C = dyn_cast<ConstantExpr>(V))
13173 if (C->getOpcode() == Instruction::AddrSpaceCast) {
13174 assert(C->getOperand(0)->getType()->getPointerAddressSpace() !=
13175 FlatAS &&
13176 "there should not be flat AS -> non-flat AS X");
13177 return C->getOperand(0);
13178 }
13179 return V;
13180 }
13181};
13182
13183struct AAAddressSpaceFloating final : AAAddressSpaceImpl {
13184 AAAddressSpaceFloating(const IRPosition &IRP, Attributor &A)
13185 : AAAddressSpaceImpl(IRP, A) {}
13186
13187 void trackStatistics() const override {
13189 }
13190};
13191
13192struct AAAddressSpaceReturned final : AAAddressSpaceImpl {
13193 AAAddressSpaceReturned(const IRPosition &IRP, Attributor &A)
13194 : AAAddressSpaceImpl(IRP, A) {}
13195
13196 /// See AbstractAttribute::initialize(...).
13197 void initialize(Attributor &A) override {
13198 // TODO: we don't rewrite function argument for now because it will need to
13199 // rewrite the function signature and all call sites.
13200 (void)indicatePessimisticFixpoint();
13201 }
13202
13203 void trackStatistics() const override {
13204 STATS_DECLTRACK_FNRET_ATTR(addrspace);
13205 }
13206};
13207
13208struct AAAddressSpaceCallSiteReturned final : AAAddressSpaceImpl {
13209 AAAddressSpaceCallSiteReturned(const IRPosition &IRP, Attributor &A)
13210 : AAAddressSpaceImpl(IRP, A) {}
13211
13212 void trackStatistics() const override {
13213 STATS_DECLTRACK_CSRET_ATTR(addrspace);
13214 }
13215};
13216
13217struct AAAddressSpaceArgument final : AAAddressSpaceImpl {
13218 AAAddressSpaceArgument(const IRPosition &IRP, Attributor &A)
13219 : AAAddressSpaceImpl(IRP, A) {}
13220
13221 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(addrspace); }
13222};
13223
13224struct AAAddressSpaceCallSiteArgument final : AAAddressSpaceImpl {
13225 AAAddressSpaceCallSiteArgument(const IRPosition &IRP, Attributor &A)
13226 : AAAddressSpaceImpl(IRP, A) {}
13227
13228 /// See AbstractAttribute::initialize(...).
13229 void initialize(Attributor &A) override {
13230 // TODO: we don't rewrite call site argument for now because it will need to
13231 // rewrite the function signature of the callee.
13232 (void)indicatePessimisticFixpoint();
13233 }
13234
13235 void trackStatistics() const override {
13236 STATS_DECLTRACK_CSARG_ATTR(addrspace);
13237 }
13238};
13239} // namespace
13240
13241/// ------------------------ No Alias Address Space ---------------------------
13242// This attribute assumes flat address space can alias all other address space
13243
13244// TODO: this is similar to AAAddressSpace, most of the code should be merged.
13245// But merging it created failing cased on gateway test that cannot be
13246// reproduced locally. So should open a separated PR to handle the merge of
13247// AANoAliasAddrSpace and AAAddressSpace attribute
13248
13249namespace {
13250struct AANoAliasAddrSpaceImpl : public AANoAliasAddrSpace {
13251 AANoAliasAddrSpaceImpl(const IRPosition &IRP, Attributor &A)
13252 : AANoAliasAddrSpace(IRP, A) {}
13253
13254 void initialize(Attributor &A) override {
13255 assert(getAssociatedType()->isPtrOrPtrVectorTy() &&
13256 "Associated value is not a pointer");
13257
13258 resetASRanges(A);
13259
13260 std::optional<unsigned> FlatAS = A.getInfoCache().getFlatAddressSpace();
13261 if (!FlatAS.has_value()) {
13262 indicatePessimisticFixpoint();
13263 return;
13264 }
13265
13266 removeAS(*FlatAS);
13267
13268 unsigned AS = getAssociatedType()->getPointerAddressSpace();
13269 if (AS != *FlatAS) {
13270 removeAS(AS);
13271 indicateOptimisticFixpoint();
13272 }
13273 }
13274
13275 ChangeStatus updateImpl(Attributor &A) override {
13276 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
13277 uint32_t OldAssumed = getAssumed();
13278
13279 auto CheckAddressSpace = [&](Value &Obj) {
13280 if (isa<PoisonValue>(&Obj))
13281 return true;
13282
13283 unsigned AS = Obj.getType()->getPointerAddressSpace();
13284 if (AS == FlatAS)
13285 return false;
13286
13287 removeAS(Obj.getType()->getPointerAddressSpace());
13288 return true;
13289 };
13290
13291 const AAUnderlyingObjects *AUO = A.getOrCreateAAFor<AAUnderlyingObjects>(
13292 getIRPosition(), this, DepClassTy::REQUIRED);
13293 if (!AUO->forallUnderlyingObjects(CheckAddressSpace))
13294 return indicatePessimisticFixpoint();
13295
13296 return OldAssumed == getAssumed() ? ChangeStatus::UNCHANGED
13297 : ChangeStatus::CHANGED;
13298 }
13299
13300 /// See AbstractAttribute::manifest(...).
13301 ChangeStatus manifest(Attributor &A) override {
13302 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
13303
13304 unsigned AS = getAssociatedType()->getPointerAddressSpace();
13305 if (AS != FlatAS || Map.empty())
13306 return ChangeStatus::UNCHANGED;
13307
13308 LLVMContext &Ctx = getAssociatedValue().getContext();
13309 MDNode *NoAliasASNode = nullptr;
13310 MDBuilder MDB(Ctx);
13311 // Has to use iterator to get the range info.
13312 for (RangeMap::const_iterator I = Map.begin(); I != Map.end(); I++) {
13313 if (!I.value())
13314 continue;
13315 unsigned Upper = I.stop();
13316 unsigned Lower = I.start();
13317 if (!NoAliasASNode) {
13318 NoAliasASNode = MDB.createRange(APInt(32, Lower), APInt(32, Upper + 1));
13319 continue;
13320 }
13321 MDNode *ASRange = MDB.createRange(APInt(32, Lower), APInt(32, Upper + 1));
13322 NoAliasASNode = MDNode::getMostGenericRange(NoAliasASNode, ASRange);
13323 }
13324
13325 Value *AssociatedValue = &getAssociatedValue();
13326 bool Changed = false;
13327
13328 auto AddNoAliasAttr = [&](const Use &U, bool &) {
13329 if (U.get() != AssociatedValue)
13330 return true;
13331 Instruction *Inst = dyn_cast<Instruction>(U.getUser());
13332 if (!Inst || Inst->hasMetadata(LLVMContext::MD_noalias_addrspace))
13333 return true;
13334 if (!isa<LoadInst>(Inst) && !isa<StoreInst>(Inst) &&
13336 return true;
13337 if (!A.isRunOn(Inst->getFunction()))
13338 return true;
13339 Inst->setMetadata(LLVMContext::MD_noalias_addrspace, NoAliasASNode);
13340 Changed = true;
13341 return true;
13342 };
13343 (void)A.checkForAllUses(AddNoAliasAttr, *this, *AssociatedValue,
13344 /*CheckBBLivenessOnly=*/true);
13345 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
13346 }
13347
13348 /// See AbstractAttribute::getAsStr().
13349 const std::string getAsStr(Attributor *A) const override {
13350 if (!isValidState())
13351 return "<invalid>";
13352 std::string Str;
13353 raw_string_ostream OS(Str);
13354 OS << "CanNotBeAddrSpace(";
13355 for (RangeMap::const_iterator I = Map.begin(); I != Map.end(); I++) {
13356 unsigned Upper = I.stop();
13357 unsigned Lower = I.start();
13358 OS << ' ' << '[' << Upper << ',' << Lower + 1 << ')';
13359 }
13360 OS << " )";
13361 return OS.str();
13362 }
13363
13364private:
13365 void removeAS(unsigned AS) {
13366 RangeMap::iterator I = Map.find(AS);
13367
13368 if (I != Map.end()) {
13369 unsigned Upper = I.stop();
13370 unsigned Lower = I.start();
13371 I.erase();
13372 if (Upper == Lower)
13373 return;
13374 if (AS != ~((unsigned)0) && AS + 1 <= Upper)
13375 Map.insert(AS + 1, Upper, /*what ever this variable name is=*/true);
13376 if (AS != 0 && Lower <= AS - 1)
13377 Map.insert(Lower, AS - 1, true);
13378 }
13379 }
13380
13381 void resetASRanges(Attributor &A) {
13382 Map.clear();
13383 Map.insert(0, A.getInfoCache().getMaxAddrSpace(), true);
13384 }
13385};
13386
13387struct AANoAliasAddrSpaceFloating final : AANoAliasAddrSpaceImpl {
13388 AANoAliasAddrSpaceFloating(const IRPosition &IRP, Attributor &A)
13389 : AANoAliasAddrSpaceImpl(IRP, A) {}
13390
13391 void trackStatistics() const override {
13392 STATS_DECLTRACK_FLOATING_ATTR(noaliasaddrspace);
13393 }
13394};
13395
13396struct AANoAliasAddrSpaceReturned final : AANoAliasAddrSpaceImpl {
13397 AANoAliasAddrSpaceReturned(const IRPosition &IRP, Attributor &A)
13398 : AANoAliasAddrSpaceImpl(IRP, A) {}
13399
13400 void trackStatistics() const override {
13401 STATS_DECLTRACK_FNRET_ATTR(noaliasaddrspace);
13402 }
13403};
13404
13405struct AANoAliasAddrSpaceCallSiteReturned final : AANoAliasAddrSpaceImpl {
13406 AANoAliasAddrSpaceCallSiteReturned(const IRPosition &IRP, Attributor &A)
13407 : AANoAliasAddrSpaceImpl(IRP, A) {}
13408
13409 void trackStatistics() const override {
13410 STATS_DECLTRACK_CSRET_ATTR(noaliasaddrspace);
13411 }
13412};
13413
13414struct AANoAliasAddrSpaceArgument final : AANoAliasAddrSpaceImpl {
13415 AANoAliasAddrSpaceArgument(const IRPosition &IRP, Attributor &A)
13416 : AANoAliasAddrSpaceImpl(IRP, A) {}
13417
13418 void trackStatistics() const override {
13419 STATS_DECLTRACK_ARG_ATTR(noaliasaddrspace);
13420 }
13421};
13422
13423struct AANoAliasAddrSpaceCallSiteArgument final : AANoAliasAddrSpaceImpl {
13424 AANoAliasAddrSpaceCallSiteArgument(const IRPosition &IRP, Attributor &A)
13425 : AANoAliasAddrSpaceImpl(IRP, A) {}
13426
13427 void trackStatistics() const override {
13428 STATS_DECLTRACK_CSARG_ATTR(noaliasaddrspace);
13429 }
13430};
13431} // namespace
13432/// ----------- Allocation Info ----------
13433namespace {
13434struct AAAllocationInfoImpl : public AAAllocationInfo {
13435 AAAllocationInfoImpl(const IRPosition &IRP, Attributor &A)
13436 : AAAllocationInfo(IRP, A) {}
13437
13438 std::optional<TypeSize> getAllocatedSize() const override {
13439 assert(isValidState() && "the AA is invalid");
13440 return AssumedAllocatedSize;
13441 }
13442
13443 std::optional<TypeSize> findInitialAllocationSize(Instruction *I,
13444 const DataLayout &DL) {
13445
13446 // TODO: implement case for malloc like instructions
13447 switch (I->getOpcode()) {
13448 case Instruction::Alloca: {
13449 AllocaInst *AI = cast<AllocaInst>(I);
13450 return AI->getAllocationSize(DL);
13451 }
13452 default:
13453 return std::nullopt;
13454 }
13455 }
13456
13457 ChangeStatus updateImpl(Attributor &A) override {
13458
13459 const IRPosition &IRP = getIRPosition();
13460 Instruction *I = IRP.getCtxI();
13461
13462 // TODO: update check for malloc like calls
13463 if (!isa<AllocaInst>(I))
13464 return indicatePessimisticFixpoint();
13465
13466 bool IsKnownNoCapture;
13468 A, this, IRP, DepClassTy::OPTIONAL, IsKnownNoCapture))
13469 return indicatePessimisticFixpoint();
13470
13471 const AAPointerInfo *PI =
13472 A.getOrCreateAAFor<AAPointerInfo>(IRP, *this, DepClassTy::REQUIRED);
13473
13474 if (!PI)
13475 return indicatePessimisticFixpoint();
13476
13477 if (!PI->getState().isValidState() || PI->reachesReturn())
13478 return indicatePessimisticFixpoint();
13479
13480 const DataLayout &DL = A.getDataLayout();
13481 const auto AllocationSize = findInitialAllocationSize(I, DL);
13482
13483 // If allocation size is nullopt, we give up.
13484 if (!AllocationSize)
13485 return indicatePessimisticFixpoint();
13486
13487 // For zero sized allocations, we give up.
13488 // Since we can't reduce further
13489 if (*AllocationSize == 0)
13490 return indicatePessimisticFixpoint();
13491
13492 int64_t BinSize = PI->numOffsetBins();
13493
13494 // TODO: implement for multiple bins
13495 if (BinSize > 1)
13496 return indicatePessimisticFixpoint();
13497
13498 if (BinSize == 0) {
13499 auto NewAllocationSize = std::make_optional<TypeSize>(0, false);
13500 if (!changeAllocationSize(NewAllocationSize))
13501 return ChangeStatus::UNCHANGED;
13502 return ChangeStatus::CHANGED;
13503 }
13504
13505 // TODO: refactor this to be part of multiple bin case
13506 const auto &It = PI->begin();
13507
13508 // TODO: handle if Offset is not zero
13509 if (It->first.Offset != 0)
13510 return indicatePessimisticFixpoint();
13511
13512 uint64_t SizeOfBin = It->first.Offset + It->first.Size;
13513
13514 if (SizeOfBin >= *AllocationSize)
13515 return indicatePessimisticFixpoint();
13516
13517 auto NewAllocationSize = std::make_optional<TypeSize>(SizeOfBin * 8, false);
13518
13519 if (!changeAllocationSize(NewAllocationSize))
13520 return ChangeStatus::UNCHANGED;
13521
13522 return ChangeStatus::CHANGED;
13523 }
13524
13525 /// See AbstractAttribute::manifest(...).
13526 ChangeStatus manifest(Attributor &A) override {
13527
13528 assert(isValidState() &&
13529 "Manifest should only be called if the state is valid.");
13530
13531 Instruction *I = getIRPosition().getCtxI();
13532
13533 auto FixedAllocatedSizeInBits = getAllocatedSize()->getFixedValue();
13534
13535 unsigned long NumBytesToAllocate = (FixedAllocatedSizeInBits + 7) / 8;
13536
13537 switch (I->getOpcode()) {
13538 // TODO: add case for malloc like calls
13539 case Instruction::Alloca: {
13540
13541 AllocaInst *AI = cast<AllocaInst>(I);
13542
13543 Type *CharType = Type::getInt8Ty(I->getContext());
13544
13545 auto *NumBytesToValue =
13546 ConstantInt::get(I->getContext(), APInt(32, NumBytesToAllocate));
13547
13548 BasicBlock::iterator insertPt = AI->getIterator();
13549 insertPt = std::next(insertPt);
13550 AllocaInst *NewAllocaInst =
13551 new AllocaInst(CharType, AI->getAddressSpace(), NumBytesToValue,
13552 AI->getAlign(), AI->getName(), insertPt);
13553
13554 if (A.changeAfterManifest(IRPosition::inst(*AI), *NewAllocaInst))
13555 return ChangeStatus::CHANGED;
13556
13557 break;
13558 }
13559 default:
13560 break;
13561 }
13562
13563 return ChangeStatus::UNCHANGED;
13564 }
13565
13566 /// See AbstractAttribute::getAsStr().
13567 const std::string getAsStr(Attributor *A) const override {
13568 if (!isValidState())
13569 return "allocationinfo(<invalid>)";
13570 return "allocationinfo(" +
13571 (AssumedAllocatedSize == HasNoAllocationSize
13572 ? "none"
13573 : std::to_string(AssumedAllocatedSize->getFixedValue())) +
13574 ")";
13575 }
13576
13577private:
13578 std::optional<TypeSize> AssumedAllocatedSize = HasNoAllocationSize;
13579
13580 // Maintain the computed allocation size of the object.
13581 // Returns (bool) weather the size of the allocation was modified or not.
13582 bool changeAllocationSize(std::optional<TypeSize> Size) {
13583 if (AssumedAllocatedSize == HasNoAllocationSize ||
13584 AssumedAllocatedSize != Size) {
13585 AssumedAllocatedSize = Size;
13586 return true;
13587 }
13588 return false;
13589 }
13590};
13591
13592struct AAAllocationInfoFloating : AAAllocationInfoImpl {
13593 AAAllocationInfoFloating(const IRPosition &IRP, Attributor &A)
13594 : AAAllocationInfoImpl(IRP, A) {}
13595
13596 void trackStatistics() const override {
13597 STATS_DECLTRACK_FLOATING_ATTR(allocationinfo);
13598 }
13599};
13600
13601struct AAAllocationInfoReturned : AAAllocationInfoImpl {
13602 AAAllocationInfoReturned(const IRPosition &IRP, Attributor &A)
13603 : AAAllocationInfoImpl(IRP, A) {}
13604
13605 /// See AbstractAttribute::initialize(...).
13606 void initialize(Attributor &A) override {
13607 // TODO: we don't rewrite function argument for now because it will need to
13608 // rewrite the function signature and all call sites
13609 (void)indicatePessimisticFixpoint();
13610 }
13611
13612 void trackStatistics() const override {
13613 STATS_DECLTRACK_FNRET_ATTR(allocationinfo);
13614 }
13615};
13616
13617struct AAAllocationInfoCallSiteReturned : AAAllocationInfoImpl {
13618 AAAllocationInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
13619 : AAAllocationInfoImpl(IRP, A) {}
13620
13621 void trackStatistics() const override {
13622 STATS_DECLTRACK_CSRET_ATTR(allocationinfo);
13623 }
13624};
13625
13626struct AAAllocationInfoArgument : AAAllocationInfoImpl {
13627 AAAllocationInfoArgument(const IRPosition &IRP, Attributor &A)
13628 : AAAllocationInfoImpl(IRP, A) {}
13629
13630 void trackStatistics() const override {
13631 STATS_DECLTRACK_ARG_ATTR(allocationinfo);
13632 }
13633};
13634
13635struct AAAllocationInfoCallSiteArgument : AAAllocationInfoImpl {
13636 AAAllocationInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
13637 : AAAllocationInfoImpl(IRP, A) {}
13638
13639 /// See AbstractAttribute::initialize(...).
13640 void initialize(Attributor &A) override {
13641
13642 (void)indicatePessimisticFixpoint();
13643 }
13644
13645 void trackStatistics() const override {
13646 STATS_DECLTRACK_CSARG_ATTR(allocationinfo);
13647 }
13648};
13649} // namespace
13650
13651const char AANoUnwind::ID = 0;
13652const char AANoSync::ID = 0;
13653const char AANoFree::ID = 0;
13654const char AANonNull::ID = 0;
13655const char AAMustProgress::ID = 0;
13656const char AANoRecurse::ID = 0;
13657const char AANonConvergent::ID = 0;
13658const char AAWillReturn::ID = 0;
13659const char AAUndefinedBehavior::ID = 0;
13660const char AANoAlias::ID = 0;
13661const char AAIntraFnReachability::ID = 0;
13662const char AANoReturn::ID = 0;
13663const char AAIsDead::ID = 0;
13664const char AADereferenceable::ID = 0;
13665const char AAAlign::ID = 0;
13666const char AAInstanceInfo::ID = 0;
13667const char AANoCapture::ID = 0;
13668const char AAValueSimplify::ID = 0;
13669const char AAHeapToStack::ID = 0;
13670const char AAPrivatizablePtr::ID = 0;
13671const char AAMemoryBehavior::ID = 0;
13672const char AAMemoryLocation::ID = 0;
13673const char AAValueConstantRange::ID = 0;
13674const char AAPotentialConstantValues::ID = 0;
13675const char AAPotentialValues::ID = 0;
13676const char AANoUndef::ID = 0;
13677const char AANoFPClass::ID = 0;
13678const char AACallEdges::ID = 0;
13679const char AAInterFnReachability::ID = 0;
13680const char AAPointerInfo::ID = 0;
13681const char AAAssumptionInfo::ID = 0;
13682const char AAUnderlyingObjects::ID = 0;
13683const char AAInvariantLoadPointer::ID = 0;
13684const char AAAddressSpace::ID = 0;
13685const char AANoAliasAddrSpace::ID = 0;
13686const char AAAllocationInfo::ID = 0;
13687const char AAIndirectCallInfo::ID = 0;
13688const char AAGlobalValueInfo::ID = 0;
13689const char AADenormalFPMath::ID = 0;
13690
13691// Macro magic to create the static generator function for attributes that
13692// follow the naming scheme.
13693
13694#define SWITCH_PK_INV(CLASS, PK, POS_NAME) \
13695 case IRPosition::PK: \
13696 llvm_unreachable("Cannot create " #CLASS " for a " POS_NAME " position!");
13697
13698#define SWITCH_PK_CREATE(CLASS, IRP, PK, SUFFIX) \
13699 case IRPosition::PK: \
13700 AA = new (A.Allocator) CLASS##SUFFIX(IRP, A); \
13701 ++NumAAs; \
13702 break;
13703
13704#define CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13705 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13706 CLASS *AA = nullptr; \
13707 switch (IRP.getPositionKind()) { \
13708 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13709 SWITCH_PK_INV(CLASS, IRP_FLOAT, "floating") \
13710 SWITCH_PK_INV(CLASS, IRP_ARGUMENT, "argument") \
13711 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13712 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_RETURNED, "call site returned") \
13713 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_ARGUMENT, "call site argument") \
13714 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13715 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13716 } \
13717 return *AA; \
13718 }
13719
13720#define CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13721 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13722 CLASS *AA = nullptr; \
13723 switch (IRP.getPositionKind()) { \
13724 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13725 SWITCH_PK_INV(CLASS, IRP_FUNCTION, "function") \
13726 SWITCH_PK_INV(CLASS, IRP_CALL_SITE, "call site") \
13727 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13728 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13729 SWITCH_PK_CREATE(CLASS, IRP, IRP_RETURNED, Returned) \
13730 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13731 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13732 } \
13733 return *AA; \
13734 }
13735
13736#define CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION(POS, SUFFIX, CLASS) \
13737 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13738 CLASS *AA = nullptr; \
13739 switch (IRP.getPositionKind()) { \
13740 SWITCH_PK_CREATE(CLASS, IRP, POS, SUFFIX) \
13741 default: \
13742 llvm_unreachable("Cannot create " #CLASS " for position otherthan " #POS \
13743 " position!"); \
13744 } \
13745 return *AA; \
13746 }
13747
13748#define CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13749 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13750 CLASS *AA = nullptr; \
13751 switch (IRP.getPositionKind()) { \
13752 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13753 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13754 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13755 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13756 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13757 SWITCH_PK_CREATE(CLASS, IRP, IRP_RETURNED, Returned) \
13758 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13759 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13760 } \
13761 return *AA; \
13762 }
13763
13764#define CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13765 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13766 CLASS *AA = nullptr; \
13767 switch (IRP.getPositionKind()) { \
13768 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13769 SWITCH_PK_INV(CLASS, IRP_ARGUMENT, "argument") \
13770 SWITCH_PK_INV(CLASS, IRP_FLOAT, "floating") \
13771 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13772 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_RETURNED, "call site returned") \
13773 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_ARGUMENT, "call site argument") \
13774 SWITCH_PK_INV(CLASS, IRP_CALL_SITE, "call site") \
13775 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13776 } \
13777 return *AA; \
13778 }
13779
13780#define CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13781 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13782 CLASS *AA = nullptr; \
13783 switch (IRP.getPositionKind()) { \
13784 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13785 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13786 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13787 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13788 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13789 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13790 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13791 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13792 } \
13793 return *AA; \
13794 }
13795
13805
13823
13828
13833
13840
13842
13843#undef CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION
13844#undef CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION
13845#undef CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION
13846#undef CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION
13847#undef CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION
13848#undef CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION
13849#undef SWITCH_PK_CREATE
13850#undef SWITCH_PK_INV
#define Success
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
aarch64 promote const
AMDGPU Register Bank Select
Rewrite undef for PHI
This file implements a class to represent arbitrary precision integral constant values and operations...
ReachingDefInfo InstSet & ToRemove
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis false
This file contains the simple types necessary to represent the attributes associated with functions a...
#define STATS_DECLTRACK(NAME, TYPE, MSG)
static std::optional< Constant * > askForAssumedConstant(Attributor &A, const AbstractAttribute &QueryingAA, const IRPosition &IRP, Type &Ty)
static cl::opt< unsigned, true > MaxPotentialValues("attributor-max-potential-values", cl::Hidden, cl::desc("Maximum number of potential values to be " "tracked for each position."), cl::location(llvm::PotentialConstantIntValuesState::MaxPotentialValues), cl::init(7))
static void clampReturnedValueStates(Attributor &A, const AAType &QueryingAA, StateType &S, const IRPosition::CallBaseContext *CBContext=nullptr)
Clamp the information known for all returned values of a function (identified by QueryingAA) into S.
#define STATS_DECLTRACK_FN_ATTR(NAME)
#define CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static cl::opt< int > MaxPotentialValuesIterations("attributor-max-potential-values-iterations", cl::Hidden, cl::desc("Maximum number of iterations we keep dismantling potential values."), cl::init(64))
#define STATS_DECLTRACK_CS_ATTR(NAME)
#define PIPE_OPERATOR(CLASS)
#define DefineKeys(ToTy)
static bool mayBeInCycle(const CycleInfo *CI, const Instruction *I, bool HeaderOnly, Cycle **CPtr=nullptr)
#define STATS_DECLTRACK_ARG_ATTR(NAME)
static const Value * stripAndAccumulateOffsets(Attributor &A, const AbstractAttribute &QueryingAA, const Value *Val, const DataLayout &DL, APInt &Offset, bool GetMinOffset, bool AllowNonInbounds, bool UseAssumed=false)
#define STATS_DECLTRACK_CSRET_ATTR(NAME)
static cl::opt< bool > ManifestInternal("attributor-manifest-internal", cl::Hidden, cl::desc("Manifest Attributor internal string attributes."), cl::init(false))
static Value * constructPointer(Value *Ptr, int64_t Offset, IRBuilder< NoFolder > &IRB)
Helper function to create a pointer based on Ptr, and advanced by Offset bytes.
#define CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
#define BUILD_STAT_NAME(NAME, TYPE)
static bool isDenselyPacked(Type *Ty, const DataLayout &DL)
Checks if a type could have padding bytes.
#define CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static const Value * getMinimalBaseOfPointer(Attributor &A, const AbstractAttribute &QueryingAA, const Value *Ptr, int64_t &BytesOffset, const DataLayout &DL, bool AllowNonInbounds=false)
#define STATS_DECLTRACK_FNRET_ATTR(NAME)
#define STATS_DECLTRACK_CSARG_ATTR(NAME)
#define CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION(POS, SUFFIX, CLASS)
#define CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static cl::opt< int > MaxHeapToStackSize("max-heap-to-stack-size", cl::init(128), cl::Hidden)
#define CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
#define STATS_DECLTRACK_FLOATING_ATTR(NAME)
#define STATS_DECL(NAME, TYPE, MSG)
basic Basic Alias true
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static bool isReachableImpl(SmallVectorImpl< BasicBlock * > &Worklist, const StopSetT &StopSet, const SmallPtrSetImpl< BasicBlock * > *ExclusionSet, const DominatorTree *DT, const LoopInfo *LI)
Definition CFG.cpp:135
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file declares an analysis pass that computes CycleInfo for LLVM IR, specialized from GenericCycl...
DXIL Forward Handle Accesses
DXIL Resource Access
dxil translate DXIL Translate Metadata
This file defines DenseMapInfo traits for DenseMap.
#define Check(C,...)
Hexagon Common GEP
IRTranslator LLVM IR MI
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
Machine Check Debug Module
This file implements a map that provides insertion order iteration.
#define T
#define T1
static unsigned getAddressSpace(const Value *V, unsigned MaxLookup)
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
uint64_t IntrinsicInst * II
if(PassOpts->AAPipeline)
static StringRef getName(Value *V)
Basic Register Allocator
dot regions Print regions of function to dot true view regions View regions of function(with no function bodies)"
Remove Loads Into Fake Uses
This builds on the llvm/ADT/GraphTraits.h file to find the strongly connected components (SCCs) of a ...
bool IsDead
std::pair< BasicBlock *, BasicBlock * > Edge
This file contains some templates that are useful if you are working with the STL at all.
BaseType
A given derived pointer can have multiple base pointers through phi/selects.
This file defines generic set operations that may be used on set's of different types,...
This file implements a set that has insertion order iteration characteristics.
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition Statistic.h:171
This file contains some functions that are useful when dealing with strings.
#define LLVM_DEBUG(...)
Definition Debug.h:114
static SymbolRef::Type getType(const Symbol *Sym)
Definition TapiFile.cpp:39
static void initialize(TargetLibraryInfoImpl &TLI, const Triple &T, const llvm::StringTable &StandardNames, VectorLibrary VecLib)
Initialize the set of available library functions based on the specified target triple.
This pass exposes codegen information to IR-level passes.
static unsigned getBitWidth(Type *Ty, const DataLayout &DL)
Returns the bitwidth of the given scalar or pointer type.
Value * RHS
Value * LHS
static unsigned getSize(unsigned Kind)
LLVM_ABI AACallGraphNode * operator*() const
bool isNoAlias(const MemoryLocation &LocA, const MemoryLocation &LocB)
A trivial helper function to check to see if the specified pointers are no-alias.
Class for arbitrary precision integers.
Definition APInt.h:78
int64_t getSExtValue() const
Get sign extended value.
Definition APInt.h:1577
CallBase * getInstruction() const
Return the underlying instruction.
bool isCallbackCall() const
Return true if this ACS represents a callback call.
bool isDirectCall() const
Return true if this ACS represents a direct call.
static LLVM_ABI void getCallbackUses(const CallBase &CB, SmallVectorImpl< const Use * > &CallbackUses)
Add operand uses of CB that represent callback uses into CallbackUses.
int getCallArgOperandNo(Argument &Arg) const
Return the operand index of the underlying instruction associated with Arg.
Align getAlign() const
Return the alignment of the memory that is being allocated by the instruction.
unsigned getAddressSpace() const
Return the address space for the allocation.
LLVM_ABI std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
This class represents an incoming formal argument to a Function.
Definition Argument.h:32
LLVM_ABI bool hasNoAliasAttr() const
Return true if this argument has the noalias attribute.
Definition Function.cpp:272
LLVM_ABI bool onlyReadsMemory() const
Return true if this argument has the readonly or readnone attribute.
Definition Function.cpp:308
LLVM_ABI bool hasPointeeInMemoryValueAttr() const
Return true if this argument has the byval, sret, inalloca, preallocated, or byref attribute.
Definition Function.cpp:171
LLVM_ABI bool hasReturnedAttr() const
Return true if this argument has the returned attribute.
Definition Function.cpp:296
LLVM_ABI bool hasByValAttr() const
Return true if this argument has the byval attribute.
Definition Function.cpp:128
const Function * getParent() const
Definition Argument.h:44
unsigned getArgNo() const
Return the index of this formal argument in its containing function.
Definition Argument.h:50
A function analysis which provides an AssumptionCache.
A cache of @llvm.assume calls within a function.
Functions, function parameters, and return types can have attributes to indicate how they should be t...
Definition Attributes.h:105
static LLVM_ABI Attribute get(LLVMContext &Context, AttrKind Kind, uint64_t Val=0)
Return a uniquified Attribute object.
LLVM_ABI FPClassTest getNoFPClass() const
Return the FPClassTest for nofpclass.
LLVM_ABI Attribute::AttrKind getKindAsEnum() const
Return the attribute's kind as an enum (Attribute::AttrKind).
LLVM_ABI MemoryEffects getMemoryEffects() const
Returns memory effects.
AttrKind
This enumeration lists the attributes that can be associated with parameters, function results,...
Definition Attributes.h:124
static LLVM_ABI Attribute getWithCaptureInfo(LLVMContext &Context, CaptureInfo CI)
static bool isEnumAttrKind(AttrKind Kind)
Definition Attributes.h:137
LLVM_ABI CaptureInfo getCaptureInfo() const
Returns information from captures attribute.
LLVM Basic Block Representation.
Definition BasicBlock.h:62
LLVM_ABI const_iterator getFirstInsertionPt() const
Returns an iterator to the first instruction in this block that is suitable for inserting a non-PHI i...
const Function * getParent() const
Return the enclosing method, or null if none.
Definition BasicBlock.h:213
static BasicBlock * Create(LLVMContext &Context, const Twine &Name="", Function *Parent=nullptr, BasicBlock *InsertBefore=nullptr)
Creates a new BasicBlock.
Definition BasicBlock.h:206
const Instruction & front() const
Definition BasicBlock.h:493
InstListType::iterator iterator
Instruction iterators...
Definition BasicBlock.h:170
const Instruction * getTerminator() const LLVM_READONLY
Returns the terminator instruction if the block is well formed or null if the block is not well forme...
Definition BasicBlock.h:233
BinaryOps getOpcode() const
Definition InstrTypes.h:374
Conditional or Unconditional Branch instruction.
unsigned getNumSuccessors() const
static BranchInst * Create(BasicBlock *IfTrue, InsertPosition InsertBefore=nullptr)
BasicBlock * getSuccessor(unsigned i) const
Value * getCondition() const
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
Function * getCalledFunction() const
Returns the function called, or null if this is an indirect function invocation or the function signa...
LLVM_ABI bool isMustTailCall() const
Tests if this call site must be tail call optimized.
LLVM_ABI bool isIndirectCall() const
Return true if the callsite is an indirect call.
bool isCallee(Value::const_user_iterator UI) const
Determine whether the passed iterator points to the callee operand's Use.
Value * getCalledOperand() const
const Use & getCalledOperandUse() const
const Use & getArgOperandUse(unsigned i) const
Wrappers for getting the Use of a call argument.
LLVM_ABI std::optional< ConstantRange > getRange() const
If this return value has a range attribute, return the value range of the argument.
Value * getArgOperand(unsigned i) const
bool isBundleOperand(unsigned Idx) const
Return true if the operand at index Idx is a bundle operand.
bool isConvergent() const
Determine if the invoke is convergent.
FunctionType * getFunctionType() const
LLVM_ABI Intrinsic::ID getIntrinsicID() const
Returns the intrinsic ID of the intrinsic called or Intrinsic::not_intrinsic if the called function i...
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
unsigned getArgOperandNo(const Use *U) const
Given a use for a arg operand, get the arg operand number that corresponds to it.
unsigned arg_size() const
bool isArgOperand(const Use *U) const
LLVM_ABI Function * getCaller()
Helper to get the caller (the parent function).
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
static CaptureInfo none()
Create CaptureInfo that does not capture any components of the pointer.
Definition ModRef.h:384
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition InstrTypes.h:610
LLVM_ABI bool isIntegerCast() const
There are several places where we need to know if a cast instruction only deals with integer source a...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition InstrTypes.h:617
bool isEquality() const
Determine if this is an equals/not equals predicate.
Definition InstrTypes.h:915
bool isFalseWhenEqual() const
This is just a convenience.
Definition InstrTypes.h:948
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ ICMP_NE
not equal
Definition InstrTypes.h:698
bool isTrueWhenEqual() const
This is just a convenience.
Definition InstrTypes.h:942
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:765
static ConstantAsMetadata * get(Constant *C)
Definition Metadata.h:537
static LLVM_ABI Constant * getExtractElement(Constant *Vec, Constant *Idx, Type *OnlyIfReducedTy=nullptr)
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
This class represents a range of values.
const APInt & getLower() const
Return the lower value for this range.
LLVM_ABI bool isFullSet() const
Return true if this set contains all of the elements possible for this data-type.
LLVM_ABI bool isEmptySet() const
Return true if this set contains no members.
bool isSingleElement() const
Return true if this set contains exactly one member.
static LLVM_ABI ConstantRange makeAllowedICmpRegion(CmpInst::Predicate Pred, const ConstantRange &Other)
Produce the smallest range such that all values that may satisfy the given predicate with any value c...
const APInt & getUpper() const
Return the upper value for this range.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
iterator find(const_arg_type_t< KeyT > Val)
Definition DenseMap.h:178
std::pair< iterator, bool > try_emplace(KeyT &&Key, Ts &&...Args)
Definition DenseMap.h:256
size_type count(const_arg_type_t< KeyT > Val) const
Return 1 if the specified key is in the map, 0 otherwise.
Definition DenseMap.h:174
iterator end()
Definition DenseMap.h:81
bool contains(const_arg_type_t< KeyT > Val) const
Return true if the specified key is in the map, false otherwise.
Definition DenseMap.h:169
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition DenseMap.h:241
Analysis pass which computes a DominatorTree.
Definition Dominators.h:283
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition Dominators.h:164
LLVM_ABI bool dominates(const BasicBlock *BB, const Use &U) const
Return true if the (end of the) basic block BB dominates the use U.
const BasicBlock & getEntryBlock() const
Definition Function.h:809
Argument * arg_iterator
Definition Function.h:73
iterator_range< arg_iterator > args()
Definition Function.h:892
const Function & getFunction() const
Definition Function.h:166
size_t arg_size() const
Definition Function.h:901
Argument * getArg(unsigned i) const
Definition Function.h:886
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
Definition Function.cpp:729
CycleT * getCycle(const BlockT *Block) const
Find the innermost cycle containing a given block.
LLVM_ABI bool isDeclaration() const
Return true if the primary definition of this global value is outside of the current translation unit...
Definition Globals.cpp:329
bool hasLocalLinkage() const
static LLVM_ABI bool compare(const APInt &LHS, const APInt &RHS, ICmpInst::Predicate Pred)
Return result of LHS Pred RHS comparison.
Value * CreatePtrAdd(Value *Ptr, Value *Offset, const Twine &Name="", GEPNoWrapFlags NW=GEPNoWrapFlags::none())
Definition IRBuilder.h:2025
ConstantInt * getInt64(uint64_t C)
Get a constant 64-bit value.
Definition IRBuilder.h:527
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
Definition IRBuilder.h:2788
LLVM_ABI Instruction * clone() const
Create a copy of 'this' instruction that is identical in all ways except the following:
LLVM_ABI bool isLifetimeStartOrEnd() const LLVM_READONLY
Return true if the instruction is a llvm.lifetime.start or llvm.lifetime.end marker.
bool mayReadOrWriteMemory() const
Return true if this instruction may read or write memory.
LLVM_ABI bool mayWriteToMemory() const LLVM_READONLY
Return true if this instruction may modify memory.
bool hasMetadata() const
Return true if this instruction has any metadata attached to it.
LLVM_ABI void insertBefore(InstListType::iterator InsertPos)
Insert an unlinked instruction into a basic block immediately before the specified position.
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
LLVM_ABI const Function * getFunction() const
Return the function this instruction belongs to.
LLVM_ABI BasicBlock * getSuccessor(unsigned Idx) const LLVM_READONLY
Return the specified successor. This instruction must be a terminator.
LLVM_ABI bool mayHaveSideEffects() const LLVM_READONLY
Return true if the instruction may have side effects.
bool isTerminator() const
LLVM_ABI bool mayReadFromMemory() const LLVM_READONLY
Return true if this instruction may read memory.
LLVM_ABI void setMetadata(unsigned KindID, MDNode *Node)
Set the metadata of the specified kind to the specified node.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
void setDebugLoc(DebugLoc Loc)
Set the debug location information for this instruction.
LLVM_ABI const DataLayout & getDataLayout() const
Get the data layout of the module this instruction belongs to.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
ConstantRange getConstantRange(Value *V, Instruction *CxtI, bool UndefAllowed)
Return the ConstantRange constraint that is known to hold for the specified value at the specified in...
LoopT * getLoopFor(const BlockT *BB) const
Return the inner most loop that BB lives in.
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
Definition Metadata.h:1572
static LLVM_ABI MDNode * getMostGenericRange(MDNode *A, MDNode *B)
bool empty() const
Definition MapVector.h:77
static MemoryEffectsBase readOnly()
Definition ModRef.h:133
bool doesNotAccessMemory() const
Whether this function accesses no memory.
Definition ModRef.h:223
static MemoryEffectsBase argMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Definition ModRef.h:143
static MemoryEffectsBase inaccessibleMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Definition ModRef.h:149
bool onlyAccessesInaccessibleMem() const
Whether this function only (at most) accesses inaccessible memory.
Definition ModRef.h:242
ModRefInfo getModRef(Location Loc) const
Get ModRefInfo for the given Location.
Definition ModRef.h:196
bool onlyAccessesArgPointees() const
Whether this function only (at most) accesses argument memory.
Definition ModRef.h:232
bool onlyReadsMemory() const
Whether this function only (at most) reads memory.
Definition ModRef.h:226
static MemoryEffectsBase writeOnly()
Definition ModRef.h:138
static MemoryEffectsBase inaccessibleOrArgMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Definition ModRef.h:166
static MemoryEffectsBase none()
Definition ModRef.h:128
bool onlyAccessesInaccessibleOrArgMem() const
Whether this function only (at most) accesses argument and inaccessible memory.
Definition ModRef.h:262
static MemoryEffectsBase unknown()
Definition ModRef.h:123
static LLVM_ABI std::optional< MemoryLocation > getOrNone(const Instruction *Inst)
static SizeOffsetValue unknown()
static PHINode * Create(Type *Ty, unsigned NumReservedValues, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
Constructors - NumReservedValues is a hint for the number of incoming edges that this phi node will h...
static LLVM_ABI PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
Value * getReturnValue() const
Convenience accessor. Returns null if there is no return value.
LLVM_ABI const SCEV * getSCEVAtScope(const SCEV *S, const Loop *L)
Return a SCEV expression for the specified value at the specified scope in the program.
LLVM_ABI const SCEV * getSCEV(Value *V)
Return a SCEV expression for the full generality of the specified expression.
LLVM_ABI unsigned getSmallConstantMaxTripCount(const Loop *L, SmallVectorImpl< const SCEVPredicate * > *Predicates=nullptr)
Returns the upper bound of the loop trip count as a normal unsigned value.
ConstantRange getUnsignedRange(const SCEV *S)
Determine the unsigned range for a particular SCEV.
A vector that has set insertion semantics.
Definition SetVector.h:57
size_type size() const
Determine the number of elements in the SetVector.
Definition SetVector.h:103
bool insert(const value_type &X)
Insert a new element into the SetVector.
Definition SetVector.h:151
bool erase(PtrType Ptr)
Remove pointer from the set.
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
std::pair< const_iterator, bool > insert(const T &V)
insert - Insert an element into the set if it isn't already there.
Definition SmallSet.h:184
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Used to lazily calculate structure layout information for a target machine, based on the DataLayout s...
Definition DataLayout.h:723
TypeSize getElementOffset(unsigned Idx) const
Definition DataLayout.h:754
TypeSize getElementOffsetInBits(unsigned Idx) const
Definition DataLayout.h:759
Class to represent struct types.
unsigned getNumElements() const
Random access to the elements.
Type * getElementType(unsigned N) const
LLVM_ABI bool areTypesABICompatible(const Function *Caller, const Function *Callee, ArrayRef< Type * > Types) const
LLVM_ABI unsigned getAssumedAddrSpace(const Value *V) const
LLVM_ABI bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const
Return true if the given instruction (assumed to be a memory access instruction) has a volatile varia...
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
LLVM_ABI unsigned getIntegerBitWidth() const
bool isPointerTy() const
True if this is an instance of PointerType.
Definition Type.h:267
LLVM_ABI unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
bool isPtrOrPtrVectorTy() const
Return true if this is a pointer type or a vector of pointer types.
Definition Type.h:270
bool isIntegerTy() const
True if this is an instance of IntegerType.
Definition Type.h:240
bool isVoidTy() const
Return true if this is 'void'.
Definition Type.h:139
static LLVM_ABI UndefValue * get(Type *T)
Static factory methods - Return an 'undef' object of the specified type.
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
User * getUser() const
Returns the User that contains this Use.
Definition Use.h:61
Value * get() const
Definition Use.h:55
const Use & getOperandUse(unsigned i) const
Definition User.h:220
LLVM_ABI bool isDroppable() const
A droppable user is a user for which uses can be dropped without affecting correctness and should be ...
Definition User.cpp:119
LLVM_ABI bool replaceUsesOfWith(Value *From, Value *To)
Replace uses of one Value with another.
Definition User.cpp:25
Value * getOperand(unsigned i) const
Definition User.h:207
unsigned getNumOperands() const
Definition User.h:229
ValueT lookup(const KeyT &Val) const
lookup - Return the entry for the specified key, or a default constructed value if no such entry exis...
Definition ValueMap.h:167
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
static constexpr uint64_t MaximumAlignment
Definition Value.h:832
LLVM_ABI void replaceAllUsesWith(Value *V)
Change all uses of this to point to a new Value.
Definition Value.cpp:553
LLVMContext & getContext() const
All values hold a context through their type.
Definition Value.h:259
iterator_range< user_iterator > users()
Definition Value.h:427
LLVM_ABI const Value * stripAndAccumulateConstantOffsets(const DataLayout &DL, APInt &Offset, bool AllowNonInbounds, bool AllowInvariantGroup=false, function_ref< bool(Value &Value, APInt &Offset)> ExternalAnalysis=nullptr, bool LookThroughIntToPtr=false) const
Accumulate the constant offset this value has compared to a base pointer.
bool use_empty() const
Definition Value.h:347
static constexpr unsigned MaxAlignmentExponent
The maximum alignment for instructions.
Definition Value.h:831
iterator_range< use_iterator > uses()
Definition Value.h:381
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:322
std::pair< iterator, bool > insert(const ValueT &V)
Definition DenseSet.h:202
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
const ParentTy * getParent() const
Definition ilist_node.h:34
self_iterator getIterator()
Definition ilist_node.h:123
NodeTy * getNextNode()
Get the next node, or nullptr for the list tail.
Definition ilist_node.h:348
This class implements an extremely fast bulk output stream that can only output to a stream.
Definition raw_ostream.h:53
A raw_ostream that writes to an std::string.
Changed
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
Abstract Attribute helper functions.
Definition Attributor.h:165
LLVM_ABI bool isAssumedReadNone(Attributor &A, const IRPosition &IRP, const AbstractAttribute &QueryingAA, bool &IsKnown)
Return true if IRP is readnone.
LLVM_ABI bool isAssumedReadOnly(Attributor &A, const IRPosition &IRP, const AbstractAttribute &QueryingAA, bool &IsKnown)
Return true if IRP is readonly.
raw_ostream & operator<<(raw_ostream &OS, const RangeTy &R)
Definition Attributor.h:323
LLVM_ABI std::optional< Value * > combineOptionalValuesInAAValueLatice(const std::optional< Value * > &A, const std::optional< Value * > &B, Type *Ty)
Return the combination of A and B such that the result is a possible value of both.
LLVM_ABI bool isValidAtPosition(const ValueAndContext &VAC, InformationCache &InfoCache)
Return true if the value of VAC is a valid at the position of VAC, that is a constant,...
LLVM_ABI bool isAssumedThreadLocalObject(Attributor &A, Value &Obj, const AbstractAttribute &QueryingAA)
Return true if Obj is assumed to be a thread local object.
LLVM_ABI bool isDynamicallyUnique(Attributor &A, const AbstractAttribute &QueryingAA, const Value &V, bool ForAnalysisOnly=true)
Return true if V is dynamically unique, that is, there are no two "instances" of V at runtime with di...
LLVM_ABI bool getPotentialCopiesOfStoredValue(Attributor &A, StoreInst &SI, SmallSetVector< Value *, 4 > &PotentialCopies, const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation, bool OnlyExact=false)
Collect all potential values of the one stored by SI into PotentialCopies.
SmallPtrSet< Instruction *, 4 > InstExclusionSetTy
Definition Attributor.h:166
LLVM_ABI bool isGPU(const Module &M)
Return true iff M target a GPU (and we can use GPU AS reasoning).
ValueScope
Flags to distinguish intra-procedural queries from potentially inter-procedural queries.
Definition Attributor.h:182
@ Intraprocedural
Definition Attributor.h:183
@ Interprocedural
Definition Attributor.h:184
LLVM_ABI bool isValidInScope(const Value &V, const Function *Scope)
Return true if V is a valid value in Scope, that is a constant or an instruction/argument of Scope.
LLVM_ABI bool isPotentiallyReachable(Attributor &A, const Instruction &FromI, const Instruction &ToI, const AbstractAttribute &QueryingAA, const AA::InstExclusionSetTy *ExclusionSet=nullptr, std::function< bool(const Function &F)> GoBackwardsCB=nullptr)
Return true if ToI is potentially reachable from FromI without running into any instruction in Exclus...
LLVM_ABI bool isNoSyncInst(Attributor &A, const Instruction &I, const AbstractAttribute &QueryingAA)
Return true if I is a nosync instruction.
bool hasAssumedIRAttr(Attributor &A, const AbstractAttribute *QueryingAA, const IRPosition &IRP, DepClassTy DepClass, bool &IsKnown, bool IgnoreSubsumingPositions=false, const AAType **AAPtr=nullptr)
Helper to avoid creating an AA for IR Attributes that might already be set.
LLVM_ABI bool getPotentiallyLoadedValues(Attributor &A, LoadInst &LI, SmallSetVector< Value *, 4 > &PotentialValues, SmallSetVector< Instruction *, 4 > &PotentialValueOrigins, const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation, bool OnlyExact=false)
Collect all potential values LI could read into PotentialValues.
LLVM_ABI Value * getWithType(Value &V, Type &Ty)
Try to convert V to type Ty without introducing new instructions.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr char Attrs[]
Key for Kernel::Metadata::mAttrs.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
@ BasicBlock
Various leaf nodes.
Definition ISDOpcodes.h:81
@ Unsupported
This operation is completely unsupported on the target.
Offsets
Offsets in bytes from the start of the input buffer.
@ SingleThread
Synchronized with respect to signal handlers executing in the same thread.
Definition LLVMContext.h:55
@ CE
Windows NT (Windows on ARM)
Definition MCAsmInfo.h:48
@ Valid
The data is already valid.
initializer< Ty > init(const Ty &Val)
LocationClass< Ty > location(Ty &L)
unsigned combineHashValue(unsigned a, unsigned b)
Simplistic combination of 32-bit hash values into 32-bit hash values.
ElementType
The element type of an SRV or UAV resource.
Definition DXILABI.h:68
std::enable_if_t< detail::IsValidPointer< X, Y >::value, X * > dyn_extract_or_null(Y &&MD)
Extract a Value from Metadata, if any, allowing null.
Definition Metadata.h:709
std::enable_if_t< detail::IsValidPointer< X, Y >::value, X * > extract(Y &&MD)
Extract a Value from Metadata.
Definition Metadata.h:668
constexpr double e
@ User
could "use" a pointer
NodeAddr< UseNode * > Use
Definition RDFGraph.h:385
Context & getContext() const
Definition BasicBlock.h:99
iterator end() const
Definition BasicBlock.h:89
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
LLVM_ABI iterator begin() const
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
@ Offset
Definition DWP.cpp:532
@ Length
Definition DWP.cpp:532
bool operator<(int64_t V1, const APSInt &V2)
Definition APSInt.h:360
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
LLVM_ATTRIBUTE_ALWAYS_INLINE DynamicAPInt gcd(const DynamicAPInt &A, const DynamicAPInt &B)
LLVM_ABI KnownFPClass computeKnownFPClass(const Value *V, const APInt &DemandedElts, FPClassTest InterestedClasses, const SimplifyQuery &SQ, unsigned Depth=0)
Determine which floating-point classes are valid for V, and return them in KnownFPClass bit sets.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1739
LLVM_ABI bool isLegalToPromote(const CallBase &CB, Function *Callee, const char **FailureReason=nullptr)
Return true if the given indirect call site can be made to call Callee.
LLVM_ABI Constant * getInitialValueOfAllocation(const Value *V, const TargetLibraryInfo *TLI, Type *Ty)
If this is a call to an allocation function that initializes memory to a fixed value,...
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
Definition STLExtras.h:1669
@ Undef
Value of the register doesn't matter.
auto pred_end(const MachineBasicBlock *BB)
unsigned getPointerAddressSpace(const Type *T)
Definition SPIRVUtils.h:370
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
FunctionAddr VTableAddr uintptr_t uintptr_t Int32Ty
Definition InstrProf.h:296
auto successors(const MachineBasicBlock *BB)
LLVM_ABI bool isRemovableAlloc(const CallBase *V, const TargetLibraryInfo *TLI)
Return true if this is a call to an allocation function that does not have side effects that we are r...
APFloat abs(APFloat X)
Returns the absolute value of the argument.
Definition APFloat.h:1626
LLVM_ABI raw_fd_ostream & outs()
This returns a reference to a raw_fd_ostream for standard output.
LLVM_ABI Value * getAllocAlignment(const CallBase *V, const TargetLibraryInfo *TLI)
Gets the alignment argument for an aligned_alloc-like function, using either built-in knowledge based...
auto dyn_cast_if_present(const Y &Val)
dyn_cast_if_present<X> - Functionally identical to dyn_cast, except that a null (or none in the case ...
Definition Casting.h:732
LLVM_ABI Value * simplifyInstructionWithOperands(Instruction *I, ArrayRef< Value * > NewOps, const SimplifyQuery &Q)
Like simplifyInstruction but the operands of I are replaced with NewOps.
Value * GetPointerBaseWithConstantOffset(Value *Ptr, int64_t &Offset, const DataLayout &DL, bool AllowNonInbounds=true)
Analyze the specified pointer to see if it can be expressed as a base pointer plus a constant offset.
scc_iterator< T > scc_begin(const T &G)
Construct the begin iterator for a deduced graph type T.
LLVM_ABI bool isNoAliasCall(const Value *V)
Return true if this pointer is returned by a noalias function.
MemoryEffectsBase< IRMemLocation > MemoryEffects
Summary of how a function affects memory in the program.
Definition ModRef.h:313
raw_ostream & WriteGraph(raw_ostream &O, const GraphType &G, bool ShortNames=false, const Twine &Title="")
LLVM_ABI bool isSafeToSpeculativelyExecute(const Instruction *I, const Instruction *CtxI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr, const TargetLibraryInfo *TLI=nullptr, bool UseVariableInfo=true, bool IgnoreUBImplyingAttrs=true)
Return true if the instruction does not have any effects besides calculating the result and does not ...
bool isa_and_nonnull(const Y &Val)
Definition Casting.h:676
bool operator==(const AddressRangeValuePair &LHS, const AddressRangeValuePair &RHS)
LLVM_ABI ConstantRange getConstantRangeFromMetadata(const MDNode &RangeMD)
Parse out a conservative ConstantRange from !range metadata.
auto map_range(ContainerTy &&C, FuncTy F)
Return a range that applies F to the elements of C.
Definition STLExtras.h:366
const Value * getPointerOperand(const Value *V)
A helper function that returns the pointer operand of a load, store or GEP instruction.
LLVM_ABI Value * simplifyInstruction(Instruction *I, const SimplifyQuery &Q)
See if we can compute a simplified version of this instruction.
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
CycleInfo::CycleT Cycle
Definition CycleInfo.h:24
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1746
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
LLVM_ABI bool isIntrinsicReturningPointerAliasingArgumentWithoutCapturing(const CallBase *Call, bool MustPreserveNullness)
{launder,strip}.invariant.group returns pointer that aliases its argument, and it only captures point...
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
PotentialValuesState< std::pair< AA::ValueAndContext, AA::ValueScope > > PotentialLLVMValuesState
void sort(IteratorTy Start, IteratorTy End)
Definition STLExtras.h:1636
LLVM_ABI bool NullPointerIsDefined(const Function *F, unsigned AS=0)
Check whether null pointer dereferencing is considered undefined behavior for a given function or an ...
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
bool isPointerTy(const Type *T)
Definition SPIRVUtils.h:364
LLVM_ABI bool wouldInstructionBeTriviallyDead(const Instruction *I, const TargetLibraryInfo *TLI=nullptr)
Return true if the result produced by the instruction would have no side effects if it was not used.
Definition Local.cpp:425
bool set_union(S1Ty &S1, const S2Ty &S2)
set_union(A, B) - Compute A := A u B, return whether A changed.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
LLVM_ABI CallBase & promoteCall(CallBase &CB, Function *Callee, CastInst **RetBitCast=nullptr)
Promote the given indirect call site to unconditionally call Callee.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool hasAssumption(const Function &F, const KnownAssumptionString &AssumptionStr)
Return true if F has the assumption AssumptionStr attached.
LLVM_ABI RetainedKnowledge getKnowledgeFromUse(const Use *U, ArrayRef< Attribute::AttrKind > AttrKinds)
Return a valid Knowledge associated to the Use U if its Attribute kind is in AttrKinds.
@ Success
The lock was released successfully.
LLVM_ATTRIBUTE_VISIBILITY_DEFAULT AnalysisKey InnerAnalysisManagerProxy< AnalysisManagerT, IRUnitT, ExtraArgTs... >::Key
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ Other
Any other memory.
Definition ModRef.h:68
PotentialValuesState< APInt > PotentialConstantIntValuesState
TargetTransformInfo TTI
std::string join(IteratorT Begin, IteratorT End, StringRef Separator)
Joins the strings in the range [Begin, End), adding Separator between the elements.
IRBuilder(LLVMContext &, FolderTy, InserterTy, MDNode *, ArrayRef< OperandBundleDef >) -> IRBuilder< FolderTy, InserterTy >
InterleavedRange< Range > interleaved_array(const Range &R, StringRef Separator=", ")
Output range R as an array of interleaved elements.
ChangeStatus clampStateAndIndicateChange< DerefState >(DerefState &S, const DerefState &R)
void RemapInstruction(Instruction *I, ValueToValueMapTy &VM, RemapFlags Flags=RF_None, ValueMapTypeRemapper *TypeMapper=nullptr, ValueMaterializer *Materializer=nullptr, const MetadataPredicate *IdentityMD=nullptr)
Convert the instruction operands from referencing the current values into those specified by VM.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
ArrayRef(const T &OneElt) -> ArrayRef< T >
LLVM_ABI Value * getFreedOperand(const CallBase *CB, const TargetLibraryInfo *TLI)
If this if a call to a free function, return the freed operand.
ChangeStatus clampStateAndIndicateChange(StateType &S, const StateType &R)
Helper function to clamp a state S of type StateType with the information in R and indicate/return if...
constexpr unsigned BitWidth
ValueMap< const Value *, WeakTrackingVH > ValueToValueMapTy
auto pred_begin(const MachineBasicBlock *BB)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
iterator_range< pointer_iterator< WrappedIteratorT > > make_pointer_range(RangeT &&Range)
Definition iterator.h:368
ChangeStatus
{
Definition Attributor.h:496
LLVM_ABI std::optional< APInt > getAllocSize(const CallBase *CB, const TargetLibraryInfo *TLI, function_ref< const Value *(const Value *)> Mapper=[](const Value *V) { return V;})
Return the size of the requested allocation.
LLVM_ABI DenseSet< StringRef > getAssumptions(const Function &F)
Return the set of all assumptions for the function F.
Align assumeAligned(uint64_t Value)
Treats the value 0 as a 1, so Align is always at least 1.
Definition Alignment.h:100
LLVM_ABI Instruction * SplitBlockAndInsertIfThen(Value *Cond, BasicBlock::iterator SplitBefore, bool Unreachable, MDNode *BranchWeights=nullptr, DomTreeUpdater *DTU=nullptr, LoopInfo *LI=nullptr, BasicBlock *ThenBlock=nullptr)
Split the containing block at the specified instruction - everything before SplitBefore stays in the ...
@ OPTIONAL
The target may be valid if the source is not.
Definition Attributor.h:508
@ NONE
Do not track a dependence between source and target.
Definition Attributor.h:509
@ REQUIRED
The target cannot be valid if the source is not.
Definition Attributor.h:507
LLVM_ABI UseCaptureInfo DetermineUseCaptureKind(const Use &U, const Value *Base)
Determine what kind of capture behaviour U may exhibit.
LLVM_ABI Value * simplifyCmpInst(CmpPredicate Predicate, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a CmpInst, fold the result or return null.
LLVM_ABI bool mayContainIrreducibleControl(const Function &F, const LoopInfo *LI)
BumpPtrAllocatorImpl<> BumpPtrAllocator
The standard BumpPtrAllocator which just uses the default template parameters.
Definition Allocator.h:383
T bit_floor(T Value)
Returns the largest integral power of two no greater than Value if Value is nonzero.
Definition bit.h:330
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
bool capturesNothing(CaptureComponents CC)
Definition ModRef.h:332
LLVM_ABI bool isIdentifiedObject(const Value *V)
Return true if this pointer refers to a distinct and identifiable object.
constexpr StringRef AssumptionAttrKey
The key we use for assumption attributes.
Definition Assumptions.h:29
constexpr bool isCallableCC(CallingConv::ID CC)
GenericCycleInfo< SSAContext > CycleInfo
Definition CycleInfo.h:23
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
A type to track pointer/struct usage and accesses for AAPointerInfo.
bool forallInterferingAccesses(AA::RangeTy Range, F CB) const
See AAPointerInfo::forallInterferingAccesses.
AAPointerInfo::const_bin_iterator end() const
ChangeStatus addAccess(Attributor &A, const AAPointerInfo::RangeList &Ranges, Instruction &I, std::optional< Value * > Content, AAPointerInfo::AccessKind Kind, Type *Ty, Instruction *RemoteI=nullptr)
Add a new Access to the state at offset Offset and with size Size.
DenseMap< const Instruction *, SmallVector< unsigned > > RemoteIMap
AAPointerInfo::const_bin_iterator begin() const
AAPointerInfo::OffsetInfo ReturnedOffsets
Flag to determine if the underlying pointer is reaching a return statement in the associated function...
State(State &&SIS)=default
const AAPointerInfo::Access & getAccess(unsigned Index) const
SmallVector< AAPointerInfo::Access > AccessList
bool isAtFixpoint() const override
See AbstractState::isAtFixpoint().
bool forallInterferingAccesses(Instruction &I, F CB, AA::RangeTy &Range) const
See AAPointerInfo::forallInterferingAccesses.
static State getWorstState(const State &SIS)
Return the worst possible representable state.
AAPointerInfo::OffsetBinsTy OffsetBins
ChangeStatus indicateOptimisticFixpoint() override
See AbstractState::indicateOptimisticFixpoint().
ChangeStatus indicatePessimisticFixpoint() override
See AbstractState::indicatePessimisticFixpoint().
static State getBestState(const State &SIS)
Return the best possible representable state.
bool isValidState() const override
See AbstractState::isValidState().
----------------—AAIntraFnReachability Attribute-----------------------—
ReachabilityQueryInfo(const ReachabilityQueryInfo &RQI)
unsigned Hash
Precomputed hash for this RQI.
const Instruction * From
Start here,.
Reachable Result
and remember if it worked:
ReachabilityQueryInfo(const Instruction *From, const ToTy *To)
ReachabilityQueryInfo(Attributor &A, const Instruction &From, const ToTy &To, const AA::InstExclusionSetTy *ES, bool MakeUnique)
Constructor replacement to ensure unique and stable sets are used for the cache.
const ToTy * To
reach this place,
const AA::InstExclusionSetTy * ExclusionSet
without going through any of these instructions,
An abstract interface for address space information.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for all align attributes.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
Align getKnownAlign() const
Return known alignment.
static LLVM_ABI const char ID
An abstract attribute for getting assumption information.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract state for querying live call edges.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract Attribute for specializing "dynamic" components of denormal_fpenv to a known denormal mod...
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for all dereferenceable attribute.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for llvm::GlobalValue information interference.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for indirect call information interference.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface to track if a value leaves it's defining function instance.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract Attribute for computing reachability between functions.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
bool canReach(Attributor &A, const Function &Fn) const
If the function represented by this possition can reach Fn.
virtual bool instructionCanReach(Attributor &A, const Instruction &Inst, const Function &Fn, const AA::InstExclusionSetTy *ExclusionSet=nullptr) const =0
Can Inst reach Fn.
An abstract interface to determine reachability of point A to B.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for identifying pointers from which loads can be marked invariant.
static LLVM_ABI const char ID
Unique ID (due to the unique address).
An abstract interface for liveness abstract attribute.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for memory access kind related attributes (readnone/readonly/writeonly).
bool isAssumedReadOnly() const
Return true if we assume that the underlying value is not accessed (=written) in its respective scope...
bool isKnownReadNone() const
Return true if we know that the underlying value is not read or accessed in its respective scope.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
bool isAssumedReadNone() const
Return true if we assume that the underlying value is not read or accessed in its respective scope.
An abstract interface for all memory location attributes (readnone/argmemonly/inaccessiblememonly/ina...
static LLVM_ABI std::string getMemoryLocationsAsStr(MemoryLocationsKind MLK)
Return the locations encoded by MLK as a readable string.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
StateType::base_t MemoryLocationsKind
An abstract interface for all nonnull attributes.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for potential address space information.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for all noalias attributes.
static LLVM_ABI bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for all nocapture attributes.
@ NO_CAPTURE_MAYBE_RETURNED
If we do not capture the value in memory or through integers we can only communicate it back as a der...
@ NO_CAPTURE
If we do not capture the value in memory, through integers, or as a derived pointer we know it is not...
static LLVM_ABI const char ID
Unique ID (due to the unique address)
bool isAssumedNoCaptureMaybeReturned() const
Return true if we assume that the underlying value is not captured in its respective scope but we all...
static LLVM_ABI bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static LLVM_ABI void determineFunctionCaptureCapabilities(const IRPosition &IRP, const Function &F, BitIntegerState &State)
Update State according to the capture capabilities of F for position IRP.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An AbstractAttribute for nofree.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract attribute for norecurse.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An AbstractAttribute for noreturn.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI bool isAlignedBarrier(const CallBase &CB, bool ExecutedAligned)
Helper function to determine if CB is an aligned (GPU) barrier.
static LLVM_ABI bool isNonRelaxedAtomic(const Instruction *I)
Helper function used to determine whether an instruction is non-relaxed atomic.
static LLVM_ABI bool isNoSyncIntrinsic(const Instruction *I)
Helper function specific for intrinsics which are potentially volatile.
An abstract interface for all noundef attributes.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract Attribute for determining the necessity of the convergent attribute.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for all nonnull attributes.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See AbstractAttribute::isImpliedByIR(...).
An access description.
A helper containing a list of offsets computed for a Use.
A container for a list of ranges.
static void set_difference(const RangeList &L, const RangeList &R, RangeList &D)
Copy ranges from L that are not in R, into D.
An abstract interface for struct information.
virtual bool reachesReturn() const =0
OffsetBinsTy::const_iterator const_bin_iterator
virtual const_bin_iterator begin() const =0
DenseMap< AA::RangeTy, SmallSet< unsigned, 4 > > OffsetBinsTy
static LLVM_ABI const char ID
Unique ID (due to the unique address)
virtual int64_t numOffsetBins() const =0
An abstract interface for potential values analysis.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
friend struct Attributor
static LLVM_ABI const char ID
Unique ID (due to the unique address)
static LLVM_ABI Value * getSingleValue(Attributor &A, const AbstractAttribute &AA, const IRPosition &IRP, SmallVectorImpl< AA::ValueAndContext > &Values)
Extract the single value in Values if any.
An abstract interface for privatizability.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract attribute for undefined behavior.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract attribute for getting all assumption underlying objects.
virtual bool forallUnderlyingObjects(function_ref< bool(Value &)> Pred, AA::ValueScope Scope=AA::Interprocedural) const =0
Check Pred on all underlying objects in Scope collected so far.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for range value analysis.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract interface for value simplify abstract attribute.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
An abstract attribute for willreturn.
static LLVM_ABI const char ID
Unique ID (due to the unique address)
Helper to represent an access offset and size, with logic to deal with uncertainty and check for over...
Definition Attributor.h:241
static constexpr int64_t Unknown
Definition Attributor.h:320
static RangeTy getUnknown()
Definition Attributor.h:247
Value * getValue() const
Definition Attributor.h:194
const Instruction * getCtxI() const
Definition Attributor.h:195
Base struct for all "concrete attribute" deductions.
void print(raw_ostream &OS) const
Helper functions, for debug purposes only.
virtual StateType & getState()=0
Return the internal abstract state for inspection.
AbstractState StateType
An interface to query the internal state of an abstract attribute.
virtual bool isAtFixpoint() const =0
Return if this abstract state is fixed, thus does not need to be updated if information changes as it...
virtual bool isValidState() const =0
Return if this abstract state is in a valid state.
Helper for AA::PointerInfo::Access DenseMap/Set usage ignoring everythign but the instruction.
static unsigned getHashValue(const Access &A)
static Access getTombstoneKey()
DenseMapInfo< Instruction * > Base
static bool isEqual(const Access &LHS, const Access &RHS)
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
Definition Alignment.h:77
std::function< void( const ArgumentReplacementInfo &, Function &, Function::arg_iterator)> CalleeRepairCBTy
Callee repair callback type.
const Argument & getReplacedArg() const
std::function< void(const ArgumentReplacementInfo &, AbstractCallSite, SmallVectorImpl< Value * > &)> ACSRepairCBTy
Abstract call site (ACS) repair callback type.
The fixpoint analysis framework that orchestrates the attribute deduction.
std::function< std::optional< Value * >( const IRPosition &, const AbstractAttribute *, bool &)> SimplifictionCallbackTy
Register CB as a simplification callback.
Specialization of the integer state for a bit-wise encoding.
BitIntegerState & addKnownBits(base_t Bits)
Add the bits in BitsEncoding to the "known bits".
Simple wrapper for a single bit (boolean) state.
static constexpr DenormalFPEnv getDefault()
static unsigned getHashValue(const Access &A)
static bool isEqual(const Access &LHS, const Access &RHS)
static bool isEqual(const AA::RangeTy &A, const AA::RangeTy B)
static unsigned getHashValue(const AA::RangeTy &Range)
DenseMapInfo< std::pair< const Instruction *, const ToTy * > > PairDMI
static ReachabilityQueryInfo< ToTy > * getTombstoneKey()
static bool isEqual(const ReachabilityQueryInfo< ToTy > *LHS, const ReachabilityQueryInfo< ToTy > *RHS)
DenseMapInfo< const AA::InstExclusionSetTy * > InstSetDMI
static unsigned getHashValue(const ReachabilityQueryInfo< ToTy > *RQI)
An information struct used to provide DenseMap with the various necessary components for a given valu...
State for dereferenceable attribute.
IncIntegerState DerefBytesState
State representing for dereferenceable bytes.
ChangeStatus manifest(Attributor &A) override
See AbstractAttribute::manifest(...).
Helper to describe and deal with positions in the LLVM-IR.
Definition Attributor.h:593
Function * getAssociatedFunction() const
Return the associated function, if any.
Definition Attributor.h:724
static const IRPosition callsite_returned(const CallBase &CB)
Create a position describing the returned value of CB.
Definition Attributor.h:661
static const IRPosition returned(const Function &F, const CallBaseContext *CBContext=nullptr)
Create a position describing the returned value of F.
Definition Attributor.h:643
LLVM_ABI Argument * getAssociatedArgument() const
Return the associated argument, if any.
static const IRPosition value(const Value &V, const CallBaseContext *CBContext=nullptr)
Create a position describing the value of V.
Definition Attributor.h:617
CallBase CallBaseContext
Definition Attributor.h:596
int getCalleeArgNo() const
Return the callee argument number of the associated value if it is an argument or call site argument,...
Definition Attributor.h:811
static const IRPosition inst(const Instruction &I, const CallBaseContext *CBContext=nullptr)
Create a position describing the instruction I.
Definition Attributor.h:629
static const IRPosition callsite_argument(const CallBase &CB, unsigned ArgNo)
Create a position describing the argument of CB at position ArgNo.
Definition Attributor.h:666
@ IRP_ARGUMENT
An attribute for a function argument.
Definition Attributor.h:607
@ IRP_RETURNED
An attribute for the function return value.
Definition Attributor.h:603
@ IRP_CALL_SITE
An attribute for a call site (function scope).
Definition Attributor.h:606
@ IRP_CALL_SITE_RETURNED
An attribute for a call site return value.
Definition Attributor.h:604
@ IRP_FUNCTION
An attribute for a function (scope).
Definition Attributor.h:605
@ IRP_CALL_SITE_ARGUMENT
An attribute for a call site argument.
Definition Attributor.h:608
@ IRP_INVALID
An invalid position.
Definition Attributor.h:600
Instruction * getCtxI() const
Return the context instruction, if any.
Definition Attributor.h:777
static const IRPosition argument(const Argument &Arg, const CallBaseContext *CBContext=nullptr)
Create a position describing the argument Arg.
Definition Attributor.h:650
Type * getAssociatedType() const
Return the type this abstract attribute is associated with.
Definition Attributor.h:800
static const IRPosition function(const Function &F, const CallBaseContext *CBContext=nullptr)
Create a position describing the function scope of F.
Definition Attributor.h:636
const CallBaseContext * getCallBaseContext() const
Get the call base context from the position.
Definition Attributor.h:939
Value & getAssociatedValue() const
Return the value this abstract attribute is associated with.
Definition Attributor.h:791
Value & getAnchorValue() const
Return the value this abstract attribute is anchored with.
Definition Attributor.h:710
int getCallSiteArgNo() const
Return the call site argument number of the associated value if it is an argument or call site argume...
Definition Attributor.h:820
static const IRPosition function_scope(const IRPosition &IRP, const CallBaseContext *CBContext=nullptr)
Create a position with function scope matching the "context" of IRP.
Definition Attributor.h:689
Kind getPositionKind() const
Return the associated position kind.
Definition Attributor.h:889
bool isArgumentPosition() const
Return true if the position is an argument or call site argument.
Definition Attributor.h:921
static const IRPosition callsite_function(const CallBase &CB)
Create a position describing the function scope of CB.
Definition Attributor.h:656
Function * getAnchorScope() const
Return the Function surrounding the anchor value.
Definition Attributor.h:765
Data structure to hold cached (LLVM-IR) information.
TargetLibraryInfo * getTargetLibraryInfoForFunction(const Function &F)
Return TargetLibraryInfo for function F.
bool isOnlyUsedByAssume(const Instruction &I) const
AP::Result * getAnalysisResultForFunction(const Function &F, bool CachedOnly=false)
Return the analysis result from a pass AP for function F.
ConstantRange getKnown() const
Return the known state encoding.
ConstantRange getAssumed() const
Return the assumed state encoding.
base_t getAssumed() const
Return the assumed state encoding.
Helper that allows to insert a new assumption string in the known assumption set by creating a (stati...
Definition Assumptions.h:37
FPClassTest KnownFPClasses
Floating-point classes the value could be one of.
A "must be executed context" for a given program point PP is the set of instructions,...
iterator & end()
Return an universal end iterator.
bool findInContextOf(const Instruction *I, const Instruction *PP)
Helper to look for I in the context of PP.
iterator & begin(const Instruction *PP)
Return an iterator to explore the context around PP.
bool checkForAllContext(const Instruction *PP, function_ref< bool(const Instruction *)> Pred)
}
Helper to tie a abstract state implementation to an abstract attribute.
StateType & getState() override
See AbstractAttribute::getState(...).
LLVM_ABI bool unionAssumed(std::optional< Value * > Other)
Merge Other into the currently assumed simplified value.
std::optional< Value * > SimplifiedAssociatedValue
An assumed simplified value.
Type * Ty
The type of the original value.