LLVM 20.0.0git
AttributorAttributes.cpp
Go to the documentation of this file.
1//===- AttributorAttributes.cpp - Attributes for Attributor deduction -----===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// See the Attributor.h file comment and the class descriptions in that file for
10// more information.
11//
12//===----------------------------------------------------------------------===//
13
15
16#include "llvm/ADT/APInt.h"
17#include "llvm/ADT/ArrayRef.h"
19#include "llvm/ADT/MapVector.h"
21#include "llvm/ADT/STLExtras.h"
23#include "llvm/ADT/SetVector.h"
26#include "llvm/ADT/Statistic.h"
39#include "llvm/IR/Argument.h"
40#include "llvm/IR/Assumptions.h"
41#include "llvm/IR/Attributes.h"
42#include "llvm/IR/BasicBlock.h"
43#include "llvm/IR/Constant.h"
44#include "llvm/IR/Constants.h"
45#include "llvm/IR/DataLayout.h"
47#include "llvm/IR/GlobalValue.h"
48#include "llvm/IR/IRBuilder.h"
49#include "llvm/IR/InlineAsm.h"
50#include "llvm/IR/InstrTypes.h"
51#include "llvm/IR/Instruction.h"
54#include "llvm/IR/IntrinsicsAMDGPU.h"
55#include "llvm/IR/IntrinsicsNVPTX.h"
56#include "llvm/IR/LLVMContext.h"
57#include "llvm/IR/MDBuilder.h"
58#include "llvm/IR/NoFolder.h"
59#include "llvm/IR/Value.h"
60#include "llvm/IR/ValueHandle.h"
73#include <cassert>
74#include <numeric>
75#include <optional>
76#include <string>
77
78using namespace llvm;
79
80#define DEBUG_TYPE "attributor"
81
83 "attributor-manifest-internal", cl::Hidden,
84 cl::desc("Manifest Attributor internal string attributes."),
85 cl::init(false));
86
87static cl::opt<int> MaxHeapToStackSize("max-heap-to-stack-size", cl::init(128),
89
90template <>
92
94
96 "attributor-max-potential-values", cl::Hidden,
97 cl::desc("Maximum number of potential values to be "
98 "tracked for each position."),
100 cl::init(7));
101
103 "attributor-max-potential-values-iterations", cl::Hidden,
104 cl::desc(
105 "Maximum number of iterations we keep dismantling potential values."),
106 cl::init(64));
107
108STATISTIC(NumAAs, "Number of abstract attributes created");
109STATISTIC(NumIndirectCallsPromoted, "Number of indirect calls promoted");
110
111// Some helper macros to deal with statistics tracking.
112//
113// Usage:
114// For simple IR attribute tracking overload trackStatistics in the abstract
115// attribute and choose the right STATS_DECLTRACK_********* macro,
116// e.g.,:
117// void trackStatistics() const override {
118// STATS_DECLTRACK_ARG_ATTR(returned)
119// }
120// If there is a single "increment" side one can use the macro
121// STATS_DECLTRACK with a custom message. If there are multiple increment
122// sides, STATS_DECL and STATS_TRACK can also be used separately.
123//
124#define BUILD_STAT_MSG_IR_ATTR(TYPE, NAME) \
125 ("Number of " #TYPE " marked '" #NAME "'")
126#define BUILD_STAT_NAME(NAME, TYPE) NumIR##TYPE##_##NAME
127#define STATS_DECL_(NAME, MSG) STATISTIC(NAME, MSG);
128#define STATS_DECL(NAME, TYPE, MSG) \
129 STATS_DECL_(BUILD_STAT_NAME(NAME, TYPE), MSG);
130#define STATS_TRACK(NAME, TYPE) ++(BUILD_STAT_NAME(NAME, TYPE));
131#define STATS_DECLTRACK(NAME, TYPE, MSG) \
132 { \
133 STATS_DECL(NAME, TYPE, MSG) \
134 STATS_TRACK(NAME, TYPE) \
135 }
136#define STATS_DECLTRACK_ARG_ATTR(NAME) \
137 STATS_DECLTRACK(NAME, Arguments, BUILD_STAT_MSG_IR_ATTR(arguments, NAME))
138#define STATS_DECLTRACK_CSARG_ATTR(NAME) \
139 STATS_DECLTRACK(NAME, CSArguments, \
140 BUILD_STAT_MSG_IR_ATTR(call site arguments, NAME))
141#define STATS_DECLTRACK_FN_ATTR(NAME) \
142 STATS_DECLTRACK(NAME, Function, BUILD_STAT_MSG_IR_ATTR(functions, NAME))
143#define STATS_DECLTRACK_CS_ATTR(NAME) \
144 STATS_DECLTRACK(NAME, CS, BUILD_STAT_MSG_IR_ATTR(call site, NAME))
145#define STATS_DECLTRACK_FNRET_ATTR(NAME) \
146 STATS_DECLTRACK(NAME, FunctionReturn, \
147 BUILD_STAT_MSG_IR_ATTR(function returns, NAME))
148#define STATS_DECLTRACK_CSRET_ATTR(NAME) \
149 STATS_DECLTRACK(NAME, CSReturn, \
150 BUILD_STAT_MSG_IR_ATTR(call site returns, NAME))
151#define STATS_DECLTRACK_FLOATING_ATTR(NAME) \
152 STATS_DECLTRACK(NAME, Floating, \
153 ("Number of floating values known to be '" #NAME "'"))
154
155// Specialization of the operator<< for abstract attributes subclasses. This
156// disambiguates situations where multiple operators are applicable.
157namespace llvm {
158#define PIPE_OPERATOR(CLASS) \
159 raw_ostream &operator<<(raw_ostream &OS, const CLASS &AA) { \
160 return OS << static_cast<const AbstractAttribute &>(AA); \
161 }
162
200
201#undef PIPE_OPERATOR
202
203template <>
205 const DerefState &R) {
206 ChangeStatus CS0 =
207 clampStateAndIndicateChange(S.DerefBytesState, R.DerefBytesState);
208 ChangeStatus CS1 = clampStateAndIndicateChange(S.GlobalState, R.GlobalState);
209 return CS0 | CS1;
210}
211
212} // namespace llvm
213
214static bool mayBeInCycle(const CycleInfo *CI, const Instruction *I,
215 bool HeaderOnly, Cycle **CPtr = nullptr) {
216 if (!CI)
217 return true;
218 auto *BB = I->getParent();
219 auto *C = CI->getCycle(BB);
220 if (!C)
221 return false;
222 if (CPtr)
223 *CPtr = C;
224 return !HeaderOnly || BB == C->getHeader();
225}
226
227/// Checks if a type could have padding bytes.
228static bool isDenselyPacked(Type *Ty, const DataLayout &DL) {
229 // There is no size information, so be conservative.
230 if (!Ty->isSized())
231 return false;
232
233 // If the alloc size is not equal to the storage size, then there are padding
234 // bytes. For x86_fp80 on x86-64, size: 80 alloc size: 128.
235 if (DL.getTypeSizeInBits(Ty) != DL.getTypeAllocSizeInBits(Ty))
236 return false;
237
238 // FIXME: This isn't the right way to check for padding in vectors with
239 // non-byte-size elements.
240 if (VectorType *SeqTy = dyn_cast<VectorType>(Ty))
241 return isDenselyPacked(SeqTy->getElementType(), DL);
242
243 // For array types, check for padding within members.
244 if (ArrayType *SeqTy = dyn_cast<ArrayType>(Ty))
245 return isDenselyPacked(SeqTy->getElementType(), DL);
246
247 if (!isa<StructType>(Ty))
248 return true;
249
250 // Check for padding within and between elements of a struct.
251 StructType *StructTy = cast<StructType>(Ty);
252 const StructLayout *Layout = DL.getStructLayout(StructTy);
253 uint64_t StartPos = 0;
254 for (unsigned I = 0, E = StructTy->getNumElements(); I < E; ++I) {
255 Type *ElTy = StructTy->getElementType(I);
256 if (!isDenselyPacked(ElTy, DL))
257 return false;
258 if (StartPos != Layout->getElementOffsetInBits(I))
259 return false;
260 StartPos += DL.getTypeAllocSizeInBits(ElTy);
261 }
262
263 return true;
264}
265
266/// Get pointer operand of memory accessing instruction. If \p I is
267/// not a memory accessing instruction, return nullptr. If \p AllowVolatile,
268/// is set to false and the instruction is volatile, return nullptr.
270 bool AllowVolatile) {
271 if (!AllowVolatile && I->isVolatile())
272 return nullptr;
273
274 if (auto *LI = dyn_cast<LoadInst>(I)) {
275 return LI->getPointerOperand();
276 }
277
278 if (auto *SI = dyn_cast<StoreInst>(I)) {
279 return SI->getPointerOperand();
280 }
281
282 if (auto *CXI = dyn_cast<AtomicCmpXchgInst>(I)) {
283 return CXI->getPointerOperand();
284 }
285
286 if (auto *RMWI = dyn_cast<AtomicRMWInst>(I)) {
287 return RMWI->getPointerOperand();
288 }
289
290 return nullptr;
291}
292
293/// Helper function to create a pointer based on \p Ptr, and advanced by \p
294/// Offset bytes.
296 IRBuilder<NoFolder> &IRB) {
297 LLVM_DEBUG(dbgs() << "Construct pointer: " << *Ptr << " + " << Offset
298 << "-bytes\n");
299
300 if (Offset)
301 Ptr = IRB.CreatePtrAdd(Ptr, IRB.getInt64(Offset),
302 Ptr->getName() + ".b" + Twine(Offset));
303 return Ptr;
304}
305
306static const Value *
308 const Value *Val, const DataLayout &DL, APInt &Offset,
309 bool GetMinOffset, bool AllowNonInbounds,
310 bool UseAssumed = false) {
311
312 auto AttributorAnalysis = [&](Value &V, APInt &ROffset) -> bool {
313 const IRPosition &Pos = IRPosition::value(V);
314 // Only track dependence if we are going to use the assumed info.
315 const AAValueConstantRange *ValueConstantRangeAA =
316 A.getAAFor<AAValueConstantRange>(QueryingAA, Pos,
317 UseAssumed ? DepClassTy::OPTIONAL
318 : DepClassTy::NONE);
319 if (!ValueConstantRangeAA)
320 return false;
321 ConstantRange Range = UseAssumed ? ValueConstantRangeAA->getAssumed()
322 : ValueConstantRangeAA->getKnown();
323 if (Range.isFullSet())
324 return false;
325
326 // We can only use the lower part of the range because the upper part can
327 // be higher than what the value can really be.
328 if (GetMinOffset)
329 ROffset = Range.getSignedMin();
330 else
331 ROffset = Range.getSignedMax();
332 return true;
333 };
334
335 return Val->stripAndAccumulateConstantOffsets(DL, Offset, AllowNonInbounds,
336 /* AllowInvariant */ true,
337 AttributorAnalysis);
338}
339
340static const Value *
342 const Value *Ptr, int64_t &BytesOffset,
343 const DataLayout &DL, bool AllowNonInbounds = false) {
344 APInt OffsetAPInt(DL.getIndexTypeSizeInBits(Ptr->getType()), 0);
345 const Value *Base =
346 stripAndAccumulateOffsets(A, QueryingAA, Ptr, DL, OffsetAPInt,
347 /* GetMinOffset */ true, AllowNonInbounds);
348
349 BytesOffset = OffsetAPInt.getSExtValue();
350 return Base;
351}
352
353/// Clamp the information known for all returned values of a function
354/// (identified by \p QueryingAA) into \p S.
355template <typename AAType, typename StateType = typename AAType::StateType,
356 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind,
357 bool RecurseForSelectAndPHI = true>
359 Attributor &A, const AAType &QueryingAA, StateType &S,
360 const IRPosition::CallBaseContext *CBContext = nullptr) {
361 LLVM_DEBUG(dbgs() << "[Attributor] Clamp return value states for "
362 << QueryingAA << " into " << S << "\n");
363
364 assert((QueryingAA.getIRPosition().getPositionKind() ==
366 QueryingAA.getIRPosition().getPositionKind() ==
368 "Can only clamp returned value states for a function returned or call "
369 "site returned position!");
370
371 // Use an optional state as there might not be any return values and we want
372 // to join (IntegerState::operator&) the state of all there are.
373 std::optional<StateType> T;
374
375 // Callback for each possibly returned value.
376 auto CheckReturnValue = [&](Value &RV) -> bool {
377 const IRPosition &RVPos = IRPosition::value(RV, CBContext);
378 // If possible, use the hasAssumedIRAttr interface.
379 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
380 bool IsKnown;
381 return AA::hasAssumedIRAttr<IRAttributeKind>(
382 A, &QueryingAA, RVPos, DepClassTy::REQUIRED, IsKnown);
383 }
384
385 const AAType *AA =
386 A.getAAFor<AAType>(QueryingAA, RVPos, DepClassTy::REQUIRED);
387 if (!AA)
388 return false;
389 LLVM_DEBUG(dbgs() << "[Attributor] RV: " << RV
390 << " AA: " << AA->getAsStr(&A) << " @ " << RVPos << "\n");
391 const StateType &AAS = AA->getState();
392 if (!T)
393 T = StateType::getBestState(AAS);
394 *T &= AAS;
395 LLVM_DEBUG(dbgs() << "[Attributor] AA State: " << AAS << " RV State: " << T
396 << "\n");
397 return T->isValidState();
398 };
399
400 if (!A.checkForAllReturnedValues(CheckReturnValue, QueryingAA,
401 AA::ValueScope::Intraprocedural,
402 RecurseForSelectAndPHI))
403 S.indicatePessimisticFixpoint();
404 else if (T)
405 S ^= *T;
406}
407
408namespace {
409/// Helper class for generic deduction: return value -> returned position.
410template <typename AAType, typename BaseType,
411 typename StateType = typename BaseType::StateType,
412 bool PropagateCallBaseContext = false,
413 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind,
414 bool RecurseForSelectAndPHI = true>
415struct AAReturnedFromReturnedValues : public BaseType {
416 AAReturnedFromReturnedValues(const IRPosition &IRP, Attributor &A)
417 : BaseType(IRP, A) {}
418
419 /// See AbstractAttribute::updateImpl(...).
420 ChangeStatus updateImpl(Attributor &A) override {
421 StateType S(StateType::getBestState(this->getState()));
422 clampReturnedValueStates<AAType, StateType, IRAttributeKind,
423 RecurseForSelectAndPHI>(
424 A, *this, S,
425 PropagateCallBaseContext ? this->getCallBaseContext() : nullptr);
426 // TODO: If we know we visited all returned values, thus no are assumed
427 // dead, we can take the known information from the state T.
428 return clampStateAndIndicateChange<StateType>(this->getState(), S);
429 }
430};
431
432/// Clamp the information known at all call sites for a given argument
433/// (identified by \p QueryingAA) into \p S.
434template <typename AAType, typename StateType = typename AAType::StateType,
435 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
436static void clampCallSiteArgumentStates(Attributor &A, const AAType &QueryingAA,
437 StateType &S) {
438 LLVM_DEBUG(dbgs() << "[Attributor] Clamp call site argument states for "
439 << QueryingAA << " into " << S << "\n");
440
441 assert(QueryingAA.getIRPosition().getPositionKind() ==
443 "Can only clamp call site argument states for an argument position!");
444
445 // Use an optional state as there might not be any return values and we want
446 // to join (IntegerState::operator&) the state of all there are.
447 std::optional<StateType> T;
448
449 // The argument number which is also the call site argument number.
450 unsigned ArgNo = QueryingAA.getIRPosition().getCallSiteArgNo();
451
452 auto CallSiteCheck = [&](AbstractCallSite ACS) {
453 const IRPosition &ACSArgPos = IRPosition::callsite_argument(ACS, ArgNo);
454 // Check if a coresponding argument was found or if it is on not associated
455 // (which can happen for callback calls).
456 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
457 return false;
458
459 // If possible, use the hasAssumedIRAttr interface.
460 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
461 bool IsKnown;
462 return AA::hasAssumedIRAttr<IRAttributeKind>(
463 A, &QueryingAA, ACSArgPos, DepClassTy::REQUIRED, IsKnown);
464 }
465
466 const AAType *AA =
467 A.getAAFor<AAType>(QueryingAA, ACSArgPos, DepClassTy::REQUIRED);
468 if (!AA)
469 return false;
470 LLVM_DEBUG(dbgs() << "[Attributor] ACS: " << *ACS.getInstruction()
471 << " AA: " << AA->getAsStr(&A) << " @" << ACSArgPos
472 << "\n");
473 const StateType &AAS = AA->getState();
474 if (!T)
475 T = StateType::getBestState(AAS);
476 *T &= AAS;
477 LLVM_DEBUG(dbgs() << "[Attributor] AA State: " << AAS << " CSA State: " << T
478 << "\n");
479 return T->isValidState();
480 };
481
482 bool UsedAssumedInformation = false;
483 if (!A.checkForAllCallSites(CallSiteCheck, QueryingAA, true,
484 UsedAssumedInformation))
485 S.indicatePessimisticFixpoint();
486 else if (T)
487 S ^= *T;
488}
489
490/// This function is the bridge between argument position and the call base
491/// context.
492template <typename AAType, typename BaseType,
493 typename StateType = typename AAType::StateType,
494 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
495bool getArgumentStateFromCallBaseContext(Attributor &A,
496 BaseType &QueryingAttribute,
497 IRPosition &Pos, StateType &State) {
499 "Expected an 'argument' position !");
500 const CallBase *CBContext = Pos.getCallBaseContext();
501 if (!CBContext)
502 return false;
503
504 int ArgNo = Pos.getCallSiteArgNo();
505 assert(ArgNo >= 0 && "Invalid Arg No!");
506 const IRPosition CBArgPos = IRPosition::callsite_argument(*CBContext, ArgNo);
507
508 // If possible, use the hasAssumedIRAttr interface.
509 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
510 bool IsKnown;
511 return AA::hasAssumedIRAttr<IRAttributeKind>(
512 A, &QueryingAttribute, CBArgPos, DepClassTy::REQUIRED, IsKnown);
513 }
514
515 const auto *AA =
516 A.getAAFor<AAType>(QueryingAttribute, CBArgPos, DepClassTy::REQUIRED);
517 if (!AA)
518 return false;
519 const StateType &CBArgumentState =
520 static_cast<const StateType &>(AA->getState());
521
522 LLVM_DEBUG(dbgs() << "[Attributor] Briding Call site context to argument"
523 << "Position:" << Pos << "CB Arg state:" << CBArgumentState
524 << "\n");
525
526 // NOTE: If we want to do call site grouping it should happen here.
527 State ^= CBArgumentState;
528 return true;
529}
530
531/// Helper class for generic deduction: call site argument -> argument position.
532template <typename AAType, typename BaseType,
533 typename StateType = typename AAType::StateType,
534 bool BridgeCallBaseContext = false,
535 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
536struct AAArgumentFromCallSiteArguments : public BaseType {
537 AAArgumentFromCallSiteArguments(const IRPosition &IRP, Attributor &A)
538 : BaseType(IRP, A) {}
539
540 /// See AbstractAttribute::updateImpl(...).
541 ChangeStatus updateImpl(Attributor &A) override {
542 StateType S = StateType::getBestState(this->getState());
543
544 if (BridgeCallBaseContext) {
545 bool Success =
546 getArgumentStateFromCallBaseContext<AAType, BaseType, StateType,
547 IRAttributeKind>(
548 A, *this, this->getIRPosition(), S);
549 if (Success)
550 return clampStateAndIndicateChange<StateType>(this->getState(), S);
551 }
552 clampCallSiteArgumentStates<AAType, StateType, IRAttributeKind>(A, *this,
553 S);
554
555 // TODO: If we know we visited all incoming values, thus no are assumed
556 // dead, we can take the known information from the state T.
557 return clampStateAndIndicateChange<StateType>(this->getState(), S);
558 }
559};
560
561/// Helper class for generic replication: function returned -> cs returned.
562template <typename AAType, typename BaseType,
563 typename StateType = typename BaseType::StateType,
564 bool IntroduceCallBaseContext = false,
565 Attribute::AttrKind IRAttributeKind = AAType::IRAttributeKind>
566struct AACalleeToCallSite : public BaseType {
567 AACalleeToCallSite(const IRPosition &IRP, Attributor &A) : BaseType(IRP, A) {}
568
569 /// See AbstractAttribute::updateImpl(...).
570 ChangeStatus updateImpl(Attributor &A) override {
571 auto IRPKind = this->getIRPosition().getPositionKind();
573 IRPKind == IRPosition::IRP_CALL_SITE) &&
574 "Can only wrap function returned positions for call site "
575 "returned positions!");
576 auto &S = this->getState();
577
578 CallBase &CB = cast<CallBase>(this->getAnchorValue());
579 if (IntroduceCallBaseContext)
580 LLVM_DEBUG(dbgs() << "[Attributor] Introducing call base context:" << CB
581 << "\n");
582
583 ChangeStatus Changed = ChangeStatus::UNCHANGED;
584 auto CalleePred = [&](ArrayRef<const Function *> Callees) {
585 for (const Function *Callee : Callees) {
586 IRPosition FnPos =
588 ? IRPosition::returned(*Callee,
589 IntroduceCallBaseContext ? &CB : nullptr)
591 *Callee, IntroduceCallBaseContext ? &CB : nullptr);
592 // If possible, use the hasAssumedIRAttr interface.
593 if (Attribute::isEnumAttrKind(IRAttributeKind)) {
594 bool IsKnown;
595 if (!AA::hasAssumedIRAttr<IRAttributeKind>(
596 A, this, FnPos, DepClassTy::REQUIRED, IsKnown))
597 return false;
598 continue;
599 }
600
601 const AAType *AA =
602 A.getAAFor<AAType>(*this, FnPos, DepClassTy::REQUIRED);
603 if (!AA)
604 return false;
605 Changed |= clampStateAndIndicateChange(S, AA->getState());
606 if (S.isAtFixpoint())
607 return S.isValidState();
608 }
609 return true;
610 };
611 if (!A.checkForAllCallees(CalleePred, *this, CB))
612 return S.indicatePessimisticFixpoint();
613 return Changed;
614 }
615};
616
617/// Helper function to accumulate uses.
618template <class AAType, typename StateType = typename AAType::StateType>
619static void followUsesInContext(AAType &AA, Attributor &A,
621 const Instruction *CtxI,
623 StateType &State) {
624 auto EIt = Explorer.begin(CtxI), EEnd = Explorer.end(CtxI);
625 for (unsigned u = 0; u < Uses.size(); ++u) {
626 const Use *U = Uses[u];
627 if (const Instruction *UserI = dyn_cast<Instruction>(U->getUser())) {
628 bool Found = Explorer.findInContextOf(UserI, EIt, EEnd);
629 if (Found && AA.followUseInMBEC(A, U, UserI, State))
630 for (const Use &Us : UserI->uses())
631 Uses.insert(&Us);
632 }
633 }
634}
635
636/// Use the must-be-executed-context around \p I to add information into \p S.
637/// The AAType class is required to have `followUseInMBEC` method with the
638/// following signature and behaviour:
639///
640/// bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I)
641/// U - Underlying use.
642/// I - The user of the \p U.
643/// Returns true if the value should be tracked transitively.
644///
645template <class AAType, typename StateType = typename AAType::StateType>
646static void followUsesInMBEC(AAType &AA, Attributor &A, StateType &S,
647 Instruction &CtxI) {
649 A.getInfoCache().getMustBeExecutedContextExplorer();
650 if (!Explorer)
651 return;
652
653 // Container for (transitive) uses of the associated value.
655 for (const Use &U : AA.getIRPosition().getAssociatedValue().uses())
656 Uses.insert(&U);
657
658 followUsesInContext<AAType>(AA, A, *Explorer, &CtxI, Uses, S);
659
660 if (S.isAtFixpoint())
661 return;
662
664 auto Pred = [&](const Instruction *I) {
665 if (const BranchInst *Br = dyn_cast<BranchInst>(I))
666 if (Br->isConditional())
667 BrInsts.push_back(Br);
668 return true;
669 };
670
671 // Here, accumulate conditional branch instructions in the context. We
672 // explore the child paths and collect the known states. The disjunction of
673 // those states can be merged to its own state. Let ParentState_i be a state
674 // to indicate the known information for an i-th branch instruction in the
675 // context. ChildStates are created for its successors respectively.
676 //
677 // ParentS_1 = ChildS_{1, 1} /\ ChildS_{1, 2} /\ ... /\ ChildS_{1, n_1}
678 // ParentS_2 = ChildS_{2, 1} /\ ChildS_{2, 2} /\ ... /\ ChildS_{2, n_2}
679 // ...
680 // ParentS_m = ChildS_{m, 1} /\ ChildS_{m, 2} /\ ... /\ ChildS_{m, n_m}
681 //
682 // Known State |= ParentS_1 \/ ParentS_2 \/... \/ ParentS_m
683 //
684 // FIXME: Currently, recursive branches are not handled. For example, we
685 // can't deduce that ptr must be dereferenced in below function.
686 //
687 // void f(int a, int c, int *ptr) {
688 // if(a)
689 // if (b) {
690 // *ptr = 0;
691 // } else {
692 // *ptr = 1;
693 // }
694 // else {
695 // if (b) {
696 // *ptr = 0;
697 // } else {
698 // *ptr = 1;
699 // }
700 // }
701 // }
702
703 Explorer->checkForAllContext(&CtxI, Pred);
704 for (const BranchInst *Br : BrInsts) {
705 StateType ParentState;
706
707 // The known state of the parent state is a conjunction of children's
708 // known states so it is initialized with a best state.
709 ParentState.indicateOptimisticFixpoint();
710
711 for (const BasicBlock *BB : Br->successors()) {
712 StateType ChildState;
713
714 size_t BeforeSize = Uses.size();
715 followUsesInContext(AA, A, *Explorer, &BB->front(), Uses, ChildState);
716
717 // Erase uses which only appear in the child.
718 for (auto It = Uses.begin() + BeforeSize; It != Uses.end();)
719 It = Uses.erase(It);
720
721 ParentState &= ChildState;
722 }
723
724 // Use only known state.
725 S += ParentState;
726 }
727}
728} // namespace
729
730/// ------------------------ PointerInfo ---------------------------------------
731
732namespace llvm {
733namespace AA {
734namespace PointerInfo {
735
736struct State;
737
738} // namespace PointerInfo
739} // namespace AA
740
741/// Helper for AA::PointerInfo::Access DenseMap/Set usage.
742template <>
745 static inline Access getEmptyKey();
746 static inline Access getTombstoneKey();
747 static unsigned getHashValue(const Access &A);
748 static bool isEqual(const Access &LHS, const Access &RHS);
749};
750
751/// Helper that allows RangeTy as a key in a DenseMap.
752template <> struct DenseMapInfo<AA::RangeTy> {
753 static inline AA::RangeTy getEmptyKey() {
754 auto EmptyKey = DenseMapInfo<int64_t>::getEmptyKey();
755 return AA::RangeTy{EmptyKey, EmptyKey};
756 }
757
758 static inline AA::RangeTy getTombstoneKey() {
759 auto TombstoneKey = DenseMapInfo<int64_t>::getTombstoneKey();
760 return AA::RangeTy{TombstoneKey, TombstoneKey};
761 }
762
763 static unsigned getHashValue(const AA::RangeTy &Range) {
767 }
768
769 static bool isEqual(const AA::RangeTy &A, const AA::RangeTy B) {
770 return A == B;
771 }
772};
773
774/// Helper for AA::PointerInfo::Access DenseMap/Set usage ignoring everythign
775/// but the instruction
776struct AccessAsInstructionInfo : DenseMapInfo<Instruction *> {
779 static inline Access getEmptyKey();
780 static inline Access getTombstoneKey();
781 static unsigned getHashValue(const Access &A);
782 static bool isEqual(const Access &LHS, const Access &RHS);
783};
784
785} // namespace llvm
786
787/// A type to track pointer/struct usage and accesses for AAPointerInfo.
789 /// Return the best possible representable state.
790 static State getBestState(const State &SIS) { return State(); }
791
792 /// Return the worst possible representable state.
793 static State getWorstState(const State &SIS) {
794 State R;
795 R.indicatePessimisticFixpoint();
796 return R;
797 }
798
799 State() = default;
800 State(State &&SIS) = default;
801
802 const State &getAssumed() const { return *this; }
803
804 /// See AbstractState::isValidState().
805 bool isValidState() const override { return BS.isValidState(); }
806
807 /// See AbstractState::isAtFixpoint().
808 bool isAtFixpoint() const override { return BS.isAtFixpoint(); }
809
810 /// See AbstractState::indicateOptimisticFixpoint().
814 }
815
816 /// See AbstractState::indicatePessimisticFixpoint().
820 }
821
822 State &operator=(const State &R) {
823 if (this == &R)
824 return *this;
825 BS = R.BS;
826 AccessList = R.AccessList;
827 OffsetBins = R.OffsetBins;
828 RemoteIMap = R.RemoteIMap;
829 ReturnedOffsets = R.ReturnedOffsets;
830 return *this;
831 }
832
834 if (this == &R)
835 return *this;
836 std::swap(BS, R.BS);
837 std::swap(AccessList, R.AccessList);
838 std::swap(OffsetBins, R.OffsetBins);
839 std::swap(RemoteIMap, R.RemoteIMap);
840 std::swap(ReturnedOffsets, R.ReturnedOffsets);
841 return *this;
842 }
843
844 /// Add a new Access to the state at offset \p Offset and with size \p Size.
845 /// The access is associated with \p I, writes \p Content (if anything), and
846 /// is of kind \p Kind. If an Access already exists for the same \p I and same
847 /// \p RemoteI, the two are combined, potentially losing information about
848 /// offset and size. The resulting access must now be moved from its original
849 /// OffsetBin to the bin for its new offset.
850 ///
851 /// \Returns CHANGED, if the state changed, UNCHANGED otherwise.
853 Instruction &I, std::optional<Value *> Content,
855 Instruction *RemoteI = nullptr);
856
859 int64_t numOffsetBins() const { return OffsetBins.size(); }
860
861 const AAPointerInfo::Access &getAccess(unsigned Index) const {
862 return AccessList[Index];
863 }
864
865protected:
866 // Every memory instruction results in an Access object. We maintain a list of
867 // all Access objects that we own, along with the following maps:
868 //
869 // - OffsetBins: RangeTy -> { Access }
870 // - RemoteIMap: RemoteI x LocalI -> Access
871 //
872 // A RemoteI is any instruction that accesses memory. RemoteI is different
873 // from LocalI if and only if LocalI is a call; then RemoteI is some
874 // instruction in the callgraph starting from LocalI. Multiple paths in the
875 // callgraph from LocalI to RemoteI may produce multiple accesses, but these
876 // are all combined into a single Access object. This may result in loss of
877 // information in RangeTy in the Access object.
881
882 /// Flag to determine if the underlying pointer is reaching a return statement
883 /// in the associated function or not. Returns in other functions cause
884 /// invalidation.
886
887 /// See AAPointerInfo::forallInterferingAccesses.
888 template <typename F>
891 return false;
892
893 for (const auto &It : OffsetBins) {
894 AA::RangeTy ItRange = It.getFirst();
895 if (!Range.mayOverlap(ItRange))
896 continue;
897 bool IsExact = Range == ItRange && !Range.offsetOrSizeAreUnknown();
898 for (auto Index : It.getSecond()) {
899 auto &Access = AccessList[Index];
900 if (!CB(Access, IsExact))
901 return false;
902 }
903 }
904 return true;
905 }
906
907 /// See AAPointerInfo::forallInterferingAccesses.
908 template <typename F>
910 AA::RangeTy &Range) const {
912 return false;
913
914 auto LocalList = RemoteIMap.find(&I);
915 if (LocalList == RemoteIMap.end()) {
916 return true;
917 }
918
919 for (unsigned Index : LocalList->getSecond()) {
920 for (auto &R : AccessList[Index]) {
921 Range &= R;
922 if (Range.offsetAndSizeAreUnknown())
923 break;
924 }
925 }
927 }
928
929private:
930 /// State to track fixpoint and validity.
931 BooleanState BS;
932};
933
936 std::optional<Value *> Content, AAPointerInfo::AccessKind Kind, Type *Ty,
937 Instruction *RemoteI) {
938 RemoteI = RemoteI ? RemoteI : &I;
939
940 // Check if we have an access for this instruction, if not, simply add it.
941 auto &LocalList = RemoteIMap[RemoteI];
942 bool AccExists = false;
943 unsigned AccIndex = AccessList.size();
944 for (auto Index : LocalList) {
945 auto &A = AccessList[Index];
946 if (A.getLocalInst() == &I) {
947 AccExists = true;
948 AccIndex = Index;
949 break;
950 }
951 }
952
953 auto AddToBins = [&](const AAPointerInfo::RangeList &ToAdd) {
954 LLVM_DEBUG(if (ToAdd.size()) dbgs()
955 << "[AAPointerInfo] Inserting access in new offset bins\n";);
956
957 for (auto Key : ToAdd) {
958 LLVM_DEBUG(dbgs() << " key " << Key << "\n");
959 OffsetBins[Key].insert(AccIndex);
960 }
961 };
962
963 if (!AccExists) {
964 AccessList.emplace_back(&I, RemoteI, Ranges, Content, Kind, Ty);
965 assert((AccessList.size() == AccIndex + 1) &&
966 "New Access should have been at AccIndex");
967 LocalList.push_back(AccIndex);
968 AddToBins(AccessList[AccIndex].getRanges());
970 }
971
972 // Combine the new Access with the existing Access, and then update the
973 // mapping in the offset bins.
974 AAPointerInfo::Access Acc(&I, RemoteI, Ranges, Content, Kind, Ty);
975 auto &Current = AccessList[AccIndex];
976 auto Before = Current;
977 Current &= Acc;
978 if (Current == Before)
980
981 auto &ExistingRanges = Before.getRanges();
982 auto &NewRanges = Current.getRanges();
983
984 // Ranges that are in the old access but not the new access need to be removed
985 // from the offset bins.
987 AAPointerInfo::RangeList::set_difference(ExistingRanges, NewRanges, ToRemove);
988 LLVM_DEBUG(if (ToRemove.size()) dbgs()
989 << "[AAPointerInfo] Removing access from old offset bins\n";);
990
991 for (auto Key : ToRemove) {
992 LLVM_DEBUG(dbgs() << " key " << Key << "\n");
993 assert(OffsetBins.count(Key) && "Existing Access must be in some bin.");
994 auto &Bin = OffsetBins[Key];
995 assert(Bin.count(AccIndex) &&
996 "Expected bin to actually contain the Access.");
997 Bin.erase(AccIndex);
998 }
999
1000 // Ranges that are in the new access but not the old access need to be added
1001 // to the offset bins.
1003 AAPointerInfo::RangeList::set_difference(NewRanges, ExistingRanges, ToAdd);
1004 AddToBins(ToAdd);
1005 return ChangeStatus::CHANGED;
1006}
1007
1008namespace {
1009
1010#ifndef NDEBUG
1012 const AAPointerInfo::OffsetInfo &OI) {
1013 ListSeparator LS;
1014 OS << "[";
1015 for (auto Offset : OI) {
1016 OS << LS << Offset;
1017 }
1018 OS << "]";
1019 return OS;
1020}
1021#endif // NDEBUG
1022
1023struct AAPointerInfoImpl
1024 : public StateWrapper<AA::PointerInfo::State, AAPointerInfo> {
1026 AAPointerInfoImpl(const IRPosition &IRP, Attributor &A) : BaseTy(IRP) {}
1027
1028 /// See AbstractAttribute::getAsStr().
1029 const std::string getAsStr(Attributor *A) const override {
1030 return std::string("PointerInfo ") +
1031 (isValidState() ? (std::string("#") +
1032 std::to_string(OffsetBins.size()) + " bins")
1033 : "<invalid>") +
1034 (reachesReturn()
1035 ? (" (returned:" +
1036 join(map_range(ReturnedOffsets,
1037 [](int64_t O) { return std::to_string(O); }),
1038 ", ") +
1039 ")")
1040 : "");
1041 }
1042
1043 /// See AbstractAttribute::manifest(...).
1044 ChangeStatus manifest(Attributor &A) override {
1045 return AAPointerInfo::manifest(A);
1046 }
1047
1048 virtual const_bin_iterator begin() const override { return State::begin(); }
1049 virtual const_bin_iterator end() const override { return State::end(); }
1050 virtual int64_t numOffsetBins() const override {
1051 return State::numOffsetBins();
1052 }
1053 virtual bool reachesReturn() const override {
1054 return !ReturnedOffsets.isUnassigned();
1055 }
1056 virtual void addReturnedOffsetsTo(OffsetInfo &OI) const override {
1057 if (ReturnedOffsets.isUnknown()) {
1058 OI.setUnknown();
1059 return;
1060 }
1061
1062 OffsetInfo MergedOI;
1063 for (auto Offset : ReturnedOffsets) {
1064 OffsetInfo TmpOI = OI;
1065 TmpOI.addToAll(Offset);
1066 MergedOI.merge(TmpOI);
1067 }
1068 OI = std::move(MergedOI);
1069 }
1070
1071 ChangeStatus setReachesReturn(const OffsetInfo &ReachedReturnedOffsets) {
1072 if (ReturnedOffsets.isUnknown())
1073 return ChangeStatus::UNCHANGED;
1074 if (ReachedReturnedOffsets.isUnknown()) {
1075 ReturnedOffsets.setUnknown();
1076 return ChangeStatus::CHANGED;
1077 }
1078 if (ReturnedOffsets.merge(ReachedReturnedOffsets))
1079 return ChangeStatus::CHANGED;
1080 return ChangeStatus::UNCHANGED;
1081 }
1082
1083 bool forallInterferingAccesses(
1085 function_ref<bool(const AAPointerInfo::Access &, bool)> CB)
1086 const override {
1087 return State::forallInterferingAccesses(Range, CB);
1088 }
1089
1090 bool forallInterferingAccesses(
1091 Attributor &A, const AbstractAttribute &QueryingAA, Instruction &I,
1092 bool FindInterferingWrites, bool FindInterferingReads,
1093 function_ref<bool(const Access &, bool)> UserCB, bool &HasBeenWrittenTo,
1095 function_ref<bool(const Access &)> SkipCB) const override {
1096 HasBeenWrittenTo = false;
1097
1098 SmallPtrSet<const Access *, 8> DominatingWrites;
1099 SmallVector<std::pair<const Access *, bool>, 8> InterferingAccesses;
1100
1101 Function &Scope = *I.getFunction();
1102 bool IsKnownNoSync;
1103 bool IsAssumedNoSync = AA::hasAssumedIRAttr<Attribute::NoSync>(
1104 A, &QueryingAA, IRPosition::function(Scope), DepClassTy::OPTIONAL,
1105 IsKnownNoSync);
1106 const auto *ExecDomainAA = A.lookupAAFor<AAExecutionDomain>(
1107 IRPosition::function(Scope), &QueryingAA, DepClassTy::NONE);
1108 bool AllInSameNoSyncFn = IsAssumedNoSync;
1109 bool InstIsExecutedByInitialThreadOnly =
1110 ExecDomainAA && ExecDomainAA->isExecutedByInitialThreadOnly(I);
1111
1112 // If the function is not ending in aligned barriers, we need the stores to
1113 // be in aligned barriers. The load being in one is not sufficient since the
1114 // store might be executed by a thread that disappears after, causing the
1115 // aligned barrier guarding the load to unblock and the load to read a value
1116 // that has no CFG path to the load.
1117 bool InstIsExecutedInAlignedRegion =
1118 FindInterferingReads && ExecDomainAA &&
1119 ExecDomainAA->isExecutedInAlignedRegion(A, I);
1120
1121 if (InstIsExecutedInAlignedRegion || InstIsExecutedByInitialThreadOnly)
1122 A.recordDependence(*ExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1123
1124 InformationCache &InfoCache = A.getInfoCache();
1125 bool IsThreadLocalObj =
1126 AA::isAssumedThreadLocalObject(A, getAssociatedValue(), *this);
1127
1128 // Helper to determine if we need to consider threading, which we cannot
1129 // right now. However, if the function is (assumed) nosync or the thread
1130 // executing all instructions is the main thread only we can ignore
1131 // threading. Also, thread-local objects do not require threading reasoning.
1132 // Finally, we can ignore threading if either access is executed in an
1133 // aligned region.
1134 auto CanIgnoreThreadingForInst = [&](const Instruction &I) -> bool {
1135 if (IsThreadLocalObj || AllInSameNoSyncFn)
1136 return true;
1137 const auto *FnExecDomainAA =
1138 I.getFunction() == &Scope
1139 ? ExecDomainAA
1140 : A.lookupAAFor<AAExecutionDomain>(
1141 IRPosition::function(*I.getFunction()), &QueryingAA,
1142 DepClassTy::NONE);
1143 if (!FnExecDomainAA)
1144 return false;
1145 if (InstIsExecutedInAlignedRegion ||
1146 (FindInterferingWrites &&
1147 FnExecDomainAA->isExecutedInAlignedRegion(A, I))) {
1148 A.recordDependence(*FnExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1149 return true;
1150 }
1151 if (InstIsExecutedByInitialThreadOnly &&
1152 FnExecDomainAA->isExecutedByInitialThreadOnly(I)) {
1153 A.recordDependence(*FnExecDomainAA, QueryingAA, DepClassTy::OPTIONAL);
1154 return true;
1155 }
1156 return false;
1157 };
1158
1159 // Helper to determine if the access is executed by the same thread as the
1160 // given instruction, for now it is sufficient to avoid any potential
1161 // threading effects as we cannot deal with them anyway.
1162 auto CanIgnoreThreading = [&](const Access &Acc) -> bool {
1163 return CanIgnoreThreadingForInst(*Acc.getRemoteInst()) ||
1164 (Acc.getRemoteInst() != Acc.getLocalInst() &&
1165 CanIgnoreThreadingForInst(*Acc.getLocalInst()));
1166 };
1167
1168 // TODO: Use inter-procedural reachability and dominance.
1169 bool IsKnownNoRecurse;
1170 AA::hasAssumedIRAttr<Attribute::NoRecurse>(
1171 A, this, IRPosition::function(Scope), DepClassTy::OPTIONAL,
1172 IsKnownNoRecurse);
1173
1174 // TODO: Use reaching kernels from AAKernelInfo (or move it to
1175 // AAExecutionDomain) such that we allow scopes other than kernels as long
1176 // as the reaching kernels are disjoint.
1177 bool InstInKernel = A.getInfoCache().isKernel(Scope);
1178 bool ObjHasKernelLifetime = false;
1179 const bool UseDominanceReasoning =
1180 FindInterferingWrites && IsKnownNoRecurse;
1181 const DominatorTree *DT =
1183
1184 // Helper to check if a value has "kernel lifetime", that is it will not
1185 // outlive a GPU kernel. This is true for shared, constant, and local
1186 // globals on AMD and NVIDIA GPUs.
1187 auto HasKernelLifetime = [&](Value *V, Module &M) {
1188 if (!AA::isGPU(M))
1189 return false;
1190 switch (AA::GPUAddressSpace(V->getType()->getPointerAddressSpace())) {
1191 case AA::GPUAddressSpace::Shared:
1192 case AA::GPUAddressSpace::Constant:
1193 case AA::GPUAddressSpace::Local:
1194 return true;
1195 default:
1196 return false;
1197 };
1198 };
1199
1200 // The IsLiveInCalleeCB will be used by the AA::isPotentiallyReachable query
1201 // to determine if we should look at reachability from the callee. For
1202 // certain pointers we know the lifetime and we do not have to step into the
1203 // callee to determine reachability as the pointer would be dead in the
1204 // callee. See the conditional initialization below.
1205 std::function<bool(const Function &)> IsLiveInCalleeCB;
1206
1207 if (auto *AI = dyn_cast<AllocaInst>(&getAssociatedValue())) {
1208 // If the alloca containing function is not recursive the alloca
1209 // must be dead in the callee.
1210 const Function *AIFn = AI->getFunction();
1211 ObjHasKernelLifetime = A.getInfoCache().isKernel(*AIFn);
1212 bool IsKnownNoRecurse;
1213 if (AA::hasAssumedIRAttr<Attribute::NoRecurse>(
1214 A, this, IRPosition::function(*AIFn), DepClassTy::OPTIONAL,
1215 IsKnownNoRecurse)) {
1216 IsLiveInCalleeCB = [AIFn](const Function &Fn) { return AIFn != &Fn; };
1217 }
1218 } else if (auto *GV = dyn_cast<GlobalValue>(&getAssociatedValue())) {
1219 // If the global has kernel lifetime we can stop if we reach a kernel
1220 // as it is "dead" in the (unknown) callees.
1221 ObjHasKernelLifetime = HasKernelLifetime(GV, *GV->getParent());
1222 if (ObjHasKernelLifetime)
1223 IsLiveInCalleeCB = [&A](const Function &Fn) {
1224 return !A.getInfoCache().isKernel(Fn);
1225 };
1226 }
1227
1228 // Set of accesses/instructions that will overwrite the result and are
1229 // therefore blockers in the reachability traversal.
1230 AA::InstExclusionSetTy ExclusionSet;
1231
1232 auto AccessCB = [&](const Access &Acc, bool Exact) {
1233 Function *AccScope = Acc.getRemoteInst()->getFunction();
1234 bool AccInSameScope = AccScope == &Scope;
1235
1236 // If the object has kernel lifetime we can ignore accesses only reachable
1237 // by other kernels. For now we only skip accesses *in* other kernels.
1238 if (InstInKernel && ObjHasKernelLifetime && !AccInSameScope &&
1239 A.getInfoCache().isKernel(*AccScope))
1240 return true;
1241
1242 if (Exact && Acc.isMustAccess() && Acc.getRemoteInst() != &I) {
1243 if (Acc.isWrite() || (isa<LoadInst>(I) && Acc.isWriteOrAssumption()))
1244 ExclusionSet.insert(Acc.getRemoteInst());
1245 }
1246
1247 if ((!FindInterferingWrites || !Acc.isWriteOrAssumption()) &&
1248 (!FindInterferingReads || !Acc.isRead()))
1249 return true;
1250
1251 bool Dominates = FindInterferingWrites && DT && Exact &&
1252 Acc.isMustAccess() && AccInSameScope &&
1253 DT->dominates(Acc.getRemoteInst(), &I);
1254 if (Dominates)
1255 DominatingWrites.insert(&Acc);
1256
1257 // Track if all interesting accesses are in the same `nosync` function as
1258 // the given instruction.
1259 AllInSameNoSyncFn &= Acc.getRemoteInst()->getFunction() == &Scope;
1260
1261 InterferingAccesses.push_back({&Acc, Exact});
1262 return true;
1263 };
1264 if (!State::forallInterferingAccesses(I, AccessCB, Range))
1265 return false;
1266
1267 HasBeenWrittenTo = !DominatingWrites.empty();
1268
1269 // Dominating writes form a chain, find the least/lowest member.
1270 Instruction *LeastDominatingWriteInst = nullptr;
1271 for (const Access *Acc : DominatingWrites) {
1272 if (!LeastDominatingWriteInst) {
1273 LeastDominatingWriteInst = Acc->getRemoteInst();
1274 } else if (DT->dominates(LeastDominatingWriteInst,
1275 Acc->getRemoteInst())) {
1276 LeastDominatingWriteInst = Acc->getRemoteInst();
1277 }
1278 }
1279
1280 // Helper to determine if we can skip a specific write access.
1281 auto CanSkipAccess = [&](const Access &Acc, bool Exact) {
1282 if (SkipCB && SkipCB(Acc))
1283 return true;
1284 if (!CanIgnoreThreading(Acc))
1285 return false;
1286
1287 // Check read (RAW) dependences and write (WAR) dependences as necessary.
1288 // If we successfully excluded all effects we are interested in, the
1289 // access can be skipped.
1290 bool ReadChecked = !FindInterferingReads;
1291 bool WriteChecked = !FindInterferingWrites;
1292
1293 // If the instruction cannot reach the access, the former does not
1294 // interfere with what the access reads.
1295 if (!ReadChecked) {
1296 if (!AA::isPotentiallyReachable(A, I, *Acc.getRemoteInst(), QueryingAA,
1297 &ExclusionSet, IsLiveInCalleeCB))
1298 ReadChecked = true;
1299 }
1300 // If the instruction cannot be reach from the access, the latter does not
1301 // interfere with what the instruction reads.
1302 if (!WriteChecked) {
1303 if (!AA::isPotentiallyReachable(A, *Acc.getRemoteInst(), I, QueryingAA,
1304 &ExclusionSet, IsLiveInCalleeCB))
1305 WriteChecked = true;
1306 }
1307
1308 // If we still might be affected by the write of the access but there are
1309 // dominating writes in the function of the instruction
1310 // (HasBeenWrittenTo), we can try to reason that the access is overwritten
1311 // by them. This would have happend above if they are all in the same
1312 // function, so we only check the inter-procedural case. Effectively, we
1313 // want to show that there is no call after the dominting write that might
1314 // reach the access, and when it returns reach the instruction with the
1315 // updated value. To this end, we iterate all call sites, check if they
1316 // might reach the instruction without going through another access
1317 // (ExclusionSet) and at the same time might reach the access. However,
1318 // that is all part of AAInterFnReachability.
1319 if (!WriteChecked && HasBeenWrittenTo &&
1320 Acc.getRemoteInst()->getFunction() != &Scope) {
1321
1322 const auto *FnReachabilityAA = A.getAAFor<AAInterFnReachability>(
1323 QueryingAA, IRPosition::function(Scope), DepClassTy::OPTIONAL);
1324 if (FnReachabilityAA) {
1325 // Without going backwards in the call tree, can we reach the access
1326 // from the least dominating write. Do not allow to pass the
1327 // instruction itself either.
1328 bool Inserted = ExclusionSet.insert(&I).second;
1329
1330 if (!FnReachabilityAA->instructionCanReach(
1331 A, *LeastDominatingWriteInst,
1332 *Acc.getRemoteInst()->getFunction(), &ExclusionSet))
1333 WriteChecked = true;
1334
1335 if (Inserted)
1336 ExclusionSet.erase(&I);
1337 }
1338 }
1339
1340 if (ReadChecked && WriteChecked)
1341 return true;
1342
1343 if (!DT || !UseDominanceReasoning)
1344 return false;
1345 if (!DominatingWrites.count(&Acc))
1346 return false;
1347 return LeastDominatingWriteInst != Acc.getRemoteInst();
1348 };
1349
1350 // Run the user callback on all accesses we cannot skip and return if
1351 // that succeeded for all or not.
1352 for (auto &It : InterferingAccesses) {
1353 if ((!AllInSameNoSyncFn && !IsThreadLocalObj && !ExecDomainAA) ||
1354 !CanSkipAccess(*It.first, It.second)) {
1355 if (!UserCB(*It.first, It.second))
1356 return false;
1357 }
1358 }
1359 return true;
1360 }
1361
1362 ChangeStatus translateAndAddStateFromCallee(Attributor &A,
1363 const AAPointerInfo &OtherAA,
1364 CallBase &CB) {
1365 using namespace AA::PointerInfo;
1366 if (!OtherAA.getState().isValidState() || !isValidState())
1367 return indicatePessimisticFixpoint();
1368
1369 ChangeStatus Changed = ChangeStatus::UNCHANGED;
1370 const auto &OtherAAImpl = static_cast<const AAPointerInfoImpl &>(OtherAA);
1371 bool IsByval = OtherAAImpl.getAssociatedArgument()->hasByValAttr();
1372 Changed |= setReachesReturn(OtherAAImpl.ReturnedOffsets);
1373
1374 // Combine the accesses bin by bin.
1375 const auto &State = OtherAAImpl.getState();
1376 for (const auto &It : State) {
1377 for (auto Index : It.getSecond()) {
1378 const auto &RAcc = State.getAccess(Index);
1379 if (IsByval && !RAcc.isRead())
1380 continue;
1381 bool UsedAssumedInformation = false;
1382 AccessKind AK = RAcc.getKind();
1383 auto Content = A.translateArgumentToCallSiteContent(
1384 RAcc.getContent(), CB, *this, UsedAssumedInformation);
1385 AK = AccessKind(AK & (IsByval ? AccessKind::AK_R : AccessKind::AK_RW));
1386 AK = AccessKind(AK | (RAcc.isMayAccess() ? AK_MAY : AK_MUST));
1387
1388 Changed |= addAccess(A, RAcc.getRanges(), CB, Content, AK,
1389 RAcc.getType(), RAcc.getRemoteInst());
1390 }
1391 }
1392 return Changed;
1393 }
1394
1395 ChangeStatus translateAndAddState(Attributor &A, const AAPointerInfo &OtherAA,
1396 const OffsetInfo &Offsets, CallBase &CB,
1397 bool IsMustAcc) {
1398 using namespace AA::PointerInfo;
1399 if (!OtherAA.getState().isValidState() || !isValidState())
1400 return indicatePessimisticFixpoint();
1401
1402 const auto &OtherAAImpl = static_cast<const AAPointerInfoImpl &>(OtherAA);
1403
1404 // Combine the accesses bin by bin.
1405 ChangeStatus Changed = ChangeStatus::UNCHANGED;
1406 const auto &State = OtherAAImpl.getState();
1407 for (const auto &It : State) {
1408 for (auto Index : It.getSecond()) {
1409 const auto &RAcc = State.getAccess(Index);
1410 if (!IsMustAcc && RAcc.isAssumption())
1411 continue;
1412 for (auto Offset : Offsets) {
1413 auto NewRanges = Offset == AA::RangeTy::Unknown
1415 : RAcc.getRanges();
1416 if (!NewRanges.isUnknown()) {
1417 NewRanges.addToAllOffsets(Offset);
1418 }
1419 AccessKind AK = RAcc.getKind();
1420 if (!IsMustAcc)
1421 AK = AccessKind((AK & ~AK_MUST) | AK_MAY);
1422 Changed |= addAccess(A, NewRanges, CB, RAcc.getContent(), AK,
1423 RAcc.getType(), RAcc.getRemoteInst());
1424 }
1425 }
1426 }
1427 return Changed;
1428 }
1429
1430 /// Statistic tracking for all AAPointerInfo implementations.
1431 /// See AbstractAttribute::trackStatistics().
1432 void trackPointerInfoStatistics(const IRPosition &IRP) const {}
1433
1434 /// Dump the state into \p O.
1435 void dumpState(raw_ostream &O) {
1436 for (auto &It : OffsetBins) {
1437 O << "[" << It.first.Offset << "-" << It.first.Offset + It.first.Size
1438 << "] : " << It.getSecond().size() << "\n";
1439 for (auto AccIndex : It.getSecond()) {
1440 auto &Acc = AccessList[AccIndex];
1441 O << " - " << Acc.getKind() << " - " << *Acc.getLocalInst() << "\n";
1442 if (Acc.getLocalInst() != Acc.getRemoteInst())
1443 O << " --> " << *Acc.getRemoteInst()
1444 << "\n";
1445 if (!Acc.isWrittenValueYetUndetermined()) {
1446 if (isa_and_nonnull<Function>(Acc.getWrittenValue()))
1447 O << " - c: func " << Acc.getWrittenValue()->getName()
1448 << "\n";
1449 else if (Acc.getWrittenValue())
1450 O << " - c: " << *Acc.getWrittenValue() << "\n";
1451 else
1452 O << " - c: <unknown>\n";
1453 }
1454 }
1455 }
1456 }
1457};
1458
1459struct AAPointerInfoFloating : public AAPointerInfoImpl {
1461 AAPointerInfoFloating(const IRPosition &IRP, Attributor &A)
1462 : AAPointerInfoImpl(IRP, A) {}
1463
1464 /// Deal with an access and signal if it was handled successfully.
1465 bool handleAccess(Attributor &A, Instruction &I,
1466 std::optional<Value *> Content, AccessKind Kind,
1467 OffsetInfo::VecTy &Offsets, ChangeStatus &Changed,
1468 Type &Ty) {
1469 using namespace AA::PointerInfo;
1471 const DataLayout &DL = A.getDataLayout();
1472 TypeSize AccessSize = DL.getTypeStoreSize(&Ty);
1473 if (!AccessSize.isScalable())
1474 Size = AccessSize.getFixedValue();
1475
1476 // Make a strictly ascending list of offsets as required by addAccess()
1477 SmallVector<int64_t> OffsetsSorted(Offsets.begin(), Offsets.end());
1478 llvm::sort(OffsetsSorted);
1479
1480 VectorType *VT = dyn_cast<VectorType>(&Ty);
1481 if (!VT || VT->getElementCount().isScalable() ||
1482 !Content.value_or(nullptr) || !isa<Constant>(*Content) ||
1483 (*Content)->getType() != VT ||
1484 DL.getTypeStoreSize(VT->getElementType()).isScalable()) {
1485 Changed =
1486 Changed | addAccess(A, {OffsetsSorted, Size}, I, Content, Kind, &Ty);
1487 } else {
1488 // Handle vector stores with constant content element-wise.
1489 // TODO: We could look for the elements or create instructions
1490 // representing them.
1491 // TODO: We need to push the Content into the range abstraction
1492 // (AA::RangeTy) to allow different content values for different
1493 // ranges. ranges. Hence, support vectors storing different values.
1494 Type *ElementType = VT->getElementType();
1495 int64_t ElementSize = DL.getTypeStoreSize(ElementType).getFixedValue();
1496 auto *ConstContent = cast<Constant>(*Content);
1497 Type *Int32Ty = Type::getInt32Ty(ElementType->getContext());
1498 SmallVector<int64_t> ElementOffsets(Offsets.begin(), Offsets.end());
1499
1500 for (int i = 0, e = VT->getElementCount().getFixedValue(); i != e; ++i) {
1501 Value *ElementContent = ConstantExpr::getExtractElement(
1502 ConstContent, ConstantInt::get(Int32Ty, i));
1503
1504 // Add the element access.
1505 Changed = Changed | addAccess(A, {ElementOffsets, ElementSize}, I,
1506 ElementContent, Kind, ElementType);
1507
1508 // Advance the offsets for the next element.
1509 for (auto &ElementOffset : ElementOffsets)
1510 ElementOffset += ElementSize;
1511 }
1512 }
1513 return true;
1514 };
1515
1516 /// See AbstractAttribute::updateImpl(...).
1517 ChangeStatus updateImpl(Attributor &A) override;
1518
1519 /// If the indices to \p GEP can be traced to constants, incorporate all
1520 /// of these into \p UsrOI.
1521 ///
1522 /// \return true iff \p UsrOI is updated.
1523 bool collectConstantsForGEP(Attributor &A, const DataLayout &DL,
1524 OffsetInfo &UsrOI, const OffsetInfo &PtrOI,
1525 const GEPOperator *GEP);
1526
1527 /// See AbstractAttribute::trackStatistics()
1528 void trackStatistics() const override {
1529 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1530 }
1531};
1532
1533bool AAPointerInfoFloating::collectConstantsForGEP(Attributor &A,
1534 const DataLayout &DL,
1535 OffsetInfo &UsrOI,
1536 const OffsetInfo &PtrOI,
1537 const GEPOperator *GEP) {
1538 unsigned BitWidth = DL.getIndexTypeSizeInBits(GEP->getType());
1539 SmallMapVector<Value *, APInt, 4> VariableOffsets;
1540 APInt ConstantOffset(BitWidth, 0);
1541
1542 assert(!UsrOI.isUnknown() && !PtrOI.isUnknown() &&
1543 "Don't look for constant values if the offset has already been "
1544 "determined to be unknown.");
1545
1546 if (!GEP->collectOffset(DL, BitWidth, VariableOffsets, ConstantOffset)) {
1547 UsrOI.setUnknown();
1548 return true;
1549 }
1550
1551 LLVM_DEBUG(dbgs() << "[AAPointerInfo] GEP offset is "
1552 << (VariableOffsets.empty() ? "" : "not") << " constant "
1553 << *GEP << "\n");
1554
1555 auto Union = PtrOI;
1556 Union.addToAll(ConstantOffset.getSExtValue());
1557
1558 // Each VI in VariableOffsets has a set of potential constant values. Every
1559 // combination of elements, picked one each from these sets, is separately
1560 // added to the original set of offsets, thus resulting in more offsets.
1561 for (const auto &VI : VariableOffsets) {
1562 auto *PotentialConstantsAA = A.getAAFor<AAPotentialConstantValues>(
1563 *this, IRPosition::value(*VI.first), DepClassTy::OPTIONAL);
1564 if (!PotentialConstantsAA || !PotentialConstantsAA->isValidState()) {
1565 UsrOI.setUnknown();
1566 return true;
1567 }
1568
1569 // UndefValue is treated as a zero, which leaves Union as is.
1570 if (PotentialConstantsAA->undefIsContained())
1571 continue;
1572
1573 // We need at least one constant in every set to compute an actual offset.
1574 // Otherwise, we end up pessimizing AAPointerInfo by respecting offsets that
1575 // don't actually exist. In other words, the absence of constant values
1576 // implies that the operation can be assumed dead for now.
1577 auto &AssumedSet = PotentialConstantsAA->getAssumedSet();
1578 if (AssumedSet.empty())
1579 return false;
1580
1581 OffsetInfo Product;
1582 for (const auto &ConstOffset : AssumedSet) {
1583 auto CopyPerOffset = Union;
1584 CopyPerOffset.addToAll(ConstOffset.getSExtValue() *
1585 VI.second.getZExtValue());
1586 Product.merge(CopyPerOffset);
1587 }
1588 Union = Product;
1589 }
1590
1591 UsrOI = std::move(Union);
1592 return true;
1593}
1594
1595ChangeStatus AAPointerInfoFloating::updateImpl(Attributor &A) {
1596 using namespace AA::PointerInfo;
1598 const DataLayout &DL = A.getDataLayout();
1599 Value &AssociatedValue = getAssociatedValue();
1600
1601 DenseMap<Value *, OffsetInfo> OffsetInfoMap;
1602 OffsetInfoMap[&AssociatedValue].insert(0);
1603
1604 auto HandlePassthroughUser = [&](Value *Usr, Value *CurPtr, bool &Follow) {
1605 // One does not simply walk into a map and assign a reference to a possibly
1606 // new location. That can cause an invalidation before the assignment
1607 // happens, like so:
1608 //
1609 // OffsetInfoMap[Usr] = OffsetInfoMap[CurPtr]; /* bad idea! */
1610 //
1611 // The RHS is a reference that may be invalidated by an insertion caused by
1612 // the LHS. So we ensure that the side-effect of the LHS happens first.
1613
1614 assert(OffsetInfoMap.contains(CurPtr) &&
1615 "CurPtr does not exist in the map!");
1616
1617 auto &UsrOI = OffsetInfoMap[Usr];
1618 auto &PtrOI = OffsetInfoMap[CurPtr];
1619 assert(!PtrOI.isUnassigned() &&
1620 "Cannot pass through if the input Ptr was not visited!");
1621 UsrOI.merge(PtrOI);
1622 Follow = true;
1623 return true;
1624 };
1625
1626 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
1627 Value *CurPtr = U.get();
1628 User *Usr = U.getUser();
1629 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Analyze " << *CurPtr << " in " << *Usr
1630 << "\n");
1631 assert(OffsetInfoMap.count(CurPtr) &&
1632 "The current pointer offset should have been seeded!");
1633 assert(!OffsetInfoMap[CurPtr].isUnassigned() &&
1634 "Current pointer should be assigned");
1635
1636 if (ConstantExpr *CE = dyn_cast<ConstantExpr>(Usr)) {
1637 if (CE->isCast())
1638 return HandlePassthroughUser(Usr, CurPtr, Follow);
1639 if (!isa<GEPOperator>(CE)) {
1640 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Unhandled constant user " << *CE
1641 << "\n");
1642 return false;
1643 }
1644 }
1645 if (auto *GEP = dyn_cast<GEPOperator>(Usr)) {
1646 // Note the order here, the Usr access might change the map, CurPtr is
1647 // already in it though.
1648 auto &UsrOI = OffsetInfoMap[Usr];
1649 auto &PtrOI = OffsetInfoMap[CurPtr];
1650
1651 if (UsrOI.isUnknown())
1652 return true;
1653
1654 if (PtrOI.isUnknown()) {
1655 Follow = true;
1656 UsrOI.setUnknown();
1657 return true;
1658 }
1659
1660 Follow = collectConstantsForGEP(A, DL, UsrOI, PtrOI, GEP);
1661 return true;
1662 }
1663 if (isa<PtrToIntInst>(Usr))
1664 return false;
1665 if (isa<CastInst>(Usr) || isa<SelectInst>(Usr))
1666 return HandlePassthroughUser(Usr, CurPtr, Follow);
1667 // Returns are allowed if they are in the associated functions. Users can
1668 // then check the call site return. Returns from other functions can't be
1669 // tracked and are cause for invalidation.
1670 if (auto *RI = dyn_cast<ReturnInst>(Usr)) {
1671 if (RI->getFunction() == getAssociatedFunction()) {
1672 auto &PtrOI = OffsetInfoMap[CurPtr];
1673 Changed |= setReachesReturn(PtrOI);
1674 return true;
1675 }
1676 return false;
1677 }
1678
1679 // For PHIs we need to take care of the recurrence explicitly as the value
1680 // might change while we iterate through a loop. For now, we give up if
1681 // the PHI is not invariant.
1682 if (auto *PHI = dyn_cast<PHINode>(Usr)) {
1683 // Note the order here, the Usr access might change the map, CurPtr is
1684 // already in it though.
1685 bool IsFirstPHIUser = !OffsetInfoMap.count(PHI);
1686 auto &UsrOI = OffsetInfoMap[PHI];
1687 auto &PtrOI = OffsetInfoMap[CurPtr];
1688
1689 // Check if the PHI operand has already an unknown offset as we can't
1690 // improve on that anymore.
1691 if (PtrOI.isUnknown()) {
1692 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand offset unknown "
1693 << *CurPtr << " in " << *PHI << "\n");
1694 Follow = !UsrOI.isUnknown();
1695 UsrOI.setUnknown();
1696 return true;
1697 }
1698
1699 // Check if the PHI is invariant (so far).
1700 if (UsrOI == PtrOI) {
1701 assert(!PtrOI.isUnassigned() &&
1702 "Cannot assign if the current Ptr was not visited!");
1703 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI is invariant (so far)");
1704 return true;
1705 }
1706
1707 // Check if the PHI operand can be traced back to AssociatedValue.
1708 APInt Offset(
1709 DL.getIndexSizeInBits(CurPtr->getType()->getPointerAddressSpace()),
1710 0);
1711 Value *CurPtrBase = CurPtr->stripAndAccumulateConstantOffsets(
1712 DL, Offset, /* AllowNonInbounds */ true);
1713 auto It = OffsetInfoMap.find(CurPtrBase);
1714 if (It == OffsetInfoMap.end()) {
1715 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand is too complex "
1716 << *CurPtr << " in " << *PHI
1717 << " (base: " << *CurPtrBase << ")\n");
1718 UsrOI.setUnknown();
1719 Follow = true;
1720 return true;
1721 }
1722
1723 // Check if the PHI operand is not dependent on the PHI itself. Every
1724 // recurrence is a cyclic net of PHIs in the data flow, and has an
1725 // equivalent Cycle in the control flow. One of those PHIs must be in the
1726 // header of that control flow Cycle. This is independent of the choice of
1727 // Cycles reported by CycleInfo. It is sufficient to check the PHIs in
1728 // every Cycle header; if such a node is marked unknown, this will
1729 // eventually propagate through the whole net of PHIs in the recurrence.
1730 const auto *CI =
1731 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
1732 *PHI->getFunction());
1733 if (mayBeInCycle(CI, cast<Instruction>(Usr), /* HeaderOnly */ true)) {
1734 auto BaseOI = It->getSecond();
1735 BaseOI.addToAll(Offset.getZExtValue());
1736 if (IsFirstPHIUser || BaseOI == UsrOI) {
1737 LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI is invariant " << *CurPtr
1738 << " in " << *Usr << "\n");
1739 return HandlePassthroughUser(Usr, CurPtr, Follow);
1740 }
1741
1742 LLVM_DEBUG(
1743 dbgs() << "[AAPointerInfo] PHI operand pointer offset mismatch "
1744 << *CurPtr << " in " << *PHI << "\n");
1745 UsrOI.setUnknown();
1746 Follow = true;
1747 return true;
1748 }
1749
1750 UsrOI.merge(PtrOI);
1751 Follow = true;
1752 return true;
1753 }
1754
1755 if (auto *LoadI = dyn_cast<LoadInst>(Usr)) {
1756 // If the access is to a pointer that may or may not be the associated
1757 // value, e.g. due to a PHI, we cannot assume it will be read.
1758 AccessKind AK = AccessKind::AK_R;
1759 if (getUnderlyingObject(CurPtr) == &AssociatedValue)
1760 AK = AccessKind(AK | AccessKind::AK_MUST);
1761 else
1762 AK = AccessKind(AK | AccessKind::AK_MAY);
1763 if (!handleAccess(A, *LoadI, /* Content */ nullptr, AK,
1764 OffsetInfoMap[CurPtr].Offsets, Changed,
1765 *LoadI->getType()))
1766 return false;
1767
1768 auto IsAssumption = [](Instruction &I) {
1769 if (auto *II = dyn_cast<IntrinsicInst>(&I))
1770 return II->isAssumeLikeIntrinsic();
1771 return false;
1772 };
1773
1774 auto IsImpactedInRange = [&](Instruction *FromI, Instruction *ToI) {
1775 // Check if the assumption and the load are executed together without
1776 // memory modification.
1777 do {
1778 if (FromI->mayWriteToMemory() && !IsAssumption(*FromI))
1779 return true;
1780 FromI = FromI->getNextNonDebugInstruction();
1781 } while (FromI && FromI != ToI);
1782 return false;
1783 };
1784
1785 BasicBlock *BB = LoadI->getParent();
1786 auto IsValidAssume = [&](IntrinsicInst &IntrI) {
1787 if (IntrI.getIntrinsicID() != Intrinsic::assume)
1788 return false;
1789 BasicBlock *IntrBB = IntrI.getParent();
1790 if (IntrI.getParent() == BB) {
1791 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(), &IntrI))
1792 return false;
1793 } else {
1794 auto PredIt = pred_begin(IntrBB);
1795 if (PredIt == pred_end(IntrBB))
1796 return false;
1797 if ((*PredIt) != BB)
1798 return false;
1799 if (++PredIt != pred_end(IntrBB))
1800 return false;
1801 for (auto *SuccBB : successors(BB)) {
1802 if (SuccBB == IntrBB)
1803 continue;
1804 if (isa<UnreachableInst>(SuccBB->getTerminator()))
1805 continue;
1806 return false;
1807 }
1808 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(),
1809 BB->getTerminator()))
1810 return false;
1811 if (IsImpactedInRange(&IntrBB->front(), &IntrI))
1812 return false;
1813 }
1814 return true;
1815 };
1816
1817 std::pair<Value *, IntrinsicInst *> Assumption;
1818 for (const Use &LoadU : LoadI->uses()) {
1819 if (auto *CmpI = dyn_cast<CmpInst>(LoadU.getUser())) {
1820 if (!CmpI->isEquality() || !CmpI->isTrueWhenEqual())
1821 continue;
1822 for (const Use &CmpU : CmpI->uses()) {
1823 if (auto *IntrI = dyn_cast<IntrinsicInst>(CmpU.getUser())) {
1824 if (!IsValidAssume(*IntrI))
1825 continue;
1826 int Idx = CmpI->getOperandUse(0) == LoadU;
1827 Assumption = {CmpI->getOperand(Idx), IntrI};
1828 break;
1829 }
1830 }
1831 }
1832 if (Assumption.first)
1833 break;
1834 }
1835
1836 // Check if we found an assumption associated with this load.
1837 if (!Assumption.first || !Assumption.second)
1838 return true;
1839
1840 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Assumption found "
1841 << *Assumption.second << ": " << *LoadI
1842 << " == " << *Assumption.first << "\n");
1843 bool UsedAssumedInformation = false;
1844 std::optional<Value *> Content = nullptr;
1845 if (Assumption.first)
1846 Content =
1847 A.getAssumedSimplified(*Assumption.first, *this,
1848 UsedAssumedInformation, AA::Interprocedural);
1849 return handleAccess(
1850 A, *Assumption.second, Content, AccessKind::AK_ASSUMPTION,
1851 OffsetInfoMap[CurPtr].Offsets, Changed, *LoadI->getType());
1852 }
1853
1854 auto HandleStoreLike = [&](Instruction &I, Value *ValueOp, Type &ValueTy,
1855 ArrayRef<Value *> OtherOps, AccessKind AK) {
1856 for (auto *OtherOp : OtherOps) {
1857 if (OtherOp == CurPtr) {
1858 LLVM_DEBUG(
1859 dbgs()
1860 << "[AAPointerInfo] Escaping use in store like instruction " << I
1861 << "\n");
1862 return false;
1863 }
1864 }
1865
1866 // If the access is to a pointer that may or may not be the associated
1867 // value, e.g. due to a PHI, we cannot assume it will be written.
1868 if (getUnderlyingObject(CurPtr) == &AssociatedValue)
1869 AK = AccessKind(AK | AccessKind::AK_MUST);
1870 else
1871 AK = AccessKind(AK | AccessKind::AK_MAY);
1872 bool UsedAssumedInformation = false;
1873 std::optional<Value *> Content = nullptr;
1874 if (ValueOp)
1875 Content = A.getAssumedSimplified(
1876 *ValueOp, *this, UsedAssumedInformation, AA::Interprocedural);
1877 return handleAccess(A, I, Content, AK, OffsetInfoMap[CurPtr].Offsets,
1878 Changed, ValueTy);
1879 };
1880
1881 if (auto *StoreI = dyn_cast<StoreInst>(Usr))
1882 return HandleStoreLike(*StoreI, StoreI->getValueOperand(),
1883 *StoreI->getValueOperand()->getType(),
1884 {StoreI->getValueOperand()}, AccessKind::AK_W);
1885 if (auto *RMWI = dyn_cast<AtomicRMWInst>(Usr))
1886 return HandleStoreLike(*RMWI, nullptr, *RMWI->getValOperand()->getType(),
1887 {RMWI->getValOperand()}, AccessKind::AK_RW);
1888 if (auto *CXI = dyn_cast<AtomicCmpXchgInst>(Usr))
1889 return HandleStoreLike(
1890 *CXI, nullptr, *CXI->getNewValOperand()->getType(),
1891 {CXI->getCompareOperand(), CXI->getNewValOperand()},
1892 AccessKind::AK_RW);
1893
1894 if (auto *CB = dyn_cast<CallBase>(Usr)) {
1895 if (CB->isLifetimeStartOrEnd())
1896 return true;
1897 const auto *TLI =
1898 A.getInfoCache().getTargetLibraryInfoForFunction(*CB->getFunction());
1899 if (getFreedOperand(CB, TLI) == U)
1900 return true;
1901 if (CB->isArgOperand(&U)) {
1902 unsigned ArgNo = CB->getArgOperandNo(&U);
1903 const auto *CSArgPI = A.getAAFor<AAPointerInfo>(
1904 *this, IRPosition::callsite_argument(*CB, ArgNo),
1906 if (!CSArgPI)
1907 return false;
1908 bool IsArgMustAcc = (getUnderlyingObject(CurPtr) == &AssociatedValue);
1909 Changed = translateAndAddState(A, *CSArgPI, OffsetInfoMap[CurPtr], *CB,
1910 IsArgMustAcc) |
1911 Changed;
1912 if (!CSArgPI->reachesReturn())
1913 return isValidState();
1914
1916 if (!Callee || Callee->arg_size() <= ArgNo)
1917 return false;
1918 bool UsedAssumedInformation = false;
1919 auto ReturnedValue = A.getAssumedSimplified(
1920 IRPosition::returned(*Callee), *this, UsedAssumedInformation,
1922 auto *ReturnedArg =
1923 dyn_cast_or_null<Argument>(ReturnedValue.value_or(nullptr));
1924 auto *Arg = Callee->getArg(ArgNo);
1925 if (ReturnedArg && Arg != ReturnedArg)
1926 return true;
1927 bool IsRetMustAcc = IsArgMustAcc && (ReturnedArg == Arg);
1928 const auto *CSRetPI = A.getAAFor<AAPointerInfo>(
1930 if (!CSRetPI)
1931 return false;
1932 OffsetInfo OI = OffsetInfoMap[CurPtr];
1933 CSArgPI->addReturnedOffsetsTo(OI);
1934 Changed =
1935 translateAndAddState(A, *CSRetPI, OI, *CB, IsRetMustAcc) | Changed;
1936 return isValidState();
1937 }
1938 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Call user not handled " << *CB
1939 << "\n");
1940 return false;
1941 }
1942
1943 LLVM_DEBUG(dbgs() << "[AAPointerInfo] User not handled " << *Usr << "\n");
1944 return false;
1945 };
1946 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
1947 assert(OffsetInfoMap.count(OldU) && "Old use should be known already!");
1948 assert(!OffsetInfoMap[OldU].isUnassigned() && "Old use should be assinged");
1949 if (OffsetInfoMap.count(NewU)) {
1950 LLVM_DEBUG({
1951 if (!(OffsetInfoMap[NewU] == OffsetInfoMap[OldU])) {
1952 dbgs() << "[AAPointerInfo] Equivalent use callback failed: "
1953 << OffsetInfoMap[NewU] << " vs " << OffsetInfoMap[OldU]
1954 << "\n";
1955 }
1956 });
1957 return OffsetInfoMap[NewU] == OffsetInfoMap[OldU];
1958 }
1959 bool Unused;
1960 return HandlePassthroughUser(NewU.get(), OldU.get(), Unused);
1961 };
1962 if (!A.checkForAllUses(UsePred, *this, AssociatedValue,
1963 /* CheckBBLivenessOnly */ true, DepClassTy::OPTIONAL,
1964 /* IgnoreDroppableUses */ true, EquivalentUseCB)) {
1965 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Check for all uses failed, abort!\n");
1966 return indicatePessimisticFixpoint();
1967 }
1968
1969 LLVM_DEBUG({
1970 dbgs() << "Accesses by bin after update:\n";
1971 dumpState(dbgs());
1972 });
1973
1974 return Changed;
1975}
1976
1977struct AAPointerInfoReturned final : AAPointerInfoImpl {
1978 AAPointerInfoReturned(const IRPosition &IRP, Attributor &A)
1979 : AAPointerInfoImpl(IRP, A) {}
1980
1981 /// See AbstractAttribute::updateImpl(...).
1982 ChangeStatus updateImpl(Attributor &A) override {
1983 return indicatePessimisticFixpoint();
1984 }
1985
1986 /// See AbstractAttribute::trackStatistics()
1987 void trackStatistics() const override {
1988 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1989 }
1990};
1991
1992struct AAPointerInfoArgument final : AAPointerInfoFloating {
1993 AAPointerInfoArgument(const IRPosition &IRP, Attributor &A)
1994 : AAPointerInfoFloating(IRP, A) {}
1995
1996 /// See AbstractAttribute::trackStatistics()
1997 void trackStatistics() const override {
1998 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
1999 }
2000};
2001
2002struct AAPointerInfoCallSiteArgument final : AAPointerInfoFloating {
2003 AAPointerInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
2004 : AAPointerInfoFloating(IRP, A) {}
2005
2006 /// See AbstractAttribute::updateImpl(...).
2007 ChangeStatus updateImpl(Attributor &A) override {
2008 using namespace AA::PointerInfo;
2009 // We handle memory intrinsics explicitly, at least the first (=
2010 // destination) and second (=source) arguments as we know how they are
2011 // accessed.
2012 if (auto *MI = dyn_cast_or_null<MemIntrinsic>(getCtxI())) {
2013 ConstantInt *Length = dyn_cast<ConstantInt>(MI->getLength());
2014 int64_t LengthVal = AA::RangeTy::Unknown;
2015 if (Length)
2016 LengthVal = Length->getSExtValue();
2017 unsigned ArgNo = getIRPosition().getCallSiteArgNo();
2018 ChangeStatus Changed = ChangeStatus::UNCHANGED;
2019 if (ArgNo > 1) {
2020 LLVM_DEBUG(dbgs() << "[AAPointerInfo] Unhandled memory intrinsic "
2021 << *MI << "\n");
2022 return indicatePessimisticFixpoint();
2023 } else {
2024 auto Kind =
2025 ArgNo == 0 ? AccessKind::AK_MUST_WRITE : AccessKind::AK_MUST_READ;
2026 Changed =
2027 Changed | addAccess(A, {0, LengthVal}, *MI, nullptr, Kind, nullptr);
2028 }
2029 LLVM_DEBUG({
2030 dbgs() << "Accesses by bin after update:\n";
2031 dumpState(dbgs());
2032 });
2033
2034 return Changed;
2035 }
2036
2037 // TODO: Once we have call site specific value information we can provide
2038 // call site specific liveness information and then it makes
2039 // sense to specialize attributes for call sites arguments instead of
2040 // redirecting requests to the callee argument.
2041 Argument *Arg = getAssociatedArgument();
2042 if (Arg) {
2043 const IRPosition &ArgPos = IRPosition::argument(*Arg);
2044 auto *ArgAA =
2045 A.getAAFor<AAPointerInfo>(*this, ArgPos, DepClassTy::REQUIRED);
2046 if (ArgAA && ArgAA->getState().isValidState())
2047 return translateAndAddStateFromCallee(A, *ArgAA,
2048 *cast<CallBase>(getCtxI()));
2049 if (!Arg->getParent()->isDeclaration())
2050 return indicatePessimisticFixpoint();
2051 }
2052
2053 bool IsKnownNoCapture;
2054 if (!AA::hasAssumedIRAttr<Attribute::NoCapture>(
2055 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNoCapture))
2056 return indicatePessimisticFixpoint();
2057
2058 bool IsKnown = false;
2059 if (AA::isAssumedReadNone(A, getIRPosition(), *this, IsKnown))
2060 return ChangeStatus::UNCHANGED;
2061 bool ReadOnly = AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown);
2062 auto Kind =
2063 ReadOnly ? AccessKind::AK_MAY_READ : AccessKind::AK_MAY_READ_WRITE;
2064 return addAccess(A, AA::RangeTy::getUnknown(), *getCtxI(), nullptr, Kind,
2065 nullptr);
2066 }
2067
2068 /// See AbstractAttribute::trackStatistics()
2069 void trackStatistics() const override {
2070 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
2071 }
2072};
2073
2074struct AAPointerInfoCallSiteReturned final : AAPointerInfoFloating {
2075 AAPointerInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
2076 : AAPointerInfoFloating(IRP, A) {}
2077
2078 /// See AbstractAttribute::trackStatistics()
2079 void trackStatistics() const override {
2080 AAPointerInfoImpl::trackPointerInfoStatistics(getIRPosition());
2081 }
2082};
2083} // namespace
2084
2085/// -----------------------NoUnwind Function Attribute--------------------------
2086
2087namespace {
2088struct AANoUnwindImpl : AANoUnwind {
2089 AANoUnwindImpl(const IRPosition &IRP, Attributor &A) : AANoUnwind(IRP, A) {}
2090
2091 /// See AbstractAttribute::initialize(...).
2092 void initialize(Attributor &A) override {
2093 bool IsKnown;
2094 assert(!AA::hasAssumedIRAttr<Attribute::NoUnwind>(
2095 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2096 (void)IsKnown;
2097 }
2098
2099 const std::string getAsStr(Attributor *A) const override {
2100 return getAssumed() ? "nounwind" : "may-unwind";
2101 }
2102
2103 /// See AbstractAttribute::updateImpl(...).
2104 ChangeStatus updateImpl(Attributor &A) override {
2105 auto Opcodes = {
2106 (unsigned)Instruction::Invoke, (unsigned)Instruction::CallBr,
2107 (unsigned)Instruction::Call, (unsigned)Instruction::CleanupRet,
2108 (unsigned)Instruction::CatchSwitch, (unsigned)Instruction::Resume};
2109
2110 auto CheckForNoUnwind = [&](Instruction &I) {
2111 if (!I.mayThrow(/* IncludePhaseOneUnwind */ true))
2112 return true;
2113
2114 if (const auto *CB = dyn_cast<CallBase>(&I)) {
2115 bool IsKnownNoUnwind;
2116 return AA::hasAssumedIRAttr<Attribute::NoUnwind>(
2117 A, this, IRPosition::callsite_function(*CB), DepClassTy::REQUIRED,
2118 IsKnownNoUnwind);
2119 }
2120 return false;
2121 };
2122
2123 bool UsedAssumedInformation = false;
2124 if (!A.checkForAllInstructions(CheckForNoUnwind, *this, Opcodes,
2125 UsedAssumedInformation))
2126 return indicatePessimisticFixpoint();
2127
2128 return ChangeStatus::UNCHANGED;
2129 }
2130};
2131
2132struct AANoUnwindFunction final : public AANoUnwindImpl {
2133 AANoUnwindFunction(const IRPosition &IRP, Attributor &A)
2134 : AANoUnwindImpl(IRP, A) {}
2135
2136 /// See AbstractAttribute::trackStatistics()
2137 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nounwind) }
2138};
2139
2140/// NoUnwind attribute deduction for a call sites.
2141struct AANoUnwindCallSite final
2142 : AACalleeToCallSite<AANoUnwind, AANoUnwindImpl> {
2143 AANoUnwindCallSite(const IRPosition &IRP, Attributor &A)
2144 : AACalleeToCallSite<AANoUnwind, AANoUnwindImpl>(IRP, A) {}
2145
2146 /// See AbstractAttribute::trackStatistics()
2147 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nounwind); }
2148};
2149} // namespace
2150
2151/// ------------------------ NoSync Function Attribute -------------------------
2152
2153bool AANoSync::isAlignedBarrier(const CallBase &CB, bool ExecutedAligned) {
2154 switch (CB.getIntrinsicID()) {
2155 case Intrinsic::nvvm_barrier0:
2156 case Intrinsic::nvvm_barrier0_and:
2157 case Intrinsic::nvvm_barrier0_or:
2158 case Intrinsic::nvvm_barrier0_popc:
2159 return true;
2160 case Intrinsic::amdgcn_s_barrier:
2161 if (ExecutedAligned)
2162 return true;
2163 break;
2164 default:
2165 break;
2166 }
2167 return hasAssumption(CB, KnownAssumptionString("ompx_aligned_barrier"));
2168}
2169
2171 if (!I->isAtomic())
2172 return false;
2173
2174 if (auto *FI = dyn_cast<FenceInst>(I))
2175 // All legal orderings for fence are stronger than monotonic.
2176 return FI->getSyncScopeID() != SyncScope::SingleThread;
2177 if (auto *AI = dyn_cast<AtomicCmpXchgInst>(I)) {
2178 // Unordered is not a legal ordering for cmpxchg.
2179 return (AI->getSuccessOrdering() != AtomicOrdering::Monotonic ||
2180 AI->getFailureOrdering() != AtomicOrdering::Monotonic);
2181 }
2182
2183 AtomicOrdering Ordering;
2184 switch (I->getOpcode()) {
2185 case Instruction::AtomicRMW:
2186 Ordering = cast<AtomicRMWInst>(I)->getOrdering();
2187 break;
2188 case Instruction::Store:
2189 Ordering = cast<StoreInst>(I)->getOrdering();
2190 break;
2191 case Instruction::Load:
2192 Ordering = cast<LoadInst>(I)->getOrdering();
2193 break;
2194 default:
2196 "New atomic operations need to be known in the attributor.");
2197 }
2198
2199 return (Ordering != AtomicOrdering::Unordered &&
2200 Ordering != AtomicOrdering::Monotonic);
2201}
2202
2203/// Return true if this intrinsic is nosync. This is only used for intrinsics
2204/// which would be nosync except that they have a volatile flag. All other
2205/// intrinsics are simply annotated with the nosync attribute in Intrinsics.td.
2207 if (auto *MI = dyn_cast<MemIntrinsic>(I))
2208 return !MI->isVolatile();
2209 return false;
2210}
2211
2212namespace {
2213struct AANoSyncImpl : AANoSync {
2214 AANoSyncImpl(const IRPosition &IRP, Attributor &A) : AANoSync(IRP, A) {}
2215
2216 /// See AbstractAttribute::initialize(...).
2217 void initialize(Attributor &A) override {
2218 bool IsKnown;
2219 assert(!AA::hasAssumedIRAttr<Attribute::NoSync>(A, nullptr, getIRPosition(),
2220 DepClassTy::NONE, IsKnown));
2221 (void)IsKnown;
2222 }
2223
2224 const std::string getAsStr(Attributor *A) const override {
2225 return getAssumed() ? "nosync" : "may-sync";
2226 }
2227
2228 /// See AbstractAttribute::updateImpl(...).
2229 ChangeStatus updateImpl(Attributor &A) override;
2230};
2231
2232ChangeStatus AANoSyncImpl::updateImpl(Attributor &A) {
2233
2234 auto CheckRWInstForNoSync = [&](Instruction &I) {
2235 return AA::isNoSyncInst(A, I, *this);
2236 };
2237
2238 auto CheckForNoSync = [&](Instruction &I) {
2239 // At this point we handled all read/write effects and they are all
2240 // nosync, so they can be skipped.
2241 if (I.mayReadOrWriteMemory())
2242 return true;
2243
2244 bool IsKnown;
2245 CallBase &CB = cast<CallBase>(I);
2246 if (AA::hasAssumedIRAttr<Attribute::NoSync>(
2248 IsKnown))
2249 return true;
2250
2251 // non-convergent and readnone imply nosync.
2252 return !CB.isConvergent();
2253 };
2254
2255 bool UsedAssumedInformation = false;
2256 if (!A.checkForAllReadWriteInstructions(CheckRWInstForNoSync, *this,
2257 UsedAssumedInformation) ||
2258 !A.checkForAllCallLikeInstructions(CheckForNoSync, *this,
2259 UsedAssumedInformation))
2260 return indicatePessimisticFixpoint();
2261
2263}
2264
2265struct AANoSyncFunction final : public AANoSyncImpl {
2266 AANoSyncFunction(const IRPosition &IRP, Attributor &A)
2267 : AANoSyncImpl(IRP, A) {}
2268
2269 /// See AbstractAttribute::trackStatistics()
2270 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nosync) }
2271};
2272
2273/// NoSync attribute deduction for a call sites.
2274struct AANoSyncCallSite final : AACalleeToCallSite<AANoSync, AANoSyncImpl> {
2275 AANoSyncCallSite(const IRPosition &IRP, Attributor &A)
2276 : AACalleeToCallSite<AANoSync, AANoSyncImpl>(IRP, A) {}
2277
2278 /// See AbstractAttribute::trackStatistics()
2279 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nosync); }
2280};
2281} // namespace
2282
2283/// ------------------------ No-Free Attributes ----------------------------
2284
2285namespace {
2286struct AANoFreeImpl : public AANoFree {
2287 AANoFreeImpl(const IRPosition &IRP, Attributor &A) : AANoFree(IRP, A) {}
2288
2289 /// See AbstractAttribute::initialize(...).
2290 void initialize(Attributor &A) override {
2291 bool IsKnown;
2292 assert(!AA::hasAssumedIRAttr<Attribute::NoFree>(A, nullptr, getIRPosition(),
2293 DepClassTy::NONE, IsKnown));
2294 (void)IsKnown;
2295 }
2296
2297 /// See AbstractAttribute::updateImpl(...).
2298 ChangeStatus updateImpl(Attributor &A) override {
2299 auto CheckForNoFree = [&](Instruction &I) {
2300 bool IsKnown;
2301 return AA::hasAssumedIRAttr<Attribute::NoFree>(
2302 A, this, IRPosition::callsite_function(cast<CallBase>(I)),
2303 DepClassTy::REQUIRED, IsKnown);
2304 };
2305
2306 bool UsedAssumedInformation = false;
2307 if (!A.checkForAllCallLikeInstructions(CheckForNoFree, *this,
2308 UsedAssumedInformation))
2309 return indicatePessimisticFixpoint();
2310 return ChangeStatus::UNCHANGED;
2311 }
2312
2313 /// See AbstractAttribute::getAsStr().
2314 const std::string getAsStr(Attributor *A) const override {
2315 return getAssumed() ? "nofree" : "may-free";
2316 }
2317};
2318
2319struct AANoFreeFunction final : public AANoFreeImpl {
2320 AANoFreeFunction(const IRPosition &IRP, Attributor &A)
2321 : AANoFreeImpl(IRP, A) {}
2322
2323 /// See AbstractAttribute::trackStatistics()
2324 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(nofree) }
2325};
2326
2327/// NoFree attribute deduction for a call sites.
2328struct AANoFreeCallSite final : AACalleeToCallSite<AANoFree, AANoFreeImpl> {
2329 AANoFreeCallSite(const IRPosition &IRP, Attributor &A)
2330 : AACalleeToCallSite<AANoFree, AANoFreeImpl>(IRP, A) {}
2331
2332 /// See AbstractAttribute::trackStatistics()
2333 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(nofree); }
2334};
2335
2336/// NoFree attribute for floating values.
2337struct AANoFreeFloating : AANoFreeImpl {
2338 AANoFreeFloating(const IRPosition &IRP, Attributor &A)
2339 : AANoFreeImpl(IRP, A) {}
2340
2341 /// See AbstractAttribute::trackStatistics()
2342 void trackStatistics() const override{STATS_DECLTRACK_FLOATING_ATTR(nofree)}
2343
2344 /// See Abstract Attribute::updateImpl(...).
2345 ChangeStatus updateImpl(Attributor &A) override {
2346 const IRPosition &IRP = getIRPosition();
2347
2348 bool IsKnown;
2349 if (AA::hasAssumedIRAttr<Attribute::NoFree>(A, this,
2351 DepClassTy::OPTIONAL, IsKnown))
2352 return ChangeStatus::UNCHANGED;
2353
2354 Value &AssociatedValue = getIRPosition().getAssociatedValue();
2355 auto Pred = [&](const Use &U, bool &Follow) -> bool {
2356 Instruction *UserI = cast<Instruction>(U.getUser());
2357 if (auto *CB = dyn_cast<CallBase>(UserI)) {
2358 if (CB->isBundleOperand(&U))
2359 return false;
2360 if (!CB->isArgOperand(&U))
2361 return true;
2362 unsigned ArgNo = CB->getArgOperandNo(&U);
2363
2364 bool IsKnown;
2365 return AA::hasAssumedIRAttr<Attribute::NoFree>(
2366 A, this, IRPosition::callsite_argument(*CB, ArgNo),
2367 DepClassTy::REQUIRED, IsKnown);
2368 }
2369
2370 if (isa<GetElementPtrInst>(UserI) || isa<PHINode>(UserI) ||
2371 isa<SelectInst>(UserI)) {
2372 Follow = true;
2373 return true;
2374 }
2375 if (isa<StoreInst>(UserI) || isa<LoadInst>(UserI))
2376 return true;
2377
2378 if (isa<ReturnInst>(UserI) && getIRPosition().isArgumentPosition())
2379 return true;
2380
2381 // Unknown user.
2382 return false;
2383 };
2384 if (!A.checkForAllUses(Pred, *this, AssociatedValue))
2385 return indicatePessimisticFixpoint();
2386
2387 return ChangeStatus::UNCHANGED;
2388 }
2389};
2390
2391/// NoFree attribute for a call site argument.
2392struct AANoFreeArgument final : AANoFreeFloating {
2393 AANoFreeArgument(const IRPosition &IRP, Attributor &A)
2394 : AANoFreeFloating(IRP, A) {}
2395
2396 /// See AbstractAttribute::trackStatistics()
2397 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nofree) }
2398};
2399
2400/// NoFree attribute for call site arguments.
2401struct AANoFreeCallSiteArgument final : AANoFreeFloating {
2402 AANoFreeCallSiteArgument(const IRPosition &IRP, Attributor &A)
2403 : AANoFreeFloating(IRP, A) {}
2404
2405 /// See AbstractAttribute::updateImpl(...).
2406 ChangeStatus updateImpl(Attributor &A) override {
2407 // TODO: Once we have call site specific value information we can provide
2408 // call site specific liveness information and then it makes
2409 // sense to specialize attributes for call sites arguments instead of
2410 // redirecting requests to the callee argument.
2411 Argument *Arg = getAssociatedArgument();
2412 if (!Arg)
2413 return indicatePessimisticFixpoint();
2414 const IRPosition &ArgPos = IRPosition::argument(*Arg);
2415 bool IsKnown;
2416 if (AA::hasAssumedIRAttr<Attribute::NoFree>(A, this, ArgPos,
2417 DepClassTy::REQUIRED, IsKnown))
2418 return ChangeStatus::UNCHANGED;
2419 return indicatePessimisticFixpoint();
2420 }
2421
2422 /// See AbstractAttribute::trackStatistics()
2423 void trackStatistics() const override{STATS_DECLTRACK_CSARG_ATTR(nofree)};
2424};
2425
2426/// NoFree attribute for function return value.
2427struct AANoFreeReturned final : AANoFreeFloating {
2428 AANoFreeReturned(const IRPosition &IRP, Attributor &A)
2429 : AANoFreeFloating(IRP, A) {
2430 llvm_unreachable("NoFree is not applicable to function returns!");
2431 }
2432
2433 /// See AbstractAttribute::initialize(...).
2434 void initialize(Attributor &A) override {
2435 llvm_unreachable("NoFree is not applicable to function returns!");
2436 }
2437
2438 /// See AbstractAttribute::updateImpl(...).
2439 ChangeStatus updateImpl(Attributor &A) override {
2440 llvm_unreachable("NoFree is not applicable to function returns!");
2441 }
2442
2443 /// See AbstractAttribute::trackStatistics()
2444 void trackStatistics() const override {}
2445};
2446
2447/// NoFree attribute deduction for a call site return value.
2448struct AANoFreeCallSiteReturned final : AANoFreeFloating {
2449 AANoFreeCallSiteReturned(const IRPosition &IRP, Attributor &A)
2450 : AANoFreeFloating(IRP, A) {}
2451
2452 ChangeStatus manifest(Attributor &A) override {
2453 return ChangeStatus::UNCHANGED;
2454 }
2455 /// See AbstractAttribute::trackStatistics()
2456 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(nofree) }
2457};
2458} // namespace
2459
2460/// ------------------------ NonNull Argument Attribute ------------------------
2461
2463 Attribute::AttrKind ImpliedAttributeKind,
2464 bool IgnoreSubsumingPositions) {
2466 AttrKinds.push_back(Attribute::NonNull);
2469 AttrKinds.push_back(Attribute::Dereferenceable);
2470 if (A.hasAttr(IRP, AttrKinds, IgnoreSubsumingPositions, Attribute::NonNull))
2471 return true;
2472
2473 DominatorTree *DT = nullptr;
2474 AssumptionCache *AC = nullptr;
2475 InformationCache &InfoCache = A.getInfoCache();
2476 if (const Function *Fn = IRP.getAnchorScope()) {
2477 if (!Fn->isDeclaration()) {
2480 }
2481 }
2482
2484 if (IRP.getPositionKind() != IRP_RETURNED) {
2485 Worklist.push_back({IRP.getAssociatedValue(), IRP.getCtxI()});
2486 } else {
2487 bool UsedAssumedInformation = false;
2488 if (!A.checkForAllInstructions(
2489 [&](Instruction &I) {
2490 Worklist.push_back({*cast<ReturnInst>(I).getReturnValue(), &I});
2491 return true;
2492 },
2493 IRP.getAssociatedFunction(), nullptr, {Instruction::Ret},
2494 UsedAssumedInformation, false, /*CheckPotentiallyDead=*/true))
2495 return false;
2496 }
2497
2498 if (llvm::any_of(Worklist, [&](AA::ValueAndContext VAC) {
2499 return !isKnownNonZero(
2500 VAC.getValue(),
2501 SimplifyQuery(A.getDataLayout(), DT, AC, VAC.getCtxI()));
2502 }))
2503 return false;
2504
2505 A.manifestAttrs(IRP, {Attribute::get(IRP.getAnchorValue().getContext(),
2506 Attribute::NonNull)});
2507 return true;
2508}
2509
2510namespace {
2511static int64_t getKnownNonNullAndDerefBytesForUse(
2512 Attributor &A, const AbstractAttribute &QueryingAA, Value &AssociatedValue,
2513 const Use *U, const Instruction *I, bool &IsNonNull, bool &TrackUse) {
2514 TrackUse = false;
2515
2516 const Value *UseV = U->get();
2517 if (!UseV->getType()->isPointerTy())
2518 return 0;
2519
2520 // We need to follow common pointer manipulation uses to the accesses they
2521 // feed into. We can try to be smart to avoid looking through things we do not
2522 // like for now, e.g., non-inbounds GEPs.
2523 if (isa<CastInst>(I)) {
2524 TrackUse = true;
2525 return 0;
2526 }
2527
2528 if (isa<GetElementPtrInst>(I)) {
2529 TrackUse = true;
2530 return 0;
2531 }
2532
2533 Type *PtrTy = UseV->getType();
2534 const Function *F = I->getFunction();
2537 const DataLayout &DL = A.getInfoCache().getDL();
2538 if (const auto *CB = dyn_cast<CallBase>(I)) {
2539 if (CB->isBundleOperand(U)) {
2541 U, {Attribute::NonNull, Attribute::Dereferenceable})) {
2542 IsNonNull |=
2543 (RK.AttrKind == Attribute::NonNull || !NullPointerIsDefined);
2544 return RK.ArgValue;
2545 }
2546 return 0;
2547 }
2548
2549 if (CB->isCallee(U)) {
2550 IsNonNull |= !NullPointerIsDefined;
2551 return 0;
2552 }
2553
2554 unsigned ArgNo = CB->getArgOperandNo(U);
2555 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
2556 // As long as we only use known information there is no need to track
2557 // dependences here.
2558 bool IsKnownNonNull;
2559 AA::hasAssumedIRAttr<Attribute::NonNull>(A, &QueryingAA, IRP,
2560 DepClassTy::NONE, IsKnownNonNull);
2561 IsNonNull |= IsKnownNonNull;
2562 auto *DerefAA =
2563 A.getAAFor<AADereferenceable>(QueryingAA, IRP, DepClassTy::NONE);
2564 return DerefAA ? DerefAA->getKnownDereferenceableBytes() : 0;
2565 }
2566
2567 std::optional<MemoryLocation> Loc = MemoryLocation::getOrNone(I);
2568 if (!Loc || Loc->Ptr != UseV || !Loc->Size.isPrecise() ||
2569 Loc->Size.isScalable() || I->isVolatile())
2570 return 0;
2571
2572 int64_t Offset;
2573 const Value *Base =
2574 getMinimalBaseOfPointer(A, QueryingAA, Loc->Ptr, Offset, DL);
2575 if (Base && Base == &AssociatedValue) {
2576 int64_t DerefBytes = Loc->Size.getValue() + Offset;
2577 IsNonNull |= !NullPointerIsDefined;
2578 return std::max(int64_t(0), DerefBytes);
2579 }
2580
2581 /// Corner case when an offset is 0.
2583 /*AllowNonInbounds*/ true);
2584 if (Base && Base == &AssociatedValue && Offset == 0) {
2585 int64_t DerefBytes = Loc->Size.getValue();
2586 IsNonNull |= !NullPointerIsDefined;
2587 return std::max(int64_t(0), DerefBytes);
2588 }
2589
2590 return 0;
2591}
2592
2593struct AANonNullImpl : AANonNull {
2594 AANonNullImpl(const IRPosition &IRP, Attributor &A) : AANonNull(IRP, A) {}
2595
2596 /// See AbstractAttribute::initialize(...).
2597 void initialize(Attributor &A) override {
2598 Value &V = *getAssociatedValue().stripPointerCasts();
2599 if (isa<ConstantPointerNull>(V)) {
2600 indicatePessimisticFixpoint();
2601 return;
2602 }
2603
2604 if (Instruction *CtxI = getCtxI())
2605 followUsesInMBEC(*this, A, getState(), *CtxI);
2606 }
2607
2608 /// See followUsesInMBEC
2609 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
2610 AANonNull::StateType &State) {
2611 bool IsNonNull = false;
2612 bool TrackUse = false;
2613 getKnownNonNullAndDerefBytesForUse(A, *this, getAssociatedValue(), U, I,
2614 IsNonNull, TrackUse);
2615 State.setKnown(IsNonNull);
2616 return TrackUse;
2617 }
2618
2619 /// See AbstractAttribute::getAsStr().
2620 const std::string getAsStr(Attributor *A) const override {
2621 return getAssumed() ? "nonnull" : "may-null";
2622 }
2623};
2624
2625/// NonNull attribute for a floating value.
2626struct AANonNullFloating : public AANonNullImpl {
2627 AANonNullFloating(const IRPosition &IRP, Attributor &A)
2628 : AANonNullImpl(IRP, A) {}
2629
2630 /// See AbstractAttribute::updateImpl(...).
2631 ChangeStatus updateImpl(Attributor &A) override {
2632 auto CheckIRP = [&](const IRPosition &IRP) {
2633 bool IsKnownNonNull;
2634 return AA::hasAssumedIRAttr<Attribute::NonNull>(
2635 A, *this, IRP, DepClassTy::OPTIONAL, IsKnownNonNull);
2636 };
2637
2638 bool Stripped;
2639 bool UsedAssumedInformation = false;
2640 Value *AssociatedValue = &getAssociatedValue();
2642 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
2643 AA::AnyScope, UsedAssumedInformation))
2644 Stripped = false;
2645 else
2646 Stripped =
2647 Values.size() != 1 || Values.front().getValue() != AssociatedValue;
2648
2649 if (!Stripped) {
2650 bool IsKnown;
2651 if (auto *PHI = dyn_cast<PHINode>(AssociatedValue))
2652 if (llvm::all_of(PHI->incoming_values(), [&](Value *Op) {
2653 return AA::hasAssumedIRAttr<Attribute::NonNull>(
2654 A, this, IRPosition::value(*Op), DepClassTy::OPTIONAL,
2655 IsKnown);
2656 }))
2657 return ChangeStatus::UNCHANGED;
2658 if (auto *Select = dyn_cast<SelectInst>(AssociatedValue))
2659 if (AA::hasAssumedIRAttr<Attribute::NonNull>(
2660 A, this, IRPosition::value(*Select->getFalseValue()),
2661 DepClassTy::OPTIONAL, IsKnown) &&
2662 AA::hasAssumedIRAttr<Attribute::NonNull>(
2663 A, this, IRPosition::value(*Select->getTrueValue()),
2664 DepClassTy::OPTIONAL, IsKnown))
2665 return ChangeStatus::UNCHANGED;
2666
2667 // If we haven't stripped anything we might still be able to use a
2668 // different AA, but only if the IRP changes. Effectively when we
2669 // interpret this not as a call site value but as a floating/argument
2670 // value.
2671 const IRPosition AVIRP = IRPosition::value(*AssociatedValue);
2672 if (AVIRP == getIRPosition() || !CheckIRP(AVIRP))
2673 return indicatePessimisticFixpoint();
2674 return ChangeStatus::UNCHANGED;
2675 }
2676
2677 for (const auto &VAC : Values)
2678 if (!CheckIRP(IRPosition::value(*VAC.getValue())))
2679 return indicatePessimisticFixpoint();
2680
2681 return ChangeStatus::UNCHANGED;
2682 }
2683
2684 /// See AbstractAttribute::trackStatistics()
2685 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(nonnull) }
2686};
2687
2688/// NonNull attribute for function return value.
2689struct AANonNullReturned final
2690 : AAReturnedFromReturnedValues<AANonNull, AANonNull, AANonNull::StateType,
2691 false, AANonNull::IRAttributeKind, false> {
2692 AANonNullReturned(const IRPosition &IRP, Attributor &A)
2693 : AAReturnedFromReturnedValues<AANonNull, AANonNull, AANonNull::StateType,
2694 false, Attribute::NonNull, false>(IRP, A) {
2695 }
2696
2697 /// See AbstractAttribute::getAsStr().
2698 const std::string getAsStr(Attributor *A) const override {
2699 return getAssumed() ? "nonnull" : "may-null";
2700 }
2701
2702 /// See AbstractAttribute::trackStatistics()
2703 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(nonnull) }
2704};
2705
2706/// NonNull attribute for function argument.
2707struct AANonNullArgument final
2708 : AAArgumentFromCallSiteArguments<AANonNull, AANonNullImpl> {
2709 AANonNullArgument(const IRPosition &IRP, Attributor &A)
2710 : AAArgumentFromCallSiteArguments<AANonNull, AANonNullImpl>(IRP, A) {}
2711
2712 /// See AbstractAttribute::trackStatistics()
2713 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nonnull) }
2714};
2715
2716struct AANonNullCallSiteArgument final : AANonNullFloating {
2717 AANonNullCallSiteArgument(const IRPosition &IRP, Attributor &A)
2718 : AANonNullFloating(IRP, A) {}
2719
2720 /// See AbstractAttribute::trackStatistics()
2721 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(nonnull) }
2722};
2723
2724/// NonNull attribute for a call site return position.
2725struct AANonNullCallSiteReturned final
2726 : AACalleeToCallSite<AANonNull, AANonNullImpl> {
2727 AANonNullCallSiteReturned(const IRPosition &IRP, Attributor &A)
2728 : AACalleeToCallSite<AANonNull, AANonNullImpl>(IRP, A) {}
2729
2730 /// See AbstractAttribute::trackStatistics()
2731 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(nonnull) }
2732};
2733} // namespace
2734
2735/// ------------------------ Must-Progress Attributes --------------------------
2736namespace {
2737struct AAMustProgressImpl : public AAMustProgress {
2738 AAMustProgressImpl(const IRPosition &IRP, Attributor &A)
2739 : AAMustProgress(IRP, A) {}
2740
2741 /// See AbstractAttribute::initialize(...).
2742 void initialize(Attributor &A) override {
2743 bool IsKnown;
2744 assert(!AA::hasAssumedIRAttr<Attribute::MustProgress>(
2745 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2746 (void)IsKnown;
2747 }
2748
2749 /// See AbstractAttribute::getAsStr()
2750 const std::string getAsStr(Attributor *A) const override {
2751 return getAssumed() ? "mustprogress" : "may-not-progress";
2752 }
2753};
2754
2755struct AAMustProgressFunction final : AAMustProgressImpl {
2756 AAMustProgressFunction(const IRPosition &IRP, Attributor &A)
2757 : AAMustProgressImpl(IRP, A) {}
2758
2759 /// See AbstractAttribute::updateImpl(...).
2760 ChangeStatus updateImpl(Attributor &A) override {
2761 bool IsKnown;
2762 if (AA::hasAssumedIRAttr<Attribute::WillReturn>(
2763 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnown)) {
2764 if (IsKnown)
2765 return indicateOptimisticFixpoint();
2766 return ChangeStatus::UNCHANGED;
2767 }
2768
2769 auto CheckForMustProgress = [&](AbstractCallSite ACS) {
2770 IRPosition IPos = IRPosition::callsite_function(*ACS.getInstruction());
2771 bool IsKnownMustProgress;
2772 return AA::hasAssumedIRAttr<Attribute::MustProgress>(
2773 A, this, IPos, DepClassTy::REQUIRED, IsKnownMustProgress,
2774 /* IgnoreSubsumingPositions */ true);
2775 };
2776
2777 bool AllCallSitesKnown = true;
2778 if (!A.checkForAllCallSites(CheckForMustProgress, *this,
2779 /* RequireAllCallSites */ true,
2780 AllCallSitesKnown))
2781 return indicatePessimisticFixpoint();
2782
2783 return ChangeStatus::UNCHANGED;
2784 }
2785
2786 /// See AbstractAttribute::trackStatistics()
2787 void trackStatistics() const override {
2788 STATS_DECLTRACK_FN_ATTR(mustprogress)
2789 }
2790};
2791
2792/// MustProgress attribute deduction for a call sites.
2793struct AAMustProgressCallSite final : AAMustProgressImpl {
2794 AAMustProgressCallSite(const IRPosition &IRP, Attributor &A)
2795 : AAMustProgressImpl(IRP, A) {}
2796
2797 /// See AbstractAttribute::updateImpl(...).
2798 ChangeStatus updateImpl(Attributor &A) override {
2799 // TODO: Once we have call site specific value information we can provide
2800 // call site specific liveness information and then it makes
2801 // sense to specialize attributes for call sites arguments instead of
2802 // redirecting requests to the callee argument.
2803 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
2804 bool IsKnownMustProgress;
2805 if (!AA::hasAssumedIRAttr<Attribute::MustProgress>(
2806 A, this, FnPos, DepClassTy::REQUIRED, IsKnownMustProgress))
2807 return indicatePessimisticFixpoint();
2808 return ChangeStatus::UNCHANGED;
2809 }
2810
2811 /// See AbstractAttribute::trackStatistics()
2812 void trackStatistics() const override {
2813 STATS_DECLTRACK_CS_ATTR(mustprogress);
2814 }
2815};
2816} // namespace
2817
2818/// ------------------------ No-Recurse Attributes ----------------------------
2819
2820namespace {
2821struct AANoRecurseImpl : public AANoRecurse {
2822 AANoRecurseImpl(const IRPosition &IRP, Attributor &A) : AANoRecurse(IRP, A) {}
2823
2824 /// See AbstractAttribute::initialize(...).
2825 void initialize(Attributor &A) override {
2826 bool IsKnown;
2827 assert(!AA::hasAssumedIRAttr<Attribute::NoRecurse>(
2828 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
2829 (void)IsKnown;
2830 }
2831
2832 /// See AbstractAttribute::getAsStr()
2833 const std::string getAsStr(Attributor *A) const override {
2834 return getAssumed() ? "norecurse" : "may-recurse";
2835 }
2836};
2837
2838struct AANoRecurseFunction final : AANoRecurseImpl {
2839 AANoRecurseFunction(const IRPosition &IRP, Attributor &A)
2840 : AANoRecurseImpl(IRP, A) {}
2841
2842 /// See AbstractAttribute::updateImpl(...).
2843 ChangeStatus updateImpl(Attributor &A) override {
2844
2845 // If all live call sites are known to be no-recurse, we are as well.
2846 auto CallSitePred = [&](AbstractCallSite ACS) {
2847 bool IsKnownNoRecurse;
2848 if (!AA::hasAssumedIRAttr<Attribute::NoRecurse>(
2849 A, this,
2850 IRPosition::function(*ACS.getInstruction()->getFunction()),
2851 DepClassTy::NONE, IsKnownNoRecurse))
2852 return false;
2853 return IsKnownNoRecurse;
2854 };
2855 bool UsedAssumedInformation = false;
2856 if (A.checkForAllCallSites(CallSitePred, *this, true,
2857 UsedAssumedInformation)) {
2858 // If we know all call sites and all are known no-recurse, we are done.
2859 // If all known call sites, which might not be all that exist, are known
2860 // to be no-recurse, we are not done but we can continue to assume
2861 // no-recurse. If one of the call sites we have not visited will become
2862 // live, another update is triggered.
2863 if (!UsedAssumedInformation)
2864 indicateOptimisticFixpoint();
2865 return ChangeStatus::UNCHANGED;
2866 }
2867
2868 const AAInterFnReachability *EdgeReachability =
2869 A.getAAFor<AAInterFnReachability>(*this, getIRPosition(),
2870 DepClassTy::REQUIRED);
2871 if (EdgeReachability && EdgeReachability->canReach(A, *getAnchorScope()))
2872 return indicatePessimisticFixpoint();
2873 return ChangeStatus::UNCHANGED;
2874 }
2875
2876 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(norecurse) }
2877};
2878
2879/// NoRecurse attribute deduction for a call sites.
2880struct AANoRecurseCallSite final
2881 : AACalleeToCallSite<AANoRecurse, AANoRecurseImpl> {
2882 AANoRecurseCallSite(const IRPosition &IRP, Attributor &A)
2883 : AACalleeToCallSite<AANoRecurse, AANoRecurseImpl>(IRP, A) {}
2884
2885 /// See AbstractAttribute::trackStatistics()
2886 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(norecurse); }
2887};
2888} // namespace
2889
2890/// ------------------------ No-Convergent Attribute --------------------------
2891
2892namespace {
2893struct AANonConvergentImpl : public AANonConvergent {
2894 AANonConvergentImpl(const IRPosition &IRP, Attributor &A)
2895 : AANonConvergent(IRP, A) {}
2896
2897 /// See AbstractAttribute::getAsStr()
2898 const std::string getAsStr(Attributor *A) const override {
2899 return getAssumed() ? "non-convergent" : "may-be-convergent";
2900 }
2901};
2902
2903struct AANonConvergentFunction final : AANonConvergentImpl {
2904 AANonConvergentFunction(const IRPosition &IRP, Attributor &A)
2905 : AANonConvergentImpl(IRP, A) {}
2906
2907 /// See AbstractAttribute::updateImpl(...).
2908 ChangeStatus updateImpl(Attributor &A) override {
2909 // If all function calls are known to not be convergent, we are not
2910 // convergent.
2911 auto CalleeIsNotConvergent = [&](Instruction &Inst) {
2912 CallBase &CB = cast<CallBase>(Inst);
2913 auto *Callee = dyn_cast_if_present<Function>(CB.getCalledOperand());
2914 if (!Callee || Callee->isIntrinsic()) {
2915 return false;
2916 }
2917 if (Callee->isDeclaration()) {
2918 return !Callee->hasFnAttribute(Attribute::Convergent);
2919 }
2920 const auto *ConvergentAA = A.getAAFor<AANonConvergent>(
2921 *this, IRPosition::function(*Callee), DepClassTy::REQUIRED);
2922 return ConvergentAA && ConvergentAA->isAssumedNotConvergent();
2923 };
2924
2925 bool UsedAssumedInformation = false;
2926 if (!A.checkForAllCallLikeInstructions(CalleeIsNotConvergent, *this,
2927 UsedAssumedInformation)) {
2928 return indicatePessimisticFixpoint();
2929 }
2930 return ChangeStatus::UNCHANGED;
2931 }
2932
2933 ChangeStatus manifest(Attributor &A) override {
2934 if (isKnownNotConvergent() &&
2935 A.hasAttr(getIRPosition(), Attribute::Convergent)) {
2936 A.removeAttrs(getIRPosition(), {Attribute::Convergent});
2937 return ChangeStatus::CHANGED;
2938 }
2939 return ChangeStatus::UNCHANGED;
2940 }
2941
2942 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(convergent) }
2943};
2944} // namespace
2945
2946/// -------------------- Undefined-Behavior Attributes ------------------------
2947
2948namespace {
2949struct AAUndefinedBehaviorImpl : public AAUndefinedBehavior {
2950 AAUndefinedBehaviorImpl(const IRPosition &IRP, Attributor &A)
2951 : AAUndefinedBehavior(IRP, A) {}
2952
2953 /// See AbstractAttribute::updateImpl(...).
2954 // through a pointer (i.e. also branches etc.)
2955 ChangeStatus updateImpl(Attributor &A) override {
2956 const size_t UBPrevSize = KnownUBInsts.size();
2957 const size_t NoUBPrevSize = AssumedNoUBInsts.size();
2958
2959 auto InspectMemAccessInstForUB = [&](Instruction &I) {
2960 // Lang ref now states volatile store is not UB, let's skip them.
2961 if (I.isVolatile() && I.mayWriteToMemory())
2962 return true;
2963
2964 // Skip instructions that are already saved.
2965 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
2966 return true;
2967
2968 // If we reach here, we know we have an instruction
2969 // that accesses memory through a pointer operand,
2970 // for which getPointerOperand() should give it to us.
2971 Value *PtrOp =
2972 const_cast<Value *>(getPointerOperand(&I, /* AllowVolatile */ true));
2973 assert(PtrOp &&
2974 "Expected pointer operand of memory accessing instruction");
2975
2976 // Either we stopped and the appropriate action was taken,
2977 // or we got back a simplified value to continue.
2978 std::optional<Value *> SimplifiedPtrOp =
2979 stopOnUndefOrAssumed(A, PtrOp, &I);
2980 if (!SimplifiedPtrOp || !*SimplifiedPtrOp)
2981 return true;
2982 const Value *PtrOpVal = *SimplifiedPtrOp;
2983
2984 // A memory access through a pointer is considered UB
2985 // only if the pointer has constant null value.
2986 // TODO: Expand it to not only check constant values.
2987 if (!isa<ConstantPointerNull>(PtrOpVal)) {
2988 AssumedNoUBInsts.insert(&I);
2989 return true;
2990 }
2991 const Type *PtrTy = PtrOpVal->getType();
2992
2993 // Because we only consider instructions inside functions,
2994 // assume that a parent function exists.
2995 const Function *F = I.getFunction();
2996
2997 // A memory access using constant null pointer is only considered UB
2998 // if null pointer is _not_ defined for the target platform.
3000 AssumedNoUBInsts.insert(&I);
3001 else
3002 KnownUBInsts.insert(&I);
3003 return true;
3004 };
3005
3006 auto InspectBrInstForUB = [&](Instruction &I) {
3007 // A conditional branch instruction is considered UB if it has `undef`
3008 // condition.
3009
3010 // Skip instructions that are already saved.
3011 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
3012 return true;
3013
3014 // We know we have a branch instruction.
3015 auto *BrInst = cast<BranchInst>(&I);
3016
3017 // Unconditional branches are never considered UB.
3018 if (BrInst->isUnconditional())
3019 return true;
3020
3021 // Either we stopped and the appropriate action was taken,
3022 // or we got back a simplified value to continue.
3023 std::optional<Value *> SimplifiedCond =
3024 stopOnUndefOrAssumed(A, BrInst->getCondition(), BrInst);
3025 if (!SimplifiedCond || !*SimplifiedCond)
3026 return true;
3027 AssumedNoUBInsts.insert(&I);
3028 return true;
3029 };
3030
3031 auto InspectCallSiteForUB = [&](Instruction &I) {
3032 // Check whether a callsite always cause UB or not
3033
3034 // Skip instructions that are already saved.
3035 if (AssumedNoUBInsts.count(&I) || KnownUBInsts.count(&I))
3036 return true;
3037
3038 // Check nonnull and noundef argument attribute violation for each
3039 // callsite.
3040 CallBase &CB = cast<CallBase>(I);
3041 auto *Callee = dyn_cast_if_present<Function>(CB.getCalledOperand());
3042 if (!Callee)
3043 return true;
3044 for (unsigned idx = 0; idx < CB.arg_size(); idx++) {
3045 // If current argument is known to be simplified to null pointer and the
3046 // corresponding argument position is known to have nonnull attribute,
3047 // the argument is poison. Furthermore, if the argument is poison and
3048 // the position is known to have noundef attriubte, this callsite is
3049 // considered UB.
3050 if (idx >= Callee->arg_size())
3051 break;
3052 Value *ArgVal = CB.getArgOperand(idx);
3053 if (!ArgVal)
3054 continue;
3055 // Here, we handle three cases.
3056 // (1) Not having a value means it is dead. (we can replace the value
3057 // with undef)
3058 // (2) Simplified to undef. The argument violate noundef attriubte.
3059 // (3) Simplified to null pointer where known to be nonnull.
3060 // The argument is a poison value and violate noundef attribute.
3061 IRPosition CalleeArgumentIRP = IRPosition::callsite_argument(CB, idx);
3062 bool IsKnownNoUndef;
3063 AA::hasAssumedIRAttr<Attribute::NoUndef>(
3064 A, this, CalleeArgumentIRP, DepClassTy::NONE, IsKnownNoUndef);
3065 if (!IsKnownNoUndef)
3066 continue;
3067 bool UsedAssumedInformation = false;
3068 std::optional<Value *> SimplifiedVal =
3069 A.getAssumedSimplified(IRPosition::value(*ArgVal), *this,
3070 UsedAssumedInformation, AA::Interprocedural);
3071 if (UsedAssumedInformation)
3072 continue;
3073 if (SimplifiedVal && !*SimplifiedVal)
3074 return true;
3075 if (!SimplifiedVal || isa<UndefValue>(**SimplifiedVal)) {
3076 KnownUBInsts.insert(&I);
3077 continue;
3078 }
3079 if (!ArgVal->getType()->isPointerTy() ||
3080 !isa<ConstantPointerNull>(**SimplifiedVal))
3081 continue;
3082 bool IsKnownNonNull;
3083 AA::hasAssumedIRAttr<Attribute::NonNull>(
3084 A, this, CalleeArgumentIRP, DepClassTy::NONE, IsKnownNonNull);
3085 if (IsKnownNonNull)
3086 KnownUBInsts.insert(&I);
3087 }
3088 return true;
3089 };
3090
3091 auto InspectReturnInstForUB = [&](Instruction &I) {
3092 auto &RI = cast<ReturnInst>(I);
3093 // Either we stopped and the appropriate action was taken,
3094 // or we got back a simplified return value to continue.
3095 std::optional<Value *> SimplifiedRetValue =
3096 stopOnUndefOrAssumed(A, RI.getReturnValue(), &I);
3097 if (!SimplifiedRetValue || !*SimplifiedRetValue)
3098 return true;
3099
3100 // Check if a return instruction always cause UB or not
3101 // Note: It is guaranteed that the returned position of the anchor
3102 // scope has noundef attribute when this is called.
3103 // We also ensure the return position is not "assumed dead"
3104 // because the returned value was then potentially simplified to
3105 // `undef` in AAReturnedValues without removing the `noundef`
3106 // attribute yet.
3107
3108 // When the returned position has noundef attriubte, UB occurs in the
3109 // following cases.
3110 // (1) Returned value is known to be undef.
3111 // (2) The value is known to be a null pointer and the returned
3112 // position has nonnull attribute (because the returned value is
3113 // poison).
3114 if (isa<ConstantPointerNull>(*SimplifiedRetValue)) {
3115 bool IsKnownNonNull;
3116 AA::hasAssumedIRAttr<Attribute::NonNull>(
3117 A, this, IRPosition::returned(*getAnchorScope()), DepClassTy::NONE,
3118 IsKnownNonNull);
3119 if (IsKnownNonNull)
3120 KnownUBInsts.insert(&I);
3121 }
3122
3123 return true;
3124 };
3125
3126 bool UsedAssumedInformation = false;
3127 A.checkForAllInstructions(InspectMemAccessInstForUB, *this,
3128 {Instruction::Load, Instruction::Store,
3129 Instruction::AtomicCmpXchg,
3130 Instruction::AtomicRMW},
3131 UsedAssumedInformation,
3132 /* CheckBBLivenessOnly */ true);
3133 A.checkForAllInstructions(InspectBrInstForUB, *this, {Instruction::Br},
3134 UsedAssumedInformation,
3135 /* CheckBBLivenessOnly */ true);
3136 A.checkForAllCallLikeInstructions(InspectCallSiteForUB, *this,
3137 UsedAssumedInformation);
3138
3139 // If the returned position of the anchor scope has noundef attriubte, check
3140 // all returned instructions.
3141 if (!getAnchorScope()->getReturnType()->isVoidTy()) {
3142 const IRPosition &ReturnIRP = IRPosition::returned(*getAnchorScope());
3143 if (!A.isAssumedDead(ReturnIRP, this, nullptr, UsedAssumedInformation)) {
3144 bool IsKnownNoUndef;
3145 AA::hasAssumedIRAttr<Attribute::NoUndef>(
3146 A, this, ReturnIRP, DepClassTy::NONE, IsKnownNoUndef);
3147 if (IsKnownNoUndef)
3148 A.checkForAllInstructions(InspectReturnInstForUB, *this,
3149 {Instruction::Ret}, UsedAssumedInformation,
3150 /* CheckBBLivenessOnly */ true);
3151 }
3152 }
3153
3154 if (NoUBPrevSize != AssumedNoUBInsts.size() ||
3155 UBPrevSize != KnownUBInsts.size())
3156 return ChangeStatus::CHANGED;
3157 return ChangeStatus::UNCHANGED;
3158 }
3159
3160 bool isKnownToCauseUB(Instruction *I) const override {
3161 return KnownUBInsts.count(I);
3162 }
3163
3164 bool isAssumedToCauseUB(Instruction *I) const override {
3165 // In simple words, if an instruction is not in the assumed to _not_
3166 // cause UB, then it is assumed UB (that includes those
3167 // in the KnownUBInsts set). The rest is boilerplate
3168 // is to ensure that it is one of the instructions we test
3169 // for UB.
3170
3171 switch (I->getOpcode()) {
3172 case Instruction::Load:
3173 case Instruction::Store:
3174 case Instruction::AtomicCmpXchg:
3175 case Instruction::AtomicRMW:
3176 return !AssumedNoUBInsts.count(I);
3177 case Instruction::Br: {
3178 auto *BrInst = cast<BranchInst>(I);
3179 if (BrInst->isUnconditional())
3180 return false;
3181 return !AssumedNoUBInsts.count(I);
3182 } break;
3183 default:
3184 return false;
3185 }
3186 return false;
3187 }
3188
3189 ChangeStatus manifest(Attributor &A) override {
3190 if (KnownUBInsts.empty())
3191 return ChangeStatus::UNCHANGED;
3192 for (Instruction *I : KnownUBInsts)
3193 A.changeToUnreachableAfterManifest(I);
3194 return ChangeStatus::CHANGED;
3195 }
3196
3197 /// See AbstractAttribute::getAsStr()
3198 const std::string getAsStr(Attributor *A) const override {
3199 return getAssumed() ? "undefined-behavior" : "no-ub";
3200 }
3201
3202 /// Note: The correctness of this analysis depends on the fact that the
3203 /// following 2 sets will stop changing after some point.
3204 /// "Change" here means that their size changes.
3205 /// The size of each set is monotonically increasing
3206 /// (we only add items to them) and it is upper bounded by the number of
3207 /// instructions in the processed function (we can never save more
3208 /// elements in either set than this number). Hence, at some point,
3209 /// they will stop increasing.
3210 /// Consequently, at some point, both sets will have stopped
3211 /// changing, effectively making the analysis reach a fixpoint.
3212
3213 /// Note: These 2 sets are disjoint and an instruction can be considered
3214 /// one of 3 things:
3215 /// 1) Known to cause UB (AAUndefinedBehavior could prove it) and put it in
3216 /// the KnownUBInsts set.
3217 /// 2) Assumed to cause UB (in every updateImpl, AAUndefinedBehavior
3218 /// has a reason to assume it).
3219 /// 3) Assumed to not cause UB. very other instruction - AAUndefinedBehavior
3220 /// could not find a reason to assume or prove that it can cause UB,
3221 /// hence it assumes it doesn't. We have a set for these instructions
3222 /// so that we don't reprocess them in every update.
3223 /// Note however that instructions in this set may cause UB.
3224
3225protected:
3226 /// A set of all live instructions _known_ to cause UB.
3227 SmallPtrSet<Instruction *, 8> KnownUBInsts;
3228
3229private:
3230 /// A set of all the (live) instructions that are assumed to _not_ cause UB.
3231 SmallPtrSet<Instruction *, 8> AssumedNoUBInsts;
3232
3233 // Should be called on updates in which if we're processing an instruction
3234 // \p I that depends on a value \p V, one of the following has to happen:
3235 // - If the value is assumed, then stop.
3236 // - If the value is known but undef, then consider it UB.
3237 // - Otherwise, do specific processing with the simplified value.
3238 // We return std::nullopt in the first 2 cases to signify that an appropriate
3239 // action was taken and the caller should stop.
3240 // Otherwise, we return the simplified value that the caller should
3241 // use for specific processing.
3242 std::optional<Value *> stopOnUndefOrAssumed(Attributor &A, Value *V,
3243 Instruction *I) {
3244 bool UsedAssumedInformation = false;
3245 std::optional<Value *> SimplifiedV =
3246 A.getAssumedSimplified(IRPosition::value(*V), *this,
3247 UsedAssumedInformation, AA::Interprocedural);
3248 if (!UsedAssumedInformation) {
3249 // Don't depend on assumed values.
3250 if (!SimplifiedV) {
3251 // If it is known (which we tested above) but it doesn't have a value,
3252 // then we can assume `undef` and hence the instruction is UB.
3253 KnownUBInsts.insert(I);
3254 return std::nullopt;
3255 }
3256 if (!*SimplifiedV)
3257 return nullptr;
3258 V = *SimplifiedV;
3259 }
3260 if (isa<UndefValue>(V)) {
3261 KnownUBInsts.insert(I);
3262 return std::nullopt;
3263 }
3264 return V;
3265 }
3266};
3267
3268struct AAUndefinedBehaviorFunction final : AAUndefinedBehaviorImpl {
3269 AAUndefinedBehaviorFunction(const IRPosition &IRP, Attributor &A)
3270 : AAUndefinedBehaviorImpl(IRP, A) {}
3271
3272 /// See AbstractAttribute::trackStatistics()
3273 void trackStatistics() const override {
3274 STATS_DECL(UndefinedBehaviorInstruction, Instruction,
3275 "Number of instructions known to have UB");
3276 BUILD_STAT_NAME(UndefinedBehaviorInstruction, Instruction) +=
3277 KnownUBInsts.size();
3278 }
3279};
3280} // namespace
3281
3282/// ------------------------ Will-Return Attributes ----------------------------
3283
3284namespace {
3285// Helper function that checks whether a function has any cycle which we don't
3286// know if it is bounded or not.
3287// Loops with maximum trip count are considered bounded, any other cycle not.
3288static bool mayContainUnboundedCycle(Function &F, Attributor &A) {
3289 ScalarEvolution *SE =
3290 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(F);
3291 LoopInfo *LI = A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(F);
3292 // If either SCEV or LoopInfo is not available for the function then we assume
3293 // any cycle to be unbounded cycle.
3294 // We use scc_iterator which uses Tarjan algorithm to find all the maximal
3295 // SCCs.To detect if there's a cycle, we only need to find the maximal ones.
3296 if (!SE || !LI) {
3297 for (scc_iterator<Function *> SCCI = scc_begin(&F); !SCCI.isAtEnd(); ++SCCI)
3298 if (SCCI.hasCycle())
3299 return true;
3300 return false;
3301 }
3302
3303 // If there's irreducible control, the function may contain non-loop cycles.
3305 return true;
3306
3307 // Any loop that does not have a max trip count is considered unbounded cycle.
3308 for (auto *L : LI->getLoopsInPreorder()) {
3309 if (!SE->getSmallConstantMaxTripCount(L))
3310 return true;
3311 }
3312 return false;
3313}
3314
3315struct AAWillReturnImpl : public AAWillReturn {
3316 AAWillReturnImpl(const IRPosition &IRP, Attributor &A)
3317 : AAWillReturn(IRP, A) {}
3318
3319 /// See AbstractAttribute::initialize(...).
3320 void initialize(Attributor &A) override {
3321 bool IsKnown;
3322 assert(!AA::hasAssumedIRAttr<Attribute::WillReturn>(
3323 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
3324 (void)IsKnown;
3325 }
3326
3327 /// Check for `mustprogress` and `readonly` as they imply `willreturn`.
3328 bool isImpliedByMustprogressAndReadonly(Attributor &A, bool KnownOnly) {
3329 if (!A.hasAttr(getIRPosition(), {Attribute::MustProgress}))
3330 return false;
3331
3332 bool IsKnown;
3333 if (AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
3334 return IsKnown || !KnownOnly;
3335 return false;
3336 }
3337
3338 /// See AbstractAttribute::updateImpl(...).
3339 ChangeStatus updateImpl(Attributor &A) override {
3340 if (isImpliedByMustprogressAndReadonly(A, /* KnownOnly */ false))
3341 return ChangeStatus::UNCHANGED;
3342
3343 auto CheckForWillReturn = [&](Instruction &I) {
3344 IRPosition IPos = IRPosition::callsite_function(cast<CallBase>(I));
3345 bool IsKnown;
3346 if (AA::hasAssumedIRAttr<Attribute::WillReturn>(
3347 A, this, IPos, DepClassTy::REQUIRED, IsKnown)) {
3348 if (IsKnown)
3349 return true;
3350 } else {
3351 return false;
3352 }
3353 bool IsKnownNoRecurse;
3354 return AA::hasAssumedIRAttr<Attribute::NoRecurse>(
3355 A, this, IPos, DepClassTy::REQUIRED, IsKnownNoRecurse);
3356 };
3357
3358 bool UsedAssumedInformation = false;
3359 if (!A.checkForAllCallLikeInstructions(CheckForWillReturn, *this,
3360 UsedAssumedInformation))
3361 return indicatePessimisticFixpoint();
3362
3363 return ChangeStatus::UNCHANGED;
3364 }
3365
3366 /// See AbstractAttribute::getAsStr()
3367 const std::string getAsStr(Attributor *A) const override {
3368 return getAssumed() ? "willreturn" : "may-noreturn";
3369 }
3370};
3371
3372struct AAWillReturnFunction final : AAWillReturnImpl {
3373 AAWillReturnFunction(const IRPosition &IRP, Attributor &A)
3374 : AAWillReturnImpl(IRP, A) {}
3375
3376 /// See AbstractAttribute::initialize(...).
3377 void initialize(Attributor &A) override {
3378 AAWillReturnImpl::initialize(A);
3379
3380 Function *F = getAnchorScope();
3381 assert(F && "Did expect an anchor function");
3382 if (F->isDeclaration() || mayContainUnboundedCycle(*F, A))
3383 indicatePessimisticFixpoint();
3384 }
3385
3386 /// See AbstractAttribute::trackStatistics()
3387 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(willreturn) }
3388};
3389
3390/// WillReturn attribute deduction for a call sites.
3391struct AAWillReturnCallSite final
3392 : AACalleeToCallSite<AAWillReturn, AAWillReturnImpl> {
3393 AAWillReturnCallSite(const IRPosition &IRP, Attributor &A)
3394 : AACalleeToCallSite<AAWillReturn, AAWillReturnImpl>(IRP, A) {}
3395
3396 /// See AbstractAttribute::updateImpl(...).
3397 ChangeStatus updateImpl(Attributor &A) override {
3398 if (isImpliedByMustprogressAndReadonly(A, /* KnownOnly */ false))
3399 return ChangeStatus::UNCHANGED;
3400
3401 return AACalleeToCallSite::updateImpl(A);
3402 }
3403
3404 /// See AbstractAttribute::trackStatistics()
3405 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(willreturn); }
3406};
3407} // namespace
3408
3409/// -------------------AAIntraFnReachability Attribute--------------------------
3410
3411/// All information associated with a reachability query. This boilerplate code
3412/// is used by both AAIntraFnReachability and AAInterFnReachability, with
3413/// different \p ToTy values.
3414template <typename ToTy> struct ReachabilityQueryInfo {
3415 enum class Reachable {
3416 No,
3417 Yes,
3418 };
3419
3420 /// Start here,
3421 const Instruction *From = nullptr;
3422 /// reach this place,
3423 const ToTy *To = nullptr;
3424 /// without going through any of these instructions,
3425 const AA::InstExclusionSetTy *ExclusionSet = nullptr;
3426 /// and remember if it worked:
3427 Reachable Result = Reachable::No;
3428
3429 /// Precomputed hash for this RQI.
3430 unsigned Hash = 0;
3431
3432 unsigned computeHashValue() const {
3433 assert(Hash == 0 && "Computed hash twice!");
3436 return const_cast<ReachabilityQueryInfo<ToTy> *>(this)->Hash =
3437 detail::combineHashValue(PairDMI ::getHashValue({From, To}),
3438 InstSetDMI::getHashValue(ExclusionSet));
3439 }
3440
3442 : From(From), To(To) {}
3443
3444 /// Constructor replacement to ensure unique and stable sets are used for the
3445 /// cache.
3447 const AA::InstExclusionSetTy *ES, bool MakeUnique)
3448 : From(&From), To(&To), ExclusionSet(ES) {
3449
3450 if (!ES || ES->empty()) {
3451 ExclusionSet = nullptr;
3452 } else if (MakeUnique) {
3453 ExclusionSet = A.getInfoCache().getOrCreateUniqueBlockExecutionSet(ES);
3454 }
3455 }
3456
3458 : From(RQI.From), To(RQI.To), ExclusionSet(RQI.ExclusionSet) {}
3459};
3460
3461namespace llvm {
3462template <typename ToTy> struct DenseMapInfo<ReachabilityQueryInfo<ToTy> *> {
3465
3468
3469 static inline ReachabilityQueryInfo<ToTy> *getEmptyKey() { return &EmptyKey; }
3471 return &TombstoneKey;
3472 }
3473 static unsigned getHashValue(const ReachabilityQueryInfo<ToTy> *RQI) {
3474 return RQI->Hash ? RQI->Hash : RQI->computeHashValue();
3475 }
3478 if (!PairDMI::isEqual({LHS->From, LHS->To}, {RHS->From, RHS->To}))
3479 return false;
3480 return InstSetDMI::isEqual(LHS->ExclusionSet, RHS->ExclusionSet);
3481 }
3482};
3483
3484#define DefineKeys(ToTy) \
3485 template <> \
3486 ReachabilityQueryInfo<ToTy> \
3487 DenseMapInfo<ReachabilityQueryInfo<ToTy> *>::EmptyKey = \
3488 ReachabilityQueryInfo<ToTy>( \
3489 DenseMapInfo<const Instruction *>::getEmptyKey(), \
3490 DenseMapInfo<const ToTy *>::getEmptyKey()); \
3491 template <> \
3492 ReachabilityQueryInfo<ToTy> \
3493 DenseMapInfo<ReachabilityQueryInfo<ToTy> *>::TombstoneKey = \
3494 ReachabilityQueryInfo<ToTy>( \
3495 DenseMapInfo<const Instruction *>::getTombstoneKey(), \
3496 DenseMapInfo<const ToTy *>::getTombstoneKey());
3497
3499#undef DefineKeys
3500
3501} // namespace llvm
3502
3503namespace {
3504
3505template <typename BaseTy, typename ToTy>
3506struct CachedReachabilityAA : public BaseTy {
3507 using RQITy = ReachabilityQueryInfo<ToTy>;
3508
3509 CachedReachabilityAA(const IRPosition &IRP, Attributor &A) : BaseTy(IRP, A) {}
3510
3511 /// See AbstractAttribute::isQueryAA.
3512 bool isQueryAA() const override { return true; }
3513
3514 /// See AbstractAttribute::updateImpl(...).
3515 ChangeStatus updateImpl(Attributor &A) override {
3516 ChangeStatus Changed = ChangeStatus::UNCHANGED;
3517 for (unsigned u = 0, e = QueryVector.size(); u < e; ++u) {
3518 RQITy *RQI = QueryVector[u];
3519 if (RQI->Result == RQITy::Reachable::No &&
3520 isReachableImpl(A, *RQI, /*IsTemporaryRQI=*/false))
3521 Changed = ChangeStatus::CHANGED;
3522 }
3523 return Changed;
3524 }
3525
3526 virtual bool isReachableImpl(Attributor &A, RQITy &RQI,
3527 bool IsTemporaryRQI) = 0;
3528
3529 bool rememberResult(Attributor &A, typename RQITy::Reachable Result,
3530 RQITy &RQI, bool UsedExclusionSet, bool IsTemporaryRQI) {
3531 RQI.Result = Result;
3532
3533 // Remove the temporary RQI from the cache.
3534 if (IsTemporaryRQI)
3535 QueryCache.erase(&RQI);
3536
3537 // Insert a plain RQI (w/o exclusion set) if that makes sense. Two options:
3538 // 1) If it is reachable, it doesn't matter if we have an exclusion set for
3539 // this query. 2) We did not use the exclusion set, potentially because
3540 // there is none.
3541 if (Result == RQITy::Reachable::Yes || !UsedExclusionSet) {
3542 RQITy PlainRQI(RQI.From, RQI.To);
3543 if (!QueryCache.count(&PlainRQI)) {
3544 RQITy *RQIPtr = new (A.Allocator) RQITy(RQI.From, RQI.To);
3545 RQIPtr->Result = Result;
3546 QueryVector.push_back(RQIPtr);
3547 QueryCache.insert(RQIPtr);
3548 }
3549 }
3550
3551 // Check if we need to insert a new permanent RQI with the exclusion set.
3552 if (IsTemporaryRQI && Result != RQITy::Reachable::Yes && UsedExclusionSet) {
3553 assert((!RQI.ExclusionSet || !RQI.ExclusionSet->empty()) &&
3554 "Did not expect empty set!");
3555 RQITy *RQIPtr = new (A.Allocator)
3556 RQITy(A, *RQI.From, *RQI.To, RQI.ExclusionSet, true);
3557 assert(RQIPtr->Result == RQITy::Reachable::No && "Already reachable?");
3558 RQIPtr->Result = Result;
3559 assert(!QueryCache.count(RQIPtr));
3560 QueryVector.push_back(RQIPtr);
3561 QueryCache.insert(RQIPtr);
3562 }
3563
3564 if (Result == RQITy::Reachable::No && IsTemporaryRQI)
3565 A.registerForUpdate(*this);
3566 return Result == RQITy::Reachable::Yes;
3567 }
3568
3569 const std::string getAsStr(Attributor *A) const override {
3570 // TODO: Return the number of reachable queries.
3571 return "#queries(" + std::to_string(QueryVector.size()) + ")";
3572 }
3573
3574 bool checkQueryCache(Attributor &A, RQITy &StackRQI,
3575 typename RQITy::Reachable &Result) {
3576 if (!this->getState().isValidState()) {
3577 Result = RQITy::Reachable::Yes;
3578 return true;
3579 }
3580
3581 // If we have an exclusion set we might be able to find our answer by
3582 // ignoring it first.
3583 if (StackRQI.ExclusionSet) {
3584 RQITy PlainRQI(StackRQI.From, StackRQI.To);
3585 auto It = QueryCache.find(&PlainRQI);
3586 if (It != QueryCache.end() && (*It)->Result == RQITy::Reachable::No) {
3587 Result = RQITy::Reachable::No;
3588 return true;
3589 }
3590 }
3591
3592 auto It = QueryCache.find(&StackRQI);
3593 if (It != QueryCache.end()) {
3594 Result = (*It)->Result;
3595 return true;
3596 }
3597
3598 // Insert a temporary for recursive queries. We will replace it with a
3599 // permanent entry later.
3600 QueryCache.insert(&StackRQI);
3601 return false;
3602 }
3603
3604private:
3605 SmallVector<RQITy *> QueryVector;
3606 DenseSet<RQITy *> QueryCache;
3607};
3608
3609struct AAIntraFnReachabilityFunction final
3610 : public CachedReachabilityAA<AAIntraFnReachability, Instruction> {
3611 using Base = CachedReachabilityAA<AAIntraFnReachability, Instruction>;
3612 AAIntraFnReachabilityFunction(const IRPosition &IRP, Attributor &A)
3613 : Base(IRP, A) {
3614 DT = A.getInfoCache().getAnalysisResultForFunction<DominatorTreeAnalysis>(
3615 *IRP.getAssociatedFunction());
3616 }
3617
3618 bool isAssumedReachable(
3619 Attributor &A, const Instruction &From, const Instruction &To,
3620 const AA::InstExclusionSetTy *ExclusionSet) const override {
3621 auto *NonConstThis = const_cast<AAIntraFnReachabilityFunction *>(this);
3622 if (&From == &To)
3623 return true;
3624
3625 RQITy StackRQI(A, From, To, ExclusionSet, false);
3626 typename RQITy::Reachable Result;
3627 if (!NonConstThis->checkQueryCache(A, StackRQI, Result))
3628 return NonConstThis->isReachableImpl(A, StackRQI,
3629 /*IsTemporaryRQI=*/true);
3630 return Result == RQITy::Reachable::Yes;
3631 }
3632
3633 ChangeStatus updateImpl(Attributor &A) override {
3634 // We only depend on liveness. DeadEdges is all we care about, check if any
3635 // of them changed.
3636 auto *LivenessAA =
3637 A.getAAFor<AAIsDead>(*this, getIRPosition(), DepClassTy::OPTIONAL);
3638 if (LivenessAA &&
3639 llvm::all_of(DeadEdges,
3640 [&](const auto &DeadEdge) {
3641 return LivenessAA->isEdgeDead(DeadEdge.first,
3642 DeadEdge.second);
3643 }) &&
3644 llvm::all_of(DeadBlocks, [&](const BasicBlock *BB) {
3645 return LivenessAA->isAssumedDead(BB);
3646 })) {
3647 return ChangeStatus::UNCHANGED;
3648 }
3649 DeadEdges.clear();
3650 DeadBlocks.clear();
3651 return Base::updateImpl(A);
3652 }
3653
3654 bool isReachableImpl(Attributor &A, RQITy &RQI,
3655 bool IsTemporaryRQI) override {
3656 const Instruction *Origin = RQI.From;
3657 bool UsedExclusionSet = false;
3658
3659 auto WillReachInBlock = [&](const Instruction &From, const Instruction &To,
3660 const AA::InstExclusionSetTy *ExclusionSet) {
3661 const Instruction *IP = &From;
3662 while (IP && IP != &To) {
3663 if (ExclusionSet && IP != Origin && ExclusionSet->count(IP)) {
3664 UsedExclusionSet = true;
3665 break;
3666 }
3667 IP = IP->getNextNode();
3668 }
3669 return IP == &To;
3670 };
3671
3672 const BasicBlock *FromBB = RQI.From->getParent();
3673 const BasicBlock *ToBB = RQI.To->getParent();
3674 assert(FromBB->getParent() == ToBB->getParent() &&
3675 "Not an intra-procedural query!");
3676
3677 // Check intra-block reachability, however, other reaching paths are still
3678 // possible.
3679 if (FromBB == ToBB &&
3680 WillReachInBlock(*RQI.From, *RQI.To, RQI.ExclusionSet))
3681 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3682 IsTemporaryRQI);
3683
3684 // Check if reaching the ToBB block is sufficient or if even that would not
3685 // ensure reaching the target. In the latter case we are done.
3686 if (!WillReachInBlock(ToBB->front(), *RQI.To, RQI.ExclusionSet))
3687 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3688 IsTemporaryRQI);
3689
3690 const Function *Fn = FromBB->getParent();
3692 if (RQI.ExclusionSet)
3693 for (auto *I : *RQI.ExclusionSet)
3694 if (I->getFunction() == Fn)
3695 ExclusionBlocks.insert(I->getParent());
3696
3697 // Check if we make it out of the FromBB block at all.
3698 if (ExclusionBlocks.count(FromBB) &&
3699 !WillReachInBlock(*RQI.From, *FromBB->getTerminator(),
3700 RQI.ExclusionSet))
3701 return rememberResult(A, RQITy::Reachable::No, RQI, true, IsTemporaryRQI);
3702
3703 auto *LivenessAA =
3704 A.getAAFor<AAIsDead>(*this, getIRPosition(), DepClassTy::OPTIONAL);
3705 if (LivenessAA && LivenessAA->isAssumedDead(ToBB)) {
3706 DeadBlocks.insert(ToBB);
3707 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3708 IsTemporaryRQI);
3709 }
3710
3713 Worklist.push_back(FromBB);
3714
3716 while (!Worklist.empty()) {
3717 const BasicBlock *BB = Worklist.pop_back_val();
3718 if (!Visited.insert(BB).second)
3719 continue;
3720 for (const BasicBlock *SuccBB : successors(BB)) {
3721 if (LivenessAA && LivenessAA->isEdgeDead(BB, SuccBB)) {
3722 LocalDeadEdges.insert({BB, SuccBB});
3723 continue;
3724 }
3725 // We checked before if we just need to reach the ToBB block.
3726 if (SuccBB == ToBB)
3727 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3728 IsTemporaryRQI);
3729 if (DT && ExclusionBlocks.empty() && DT->dominates(BB, ToBB))
3730 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
3731 IsTemporaryRQI);
3732
3733 if (ExclusionBlocks.count(SuccBB)) {
3734 UsedExclusionSet = true;
3735 continue;
3736 }
3737 Worklist.push_back(SuccBB);
3738 }
3739 }
3740
3741 DeadEdges.insert(LocalDeadEdges.begin(), LocalDeadEdges.end());
3742 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
3743 IsTemporaryRQI);
3744 }
3745
3746 /// See AbstractAttribute::trackStatistics()
3747 void trackStatistics() const override {}
3748
3749private:
3750 // Set of assumed dead blocks we used in the last query. If any changes we
3751 // update the state.
3753
3754 // Set of assumed dead edges we used in the last query. If any changes we
3755 // update the state.
3757
3758 /// The dominator tree of the function to short-circuit reasoning.
3759 const DominatorTree *DT = nullptr;
3760};
3761} // namespace
3762
3763/// ------------------------ NoAlias Argument Attribute ------------------------
3764
3766 Attribute::AttrKind ImpliedAttributeKind,
3767 bool IgnoreSubsumingPositions) {
3768 assert(ImpliedAttributeKind == Attribute::NoAlias &&
3769 "Unexpected attribute kind");
3770 Value *Val = &IRP.getAssociatedValue();
3771 if (IRP.getPositionKind() != IRP_CALL_SITE_ARGUMENT) {
3772 if (isa<AllocaInst>(Val))
3773 return true;
3774 } else {
3775 IgnoreSubsumingPositions = true;
3776 }
3777
3778 if (isa<UndefValue>(Val))
3779 return true;
3780
3781 if (isa<ConstantPointerNull>(Val) &&
3784 return true;
3785
3786 if (A.hasAttr(IRP, {Attribute::ByVal, Attribute::NoAlias},
3787 IgnoreSubsumingPositions, Attribute::NoAlias))
3788 return true;
3789
3790 return false;
3791}
3792
3793namespace {
3794struct AANoAliasImpl : AANoAlias {
3795 AANoAliasImpl(const IRPosition &IRP, Attributor &A) : AANoAlias(IRP, A) {
3796 assert(getAssociatedType()->isPointerTy() &&
3797 "Noalias is a pointer attribute");
3798 }
3799
3800 const std::string getAsStr(Attributor *A) const override {
3801 return getAssumed() ? "noalias" : "may-alias";
3802 }
3803};
3804
3805/// NoAlias attribute for a floating value.
3806struct AANoAliasFloating final : AANoAliasImpl {
3807 AANoAliasFloating(const IRPosition &IRP, Attributor &A)
3808 : AANoAliasImpl(IRP, A) {}
3809
3810 /// See AbstractAttribute::updateImpl(...).
3811 ChangeStatus updateImpl(Attributor &A) override {
3812 // TODO: Implement this.
3813 return indicatePessimisticFixpoint();
3814 }
3815
3816 /// See AbstractAttribute::trackStatistics()
3817 void trackStatistics() const override {
3819 }
3820};
3821
3822/// NoAlias attribute for an argument.
3823struct AANoAliasArgument final
3824 : AAArgumentFromCallSiteArguments<AANoAlias, AANoAliasImpl> {
3825 using Base = AAArgumentFromCallSiteArguments<AANoAlias, AANoAliasImpl>;
3826 AANoAliasArgument(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
3827
3828 /// See AbstractAttribute::update(...).
3829 ChangeStatus updateImpl(Attributor &A) override {
3830 // We have to make sure no-alias on the argument does not break
3831 // synchronization when this is a callback argument, see also [1] below.
3832 // If synchronization cannot be affected, we delegate to the base updateImpl
3833 // function, otherwise we give up for now.
3834
3835 // If the function is no-sync, no-alias cannot break synchronization.
3836 bool IsKnownNoSycn;
3837 if (AA::hasAssumedIRAttr<Attribute::NoSync>(
3838 A, this, IRPosition::function_scope(getIRPosition()),
3839 DepClassTy::OPTIONAL, IsKnownNoSycn))
3840 return Base::updateImpl(A);
3841
3842 // If the argument is read-only, no-alias cannot break synchronization.
3843 bool IsKnown;
3844 if (AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
3845 return Base::updateImpl(A);
3846
3847 // If the argument is never passed through callbacks, no-alias cannot break
3848 // synchronization.
3849 bool UsedAssumedInformation = false;
3850 if (A.checkForAllCallSites(
3851 [](AbstractCallSite ACS) { return !ACS.isCallbackCall(); }, *this,
3852 true, UsedAssumedInformation))
3853 return Base::updateImpl(A);
3854
3855 // TODO: add no-alias but make sure it doesn't break synchronization by
3856 // introducing fake uses. See:
3857 // [1] Compiler Optimizations for OpenMP, J. Doerfert and H. Finkel,
3858 // International Workshop on OpenMP 2018,
3859 // http://compilers.cs.uni-saarland.de/people/doerfert/par_opt18.pdf
3860
3861 return indicatePessimisticFixpoint();
3862 }
3863
3864 /// See AbstractAttribute::trackStatistics()
3865 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(noalias) }
3866};
3867
3868struct AANoAliasCallSiteArgument final : AANoAliasImpl {
3869 AANoAliasCallSiteArgument(const IRPosition &IRP, Attributor &A)
3870 : AANoAliasImpl(IRP, A) {}
3871
3872 /// Determine if the underlying value may alias with the call site argument
3873 /// \p OtherArgNo of \p ICS (= the underlying call site).
3874 bool mayAliasWithArgument(Attributor &A, AAResults *&AAR,
3875 const AAMemoryBehavior &MemBehaviorAA,
3876 const CallBase &CB, unsigned OtherArgNo) {
3877 // We do not need to worry about aliasing with the underlying IRP.
3878 if (this->getCalleeArgNo() == (int)OtherArgNo)
3879 return false;
3880
3881 // If it is not a pointer or pointer vector we do not alias.
3882 const Value *ArgOp = CB.getArgOperand(OtherArgNo);
3883 if (!ArgOp->getType()->isPtrOrPtrVectorTy())
3884 return false;
3885
3886 auto *CBArgMemBehaviorAA = A.getAAFor<AAMemoryBehavior>(
3887 *this, IRPosition::callsite_argument(CB, OtherArgNo), DepClassTy::NONE);
3888
3889 // If the argument is readnone, there is no read-write aliasing.
3890 if (CBArgMemBehaviorAA && CBArgMemBehaviorAA->isAssumedReadNone()) {
3891 A.recordDependence(*CBArgMemBehaviorAA, *this, DepClassTy::OPTIONAL);
3892 return false;
3893 }
3894
3895 // If the argument is readonly and the underlying value is readonly, there
3896 // is no read-write aliasing.
3897 bool IsReadOnly = MemBehaviorAA.isAssumedReadOnly();
3898 if (CBArgMemBehaviorAA && CBArgMemBehaviorAA->isAssumedReadOnly() &&
3899 IsReadOnly) {
3900 A.recordDependence(MemBehaviorAA, *this, DepClassTy::OPTIONAL);
3901 A.recordDependence(*CBArgMemBehaviorAA, *this, DepClassTy::OPTIONAL);
3902 return false;
3903 }
3904
3905 // We have to utilize actual alias analysis queries so we need the object.
3906 if (!AAR)
3907 AAR = A.getInfoCache().getAnalysisResultForFunction<AAManager>(
3908 *getAnchorScope());
3909
3910 // Try to rule it out at the call site.
3911 bool IsAliasing = !AAR || !AAR->isNoAlias(&getAssociatedValue(), ArgOp);
3912 LLVM_DEBUG(dbgs() << "[NoAliasCSArg] Check alias between "
3913 "callsite arguments: "
3914 << getAssociatedValue() << " " << *ArgOp << " => "
3915 << (IsAliasing ? "" : "no-") << "alias \n");
3916
3917 return IsAliasing;
3918 }
3919
3920 bool isKnownNoAliasDueToNoAliasPreservation(
3921 Attributor &A, AAResults *&AAR, const AAMemoryBehavior &MemBehaviorAA) {
3922 // We can deduce "noalias" if the following conditions hold.
3923 // (i) Associated value is assumed to be noalias in the definition.
3924 // (ii) Associated value is assumed to be no-capture in all the uses
3925 // possibly executed before this callsite.
3926 // (iii) There is no other pointer argument which could alias with the
3927 // value.
3928
3929 auto IsDereferenceableOrNull = [&](Value *O, const DataLayout &DL) {
3930 const auto *DerefAA = A.getAAFor<AADereferenceable>(
3931 *this, IRPosition::value(*O), DepClassTy::OPTIONAL);
3932 return DerefAA ? DerefAA->getAssumedDereferenceableBytes() : 0;
3933 };
3934
3935 const IRPosition &VIRP = IRPosition::value(getAssociatedValue());
3936 const Function *ScopeFn = VIRP.getAnchorScope();
3937 // Check whether the value is captured in the scope using AANoCapture.
3938 // Look at CFG and check only uses possibly executed before this
3939 // callsite.
3940 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
3941 Instruction *UserI = cast<Instruction>(U.getUser());
3942
3943 // If UserI is the curr instruction and there is a single potential use of
3944 // the value in UserI we allow the use.
3945 // TODO: We should inspect the operands and allow those that cannot alias
3946 // with the value.
3947 if (UserI == getCtxI() && UserI->getNumOperands() == 1)
3948 return true;
3949
3950 if (ScopeFn) {
3951 if (auto *CB = dyn_cast<CallBase>(UserI)) {
3952 if (CB->isArgOperand(&U)) {
3953
3954 unsigned ArgNo = CB->getArgOperandNo(&U);
3955
3956 bool IsKnownNoCapture;
3957 if (AA::hasAssumedIRAttr<Attribute::NoCapture>(
3958 A, this, IRPosition::callsite_argument(*CB, ArgNo),
3959 DepClassTy::OPTIONAL, IsKnownNoCapture))
3960 return true;
3961 }
3962 }
3963
3965 A, *UserI, *getCtxI(), *this, /* ExclusionSet */ nullptr,
3966 [ScopeFn](const Function &Fn) { return &Fn != ScopeFn; }))
3967 return true;
3968 }
3969
3970 // TODO: We should track the capturing uses in AANoCapture but the problem
3971 // is CGSCC runs. For those we would need to "allow" AANoCapture for
3972 // a value in the module slice.
3973 switch (DetermineUseCaptureKind(U, IsDereferenceableOrNull)) {
3974 case UseCaptureKind::NO_CAPTURE:
3975 return true;
3976 case UseCaptureKind::MAY_CAPTURE:
3977 LLVM_DEBUG(dbgs() << "[AANoAliasCSArg] Unknown user: " << *UserI
3978 << "\n");
3979 return false;
3980 case UseCaptureKind::PASSTHROUGH:
3981 Follow = true;
3982 return true;
3983 }
3984 llvm_unreachable("unknown UseCaptureKind");
3985 };
3986
3987 bool IsKnownNoCapture;
3988 const AANoCapture *NoCaptureAA = nullptr;
3989 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
3990 A, this, VIRP, DepClassTy::NONE, IsKnownNoCapture, false, &NoCaptureAA);
3991 if (!IsAssumedNoCapture &&
3992 (!NoCaptureAA || !NoCaptureAA->isAssumedNoCaptureMaybeReturned())) {
3993 if (!A.checkForAllUses(UsePred, *this, getAssociatedValue())) {
3994 LLVM_DEBUG(
3995 dbgs() << "[AANoAliasCSArg] " << getAssociatedValue()
3996 << " cannot be noalias as it is potentially captured\n");
3997 return false;
3998 }
3999 }
4000 if (NoCaptureAA)
4001 A.recordDependence(*NoCaptureAA, *this, DepClassTy::OPTIONAL);
4002
4003 // Check there is no other pointer argument which could alias with the
4004 // value passed at this call site.
4005 // TODO: AbstractCallSite
4006 const auto &CB = cast<CallBase>(getAnchorValue());
4007 for (unsigned OtherArgNo = 0; OtherArgNo < CB.arg_size(); OtherArgNo++)
4008 if (mayAliasWithArgument(A, AAR, MemBehaviorAA, CB, OtherArgNo))
4009 return false;
4010
4011 return true;
4012 }
4013
4014 /// See AbstractAttribute::updateImpl(...).
4015 ChangeStatus updateImpl(Attributor &A) override {
4016 // If the argument is readnone we are done as there are no accesses via the
4017 // argument.
4018 auto *MemBehaviorAA =
4019 A.getAAFor<AAMemoryBehavior>(*this, getIRPosition(), DepClassTy::NONE);
4020 if (MemBehaviorAA && MemBehaviorAA->isAssumedReadNone()) {
4021 A.recordDependence(*MemBehaviorAA, *this, DepClassTy::OPTIONAL);
4022 return ChangeStatus::UNCHANGED;
4023 }
4024
4025 bool IsKnownNoAlias;
4026 const IRPosition &VIRP = IRPosition::value(getAssociatedValue());
4027 if (!AA::hasAssumedIRAttr<Attribute::NoAlias>(
4028 A, this, VIRP, DepClassTy::REQUIRED, IsKnownNoAlias)) {
4029 LLVM_DEBUG(dbgs() << "[AANoAlias] " << getAssociatedValue()
4030 << " is not no-alias at the definition\n");
4031 return indicatePessimisticFixpoint();
4032 }
4033
4034 AAResults *AAR = nullptr;
4035 if (MemBehaviorAA &&
4036 isKnownNoAliasDueToNoAliasPreservation(A, AAR, *MemBehaviorAA)) {
4037 LLVM_DEBUG(
4038 dbgs() << "[AANoAlias] No-Alias deduced via no-alias preservation\n");
4039 return ChangeStatus::UNCHANGED;
4040 }
4041
4042 return indicatePessimisticFixpoint();
4043 }
4044
4045 /// See AbstractAttribute::trackStatistics()
4046 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(noalias) }
4047};
4048
4049/// NoAlias attribute for function return value.
4050struct AANoAliasReturned final : AANoAliasImpl {
4051 AANoAliasReturned(const IRPosition &IRP, Attributor &A)
4052 : AANoAliasImpl(IRP, A) {}
4053
4054 /// See AbstractAttribute::updateImpl(...).
4055 ChangeStatus updateImpl(Attributor &A) override {
4056
4057 auto CheckReturnValue = [&](Value &RV) -> bool {
4058 if (Constant *C = dyn_cast<Constant>(&RV))
4059 if (C->isNullValue() || isa<UndefValue>(C))
4060 return true;
4061
4062 /// For now, we can only deduce noalias if we have call sites.
4063 /// FIXME: add more support.
4064 if (!isa<CallBase>(&RV))
4065 return false;
4066
4067 const IRPosition &RVPos = IRPosition::value(RV);
4068 bool IsKnownNoAlias;
4069 if (!AA::hasAssumedIRAttr<Attribute::NoAlias>(
4070 A, this, RVPos, DepClassTy::REQUIRED, IsKnownNoAlias))
4071 return false;
4072
4073 bool IsKnownNoCapture;
4074 const AANoCapture *NoCaptureAA = nullptr;
4075 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
4076 A, this, RVPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
4077 &NoCaptureAA);
4078 return IsAssumedNoCapture ||
4079 (NoCaptureAA && NoCaptureAA->isAssumedNoCaptureMaybeReturned());
4080 };
4081
4082 if (!A.checkForAllReturnedValues(CheckReturnValue, *this))
4083 return indicatePessimisticFixpoint();
4084
4085 return ChangeStatus::UNCHANGED;
4086 }
4087
4088 /// See AbstractAttribute::trackStatistics()
4089 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noalias) }
4090};
4091
4092/// NoAlias attribute deduction for a call site return value.
4093struct AANoAliasCallSiteReturned final
4094 : AACalleeToCallSite<AANoAlias, AANoAliasImpl> {
4095 AANoAliasCallSiteReturned(const IRPosition &IRP, Attributor &A)
4096 : AACalleeToCallSite<AANoAlias, AANoAliasImpl>(IRP, A) {}
4097
4098 /// See AbstractAttribute::trackStatistics()
4099 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(noalias); }
4100};
4101} // namespace
4102
4103/// -------------------AAIsDead Function Attribute-----------------------
4104
4105namespace {
4106struct AAIsDeadValueImpl : public AAIsDead {
4107 AAIsDeadValueImpl(const IRPosition &IRP, Attributor &A) : AAIsDead(IRP, A) {}
4108
4109 /// See AAIsDead::isAssumedDead().
4110 bool isAssumedDead() const override { return isAssumed(IS_DEAD); }
4111
4112 /// See AAIsDead::isKnownDead().
4113 bool isKnownDead() const override { return isKnown(IS_DEAD); }
4114
4115 /// See AAIsDead::isAssumedDead(BasicBlock *).
4116 bool isAssumedDead(const BasicBlock *BB) const override { return false; }
4117
4118 /// See AAIsDead::isKnownDead(BasicBlock *).
4119 bool isKnownDead(const BasicBlock *BB) const override { return false; }
4120
4121 /// See AAIsDead::isAssumedDead(Instruction *I).
4122 bool isAssumedDead(const Instruction *I) const override {
4123 return I == getCtxI() && isAssumedDead();
4124 }
4125
4126 /// See AAIsDead::isKnownDead(Instruction *I).
4127 bool isKnownDead(const Instruction *I) const override {
4128 return isAssumedDead(I) && isKnownDead();
4129 }
4130
4131 /// See AbstractAttribute::getAsStr().
4132 const std::string getAsStr(Attributor *A) const override {
4133 return isAssumedDead() ? "assumed-dead" : "assumed-live";
4134 }
4135
4136 /// Check if all uses are assumed dead.
4137 bool areAllUsesAssumedDead(Attributor &A, Value &V) {
4138 // Callers might not check the type, void has no uses.
4139 if (V.getType()->isVoidTy() || V.use_empty())
4140 return true;
4141
4142 // If we replace a value with a constant there are no uses left afterwards.
4143 if (!isa<Constant>(V)) {
4144 if (auto *I = dyn_cast<Instruction>(&V))
4145 if (!A.isRunOn(*I->getFunction()))
4146 return false;
4147 bool UsedAssumedInformation = false;
4148 std::optional<Constant *> C =
4149 A.getAssumedConstant(V, *this, UsedAssumedInformation);
4150 if (!C || *C)
4151 return true;
4152 }
4153
4154 auto UsePred = [&](const Use &U, bool &Follow) { return false; };
4155 // Explicitly set the dependence class to required because we want a long
4156 // chain of N dependent instructions to be considered live as soon as one is
4157 // without going through N update cycles. This is not required for
4158 // correctness.
4159 return A.checkForAllUses(UsePred, *this, V, /* CheckBBLivenessOnly */ false,
4160 DepClassTy::REQUIRED,
4161 /* IgnoreDroppableUses */ false);
4162 }
4163
4164 /// Determine if \p I is assumed to be side-effect free.
4165 bool isAssumedSideEffectFree(Attributor &A, Instruction *I) {
4167 return true;
4168
4169 auto *CB = dyn_cast<CallBase>(I);
4170 if (!CB || isa<IntrinsicInst>(CB))
4171 return false;
4172
4173 const IRPosition &CallIRP = IRPosition::callsite_function(*CB);
4174
4175 bool IsKnownNoUnwind;
4176 if (!AA::hasAssumedIRAttr<Attribute::NoUnwind>(
4177 A, this, CallIRP, DepClassTy::OPTIONAL, IsKnownNoUnwind))
4178 return false;
4179
4180 bool IsKnown;
4181 return AA::isAssumedReadOnly(A, CallIRP, *this, IsKnown);
4182 }
4183};
4184
4185struct AAIsDeadFloating : public AAIsDeadValueImpl {
4186 AAIsDeadFloating(const IRPosition &IRP, Attributor &A)
4187 : AAIsDeadValueImpl(IRP, A) {}
4188
4189 /// See AbstractAttribute::initialize(...).
4190 void initialize(Attributor &A) override {
4191 AAIsDeadValueImpl::initialize(A);
4192
4193 if (isa<UndefValue>(getAssociatedValue())) {
4194 indicatePessimisticFixpoint();
4195 return;
4196 }
4197
4198 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4199 if (!isAssumedSideEffectFree(A, I)) {
4200 if (!isa_and_nonnull<StoreInst>(I) && !isa_and_nonnull<FenceInst>(I))
4201 indicatePessimisticFixpoint();
4202 else
4203 removeAssumedBits(HAS_NO_EFFECT);
4204 }
4205 }
4206
4207 bool isDeadFence(Attributor &A, FenceInst &FI) {
4208 const auto *ExecDomainAA = A.lookupAAFor<AAExecutionDomain>(
4209 IRPosition::function(*FI.getFunction()), *this, DepClassTy::NONE);
4210 if (!ExecDomainAA || !ExecDomainAA->isNoOpFence(FI))
4211 return false;
4212 A.recordDependence(*ExecDomainAA, *this, DepClassTy::OPTIONAL);
4213 return true;
4214 }
4215
4216 bool isDeadStore(Attributor &A, StoreInst &SI,
4217 SmallSetVector<Instruction *, 8> *AssumeOnlyInst = nullptr) {
4218 // Lang ref now states volatile store is not UB/dead, let's skip them.
4219 if (SI.isVolatile())
4220 return false;
4221
4222 // If we are collecting assumes to be deleted we are in the manifest stage.
4223 // It's problematic to collect the potential copies again now so we use the
4224 // cached ones.
4225 bool UsedAssumedInformation = false;
4226 if (!AssumeOnlyInst) {
4227 PotentialCopies.clear();
4228 if (!AA::getPotentialCopiesOfStoredValue(A, SI, PotentialCopies, *this,
4229 UsedAssumedInformation)) {
4230 LLVM_DEBUG(
4231 dbgs()
4232 << "[AAIsDead] Could not determine potential copies of store!\n");
4233 return false;
4234 }
4235 }
4236 LLVM_DEBUG(dbgs() << "[AAIsDead] Store has " << PotentialCopies.size()
4237 << " potential copies.\n");
4238
4239 InformationCache &InfoCache = A.getInfoCache();
4240 return llvm::all_of(PotentialCopies, [&](Value *V) {
4241 if (A.isAssumedDead(IRPosition::value(*V), this, nullptr,
4242 UsedAssumedInformation))
4243 return true;
4244 if (auto *LI = dyn_cast<LoadInst>(V)) {
4245 if (llvm::all_of(LI->uses(), [&](const Use &U) {
4246 auto &UserI = cast<Instruction>(*U.getUser());
4247 if (InfoCache.isOnlyUsedByAssume(UserI)) {
4248 if (AssumeOnlyInst)
4249 AssumeOnlyInst->insert(&UserI);
4250 return true;
4251 }
4252 return A.isAssumedDead(U, this, nullptr, UsedAssumedInformation);
4253 })) {
4254 return true;
4255 }
4256 }
4257 LLVM_DEBUG(dbgs() << "[AAIsDead] Potential copy " << *V
4258 << " is assumed live!\n");
4259 return false;
4260 });
4261 }
4262
4263 /// See AbstractAttribute::getAsStr().
4264 const std::string getAsStr(Attributor *A) const override {
4265 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4266 if (isa_and_nonnull<StoreInst>(I))
4267 if (isValidState())
4268 return "assumed-dead-store";
4269 if (isa_and_nonnull<FenceInst>(I))
4270 if (isValidState())
4271 return "assumed-dead-fence";
4272 return AAIsDeadValueImpl::getAsStr(A);
4273 }
4274
4275 /// See AbstractAttribute::updateImpl(...).
4276 ChangeStatus updateImpl(Attributor &A) override {
4277 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
4278 if (auto *SI = dyn_cast_or_null<StoreInst>(I)) {
4279 if (!isDeadStore(A, *SI))
4280 return indicatePessimisticFixpoint();
4281 } else if (auto *FI = dyn_cast_or_null<FenceInst>(I)) {
4282 if (!isDeadFence(A, *FI))
4283 return indicatePessimisticFixpoint();
4284 } else {
4285 if (!isAssumedSideEffectFree(A, I))
4286 return indicatePessimisticFixpoint();
4287 if (!areAllUsesAssumedDead(A, getAssociatedValue()))
4288 return indicatePessimisticFixpoint();
4289 }
4291 }
4292
4293 bool isRemovableStore() const override {
4294 return isAssumed(IS_REMOVABLE) && isa<StoreInst>(&getAssociatedValue());
4295 }
4296
4297 /// See AbstractAttribute::manifest(...).
4298 ChangeStatus manifest(Attributor &A) override {
4299 Value &V = getAssociatedValue();
4300 if (auto *I = dyn_cast<Instruction>(&V)) {
4301 // If we get here we basically know the users are all dead. We check if
4302 // isAssumedSideEffectFree returns true here again because it might not be
4303 // the case and only the users are dead but the instruction (=call) is
4304 // still needed.
4305 if (auto *SI = dyn_cast<StoreInst>(I)) {
4306 SmallSetVector<Instruction *, 8> AssumeOnlyInst;
4307 bool IsDead = isDeadStore(A, *SI, &AssumeOnlyInst);
4308 (void)IsDead;
4309 assert(IsDead && "Store was assumed to be dead!");
4310 A.deleteAfterManifest(*I);
4311 for (size_t i = 0; i < AssumeOnlyInst.size(); ++i) {
4312 Instruction *AOI = AssumeOnlyInst[i];
4313 for (auto *Usr : AOI->users())
4314 AssumeOnlyInst.insert(cast<Instruction>(Usr));
4315 A.deleteAfterManifest(*AOI);
4316 }
4317 return ChangeStatus::CHANGED;
4318 }
4319 if (auto *FI = dyn_cast<FenceInst>(I)) {
4320 assert(isDeadFence(A, *FI));
4321 A.deleteAfterManifest(*FI);
4322 return ChangeStatus::CHANGED;
4323 }
4324 if (isAssumedSideEffectFree(A, I) && !isa<InvokeInst>(I)) {
4325 A.deleteAfterManifest(*I);
4326 return ChangeStatus::CHANGED;
4327 }
4328 }
4330 }
4331
4332 /// See AbstractAttribute::trackStatistics()
4333 void trackStatistics() const override {
4335 }
4336
4337private:
4338 // The potential copies of a dead store, used for deletion during manifest.
4339 SmallSetVector<Value *, 4> PotentialCopies;
4340};
4341
4342struct AAIsDeadArgument : public AAIsDeadFloating {
4343 AAIsDeadArgument(const IRPosition &IRP, Attributor &A)
4344 : AAIsDeadFloating(IRP, A) {}
4345
4346 /// See AbstractAttribute::manifest(...).
4347 ChangeStatus manifest(Attributor &A) override {
4348 Argument &Arg = *getAssociatedArgument();
4349 if (A.isValidFunctionSignatureRewrite(Arg, /* ReplacementTypes */ {}))
4350 if (A.registerFunctionSignatureRewrite(
4351 Arg, /* ReplacementTypes */ {},
4354 return ChangeStatus::CHANGED;
4355 }
4356 return ChangeStatus::UNCHANGED;
4357 }
4358
4359 /// See AbstractAttribute::trackStatistics()
4360 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(IsDead) }
4361};
4362
4363struct AAIsDeadCallSiteArgument : public AAIsDeadValueImpl {
4364 AAIsDeadCallSiteArgument(const IRPosition &IRP, Attributor &A)
4365 : AAIsDeadValueImpl(IRP, A) {}
4366
4367 /// See AbstractAttribute::initialize(...).
4368 void initialize(Attributor &A) override {
4369 AAIsDeadValueImpl::initialize(A);
4370 if (isa<UndefValue>(getAssociatedValue()))
4371 indicatePessimisticFixpoint();
4372 }
4373
4374 /// See AbstractAttribute::updateImpl(...).
4375 ChangeStatus updateImpl(Attributor &A) override {
4376 // TODO: Once we have call site specific value information we can provide
4377 // call site specific liveness information and then it makes
4378 // sense to specialize attributes for call sites arguments instead of
4379 // redirecting requests to the callee argument.
4380 Argument *Arg = getAssociatedArgument();
4381 if (!Arg)
4382 return indicatePessimisticFixpoint();
4383 const IRPosition &ArgPos = IRPosition::argument(*Arg);
4384 auto *ArgAA = A.getAAFor<AAIsDead>(*this, ArgPos, DepClassTy::REQUIRED);
4385 if (!ArgAA)
4386 return indicatePessimisticFixpoint();
4387 return clampStateAndIndicateChange(getState(), ArgAA->getState());
4388 }
4389
4390 /// See AbstractAttribute::manifest(...).
4391 ChangeStatus manifest(Attributor &A) override {
4392 CallBase &CB = cast<CallBase>(getAnchorValue());
4393 Use &U = CB.getArgOperandUse(getCallSiteArgNo());
4394 assert(!isa<UndefValue>(U.get()) &&
4395 "Expected undef values to be filtered out!");
4396 UndefValue &UV = *UndefValue::get(U->getType());
4397 if (A.changeUseAfterManifest(U, UV))
4398 return ChangeStatus::CHANGED;
4399 return ChangeStatus::UNCHANGED;
4400 }
4401
4402 /// See AbstractAttribute::trackStatistics()
4403 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(IsDead) }
4404};
4405
4406struct AAIsDeadCallSiteReturned : public AAIsDeadFloating {
4407 AAIsDeadCallSiteReturned(const IRPosition &IRP, Attributor &A)
4408 : AAIsDeadFloating(IRP, A) {}
4409
4410 /// See AAIsDead::isAssumedDead().
4411 bool isAssumedDead() const override {
4412 return AAIsDeadFloating::isAssumedDead() && IsAssumedSideEffectFree;
4413 }
4414
4415 /// See AbstractAttribute::initialize(...).
4416 void initialize(Attributor &A) override {
4417 AAIsDeadFloating::initialize(A);
4418 if (isa<UndefValue>(getAssociatedValue())) {
4419 indicatePessimisticFixpoint();
4420 return;
4421 }
4422
4423 // We track this separately as a secondary state.
4424 IsAssumedSideEffectFree = isAssumedSideEffectFree(A, getCtxI());
4425 }
4426
4427 /// See AbstractAttribute::updateImpl(...).
4428 ChangeStatus updateImpl(Attributor &A) override {
4429 ChangeStatus Changed = ChangeStatus::UNCHANGED;
4430 if (IsAssumedSideEffectFree && !isAssumedSideEffectFree(A, getCtxI())) {
4431 IsAssumedSideEffectFree = false;
4432 Changed = ChangeStatus::CHANGED;
4433 }
4434 if (!areAllUsesAssumedDead(A, getAssociatedValue()))
4435 return indicatePessimisticFixpoint();
4436 return Changed;
4437 }
4438
4439 /// See AbstractAttribute::trackStatistics()
4440 void trackStatistics() const override {
4441 if (IsAssumedSideEffectFree)
4443 else
4444 STATS_DECLTRACK_CSRET_ATTR(UnusedResult)
4445 }
4446
4447 /// See AbstractAttribute::getAsStr().
4448 const std::string getAsStr(Attributor *A) const override {
4449 return isAssumedDead()
4450 ? "assumed-dead"
4451 : (getAssumed() ? "assumed-dead-users" : "assumed-live");
4452 }
4453
4454private:
4455 bool IsAssumedSideEffectFree = true;
4456};
4457
4458struct AAIsDeadReturned : public AAIsDeadValueImpl {
4459 AAIsDeadReturned(const IRPosition &IRP, Attributor &A)
4460 : AAIsDeadValueImpl(IRP, A) {}
4461
4462 /// See AbstractAttribute::updateImpl(...).
4463 ChangeStatus updateImpl(Attributor &A) override {
4464
4465 bool UsedAssumedInformation = false;
4466 A.checkForAllInstructions([](Instruction &) { return true; }, *this,
4467 {Instruction::Ret}, UsedAssumedInformation);
4468
4469 auto PredForCallSite = [&](AbstractCallSite ACS) {
4470 if (ACS.isCallbackCall() || !ACS.getInstruction())
4471 return false;
4472 return areAllUsesAssumedDead(A, *ACS.getInstruction());
4473 };
4474
4475 if (!A.checkForAllCallSites(PredForCallSite, *this, true,
4476 UsedAssumedInformation))
4477 return indicatePessimisticFixpoint();
4478
4479 return ChangeStatus::UNCHANGED;
4480 }
4481
4482 /// See AbstractAttribute::manifest(...).
4483 ChangeStatus manifest(Attributor &A) override {
4484 // TODO: Rewrite the signature to return void?
4485 bool AnyChange = false;
4486 UndefValue &UV = *UndefValue::get(getAssociatedFunction()->getReturnType());
4487 auto RetInstPred = [&](Instruction &I) {
4488 ReturnInst &RI = cast<ReturnInst>(I);
4489 if (!isa<UndefValue>(RI.getReturnValue()))
4490 AnyChange |= A.changeUseAfterManifest(RI.getOperandUse(0), UV);
4491 return true;
4492 };
4493 bool UsedAssumedInformation = false;
4494 A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
4495 UsedAssumedInformation);
4496 return AnyChange ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
4497 }
4498
4499 /// See AbstractAttribute::trackStatistics()
4500 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(IsDead) }
4501};
4502
4503struct AAIsDeadFunction : public AAIsDead {
4504 AAIsDeadFunction(const IRPosition &IRP, Attributor &A) : AAIsDead(IRP, A) {}
4505
4506 /// See AbstractAttribute::initialize(...).
4507 void initialize(Attributor &A) override {
4508 Function *F = getAnchorScope();
4509 assert(F && "Did expect an anchor function");
4510 if (!isAssumedDeadInternalFunction(A)) {
4511 ToBeExploredFrom.insert(&F->getEntryBlock().front());
4512 assumeLive(A, F->getEntryBlock());
4513 }
4514 }
4515
4516 bool isAssumedDeadInternalFunction(Attributor &A) {
4517 if (!getAnchorScope()->hasLocalLinkage())
4518 return false;
4519 bool UsedAssumedInformation = false;
4520 return A.checkForAllCallSites([](AbstractCallSite) { return false; }, *this,
4521 true, UsedAssumedInformation);
4522 }
4523
4524 /// See AbstractAttribute::getAsStr().
4525 const std::string getAsStr(Attributor *A) const override {
4526 return "Live[#BB " + std::to_string(AssumedLiveBlocks.size()) + "/" +
4527 std::to_string(getAnchorScope()->size()) + "][#TBEP " +
4528 std::to_string(ToBeExploredFrom.size()) + "][#KDE " +
4529 std::to_string(KnownDeadEnds.size()) + "]";
4530 }
4531
4532 /// See AbstractAttribute::manifest(...).
4533 ChangeStatus manifest(Attributor &A) override {
4534 assert(getState().isValidState() &&
4535 "Attempted to manifest an invalid state!");
4536
4537 ChangeStatus HasChanged = ChangeStatus::UNCHANGED;
4538 Function &F = *getAnchorScope();
4539
4540 if (AssumedLiveBlocks.empty()) {
4541 A.deleteAfterManifest(F);
4542 return ChangeStatus::CHANGED;
4543 }
4544
4545 // Flag to determine if we can change an invoke to a call assuming the
4546 // callee is nounwind. This is not possible if the personality of the
4547 // function allows to catch asynchronous exceptions.
4548 bool Invoke2CallAllowed = !mayCatchAsynchronousExceptions(F);
4549
4550 KnownDeadEnds.set_union(ToBeExploredFrom);
4551 for (const Instruction *DeadEndI : KnownDeadEnds) {
4552 auto *CB = dyn_cast<CallBase>(DeadEndI);
4553 if (!CB)
4554 continue;
4555 bool IsKnownNoReturn;
4556 bool MayReturn = !AA::hasAssumedIRAttr<Attribute::NoReturn>(
4557 A, this, IRPosition::callsite_function(*CB), DepClassTy::OPTIONAL,
4558 IsKnownNoReturn);
4559 if (MayReturn && (!Invoke2CallAllowed || !isa<InvokeInst>(CB)))
4560 continue;
4561
4562 if (auto *II = dyn_cast<InvokeInst>(DeadEndI))
4563 A.registerInvokeWithDeadSuccessor(const_cast<InvokeInst &>(*II));
4564 else
4565 A.changeToUnreachableAfterManifest(
4566 const_cast<Instruction *>(DeadEndI->getNextNode()));
4567 HasChanged = ChangeStatus::CHANGED;
4568 }
4569
4570 STATS_DECL(AAIsDead, BasicBlock, "Number of dead basic blocks deleted.");
4571 for (BasicBlock &BB : F)
4572 if (!AssumedLiveBlocks.count(&BB)) {
4573 A.deleteAfterManifest(BB);
4575 HasChanged = ChangeStatus::CHANGED;
4576 }
4577
4578 return HasChanged;
4579 }
4580
4581 /// See AbstractAttribute::updateImpl(...).
4582 ChangeStatus updateImpl(Attributor &A) override;
4583
4584 bool isEdgeDead(const BasicBlock *From, const BasicBlock *To) const override {
4585 assert(From->getParent() == getAnchorScope() &&
4586 To->getParent() == getAnchorScope() &&
4587 "Used AAIsDead of the wrong function");
4588 return isValidState() && !AssumedLiveEdges.count(std::make_pair(From, To));
4589 }
4590
4591 /// See AbstractAttribute::trackStatistics()
4592 void trackStatistics() const override {}
4593
4594 /// Returns true if the function is assumed dead.
4595 bool isAssumedDead() const override { return false; }
4596
4597 /// See AAIsDead::isKnownDead().
4598 bool isKnownDead() const override { return false; }
4599
4600 /// See AAIsDead::isAssumedDead(BasicBlock *).
4601 bool isAssumedDead(const BasicBlock *BB) const override {
4602 assert(BB->getParent() == getAnchorScope() &&
4603 "BB must be in the same anchor scope function.");
4604
4605 if (!getAssumed())
4606 return false;
4607 return !AssumedLiveBlocks.count(BB);
4608 }
4609
4610 /// See AAIsDead::isKnownDead(BasicBlock *).
4611 bool isKnownDead(const BasicBlock *BB) const override {
4612 return getKnown() && isAssumedDead(BB);
4613 }
4614
4615 /// See AAIsDead::isAssumed(Instruction *I).
4616 bool isAssumedDead(const Instruction *I) const override {
4617 assert(I->getParent()->getParent() == getAnchorScope() &&
4618 "Instruction must be in the same anchor scope function.");
4619
4620 if (!getAssumed())
4621 return false;
4622
4623 // If it is not in AssumedLiveBlocks then it for sure dead.
4624 // Otherwise, it can still be after noreturn call in a live block.
4625 if (!AssumedLiveBlocks.count(I->getParent()))
4626 return true;
4627
4628 // If it is not after a liveness barrier it is live.
4629 const Instruction *PrevI = I->getPrevNode();
4630 while (PrevI) {
4631 if (KnownDeadEnds.count(PrevI) || ToBeExploredFrom.count(PrevI))
4632 return true;
4633 PrevI = PrevI->getPrevNode();
4634 }
4635 return false;
4636 }
4637
4638 /// See AAIsDead::isKnownDead(Instruction *I).
4639 bool isKnownDead(const Instruction *I) const override {
4640 return getKnown() && isAssumedDead(I);
4641 }
4642
4643 /// Assume \p BB is (partially) live now and indicate to the Attributor \p A
4644 /// that internal function called from \p BB should now be looked at.
4645 bool assumeLive(Attributor &A, const BasicBlock &BB) {
4646 if (!AssumedLiveBlocks.insert(&BB).second)
4647 return false;
4648
4649 // We assume that all of BB is (probably) live now and if there are calls to
4650 // internal functions we will assume that those are now live as well. This
4651 // is a performance optimization for blocks with calls to a lot of internal
4652 // functions. It can however cause dead functions to be treated as live.
4653 for (const Instruction &I : BB)
4654 if (const auto *CB = dyn_cast<CallBase>(&I))
4655 if (auto *F = dyn_cast_if_present<Function>(CB->getCalledOperand()))
4656 if (F->hasLocalLinkage())
4657 A.markLiveInternalFunction(*F);
4658 return true;
4659 }
4660
4661 /// Collection of instructions that need to be explored again, e.g., we
4662 /// did assume they do not transfer control to (one of their) successors.
4664
4665 /// Collection of instructions that are known to not transfer control.
4667
4668 /// Collection of all assumed live edges
4670
4671 /// Collection of all assumed live BasicBlocks.
4672 DenseSet<const BasicBlock *> AssumedLiveBlocks;
4673};
4674
4675static bool
4676identifyAliveSuccessors(Attributor &A, const CallBase &CB,
4678 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4679 const IRPosition &IPos = IRPosition::callsite_function(CB);
4680
4681 bool IsKnownNoReturn;
4682 if (AA::hasAssumedIRAttr<Attribute::NoReturn>(
4683 A, &AA, IPos, DepClassTy::OPTIONAL, IsKnownNoReturn))
4684 return !IsKnownNoReturn;
4685 if (CB.isTerminator())
4686 AliveSuccessors.push_back(&CB.getSuccessor(0)->front());
4687 else
4688 AliveSuccessors.push_back(CB.getNextNode());
4689 return false;
4690}
4691
4692static bool
4693identifyAliveSuccessors(Attributor &A, const InvokeInst &II,
4695 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4696 bool UsedAssumedInformation =
4697 identifyAliveSuccessors(A, cast<CallBase>(II), AA, AliveSuccessors);
4698
4699 // First, determine if we can change an invoke to a call assuming the
4700 // callee is nounwind. This is not possible if the personality of the
4701 // function allows to catch asynchronous exceptions.
4702 if (AAIsDeadFunction::mayCatchAsynchronousExceptions(*II.getFunction())) {
4703 AliveSuccessors.push_back(&II.getUnwindDest()->front());
4704 } else {
4706
4707 bool IsKnownNoUnwind;
4708 if (AA::hasAssumedIRAttr<Attribute::NoUnwind>(
4709 A, &AA, IPos, DepClassTy::OPTIONAL, IsKnownNoUnwind)) {
4710 UsedAssumedInformation |= !IsKnownNoUnwind;
4711 } else {
4712 AliveSuccessors.push_back(&II.getUnwindDest()->front());
4713 }
4714 }
4715 return UsedAssumedInformation;
4716}
4717
4718static bool
4719identifyAliveSuccessors(Attributor &A, const BranchInst &BI,
4721 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4722 bool UsedAssumedInformation = false;
4723 if (BI.getNumSuccessors() == 1) {
4724 AliveSuccessors.push_back(&BI.getSuccessor(0)->front());
4725 } else {
4726 std::optional<Constant *> C =
4727 A.getAssumedConstant(*BI.getCondition(), AA, UsedAssumedInformation);
4728 if (!C || isa_and_nonnull<UndefValue>(*C)) {
4729 // No value yet, assume both edges are dead.
4730 } else if (isa_and_nonnull<ConstantInt>(*C)) {
4731 const BasicBlock *SuccBB =
4732 BI.getSuccessor(1 - cast<ConstantInt>(*C)->getValue().getZExtValue());
4733 AliveSuccessors.push_back(&SuccBB->front());
4734 } else {
4735 AliveSuccessors.push_back(&BI.getSuccessor(0)->front());
4736 AliveSuccessors.push_back(&BI.getSuccessor(1)->front());
4737 UsedAssumedInformation = false;
4738 }
4739 }
4740 return UsedAssumedInformation;
4741}
4742
4743static bool
4744identifyAliveSuccessors(Attributor &A, const SwitchInst &SI,
4746 SmallVectorImpl<const Instruction *> &AliveSuccessors) {
4747 bool UsedAssumedInformation = false;
4749 if (!A.getAssumedSimplifiedValues(IRPosition::value(*SI.getCondition()), &AA,
4750 Values, AA::AnyScope,
4751 UsedAssumedInformation)) {
4752 // Something went wrong, assume all successors are live.
4753 for (const BasicBlock *SuccBB : successors(SI.getParent()))
4754 AliveSuccessors.push_back(&SuccBB->front());
4755 return false;
4756 }
4757
4758 if (Values.empty() ||
4759 (Values.size() == 1 &&
4760 isa_and_nonnull<UndefValue>(Values.front().getValue()))) {
4761 // No valid value yet, assume all edges are dead.
4762 return UsedAssumedInformation;
4763 }
4764
4765 Type &Ty = *SI.getCondition()->getType();
4767 auto CheckForConstantInt = [&](Value *V) {
4768 if (auto *CI = dyn_cast_if_present<ConstantInt>(AA::getWithType(*V, Ty))) {
4769 Constants.insert(CI);
4770 return true;
4771 }
4772 return false;
4773 };
4774
4775 if (!all_of(Values, [&](AA::ValueAndContext &VAC) {
4776 return CheckForConstantInt(VAC.getValue());
4777 })) {
4778 for (const BasicBlock *SuccBB : successors(SI.getParent()))
4779 AliveSuccessors.push_back(&SuccBB->front());
4780 return UsedAssumedInformation;
4781 }
4782
4783 unsigned MatchedCases = 0;
4784 for (const auto &CaseIt : SI.cases()) {
4785 if (Constants.count(CaseIt.getCaseValue())) {
4786 ++MatchedCases;
4787 AliveSuccessors.push_back(&CaseIt.getCaseSuccessor()->front());
4788 }
4789 }
4790
4791 // If all potential values have been matched, we will not visit the default
4792 // case.
4793 if (MatchedCases < Constants.size())
4794 AliveSuccessors.push_back(&SI.getDefaultDest()->front());
4795 return UsedAssumedInformation;
4796}
4797
4798ChangeStatus AAIsDeadFunction::updateImpl(Attributor &A) {
4800
4801 if (AssumedLiveBlocks.empty()) {
4802 if (isAssumedDeadInternalFunction(A))
4804
4805 Function *F = getAnchorScope();
4806 ToBeExploredFrom.insert(&F->getEntryBlock().front());
4807 assumeLive(A, F->getEntryBlock());
4808 Change = ChangeStatus::CHANGED;
4809 }
4810
4811 LLVM_DEBUG(dbgs() << "[AAIsDead] Live [" << AssumedLiveBlocks.size() << "/"
4812 << getAnchorScope()->size() << "] BBs and "
4813 << ToBeExploredFrom.size() << " exploration points and "
4814 << KnownDeadEnds.size() << " known dead ends\n");
4815
4816 // Copy and clear the list of instructions we need to explore from. It is
4817 // refilled with instructions the next update has to look at.
4818 SmallVector<const Instruction *, 8> Worklist(ToBeExploredFrom.begin(),
4819 ToBeExploredFrom.end());
4820 decltype(ToBeExploredFrom) NewToBeExploredFrom;
4821
4823 while (!Worklist.empty()) {
4824 const Instruction *I = Worklist.pop_back_val();
4825 LLVM_DEBUG(dbgs() << "[AAIsDead] Exploration inst: " << *I << "\n");
4826
4827 // Fast forward for uninteresting instructions. We could look for UB here
4828 // though.
4829 while (!I->isTerminator() && !isa<CallBase>(I))
4830 I = I->getNextNode();
4831
4832 AliveSuccessors.clear();
4833
4834 bool UsedAssumedInformation = false;
4835 switch (I->getOpcode()) {
4836 // TODO: look for (assumed) UB to backwards propagate "deadness".
4837 default:
4838 assert(I->isTerminator() &&
4839 "Expected non-terminators to be handled already!");
4840 for (const BasicBlock *SuccBB : successors(I->getParent()))
4841 AliveSuccessors.push_back(&SuccBB->front());
4842 break;
4843 case Instruction::Call:
4844 UsedAssumedInformation = identifyAliveSuccessors(A, cast<CallInst>(*I),
4845 *this, AliveSuccessors);
4846 break;
4847 case Instruction::Invoke:
4848 UsedAssumedInformation = identifyAliveSuccessors(A, cast<InvokeInst>(*I),
4849 *this, AliveSuccessors);
4850 break;
4851 case Instruction::Br:
4852 UsedAssumedInformation = identifyAliveSuccessors(A, cast<BranchInst>(*I),
4853 *this, AliveSuccessors);
4854 break;
4855 case Instruction::Switch:
4856 UsedAssumedInformation = identifyAliveSuccessors(A, cast<SwitchInst>(*I),
4857 *this, AliveSuccessors);
4858 break;
4859 }
4860
4861 if (UsedAssumedInformation) {
4862 NewToBeExploredFrom.insert(I);
4863 } else if (AliveSuccessors.empty() ||
4864 (I->isTerminator() &&
4865 AliveSuccessors.size() < I->getNumSuccessors())) {
4866 if (KnownDeadEnds.insert(I))
4867 Change = ChangeStatus::CHANGED;
4868 }
4869
4870 LLVM_DEBUG(dbgs() << "[AAIsDead] #AliveSuccessors: "
4871 << AliveSuccessors.size() << " UsedAssumedInformation: "
4872 << UsedAssumedInformation << "\n");
4873
4874 for (const Instruction *AliveSuccessor : AliveSuccessors) {
4875 if (!I->isTerminator()) {
4876 assert(AliveSuccessors.size() == 1 &&
4877 "Non-terminator expected to have a single successor!");
4878 Worklist.push_back(AliveSuccessor);
4879 } else {
4880 // record the assumed live edge
4881 auto Edge = std::make_pair(I->getParent(), AliveSuccessor->getParent());
4882 if (AssumedLiveEdges.insert(Edge).second)
4883 Change = ChangeStatus::CHANGED;
4884 if (assumeLive(A, *AliveSuccessor->getParent()))
4885 Worklist.push_back(AliveSuccessor);
4886 }
4887 }
4888 }
4889
4890 // Check if the content of ToBeExploredFrom changed, ignore the order.
4891 if (NewToBeExploredFrom.size() != ToBeExploredFrom.size() ||
4892 llvm::any_of(NewToBeExploredFrom, [&](const Instruction *I) {
4893 return !ToBeExploredFrom.count(I);
4894 })) {
4895 Change = ChangeStatus::CHANGED;
4896 ToBeExploredFrom = std::move(NewToBeExploredFrom);
4897 }
4898
4899 // If we know everything is live there is no need to query for liveness.
4900 // Instead, indicating a pessimistic fixpoint will cause the state to be
4901 // "invalid" and all queries to be answered conservatively without lookups.
4902 // To be in this state we have to (1) finished the exploration and (3) not
4903 // discovered any non-trivial dead end and (2) not ruled unreachable code
4904 // dead.
4905 if (ToBeExploredFrom.empty() &&
4906 getAnchorScope()->size() == AssumedLiveBlocks.size() &&
4907 llvm::all_of(KnownDeadEnds, [](const Instruction *DeadEndI) {
4908 return DeadEndI->isTerminator() && DeadEndI->getNumSuccessors() == 0;
4909 }))
4910 return indicatePessimisticFixpoint();
4911 return Change;
4912}
4913
4914/// Liveness information for a call sites.
4915struct AAIsDeadCallSite final : AAIsDeadFunction {
4916 AAIsDeadCallSite(const IRPosition &IRP, Attributor &A)
4917 : AAIsDeadFunction(IRP, A) {}
4918
4919 /// See AbstractAttribute::initialize(...).
4920 void initialize(Attributor &A) override {
4921 // TODO: Once we have call site specific value information we can provide
4922 // call site specific liveness information and then it makes
4923 // sense to specialize attributes for call sites instead of
4924 // redirecting requests to the callee.
4925 llvm_unreachable("Abstract attributes for liveness are not "
4926 "supported for call sites yet!");
4927 }
4928
4929 /// See AbstractAttribute::updateImpl(...).
4930 ChangeStatus updateImpl(Attributor &A) override {
4931 return indicatePessimisticFixpoint();
4932 }
4933
4934 /// See AbstractAttribute::trackStatistics()
4935 void trackStatistics() const override {}
4936};
4937} // namespace
4938
4939/// -------------------- Dereferenceable Argument Attribute --------------------
4940
4941namespace {
4942struct AADereferenceableImpl : AADereferenceable {
4943 AADereferenceableImpl(const IRPosition &IRP, Attributor &A)
4944 : AADereferenceable(IRP, A) {}
4945 using StateType = DerefState;
4946
4947 /// See AbstractAttribute::initialize(...).
4948 void initialize(Attributor &A) override {
4949 Value &V = *getAssociatedValue().stripPointerCasts();
4951 A.getAttrs(getIRPosition(),
4952 {Attribute::Dereferenceable, Attribute::DereferenceableOrNull},
4953 Attrs, /* IgnoreSubsumingPositions */ false);
4954 for (const Attribute &Attr : Attrs)
4955 takeKnownDerefBytesMaximum(Attr.getValueAsInt());
4956
4957 // Ensure we initialize the non-null AA (if necessary).
4958 bool IsKnownNonNull;
4959 AA::hasAssumedIRAttr<Attribute::NonNull>(
4960 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNonNull);
4961
4962 bool CanBeNull, CanBeFreed;
4963 takeKnownDerefBytesMaximum(V.getPointerDereferenceableBytes(
4964 A.getDataLayout(), CanBeNull, CanBeFreed));
4965
4966 if (Instruction *CtxI = getCtxI())
4967 followUsesInMBEC(*this, A, getState(), *CtxI);
4968 }
4969
4970 /// See AbstractAttribute::getState()
4971 /// {
4972 StateType &getState() override { return *this; }
4973 const StateType &getState() const override { return *this; }
4974 /// }
4975
4976 /// Helper function for collecting accessed bytes in must-be-executed-context
4977 void addAccessedBytesForUse(Attributor &A, const Use *U, const Instruction *I,
4978 DerefState &State) {
4979 const Value *UseV = U->get();
4980 if (!UseV->getType()->isPointerTy())
4981 return;
4982
4983 std::optional<MemoryLocation> Loc = MemoryLocation::getOrNone(I);
4984 if (!Loc || Loc->Ptr != UseV || !Loc->Size.isPrecise() || I->isVolatile())
4985 return;
4986
4987 int64_t Offset;
4989 Loc->Ptr, Offset, A.getDataLayout(), /*AllowNonInbounds*/ true);
4990 if (Base && Base == &getAssociatedValue())
4991 State.addAccessedBytes(Offset, Loc->Size.getValue());
4992 }
4993
4994 /// See followUsesInMBEC
4995 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
4997 bool IsNonNull = false;
4998 bool TrackUse = false;
4999 int64_t DerefBytes = getKnownNonNullAndDerefBytesForUse(
5000 A, *this, getAssociatedValue(), U, I, IsNonNull, TrackUse);
5001 LLVM_DEBUG(dbgs() << "[AADereferenceable] Deref bytes: " << DerefBytes
5002 << " for instruction " << *I << "\n");
5003
5004 addAccessedBytesForUse(A, U, I, State);
5005 State.takeKnownDerefBytesMaximum(DerefBytes);
5006 return TrackUse;
5007 }
5008
5009 /// See AbstractAttribute::manifest(...).
5010 ChangeStatus manifest(Attributor &A) override {
5011 ChangeStatus Change = AADereferenceable::manifest(A);
5012 bool IsKnownNonNull;
5013 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
5014 A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5015 if (IsAssumedNonNull &&
5016 A.hasAttr(getIRPosition(), Attribute::DereferenceableOrNull)) {
5017 A.removeAttrs(getIRPosition(), {Attribute::DereferenceableOrNull});
5018 return ChangeStatus::CHANGED;
5019 }
5020 return Change;
5021 }
5022
5023 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5024 SmallVectorImpl<Attribute> &Attrs) const override {
5025 // TODO: Add *_globally support
5026 bool IsKnownNonNull;
5027 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
5028 A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5029 if (IsAssumedNonNull)
5031 Ctx, getAssumedDereferenceableBytes()));
5032 else
5034 Ctx, getAssumedDereferenceableBytes()));
5035 }
5036
5037 /// See AbstractAttribute::getAsStr().
5038 const std::string getAsStr(Attributor *A) const override {
5039 if (!getAssumedDereferenceableBytes())
5040 return "unknown-dereferenceable";
5041 bool IsKnownNonNull;
5042 bool IsAssumedNonNull = false;
5043 if (A)
5044 IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
5045 *A, this, getIRPosition(), DepClassTy::NONE, IsKnownNonNull);
5046 return std::string("dereferenceable") +
5047 (IsAssumedNonNull ? "" : "_or_null") +
5048 (isAssumedGlobal() ? "_globally" : "") + "<" +
5049 std::to_string(getKnownDereferenceableBytes()) + "-" +
5050 std::to_string(getAssumedDereferenceableBytes()) + ">" +
5051 (!A ? " [non-null is unknown]" : "");
5052 }
5053};
5054
5055/// Dereferenceable attribute for a floating value.
5056struct AADereferenceableFloating : AADereferenceableImpl {
5057 AADereferenceableFloating(const IRPosition &IRP, Attributor &A)
5058 : AADereferenceableImpl(IRP, A) {}
5059
5060 /// See AbstractAttribute::updateImpl(...).
5061 ChangeStatus updateImpl(Attributor &A) override {
5062 bool Stripped;
5063 bool UsedAssumedInformation = false;
5065 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
5066 AA::AnyScope, UsedAssumedInformation)) {
5067 Values.push_back({getAssociatedValue(), getCtxI()});
5068 Stripped = false;
5069 } else {
5070 Stripped = Values.size() != 1 ||
5071 Values.front().getValue() != &getAssociatedValue();
5072 }
5073
5074 const DataLayout &DL = A.getDataLayout();
5075 DerefState T;
5076
5077 auto VisitValueCB = [&](const Value &V) -> bool {
5078 unsigned IdxWidth =
5079 DL.getIndexSizeInBits(V.getType()->getPointerAddressSpace());
5080 APInt Offset(IdxWidth, 0);
5082 A, *this, &V, DL, Offset, /* GetMinOffset */ false,
5083 /* AllowNonInbounds */ true);
5084
5085 const auto *AA = A.getAAFor<AADereferenceable>(
5086 *this, IRPosition::value(*Base), DepClassTy::REQUIRED);
5087 int64_t DerefBytes = 0;
5088 if (!AA || (!Stripped && this == AA)) {
5089 // Use IR information if we did not strip anything.
5090 // TODO: track globally.
5091 bool CanBeNull, CanBeFreed;
5092 DerefBytes =
5093 Base->getPointerDereferenceableBytes(DL, CanBeNull, CanBeFreed);
5094 T.GlobalState.indicatePessimisticFixpoint();
5095 } else {
5096 const DerefState &DS = AA->getState();
5097 DerefBytes = DS.DerefBytesState.getAssumed();
5098 T.GlobalState &= DS.GlobalState;
5099 }
5100
5101 // For now we do not try to "increase" dereferenceability due to negative
5102 // indices as we first have to come up with code to deal with loops and
5103 // for overflows of the dereferenceable bytes.
5104 int64_t OffsetSExt = Offset.getSExtValue();
5105 if (OffsetSExt < 0)
5106 OffsetSExt = 0;
5107
5108 T.takeAssumedDerefBytesMinimum(
5109 std::max(int64_t(0), DerefBytes - OffsetSExt));
5110
5111 if (this == AA) {
5112 if (!Stripped) {
5113 // If nothing was stripped IR information is all we got.
5114 T.takeKnownDerefBytesMaximum(
5115 std::max(int64_t(0), DerefBytes - OffsetSExt));
5116 T.indicatePessimisticFixpoint();
5117 } else if (OffsetSExt > 0) {
5118 // If something was stripped but there is circular reasoning we look
5119 // for the offset. If it is positive we basically decrease the
5120 // dereferenceable bytes in a circular loop now, which will simply
5121 // drive them down to the known value in a very slow way which we
5122 // can accelerate.
5123 T.indicatePessimisticFixpoint();
5124 }
5125 }
5126
5127 return T.isValidState();
5128 };
5129
5130 for (const auto &VAC : Values)
5131 if (!VisitValueCB(*VAC.getValue()))
5132 return indicatePessimisticFixpoint();
5133
5134 return clampStateAndIndicateChange(getState(), T);
5135 }
5136
5137 /// See AbstractAttribute::trackStatistics()
5138 void trackStatistics() const override {
5139 STATS_DECLTRACK_FLOATING_ATTR(dereferenceable)
5140 }
5141};
5142
5143/// Dereferenceable attribute for a return value.
5144struct AADereferenceableReturned final
5145 : AAReturnedFromReturnedValues<AADereferenceable, AADereferenceableImpl> {
5146 using Base =
5147 AAReturnedFromReturnedValues<AADereferenceable, AADereferenceableImpl>;
5148 AADereferenceableReturned(const IRPosition &IRP, Attributor &A)
5149 : Base(IRP, A) {}
5150
5151 /// See AbstractAttribute::trackStatistics()
5152 void trackStatistics() const override {
5153 STATS_DECLTRACK_FNRET_ATTR(dereferenceable)
5154 }
5155};
5156
5157/// Dereferenceable attribute for an argument
5158struct AADereferenceableArgument final
5159 : AAArgumentFromCallSiteArguments<AADereferenceable,
5160 AADereferenceableImpl> {
5161 using Base =
5162 AAArgumentFromCallSiteArguments<AADereferenceable, AADereferenceableImpl>;
5163 AADereferenceableArgument(const IRPosition &IRP, Attributor &A)
5164 : Base(IRP, A) {}
5165
5166 /// See AbstractAttribute::trackStatistics()
5167 void trackStatistics() const override {
5168 STATS_DECLTRACK_ARG_ATTR(dereferenceable)
5169 }
5170};
5171
5172/// Dereferenceable attribute for a call site argument.
5173struct AADereferenceableCallSiteArgument final : AADereferenceableFloating {
5174 AADereferenceableCallSiteArgument(const IRPosition &IRP, Attributor &A)
5175 : AADereferenceableFloating(IRP, A) {}
5176
5177 /// See AbstractAttribute::trackStatistics()
5178 void trackStatistics() const override {
5179 STATS_DECLTRACK_CSARG_ATTR(dereferenceable)
5180 }
5181};
5182
5183/// Dereferenceable attribute deduction for a call site return value.
5184struct AADereferenceableCallSiteReturned final
5185 : AACalleeToCallSite<AADereferenceable, AADereferenceableImpl> {
5186 using Base = AACalleeToCallSite<AADereferenceable, AADereferenceableImpl>;
5187 AADereferenceableCallSiteReturned(const IRPosition &IRP, Attributor &A)
5188 : Base(IRP, A) {}
5189
5190 /// See AbstractAttribute::trackStatistics()
5191 void trackStatistics() const override {
5192 STATS_DECLTRACK_CS_ATTR(dereferenceable);
5193 }
5194};
5195} // namespace
5196
5197// ------------------------ Align Argument Attribute ------------------------
5198
5199namespace {
5200static unsigned getKnownAlignForUse(Attributor &A, AAAlign &QueryingAA,
5201 Value &AssociatedValue, const Use *U,
5202 const Instruction *I, bool &TrackUse) {
5203 // We need to follow common pointer manipulation uses to the accesses they
5204 // feed into.
5205 if (isa<CastInst>(I)) {
5206 // Follow all but ptr2int casts.
5207 TrackUse = !isa<PtrToIntInst>(I);
5208 return 0;
5209 }
5210 if (auto *GEP = dyn_cast<GetElementPtrInst>(I)) {
5211 if (GEP->hasAllConstantIndices())
5212 TrackUse = true;
5213 return 0;
5214 }
5215
5216 MaybeAlign MA;
5217 if (const auto *CB = dyn_cast<CallBase>(I)) {
5218 if (CB->isBundleOperand(U) || CB->isCallee(U))
5219 return 0;
5220
5221 unsigned ArgNo = CB->getArgOperandNo(U);
5222 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
5223 // As long as we only use known information there is no need to track
5224 // dependences here.
5225 auto *AlignAA = A.getAAFor<AAAlign>(QueryingAA, IRP, DepClassTy::NONE);
5226 if (AlignAA)
5227 MA = MaybeAlign(AlignAA->getKnownAlign());
5228 }
5229
5230 const DataLayout &DL = A.getDataLayout();
5231 const Value *UseV = U->get();
5232 if (auto *SI = dyn_cast<StoreInst>(I)) {
5233 if (SI->getPointerOperand() == UseV)
5234 MA = SI->getAlign();
5235 } else if (auto *LI = dyn_cast<LoadInst>(I)) {
5236 if (LI->getPointerOperand() == UseV)
5237 MA = LI->getAlign();
5238 } else if (auto *AI = dyn_cast<AtomicRMWInst>(I)) {
5239 if (AI->getPointerOperand() == UseV)
5240 MA = AI->getAlign();
5241 } else if (auto *AI = dyn_cast<AtomicCmpXchgInst>(I)) {
5242 if (AI->getPointerOperand() == UseV)
5243 MA = AI->getAlign();
5244 }
5245
5246 if (!MA || *MA <= QueryingAA.getKnownAlign())
5247 return 0;
5248
5249 unsigned Alignment = MA->value();
5250 int64_t Offset;
5251
5252 if (const Value *Base = GetPointerBaseWithConstantOffset(UseV, Offset, DL)) {
5253 if (Base == &AssociatedValue) {
5254 // BasePointerAddr + Offset = Alignment * Q for some integer Q.
5255 // So we can say that the maximum power of two which is a divisor of
5256 // gcd(Offset, Alignment) is an alignment.
5257
5258 uint32_t gcd = std::gcd(uint32_t(abs((int32_t)Offset)), Alignment);
5259 Alignment = llvm::bit_floor(gcd);
5260 }
5261 }
5262
5263 return Alignment;
5264}
5265
5266struct AAAlignImpl : AAAlign {
5267 AAAlignImpl(const IRPosition &IRP, Attributor &A) : AAAlign(IRP, A) {}
5268
5269 /// See AbstractAttribute::initialize(...).
5270 void initialize(Attributor &A) override {
5272 A.getAttrs(getIRPosition(), {Attribute::Alignment}, Attrs);
5273 for (const Attribute &Attr : Attrs)
5274 takeKnownMaximum(Attr.getValueAsInt());
5275
5276 Value &V = *getAssociatedValue().stripPointerCasts();
5277 takeKnownMaximum(V.getPointerAlignment(A.getDataLayout()).value());
5278
5279 if (Instruction *CtxI = getCtxI())
5280 followUsesInMBEC(*this, A, getState(), *CtxI);
5281 }
5282
5283 /// See AbstractAttribute::manifest(...).
5284 ChangeStatus manifest(Attributor &A) override {
5285 ChangeStatus LoadStoreChanged = ChangeStatus::UNCHANGED;
5286
5287 // Check for users that allow alignment annotations.
5288 Value &AssociatedValue = getAssociatedValue();
5289 for (const Use &U : AssociatedValue.uses()) {
5290 if (auto *SI = dyn_cast<StoreInst>(U.getUser())) {
5291 if (SI->getPointerOperand() == &AssociatedValue)
5292 if (SI->getAlign() < getAssumedAlign()) {
5293 STATS_DECLTRACK(AAAlign, Store,
5294 "Number of times alignment added to a store");
5295 SI->setAlignment(getAssumedAlign());
5296 LoadStoreChanged = ChangeStatus::CHANGED;
5297 }
5298 } else if (auto *LI = dyn_cast<LoadInst>(U.getUser())) {
5299 if (LI->getPointerOperand() == &AssociatedValue)
5300 if (LI->getAlign() < getAssumedAlign()) {
5301 LI->setAlignment(getAssumedAlign());
5303 "Number of times alignment added to a load");
5304 LoadStoreChanged = ChangeStatus::CHANGED;
5305 }
5306 }
5307 }
5308
5309 ChangeStatus Changed = AAAlign::manifest(A);
5310
5311 Align InheritAlign =
5312 getAssociatedValue().getPointerAlignment(A.getDataLayout());
5313 if (InheritAlign >= getAssumedAlign())
5314 return LoadStoreChanged;
5315 return Changed | LoadStoreChanged;
5316 }
5317
5318 // TODO: Provide a helper to determine the implied ABI alignment and check in
5319 // the existing manifest method and a new one for AAAlignImpl that value
5320 // to avoid making the alignment explicit if it did not improve.
5321
5322 /// See AbstractAttribute::getDeducedAttributes
5323 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5324 SmallVectorImpl<Attribute> &Attrs) const override {
5325 if (getAssumedAlign() > 1)
5326 Attrs.emplace_back(
5327 Attribute::getWithAlignment(Ctx, Align(getAssumedAlign())));
5328 }
5329
5330 /// See followUsesInMBEC
5331 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
5332 AAAlign::StateType &State) {
5333 bool TrackUse = false;
5334
5335 unsigned int KnownAlign =
5336 getKnownAlignForUse(A, *this, getAssociatedValue(), U, I, TrackUse);
5337 State.takeKnownMaximum(KnownAlign);
5338
5339 return TrackUse;
5340 }
5341
5342 /// See AbstractAttribute::getAsStr().
5343 const std::string getAsStr(Attributor *A) const override {
5344 return "align<" + std::to_string(getKnownAlign().value()) + "-" +
5345 std::to_string(getAssumedAlign().value()) + ">";
5346 }
5347};
5348
5349/// Align attribute for a floating value.
5350struct AAAlignFloating : AAAlignImpl {
5351 AAAlignFloating(const IRPosition &IRP, Attributor &A) : AAAlignImpl(IRP, A) {}
5352
5353 /// See AbstractAttribute::updateImpl(...).
5354 ChangeStatus updateImpl(Attributor &A) override {
5355 const DataLayout &DL = A.getDataLayout();
5356
5357 bool Stripped;
5358 bool UsedAssumedInformation = false;
5360 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
5361 AA::AnyScope, UsedAssumedInformation)) {
5362 Values.push_back({getAssociatedValue(), getCtxI()});
5363 Stripped = false;
5364 } else {
5365 Stripped = Values.size() != 1 ||
5366 Values.front().getValue() != &getAssociatedValue();
5367 }
5368
5369 StateType T;
5370 auto VisitValueCB = [&](Value &V) -> bool {
5371 if (isa<UndefValue>(V) || isa<ConstantPointerNull>(V))
5372 return true;
5373 const auto *AA = A.getAAFor<AAAlign>(*this, IRPosition::value(V),
5374 DepClassTy::REQUIRED);
5375 if (!AA || (!Stripped && this == AA)) {
5376 int64_t Offset;
5377 unsigned Alignment = 1;
5378 if (const Value *Base =
5380 // TODO: Use AAAlign for the base too.
5381 Align PA = Base->getPointerAlignment(DL);
5382 // BasePointerAddr + Offset = Alignment * Q for some integer Q.
5383 // So we can say that the maximum power of two which is a divisor of
5384 // gcd(Offset, Alignment) is an alignment.
5385
5386 uint32_t gcd =
5387 std::gcd(uint32_t(abs((int32_t)Offset)), uint32_t(PA.value()));
5388 Alignment = llvm::bit_floor(gcd);
5389 } else {
5390 Alignment = V.getPointerAlignment(DL).value();
5391 }
5392 // Use only IR information if we did not strip anything.
5393 T.takeKnownMaximum(Alignment);
5394 T.indicatePessimisticFixpoint();
5395 } else {
5396 // Use abstract attribute information.
5397 const AAAlign::StateType &DS = AA->getState();
5398 T ^= DS;
5399 }
5400 return T.isValidState();
5401 };
5402
5403 for (const auto &VAC : Values) {
5404 if (!VisitValueCB(*VAC.getValue()))
5405 return indicatePessimisticFixpoint();
5406 }
5407
5408 // TODO: If we know we visited all incoming values, thus no are assumed
5409 // dead, we can take the known information from the state T.
5410 return clampStateAndIndicateChange(getState(), T);
5411 }
5412
5413 /// See AbstractAttribute::trackStatistics()
5414 void trackStatistics() const override { STATS_DECLTRACK_FLOATING_ATTR(align) }
5415};
5416
5417/// Align attribute for function return value.
5418struct AAAlignReturned final
5419 : AAReturnedFromReturnedValues<AAAlign, AAAlignImpl> {
5420 using Base = AAReturnedFromReturnedValues<AAAlign, AAAlignImpl>;
5421 AAAlignReturned(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
5422
5423 /// See AbstractAttribute::trackStatistics()
5424 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(aligned) }
5425};
5426
5427/// Align attribute for function argument.
5428struct AAAlignArgument final
5429 : AAArgumentFromCallSiteArguments<AAAlign, AAAlignImpl> {
5430 using Base = AAArgumentFromCallSiteArguments<AAAlign, AAAlignImpl>;
5431 AAAlignArgument(const IRPosition &IRP, Attributor &A) : Base(IRP, A) {}
5432
5433 /// See AbstractAttribute::manifest(...).
5434 ChangeStatus manifest(Attributor &A) override {
5435 // If the associated argument is involved in a must-tail call we give up
5436 // because we would need to keep the argument alignments of caller and
5437 // callee in-sync. Just does not seem worth the trouble right now.
5438 if (A.getInfoCache().isInvolvedInMustTailCall(*getAssociatedArgument()))
5439 return ChangeStatus::UNCHANGED;
5440 return Base::manifest(A);
5441 }
5442
5443 /// See AbstractAttribute::trackStatistics()
5444 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(aligned) }
5445};
5446
5447struct AAAlignCallSiteArgument final : AAAlignFloating {
5448 AAAlignCallSiteArgument(const IRPosition &IRP, Attributor &A)
5449 : AAAlignFloating(IRP, A) {}
5450
5451 /// See AbstractAttribute::manifest(...).
5452 ChangeStatus manifest(Attributor &A) override {
5453 // If the associated argument is involved in a must-tail call we give up
5454 // because we would need to keep the argument alignments of caller and
5455 // callee in-sync. Just does not seem worth the trouble right now.
5456 if (Argument *Arg = getAssociatedArgument())
5457 if (A.getInfoCache().isInvolvedInMustTailCall(*Arg))
5458 return ChangeStatus::UNCHANGED;
5459 ChangeStatus Changed = AAAlignImpl::manifest(A);
5460 Align InheritAlign =
5461 getAssociatedValue().getPointerAlignment(A.getDataLayout());
5462 if (InheritAlign >= getAssumedAlign())
5463 Changed = ChangeStatus::UNCHANGED;
5464 return Changed;
5465 }
5466
5467 /// See AbstractAttribute::updateImpl(Attributor &A).
5468 ChangeStatus updateImpl(Attributor &A) override {
5469 ChangeStatus Changed = AAAlignFloating::updateImpl(A);
5470 if (Argument *Arg = getAssociatedArgument()) {
5471 // We only take known information from the argument
5472 // so we do not need to track a dependence.
5473 const auto *ArgAlignAA = A.getAAFor<AAAlign>(
5474 *this, IRPosition::argument(*Arg), DepClassTy::NONE);
5475 if (ArgAlignAA)
5476 takeKnownMaximum(ArgAlignAA->getKnownAlign().value());
5477 }
5478 return Changed;
5479 }
5480
5481 /// See AbstractAttribute::trackStatistics()
5482 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(aligned) }
5483};
5484
5485/// Align attribute deduction for a call site return value.
5486struct AAAlignCallSiteReturned final
5487 : AACalleeToCallSite<AAAlign, AAAlignImpl> {
5488 using Base = AACalleeToCallSite<AAAlign, AAAlignImpl>;
5489 AAAlignCallSiteReturned(const IRPosition &IRP, Attributor &A)
5490 : Base(IRP, A) {}
5491
5492 /// See AbstractAttribute::trackStatistics()
5493 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(align); }
5494};
5495} // namespace
5496
5497/// ------------------ Function No-Return Attribute ----------------------------
5498namespace {
5499struct AANoReturnImpl : public AANoReturn {
5500 AANoReturnImpl(const IRPosition &IRP, Attributor &A) : AANoReturn(IRP, A) {}
5501
5502 /// See AbstractAttribute::initialize(...).
5503 void initialize(Attributor &A) override {
5504 bool IsKnown;
5505 assert(!AA::hasAssumedIRAttr<Attribute::NoReturn>(
5506 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
5507 (void)IsKnown;
5508 }
5509
5510 /// See AbstractAttribute::getAsStr().
5511 const std::string getAsStr(Attributor *A) const override {
5512 return getAssumed() ? "noreturn" : "may-return";
5513 }
5514
5515 /// See AbstractAttribute::updateImpl(Attributor &A).
5516 ChangeStatus updateImpl(Attributor &A) override {
5517 auto CheckForNoReturn = [](Instruction &) { return false; };
5518 bool UsedAssumedInformation = false;
5519 if (!A.checkForAllInstructions(CheckForNoReturn, *this,
5520 {(unsigned)Instruction::Ret},
5521 UsedAssumedInformation))
5522 return indicatePessimisticFixpoint();
5523 return ChangeStatus::UNCHANGED;
5524 }
5525};
5526
5527struct AANoReturnFunction final : AANoReturnImpl {
5528 AANoReturnFunction(const IRPosition &IRP, Attributor &A)
5529 : AANoReturnImpl(IRP, A) {}
5530
5531 /// See AbstractAttribute::trackStatistics()
5532 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(noreturn) }
5533};
5534
5535/// NoReturn attribute deduction for a call sites.
5536struct AANoReturnCallSite final
5537 : AACalleeToCallSite<AANoReturn, AANoReturnImpl> {
5538 AANoReturnCallSite(const IRPosition &IRP, Attributor &A)
5539 : AACalleeToCallSite<AANoReturn, AANoReturnImpl>(IRP, A) {}
5540
5541 /// See AbstractAttribute::trackStatistics()
5542 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(noreturn); }
5543};
5544} // namespace
5545
5546/// ----------------------- Instance Info ---------------------------------
5547
5548namespace {
5549/// A class to hold the state of for no-capture attributes.
5550struct AAInstanceInfoImpl : public AAInstanceInfo {
5551 AAInstanceInfoImpl(const IRPosition &IRP, Attributor &A)
5552 : AAInstanceInfo(IRP, A) {}
5553
5554 /// See AbstractAttribute::initialize(...).
5555 void initialize(Attributor &A) override {
5556 Value &V = getAssociatedValue();
5557 if (auto *C = dyn_cast<Constant>(&V)) {
5558 if (C->isThreadDependent())
5559 indicatePessimisticFixpoint();
5560 else
5561 indicateOptimisticFixpoint();
5562 return;
5563 }
5564 if (auto *CB = dyn_cast<CallBase>(&V))
5565 if (CB->arg_size() == 0 && !CB->mayHaveSideEffects() &&
5566 !CB->mayReadFromMemory()) {
5567 indicateOptimisticFixpoint();
5568 return;
5569 }
5570 if (auto *I = dyn_cast<Instruction>(&V)) {
5571 const auto *CI =
5572 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
5573 *I->getFunction());
5574 if (mayBeInCycle(CI, I, /* HeaderOnly */ false)) {
5575 indicatePessimisticFixpoint();
5576 return;
5577 }
5578 }
5579 }
5580
5581 /// See AbstractAttribute::updateImpl(...).
5582 ChangeStatus updateImpl(Attributor &A) override {
5583 ChangeStatus Changed = ChangeStatus::UNCHANGED;
5584
5585 Value &V = getAssociatedValue();
5586 const Function *Scope = nullptr;
5587 if (auto *I = dyn_cast<Instruction>(&V))
5588 Scope = I->getFunction();
5589 if (auto *A = dyn_cast<Argument>(&V)) {
5590 Scope = A->getParent();
5591 if (!Scope->hasLocalLinkage())
5592 return Changed;
5593 }
5594 if (!Scope)
5595 return indicateOptimisticFixpoint();
5596
5597 bool IsKnownNoRecurse;
5598 if (AA::hasAssumedIRAttr<Attribute::NoRecurse>(
5599 A, this, IRPosition::function(*Scope), DepClassTy::OPTIONAL,
5600 IsKnownNoRecurse))
5601 return Changed;
5602
5603 auto UsePred = [&](const Use &U, bool &Follow) {
5604 const Instruction *UserI = dyn_cast<Instruction>(U.getUser());
5605 if (!UserI || isa<GetElementPtrInst>(UserI) || isa<CastInst>(UserI) ||
5606 isa<PHINode>(UserI) || isa<SelectInst>(UserI)) {
5607 Follow = true;
5608 return true;
5609 }
5610 if (isa<LoadInst>(UserI) || isa<CmpInst>(UserI) ||
5611 (isa<StoreInst>(UserI) &&
5612 cast<StoreInst>(UserI)->getValueOperand() != U.get()))
5613 return true;
5614 if (auto *CB = dyn_cast<CallBase>(UserI)) {
5615 // This check is not guaranteeing uniqueness but for now that we cannot
5616 // end up with two versions of \p U thinking it was one.
5617 auto *Callee = dyn_cast_if_present<Function>(CB->getCalledOperand());
5618 if (!Callee || !Callee->hasLocalLinkage())
5619 return true;
5620 if (!CB->isArgOperand(&U))
5621 return false;
5622 const auto *ArgInstanceInfoAA = A.getAAFor<AAInstanceInfo>(
5624 DepClassTy::OPTIONAL);
5625 if (!ArgInstanceInfoAA ||
5626 !ArgInstanceInfoAA->isAssumedUniqueForAnalysis())
5627 return false;
5628 // If this call base might reach the scope again we might forward the
5629 // argument back here. This is very conservative.
5631 A, *CB, *Scope, *this, /* ExclusionSet */ nullptr,
5632 [Scope](const Function &Fn) { return &Fn != Scope; }))
5633 return false;
5634 return true;
5635 }
5636 return false;
5637 };
5638
5639 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
5640 if (auto *SI = dyn_cast<StoreInst>(OldU.getUser())) {
5641 auto *Ptr = SI->getPointerOperand()->stripPointerCasts();
5642 if ((isa<AllocaInst>(Ptr) || isNoAliasCall(Ptr)) &&
5643 AA::isDynamicallyUnique(A, *this, *Ptr))
5644 return true;
5645 }
5646 return false;
5647 };
5648
5649 if (!A.checkForAllUses(UsePred, *this, V, /* CheckBBLivenessOnly */ true,
5650 DepClassTy::OPTIONAL,
5651 /* IgnoreDroppableUses */ true, EquivalentUseCB))
5652 return indicatePessimisticFixpoint();
5653
5654 return Changed;
5655 }
5656
5657 /// See AbstractState::getAsStr().
5658 const std::string getAsStr(Attributor *A) const override {
5659 return isAssumedUniqueForAnalysis() ? "<unique [fAa]>" : "<unknown>";
5660 }
5661
5662 /// See AbstractAttribute::trackStatistics()
5663 void trackStatistics() const override {}
5664};
5665
5666/// InstanceInfo attribute for floating values.
5667struct AAInstanceInfoFloating : AAInstanceInfoImpl {
5668 AAInstanceInfoFloating(const IRPosition &IRP, Attributor &A)
5669 : AAInstanceInfoImpl(IRP, A) {}
5670};
5671
5672/// NoCapture attribute for function arguments.
5673struct AAInstanceInfoArgument final : AAInstanceInfoFloating {
5674 AAInstanceInfoArgument(const IRPosition &IRP, Attributor &A)
5675 : AAInstanceInfoFloating(IRP, A) {}
5676};
5677
5678/// InstanceInfo attribute for call site arguments.
5679struct AAInstanceInfoCallSiteArgument final : AAInstanceInfoImpl {
5680 AAInstanceInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
5681 : AAInstanceInfoImpl(IRP, A) {}
5682
5683 /// See AbstractAttribute::updateImpl(...).
5684 ChangeStatus updateImpl(Attributor &A) override {
5685 // TODO: Once we have call site specific value information we can provide
5686 // call site specific liveness information and then it makes
5687 // sense to specialize attributes for call sites arguments instead of
5688 // redirecting requests to the callee argument.
5689 Argument *Arg = getAssociatedArgument();
5690 if (!Arg)
5691 return indicatePessimisticFixpoint();
5692 const IRPosition &ArgPos = IRPosition::argument(*Arg);
5693 auto *ArgAA =
5694 A.getAAFor<AAInstanceInfo>(*this, ArgPos, DepClassTy::REQUIRED);
5695 if (!ArgAA)
5696 return indicatePessimisticFixpoint();
5697 return clampStateAndIndicateChange(getState(), ArgAA->getState());
5698 }
5699};
5700
5701/// InstanceInfo attribute for function return value.
5702struct AAInstanceInfoReturned final : AAInstanceInfoImpl {
5703 AAInstanceInfoReturned(const IRPosition &IRP, Attributor &A)
5704 : AAInstanceInfoImpl(IRP, A) {
5705 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5706 }
5707
5708 /// See AbstractAttribute::initialize(...).
5709 void initialize(Attributor &A) override {
5710 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5711 }
5712
5713 /// See AbstractAttribute::updateImpl(...).
5714 ChangeStatus updateImpl(Attributor &A) override {
5715 llvm_unreachable("InstanceInfo is not applicable to function returns!");
5716 }
5717};
5718
5719/// InstanceInfo attribute deduction for a call site return value.
5720struct AAInstanceInfoCallSiteReturned final : AAInstanceInfoFloating {
5721 AAInstanceInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
5722 : AAInstanceInfoFloating(IRP, A) {}
5723};
5724} // namespace
5725
5726/// ----------------------- Variable Capturing ---------------------------------
5728 Attribute::AttrKind ImpliedAttributeKind,
5729 bool IgnoreSubsumingPositions) {
5730 assert(ImpliedAttributeKind == Attribute::NoCapture &&
5731 "Unexpected attribute kind");
5732 Value &V = IRP.getAssociatedValue();
5733 if (!IRP.isArgumentPosition())
5734 return V.use_empty();
5735
5736 // You cannot "capture" null in the default address space.
5737 //
5738 // FIXME: This should use NullPointerIsDefined to account for the function
5739 // attribute.
5740 if (isa<UndefValue>(V) || (isa<ConstantPointerNull>(V) &&
5741 V.getType()->getPointerAddressSpace() == 0)) {
5742 return true;
5743 }
5744
5745 if (A.hasAttr(IRP, {Attribute::NoCapture},
5746 /* IgnoreSubsumingPositions */ true, Attribute::NoCapture))
5747 return true;
5748
5749 if (IRP.getPositionKind() == IRP_CALL_SITE_ARGUMENT)
5750 if (Argument *Arg = IRP.getAssociatedArgument())
5751 if (A.hasAttr(IRPosition::argument(*Arg),
5752 {Attribute::NoCapture, Attribute::ByVal},
5753 /* IgnoreSubsumingPositions */ true)) {
5754 A.manifestAttrs(IRP,
5755 Attribute::get(V.getContext(), Attribute::NoCapture));
5756 return true;
5757 }
5758
5759 if (const Function *F = IRP.getAssociatedFunction()) {
5760 // Check what state the associated function can actually capture.
5762 determineFunctionCaptureCapabilities(IRP, *F, State);
5763 if (State.isKnown(NO_CAPTURE)) {
5764 A.manifestAttrs(IRP,
5765 Attribute::get(V.getContext(), Attribute::NoCapture));
5766 return true;
5767 }
5768 }
5769
5770 return false;
5771}
5772
5773/// Set the NOT_CAPTURED_IN_MEM and NOT_CAPTURED_IN_RET bits in \p Known
5774/// depending on the ability of the function associated with \p IRP to capture
5775/// state in memory and through "returning/throwing", respectively.
5777 const Function &F,
5778 BitIntegerState &State) {
5779 // TODO: Once we have memory behavior attributes we should use them here.
5780
5781 // If we know we cannot communicate or write to memory, we do not care about
5782 // ptr2int anymore.
5783 bool ReadOnly = F.onlyReadsMemory();
5784 bool NoThrow = F.doesNotThrow();
5785 bool IsVoidReturn = F.getReturnType()->isVoidTy();
5786 if (ReadOnly && NoThrow && IsVoidReturn) {
5787 State.addKnownBits(NO_CAPTURE);
5788 return;
5789 }
5790
5791 // A function cannot capture state in memory if it only reads memory, it can
5792 // however return/throw state and the state might be influenced by the
5793 // pointer value, e.g., loading from a returned pointer might reveal a bit.
5794 if (ReadOnly)
5795 State.addKnownBits(NOT_CAPTURED_IN_MEM);
5796
5797 // A function cannot communicate state back if it does not through
5798 // exceptions and doesn not return values.
5799 if (NoThrow && IsVoidReturn)
5800 State.addKnownBits(NOT_CAPTURED_IN_RET);
5801
5802 // Check existing "returned" attributes.
5803 int ArgNo = IRP.getCalleeArgNo();
5804 if (!NoThrow || ArgNo < 0 ||
5805 !F.getAttributes().hasAttrSomewhere(Attribute::Returned))
5806 return;
5807
5808 for (unsigned U = 0, E = F.arg_size(); U < E; ++U)
5809 if (F.hasParamAttribute(U, Attribute::Returned)) {
5810 if (U == unsigned(ArgNo))
5811 State.removeAssumedBits(NOT_CAPTURED_IN_RET);
5812 else if (ReadOnly)
5813 State.addKnownBits(NO_CAPTURE);
5814 else
5815 State.addKnownBits(NOT_CAPTURED_IN_RET);
5816 break;
5817 }
5818}
5819
5820namespace {
5821/// A class to hold the state of for no-capture attributes.
5822struct AANoCaptureImpl : public AANoCapture {
5823 AANoCaptureImpl(const IRPosition &IRP, Attributor &A) : AANoCapture(IRP, A) {}
5824
5825 /// See AbstractAttribute::initialize(...).
5826 void initialize(Attributor &A) override {
5827 bool IsKnown;
5828 assert(!AA::hasAssumedIRAttr<Attribute::NoCapture>(
5829 A, nullptr, getIRPosition(), DepClassTy::NONE, IsKnown));
5830 (void)IsKnown;
5831 }
5832
5833 /// See AbstractAttribute::updateImpl(...).
5834 ChangeStatus updateImpl(Attributor &A) override;
5835
5836 /// see AbstractAttribute::isAssumedNoCaptureMaybeReturned(...).
5837 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
5838 SmallVectorImpl<Attribute> &Attrs) const override {
5839 if (!isAssumedNoCaptureMaybeReturned())
5840 return;
5841
5842 if (isArgumentPosition()) {
5843 if (isAssumedNoCapture())
5844 Attrs.emplace_back(Attribute::get(Ctx, Attribute::NoCapture));
5845 else if (ManifestInternal)
5846 Attrs.emplace_back(Attribute::get(Ctx, "no-capture-maybe-returned"));
5847 }
5848 }
5849
5850 /// See AbstractState::getAsStr().
5851 const std::string getAsStr(Attributor *A) const override {
5852 if (isKnownNoCapture())
5853 return "known not-captured";
5854 if (isAssumedNoCapture())
5855 return "assumed not-captured";
5856 if (isKnownNoCaptureMaybeReturned())
5857 return "known not-captured-maybe-returned";
5858 if (isAssumedNoCaptureMaybeReturned())
5859 return "assumed not-captured-maybe-returned";
5860 return "assumed-captured";
5861 }
5862
5863 /// Check the use \p U and update \p State accordingly. Return true if we
5864 /// should continue to update the state.
5865 bool checkUse(Attributor &A, AANoCapture::StateType &State, const Use &U,
5866 bool &Follow) {
5867 Instruction *UInst = cast<Instruction>(U.getUser());
5868 LLVM_DEBUG(dbgs() << "[AANoCapture] Check use: " << *U.get() << " in "
5869 << *UInst << "\n");
5870
5871 // Deal with ptr2int by following uses.
5872 if (isa<PtrToIntInst>(UInst)) {
5873 LLVM_DEBUG(dbgs() << " - ptr2int assume the worst!\n");
5874 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5875 /* Return */ true);
5876 }
5877
5878 // For stores we already checked if we can follow them, if they make it
5879 // here we give up.
5880 if (isa<StoreInst>(UInst))
5881 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5882 /* Return */ true);
5883
5884 // Explicitly catch return instructions.
5885 if (isa<ReturnInst>(UInst)) {
5886 if (UInst->getFunction() == getAnchorScope())
5887 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
5888 /* Return */ true);
5889 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5890 /* Return */ true);
5891 }
5892
5893 // For now we only use special logic for call sites. However, the tracker
5894 // itself knows about a lot of other non-capturing cases already.
5895 auto *CB = dyn_cast<CallBase>(UInst);
5896 if (!CB || !CB->isArgOperand(&U))
5897 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5898 /* Return */ true);
5899
5900 unsigned ArgNo = CB->getArgOperandNo(&U);
5901 const IRPosition &CSArgPos = IRPosition::callsite_argument(*CB, ArgNo);
5902 // If we have a abstract no-capture attribute for the argument we can use
5903 // it to justify a non-capture attribute here. This allows recursion!
5904 bool IsKnownNoCapture;
5905 const AANoCapture *ArgNoCaptureAA = nullptr;
5906 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
5907 A, this, CSArgPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
5908 &ArgNoCaptureAA);
5909 if (IsAssumedNoCapture)
5910 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
5911 /* Return */ false);
5912 if (ArgNoCaptureAA && ArgNoCaptureAA->isAssumedNoCaptureMaybeReturned()) {
5913 Follow = true;
5914 return isCapturedIn(State, /* Memory */ false, /* Integer */ false,
5915 /* Return */ false);
5916 }
5917
5918 // Lastly, we could not find a reason no-capture can be assumed so we don't.
5919 return isCapturedIn(State, /* Memory */ true, /* Integer */ true,
5920 /* Return */ true);
5921 }
5922
5923 /// Update \p State according to \p CapturedInMem, \p CapturedInInt, and
5924 /// \p CapturedInRet, then return true if we should continue updating the
5925 /// state.
5926 static bool isCapturedIn(AANoCapture::StateType &State, bool CapturedInMem,
5927 bool CapturedInInt, bool CapturedInRet) {
5928 LLVM_DEBUG(dbgs() << " - captures [Mem " << CapturedInMem << "|Int "
5929 << CapturedInInt << "|Ret " << CapturedInRet << "]\n");
5930 if (CapturedInMem)
5931 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_MEM);
5932 if (CapturedInInt)
5933 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_INT);
5934 if (CapturedInRet)
5935 State.removeAssumedBits(AANoCapture::NOT_CAPTURED_IN_RET);
5936 return State.isAssumed(AANoCapture::NO_CAPTURE_MAYBE_RETURNED);
5937 }
5938};
5939
5940ChangeStatus AANoCaptureImpl::updateImpl(Attributor &A) {
5941 const IRPosition &IRP = getIRPosition();
5942 Value *V = isArgumentPosition() ? IRP.getAssociatedArgument()
5943 : &IRP.getAssociatedValue();
5944 if (!V)
5945 return indicatePessimisticFixpoint();
5946
5947 const Function *F =
5948 isArgumentPosition() ? IRP.getAssociatedFunction() : IRP.getAnchorScope();
5949
5950 // TODO: Is the checkForAllUses below useful for constants?
5951 if (!F)
5952 return indicatePessimisticFixpoint();
5953
5955 const IRPosition &FnPos = IRPosition::function(*F);
5956
5957 // Readonly means we cannot capture through memory.
5958 bool IsKnown;
5959 if (AA::isAssumedReadOnly(A, FnPos, *this, IsKnown)) {
5960 T.addKnownBits(NOT_CAPTURED_IN_MEM);
5961 if (IsKnown)
5962 addKnownBits(NOT_CAPTURED_IN_MEM);
5963 }
5964
5965 // Make sure all returned values are different than the underlying value.
5966 // TODO: we could do this in a more sophisticated way inside
5967 // AAReturnedValues, e.g., track all values that escape through returns
5968 // directly somehow.
5969 auto CheckReturnedArgs = [&](bool &UsedAssumedInformation) {
5971 if (!A.getAssumedSimplifiedValues(IRPosition::returned(*F), this, Values,
5973 UsedAssumedInformation))
5974 return false;
5975 bool SeenConstant = false;
5976 for (const AA::ValueAndContext &VAC : Values) {
5977 if (isa<Constant>(VAC.getValue())) {
5978 if (SeenConstant)
5979 return false;
5980 SeenConstant = true;
5981 } else if (!isa<Argument>(VAC.getValue()) ||
5982 VAC.getValue() == getAssociatedArgument())
5983 return false;
5984 }
5985 return true;
5986 };
5987
5988 bool IsKnownNoUnwind;
5989 if (AA::hasAssumedIRAttr<Attribute::NoUnwind>(
5990 A, this, FnPos, DepClassTy::OPTIONAL, IsKnownNoUnwind)) {
5991 bool IsVoidTy = F->getReturnType()->isVoidTy();
5992 bool UsedAssumedInformation = false;
5993 if (IsVoidTy || CheckReturnedArgs(UsedAssumedInformation)) {
5994 T.addKnownBits(NOT_CAPTURED_IN_RET);
5995 if (T.isKnown(NOT_CAPTURED_IN_MEM))
5997 if (IsKnownNoUnwind && (IsVoidTy || !UsedAssumedInformation)) {
5998 addKnownBits(NOT_CAPTURED_IN_RET);
5999 if (isKnown(NOT_CAPTURED_IN_MEM))
6000 return indicateOptimisticFixpoint();
6001 }
6002 }
6003 }
6004
6005 auto IsDereferenceableOrNull = [&](Value *O, const DataLayout &DL) {
6006 const auto *DerefAA = A.getAAFor<AADereferenceable>(
6008 return DerefAA && DerefAA->getAssumedDereferenceableBytes();
6009 };
6010
6011 auto UseCheck = [&](const Use &U, bool &Follow) -> bool {
6012 switch (DetermineUseCaptureKind(U, IsDereferenceableOrNull)) {
6014 return true;
6016 return checkUse(A, T, U, Follow);
6018 Follow = true;
6019 return true;
6020 }
6021 llvm_unreachable("Unexpected use capture kind!");
6022 };
6023
6024 if (!A.checkForAllUses(UseCheck, *this, *V))
6025 return indicatePessimisticFixpoint();
6026
6027 AANoCapture::StateType &S = getState();
6028 auto Assumed = S.getAssumed();
6029 S.intersectAssumedBits(T.getAssumed());
6030 if (!isAssumedNoCaptureMaybeReturned())
6031 return indicatePessimisticFixpoint();
6032 return Assumed == S.getAssumed() ? ChangeStatus::UNCHANGED
6034}
6035
6036/// NoCapture attribute for function arguments.
6037struct AANoCaptureArgument final : AANoCaptureImpl {
6038 AANoCaptureArgument(const IRPosition &IRP, Attributor &A)
6039 : AANoCaptureImpl(IRP, A) {}
6040
6041 /// See AbstractAttribute::trackStatistics()
6042 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nocapture) }
6043};
6044
6045/// NoCapture attribute for call site arguments.
6046struct AANoCaptureCallSiteArgument final : AANoCaptureImpl {
6047 AANoCaptureCallSiteArgument(const IRPosition &IRP, Attributor &A)
6048 : AANoCaptureImpl(IRP, A) {}
6049
6050 /// See AbstractAttribute::updateImpl(...).
6051 ChangeStatus updateImpl(Attributor &A) override {
6052 // TODO: Once we have call site specific value information we can provide
6053 // call site specific liveness information and then it makes
6054 // sense to specialize attributes for call sites arguments instead of
6055 // redirecting requests to the callee argument.
6056 Argument *Arg = getAssociatedArgument();
6057 if (!Arg)
6058 return indicatePessimisticFixpoint();
6059 const IRPosition &ArgPos = IRPosition::argument(*Arg);
6060 bool IsKnownNoCapture;
6061 const AANoCapture *ArgAA = nullptr;
6062 if (AA::hasAssumedIRAttr<Attribute::NoCapture>(
6063 A, this, ArgPos, DepClassTy::REQUIRED, IsKnownNoCapture, false,
6064 &ArgAA))
6065 return ChangeStatus::UNCHANGED;
6066 if (!ArgAA || !ArgAA->isAssumedNoCaptureMaybeReturned())
6067 return indicatePessimisticFixpoint();
6068 return clampStateAndIndicateChange(getState(), ArgAA->getState());
6069 }
6070
6071 /// See AbstractAttribute::trackStatistics()
6072 void trackStatistics() const override{STATS_DECLTRACK_CSARG_ATTR(nocapture)};
6073};
6074
6075/// NoCapture attribute for floating values.
6076struct AANoCaptureFloating final : AANoCaptureImpl {
6077 AANoCaptureFloating(const IRPosition &IRP, Attributor &A)
6078 : AANoCaptureImpl(IRP, A) {}
6079
6080 /// See AbstractAttribute::trackStatistics()
6081 void trackStatistics() const override {
6083 }
6084};
6085
6086/// NoCapture attribute for function return value.
6087struct AANoCaptureReturned final : AANoCaptureImpl {
6088 AANoCaptureReturned(const IRPosition &IRP, Attributor &A)
6089 : AANoCaptureImpl(IRP, A) {
6090 llvm_unreachable("NoCapture is not applicable to function returns!");
6091 }
6092
6093 /// See AbstractAttribute::initialize(...).
6094 void initialize(Attributor &A) override {
6095 llvm_unreachable("NoCapture is not applicable to function returns!");
6096 }
6097
6098 /// See AbstractAttribute::updateImpl(...).
6099 ChangeStatus updateImpl(Attributor &A) override {
6100 llvm_unreachable("NoCapture is not applicable to function returns!");
6101 }
6102
6103 /// See AbstractAttribute::trackStatistics()
6104 void trackStatistics() const override {}
6105};
6106
6107/// NoCapture attribute deduction for a call site return value.
6108struct AANoCaptureCallSiteReturned final : AANoCaptureImpl {
6109 AANoCaptureCallSiteReturned(const IRPosition &IRP, Attributor &A)
6110 : AANoCaptureImpl(IRP, A) {}
6111
6112 /// See AbstractAttribute::initialize(...).
6113 void initialize(Attributor &A) override {
6114 const Function *F = getAnchorScope();
6115 // Check what state the associated function can actually capture.
6116 determineFunctionCaptureCapabilities(getIRPosition(), *F, *this);
6117 }
6118
6119 /// See AbstractAttribute::trackStatistics()
6120 void trackStatistics() const override {
6122 }
6123};
6124} // namespace
6125
6126/// ------------------ Value Simplify Attribute ----------------------------
6127
6128bool ValueSimplifyStateType::unionAssumed(std::optional<Value *> Other) {
6129 // FIXME: Add a typecast support.
6130 SimplifiedAssociatedValue = AA::combineOptionalValuesInAAValueLatice(
6131 SimplifiedAssociatedValue, Other, Ty);
6132 if (SimplifiedAssociatedValue == std::optional<Value *>(nullptr))
6133 return false;
6134
6135 LLVM_DEBUG({
6136 if (SimplifiedAssociatedValue)
6137 dbgs() << "[ValueSimplify] is assumed to be "
6138 << **SimplifiedAssociatedValue << "\n";
6139 else
6140 dbgs() << "[ValueSimplify] is assumed to be <none>\n";
6141 });
6142 return true;
6143}
6144
6145namespace {
6146struct AAValueSimplifyImpl : AAValueSimplify {
6147 AAValueSimplifyImpl(const IRPosition &IRP, Attributor &A)
6148 : AAValueSimplify(IRP, A) {}
6149
6150 /// See AbstractAttribute::initialize(...).
6151 void initialize(Attributor &A) override {
6152 if (getAssociatedValue().getType()->isVoidTy())
6153 indicatePessimisticFixpoint();
6154 if (A.hasSimplificationCallback(getIRPosition()))
6155 indicatePessimisticFixpoint();
6156 }
6157
6158 /// See AbstractAttribute::getAsStr().
6159 const std::string getAsStr(Attributor *A) const override {
6160 LLVM_DEBUG({
6161 dbgs() << "SAV: " << (bool)SimplifiedAssociatedValue << " ";
6162 if (SimplifiedAssociatedValue && *SimplifiedAssociatedValue)
6163 dbgs() << "SAV: " << **SimplifiedAssociatedValue << " ";
6164 });
6165 return isValidState() ? (isAtFixpoint() ? "simplified" : "maybe-simple")
6166 : "not-simple";
6167 }
6168
6169 /// See AbstractAttribute::trackStatistics()
6170 void trackStatistics() const override {}
6171
6172 /// See AAValueSimplify::getAssumedSimplifiedValue()
6173 std::optional<Value *>
6174 getAssumedSimplifiedValue(Attributor &A) const override {
6175 return SimplifiedAssociatedValue;
6176 }
6177
6178 /// Ensure the return value is \p V with type \p Ty, if not possible return
6179 /// nullptr. If \p Check is true we will only verify such an operation would
6180 /// suceed and return a non-nullptr value if that is the case. No IR is
6181 /// generated or modified.
6182 static Value *ensureType(Attributor &A, Value &V, Type &Ty, Instruction *CtxI,
6183 bool Check) {
6184 if (auto *TypedV = AA::getWithType(V, Ty))
6185 return TypedV;
6186 if (CtxI && V.getType()->canLosslesslyBitCastTo(&Ty))
6187 return Check ? &V
6188 : BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
6189 &V, &Ty, "", CtxI->getIterator());
6190 return nullptr;
6191 }
6192
6193 /// Reproduce \p I with type \p Ty or return nullptr if that is not posisble.
6194 /// If \p Check is true we will only verify such an operation would suceed and
6195 /// return a non-nullptr value if that is the case. No IR is generated or
6196 /// modified.
6197 static Value *reproduceInst(Attributor &A,
6198 const AbstractAttribute &QueryingAA,
6199 Instruction &I, Type &Ty, Instruction *CtxI,
6200 bool Check, ValueToValueMapTy &VMap) {
6201 assert(CtxI && "Cannot reproduce an instruction without context!");
6202 if (Check && (I.mayReadFromMemory() ||
6203 !isSafeToSpeculativelyExecute(&I, CtxI, /* DT */ nullptr,
6204 /* TLI */ nullptr)))
6205 return nullptr;
6206 for (Value *Op : I.operands()) {
6207 Value *NewOp = reproduceValue(A, QueryingAA, *Op, Ty, CtxI, Check, VMap);
6208 if (!NewOp) {
6209 assert(Check && "Manifest of new value unexpectedly failed!");
6210 return nullptr;
6211 }
6212 if (!Check)
6213 VMap[Op] = NewOp;
6214 }
6215 if (Check)
6216 return &I;
6217
6218 Instruction *CloneI = I.clone();
6219 // TODO: Try to salvage debug information here.
6220 CloneI->setDebugLoc(DebugLoc());
6221 VMap[&I] = CloneI;
6222 CloneI->insertBefore(CtxI);
6223 RemapInstruction(CloneI, VMap);
6224 return CloneI;
6225 }
6226
6227 /// Reproduce \p V with type \p Ty or return nullptr if that is not posisble.
6228 /// If \p Check is true we will only verify such an operation would suceed and
6229 /// return a non-nullptr value if that is the case. No IR is generated or
6230 /// modified.
6231 static Value *reproduceValue(Attributor &A,
6232 const AbstractAttribute &QueryingAA, Value &V,
6233 Type &Ty, Instruction *CtxI, bool Check,
6234 ValueToValueMapTy &VMap) {
6235 if (const auto &NewV = VMap.lookup(&V))
6236 return NewV;
6237 bool UsedAssumedInformation = false;
6238 std::optional<Value *> SimpleV = A.getAssumedSimplified(
6239 V, QueryingAA, UsedAssumedInformation, AA::Interprocedural);
6240 if (!SimpleV.has_value())
6241 return PoisonValue::get(&Ty);
6242 Value *EffectiveV = &V;
6243 if (*SimpleV)
6244 EffectiveV = *SimpleV;
6245 if (auto *C = dyn_cast<Constant>(EffectiveV))
6246 return C;
6247 if (CtxI && AA::isValidAtPosition(AA::ValueAndContext(*EffectiveV, *CtxI),
6248 A.getInfoCache()))
6249 return ensureType(A, *EffectiveV, Ty, CtxI, Check);
6250 if (auto *I = dyn_cast<Instruction>(EffectiveV))
6251 if (Value *NewV = reproduceInst(A, QueryingAA, *I, Ty, CtxI, Check, VMap))
6252 return ensureType(A, *NewV, Ty, CtxI, Check);
6253 return nullptr;
6254 }
6255
6256 /// Return a value we can use as replacement for the associated one, or
6257 /// nullptr if we don't have one that makes sense.
6258 Value *manifestReplacementValue(Attributor &A, Instruction *CtxI) const {
6259 Value *NewV = SimplifiedAssociatedValue
6260 ? *SimplifiedAssociatedValue
6261 : UndefValue::get(getAssociatedType());
6262 if (NewV && NewV != &getAssociatedValue()) {
6263 ValueToValueMapTy VMap;
6264 // First verify we can reprduce the value with the required type at the
6265 // context location before we actually start modifying the IR.
6266 if (reproduceValue(A, *this, *NewV, *getAssociatedType(), CtxI,
6267 /* CheckOnly */ true, VMap))
6268 return reproduceValue(A, *this, *NewV, *getAssociatedType(), CtxI,
6269 /* CheckOnly */ false, VMap);
6270 }
6271 return nullptr;
6272 }
6273
6274 /// Helper function for querying AAValueSimplify and updating candidate.
6275 /// \param IRP The value position we are trying to unify with SimplifiedValue
6276 bool checkAndUpdate(Attributor &A, const AbstractAttribute &QueryingAA,
6277 const IRPosition &IRP, bool Simplify = true) {
6278 bool UsedAssumedInformation = false;
6279 std::optional<Value *> QueryingValueSimplified = &IRP.getAssociatedValue();
6280 if (Simplify)
6281 QueryingValueSimplified = A.getAssumedSimplified(
6282 IRP, QueryingAA, UsedAssumedInformation, AA::Interprocedural);
6283 return unionAssumed(QueryingValueSimplified);
6284 }
6285
6286 /// Returns a candidate is found or not
6287 template <typename AAType> bool askSimplifiedValueFor(Attributor &A) {
6288 if (!getAssociatedValue().getType()->isIntegerTy())
6289 return false;
6290
6291 // This will also pass the call base context.
6292 const auto *AA =
6293 A.getAAFor<AAType>(*this, getIRPosition(), DepClassTy::NONE);
6294 if (!AA)
6295 return false;
6296
6297 std::optional<Constant *> COpt = AA->getAssumedConstant(A);
6298
6299 if (!COpt) {
6300 SimplifiedAssociatedValue = std::nullopt;
6301 A.recordDependence(*AA, *this, DepClassTy::OPTIONAL);
6302 return true;
6303 }
6304 if (auto *C = *COpt) {
6305 SimplifiedAssociatedValue = C;
6306 A.recordDependence(*AA, *this, DepClassTy::OPTIONAL);
6307 return true;
6308 }
6309 return false;
6310 }
6311
6312 bool askSimplifiedValueForOtherAAs(Attributor &A) {
6313 if (askSimplifiedValueFor<AAValueConstantRange>(A))
6314 return true;
6315 if (askSimplifiedValueFor<AAPotentialConstantValues>(A))
6316 return true;
6317 return false;
6318 }
6319
6320 /// See AbstractAttribute::manifest(...).
6321 ChangeStatus manifest(Attributor &A) override {
6322 ChangeStatus Changed = ChangeStatus::UNCHANGED;
6323 for (auto &U : getAssociatedValue().uses()) {
6324 // Check if we need to adjust the insertion point to make sure the IR is
6325 // valid.
6326 Instruction *IP = dyn_cast<Instruction>(U.getUser());
6327 if (auto *PHI = dyn_cast_or_null<PHINode>(IP))
6328 IP = PHI->getIncomingBlock(U)->getTerminator();
6329 if (auto *NewV = manifestReplacementValue(A, IP)) {
6330 LLVM_DEBUG(dbgs() << "[ValueSimplify] " << getAssociatedValue()
6331 << " -> " << *NewV << " :: " << *this << "\n");
6332 if (A.changeUseAfterManifest(U, *NewV))
6333 Changed = ChangeStatus::CHANGED;
6334 }
6335 }
6336
6337 return Changed | AAValueSimplify::manifest(A);
6338 }
6339
6340 /// See AbstractState::indicatePessimisticFixpoint(...).
6341 ChangeStatus indicatePessimisticFixpoint() override {
6342 SimplifiedAssociatedValue = &getAssociatedValue();
6343 return AAValueSimplify::indicatePessimisticFixpoint();
6344 }
6345};
6346
6347struct AAValueSimplifyArgument final : AAValueSimplifyImpl {
6348 AAValueSimplifyArgument(const IRPosition &IRP, Attributor &A)
6349 : AAValueSimplifyImpl(IRP, A) {}
6350
6351 void initialize(Attributor &A) override {
6352 AAValueSimplifyImpl::initialize(A);
6353 if (A.hasAttr(getIRPosition(),
6354 {Attribute::InAlloca, Attribute::Preallocated,
6355 Attribute::StructRet, Attribute::Nest, Attribute::ByVal},
6356 /* IgnoreSubsumingPositions */ true))
6357 indicatePessimisticFixpoint();
6358 }
6359
6360 /// See AbstractAttribute::updateImpl(...).
6361 ChangeStatus updateImpl(Attributor &A) override {
6362 // Byval is only replacable if it is readonly otherwise we would write into
6363 // the replaced value and not the copy that byval creates implicitly.
6364 Argument *Arg = getAssociatedArgument();
6365 if (Arg->hasByValAttr()) {
6366 // TODO: We probably need to verify synchronization is not an issue, e.g.,
6367 // there is no race by not copying a constant byval.
6368 bool IsKnown;
6369 if (!AA::isAssumedReadOnly(A, getIRPosition(), *this, IsKnown))
6370 return indicatePessimisticFixpoint();
6371 }
6372
6373 auto Before = SimplifiedAssociatedValue;
6374
6375 auto PredForCallSite = [&](AbstractCallSite ACS) {
6376 const IRPosition &ACSArgPos =
6377 IRPosition::callsite_argument(ACS, getCallSiteArgNo());
6378 // Check if a coresponding argument was found or if it is on not
6379 // associated (which can happen for callback calls).
6380 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
6381 return false;
6382
6383 // Simplify the argument operand explicitly and check if the result is
6384 // valid in the current scope. This avoids refering to simplified values
6385 // in other functions, e.g., we don't want to say a an argument in a
6386 // static function is actually an argument in a different function.
6387 bool UsedAssumedInformation = false;
6388 std::optional<Constant *> SimpleArgOp =
6389 A.getAssumedConstant(ACSArgPos, *this, UsedAssumedInformation);
6390 if (!SimpleArgOp)
6391 return true;
6392 if (!*SimpleArgOp)
6393 return false;
6394 if (!AA::isDynamicallyUnique(A, *this, **SimpleArgOp))
6395 return false;
6396 return unionAssumed(*SimpleArgOp);
6397 };
6398
6399 // Generate a answer specific to a call site context.
6400 bool Success;
6401 bool UsedAssumedInformation = false;
6402 if (hasCallBaseContext() &&
6403 getCallBaseContext()->getCalledOperand() == Arg->getParent())
6404 Success = PredForCallSite(
6405 AbstractCallSite(&getCallBaseContext()->getCalledOperandUse()));
6406 else
6407 Success = A.checkForAllCallSites(PredForCallSite, *this, true,
6408 UsedAssumedInformation);
6409
6410 if (!Success)
6411 if (!askSimplifiedValueForOtherAAs(A))
6412 return indicatePessimisticFixpoint();
6413
6414 // If a candidate was found in this update, return CHANGED.
6415 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6416 : ChangeStatus ::CHANGED;
6417 }
6418
6419 /// See AbstractAttribute::trackStatistics()
6420 void trackStatistics() const override {
6421 STATS_DECLTRACK_ARG_ATTR(value_simplify)
6422 }
6423};
6424
6425struct AAValueSimplifyReturned : AAValueSimplifyImpl {
6426 AAValueSimplifyReturned(const IRPosition &IRP, Attributor &A)
6427 : AAValueSimplifyImpl(IRP, A) {}
6428
6429 /// See AAValueSimplify::getAssumedSimplifiedValue()
6430 std::optional<Value *>
6431 getAssumedSimplifiedValue(Attributor &A) const override {
6432 if (!isValidState())
6433 return nullptr;
6434 return SimplifiedAssociatedValue;
6435 }
6436
6437 /// See AbstractAttribute::updateImpl(...).
6438 ChangeStatus updateImpl(Attributor &A) override {
6439 auto Before = SimplifiedAssociatedValue;
6440
6441 auto ReturnInstCB = [&](Instruction &I) {
6442 auto &RI = cast<ReturnInst>(I);
6443 return checkAndUpdate(
6444 A, *this,
6445 IRPosition::value(*RI.getReturnValue(), getCallBaseContext()));
6446 };
6447
6448 bool UsedAssumedInformation = false;
6449 if (!A.checkForAllInstructions(ReturnInstCB, *this, {Instruction::Ret},
6450 UsedAssumedInformation))
6451 if (!askSimplifiedValueForOtherAAs(A))
6452 return indicatePessimisticFixpoint();
6453
6454 // If a candidate was found in this update, return CHANGED.
6455 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6456 : ChangeStatus ::CHANGED;
6457 }
6458
6459 ChangeStatus manifest(Attributor &A) override {
6460 // We queried AAValueSimplify for the returned values so they will be
6461 // replaced if a simplified form was found. Nothing to do here.
6462 return ChangeStatus::UNCHANGED;
6463 }
6464
6465 /// See AbstractAttribute::trackStatistics()
6466 void trackStatistics() const override {
6467 STATS_DECLTRACK_FNRET_ATTR(value_simplify)
6468 }
6469};
6470
6471struct AAValueSimplifyFloating : AAValueSimplifyImpl {
6472 AAValueSimplifyFloating(const IRPosition &IRP, Attributor &A)
6473 : AAValueSimplifyImpl(IRP, A) {}
6474
6475 /// See AbstractAttribute::initialize(...).
6476 void initialize(Attributor &A) override {
6477 AAValueSimplifyImpl::initialize(A);
6478 Value &V = getAnchorValue();
6479
6480 // TODO: add other stuffs
6481 if (isa<Constant>(V))
6482 indicatePessimisticFixpoint();
6483 }
6484
6485 /// See AbstractAttribute::updateImpl(...).
6486 ChangeStatus updateImpl(Attributor &A) override {
6487 auto Before = SimplifiedAssociatedValue;
6488 if (!askSimplifiedValueForOtherAAs(A))
6489 return indicatePessimisticFixpoint();
6490
6491 // If a candidate was found in this update, return CHANGED.
6492 return Before == SimplifiedAssociatedValue ? ChangeStatus::UNCHANGED
6493 : ChangeStatus ::CHANGED;
6494 }
6495
6496 /// See AbstractAttribute::trackStatistics()
6497 void trackStatistics() const override {
6498 STATS_DECLTRACK_FLOATING_ATTR(value_simplify)
6499 }
6500};
6501
6502struct AAValueSimplifyFunction : AAValueSimplifyImpl {
6503 AAValueSimplifyFunction(const IRPosition &IRP, Attributor &A)
6504 : AAValueSimplifyImpl(IRP, A) {}
6505
6506 /// See AbstractAttribute::initialize(...).
6507 void initialize(Attributor &A) override {
6508 SimplifiedAssociatedValue = nullptr;
6509 indicateOptimisticFixpoint();
6510 }
6511 /// See AbstractAttribute::initialize(...).
6512 ChangeStatus updateImpl(Attributor &A) override {
6514 "AAValueSimplify(Function|CallSite)::updateImpl will not be called");
6515 }
6516 /// See AbstractAttribute::trackStatistics()
6517 void trackStatistics() const override {
6518 STATS_DECLTRACK_FN_ATTR(value_simplify)
6519 }
6520};
6521
6522struct AAValueSimplifyCallSite : AAValueSimplifyFunction {
6523 AAValueSimplifyCallSite(const IRPosition &IRP, Attributor &A)
6524 : AAValueSimplifyFunction(IRP, A) {}
6525 /// See AbstractAttribute::trackStatistics()
6526 void trackStatistics() const override {
6527 STATS_DECLTRACK_CS_ATTR(value_simplify)
6528 }
6529};
6530
6531struct AAValueSimplifyCallSiteReturned : AAValueSimplifyImpl {
6532 AAValueSimplifyCallSiteReturned(const IRPosition &IRP, Attributor &A)
6533 : AAValueSimplifyImpl(IRP, A) {}
6534
6535 void initialize(Attributor &A) override {
6536 AAValueSimplifyImpl::initialize(A);
6537 Function *Fn = getAssociatedFunction();
6538 assert(Fn && "Did expect an associted function");
6539 for (Argument &Arg : Fn->args()) {
6540 if (Arg.hasReturnedAttr()) {
6541 auto IRP = IRPosition::callsite_argument(*cast<CallBase>(getCtxI()),
6542 Arg.getArgNo());
6544 checkAndUpdate(A, *this, IRP))
6545 indicateOptimisticFixpoint();
6546 else
6547 indicatePessimisticFixpoint();
6548 return;
6549 }
6550 }
6551 }
6552
6553 /// See AbstractAttribute::updateImpl(...).
6554 ChangeStatus updateImpl(Attributor &A) override {
6555 return indicatePessimisticFixpoint();
6556 }
6557
6558 void trackStatistics() const override {
6559 STATS_DECLTRACK_CSRET_ATTR(value_simplify)
6560 }
6561};
6562
6563struct AAValueSimplifyCallSiteArgument : AAValueSimplifyFloating {
6564 AAValueSimplifyCallSiteArgument(const IRPosition &IRP, Attributor &A)
6565 : AAValueSimplifyFloating(IRP, A) {}
6566
6567 /// See AbstractAttribute::manifest(...).
6568 ChangeStatus manifest(Attributor &A) override {
6569 ChangeStatus Changed = ChangeStatus::UNCHANGED;
6570 // TODO: We should avoid simplification duplication to begin with.
6571 auto *FloatAA = A.lookupAAFor<AAValueSimplify>(
6572 IRPosition::value(getAssociatedValue()), this, DepClassTy::NONE);
6573 if (FloatAA && FloatAA->getState().isValidState())
6574 return Changed;
6575
6576 if (auto *NewV = manifestReplacementValue(A, getCtxI())) {
6577 Use &U = cast<CallBase>(&getAnchorValue())
6578 ->getArgOperandUse(getCallSiteArgNo());
6579 if (A.changeUseAfterManifest(U, *NewV))
6580 Changed = ChangeStatus::CHANGED;
6581 }
6582
6583 return Changed | AAValueSimplify::manifest(A);
6584 }
6585
6586 void trackStatistics() const override {
6587 STATS_DECLTRACK_CSARG_ATTR(value_simplify)
6588 }
6589};
6590} // namespace
6591
6592/// ----------------------- Heap-To-Stack Conversion ---------------------------
6593namespace {
6594struct AAHeapToStackFunction final : public AAHeapToStack {
6595
6596 struct AllocationInfo {
6597 /// The call that allocates the memory.
6598 CallBase *const CB;
6599
6600 /// The library function id for the allocation.
6601 LibFunc LibraryFunctionId = NotLibFunc;
6602
6603 /// The status wrt. a rewrite.
6604 enum {
6605 STACK_DUE_TO_USE,
6606 STACK_DUE_TO_FREE,
6607 INVALID,
6608 } Status = STACK_DUE_TO_USE;
6609
6610 /// Flag to indicate if we encountered a use that might free this allocation
6611 /// but which is not in the deallocation infos.
6612 bool HasPotentiallyFreeingUnknownUses = false;
6613
6614 /// Flag to indicate that we should place the new alloca in the function
6615 /// entry block rather than where the call site (CB) is.
6616 bool MoveAllocaIntoEntry = true;
6617
6618 /// The set of free calls that use this allocation.
6619 SmallSetVector<CallBase *, 1> PotentialFreeCalls{};
6620 };
6621
6622 struct DeallocationInfo {
6623 /// The call that deallocates the memory.
6624 CallBase *const CB;
6625 /// The value freed by the call.
6626 Value *FreedOp;
6627
6628 /// Flag to indicate if we don't know all objects this deallocation might
6629 /// free.
6630 bool MightFreeUnknownObjects = false;
6631
6632 /// The set of allocation calls that are potentially freed.
6633 SmallSetVector<CallBase *, 1> PotentialAllocationCalls{};
6634 };
6635
6636 AAHeapToStackFunction(const IRPosition &IRP, Attributor &A)
6637 : AAHeapToStack(IRP, A) {}
6638
6639 ~AAHeapToStackFunction() {
6640 // Ensure we call the destructor so we release any memory allocated in the
6641 // sets.
6642 for (auto &It : AllocationInfos)
6643 It.second->~AllocationInfo();
6644 for (auto &It : DeallocationInfos)
6645 It.second->~DeallocationInfo();
6646 }
6647
6648 void initialize(Attributor &A) override {
6649 AAHeapToStack::initialize(A);
6650
6651 const Function *F = getAnchorScope();
6652 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6653
6654 auto AllocationIdentifierCB = [&](Instruction &I) {
6655 CallBase *CB = dyn_cast<CallBase>(&I);
6656 if (!CB)
6657 return true;
6658 if (Value *FreedOp = getFreedOperand(CB, TLI)) {
6659 DeallocationInfos[CB] = new (A.Allocator) DeallocationInfo{CB, FreedOp};
6660 return true;
6661 }
6662 // To do heap to stack, we need to know that the allocation itself is
6663 // removable once uses are rewritten, and that we can initialize the
6664 // alloca to the same pattern as the original allocation result.
6665 if (isRemovableAlloc(CB, TLI)) {
6666 auto *I8Ty = Type::getInt8Ty(CB->getParent()->getContext());
6667 if (nullptr != getInitialValueOfAllocation(CB, TLI, I8Ty)) {
6668 AllocationInfo *AI = new (A.Allocator) AllocationInfo{CB};
6669 AllocationInfos[CB] = AI;
6670 if (TLI)
6671 TLI->getLibFunc(*CB, AI->LibraryFunctionId);
6672 }
6673 }
6674 return true;
6675 };
6676
6677 bool UsedAssumedInformation = false;
6678 bool Success = A.checkForAllCallLikeInstructions(
6679 AllocationIdentifierCB, *this, UsedAssumedInformation,
6680 /* CheckBBLivenessOnly */ false,
6681 /* CheckPotentiallyDead */ true);
6682 (void)Success;
6683 assert(Success && "Did not expect the call base visit callback to fail!");
6684
6686 [](const IRPosition &, const AbstractAttribute *,
6687 bool &) -> std::optional<Value *> { return nullptr; };
6688 for (const auto &It : AllocationInfos)
6689 A.registerSimplificationCallback(IRPosition::callsite_returned(*It.first),
6690 SCB);
6691 for (const auto &It : DeallocationInfos)
6692 A.registerSimplificationCallback(IRPosition::callsite_returned(*It.first),
6693 SCB);
6694 }
6695
6696 const std::string getAsStr(Attributor *A) const override {
6697 unsigned NumH2SMallocs = 0, NumInvalidMallocs = 0;
6698 for (const auto &It : AllocationInfos) {
6699 if (It.second->Status == AllocationInfo::INVALID)
6700 ++NumInvalidMallocs;
6701 else
6702 ++NumH2SMallocs;
6703 }
6704 return "[H2S] Mallocs Good/Bad: " + std::to_string(NumH2SMallocs) + "/" +
6705 std::to_string(NumInvalidMallocs);
6706 }
6707
6708 /// See AbstractAttribute::trackStatistics().
6709 void trackStatistics() const override {
6710 STATS_DECL(
6711 MallocCalls, Function,
6712 "Number of malloc/calloc/aligned_alloc calls converted to allocas");
6713 for (const auto &It : AllocationInfos)
6714 if (It.second->Status != AllocationInfo::INVALID)
6715 ++BUILD_STAT_NAME(MallocCalls, Function);
6716 }
6717
6718 bool isAssumedHeapToStack(const CallBase &CB) const override {
6719 if (isValidState())
6720 if (AllocationInfo *AI =
6721 AllocationInfos.lookup(const_cast<CallBase *>(&CB)))
6722 return AI->Status != AllocationInfo::INVALID;
6723 return false;
6724 }
6725
6726 bool isAssumedHeapToStackRemovedFree(CallBase &CB) const override {
6727 if (!isValidState())
6728 return false;
6729
6730 for (const auto &It : AllocationInfos) {
6731 AllocationInfo &AI = *It.second;
6732 if (AI.Status == AllocationInfo::INVALID)
6733 continue;
6734
6735 if (AI.PotentialFreeCalls.count(&CB))
6736 return true;
6737 }
6738
6739 return false;
6740 }
6741
6742 ChangeStatus manifest(Attributor &A) override {
6743 assert(getState().isValidState() &&
6744 "Attempted to manifest an invalid state!");
6745
6746 ChangeStatus HasChanged = ChangeStatus::UNCHANGED;
6747 Function *F = getAnchorScope();
6748 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6749
6750 for (auto &It : AllocationInfos) {
6751 AllocationInfo &AI = *It.second;
6752 if (AI.Status == AllocationInfo::INVALID)
6753 continue;
6754
6755 for (CallBase *FreeCall : AI.PotentialFreeCalls) {
6756 LLVM_DEBUG(dbgs() << "H2S: Removing free call: " << *FreeCall << "\n");
6757 A.deleteAfterManifest(*FreeCall);
6758 HasChanged = ChangeStatus::CHANGED;
6759 }
6760
6761 LLVM_DEBUG(dbgs() << "H2S: Removing malloc-like call: " << *AI.CB
6762 << "\n");
6763
6764 auto Remark = [&](OptimizationRemark OR) {
6765 LibFunc IsAllocShared;
6766 if (TLI->getLibFunc(*AI.CB, IsAllocShared))
6767 if (IsAllocShared == LibFunc___kmpc_alloc_shared)
6768 return OR << "Moving globalized variable to the stack.";
6769 return OR << "Moving memory allocation from the heap to the stack.";
6770 };
6771 if (AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared)
6772 A.emitRemark<OptimizationRemark>(AI.CB, "OMP110", Remark);
6773 else
6774 A.emitRemark<OptimizationRemark>(AI.CB, "HeapToStack", Remark);
6775
6776 const DataLayout &DL = A.getInfoCache().getDL();
6777 Value *Size;
6778 std::optional<APInt> SizeAPI = getSize(A, *this, AI);
6779 if (SizeAPI) {
6780 Size = ConstantInt::get(AI.CB->getContext(), *SizeAPI);
6781 } else {
6782 LLVMContext &Ctx = AI.CB->getContext();
6783 ObjectSizeOpts Opts;
6784 ObjectSizeOffsetEvaluator Eval(DL, TLI, Ctx, Opts);
6785 SizeOffsetValue SizeOffsetPair = Eval.compute(AI.CB);
6786 assert(SizeOffsetPair != ObjectSizeOffsetEvaluator::unknown() &&
6787 cast<ConstantInt>(SizeOffsetPair.Offset)->isZero());
6788 Size = SizeOffsetPair.Size;
6789 }
6790
6791 BasicBlock::iterator IP = AI.MoveAllocaIntoEntry
6792 ? F->getEntryBlock().begin()
6793 : AI.CB->getIterator();
6794
6795 Align Alignment(1);
6796 if (MaybeAlign RetAlign = AI.CB->getRetAlign())
6797 Alignment = std::max(Alignment, *RetAlign);
6798 if (Value *Align = getAllocAlignment(AI.CB, TLI)) {
6799 std::optional<APInt> AlignmentAPI = getAPInt(A, *this, *Align);
6800 assert(AlignmentAPI && AlignmentAPI->getZExtValue() > 0 &&
6801 "Expected an alignment during manifest!");
6802 Alignment =
6803 std::max(Alignment, assumeAligned(AlignmentAPI->getZExtValue()));
6804 }
6805
6806 // TODO: Hoist the alloca towards the function entry.
6807 unsigned AS = DL.getAllocaAddrSpace();
6808 Instruction *Alloca =
6809 new AllocaInst(Type::getInt8Ty(F->getContext()), AS, Size, Alignment,
6810 AI.CB->getName() + ".h2s", IP);
6811
6812 if (Alloca->getType() != AI.CB->getType())
6813 Alloca = BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
6814 Alloca, AI.CB->getType(), "malloc_cast", AI.CB->getIterator());
6815
6816 auto *I8Ty = Type::getInt8Ty(F->getContext());
6817 auto *InitVal = getInitialValueOfAllocation(AI.CB, TLI, I8Ty);
6818 assert(InitVal &&
6819 "Must be able to materialize initial memory state of allocation");
6820
6821 A.changeAfterManifest(IRPosition::inst(*AI.CB), *Alloca);
6822
6823 if (auto *II = dyn_cast<InvokeInst>(AI.CB)) {
6824 auto *NBB = II->getNormalDest();
6825 BranchInst::Create(NBB, AI.CB->getParent());
6826 A.deleteAfterManifest(*AI.CB);
6827 } else {
6828 A.deleteAfterManifest(*AI.CB);
6829 }
6830
6831 // Initialize the alloca with the same value as used by the allocation
6832 // function. We can skip undef as the initial value of an alloc is
6833 // undef, and the memset would simply end up being DSEd.
6834 if (!isa<UndefValue>(InitVal)) {
6835 IRBuilder<> Builder(Alloca->getNextNode());
6836 // TODO: Use alignment above if align!=1
6837 Builder.CreateMemSet(Alloca, InitVal, Size, std::nullopt);
6838 }
6839 HasChanged = ChangeStatus::CHANGED;
6840 }
6841
6842 return HasChanged;
6843 }
6844
6845 std::optional<APInt> getAPInt(Attributor &A, const AbstractAttribute &AA,
6846 Value &V) {
6847 bool UsedAssumedInformation = false;
6848 std::optional<Constant *> SimpleV =
6849 A.getAssumedConstant(V, AA, UsedAssumedInformation);
6850 if (!SimpleV)
6851 return APInt(64, 0);
6852 if (auto *CI = dyn_cast_or_null<ConstantInt>(*SimpleV))
6853 return CI->getValue();
6854 return std::nullopt;
6855 }
6856
6857 std::optional<APInt> getSize(Attributor &A, const AbstractAttribute &AA,
6858 AllocationInfo &AI) {
6859 auto Mapper = [&](const Value *V) -> const Value * {
6860 bool UsedAssumedInformation = false;
6861 if (std::optional<Constant *> SimpleV =
6862 A.getAssumedConstant(*V, AA, UsedAssumedInformation))
6863 if (*SimpleV)
6864 return *SimpleV;
6865 return V;
6866 };
6867
6868 const Function *F = getAnchorScope();
6869 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6870 return getAllocSize(AI.CB, TLI, Mapper);
6871 }
6872
6873 /// Collection of all malloc-like calls in a function with associated
6874 /// information.
6876
6877 /// Collection of all free-like calls in a function with associated
6878 /// information.
6880
6881 ChangeStatus updateImpl(Attributor &A) override;
6882};
6883
6884ChangeStatus AAHeapToStackFunction::updateImpl(Attributor &A) {
6886 const Function *F = getAnchorScope();
6887 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
6888
6889 const auto *LivenessAA =
6890 A.getAAFor<AAIsDead>(*this, IRPosition::function(*F), DepClassTy::NONE);
6891
6893 A.getInfoCache().getMustBeExecutedContextExplorer();
6894
6895 bool StackIsAccessibleByOtherThreads =
6896 A.getInfoCache().stackIsAccessibleByOtherThreads();
6897
6898 LoopInfo *LI =
6899 A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(*F);
6900 std::optional<bool> MayContainIrreducibleControl;
6901 auto IsInLoop = [&](BasicBlock &BB) {
6902 if (&F->getEntryBlock() == &BB)
6903 return false;
6904 if (!MayContainIrreducibleControl.has_value())
6905 MayContainIrreducibleControl = mayContainIrreducibleControl(*F, LI);
6906 if (*MayContainIrreducibleControl)
6907 return true;
6908 if (!LI)
6909 return true;
6910 return LI->getLoopFor(&BB) != nullptr;
6911 };
6912
6913 // Flag to ensure we update our deallocation information at most once per
6914 // updateImpl call and only if we use the free check reasoning.
6915 bool HasUpdatedFrees = false;
6916
6917 auto UpdateFrees = [&]() {
6918 HasUpdatedFrees = true;
6919
6920 for (auto &It : DeallocationInfos) {
6921 DeallocationInfo &DI = *It.second;
6922 // For now we cannot use deallocations that have unknown inputs, skip
6923 // them.
6924 if (DI.MightFreeUnknownObjects)
6925 continue;
6926
6927 // No need to analyze dead calls, ignore them instead.
6928 bool UsedAssumedInformation = false;
6929 if (A.isAssumedDead(*DI.CB, this, LivenessAA, UsedAssumedInformation,
6930 /* CheckBBLivenessOnly */ true))
6931 continue;
6932
6933 // Use the non-optimistic version to get the freed object.
6934 Value *Obj = getUnderlyingObject(DI.FreedOp);
6935 if (!Obj) {
6936 LLVM_DEBUG(dbgs() << "[H2S] Unknown underlying object for free!\n");
6937 DI.MightFreeUnknownObjects = true;
6938 continue;
6939 }
6940
6941 // Free of null and undef can be ignored as no-ops (or UB in the latter
6942 // case).
6943 if (isa<ConstantPointerNull>(Obj) || isa<UndefValue>(Obj))
6944 continue;
6945
6946 CallBase *ObjCB = dyn_cast<CallBase>(Obj);
6947 if (!ObjCB) {
6948 LLVM_DEBUG(dbgs() << "[H2S] Free of a non-call object: " << *Obj
6949 << "\n");
6950 DI.MightFreeUnknownObjects = true;
6951 continue;
6952 }
6953
6954 AllocationInfo *AI = AllocationInfos.lookup(ObjCB);
6955 if (!AI) {
6956 LLVM_DEBUG(dbgs() << "[H2S] Free of a non-allocation object: " << *Obj
6957 << "\n");
6958 DI.MightFreeUnknownObjects = true;
6959 continue;
6960 }
6961
6962 DI.PotentialAllocationCalls.insert(ObjCB);
6963 }
6964 };
6965
6966 auto FreeCheck = [&](AllocationInfo &AI) {
6967 // If the stack is not accessible by other threads, the "must-free" logic
6968 // doesn't apply as the pointer could be shared and needs to be places in
6969 // "shareable" memory.
6970 if (!StackIsAccessibleByOtherThreads) {
6971 bool IsKnownNoSycn;
6972 if (!AA::hasAssumedIRAttr<Attribute::NoSync>(
6973 A, this, getIRPosition(), DepClassTy::OPTIONAL, IsKnownNoSycn)) {
6974 LLVM_DEBUG(
6975 dbgs() << "[H2S] found an escaping use, stack is not accessible by "
6976 "other threads and function is not nosync:\n");
6977 return false;
6978 }
6979 }
6980 if (!HasUpdatedFrees)
6981 UpdateFrees();
6982
6983 // TODO: Allow multi exit functions that have different free calls.
6984 if (AI.PotentialFreeCalls.size() != 1) {
6985 LLVM_DEBUG(dbgs() << "[H2S] did not find one free call but "
6986 << AI.PotentialFreeCalls.size() << "\n");
6987 return false;
6988 }
6989 CallBase *UniqueFree = *AI.PotentialFreeCalls.begin();
6990 DeallocationInfo *DI = DeallocationInfos.lookup(UniqueFree);
6991 if (!DI) {
6992 LLVM_DEBUG(
6993 dbgs() << "[H2S] unique free call was not known as deallocation call "
6994 << *UniqueFree << "\n");
6995 return false;
6996 }
6997 if (DI->MightFreeUnknownObjects) {
6998 LLVM_DEBUG(
6999 dbgs() << "[H2S] unique free call might free unknown allocations\n");
7000 return false;
7001 }
7002 if (DI->PotentialAllocationCalls.empty())
7003 return true;
7004 if (DI->PotentialAllocationCalls.size() > 1) {
7005 LLVM_DEBUG(dbgs() << "[H2S] unique free call might free "
7006 << DI->PotentialAllocationCalls.size()
7007 << " different allocations\n");
7008 return false;
7009 }
7010 if (*DI->PotentialAllocationCalls.begin() != AI.CB) {
7011 LLVM_DEBUG(
7012 dbgs()
7013 << "[H2S] unique free call not known to free this allocation but "
7014 << **DI->PotentialAllocationCalls.begin() << "\n");
7015 return false;
7016 }
7017
7018 // __kmpc_alloc_shared and __kmpc_alloc_free are by construction matched.
7019 if (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared) {
7020 Instruction *CtxI = isa<InvokeInst>(AI.CB) ? AI.CB : AI.CB->getNextNode();
7021 if (!Explorer || !Explorer->findInContextOf(UniqueFree, CtxI)) {
7022 LLVM_DEBUG(dbgs() << "[H2S] unique free call might not be executed "
7023 "with the allocation "
7024 << *UniqueFree << "\n");
7025 return false;
7026 }
7027 }
7028 return true;
7029 };
7030
7031 auto UsesCheck = [&](AllocationInfo &AI) {
7032 bool ValidUsesOnly = true;
7033
7034 auto Pred = [&](const Use &U, bool &Follow) -> bool {
7035 Instruction *UserI = cast<Instruction>(U.getUser());
7036 if (isa<LoadInst>(UserI))
7037 return true;
7038 if (auto *SI = dyn_cast<StoreInst>(UserI)) {
7039 if (SI->getValueOperand() == U.get()) {
7041 << "[H2S] escaping store to memory: " << *UserI << "\n");
7042 ValidUsesOnly = false;
7043 } else {
7044 // A store into the malloc'ed memory is fine.
7045 }
7046 return true;
7047 }
7048 if (auto *CB = dyn_cast<CallBase>(UserI)) {
7049 if (!CB->isArgOperand(&U) || CB->isLifetimeStartOrEnd())
7050 return true;
7051 if (DeallocationInfos.count(CB)) {
7052 AI.PotentialFreeCalls.insert(CB);
7053 return true;
7054 }
7055
7056 unsigned ArgNo = CB->getArgOperandNo(&U);
7057 auto CBIRP = IRPosition::callsite_argument(*CB, ArgNo);
7058
7059 bool IsKnownNoCapture;
7060 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
7061 A, this, CBIRP, DepClassTy::OPTIONAL, IsKnownNoCapture);
7062
7063 // If a call site argument use is nofree, we are fine.
7064 bool IsKnownNoFree;
7065 bool IsAssumedNoFree = AA::hasAssumedIRAttr<Attribute::NoFree>(
7066 A, this, CBIRP, DepClassTy::OPTIONAL, IsKnownNoFree);
7067
7068 if (!IsAssumedNoCapture ||
7069 (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared &&
7070 !IsAssumedNoFree)) {
7071 AI.HasPotentiallyFreeingUnknownUses |= !IsAssumedNoFree;
7072
7073 // Emit a missed remark if this is missed OpenMP globalization.
7074 auto Remark = [&](OptimizationRemarkMissed ORM) {
7075 return ORM
7076 << "Could not move globalized variable to the stack. "
7077 "Variable is potentially captured in call. Mark "
7078 "parameter as `__attribute__((noescape))` to override.";
7079 };
7080
7081 if (ValidUsesOnly &&
7082 AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared)
7083 A.emitRemark<OptimizationRemarkMissed>(CB, "OMP113", Remark);
7084
7085 LLVM_DEBUG(dbgs() << "[H2S] Bad user: " << *UserI << "\n");
7086 ValidUsesOnly = false;
7087 }
7088 return true;
7089 }
7090
7091 if (isa<GetElementPtrInst>(UserI) || isa<BitCastInst>(UserI) ||
7092 isa<PHINode>(UserI) || isa<SelectInst>(UserI)) {
7093 Follow = true;
7094 return true;
7095 }
7096 // Unknown user for which we can not track uses further (in a way that
7097 // makes sense).
7098 LLVM_DEBUG(dbgs() << "[H2S] Unknown user: " << *UserI << "\n");
7099 ValidUsesOnly = false;
7100 return true;
7101 };
7102 if (!A.checkForAllUses(Pred, *this, *AI.CB, /* CheckBBLivenessOnly */ false,
7103 DepClassTy::OPTIONAL, /* IgnoreDroppableUses */ true,
7104 [&](const Use &OldU, const Use &NewU) {
7105 auto *SI = dyn_cast<StoreInst>(OldU.getUser());
7106 return !SI || StackIsAccessibleByOtherThreads ||
7107 AA::isAssumedThreadLocalObject(
7108 A, *SI->getPointerOperand(), *this);
7109 }))
7110 return false;
7111 return ValidUsesOnly;
7112 };
7113
7114 // The actual update starts here. We look at all allocations and depending on
7115 // their status perform the appropriate check(s).
7116 for (auto &It : AllocationInfos) {
7117 AllocationInfo &AI = *It.second;
7118 if (AI.Status == AllocationInfo::INVALID)
7119 continue;
7120
7121 if (Value *Align = getAllocAlignment(AI.CB, TLI)) {
7122 std::optional<APInt> APAlign = getAPInt(A, *this, *Align);
7123 if (!APAlign) {
7124 // Can't generate an alloca which respects the required alignment
7125 // on the allocation.
7126 LLVM_DEBUG(dbgs() << "[H2S] Unknown allocation alignment: " << *AI.CB
7127 << "\n");
7128 AI.Status = AllocationInfo::INVALID;
7129 Changed = ChangeStatus::CHANGED;
7130 continue;
7131 }
7132 if (APAlign->ugt(llvm::Value::MaximumAlignment) ||
7133 !APAlign->isPowerOf2()) {
7134 LLVM_DEBUG(dbgs() << "[H2S] Invalid allocation alignment: " << APAlign
7135 << "\n");
7136 AI.Status = AllocationInfo::INVALID;
7137 Changed = ChangeStatus::CHANGED;
7138 continue;
7139 }
7140 }
7141
7142 std::optional<APInt> Size = getSize(A, *this, AI);
7143 if (AI.LibraryFunctionId != LibFunc___kmpc_alloc_shared &&
7144 MaxHeapToStackSize != -1) {
7145 if (!Size || Size->ugt(MaxHeapToStackSize)) {
7146 LLVM_DEBUG({
7147 if (!Size)
7148 dbgs() << "[H2S] Unknown allocation size: " << *AI.CB << "\n";
7149 else
7150 dbgs() << "[H2S] Allocation size too large: " << *AI.CB << " vs. "
7151 << MaxHeapToStackSize << "\n";
7152 });
7153
7154 AI.Status = AllocationInfo::INVALID;
7155 Changed = ChangeStatus::CHANGED;
7156 continue;
7157 }
7158 }
7159
7160 switch (AI.Status) {
7161 case AllocationInfo::STACK_DUE_TO_USE:
7162 if (UsesCheck(AI))
7163 break;
7164 AI.Status = AllocationInfo::STACK_DUE_TO_FREE;
7165 [[fallthrough]];
7166 case AllocationInfo::STACK_DUE_TO_FREE:
7167 if (FreeCheck(AI))
7168 break;
7169 AI.Status = AllocationInfo::INVALID;
7170 Changed = ChangeStatus::CHANGED;
7171 break;
7172 case AllocationInfo::INVALID:
7173 llvm_unreachable("Invalid allocations should never reach this point!");
7174 };
7175
7176 // Check if we still think we can move it into the entry block. If the
7177 // alloca comes from a converted __kmpc_alloc_shared then we can usually
7178 // ignore the potential compilations associated with loops.
7179 bool IsGlobalizedLocal =
7180 AI.LibraryFunctionId == LibFunc___kmpc_alloc_shared;
7181 if (AI.MoveAllocaIntoEntry &&
7182 (!Size.has_value() ||
7183 (!IsGlobalizedLocal && IsInLoop(*AI.CB->getParent()))))
7184 AI.MoveAllocaIntoEntry = false;
7185 }
7186
7187 return Changed;
7188}
7189} // namespace
7190
7191/// ----------------------- Privatizable Pointers ------------------------------
7192namespace {
7193struct AAPrivatizablePtrImpl : public AAPrivatizablePtr {
7194 AAPrivatizablePtrImpl(const IRPosition &IRP, Attributor &A)
7195 : AAPrivatizablePtr(IRP, A), PrivatizableType(std::nullopt) {}
7196
7197 ChangeStatus indicatePessimisticFixpoint() override {
7198 AAPrivatizablePtr::indicatePessimisticFixpoint();
7199 PrivatizableType = nullptr;
7200 return ChangeStatus::CHANGED;
7201 }
7202
7203 /// Identify the type we can chose for a private copy of the underlying
7204 /// argument. std::nullopt means it is not clear yet, nullptr means there is
7205 /// none.
7206 virtual std::optional<Type *> identifyPrivatizableType(Attributor &A) = 0;
7207
7208 /// Return a privatizable type that encloses both T0 and T1.
7209 /// TODO: This is merely a stub for now as we should manage a mapping as well.
7210 std::optional<Type *> combineTypes(std::optional<Type *> T0,
7211 std::optional<Type *> T1) {
7212 if (!T0)
7213 return T1;
7214 if (!T1)
7215 return T0;
7216 if (T0 == T1)
7217 return T0;
7218 return nullptr;
7219 }
7220
7221 std::optional<Type *> getPrivatizableType() const override {
7222 return PrivatizableType;
7223 }
7224
7225 const std::string getAsStr(Attributor *A) const override {
7226 return isAssumedPrivatizablePtr() ? "[priv]" : "[no-priv]";
7227 }
7228
7229protected:
7230 std::optional<Type *> PrivatizableType;
7231};
7232
7233// TODO: Do this for call site arguments (probably also other values) as well.
7234
7235struct AAPrivatizablePtrArgument final : public AAPrivatizablePtrImpl {
7236 AAPrivatizablePtrArgument(const IRPosition &IRP, Attributor &A)
7237 : AAPrivatizablePtrImpl(IRP, A) {}
7238
7239 /// See AAPrivatizablePtrImpl::identifyPrivatizableType(...)
7240 std::optional<Type *> identifyPrivatizableType(Attributor &A) override {
7241 // If this is a byval argument and we know all the call sites (so we can
7242 // rewrite them), there is no need to check them explicitly.
7243 bool UsedAssumedInformation = false;
7245 A.getAttrs(getIRPosition(), {Attribute::ByVal}, Attrs,
7246 /* IgnoreSubsumingPositions */ true);
7247 if (!Attrs.empty() &&
7248 A.checkForAllCallSites([](AbstractCallSite ACS) { return true; }, *this,
7249 true, UsedAssumedInformation))
7250 return Attrs[0].getValueAsType();
7251
7252 std::optional<Type *> Ty;
7253 unsigned ArgNo = getIRPosition().getCallSiteArgNo();
7254
7255 // Make sure the associated call site argument has the same type at all call
7256 // sites and it is an allocation we know is safe to privatize, for now that
7257 // means we only allow alloca instructions.
7258 // TODO: We can additionally analyze the accesses in the callee to create
7259 // the type from that information instead. That is a little more
7260 // involved and will be done in a follow up patch.
7261 auto CallSiteCheck = [&](AbstractCallSite ACS) {
7262 IRPosition ACSArgPos = IRPosition::callsite_argument(ACS, ArgNo);
7263 // Check if a coresponding argument was found or if it is one not
7264 // associated (which can happen for callback calls).
7265 if (ACSArgPos.getPositionKind() == IRPosition::IRP_INVALID)
7266 return false;
7267
7268 // Check that all call sites agree on a type.
7269 auto *PrivCSArgAA =
7270 A.getAAFor<AAPrivatizablePtr>(*this, ACSArgPos, DepClassTy::REQUIRED);
7271 if (!PrivCSArgAA)
7272 return false;
7273 std::optional<Type *> CSTy = PrivCSArgAA->getPrivatizableType();
7274
7275 LLVM_DEBUG({
7276 dbgs() << "[AAPrivatizablePtr] ACSPos: " << ACSArgPos << ", CSTy: ";
7277 if (CSTy && *CSTy)
7278 (*CSTy)->print(dbgs());
7279 else if (CSTy)
7280 dbgs() << "<nullptr>";
7281 else
7282 dbgs() << "<none>";
7283 });
7284
7285 Ty = combineTypes(Ty, CSTy);
7286
7287 LLVM_DEBUG({
7288 dbgs() << " : New Type: ";
7289 if (Ty && *Ty)
7290 (*Ty)->print(dbgs());
7291 else if (Ty)
7292 dbgs() << "<nullptr>";
7293 else
7294 dbgs() << "<none>";
7295 dbgs() << "\n";
7296 });
7297
7298 return !Ty || *Ty;
7299 };
7300
7301 if (!A.checkForAllCallSites(CallSiteCheck, *this, true,
7302 UsedAssumedInformation))
7303 return nullptr;
7304 return Ty;
7305 }
7306
7307 /// See AbstractAttribute::updateImpl(...).
7308 ChangeStatus updateImpl(Attributor &A) override {
7309 PrivatizableType = identifyPrivatizableType(A);
7310 if (!PrivatizableType)
7311 return ChangeStatus::UNCHANGED;
7312 if (!*PrivatizableType)
7313 return indicatePessimisticFixpoint();
7314
7315 // The dependence is optional so we don't give up once we give up on the
7316 // alignment.
7317 A.getAAFor<AAAlign>(*this, IRPosition::value(getAssociatedValue()),
7318 DepClassTy::OPTIONAL);
7319
7320 // Avoid arguments with padding for now.
7321 if (!A.hasAttr(getIRPosition(), Attribute::ByVal) &&
7322 !isDenselyPacked(*PrivatizableType, A.getInfoCache().getDL())) {
7323 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Padding detected\n");
7324 return indicatePessimisticFixpoint();
7325 }
7326
7327 // Collect the types that will replace the privatizable type in the function
7328 // signature.
7329 SmallVector<Type *, 16> ReplacementTypes;
7330 identifyReplacementTypes(*PrivatizableType, ReplacementTypes);
7331
7332 // Verify callee and caller agree on how the promoted argument would be
7333 // passed.
7334 Function &Fn = *getIRPosition().getAnchorScope();
7335 const auto *TTI =
7336 A.getInfoCache().getAnalysisResultForFunction<TargetIRAnalysis>(Fn);
7337 if (!TTI) {
7338 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Missing TTI for function "
7339 << Fn.getName() << "\n");
7340 return indicatePessimisticFixpoint();
7341 }
7342
7343 auto CallSiteCheck = [&](AbstractCallSite ACS) {
7344 CallBase *CB = ACS.getInstruction();
7345 return TTI->areTypesABICompatible(
7346 CB->getCaller(),
7347 dyn_cast_if_present<Function>(CB->getCalledOperand()),
7348 ReplacementTypes);
7349 };
7350 bool UsedAssumedInformation = false;
7351 if (!A.checkForAllCallSites(CallSiteCheck, *this, true,
7352 UsedAssumedInformation)) {
7353 LLVM_DEBUG(
7354 dbgs() << "[AAPrivatizablePtr] ABI incompatibility detected for "
7355 << Fn.getName() << "\n");
7356 return indicatePessimisticFixpoint();
7357 }
7358
7359 // Register a rewrite of the argument.
7360 Argument *Arg = getAssociatedArgument();
7361 if (!A.isValidFunctionSignatureRewrite(*Arg, ReplacementTypes)) {
7362 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Rewrite not valid\n");
7363 return indicatePessimisticFixpoint();
7364 }
7365
7366 unsigned ArgNo = Arg->getArgNo();
7367
7368 // Helper to check if for the given call site the associated argument is
7369 // passed to a callback where the privatization would be different.
7370 auto IsCompatiblePrivArgOfCallback = [&](CallBase &CB) {
7371 SmallVector<const Use *, 4> CallbackUses;
7372 AbstractCallSite::getCallbackUses(CB, CallbackUses);
7373 for (const Use *U : CallbackUses) {
7374 AbstractCallSite CBACS(U);
7375 assert(CBACS && CBACS.isCallbackCall());
7376 for (Argument &CBArg : CBACS.getCalledFunction()->args()) {
7377 int CBArgNo = CBACS.getCallArgOperandNo(CBArg);
7378
7379 LLVM_DEBUG({
7380 dbgs()
7381 << "[AAPrivatizablePtr] Argument " << *Arg
7382 << "check if can be privatized in the context of its parent ("
7383 << Arg->getParent()->getName()
7384 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7385 "callback ("
7386 << CBArgNo << "@" << CBACS.getCalledFunction()->getName()
7387 << ")\n[AAPrivatizablePtr] " << CBArg << " : "
7388 << CBACS.getCallArgOperand(CBArg) << " vs "
7389 << CB.getArgOperand(ArgNo) << "\n"
7390 << "[AAPrivatizablePtr] " << CBArg << " : "
7391 << CBACS.getCallArgOperandNo(CBArg) << " vs " << ArgNo << "\n";
7392 });
7393
7394 if (CBArgNo != int(ArgNo))
7395 continue;
7396 const auto *CBArgPrivAA = A.getAAFor<AAPrivatizablePtr>(
7397 *this, IRPosition::argument(CBArg), DepClassTy::REQUIRED);
7398 if (CBArgPrivAA && CBArgPrivAA->isValidState()) {
7399 auto CBArgPrivTy = CBArgPrivAA->getPrivatizableType();
7400 if (!CBArgPrivTy)
7401 continue;
7402 if (*CBArgPrivTy == PrivatizableType)
7403 continue;
7404 }
7405
7406 LLVM_DEBUG({
7407 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7408 << " cannot be privatized in the context of its parent ("
7409 << Arg->getParent()->getName()
7410 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7411 "callback ("
7412 << CBArgNo << "@" << CBACS.getCalledFunction()->getName()
7413 << ").\n[AAPrivatizablePtr] for which the argument "
7414 "privatization is not compatible.\n";
7415 });
7416 return false;
7417 }
7418 }
7419 return true;
7420 };
7421
7422 // Helper to check if for the given call site the associated argument is
7423 // passed to a direct call where the privatization would be different.
7424 auto IsCompatiblePrivArgOfDirectCS = [&](AbstractCallSite ACS) {
7425 CallBase *DC = cast<CallBase>(ACS.getInstruction());
7426 int DCArgNo = ACS.getCallArgOperandNo(ArgNo);
7427 assert(DCArgNo >= 0 && unsigned(DCArgNo) < DC->arg_size() &&
7428 "Expected a direct call operand for callback call operand");
7429
7430 Function *DCCallee =
7431 dyn_cast_if_present<Function>(DC->getCalledOperand());
7432 LLVM_DEBUG({
7433 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7434 << " check if be privatized in the context of its parent ("
7435 << Arg->getParent()->getName()
7436 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7437 "direct call of ("
7438 << DCArgNo << "@" << DCCallee->getName() << ").\n";
7439 });
7440
7441 if (unsigned(DCArgNo) < DCCallee->arg_size()) {
7442 const auto *DCArgPrivAA = A.getAAFor<AAPrivatizablePtr>(
7443 *this, IRPosition::argument(*DCCallee->getArg(DCArgNo)),
7444 DepClassTy::REQUIRED);
7445 if (DCArgPrivAA && DCArgPrivAA->isValidState()) {
7446 auto DCArgPrivTy = DCArgPrivAA->getPrivatizableType();
7447 if (!DCArgPrivTy)
7448 return true;
7449 if (*DCArgPrivTy == PrivatizableType)
7450 return true;
7451 }
7452 }
7453
7454 LLVM_DEBUG({
7455 dbgs() << "[AAPrivatizablePtr] Argument " << *Arg
7456 << " cannot be privatized in the context of its parent ("
7457 << Arg->getParent()->getName()
7458 << ")\n[AAPrivatizablePtr] because it is an argument in a "
7459 "direct call of ("
7461 << ").\n[AAPrivatizablePtr] for which the argument "
7462 "privatization is not compatible.\n";
7463 });
7464 return false;
7465 };
7466
7467 // Helper to check if the associated argument is used at the given abstract
7468 // call site in a way that is incompatible with the privatization assumed
7469 // here.
7470 auto IsCompatiblePrivArgOfOtherCallSite = [&](AbstractCallSite ACS) {
7471 if (ACS.isDirectCall())
7472 return IsCompatiblePrivArgOfCallback(*ACS.getInstruction());
7473 if (ACS.isCallbackCall())
7474 return IsCompatiblePrivArgOfDirectCS(ACS);
7475 return false;
7476 };
7477
7478 if (!A.checkForAllCallSites(IsCompatiblePrivArgOfOtherCallSite, *this, true,
7479 UsedAssumedInformation))
7480 return indicatePessimisticFixpoint();
7481
7482 return ChangeStatus::UNCHANGED;
7483 }
7484
7485 /// Given a type to private \p PrivType, collect the constituates (which are
7486 /// used) in \p ReplacementTypes.
7487 static void
7488 identifyReplacementTypes(Type *PrivType,
7489 SmallVectorImpl<Type *> &ReplacementTypes) {
7490 // TODO: For now we expand the privatization type to the fullest which can
7491 // lead to dead arguments that need to be removed later.
7492 assert(PrivType && "Expected privatizable type!");
7493
7494 // Traverse the type, extract constituate types on the outermost level.
7495 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7496 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++)
7497 ReplacementTypes.push_back(PrivStructType->getElementType(u));
7498 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7499 ReplacementTypes.append(PrivArrayType->getNumElements(),
7500 PrivArrayType->getElementType());
7501 } else {
7502 ReplacementTypes.push_back(PrivType);
7503 }
7504 }
7505
7506 /// Initialize \p Base according to the type \p PrivType at position \p IP.
7507 /// The values needed are taken from the arguments of \p F starting at
7508 /// position \p ArgNo.
7509 static void createInitialization(Type *PrivType, Value &Base, Function &F,
7510 unsigned ArgNo, BasicBlock::iterator IP) {
7511 assert(PrivType && "Expected privatizable type!");
7512
7513 IRBuilder<NoFolder> IRB(IP->getParent(), IP);
7514 const DataLayout &DL = F.getDataLayout();
7515
7516 // Traverse the type, build GEPs and stores.
7517 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7518 const StructLayout *PrivStructLayout = DL.getStructLayout(PrivStructType);
7519 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++) {
7520 Value *Ptr =
7521 constructPointer(&Base, PrivStructLayout->getElementOffset(u), IRB);
7522 new StoreInst(F.getArg(ArgNo + u), Ptr, IP);
7523 }
7524 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7525 Type *PointeeTy = PrivArrayType->getElementType();
7526 uint64_t PointeeTySize = DL.getTypeStoreSize(PointeeTy);
7527 for (unsigned u = 0, e = PrivArrayType->getNumElements(); u < e; u++) {
7528 Value *Ptr = constructPointer(&Base, u * PointeeTySize, IRB);
7529 new StoreInst(F.getArg(ArgNo + u), Ptr, IP);
7530 }
7531 } else {
7532 new StoreInst(F.getArg(ArgNo), &Base, IP);
7533 }
7534 }
7535
7536 /// Extract values from \p Base according to the type \p PrivType at the
7537 /// call position \p ACS. The values are appended to \p ReplacementValues.
7538 void createReplacementValues(Align Alignment, Type *PrivType,
7540 SmallVectorImpl<Value *> &ReplacementValues) {
7541 assert(Base && "Expected base value!");
7542 assert(PrivType && "Expected privatizable type!");
7543 Instruction *IP = ACS.getInstruction();
7544
7545 IRBuilder<NoFolder> IRB(IP);
7546 const DataLayout &DL = IP->getDataLayout();
7547
7548 // Traverse the type, build GEPs and loads.
7549 if (auto *PrivStructType = dyn_cast<StructType>(PrivType)) {
7550 const StructLayout *PrivStructLayout = DL.getStructLayout(PrivStructType);
7551 for (unsigned u = 0, e = PrivStructType->getNumElements(); u < e; u++) {
7552 Type *PointeeTy = PrivStructType->getElementType(u);
7553 Value *Ptr =
7554 constructPointer(Base, PrivStructLayout->getElementOffset(u), IRB);
7555 LoadInst *L = new LoadInst(PointeeTy, Ptr, "", IP->getIterator());
7556 L->setAlignment(Alignment);
7557 ReplacementValues.push_back(L);
7558 }
7559 } else if (auto *PrivArrayType = dyn_cast<ArrayType>(PrivType)) {
7560 Type *PointeeTy = PrivArrayType->getElementType();
7561 uint64_t PointeeTySize = DL.getTypeStoreSize(PointeeTy);
7562 for (unsigned u = 0, e = PrivArrayType->getNumElements(); u < e; u++) {
7563 Value *Ptr = constructPointer(Base, u * PointeeTySize, IRB);
7564 LoadInst *L = new LoadInst(PointeeTy, Ptr, "", IP->getIterator());
7565 L->setAlignment(Alignment);
7566 ReplacementValues.push_back(L);
7567 }
7568 } else {
7569 LoadInst *L = new LoadInst(PrivType, Base, "", IP->getIterator());
7570 L->setAlignment(Alignment);
7571 ReplacementValues.push_back(L);
7572 }
7573 }
7574
7575 /// See AbstractAttribute::manifest(...)
7576 ChangeStatus manifest(Attributor &A) override {
7577 if (!PrivatizableType)
7578 return ChangeStatus::UNCHANGED;
7579 assert(*PrivatizableType && "Expected privatizable type!");
7580
7581 // Collect all tail calls in the function as we cannot allow new allocas to
7582 // escape into tail recursion.
7583 // TODO: Be smarter about new allocas escaping into tail calls.
7585 bool UsedAssumedInformation = false;
7586 if (!A.checkForAllInstructions(
7587 [&](Instruction &I) {
7588 CallInst &CI = cast<CallInst>(I);
7589 if (CI.isTailCall())
7590 TailCalls.push_back(&CI);
7591 return true;
7592 },
7593 *this, {Instruction::Call}, UsedAssumedInformation))
7594 return ChangeStatus::UNCHANGED;
7595
7596 Argument *Arg = getAssociatedArgument();
7597 // Query AAAlign attribute for alignment of associated argument to
7598 // determine the best alignment of loads.
7599 const auto *AlignAA =
7600 A.getAAFor<AAAlign>(*this, IRPosition::value(*Arg), DepClassTy::NONE);
7601
7602 // Callback to repair the associated function. A new alloca is placed at the
7603 // beginning and initialized with the values passed through arguments. The
7604 // new alloca replaces the use of the old pointer argument.
7607 Function &ReplacementFn, Function::arg_iterator ArgIt) {
7608 BasicBlock &EntryBB = ReplacementFn.getEntryBlock();
7610 const DataLayout &DL = IP->getDataLayout();
7611 unsigned AS = DL.getAllocaAddrSpace();
7612 Instruction *AI = new AllocaInst(*PrivatizableType, AS,
7613 Arg->getName() + ".priv", IP);
7614 createInitialization(*PrivatizableType, *AI, ReplacementFn,
7615 ArgIt->getArgNo(), IP);
7616
7617 if (AI->getType() != Arg->getType())
7618 AI = BitCastInst::CreatePointerBitCastOrAddrSpaceCast(
7619 AI, Arg->getType(), "", IP);
7620 Arg->replaceAllUsesWith(AI);
7621
7622 for (CallInst *CI : TailCalls)
7623 CI->setTailCall(false);
7624 };
7625
7626 // Callback to repair a call site of the associated function. The elements
7627 // of the privatizable type are loaded prior to the call and passed to the
7628 // new function version.
7631 AbstractCallSite ACS, SmallVectorImpl<Value *> &NewArgOperands) {
7632 // When no alignment is specified for the load instruction,
7633 // natural alignment is assumed.
7634 createReplacementValues(
7635 AlignAA ? AlignAA->getAssumedAlign() : Align(0),
7636 *PrivatizableType, ACS,
7637 ACS.getCallArgOperand(ARI.getReplacedArg().getArgNo()),
7638 NewArgOperands);
7639 };
7640
7641 // Collect the types that will replace the privatizable type in the function
7642 // signature.
7643 SmallVector<Type *, 16> ReplacementTypes;
7644 identifyReplacementTypes(*PrivatizableType, ReplacementTypes);
7645
7646 // Register a rewrite of the argument.
7647 if (A.registerFunctionSignatureRewrite(*Arg, ReplacementTypes,
7648 std::move(FnRepairCB),
7649 std::move(ACSRepairCB)))
7650 return ChangeStatus::CHANGED;
7651 return ChangeStatus::UNCHANGED;
7652 }
7653
7654 /// See AbstractAttribute::trackStatistics()
7655 void trackStatistics() const override {
7656 STATS_DECLTRACK_ARG_ATTR(privatizable_ptr);
7657 }
7658};
7659
7660struct AAPrivatizablePtrFloating : public AAPrivatizablePtrImpl {
7661 AAPrivatizablePtrFloating(const IRPosition &IRP, Attributor &A)
7662 : AAPrivatizablePtrImpl(IRP, A) {}
7663
7664 /// See AbstractAttribute::initialize(...).
7665 void initialize(Attributor &A) override {
7666 // TODO: We can privatize more than arguments.
7667 indicatePessimisticFixpoint();
7668 }
7669
7670 ChangeStatus updateImpl(Attributor &A) override {
7671 llvm_unreachable("AAPrivatizablePtr(Floating|Returned|CallSiteReturned)::"
7672 "updateImpl will not be called");
7673 }
7674
7675 /// See AAPrivatizablePtrImpl::identifyPrivatizableType(...)
7676 std::optional<Type *> identifyPrivatizableType(Attributor &A) override {
7677 Value *Obj = getUnderlyingObject(&getAssociatedValue());
7678 if (!Obj) {
7679 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] No underlying object found!\n");
7680 return nullptr;
7681 }
7682
7683 if (auto *AI = dyn_cast<AllocaInst>(Obj))
7684 if (auto *CI = dyn_cast<ConstantInt>(AI->getArraySize()))
7685 if (CI->isOne())
7686 return AI->getAllocatedType();
7687 if (auto *Arg = dyn_cast<Argument>(Obj)) {
7688 auto *PrivArgAA = A.getAAFor<AAPrivatizablePtr>(
7689 *this, IRPosition::argument(*Arg), DepClassTy::REQUIRED);
7690 if (PrivArgAA && PrivArgAA->isAssumedPrivatizablePtr())
7691 return PrivArgAA->getPrivatizableType();
7692 }
7693
7694 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] Underlying object neither valid "
7695 "alloca nor privatizable argument: "
7696 << *Obj << "!\n");
7697 return nullptr;
7698 }
7699
7700 /// See AbstractAttribute::trackStatistics()
7701 void trackStatistics() const override {
7702 STATS_DECLTRACK_FLOATING_ATTR(privatizable_ptr);
7703 }
7704};
7705
7706struct AAPrivatizablePtrCallSiteArgument final
7707 : public AAPrivatizablePtrFloating {
7708 AAPrivatizablePtrCallSiteArgument(const IRPosition &IRP, Attributor &A)
7709 : AAPrivatizablePtrFloating(IRP, A) {}
7710
7711 /// See AbstractAttribute::initialize(...).
7712 void initialize(Attributor &A) override {
7713 if (A.hasAttr(getIRPosition(), Attribute::ByVal))
7714 indicateOptimisticFixpoint();
7715 }
7716
7717 /// See AbstractAttribute::updateImpl(...).
7718 ChangeStatus updateImpl(Attributor &A) override {
7719 PrivatizableType = identifyPrivatizableType(A);
7720 if (!PrivatizableType)
7721 return ChangeStatus::UNCHANGED;
7722 if (!*PrivatizableType)
7723 return indicatePessimisticFixpoint();
7724
7725 const IRPosition &IRP = getIRPosition();
7726 bool IsKnownNoCapture;
7727 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
7728 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoCapture);
7729 if (!IsAssumedNoCapture) {
7730 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer might be captured!\n");
7731 return indicatePessimisticFixpoint();
7732 }
7733
7734 bool IsKnownNoAlias;
7735 if (!AA::hasAssumedIRAttr<Attribute::NoAlias>(
7736 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoAlias)) {
7737 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer might alias!\n");
7738 return indicatePessimisticFixpoint();
7739 }
7740
7741 bool IsKnown;
7742 if (!AA::isAssumedReadOnly(A, IRP, *this, IsKnown)) {
7743 LLVM_DEBUG(dbgs() << "[AAPrivatizablePtr] pointer is written!\n");
7744 return indicatePessimisticFixpoint();
7745 }
7746
7747 return ChangeStatus::UNCHANGED;
7748 }
7749
7750 /// See AbstractAttribute::trackStatistics()
7751 void trackStatistics() const override {
7752 STATS_DECLTRACK_CSARG_ATTR(privatizable_ptr);
7753 }
7754};
7755
7756struct AAPrivatizablePtrCallSiteReturned final
7757 : public AAPrivatizablePtrFloating {
7758 AAPrivatizablePtrCallSiteReturned(const IRPosition &IRP, Attributor &A)
7759 : AAPrivatizablePtrFloating(IRP, A) {}
7760
7761 /// See AbstractAttribute::initialize(...).
7762 void initialize(Attributor &A) override {
7763 // TODO: We can privatize more than arguments.
7764 indicatePessimisticFixpoint();
7765 }
7766
7767 /// See AbstractAttribute::trackStatistics()
7768 void trackStatistics() const override {
7769 STATS_DECLTRACK_CSRET_ATTR(privatizable_ptr);
7770 }
7771};
7772
7773struct AAPrivatizablePtrReturned final : public AAPrivatizablePtrFloating {
7774 AAPrivatizablePtrReturned(const IRPosition &IRP, Attributor &A)
7775 : AAPrivatizablePtrFloating(IRP, A) {}
7776
7777 /// See AbstractAttribute::initialize(...).
7778 void initialize(Attributor &A) override {
7779 // TODO: We can privatize more than arguments.
7780 indicatePessimisticFixpoint();
7781 }
7782
7783 /// See AbstractAttribute::trackStatistics()
7784 void trackStatistics() const override {
7785 STATS_DECLTRACK_FNRET_ATTR(privatizable_ptr);
7786 }
7787};
7788} // namespace
7789
7790/// -------------------- Memory Behavior Attributes ----------------------------
7791/// Includes read-none, read-only, and write-only.
7792/// ----------------------------------------------------------------------------
7793namespace {
7794struct AAMemoryBehaviorImpl : public AAMemoryBehavior {
7795 AAMemoryBehaviorImpl(const IRPosition &IRP, Attributor &A)
7796 : AAMemoryBehavior(IRP, A) {}
7797
7798 /// See AbstractAttribute::initialize(...).
7799 void initialize(Attributor &A) override {
7800 intersectAssumedBits(BEST_STATE);
7801 getKnownStateFromValue(A, getIRPosition(), getState());
7802 AAMemoryBehavior::initialize(A);
7803 }
7804
7805 /// Return the memory behavior information encoded in the IR for \p IRP.
7806 static void getKnownStateFromValue(Attributor &A, const IRPosition &IRP,
7807 BitIntegerState &State,
7808 bool IgnoreSubsumingPositions = false) {
7810 A.getAttrs(IRP, AttrKinds, Attrs, IgnoreSubsumingPositions);
7811 for (const Attribute &Attr : Attrs) {
7812 switch (Attr.getKindAsEnum()) {
7813 case Attribute::ReadNone:
7814 State.addKnownBits(NO_ACCESSES);
7815 break;
7816 case Attribute::ReadOnly:
7817 State.addKnownBits(NO_WRITES);
7818 break;
7819 case Attribute::WriteOnly:
7820 State.addKnownBits(NO_READS);
7821 break;
7822 default:
7823 llvm_unreachable("Unexpected attribute!");
7824 }
7825 }
7826
7827 if (auto *I = dyn_cast<Instruction>(&IRP.getAnchorValue())) {
7828 if (!I->mayReadFromMemory())
7829 State.addKnownBits(NO_READS);
7830 if (!I->mayWriteToMemory())
7831 State.addKnownBits(NO_WRITES);
7832 }
7833 }
7834
7835 /// See AbstractAttribute::getDeducedAttributes(...).
7836 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
7837 SmallVectorImpl<Attribute> &Attrs) const override {
7838 assert(Attrs.size() == 0);
7839 if (isAssumedReadNone())
7840 Attrs.push_back(Attribute::get(Ctx, Attribute::ReadNone));
7841 else if (isAssumedReadOnly())
7842 Attrs.push_back(Attribute::get(Ctx, Attribute::ReadOnly));
7843 else if (isAssumedWriteOnly())
7844 Attrs.push_back(Attribute::get(Ctx, Attribute::WriteOnly));
7845 assert(Attrs.size() <= 1);
7846 }
7847
7848 /// See AbstractAttribute::manifest(...).
7849 ChangeStatus manifest(Attributor &A) override {
7850 const IRPosition &IRP = getIRPosition();
7851
7852 if (A.hasAttr(IRP, Attribute::ReadNone,
7853 /* IgnoreSubsumingPositions */ true))
7854 return ChangeStatus::UNCHANGED;
7855
7856 // Check if we would improve the existing attributes first.
7857 SmallVector<Attribute, 4> DeducedAttrs;
7858 getDeducedAttributes(A, IRP.getAnchorValue().getContext(), DeducedAttrs);
7859 if (llvm::all_of(DeducedAttrs, [&](const Attribute &Attr) {
7860 return A.hasAttr(IRP, Attr.getKindAsEnum(),
7861 /* IgnoreSubsumingPositions */ true);
7862 }))
7863 return ChangeStatus::UNCHANGED;
7864
7865 // Clear existing attributes.
7866 A.removeAttrs(IRP, AttrKinds);
7867 // Clear conflicting writable attribute.
7868 if (isAssumedReadOnly())
7869 A.removeAttrs(IRP, Attribute::Writable);
7870
7871 // Use the generic manifest method.
7872 return IRAttribute::manifest(A);
7873 }
7874
7875 /// See AbstractState::getAsStr().
7876 const std::string getAsStr(Attributor *A) const override {
7877 if (isAssumedReadNone())
7878 return "readnone";
7879 if (isAssumedReadOnly())
7880 return "readonly";
7881 if (isAssumedWriteOnly())
7882 return "writeonly";
7883 return "may-read/write";
7884 }
7885
7886 /// The set of IR attributes AAMemoryBehavior deals with.
7887 static const Attribute::AttrKind AttrKinds[3];
7888};
7889
7890const Attribute::AttrKind AAMemoryBehaviorImpl::AttrKinds[] = {
7891 Attribute::ReadNone, Attribute::ReadOnly, Attribute::WriteOnly};
7892
7893/// Memory behavior attribute for a floating value.
7894struct AAMemoryBehaviorFloating : AAMemoryBehaviorImpl {
7895 AAMemoryBehaviorFloating(const IRPosition &IRP, Attributor &A)
7896 : AAMemoryBehaviorImpl(IRP, A) {}
7897
7898 /// See AbstractAttribute::updateImpl(...).
7899 ChangeStatus updateImpl(Attributor &A) override;
7900
7901 /// See AbstractAttribute::trackStatistics()
7902 void trackStatistics() const override {
7903 if (isAssumedReadNone())
7905 else if (isAssumedReadOnly())
7907 else if (isAssumedWriteOnly())
7909 }
7910
7911private:
7912 /// Return true if users of \p UserI might access the underlying
7913 /// variable/location described by \p U and should therefore be analyzed.
7914 bool followUsersOfUseIn(Attributor &A, const Use &U,
7915 const Instruction *UserI);
7916
7917 /// Update the state according to the effect of use \p U in \p UserI.
7918 void analyzeUseIn(Attributor &A, const Use &U, const Instruction *UserI);
7919};
7920
7921/// Memory behavior attribute for function argument.
7922struct AAMemoryBehaviorArgument : AAMemoryBehaviorFloating {
7923 AAMemoryBehaviorArgument(const IRPosition &IRP, Attributor &A)
7924 : AAMemoryBehaviorFloating(IRP, A) {}
7925
7926 /// See AbstractAttribute::initialize(...).
7927 void initialize(Attributor &A) override {
7928 intersectAssumedBits(BEST_STATE);
7929 const IRPosition &IRP = getIRPosition();
7930 // TODO: Make IgnoreSubsumingPositions a property of an IRAttribute so we
7931 // can query it when we use has/getAttr. That would allow us to reuse the
7932 // initialize of the base class here.
7933 bool HasByVal = A.hasAttr(IRP, {Attribute::ByVal},
7934 /* IgnoreSubsumingPositions */ true);
7935 getKnownStateFromValue(A, IRP, getState(),
7936 /* IgnoreSubsumingPositions */ HasByVal);
7937 }
7938
7939 ChangeStatus manifest(Attributor &A) override {
7940 // TODO: Pointer arguments are not supported on vectors of pointers yet.
7941 if (!getAssociatedValue().getType()->isPointerTy())
7942 return ChangeStatus::UNCHANGED;
7943
7944 // TODO: From readattrs.ll: "inalloca parameters are always
7945 // considered written"
7946 if (A.hasAttr(getIRPosition(),
7947 {Attribute::InAlloca, Attribute::Preallocated})) {
7948 removeKnownBits(NO_WRITES);
7949 removeAssumedBits(NO_WRITES);
7950 }
7951 A.removeAttrs(getIRPosition(), AttrKinds);
7952 return AAMemoryBehaviorFloating::manifest(A);
7953 }
7954
7955 /// See AbstractAttribute::trackStatistics()
7956 void trackStatistics() const override {
7957 if (isAssumedReadNone())
7958 STATS_DECLTRACK_ARG_ATTR(readnone)
7959 else if (isAssumedReadOnly())
7960 STATS_DECLTRACK_ARG_ATTR(readonly)
7961 else if (isAssumedWriteOnly())
7962 STATS_DECLTRACK_ARG_ATTR(writeonly)
7963 }
7964};
7965
7966struct AAMemoryBehaviorCallSiteArgument final : AAMemoryBehaviorArgument {
7967 AAMemoryBehaviorCallSiteArgument(const IRPosition &IRP, Attributor &A)
7968 : AAMemoryBehaviorArgument(IRP, A) {}
7969
7970 /// See AbstractAttribute::initialize(...).
7971 void initialize(Attributor &A) override {
7972 // If we don't have an associated attribute this is either a variadic call
7973 // or an indirect call, either way, nothing to do here.
7974 Argument *Arg = getAssociatedArgument();
7975 if (!Arg) {
7976 indicatePessimisticFixpoint();
7977 return;
7978 }
7979 if (Arg->hasByValAttr()) {
7980 addKnownBits(NO_WRITES);
7981 removeKnownBits(NO_READS);
7982 removeAssumedBits(NO_READS);
7983 }
7984 AAMemoryBehaviorArgument::initialize(A);
7985 if (getAssociatedFunction()->isDeclaration())
7986 indicatePessimisticFixpoint();
7987 }
7988
7989 /// See AbstractAttribute::updateImpl(...).
7990 ChangeStatus updateImpl(Attributor &A) override {
7991 // TODO: Once we have call site specific value information we can provide
7992 // call site specific liveness liveness information and then it makes
7993 // sense to specialize attributes for call sites arguments instead of
7994 // redirecting requests to the callee argument.
7995 Argument *Arg = getAssociatedArgument();
7996 const IRPosition &ArgPos = IRPosition::argument(*Arg);
7997 auto *ArgAA =
7998 A.getAAFor<AAMemoryBehavior>(*this, ArgPos, DepClassTy::REQUIRED);
7999 if (!ArgAA)
8000 return indicatePessimisticFixpoint();
8001 return clampStateAndIndicateChange(getState(), ArgAA->getState());
8002 }
8003
8004 /// See AbstractAttribute::trackStatistics()
8005 void trackStatistics() const override {
8006 if (isAssumedReadNone())
8008 else if (isAssumedReadOnly())
8010 else if (isAssumedWriteOnly())
8012 }
8013};
8014
8015/// Memory behavior attribute for a call site return position.
8016struct AAMemoryBehaviorCallSiteReturned final : AAMemoryBehaviorFloating {
8017 AAMemoryBehaviorCallSiteReturned(const IRPosition &IRP, Attributor &A)
8018 : AAMemoryBehaviorFloating(IRP, A) {}
8019
8020 /// See AbstractAttribute::initialize(...).
8021 void initialize(Attributor &A) override {
8022 AAMemoryBehaviorImpl::initialize(A);
8023 }
8024 /// See AbstractAttribute::manifest(...).
8025 ChangeStatus manifest(Attributor &A) override {
8026 // We do not annotate returned values.
8027 return ChangeStatus::UNCHANGED;
8028 }
8029
8030 /// See AbstractAttribute::trackStatistics()
8031 void trackStatistics() const override {}
8032};
8033
8034/// An AA to represent the memory behavior function attributes.
8035struct AAMemoryBehaviorFunction final : public AAMemoryBehaviorImpl {
8036 AAMemoryBehaviorFunction(const IRPosition &IRP, Attributor &A)
8037 : AAMemoryBehaviorImpl(IRP, A) {}
8038
8039 /// See AbstractAttribute::updateImpl(Attributor &A).
8040 ChangeStatus updateImpl(Attributor &A) override;
8041
8042 /// See AbstractAttribute::manifest(...).
8043 ChangeStatus manifest(Attributor &A) override {
8044 // TODO: It would be better to merge this with AAMemoryLocation, so that
8045 // we could determine read/write per location. This would also have the
8046 // benefit of only one place trying to manifest the memory attribute.
8047 Function &F = cast<Function>(getAnchorValue());
8049 if (isAssumedReadNone())
8050 ME = MemoryEffects::none();
8051 else if (isAssumedReadOnly())
8053 else if (isAssumedWriteOnly())
8055
8056 A.removeAttrs(getIRPosition(), AttrKinds);
8057 // Clear conflicting writable attribute.
8058 if (ME.onlyReadsMemory())
8059 for (Argument &Arg : F.args())
8060 A.removeAttrs(IRPosition::argument(Arg), Attribute::Writable);
8061 return A.manifestAttrs(getIRPosition(),
8062 Attribute::getWithMemoryEffects(F.getContext(), ME));
8063 }
8064
8065 /// See AbstractAttribute::trackStatistics()
8066 void trackStatistics() const override {
8067 if (isAssumedReadNone())
8068 STATS_DECLTRACK_FN_ATTR(readnone)
8069 else if (isAssumedReadOnly())
8070 STATS_DECLTRACK_FN_ATTR(readonly)
8071 else if (isAssumedWriteOnly())
8072 STATS_DECLTRACK_FN_ATTR(writeonly)
8073 }
8074};
8075
8076/// AAMemoryBehavior attribute for call sites.
8077struct AAMemoryBehaviorCallSite final
8078 : AACalleeToCallSite<AAMemoryBehavior, AAMemoryBehaviorImpl> {
8079 AAMemoryBehaviorCallSite(const IRPosition &IRP, Attributor &A)
8080 : AACalleeToCallSite<AAMemoryBehavior, AAMemoryBehaviorImpl>(IRP, A) {}
8081
8082 /// See AbstractAttribute::manifest(...).
8083 ChangeStatus manifest(Attributor &A) override {
8084 // TODO: Deduplicate this with AAMemoryBehaviorFunction.
8085 CallBase &CB = cast<CallBase>(getAnchorValue());
8087 if (isAssumedReadNone())
8088 ME = MemoryEffects::none();
8089 else if (isAssumedReadOnly())
8091 else if (isAssumedWriteOnly())
8093
8094 A.removeAttrs(getIRPosition(), AttrKinds);
8095 // Clear conflicting writable attribute.
8096 if (ME.onlyReadsMemory())
8097 for (Use &U : CB.args())
8098 A.removeAttrs(IRPosition::callsite_argument(CB, U.getOperandNo()),
8099 Attribute::Writable);
8100 return A.manifestAttrs(
8101 getIRPosition(), Attribute::getWithMemoryEffects(CB.getContext(), ME));
8102 }
8103
8104 /// See AbstractAttribute::trackStatistics()
8105 void trackStatistics() const override {
8106 if (isAssumedReadNone())
8107 STATS_DECLTRACK_CS_ATTR(readnone)
8108 else if (isAssumedReadOnly())
8109 STATS_DECLTRACK_CS_ATTR(readonly)
8110 else if (isAssumedWriteOnly())
8111 STATS_DECLTRACK_CS_ATTR(writeonly)
8112 }
8113};
8114
8115ChangeStatus AAMemoryBehaviorFunction::updateImpl(Attributor &A) {
8116
8117 // The current assumed state used to determine a change.
8118 auto AssumedState = getAssumed();
8119
8120 auto CheckRWInst = [&](Instruction &I) {
8121 // If the instruction has an own memory behavior state, use it to restrict
8122 // the local state. No further analysis is required as the other memory
8123 // state is as optimistic as it gets.
8124 if (const auto *CB = dyn_cast<CallBase>(&I)) {
8125 const auto *MemBehaviorAA = A.getAAFor<AAMemoryBehavior>(
8127 if (MemBehaviorAA) {
8128 intersectAssumedBits(MemBehaviorAA->getAssumed());
8129 return !isAtFixpoint();
8130 }
8131 }
8132
8133 // Remove access kind modifiers if necessary.
8134 if (I.mayReadFromMemory())
8135 removeAssumedBits(NO_READS);
8136 if (I.mayWriteToMemory())
8137 removeAssumedBits(NO_WRITES);
8138 return !isAtFixpoint();
8139 };
8140
8141 bool UsedAssumedInformation = false;
8142 if (!A.checkForAllReadWriteInstructions(CheckRWInst, *this,
8143 UsedAssumedInformation))
8144 return indicatePessimisticFixpoint();
8145
8146 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8148}
8149
8150ChangeStatus AAMemoryBehaviorFloating::updateImpl(Attributor &A) {
8151
8152 const IRPosition &IRP = getIRPosition();
8153 const IRPosition &FnPos = IRPosition::function_scope(IRP);
8154 AAMemoryBehavior::StateType &S = getState();
8155
8156 // First, check the function scope. We take the known information and we avoid
8157 // work if the assumed information implies the current assumed information for
8158 // this attribute. This is a valid for all but byval arguments.
8159 Argument *Arg = IRP.getAssociatedArgument();
8160 AAMemoryBehavior::base_t FnMemAssumedState =
8162 if (!Arg || !Arg->hasByValAttr()) {
8163 const auto *FnMemAA =
8164 A.getAAFor<AAMemoryBehavior>(*this, FnPos, DepClassTy::OPTIONAL);
8165 if (FnMemAA) {
8166 FnMemAssumedState = FnMemAA->getAssumed();
8167 S.addKnownBits(FnMemAA->getKnown());
8168 if ((S.getAssumed() & FnMemAA->getAssumed()) == S.getAssumed())
8170 }
8171 }
8172
8173 // The current assumed state used to determine a change.
8174 auto AssumedState = S.getAssumed();
8175
8176 // Make sure the value is not captured (except through "return"), if
8177 // it is, any information derived would be irrelevant anyway as we cannot
8178 // check the potential aliases introduced by the capture. However, no need
8179 // to fall back to anythign less optimistic than the function state.
8180 bool IsKnownNoCapture;
8181 const AANoCapture *ArgNoCaptureAA = nullptr;
8182 bool IsAssumedNoCapture = AA::hasAssumedIRAttr<Attribute::NoCapture>(
8183 A, this, IRP, DepClassTy::OPTIONAL, IsKnownNoCapture, false,
8184 &ArgNoCaptureAA);
8185
8186 if (!IsAssumedNoCapture &&
8187 (!ArgNoCaptureAA || !ArgNoCaptureAA->isAssumedNoCaptureMaybeReturned())) {
8188 S.intersectAssumedBits(FnMemAssumedState);
8189 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8191 }
8192
8193 // Visit and expand uses until all are analyzed or a fixpoint is reached.
8194 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
8195 Instruction *UserI = cast<Instruction>(U.getUser());
8196 LLVM_DEBUG(dbgs() << "[AAMemoryBehavior] Use: " << *U << " in " << *UserI
8197 << " \n");
8198
8199 // Droppable users, e.g., llvm::assume does not actually perform any action.
8200 if (UserI->isDroppable())
8201 return true;
8202
8203 // Check if the users of UserI should also be visited.
8204 Follow = followUsersOfUseIn(A, U, UserI);
8205
8206 // If UserI might touch memory we analyze the use in detail.
8207 if (UserI->mayReadOrWriteMemory())
8208 analyzeUseIn(A, U, UserI);
8209
8210 return !isAtFixpoint();
8211 };
8212
8213 if (!A.checkForAllUses(UsePred, *this, getAssociatedValue()))
8214 return indicatePessimisticFixpoint();
8215
8216 return (AssumedState != getAssumed()) ? ChangeStatus::CHANGED
8218}
8219
8220bool AAMemoryBehaviorFloating::followUsersOfUseIn(Attributor &A, const Use &U,
8221 const Instruction *UserI) {
8222 // The loaded value is unrelated to the pointer argument, no need to
8223 // follow the users of the load.
8224 if (isa<LoadInst>(UserI) || isa<ReturnInst>(UserI))
8225 return false;
8226
8227 // By default we follow all uses assuming UserI might leak information on U,
8228 // we have special handling for call sites operands though.
8229 const auto *CB = dyn_cast<CallBase>(UserI);
8230 if (!CB || !CB->isArgOperand(&U))
8231 return true;
8232
8233 // If the use is a call argument known not to be captured, the users of
8234 // the call do not need to be visited because they have to be unrelated to
8235 // the input. Note that this check is not trivial even though we disallow
8236 // general capturing of the underlying argument. The reason is that the
8237 // call might the argument "through return", which we allow and for which we
8238 // need to check call users.
8239 if (U.get()->getType()->isPointerTy()) {
8240 unsigned ArgNo = CB->getArgOperandNo(&U);
8241 bool IsKnownNoCapture;
8242 return !AA::hasAssumedIRAttr<Attribute::NoCapture>(
8243 A, this, IRPosition::callsite_argument(*CB, ArgNo),
8244 DepClassTy::OPTIONAL, IsKnownNoCapture);
8245 }
8246
8247 return true;
8248}
8249
8250void AAMemoryBehaviorFloating::analyzeUseIn(Attributor &A, const Use &U,
8251 const Instruction *UserI) {
8252 assert(UserI->mayReadOrWriteMemory());
8253
8254 switch (UserI->getOpcode()) {
8255 default:
8256 // TODO: Handle all atomics and other side-effect operations we know of.
8257 break;
8258 case Instruction::Load:
8259 // Loads cause the NO_READS property to disappear.
8260 removeAssumedBits(NO_READS);
8261 return;
8262
8263 case Instruction::Store:
8264 // Stores cause the NO_WRITES property to disappear if the use is the
8265 // pointer operand. Note that while capturing was taken care of somewhere
8266 // else we need to deal with stores of the value that is not looked through.
8267 if (cast<StoreInst>(UserI)->getPointerOperand() == U.get())
8268 removeAssumedBits(NO_WRITES);
8269 else
8270 indicatePessimisticFixpoint();
8271 return;
8272
8273 case Instruction::Call:
8274 case Instruction::CallBr:
8275 case Instruction::Invoke: {
8276 // For call sites we look at the argument memory behavior attribute (this
8277 // could be recursive!) in order to restrict our own state.
8278 const auto *CB = cast<CallBase>(UserI);
8279
8280 // Give up on operand bundles.
8281 if (CB->isBundleOperand(&U)) {
8282 indicatePessimisticFixpoint();
8283 return;
8284 }
8285
8286 // Calling a function does read the function pointer, maybe write it if the
8287 // function is self-modifying.
8288 if (CB->isCallee(&U)) {
8289 removeAssumedBits(NO_READS);
8290 break;
8291 }
8292
8293 // Adjust the possible access behavior based on the information on the
8294 // argument.
8295 IRPosition Pos;
8296 if (U.get()->getType()->isPointerTy())
8298 else
8300 const auto *MemBehaviorAA =
8301 A.getAAFor<AAMemoryBehavior>(*this, Pos, DepClassTy::OPTIONAL);
8302 if (!MemBehaviorAA)
8303 break;
8304 // "assumed" has at most the same bits as the MemBehaviorAA assumed
8305 // and at least "known".
8306 intersectAssumedBits(MemBehaviorAA->getAssumed());
8307 return;
8308 }
8309 };
8310
8311 // Generally, look at the "may-properties" and adjust the assumed state if we
8312 // did not trigger special handling before.
8313 if (UserI->mayReadFromMemory())
8314 removeAssumedBits(NO_READS);
8315 if (UserI->mayWriteToMemory())
8316 removeAssumedBits(NO_WRITES);
8317}
8318} // namespace
8319
8320/// -------------------- Memory Locations Attributes ---------------------------
8321/// Includes read-none, argmemonly, inaccessiblememonly,
8322/// inaccessiblememorargmemonly
8323/// ----------------------------------------------------------------------------
8324
8327 if (0 == (MLK & AAMemoryLocation::NO_LOCATIONS))
8328 return "all memory";
8330 return "no memory";
8331 std::string S = "memory:";
8332 if (0 == (MLK & AAMemoryLocation::NO_LOCAL_MEM))
8333 S += "stack,";
8334 if (0 == (MLK & AAMemoryLocation::NO_CONST_MEM))
8335 S += "constant,";
8337 S += "internal global,";
8339 S += "external global,";
8340 if (0 == (MLK & AAMemoryLocation::NO_ARGUMENT_MEM))
8341 S += "argument,";
8343 S += "inaccessible,";
8344 if (0 == (MLK & AAMemoryLocation::NO_MALLOCED_MEM))
8345 S += "malloced,";
8346 if (0 == (MLK & AAMemoryLocation::NO_UNKOWN_MEM))
8347 S += "unknown,";
8348 S.pop_back();
8349 return S;
8350}
8351
8352namespace {
8353struct AAMemoryLocationImpl : public AAMemoryLocation {
8354
8355 AAMemoryLocationImpl(const IRPosition &IRP, Attributor &A)
8357 AccessKind2Accesses.fill(nullptr);
8358 }
8359
8360 ~AAMemoryLocationImpl() {
8361 // The AccessSets are allocated via a BumpPtrAllocator, we call
8362 // the destructor manually.
8363 for (AccessSet *AS : AccessKind2Accesses)
8364 if (AS)
8365 AS->~AccessSet();
8366 }
8367
8368 /// See AbstractAttribute::initialize(...).
8369 void initialize(Attributor &A) override {
8370 intersectAssumedBits(BEST_STATE);
8371 getKnownStateFromValue(A, getIRPosition(), getState());
8372 AAMemoryLocation::initialize(A);
8373 }
8374
8375 /// Return the memory behavior information encoded in the IR for \p IRP.
8376 static void getKnownStateFromValue(Attributor &A, const IRPosition &IRP,
8377 BitIntegerState &State,
8378 bool IgnoreSubsumingPositions = false) {
8379 // For internal functions we ignore `argmemonly` and
8380 // `inaccessiblememorargmemonly` as we might break it via interprocedural
8381 // constant propagation. It is unclear if this is the best way but it is
8382 // unlikely this will cause real performance problems. If we are deriving
8383 // attributes for the anchor function we even remove the attribute in
8384 // addition to ignoring it.
8385 // TODO: A better way to handle this would be to add ~NO_GLOBAL_MEM /
8386 // MemoryEffects::Other as a possible location.
8387 bool UseArgMemOnly = true;
8388 Function *AnchorFn = IRP.getAnchorScope();
8389 if (AnchorFn && A.isRunOn(*AnchorFn))
8390 UseArgMemOnly = !AnchorFn->hasLocalLinkage();
8391
8393 A.getAttrs(IRP, {Attribute::Memory}, Attrs, IgnoreSubsumingPositions);
8394 for (const Attribute &Attr : Attrs) {
8395 // TODO: We can map MemoryEffects to Attributor locations more precisely.
8396 MemoryEffects ME = Attr.getMemoryEffects();
8397 if (ME.doesNotAccessMemory()) {
8398 State.addKnownBits(NO_LOCAL_MEM | NO_CONST_MEM);
8399 continue;
8400 }
8401 if (ME.onlyAccessesInaccessibleMem()) {
8402 State.addKnownBits(inverseLocation(NO_INACCESSIBLE_MEM, true, true));
8403 continue;
8404 }
8405 if (ME.onlyAccessesArgPointees()) {
8406 if (UseArgMemOnly)
8407 State.addKnownBits(inverseLocation(NO_ARGUMENT_MEM, true, true));
8408 else {
8409 // Remove location information, only keep read/write info.
8410 ME = MemoryEffects(ME.getModRef());
8411 A.manifestAttrs(IRP,
8413 IRP.getAnchorValue().getContext(), ME),
8414 /*ForceReplace*/ true);
8415 }
8416 continue;
8417 }
8419 if (UseArgMemOnly)
8420 State.addKnownBits(inverseLocation(
8421 NO_INACCESSIBLE_MEM | NO_ARGUMENT_MEM, true, true));
8422 else {
8423 // Remove location information, only keep read/write info.
8424 ME = MemoryEffects(ME.getModRef());
8425 A.manifestAttrs(IRP,
8427 IRP.getAnchorValue().getContext(), ME),
8428 /*ForceReplace*/ true);
8429 }
8430 continue;
8431 }
8432 }
8433 }
8434
8435 /// See AbstractAttribute::getDeducedAttributes(...).
8436 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
8437 SmallVectorImpl<Attribute> &Attrs) const override {
8438 // TODO: We can map Attributor locations to MemoryEffects more precisely.
8439 assert(Attrs.size() == 0);
8440 if (getIRPosition().getPositionKind() == IRPosition::IRP_FUNCTION) {
8441 if (isAssumedReadNone())
8442 Attrs.push_back(
8444 else if (isAssumedInaccessibleMemOnly())
8447 else if (isAssumedArgMemOnly())
8448 Attrs.push_back(
8450 else if (isAssumedInaccessibleOrArgMemOnly())
8453 }
8454 assert(Attrs.size() <= 1);
8455 }
8456
8457 /// See AbstractAttribute::manifest(...).
8458 ChangeStatus manifest(Attributor &A) override {
8459 // TODO: If AAMemoryLocation and AAMemoryBehavior are merged, we could
8460 // provide per-location modref information here.
8461 const IRPosition &IRP = getIRPosition();
8462
8463 SmallVector<Attribute, 1> DeducedAttrs;
8464 getDeducedAttributes(A, IRP.getAnchorValue().getContext(), DeducedAttrs);
8465 if (DeducedAttrs.size() != 1)
8466 return ChangeStatus::UNCHANGED;
8467 MemoryEffects ME = DeducedAttrs[0].getMemoryEffects();
8468
8469 return A.manifestAttrs(IRP, Attribute::getWithMemoryEffects(
8470 IRP.getAnchorValue().getContext(), ME));
8471 }
8472
8473 /// See AAMemoryLocation::checkForAllAccessesToMemoryKind(...).
8474 bool checkForAllAccessesToMemoryKind(
8475 function_ref<bool(const Instruction *, const Value *, AccessKind,
8476 MemoryLocationsKind)>
8477 Pred,
8478 MemoryLocationsKind RequestedMLK) const override {
8479 if (!isValidState())
8480 return false;
8481
8482 MemoryLocationsKind AssumedMLK = getAssumedNotAccessedLocation();
8483 if (AssumedMLK == NO_LOCATIONS)
8484 return true;
8485
8486 unsigned Idx = 0;
8487 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS;
8488 CurMLK *= 2, ++Idx) {
8489 if (CurMLK & RequestedMLK)
8490 continue;
8491
8492 if (const AccessSet *Accesses = AccessKind2Accesses[Idx])
8493 for (const AccessInfo &AI : *Accesses)
8494 if (!Pred(AI.I, AI.Ptr, AI.Kind, CurMLK))
8495 return false;
8496 }
8497
8498 return true;
8499 }
8500
8501 ChangeStatus indicatePessimisticFixpoint() override {
8502 // If we give up and indicate a pessimistic fixpoint this instruction will
8503 // become an access for all potential access kinds:
8504 // TODO: Add pointers for argmemonly and globals to improve the results of
8505 // checkForAllAccessesToMemoryKind.
8506 bool Changed = false;
8507 MemoryLocationsKind KnownMLK = getKnown();
8508 Instruction *I = dyn_cast<Instruction>(&getAssociatedValue());
8509 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS; CurMLK *= 2)
8510 if (!(CurMLK & KnownMLK))
8511 updateStateAndAccessesMap(getState(), CurMLK, I, nullptr, Changed,
8512 getAccessKindFromInst(I));
8513 return AAMemoryLocation::indicatePessimisticFixpoint();
8514 }
8515
8516protected:
8517 /// Helper struct to tie together an instruction that has a read or write
8518 /// effect with the pointer it accesses (if any).
8519 struct AccessInfo {
8520
8521 /// The instruction that caused the access.
8522 const Instruction *I;
8523
8524 /// The base pointer that is accessed, or null if unknown.
8525 const Value *Ptr;
8526
8527 /// The kind of access (read/write/read+write).
8529
8530 bool operator==(const AccessInfo &RHS) const {
8531 return I == RHS.I && Ptr == RHS.Ptr && Kind == RHS.Kind;
8532 }
8533 bool operator()(const AccessInfo &LHS, const AccessInfo &RHS) const {
8534 if (LHS.I != RHS.I)
8535 return LHS.I < RHS.I;
8536 if (LHS.Ptr != RHS.Ptr)
8537 return LHS.Ptr < RHS.Ptr;
8538 if (LHS.Kind != RHS.Kind)
8539 return LHS.Kind < RHS.Kind;
8540 return false;
8541 }
8542 };
8543
8544 /// Mapping from *single* memory location kinds, e.g., LOCAL_MEM with the
8545 /// value of NO_LOCAL_MEM, to the accesses encountered for this memory kind.
8546 using AccessSet = SmallSet<AccessInfo, 2, AccessInfo>;
8547 std::array<AccessSet *, llvm::CTLog2<VALID_STATE>()> AccessKind2Accesses;
8548
8549 /// Categorize the pointer arguments of CB that might access memory in
8550 /// AccessedLoc and update the state and access map accordingly.
8551 void
8552 categorizeArgumentPointerLocations(Attributor &A, CallBase &CB,
8553 AAMemoryLocation::StateType &AccessedLocs,
8554 bool &Changed);
8555
8556 /// Return the kind(s) of location that may be accessed by \p V.
8558 categorizeAccessedLocations(Attributor &A, Instruction &I, bool &Changed);
8559
8560 /// Return the access kind as determined by \p I.
8561 AccessKind getAccessKindFromInst(const Instruction *I) {
8562 AccessKind AK = READ_WRITE;
8563 if (I) {
8564 AK = I->mayReadFromMemory() ? READ : NONE;
8565 AK = AccessKind(AK | (I->mayWriteToMemory() ? WRITE : NONE));
8566 }
8567 return AK;
8568 }
8569
8570 /// Update the state \p State and the AccessKind2Accesses given that \p I is
8571 /// an access of kind \p AK to a \p MLK memory location with the access
8572 /// pointer \p Ptr.
8573 void updateStateAndAccessesMap(AAMemoryLocation::StateType &State,
8574 MemoryLocationsKind MLK, const Instruction *I,
8575 const Value *Ptr, bool &Changed,
8576 AccessKind AK = READ_WRITE) {
8577
8578 assert(isPowerOf2_32(MLK) && "Expected a single location set!");
8579 auto *&Accesses = AccessKind2Accesses[llvm::Log2_32(MLK)];
8580 if (!Accesses)
8581 Accesses = new (Allocator) AccessSet();
8582 Changed |= Accesses->insert(AccessInfo{I, Ptr, AK}).second;
8583 if (MLK == NO_UNKOWN_MEM)
8584 MLK = NO_LOCATIONS;
8585 State.removeAssumedBits(MLK);
8586 }
8587
8588 /// Determine the underlying locations kinds for \p Ptr, e.g., globals or
8589 /// arguments, and update the state and access map accordingly.
8590 void categorizePtrValue(Attributor &A, const Instruction &I, const Value &Ptr,
8591 AAMemoryLocation::StateType &State, bool &Changed,
8592 unsigned AccessAS = 0);
8593
8594 /// Used to allocate access sets.
8596};
8597
8598void AAMemoryLocationImpl::categorizePtrValue(
8599 Attributor &A, const Instruction &I, const Value &Ptr,
8600 AAMemoryLocation::StateType &State, bool &Changed, unsigned AccessAS) {
8601 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize pointer locations for "
8602 << Ptr << " ["
8603 << getMemoryLocationsAsStr(State.getAssumed()) << "]\n");
8604
8605 auto Pred = [&](Value &Obj) {
8606 unsigned ObjectAS = Obj.getType()->getPointerAddressSpace();
8607 // TODO: recognize the TBAA used for constant accesses.
8608 MemoryLocationsKind MLK = NO_LOCATIONS;
8609
8610 // Filter accesses to constant (GPU) memory if we have an AS at the access
8611 // site or the object is known to actually have the associated AS.
8612 if ((AccessAS == (unsigned)AA::GPUAddressSpace::Constant ||
8613 (ObjectAS == (unsigned)AA::GPUAddressSpace::Constant &&
8614 isIdentifiedObject(&Obj))) &&
8615 AA::isGPU(*I.getModule()))
8616 return true;
8617
8618 if (isa<UndefValue>(&Obj))
8619 return true;
8620 if (isa<Argument>(&Obj)) {
8621 // TODO: For now we do not treat byval arguments as local copies performed
8622 // on the call edge, though, we should. To make that happen we need to
8623 // teach various passes, e.g., DSE, about the copy effect of a byval. That
8624 // would also allow us to mark functions only accessing byval arguments as
8625 // readnone again, arguably their accesses have no effect outside of the
8626 // function, like accesses to allocas.
8627 MLK = NO_ARGUMENT_MEM;
8628 } else if (auto *GV = dyn_cast<GlobalValue>(&Obj)) {
8629 // Reading constant memory is not treated as a read "effect" by the
8630 // function attr pass so we won't neither. Constants defined by TBAA are
8631 // similar. (We know we do not write it because it is constant.)
8632 if (auto *GVar = dyn_cast<GlobalVariable>(GV))
8633 if (GVar->isConstant())
8634 return true;
8635
8636 if (GV->hasLocalLinkage())
8637 MLK = NO_GLOBAL_INTERNAL_MEM;
8638 else
8639 MLK = NO_GLOBAL_EXTERNAL_MEM;
8640 } else if (isa<ConstantPointerNull>(&Obj) &&
8641 (!NullPointerIsDefined(getAssociatedFunction(), AccessAS) ||
8642 !NullPointerIsDefined(getAssociatedFunction(), ObjectAS))) {
8643 return true;
8644 } else if (isa<AllocaInst>(&Obj)) {
8645 MLK = NO_LOCAL_MEM;
8646 } else if (const auto *CB = dyn_cast<CallBase>(&Obj)) {
8647 bool IsKnownNoAlias;
8648 if (AA::hasAssumedIRAttr<Attribute::NoAlias>(
8650 IsKnownNoAlias))
8651 MLK = NO_MALLOCED_MEM;
8652 else
8653 MLK = NO_UNKOWN_MEM;
8654 } else {
8655 MLK = NO_UNKOWN_MEM;
8656 }
8657
8658 assert(MLK != NO_LOCATIONS && "No location specified!");
8659 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Ptr value can be categorized: "
8660 << Obj << " -> " << getMemoryLocationsAsStr(MLK) << "\n");
8661 updateStateAndAccessesMap(State, MLK, &I, &Obj, Changed,
8662 getAccessKindFromInst(&I));
8663
8664 return true;
8665 };
8666
8667 const auto *AA = A.getAAFor<AAUnderlyingObjects>(
8669 if (!AA || !AA->forallUnderlyingObjects(Pred, AA::Intraprocedural)) {
8670 LLVM_DEBUG(
8671 dbgs() << "[AAMemoryLocation] Pointer locations not categorized\n");
8672 updateStateAndAccessesMap(State, NO_UNKOWN_MEM, &I, nullptr, Changed,
8673 getAccessKindFromInst(&I));
8674 return;
8675 }
8676
8677 LLVM_DEBUG(
8678 dbgs() << "[AAMemoryLocation] Accessed locations with pointer locations: "
8679 << getMemoryLocationsAsStr(State.getAssumed()) << "\n");
8680}
8681
8682void AAMemoryLocationImpl::categorizeArgumentPointerLocations(
8683 Attributor &A, CallBase &CB, AAMemoryLocation::StateType &AccessedLocs,
8684 bool &Changed) {
8685 for (unsigned ArgNo = 0, E = CB.arg_size(); ArgNo < E; ++ArgNo) {
8686
8687 // Skip non-pointer arguments.
8688 const Value *ArgOp = CB.getArgOperand(ArgNo);
8689 if (!ArgOp->getType()->isPtrOrPtrVectorTy())
8690 continue;
8691
8692 // Skip readnone arguments.
8693 const IRPosition &ArgOpIRP = IRPosition::callsite_argument(CB, ArgNo);
8694 const auto *ArgOpMemLocationAA =
8695 A.getAAFor<AAMemoryBehavior>(*this, ArgOpIRP, DepClassTy::OPTIONAL);
8696
8697 if (ArgOpMemLocationAA && ArgOpMemLocationAA->isAssumedReadNone())
8698 continue;
8699
8700 // Categorize potentially accessed pointer arguments as if there was an
8701 // access instruction with them as pointer.
8702 categorizePtrValue(A, CB, *ArgOp, AccessedLocs, Changed);
8703 }
8704}
8705
8707AAMemoryLocationImpl::categorizeAccessedLocations(Attributor &A, Instruction &I,
8708 bool &Changed) {
8709 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize accessed locations for "
8710 << I << "\n");
8711
8712 AAMemoryLocation::StateType AccessedLocs;
8713 AccessedLocs.intersectAssumedBits(NO_LOCATIONS);
8714
8715 if (auto *CB = dyn_cast<CallBase>(&I)) {
8716
8717 // First check if we assume any memory is access is visible.
8718 const auto *CBMemLocationAA = A.getAAFor<AAMemoryLocation>(
8720 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Categorize call site: " << I
8721 << " [" << CBMemLocationAA << "]\n");
8722 if (!CBMemLocationAA) {
8723 updateStateAndAccessesMap(AccessedLocs, NO_UNKOWN_MEM, &I, nullptr,
8724 Changed, getAccessKindFromInst(&I));
8725 return NO_UNKOWN_MEM;
8726 }
8727
8728 if (CBMemLocationAA->isAssumedReadNone())
8729 return NO_LOCATIONS;
8730
8731 if (CBMemLocationAA->isAssumedInaccessibleMemOnly()) {
8732 updateStateAndAccessesMap(AccessedLocs, NO_INACCESSIBLE_MEM, &I, nullptr,
8733 Changed, getAccessKindFromInst(&I));
8734 return AccessedLocs.getAssumed();
8735 }
8736
8737 uint32_t CBAssumedNotAccessedLocs =
8738 CBMemLocationAA->getAssumedNotAccessedLocation();
8739
8740 // Set the argmemonly and global bit as we handle them separately below.
8741 uint32_t CBAssumedNotAccessedLocsNoArgMem =
8742 CBAssumedNotAccessedLocs | NO_ARGUMENT_MEM | NO_GLOBAL_MEM;
8743
8744 for (MemoryLocationsKind CurMLK = 1; CurMLK < NO_LOCATIONS; CurMLK *= 2) {
8745 if (CBAssumedNotAccessedLocsNoArgMem & CurMLK)
8746 continue;
8747 updateStateAndAccessesMap(AccessedLocs, CurMLK, &I, nullptr, Changed,
8748 getAccessKindFromInst(&I));
8749 }
8750
8751 // Now handle global memory if it might be accessed. This is slightly tricky
8752 // as NO_GLOBAL_MEM has multiple bits set.
8753 bool HasGlobalAccesses = ((~CBAssumedNotAccessedLocs) & NO_GLOBAL_MEM);
8754 if (HasGlobalAccesses) {
8755 auto AccessPred = [&](const Instruction *, const Value *Ptr,
8756 AccessKind Kind, MemoryLocationsKind MLK) {
8757 updateStateAndAccessesMap(AccessedLocs, MLK, &I, Ptr, Changed,
8758 getAccessKindFromInst(&I));
8759 return true;
8760 };
8761 if (!CBMemLocationAA->checkForAllAccessesToMemoryKind(
8762 AccessPred, inverseLocation(NO_GLOBAL_MEM, false, false)))
8763 return AccessedLocs.getWorstState();
8764 }
8765
8766 LLVM_DEBUG(
8767 dbgs() << "[AAMemoryLocation] Accessed state before argument handling: "
8768 << getMemoryLocationsAsStr(AccessedLocs.getAssumed()) << "\n");
8769
8770 // Now handle argument memory if it might be accessed.
8771 bool HasArgAccesses = ((~CBAssumedNotAccessedLocs) & NO_ARGUMENT_MEM);
8772 if (HasArgAccesses)
8773 categorizeArgumentPointerLocations(A, *CB, AccessedLocs, Changed);
8774
8775 LLVM_DEBUG(
8776 dbgs() << "[AAMemoryLocation] Accessed state after argument handling: "
8777 << getMemoryLocationsAsStr(AccessedLocs.getAssumed()) << "\n");
8778
8779 return AccessedLocs.getAssumed();
8780 }
8781
8782 if (const Value *Ptr = getPointerOperand(&I, /* AllowVolatile */ true)) {
8783 LLVM_DEBUG(
8784 dbgs() << "[AAMemoryLocation] Categorize memory access with pointer: "
8785 << I << " [" << *Ptr << "]\n");
8786 categorizePtrValue(A, I, *Ptr, AccessedLocs, Changed,
8787 Ptr->getType()->getPointerAddressSpace());
8788 return AccessedLocs.getAssumed();
8789 }
8790
8791 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Failed to categorize instruction: "
8792 << I << "\n");
8793 updateStateAndAccessesMap(AccessedLocs, NO_UNKOWN_MEM, &I, nullptr, Changed,
8794 getAccessKindFromInst(&I));
8795 return AccessedLocs.getAssumed();
8796}
8797
8798/// An AA to represent the memory behavior function attributes.
8799struct AAMemoryLocationFunction final : public AAMemoryLocationImpl {
8800 AAMemoryLocationFunction(const IRPosition &IRP, Attributor &A)
8801 : AAMemoryLocationImpl(IRP, A) {}
8802
8803 /// See AbstractAttribute::updateImpl(Attributor &A).
8804 ChangeStatus updateImpl(Attributor &A) override {
8805
8806 const auto *MemBehaviorAA =
8807 A.getAAFor<AAMemoryBehavior>(*this, getIRPosition(), DepClassTy::NONE);
8808 if (MemBehaviorAA && MemBehaviorAA->isAssumedReadNone()) {
8809 if (MemBehaviorAA->isKnownReadNone())
8810 return indicateOptimisticFixpoint();
8812 "AAMemoryLocation was not read-none but AAMemoryBehavior was!");
8813 A.recordDependence(*MemBehaviorAA, *this, DepClassTy::OPTIONAL);
8814 return ChangeStatus::UNCHANGED;
8815 }
8816
8817 // The current assumed state used to determine a change.
8818 auto AssumedState = getAssumed();
8819 bool Changed = false;
8820
8821 auto CheckRWInst = [&](Instruction &I) {
8822 MemoryLocationsKind MLK = categorizeAccessedLocations(A, I, Changed);
8823 LLVM_DEBUG(dbgs() << "[AAMemoryLocation] Accessed locations for " << I
8824 << ": " << getMemoryLocationsAsStr(MLK) << "\n");
8825 removeAssumedBits(inverseLocation(MLK, false, false));
8826 // Stop once only the valid bit set in the *not assumed location*, thus
8827 // once we don't actually exclude any memory locations in the state.
8828 return getAssumedNotAccessedLocation() != VALID_STATE;
8829 };
8830
8831 bool UsedAssumedInformation = false;
8832 if (!A.checkForAllReadWriteInstructions(CheckRWInst, *this,
8833 UsedAssumedInformation))
8834 return indicatePessimisticFixpoint();
8835
8836 Changed |= AssumedState != getAssumed();
8837 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
8838 }
8839
8840 /// See AbstractAttribute::trackStatistics()
8841 void trackStatistics() const override {
8842 if (isAssumedReadNone())
8843 STATS_DECLTRACK_FN_ATTR(readnone)
8844 else if (isAssumedArgMemOnly())
8845 STATS_DECLTRACK_FN_ATTR(argmemonly)
8846 else if (isAssumedInaccessibleMemOnly())
8847 STATS_DECLTRACK_FN_ATTR(inaccessiblememonly)
8848 else if (isAssumedInaccessibleOrArgMemOnly())
8849 STATS_DECLTRACK_FN_ATTR(inaccessiblememorargmemonly)
8850 }
8851};
8852
8853/// AAMemoryLocation attribute for call sites.
8854struct AAMemoryLocationCallSite final : AAMemoryLocationImpl {
8855 AAMemoryLocationCallSite(const IRPosition &IRP, Attributor &A)
8856 : AAMemoryLocationImpl(IRP, A) {}
8857
8858 /// See AbstractAttribute::updateImpl(...).
8859 ChangeStatus updateImpl(Attributor &A) override {
8860 // TODO: Once we have call site specific value information we can provide
8861 // call site specific liveness liveness information and then it makes
8862 // sense to specialize attributes for call sites arguments instead of
8863 // redirecting requests to the callee argument.
8864 Function *F = getAssociatedFunction();
8865 const IRPosition &FnPos = IRPosition::function(*F);
8866 auto *FnAA =
8867 A.getAAFor<AAMemoryLocation>(*this, FnPos, DepClassTy::REQUIRED);
8868 if (!FnAA)
8869 return indicatePessimisticFixpoint();
8870 bool Changed = false;
8871 auto AccessPred = [&](const Instruction *I, const Value *Ptr,
8872 AccessKind Kind, MemoryLocationsKind MLK) {
8873 updateStateAndAccessesMap(getState(), MLK, I, Ptr, Changed,
8874 getAccessKindFromInst(I));
8875 return true;
8876 };
8877 if (!FnAA->checkForAllAccessesToMemoryKind(AccessPred, ALL_LOCATIONS))
8878 return indicatePessimisticFixpoint();
8879 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
8880 }
8881
8882 /// See AbstractAttribute::trackStatistics()
8883 void trackStatistics() const override {
8884 if (isAssumedReadNone())
8885 STATS_DECLTRACK_CS_ATTR(readnone)
8886 }
8887};
8888} // namespace
8889
8890/// ------------------ denormal-fp-math Attribute -------------------------
8891
8892namespace {
8893struct AADenormalFPMathImpl : public AADenormalFPMath {
8894 AADenormalFPMathImpl(const IRPosition &IRP, Attributor &A)
8895 : AADenormalFPMath(IRP, A) {}
8896
8897 const std::string getAsStr(Attributor *A) const override {
8898 std::string Str("AADenormalFPMath[");
8900
8901 DenormalState Known = getKnown();
8902 if (Known.Mode.isValid())
8903 OS << "denormal-fp-math=" << Known.Mode;
8904 else
8905 OS << "invalid";
8906
8907 if (Known.ModeF32.isValid())
8908 OS << " denormal-fp-math-f32=" << Known.ModeF32;
8909 OS << ']';
8910 return Str;
8911 }
8912};
8913
8914struct AADenormalFPMathFunction final : AADenormalFPMathImpl {
8915 AADenormalFPMathFunction(const IRPosition &IRP, Attributor &A)
8916 : AADenormalFPMathImpl(IRP, A) {}
8917
8918 void initialize(Attributor &A) override {
8919 const Function *F = getAnchorScope();
8920 DenormalMode Mode = F->getDenormalModeRaw();
8921 DenormalMode ModeF32 = F->getDenormalModeF32Raw();
8922
8923 // TODO: Handling this here prevents handling the case where a callee has a
8924 // fixed denormal-fp-math with dynamic denormal-fp-math-f32, but called from
8925 // a function with a fully fixed mode.
8926 if (ModeF32 == DenormalMode::getInvalid())
8927 ModeF32 = Mode;
8928 Known = DenormalState{Mode, ModeF32};
8929 if (isModeFixed())
8930 indicateFixpoint();
8931 }
8932
8933 ChangeStatus updateImpl(Attributor &A) override {
8934 ChangeStatus Change = ChangeStatus::UNCHANGED;
8935
8936 auto CheckCallSite = [=, &Change, &A](AbstractCallSite CS) {
8937 Function *Caller = CS.getInstruction()->getFunction();
8938 LLVM_DEBUG(dbgs() << "[AADenormalFPMath] Call " << Caller->getName()
8939 << "->" << getAssociatedFunction()->getName() << '\n');
8940
8941 const auto *CallerInfo = A.getAAFor<AADenormalFPMath>(
8942 *this, IRPosition::function(*Caller), DepClassTy::REQUIRED);
8943 if (!CallerInfo)
8944 return false;
8945
8946 Change = Change | clampStateAndIndicateChange(this->getState(),
8947 CallerInfo->getState());
8948 return true;
8949 };
8950
8951 bool AllCallSitesKnown = true;
8952 if (!A.checkForAllCallSites(CheckCallSite, *this, true, AllCallSitesKnown))
8953 return indicatePessimisticFixpoint();
8954
8955 if (Change == ChangeStatus::CHANGED && isModeFixed())
8956 indicateFixpoint();
8957 return Change;
8958 }
8959
8960 ChangeStatus manifest(Attributor &A) override {
8961 LLVMContext &Ctx = getAssociatedFunction()->getContext();
8962
8963 SmallVector<Attribute, 2> AttrToAdd;
8964 SmallVector<StringRef, 2> AttrToRemove;
8965 if (Known.Mode == DenormalMode::getDefault()) {
8966 AttrToRemove.push_back("denormal-fp-math");
8967 } else {
8968 AttrToAdd.push_back(
8969 Attribute::get(Ctx, "denormal-fp-math", Known.Mode.str()));
8970 }
8971
8972 if (Known.ModeF32 != Known.Mode) {
8973 AttrToAdd.push_back(
8974 Attribute::get(Ctx, "denormal-fp-math-f32", Known.ModeF32.str()));
8975 } else {
8976 AttrToRemove.push_back("denormal-fp-math-f32");
8977 }
8978
8979 auto &IRP = getIRPosition();
8980
8981 // TODO: There should be a combined add and remove API.
8982 return A.removeAttrs(IRP, AttrToRemove) |
8983 A.manifestAttrs(IRP, AttrToAdd, /*ForceReplace=*/true);
8984 }
8985
8986 void trackStatistics() const override {
8987 STATS_DECLTRACK_FN_ATTR(denormal_fp_math)
8988 }
8989};
8990} // namespace
8991
8992/// ------------------ Value Constant Range Attribute -------------------------
8993
8994namespace {
8995struct AAValueConstantRangeImpl : AAValueConstantRange {
8996 using StateType = IntegerRangeState;
8997 AAValueConstantRangeImpl(const IRPosition &IRP, Attributor &A)
8998 : AAValueConstantRange(IRP, A) {}
8999
9000 /// See AbstractAttribute::initialize(..).
9001 void initialize(Attributor &A) override {
9002 if (A.hasSimplificationCallback(getIRPosition())) {
9003 indicatePessimisticFixpoint();
9004 return;
9005 }
9006
9007 // Intersect a range given by SCEV.
9008 intersectKnown(getConstantRangeFromSCEV(A, getCtxI()));
9009
9010 // Intersect a range given by LVI.
9011 intersectKnown(getConstantRangeFromLVI(A, getCtxI()));
9012 }
9013
9014 /// See AbstractAttribute::getAsStr().
9015 const std::string getAsStr(Attributor *A) const override {
9016 std::string Str;
9018 OS << "range(" << getBitWidth() << ")<";
9019 getKnown().print(OS);
9020 OS << " / ";
9021 getAssumed().print(OS);
9022 OS << ">";
9023 return Str;
9024 }
9025
9026 /// Helper function to get a SCEV expr for the associated value at program
9027 /// point \p I.
9028 const SCEV *getSCEV(Attributor &A, const Instruction *I = nullptr) const {
9029 if (!getAnchorScope())
9030 return nullptr;
9031
9032 ScalarEvolution *SE =
9033 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(
9034 *getAnchorScope());
9035
9036 LoopInfo *LI = A.getInfoCache().getAnalysisResultForFunction<LoopAnalysis>(
9037 *getAnchorScope());
9038
9039 if (!SE || !LI)
9040 return nullptr;
9041
9042 const SCEV *S = SE->getSCEV(&getAssociatedValue());
9043 if (!I)
9044 return S;
9045
9046 return SE->getSCEVAtScope(S, LI->getLoopFor(I->getParent()));
9047 }
9048
9049 /// Helper function to get a range from SCEV for the associated value at
9050 /// program point \p I.
9051 ConstantRange getConstantRangeFromSCEV(Attributor &A,
9052 const Instruction *I = nullptr) const {
9053 if (!getAnchorScope())
9054 return getWorstState(getBitWidth());
9055
9056 ScalarEvolution *SE =
9057 A.getInfoCache().getAnalysisResultForFunction<ScalarEvolutionAnalysis>(
9058 *getAnchorScope());
9059
9060 const SCEV *S = getSCEV(A, I);
9061 if (!SE || !S)
9062 return getWorstState(getBitWidth());
9063
9064 return SE->getUnsignedRange(S);
9065 }
9066
9067 /// Helper function to get a range from LVI for the associated value at
9068 /// program point \p I.
9070 getConstantRangeFromLVI(Attributor &A,
9071 const Instruction *CtxI = nullptr) const {
9072 if (!getAnchorScope())
9073 return getWorstState(getBitWidth());
9074
9075 LazyValueInfo *LVI =
9076 A.getInfoCache().getAnalysisResultForFunction<LazyValueAnalysis>(
9077 *getAnchorScope());
9078
9079 if (!LVI || !CtxI)
9080 return getWorstState(getBitWidth());
9081 return LVI->getConstantRange(&getAssociatedValue(),
9082 const_cast<Instruction *>(CtxI),
9083 /*UndefAllowed*/ false);
9084 }
9085
9086 /// Return true if \p CtxI is valid for querying outside analyses.
9087 /// This basically makes sure we do not ask intra-procedural analysis
9088 /// about a context in the wrong function or a context that violates
9089 /// dominance assumptions they might have. The \p AllowAACtxI flag indicates
9090 /// if the original context of this AA is OK or should be considered invalid.
9091 bool isValidCtxInstructionForOutsideAnalysis(Attributor &A,
9092 const Instruction *CtxI,
9093 bool AllowAACtxI) const {
9094 if (!CtxI || (!AllowAACtxI && CtxI == getCtxI()))
9095 return false;
9096
9097 // Our context might be in a different function, neither intra-procedural
9098 // analysis (ScalarEvolution nor LazyValueInfo) can handle that.
9099 if (!AA::isValidInScope(getAssociatedValue(), CtxI->getFunction()))
9100 return false;
9101
9102 // If the context is not dominated by the value there are paths to the
9103 // context that do not define the value. This cannot be handled by
9104 // LazyValueInfo so we need to bail.
9105 if (auto *I = dyn_cast<Instruction>(&getAssociatedValue())) {
9106 InformationCache &InfoCache = A.getInfoCache();
9107 const DominatorTree *DT =
9109 *I->getFunction());
9110 return DT && DT->dominates(I, CtxI);
9111 }
9112
9113 return true;
9114 }
9115
9116 /// See AAValueConstantRange::getKnownConstantRange(..).
9118 getKnownConstantRange(Attributor &A,
9119 const Instruction *CtxI = nullptr) const override {
9120 if (!isValidCtxInstructionForOutsideAnalysis(A, CtxI,
9121 /* AllowAACtxI */ false))
9122 return getKnown();
9123
9124 ConstantRange LVIR = getConstantRangeFromLVI(A, CtxI);
9125 ConstantRange SCEVR = getConstantRangeFromSCEV(A, CtxI);
9126 return getKnown().intersectWith(SCEVR).intersectWith(LVIR);
9127 }
9128
9129 /// See AAValueConstantRange::getAssumedConstantRange(..).
9131 getAssumedConstantRange(Attributor &A,
9132 const Instruction *CtxI = nullptr) const override {
9133 // TODO: Make SCEV use Attributor assumption.
9134 // We may be able to bound a variable range via assumptions in
9135 // Attributor. ex.) If x is assumed to be in [1, 3] and y is known to
9136 // evolve to x^2 + x, then we can say that y is in [2, 12].
9137 if (!isValidCtxInstructionForOutsideAnalysis(A, CtxI,
9138 /* AllowAACtxI */ false))
9139 return getAssumed();
9140
9141 ConstantRange LVIR = getConstantRangeFromLVI(A, CtxI);
9142 ConstantRange SCEVR = getConstantRangeFromSCEV(A, CtxI);
9143 return getAssumed().intersectWith(SCEVR).intersectWith(LVIR);
9144 }
9145
9146 /// Helper function to create MDNode for range metadata.
9147 static MDNode *
9148 getMDNodeForConstantRange(Type *Ty, LLVMContext &Ctx,
9149 const ConstantRange &AssumedConstantRange) {
9150 Metadata *LowAndHigh[] = {ConstantAsMetadata::get(ConstantInt::get(
9151 Ty, AssumedConstantRange.getLower())),
9152 ConstantAsMetadata::get(ConstantInt::get(
9153 Ty, AssumedConstantRange.getUpper()))};
9154 return MDNode::get(Ctx, LowAndHigh);
9155 }
9156
9157 /// Return true if \p Assumed is included in \p KnownRanges.
9158 static bool isBetterRange(const ConstantRange &Assumed, MDNode *KnownRanges) {
9159
9160 if (Assumed.isFullSet())
9161 return false;
9162
9163 if (!KnownRanges)
9164 return true;
9165
9166 // If multiple ranges are annotated in IR, we give up to annotate assumed
9167 // range for now.
9168
9169 // TODO: If there exists a known range which containts assumed range, we
9170 // can say assumed range is better.
9171 if (KnownRanges->getNumOperands() > 2)
9172 return false;
9173
9175 mdconst::extract<ConstantInt>(KnownRanges->getOperand(0));
9177 mdconst::extract<ConstantInt>(KnownRanges->getOperand(1));
9178
9179 ConstantRange Known(Lower->getValue(), Upper->getValue());
9180 return Known.contains(Assumed) && Known != Assumed;
9181 }
9182
9183 /// Helper function to set range metadata.
9184 static bool
9185 setRangeMetadataIfisBetterRange(Instruction *I,
9186 const ConstantRange &AssumedConstantRange) {
9187 auto *OldRangeMD = I->getMetadata(LLVMContext::MD_range);
9188 if (isBetterRange(AssumedConstantRange, OldRangeMD)) {
9189 if (!AssumedConstantRange.isEmptySet()) {
9190 I->setMetadata(LLVMContext::MD_range,
9191 getMDNodeForConstantRange(I->getType(), I->getContext(),
9192 AssumedConstantRange));
9193 return true;
9194 }
9195 }
9196 return false;
9197 }
9198
9199 /// See AbstractAttribute::manifest()
9200 ChangeStatus manifest(Attributor &A) override {
9201 ChangeStatus Changed = ChangeStatus::UNCHANGED;
9202 ConstantRange AssumedConstantRange = getAssumedConstantRange(A);
9203 assert(!AssumedConstantRange.isFullSet() && "Invalid state");
9204
9205 auto &V = getAssociatedValue();
9206 if (!AssumedConstantRange.isEmptySet() &&
9207 !AssumedConstantRange.isSingleElement()) {
9208 if (Instruction *I = dyn_cast<Instruction>(&V)) {
9209 assert(I == getCtxI() && "Should not annotate an instruction which is "
9210 "not the context instruction");
9211 if (isa<CallInst>(I) || isa<LoadInst>(I))
9212 if (setRangeMetadataIfisBetterRange(I, AssumedConstantRange))
9213 Changed = ChangeStatus::CHANGED;
9214 }
9215 }
9216
9217 return Changed;
9218 }
9219};
9220
9221struct AAValueConstantRangeArgument final
9222 : AAArgumentFromCallSiteArguments<
9223 AAValueConstantRange, AAValueConstantRangeImpl, IntegerRangeState,
9224 true /* BridgeCallBaseContext */> {
9225 using Base = AAArgumentFromCallSiteArguments<
9226 AAValueConstantRange, AAValueConstantRangeImpl, IntegerRangeState,
9227 true /* BridgeCallBaseContext */>;
9228 AAValueConstantRangeArgument(const IRPosition &IRP, Attributor &A)
9229 : Base(IRP, A) {}
9230
9231 /// See AbstractAttribute::trackStatistics()
9232 void trackStatistics() const override {
9233 STATS_DECLTRACK_ARG_ATTR(value_range)
9234 }
9235};
9236
9237struct AAValueConstantRangeReturned
9238 : AAReturnedFromReturnedValues<AAValueConstantRange,
9239 AAValueConstantRangeImpl,
9240 AAValueConstantRangeImpl::StateType,
9241 /* PropagateCallBaseContext */ true> {
9242 using Base =
9243 AAReturnedFromReturnedValues<AAValueConstantRange,
9244 AAValueConstantRangeImpl,
9246 /* PropagateCallBaseContext */ true>;
9247 AAValueConstantRangeReturned(const IRPosition &IRP, Attributor &A)
9248 : Base(IRP, A) {}
9249
9250 /// See AbstractAttribute::initialize(...).
9251 void initialize(Attributor &A) override {
9252 if (!A.isFunctionIPOAmendable(*getAssociatedFunction()))
9253 indicatePessimisticFixpoint();
9254 }
9255
9256 /// See AbstractAttribute::trackStatistics()
9257 void trackStatistics() const override {
9258 STATS_DECLTRACK_FNRET_ATTR(value_range)
9259 }
9260};
9261
9262struct AAValueConstantRangeFloating : AAValueConstantRangeImpl {
9263 AAValueConstantRangeFloating(const IRPosition &IRP, Attributor &A)
9264 : AAValueConstantRangeImpl(IRP, A) {}
9265
9266 /// See AbstractAttribute::initialize(...).
9267 void initialize(Attributor &A) override {
9268 AAValueConstantRangeImpl::initialize(A);
9269 if (isAtFixpoint())
9270 return;
9271
9272 Value &V = getAssociatedValue();
9273
9274 if (auto *C = dyn_cast<ConstantInt>(&V)) {
9275 unionAssumed(ConstantRange(C->getValue()));
9276 indicateOptimisticFixpoint();
9277 return;
9278 }
9279
9280 if (isa<UndefValue>(&V)) {
9281 // Collapse the undef state to 0.
9282 unionAssumed(ConstantRange(APInt(getBitWidth(), 0)));
9283 indicateOptimisticFixpoint();
9284 return;
9285 }
9286
9287 if (isa<CallBase>(&V))
9288 return;
9289
9290 if (isa<BinaryOperator>(&V) || isa<CmpInst>(&V) || isa<CastInst>(&V))
9291 return;
9292
9293 // If it is a load instruction with range metadata, use it.
9294 if (LoadInst *LI = dyn_cast<LoadInst>(&V))
9295 if (auto *RangeMD = LI->getMetadata(LLVMContext::MD_range)) {
9296 intersectKnown(getConstantRangeFromMetadata(*RangeMD));
9297 return;
9298 }
9299
9300 // We can work with PHI and select instruction as we traverse their operands
9301 // during update.
9302 if (isa<SelectInst>(V) || isa<PHINode>(V))
9303 return;
9304
9305 // Otherwise we give up.
9306 indicatePessimisticFixpoint();
9307
9308 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] We give up: "
9309 << getAssociatedValue() << "\n");
9310 }
9311
9312 bool calculateBinaryOperator(
9314 const Instruction *CtxI,
9316 Value *LHS = BinOp->getOperand(0);
9317 Value *RHS = BinOp->getOperand(1);
9318
9319 // Simplify the operands first.
9320 bool UsedAssumedInformation = false;
9321 const auto &SimplifiedLHS = A.getAssumedSimplified(
9322 IRPosition::value(*LHS, getCallBaseContext()), *this,
9323 UsedAssumedInformation, AA::Interprocedural);
9324 if (!SimplifiedLHS.has_value())
9325 return true;
9326 if (!*SimplifiedLHS)
9327 return false;
9328 LHS = *SimplifiedLHS;
9329
9330 const auto &SimplifiedRHS = A.getAssumedSimplified(
9331 IRPosition::value(*RHS, getCallBaseContext()), *this,
9332 UsedAssumedInformation, AA::Interprocedural);
9333 if (!SimplifiedRHS.has_value())
9334 return true;
9335 if (!*SimplifiedRHS)
9336 return false;
9337 RHS = *SimplifiedRHS;
9338
9339 // TODO: Allow non integers as well.
9340 if (!LHS->getType()->isIntegerTy() || !RHS->getType()->isIntegerTy())
9341 return false;
9342
9343 auto *LHSAA = A.getAAFor<AAValueConstantRange>(
9344 *this, IRPosition::value(*LHS, getCallBaseContext()),
9345 DepClassTy::REQUIRED);
9346 if (!LHSAA)
9347 return false;
9348 QuerriedAAs.push_back(LHSAA);
9349 auto LHSAARange = LHSAA->getAssumedConstantRange(A, CtxI);
9350
9351 auto *RHSAA = A.getAAFor<AAValueConstantRange>(
9352 *this, IRPosition::value(*RHS, getCallBaseContext()),
9353 DepClassTy::REQUIRED);
9354 if (!RHSAA)
9355 return false;
9356 QuerriedAAs.push_back(RHSAA);
9357 auto RHSAARange = RHSAA->getAssumedConstantRange(A, CtxI);
9358
9359 auto AssumedRange = LHSAARange.binaryOp(BinOp->getOpcode(), RHSAARange);
9360
9361 T.unionAssumed(AssumedRange);
9362
9363 // TODO: Track a known state too.
9364
9365 return T.isValidState();
9366 }
9367
9368 bool calculateCastInst(
9370 const Instruction *CtxI,
9372 assert(CastI->getNumOperands() == 1 && "Expected cast to be unary!");
9373 // TODO: Allow non integers as well.
9374 Value *OpV = CastI->getOperand(0);
9375
9376 // Simplify the operand first.
9377 bool UsedAssumedInformation = false;
9378 const auto &SimplifiedOpV = A.getAssumedSimplified(
9379 IRPosition::value(*OpV, getCallBaseContext()), *this,
9380 UsedAssumedInformation, AA::Interprocedural);
9381 if (!SimplifiedOpV.has_value())
9382 return true;
9383 if (!*SimplifiedOpV)
9384 return false;
9385 OpV = *SimplifiedOpV;
9386
9387 if (!OpV->getType()->isIntegerTy())
9388 return false;
9389
9390 auto *OpAA = A.getAAFor<AAValueConstantRange>(
9391 *this, IRPosition::value(*OpV, getCallBaseContext()),
9392 DepClassTy::REQUIRED);
9393 if (!OpAA)
9394 return false;
9395 QuerriedAAs.push_back(OpAA);
9396 T.unionAssumed(OpAA->getAssumed().castOp(CastI->getOpcode(),
9397 getState().getBitWidth()));
9398 return T.isValidState();
9399 }
9400
9401 bool
9402 calculateCmpInst(Attributor &A, CmpInst *CmpI, IntegerRangeState &T,
9403 const Instruction *CtxI,
9405 Value *LHS = CmpI->getOperand(0);
9406 Value *RHS = CmpI->getOperand(1);
9407
9408 // Simplify the operands first.
9409 bool UsedAssumedInformation = false;
9410 const auto &SimplifiedLHS = A.getAssumedSimplified(
9411 IRPosition::value(*LHS, getCallBaseContext()), *this,
9412 UsedAssumedInformation, AA::Interprocedural);
9413 if (!SimplifiedLHS.has_value())
9414 return true;
9415 if (!*SimplifiedLHS)
9416 return false;
9417 LHS = *SimplifiedLHS;
9418
9419 const auto &SimplifiedRHS = A.getAssumedSimplified(
9420 IRPosition::value(*RHS, getCallBaseContext()), *this,
9421 UsedAssumedInformation, AA::Interprocedural);
9422 if (!SimplifiedRHS.has_value())
9423 return true;
9424 if (!*SimplifiedRHS)
9425 return false;
9426 RHS = *SimplifiedRHS;
9427
9428 // TODO: Allow non integers as well.
9429 if (!LHS->getType()->isIntegerTy() || !RHS->getType()->isIntegerTy())
9430 return false;
9431
9432 auto *LHSAA = A.getAAFor<AAValueConstantRange>(
9433 *this, IRPosition::value(*LHS, getCallBaseContext()),
9434 DepClassTy::REQUIRED);
9435 if (!LHSAA)
9436 return false;
9437 QuerriedAAs.push_back(LHSAA);
9438 auto *RHSAA = A.getAAFor<AAValueConstantRange>(
9439 *this, IRPosition::value(*RHS, getCallBaseContext()),
9440 DepClassTy::REQUIRED);
9441 if (!RHSAA)
9442 return false;
9443 QuerriedAAs.push_back(RHSAA);
9444 auto LHSAARange = LHSAA->getAssumedConstantRange(A, CtxI);
9445 auto RHSAARange = RHSAA->getAssumedConstantRange(A, CtxI);
9446
9447 // If one of them is empty set, we can't decide.
9448 if (LHSAARange.isEmptySet() || RHSAARange.isEmptySet())
9449 return true;
9450
9451 bool MustTrue = false, MustFalse = false;
9452
9453 auto AllowedRegion =
9455
9456 if (AllowedRegion.intersectWith(LHSAARange).isEmptySet())
9457 MustFalse = true;
9458
9459 if (LHSAARange.icmp(CmpI->getPredicate(), RHSAARange))
9460 MustTrue = true;
9461
9462 assert((!MustTrue || !MustFalse) &&
9463 "Either MustTrue or MustFalse should be false!");
9464
9465 if (MustTrue)
9466 T.unionAssumed(ConstantRange(APInt(/* numBits */ 1, /* val */ 1)));
9467 else if (MustFalse)
9468 T.unionAssumed(ConstantRange(APInt(/* numBits */ 1, /* val */ 0)));
9469 else
9470 T.unionAssumed(ConstantRange(/* BitWidth */ 1, /* isFullSet */ true));
9471
9472 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] " << *CmpI << " after "
9473 << (MustTrue ? "true" : (MustFalse ? "false" : "unknown"))
9474 << ": " << T << "\n\t" << *LHSAA << "\t<op>\n\t"
9475 << *RHSAA);
9476
9477 // TODO: Track a known state too.
9478 return T.isValidState();
9479 }
9480
9481 /// See AbstractAttribute::updateImpl(...).
9482 ChangeStatus updateImpl(Attributor &A) override {
9483
9485 auto VisitValueCB = [&](Value &V, const Instruction *CtxI) -> bool {
9486 Instruction *I = dyn_cast<Instruction>(&V);
9487 if (!I || isa<CallBase>(I)) {
9488
9489 // Simplify the operand first.
9490 bool UsedAssumedInformation = false;
9491 const auto &SimplifiedOpV = A.getAssumedSimplified(
9492 IRPosition::value(V, getCallBaseContext()), *this,
9493 UsedAssumedInformation, AA::Interprocedural);
9494 if (!SimplifiedOpV.has_value())
9495 return true;
9496 if (!*SimplifiedOpV)
9497 return false;
9498 Value *VPtr = *SimplifiedOpV;
9499
9500 // If the value is not instruction, we query AA to Attributor.
9501 const auto *AA = A.getAAFor<AAValueConstantRange>(
9502 *this, IRPosition::value(*VPtr, getCallBaseContext()),
9503 DepClassTy::REQUIRED);
9504
9505 // Clamp operator is not used to utilize a program point CtxI.
9506 if (AA)
9507 T.unionAssumed(AA->getAssumedConstantRange(A, CtxI));
9508 else
9509 return false;
9510
9511 return T.isValidState();
9512 }
9513
9515 if (auto *BinOp = dyn_cast<BinaryOperator>(I)) {
9516 if (!calculateBinaryOperator(A, BinOp, T, CtxI, QuerriedAAs))
9517 return false;
9518 } else if (auto *CmpI = dyn_cast<CmpInst>(I)) {
9519 if (!calculateCmpInst(A, CmpI, T, CtxI, QuerriedAAs))
9520 return false;
9521 } else if (auto *CastI = dyn_cast<CastInst>(I)) {
9522 if (!calculateCastInst(A, CastI, T, CtxI, QuerriedAAs))
9523 return false;
9524 } else {
9525 // Give up with other instructions.
9526 // TODO: Add other instructions
9527
9528 T.indicatePessimisticFixpoint();
9529 return false;
9530 }
9531
9532 // Catch circular reasoning in a pessimistic way for now.
9533 // TODO: Check how the range evolves and if we stripped anything, see also
9534 // AADereferenceable or AAAlign for similar situations.
9535 for (const AAValueConstantRange *QueriedAA : QuerriedAAs) {
9536 if (QueriedAA != this)
9537 continue;
9538 // If we are in a stady state we do not need to worry.
9539 if (T.getAssumed() == getState().getAssumed())
9540 continue;
9541 T.indicatePessimisticFixpoint();
9542 }
9543
9544 return T.isValidState();
9545 };
9546
9547 if (!VisitValueCB(getAssociatedValue(), getCtxI()))
9548 return indicatePessimisticFixpoint();
9549
9550 // Ensure that long def-use chains can't cause circular reasoning either by
9551 // introducing a cutoff below.
9552 if (clampStateAndIndicateChange(getState(), T) == ChangeStatus::UNCHANGED)
9553 return ChangeStatus::UNCHANGED;
9554 if (++NumChanges > MaxNumChanges) {
9555 LLVM_DEBUG(dbgs() << "[AAValueConstantRange] performed " << NumChanges
9556 << " but only " << MaxNumChanges
9557 << " are allowed to avoid cyclic reasoning.");
9558 return indicatePessimisticFixpoint();
9559 }
9560 return ChangeStatus::CHANGED;
9561 }
9562
9563 /// See AbstractAttribute::trackStatistics()
9564 void trackStatistics() const override {
9566 }
9567
9568 /// Tracker to bail after too many widening steps of the constant range.
9569 int NumChanges = 0;
9570
9571 /// Upper bound for the number of allowed changes (=widening steps) for the
9572 /// constant range before we give up.
9573 static constexpr int MaxNumChanges = 5;
9574};
9575
9576struct AAValueConstantRangeFunction : AAValueConstantRangeImpl {
9577 AAValueConstantRangeFunction(const IRPosition &IRP, Attributor &A)
9578 : AAValueConstantRangeImpl(IRP, A) {}
9579
9580 /// See AbstractAttribute::initialize(...).
9581 ChangeStatus updateImpl(Attributor &A) override {
9582 llvm_unreachable("AAValueConstantRange(Function|CallSite)::updateImpl will "
9583 "not be called");
9584 }
9585
9586 /// See AbstractAttribute::trackStatistics()
9587 void trackStatistics() const override { STATS_DECLTRACK_FN_ATTR(value_range) }
9588};
9589
9590struct AAValueConstantRangeCallSite : AAValueConstantRangeFunction {
9591 AAValueConstantRangeCallSite(const IRPosition &IRP, Attributor &A)
9592 : AAValueConstantRangeFunction(IRP, A) {}
9593
9594 /// See AbstractAttribute::trackStatistics()
9595 void trackStatistics() const override { STATS_DECLTRACK_CS_ATTR(value_range) }
9596};
9597
9598struct AAValueConstantRangeCallSiteReturned
9599 : AACalleeToCallSite<AAValueConstantRange, AAValueConstantRangeImpl,
9600 AAValueConstantRangeImpl::StateType,
9601 /* IntroduceCallBaseContext */ true> {
9602 AAValueConstantRangeCallSiteReturned(const IRPosition &IRP, Attributor &A)
9603 : AACalleeToCallSite<AAValueConstantRange, AAValueConstantRangeImpl,
9604 AAValueConstantRangeImpl::StateType,
9605 /* IntroduceCallBaseContext */ true>(IRP, A) {}
9606
9607 /// See AbstractAttribute::initialize(...).
9608 void initialize(Attributor &A) override {
9609 // If it is a load instruction with range metadata, use the metadata.
9610 if (CallInst *CI = dyn_cast<CallInst>(&getAssociatedValue()))
9611 if (auto *RangeMD = CI->getMetadata(LLVMContext::MD_range))
9612 intersectKnown(getConstantRangeFromMetadata(*RangeMD));
9613
9614 AAValueConstantRangeImpl::initialize(A);
9615 }
9616
9617 /// See AbstractAttribute::trackStatistics()
9618 void trackStatistics() const override {
9619 STATS_DECLTRACK_CSRET_ATTR(value_range)
9620 }
9621};
9622struct AAValueConstantRangeCallSiteArgument : AAValueConstantRangeFloating {
9623 AAValueConstantRangeCallSiteArgument(const IRPosition &IRP, Attributor &A)
9624 : AAValueConstantRangeFloating(IRP, A) {}
9625
9626 /// See AbstractAttribute::manifest()
9627 ChangeStatus manifest(Attributor &A) override {
9628 return ChangeStatus::UNCHANGED;
9629 }
9630
9631 /// See AbstractAttribute::trackStatistics()
9632 void trackStatistics() const override {
9633 STATS_DECLTRACK_CSARG_ATTR(value_range)
9634 }
9635};
9636} // namespace
9637
9638/// ------------------ Potential Values Attribute -------------------------
9639
9640namespace {
9641struct AAPotentialConstantValuesImpl : AAPotentialConstantValues {
9642 using StateType = PotentialConstantIntValuesState;
9643
9644 AAPotentialConstantValuesImpl(const IRPosition &IRP, Attributor &A)
9645 : AAPotentialConstantValues(IRP, A) {}
9646
9647 /// See AbstractAttribute::initialize(..).
9648 void initialize(Attributor &A) override {
9649 if (A.hasSimplificationCallback(getIRPosition()))
9650 indicatePessimisticFixpoint();
9651 else
9652 AAPotentialConstantValues::initialize(A);
9653 }
9654
9655 bool fillSetWithConstantValues(Attributor &A, const IRPosition &IRP, SetTy &S,
9656 bool &ContainsUndef, bool ForSelf) {
9658 bool UsedAssumedInformation = false;
9659 if (!A.getAssumedSimplifiedValues(IRP, *this, Values, AA::Interprocedural,
9660 UsedAssumedInformation)) {
9661 // Avoid recursion when the caller is computing constant values for this
9662 // IRP itself.
9663 if (ForSelf)
9664 return false;
9665 if (!IRP.getAssociatedType()->isIntegerTy())
9666 return false;
9667 auto *PotentialValuesAA = A.getAAFor<AAPotentialConstantValues>(
9668 *this, IRP, DepClassTy::REQUIRED);
9669 if (!PotentialValuesAA || !PotentialValuesAA->getState().isValidState())
9670 return false;
9671 ContainsUndef = PotentialValuesAA->getState().undefIsContained();
9672 S = PotentialValuesAA->getState().getAssumedSet();
9673 return true;
9674 }
9675
9676 // Copy all the constant values, except UndefValue. ContainsUndef is true
9677 // iff Values contains only UndefValue instances. If there are other known
9678 // constants, then UndefValue is dropped.
9679 ContainsUndef = false;
9680 for (auto &It : Values) {
9681 if (isa<UndefValue>(It.getValue())) {
9682 ContainsUndef = true;
9683 continue;
9684 }
9685 auto *CI = dyn_cast<ConstantInt>(It.getValue());
9686 if (!CI)
9687 return false;
9688 S.insert(CI->getValue());
9689 }
9690 ContainsUndef &= S.empty();
9691
9692 return true;
9693 }
9694
9695 /// See AbstractAttribute::getAsStr().
9696 const std::string getAsStr(Attributor *A) const override {
9697 std::string Str;
9699 OS << getState();
9700 return Str;
9701 }
9702
9703 /// See AbstractAttribute::updateImpl(...).
9704 ChangeStatus updateImpl(Attributor &A) override {
9705 return indicatePessimisticFixpoint();
9706 }
9707};
9708
9709struct AAPotentialConstantValuesArgument final
9710 : AAArgumentFromCallSiteArguments<AAPotentialConstantValues,
9711 AAPotentialConstantValuesImpl,
9712 PotentialConstantIntValuesState> {
9713 using Base = AAArgumentFromCallSiteArguments<AAPotentialConstantValues,
9714 AAPotentialConstantValuesImpl,
9716 AAPotentialConstantValuesArgument(const IRPosition &IRP, Attributor &A)
9717 : Base(IRP, A) {}
9718
9719 /// See AbstractAttribute::trackStatistics()
9720 void trackStatistics() const override {
9721 STATS_DECLTRACK_ARG_ATTR(potential_values)
9722 }
9723};
9724
9725struct AAPotentialConstantValuesReturned
9726 : AAReturnedFromReturnedValues<AAPotentialConstantValues,
9727 AAPotentialConstantValuesImpl> {
9728 using Base = AAReturnedFromReturnedValues<AAPotentialConstantValues,
9729 AAPotentialConstantValuesImpl>;
9730 AAPotentialConstantValuesReturned(const IRPosition &IRP, Attributor &A)
9731 : Base(IRP, A) {}
9732
9733 void initialize(Attributor &A) override {
9734 if (!A.isFunctionIPOAmendable(*getAssociatedFunction()))
9735 indicatePessimisticFixpoint();
9736 Base::initialize(A);
9737 }
9738
9739 /// See AbstractAttribute::trackStatistics()
9740 void trackStatistics() const override {
9741 STATS_DECLTRACK_FNRET_ATTR(potential_values)
9742 }
9743};
9744
9745struct AAPotentialConstantValuesFloating : AAPotentialConstantValuesImpl {
9746 AAPotentialConstantValuesFloating(const IRPosition &IRP, Attributor &A)
9747 : AAPotentialConstantValuesImpl(IRP, A) {}
9748
9749 /// See AbstractAttribute::initialize(..).
9750 void initialize(Attributor &A) override {
9751 AAPotentialConstantValuesImpl::initialize(A);
9752 if (isAtFixpoint())
9753 return;
9754
9755 Value &V = getAssociatedValue();
9756
9757 if (auto *C = dyn_cast<ConstantInt>(&V)) {
9758 unionAssumed(C->getValue());
9759 indicateOptimisticFixpoint();
9760 return;
9761 }
9762
9763 if (isa<UndefValue>(&V)) {
9764 unionAssumedWithUndef();
9765 indicateOptimisticFixpoint();
9766 return;
9767 }
9768
9769 if (isa<BinaryOperator>(&V) || isa<ICmpInst>(&V) || isa<CastInst>(&V))
9770 return;
9771
9772 if (isa<SelectInst>(V) || isa<PHINode>(V) || isa<LoadInst>(V))
9773 return;
9774
9775 indicatePessimisticFixpoint();
9776
9777 LLVM_DEBUG(dbgs() << "[AAPotentialConstantValues] We give up: "
9778 << getAssociatedValue() << "\n");
9779 }
9780
9781 static bool calculateICmpInst(const ICmpInst *ICI, const APInt &LHS,
9782 const APInt &RHS) {
9783 return ICmpInst::compare(LHS, RHS, ICI->getPredicate());
9784 }
9785
9786 static APInt calculateCastInst(const CastInst *CI, const APInt &Src,
9787 uint32_t ResultBitWidth) {
9788 Instruction::CastOps CastOp = CI->getOpcode();
9789 switch (CastOp) {
9790 default:
9791 llvm_unreachable("unsupported or not integer cast");
9792 case Instruction::Trunc:
9793 return Src.trunc(ResultBitWidth);
9794 case Instruction::SExt:
9795 return Src.sext(ResultBitWidth);
9796 case Instruction::ZExt:
9797 return Src.zext(ResultBitWidth);
9798 case Instruction::BitCast:
9799 return Src;
9800 }
9801 }
9802
9803 static APInt calculateBinaryOperator(const BinaryOperator *BinOp,
9804 const APInt &LHS, const APInt &RHS,
9805 bool &SkipOperation, bool &Unsupported) {
9806 Instruction::BinaryOps BinOpcode = BinOp->getOpcode();
9807 // Unsupported is set to true when the binary operator is not supported.
9808 // SkipOperation is set to true when UB occur with the given operand pair
9809 // (LHS, RHS).
9810 // TODO: we should look at nsw and nuw keywords to handle operations
9811 // that create poison or undef value.
9812 switch (BinOpcode) {
9813 default:
9814 Unsupported = true;
9815 return LHS;
9816 case Instruction::Add:
9817 return LHS + RHS;
9818 case Instruction::Sub:
9819 return LHS - RHS;
9820 case Instruction::Mul:
9821 return LHS * RHS;
9822 case Instruction::UDiv:
9823 if (RHS.isZero()) {
9824 SkipOperation = true;
9825 return LHS;
9826 }
9827 return LHS.udiv(RHS);
9828 case Instruction::SDiv:
9829 if (RHS.isZero()) {
9830 SkipOperation = true;
9831 return LHS;
9832 }
9833 return LHS.sdiv(RHS);
9834 case Instruction::URem:
9835 if (RHS.isZero()) {
9836 SkipOperation = true;
9837 return LHS;
9838 }
9839 return LHS.urem(RHS);
9840 case Instruction::SRem:
9841 if (RHS.isZero()) {
9842 SkipOperation = true;
9843 return LHS;
9844 }
9845 return LHS.srem(RHS);
9846 case Instruction::Shl:
9847 return LHS.shl(RHS);
9848 case Instruction::LShr:
9849 return LHS.lshr(RHS);
9850 case Instruction::AShr:
9851 return LHS.ashr(RHS);
9852 case Instruction::And:
9853 return LHS & RHS;
9854 case Instruction::Or:
9855 return LHS | RHS;
9856 case Instruction::Xor:
9857 return LHS ^ RHS;
9858 }
9859 }
9860
9861 bool calculateBinaryOperatorAndTakeUnion(const BinaryOperator *BinOp,
9862 const APInt &LHS, const APInt &RHS) {
9863 bool SkipOperation = false;
9864 bool Unsupported = false;
9865 APInt Result =
9866 calculateBinaryOperator(BinOp, LHS, RHS, SkipOperation, Unsupported);
9867 if (Unsupported)
9868 return false;
9869 // If SkipOperation is true, we can ignore this operand pair (L, R).
9870 if (!SkipOperation)
9871 unionAssumed(Result);
9872 return isValidState();
9873 }
9874
9875 ChangeStatus updateWithICmpInst(Attributor &A, ICmpInst *ICI) {
9876 auto AssumedBefore = getAssumed();
9877 Value *LHS = ICI->getOperand(0);
9878 Value *RHS = ICI->getOperand(1);
9879
9880 bool LHSContainsUndef = false, RHSContainsUndef = false;
9881 SetTy LHSAAPVS, RHSAAPVS;
9882 if (!fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
9883 LHSContainsUndef, /* ForSelf */ false) ||
9884 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
9885 RHSContainsUndef, /* ForSelf */ false))
9886 return indicatePessimisticFixpoint();
9887
9888 // TODO: make use of undef flag to limit potential values aggressively.
9889 bool MaybeTrue = false, MaybeFalse = false;
9890 const APInt Zero(RHS->getType()->getIntegerBitWidth(), 0);
9891 if (LHSContainsUndef && RHSContainsUndef) {
9892 // The result of any comparison between undefs can be soundly replaced
9893 // with undef.
9894 unionAssumedWithUndef();
9895 } else if (LHSContainsUndef) {
9896 for (const APInt &R : RHSAAPVS) {
9897 bool CmpResult = calculateICmpInst(ICI, Zero, R);
9898 MaybeTrue |= CmpResult;
9899 MaybeFalse |= !CmpResult;
9900 if (MaybeTrue & MaybeFalse)
9901 return indicatePessimisticFixpoint();
9902 }
9903 } else if (RHSContainsUndef) {
9904 for (const APInt &L : LHSAAPVS) {
9905 bool CmpResult = calculateICmpInst(ICI, L, Zero);
9906 MaybeTrue |= CmpResult;
9907 MaybeFalse |= !CmpResult;
9908 if (MaybeTrue & MaybeFalse)
9909 return indicatePessimisticFixpoint();
9910 }
9911 } else {
9912 for (const APInt &L : LHSAAPVS) {
9913 for (const APInt &R : RHSAAPVS) {
9914 bool CmpResult = calculateICmpInst(ICI, L, R);
9915 MaybeTrue |= CmpResult;
9916 MaybeFalse |= !CmpResult;
9917 if (MaybeTrue & MaybeFalse)
9918 return indicatePessimisticFixpoint();
9919 }
9920 }
9921 }
9922 if (MaybeTrue)
9923 unionAssumed(APInt(/* numBits */ 1, /* val */ 1));
9924 if (MaybeFalse)
9925 unionAssumed(APInt(/* numBits */ 1, /* val */ 0));
9926 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
9927 : ChangeStatus::CHANGED;
9928 }
9929
9930 ChangeStatus updateWithSelectInst(Attributor &A, SelectInst *SI) {
9931 auto AssumedBefore = getAssumed();
9932 Value *LHS = SI->getTrueValue();
9933 Value *RHS = SI->getFalseValue();
9934
9935 bool UsedAssumedInformation = false;
9936 std::optional<Constant *> C = A.getAssumedConstant(
9937 *SI->getCondition(), *this, UsedAssumedInformation);
9938
9939 // Check if we only need one operand.
9940 bool OnlyLeft = false, OnlyRight = false;
9941 if (C && *C && (*C)->isOneValue())
9942 OnlyLeft = true;
9943 else if (C && *C && (*C)->isZeroValue())
9944 OnlyRight = true;
9945
9946 bool LHSContainsUndef = false, RHSContainsUndef = false;
9947 SetTy LHSAAPVS, RHSAAPVS;
9948 if (!OnlyRight &&
9949 !fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
9950 LHSContainsUndef, /* ForSelf */ false))
9951 return indicatePessimisticFixpoint();
9952
9953 if (!OnlyLeft &&
9954 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
9955 RHSContainsUndef, /* ForSelf */ false))
9956 return indicatePessimisticFixpoint();
9957
9958 if (OnlyLeft || OnlyRight) {
9959 // select (true/false), lhs, rhs
9960 auto *OpAA = OnlyLeft ? &LHSAAPVS : &RHSAAPVS;
9961 auto Undef = OnlyLeft ? LHSContainsUndef : RHSContainsUndef;
9962
9963 if (Undef)
9964 unionAssumedWithUndef();
9965 else {
9966 for (const auto &It : *OpAA)
9967 unionAssumed(It);
9968 }
9969
9970 } else if (LHSContainsUndef && RHSContainsUndef) {
9971 // select i1 *, undef , undef => undef
9972 unionAssumedWithUndef();
9973 } else {
9974 for (const auto &It : LHSAAPVS)
9975 unionAssumed(It);
9976 for (const auto &It : RHSAAPVS)
9977 unionAssumed(It);
9978 }
9979 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
9980 : ChangeStatus::CHANGED;
9981 }
9982
9983 ChangeStatus updateWithCastInst(Attributor &A, CastInst *CI) {
9984 auto AssumedBefore = getAssumed();
9985 if (!CI->isIntegerCast())
9986 return indicatePessimisticFixpoint();
9987 assert(CI->getNumOperands() == 1 && "Expected cast to be unary!");
9988 uint32_t ResultBitWidth = CI->getDestTy()->getIntegerBitWidth();
9989 Value *Src = CI->getOperand(0);
9990
9991 bool SrcContainsUndef = false;
9992 SetTy SrcPVS;
9993 if (!fillSetWithConstantValues(A, IRPosition::value(*Src), SrcPVS,
9994 SrcContainsUndef, /* ForSelf */ false))
9995 return indicatePessimisticFixpoint();
9996
9997 if (SrcContainsUndef)
9998 unionAssumedWithUndef();
9999 else {
10000 for (const APInt &S : SrcPVS) {
10001 APInt T = calculateCastInst(CI, S, ResultBitWidth);
10002 unionAssumed(T);
10003 }
10004 }
10005 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10006 : ChangeStatus::CHANGED;
10007 }
10008
10009 ChangeStatus updateWithBinaryOperator(Attributor &A, BinaryOperator *BinOp) {
10010 auto AssumedBefore = getAssumed();
10011 Value *LHS = BinOp->getOperand(0);
10012 Value *RHS = BinOp->getOperand(1);
10013
10014 bool LHSContainsUndef = false, RHSContainsUndef = false;
10015 SetTy LHSAAPVS, RHSAAPVS;
10016 if (!fillSetWithConstantValues(A, IRPosition::value(*LHS), LHSAAPVS,
10017 LHSContainsUndef, /* ForSelf */ false) ||
10018 !fillSetWithConstantValues(A, IRPosition::value(*RHS), RHSAAPVS,
10019 RHSContainsUndef, /* ForSelf */ false))
10020 return indicatePessimisticFixpoint();
10021
10022 const APInt Zero = APInt(LHS->getType()->getIntegerBitWidth(), 0);
10023
10024 // TODO: make use of undef flag to limit potential values aggressively.
10025 if (LHSContainsUndef && RHSContainsUndef) {
10026 if (!calculateBinaryOperatorAndTakeUnion(BinOp, Zero, Zero))
10027 return indicatePessimisticFixpoint();
10028 } else if (LHSContainsUndef) {
10029 for (const APInt &R : RHSAAPVS) {
10030 if (!calculateBinaryOperatorAndTakeUnion(BinOp, Zero, R))
10031 return indicatePessimisticFixpoint();
10032 }
10033 } else if (RHSContainsUndef) {
10034 for (const APInt &L : LHSAAPVS) {
10035 if (!calculateBinaryOperatorAndTakeUnion(BinOp, L, Zero))
10036 return indicatePessimisticFixpoint();
10037 }
10038 } else {
10039 for (const APInt &L : LHSAAPVS) {
10040 for (const APInt &R : RHSAAPVS) {
10041 if (!calculateBinaryOperatorAndTakeUnion(BinOp, L, R))
10042 return indicatePessimisticFixpoint();
10043 }
10044 }
10045 }
10046 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10047 : ChangeStatus::CHANGED;
10048 }
10049
10050 ChangeStatus updateWithInstruction(Attributor &A, Instruction *Inst) {
10051 auto AssumedBefore = getAssumed();
10052 SetTy Incoming;
10053 bool ContainsUndef;
10054 if (!fillSetWithConstantValues(A, IRPosition::value(*Inst), Incoming,
10055 ContainsUndef, /* ForSelf */ true))
10056 return indicatePessimisticFixpoint();
10057 if (ContainsUndef) {
10058 unionAssumedWithUndef();
10059 } else {
10060 for (const auto &It : Incoming)
10061 unionAssumed(It);
10062 }
10063 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10064 : ChangeStatus::CHANGED;
10065 }
10066
10067 /// See AbstractAttribute::updateImpl(...).
10068 ChangeStatus updateImpl(Attributor &A) override {
10069 Value &V = getAssociatedValue();
10070 Instruction *I = dyn_cast<Instruction>(&V);
10071
10072 if (auto *ICI = dyn_cast<ICmpInst>(I))
10073 return updateWithICmpInst(A, ICI);
10074
10075 if (auto *SI = dyn_cast<SelectInst>(I))
10076 return updateWithSelectInst(A, SI);
10077
10078 if (auto *CI = dyn_cast<CastInst>(I))
10079 return updateWithCastInst(A, CI);
10080
10081 if (auto *BinOp = dyn_cast<BinaryOperator>(I))
10082 return updateWithBinaryOperator(A, BinOp);
10083
10084 if (isa<PHINode>(I) || isa<LoadInst>(I))
10085 return updateWithInstruction(A, I);
10086
10087 return indicatePessimisticFixpoint();
10088 }
10089
10090 /// See AbstractAttribute::trackStatistics()
10091 void trackStatistics() const override {
10092 STATS_DECLTRACK_FLOATING_ATTR(potential_values)
10093 }
10094};
10095
10096struct AAPotentialConstantValuesFunction : AAPotentialConstantValuesImpl {
10097 AAPotentialConstantValuesFunction(const IRPosition &IRP, Attributor &A)
10098 : AAPotentialConstantValuesImpl(IRP, A) {}
10099
10100 /// See AbstractAttribute::initialize(...).
10101 ChangeStatus updateImpl(Attributor &A) override {
10103 "AAPotentialConstantValues(Function|CallSite)::updateImpl will "
10104 "not be called");
10105 }
10106
10107 /// See AbstractAttribute::trackStatistics()
10108 void trackStatistics() const override {
10109 STATS_DECLTRACK_FN_ATTR(potential_values)
10110 }
10111};
10112
10113struct AAPotentialConstantValuesCallSite : AAPotentialConstantValuesFunction {
10114 AAPotentialConstantValuesCallSite(const IRPosition &IRP, Attributor &A)
10115 : AAPotentialConstantValuesFunction(IRP, A) {}
10116
10117 /// See AbstractAttribute::trackStatistics()
10118 void trackStatistics() const override {
10119 STATS_DECLTRACK_CS_ATTR(potential_values)
10120 }
10121};
10122
10123struct AAPotentialConstantValuesCallSiteReturned
10124 : AACalleeToCallSite<AAPotentialConstantValues,
10125 AAPotentialConstantValuesImpl> {
10126 AAPotentialConstantValuesCallSiteReturned(const IRPosition &IRP,
10127 Attributor &A)
10128 : AACalleeToCallSite<AAPotentialConstantValues,
10129 AAPotentialConstantValuesImpl>(IRP, A) {}
10130
10131 /// See AbstractAttribute::trackStatistics()
10132 void trackStatistics() const override {
10133 STATS_DECLTRACK_CSRET_ATTR(potential_values)
10134 }
10135};
10136
10137struct AAPotentialConstantValuesCallSiteArgument
10138 : AAPotentialConstantValuesFloating {
10139 AAPotentialConstantValuesCallSiteArgument(const IRPosition &IRP,
10140 Attributor &A)
10141 : AAPotentialConstantValuesFloating(IRP, A) {}
10142
10143 /// See AbstractAttribute::initialize(..).
10144 void initialize(Attributor &A) override {
10145 AAPotentialConstantValuesImpl::initialize(A);
10146 if (isAtFixpoint())
10147 return;
10148
10149 Value &V = getAssociatedValue();
10150
10151 if (auto *C = dyn_cast<ConstantInt>(&V)) {
10152 unionAssumed(C->getValue());
10153 indicateOptimisticFixpoint();
10154 return;
10155 }
10156
10157 if (isa<UndefValue>(&V)) {
10158 unionAssumedWithUndef();
10159 indicateOptimisticFixpoint();
10160 return;
10161 }
10162 }
10163
10164 /// See AbstractAttribute::updateImpl(...).
10165 ChangeStatus updateImpl(Attributor &A) override {
10166 Value &V = getAssociatedValue();
10167 auto AssumedBefore = getAssumed();
10168 auto *AA = A.getAAFor<AAPotentialConstantValues>(
10169 *this, IRPosition::value(V), DepClassTy::REQUIRED);
10170 if (!AA)
10171 return indicatePessimisticFixpoint();
10172 const auto &S = AA->getAssumed();
10173 unionAssumed(S);
10174 return AssumedBefore == getAssumed() ? ChangeStatus::UNCHANGED
10175 : ChangeStatus::CHANGED;
10176 }
10177
10178 /// See AbstractAttribute::trackStatistics()
10179 void trackStatistics() const override {
10180 STATS_DECLTRACK_CSARG_ATTR(potential_values)
10181 }
10182};
10183} // namespace
10184
10185/// ------------------------ NoUndef Attribute ---------------------------------
10187 Attribute::AttrKind ImpliedAttributeKind,
10188 bool IgnoreSubsumingPositions) {
10189 assert(ImpliedAttributeKind == Attribute::NoUndef &&
10190 "Unexpected attribute kind");
10191 if (A.hasAttr(IRP, {Attribute::NoUndef}, IgnoreSubsumingPositions,
10192 Attribute::NoUndef))
10193 return true;
10194
10195 Value &Val = IRP.getAssociatedValue();
10198 LLVMContext &Ctx = Val.getContext();
10199 A.manifestAttrs(IRP, Attribute::get(Ctx, Attribute::NoUndef));
10200 return true;
10201 }
10202
10203 return false;
10204}
10205
10206namespace {
10207struct AANoUndefImpl : AANoUndef {
10208 AANoUndefImpl(const IRPosition &IRP, Attributor &A) : AANoUndef(IRP, A) {}
10209
10210 /// See AbstractAttribute::initialize(...).
10211 void initialize(Attributor &A) override {
10212 Value &V = getAssociatedValue();
10213 if (isa<UndefValue>(V))
10214 indicatePessimisticFixpoint();
10215 assert(!isImpliedByIR(A, getIRPosition(), Attribute::NoUndef));
10216 }
10217
10218 /// See followUsesInMBEC
10219 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
10220 AANoUndef::StateType &State) {
10221 const Value *UseV = U->get();
10222 const DominatorTree *DT = nullptr;
10223 AssumptionCache *AC = nullptr;
10224 InformationCache &InfoCache = A.getInfoCache();
10225 if (Function *F = getAnchorScope()) {
10228 }
10229 State.setKnown(isGuaranteedNotToBeUndefOrPoison(UseV, AC, I, DT));
10230 bool TrackUse = false;
10231 // Track use for instructions which must produce undef or poison bits when
10232 // at least one operand contains such bits.
10233 if (isa<CastInst>(*I) || isa<GetElementPtrInst>(*I))
10234 TrackUse = true;
10235 return TrackUse;
10236 }
10237
10238 /// See AbstractAttribute::getAsStr().
10239 const std::string getAsStr(Attributor *A) const override {
10240 return getAssumed() ? "noundef" : "may-undef-or-poison";
10241 }
10242
10243 ChangeStatus manifest(Attributor &A) override {
10244 // We don't manifest noundef attribute for dead positions because the
10245 // associated values with dead positions would be replaced with undef
10246 // values.
10247 bool UsedAssumedInformation = false;
10248 if (A.isAssumedDead(getIRPosition(), nullptr, nullptr,
10249 UsedAssumedInformation))
10250 return ChangeStatus::UNCHANGED;
10251 // A position whose simplified value does not have any value is
10252 // considered to be dead. We don't manifest noundef in such positions for
10253 // the same reason above.
10254 if (!A.getAssumedSimplified(getIRPosition(), *this, UsedAssumedInformation,
10256 .has_value())
10257 return ChangeStatus::UNCHANGED;
10258 return AANoUndef::manifest(A);
10259 }
10260};
10261
10262struct AANoUndefFloating : public AANoUndefImpl {
10263 AANoUndefFloating(const IRPosition &IRP, Attributor &A)
10264 : AANoUndefImpl(IRP, A) {}
10265
10266 /// See AbstractAttribute::initialize(...).
10267 void initialize(Attributor &A) override {
10268 AANoUndefImpl::initialize(A);
10269 if (!getState().isAtFixpoint() && getAnchorScope() &&
10270 !getAnchorScope()->isDeclaration())
10271 if (Instruction *CtxI = getCtxI())
10272 followUsesInMBEC(*this, A, getState(), *CtxI);
10273 }
10274
10275 /// See AbstractAttribute::updateImpl(...).
10276 ChangeStatus updateImpl(Attributor &A) override {
10277 auto VisitValueCB = [&](const IRPosition &IRP) -> bool {
10278 bool IsKnownNoUndef;
10279 return AA::hasAssumedIRAttr<Attribute::NoUndef>(
10280 A, this, IRP, DepClassTy::REQUIRED, IsKnownNoUndef);
10281 };
10282
10283 bool Stripped;
10284 bool UsedAssumedInformation = false;
10285 Value *AssociatedValue = &getAssociatedValue();
10287 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
10288 AA::AnyScope, UsedAssumedInformation))
10289 Stripped = false;
10290 else
10291 Stripped =
10292 Values.size() != 1 || Values.front().getValue() != AssociatedValue;
10293
10294 if (!Stripped) {
10295 // If we haven't stripped anything we might still be able to use a
10296 // different AA, but only if the IRP changes. Effectively when we
10297 // interpret this not as a call site value but as a floating/argument
10298 // value.
10299 const IRPosition AVIRP = IRPosition::value(*AssociatedValue);
10300 if (AVIRP == getIRPosition() || !VisitValueCB(AVIRP))
10301 return indicatePessimisticFixpoint();
10302 return ChangeStatus::UNCHANGED;
10303 }
10304
10305 for (const auto &VAC : Values)
10306 if (!VisitValueCB(IRPosition::value(*VAC.getValue())))
10307 return indicatePessimisticFixpoint();
10308
10309 return ChangeStatus::UNCHANGED;
10310 }
10311
10312 /// See AbstractAttribute::trackStatistics()
10313 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noundef) }
10314};
10315
10316struct AANoUndefReturned final
10317 : AAReturnedFromReturnedValues<AANoUndef, AANoUndefImpl> {
10318 AANoUndefReturned(const IRPosition &IRP, Attributor &A)
10319 : AAReturnedFromReturnedValues<AANoUndef, AANoUndefImpl>(IRP, A) {}
10320
10321 /// See AbstractAttribute::trackStatistics()
10322 void trackStatistics() const override { STATS_DECLTRACK_FNRET_ATTR(noundef) }
10323};
10324
10325struct AANoUndefArgument final
10326 : AAArgumentFromCallSiteArguments<AANoUndef, AANoUndefImpl> {
10327 AANoUndefArgument(const IRPosition &IRP, Attributor &A)
10328 : AAArgumentFromCallSiteArguments<AANoUndef, AANoUndefImpl>(IRP, A) {}
10329
10330 /// See AbstractAttribute::trackStatistics()
10331 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(noundef) }
10332};
10333
10334struct AANoUndefCallSiteArgument final : AANoUndefFloating {
10335 AANoUndefCallSiteArgument(const IRPosition &IRP, Attributor &A)
10336 : AANoUndefFloating(IRP, A) {}
10337
10338 /// See AbstractAttribute::trackStatistics()
10339 void trackStatistics() const override { STATS_DECLTRACK_CSARG_ATTR(noundef) }
10340};
10341
10342struct AANoUndefCallSiteReturned final
10343 : AACalleeToCallSite<AANoUndef, AANoUndefImpl> {
10344 AANoUndefCallSiteReturned(const IRPosition &IRP, Attributor &A)
10345 : AACalleeToCallSite<AANoUndef, AANoUndefImpl>(IRP, A) {}
10346
10347 /// See AbstractAttribute::trackStatistics()
10348 void trackStatistics() const override { STATS_DECLTRACK_CSRET_ATTR(noundef) }
10349};
10350
10351/// ------------------------ NoFPClass Attribute -------------------------------
10352
10353struct AANoFPClassImpl : AANoFPClass {
10354 AANoFPClassImpl(const IRPosition &IRP, Attributor &A) : AANoFPClass(IRP, A) {}
10355
10356 void initialize(Attributor &A) override {
10357 const IRPosition &IRP = getIRPosition();
10358
10359 Value &V = IRP.getAssociatedValue();
10360 if (isa<UndefValue>(V)) {
10361 indicateOptimisticFixpoint();
10362 return;
10363 }
10364
10366 A.getAttrs(getIRPosition(), {Attribute::NoFPClass}, Attrs, false);
10367 for (const auto &Attr : Attrs) {
10368 addKnownBits(Attr.getNoFPClass());
10369 }
10370
10371 const DataLayout &DL = A.getDataLayout();
10372 if (getPositionKind() != IRPosition::IRP_RETURNED) {
10374 addKnownBits(~KnownFPClass.KnownFPClasses);
10375 }
10376
10377 if (Instruction *CtxI = getCtxI())
10378 followUsesInMBEC(*this, A, getState(), *CtxI);
10379 }
10380
10381 /// See followUsesInMBEC
10382 bool followUseInMBEC(Attributor &A, const Use *U, const Instruction *I,
10383 AANoFPClass::StateType &State) {
10384 // TODO: Determine what instructions can be looked through.
10385 auto *CB = dyn_cast<CallBase>(I);
10386 if (!CB)
10387 return false;
10388
10389 if (!CB->isArgOperand(U))
10390 return false;
10391
10392 unsigned ArgNo = CB->getArgOperandNo(U);
10393 IRPosition IRP = IRPosition::callsite_argument(*CB, ArgNo);
10394 if (auto *NoFPAA = A.getAAFor<AANoFPClass>(*this, IRP, DepClassTy::NONE))
10395 State.addKnownBits(NoFPAA->getState().getKnown());
10396 return false;
10397 }
10398
10399 const std::string getAsStr(Attributor *A) const override {
10400 std::string Result = "nofpclass";
10401 raw_string_ostream OS(Result);
10402 OS << getKnownNoFPClass() << '/' << getAssumedNoFPClass();
10403 return Result;
10404 }
10405
10406 void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
10407 SmallVectorImpl<Attribute> &Attrs) const override {
10408 Attrs.emplace_back(Attribute::getWithNoFPClass(Ctx, getAssumedNoFPClass()));
10409 }
10410};
10411
10412struct AANoFPClassFloating : public AANoFPClassImpl {
10413 AANoFPClassFloating(const IRPosition &IRP, Attributor &A)
10414 : AANoFPClassImpl(IRP, A) {}
10415
10416 /// See AbstractAttribute::updateImpl(...).
10417 ChangeStatus updateImpl(Attributor &A) override {
10419 bool UsedAssumedInformation = false;
10420 if (!A.getAssumedSimplifiedValues(getIRPosition(), *this, Values,
10421 AA::AnyScope, UsedAssumedInformation)) {
10422 Values.push_back({getAssociatedValue(), getCtxI()});
10423 }
10424
10425 StateType T;
10426 auto VisitValueCB = [&](Value &V, const Instruction *CtxI) -> bool {
10427 const auto *AA = A.getAAFor<AANoFPClass>(*this, IRPosition::value(V),
10428 DepClassTy::REQUIRED);
10429 if (!AA || this == AA) {
10430 T.indicatePessimisticFixpoint();
10431 } else {
10432 const AANoFPClass::StateType &S =
10433 static_cast<const AANoFPClass::StateType &>(AA->getState());
10434 T ^= S;
10435 }
10436 return T.isValidState();
10437 };
10438
10439 for (const auto &VAC : Values)
10440 if (!VisitValueCB(*VAC.getValue(), VAC.getCtxI()))
10441 return indicatePessimisticFixpoint();
10442
10443 return clampStateAndIndicateChange(getState(), T);
10444 }
10445
10446 /// See AbstractAttribute::trackStatistics()
10447 void trackStatistics() const override {
10449 }
10450};
10451
10452struct AANoFPClassReturned final
10453 : AAReturnedFromReturnedValues<AANoFPClass, AANoFPClassImpl,
10454 AANoFPClassImpl::StateType, false,
10455 Attribute::None, false> {
10456 AANoFPClassReturned(const IRPosition &IRP, Attributor &A)
10457 : AAReturnedFromReturnedValues<AANoFPClass, AANoFPClassImpl,
10458 AANoFPClassImpl::StateType, false,
10459 Attribute::None, false>(IRP, A) {}
10460
10461 /// See AbstractAttribute::trackStatistics()
10462 void trackStatistics() const override {
10464 }
10465};
10466
10467struct AANoFPClassArgument final
10468 : AAArgumentFromCallSiteArguments<AANoFPClass, AANoFPClassImpl> {
10469 AANoFPClassArgument(const IRPosition &IRP, Attributor &A)
10470 : AAArgumentFromCallSiteArguments<AANoFPClass, AANoFPClassImpl>(IRP, A) {}
10471
10472 /// See AbstractAttribute::trackStatistics()
10473 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(nofpclass) }
10474};
10475
10476struct AANoFPClassCallSiteArgument final : AANoFPClassFloating {
10477 AANoFPClassCallSiteArgument(const IRPosition &IRP, Attributor &A)
10478 : AANoFPClassFloating(IRP, A) {}
10479
10480 /// See AbstractAttribute::trackStatistics()
10481 void trackStatistics() const override {
10483 }
10484};
10485
10486struct AANoFPClassCallSiteReturned final
10487 : AACalleeToCallSite<AANoFPClass, AANoFPClassImpl> {
10488 AANoFPClassCallSiteReturned(const IRPosition &IRP, Attributor &A)
10489 : AACalleeToCallSite<AANoFPClass, AANoFPClassImpl>(IRP, A) {}
10490
10491 /// See AbstractAttribute::trackStatistics()
10492 void trackStatistics() const override {
10494 }
10495};
10496
10497struct AACallEdgesImpl : public AACallEdges {
10498 AACallEdgesImpl(const IRPosition &IRP, Attributor &A) : AACallEdges(IRP, A) {}
10499
10500 const SetVector<Function *> &getOptimisticEdges() const override {
10501 return CalledFunctions;
10502 }
10503
10504 bool hasUnknownCallee() const override { return HasUnknownCallee; }
10505
10506 bool hasNonAsmUnknownCallee() const override {
10507 return HasUnknownCalleeNonAsm;
10508 }
10509
10510 const std::string getAsStr(Attributor *A) const override {
10511 return "CallEdges[" + std::to_string(HasUnknownCallee) + "," +
10512 std::to_string(CalledFunctions.size()) + "]";
10513 }
10514
10515 void trackStatistics() const override {}
10516
10517protected:
10518 void addCalledFunction(Function *Fn, ChangeStatus &Change) {
10519 if (CalledFunctions.insert(Fn)) {
10520 Change = ChangeStatus::CHANGED;
10521 LLVM_DEBUG(dbgs() << "[AACallEdges] New call edge: " << Fn->getName()
10522 << "\n");
10523 }
10524 }
10525
10526 void setHasUnknownCallee(bool NonAsm, ChangeStatus &Change) {
10527 if (!HasUnknownCallee)
10528 Change = ChangeStatus::CHANGED;
10529 if (NonAsm && !HasUnknownCalleeNonAsm)
10530 Change = ChangeStatus::CHANGED;
10531 HasUnknownCalleeNonAsm |= NonAsm;
10532 HasUnknownCallee = true;
10533 }
10534
10535private:
10536 /// Optimistic set of functions that might be called by this position.
10537 SetVector<Function *> CalledFunctions;
10538
10539 /// Is there any call with a unknown callee.
10540 bool HasUnknownCallee = false;
10541
10542 /// Is there any call with a unknown callee, excluding any inline asm.
10543 bool HasUnknownCalleeNonAsm = false;
10544};
10545
10546struct AACallEdgesCallSite : public AACallEdgesImpl {
10547 AACallEdgesCallSite(const IRPosition &IRP, Attributor &A)
10548 : AACallEdgesImpl(IRP, A) {}
10549 /// See AbstractAttribute::updateImpl(...).
10550 ChangeStatus updateImpl(Attributor &A) override {
10551 ChangeStatus Change = ChangeStatus::UNCHANGED;
10552
10553 auto VisitValue = [&](Value &V, const Instruction *CtxI) -> bool {
10554 if (Function *Fn = dyn_cast<Function>(&V)) {
10555 addCalledFunction(Fn, Change);
10556 } else {
10557 LLVM_DEBUG(dbgs() << "[AACallEdges] Unrecognized value: " << V << "\n");
10558 setHasUnknownCallee(true, Change);
10559 }
10560
10561 // Explore all values.
10562 return true;
10563 };
10564
10566 // Process any value that we might call.
10567 auto ProcessCalledOperand = [&](Value *V, Instruction *CtxI) {
10568 if (isa<Constant>(V)) {
10569 VisitValue(*V, CtxI);
10570 return;
10571 }
10572
10573 bool UsedAssumedInformation = false;
10574 Values.clear();
10575 if (!A.getAssumedSimplifiedValues(IRPosition::value(*V), *this, Values,
10576 AA::AnyScope, UsedAssumedInformation)) {
10577 Values.push_back({*V, CtxI});
10578 }
10579 for (auto &VAC : Values)
10580 VisitValue(*VAC.getValue(), VAC.getCtxI());
10581 };
10582
10583 CallBase *CB = cast<CallBase>(getCtxI());
10584
10585 if (auto *IA = dyn_cast<InlineAsm>(CB->getCalledOperand())) {
10586 if (IA->hasSideEffects() &&
10587 !hasAssumption(*CB->getCaller(), "ompx_no_call_asm") &&
10588 !hasAssumption(*CB, "ompx_no_call_asm")) {
10589 setHasUnknownCallee(false, Change);
10590 }
10591 return Change;
10592 }
10593
10594 if (CB->isIndirectCall())
10595 if (auto *IndirectCallAA = A.getAAFor<AAIndirectCallInfo>(
10596 *this, getIRPosition(), DepClassTy::OPTIONAL))
10597 if (IndirectCallAA->foreachCallee(
10598 [&](Function *Fn) { return VisitValue(*Fn, CB); }))
10599 return Change;
10600
10601 // The most simple case.
10602 ProcessCalledOperand(CB->getCalledOperand(), CB);
10603
10604 // Process callback functions.
10605 SmallVector<const Use *, 4u> CallbackUses;
10606 AbstractCallSite::getCallbackUses(*CB, CallbackUses);
10607 for (const Use *U : CallbackUses)
10608 ProcessCalledOperand(U->get(), CB);
10609
10610 return Change;
10611 }
10612};
10613
10614struct AACallEdgesFunction : public AACallEdgesImpl {
10615 AACallEdgesFunction(const IRPosition &IRP, Attributor &A)
10616 : AACallEdgesImpl(IRP, A) {}
10617
10618 /// See AbstractAttribute::updateImpl(...).
10619 ChangeStatus updateImpl(Attributor &A) override {
10620 ChangeStatus Change = ChangeStatus::UNCHANGED;
10621
10622 auto ProcessCallInst = [&](Instruction &Inst) {
10623 CallBase &CB = cast<CallBase>(Inst);
10624
10625 auto *CBEdges = A.getAAFor<AACallEdges>(
10626 *this, IRPosition::callsite_function(CB), DepClassTy::REQUIRED);
10627 if (!CBEdges)
10628 return false;
10629 if (CBEdges->hasNonAsmUnknownCallee())
10630 setHasUnknownCallee(true, Change);
10631 if (CBEdges->hasUnknownCallee())
10632 setHasUnknownCallee(false, Change);
10633
10634 for (Function *F : CBEdges->getOptimisticEdges())
10635 addCalledFunction(F, Change);
10636
10637 return true;
10638 };
10639
10640 // Visit all callable instructions.
10641 bool UsedAssumedInformation = false;
10642 if (!A.checkForAllCallLikeInstructions(ProcessCallInst, *this,
10643 UsedAssumedInformation,
10644 /* CheckBBLivenessOnly */ true)) {
10645 // If we haven't looked at all call like instructions, assume that there
10646 // are unknown callees.
10647 setHasUnknownCallee(true, Change);
10648 }
10649
10650 return Change;
10651 }
10652};
10653
10654/// -------------------AAInterFnReachability Attribute--------------------------
10655
10656struct AAInterFnReachabilityFunction
10657 : public CachedReachabilityAA<AAInterFnReachability, Function> {
10658 using Base = CachedReachabilityAA<AAInterFnReachability, Function>;
10659 AAInterFnReachabilityFunction(const IRPosition &IRP, Attributor &A)
10660 : Base(IRP, A) {}
10661
10662 bool instructionCanReach(
10663 Attributor &A, const Instruction &From, const Function &To,
10664 const AA::InstExclusionSetTy *ExclusionSet) const override {
10665 assert(From.getFunction() == getAnchorScope() && "Queried the wrong AA!");
10666 auto *NonConstThis = const_cast<AAInterFnReachabilityFunction *>(this);
10667
10668 RQITy StackRQI(A, From, To, ExclusionSet, false);
10669 typename RQITy::Reachable Result;
10670 if (!NonConstThis->checkQueryCache(A, StackRQI, Result))
10671 return NonConstThis->isReachableImpl(A, StackRQI,
10672 /*IsTemporaryRQI=*/true);
10673 return Result == RQITy::Reachable::Yes;
10674 }
10675
10676 bool isReachableImpl(Attributor &A, RQITy &RQI,
10677 bool IsTemporaryRQI) override {
10678 const Instruction *EntryI =
10679 &RQI.From->getFunction()->getEntryBlock().front();
10680 if (EntryI != RQI.From &&
10681 !instructionCanReach(A, *EntryI, *RQI.To, nullptr))
10682 return rememberResult(A, RQITy::Reachable::No, RQI, false,
10683 IsTemporaryRQI);
10684
10685 auto CheckReachableCallBase = [&](CallBase *CB) {
10686 auto *CBEdges = A.getAAFor<AACallEdges>(
10687 *this, IRPosition::callsite_function(*CB), DepClassTy::OPTIONAL);
10688 if (!CBEdges || !CBEdges->getState().isValidState())
10689 return false;
10690 // TODO Check To backwards in this case.
10691 if (CBEdges->hasUnknownCallee())
10692 return false;
10693
10694 for (Function *Fn : CBEdges->getOptimisticEdges()) {
10695 if (Fn == RQI.To)
10696 return false;
10697
10698 if (Fn->isDeclaration()) {
10699 if (Fn->hasFnAttribute(Attribute::NoCallback))
10700 continue;
10701 // TODO Check To backwards in this case.
10702 return false;
10703 }
10704
10705 if (Fn == getAnchorScope()) {
10706 if (EntryI == RQI.From)
10707 continue;
10708 return false;
10709 }
10710
10711 const AAInterFnReachability *InterFnReachability =
10712 A.getAAFor<AAInterFnReachability>(*this, IRPosition::function(*Fn),
10713 DepClassTy::OPTIONAL);
10714
10715 const Instruction &FnFirstInst = Fn->getEntryBlock().front();
10716 if (!InterFnReachability ||
10717 InterFnReachability->instructionCanReach(A, FnFirstInst, *RQI.To,
10718 RQI.ExclusionSet))
10719 return false;
10720 }
10721 return true;
10722 };
10723
10724 const auto *IntraFnReachability = A.getAAFor<AAIntraFnReachability>(
10725 *this, IRPosition::function(*RQI.From->getFunction()),
10726 DepClassTy::OPTIONAL);
10727
10728 // Determine call like instructions that we can reach from the inst.
10729 auto CheckCallBase = [&](Instruction &CBInst) {
10730 // There are usually less nodes in the call graph, check inter function
10731 // reachability first.
10732 if (CheckReachableCallBase(cast<CallBase>(&CBInst)))
10733 return true;
10734 return IntraFnReachability && !IntraFnReachability->isAssumedReachable(
10735 A, *RQI.From, CBInst, RQI.ExclusionSet);
10736 };
10737
10738 bool UsedExclusionSet = /* conservative */ true;
10739 bool UsedAssumedInformation = false;
10740 if (!A.checkForAllCallLikeInstructions(CheckCallBase, *this,
10741 UsedAssumedInformation,
10742 /* CheckBBLivenessOnly */ true))
10743 return rememberResult(A, RQITy::Reachable::Yes, RQI, UsedExclusionSet,
10744 IsTemporaryRQI);
10745
10746 return rememberResult(A, RQITy::Reachable::No, RQI, UsedExclusionSet,
10747 IsTemporaryRQI);
10748 }
10749
10750 void trackStatistics() const override {}
10751};
10752} // namespace
10753
10754template <typename AAType>
10755static std::optional<Constant *>
10757 const IRPosition &IRP, Type &Ty) {
10758 if (!Ty.isIntegerTy())
10759 return nullptr;
10760
10761 // This will also pass the call base context.
10762 const auto *AA = A.getAAFor<AAType>(QueryingAA, IRP, DepClassTy::NONE);
10763 if (!AA)
10764 return nullptr;
10765
10766 std::optional<Constant *> COpt = AA->getAssumedConstant(A);
10767
10768 if (!COpt.has_value()) {
10769 A.recordDependence(*AA, QueryingAA, DepClassTy::OPTIONAL);
10770 return std::nullopt;
10771 }
10772 if (auto *C = *COpt) {
10773 A.recordDependence(*AA, QueryingAA, DepClassTy::OPTIONAL);
10774 return C;
10775 }
10776 return nullptr;
10777}
10778
10780 Attributor &A, const AbstractAttribute &AA, const IRPosition &IRP,
10782 Type &Ty = *IRP.getAssociatedType();
10783 std::optional<Value *> V;
10784 for (auto &It : Values) {
10785 V = AA::combineOptionalValuesInAAValueLatice(V, It.getValue(), &Ty);
10786 if (V.has_value() && !*V)
10787 break;
10788 }
10789 if (!V.has_value())
10790 return UndefValue::get(&Ty);
10791 return *V;
10792}
10793
10794namespace {
10795struct AAPotentialValuesImpl : AAPotentialValues {
10796 using StateType = PotentialLLVMValuesState;
10797
10798 AAPotentialValuesImpl(const IRPosition &IRP, Attributor &A)
10799 : AAPotentialValues(IRP, A) {}
10800
10801 /// See AbstractAttribute::initialize(..).
10802 void initialize(Attributor &A) override {
10803 if (A.hasSimplificationCallback(getIRPosition())) {
10804 indicatePessimisticFixpoint();
10805 return;
10806 }
10807 Value *Stripped = getAssociatedValue().stripPointerCasts();
10808 if (isa<Constant>(Stripped) && !isa<ConstantExpr>(Stripped)) {
10809 addValue(A, getState(), *Stripped, getCtxI(), AA::AnyScope,
10810 getAnchorScope());
10811 indicateOptimisticFixpoint();
10812 return;
10813 }
10814 AAPotentialValues::initialize(A);
10815 }
10816
10817 /// See AbstractAttribute::getAsStr().
10818 const std::string getAsStr(Attributor *A) const override {
10819 std::string Str;
10821 OS << getState();
10822 return Str;
10823 }
10824
10825 template <typename AAType>
10826 static std::optional<Value *> askOtherAA(Attributor &A,
10827 const AbstractAttribute &AA,
10828 const IRPosition &IRP, Type &Ty) {
10829 if (isa<Constant>(IRP.getAssociatedValue()))
10830 return &IRP.getAssociatedValue();
10831 std::optional<Constant *> C = askForAssumedConstant<AAType>(A, AA, IRP, Ty);
10832 if (!C)
10833 return std::nullopt;
10834 if (*C)
10835 if (auto *CC = AA::getWithType(**C, Ty))
10836 return CC;
10837 return nullptr;
10838 }
10839
10840 virtual void addValue(Attributor &A, StateType &State, Value &V,
10841 const Instruction *CtxI, AA::ValueScope S,
10842 Function *AnchorScope) const {
10843
10844 IRPosition ValIRP = IRPosition::value(V);
10845 if (auto *CB = dyn_cast_or_null<CallBase>(CtxI)) {
10846 for (const auto &U : CB->args()) {
10847 if (U.get() != &V)
10848 continue;
10849 ValIRP = IRPosition::callsite_argument(*CB, CB->getArgOperandNo(&U));
10850 break;
10851 }
10852 }
10853
10854 Value *VPtr = &V;
10855 if (ValIRP.getAssociatedType()->isIntegerTy()) {
10856 Type &Ty = *getAssociatedType();
10857 std::optional<Value *> SimpleV =
10858 askOtherAA<AAValueConstantRange>(A, *this, ValIRP, Ty);
10859 if (SimpleV.has_value() && !*SimpleV) {
10860 auto *PotentialConstantsAA = A.getAAFor<AAPotentialConstantValues>(
10861 *this, ValIRP, DepClassTy::OPTIONAL);
10862 if (PotentialConstantsAA && PotentialConstantsAA->isValidState()) {
10863 for (const auto &It : PotentialConstantsAA->getAssumedSet())
10864 State.unionAssumed({{*ConstantInt::get(&Ty, It), nullptr}, S});
10865 if (PotentialConstantsAA->undefIsContained())
10866 State.unionAssumed({{*UndefValue::get(&Ty), nullptr}, S});
10867 return;
10868 }
10869 }
10870 if (!SimpleV.has_value())
10871 return;
10872
10873 if (*SimpleV)
10874 VPtr = *SimpleV;
10875 }
10876
10877 if (isa<ConstantInt>(VPtr))
10878 CtxI = nullptr;
10879 if (!AA::isValidInScope(*VPtr, AnchorScope))
10881
10882 State.unionAssumed({{*VPtr, CtxI}, S});
10883 }
10884
10885 /// Helper struct to tie a value+context pair together with the scope for
10886 /// which this is the simplified version.
10887 struct ItemInfo {
10890
10891 bool operator==(const ItemInfo &II) const {
10892 return II.I == I && II.S == S;
10893 };
10894 bool operator<(const ItemInfo &II) const {
10895 if (I == II.I)
10896 return S < II.S;
10897 return I < II.I;
10898 };
10899 };
10900
10901 bool recurseForValue(Attributor &A, const IRPosition &IRP, AA::ValueScope S) {
10903 for (auto CS : {AA::Intraprocedural, AA::Interprocedural}) {
10904 if (!(CS & S))
10905 continue;
10906
10907 bool UsedAssumedInformation = false;
10909 if (!A.getAssumedSimplifiedValues(IRP, this, Values, CS,
10910 UsedAssumedInformation))
10911 return false;
10912
10913 for (auto &It : Values)
10914 ValueScopeMap[It] += CS;
10915 }
10916 for (auto &It : ValueScopeMap)
10917 addValue(A, getState(), *It.first.getValue(), It.first.getCtxI(),
10918 AA::ValueScope(It.second), getAnchorScope());
10919
10920 return true;
10921 }
10922
10923 void giveUpOnIntraprocedural(Attributor &A) {
10924 auto NewS = StateType::getBestState(getState());
10925 for (const auto &It : getAssumedSet()) {
10926 if (It.second == AA::Intraprocedural)
10927 continue;
10928 addValue(A, NewS, *It.first.getValue(), It.first.getCtxI(),
10929 AA::Interprocedural, getAnchorScope());
10930 }
10931 assert(!undefIsContained() && "Undef should be an explicit value!");
10932 addValue(A, NewS, getAssociatedValue(), getCtxI(), AA::Intraprocedural,
10933 getAnchorScope());
10934 getState() = NewS;
10935 }
10936
10937 /// See AbstractState::indicatePessimisticFixpoint(...).
10938 ChangeStatus indicatePessimisticFixpoint() override {
10939 getState() = StateType::getBestState(getState());
10940 getState().unionAssumed({{getAssociatedValue(), getCtxI()}, AA::AnyScope});
10942 return ChangeStatus::CHANGED;
10943 }
10944
10945 /// See AbstractAttribute::updateImpl(...).
10946 ChangeStatus updateImpl(Attributor &A) override {
10947 return indicatePessimisticFixpoint();
10948 }
10949
10950 /// See AbstractAttribute::manifest(...).
10951 ChangeStatus manifest(Attributor &A) override {
10954 Values.clear();
10955 if (!getAssumedSimplifiedValues(A, Values, S))
10956 continue;
10957 Value &OldV = getAssociatedValue();
10958 if (isa<UndefValue>(OldV))
10959 continue;
10960 Value *NewV = getSingleValue(A, *this, getIRPosition(), Values);
10961 if (!NewV || NewV == &OldV)
10962 continue;
10963 if (getCtxI() &&
10964 !AA::isValidAtPosition({*NewV, *getCtxI()}, A.getInfoCache()))
10965 continue;
10966 if (A.changeAfterManifest(getIRPosition(), *NewV))
10967 return ChangeStatus::CHANGED;
10968 }
10970 }
10971
10972 bool getAssumedSimplifiedValues(
10974 AA::ValueScope S, bool RecurseForSelectAndPHI = false) const override {
10975 if (!isValidState())
10976 return false;
10977 bool UsedAssumedInformation = false;
10978 for (const auto &It : getAssumedSet())
10979 if (It.second & S) {
10980 if (RecurseForSelectAndPHI && (isa<PHINode>(It.first.getValue()) ||
10981 isa<SelectInst>(It.first.getValue()))) {
10982 if (A.getAssumedSimplifiedValues(
10983 IRPosition::inst(*cast<Instruction>(It.first.getValue())),
10984 this, Values, S, UsedAssumedInformation))
10985 continue;
10986 }
10987 Values.push_back(It.first);
10988 }
10989 assert(!undefIsContained() && "Undef should be an explicit value!");
10990 return true;
10991 }
10992};
10993
10994struct AAPotentialValuesFloating : AAPotentialValuesImpl {
10995 AAPotentialValuesFloating(const IRPosition &IRP, Attributor &A)
10996 : AAPotentialValuesImpl(IRP, A) {}
10997
10998 /// See AbstractAttribute::updateImpl(...).
10999 ChangeStatus updateImpl(Attributor &A) override {
11000 auto AssumedBefore = getAssumed();
11001
11002 genericValueTraversal(A, &getAssociatedValue());
11003
11004 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11005 : ChangeStatus::CHANGED;
11006 }
11007
11008 /// Helper struct to remember which AAIsDead instances we actually used.
11009 struct LivenessInfo {
11010 const AAIsDead *LivenessAA = nullptr;
11011 bool AnyDead = false;
11012 };
11013
11014 /// Check if \p Cmp is a comparison we can simplify.
11015 ///
11016 /// We handle multiple cases, one in which at least one operand is an
11017 /// (assumed) nullptr. If so, try to simplify it using AANonNull on the other
11018 /// operand. Return true if successful, in that case Worklist will be updated.
11019 bool handleCmp(Attributor &A, Value &Cmp, Value *LHS, Value *RHS,
11020 CmpInst::Predicate Pred, ItemInfo II,
11021 SmallVectorImpl<ItemInfo> &Worklist) {
11022
11023 // Simplify the operands first.
11024 bool UsedAssumedInformation = false;
11025 SmallVector<AA::ValueAndContext> LHSValues, RHSValues;
11026 auto GetSimplifiedValues = [&](Value &V,
11028 if (!A.getAssumedSimplifiedValues(
11029 IRPosition::value(V, getCallBaseContext()), this, Values,
11030 AA::Intraprocedural, UsedAssumedInformation)) {
11031 Values.clear();
11032 Values.push_back(AA::ValueAndContext{V, II.I.getCtxI()});
11033 }
11034 return Values.empty();
11035 };
11036 if (GetSimplifiedValues(*LHS, LHSValues))
11037 return true;
11038 if (GetSimplifiedValues(*RHS, RHSValues))
11039 return true;
11040
11041 LLVMContext &Ctx = LHS->getContext();
11042
11043 InformationCache &InfoCache = A.getInfoCache();
11044 Instruction *CmpI = dyn_cast<Instruction>(&Cmp);
11045 Function *F = CmpI ? CmpI->getFunction() : nullptr;
11046 const auto *DT =
11048 : nullptr;
11049 const auto *TLI =
11050 F ? A.getInfoCache().getTargetLibraryInfoForFunction(*F) : nullptr;
11051 auto *AC =
11053 : nullptr;
11054
11055 const DataLayout &DL = A.getDataLayout();
11056 SimplifyQuery Q(DL, TLI, DT, AC, CmpI);
11057
11058 auto CheckPair = [&](Value &LHSV, Value &RHSV) {
11059 if (isa<UndefValue>(LHSV) || isa<UndefValue>(RHSV)) {
11060 addValue(A, getState(), *UndefValue::get(Cmp.getType()),
11061 /* CtxI */ nullptr, II.S, getAnchorScope());
11062 return true;
11063 }
11064
11065 // Handle the trivial case first in which we don't even need to think
11066 // about null or non-null.
11067 if (&LHSV == &RHSV &&
11069 Constant *NewV = ConstantInt::get(Type::getInt1Ty(Ctx),
11071 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11072 getAnchorScope());
11073 return true;
11074 }
11075
11076 auto *TypedLHS = AA::getWithType(LHSV, *LHS->getType());
11077 auto *TypedRHS = AA::getWithType(RHSV, *RHS->getType());
11078 if (TypedLHS && TypedRHS) {
11079 Value *NewV = simplifyCmpInst(Pred, TypedLHS, TypedRHS, Q);
11080 if (NewV && NewV != &Cmp) {
11081 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11082 getAnchorScope());
11083 return true;
11084 }
11085 }
11086
11087 // From now on we only handle equalities (==, !=).
11088 if (!CmpInst::isEquality(Pred))
11089 return false;
11090
11091 bool LHSIsNull = isa<ConstantPointerNull>(LHSV);
11092 bool RHSIsNull = isa<ConstantPointerNull>(RHSV);
11093 if (!LHSIsNull && !RHSIsNull)
11094 return false;
11095
11096 // Left is the nullptr ==/!= non-nullptr case. We'll use AANonNull on the
11097 // non-nullptr operand and if we assume it's non-null we can conclude the
11098 // result of the comparison.
11099 assert((LHSIsNull || RHSIsNull) &&
11100 "Expected nullptr versus non-nullptr comparison at this point");
11101
11102 // The index is the operand that we assume is not null.
11103 unsigned PtrIdx = LHSIsNull;
11104 bool IsKnownNonNull;
11105 bool IsAssumedNonNull = AA::hasAssumedIRAttr<Attribute::NonNull>(
11106 A, this, IRPosition::value(*(PtrIdx ? &RHSV : &LHSV)),
11107 DepClassTy::REQUIRED, IsKnownNonNull);
11108 if (!IsAssumedNonNull)
11109 return false;
11110
11111 // The new value depends on the predicate, true for != and false for ==.
11112 Constant *NewV =
11113 ConstantInt::get(Type::getInt1Ty(Ctx), Pred == CmpInst::ICMP_NE);
11114 addValue(A, getState(), *NewV, /* CtxI */ nullptr, II.S,
11115 getAnchorScope());
11116 return true;
11117 };
11118
11119 for (auto &LHSValue : LHSValues)
11120 for (auto &RHSValue : RHSValues)
11121 if (!CheckPair(*LHSValue.getValue(), *RHSValue.getValue()))
11122 return false;
11123 return true;
11124 }
11125
11126 bool handleSelectInst(Attributor &A, SelectInst &SI, ItemInfo II,
11127 SmallVectorImpl<ItemInfo> &Worklist) {
11128 const Instruction *CtxI = II.I.getCtxI();
11129 bool UsedAssumedInformation = false;
11130
11131 std::optional<Constant *> C =
11132 A.getAssumedConstant(*SI.getCondition(), *this, UsedAssumedInformation);
11133 bool NoValueYet = !C.has_value();
11134 if (NoValueYet || isa_and_nonnull<UndefValue>(*C))
11135 return true;
11136 if (auto *CI = dyn_cast_or_null<ConstantInt>(*C)) {
11137 if (CI->isZero())
11138 Worklist.push_back({{*SI.getFalseValue(), CtxI}, II.S});
11139 else
11140 Worklist.push_back({{*SI.getTrueValue(), CtxI}, II.S});
11141 } else if (&SI == &getAssociatedValue()) {
11142 // We could not simplify the condition, assume both values.
11143 Worklist.push_back({{*SI.getTrueValue(), CtxI}, II.S});
11144 Worklist.push_back({{*SI.getFalseValue(), CtxI}, II.S});
11145 } else {
11146 std::optional<Value *> SimpleV = A.getAssumedSimplified(
11147 IRPosition::inst(SI), *this, UsedAssumedInformation, II.S);
11148 if (!SimpleV.has_value())
11149 return true;
11150 if (*SimpleV) {
11151 addValue(A, getState(), **SimpleV, CtxI, II.S, getAnchorScope());
11152 return true;
11153 }
11154 return false;
11155 }
11156 return true;
11157 }
11158
11159 bool handleLoadInst(Attributor &A, LoadInst &LI, ItemInfo II,
11160 SmallVectorImpl<ItemInfo> &Worklist) {
11161 SmallSetVector<Value *, 4> PotentialCopies;
11162 SmallSetVector<Instruction *, 4> PotentialValueOrigins;
11163 bool UsedAssumedInformation = false;
11164 if (!AA::getPotentiallyLoadedValues(A, LI, PotentialCopies,
11165 PotentialValueOrigins, *this,
11166 UsedAssumedInformation,
11167 /* OnlyExact */ true)) {
11168 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Failed to get potentially "
11169 "loaded values for load instruction "
11170 << LI << "\n");
11171 return false;
11172 }
11173
11174 // Do not simplify loads that are only used in llvm.assume if we cannot also
11175 // remove all stores that may feed into the load. The reason is that the
11176 // assume is probably worth something as long as the stores are around.
11177 InformationCache &InfoCache = A.getInfoCache();
11178 if (InfoCache.isOnlyUsedByAssume(LI)) {
11179 if (!llvm::all_of(PotentialValueOrigins, [&](Instruction *I) {
11180 if (!I || isa<AssumeInst>(I))
11181 return true;
11182 if (auto *SI = dyn_cast<StoreInst>(I))
11183 return A.isAssumedDead(SI->getOperandUse(0), this,
11184 /* LivenessAA */ nullptr,
11185 UsedAssumedInformation,
11186 /* CheckBBLivenessOnly */ false);
11187 return A.isAssumedDead(*I, this, /* LivenessAA */ nullptr,
11188 UsedAssumedInformation,
11189 /* CheckBBLivenessOnly */ false);
11190 })) {
11191 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Load is onl used by assumes "
11192 "and we cannot delete all the stores: "
11193 << LI << "\n");
11194 return false;
11195 }
11196 }
11197
11198 // Values have to be dynamically unique or we loose the fact that a
11199 // single llvm::Value might represent two runtime values (e.g.,
11200 // stack locations in different recursive calls).
11201 const Instruction *CtxI = II.I.getCtxI();
11202 bool ScopeIsLocal = (II.S & AA::Intraprocedural);
11203 bool AllLocal = ScopeIsLocal;
11204 bool DynamicallyUnique = llvm::all_of(PotentialCopies, [&](Value *PC) {
11205 AllLocal &= AA::isValidInScope(*PC, getAnchorScope());
11206 return AA::isDynamicallyUnique(A, *this, *PC);
11207 });
11208 if (!DynamicallyUnique) {
11209 LLVM_DEBUG(dbgs() << "[AAPotentialValues] Not all potentially loaded "
11210 "values are dynamically unique: "
11211 << LI << "\n");
11212 return false;
11213 }
11214
11215 for (auto *PotentialCopy : PotentialCopies) {
11216 if (AllLocal) {
11217 Worklist.push_back({{*PotentialCopy, CtxI}, II.S});
11218 } else {
11219 Worklist.push_back({{*PotentialCopy, CtxI}, AA::Interprocedural});
11220 }
11221 }
11222 if (!AllLocal && ScopeIsLocal)
11223 addValue(A, getState(), LI, CtxI, AA::Intraprocedural, getAnchorScope());
11224 return true;
11225 }
11226
11227 bool handlePHINode(
11228 Attributor &A, PHINode &PHI, ItemInfo II,
11229 SmallVectorImpl<ItemInfo> &Worklist,
11231 auto GetLivenessInfo = [&](const Function &F) -> LivenessInfo & {
11232 LivenessInfo &LI = LivenessAAs[&F];
11233 if (!LI.LivenessAA)
11234 LI.LivenessAA = A.getAAFor<AAIsDead>(*this, IRPosition::function(F),
11235 DepClassTy::NONE);
11236 return LI;
11237 };
11238
11239 if (&PHI == &getAssociatedValue()) {
11240 LivenessInfo &LI = GetLivenessInfo(*PHI.getFunction());
11241 const auto *CI =
11242 A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(
11243 *PHI.getFunction());
11244
11245 Cycle *C = nullptr;
11246 bool CyclePHI = mayBeInCycle(CI, &PHI, /* HeaderOnly */ true, &C);
11247 for (unsigned u = 0, e = PHI.getNumIncomingValues(); u < e; u++) {
11248 BasicBlock *IncomingBB = PHI.getIncomingBlock(u);
11249 if (LI.LivenessAA &&
11250 LI.LivenessAA->isEdgeDead(IncomingBB, PHI.getParent())) {
11251 LI.AnyDead = true;
11252 continue;
11253 }
11254 Value *V = PHI.getIncomingValue(u);
11255 if (V == &PHI)
11256 continue;
11257
11258 // If the incoming value is not the PHI but an instruction in the same
11259 // cycle we might have multiple versions of it flying around.
11260 if (CyclePHI && isa<Instruction>(V) &&
11261 (!C || C->contains(cast<Instruction>(V)->getParent())))
11262 return false;
11263
11264 Worklist.push_back({{*V, IncomingBB->getTerminator()}, II.S});
11265 }
11266 return true;
11267 }
11268
11269 bool UsedAssumedInformation = false;
11270 std::optional<Value *> SimpleV = A.getAssumedSimplified(
11271 IRPosition::inst(PHI), *this, UsedAssumedInformation, II.S);
11272 if (!SimpleV.has_value())
11273 return true;
11274 if (!(*SimpleV))
11275 return false;
11276 addValue(A, getState(), **SimpleV, &PHI, II.S, getAnchorScope());
11277 return true;
11278 }
11279
11280 /// Use the generic, non-optimistic InstSimplfy functionality if we managed to
11281 /// simplify any operand of the instruction \p I. Return true if successful,
11282 /// in that case Worklist will be updated.
11283 bool handleGenericInst(Attributor &A, Instruction &I, ItemInfo II,
11284 SmallVectorImpl<ItemInfo> &Worklist) {
11285 bool SomeSimplified = false;
11286 bool UsedAssumedInformation = false;
11287
11288 SmallVector<Value *, 8> NewOps(I.getNumOperands());
11289 int Idx = 0;
11290 for (Value *Op : I.operands()) {
11291 const auto &SimplifiedOp = A.getAssumedSimplified(
11292 IRPosition::value(*Op, getCallBaseContext()), *this,
11293 UsedAssumedInformation, AA::Intraprocedural);
11294 // If we are not sure about any operand we are not sure about the entire
11295 // instruction, we'll wait.
11296 if (!SimplifiedOp.has_value())
11297 return true;
11298
11299 if (*SimplifiedOp)
11300 NewOps[Idx] = *SimplifiedOp;
11301 else
11302 NewOps[Idx] = Op;
11303
11304 SomeSimplified |= (NewOps[Idx] != Op);
11305 ++Idx;
11306 }
11307
11308 // We won't bother with the InstSimplify interface if we didn't simplify any
11309 // operand ourselves.
11310 if (!SomeSimplified)
11311 return false;
11312
11313 InformationCache &InfoCache = A.getInfoCache();
11314 Function *F = I.getFunction();
11315 const auto *DT =
11317 const auto *TLI = A.getInfoCache().getTargetLibraryInfoForFunction(*F);
11318 auto *AC = InfoCache.getAnalysisResultForFunction<AssumptionAnalysis>(*F);
11319
11320 const DataLayout &DL = I.getDataLayout();
11321 SimplifyQuery Q(DL, TLI, DT, AC, &I);
11322 Value *NewV = simplifyInstructionWithOperands(&I, NewOps, Q);
11323 if (!NewV || NewV == &I)
11324 return false;
11325
11326 LLVM_DEBUG(dbgs() << "Generic inst " << I << " assumed simplified to "
11327 << *NewV << "\n");
11328 Worklist.push_back({{*NewV, II.I.getCtxI()}, II.S});
11329 return true;
11330 }
11331
11333 Attributor &A, Instruction &I, ItemInfo II,
11334 SmallVectorImpl<ItemInfo> &Worklist,
11336 if (auto *CI = dyn_cast<CmpInst>(&I))
11337 return handleCmp(A, *CI, CI->getOperand(0), CI->getOperand(1),
11338 CI->getPredicate(), II, Worklist);
11339
11340 switch (I.getOpcode()) {
11341 case Instruction::Select:
11342 return handleSelectInst(A, cast<SelectInst>(I), II, Worklist);
11343 case Instruction::PHI:
11344 return handlePHINode(A, cast<PHINode>(I), II, Worklist, LivenessAAs);
11345 case Instruction::Load:
11346 return handleLoadInst(A, cast<LoadInst>(I), II, Worklist);
11347 default:
11348 return handleGenericInst(A, I, II, Worklist);
11349 };
11350 return false;
11351 }
11352
11353 void genericValueTraversal(Attributor &A, Value *InitialV) {
11355
11356 SmallSet<ItemInfo, 16> Visited;
11358 Worklist.push_back({{*InitialV, getCtxI()}, AA::AnyScope});
11359
11360 int Iteration = 0;
11361 do {
11362 ItemInfo II = Worklist.pop_back_val();
11363 Value *V = II.I.getValue();
11364 assert(V);
11365 const Instruction *CtxI = II.I.getCtxI();
11366 AA::ValueScope S = II.S;
11367
11368 // Check if we should process the current value. To prevent endless
11369 // recursion keep a record of the values we followed!
11370 if (!Visited.insert(II).second)
11371 continue;
11372
11373 // Make sure we limit the compile time for complex expressions.
11374 if (Iteration++ >= MaxPotentialValuesIterations) {
11375 LLVM_DEBUG(dbgs() << "Generic value traversal reached iteration limit: "
11376 << Iteration << "!\n");
11377 addValue(A, getState(), *V, CtxI, S, getAnchorScope());
11378 continue;
11379 }
11380
11381 // Explicitly look through calls with a "returned" attribute if we do
11382 // not have a pointer as stripPointerCasts only works on them.
11383 Value *NewV = nullptr;
11384 if (V->getType()->isPointerTy()) {
11385 NewV = AA::getWithType(*V->stripPointerCasts(), *V->getType());
11386 } else {
11387 if (auto *CB = dyn_cast<CallBase>(V))
11388 if (auto *Callee =
11389 dyn_cast_if_present<Function>(CB->getCalledOperand())) {
11390 for (Argument &Arg : Callee->args())
11391 if (Arg.hasReturnedAttr()) {
11392 NewV = CB->getArgOperand(Arg.getArgNo());
11393 break;
11394 }
11395 }
11396 }
11397 if (NewV && NewV != V) {
11398 Worklist.push_back({{*NewV, CtxI}, S});
11399 continue;
11400 }
11401
11402 if (auto *I = dyn_cast<Instruction>(V)) {
11403 if (simplifyInstruction(A, *I, II, Worklist, LivenessAAs))
11404 continue;
11405 }
11406
11407 if (V != InitialV || isa<Argument>(V))
11408 if (recurseForValue(A, IRPosition::value(*V), II.S))
11409 continue;
11410
11411 // If we haven't stripped anything we give up.
11412 if (V == InitialV && CtxI == getCtxI()) {
11413 indicatePessimisticFixpoint();
11414 return;
11415 }
11416
11417 addValue(A, getState(), *V, CtxI, S, getAnchorScope());
11418 } while (!Worklist.empty());
11419
11420 // If we actually used liveness information so we have to record a
11421 // dependence.
11422 for (auto &It : LivenessAAs)
11423 if (It.second.AnyDead)
11424 A.recordDependence(*It.second.LivenessAA, *this, DepClassTy::OPTIONAL);
11425 }
11426
11427 /// See AbstractAttribute::trackStatistics()
11428 void trackStatistics() const override {
11429 STATS_DECLTRACK_FLOATING_ATTR(potential_values)
11430 }
11431};
11432
11433struct AAPotentialValuesArgument final : AAPotentialValuesImpl {
11434 using Base = AAPotentialValuesImpl;
11435 AAPotentialValuesArgument(const IRPosition &IRP, Attributor &A)
11436 : Base(IRP, A) {}
11437
11438 /// See AbstractAttribute::initialize(..).
11439 void initialize(Attributor &A) override {
11440 auto &Arg = cast<Argument>(getAssociatedValue());
11442 indicatePessimisticFixpoint();
11443 }
11444
11445 /// See AbstractAttribute::updateImpl(...).
11446 ChangeStatus updateImpl(Attributor &A) override {
11447 auto AssumedBefore = getAssumed();
11448
11449 unsigned ArgNo = getCalleeArgNo();
11450
11451 bool UsedAssumedInformation = false;
11453 auto CallSitePred = [&](AbstractCallSite ACS) {
11454 const auto CSArgIRP = IRPosition::callsite_argument(ACS, ArgNo);
11455 if (CSArgIRP.getPositionKind() == IRP_INVALID)
11456 return false;
11457
11458 if (!A.getAssumedSimplifiedValues(CSArgIRP, this, Values,
11460 UsedAssumedInformation))
11461 return false;
11462
11463 return isValidState();
11464 };
11465
11466 if (!A.checkForAllCallSites(CallSitePred, *this,
11467 /* RequireAllCallSites */ true,
11468 UsedAssumedInformation))
11469 return indicatePessimisticFixpoint();
11470
11471 Function *Fn = getAssociatedFunction();
11472 bool AnyNonLocal = false;
11473 for (auto &It : Values) {
11474 if (isa<Constant>(It.getValue())) {
11475 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::AnyScope,
11476 getAnchorScope());
11477 continue;
11478 }
11479 if (!AA::isDynamicallyUnique(A, *this, *It.getValue()))
11480 return indicatePessimisticFixpoint();
11481
11482 if (auto *Arg = dyn_cast<Argument>(It.getValue()))
11483 if (Arg->getParent() == Fn) {
11484 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::AnyScope,
11485 getAnchorScope());
11486 continue;
11487 }
11488 addValue(A, getState(), *It.getValue(), It.getCtxI(), AA::Interprocedural,
11489 getAnchorScope());
11490 AnyNonLocal = true;
11491 }
11492 assert(!undefIsContained() && "Undef should be an explicit value!");
11493 if (AnyNonLocal)
11494 giveUpOnIntraprocedural(A);
11495
11496 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11497 : ChangeStatus::CHANGED;
11498 }
11499
11500 /// See AbstractAttribute::trackStatistics()
11501 void trackStatistics() const override {
11502 STATS_DECLTRACK_ARG_ATTR(potential_values)
11503 }
11504};
11505
11506struct AAPotentialValuesReturned : public AAPotentialValuesFloating {
11507 using Base = AAPotentialValuesFloating;
11508 AAPotentialValuesReturned(const IRPosition &IRP, Attributor &A)
11509 : Base(IRP, A) {}
11510
11511 /// See AbstractAttribute::initialize(..).
11512 void initialize(Attributor &A) override {
11513 Function *F = getAssociatedFunction();
11514 if (!F || F->isDeclaration() || F->getReturnType()->isVoidTy()) {
11515 indicatePessimisticFixpoint();
11516 return;
11517 }
11518
11519 for (Argument &Arg : F->args())
11520 if (Arg.hasReturnedAttr()) {
11521 addValue(A, getState(), Arg, nullptr, AA::AnyScope, F);
11522 ReturnedArg = &Arg;
11523 break;
11524 }
11525 if (!A.isFunctionIPOAmendable(*F) ||
11526 A.hasSimplificationCallback(getIRPosition())) {
11527 if (!ReturnedArg)
11528 indicatePessimisticFixpoint();
11529 else
11530 indicateOptimisticFixpoint();
11531 }
11532 }
11533
11534 /// See AbstractAttribute::updateImpl(...).
11535 ChangeStatus updateImpl(Attributor &A) override {
11536 auto AssumedBefore = getAssumed();
11537 bool UsedAssumedInformation = false;
11538
11540 Function *AnchorScope = getAnchorScope();
11541 auto HandleReturnedValue = [&](Value &V, Instruction *CtxI,
11542 bool AddValues) {
11544 Values.clear();
11545 if (!A.getAssumedSimplifiedValues(IRPosition::value(V), this, Values, S,
11546 UsedAssumedInformation,
11547 /* RecurseForSelectAndPHI */ true))
11548 return false;
11549 if (!AddValues)
11550 continue;
11551
11552 bool AllInterAreIntra = false;
11553 if (S == AA::Interprocedural)
11554 AllInterAreIntra =
11555 llvm::all_of(Values, [&](const AA::ValueAndContext &VAC) {
11556 return AA::isValidInScope(*VAC.getValue(), AnchorScope);
11557 });
11558
11559 for (const AA::ValueAndContext &VAC : Values) {
11560 addValue(A, getState(), *VAC.getValue(),
11561 VAC.getCtxI() ? VAC.getCtxI() : CtxI,
11562 AllInterAreIntra ? AA::AnyScope : S, AnchorScope);
11563 }
11564 if (AllInterAreIntra)
11565 break;
11566 }
11567 return true;
11568 };
11569
11570 if (ReturnedArg) {
11571 HandleReturnedValue(*ReturnedArg, nullptr, true);
11572 } else {
11573 auto RetInstPred = [&](Instruction &RetI) {
11574 bool AddValues = true;
11575 if (isa<PHINode>(RetI.getOperand(0)) ||
11576 isa<SelectInst>(RetI.getOperand(0))) {
11577 addValue(A, getState(), *RetI.getOperand(0), &RetI, AA::AnyScope,
11578 AnchorScope);
11579 AddValues = false;
11580 }
11581 return HandleReturnedValue(*RetI.getOperand(0), &RetI, AddValues);
11582 };
11583
11584 if (!A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
11585 UsedAssumedInformation,
11586 /* CheckBBLivenessOnly */ true))
11587 return indicatePessimisticFixpoint();
11588 }
11589
11590 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11591 : ChangeStatus::CHANGED;
11592 }
11593
11594 ChangeStatus manifest(Attributor &A) override {
11595 if (ReturnedArg)
11596 return ChangeStatus::UNCHANGED;
11598 if (!getAssumedSimplifiedValues(A, Values, AA::ValueScope::Intraprocedural,
11599 /* RecurseForSelectAndPHI */ true))
11600 return ChangeStatus::UNCHANGED;
11601 Value *NewVal = getSingleValue(A, *this, getIRPosition(), Values);
11602 if (!NewVal)
11603 return ChangeStatus::UNCHANGED;
11604
11605 ChangeStatus Changed = ChangeStatus::UNCHANGED;
11606 if (auto *Arg = dyn_cast<Argument>(NewVal)) {
11607 STATS_DECLTRACK(UniqueReturnValue, FunctionReturn,
11608 "Number of function with unique return");
11609 Changed |= A.manifestAttrs(
11611 {Attribute::get(Arg->getContext(), Attribute::Returned)});
11612 STATS_DECLTRACK_ARG_ATTR(returned);
11613 }
11614
11615 auto RetInstPred = [&](Instruction &RetI) {
11616 Value *RetOp = RetI.getOperand(0);
11617 if (isa<UndefValue>(RetOp) || RetOp == NewVal)
11618 return true;
11619 if (AA::isValidAtPosition({*NewVal, RetI}, A.getInfoCache()))
11620 if (A.changeUseAfterManifest(RetI.getOperandUse(0), *NewVal))
11621 Changed = ChangeStatus::CHANGED;
11622 return true;
11623 };
11624 bool UsedAssumedInformation = false;
11625 (void)A.checkForAllInstructions(RetInstPred, *this, {Instruction::Ret},
11626 UsedAssumedInformation,
11627 /* CheckBBLivenessOnly */ true);
11628 return Changed;
11629 }
11630
11631 ChangeStatus indicatePessimisticFixpoint() override {
11633 }
11634
11635 /// See AbstractAttribute::trackStatistics()
11636 void trackStatistics() const override{
11637 STATS_DECLTRACK_FNRET_ATTR(potential_values)}
11638
11639 /// The argumented with an existing `returned` attribute.
11640 Argument *ReturnedArg = nullptr;
11641};
11642
11643struct AAPotentialValuesFunction : AAPotentialValuesImpl {
11644 AAPotentialValuesFunction(const IRPosition &IRP, Attributor &A)
11645 : AAPotentialValuesImpl(IRP, A) {}
11646
11647 /// See AbstractAttribute::updateImpl(...).
11648 ChangeStatus updateImpl(Attributor &A) override {
11649 llvm_unreachable("AAPotentialValues(Function|CallSite)::updateImpl will "
11650 "not be called");
11651 }
11652
11653 /// See AbstractAttribute::trackStatistics()
11654 void trackStatistics() const override {
11655 STATS_DECLTRACK_FN_ATTR(potential_values)
11656 }
11657};
11658
11659struct AAPotentialValuesCallSite : AAPotentialValuesFunction {
11660 AAPotentialValuesCallSite(const IRPosition &IRP, Attributor &A)
11661 : AAPotentialValuesFunction(IRP, A) {}
11662
11663 /// See AbstractAttribute::trackStatistics()
11664 void trackStatistics() const override {
11665 STATS_DECLTRACK_CS_ATTR(potential_values)
11666 }
11667};
11668
11669struct AAPotentialValuesCallSiteReturned : AAPotentialValuesImpl {
11670 AAPotentialValuesCallSiteReturned(const IRPosition &IRP, Attributor &A)
11671 : AAPotentialValuesImpl(IRP, A) {}
11672
11673 /// See AbstractAttribute::updateImpl(...).
11674 ChangeStatus updateImpl(Attributor &A) override {
11675 auto AssumedBefore = getAssumed();
11676
11677 Function *Callee = getAssociatedFunction();
11678 if (!Callee)
11679 return indicatePessimisticFixpoint();
11680
11681 bool UsedAssumedInformation = false;
11682 auto *CB = cast<CallBase>(getCtxI());
11683 if (CB->isMustTailCall() &&
11684 !A.isAssumedDead(IRPosition::inst(*CB), this, nullptr,
11685 UsedAssumedInformation))
11686 return indicatePessimisticFixpoint();
11687
11688 Function *Caller = CB->getCaller();
11689
11690 auto AddScope = [&](AA::ValueScope S) {
11692 if (!A.getAssumedSimplifiedValues(IRPosition::returned(*Callee), this,
11693 Values, S, UsedAssumedInformation))
11694 return false;
11695
11696 for (auto &It : Values) {
11697 Value *V = It.getValue();
11698 std::optional<Value *> CallerV = A.translateArgumentToCallSiteContent(
11699 V, *CB, *this, UsedAssumedInformation);
11700 if (!CallerV.has_value()) {
11701 // Nothing to do as long as no value was determined.
11702 continue;
11703 }
11704 V = *CallerV ? *CallerV : V;
11705 if (*CallerV && AA::isDynamicallyUnique(A, *this, *V)) {
11706 if (recurseForValue(A, IRPosition::value(*V), S))
11707 continue;
11708 }
11709 if (S == AA::Intraprocedural && !AA::isValidInScope(*V, Caller)) {
11710 giveUpOnIntraprocedural(A);
11711 return true;
11712 }
11713 addValue(A, getState(), *V, CB, S, getAnchorScope());
11714 }
11715 return true;
11716 };
11717 if (!AddScope(AA::Intraprocedural))
11718 return indicatePessimisticFixpoint();
11719 if (!AddScope(AA::Interprocedural))
11720 return indicatePessimisticFixpoint();
11721 return (AssumedBefore == getAssumed()) ? ChangeStatus::UNCHANGED
11722 : ChangeStatus::CHANGED;
11723 }
11724
11725 ChangeStatus indicatePessimisticFixpoint() override {
11726 return AAPotentialValues::indicatePessimisticFixpoint();
11727 }
11728
11729 /// See AbstractAttribute::trackStatistics()
11730 void trackStatistics() const override {
11731 STATS_DECLTRACK_CSRET_ATTR(potential_values)
11732 }
11733};
11734
11735struct AAPotentialValuesCallSiteArgument : AAPotentialValuesFloating {
11736 AAPotentialValuesCallSiteArgument(const IRPosition &IRP, Attributor &A)
11737 : AAPotentialValuesFloating(IRP, A) {}
11738
11739 /// See AbstractAttribute::trackStatistics()
11740 void trackStatistics() const override {
11741 STATS_DECLTRACK_CSARG_ATTR(potential_values)
11742 }
11743};
11744} // namespace
11745
11746/// ---------------------- Assumption Propagation ------------------------------
11747namespace {
11748struct AAAssumptionInfoImpl : public AAAssumptionInfo {
11749 AAAssumptionInfoImpl(const IRPosition &IRP, Attributor &A,
11750 const DenseSet<StringRef> &Known)
11751 : AAAssumptionInfo(IRP, A, Known) {}
11752
11753 /// See AbstractAttribute::manifest(...).
11754 ChangeStatus manifest(Attributor &A) override {
11755 // Don't manifest a universal set if it somehow made it here.
11756 if (getKnown().isUniversal())
11757 return ChangeStatus::UNCHANGED;
11758
11759 const IRPosition &IRP = getIRPosition();
11760 SmallVector<StringRef, 0> Set(getAssumed().getSet().begin(),
11761 getAssumed().getSet().end());
11762 llvm::sort(Set);
11763 return A.manifestAttrs(IRP,
11766 llvm::join(Set, ",")),
11767 /*ForceReplace=*/true);
11768 }
11769
11770 bool hasAssumption(const StringRef Assumption) const override {
11771 return isValidState() && setContains(Assumption);
11772 }
11773
11774 /// See AbstractAttribute::getAsStr()
11775 const std::string getAsStr(Attributor *A) const override {
11776 const SetContents &Known = getKnown();
11777 const SetContents &Assumed = getAssumed();
11778
11779 SmallVector<StringRef, 0> Set(Known.getSet().begin(), Known.getSet().end());
11780 llvm::sort(Set);
11781 const std::string KnownStr = llvm::join(Set, ",");
11782
11783 std::string AssumedStr = "Universal";
11784 if (!Assumed.isUniversal()) {
11785 Set.assign(Assumed.getSet().begin(), Assumed.getSet().end());
11786 AssumedStr = llvm::join(Set, ",");
11787 }
11788 return "Known [" + KnownStr + "]," + " Assumed [" + AssumedStr + "]";
11789 }
11790};
11791
11792/// Propagates assumption information from parent functions to all of their
11793/// successors. An assumption can be propagated if the containing function
11794/// dominates the called function.
11795///
11796/// We start with a "known" set of assumptions already valid for the associated
11797/// function and an "assumed" set that initially contains all possible
11798/// assumptions. The assumed set is inter-procedurally updated by narrowing its
11799/// contents as concrete values are known. The concrete values are seeded by the
11800/// first nodes that are either entries into the call graph, or contains no
11801/// assumptions. Each node is updated as the intersection of the assumed state
11802/// with all of its predecessors.
11803struct AAAssumptionInfoFunction final : AAAssumptionInfoImpl {
11804 AAAssumptionInfoFunction(const IRPosition &IRP, Attributor &A)
11805 : AAAssumptionInfoImpl(IRP, A,
11806 getAssumptions(*IRP.getAssociatedFunction())) {}
11807
11808 /// See AbstractAttribute::updateImpl(...).
11809 ChangeStatus updateImpl(Attributor &A) override {
11810 bool Changed = false;
11811
11812 auto CallSitePred = [&](AbstractCallSite ACS) {
11813 const auto *AssumptionAA = A.getAAFor<AAAssumptionInfo>(
11814 *this, IRPosition::callsite_function(*ACS.getInstruction()),
11815 DepClassTy::REQUIRED);
11816 if (!AssumptionAA)
11817 return false;
11818 // Get the set of assumptions shared by all of this function's callers.
11819 Changed |= getIntersection(AssumptionAA->getAssumed());
11820 return !getAssumed().empty() || !getKnown().empty();
11821 };
11822
11823 bool UsedAssumedInformation = false;
11824 // Get the intersection of all assumptions held by this node's predecessors.
11825 // If we don't know all the call sites then this is either an entry into the
11826 // call graph or an empty node. This node is known to only contain its own
11827 // assumptions and can be propagated to its successors.
11828 if (!A.checkForAllCallSites(CallSitePred, *this, true,
11829 UsedAssumedInformation))
11830 return indicatePessimisticFixpoint();
11831
11832 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
11833 }
11834
11835 void trackStatistics() const override {}
11836};
11837
11838/// Assumption Info defined for call sites.
11839struct AAAssumptionInfoCallSite final : AAAssumptionInfoImpl {
11840
11841 AAAssumptionInfoCallSite(const IRPosition &IRP, Attributor &A)
11842 : AAAssumptionInfoImpl(IRP, A, getInitialAssumptions(IRP)) {}
11843
11844 /// See AbstractAttribute::initialize(...).
11845 void initialize(Attributor &A) override {
11846 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
11847 A.getAAFor<AAAssumptionInfo>(*this, FnPos, DepClassTy::REQUIRED);
11848 }
11849
11850 /// See AbstractAttribute::updateImpl(...).
11851 ChangeStatus updateImpl(Attributor &A) override {
11852 const IRPosition &FnPos = IRPosition::function(*getAnchorScope());
11853 auto *AssumptionAA =
11854 A.getAAFor<AAAssumptionInfo>(*this, FnPos, DepClassTy::REQUIRED);
11855 if (!AssumptionAA)
11856 return indicatePessimisticFixpoint();
11857 bool Changed = getIntersection(AssumptionAA->getAssumed());
11858 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
11859 }
11860
11861 /// See AbstractAttribute::trackStatistics()
11862 void trackStatistics() const override {}
11863
11864private:
11865 /// Helper to initialized the known set as all the assumptions this call and
11866 /// the callee contain.
11867 DenseSet<StringRef> getInitialAssumptions(const IRPosition &IRP) {
11868 const CallBase &CB = cast<CallBase>(IRP.getAssociatedValue());
11869 auto Assumptions = getAssumptions(CB);
11870 if (const Function *F = CB.getCaller())
11871 set_union(Assumptions, getAssumptions(*F));
11872 if (Function *F = IRP.getAssociatedFunction())
11873 set_union(Assumptions, getAssumptions(*F));
11874 return Assumptions;
11875 }
11876};
11877} // namespace
11878
11880 return static_cast<AACallGraphNode *>(const_cast<AACallEdges *>(
11881 A.getOrCreateAAFor<AACallEdges>(IRPosition::function(**I))));
11882}
11883
11885
11886/// ------------------------ UnderlyingObjects ---------------------------------
11887
11888namespace {
11889struct AAUnderlyingObjectsImpl
11890 : StateWrapper<BooleanState, AAUnderlyingObjects> {
11892 AAUnderlyingObjectsImpl(const IRPosition &IRP, Attributor &A) : BaseTy(IRP) {}
11893
11894 /// See AbstractAttribute::getAsStr().
11895 const std::string getAsStr(Attributor *A) const override {
11896 if (!isValidState())
11897 return "<invalid>";
11898 std::string Str;
11900 OS << "underlying objects: inter " << InterAssumedUnderlyingObjects.size()
11901 << " objects, intra " << IntraAssumedUnderlyingObjects.size()
11902 << " objects.\n";
11903 if (!InterAssumedUnderlyingObjects.empty()) {
11904 OS << "inter objects:\n";
11905 for (auto *Obj : InterAssumedUnderlyingObjects)
11906 OS << *Obj << '\n';
11907 }
11908 if (!IntraAssumedUnderlyingObjects.empty()) {
11909 OS << "intra objects:\n";
11910 for (auto *Obj : IntraAssumedUnderlyingObjects)
11911 OS << *Obj << '\n';
11912 }
11913 return Str;
11914 }
11915
11916 /// See AbstractAttribute::trackStatistics()
11917 void trackStatistics() const override {}
11918
11919 /// See AbstractAttribute::updateImpl(...).
11920 ChangeStatus updateImpl(Attributor &A) override {
11921 auto &Ptr = getAssociatedValue();
11922
11923 bool UsedAssumedInformation = false;
11924 auto DoUpdate = [&](SmallSetVector<Value *, 8> &UnderlyingObjects,
11926 SmallPtrSet<Value *, 8> SeenObjects;
11928
11929 if (!A.getAssumedSimplifiedValues(IRPosition::value(Ptr), *this, Values,
11930 Scope, UsedAssumedInformation))
11931 return UnderlyingObjects.insert(&Ptr);
11932
11933 bool Changed = false;
11934
11935 for (unsigned I = 0; I < Values.size(); ++I) {
11936 auto &VAC = Values[I];
11937 auto *Obj = VAC.getValue();
11938 Value *UO = getUnderlyingObject(Obj);
11939 if (!SeenObjects.insert(UO ? UO : Obj).second)
11940 continue;
11941 if (UO && UO != Obj) {
11942 if (isa<AllocaInst>(UO) || isa<GlobalValue>(UO)) {
11943 Changed |= UnderlyingObjects.insert(UO);
11944 continue;
11945 }
11946
11947 const auto *OtherAA = A.getAAFor<AAUnderlyingObjects>(
11948 *this, IRPosition::value(*UO), DepClassTy::OPTIONAL);
11949 auto Pred = [&](Value &V) {
11950 if (&V == UO)
11951 Changed |= UnderlyingObjects.insert(UO);
11952 else
11953 Values.emplace_back(V, nullptr);
11954 return true;
11955 };
11956
11957 if (!OtherAA || !OtherAA->forallUnderlyingObjects(Pred, Scope))
11959 "The forall call should not return false at this position");
11960 UsedAssumedInformation |= !OtherAA->getState().isAtFixpoint();
11961 continue;
11962 }
11963
11964 if (isa<SelectInst>(Obj)) {
11965 Changed |= handleIndirect(A, *Obj, UnderlyingObjects, Scope,
11966 UsedAssumedInformation);
11967 continue;
11968 }
11969 if (auto *PHI = dyn_cast<PHINode>(Obj)) {
11970 // Explicitly look through PHIs as we do not care about dynamically
11971 // uniqueness.
11972 for (unsigned u = 0, e = PHI->getNumIncomingValues(); u < e; u++) {
11973 Changed |=
11974 handleIndirect(A, *PHI->getIncomingValue(u), UnderlyingObjects,
11975 Scope, UsedAssumedInformation);
11976 }
11977 continue;
11978 }
11979
11980 Changed |= UnderlyingObjects.insert(Obj);
11981 }
11982
11983 return Changed;
11984 };
11985
11986 bool Changed = false;
11987 Changed |= DoUpdate(IntraAssumedUnderlyingObjects, AA::Intraprocedural);
11988 Changed |= DoUpdate(InterAssumedUnderlyingObjects, AA::Interprocedural);
11989 if (!UsedAssumedInformation)
11990 indicateOptimisticFixpoint();
11991 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
11992 }
11993
11994 bool forallUnderlyingObjects(
11995 function_ref<bool(Value &)> Pred,
11996 AA::ValueScope Scope = AA::Interprocedural) const override {
11997 if (!isValidState())
11998 return Pred(getAssociatedValue());
11999
12000 auto &AssumedUnderlyingObjects = Scope == AA::Intraprocedural
12001 ? IntraAssumedUnderlyingObjects
12002 : InterAssumedUnderlyingObjects;
12003 for (Value *Obj : AssumedUnderlyingObjects)
12004 if (!Pred(*Obj))
12005 return false;
12006
12007 return true;
12008 }
12009
12010private:
12011 /// Handle the case where the value is not the actual underlying value, such
12012 /// as a phi node or a select instruction.
12013 bool handleIndirect(Attributor &A, Value &V,
12014 SmallSetVector<Value *, 8> &UnderlyingObjects,
12015 AA::ValueScope Scope, bool &UsedAssumedInformation) {
12016 bool Changed = false;
12017 const auto *AA = A.getAAFor<AAUnderlyingObjects>(
12018 *this, IRPosition::value(V), DepClassTy::OPTIONAL);
12019 auto Pred = [&](Value &V) {
12020 Changed |= UnderlyingObjects.insert(&V);
12021 return true;
12022 };
12023 if (!AA || !AA->forallUnderlyingObjects(Pred, Scope))
12025 "The forall call should not return false at this position");
12026 UsedAssumedInformation |= !AA->getState().isAtFixpoint();
12027 return Changed;
12028 }
12029
12030 /// All the underlying objects collected so far via intra procedural scope.
12031 SmallSetVector<Value *, 8> IntraAssumedUnderlyingObjects;
12032 /// All the underlying objects collected so far via inter procedural scope.
12033 SmallSetVector<Value *, 8> InterAssumedUnderlyingObjects;
12034};
12035
12036struct AAUnderlyingObjectsFloating final : AAUnderlyingObjectsImpl {
12037 AAUnderlyingObjectsFloating(const IRPosition &IRP, Attributor &A)
12038 : AAUnderlyingObjectsImpl(IRP, A) {}
12039};
12040
12041struct AAUnderlyingObjectsArgument final : AAUnderlyingObjectsImpl {
12042 AAUnderlyingObjectsArgument(const IRPosition &IRP, Attributor &A)
12043 : AAUnderlyingObjectsImpl(IRP, A) {}
12044};
12045
12046struct AAUnderlyingObjectsCallSite final : AAUnderlyingObjectsImpl {
12047 AAUnderlyingObjectsCallSite(const IRPosition &IRP, Attributor &A)
12048 : AAUnderlyingObjectsImpl(IRP, A) {}
12049};
12050
12051struct AAUnderlyingObjectsCallSiteArgument final : AAUnderlyingObjectsImpl {
12052 AAUnderlyingObjectsCallSiteArgument(const IRPosition &IRP, Attributor &A)
12053 : AAUnderlyingObjectsImpl(IRP, A) {}
12054};
12055
12056struct AAUnderlyingObjectsReturned final : AAUnderlyingObjectsImpl {
12057 AAUnderlyingObjectsReturned(const IRPosition &IRP, Attributor &A)
12058 : AAUnderlyingObjectsImpl(IRP, A) {}
12059};
12060
12061struct AAUnderlyingObjectsCallSiteReturned final : AAUnderlyingObjectsImpl {
12062 AAUnderlyingObjectsCallSiteReturned(const IRPosition &IRP, Attributor &A)
12063 : AAUnderlyingObjectsImpl(IRP, A) {}
12064};
12065
12066struct AAUnderlyingObjectsFunction final : AAUnderlyingObjectsImpl {
12067 AAUnderlyingObjectsFunction(const IRPosition &IRP, Attributor &A)
12068 : AAUnderlyingObjectsImpl(IRP, A) {}
12069};
12070} // namespace
12071
12072/// ------------------------ Global Value Info -------------------------------
12073namespace {
12074struct AAGlobalValueInfoFloating : public AAGlobalValueInfo {
12075 AAGlobalValueInfoFloating(const IRPosition &IRP, Attributor &A)
12076 : AAGlobalValueInfo(IRP, A) {}
12077
12078 /// See AbstractAttribute::initialize(...).
12079 void initialize(Attributor &A) override {}
12080
12081 bool checkUse(Attributor &A, const Use &U, bool &Follow,
12083 Instruction *UInst = dyn_cast<Instruction>(U.getUser());
12084 if (!UInst) {
12085 Follow = true;
12086 return true;
12087 }
12088
12089 LLVM_DEBUG(dbgs() << "[AAGlobalValueInfo] Check use: " << *U.get() << " in "
12090 << *UInst << "\n");
12091
12092 if (auto *Cmp = dyn_cast<ICmpInst>(U.getUser())) {
12093 int Idx = &Cmp->getOperandUse(0) == &U;
12094 if (isa<Constant>(Cmp->getOperand(Idx)))
12095 return true;
12096 return U == &getAnchorValue();
12097 }
12098
12099 // Explicitly catch return instructions.
12100 if (isa<ReturnInst>(UInst)) {
12101 auto CallSitePred = [&](AbstractCallSite ACS) {
12102 Worklist.push_back(ACS.getInstruction());
12103 return true;
12104 };
12105 bool UsedAssumedInformation = false;
12106 // TODO: We should traverse the uses or add a "non-call-site" CB.
12107 if (!A.checkForAllCallSites(CallSitePred, *UInst->getFunction(),
12108 /*RequireAllCallSites=*/true, this,
12109 UsedAssumedInformation))
12110 return false;
12111 return true;
12112 }
12113
12114 // For now we only use special logic for call sites. However, the tracker
12115 // itself knows about a lot of other non-capturing cases already.
12116 auto *CB = dyn_cast<CallBase>(UInst);
12117 if (!CB)
12118 return false;
12119 // Direct calls are OK uses.
12120 if (CB->isCallee(&U))
12121 return true;
12122 // Non-argument uses are scary.
12123 if (!CB->isArgOperand(&U))
12124 return false;
12125 // TODO: Iterate callees.
12126 auto *Fn = dyn_cast<Function>(CB->getCalledOperand());
12127 if (!Fn || !A.isFunctionIPOAmendable(*Fn))
12128 return false;
12129
12130 unsigned ArgNo = CB->getArgOperandNo(&U);
12131 Worklist.push_back(Fn->getArg(ArgNo));
12132 return true;
12133 }
12134
12135 ChangeStatus updateImpl(Attributor &A) override {
12136 unsigned NumUsesBefore = Uses.size();
12137
12140 Worklist.push_back(&getAnchorValue());
12141
12142 auto UsePred = [&](const Use &U, bool &Follow) -> bool {
12143 Uses.insert(&U);
12144 switch (DetermineUseCaptureKind(U, nullptr)) {
12145 case UseCaptureKind::NO_CAPTURE:
12146 return checkUse(A, U, Follow, Worklist);
12147 case UseCaptureKind::MAY_CAPTURE:
12148 return checkUse(A, U, Follow, Worklist);
12149 case UseCaptureKind::PASSTHROUGH:
12150 Follow = true;
12151 return true;
12152 }
12153 return true;
12154 };
12155 auto EquivalentUseCB = [&](const Use &OldU, const Use &NewU) {
12156 Uses.insert(&OldU);
12157 return true;
12158 };
12159
12160 while (!Worklist.empty()) {
12161 const Value *V = Worklist.pop_back_val();
12162 if (!Visited.insert(V).second)
12163 continue;
12164 if (!A.checkForAllUses(UsePred, *this, *V,
12165 /* CheckBBLivenessOnly */ true,
12166 DepClassTy::OPTIONAL,
12167 /* IgnoreDroppableUses */ true, EquivalentUseCB)) {
12168 return indicatePessimisticFixpoint();
12169 }
12170 }
12171
12172 return Uses.size() == NumUsesBefore ? ChangeStatus::UNCHANGED
12173 : ChangeStatus::CHANGED;
12174 }
12175
12176 bool isPotentialUse(const Use &U) const override {
12177 return !isValidState() || Uses.contains(&U);
12178 }
12179
12180 /// See AbstractAttribute::manifest(...).
12181 ChangeStatus manifest(Attributor &A) override {
12182 return ChangeStatus::UNCHANGED;
12183 }
12184
12185 /// See AbstractAttribute::getAsStr().
12186 const std::string getAsStr(Attributor *A) const override {
12187 return "[" + std::to_string(Uses.size()) + " uses]";
12188 }
12189
12190 void trackStatistics() const override {
12191 STATS_DECLTRACK_FLOATING_ATTR(GlobalValuesTracked);
12192 }
12193
12194private:
12195 /// Set of (transitive) uses of this GlobalValue.
12197};
12198} // namespace
12199
12200/// ------------------------ Indirect Call Info -------------------------------
12201namespace {
12202struct AAIndirectCallInfoCallSite : public AAIndirectCallInfo {
12203 AAIndirectCallInfoCallSite(const IRPosition &IRP, Attributor &A)
12204 : AAIndirectCallInfo(IRP, A) {}
12205
12206 /// See AbstractAttribute::initialize(...).
12207 void initialize(Attributor &A) override {
12208 auto *MD = getCtxI()->getMetadata(LLVMContext::MD_callees);
12209 if (!MD && !A.isClosedWorldModule())
12210 return;
12211
12212 if (MD) {
12213 for (const auto &Op : MD->operands())
12214 if (Function *Callee = mdconst::dyn_extract_or_null<Function>(Op))
12215 PotentialCallees.insert(Callee);
12216 } else if (A.isClosedWorldModule()) {
12217 ArrayRef<Function *> IndirectlyCallableFunctions =
12218 A.getInfoCache().getIndirectlyCallableFunctions(A);
12219 PotentialCallees.insert(IndirectlyCallableFunctions.begin(),
12220 IndirectlyCallableFunctions.end());
12221 }
12222
12223 if (PotentialCallees.empty())
12224 indicateOptimisticFixpoint();
12225 }
12226
12227 ChangeStatus updateImpl(Attributor &A) override {
12228 CallBase *CB = cast<CallBase>(getCtxI());
12229 const Use &CalleeUse = CB->getCalledOperandUse();
12230 Value *FP = CB->getCalledOperand();
12231
12232 SmallSetVector<Function *, 4> AssumedCalleesNow;
12233 bool AllCalleesKnownNow = AllCalleesKnown;
12234
12235 auto CheckPotentialCalleeUse = [&](Function &PotentialCallee,
12236 bool &UsedAssumedInformation) {
12237 const auto *GIAA = A.getAAFor<AAGlobalValueInfo>(
12238 *this, IRPosition::value(PotentialCallee), DepClassTy::OPTIONAL);
12239 if (!GIAA || GIAA->isPotentialUse(CalleeUse))
12240 return true;
12241 UsedAssumedInformation = !GIAA->isAtFixpoint();
12242 return false;
12243 };
12244
12245 auto AddPotentialCallees = [&]() {
12246 for (auto *PotentialCallee : PotentialCallees) {
12247 bool UsedAssumedInformation = false;
12248 if (CheckPotentialCalleeUse(*PotentialCallee, UsedAssumedInformation))
12249 AssumedCalleesNow.insert(PotentialCallee);
12250 }
12251 };
12252
12253 // Use simplification to find potential callees, if !callees was present,
12254 // fallback to that set if necessary.
12255 bool UsedAssumedInformation = false;
12257 if (!A.getAssumedSimplifiedValues(IRPosition::value(*FP), this, Values,
12258 AA::ValueScope::AnyScope,
12259 UsedAssumedInformation)) {
12260 if (PotentialCallees.empty())
12261 return indicatePessimisticFixpoint();
12262 AddPotentialCallees();
12263 }
12264
12265 // Try to find a reason for \p Fn not to be a potential callee. If none was
12266 // found, add it to the assumed callees set.
12267 auto CheckPotentialCallee = [&](Function &Fn) {
12268 if (!PotentialCallees.empty() && !PotentialCallees.count(&Fn))
12269 return false;
12270
12271 auto &CachedResult = FilterResults[&Fn];
12272 if (CachedResult.has_value())
12273 return CachedResult.value();
12274
12275 bool UsedAssumedInformation = false;
12276 if (!CheckPotentialCalleeUse(Fn, UsedAssumedInformation)) {
12277 if (!UsedAssumedInformation)
12278 CachedResult = false;
12279 return false;
12280 }
12281
12282 int NumFnArgs = Fn.arg_size();
12283 int NumCBArgs = CB->arg_size();
12284
12285 // Check if any excess argument (which we fill up with poison) is known to
12286 // be UB on undef.
12287 for (int I = NumCBArgs; I < NumFnArgs; ++I) {
12288 bool IsKnown = false;
12289 if (AA::hasAssumedIRAttr<Attribute::NoUndef>(
12290 A, this, IRPosition::argument(*Fn.getArg(I)),
12291 DepClassTy::OPTIONAL, IsKnown)) {
12292 if (IsKnown)
12293 CachedResult = false;
12294 return false;
12295 }
12296 }
12297
12298 CachedResult = true;
12299 return true;
12300 };
12301
12302 // Check simplification result, prune known UB callees, also restrict it to
12303 // the !callees set, if present.
12304 for (auto &VAC : Values) {
12305 if (isa<UndefValue>(VAC.getValue()))
12306 continue;
12307 if (isa<ConstantPointerNull>(VAC.getValue()) &&
12308 VAC.getValue()->getType()->getPointerAddressSpace() == 0)
12309 continue;
12310 // TODO: Check for known UB, e.g., poison + noundef.
12311 if (auto *VACFn = dyn_cast<Function>(VAC.getValue())) {
12312 if (CheckPotentialCallee(*VACFn))
12313 AssumedCalleesNow.insert(VACFn);
12314 continue;
12315 }
12316 if (!PotentialCallees.empty()) {
12317 AddPotentialCallees();
12318 break;
12319 }
12320 AllCalleesKnownNow = false;
12321 }
12322
12323 if (AssumedCalleesNow == AssumedCallees &&
12324 AllCalleesKnown == AllCalleesKnownNow)
12325 return ChangeStatus::UNCHANGED;
12326
12327 std::swap(AssumedCallees, AssumedCalleesNow);
12328 AllCalleesKnown = AllCalleesKnownNow;
12329 return ChangeStatus::CHANGED;
12330 }
12331
12332 /// See AbstractAttribute::manifest(...).
12333 ChangeStatus manifest(Attributor &A) override {
12334 // If we can't specialize at all, give up now.
12335 if (!AllCalleesKnown && AssumedCallees.empty())
12336 return ChangeStatus::UNCHANGED;
12337
12338 CallBase *CB = cast<CallBase>(getCtxI());
12339 bool UsedAssumedInformation = false;
12340 if (A.isAssumedDead(*CB, this, /*LivenessAA=*/nullptr,
12341 UsedAssumedInformation))
12342 return ChangeStatus::UNCHANGED;
12343
12344 ChangeStatus Changed = ChangeStatus::UNCHANGED;
12345 Value *FP = CB->getCalledOperand();
12346 if (FP->getType()->getPointerAddressSpace())
12347 FP = new AddrSpaceCastInst(FP, PointerType::get(FP->getType(), 0),
12348 FP->getName() + ".as0", CB->getIterator());
12349
12350 bool CBIsVoid = CB->getType()->isVoidTy();
12352 FunctionType *CSFT = CB->getFunctionType();
12353 SmallVector<Value *> CSArgs(CB->args());
12354
12355 // If we know all callees and there are none, the call site is (effectively)
12356 // dead (or UB).
12357 if (AssumedCallees.empty()) {
12358 assert(AllCalleesKnown &&
12359 "Expected all callees to be known if there are none.");
12360 A.changeToUnreachableAfterManifest(CB);
12361 return ChangeStatus::CHANGED;
12362 }
12363
12364 // Special handling for the single callee case.
12365 if (AllCalleesKnown && AssumedCallees.size() == 1) {
12366 auto *NewCallee = AssumedCallees.front();
12367 if (isLegalToPromote(*CB, NewCallee)) {
12368 promoteCall(*CB, NewCallee, nullptr);
12369 NumIndirectCallsPromoted++;
12370 return ChangeStatus::CHANGED;
12371 }
12372 Instruction *NewCall =
12373 CallInst::Create(FunctionCallee(CSFT, NewCallee), CSArgs,
12374 CB->getName(), CB->getIterator());
12375 if (!CBIsVoid)
12376 A.changeAfterManifest(IRPosition::callsite_returned(*CB), *NewCall);
12377 A.deleteAfterManifest(*CB);
12378 return ChangeStatus::CHANGED;
12379 }
12380
12381 // For each potential value we create a conditional
12382 //
12383 // ```
12384 // if (ptr == value) value(args);
12385 // else ...
12386 // ```
12387 //
12388 bool SpecializedForAnyCallees = false;
12389 bool SpecializedForAllCallees = AllCalleesKnown;
12390 ICmpInst *LastCmp = nullptr;
12391 SmallVector<Function *, 8> SkippedAssumedCallees;
12393 for (Function *NewCallee : AssumedCallees) {
12394 if (!A.shouldSpecializeCallSiteForCallee(*this, *CB, *NewCallee,
12395 AssumedCallees.size())) {
12396 SkippedAssumedCallees.push_back(NewCallee);
12397 SpecializedForAllCallees = false;
12398 continue;
12399 }
12400 SpecializedForAnyCallees = true;
12401
12402 LastCmp = new ICmpInst(IP, llvm::CmpInst::ICMP_EQ, FP, NewCallee);
12403 Instruction *ThenTI =
12404 SplitBlockAndInsertIfThen(LastCmp, IP, /* Unreachable */ false);
12405 BasicBlock *CBBB = CB->getParent();
12406 A.registerManifestAddedBasicBlock(*ThenTI->getParent());
12407 A.registerManifestAddedBasicBlock(*IP->getParent());
12408 auto *SplitTI = cast<BranchInst>(LastCmp->getNextNode());
12409 BasicBlock *ElseBB;
12410 if (&*IP == CB) {
12411 ElseBB = BasicBlock::Create(ThenTI->getContext(), "",
12412 ThenTI->getFunction(), CBBB);
12413 A.registerManifestAddedBasicBlock(*ElseBB);
12414 IP = BranchInst::Create(CBBB, ElseBB)->getIterator();
12415 SplitTI->replaceUsesOfWith(CBBB, ElseBB);
12416 } else {
12417 ElseBB = IP->getParent();
12418 ThenTI->replaceUsesOfWith(ElseBB, CBBB);
12419 }
12420 CastInst *RetBC = nullptr;
12421 CallInst *NewCall = nullptr;
12422 if (isLegalToPromote(*CB, NewCallee)) {
12423 auto *CBClone = cast<CallBase>(CB->clone());
12424 CBClone->insertBefore(ThenTI);
12425 NewCall = &cast<CallInst>(promoteCall(*CBClone, NewCallee, &RetBC));
12426 NumIndirectCallsPromoted++;
12427 } else {
12428 NewCall = CallInst::Create(FunctionCallee(CSFT, NewCallee), CSArgs,
12429 CB->getName(), ThenTI->getIterator());
12430 }
12431 NewCalls.push_back({NewCall, RetBC});
12432 }
12433
12434 auto AttachCalleeMetadata = [&](CallBase &IndirectCB) {
12435 if (!AllCalleesKnown)
12436 return ChangeStatus::UNCHANGED;
12437 MDBuilder MDB(IndirectCB.getContext());
12438 MDNode *Callees = MDB.createCallees(SkippedAssumedCallees);
12439 IndirectCB.setMetadata(LLVMContext::MD_callees, Callees);
12440 return ChangeStatus::CHANGED;
12441 };
12442
12443 if (!SpecializedForAnyCallees)
12444 return AttachCalleeMetadata(*CB);
12445
12446 // Check if we need the fallback indirect call still.
12447 if (SpecializedForAllCallees) {
12449 LastCmp->eraseFromParent();
12450 new UnreachableInst(IP->getContext(), IP);
12451 IP->eraseFromParent();
12452 } else {
12453 auto *CBClone = cast<CallInst>(CB->clone());
12454 CBClone->setName(CB->getName());
12455 CBClone->insertBefore(*IP->getParent(), IP);
12456 NewCalls.push_back({CBClone, nullptr});
12457 AttachCalleeMetadata(*CBClone);
12458 }
12459
12460 // Check if we need a PHI to merge the results.
12461 if (!CBIsVoid) {
12462 auto *PHI = PHINode::Create(CB->getType(), NewCalls.size(),
12463 CB->getName() + ".phi",
12464 CB->getParent()->getFirstInsertionPt());
12465 for (auto &It : NewCalls) {
12466 CallBase *NewCall = It.first;
12467 Instruction *CallRet = It.second ? It.second : It.first;
12468 if (CallRet->getType() == CB->getType())
12469 PHI->addIncoming(CallRet, CallRet->getParent());
12470 else if (NewCall->getType()->isVoidTy())
12471 PHI->addIncoming(PoisonValue::get(CB->getType()),
12472 NewCall->getParent());
12473 else
12474 llvm_unreachable("Call return should match or be void!");
12475 }
12476 A.changeAfterManifest(IRPosition::callsite_returned(*CB), *PHI);
12477 }
12478
12479 A.deleteAfterManifest(*CB);
12480 Changed = ChangeStatus::CHANGED;
12481
12482 return Changed;
12483 }
12484
12485 /// See AbstractAttribute::getAsStr().
12486 const std::string getAsStr(Attributor *A) const override {
12487 return std::string(AllCalleesKnown ? "eliminate" : "specialize") +
12488 " indirect call site with " + std::to_string(AssumedCallees.size()) +
12489 " functions";
12490 }
12491
12492 void trackStatistics() const override {
12493 if (AllCalleesKnown) {
12495 Eliminated, CallSites,
12496 "Number of indirect call sites eliminated via specialization")
12497 } else {
12498 STATS_DECLTRACK(Specialized, CallSites,
12499 "Number of indirect call sites specialized")
12500 }
12501 }
12502
12503 bool foreachCallee(function_ref<bool(Function *)> CB) const override {
12504 return isValidState() && AllCalleesKnown && all_of(AssumedCallees, CB);
12505 }
12506
12507private:
12508 /// Map to remember filter results.
12510
12511 /// If the !callee metadata was present, this set will contain all potential
12512 /// callees (superset).
12513 SmallSetVector<Function *, 4> PotentialCallees;
12514
12515 /// This set contains all currently assumed calllees, which might grow over
12516 /// time.
12517 SmallSetVector<Function *, 4> AssumedCallees;
12518
12519 /// Flag to indicate if all possible callees are in the AssumedCallees set or
12520 /// if there could be others.
12521 bool AllCalleesKnown = true;
12522};
12523} // namespace
12524
12525/// ------------------------ Address Space ------------------------------------
12526namespace {
12527
12528template <typename InstType>
12529static bool makeChange(Attributor &A, InstType *MemInst, const Use &U,
12530 Value *OriginalValue, PointerType *NewPtrTy,
12531 bool UseOriginalValue) {
12532 if (U.getOperandNo() != InstType::getPointerOperandIndex())
12533 return false;
12534
12535 if (MemInst->isVolatile()) {
12536 auto *TTI = A.getInfoCache().getAnalysisResultForFunction<TargetIRAnalysis>(
12537 *MemInst->getFunction());
12538 unsigned NewAS = NewPtrTy->getPointerAddressSpace();
12539 if (!TTI || !TTI->hasVolatileVariant(MemInst, NewAS))
12540 return false;
12541 }
12542
12543 if (UseOriginalValue) {
12544 A.changeUseAfterManifest(const_cast<Use &>(U), *OriginalValue);
12545 return true;
12546 }
12547
12548 Instruction *CastInst = new AddrSpaceCastInst(OriginalValue, NewPtrTy);
12549 CastInst->insertBefore(MemInst);
12550 A.changeUseAfterManifest(const_cast<Use &>(U), *CastInst);
12551 return true;
12552}
12553
12554struct AAAddressSpaceImpl : public AAAddressSpace {
12555 AAAddressSpaceImpl(const IRPosition &IRP, Attributor &A)
12556 : AAAddressSpace(IRP, A) {}
12557
12558 uint32_t getAddressSpace() const override {
12559 assert(isValidState() && "the AA is invalid");
12560 return AssumedAddressSpace;
12561 }
12562
12563 /// See AbstractAttribute::initialize(...).
12564 void initialize(Attributor &A) override {
12565 assert(getAssociatedType()->isPtrOrPtrVectorTy() &&
12566 "Associated value is not a pointer");
12567
12568 if (!A.getInfoCache().getFlatAddressSpace().has_value()) {
12569 indicatePessimisticFixpoint();
12570 return;
12571 }
12572
12573 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
12574 unsigned AS = getAssociatedType()->getPointerAddressSpace();
12575 if (AS != FlatAS) {
12576 [[maybe_unused]] bool R = takeAddressSpace(AS);
12577 assert(R && "The take should happen");
12578 indicateOptimisticFixpoint();
12579 }
12580 }
12581
12582 ChangeStatus updateImpl(Attributor &A) override {
12583 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
12584 uint32_t OldAddressSpace = AssumedAddressSpace;
12585
12586 auto CheckAddressSpace = [&](Value &Obj) {
12587 if (isa<UndefValue>(&Obj))
12588 return true;
12589 // If an argument in flat address space only has addrspace cast uses, and
12590 // those casts are same, then we take the dst addrspace.
12591 if (auto *Arg = dyn_cast<Argument>(&Obj)) {
12592 if (Arg->getType()->getPointerAddressSpace() == FlatAS) {
12593 unsigned CastAddrSpace = FlatAS;
12594 for (auto *U : Arg->users()) {
12595 auto *ASCI = dyn_cast<AddrSpaceCastInst>(U);
12596 if (!ASCI)
12597 return takeAddressSpace(Obj.getType()->getPointerAddressSpace());
12598 if (CastAddrSpace != FlatAS &&
12599 CastAddrSpace != ASCI->getDestAddressSpace())
12600 return false;
12601 CastAddrSpace = ASCI->getDestAddressSpace();
12602 }
12603 if (CastAddrSpace != FlatAS)
12604 return takeAddressSpace(CastAddrSpace);
12605 }
12606 }
12607 return takeAddressSpace(Obj.getType()->getPointerAddressSpace());
12608 };
12609
12610 auto *AUO = A.getOrCreateAAFor<AAUnderlyingObjects>(getIRPosition(), this,
12611 DepClassTy::REQUIRED);
12612 if (!AUO->forallUnderlyingObjects(CheckAddressSpace))
12613 return indicatePessimisticFixpoint();
12614
12615 return OldAddressSpace == AssumedAddressSpace ? ChangeStatus::UNCHANGED
12616 : ChangeStatus::CHANGED;
12617 }
12618
12619 /// See AbstractAttribute::manifest(...).
12620 ChangeStatus manifest(Attributor &A) override {
12621 unsigned NewAS = getAddressSpace();
12622
12623 if (NewAS == InvalidAddressSpace ||
12624 NewAS == getAssociatedType()->getPointerAddressSpace())
12625 return ChangeStatus::UNCHANGED;
12626
12627 unsigned FlatAS = A.getInfoCache().getFlatAddressSpace().value();
12628
12629 Value *AssociatedValue = &getAssociatedValue();
12630 Value *OriginalValue = peelAddrspacecast(AssociatedValue, FlatAS);
12631
12632 PointerType *NewPtrTy =
12633 PointerType::get(getAssociatedType()->getContext(), NewAS);
12634 bool UseOriginalValue =
12635 OriginalValue->getType()->getPointerAddressSpace() == NewAS;
12636
12637 bool Changed = false;
12638
12639 auto Pred = [&](const Use &U, bool &) {
12640 if (U.get() != AssociatedValue)
12641 return true;
12642 auto *Inst = dyn_cast<Instruction>(U.getUser());
12643 if (!Inst)
12644 return true;
12645 // This is a WA to make sure we only change uses from the corresponding
12646 // CGSCC if the AA is run on CGSCC instead of the entire module.
12647 if (!A.isRunOn(Inst->getFunction()))
12648 return true;
12649 if (auto *LI = dyn_cast<LoadInst>(Inst)) {
12650 Changed |=
12651 makeChange(A, LI, U, OriginalValue, NewPtrTy, UseOriginalValue);
12652 } else if (auto *SI = dyn_cast<StoreInst>(Inst)) {
12653 Changed |=
12654 makeChange(A, SI, U, OriginalValue, NewPtrTy, UseOriginalValue);
12655 } else if (auto *RMW = dyn_cast<AtomicRMWInst>(Inst)) {
12656 Changed |=
12657 makeChange(A, RMW, U, OriginalValue, NewPtrTy, UseOriginalValue);
12658 } else if (auto *CmpX = dyn_cast<AtomicCmpXchgInst>(Inst)) {
12659 Changed |=
12660 makeChange(A, CmpX, U, OriginalValue, NewPtrTy, UseOriginalValue);
12661 }
12662 return true;
12663 };
12664
12665 // It doesn't matter if we can't check all uses as we can simply
12666 // conservatively ignore those that can not be visited.
12667 (void)A.checkForAllUses(Pred, *this, getAssociatedValue(),
12668 /* CheckBBLivenessOnly */ true);
12669
12670 return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
12671 }
12672
12673 /// See AbstractAttribute::getAsStr().
12674 const std::string getAsStr(Attributor *A) const override {
12675 if (!isValidState())
12676 return "addrspace(<invalid>)";
12677 return "addrspace(" +
12678 (AssumedAddressSpace == InvalidAddressSpace
12679 ? "none"
12680 : std::to_string(AssumedAddressSpace)) +
12681 ")";
12682 }
12683
12684private:
12685 uint32_t AssumedAddressSpace = InvalidAddressSpace;
12686
12687 bool takeAddressSpace(uint32_t AS) {
12688 if (AssumedAddressSpace == InvalidAddressSpace) {
12689 AssumedAddressSpace = AS;
12690 return true;
12691 }
12692 return AssumedAddressSpace == AS;
12693 }
12694
12695 static Value *peelAddrspacecast(Value *V, unsigned FlatAS) {
12696 if (auto *I = dyn_cast<AddrSpaceCastInst>(V)) {
12697 assert(I->getSrcAddressSpace() != FlatAS &&
12698 "there should not be flat AS -> non-flat AS");
12699 return I->getPointerOperand();
12700 }
12701 if (auto *C = dyn_cast<ConstantExpr>(V))
12702 if (C->getOpcode() == Instruction::AddrSpaceCast) {
12703 assert(C->getOperand(0)->getType()->getPointerAddressSpace() !=
12704 FlatAS &&
12705 "there should not be flat AS -> non-flat AS X");
12706 return C->getOperand(0);
12707 }
12708 return V;
12709 }
12710};
12711
12712struct AAAddressSpaceFloating final : AAAddressSpaceImpl {
12713 AAAddressSpaceFloating(const IRPosition &IRP, Attributor &A)
12714 : AAAddressSpaceImpl(IRP, A) {}
12715
12716 void trackStatistics() const override {
12718 }
12719};
12720
12721struct AAAddressSpaceReturned final : AAAddressSpaceImpl {
12722 AAAddressSpaceReturned(const IRPosition &IRP, Attributor &A)
12723 : AAAddressSpaceImpl(IRP, A) {}
12724
12725 /// See AbstractAttribute::initialize(...).
12726 void initialize(Attributor &A) override {
12727 // TODO: we don't rewrite function argument for now because it will need to
12728 // rewrite the function signature and all call sites.
12729 (void)indicatePessimisticFixpoint();
12730 }
12731
12732 void trackStatistics() const override {
12733 STATS_DECLTRACK_FNRET_ATTR(addrspace);
12734 }
12735};
12736
12737struct AAAddressSpaceCallSiteReturned final : AAAddressSpaceImpl {
12738 AAAddressSpaceCallSiteReturned(const IRPosition &IRP, Attributor &A)
12739 : AAAddressSpaceImpl(IRP, A) {}
12740
12741 void trackStatistics() const override {
12742 STATS_DECLTRACK_CSRET_ATTR(addrspace);
12743 }
12744};
12745
12746struct AAAddressSpaceArgument final : AAAddressSpaceImpl {
12747 AAAddressSpaceArgument(const IRPosition &IRP, Attributor &A)
12748 : AAAddressSpaceImpl(IRP, A) {}
12749
12750 void trackStatistics() const override { STATS_DECLTRACK_ARG_ATTR(addrspace); }
12751};
12752
12753struct AAAddressSpaceCallSiteArgument final : AAAddressSpaceImpl {
12754 AAAddressSpaceCallSiteArgument(const IRPosition &IRP, Attributor &A)
12755 : AAAddressSpaceImpl(IRP, A) {}
12756
12757 /// See AbstractAttribute::initialize(...).
12758 void initialize(Attributor &A) override {
12759 // TODO: we don't rewrite call site argument for now because it will need to
12760 // rewrite the function signature of the callee.
12761 (void)indicatePessimisticFixpoint();
12762 }
12763
12764 void trackStatistics() const override {
12765 STATS_DECLTRACK_CSARG_ATTR(addrspace);
12766 }
12767};
12768} // namespace
12769
12770/// ----------- Allocation Info ----------
12771namespace {
12772struct AAAllocationInfoImpl : public AAAllocationInfo {
12773 AAAllocationInfoImpl(const IRPosition &IRP, Attributor &A)
12774 : AAAllocationInfo(IRP, A) {}
12775
12776 std::optional<TypeSize> getAllocatedSize() const override {
12777 assert(isValidState() && "the AA is invalid");
12778 return AssumedAllocatedSize;
12779 }
12780
12781 std::optional<TypeSize> findInitialAllocationSize(Instruction *I,
12782 const DataLayout &DL) {
12783
12784 // TODO: implement case for malloc like instructions
12785 switch (I->getOpcode()) {
12786 case Instruction::Alloca: {
12787 AllocaInst *AI = cast<AllocaInst>(I);
12788 return AI->getAllocationSize(DL);
12789 }
12790 default:
12791 return std::nullopt;
12792 }
12793 }
12794
12795 ChangeStatus updateImpl(Attributor &A) override {
12796
12797 const IRPosition &IRP = getIRPosition();
12798 Instruction *I = IRP.getCtxI();
12799
12800 // TODO: update check for malloc like calls
12801 if (!isa<AllocaInst>(I))
12802 return indicatePessimisticFixpoint();
12803
12804 bool IsKnownNoCapture;
12805 if (!AA::hasAssumedIRAttr<Attribute::NoCapture>(
12806 A, this, IRP, DepClassTy::OPTIONAL, IsKnownNoCapture))
12807 return indicatePessimisticFixpoint();
12808
12809 const AAPointerInfo *PI =
12810 A.getOrCreateAAFor<AAPointerInfo>(IRP, *this, DepClassTy::REQUIRED);
12811
12812 if (!PI)
12813 return indicatePessimisticFixpoint();
12814
12815 if (!PI->getState().isValidState() || PI->reachesReturn())
12816 return indicatePessimisticFixpoint();
12817
12818 const DataLayout &DL = A.getDataLayout();
12819 const auto AllocationSize = findInitialAllocationSize(I, DL);
12820
12821 // If allocation size is nullopt, we give up.
12822 if (!AllocationSize)
12823 return indicatePessimisticFixpoint();
12824
12825 // For zero sized allocations, we give up.
12826 // Since we can't reduce further
12827 if (*AllocationSize == 0)
12828 return indicatePessimisticFixpoint();
12829
12830 int64_t BinSize = PI->numOffsetBins();
12831
12832 // TODO: implement for multiple bins
12833 if (BinSize > 1)
12834 return indicatePessimisticFixpoint();
12835
12836 if (BinSize == 0) {
12837 auto NewAllocationSize = std::optional<TypeSize>(TypeSize(0, false));
12838 if (!changeAllocationSize(NewAllocationSize))
12839 return ChangeStatus::UNCHANGED;
12840 return ChangeStatus::CHANGED;
12841 }
12842
12843 // TODO: refactor this to be part of multiple bin case
12844 const auto &It = PI->begin();
12845
12846 // TODO: handle if Offset is not zero
12847 if (It->first.Offset != 0)
12848 return indicatePessimisticFixpoint();
12849
12850 uint64_t SizeOfBin = It->first.Offset + It->first.Size;
12851
12852 if (SizeOfBin >= *AllocationSize)
12853 return indicatePessimisticFixpoint();
12854
12855 auto NewAllocationSize =
12856 std::optional<TypeSize>(TypeSize(SizeOfBin * 8, false));
12857
12858 if (!changeAllocationSize(NewAllocationSize))
12859 return ChangeStatus::UNCHANGED;
12860
12861 return ChangeStatus::CHANGED;
12862 }
12863
12864 /// See AbstractAttribute::manifest(...).
12865 ChangeStatus manifest(Attributor &A) override {
12866
12867 assert(isValidState() &&
12868 "Manifest should only be called if the state is valid.");
12869
12870 Instruction *I = getIRPosition().getCtxI();
12871
12872 auto FixedAllocatedSizeInBits = getAllocatedSize()->getFixedValue();
12873
12874 unsigned long NumBytesToAllocate = (FixedAllocatedSizeInBits + 7) / 8;
12875
12876 switch (I->getOpcode()) {
12877 // TODO: add case for malloc like calls
12878 case Instruction::Alloca: {
12879
12880 AllocaInst *AI = cast<AllocaInst>(I);
12881
12882 Type *CharType = Type::getInt8Ty(I->getContext());
12883
12884 auto *NumBytesToValue =
12885 ConstantInt::get(I->getContext(), APInt(32, NumBytesToAllocate));
12886
12887 BasicBlock::iterator insertPt = AI->getIterator();
12888 insertPt = std::next(insertPt);
12889 AllocaInst *NewAllocaInst =
12890 new AllocaInst(CharType, AI->getAddressSpace(), NumBytesToValue,
12891 AI->getAlign(), AI->getName(), insertPt);
12892
12893 if (A.changeAfterManifest(IRPosition::inst(*AI), *NewAllocaInst))
12894 return ChangeStatus::CHANGED;
12895
12896 break;
12897 }
12898 default:
12899 break;
12900 }
12901
12902 return ChangeStatus::UNCHANGED;
12903 }
12904
12905 /// See AbstractAttribute::getAsStr().
12906 const std::string getAsStr(Attributor *A) const override {
12907 if (!isValidState())
12908 return "allocationinfo(<invalid>)";
12909 return "allocationinfo(" +
12910 (AssumedAllocatedSize == HasNoAllocationSize
12911 ? "none"
12912 : std::to_string(AssumedAllocatedSize->getFixedValue())) +
12913 ")";
12914 }
12915
12916private:
12917 std::optional<TypeSize> AssumedAllocatedSize = HasNoAllocationSize;
12918
12919 // Maintain the computed allocation size of the object.
12920 // Returns (bool) weather the size of the allocation was modified or not.
12921 bool changeAllocationSize(std::optional<TypeSize> Size) {
12922 if (AssumedAllocatedSize == HasNoAllocationSize ||
12923 AssumedAllocatedSize != Size) {
12924 AssumedAllocatedSize = Size;
12925 return true;
12926 }
12927 return false;
12928 }
12929};
12930
12931struct AAAllocationInfoFloating : AAAllocationInfoImpl {
12932 AAAllocationInfoFloating(const IRPosition &IRP, Attributor &A)
12933 : AAAllocationInfoImpl(IRP, A) {}
12934
12935 void trackStatistics() const override {
12936 STATS_DECLTRACK_FLOATING_ATTR(allocationinfo);
12937 }
12938};
12939
12940struct AAAllocationInfoReturned : AAAllocationInfoImpl {
12941 AAAllocationInfoReturned(const IRPosition &IRP, Attributor &A)
12942 : AAAllocationInfoImpl(IRP, A) {}
12943
12944 /// See AbstractAttribute::initialize(...).
12945 void initialize(Attributor &A) override {
12946 // TODO: we don't rewrite function argument for now because it will need to
12947 // rewrite the function signature and all call sites
12948 (void)indicatePessimisticFixpoint();
12949 }
12950
12951 void trackStatistics() const override {
12952 STATS_DECLTRACK_FNRET_ATTR(allocationinfo);
12953 }
12954};
12955
12956struct AAAllocationInfoCallSiteReturned : AAAllocationInfoImpl {
12957 AAAllocationInfoCallSiteReturned(const IRPosition &IRP, Attributor &A)
12958 : AAAllocationInfoImpl(IRP, A) {}
12959
12960 void trackStatistics() const override {
12961 STATS_DECLTRACK_CSRET_ATTR(allocationinfo);
12962 }
12963};
12964
12965struct AAAllocationInfoArgument : AAAllocationInfoImpl {
12966 AAAllocationInfoArgument(const IRPosition &IRP, Attributor &A)
12967 : AAAllocationInfoImpl(IRP, A) {}
12968
12969 void trackStatistics() const override {
12970 STATS_DECLTRACK_ARG_ATTR(allocationinfo);
12971 }
12972};
12973
12974struct AAAllocationInfoCallSiteArgument : AAAllocationInfoImpl {
12975 AAAllocationInfoCallSiteArgument(const IRPosition &IRP, Attributor &A)
12976 : AAAllocationInfoImpl(IRP, A) {}
12977
12978 /// See AbstractAttribute::initialize(...).
12979 void initialize(Attributor &A) override {
12980
12981 (void)indicatePessimisticFixpoint();
12982 }
12983
12984 void trackStatistics() const override {
12985 STATS_DECLTRACK_CSARG_ATTR(allocationinfo);
12986 }
12987};
12988} // namespace
12989
12990const char AANoUnwind::ID = 0;
12991const char AANoSync::ID = 0;
12992const char AANoFree::ID = 0;
12993const char AANonNull::ID = 0;
12994const char AAMustProgress::ID = 0;
12995const char AANoRecurse::ID = 0;
12996const char AANonConvergent::ID = 0;
12997const char AAWillReturn::ID = 0;
12998const char AAUndefinedBehavior::ID = 0;
12999const char AANoAlias::ID = 0;
13000const char AAIntraFnReachability::ID = 0;
13001const char AANoReturn::ID = 0;
13002const char AAIsDead::ID = 0;
13003const char AADereferenceable::ID = 0;
13004const char AAAlign::ID = 0;
13005const char AAInstanceInfo::ID = 0;
13006const char AANoCapture::ID = 0;
13007const char AAValueSimplify::ID = 0;
13008const char AAHeapToStack::ID = 0;
13009const char AAPrivatizablePtr::ID = 0;
13010const char AAMemoryBehavior::ID = 0;
13011const char AAMemoryLocation::ID = 0;
13012const char AAValueConstantRange::ID = 0;
13013const char AAPotentialConstantValues::ID = 0;
13014const char AAPotentialValues::ID = 0;
13015const char AANoUndef::ID = 0;
13016const char AANoFPClass::ID = 0;
13017const char AACallEdges::ID = 0;
13018const char AAInterFnReachability::ID = 0;
13019const char AAPointerInfo::ID = 0;
13020const char AAAssumptionInfo::ID = 0;
13021const char AAUnderlyingObjects::ID = 0;
13022const char AAAddressSpace::ID = 0;
13023const char AAAllocationInfo::ID = 0;
13024const char AAIndirectCallInfo::ID = 0;
13025const char AAGlobalValueInfo::ID = 0;
13026const char AADenormalFPMath::ID = 0;
13027
13028// Macro magic to create the static generator function for attributes that
13029// follow the naming scheme.
13030
13031#define SWITCH_PK_INV(CLASS, PK, POS_NAME) \
13032 case IRPosition::PK: \
13033 llvm_unreachable("Cannot create " #CLASS " for a " POS_NAME " position!");
13034
13035#define SWITCH_PK_CREATE(CLASS, IRP, PK, SUFFIX) \
13036 case IRPosition::PK: \
13037 AA = new (A.Allocator) CLASS##SUFFIX(IRP, A); \
13038 ++NumAAs; \
13039 break;
13040
13041#define CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13042 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13043 CLASS *AA = nullptr; \
13044 switch (IRP.getPositionKind()) { \
13045 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13046 SWITCH_PK_INV(CLASS, IRP_FLOAT, "floating") \
13047 SWITCH_PK_INV(CLASS, IRP_ARGUMENT, "argument") \
13048 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13049 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_RETURNED, "call site returned") \
13050 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_ARGUMENT, "call site argument") \
13051 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13052 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13053 } \
13054 return *AA; \
13055 }
13056
13057#define CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13058 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13059 CLASS *AA = nullptr; \
13060 switch (IRP.getPositionKind()) { \
13061 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13062 SWITCH_PK_INV(CLASS, IRP_FUNCTION, "function") \
13063 SWITCH_PK_INV(CLASS, IRP_CALL_SITE, "call site") \
13064 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13065 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13066 SWITCH_PK_CREATE(CLASS, IRP, IRP_RETURNED, Returned) \
13067 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13068 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13069 } \
13070 return *AA; \
13071 }
13072
13073#define CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION(POS, SUFFIX, CLASS) \
13074 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13075 CLASS *AA = nullptr; \
13076 switch (IRP.getPositionKind()) { \
13077 SWITCH_PK_CREATE(CLASS, IRP, POS, SUFFIX) \
13078 default: \
13079 llvm_unreachable("Cannot create " #CLASS " for position otherthan " #POS \
13080 " position!"); \
13081 } \
13082 return *AA; \
13083 }
13084
13085#define CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13086 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13087 CLASS *AA = nullptr; \
13088 switch (IRP.getPositionKind()) { \
13089 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13090 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13091 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13092 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13093 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13094 SWITCH_PK_CREATE(CLASS, IRP, IRP_RETURNED, Returned) \
13095 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13096 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13097 } \
13098 return *AA; \
13099 }
13100
13101#define CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13102 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13103 CLASS *AA = nullptr; \
13104 switch (IRP.getPositionKind()) { \
13105 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13106 SWITCH_PK_INV(CLASS, IRP_ARGUMENT, "argument") \
13107 SWITCH_PK_INV(CLASS, IRP_FLOAT, "floating") \
13108 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13109 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_RETURNED, "call site returned") \
13110 SWITCH_PK_INV(CLASS, IRP_CALL_SITE_ARGUMENT, "call site argument") \
13111 SWITCH_PK_INV(CLASS, IRP_CALL_SITE, "call site") \
13112 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13113 } \
13114 return *AA; \
13115 }
13116
13117#define CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS) \
13118 CLASS &CLASS::createForPosition(const IRPosition &IRP, Attributor &A) { \
13119 CLASS *AA = nullptr; \
13120 switch (IRP.getPositionKind()) { \
13121 SWITCH_PK_INV(CLASS, IRP_INVALID, "invalid") \
13122 SWITCH_PK_INV(CLASS, IRP_RETURNED, "returned") \
13123 SWITCH_PK_CREATE(CLASS, IRP, IRP_FUNCTION, Function) \
13124 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE, CallSite) \
13125 SWITCH_PK_CREATE(CLASS, IRP, IRP_FLOAT, Floating) \
13126 SWITCH_PK_CREATE(CLASS, IRP, IRP_ARGUMENT, Argument) \
13127 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_RETURNED, CallSiteReturned) \
13128 SWITCH_PK_CREATE(CLASS, IRP, IRP_CALL_SITE_ARGUMENT, CallSiteArgument) \
13129 } \
13130 return *AA; \
13131 }
13132
13142
13158
13163
13168
13175
13177
13178#undef CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION
13179#undef CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION
13180#undef CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION
13181#undef CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION
13182#undef CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION
13183#undef CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION
13184#undef SWITCH_PK_CREATE
13185#undef SWITCH_PK_INV
#define Success
AMDGPU Register Bank Select
Rewrite undef for PHI
This file implements a class to represent arbitrary precision integral constant values and operations...
ReachingDefAnalysis InstSet & ToRemove
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file contains the simple types necessary to represent the attributes associated with functions a...
#define STATS_DECLTRACK(NAME, TYPE, MSG)
static std::optional< Constant * > askForAssumedConstant(Attributor &A, const AbstractAttribute &QueryingAA, const IRPosition &IRP, Type &Ty)
static cl::opt< unsigned, true > MaxPotentialValues("attributor-max-potential-values", cl::Hidden, cl::desc("Maximum number of potential values to be " "tracked for each position."), cl::location(llvm::PotentialConstantIntValuesState::MaxPotentialValues), cl::init(7))
static const Value * getPointerOperand(const Instruction *I, bool AllowVolatile)
Get pointer operand of memory accessing instruction.
static void clampReturnedValueStates(Attributor &A, const AAType &QueryingAA, StateType &S, const IRPosition::CallBaseContext *CBContext=nullptr)
Clamp the information known for all returned values of a function (identified by QueryingAA) into S.
#define STATS_DECLTRACK_FN_ATTR(NAME)
#define CREATE_FUNCTION_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static cl::opt< int > MaxPotentialValuesIterations("attributor-max-potential-values-iterations", cl::Hidden, cl::desc("Maximum number of iterations we keep dismantling potential values."), cl::init(64))
#define STATS_DECLTRACK_CS_ATTR(NAME)
#define PIPE_OPERATOR(CLASS)
#define DefineKeys(ToTy)
static bool mayBeInCycle(const CycleInfo *CI, const Instruction *I, bool HeaderOnly, Cycle **CPtr=nullptr)
#define STATS_DECLTRACK_ARG_ATTR(NAME)
static const Value * stripAndAccumulateOffsets(Attributor &A, const AbstractAttribute &QueryingAA, const Value *Val, const DataLayout &DL, APInt &Offset, bool GetMinOffset, bool AllowNonInbounds, bool UseAssumed=false)
#define STATS_DECLTRACK_CSRET_ATTR(NAME)
static cl::opt< bool > ManifestInternal("attributor-manifest-internal", cl::Hidden, cl::desc("Manifest Attributor internal string attributes."), cl::init(false))
static Value * constructPointer(Value *Ptr, int64_t Offset, IRBuilder< NoFolder > &IRB)
Helper function to create a pointer based on Ptr, and advanced by Offset bytes.
#define CREATE_NON_RET_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
#define BUILD_STAT_NAME(NAME, TYPE)
static bool isDenselyPacked(Type *Ty, const DataLayout &DL)
Checks if a type could have padding bytes.
#define CREATE_VALUE_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static const Value * getMinimalBaseOfPointer(Attributor &A, const AbstractAttribute &QueryingAA, const Value *Ptr, int64_t &BytesOffset, const DataLayout &DL, bool AllowNonInbounds=false)
#define STATS_DECLTRACK_FNRET_ATTR(NAME)
#define STATS_DECLTRACK_CSARG_ATTR(NAME)
#define CREATE_ABSTRACT_ATTRIBUTE_FOR_ONE_POSITION(POS, SUFFIX, CLASS)
#define CREATE_ALL_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
static cl::opt< int > MaxHeapToStackSize("max-heap-to-stack-size", cl::init(128), cl::Hidden)
#define CREATE_FUNCTION_ONLY_ABSTRACT_ATTRIBUTE_FOR_POSITION(CLASS)
#define STATS_DECLTRACK_FLOATING_ATTR(NAME)
#define STATS_DECL(NAME, TYPE, MSG)
basic Basic Alias true
BlockVerifier::State From
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static bool isReachableImpl(SmallVectorImpl< BasicBlock * > &Worklist, const StopSetT &StopSet, const SmallPtrSetImpl< BasicBlock * > *ExclusionSet, const DominatorTree *DT, const LoopInfo *LI)
Definition: CFG.cpp:134
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file declares an analysis pass that computes CycleInfo for LLVM IR, specialized from GenericCycl...
DXIL Resource Access
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
Given that RA is a live propagate it s liveness to any other values it uses(according to Uses). void DeadArgumentEliminationPass
Performs the initial survey of the specified function
Given that RA is a live value
#define LLVM_DEBUG(...)
Definition: Debug.h:106
This file defines DenseMapInfo traits for DenseMap.
T Content
uint32_t Index
uint64_t Size
#define Check(C,...)
Hexagon Common GEP
IRTranslator LLVM IR MI
#define F(x, y, z)
Definition: MD5.cpp:55
#define I(x, y, z)
Definition: MD5.cpp:58
This file implements a map that provides insertion order iteration.
#define T
#define T1
static unsigned getAddressSpace(const Value *V, unsigned MaxLookup)
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
uint64_t IntrinsicInst * II
if(PassOpts->AAPipeline)
static StringRef getName(Value *V)
Basic Register Allocator
static cl::opt< RegAllocEvictionAdvisorAnalysis::AdvisorMode > Mode("regalloc-enable-advisor", cl::Hidden, cl::init(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default), cl::desc("Enable regalloc advisor mode"), cl::values(clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default, "default", "Default"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Release, "release", "precompiled"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Development, "development", "for training")))
Remove Loads Into Fake Uses
This builds on the llvm/ADT/GraphTraits.h file to find the strongly connected components (SCCs) of a ...
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
bool IsDead
This file contains some templates that are useful if you are working with the STL at all.
raw_pwrite_stream & OS
This file defines generic set operations that may be used on set's of different types,...
This file implements a set that has insertion order iteration characteristics.
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition: Statistic.h:166
This file contains some functions that are useful when dealing with strings.
static SymbolRef::Type getType(const Symbol *Sym)
Definition: TapiFile.cpp:39
static void initialize(TargetLibraryInfoImpl &TLI, const Triple &T, ArrayRef< StringLiteral > StandardNames)
Initialize the set of available library functions based on the specified target triple.
@ Floating
This pass exposes codegen information to IR-level passes.
static unsigned getBitWidth(Type *Ty, const DataLayout &DL)
Returns the bitwidth of the given scalar or pointer type.
Value * RHS
Value * LHS
static unsigned getSize(unsigned Kind)
AACallGraphNode * operator*() const
A manager for alias analyses.
bool isNoAlias(const MemoryLocation &LocA, const MemoryLocation &LocB)
A trivial helper function to check to see if the specified pointers are no-alias.
Class for arbitrary precision integers.
Definition: APInt.h:78
int64_t getSExtValue() const
Get sign extended value.
Definition: APInt.h:1542
AbstractCallSite.
CallBase * getInstruction() const
Return the underlying instruction.
bool isCallbackCall() const
Return true if this ACS represents a callback call.
bool isDirectCall() const
Return true if this ACS represents a direct call.
static void getCallbackUses(const CallBase &CB, SmallVectorImpl< const Use * > &CallbackUses)
Add operand uses of CB that represent callback uses into CallbackUses.
int getCallArgOperandNo(Argument &Arg) const
Return the operand index of the underlying instruction associated with Arg.
This class represents a conversion between pointers from one address space to another.
an instruction to allocate memory on the stack
Definition: Instructions.h:63
Align getAlign() const
Return the alignment of the memory that is being allocated by the instruction.
Definition: Instructions.h:124
unsigned getAddressSpace() const
Return the address space for the allocation.
Definition: Instructions.h:104
std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
This class represents an incoming formal argument to a Function.
Definition: Argument.h:31
bool hasPointeeInMemoryValueAttr() const
Return true if this argument has the byval, sret, inalloca, preallocated, or byref attribute.
Definition: Function.cpp:182
bool hasReturnedAttr() const
Return true if this argument has the returned attribute.
Definition: Function.cpp:307
bool hasByValAttr() const
Return true if this argument has the byval attribute.
Definition: Function.cpp:144
const Function * getParent() const
Definition: Argument.h:43
unsigned getArgNo() const
Return the index of this formal argument in its containing function.
Definition: Argument.h:49
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition: ArrayRef.h:41
iterator end() const
Definition: ArrayRef.h:157
iterator begin() const
Definition: ArrayRef.h:156
A function analysis which provides an AssumptionCache.
A cache of @llvm.assume calls within a function.
static Attribute get(LLVMContext &Context, AttrKind Kind, uint64_t Val=0)
Return a uniquified Attribute object.
Definition: Attributes.cpp:95
static Attribute getWithDereferenceableBytes(LLVMContext &Context, uint64_t Bytes)
Definition: Attributes.cpp:244
FPClassTest getNoFPClass() const
Return the FPClassTest for nofpclass.
Definition: Attributes.cpp:496
Attribute::AttrKind getKindAsEnum() const
Return the attribute's kind as an enum (Attribute::AttrKind).
Definition: Attributes.cpp:364
MemoryEffects getMemoryEffects() const
Returns memory effects.
Definition: Attributes.cpp:484
static Attribute getWithDereferenceableOrNullBytes(LLVMContext &Context, uint64_t Bytes)
Definition: Attributes.cpp:250
static Attribute getWithNoFPClass(LLVMContext &Context, FPClassTest Mask)
Definition: Attributes.cpp:286
AttrKind
This enumeration lists the attributes that can be associated with parameters, function results,...
Definition: Attributes.h:86
static bool isEnumAttrKind(AttrKind Kind)
Definition: Attributes.h:99
static Attribute getWithMemoryEffects(LLVMContext &Context, MemoryEffects ME)
Definition: Attributes.cpp:281
static Attribute getWithAlignment(LLVMContext &Context, Align Alignment)
Return a uniquified Attribute object that has the specific alignment set.
Definition: Attributes.cpp:234
LLVM Basic Block Representation.
Definition: BasicBlock.h:61
const_iterator getFirstInsertionPt() const
Returns an iterator to the first instruction in this block that is suitable for inserting a non-PHI i...
Definition: BasicBlock.cpp:416
const Instruction & front() const
Definition: BasicBlock.h:471
static BasicBlock * Create(LLVMContext &Context, const Twine &Name="", Function *Parent=nullptr, BasicBlock *InsertBefore=nullptr)
Creates a new BasicBlock.
Definition: BasicBlock.h:212
const Function * getParent() const
Return the enclosing method, or null if none.
Definition: BasicBlock.h:219
InstListType::iterator iterator
Instruction iterators...
Definition: BasicBlock.h:177
const Instruction * getTerminator() const LLVM_READONLY
Returns the terminator instruction if the block is well formed or null if the block is not well forme...
Definition: BasicBlock.h:239
BinaryOps getOpcode() const
Definition: InstrTypes.h:370
Conditional or Unconditional Branch instruction.
unsigned getNumSuccessors() const
static BranchInst * Create(BasicBlock *IfTrue, InsertPosition InsertBefore=nullptr)
BasicBlock * getSuccessor(unsigned i) const
Value * getCondition() const
Allocate memory in an ever growing pool, as if by bump-pointer.
Definition: Allocator.h:66
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
Definition: InstrTypes.h:1112
Function * getCalledFunction() const
Returns the function called, or null if this is an indirect function invocation or the function signa...
Definition: InstrTypes.h:1341
bool isMustTailCall() const
Tests if this call site must be tail call optimized.
bool isIndirectCall() const
Return true if the callsite is an indirect call.
bool isCallee(Value::const_user_iterator UI) const
Determine whether the passed iterator points to the callee operand's Use.
Definition: InstrTypes.h:1352
Value * getCalledOperand() const
Definition: InstrTypes.h:1334
const Use & getCalledOperandUse() const
Definition: InstrTypes.h:1336
const Use & getArgOperandUse(unsigned i) const
Wrappers for getting the Use of a call argument.
Definition: InstrTypes.h:1297
Value * getArgOperand(unsigned i) const
Definition: InstrTypes.h:1286
bool isBundleOperand(unsigned Idx) const
Return true if the operand at index Idx is a bundle operand.
Definition: InstrTypes.h:1981
bool isConvergent() const
Determine if the invoke is convergent.
Definition: InstrTypes.h:1932
FunctionType * getFunctionType() const
Definition: InstrTypes.h:1199
Intrinsic::ID getIntrinsicID() const
Returns the intrinsic ID of the intrinsic called or Intrinsic::not_intrinsic if the called function i...
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
Definition: InstrTypes.h:1277
unsigned getArgOperandNo(const Use *U) const
Given a use for a arg operand, get the arg operand number that corresponds to it.
Definition: InstrTypes.h:1317
unsigned arg_size() const
Definition: InstrTypes.h:1284
bool isArgOperand(const Use *U) const
Definition: InstrTypes.h:1306
Function * getCaller()
Helper to get the caller (the parent function).
This class represents a function call, abstracting a target machine's calling convention.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Definition: InstrTypes.h:444
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition: InstrTypes.h:608
bool isIntegerCast() const
There are several places where we need to know if a cast instruction only deals with integer source a...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition: InstrTypes.h:615
This class is the base class for the comparison instructions.
Definition: InstrTypes.h:661
bool isEquality() const
Determine if this is an equals/not equals predicate.
Definition: InstrTypes.h:913
bool isFalseWhenEqual() const
This is just a convenience.
Definition: InstrTypes.h:946
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition: InstrTypes.h:673
@ ICMP_EQ
equal
Definition: InstrTypes.h:694
@ ICMP_NE
not equal
Definition: InstrTypes.h:695
bool isTrueWhenEqual() const
This is just a convenience.
Definition: InstrTypes.h:940
Predicate getPredicate() const
Return the predicate for this instruction.
Definition: InstrTypes.h:763
static ConstantAsMetadata * get(Constant *C)
Definition: Metadata.h:528
A constant value that is initialized with an expression using other constant values.
Definition: Constants.h:1108
static Constant * getExtractElement(Constant *Vec, Constant *Idx, Type *OnlyIfReducedTy=nullptr)
Definition: Constants.cpp:2554
This is the shared class of boolean and integer constants.
Definition: Constants.h:83
static ConstantInt * getTrue(LLVMContext &Context)
Definition: Constants.cpp:866
This class represents a range of values.
Definition: ConstantRange.h:47
const APInt & getLower() const
Return the lower value for this range.
bool isFullSet() const
Return true if this set contains all of the elements possible for this data-type.
bool isEmptySet() const
Return true if this set contains no members.
APInt getSignedMin() const
Return the smallest signed value contained in the ConstantRange.
bool isSingleElement() const
Return true if this set contains exactly one member.
static ConstantRange makeAllowedICmpRegion(CmpInst::Predicate Pred, const ConstantRange &Other)
Produce the smallest range such that all values that may satisfy the given predicate with any value c...
const APInt & getUpper() const
Return the upper value for this range.
APInt getSignedMax() const
Return the largest signed value contained in the ConstantRange.
This is an important base class in LLVM.
Definition: Constant.h:42
Analysis pass which computes a CycleInfo.
Definition: CycleAnalysis.h:46
This class represents an Operation in the Expression.
A parsed version of the target data layout string in and methods for querying it.
Definition: DataLayout.h:63
A debug info location.
Definition: DebugLoc.h:33
iterator find(const_arg_type_t< KeyT > Val)
Definition: DenseMap.h:156
unsigned size() const
Definition: DenseMap.h:99
iterator begin()
Definition: DenseMap.h:75
size_type count(const_arg_type_t< KeyT > Val) const
Return 1 if the specified key is in the map, 0 otherwise.
Definition: DenseMap.h:152
iterator end()
Definition: DenseMap.h:84
bool contains(const_arg_type_t< KeyT > Val) const
Return true if the specified key is in the map, false otherwise.
Definition: DenseMap.h:147
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition: DenseMap.h:211
Implements a dense probed hash-table based set.
Definition: DenseSet.h:278
Analysis pass which computes a DominatorTree.
Definition: Dominators.h:279
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition: Dominators.h:162
bool dominates(const BasicBlock *BB, const Use &U) const
Return true if the (end of the) basic block BB dominates the use U.
Definition: Dominators.cpp:122
An instruction for ordering other memory operations.
Definition: Instructions.h:424
A handy container for a FunctionType+Callee-pointer pair, which can be passed around as a single enti...
Definition: DerivedTypes.h:170
const BasicBlock & getEntryBlock() const
Definition: Function.h:809
iterator_range< arg_iterator > args()
Definition: Function.h:892
const Function & getFunction() const
Definition: Function.h:171
size_t arg_size() const
Definition: Function.h:901
Argument * getArg(unsigned i) const
Definition: Function.h:886
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
Definition: Function.cpp:731
CycleT * getCycle(const BlockT *Block) const
Find the innermost cycle containing a given block.
A possibly irreducible generalization of a Loop.
bool isDeclaration() const
Return true if the primary definition of this global value is outside of the current translation unit...
Definition: Globals.cpp:296
bool hasLocalLinkage() const
Definition: GlobalValue.h:528
This instruction compares its operands according to the predicate given to the constructor.
static bool compare(const APInt &LHS, const APInt &RHS, ICmpInst::Predicate Pred)
Return result of LHS Pred RHS comparison.
Value * CreatePtrAdd(Value *Ptr, Value *Offset, const Twine &Name="", GEPNoWrapFlags NW=GEPNoWrapFlags::none())
Definition: IRBuilder.h:1986
ConstantInt * getInt64(uint64_t C)
Get a constant 64-bit value.
Definition: IRBuilder.h:510
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
Definition: IRBuilder.h:2704
Instruction * clone() const
Create a copy of 'this' instruction that is identical in all ways except the following:
bool isLifetimeStartOrEnd() const LLVM_READONLY
Return true if the instruction is a llvm.lifetime.start or llvm.lifetime.end marker.
bool mayReadOrWriteMemory() const
Return true if this instruction may read or write memory.
Definition: Instruction.h:763
bool mayWriteToMemory() const LLVM_READONLY
Return true if this instruction may modify memory.
void insertBefore(Instruction *InsertPos)
Insert an unlinked instruction into a basic block immediately before the specified instruction.
Definition: Instruction.cpp:99
InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
Definition: Instruction.cpp:94
const Function * getFunction() const
Return the function this instruction belongs to.
Definition: Instruction.cpp:72
BasicBlock * getSuccessor(unsigned Idx) const LLVM_READONLY
Return the specified successor. This instruction must be a terminator.
bool mayHaveSideEffects() const LLVM_READONLY
Return true if the instruction may have side effects.
bool isTerminator() const
Definition: Instruction.h:277
bool mayReadFromMemory() const LLVM_READONLY
Return true if this instruction may read memory.
const Instruction * getNextNonDebugInstruction(bool SkipPseudoOp=false) const
Return a pointer to the next non-debug instruction in the same basic block as 'this',...
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
Definition: Instruction.h:274
void setDebugLoc(DebugLoc Loc)
Set the debug location information for this instruction.
Definition: Instruction.h:472
const DataLayout & getDataLayout() const
Get the data layout of the module this instruction belongs to.
Definition: Instruction.cpp:76
A wrapper class for inspecting calls to intrinsic functions.
Definition: IntrinsicInst.h:48
Invoke instruction.
This is an important class for using LLVM in a threaded context.
Definition: LLVMContext.h:67
Analysis to compute lazy value information.
This pass computes, caches, and vends lazy value constraint information.
Definition: LazyValueInfo.h:32
ConstantRange getConstantRange(Value *V, Instruction *CxtI, bool UndefAllowed)
Return the ConstantRange constraint that is known to hold for the specified value at the specified in...
An instruction for reading from memory.
Definition: Instructions.h:176
Analysis pass that exposes the LoopInfo for a function.
Definition: LoopInfo.h:566
LoopT * getLoopFor(const BlockT *BB) const
Return the inner most loop that BB lives in.
Metadata node.
Definition: Metadata.h:1069
const MDOperand & getOperand(unsigned I) const
Definition: Metadata.h:1430
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
Definition: Metadata.h:1545
unsigned getNumOperands() const
Return number of MDNode operands.
Definition: Metadata.h:1436
This class implements a map that also provides access to all stored values in a deterministic order.
Definition: MapVector.h:36
bool empty() const
Definition: MapVector.h:79
static MemoryEffectsBase readOnly()
Create MemoryEffectsBase that can read any memory.
Definition: ModRef.h:122
bool doesNotAccessMemory() const
Whether this function accesses no memory.
Definition: ModRef.h:192
static MemoryEffectsBase argMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Create MemoryEffectsBase that can only access argument memory.
Definition: ModRef.h:132
static MemoryEffectsBase inaccessibleMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Create MemoryEffectsBase that can only access inaccessible memory.
Definition: ModRef.h:138
bool onlyAccessesInaccessibleMem() const
Whether this function only (at most) accesses inaccessible memory.
Definition: ModRef.h:211
ModRefInfo getModRef(Location Loc) const
Get ModRefInfo for the given Location.
Definition: ModRef.h:165
bool onlyAccessesArgPointees() const
Whether this function only (at most) accesses argument memory.
Definition: ModRef.h:201
bool onlyReadsMemory() const
Whether this function only (at most) reads memory.
Definition: ModRef.h:195
static MemoryEffectsBase writeOnly()
Create MemoryEffectsBase that can write any memory.
Definition: ModRef.h:127
static MemoryEffectsBase inaccessibleOrArgMemOnly(ModRefInfo MR=ModRefInfo::ModRef)
Create MemoryEffectsBase that can only access inaccessible or argument memory.
Definition: ModRef.h:145
static MemoryEffectsBase none()
Create MemoryEffectsBase that cannot read or write any memory.
Definition: ModRef.h:117
bool onlyAccessesInaccessibleOrArgMem() const
Whether this function only (at most) accesses argument and inaccessible memory.
Definition: ModRef.h:217
static MemoryEffectsBase unknown()
Create MemoryEffectsBase that can read and write any memory.
Definition: ModRef.h:112
static std::optional< MemoryLocation > getOrNone(const Instruction *Inst)
Root of the metadata hierarchy.
Definition: Metadata.h:62
A Module instance is used to store all the information related to an LLVM module.
Definition: Module.h:65
Evaluate the size and offset of an object pointed to by a Value*.
static SizeOffsetValue unknown()
Diagnostic information for missed-optimization remarks.
Diagnostic information for applied optimization remarks.
static PHINode * Create(Type *Ty, unsigned NumReservedValues, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
Constructors - NumReservedValues is a hint for the number of incoming edges that this phi node will h...
static PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
Definition: Constants.cpp:1878
Return a value (possibly void), from a function.
Value * getReturnValue() const
Convenience accessor. Returns null if there is no return value.
This class represents an analyzed expression in the program.
Analysis pass that exposes the ScalarEvolution for a function.
The main scalar evolution driver.
const SCEV * getSCEVAtScope(const SCEV *S, const Loop *L)
Return a SCEV expression for the specified value at the specified scope in the program.
const SCEV * getSCEV(Value *V)
Return a SCEV expression for the full generality of the specified expression.
unsigned getSmallConstantMaxTripCount(const Loop *L, SmallVectorImpl< const SCEVPredicate * > *Predicates=nullptr)
Returns the upper bound of the loop trip count as a normal unsigned value.
ConstantRange getUnsignedRange(const SCEV *S)
Determine the unsigned range for a particular SCEV.
This class represents the LLVM 'select' instruction.
A vector that has set insertion semantics.
Definition: SetVector.h:57
size_type size() const
Determine the number of elements in the SetVector.
Definition: SetVector.h:98
bool insert(const value_type &X)
Insert a new element into the SetVector.
Definition: SetVector.h:162
bool erase(PtrType Ptr)
Remove pointer from the set.
Definition: SmallPtrSet.h:401
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
Definition: SmallPtrSet.h:452
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
Definition: SmallPtrSet.h:384
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
Definition: SmallPtrSet.h:519
A SetVector that performs no allocations if smaller than a certain size.
Definition: SetVector.h:370
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
Definition: SmallSet.h:132
std::pair< const_iterator, bool > insert(const T &V)
insert - Insert an element into the set if it isn't already there.
Definition: SmallSet.h:181
bool empty() const
Definition: SmallVector.h:81
size_t size() const
Definition: SmallVector.h:78
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Definition: SmallVector.h:573
reference emplace_back(ArgTypes &&... Args)
Definition: SmallVector.h:937
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
Definition: SmallVector.h:683
void push_back(const T &Elt)
Definition: SmallVector.h:413
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1196
An instruction for storing to memory.
Definition: Instructions.h:292
StringRef - Represent a constant reference to a string, i.e.
Definition: StringRef.h:51
Used to lazily calculate structure layout information for a target machine, based on the DataLayout s...
Definition: DataLayout.h:567
TypeSize getElementOffset(unsigned Idx) const
Definition: DataLayout.h:596
TypeSize getElementOffsetInBits(unsigned Idx) const
Definition: DataLayout.h:601
Class to represent struct types.
Definition: DerivedTypes.h:218
unsigned getNumElements() const
Random access to the elements.
Definition: DerivedTypes.h:365
Type * getElementType(unsigned N) const
Definition: DerivedTypes.h:366
Multiway switch.
Analysis pass providing the TargetTransformInfo.
This pass provides access to the codegen interfaces that are needed for IR-level transformations.
bool areTypesABICompatible(const Function *Caller, const Function *Callee, const ArrayRef< Type * > &Types) const
bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const
Return true if the given instruction (assumed to be a memory access instruction) has a volatile varia...
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition: Twine.h:81
The instances of the Type class are immutable: once they are created, they are never changed.
Definition: Type.h:45
unsigned getIntegerBitWidth() const
bool isPointerTy() const
True if this is an instance of PointerType.
Definition: Type.h:264
static IntegerType * getInt1Ty(LLVMContext &C)
unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
bool isSized(SmallPtrSetImpl< Type * > *Visited=nullptr) const
Return true if it makes sense to take the size of this type.
Definition: Type.h:310
static IntegerType * getInt8Ty(LLVMContext &C)
bool isPtrOrPtrVectorTy() const
Return true if this is a pointer type or a vector of pointer types.
Definition: Type.h:267
static IntegerType * getInt32Ty(LLVMContext &C)
bool isIntegerTy() const
True if this is an instance of IntegerType.
Definition: Type.h:237
bool isVoidTy() const
Return true if this is 'void'.
Definition: Type.h:139
'undef' values are things that do not have specified contents.
Definition: Constants.h:1412
static UndefValue * get(Type *T)
Static factory methods - Return an 'undef' object of the specified type.
Definition: Constants.cpp:1859
This function has undefined behavior.
A Use represents the edge between a Value definition and its users.
Definition: Use.h:43
User * getUser() const
Returns the User that contains this Use.
Definition: Use.h:72
Value * get() const
Definition: Use.h:66
bool replaceUsesOfWith(Value *From, Value *To)
Replace uses of one Value with another.
Definition: User.cpp:21
const Use & getOperandUse(unsigned i) const
Definition: User.h:241
Value * getOperand(unsigned i) const
Definition: User.h:228
unsigned getNumOperands() const
Definition: User.h:250
bool isDroppable() const
A droppable user is a user for which uses can be dropped without affecting correctness and should be ...
Definition: User.cpp:115
ValueT lookup(const KeyT &Val) const
lookup - Return the entry for the specified key, or a default constructed value if no such entry exis...
Definition: ValueMap.h:164
LLVM Value Representation.
Definition: Value.h:74
Type * getType() const
All values are typed, get the type of this value.
Definition: Value.h:255
static constexpr uint64_t MaximumAlignment
Definition: Value.h:811
const Value * stripAndAccumulateConstantOffsets(const DataLayout &DL, APInt &Offset, bool AllowNonInbounds, bool AllowInvariantGroup=false, function_ref< bool(Value &Value, APInt &Offset)> ExternalAnalysis=nullptr) const
Accumulate the constant offset this value has compared to a base pointer.
void replaceAllUsesWith(Value *V)
Change all uses of this to point to a new Value.
Definition: Value.cpp:534
iterator_range< user_iterator > users()
Definition: Value.h:421
LLVMContext & getContext() const
All values hold a context through their type.
Definition: Value.cpp:1075
iterator_range< use_iterator > uses()
Definition: Value.h:376
StringRef getName() const
Return a constant reference to the value's name.
Definition: Value.cpp:309
std::pair< iterator, bool > insert(const ValueT &V)
Definition: DenseSet.h:213
constexpr ScalarTy getFixedValue() const
Definition: TypeSize.h:202
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition: TypeSize.h:171
An efficient, type-erasing, non-owning reference to a callable.
const ParentTy * getParent() const
Definition: ilist_node.h:32
self_iterator getIterator()
Definition: ilist_node.h:132
NodeTy * getNextNode()
Get the next node, or nullptr for the list tail.
Definition: ilist_node.h:353
This class implements an extremely fast bulk output stream that can only output to a stream.
Definition: raw_ostream.h:52
A raw_ostream that writes to an std::string.
Definition: raw_ostream.h:661
Enumerate the SCCs of a directed graph in reverse topological order of the SCC DAG.
Definition: SCCIterator.h:49
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
bool isAssumedReadNone(Attributor &A, const IRPosition &IRP, const AbstractAttribute &QueryingAA, bool &IsKnown)
Return true if IRP is readnone.
Definition: Attributor.cpp:653
bool isAssumedReadOnly(Attributor &A, const IRPosition &IRP, const AbstractAttribute &QueryingAA, bool &IsKnown)
Return true if IRP is readonly.
Definition: Attributor.cpp:648
raw_ostream & operator<<(raw_ostream &OS, const RangeTy &R)
Definition: Attributor.h:319
std::optional< Value * > combineOptionalValuesInAAValueLatice(const std::optional< Value * > &A, const std::optional< Value * > &B, Type *Ty)
Return the combination of A and B such that the result is a possible value of both.
Definition: Attributor.cpp:339
bool isValidAtPosition(const ValueAndContext &VAC, InformationCache &InfoCache)
Return true if the value of VAC is a valid at the position of VAC, that is a constant,...
Definition: Attributor.cpp:290
bool isAssumedThreadLocalObject(Attributor &A, Value &Obj, const AbstractAttribute &QueryingAA)
Return true if Obj is assumed to be a thread local object.
Definition: Attributor.cpp:835
bool isDynamicallyUnique(Attributor &A, const AbstractAttribute &QueryingAA, const Value &V, bool ForAnalysisOnly=true)
Return true if V is dynamically unique, that is, there are no two "instances" of V at runtime with di...
Definition: Attributor.cpp:231
bool getPotentialCopiesOfStoredValue(Attributor &A, StoreInst &SI, SmallSetVector< Value *, 4 > &PotentialCopies, const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation, bool OnlyExact=false)
Collect all potential values of the one stored by SI into PotentialCopies.
Definition: Attributor.cpp:599
bool isGPU(const Module &M)
Return true iff M target a GPU (and we can use GPU AS reasoning).
Definition: Attributor.cpp:200
ValueScope
Flags to distinguish intra-procedural queries from potentially inter-procedural queries.
Definition: Attributor.h:180
@ Intraprocedural
Definition: Attributor.h:181
@ Interprocedural
Definition: Attributor.h:182
bool isValidInScope(const Value &V, const Function *Scope)
Return true if V is a valid value in Scope, that is a constant or an instruction/argument of Scope.
Definition: Attributor.cpp:280
bool isPotentiallyReachable(Attributor &A, const Instruction &FromI, const Instruction &ToI, const AbstractAttribute &QueryingAA, const AA::InstExclusionSetTy *ExclusionSet=nullptr, std::function< bool(const Function &F)> GoBackwardsCB=nullptr)
Return true if ToI is potentially reachable from FromI without running into any instruction in Exclus...
Definition: Attributor.cpp:816
bool isNoSyncInst(Attributor &A, const Instruction &I, const AbstractAttribute &QueryingAA)
Return true if I is a nosync instruction.
Definition: Attributor.cpp:205
bool getPotentiallyLoadedValues(Attributor &A, LoadInst &LI, SmallSetVector< Value *, 4 > &PotentialValues, SmallSetVector< Instruction *, 4 > &PotentialValueOrigins, const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation, bool OnlyExact=false)
Collect all potential values LI could read into PotentialValues.
Definition: Attributor.cpp:589
Value * getWithType(Value &V, Type &Ty)
Try to convert V to type Ty without introducing new instructions.
Definition: Attributor.cpp:316
constexpr char Attrs[]
Key for Kernel::Metadata::mAttrs.
@ C
The default llvm calling convention, compatible with C.
Definition: CallingConv.h:34
@ Unsupported
This operation is completely unsupported on the target.
@ Undef
Value of the register doesn't matter.
Offsets
Offsets in bytes from the start of the input buffer.
Definition: SIInstrInfo.h:1609
@ SingleThread
Synchronized with respect to signal handlers executing in the same thread.
Definition: LLVMContext.h:54
@ CE
Windows NT (Windows on ARM)
initializer< Ty > init(const Ty &Val)
Definition: CommandLine.h:443
LocationClass< Ty > location(Ty &L)
Definition: CommandLine.h:463
unsigned combineHashValue(unsigned a, unsigned b)
Simplistic combination of 32-bit hash values into 32-bit hash values.
Definition: DenseMapInfo.h:39
ElementType
The element type of an SRV or UAV resource.
Definition: DXILABI.h:58
constexpr double e
Definition: MathExtras.h:47
const_iterator begin(StringRef path LLVM_LIFETIME_BOUND, Style style=Style::native)
Get begin iterator over path.
Definition: Path.cpp:226
const_iterator end(StringRef path LLVM_LIFETIME_BOUND)
Get end iterator over path.
Definition: Path.cpp:235
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Offset
Definition: DWP.cpp:480
@ Length
Definition: DWP.cpp:480
bool operator<(int64_t V1, const APSInt &V2)
Definition: APSInt.h:361
LLVM_ATTRIBUTE_ALWAYS_INLINE DynamicAPInt gcd(const DynamicAPInt &A, const DynamicAPInt &B)
Definition: DynamicAPInt.h:390
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1739
bool isLegalToPromote(const CallBase &CB, Function *Callee, const char **FailureReason=nullptr)
Return true if the given indirect call site can be made to call Callee.
Constant * getInitialValueOfAllocation(const Value *V, const TargetLibraryInfo *TLI, Type *Ty)
If this is a call to an allocation function that initializes memory to a fixed value,...
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
Definition: STLExtras.h:1697
auto pred_end(const MachineBasicBlock *BB)
unsigned getPointerAddressSpace(const Type *T)
Definition: SPIRVUtils.h:261
auto successors(const MachineBasicBlock *BB)
bool isRemovableAlloc(const CallBase *V, const TargetLibraryInfo *TLI)
Return true if this is a call to an allocation function that does not have side effects that we are r...
APFloat abs(APFloat X)
Returns the absolute value of the argument.
Definition: APFloat.h:1534
raw_fd_ostream & outs()
This returns a reference to a raw_fd_ostream for standard output.
UseCaptureKind DetermineUseCaptureKind(const Use &U, llvm::function_ref< bool(Value *, const DataLayout &)> IsDereferenceableOrNull)
Determine what kind of capture behaviour U may exhibit.
Value * getAllocAlignment(const CallBase *V, const TargetLibraryInfo *TLI)
Gets the alignment argument for an aligned_alloc-like function, using either built-in knowledge based...
Value * simplifyInstructionWithOperands(Instruction *I, ArrayRef< Value * > NewOps, const SimplifyQuery &Q)
Like simplifyInstruction but the operands of I are replaced with NewOps.
Value * GetPointerBaseWithConstantOffset(Value *Ptr, int64_t &Offset, const DataLayout &DL, bool AllowNonInbounds=true)
Analyze the specified pointer to see if it can be expressed as a base pointer plus a constant offset.
const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=6)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
scc_iterator< T > scc_begin(const T &G)
Construct the begin iterator for a deduced graph type T.
Definition: SCCIterator.h:233
PotentialValuesState< std::pair< AA::ValueAndContext, AA::ValueScope > > PotentialLLVMValuesState
Definition: Attributor.h:5231
@ NONE
Definition: Attributor.h:6484
bool isNoAliasCall(const Value *V)
Return true if this pointer is returned by a noalias function.
raw_ostream & WriteGraph(raw_ostream &O, const GraphType &G, bool ShortNames=false, const Twine &Title="")
Definition: GraphWriter.h:359
bool operator==(const AddressRangeValuePair &LHS, const AddressRangeValuePair &RHS)
ConstantRange getConstantRangeFromMetadata(const MDNode &RangeMD)
Parse out a conservative ConstantRange from !range metadata.
auto map_range(ContainerTy &&C, FuncTy F)
Definition: STLExtras.h:377
const Value * getPointerOperand(const Value *V)
A helper function that returns the pointer operand of a load, store or GEP instruction.
Value * simplifyInstruction(Instruction *I, const SimplifyQuery &Q)
See if we can compute a simplified version of this instruction.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1746
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition: MathExtras.h:340
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition: MathExtras.h:291
void sort(IteratorTy Start, IteratorTy End)
Definition: STLExtras.h:1664
MemoryEffectsBase< IRMemLocation > MemoryEffects
Summary of how a function affects memory in the program.
Definition: ModRef.h:268
@ None
Definition: CodeGenData.h:106
bool NullPointerIsDefined(const Function *F, unsigned AS=0)
Check whether null pointer dereferencing is considered undefined behavior for a given function or an ...
Definition: Function.cpp:1187
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
bool isPointerTy(const Type *T)
Definition: SPIRVUtils.h:255
bool wouldInstructionBeTriviallyDead(const Instruction *I, const TargetLibraryInfo *TLI=nullptr)
Return true if the result produced by the instruction would have no side effects if it was not used.
Definition: Local.cpp:425
bool isSafeToSpeculativelyExecute(const Instruction *I, const Instruction *CtxI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr, const TargetLibraryInfo *TLI=nullptr, bool UseVariableInfo=true)
Return true if the instruction does not have any effects besides calculating the result and does not ...
bool set_union(S1Ty &S1, const S2Ty &S2)
set_union(A, B) - Compute A := A u B, return whether A changed.
Definition: SetOperations.h:43
void RemapInstruction(Instruction *I, ValueToValueMapTy &VM, RemapFlags Flags=RF_None, ValueMapTypeRemapper *TypeMapper=nullptr, ValueMaterializer *Materializer=nullptr)
Convert the instruction operands from referencing the current values into those specified by VM.
Definition: ValueMapper.h:263
CallBase & promoteCall(CallBase &CB, Function *Callee, CastInst **RetBitCast=nullptr)
Promote the given indirect call site to unconditionally call Callee.
bool hasAssumption(const Function &F, const KnownAssumptionString &AssumptionStr)
Return true if F has the assumption AssumptionStr attached.
Definition: Assumptions.cpp:70
RetainedKnowledge getKnowledgeFromUse(const Use *U, ArrayRef< Attribute::AttrKind > AttrKinds)
Return a valid Knowledge associated to the Use U if its Attribute kind is in AttrKinds.
bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ Other
Any other memory.
ChangeStatus clampStateAndIndicateChange< DerefState >(DerefState &S, const DerefState &R)
PotentialValuesState< APInt > PotentialConstantIntValuesState
Definition: Attributor.h:5229
DWARFExpression::Operation Op
bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
Value * getFreedOperand(const CallBase *CB, const TargetLibraryInfo *TLI)
If this if a call to a free function, return the freed operand.
ChangeStatus clampStateAndIndicateChange(StateType &S, const StateType &R)
Helper function to clamp a state S of type StateType with the information in R and indicate/return if...
Definition: Attributor.h:3473
constexpr unsigned BitWidth
Definition: BitmaskEnum.h:217
auto pred_begin(const MachineBasicBlock *BB)
ChangeStatus
{
Definition: Attributor.h:489
std::optional< APInt > getAllocSize(const CallBase *CB, const TargetLibraryInfo *TLI, function_ref< const Value *(const Value *)> Mapper=[](const Value *V) { return V;})
Return the size of the requested allocation.
DenseSet< StringRef > getAssumptions(const Function &F)
Return the set of all assumptions for the function F.
Definition: Assumptions.cpp:86
Align assumeAligned(uint64_t Value)
Treats the value 0 as a 1, so Align is always at least 1.
Definition: Alignment.h:111
Instruction * SplitBlockAndInsertIfThen(Value *Cond, BasicBlock::iterator SplitBefore, bool Unreachable, MDNode *BranchWeights=nullptr, DomTreeUpdater *DTU=nullptr, LoopInfo *LI=nullptr, BasicBlock *ThenBlock=nullptr)
Split the containing block at the specified instruction - everything before SplitBefore stays in the ...
@ OPTIONAL
The target may be valid if the source is not.
@ NONE
Do not track a dependence between source and target.
@ REQUIRED
The target cannot be valid if the source is not.
Value * simplifyCmpInst(CmpPredicate Predicate, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a CmpInst, fold the result or return null.
bool mayContainIrreducibleControl(const Function &F, const LoopInfo *LI)
T bit_floor(T Value)
Returns the largest integral power of two no greater than Value if Value is nonzero.
Definition: bit.h:327
KnownFPClass computeKnownFPClass(const Value *V, const APInt &DemandedElts, FPClassTest InterestedClasses, unsigned Depth, const SimplifyQuery &SQ)
Determine which floating-point classes are valid for V, and return them in KnownFPClass bit sets.
bool isIdentifiedObject(const Value *V)
Return true if this pointer refers to a distinct and identifiable object.
constexpr StringRef AssumptionAttrKey
The key we use for assumption attributes.
Definition: Assumptions.h:28
Implement std::hash so that hash_code can be used in STL containers.
Definition: BitVector.h:858
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition: BitVector.h:860
A type to track pointer/struct usage and accesses for AAPointerInfo.
bool forallInterferingAccesses(AA::RangeTy Range, F CB) const
See AAPointerInfo::forallInterferingAccesses.
AAPointerInfo::const_bin_iterator end() const
ChangeStatus addAccess(Attributor &A, const AAPointerInfo::RangeList &Ranges, Instruction &I, std::optional< Value * > Content, AAPointerInfo::AccessKind Kind, Type *Ty, Instruction *RemoteI=nullptr)
Add a new Access to the state at offset Offset and with size Size.
DenseMap< const Instruction *, SmallVector< unsigned > > RemoteIMap
AAPointerInfo::const_bin_iterator begin() const
AAPointerInfo::OffsetInfo ReturnedOffsets
Flag to determine if the underlying pointer is reaching a return statement in the associated function...
State(State &&SIS)=default
const AAPointerInfo::Access & getAccess(unsigned Index) const
SmallVector< AAPointerInfo::Access > AccessList
bool isAtFixpoint() const override
See AbstractState::isAtFixpoint().
bool forallInterferingAccesses(Instruction &I, F CB, AA::RangeTy &Range) const
See AAPointerInfo::forallInterferingAccesses.
static State getWorstState(const State &SIS)
Return the worst possible representable state.
AAPointerInfo::OffsetBinsTy OffsetBins
ChangeStatus indicateOptimisticFixpoint() override
See AbstractState::indicateOptimisticFixpoint().
State & operator=(const State &R)
ChangeStatus indicatePessimisticFixpoint() override
See AbstractState::indicatePessimisticFixpoint().
static State getBestState(const State &SIS)
Return the best possible representable state.
bool isValidState() const override
See AbstractState::isValidState().
----------------—AAIntraFnReachability Attribute-----------------------—
ReachabilityQueryInfo(const ReachabilityQueryInfo &RQI)
unsigned Hash
Precomputed hash for this RQI.
ReachabilityQueryInfo(const Instruction *From, const ToTy *To)
ReachabilityQueryInfo(Attributor &A, const Instruction &From, const ToTy &To, const AA::InstExclusionSetTy *ES, bool MakeUnique)
Constructor replacement to ensure unique and stable sets are used for the cache.
An abstract interface for address space information.
Definition: Attributor.h:6294
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6330
An abstract interface for all align attributes.
Definition: Attributor.h:4281
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4312
Align getKnownAlign() const
Return known alignment.
Definition: Attributor.h:4295
static const char ID
Definition: Attributor.h:6369
An abstract attribute for getting assumption information.
Definition: Attributor.h:6220
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6248
An abstract state for querying live call edges.
Definition: Attributor.h:5495
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5538
An abstract Attribute for specializing "dynamic" components of "denormal-fp-math" and "denormal-fp-ma...
Definition: Attributor.h:6455
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6477
An abstract interface for all dereferenceable attribute.
Definition: Attributor.h:4227
uint32_t getKnownDereferenceableBytes() const
Return known dereferenceable bytes.
Definition: Attributor.h:4251
uint32_t getAssumedDereferenceableBytes() const
Return assumed dereferenceable bytes.
Definition: Attributor.h:4246
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4272
An abstract interface for llvm::GlobalValue information interference.
Definition: Attributor.h:6374
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6408
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4574
An abstract interface for indirect call information interference.
Definition: Attributor.h:6413
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6449
An abstract interface to track if a value leaves it's defining function instance.
Definition: Attributor.h:4319
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4352
An abstract Attribute for computing reachability between functions.
Definition: Attributor.h:5691
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5726
bool canReach(Attributor &A, const Function &Fn) const
If the function represented by this possition can reach Fn.
Definition: Attributor.h:5697
virtual bool instructionCanReach(Attributor &A, const Instruction &Inst, const Function &Fn, const AA::InstExclusionSetTy *ExclusionSet=nullptr) const =0
Can Inst reach Fn.
An abstract interface to determine reachability of point A to B.
Definition: Attributor.h:3825
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3853
virtual bool isAssumedReachable(Attributor &A, const Instruction &From, const Instruction &To, const AA::InstExclusionSetTy *ExclusionSet=nullptr) const =0
Returns true if 'From' instruction is assumed to reach, 'To' instruction.
An abstract interface for liveness abstract attribute.
Definition: Attributor.h:3985
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4075
An abstract interface for memory access kind related attributes (readnone/readonly/writeonly).
Definition: Attributor.h:4639
bool isAssumedReadOnly() const
Return true if we assume that the underlying value is not accessed (=written) in its respective scope...
Definition: Attributor.h:4677
bool isKnownReadNone() const
Return true if we know that the underlying value is not read or accessed in its respective scope.
Definition: Attributor.h:4665
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4704
bool isAssumedReadNone() const
Return true if we assume that the underlying value is not read or accessed in its respective scope.
Definition: Attributor.h:4669
An abstract interface for all memory location attributes (readnone/argmemonly/inaccessiblememonly/ina...
Definition: Attributor.h:4713
static std::string getMemoryLocationsAsStr(MemoryLocationsKind MLK)
Return the locations encoded by MLK as a readable string.
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4889
StateType::base_t MemoryLocationsKind
Definition: Attributor.h:4714
An abstract interface for all nonnull attributes.
Definition: Attributor.h:3601
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3636
An abstract interface for all noalias attributes.
Definition: Attributor.h:3860
static bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3899
An abstract interface for all nocapture attributes.
Definition: Attributor.h:4360
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4433
@ NO_CAPTURE_MAYBE_RETURNED
If we do not capture the value in memory or through integers we can only communicate it back as a der...
Definition: Attributor.h:4390
bool isAssumedNoCaptureMaybeReturned() const
Return true if we assume that the underlying value is not captured in its respective scope but we all...
Definition: Attributor.h:4414
static bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static void determineFunctionCaptureCapabilities(const IRPosition &IRP, const Function &F, BitIntegerState &State)
Update State according to the capture capabilities of F for position IRP.
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5449
An AbstractAttribute for nofree.
Definition: Attributor.h:3906
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3949
An abstract attribute for norecurse.
Definition: Attributor.h:3694
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3718
An AbstractAttribute for noreturn.
Definition: Attributor.h:3956
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3980
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3594
static bool isAlignedBarrier(const CallBase &CB, bool ExecutedAligned)
Helper function to determine if CB is an aligned (GPU) barrier.
static bool isNonRelaxedAtomic(const Instruction *I)
Helper function used to determine whether an instruction is non-relaxed atomic.
static bool isNoSyncIntrinsic(const Instruction *I)
Helper function specific for intrinsics which are potentially volatile.
An abstract interface for all noundef attributes.
Definition: Attributor.h:5363
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5398
static bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See IRAttribute::isImpliedByIR.
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3511
An abstract Attribute for determining the necessity of the convergent attribute.
Definition: Attributor.h:5731
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5759
bool isAssumedNotConvergent() const
Return true if "non-convergent" is assumed.
Definition: Attributor.h:5741
An abstract interface for all nonnull attributes.
Definition: Attributor.h:3643
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3687
static bool isImpliedByIR(Attributor &A, const IRPosition &IRP, Attribute::AttrKind ImpliedAttributeKind, bool IgnoreSubsumingPositions=false)
See AbstractAttribute::isImpliedByIR(...).
An access description.
Definition: Attributor.h:5996
A helper containing a list of offsets computed for a Use.
Definition: Attributor.h:5800
A container for a list of ranges.
Definition: Attributor.h:5845
static void set_difference(const RangeList &L, const RangeList &R, RangeList &D)
Copy ranges from L that are not in R, into D.
Definition: Attributor.h:5881
An abstract interface for struct information.
Definition: Attributor.h:5763
virtual bool reachesReturn() const =0
virtual const_bin_iterator begin() const =0
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6212
virtual int64_t numOffsetBins() const =0
An abstract interface for potential values analysis.
Definition: Attributor.h:5253
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5312
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:5349
static Value * getSingleValue(Attributor &A, const AbstractAttribute &AA, const IRPosition &IRP, SmallVectorImpl< AA::ValueAndContext > &Values)
Extract the single value in Values if any.
An abstract interface for privatizability.
Definition: Attributor.h:4588
virtual std::optional< Type * > getPrivatizableType() const =0
Return the type we can choose for a private copy of the underlying value.
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4630
An abstract attribute for undefined behavior.
Definition: Attributor.h:3787
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3820
An abstract attribute for getting all assumption underlying objects.
Definition: Attributor.h:6252
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:6282
An abstract interface for range value analysis.
Definition: Attributor.h:4894
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4957
An abstract interface for value simplify abstract attribute.
Definition: Attributor.h:4512
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:4534
An abstract attribute for willreturn.
Definition: Attributor.h:3725
static const char ID
Unique ID (due to the unique address)
Definition: Attributor.h:3782
Helper to represent an access offset and size, with logic to deal with uncertainty and check for over...
Definition: Attributor.h:237
static constexpr int64_t Unknown
Definition: Attributor.h:316
static RangeTy getUnknown()
Definition: Attributor.h:243
Base struct for all "concrete attribute" deductions.
Definition: Attributor.h:3293
void print(raw_ostream &OS) const
Helper functions, for debug purposes only.
Definition: Attributor.h:3377
virtual StateType & getState()=0
Return the internal abstract state for inspection.
An interface to query the internal state of an abstract attribute.
Definition: Attributor.h:2613
virtual ChangeStatus indicatePessimisticFixpoint()=0
Indicate that the abstract state should converge to the pessimistic state.
virtual bool isAtFixpoint() const =0
Return if this abstract state is fixed, thus does not need to be updated if information changes as it...
virtual bool isValidState() const =0
Return if this abstract state is in a valid state.
virtual ChangeStatus indicateOptimisticFixpoint()=0
Indicate that the abstract state should converge to the optimistic state.
Helper for AA::PointerInfo::Access DenseMap/Set usage ignoring everythign but the instruction.
static unsigned getHashValue(const Access &A)
static Access getTombstoneKey()
static bool isEqual(const Access &LHS, const Access &RHS)
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition: Alignment.h:39
uint64_t value() const
This is a hole in the type system and should not be abused.
Definition: Alignment.h:85
Helper struct used in the communication between an abstract attribute (AA) that wants to change the s...
Definition: Attributor.h:2221
std::function< void(const ArgumentReplacementInfo &, AbstractCallSite, SmallVectorImpl< Value * > &)> ACSRepairCBTy
Abstract call site (ACS) repair callback type.
Definition: Attributor.h:2244
const Argument & getReplacedArg() const
Definition: Attributor.h:2251
std::function< void(const ArgumentReplacementInfo &, Function &, Function::arg_iterator)> CalleeRepairCBTy
Callee repair callback type.
Definition: Attributor.h:2230
The fixpoint analysis framework that orchestrates the attribute deduction.
Definition: Attributor.h:1525
std::function< std::optional< Value * >(const IRPosition &, const AbstractAttribute *, bool &)> SimplifictionCallbackTy
Register CB as a simplification callback.
Definition: Attributor.h:2026
Specialization of the integer state for a bit-wise encoding.
Definition: Attributor.h:2754
BitIntegerState & removeAssumedBits(base_t BitsEncoding)
Remove the bits in BitsEncoding from the "assumed bits" if not known.
Definition: Attributor.h:2779
BitIntegerState & addKnownBits(base_t Bits)
Add the bits in BitsEncoding to the "known bits".
Definition: Attributor.h:2771
Simple wrapper for a single bit (boolean) state.
Definition: Attributor.h:2897
Represent subnormal handling kind for floating point instruction inputs and outputs.
static constexpr DenormalMode getDefault()
Return the assumed default mode for a function without denormal-fp-math.
static constexpr DenormalMode getInvalid()
static unsigned getHashValue(const Access &A)
static bool isEqual(const Access &LHS, const Access &RHS)
static bool isEqual(const AA::RangeTy &A, const AA::RangeTy B)
static unsigned getHashValue(const AA::RangeTy &Range)
static ReachabilityQueryInfo< ToTy > * getEmptyKey()
static ReachabilityQueryInfo< ToTy > * getTombstoneKey()
static bool isEqual(const ReachabilityQueryInfo< ToTy > *LHS, const ReachabilityQueryInfo< ToTy > *RHS)
static unsigned getHashValue(const ReachabilityQueryInfo< ToTy > *RQI)
An information struct used to provide DenseMap with the various necessary components for a given valu...
Definition: DenseMapInfo.h:52
State for dereferenceable attribute.
Definition: Attributor.h:4081
IncIntegerState DerefBytesState
State representing for dereferenceable bytes.
Definition: Attributor.h:4097
ChangeStatus manifest(Attributor &A) override
See AbstractAttribute::manifest(...).
Definition: Attributor.h:3230
Helper to describe and deal with positions in the LLVM-IR.
Definition: Attributor.h:586
Function * getAssociatedFunction() const
Return the associated function, if any.
Definition: Attributor.h:717
static const IRPosition callsite_returned(const CallBase &CB)
Create a position describing the returned value of CB.
Definition: Attributor.h:654
static const IRPosition returned(const Function &F, const CallBaseContext *CBContext=nullptr)
Create a position describing the returned value of F.
Definition: Attributor.h:636
Argument * getAssociatedArgument() const
Return the associated argument, if any.
Definition: Attributor.cpp:995
static const IRPosition value(const Value &V, const CallBaseContext *CBContext=nullptr)
Create a position describing the value of V.
Definition: Attributor.h:610
int getCalleeArgNo() const
Return the callee argument number of the associated value if it is an argument or call site argument,...
Definition: Attributor.h:804
static const IRPosition inst(const Instruction &I, const CallBaseContext *CBContext=nullptr)
Create a position describing the instruction I.
Definition: Attributor.h:622
static const IRPosition callsite_argument(const CallBase &CB, unsigned ArgNo)
Create a position describing the argument of CB at position ArgNo.
Definition: Attributor.h:659
@ IRP_ARGUMENT
An attribute for a function argument.
Definition: Attributor.h:600
@ IRP_RETURNED
An attribute for the function return value.
Definition: Attributor.h:596
@ IRP_CALL_SITE
An attribute for a call site (function scope).
Definition: Attributor.h:599
@ IRP_CALL_SITE_RETURNED
An attribute for a call site return value.
Definition: Attributor.h:597
@ IRP_FUNCTION
An attribute for a function (scope).
Definition: Attributor.h:598
@ IRP_CALL_SITE_ARGUMENT
An attribute for a call site argument.
Definition: Attributor.h:601
@ IRP_INVALID
An invalid position.
Definition: Attributor.h:593
Instruction * getCtxI() const
Return the context instruction, if any.
Definition: Attributor.h:770
static const IRPosition argument(const Argument &Arg, const CallBaseContext *CBContext=nullptr)
Create a position describing the argument Arg.
Definition: Attributor.h:643
Type * getAssociatedType() const
Return the type this abstract attribute is associated with.
Definition: Attributor.h:793
static const IRPosition function(const Function &F, const CallBaseContext *CBContext=nullptr)
Create a position describing the function scope of F.
Definition: Attributor.h:629
const CallBaseContext * getCallBaseContext() const
Get the call base context from the position.
Definition: Attributor.h:932
Value & getAssociatedValue() const
Return the value this abstract attribute is associated with.
Definition: Attributor.h:784
Value & getAnchorValue() const
Return the value this abstract attribute is anchored with.
Definition: Attributor.h:703
int getCallSiteArgNo() const
Return the call site argument number of the associated value if it is an argument or call site argume...
Definition: Attributor.h:813
static const IRPosition function_scope(const IRPosition &IRP, const CallBaseContext *CBContext=nullptr)
Create a position with function scope matching the "context" of IRP.
Definition: Attributor.h:682
Kind getPositionKind() const
Return the associated position kind.
Definition: Attributor.h:882
bool isArgumentPosition() const
Return true if the position is an argument or call site argument.
Definition: Attributor.h:914
static const IRPosition callsite_function(const CallBase &CB)
Create a position describing the function scope of CB.
Definition: Attributor.h:649
Function * getAnchorScope() const
Return the Function surrounding the anchor value.
Definition: Attributor.h:758
Incoming for lane maks phi as machine instruction, incoming register Reg and incoming block Block are...
Data structure to hold cached (LLVM-IR) information.
Definition: Attributor.h:1203
bool isOnlyUsedByAssume(const Instruction &I) const
Definition: Attributor.h:1303
AP::Result * getAnalysisResultForFunction(const Function &F, bool CachedOnly=false)
Return the analysis result from a pass AP for function F.
Definition: Attributor.h:1313
State for an integer range.
Definition: Attributor.h:2939
ConstantRange getKnown() const
Return the known state encoding.
Definition: Attributor.h:2995
ConstantRange getAssumed() const
Return the assumed state encoding.
Definition: Attributor.h:2998
bool isValidState() const override
See AbstractState::isValidState() NOTE: For now we simply pretend that the worst possible state is in...
Definition: Attributor.h:2672
bool isAtFixpoint() const override
See AbstractState::isAtFixpoint()
Definition: Attributor.h:2675
ChangeStatus indicateOptimisticFixpoint() override
See AbstractState::indicateOptimisticFixpoint(...)
Definition: Attributor.h:2678
base_t getAssumed() const
Return the assumed state encoding.
Definition: Attributor.h:2693
static constexpr base_t getWorstState()
Return the worst possible representable state.
Definition: Attributor.h:2665
ChangeStatus indicatePessimisticFixpoint() override
See AbstractState::indicatePessimisticFixpoint(...)
Definition: Attributor.h:2684
Helper that allows to insert a new assumption string in the known assumption set by creating a (stati...
Definition: Assumptions.h:36
FPClassTest KnownFPClasses
Floating-point classes the value could be one of.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Definition: Alignment.h:117
A "must be executed context" for a given program point PP is the set of instructions,...
Definition: MustExecute.h:385
iterator & end()
Return an universal end iterator.
Definition: MustExecute.h:433
bool findInContextOf(const Instruction *I, const Instruction *PP)
Helper to look for I in the context of PP.
Definition: MustExecute.h:469
iterator & begin(const Instruction *PP)
Return an iterator to explore the context around PP.
Definition: MustExecute.h:419
bool checkForAllContext(const Instruction *PP, function_ref< bool(const Instruction *)> Pred)
}
Definition: MustExecute.h:455
Various options to control the behavior of getObjectSize.
A class for a set state.
Definition: Attributor.h:4968
static unsigned MaxPotentialValues
Maximum number of potential values to be tracked.
Definition: Attributor.h:5021
void unionAssumed(const MemberTy &C)
Union assumed set with the passed value.
Definition: Attributor.h:5038
const SetTy & getAssumedSet() const
Return this set.
Definition: Attributor.h:4998
Represent one information held inside an operand bundle of an llvm.assume.
A MapVector that performs no allocations if smaller than a certain size.
Definition: MapVector.h:254
Helper to tie a abstract state implementation to an abstract attribute.
Definition: Attributor.h:3182
StateType & getState() override
See AbstractAttribute::getState(...).
Definition: Attributor.h:3190
bool unionAssumed(std::optional< Value * > Other)
Merge Other into the currently assumed simplified value.