18#ifndef LLVM_TRANSFORMS_INSTCOMBINE_INSTCOMBINER_H
19#define LLVM_TRANSFORMS_INSTCOMBINE_INSTCOMBINER_H
31#define DEBUG_TYPE "instcombine"
38class OptimizationRemarkEmitter;
39class ProfileSummaryInfo;
40class TargetLibraryInfo;
41class TargetTransformInfo;
87 bool MadeIRChange =
false;
102 :
TTI(
TTI), Builder(Builder), Worklist(Worklist),
103 MinimizeSize(MinimizeSize), AA(AA), AC(AC), TLI(TLI), DT(DT),
DL(
DL),
104 SQ(
DL, &TLI, &DT, &AC, nullptr,
true,
106 ORE(ORE), BFI(BFI), BPI(BPI), PSI(PSI), LI(LI) {}
114 if (
auto *BitCast = dyn_cast<BitCastInst>(V))
115 if (!OneUseOnly || BitCast->hasOneUse())
116 return BitCast->getOperand(0);
140 if (isa<Instruction>(V)) {
141 if (isa<CastInst>(V) ||
match(V, m_Neg(PatternMatch::m_Value())) ||
142 match(V, m_Not(PatternMatch::m_Value())) ||
147 if (isa<Argument>(V))
149 return isa<Constant>(V) ? (isa<UndefValue>(V) ? 0 : 1) : 2;
159 case CmpInst::ICMP_NE:
160 case CmpInst::ICMP_ULE:
161 case CmpInst::ICMP_SLE:
162 case CmpInst::ICMP_UGE:
163 case CmpInst::ICMP_SGE:
165 case CmpInst::FCMP_ONE:
166 case CmpInst::FCMP_OLE:
167 case CmpInst::FCMP_OGE:
176 return ConstantExpr::getAdd(
C, ConstantInt::get(
C->getType(), 1));
181 return ConstantExpr::getSub(
C, ConstantInt::get(
C->getType(), 1));
184 std::optional<std::pair<
196 return match(&
SI, PatternMatch::m_LogicalAnd(PatternMatch::m_Value(),
197 PatternMatch::m_Value())) ||
198 match(&
SI, PatternMatch::m_LogicalOr(PatternMatch::m_Value(),
199 PatternMatch::m_Value()));
209 Value *getFreelyInvertedImpl(
Value *V,
bool WillInvertAllUses,
216 return getFreelyInvertedImpl(V, WillInvertAllUses, Builder, DoesConsume,
223 return getFreelyInverted(V, WillInvertAllUses, Builder, Unused);
234 return getFreelyInverted(V, WillInvertAllUses,
nullptr,
235 DoesConsume) !=
nullptr;
240 return isFreeToInvert(V, WillInvertAllUses, Unused);
250 for (
Use &U : V->uses()) {
251 if (U.getUser() == IgnoredUser)
254 auto *
I = cast<Instruction>(U.getUser());
255 switch (
I->getOpcode()) {
256 case Instruction::Select:
257 if (U.getOperandNo() != 0)
259 if (shouldAvoidAbsorbingNotIntoSelect(*cast<SelectInst>(
I)))
262 case Instruction::Br:
263 assert(U.getOperandNo() == 0 &&
"Must be branching on that value.");
265 case Instruction::Xor:
267 if (!
match(
I, m_Not(PatternMatch::m_Value())))
285 bool IsRHSConstant) {
286 auto *InVTy = cast<FixedVectorType>(In->getType());
288 Type *EltTy = InVTy->getElementType();
289 auto *SafeC = ConstantExpr::getBinOpIdentity(Opcode, EltTy, IsRHSConstant);
295 case Instruction::SRem:
296 case Instruction::URem:
297 SafeC = ConstantInt::get(EltTy, 1);
299 case Instruction::FRem:
300 SafeC = ConstantFP::get(EltTy, 1.0);
304 "Only rem opcodes have no identity constant for RHS");
308 case Instruction::Shl:
309 case Instruction::LShr:
310 case Instruction::AShr:
311 case Instruction::SDiv:
312 case Instruction::UDiv:
313 case Instruction::SRem:
314 case Instruction::URem:
315 case Instruction::Sub:
316 case Instruction::FSub:
317 case Instruction::FDiv:
318 case Instruction::FRem:
319 SafeC = Constant::getNullValue(EltTy);
326 assert(SafeC &&
"Must have safe constant for binop");
327 unsigned NumElts = InVTy->getNumElements();
329 for (
unsigned i = 0; i != NumElts; ++i) {
330 Constant *
C = In->getAggregateElement(i);
331 Out[i] = isa<UndefValue>(
C) ? SafeC :
C;
333 return ConstantVector::get(Out);
351 std::optional<Instruction *> targetInstCombineIntrinsic(
IntrinsicInst &
II);
352 std::optional<Value *>
355 bool &KnownBitsComputed);
356 std::optional<Value *> targetSimplifyDemandedVectorEltsIntrinsic(
367 assert(New && !New->getParent() &&
368 "New instruction already inserted into a basic block!");
369 New->insertBefore(Old);
376 New->setDebugLoc(Old->getDebugLoc());
377 return InsertNewInstBefore(New, Old);
389 if (
I.use_empty())
return nullptr;
396 V = PoisonValue::get(
I.getType());
399 <<
" with " << *V <<
'\n');
402 if (V->use_empty() && isa<Instruction>(V) && !V->hasName() &&
I.hasName())
405 I.replaceAllUsesWith(V);
411 Value *OldOp =
I.getOperand(OpNum);
412 I.setOperand(OpNum, V);
465 bool IsNSW =
false)
const {
511 return SimplifyDemandedBits(
I, OpNo, DemandedMask, Known,
518 bool AllowMultipleUsers =
false) = 0;
520 bool isValidAddrSpaceCast(
unsigned FromAS,
unsigned ToAS)
const;
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
IRBuilder< TargetFolder > BuilderTy
static GCRegistry::Add< ShadowStackGC > C("shadow-stack", "Very portable GC for uncooperative code generators")
#define LLVM_LIBRARY_VISIBILITY
uint64_t IntrinsicInst * II
StandardInstrumentations SI(Mod->getContext(), Debug, VerifyEach)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
Class for arbitrary precision integers.
A cache of @llvm.assume calls within a function.
InstListType::iterator iterator
Instruction iterators...
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
Analysis providing branch probability information.
This class is the base class for the comparison instructions.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
This is an important base class in LLVM.
This class represents an Operation in the Expression.
A parsed version of the target data layout string in and methods for querying it.
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
The core instruction combiner logic.
OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
const DataLayout & getDataLayout() const
static bool isCanonicalPredicate(CmpInst::Predicate Pred)
Predicate canonicalization reduces the number of patterns that need to be matched by other transforms...
bool isFreeToInvert(Value *V, bool WillInvertAllUses)
virtual Instruction * eraseInstFromFunction(Instruction &I)=0
Combiner aware instruction erasure.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
DominatorTree & getDominatorTree() const
OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI, bool IsNSW=false) const
virtual ~InstCombiner()=default
LoopInfo * getLoopInfo() const
InstCombiner(InstructionWorklist &Worklist, BuilderTy &Builder, bool MinimizeSize, AAResults *AA, AssumptionCache &AC, TargetLibraryInfo &TLI, TargetTransformInfo &TTI, DominatorTree &DT, OptimizationRemarkEmitter &ORE, BlockFrequencyInfo *BFI, BranchProbabilityInfo *BPI, ProfileSummaryInfo *PSI, const DataLayout &DL, LoopInfo *LI)
static unsigned getComplexity(Value *V)
Assign a complexity or rank value to LLVM Values.
SmallDenseMap< BasicBlock *, SmallVector< BasicBlock * >, 8 > PredOrder
Order of predecessors to canonicalize phi nodes towards.
TargetLibraryInfo & getTargetLibraryInfo() const
BlockFrequencyInfo * getBlockFrequencyInfo() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, unsigned Depth=0, const Instruction *CxtI=nullptr)
Instruction * InsertNewInstBefore(Instruction *New, BasicBlock::iterator Old)
Inserts an instruction New before instruction Old.
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
static bool shouldAvoidAbsorbingNotIntoSelect(const SelectInst &SI)
OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
static Constant * SubOne(Constant *C)
Subtract one from a Constant.
KnownBits computeKnownBits(const Value *V, unsigned Depth, const Instruction *CxtI) const
void replaceUse(Use &U, Value *NewValue)
Replace use and add the previously used value to the worklist.
OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
Instruction * InsertNewInstWith(Instruction *New, BasicBlock::iterator Old)
Same as InsertNewInstBefore, but also sets the debug loc.
BranchProbabilityInfo * BPI
bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known)
unsigned ComputeNumSignBits(const Value *Op, unsigned Depth=0, const Instruction *CxtI=nullptr) const
virtual bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known, unsigned Depth, const SimplifyQuery &Q)=0
virtual Value * SimplifyDemandedVectorElts(Value *V, APInt DemandedElts, APInt &UndefElts, unsigned Depth=0, bool AllowMultipleUsers=false)=0
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ? InstCombine's freelyInvertA...
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder)
void addToWorklist(Instruction *I)
Instruction * replaceOperand(Instruction &I, unsigned OpNum, Value *V)
Replace operand of instruction and add old operand to the worklist.
static Constant * getSafeVectorConstantForBinop(BinaryOperator::BinaryOps Opcode, Constant *In, bool IsRHSConstant)
Some binary operators require special handling to avoid poison and undefined behavior.
OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const Instruction *CxtI) const
ProfileSummaryInfo * getProfileSummaryInfo() const
OptimizationRemarkEmitter & getOptimizationRemarkEmitter() const
SmallDenseSet< std::pair< BasicBlock *, BasicBlock * >, 8 > DeadEdges
Edges that are known to never be taken.
void computeKnownBits(const Value *V, KnownBits &Known, unsigned Depth, const Instruction *CxtI) const
AssumptionCache & getAssumptionCache() const
bool MaskedValueIsZero(const Value *V, const APInt &Mask, unsigned Depth=0, const Instruction *CxtI=nullptr) const
OptimizationRemarkEmitter & ORE
OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const Instruction *CxtI) const
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
static Constant * AddOne(Constant *C)
Add one to a Constant.
unsigned ComputeMaxSignificantBits(const Value *Op, unsigned Depth=0, const Instruction *CxtI=nullptr) const
InstructionWorklist - This is the worklist management logic for InstCombine and other simplification ...
void pushUsersToWorkList(Instruction &I)
When an instruction is simplified, add all users of the instruction to the work lists because they mi...
void add(Instruction *I)
Add instruction to the worklist.
void push(Instruction *I)
Push the instruction onto the worklist stack.
void handleUseCountDecrement(Value *V)
Should be called after decrementing the use-count on V.
A wrapper class for inspecting calls to intrinsic functions.
Analysis providing profile information.
This class represents the LLVM 'select' instruction.
Implements a dense probed hash-table based set with some number of buckets stored inline.
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Provides information about what library functions are available for the current target.
The instances of the Type class are immutable: once they are created, they are never changed.
A Use represents the edge between a Value definition and its users.
LLVM Value Representation.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
bool match(Val *V, const Pattern &P)
FNeg_match< OpTy > m_FNeg(const OpTy &X)
Match 'fneg X' as 'fsub -0.0, X'.
This is an optimization pass for GlobalISel generic memory operations.
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const SimplifyQuery &DL, unsigned Depth=0)
Return true if 'V & Mask' is known to be zero.
bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Return true if the given value is known to have exactly one bit set when defined.
OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ, bool IsNSW=false)
OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const SimplifyQuery &SQ)
void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
unsigned ComputeNumSignBits(const Value *Op, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true)
Return the number of times the sign bit of the register is replicated into the other bits.
OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const SimplifyQuery &SQ)
unsigned ComputeMaxSignificantBits(const Value *Op, const DataLayout &DL, unsigned Depth=0, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr)
Get the upper bound on bit size for this Value Op as a signed integer.
SimplifyQuery getWithInstruction(const Instruction *I) const