LLVM 23.0.0git
InstCombineInternal.h
Go to the documentation of this file.
1//===- InstCombineInternal.h - InstCombine pass internals -------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9/// \file
10///
11/// This file provides internal interfaces used to implement the InstCombine.
12//
13//===----------------------------------------------------------------------===//
14
15#ifndef LLVM_LIB_TRANSFORMS_INSTCOMBINE_INSTCOMBINEINTERNAL_H
16#define LLVM_LIB_TRANSFORMS_INSTCOMBINE_INSTCOMBINEINTERNAL_H
17
19#include "llvm/ADT/Statistic.h"
23#include "llvm/IR/IRBuilder.h"
24#include "llvm/IR/InstVisitor.h"
27#include "llvm/IR/Value.h"
28#include "llvm/Support/Debug.h"
33#include <cassert>
34
35#define DEBUG_TYPE "instcombine"
37
38// As a default, let's assume that we want to be aggressive,
39// and attempt to traverse with no limits in attempt to sink negation.
40static constexpr unsigned NegatorDefaultMaxDepth = ~0U;
41
42// Let's guesstimate that most often we will end up visiting/producing
43// fairly small number of new instructions.
44static constexpr unsigned NegatorMaxNodesSSO = 16;
45
46namespace llvm {
47
48class AAResults;
49class APInt;
50class AssumptionCache;
51class BlockFrequencyInfo;
52class DataLayout;
53class DominatorTree;
54class GEPOperator;
55class GlobalVariable;
56class OptimizationRemarkEmitter;
57class ProfileSummaryInfo;
58class TargetLibraryInfo;
59class User;
60
61/// Enum to specify how shift operations should be evaluated in
62/// canEvaluateShifted.
63/// Lossy: Allows lossy transformations
64/// Signed: Requires lossless transformation, using ashr to restore for shl,
65/// or represents ashr handling for right shifts
66/// Unsigned: Requires lossless transformation, using lshr to restore for shl,
67/// or represents lshr handling for right shifts
69
71 : public InstCombiner,
72 public InstVisitor<InstCombinerImpl, Instruction *> {
73public:
83
84 ~InstCombinerImpl() override = default;
85
86 /// Perform early cleanup and prepare the InstCombine worklist.
88
89 /// Run the combiner over the entire worklist until it is empty.
90 ///
91 /// \returns true if the IR is changed.
92 bool run();
93
94 // Visitation implementation - Implement instruction combining for different
95 // instruction types. The semantics are as follows:
96 // Return Value:
97 // null - No change was made
98 // I - Change was made, I is still valid, I may be dead though
99 // otherwise - Change was made, replace I with returned instruction
100 //
105 Value *LHS, Value *RHS, Type *Ty, bool isNUW);
122 Value *simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted);
131 BinaryOperator *Sh0, const SimplifyQuery &SQ,
132 bool AnalyzeForSignBitExtraction = false);
136 BinaryOperator &OldAShr);
160 template <typename FPToIntTy> Instruction *foldItoFPtoI(FPToIntTy &FI);
167
174 Instruction *visitFree(CallInst &FI, Value *FreedOp);
195 bool freezeOtherUses(FreezeInst &FI);
198
199 /// Specify what to return for unhandled instructions.
201
202 /// True when DB dominates all uses of DI except UI.
203 /// UI must be in the same block as DI.
204 /// The routine checks that the DI parent and DB are different.
205 bool dominatesAllUses(const Instruction *DI, const Instruction *UI,
206 const BasicBlock *DB) const;
207
208 /// Try to replace select with select operand SIOpd in SI-ICmp sequence.
209 bool replacedSelectWithOperand(SelectInst *SI, const ICmpInst *Icmp,
210 const unsigned SIOpd);
211
212 LoadInst *combineLoadToNewType(LoadInst &LI, Type *NewTy,
213 const Twine &Suffix = "");
214
215 /// Check if fmul \p MulVal, +0.0 will yield +0.0 (or signed zero is
216 /// ignorable).
218 const Instruction *CtxI) const;
219
220 std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
222
223private:
224 bool annotateAnyAllocSite(CallBase &Call, const TargetLibraryInfo *TLI);
225 bool isDesirableIntType(unsigned BitWidth) const;
226 bool shouldChangeType(unsigned FromBitWidth, unsigned ToBitWidth) const;
227 bool shouldChangeType(Type *From, Type *To) const;
228 Value *dyn_castNegVal(Value *V) const;
229
230 /// Classify whether a cast is worth optimizing.
231 ///
232 /// This is a helper to decide whether the simplification of
233 /// logic(cast(A), cast(B)) to cast(logic(A, B)) should be performed.
234 ///
235 /// \param CI The cast we are interested in.
236 ///
237 /// \return true if this cast actually results in any code being generated and
238 /// if it cannot already be eliminated by some other transformation.
239 bool shouldOptimizeCast(CastInst *CI);
240
241 /// Try to optimize a sequence of instructions checking if an operation
242 /// on LHS and RHS overflows.
243 ///
244 /// If this overflow check is done via one of the overflow check intrinsics,
245 /// then CtxI has to be the call instruction calling that intrinsic. If this
246 /// overflow check is done by arithmetic followed by a compare, then CtxI has
247 /// to be the arithmetic instruction.
248 ///
249 /// If a simplification is possible, stores the simplified result of the
250 /// operation in OperationResult and result of the overflow check in
251 /// OverflowResult, and return true. If no simplification is possible,
252 /// returns false.
253 bool OptimizeOverflowCheck(Instruction::BinaryOps BinaryOp, bool IsSigned,
254 Value *LHS, Value *RHS,
255 Instruction &CtxI, Value *&OperationResult,
257
258 Instruction *visitCallBase(CallBase &Call);
259 Instruction *tryOptimizeCall(CallInst *CI);
260 bool transformConstExprCastCall(CallBase &Call);
261 Instruction *transformCallThroughTrampoline(CallBase &Call,
262 IntrinsicInst &Tramp);
263
264 /// Try to optimize a call to the result of a ptrauth intrinsic, potentially
265 /// into the ptrauth call bundle:
266 /// - call(ptrauth.resign(p)), ["ptrauth"()] -> call p, ["ptrauth"()]
267 /// - call(ptrauth.sign(p)), ["ptrauth"()] -> call p
268 /// as long as the key/discriminator are the same in sign and auth-bundle,
269 /// and we don't change the key in the bundle (to a potentially-invalid key.)
270 Instruction *foldPtrAuthIntrinsicCallee(CallBase &Call);
271
272 /// Try to optimize a call to a ptrauth constant, into its ptrauth bundle:
273 /// call(ptrauth(f)), ["ptrauth"()] -> call f
274 /// as long as the key/discriminator are the same in constant and bundle.
275 Instruction *foldPtrAuthConstantCallee(CallBase &Call);
276
277 // Return (a, b) if (LHS, RHS) is known to be (a, b) or (b, a).
278 // Otherwise, return std::nullopt
279 // Currently it matches:
280 // - LHS = (select c, a, b), RHS = (select c, b, a)
281 // - LHS = (phi [a, BB0], [b, BB1]), RHS = (phi [b, BB0], [a, BB1])
282 // - LHS = min(a, b), RHS = max(a, b)
283 std::optional<std::pair<Value *, Value *>> matchSymmetricPair(Value *LHS,
284 Value *RHS);
285
286 Value *simplifyMaskedLoad(IntrinsicInst &II);
287 Instruction *simplifyMaskedStore(IntrinsicInst &II);
288 Instruction *simplifyMaskedGather(IntrinsicInst &II);
289 Instruction *simplifyMaskedScatter(IntrinsicInst &II);
290
291 /// Transform (zext icmp) to bitwise / integer operations in order to
292 /// eliminate it.
293 ///
294 /// \param ICI The icmp of the (zext icmp) pair we are interested in.
295 /// \parem CI The zext of the (zext icmp) pair we are interested in.
296 ///
297 /// \return null if the transformation cannot be performed. If the
298 /// transformation can be performed the new instruction that replaces the
299 /// (zext icmp) pair will be returned.
300 Instruction *transformZExtICmp(ICmpInst *Cmp, ZExtInst &Zext);
301
302 Instruction *transformSExtICmp(ICmpInst *Cmp, SExtInst &Sext);
303
304 bool willNotOverflowSignedAdd(const WithCache<const Value *> &LHS,
305 const WithCache<const Value *> &RHS,
306 const Instruction &CxtI) const {
307 return computeOverflowForSignedAdd(LHS, RHS, &CxtI) ==
309 }
310
311 bool willNotOverflowUnsignedAdd(const WithCache<const Value *> &LHS,
312 const WithCache<const Value *> &RHS,
313 const Instruction &CxtI) const {
314 return computeOverflowForUnsignedAdd(LHS, RHS, &CxtI) ==
316 }
317
318 bool willNotOverflowAdd(const Value *LHS, const Value *RHS,
319 const Instruction &CxtI, bool IsSigned) const {
320 return IsSigned ? willNotOverflowSignedAdd(LHS, RHS, CxtI)
321 : willNotOverflowUnsignedAdd(LHS, RHS, CxtI);
322 }
323
324 bool willNotOverflowSignedSub(const Value *LHS, const Value *RHS,
325 const Instruction &CxtI) const {
326 return computeOverflowForSignedSub(LHS, RHS, &CxtI) ==
327 OverflowResult::NeverOverflows;
328 }
329
330 bool willNotOverflowUnsignedSub(const Value *LHS, const Value *RHS,
331 const Instruction &CxtI) const {
332 return computeOverflowForUnsignedSub(LHS, RHS, &CxtI) ==
333 OverflowResult::NeverOverflows;
334 }
335
336 bool willNotOverflowSub(const Value *LHS, const Value *RHS,
337 const Instruction &CxtI, bool IsSigned) const {
338 return IsSigned ? willNotOverflowSignedSub(LHS, RHS, CxtI)
339 : willNotOverflowUnsignedSub(LHS, RHS, CxtI);
340 }
341
342 bool willNotOverflowSignedMul(const Value *LHS, const Value *RHS,
343 const Instruction &CxtI) const {
344 return computeOverflowForSignedMul(LHS, RHS, &CxtI) ==
345 OverflowResult::NeverOverflows;
346 }
347
348 bool willNotOverflowUnsignedMul(const Value *LHS, const Value *RHS,
349 const Instruction &CxtI,
350 bool IsNSW = false) const {
351 return computeOverflowForUnsignedMul(LHS, RHS, &CxtI, IsNSW) ==
352 OverflowResult::NeverOverflows;
353 }
354
355 bool willNotOverflowMul(const Value *LHS, const Value *RHS,
356 const Instruction &CxtI, bool IsSigned) const {
357 return IsSigned ? willNotOverflowSignedMul(LHS, RHS, CxtI)
358 : willNotOverflowUnsignedMul(LHS, RHS, CxtI);
359 }
360
361 bool willNotOverflow(BinaryOperator::BinaryOps Opcode, const Value *LHS,
362 const Value *RHS, const Instruction &CxtI,
363 bool IsSigned) const {
364 switch (Opcode) {
365 case Instruction::Add: return willNotOverflowAdd(LHS, RHS, CxtI, IsSigned);
366 case Instruction::Sub: return willNotOverflowSub(LHS, RHS, CxtI, IsSigned);
367 case Instruction::Mul: return willNotOverflowMul(LHS, RHS, CxtI, IsSigned);
368 default: llvm_unreachable("Unexpected opcode for overflow query");
369 }
370 }
371
372 Value *EmitGEPOffset(GEPOperator *GEP, bool RewriteGEP = false);
373 /// Emit sum of multiple GEP offsets. The GEPs are processed in reverse
374 /// order.
375 Value *EmitGEPOffsets(ArrayRef<GEPOperator *> GEPs, GEPNoWrapFlags NW,
376 Type *IdxTy, bool RewriteGEPs);
377 Instruction *scalarizePHI(ExtractElementInst &EI, PHINode *PN);
378 Instruction *foldBitcastExtElt(ExtractElementInst &ExtElt);
379 Instruction *foldCastedBitwiseLogic(BinaryOperator &I);
380 Instruction *foldFBinOpOfIntCasts(BinaryOperator &I);
381 // Should only be called by `foldFBinOpOfIntCasts`.
382 Instruction *foldFBinOpOfIntCastsFromSign(
383 BinaryOperator &BO, bool OpsFromSigned, std::array<Value *, 2> IntOps,
384 Constant *Op1FpC, SmallVectorImpl<WithCache<const Value *>> &OpsKnown);
385 Instruction *foldBinopOfSextBoolToSelect(BinaryOperator &I);
386 Instruction *narrowBinOp(TruncInst &Trunc);
387 Instruction *narrowMaskedBinOp(BinaryOperator &And);
388 Instruction *narrowMathIfNoOverflow(BinaryOperator &I);
389 Instruction *narrowFunnelShift(TruncInst &Trunc);
390 Instruction *optimizeBitCastFromPhi(CastInst &CI, PHINode *PN);
391 Instruction *matchSAddSubSat(IntrinsicInst &MinMax1);
392 Instruction *foldNot(BinaryOperator &I);
393 Instruction *foldBinOpOfDisplacedShifts(BinaryOperator &I);
394
395 /// Determine if a pair of casts can be replaced by a single cast.
396 ///
397 /// \param CI1 The first of a pair of casts.
398 /// \param CI2 The second of a pair of casts.
399 ///
400 /// \return 0 if the cast pair cannot be eliminated, otherwise returns an
401 /// Instruction::CastOps value for a cast that can replace the pair, casting
402 /// CI1->getSrcTy() to CI2->getDstTy().
403 ///
404 /// \see CastInst::isEliminableCastPair
405 Instruction::CastOps isEliminableCastPair(const CastInst *CI1,
406 const CastInst *CI2);
407 Value *simplifyIntToPtrRoundTripCast(Value *Val);
408
409 Value *foldAndOrOfICmps(ICmpInst *LHS, ICmpInst *RHS, Instruction &I,
410 bool IsAnd, bool IsLogical = false);
411 Value *foldXorOfICmps(ICmpInst *LHS, ICmpInst *RHS, BinaryOperator &Xor);
412
413 Value *foldEqOfParts(Value *Cmp0, Value *Cmp1, bool IsAnd);
414
415 Value *foldAndOrOfICmpsUsingRanges(ICmpInst *ICmp1, ICmpInst *ICmp2,
416 bool IsAnd);
417
418 /// Optimize (fcmp)&(fcmp) or (fcmp)|(fcmp).
419 /// NOTE: Unlike most of instcombine, this returns a Value which should
420 /// already be inserted into the function.
421 Value *foldLogicOfFCmps(FCmpInst *LHS, FCmpInst *RHS, bool IsAnd,
422 bool IsLogicalSelect = false);
423
424 Instruction *foldLogicOfIsFPClass(BinaryOperator &Operator, Value *LHS,
425 Value *RHS);
426
427 Value *foldBooleanAndOr(Value *LHS, Value *RHS, Instruction &I, bool IsAnd,
428 bool IsLogical);
429
430 Value *reassociateBooleanAndOr(Value *LHS, Value *X, Value *Y, Instruction &I,
431 bool IsAnd, bool RHSIsLogical);
432
433 Value *foldDisjointOr(Value *LHS, Value *RHS);
434
435 Value *reassociateDisjointOr(Value *LHS, Value *RHS);
436
438 canonicalizeConditionalNegationViaMathToSelect(BinaryOperator &i);
439
440 Value *matchSelectFromAndOr(Value *A, Value *B, Value *C, Value *D,
441 bool InvertFalseVal = false);
442 Value *getSelectCondition(Value *A, Value *B, bool ABIsTheSame);
443
444 bool canEvaluateShifted(Value *V, unsigned NumBits, bool IsLeftShift,
445 ShiftSemantics Semantics, Instruction *CxtI);
446 Value *getShiftedValue(Value *V, unsigned NumBits, bool IsLeftShift,
447 ShiftSemantics Semantics);
448
449 Instruction *foldLShrOverflowBit(BinaryOperator &I);
450 Instruction *foldExtractOfOverflowIntrinsic(ExtractValueInst &EV);
451 Instruction *foldIntrinsicWithOverflowCommon(IntrinsicInst *II);
452 Instruction *foldIntrinsicIsFPClass(IntrinsicInst &II);
453 Instruction *foldFPSignBitOps(BinaryOperator &I);
454 Instruction *foldFDivConstantDivisor(BinaryOperator &I);
455
456 // Optimize one of these forms:
457 // and i1 Op, SI / select i1 Op, i1 SI, i1 false (if IsAnd = true)
458 // or i1 Op, SI / select i1 Op, i1 true, i1 SI (if IsAnd = false)
459 // into simplier select instruction using isImpliedCondition.
460 Instruction *foldAndOrOfSelectUsingImpliedCond(Value *Op, SelectInst &SI,
461 bool IsAnd);
462
463 Instruction *hoistFNegAboveFMulFDiv(Value *FNegOp, Instruction &FMFSource);
464
465 /// Simplify \p V given that it is known to be non-null.
466 /// Returns the simplified value if possible, otherwise returns nullptr.
467 /// If \p HasDereferenceable is true, the simplification will not perform
468 /// same object checks.
469 Value *simplifyNonNullOperand(Value *V, bool HasDereferenceable,
470 unsigned Depth = 0);
471
472 /// Create `select C, S1, S2`. Use only when the profile cannot be calculated
473 /// from existing profile metadata: if the Function has profiles, this will
474 /// set the profile of this select to "unknown".
475 SelectInst *
476 createSelectInstWithUnknownProfile(Value *C, Value *S1, Value *S2,
477 const Twine &NameStr = "",
478 InsertPosition InsertBefore = nullptr) {
479 auto *Sel = SelectInst::Create(C, S1, S2, NameStr, InsertBefore, nullptr);
481 return Sel;
482 }
483
484public:
485 /// Create and insert the idiom we use to indicate a block is unreachable
486 /// without having to rewrite the CFG from within InstCombine.
488 auto &Ctx = InsertAt->getContext();
489 auto *SI = new StoreInst(ConstantInt::getTrue(Ctx),
491 /*isVolatile*/ false, Align(1));
492 InsertNewInstWith(SI, InsertAt->getIterator());
493 }
494
495 /// Combiner aware instruction erasure.
496 ///
497 /// When dealing with an instruction that has side effects or produces a void
498 /// value, we can't rely on DCE to delete the instruction. Instead, visit
499 /// methods should return the value returned by this function.
501 LLVM_DEBUG(dbgs() << "IC: ERASE " << I << '\n');
502 assert(I.use_empty() && "Cannot erase instruction that is used!");
504
505 // Make sure that we reprocess all operands now that we reduced their
506 // use counts.
507 SmallVector<Value *> Ops(I.operands());
508 Worklist.remove(&I);
509 DC.removeValue(&I);
510 I.eraseFromParent();
511 for (Value *Op : Ops)
512 Worklist.handleUseCountDecrement(Op);
513 MadeIRChange = true;
514 return nullptr; // Don't do anything with FI
515 }
516
517 OverflowResult computeOverflow(
518 Instruction::BinaryOps BinaryOp, bool IsSigned,
519 Value *LHS, Value *RHS, Instruction *CxtI) const;
520
521 /// Performs a few simplifications for operators which are associative
522 /// or commutative.
523 bool SimplifyAssociativeOrCommutative(BinaryOperator &I);
524
525 /// Tries to simplify binary operations which some other binary
526 /// operation distributes over.
527 ///
528 /// It does this by either by factorizing out common terms (eg "(A*B)+(A*C)"
529 /// -> "A*(B+C)") or expanding out if this results in simplifications (eg: "A
530 /// & (B | C) -> (A&B) | (A&C)" if this is a win). Returns the simplified
531 /// value, or null if it didn't simplify.
532 Value *foldUsingDistributiveLaws(BinaryOperator &I);
533
534 /// Tries to simplify add operations using the definition of remainder.
535 ///
536 /// The definition of remainder is X % C = X - (X / C ) * C. The add
537 /// expression X % C0 + (( X / C0 ) % C1) * C0 can be simplified to
538 /// X % (C0 * C1)
539 Value *SimplifyAddWithRemainder(BinaryOperator &I);
540
541 // Binary Op helper for select operations where the expression can be
542 // efficiently reorganized.
543 Value *SimplifySelectsFeedingBinaryOp(BinaryOperator &I, Value *LHS,
544 Value *RHS);
545
546 // If `I` has operand `(ctpop (not x))`, fold `I` with `(sub nuw nsw
547 // BitWidth(x), (ctpop x))`.
548 Instruction *tryFoldInstWithCtpopWithNot(Instruction *I);
549
550 // (Binop1 (Binop2 (logic_shift X, C), C1), (logic_shift Y, C))
551 // -> (logic_shift (Binop1 (Binop2 X, inv_logic_shift(C1, C)), Y), C)
552 // (Binop1 (Binop2 (logic_shift X, Amt), Mask), (logic_shift Y, Amt))
553 // -> (BinOp (logic_shift (BinOp X, Y)), Mask)
554 Instruction *foldBinOpShiftWithShift(BinaryOperator &I);
555
556 /// Tries to simplify binops of select and cast of the select condition.
557 ///
558 /// (Binop (cast C), (select C, T, F))
559 /// -> (select C, C0, C1)
560 Instruction *foldBinOpOfSelectAndCastOfSelectCondition(BinaryOperator &I);
561 /// Fold both forms of the div_ceil idiom:
562 /// (add (udiv X, Y), (zext (icmp ne (urem X, Y), 0)))
563 /// -> (udiv (add nuw X, Y-1), Y)
564 /// (add (zext (udiv X, Y)), (zext (icmp ne (urem X, Y), 0)))
565 /// -> (zext (udiv (add nuw X, Y-1), Y))
566 Instruction *foldDivCeil(BinaryOperator &I);
567
568 /// This tries to simplify binary operations by factorizing out common terms
569 /// (e. g. "(A*B)+(A*C)" -> "A*(B+C)").
570 Value *tryFactorizationFolds(BinaryOperator &I);
571
572 /// Match a select chain which produces one of three values based on whether
573 /// the LHS is less than, equal to, or greater than RHS respectively.
574 /// Return true if we matched a three way compare idiom. The LHS, RHS, Less,
575 /// Equal and Greater values are saved in the matching process and returned to
576 /// the caller.
577 bool matchThreeWayIntCompare(SelectInst *SI, Value *&LHS, Value *&RHS,
578 ConstantInt *&Less, ConstantInt *&Equal,
579 ConstantInt *&Greater);
580
581 /// Attempts to replace I with a simpler value based on the demanded
582 /// bits.
583 Value *SimplifyDemandedUseBits(Instruction *I, const APInt &DemandedMask,
584 KnownBits &Known, const SimplifyQuery &Q,
585 unsigned Depth = 0);
587 bool SimplifyDemandedBits(Instruction *I, unsigned Op,
588 const APInt &DemandedMask, KnownBits &Known,
589 const SimplifyQuery &Q,
590 unsigned Depth = 0) override;
591
592 /// Helper routine of SimplifyDemandedUseBits. It computes KnownZero/KnownOne
593 /// bits. It also tries to handle simplifications that can be done based on
594 /// DemandedMask, but without modifying the Instruction.
595 Value *SimplifyMultipleUseDemandedBits(Instruction *I,
596 const APInt &DemandedMask,
597 KnownBits &Known,
598 const SimplifyQuery &Q,
599 unsigned Depth = 0);
600
601 /// Helper routine of SimplifyDemandedUseBits. It tries to simplify demanded
602 /// bit for "r1 = shr x, c1; r2 = shl r1, c2" instruction sequence.
603 Value *simplifyShrShlDemandedBits(
604 Instruction *Shr, const APInt &ShrOp1, Instruction *Shl,
605 const APInt &ShlOp1, const APInt &DemandedMask, KnownBits &Known);
606
607 /// Tries to simplify operands to an integer instruction based on its
608 /// demanded bits.
609 bool SimplifyDemandedInstructionBits(Instruction &Inst);
610 bool SimplifyDemandedInstructionBits(Instruction &Inst, KnownBits &Known);
611
612 Value *SimplifyDemandedVectorElts(Value *V, APInt DemandedElts,
613 APInt &PoisonElts, unsigned Depth = 0,
614 bool AllowMultipleUsers = false) override;
615
616 /// Attempts to replace V with a simpler value based on the demanded
617 /// floating-point classes
618 Value *SimplifyDemandedUseFPClass(Instruction *I, FPClassTest DemandedMask,
619 KnownFPClass &Known, const SimplifyQuery &Q,
620 unsigned Depth = 0);
621 Value *SimplifyMultipleUseDemandedFPClass(Instruction *I,
622 FPClassTest DemandedMask,
623 KnownFPClass &Known,
624 const SimplifyQuery &Q,
625 unsigned Depth);
626
627 bool SimplifyDemandedFPClass(Instruction *I, unsigned Op,
628 FPClassTest DemandedMask, KnownFPClass &Known,
629 const SimplifyQuery &Q, unsigned Depth = 0);
630
631 bool SimplifyDemandedInstructionFPClass(Instruction &Inst);
632
633 /// Common transforms for add / disjoint or
634 Instruction *foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW,
635 bool NUW);
636
637 /// Canonicalize the position of binops relative to shufflevector.
638 Instruction *foldVectorBinop(BinaryOperator &Inst);
642 VectorType *NewCTy);
643
644 /// Given a binary operator, cast instruction, or select which has a PHI node
645 /// as operand #0, see if we can fold the instruction into the PHI (which is
646 /// only possible if all operands to the PHI are constants).
648 bool AllowMultipleUses = false);
649
650 /// Try to fold binary operators whose operands are simple interleaved
651 /// recurrences to a single recurrence. This is a common pattern in reduction
652 /// operations.
653 /// Example:
654 /// %phi1 = phi [init1, %BB1], [%op1, %BB2]
655 /// %phi2 = phi [init2, %BB1], [%op2, %BB2]
656 /// %op1 = binop %phi1, constant1
657 /// %op2 = binop %phi2, constant2
658 /// %rdx = binop %op1, %op2
659 /// -->
660 /// %phi_combined = phi [init_combined, %BB1], [%op_combined, %BB2]
661 /// %rdx_combined = binop %phi_combined, constant_combined
663
664 /// For a binary operator with 2 phi operands, try to hoist the binary
665 /// operation before the phi. This can result in fewer instructions in
666 /// patterns where at least one set of phi operands simplifies.
667 /// Example:
668 /// BB3: binop (phi [X, BB1], [C1, BB2]), (phi [Y, BB1], [C2, BB2])
669 /// -->
670 /// BB1: BO = binop X, Y
671 /// BB3: phi [BO, BB1], [(binop C1, C2), BB2]
673
674 /// Given an instruction with a select as one operand and a constant as the
675 /// other operand, try to fold the binary operator into the select arguments.
676 /// This also works for Cast instructions, which obviously do not have a
677 /// second operand.
679 bool FoldWithMultiUse = false,
680 bool SimplifyBothArms = false);
681
683
684 /// This is a convenience wrapper function for the above two functions.
686
688
691
692 /// Try to rotate an operation below a PHI node, using PHI nodes for
693 /// its operands.
702
703 /// If the phi is within a phi web, which is formed by the def-use chain
704 /// of phis and all the phis in the web are only used in the other phis.
705 /// In this case, these phis are dead and we will remove all of them.
706 bool foldDeadPhiWeb(PHINode &PN);
707
708 /// If an integer typed PHI has only one use which is an IntToPtr operation,
709 /// replace the PHI with an existing pointer typed PHI if it exists. Otherwise
710 /// insert a new pointer typed PHI and replace the original one.
712
713 /// Helper function for FoldPHIArgXIntoPHI() to set debug location for the
714 /// folded operation.
716
719 Instruction &I);
721 const ICmpInst &I);
722 bool foldAllocaCmp(AllocaInst *Alloca);
725 CmpInst &ICI,
726 ConstantInt *AndCst = nullptr);
728 Constant *RHSC);
733
742 const APInt &C);
745 Value *Z, CmpPredicate Pred);
751
753
755 const APInt &C);
757 ConstantInt *C);
759 const APInt &C);
761 const SimplifyQuery &Q);
763 const APInt &C);
765 const APInt &C);
767 const APInt &C);
769 const APInt &C);
771 const APInt &C);
773 const APInt &C);
775 const APInt &C);
777 const APInt &C);
779 const APInt &C);
781 const APInt &C);
783 const APInt &C);
785 const APInt &C1);
787 const APInt &C1, const APInt &C2);
789 const APInt &C);
791 const APInt &C2);
793 const APInt &C2);
794
796 BinaryOperator *BO,
797 const APInt &C);
799 BinaryOperator *BO,
800 const APInt &C);
802 const APInt &C);
804 const APInt &C);
808 ICmpInst &CxtI);
809
810 // Helpers of visitSelectInst().
819 Value *A, Value *B, Instruction &Outer,
823 Value *FalseVal);
826 unsigned Depth = 0);
827
828 Value *insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi,
829 bool isSigned, bool Inside);
831
832 /// Given an initial instruction, check to see if it is the root of a
833 /// bswap/bitreverse idiom. If so, return the equivalent bswap/bitreverse
834 /// intrinsic.
836 bool MatchBitReversals);
837
840
842
843 bool tryToSinkInstruction(Instruction *I, BasicBlock *DestBlock);
845 Instruction *I, BasicBlock::iterator InsertPos, BasicBlock *SrcBlock,
847
849 void addDeadEdge(BasicBlock *From, BasicBlock *To,
855 void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser = nullptr);
856
857 /// Take the exact integer log2 of the value. If DoFold is true, create the
858 /// actual instructions, otherwise return a non-null dummy value. Return
859 /// nullptr on failure. Note, if DoFold is true the caller must ensure that
860 /// takeLog2 will succeed, otherwise it may create stray instructions.
861 Value *takeLog2(Value *Op, unsigned Depth, bool AssumeNonZero, bool DoFold);
862
863 Value *tryGetLog2(Value *Op, bool AssumeNonZero) {
864 if (takeLog2(Op, /*Depth=*/0, AssumeNonZero, /*DoFold=*/false))
865 return takeLog2(Op, /*Depth=*/0, AssumeNonZero, /*DoFold=*/true);
866 return nullptr;
867 }
868};
869
870class Negator final {
871 /// Top-to-bottom, def-to-use negated instruction tree we produced.
873
875 BuilderTy Builder;
876
877 const DominatorTree &DT;
878
879 const bool IsTrulyNegation;
880
881 SmallDenseMap<Value *, Value *> NegationsCache;
882
883 Negator(LLVMContext &C, const DataLayout &DL, const DominatorTree &DT,
884 bool IsTrulyNegation);
885
886#if LLVM_ENABLE_STATS
887 unsigned NumValuesVisitedInThisNegator = 0;
888 ~Negator();
889#endif
890
891 using Result = std::pair<ArrayRef<Instruction *> /*NewInstructions*/,
892 Value * /*NegatedRoot*/>;
893
894 std::array<Value *, 2> getSortedOperandsOfBinOp(Instruction *I);
895
896 [[nodiscard]] Value *visitImpl(Value *V, bool IsNSW, unsigned Depth);
897
898 [[nodiscard]] Value *negate(Value *V, bool IsNSW, unsigned Depth);
899
900 /// Recurse depth-first and attempt to sink the negation.
901 /// FIXME: use worklist?
902 [[nodiscard]] std::optional<Result> run(Value *Root, bool IsNSW);
903
904 Negator(const Negator &) = delete;
905 Negator(Negator &&) = delete;
906 Negator &operator=(const Negator &) = delete;
907 Negator &operator=(Negator &&) = delete;
908
909public:
910 /// Attempt to negate \p Root. Retuns nullptr if negation can't be performed,
911 /// otherwise returns negated value.
912 [[nodiscard]] static Value *Negate(bool LHSIsZero, bool IsNSW, Value *Root,
913 InstCombinerImpl &IC);
914};
915
917 /// Common base pointer.
918 Value *Ptr = nullptr;
919 /// LHS GEPs until common base.
921 /// RHS GEPs until common base.
923 /// LHS GEP NoWrapFlags until common base.
925 /// RHS GEP NoWrapFlags until common base.
927
928 static CommonPointerBase compute(Value *LHS, Value *RHS);
929
930 /// Whether expanding the GEP chains is expensive.
931 bool isExpensive() const;
932};
933
934} // end namespace llvm
935
936#undef DEBUG_TYPE
937
938#endif // LLVM_LIB_TRANSFORMS_INSTCOMBINE_INSTCOMBINEINTERNAL_H
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
constexpr LLT S1
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
#define X(NUM, ENUM, NAME)
Definition ELF.h:851
static GCRegistry::Add< ShadowStackGC > C("shadow-stack", "Very portable GC for uncooperative code generators")
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static bool foldICmpWithDominatingICmp(CmpInst *Cmp, const TargetLowering &TLI)
For pattern like:
#define LLVM_LIBRARY_VISIBILITY
Definition Compiler.h:137
static bool willNotOverflow(BinaryOpIntrinsic *BO, LazyValueInfo *LVI)
static bool isSigned(unsigned Opcode)
#define DEBUG_TYPE
Hexagon Common GEP
IRTranslator LLVM IR MI
static constexpr unsigned NegatorMaxNodesSSO
static constexpr unsigned NegatorDefaultMaxDepth
This file provides the interface for the instcombine pass implementation.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
uint64_t IntrinsicInst * II
StandardInstrumentations SI(Mod->getContext(), Debug, VerifyEach)
This file builds on the ADT/GraphTraits.h file to build a generic graph post order iterator.
This file contains the declarations for profiling metadata utility functions.
const SmallVectorImpl< MachineOperand > & Cond
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define LLVM_DEBUG(...)
Definition Debug.h:114
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static OverflowResult computeOverflowForSignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const AddOperator *Add, const SimplifyQuery &SQ)
Value * RHS
Value * LHS
static const uint32_t IV[8]
Definition blake3_impl.h:83
Class for arbitrary precision integers.
Definition APInt.h:78
This class represents a conversion between pointers from one address space to another.
an instruction to allocate memory on the stack
This class represents any memset intrinsic.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
A cache of @llvm.assume calls within a function.
an instruction that atomically reads a memory location, combines it with another value,...
LLVM Basic Block Representation.
Definition BasicBlock.h:62
InstListType::iterator iterator
Instruction iterators...
Definition BasicBlock.h:170
This class represents a no-op cast from one type to another.
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
Analysis providing branch probability information.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
CallBr instruction, tracking function calls that may not return control but instead transfer it to a ...
This class represents a function call, abstracting a target machine's calling convention.
This is the base class for all instructions that perform data casts.
Definition InstrTypes.h:448
This class is the base class for the comparison instructions.
Definition InstrTypes.h:664
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
Conditional Branch instruction.
This is the shared class of boolean and integer constants.
Definition Constants.h:87
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
This is an important base class in LLVM.
Definition Constant.h:43
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition Dominators.h:159
This instruction extracts a single (scalar) element from a VectorType value.
This instruction extracts a struct member or array element value from an aggregate value.
This instruction compares its operands according to the predicate given to the constructor.
This class represents a cast from floating point to signed integer.
This class represents a cast from floating point to unsigned integer.
This class represents a truncation of floating point types.
Convenience struct for specifying and reasoning about fast-math flags.
Definition FMF.h:23
An instruction for ordering other memory operations.
This class represents a freeze function that returns random concrete value if an operand is either a ...
Represents flags for the getelementptr instruction/expression.
static GEPNoWrapFlags all()
an instruction for type-safe pointer arithmetic to access elements of arrays and structs
This instruction compares its operands according to the predicate given to the constructor.
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
Definition IRBuilder.h:2847
This instruction inserts a single (scalar) element into a VectorType value.
This instruction inserts a struct field of array element value into an aggregate value.
Instruction * visitMul(BinaryOperator &I)
Instruction * foldICmpShrConstant(ICmpInst &Cmp, BinaryOperator *Shr, const APInt &C)
Fold icmp ({al}shr X, Y), C.
Instruction * foldICmpWithZextOrSext(ICmpInst &ICmp)
Instruction * foldICmpSelectConstant(ICmpInst &Cmp, SelectInst *Select, ConstantInt *C)
Instruction * foldICmpSRemConstant(ICmpInst &Cmp, BinaryOperator *UDiv, const APInt &C)
Instruction * foldSelectToCmp(SelectInst &SI)
Instruction * visitAdd(BinaryOperator &I)
Instruction * visitCondBrInst(CondBrInst &BI)
bool fmulByZeroIsZero(Value *MulVal, FastMathFlags FMF, const Instruction *CtxI) const
Check if fmul MulVal, +0.0 will yield +0.0 (or signed zero is ignorable).
Instruction * foldICmpBinOpWithConstant(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Fold an icmp with BinaryOp and constant operand: icmp Pred BO, C.
Instruction * foldICmpOrConstant(ICmpInst &Cmp, BinaryOperator *Or, const APInt &C)
Fold icmp (or X, Y), C.
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldICmpTruncWithTruncOrExt(ICmpInst &Cmp, const SimplifyQuery &Q)
Fold icmp (trunc nuw/nsw X), (trunc nuw/nsw Y).
Instruction * visitLShr(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitUDiv(BinaryOperator &I)
Instruction * visitOr(BinaryOperator &I)
Instruction * foldSignBitTest(ICmpInst &I)
Fold equality-comparison between zero and any (maybe truncated) right-shift by one-less-than-bitwidth...
Instruction * foldSelectEqualityTest(SelectInst &SI)
Instruction * visitZExt(ZExtInst &Zext)
Instruction * visitGEPOfGEP(GetElementPtrInst &GEP, GEPOperator *Src)
Instruction * foldSelectValueEquivalence(SelectInst &SI, CmpInst &CI)
Instruction * visitAddrSpaceCast(AddrSpaceCastInst &CI)
Instruction * foldPHIArgInsertValueInstructionIntoPHI(PHINode &PN)
If we have something like phi [insertvalue(a,b,0), insertvalue(c,d,0)], turn this into a phi[a,...
~InstCombinerImpl() override=default
Instruction * visitSExt(SExtInst &Sext)
Instruction * visitUnreachableInst(UnreachableInst &I)
Instruction * visitURem(BinaryOperator &I)
Instruction * foldSquareSumInt(BinaryOperator &I)
Instruction * foldOpIntoPhi(Instruction &I, PHINode *PN, bool AllowMultipleUses=false)
Given a binary operator, cast instruction, or select which has a PHI node as operand #0,...
InstCombinerImpl(InstructionWorklist &Worklist, BuilderTy &Builder, Function &F, AAResults *AA, AssumptionCache &AC, TargetLibraryInfo &TLI, TargetTransformInfo &TTI, DominatorTree &DT, OptimizationRemarkEmitter &ORE, BlockFrequencyInfo *BFI, BranchProbabilityInfo *BPI, ProfileSummaryInfo *PSI, const DataLayout &DL, ReversePostOrderTraversal< BasicBlock * > &RPOT)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
void handleUnreachableFrom(Instruction *I, SmallVectorImpl< BasicBlock * > &Worklist)
Instruction * foldICmpBinOp(ICmpInst &Cmp, const SimplifyQuery &SQ)
Try to fold icmp (binop), X or icmp X, (binop).
Instruction * foldVectorSelect(SelectInst &Sel)
Instruction * foldCmpLoadFromIndexedGlobal(LoadInst *LI, GetElementPtrInst *GEP, CmpInst &ICI, ConstantInt *AndCst=nullptr)
This is called when we see this pattern: cmp pred (load (gep GV, ...)), cmpcst where GV is a global v...
Instruction * visitFreeze(FreezeInst &I)
Instruction * foldICmpSubConstant(ICmpInst &Cmp, BinaryOperator *Sub, const APInt &C)
Fold icmp (sub X, Y), C.
Instruction * foldSelectShuffle(ShuffleVectorInst &Shuf)
Try to fold shuffles that are the equivalent of a vector select.
Instruction * visitLoadInst(LoadInst &LI)
Value * takeLog2(Value *Op, unsigned Depth, bool AssumeNonZero, bool DoFold)
Take the exact integer log2 of the value.
Instruction * visitFPToSI(FPToSIInst &FI)
Instruction * foldICmpWithClamp(ICmpInst &Cmp, Value *X, MinMaxIntrinsic *Min)
Match and fold patterns like: icmp eq/ne X, min(max(X, Lo), Hi) which represents a range check and ca...
Instruction * foldICmpInstWithConstantNotInt(ICmpInst &Cmp)
Handle icmp with constant (but not simple integer constant) RHS.
Instruction * visitAtomicRMWInst(AtomicRMWInst &SI)
Instruction * visitSRem(BinaryOperator &I)
Instruction * foldSPFofSPF(Instruction *Inner, SelectPatternFlavor SPF1, Value *A, Value *B, Instruction &Outer, SelectPatternFlavor SPF2, Value *C)
Instruction * visitTrunc(TruncInst &CI)
Instruction * foldBinOpSelectBinOp(BinaryOperator &Op)
In some cases it is beneficial to fold a select into a binary operator.
Instruction * foldICmpShlConstConst(ICmpInst &I, Value *ShAmt, const APInt &C1, const APInt &C2)
Handle "(icmp eq/ne (shl AP2, A), AP1)" -> (icmp eq/ne A, TrailingZeros(AP1) - TrailingZeros(AP2)).
Instruction * foldSquareSumFP(BinaryOperator &I)
Value * reassociateShiftAmtsOfTwoSameDirectionShifts(BinaryOperator *Sh0, const SimplifyQuery &SQ, bool AnalyzeForSignBitExtraction=false)
Instruction * foldSelectOpOp(SelectInst &SI, Instruction *TI, Instruction *FI)
We have (select c, TI, FI), and we know that TI and FI have the same opcode.
Instruction * visitUIToFP(CastInst &CI)
Instruction * foldPHIArgBinOpIntoPHI(PHINode &PN)
If we have something like phi [add (a,b), add(a,c)] and if a/b/c and the adds all have a single user,...
void handlePotentiallyDeadBlocks(SmallVectorImpl< BasicBlock * > &Worklist)
bool sinkNotIntoLogicalOp(Instruction &I)
Instruction * foldICmpEqIntrinsicWithConstant(ICmpInst &ICI, IntrinsicInst *II, const APInt &C)
Fold an equality icmp with LLVM intrinsic and constant operand.
Instruction * visitPtrToInt(PtrToIntInst &CI)
bool prepareWorklist(Function &F)
Perform early cleanup and prepare the InstCombine worklist.
Instruction * foldSelectIntrinsic(SelectInst &SI)
This transforms patterns of the form: select cond, intrinsic(x, ...), intrinsic(y,...
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitFDiv(BinaryOperator &I)
Instruction * FoldOpIntoSelect(Instruction &Op, SelectInst *SI, bool FoldWithMultiUse=false, bool SimplifyBothArms=false)
Given an instruction with a select as one operand and a constant as the other operand,...
Instruction * SimplifyAnyMemSet(AnyMemSetInst *MI)
bool simplifyDivRemOfSelectWithZeroOp(BinaryOperator &I)
Fold a divide or remainder with a select instruction divisor when one of the select operands is zero.
Instruction * foldItoFPtoI(FPToIntTy &FI)
fpto{s/u}i.sat --> X or zext(X) or sext(X) or trunc(X) This is safe if the intermediate type has enou...
Instruction * visitSIToFP(CastInst &CI)
Instruction * visitSub(BinaryOperator &I)
Instruction * visitAShr(BinaryOperator &I)
bool replaceInInstruction(Value *V, Value *Old, Value *New, unsigned Depth=0)
Instruction * visitFree(CallInst &FI, Value *FreedOp)
Instruction * visitInsertValueInst(InsertValueInst &IV)
Try to find redundant insertvalue instructions, like the following ones: %0 = insertvalue { i8,...
Instruction * visitAnd(BinaryOperator &I)
Value * foldMultiplicationOverflowCheck(ICmpInst &Cmp)
Fold (-1 u/ x) u< y ((x * y) ?
Instruction * visitCallBrInst(CallBrInst &CBI)
Instruction * visitExtractValueInst(ExtractValueInst &EV)
Instruction * visitInsertElementInst(InsertElementInst &IE)
void handlePotentiallyDeadSuccessors(BasicBlock *BB, BasicBlock *LiveSucc)
Instruction * commonCastTransforms(CastInst &CI)
Implement the transforms common to all CastInst visitors.
Instruction * foldICmpWithConstant(ICmpInst &Cmp)
Fold icmp Pred X, C.
CmpInst * canonicalizeICmpPredicate(CmpInst &I)
If we have a comparison with a non-canonical predicate, if we can update all the users,...
Instruction * foldBinopWithRecurrence(BinaryOperator &BO)
Try to fold binary operators whose operands are simple interleaved recurrences to a single recurrence...
Instruction * eraseInstFromFunction(Instruction &I) override
Combiner aware instruction erasure.
Instruction * foldICmpWithZero(ICmpInst &Cmp)
Instruction * visitExtractElementInst(ExtractElementInst &EI)
Instruction * commonIDivRemTransforms(BinaryOperator &I)
Common integer divide/remainder transforms.
Value * foldReversedIntrinsicOperands(IntrinsicInst *II)
If all arguments of the intrinsic are reverses, try to pull the reverse after the intrinsic.
Instruction * visitPHINode(PHINode &PN)
Instruction * foldICmpCommutative(CmpPredicate Pred, Value *Op0, Value *Op1, ICmpInst &CxtI)
Instruction * foldICmpBinOpEqualityWithConstant(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Fold an icmp equality instruction with binary operator LHS and constant RHS: icmp eq/ne BO,...
Instruction * foldPHIArgOpIntoPHI(PHINode &PN)
Try to rotate an operation below a PHI node, using PHI nodes for its operands.
Instruction * visitLandingPadInst(LandingPadInst &LI)
Instruction * foldICmpUsingBoolRange(ICmpInst &I)
If one operand of an icmp is effectively a bool (value range of {0,1}), then try to reduce patterns b...
Instruction * foldICmpWithTrunc(ICmpInst &Cmp)
Instruction * foldCmpSelectOfConstants(CmpInst &I)
Fold fcmp/icmp pred (select C1, TV1, FV1), (select C2, TV2, FV2) where all true/false values are cons...
Instruction * foldICmpIntrinsicWithConstant(ICmpInst &ICI, IntrinsicInst *II, const APInt &C)
Fold an icmp with LLVM intrinsic and constant operand: icmp Pred II, C.
Instruction * visitFPTrunc(FPTruncInst &CI)
Instruction * visitStoreInst(StoreInst &SI)
Value * tryGetLog2(Value *Op, bool AssumeNonZero)
Instruction * foldPHIArgZextsIntoPHI(PHINode &PN)
TODO: This function could handle other cast types, but then it might require special-casing a cast fr...
Instruction * foldSelectInstWithICmp(SelectInst &SI, ICmpInst *ICI)
Instruction * visitFenceInst(FenceInst &FI)
Value * foldPtrToIntOrAddrOfGEP(Type *IntTy, Value *Ptr)
Instruction * visitFCmpInst(FCmpInst &I)
Value * OptimizePointerDifference(Value *LHS, Value *RHS, Type *Ty, bool isNUW)
Optimize pointer differences into the same array into a size.
Instruction * visitBitCast(BitCastInst &CI)
Instruction * visitReturnInst(ReturnInst &RI)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * commonIDivTransforms(BinaryOperator &I)
This function implements the transforms common to both integer division instructions (udiv and sdiv).
Instruction * foldICmpUsingKnownBits(ICmpInst &Cmp)
Try to fold the comparison based on range information we can get by checking whether bits are known t...
Instruction * foldICmpDivConstant(ICmpInst &Cmp, BinaryOperator *Div, const APInt &C)
Fold icmp ({su}div X, Y), C.
Instruction * foldIRemByPowerOfTwoToBitTest(ICmpInst &I)
If we have: icmp eq/ne (urem/srem x, y), 0 iff y is a power-of-two, we can replace this with a bit te...
Instruction * foldFCmpIntToFPConst(FCmpInst &I, Instruction *LHSI, Constant *RHSC)
Fold fcmp ([us]itofp x, cst) if possible.
Instruction * visitShl(BinaryOperator &I)
Instruction * visitSwitchInst(SwitchInst &SI)
Instruction * foldICmpUDivConstant(ICmpInst &Cmp, BinaryOperator *UDiv, const APInt &C)
Fold icmp (udiv X, Y), C.
Instruction * visitFAdd(BinaryOperator &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * visitIntToPtr(IntToPtrInst &CI)
Instruction * foldICmpAddOpConst(Value *X, const APInt &C, CmpPredicate Pred)
Fold "icmp pred (X+C), X".
Instruction * foldICmpWithCastOp(ICmpInst &ICmp)
Handle icmp (cast x), (cast or constant).
Instruction * visitFPToUI(FPToUIInst &FI)
Instruction * foldICmpTruncConstant(ICmpInst &Cmp, TruncInst *Trunc, const APInt &C)
Fold icmp (trunc X), C.
bool mergeStoreIntoSuccessor(StoreInst &SI)
Try to transform: if () { *P = v1; } else { *P = v2 } or: *P = v1; if () { *P = v2; }...
Instruction * visitPtrToAddr(PtrToAddrInst &CI)
Instruction * visitInstruction(Instruction &I)
Specify what to return for unhandled instructions.
Instruction * foldSelectIntoOp(SelectInst &SI, Value *, Value *)
Try to fold the select into one of the operands to allow further optimization.
Instruction * foldShuffledIntrinsicOperands(IntrinsicInst *II)
If all arguments of the intrinsic are unary shuffles with the same mask, try to shuffle after the int...
Instruction * foldICmpAddConstant(ICmpInst &Cmp, BinaryOperator *Add, const APInt &C)
Fold icmp (add X, Y), C.
Instruction * foldICmpMulConstant(ICmpInst &Cmp, BinaryOperator *Mul, const APInt &C)
Fold icmp (mul X, Y), C.
Instruction * visitInvokeInst(InvokeInst &II)
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * foldVariableSignZeroExtensionOfVariableHighBitExtract(BinaryOperator &OldAShr)
Instruction * visitUncondBrInst(UncondBrInst &BI)
Instruction * commonShiftTransforms(BinaryOperator &I)
Instruction * visitFRem(BinaryOperator &I)
Instruction * foldPHIArgLoadIntoPHI(PHINode &PN)
Instruction * foldICmpXorConstant(ICmpInst &Cmp, BinaryOperator *Xor, const APInt &C)
Fold icmp (xor X, Y), C.
Instruction * FoldOrOfLogicalAnds(Value *Op0, Value *Op1)
Instruction * foldSelectICmp(CmpPredicate Pred, SelectInst *SI, Value *RHS, const ICmpInst &I)
bool foldIntegerTypedPHI(PHINode &PN)
If an integer typed PHI has only one use which is an IntToPtr operation, replace the PHI with an exis...
Instruction * foldICmpInstWithConstantAllowPoison(ICmpInst &Cmp, const APInt &C)
Try to fold integer comparisons with a constant operand: icmp Pred X, C where X is some kind of instr...
bool foldDeadPhiWeb(PHINode &PN)
If the phi is within a phi web, which is formed by the def-use chain of phis and all the phis in the ...
Instruction * foldIsMultipleOfAPowerOfTwo(ICmpInst &Cmp)
Fold icmp eq (num + mask) & ~mask, num to icmp eq (and num, mask), 0 Where mask is a low bit mask.
Instruction * visitXor(BinaryOperator &I)
Value * foldSelectWithConstOpToBinOp(ICmpInst *Cmp, Value *TrueVal, Value *FalseVal)
Value * EvaluateInDifferentType(Value *V, Type *Ty, bool isSigned)
Given an expression that CanEvaluateTruncated or CanEvaluateSExtd returns true for,...
Instruction * simplifyBinOpSplats(ShuffleVectorInst &SVI)
void CreateNonTerminatorUnreachable(Instruction *InsertAt)
Create and insert the idiom we use to indicate a block is unreachable without having to rewrite the C...
Instruction * foldICmpAndShift(ICmpInst &Cmp, BinaryOperator *And, const APInt &C1, const APInt &C2)
Fold icmp (and (sh X, Y), C2), C1.
Value * pushFreezeToPreventPoisonFromPropagating(FreezeInst &FI)
Instruction * foldICmpBinOpWithConstantViaTruthTable(ICmpInst &Cmp, BinaryOperator *BO, const APInt &C)
Instruction * foldICmpInstWithConstant(ICmpInst &Cmp)
Try to fold integer comparisons with a constant operand: icmp Pred X, C where X is some kind of instr...
Instruction * visitSelectInst(SelectInst &SI)
Instruction * foldICmpXorShiftConst(ICmpInst &Cmp, BinaryOperator *Xor, const APInt &C)
For power-of-2 C: ((X s>> ShiftC) ^ X) u< C --> (X + C) u< (C << 1) ((X s>> ShiftC) ^ X) u> (C - 1) -...
Instruction * foldPHIArgIntToPtrToPHI(PHINode &PN)
Instruction * visitFPExt(CastInst &CI)
Instruction * foldICmpShlConstant(ICmpInst &Cmp, BinaryOperator *Shl, const APInt &C)
Fold icmp (shl X, Y), C.
Instruction * visitFMul(BinaryOperator &I)
Instruction * foldSelectOfBools(SelectInst &SI)
Instruction * foldSelectExtConst(SelectInst &Sel)
Instruction * foldAddWithConstant(BinaryOperator &Add)
Instruction * foldICmpAndConstant(ICmpInst &Cmp, BinaryOperator *And, const APInt &C)
Fold icmp (and X, Y), C.
bool run()
Run the combiner over the entire worklist until it is empty.
Instruction * foldFMulReassoc(BinaryOperator &I)
Instruction * SliceUpIllegalIntegerPHI(PHINode &PN)
This is an integer PHI and we know that it has an illegal type: see if it is only used by trunc or tr...
Instruction * foldAggregateConstructionIntoAggregateReuse(InsertValueInst &OrigIVI)
Look for chain of insertvalue's that fully define an aggregate, and trace back the values inserted,...
Instruction * foldICmpEquality(ICmpInst &Cmp)
bool removeInstructionsBeforeUnreachable(Instruction &I)
Instruction * foldPHIArgGEPIntoPHI(PHINode &PN)
Instruction * foldICmpWithMinMax(Instruction &I, MinMaxIntrinsic *MinMax, Value *Z, CmpPredicate Pred)
Fold icmp Pred min|max(X, Y), Z.
Instruction * visitShuffleVectorInst(ShuffleVectorInst &SVI)
Instruction * FoldShiftByConstant(Value *Op0, Constant *Op1, BinaryOperator &I)
void tryToSinkInstructionDbgVariableRecords(Instruction *I, BasicBlock::iterator InsertPos, BasicBlock *SrcBlock, BasicBlock *DestBlock, SmallVectorImpl< DbgVariableRecord * > &DPUsers)
bool foldAllocaCmp(AllocaInst *Alloca)
void addDeadEdge(BasicBlock *From, BasicBlock *To, SmallVectorImpl< BasicBlock * > &Worklist)
void PHIArgMergedDebugLoc(Instruction *Inst, PHINode &PN)
Helper function for FoldPHIArgXIntoPHI() to set debug location for the folded operation.
Instruction * visitVAEndInst(VAEndInst &I)
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
Constant * unshuffleConstant(ArrayRef< int > ShMask, Constant *C, VectorType *NewCTy)
Find a constant NewC that has property: shuffle(NewC, ShMask) = C Returns nullptr if such a constant ...
Instruction * visitAllocSite(Instruction &FI)
Instruction * visitICmpInst(ICmpInst &I)
Instruction * SimplifyAnyMemTransfer(AnyMemTransferInst *MI)
Instruction * visitGetElementPtrInst(GetElementPtrInst &GEP)
Instruction * foldPowiReassoc(BinaryOperator &I)
Instruction * foldFreezeIntoRecurrence(FreezeInst &I, PHINode *PN)
Instruction * visitSDiv(BinaryOperator &I)
bool tryToSinkInstruction(Instruction *I, BasicBlock *DestBlock)
Try to move the specified instruction from its current block into the beginning of DestBlock,...
Instruction * foldPHIArgExtractValueInstructionIntoPHI(PHINode &PN)
If we have something like phi [extractvalue(a,0), extractvalue(b,0)], turn this into a phi[a,...
Instruction * foldICmpShrConstConst(ICmpInst &I, Value *ShAmt, const APInt &C1, const APInt &C2)
Handle "(icmp eq/ne (ashr/lshr AP2, A), AP1)" -> (icmp eq/ne A, Log2(AP2/AP1)) -> (icmp eq/ne A,...
bool freezeOtherUses(FreezeInst &FI)
Instruction * visitFNeg(UnaryOperator &I)
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
Instruction * commonIRemTransforms(BinaryOperator &I)
This function implements the transforms common to both integer remainder instructions (urem and srem)...
Instruction * visitAllocaInst(AllocaInst &AI)
Instruction * visitCallInst(CallInst &CI)
CallInst simplification.
Instruction * foldICmpAndConstConst(ICmpInst &Cmp, BinaryOperator *And, const APInt &C1)
Fold icmp (and X, C2), C1.
Instruction * visitFSub(BinaryOperator &I)
Instruction * foldICmpBitCast(ICmpInst &Cmp)
Instruction * foldGEPICmp(GEPOperator *GEPLHS, Value *RHS, CmpPredicate Cond, Instruction &I)
Fold comparisons between a GEP instruction and something else.
SimplifyQuery SQ
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
BlockFrequencyInfo * BFI
TargetLibraryInfo & TLI
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
Instruction * InsertNewInstWith(Instruction *New, BasicBlock::iterator Old)
Same as InsertNewInstBefore, but also sets the debug loc.
BranchProbabilityInfo * BPI
virtual bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known, const SimplifyQuery &Q, unsigned Depth=0)=0
ReversePostOrderTraversal< BasicBlock * > & RPOT
const DataLayout & DL
DomConditionCache DC
AssumptionCache & AC
DominatorTree & DT
ProfileSummaryInfo * PSI
BuilderTy & Builder
OptimizationRemarkEmitter & ORE
InstCombiner(InstructionWorklist &Worklist, BuilderTy &Builder, Function &F, AAResults *AA, AssumptionCache &AC, TargetLibraryInfo &TLI, TargetTransformInfo &TTI, DominatorTree &DT, OptimizationRemarkEmitter &ORE, BlockFrequencyInfo *BFI, BranchProbabilityInfo *BPI, ProfileSummaryInfo *PSI, const DataLayout &DL, ReversePostOrderTraversal< BasicBlock * > &RPOT)
Base class for instruction visitors.
Definition InstVisitor.h:78
InstructionWorklist - This is the worklist management logic for InstCombine and other simplification ...
This class represents a cast from an integer to a pointer.
A wrapper class for inspecting calls to intrinsic functions.
Invoke instruction.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
The landingpad instruction holds all of the information necessary to generate correct exception handl...
An instruction for reading from memory.
This class represents min/max intrinsics.
static Value * Negate(bool LHSIsZero, bool IsNSW, Value *Root, InstCombinerImpl &IC)
Attempt to negate Root.
The optimization diagnostic interface.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
static LLVM_ABI PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
Analysis providing profile information.
This class represents a cast from a pointer to an address (non-capturing ptrtoint).
This class represents a cast from a pointer to an integer.
Return a value (possibly void), from a function.
This class represents a sign extension of integer types.
This class represents the LLVM 'select' instruction.
This instruction constructs a fixed permutation of two input vectors.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
An instruction for storing to memory.
Multiway switch.
Provides information about what library functions are available for the current target.
This pass provides access to the codegen interfaces that are needed for IR-level transformations.
This class represents a truncation of integer types.
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
Unconditional Branch instruction.
This function has undefined behavior.
This represents the llvm.va_end intrinsic.
LLVM Value Representation.
Definition Value.h:75
LLVMContext & getContext() const
All values hold a context through their type.
Definition Value.h:258
Base class of all SIMD vector types.
This class represents zero extension of integer types.
self_iterator getIterator()
Definition ilist_node.h:123
CallInst * Call
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
This is an optimization pass for GlobalISel generic memory operations.
ShiftSemantics
Enum to specify how shift operations should be evaluated in canEvaluateShifted.
@ NeverOverflows
Never overflows.
LLVM_ABI void setExplicitlyUnknownBranchWeightsIfProfiled(Instruction &I, StringRef PassName, const Function *F=nullptr)
Like setExplicitlyUnknownBranchWeights(...), but only sets unknown branch weights in the new instruct...
LLVM_ABI void salvageDebugInfo(const MachineRegisterInfo &MRI, MachineInstr &MI)
Assuming the instruction MI is going to be deleted, attempt to salvage debug users of MI by writing t...
Definition Utils.cpp:1682
LLVM_ABI OverflowResult computeOverflowForUnsignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ, bool IsNSW=false)
LLVM_ABI OverflowResult computeOverflowForSignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
SelectPatternFlavor
Specific patterns of select instructions we can match.
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
LLVM_ABI OverflowResult computeOverflowForSignedMul(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
TargetTransformInfo TTI
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
DWARFExpression::Operation Op
constexpr unsigned BitWidth
LLVM_ABI OverflowResult computeOverflowForUnsignedSub(const Value *LHS, const Value *RHS, const SimplifyQuery &SQ)
LLVM_ABI OverflowResult computeOverflowForUnsignedAdd(const WithCache< const Value * > &LHS, const WithCache< const Value * > &RHS, const SimplifyQuery &SQ)
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Value * Ptr
Common base pointer.
SmallVector< GEPOperator * > RHSGEPs
RHS GEPs until common base.
GEPNoWrapFlags LHSNW
LHS GEP NoWrapFlags until common base.
GEPNoWrapFlags RHSNW
RHS GEP NoWrapFlags until common base.
SmallVector< GEPOperator * > LHSGEPs
LHS GEPs until common base.
bool isExpensive() const
Whether expanding the GEP chains is expensive.
static CommonPointerBase compute(Value *LHS, Value *RHS)
Matching combinators.