30#define DEBUG_TYPE "instcombine"
45 return Builder.CreateICmp(NewPred,
LHS,
RHS);
55 return Builder.CreateFCmpFMF(NewPred,
LHS,
RHS, FMF);
65 "Lo is not < Hi in range emission code!");
67 Type *Ty = V->getType();
72 if (
isSigned ?
Lo.isMinSignedValue() :
Lo.isMinValue()) {
74 return Builder.CreateICmp(Pred, V, ConstantInt::get(Ty,
Hi));
80 Builder.CreateSub(V, ConstantInt::get(Ty,
Lo), V->getName() +
".off");
82 return Builder.CreateICmp(Pred, VMinusLo, HiMinusLo);
129 const APInt *ConstA =
nullptr, *ConstB =
nullptr, *ConstC =
nullptr;
134 bool IsAPow2 = ConstA && ConstA->
isPowerOf2();
135 bool IsBPow2 = ConstB && ConstB->isPowerOf2();
136 unsigned MaskVal = 0;
137 if (ConstC && ConstC->isZero()) {
156 }
else if (ConstA && ConstC && ConstC->
isSubsetOf(*ConstA)) {
166 }
else if (ConstB && ConstC && ConstC->isSubsetOf(*ConstB)) {
201 Y = ConstantInt::get(
X->getType(), Res->Mask);
202 Z = ConstantInt::get(
X->getType(), Res->C);
211static std::optional<std::pair<unsigned, unsigned>>
224 Value *L1, *L11, *L12, *L2, *L21, *L22;
226 L21 = L22 = L1 =
nullptr;
233 if (!LHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
236 PredL = LHSCMP->getPredicate();
237 L1 = LHSCMP->getOperand(0);
238 L2 = LHSCMP->getOperand(1);
259 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
262 }
else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
274 if (!RHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
277 PredR = RHSCMP->getPredicate();
279 Value *R1 = RHSCMP->getOperand(0);
280 R2 = RHSCMP->getOperand(1);
289 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
294 }
else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
312 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
316 }
else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
333 }
else if (L12 ==
A) {
336 }
else if (L21 ==
A) {
339 }
else if (L22 ==
A) {
346 return std::optional<std::pair<unsigned, unsigned>>(
347 std::make_pair(LeftType, RightType));
369 const APInt *BCst, *DCst, *OrigECst;
380 APInt ECst = *OrigECst;
386 if (*BCst == 0 || *DCst == 0)
396 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
397 Attribute::StrictFP)) {
399 if (!Ty->isIEEELikeFPTy())
405 APInt FractionBits = ~ExpBits;
407 if (*BCst != FractionBits)
432 if ((((*BCst & *DCst) & ECst) == 0) &&
433 (*BCst & (*BCst ^ *DCst)).isPowerOf2()) {
434 APInt BorD = *BCst | *DCst;
435 APInt BandBxorDorE = (*BCst & (*BCst ^ *DCst)) | ECst;
436 Value *NewMask = ConstantInt::get(
A->getType(), BorD);
437 Value *NewMaskedValue = ConstantInt::get(
A->getType(), BandBxorDorE);
438 Value *NewAnd = Builder.CreateAnd(
A, NewMask);
439 return Builder.CreateICmp(NewCC, NewAnd, NewMaskedValue);
442 auto IsSubSetOrEqual = [](
const APInt *C1,
const APInt *C2) {
443 return (*C1 & *C2) == *C1;
445 auto IsSuperSetOrEqual = [](
const APInt *C1,
const APInt *C2) {
446 return (*C1 & *C2) == *C2;
455 if (!IsSubSetOrEqual(BCst, DCst) && !IsSuperSetOrEqual(BCst, DCst))
467 if (IsSubSetOrEqual(BCst, DCst))
468 return ConstantInt::get(
LHS->getType(), !IsAnd);
478 if (IsSuperSetOrEqual(BCst, DCst)) {
481 ICmp->setSameSign(
false);
487 assert(IsSubSetOrEqual(BCst, DCst) &&
"Precondition due to above code");
488 if ((*BCst & ECst) != 0) {
491 ICmp->setSameSign(
false);
498 return ConstantInt::get(
LHS->getType(), !IsAnd);
510 "Expected equality predicates for masked type of icmps.");
522 LHS,
RHS, IsAnd,
A,
B,
D,
E, PredL, PredR, Builder)) {
527 RHS,
LHS, IsAnd,
A,
D,
B,
C, PredR, PredL, Builder)) {
540 Value *
A =
nullptr, *
B =
nullptr, *
C =
nullptr, *
D =
nullptr, *
E =
nullptr;
542 std::optional<std::pair<unsigned, unsigned>> MaskPair =
547 "Expected equality predicates for masked type of icmps.");
548 unsigned LHSMask = MaskPair->first;
549 unsigned RHSMask = MaskPair->second;
550 unsigned Mask = LHSMask & RHSMask;
555 LHS,
RHS, IsAnd,
A,
B,
C,
D,
E, PredL, PredR, LHSMask, RHSMask,
585 Value *NewOr = Builder.CreateOr(
B,
D);
586 Value *NewAnd = Builder.CreateAnd(
A, NewOr);
591 return Builder.CreateICmp(NewCC, NewAnd, Zero);
598 Value *NewOr = Builder.CreateOr(
B,
D);
599 Value *NewAnd = Builder.CreateAnd(
A, NewOr);
600 return Builder.CreateICmp(NewCC, NewAnd, NewOr);
607 Value *NewAnd1 = Builder.CreateAnd(
B,
D);
608 Value *NewAnd2 = Builder.CreateAnd(
A, NewAnd1);
609 return Builder.CreateICmp(NewCC, NewAnd2,
A);
612 const APInt *ConstB, *ConstD;
620 APInt NewMask = *ConstB & *ConstD;
621 if (NewMask == *ConstB)
623 if (NewMask == *ConstD) {
626 RHSI->dropPoisonGeneratingFlags();
637 APInt NewMask = *ConstB | *ConstD;
638 if (NewMask == *ConstB)
640 if (NewMask == *ConstD)
667 const APInt *OldConstC, *OldConstE;
673 const APInt ConstC = PredL != CC ? *ConstB ^ *OldConstC : *OldConstC;
674 const APInt ConstE = PredR != CC ? *ConstD ^ *OldConstE : *OldConstE;
676 if (((*ConstB & *ConstD) & (ConstC ^ ConstE)).getBoolValue())
677 return IsNot ? nullptr : ConstantInt::get(
LHS->getType(), !IsAnd);
680 !ConstD->isSubsetOf(*ConstB))
685 BD = *ConstB & *ConstD;
686 CE = ConstC & ConstE;
688 BD = *ConstB | *ConstD;
689 CE = ConstC | ConstE;
691 Value *NewAnd = Builder.CreateAnd(
A, BD);
692 Value *CEVal = ConstantInt::get(
A->getType(), CE);
693 return Builder.CreateICmp(CC, NewAnd, CEVal);
697 return FoldBMixed(NewCC,
false);
699 return FoldBMixed(NewCC,
true);
714 D = Builder.CreateFreeze(
D);
715 Value *Mask = Builder.CreateOr(
B,
D);
717 return Builder.CreateICmp(NewCC,
Masked, Mask);
767 default:
return nullptr;
791 if (
LHS->getPredicate() != Pred ||
RHS->getPredicate() != Pred)
816 return Builder.CreateICmp(Pred,
And,
Op);
855 auto tryToMatchSignedTruncationCheck = [](
ICmpInst *ICmp,
Value *&
X,
856 APInt &SignBitMask) ->
bool {
857 const APInt *I01, *I1;
861 I1->ugt(*I01) && I01->
shl(1) == *I1))
873 if (tryToMatchSignedTruncationCheck(ICmp1, X1, HighestBit))
875 else if (tryToMatchSignedTruncationCheck(ICmp0, X1, HighestBit))
880 assert(HighestBit.
isPowerOf2() &&
"expected to be power of two (non-zero)");
884 APInt &UnsetBitsMask) ->
bool {
893 UnsetBitsMask = Res->Mask;
903 if (!tryToDecompose(OtherICmp, X0, UnsetBitsMask))
906 assert(!UnsetBitsMask.
isZero() &&
"empty mask makes no sense.");
921 APInt SignBitsMask = ~(HighestBit - 1U);
928 if (!UnsetBitsMask.
isSubsetOf(SignBitsMask)) {
929 APInt OtherHighestBit = (~UnsetBitsMask) + 1U;
937 return Builder.CreateICmpULT(
X, ConstantInt::get(
X->getType(), HighestBit),
938 CxtI.
getName() +
".simplified");
957 CtPop->dropPoisonGeneratingAnnotations();
959 return Builder.CreateICmpUGT(CtPop, ConstantInt::get(CtPop->getType(), 1));
963 CtPop->dropPoisonGeneratingAnnotations();
965 return Builder.CreateICmpULT(CtPop, ConstantInt::get(CtPop->getType(), 2));
991 CtPop->dropPoisonGeneratingAnnotations();
993 return Builder.CreateICmpEQ(CtPop, ConstantInt::get(CtPop->getType(), 1));
1002 CtPop->dropPoisonGeneratingAnnotations();
1004 return Builder.CreateICmpNE(CtPop, ConstantInt::get(CtPop->getType(), 1));
1018 "Expected equality predicates for masked type of icmps.");
1038 const APInt *BCst, *DCst, *ECst;
1052 if (!BFVTy || !BConst || !DConst || !EConst)
1055 for (
unsigned I = 0;
I != BFVTy->getNumElements(); ++
I) {
1056 const auto *BElt = BConst->getAggregateElement(
I);
1057 const auto *DElt = DConst->getAggregateElement(
I);
1058 const auto *EElt = EConst->getAggregateElement(
I);
1060 if (!BElt || !DElt || !EElt)
1062 if (!isReducible(BElt, DElt, EElt))
1067 if (!isReducible(
B,
D,
E))
1085 Value *
A =
nullptr, *
B =
nullptr, *
C =
nullptr, *
D =
nullptr, *
E =
nullptr;
1090 std::optional<std::pair<unsigned, unsigned>> MaskPair =
1096 unsigned CmpMask0 = MaskPair->first;
1097 unsigned CmpMask1 = MaskPair->second;
1098 if ((CmpMask0 &
Mask_AllZeros) && (CmpMask1 == compareBMask)) {
1102 }
else if ((CmpMask0 == compareBMask) && (CmpMask1 &
Mask_AllZeros)) {
1113 ICmpInst *UnsignedICmp,
bool IsAnd,
1125 if (
match(UnsignedICmp,
1141 IsAnd && GetKnownNonZeroAndOther(
B,
A))
1142 return Builder.CreateICmpULT(Builder.CreateNeg(
B),
A);
1144 !IsAnd && GetKnownNonZeroAndOther(
B,
A))
1145 return Builder.CreateICmpUGE(Builder.CreateNeg(
B),
A);
1161 return std::nullopt;
1163 unsigned NumOriginalBits =
X->getType()->getScalarSizeInBits();
1164 unsigned NumExtractedBits = V->getType()->getScalarSizeInBits();
1170 Shift->
ule(NumOriginalBits - NumExtractedBits))
1172 return {{
X, 0, NumExtractedBits}};
1179 V = Builder.CreateLShr(V,
P.StartBit);
1181 if (TruncTy != V->getType())
1182 V = Builder.CreateTrunc(V, TruncTy);
1189Value *InstCombinerImpl::foldEqOfParts(
Value *Cmp0,
Value *Cmp1,
bool IsAnd) {
1194 auto GetMatchPart = [&](
Value *CmpV,
1195 unsigned OpNo) -> std::optional<IntPart> {
1204 return {{OpNo == 0 ?
X :
Y, 0, 1}};
1208 return std::nullopt;
1210 if (Pred ==
Cmp->getPredicate())
1219 return std::nullopt;
1228 return std::nullopt;
1230 return std::nullopt;
1235 return {{
I->getOperand(OpNo), From,
C->getBitWidth() - From}};
1238 std::optional<IntPart> L0 = GetMatchPart(Cmp0, 0);
1239 std::optional<IntPart> R0 = GetMatchPart(Cmp0, 1);
1240 std::optional<IntPart> L1 = GetMatchPart(Cmp1, 0);
1241 std::optional<IntPart> R1 = GetMatchPart(Cmp1, 1);
1242 if (!L0 || !R0 || !L1 || !R1)
1247 if (L0->From != L1->From || R0->From != R1->From) {
1248 if (L0->From != R1->From || R0->From != L1->From)
1255 if (L0->StartBit + L0->NumBits != L1->StartBit ||
1256 R0->StartBit + R0->NumBits != R1->StartBit) {
1257 if (L1->StartBit + L1->NumBits != L0->StartBit ||
1258 R1->StartBit + R1->NumBits != R0->StartBit)
1265 IntPart
L = {L0->From, L0->StartBit, L0->NumBits + L1->NumBits};
1266 IntPart
R = {R0->From, R0->StartBit, R0->NumBits + R1->NumBits};
1276 bool IsAnd,
bool IsLogical,
1306 if (!SubstituteCmp) {
1311 SubstituteCmp = Builder.CreateICmp(Pred1,
Y,
C);
1316 return IsAnd ? Builder.CreateLogicalAnd(Cmp0, SubstituteCmp,
"", MDFrom)
1317 : Builder.CreateLogicalOr(Cmp0, SubstituteCmp,
"", MDFrom);
1319 return Builder.CreateBinOp(IsAnd ? Instruction::And : Instruction::Or, Cmp0,
1327Value *InstCombinerImpl::foldAndOrOfICmpsUsingRanges(
ICmpInst *ICmp1,
1331 auto MatchExactRangeCheck =
1332 [](ICmpInst *ICmp) -> std::optional<std::pair<Value *, ConstantRange>> {
1335 return std::nullopt;
1337 CmpPredicate Pred = ICmp->getPredicate();
1343 C->countr_zero() >=
Mask->countr_zero()) {
1344 ConstantRange CR(*
C, *
C - *Mask);
1347 return std::make_pair(
X, CR);
1354 return std::make_pair(
X, CR.
subtract(*C1));
1355 return std::make_pair(
LHS, CR);
1358 auto RC1 = MatchExactRangeCheck(ICmp1);
1362 auto RC2 = MatchExactRangeCheck(ICmp2);
1366 auto &[V1, CR1] = *RC1;
1367 auto &[V2, CR2] = *RC2;
1373 CR1 = CR1.inverse();
1374 CR2 = CR2.inverse();
1377 Type *Ty = V1->getType();
1387 APInt LowerDiff = CR1.getLower() ^ CR2.getLower();
1388 APInt UpperDiff = (CR1.getUpper() - 1) ^ (CR2.getUpper() - 1);
1389 APInt CR1Size = CR1.getUpper() - CR1.getLower();
1390 if (!LowerDiff.
isPowerOf2() || LowerDiff != UpperDiff ||
1391 CR1Size != CR2.getUpper() - CR2.getLower())
1394 CR = CR1.getLower().ult(CR2.getLower()) ? CR1 : CR2;
1395 NewV =
Builder.CreateAnd(NewV, ConstantInt::get(Ty, ~LowerDiff));
1403 CR->getEquivalentICmp(NewPred, NewC,
Offset);
1406 NewV =
Builder.CreateAdd(NewV, ConstantInt::get(Ty,
Offset));
1407 return Builder.CreateICmp(NewPred, NewV, ConstantInt::get(Ty, NewC));
1426 Value *LHS0 =
LHS->getOperand(0), *LHS1 =
LHS->getOperand(1);
1427 Value *RHS0 =
RHS->getOperand(0), *RHS1 =
RHS->getOperand(1);
1439 bool IsAnd,
bool IsLogicalSelect) {
1440 Value *LHS0 =
LHS->getOperand(0), *LHS1 =
LHS->getOperand(1);
1441 Value *RHS0 =
RHS->getOperand(0), *RHS1 =
RHS->getOperand(1);
1444 if (LHS0 == RHS1 && RHS0 == LHS1) {
1464 if (LHS0 == RHS0 && LHS1 == RHS1) {
1467 unsigned NewPred = IsAnd ? FCmpCodeL & FCmpCodeR : FCmpCodeL | FCmpCodeR;
1476 if (!IsLogicalSelect &&
1489 return Builder.CreateFCmpFMF(PredL, LHS0, RHS0,
1495 if (!IsLogicalSelect && IsAnd &&
1511 auto [ClassValRHS, ClassMaskRHS] =
1514 auto [ClassValLHS, ClassMaskLHS] =
1516 if (ClassValLHS == ClassValRHS) {
1517 unsigned CombinedMask = IsAnd ? (ClassMaskLHS & ClassMaskRHS)
1518 : (ClassMaskLHS | ClassMaskRHS);
1519 return Builder.CreateIntrinsic(
1520 Intrinsic::is_fpclass, {ClassValLHS->getType()},
1521 {ClassValLHS,
Builder.getInt32(CombinedMask)});
1549 if (IsLessThanOrLessEqual(IsAnd ? PredR : PredL)) {
1553 if (IsLessThanOrLessEqual(IsAnd ? PredL : PredR)) {
1554 FastMathFlags NewFlag =
LHS->getFastMathFlags();
1555 if (!IsLogicalSelect)
1556 NewFlag |=
RHS->getFastMathFlags();
1560 PredL, FAbs, ConstantFP::get(LHS0->
getType(), *LHSC), NewFlag);
1572 if (!FCmp || !FCmp->hasOneUse())
1575 std::tie(ClassVal, ClassMask) =
1576 fcmpToClassTest(FCmp->getPredicate(), *FCmp->getParent()->getParent(),
1577 FCmp->getOperand(0), FCmp->getOperand(1));
1578 return ClassVal !=
nullptr;
1589 Value *ClassVal0 =
nullptr;
1590 Value *ClassVal1 =
nullptr;
1591 uint64_t ClassMask0, ClassMask1;
1607 ClassVal0 == ClassVal1) {
1608 unsigned NewClassMask;
1610 case Instruction::And:
1611 NewClassMask = ClassMask0 & ClassMask1;
1613 case Instruction::Or:
1614 NewClassMask = ClassMask0 | ClassMask1;
1616 case Instruction::Xor:
1617 NewClassMask = ClassMask0 ^ ClassMask1;
1626 1, ConstantInt::get(
II->getArgOperand(1)->getType(), NewClassMask));
1633 1, ConstantInt::get(
II->getArgOperand(1)->getType(), NewClassMask));
1637 CallInst *NewClass =
1638 Builder.CreateIntrinsic(Intrinsic::is_fpclass, {ClassVal0->
getType()},
1639 {ClassVal0,
Builder.getInt32(NewClassMask)});
1653Instruction *InstCombinerImpl::canonicalizeConditionalNegationViaMathToSelect(
1655 assert(
I.getOpcode() == BinaryOperator::Xor &&
"Only for xor!");
1660 !
Cond->getType()->isIntOrIntVectorTy(1) ||
1663 return createSelectInstWithUnknownProfile(
1674 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1675 "Expecting and/or op for fcmp transform");
1694 X->getType() !=
Y->getType())
1698 X->getType() !=
Y->getType())
1715 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1716 "Trying to match De Morgan's Laws with something other than and/or");
1720 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
1722 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
1748bool InstCombinerImpl::shouldOptimizeCast(
CastInst *CI) {
1758 if (isEliminableCastPair(PrecedingCI, CI))
1786 auto *ZExt =
new ZExtInst(NewOp, DestTy);
1787 ZExt->setNonNeg(Flags.NNeg);
1788 ZExt->andIRFlags(Cast);
1797 return new SExtInst(NewOp, DestTy);
1807 assert(
I.isBitwiseLogicOp() &&
"Unexpected opcode for bitwise logic folding");
1809 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
1815 auto FoldBitwiseICmpZeroWithICmp = [&](
Value *Op0,
1816 Value *Op1) -> Instruction * {
1831 auto *BitwiseOp =
Builder.CreateBinOp(LogicOpc, ICmpL, ICmpR);
1833 return new ZExtInst(BitwiseOp, Op0->
getType());
1836 if (
auto *Ret = FoldBitwiseICmpZeroWithICmp(Op0, Op1))
1839 if (
auto *Ret = FoldBitwiseICmpZeroWithICmp(Op1, Op0))
1848 Type *DestTy =
I.getType();
1874 unsigned XNumBits =
X->getType()->getScalarSizeInBits();
1875 unsigned YNumBits =
Y->getType()->getScalarSizeInBits();
1876 if (XNumBits != YNumBits) {
1884 if (XNumBits < YNumBits) {
1885 X =
Builder.CreateCast(CastOpcode,
X,
Y->getType());
1886 }
else if (YNumBits < XNumBits) {
1887 Y =
Builder.CreateCast(CastOpcode,
Y,
X->getType());
1892 Value *NarrowLogic =
Builder.CreateBinOp(LogicOpc,
X,
Y,
I.getName());
1895 if (Disjoint && NewDisjoint)
1896 NewDisjoint->setIsDisjoint(Disjoint->isDisjoint());
1908 if (shouldOptimizeCast(Cast0) && shouldOptimizeCast(Cast1)) {
1909 Value *NewOp =
Builder.CreateBinOp(LogicOpc, Cast0Src, Cast1Src,
1919 assert(
I.getOpcode() == Instruction::And);
1920 Value *Op0 =
I.getOperand(0);
1921 Value *Op1 =
I.getOperand(1);
1929 return BinaryOperator::CreateXor(
A,
B);
1945 assert(
I.getOpcode() == Instruction::Or);
1946 Value *Op0 =
I.getOperand(0);
1947 Value *Op1 =
I.getOperand(1);
1972 return BinaryOperator::CreateXor(
A,
B);
1992 Value *Op0 =
And.getOperand(0), *Op1 =
And.getOperand(1);
2013 if (
Opc == Instruction::LShr ||
Opc == Instruction::Shl)
2022 return new ZExtInst(
Builder.CreateAnd(NewBO,
X), Ty);
2030 assert(Opcode == Instruction::And || Opcode == Instruction::Or);
2034 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
2036 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
2043 const auto matchNotOrAnd =
2044 [Opcode, FlippedOpcode](
Value *
Op,
auto m_A,
auto m_B,
auto m_C,
2045 Value *&
X,
bool CountUses =
false) ->
bool {
2046 if (CountUses && !
Op->hasOneUse())
2052 return !CountUses ||
X->hasOneUse();
2068 return (Opcode == Instruction::Or)
2069 ? BinaryOperator::CreateAnd(
Xor, Builder.CreateNot(
A))
2078 return (Opcode == Instruction::Or)
2079 ? BinaryOperator::CreateAnd(
Xor, Builder.CreateNot(
B))
2088 Opcode, Builder.CreateBinOp(FlippedOpcode,
B,
C),
A));
2095 Opcode, Builder.CreateBinOp(FlippedOpcode,
A,
C),
B));
2101 if (Opcode == Instruction::Or && Op0->
hasOneUse() &&
2139 return (Opcode == Instruction::Or)
2141 : BinaryOperator::CreateOr(
Xor,
X);
2149 FlippedOpcode, Builder.CreateBinOp(Opcode,
C, Builder.CreateNot(
B)),
2157 FlippedOpcode, Builder.CreateBinOp(Opcode,
B, Builder.CreateNot(
C)),
2177 if (!
X->hasOneUse()) {
2178 Value *YZ = Builder.CreateBinOp(Opcode,
Y, Z);
2182 if (!
Y->hasOneUse()) {
2183 Value *XZ = Builder.CreateBinOp(Opcode,
X, Z);
2203 Type *Ty =
I.getType();
2205 Value *Op0 =
I.getOperand(0);
2206 Value *Op1 =
I.getOperand(1);
2214 unsigned Width = Ty->getScalarSizeInBits();
2218 case Instruction::And:
2219 if (
C->countl_one() < LastOneMath)
2222 case Instruction::Xor:
2223 case Instruction::Or:
2224 if (
C->countl_zero() < LastOneMath)
2231 Value *NewBinOp = Builder.CreateBinOp(OpC,
X, ConstantInt::get(Ty, *
C));
2233 ConstantInt::get(Ty, *C2), Op0);
2240 assert((
I.isBitwiseLogicOp() ||
I.getOpcode() == Instruction::Add) &&
2241 "Unexpected opcode");
2244 Constant *ShiftedC1, *ShiftedC2, *AddC;
2245 Type *Ty =
I.getType();
2261 if (!Op0Inst || !Op1Inst)
2267 if (ShiftOp != Op1Inst->getOpcode())
2271 if (
I.getOpcode() == Instruction::Add && ShiftOp != Instruction::Shl)
2275 I.getOpcode(), ShiftedC1,
Builder.CreateBinOp(ShiftOp, ShiftedC2, AddC));
2291 assert(
I.isBitwiseLogicOp() &&
"Should and/or/xor");
2292 if (!
I.getOperand(0)->hasOneUse())
2299 if (
Y && (!
Y->hasOneUse() ||
X->getIntrinsicID() !=
Y->getIntrinsicID()))
2305 if (!
Y && (!(IID == Intrinsic::bswap || IID == Intrinsic::bitreverse) ||
2310 case Intrinsic::fshl:
2311 case Intrinsic::fshr: {
2312 if (
X->getOperand(2) !=
Y->getOperand(2))
2315 Builder.CreateBinOp(
I.getOpcode(),
X->getOperand(0),
Y->getOperand(0));
2317 Builder.CreateBinOp(
I.getOpcode(),
X->getOperand(1),
Y->getOperand(1));
2322 case Intrinsic::bswap:
2323 case Intrinsic::bitreverse: {
2324 Value *NewOp0 = Builder.CreateBinOp(
2325 I.getOpcode(),
X->getOperand(0),
2326 Y ?
Y->getOperand(0)
2327 : ConstantInt::get(
I.getType(), IID == Intrinsic::bswap
2347 unsigned Depth = 0) {
2355 if (!
I || !
I->isBitwiseLogicOp() ||
Depth >= 3)
2358 if (!
I->hasOneUse())
2359 SimplifyOnly =
true;
2362 SimplifyOnly, IC,
Depth + 1);
2364 SimplifyOnly, IC,
Depth + 1);
2365 if (!NewOp0 && !NewOp1)
2369 NewOp0 =
I->getOperand(0);
2371 NewOp1 =
I->getOperand(1);
2387 bool RHSIsLogical) {
2389 Value *Folded =
nullptr;
2392 if (
Value *Res = foldBooleanAndOr(
LHS,
X,
I, IsAnd,
false))
2393 Folded = RHSIsLogical ?
Builder.CreateLogicalOp(Opcode, Res,
Y)
2394 :
Builder.CreateBinOp(Opcode, Res,
Y);
2397 else if (
Value *Res = foldBooleanAndOr(
LHS,
Y,
I, IsAnd,
false))
2398 Folded = RHSIsLogical ?
Builder.CreateLogicalOp(Opcode,
X, Res)
2399 :
Builder.CreateBinOp(Opcode,
X, Res);
2417 Type *Ty =
I.getType();
2420 SQ.getWithInstruction(&
I)))
2451 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
2460 Value *IsZero =
Builder.CreateICmpEQ(
X, ConstantInt::get(Ty, 0));
2470 return createSelectInstWithUnknownProfile(Cmp,
2480 return BinaryOperator::CreateAnd(
Builder.CreateNot(
X),
Y);
2486 Constant *NewC = ConstantInt::get(Ty, *
C & *XorC);
2489 return BinaryOperator::CreateXor(
And, NewC);
2500 APInt Together = *
C & *OrC;
2503 return BinaryOperator::CreateOr(
And, ConstantInt::get(Ty, Together));
2506 unsigned Width = Ty->getScalarSizeInBits();
2507 const APInt *ShiftC;
2509 ShiftC->
ult(Width)) {
2514 Constant *ShAmtC = ConstantInt::get(Ty, ShiftC->
zext(Width));
2515 return BinaryOperator::CreateLShr(Sext, ShAmtC);
2523 return BinaryOperator::CreateLShr(
X, ConstantInt::get(Ty, *ShiftC));
2531 if (Op0->
hasOneUse() &&
C->isPowerOf2() && (*AddC & (*
C - 1)) == 0) {
2532 assert((*
C & *AddC) != 0 &&
"Expected common bit");
2534 return BinaryOperator::CreateXor(NewAnd, Op1);
2541 switch (
B->getOpcode()) {
2542 case Instruction::Xor:
2543 case Instruction::Or:
2544 case Instruction::Mul:
2545 case Instruction::Add:
2546 case Instruction::Sub:
2562 C->isIntN(
X->getType()->getScalarSizeInBits())) {
2563 unsigned XWidth =
X->getType()->getScalarSizeInBits();
2564 Constant *TruncC1 = ConstantInt::get(
X->getType(), C1->
trunc(XWidth));
2566 ?
Builder.CreateBinOp(BOpcode,
X, TruncC1)
2567 :
Builder.CreateBinOp(BOpcode, TruncC1,
X);
2568 Constant *TruncC = ConstantInt::get(
X->getType(),
C->trunc(XWidth));
2578 C->isMask(
X->getType()->getScalarSizeInBits())) {
2580 Value *TrY =
Builder.CreateTrunc(
Y,
X->getType(),
Y->getName() +
".tr");
2588 C->isMask(
X->getType()->getScalarSizeInBits())) {
2590 Value *TrY =
Builder.CreateTrunc(
Y,
X->getType(),
Y->getName() +
".tr");
2607 Value *NewRHS =
Builder.CreateAnd(
Y, Op1,
Y->getName() +
".masked");
2613 Value *NewLHS =
Builder.CreateAnd(
X, Op1,
X->getName() +
".masked");
2622 if (
C->isPowerOf2() &&
2625 int Log2C =
C->exactLogBase2();
2628 int BitNum = IsShiftLeft ? Log2C - Log2ShiftC : Log2ShiftC - Log2C;
2629 assert(BitNum >= 0 &&
"Expected demanded bits to handle impossible mask");
2630 Value *Cmp =
Builder.CreateICmpEQ(
X, ConstantInt::get(Ty, BitNum));
2631 return createSelectInstWithUnknownProfile(Cmp, ConstantInt::get(Ty, *
C),
2651 return createSelectInstWithUnknownProfile(
2662 if (Cmp && Cmp->isNullValue()) {
2668 return createSelectInstWithUnknownProfile(
2686 !
Builder.GetInsertBlock()->getParent()->hasFnAttribute(
2687 Attribute::NoImplicitFloat)) {
2702 APInt(Ty->getScalarSizeInBits(),
2703 Ty->getScalarSizeInBits() -
2704 X->getType()->getScalarSizeInBits())))) {
2705 auto *SExt =
Builder.CreateSExt(
X, Ty,
X->getName() +
".signext");
2706 return BinaryOperator::CreateAnd(SExt, Op1);
2712 if (
I.getType()->isIntOrIntVectorTy(1)) {
2715 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0,
true))
2720 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1,
true))
2735 return BinaryOperator::CreateAnd(Op0,
B);
2738 return BinaryOperator::CreateAnd(Op1,
B);
2746 if (NotC !=
nullptr)
2747 return BinaryOperator::CreateAnd(Op0, NotC);
2756 if (NotC !=
nullptr)
2757 return BinaryOperator::CreateAnd(Op1, NotC);
2766 return BinaryOperator::CreateAnd(
A,
B);
2774 return BinaryOperator::CreateAnd(
A,
B);
2782 return BinaryOperator::CreateAnd(
Builder.CreateNot(
A),
B);
2790 return BinaryOperator::CreateAnd(
Builder.CreateNot(
A),
B);
2794 foldBooleanAndOr(Op0, Op1,
I,
true,
false))
2799 if (
auto *V = reassociateBooleanAndOr(Op0,
X,
Y,
I,
true,
2805 if (
auto *V = reassociateBooleanAndOr(Op1,
X,
Y,
I,
true,
2813 if (
Instruction *CastedAnd = foldCastedBitwiseLogic(
I))
2826 A->getType()->isIntOrIntVectorTy(1))
2832 A->getType()->isIntOrIntVectorTy(1))
2837 A->getType()->isIntOrIntVectorTy(1))
2838 return createSelectInstWithUnknownProfile(
2839 A,
Builder.CreateAnd(
B, ConstantInt::get(Ty, 1)),
2845 if (
A->getType()->isIntOrIntVectorTy(1))
2849 return createSelectInstWithUnknownProfile(
2859 *
C ==
X->getType()->getScalarSizeInBits() - 1) {
2861 return createSelectInstWithUnknownProfile(IsNeg,
Y,
2869 *
C ==
X->getType()->getScalarSizeInBits() - 1) {
2871 return createSelectInstWithUnknownProfile(IsNeg,
2881 Value *Start =
nullptr, *Step =
nullptr;
2889 return Canonicalized;
2891 if (
Instruction *Folded = foldLogicOfIsFPClass(
I, Op0, Op1))
2903 return BinaryOperator::CreateAnd(V, Op1);
2907 return BinaryOperator::CreateAnd(Op0, V);
2914 bool MatchBitReversals) {
2922 for (
auto *Inst : Insts) {
2923 Inst->setDebugLoc(
I.getDebugLoc());
2929std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
2933 assert(
Or.getOpcode() == BinaryOperator::Or &&
"Expecting or instruction");
2935 unsigned Width =
Or.getType()->getScalarSizeInBits();
2940 return std::nullopt;
2948 Value *ShVal0, *ShVal1, *ShAmt0, *ShAmt1;
2954 return std::nullopt;
2957 if (Or0->
getOpcode() == BinaryOperator::LShr) {
2963 Or1->
getOpcode() == BinaryOperator::LShr &&
2964 "Illegal or(shift,shift) pair");
2968 auto matchShiftAmount = [&](
Value *L,
Value *R,
unsigned Width) ->
Value * {
2970 const APInt *LI, *RI;
2972 if (LI->
ult(Width) && RI->
ult(Width) && (*LI + *RI) == Width)
2973 return ConstantInt::get(L->getType(), *LI);
2997 if (ShVal0 != ShVal1)
3008 unsigned Mask = Width - 1;
3032 Value *ShAmt = matchShiftAmount(ShAmt0, ShAmt1, Width);
3034 ShAmt = matchShiftAmount(ShAmt1, ShAmt0, Width);
3038 return std::nullopt;
3040 FShiftArgs = {ShVal0, ShVal1, ShAmt};
3057 const APInt *ZextHighShlAmt;
3060 return std::nullopt;
3064 return std::nullopt;
3066 unsigned HighSize =
High->getType()->getScalarSizeInBits();
3067 unsigned LowSize =
Low->getType()->getScalarSizeInBits();
3070 if (ZextHighShlAmt->
ult(LowSize) || ZextHighShlAmt->
ugt(Width - HighSize))
3071 return std::nullopt;
3081 const APInt *ZextLowShlAmt;
3088 if (*ZextLowShlAmt + *ZextHighShlAmt != Width)
3094 ZextLowShlAmt->
ule(Width - LowSize) &&
"Invalid concat");
3103 FShiftArgs = {U, U, ConstantInt::get(Or0->
getType(), *ZextHighShlAmt)};
3108 if (FShiftArgs.
empty())
3109 return std::nullopt;
3111 Intrinsic::ID IID = IsFshl ? Intrinsic::fshl : Intrinsic::fshr;
3112 return std::make_pair(IID, FShiftArgs);
3118 auto [IID, FShiftArgs] = *Opt;
3129 assert(
Or.getOpcode() == Instruction::Or &&
"bswap requires an 'or'");
3130 Value *Op0 =
Or.getOperand(0), *Op1 =
Or.getOperand(1);
3133 unsigned Width = Ty->getScalarSizeInBits();
3134 if ((Width & 1) != 0)
3136 unsigned HalfWidth = Width / 2;
3143 Value *LowerSrc, *ShlVal, *UpperSrc;
3154 Value *NewLower = Builder.CreateZExt(
Lo, Ty);
3155 Value *NewUpper = Builder.CreateZExt(
Hi, Ty);
3156 NewUpper = Builder.CreateShl(NewUpper, HalfWidth);
3157 Value *BinOp = Builder.CreateDisjointOr(NewLower, NewUpper);
3158 return Builder.CreateIntrinsic(
id, Ty, BinOp);
3163 Value *LowerBSwap, *UpperBSwap;
3166 return ConcatIntrinsicCalls(Intrinsic::bswap, UpperBSwap, LowerBSwap);
3170 Value *LowerBRev, *UpperBRev;
3173 return ConcatIntrinsicCalls(Intrinsic::bitreverse, UpperBRev, LowerBRev);
3185 return Builder.CreateSExt(
X, Ty);
3193 for (
unsigned i = 0; i != NumElts; ++i) {
3196 if (!EltC1 || !EltC2)
3215 Type *Ty =
A->getType();
3231 if (
A->getType()->isIntOrIntVectorTy()) {
3233 if (NumSignBits ==
A->getType()->getScalarSizeInBits() &&
3256 Cond->getType()->isIntOrIntVectorTy(1)) {
3282 Cond->getType()->isIntOrIntVectorTy(1) &&
3296 Value *
D,
bool InvertFalseVal) {
3302 if (
Value *
Cond = getSelectCondition(
A,
C, InvertFalseVal)) {
3307 Type *SelTy =
A->getType();
3310 unsigned Elts = VecTy->getElementCount().getKnownMinValue();
3314 Type *EltTy =
Builder.getIntNTy(SelEltSize / Elts);
3331 bool IsAnd,
bool IsLogical,
3338 IsAnd ?
LHS->getInversePredicate() :
LHS->getPredicate();
3340 IsAnd ?
RHS->getInversePredicate() :
RHS->getPredicate();
3346 !(
LHS->hasOneUse() ||
RHS->hasOneUse()))
3349 auto MatchRHSOp = [LHS0, CInt](
const Value *RHSOp) {
3352 (CInt->
isZero() && RHSOp == LHS0);
3366 return Builder.CreateICmp(
3368 Builder.CreateSub(LHS0, ConstantInt::get(LHS0->
getType(), *CInt + 1)),
3378 const SimplifyQuery Q =
SQ.getWithInstruction(&
I);
3381 Value *LHS0 =
LHS->getOperand(0), *RHS0 =
RHS->getOperand(0);
3382 Value *LHS1 =
LHS->getOperand(1), *RHS1 =
RHS->getOperand(1);
3384 const APInt *LHSC =
nullptr, *RHSC =
nullptr;
3391 if (LHS0 == RHS1 && LHS1 == RHS0) {
3395 if (LHS0 == RHS0 && LHS1 == RHS1) {
3398 bool IsSigned =
LHS->isSigned() ||
RHS->isSigned();
3421 RHS->setSameSign(
false);
3447 if (IsAnd && !IsLogical)
3473 return Builder.CreateICmp(PredL, NewOr,
3484 return Builder.CreateICmp(PredL, NewAnd,
3504 const APInt *AndC, *SmallC =
nullptr, *BigC =
nullptr;
3518 if (SmallC && BigC) {
3519 unsigned BigBitSize = BigC->getBitWidth();
3526 APInt
N = SmallC->
zext(BigBitSize) | *BigC;
3528 return Builder.CreateICmp(PredL, NewAnd, NewVal);
3538 bool TrueIfSignedL, TrueIfSignedR;
3544 if ((TrueIfSignedL && !TrueIfSignedR &&
3547 (!TrueIfSignedL && TrueIfSignedR &&
3551 return Builder.CreateIsNeg(NewXor);
3554 if ((TrueIfSignedL && !TrueIfSignedR &&
3557 (!TrueIfSignedL && TrueIfSignedR &&
3561 return Builder.CreateIsNotNeg(NewXor);
3570 if (LHS0 == RHS0 && PredL == PredR &&
3572 !
I.getFunction()->hasFnAttribute(Attribute::NoImplicitFloat) &&
3575 X->getType()->getScalarType()->isIEEELikeFPTy() &&
3576 APFloat(
X->getType()->getScalarType()->getFltSemantics(), *MaskC)
3578 ((LHSC->
isZero() && *RHSC == *MaskC) ||
3579 (RHSC->
isZero() && *LHSC == *MaskC)))
3583 return foldAndOrOfICmpsUsingRanges(
LHS,
RHS, IsAnd);
3598 SQ.getWithInstruction(&
I)))
3603 if (
Value *Res = foldAndOrOfICmps(LHSCmp, RHSCmp,
I, IsAnd, IsLogical))
3608 if (
Value *Res = foldLogicOfFCmps(LHSCmp, RHSCmp, IsAnd, IsLogical))
3619 assert(
I.getOpcode() == Instruction::Or &&
3620 "Simplification only supports or at the moment.");
3622 Value *Cmp1, *Cmp2, *Cmp3, *Cmp4;
3629 return Builder.CreateXor(Cmp1, Cmp4);
3631 return Builder.CreateXor(Cmp1, Cmp3);
3661 const unsigned EltBitWidth = EltTy->getBitWidth();
3663 if (TargetBitWidth % EltBitWidth != 0 || ShlAmt % EltBitWidth != 0)
3665 const unsigned TargetEltWidth = TargetBitWidth / EltBitWidth;
3666 const unsigned ShlEltAmt = ShlAmt / EltBitWidth;
3668 const unsigned MaskIdx =
3669 DL.isLittleEndian() ? ShlEltAmt : TargetEltWidth - ShlEltAmt - 1;
3671 VecOffset =
static_cast<int64_t
>(VecIdx) -
static_cast<int64_t
>(MaskIdx);
3672 Mask.resize(TargetEltWidth);
3686 Mask.resize(SrcTy->getNumElements());
3700 const unsigned NumVecElts = VecTy->getNumElements();
3701 bool FoundVecOffset =
false;
3702 for (
unsigned Idx = 0; Idx < ShuffleMask.size(); ++Idx) {
3705 const unsigned ShuffleIdx = ShuffleMask[Idx];
3706 if (ShuffleIdx >= NumVecElts) {
3707 const unsigned ConstIdx = ShuffleIdx - NumVecElts;
3710 if (!ConstElt || !ConstElt->isNullValue())
3715 if (FoundVecOffset) {
3716 if (VecOffset + Idx != ShuffleIdx)
3719 if (ShuffleIdx < Idx)
3721 VecOffset = ShuffleIdx - Idx;
3722 FoundVecOffset =
true;
3726 return FoundVecOffset;
3739 bool AlreadyInsertedMaskedElt = Mask.test(InsertIdx);
3741 if (!AlreadyInsertedMaskedElt)
3742 Mask.reset(InsertIdx);
3751 assert(
I.getOpcode() == Instruction::Or);
3752 Value *LhsVec, *RhsVec;
3753 int64_t LhsVecOffset, RhsVecOffset;
3761 if (LhsVec != RhsVec || LhsVecOffset != RhsVecOffset)
3765 const unsigned ZeroVecIdx =
3768 for (
unsigned Idx : Mask.set_bits()) {
3769 assert(LhsVecOffset + Idx >= 0);
3770 ShuffleMask[Idx] = LhsVecOffset + Idx;
3773 Value *MaskedVec = Builder.CreateShuffleVector(
3775 I.getName() +
".v");
3801 const APInt *ShiftedMaskConst =
nullptr;
3808 if (!
match(MaskedOp0,
3813 if (LShrAmt > ShlAmt)
3815 Offset = ShlAmt - LShrAmt;
3817 Mask = ShiftedMaskConst ? ShiftedMaskConst->
shl(LShrAmt)
3819 Int->getType()->getScalarSizeInBits(), LShrAmt);
3829 Value *LhsInt, *RhsInt;
3830 APInt LhsMask, RhsMask;
3832 bool IsLhsShlNUW, IsLhsShlNSW, IsRhsShlNUW, IsRhsShlNSW;
3839 if (LhsInt != RhsInt || LhsOffset != RhsOffset)
3842 APInt Mask = LhsMask | RhsMask;
3845 Value *Res = Builder.CreateShl(
3847 Builder.CreateAnd(LhsInt, Mask, LhsInt->
getName() +
".mask"), DestTy,
3849 ConstantInt::get(DestTy, LhsOffset),
"", IsLhsShlNUW && IsRhsShlNUW,
3850 IsLhsShlNSW && IsRhsShlNSW);
3875 return std::nullopt;
3878 Value *Original =
nullptr;
3879 const APInt *Mask =
nullptr;
3880 const APInt *MulConst =
nullptr;
3883 if (MulConst->
isZero() || Mask->isZero())
3884 return std::nullopt;
3886 return std::optional<DecomposedBitMaskMul>(
3887 {Original, *MulConst, *Mask,
3893 const APInt *EqZero =
nullptr, *NeZero =
nullptr;
3897 auto ICmpDecompose =
3900 if (!ICmpDecompose.has_value())
3901 return std::nullopt;
3904 ICmpDecompose->C.isZero());
3909 if (!EqZero->
isZero() || NeZero->isZero())
3910 return std::nullopt;
3912 if (!ICmpDecompose->Mask.isPowerOf2() || ICmpDecompose->Mask.isZero() ||
3913 NeZero->getBitWidth() != ICmpDecompose->Mask.getBitWidth())
3914 return std::nullopt;
3916 if (!NeZero->urem(ICmpDecompose->Mask).isZero())
3917 return std::nullopt;
3919 return std::optional<DecomposedBitMaskMul>(
3920 {ICmpDecompose->X, NeZero->udiv(ICmpDecompose->Mask),
3921 ICmpDecompose->Mask,
false,
false});
3924 return std::nullopt;
3940 if (Decomp0->isCombineableWith(*Decomp1)) {
3941 Value *NewAnd = Builder.CreateAnd(
3943 ConstantInt::get(Decomp0->X->getType(), Decomp0->Mask + Decomp1->Mask));
3945 return Builder.CreateMul(
3946 NewAnd, ConstantInt::get(NewAnd->
getType(), Decomp1->Factor),
"",
3947 Decomp0->NUW && Decomp1->NUW, Decomp0->NSW && Decomp1->NSW);
3966 if (
Value *Res = foldDisjointOr(
LHS,
X))
3967 return Builder.CreateDisjointOr(Res,
Y);
3968 if (
Value *Res = foldDisjointOr(
LHS,
Y))
3969 return Builder.CreateDisjointOr(Res,
X);
3973 if (
Value *Res = foldDisjointOr(
X,
RHS))
3974 return Builder.CreateDisjointOr(Res,
Y);
3975 if (
Value *Res = foldDisjointOr(
Y,
RHS))
3976 return Builder.CreateDisjointOr(Res,
X);
3990 const APInt *C1, *C2;
3999 Constant *NewC = ConstantInt::get(
X->getType(), C2->
udiv(*C1));
4020 return Builder.CreateBinaryIntrinsic(Intrinsic::abs,
X,
4021 Builder.getFalse());
4039 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4040 match(SelOp1->getTrueValue(),
4045 Value *C2 =
nullptr, *A2 =
nullptr, *B2 =
nullptr;
4054 return createSelectInstWithUnknownProfile(
C,
A,
B);
4070 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4071 match(SelOp0->getTrueValue(),
4076 Value *C2 =
nullptr, *A2 =
nullptr, *B2 =
nullptr;
4085 return createSelectInstWithUnknownProfile(
C,
B,
A);
4099 SQ.getWithInstruction(&
I)))
4135 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
4136 Type *Ty =
I.getType();
4137 if (Ty->isIntOrIntVectorTy(1)) {
4140 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0,
false))
4145 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1,
false))
4182 if (
Value *Res = foldDisjointOr(
I.getOperand(0),
I.getOperand(1)))
4185 if (
Value *Res = reassociateDisjointOr(
I.getOperand(0),
I.getOperand(1)))
4196 return BinaryOperator::CreateXor(
Or, ConstantInt::get(Ty, *CV));
4203 Value *IncrementY =
Builder.CreateAdd(
Y, ConstantInt::get(Ty, 1));
4204 return BinaryOperator::CreateMul(
X, IncrementY);
4221 if (
I.getType()->isIntOrIntVectorTy(1) &&
4234 const APInt *C0, *C1;
4240 return BinaryOperator::CreateOr(
Builder.CreateAnd(
X, *C0),
B);
4243 return BinaryOperator::CreateOr(
Builder.CreateAnd(
X, *C1),
A);
4247 return BinaryOperator::CreateXor(
Builder.CreateAnd(
X, *C0),
B);
4250 return BinaryOperator::CreateXor(
Builder.CreateAnd(
X, *C1),
A);
4253 if ((*C0 & *C1).
isZero()) {
4258 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4259 return BinaryOperator::CreateAnd(
A, C01);
4265 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4266 return BinaryOperator::CreateAnd(
B, C01);
4270 const APInt *C2, *C3;
4275 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4276 return BinaryOperator::CreateAnd(
Or, C01);
4286 if (
Value *V = matchSelectFromAndOr(
A,
C,
B,
D))
4288 if (
Value *V = matchSelectFromAndOr(
A,
C,
D,
B))
4290 if (
Value *V = matchSelectFromAndOr(
C,
A,
B,
D))
4292 if (
Value *V = matchSelectFromAndOr(
C,
A,
D,
B))
4294 if (
Value *V = matchSelectFromAndOr(
B,
D,
A,
C))
4296 if (
Value *V = matchSelectFromAndOr(
B,
D,
C,
A))
4298 if (
Value *V = matchSelectFromAndOr(
D,
B,
A,
C))
4300 if (
Value *V = matchSelectFromAndOr(
D,
B,
C,
A))
4309 if (
Value *V = matchSelectFromAndOr(
A,
C,
B,
D,
true))
4311 if (
Value *V = matchSelectFromAndOr(
A,
C,
D,
B,
true))
4313 if (
Value *V = matchSelectFromAndOr(
C,
A,
B,
D,
true))
4315 if (
Value *V = matchSelectFromAndOr(
C,
A,
D,
B,
true))
4324 return BinaryOperator::CreateOr(Op0,
C);
4331 return BinaryOperator::CreateOr(Op1,
C);
4337 bool SwappedForXor =
false;
4340 SwappedForXor =
true;
4347 return BinaryOperator::CreateOr(Op0,
B);
4349 return BinaryOperator::CreateOr(Op0,
A);
4354 return BinaryOperator::CreateOr(
A,
B);
4382 return BinaryOperator::CreateOr(Nand,
C);
4390 foldBooleanAndOr(Op0, Op1,
I,
false,
false))
4395 if (
auto *V = reassociateBooleanAndOr(Op0,
X,
Y,
I,
false,
4401 if (
auto *V = reassociateBooleanAndOr(Op1,
X,
Y,
I,
false,
4421 A->getType()->isIntOrIntVectorTy(1))
4422 return createSelectInstWithUnknownProfile(
4441 Value *Inner =
Builder.CreateOr(
A, Op1,
"", IsDisjointOuter);
4443 return IsDisjointOuter && IsDisjointInner
4444 ? BinaryOperator::CreateDisjointOr(Inner, CI)
4445 : BinaryOperator::CreateOr(Inner, CI);
4452 Value *
X =
nullptr, *
Y =
nullptr;
4471 return createSelectInstWithUnknownProfile(NewICmpInst,
AllOnes,
X);
4484 return BinaryOperator::CreateXor(
A,
B);
4500 Value *
Mul, *Ov, *MulIsNotZero, *UMulWithOv;
4518 return BinaryOperator::CreateAnd(NotNullA, NotNullB);
4527 const APInt *C1, *C2;
4542 : C2->
uadd_ov(*C1, Overflow));
4546 return BinaryOperator::CreateOr(Ov, NewCmp);
4565 ConstantInt::get(Ty, Ty->getScalarSizeInBits() - 1),
X);
4571 Value *Start =
nullptr, *Step =
nullptr;
4589 return BinaryOperator::CreateOr(
4601 return BinaryOperator::CreateOr(
4609 return Canonicalized;
4611 if (
Instruction *Folded = foldLogicOfIsFPClass(
I, Op0, Op1))
4631 !
Builder.GetInsertBlock()->getParent()->hasFnAttribute(
4632 Attribute::NoImplicitFloat)) {
4646 if ((KnownX.
One & *C2) == *C2)
4647 return BinaryOperator::CreateAnd(
X, ConstantInt::get(Ty, *C1 | *C2));
4656 return BinaryOperator::CreateOr(V, Op1);
4660 return BinaryOperator::CreateOr(Op0, V);
4676 assert(
I.getOpcode() == Instruction::Xor);
4677 Value *Op0 =
I.getOperand(0);
4678 Value *Op1 =
I.getOperand(1);
4689 return BinaryOperator::CreateXor(
A,
B);
4697 return BinaryOperator::CreateXor(
A,
B);
4705 return BinaryOperator::CreateXor(
A,
B);
4727 assert(
I.getOpcode() == Instruction::Xor &&
I.getOperand(0) ==
LHS &&
4728 I.getOperand(1) ==
RHS &&
"Should be 'xor' with these operands");
4731 Value *LHS0 =
LHS->getOperand(0), *LHS1 =
LHS->getOperand(1);
4732 Value *RHS0 =
RHS->getOperand(0), *RHS1 =
RHS->getOperand(1);
4735 if (LHS0 == RHS1 && LHS1 == RHS0) {
4739 if (LHS0 == RHS0 && LHS1 == RHS1) {
4742 bool IsSigned =
LHS->isSigned() ||
RHS->isSigned();
4747 const APInt *LC, *RC;
4756 bool TrueIfSignedL, TrueIfSignedR;
4761 return TrueIfSignedL == TrueIfSignedR ?
Builder.CreateIsNeg(XorLR) :
4762 Builder.CreateIsNotNeg(XorLR);
4772 if (CRUnion && CRIntersect)
4773 if (
auto CR = CRUnion->exactIntersectWith(CRIntersect->inverse())) {
4774 if (CR->isFullSet())
4776 if (CR->isEmptySet())
4781 CR->getEquivalentICmp(NewPred, NewC,
Offset);
4788 NewV =
Builder.CreateAdd(NewV, ConstantInt::get(Ty,
Offset));
4789 return Builder.CreateICmp(NewPred, NewV,
4790 ConstantInt::get(Ty, NewC));
4822 ICmpInst *
X =
nullptr, *
Y =
nullptr;
4823 if (OrICmp ==
LHS && AndICmp ==
RHS) {
4828 if (OrICmp ==
RHS && AndICmp ==
LHS) {
4835 Y->setPredicate(
Y->getInversePredicate());
4837 if (!
Y->hasOneUse()) {
4844 Builder.SetInsertPoint(
Y->getParent(), ++(
Y->getIterator()));
4848 Y->replaceUsesWithIf(NotY,
4849 [NotY](Use &U) {
return U.getUser() != NotY; });
4887 Value *NewA = Builder.CreateAnd(
D, NotM);
4888 return BinaryOperator::CreateXor(NewA,
X);
4894 Type *EltTy =
C->getType()->getScalarType();
4898 Value *NotC = Builder.CreateNot(
C);
4899 Value *
RHS = Builder.CreateAnd(
B, NotC);
4900 return BinaryOperator::CreateOr(
LHS,
RHS);
4915 return A ==
C ||
A ==
D ||
B ==
C ||
B ==
D;
4923 Value *NotY = Builder.CreateNot(
Y);
4924 return BinaryOperator::CreateOr(
X, NotY);
4931 Value *NotX = Builder.CreateNot(
X);
4932 return BinaryOperator::CreateOr(
Y, NotX);
4942 assert(
Xor.getOpcode() == Instruction::Xor &&
"Expected an xor instruction.");
4948 Value *Op0 =
Xor.getOperand(0), *Op1 =
Xor.getOperand(1);
4956 Op1->
hasNUses(2) && *ShAmt == Ty->getScalarSizeInBits() - 1 &&
4961 Value *IsNeg = Builder.CreateIsNeg(
A);
4964 Value *NegA =
Add->hasNoUnsignedWrap()
4966 : Builder.CreateNeg(
A,
"",
Add->hasNoSignedWrap());
4984 Op->replaceUsesWithIf(NotOp,
4985 [NotOp](
Use &U) {
return U.getUser() != NotOp; });
5026 Builder.SetInsertPoint(*
I.getInsertionPointAfterDef());
5029 NewLogicOp =
Builder.CreateBinOp(NewOpc, Op0, Op1,
I.getName() +
".not");
5032 Builder.CreateLogicalOp(NewOpc, Op0, Op1,
I.getName() +
".not",
5035 SI->swapProfMetadata();
5059 Value *NotOp0 =
nullptr;
5060 Value *NotOp1 =
nullptr;
5061 Value **OpToInvert =
nullptr;
5078 Builder.SetInsertPoint(*
I.getInsertionPointAfterDef());
5081 NewBinOp =
Builder.CreateBinOp(NewOpc, Op0, Op1,
I.getName() +
".not");
5083 NewBinOp =
Builder.CreateLogicalOp(NewOpc, Op0, Op1,
I.getName() +
".not");
5106 Type *Ty =
I.getType();
5109 Value *NotY = Builder.CreateNot(
Y,
Y->getName() +
".not");
5110 return BinaryOperator::CreateOr(
X, NotY);
5113 Value *NotY = Builder.CreateNot(
Y,
Y->getName() +
".not");
5117 SI->swapProfMetadata();
5125 return BinaryOperator::CreateAnd(
X, NotY);
5132 SI->swapProfMetadata();
5137 BinaryOperator *NotVal;
5144 return BinaryOperator::CreateAnd(DecX, NotY);
5149 return BinaryOperator::CreateAShr(
X,
Y);
5155 return BinaryOperator::CreateAShr(
X,
Y);
5162 return new SExtInst(IsNotNeg, Ty);
5189 return BinaryOperator::CreateAdd(
Builder.CreateNot(
X),
Y);
5212 return new BitCastInst(
X, Ty);
5218 X->getType()->isIntOrIntVectorTy(1)) {
5222 return new BitCastInst(Sext, Ty);
5233 if (
II &&
II->hasOneUse()) {
5237 Value *InvMaxMin =
Builder.CreateBinaryIntrinsic(InvID,
X, NotY);
5241 if (
II->getIntrinsicID() == Intrinsic::is_fpclass) {
5244 1, ConstantInt::get(ClassMask->
getType(),
5260 Value *TV = Sel->getTrueValue();
5261 Value *FV = Sel->getFalseValue();
5264 bool InvertibleT = (CmpT && CmpT->hasOneUse()) ||
isa<Constant>(TV);
5265 bool InvertibleF = (CmpF && CmpF->hasOneUse()) ||
isa<Constant>(FV);
5266 if (InvertibleT && InvertibleF) {
5268 CmpT->setPredicate(CmpT->getInversePredicate());
5272 CmpF->setPredicate(CmpF->getInversePredicate());
5303 Value *NotC = Builder.CreateNot(AddC);
5306 return BinaryOperator::CreateAnd(NewSub, Mask);
5317 SQ.getWithInstruction(&
I)))
5347 Value *Op0 =
I.getOperand(0), *Op1 =
I.getOperand(1);
5355 return BinaryOperator::CreateXor(XorAC,
Y);
5358 return BinaryOperator::CreateXor(XorBC,
X);
5368 return BinaryOperator::CreateDisjointOr(Op0, Op1);
5370 return BinaryOperator::CreateOr(Op0, Op1);
5387 return BinaryOperator::CreateXor(
5410 *CA ==
X->getType()->getScalarSizeInBits() - 1 &&
5414 return createSelectInstWithUnknownProfile(IsNotNeg, Op1,
5419 Type *Ty =
I.getType();
5427 return BinaryOperator::CreateSub(ConstantInt::get(Ty, *
C + *RHSC),
X);
5431 return BinaryOperator::CreateAdd(
X, ConstantInt::get(Ty, *
C + *RHSC));
5436 return BinaryOperator::CreateXor(
X, ConstantInt::get(Ty, *
C ^ *RHSC));
5442 if (
II &&
II->hasOneUse() && *RHSC == Ty->getScalarSizeInBits() - 1) {
5444 if ((IID == Intrinsic::ctlz || IID == Intrinsic::cttz) &&
5447 IID = (IID == Intrinsic::ctlz) ? Intrinsic::cttz : Intrinsic::ctlz;
5460 return BinaryOperator::CreateShl(NotX, ConstantInt::get(Ty, *
C));
5466 return BinaryOperator::CreateLShr(NotX, ConstantInt::get(Ty, *
C));
5484 !
Builder.GetInsertBlock()->getParent()->hasFnAttribute(
5485 Attribute::NoImplicitFloat)) {
5508 auto *Opnd0 =
Builder.CreateLShr(
X, C2);
5509 Opnd0->takeName(Op0);
5510 return BinaryOperator::CreateXor(Opnd0, ConstantInt::get(Ty, FoldConst));
5523 return BinaryOperator::CreateAnd(
X,
Builder.CreateNot(Op0));
5527 return BinaryOperator::CreateAnd(
X,
Builder.CreateNot(Op1));
5532 return BinaryOperator::CreateAnd(Op0,
Builder.CreateNot(
X));
5540 return BinaryOperator::CreateAnd(Op1,
Builder.CreateNot(
X));
5546 return BinaryOperator::CreateXor(
5552 return BinaryOperator::CreateXor(
5558 return BinaryOperator::CreateOr(
A,
B);
5562 return BinaryOperator::CreateOr(
A,
B);
5572 return BinaryOperator::CreateOr(
A,
B);
5587 if (
B ==
C ||
B ==
D)
5593 return BinaryOperator::CreateAnd(
Builder.CreateXor(
B,
C), NotA);
5598 if (
I.getType()->isIntOrIntVectorTy(1) &&
5603 if (
B ==
C ||
B ==
D) {
5614 ? createSelectInstWithUnknownProfile(
A, NotB,
C)
5621 if (
Value *V = foldXorOfICmps(LHS, RHS,
I))
5624 if (
Instruction *CastedXor = foldCastedBitwiseLogic(
I))
5637 return BinaryOperator::CreateXor(
Builder.CreateXor(
X,
Y), C1);
5643 return Canonicalized;
5645 if (
Instruction *Folded = foldLogicOfIsFPClass(
I, Op0, Op1))
5648 if (
Instruction *Folded = canonicalizeConditionalNegationViaMathToSelect(
I))
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static bool isSigned(unsigned Opcode)
static Value * foldAndOrOfICmpsWithConstEq(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q, Instruction &I)
Reduce logic-of-compares with equality to a constant by substituting a common operand with the consta...
static Value * foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and fold (icmp ne ctpop(X) 1) & ...
static Value * foldBitmaskMul(Value *Op0, Value *Op1, InstCombiner::BuilderTy &Builder)
(A & N) * C + (A & M) * C -> (A & (N + M)) & C This also accepts the equivalent select form of (A & N...
static unsigned conjugateICmpMask(unsigned Mask)
Convert an analysis of a masked ICmp into its equivalent if all boolean operations had the opposite s...
static Instruction * foldNotXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * foldLogOpOfMaskedICmps(Value *LHS, Value *RHS, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * getFCmpValue(unsigned Code, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder, FMFSource FMF)
This is the complement of getFCmpCode, which turns an opcode and two operands into either a FCmp inst...
static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal, uint64_t &ClassMask)
Match an fcmp against a special value that performs a test possible by llvm.is.fpclass.
static Value * foldSignedTruncationCheck(ICmpInst *ICmp0, ICmpInst *ICmp1, Instruction &CxtI, InstCombiner::BuilderTy &Builder)
General pattern: X & Y.
static Instruction * visitMaskedMerge(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
If we have a masked merge, in the canonical form of: (assuming that A only has one use....
static Instruction * canonicalizeAbs(BinaryOperator &Xor, InstCombiner::BuilderTy &Builder)
Canonicalize a shifty way to code absolute value to the more common pattern that uses negation and se...
static Value * foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Reduce a pair of compares that check if a value has exactly 1 bit set.
static Value * foldUnsignedUnderflowCheck(ICmpInst *ZeroICmp, ICmpInst *UnsignedICmp, bool IsAnd, const SimplifyQuery &Q, InstCombiner::BuilderTy &Builder)
Commuted variants are assumed to be handled by calling this function again with the parameters swappe...
static Instruction * foldOrToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * simplifyAndOrWithOpReplaced(Value *V, Value *Op, Value *RepOp, bool SimplifyOnly, InstCombinerImpl &IC, unsigned Depth=0)
static Instruction * matchDeMorgansLaws(BinaryOperator &I, InstCombiner &IC)
Match variations of De Morgan's Laws: (~A & ~B) == (~(A | B)) (~A | ~B) == (~(A & B))
static Value * foldLogOpOfMaskedICmpsAsymmetric(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * FoldOrOfSelectSmaxToAbs(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X) select(X <s 0, -X, 0) | smax(X,...
static Instruction * foldAndToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static unsigned getMaskedICmpType(Value *A, Value *B, Value *C, ICmpInst::Predicate Pred)
Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C) satisfies.
static Instruction * foldXorToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
A ^ B can be specified using other logic ops in a variety of patterns.
static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth)
Return true if a constant shift amount is always less than the specified bit-width.
static Instruction * foldLogicCastConstant(BinaryOperator &Logic, CastInst *Cast, InstCombinerImpl &IC)
Fold {and,or,xor} (cast X), C.
static Value * foldAndOrOfICmpEqConstantAndICmp(ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, bool IsLogical, IRBuilderBase &Builder)
static bool canFreelyInvert(InstCombiner &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldNegativePower2AndShiftedMask(Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff B is a contiguous set of o...
static Value * matchIsFiniteTest(InstCombiner::BuilderTy &Builder, FCmpInst *LHS, FCmpInst *RHS)
and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
static Value * foldPowerOf2AndShiftedMask(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder)
Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) & (icmp(X & M) !...
static Value * foldOrUnsignedUMulOverflowICmp(BinaryOperator &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2)) --> (ugt x (c2/c1)).
static Value * freelyInvert(InstCombinerImpl &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldLogOpOfMaskedICmps_NotAllZeros_BMask_Mixed(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static std::optional< IntPart > matchIntPart(Value *V)
Match an extraction of bits from an integer.
static Instruction * canonicalizeLogicFirst(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * reassociateFCmps(BinaryOperator &BO, InstCombiner::BuilderTy &Builder)
This a limited reassociation for a special case (see above) where we are checking if two values are e...
static Value * getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder)
This is the complement of getICmpCode, which turns an opcode and two operands into either a constant ...
static Value * extractIntPart(const IntPart &P, IRBuilderBase &Builder)
Materialize an extraction of bits from an integer in IR.
static bool matchUnorderedInfCompare(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches fcmp u__ x, +/-inf.
static bool matchIsNotNaN(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches canonical form of isnan, fcmp ord x, 0.
static bool areInverseVectorBitmasks(Constant *C1, Constant *C2)
If all elements of two constant vectors are 0/-1 and inverses, return true.
MaskedICmpType
Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns that can be simplified.
static Instruction * foldComplexAndOrPatterns(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Try folding relatively complex patterns for both And and Or operations with all And and Or swapped.
static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask, uint64_t &Offset, bool &IsShlNUW, bool &IsShlNSW)
Match V as "lshr -> mask -> zext -> shl".
static std::optional< DecomposedBitMaskMul > matchBitmaskMul(Value *V)
static Value * foldOrOfInversions(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static bool matchSubIntegerPackFromVector(Value *V, Value *&Vec, int64_t &VecOffset, SmallBitVector &Mask, const DataLayout &DL)
Match V as "shufflevector -> bitcast" or "extractelement -> zext -> shl" patterns,...
static Instruction * matchFunnelShift(Instruction &Or, InstCombinerImpl &IC)
Match UB-safe variants of the funnel shift intrinsic.
static Instruction * reassociateForUses(BinaryOperator &BO, InstCombinerImpl::BuilderTy &Builder)
Try to reassociate a pair of binops so that values with one use only are part of the same instruction...
static Value * matchOrConcat(Instruction &Or, InstCombiner::BuilderTy &Builder)
Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
static Value * foldAndOrOfICmpsWithPow2AndWithZero(InstCombiner::BuilderTy &Builder, ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, const SimplifyQuery &Q)
static Instruction * foldMaskedAddXorPattern(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * foldBitwiseLogicWithIntrinsics(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static std::optional< std::pair< unsigned, unsigned > > getMaskedTypeForICmpPair(Value *&A, Value *&B, Value *&C, Value *&D, Value *&E, Value *LHS, Value *RHS, ICmpInst::Predicate &PredL, ICmpInst::Predicate &PredR)
Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
static Instruction * foldIntegerPackFromVector(Instruction &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Try to fold the join of two scalar integers whose contents are packed elements of the same vector.
static Value * foldIntegerRepackThroughZExt(Value *Lhs, Value *Rhs, InstCombiner::BuilderTy &Builder)
Try to fold the join of two scalar integers whose bits are unpacked and zexted from the same source i...
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
uint64_t IntrinsicInst * II
This file contains the declarations for profiling metadata utility functions.
const SmallVectorImpl< MachineOperand > & Cond
This file implements the SmallBitVector class.
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static constexpr int Concat[]
static LLVM_ABI bool hasSignBitInMSB(const fltSemantics &)
bool bitwiseIsEqual(const APFloat &RHS) const
APInt bitcastToAPInt() const
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Class for arbitrary precision integers.
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
uint64_t getZExtValue() const
Get zero extended value.
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
unsigned countLeadingOnes() const
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
unsigned getBitWidth() const
Return the number of bits in the APInt.
bool ult(const APInt &RHS) const
Unsigned less than comparison.
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
bool intersects(const APInt &RHS) const
This operation tests if there are any pairs of corresponding bits between this APInt and RHS that are...
int32_t exactLogBase2() const
LLVM_ABI APInt reverseBits() const
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
unsigned countr_zero() const
Count the number of trailing zero bits.
unsigned countLeadingZeros() const
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
APInt shl(unsigned shiftAmt) const
Left-shift function.
LLVM_ABI APInt byteSwap() const
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
void clearSignBit()
Set the sign bit to 0.
Represent a constant reference to an array (0 or more elements consecutively in memory),...
LLVM_ABI bool isSigned() const
Whether the intrinsic is signed or unsigned.
LLVM_ABI Instruction::BinaryOps getBinaryOp() const
Returns the binary operation underlying the intrinsic.
BinaryOps getOpcode() const
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
static BinaryOperator * CreateWithCopiedFlags(BinaryOps Opc, Value *V1, Value *V2, Value *CopyO, const Twine &Name="", InsertPosition InsertBefore=nullptr)
This class represents a no-op cast from one type to another.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Type * getSrcTy() const
Return the source type, as a convenience.
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
static LLVM_ABI CastInst * Create(Instruction::CastOps, Value *S, Type *Ty, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Provides a way to construct any of the CastInst subclasses using an opcode instead of the subclass's ...
Type * getDestTy() const
Return the destination type, as a convenience.
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
@ ICMP_SLT
signed less than
@ ICMP_SLE
signed less or equal
@ FCMP_OLT
0 1 0 0 True if ordered and less than
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
@ ICMP_UGE
unsigned greater or equal
@ ICMP_UGT
unsigned greater than
@ ICMP_SGT
signed greater than
@ FCMP_ULT
1 1 0 0 True if unordered or less than
@ ICMP_ULT
unsigned less than
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
@ ICMP_SGE
signed greater or equal
@ ICMP_ULE
unsigned less or equal
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Predicate getPredicate() const
Return the predicate for this instruction.
static LLVM_ABI bool isUnordered(Predicate predicate)
Determine if the predicate is an unordered operation.
static Predicate getOrderedPredicate(Predicate Pred)
Returns the ordered variant of a floating point compare.
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getAdd(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getTrunc(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getExactLogBase2(Constant *C)
If C is a scalar/fixed width vector of known powers of 2, then this function returns a new scalar/fix...
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
bool isMinusOne() const
This function will return true iff every bit in this constant is set to true.
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
const APInt & getValue() const
Return the constant as an APInt value reference.
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
This is an important base class in LLVM.
static LLVM_ABI Constant * replaceUndefsWith(Constant *C, Constant *Replacement)
Try to replace undefined constant C or undefined elements in C with Replacement.
static LLVM_ABI Constant * mergeUndefsWith(Constant *C, Constant *Other)
Merges undefs of a Constant with another Constant, along with the undefs already present.
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
A parsed version of the target data layout string in and methods for querying it.
This instruction compares its operands according to the predicate given to the constructor.
This provides a helper for copying FMF from an instruction or setting specified flags.
static FMFSource intersect(Value *A, Value *B)
Intersect the FMF from two instructions.
This instruction compares its operands according to the predicate given to the constructor.
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
Common base class shared among various IRBuilders.
Value * CreateNot(Value *V, const Twine &Name="")
Value * CreateBinOp(Instruction::BinaryOps Opc, Value *LHS, Value *RHS, const Twine &Name="", MDNode *FPMathTag=nullptr)
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitOr(BinaryOperator &I)
bool SimplifyAssociativeOrCommutative(BinaryOperator &I)
Performs a few simplifications for operators which are associative or commutative.
Value * foldUsingDistributiveLaws(BinaryOperator &I)
Tries to simplify binary operations which some other binary operation distributes over.
Instruction * foldBinOpShiftWithShift(BinaryOperator &I)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
Instruction * foldBinOpSelectBinOp(BinaryOperator &Op)
In some cases it is beneficial to fold a select into a binary operator.
bool sinkNotIntoLogicalOp(Instruction &I)
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitAnd(BinaryOperator &I)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW, bool NUW)
Common transforms for add / disjoint or.
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Instruction * FoldOrOfLogicalAnds(Value *Op0, Value *Op1)
Value * SimplifyAddWithRemainder(BinaryOperator &I)
Tries to simplify add operations using the definition of remainder.
Instruction * visitXor(BinaryOperator &I)
bool SimplifyDemandedInstructionBits(Instruction &Inst)
Tries to simplify operands to an integer instruction based on its demanded bits.
Instruction * foldVectorBinop(BinaryOperator &Inst)
Canonicalize the position of binops relative to shufflevector.
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
The core instruction combiner logic.
const DataLayout & getDataLayout() const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
unsigned ComputeNumSignBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI void removeFromParent()
This method unlinks 'this' from the containing basic block, but does not delete it.
LLVM_ABI bool hasNoUnsignedWrap() const LLVM_READONLY
Determine whether the no unsigned wrap flag is set.
LLVM_ABI bool hasNoSignedWrap() const LLVM_READONLY
Determine whether the no signed wrap flag is set.
LLVM_ABI void swapProfMetadata()
If the instruction has "branch_weights" MD_prof metadata and the MDNode has three operands (including...
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
A wrapper class for inspecting calls to intrinsic functions.
This class represents a sign extension of integer types.
This class represents the LLVM 'select' instruction.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, const Instruction *MDFrom=nullptr)
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
LLVM_ABI unsigned getIntegerBitWidth() const
bool isVectorTy() const
True if this is an instance of VectorType.
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
LLVM_ABI const fltSemantics & getFltSemantics() const
A Use represents the edge between a Value definition and its users.
Value * getOperand(unsigned i) const
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
bool hasOneUse() const
Return true if there is exactly one use of this value.
iterator_range< user_iterator > users()
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
LLVM_ABI bool hasNUses(unsigned N) const
Return true if this Value has exactly N uses.
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
Represents an op.with.overflow intrinsic.
This class represents zero extension of integer types.
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
const APInt & umin(const APInt &A, const APInt &B)
Determine the smaller of two APInts considered to be unsigned.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
match_combine_or< Ty... > m_CombineOr(const Ty &...Ps)
Combine pattern matchers matching any of Ps patterns.
match_combine_and< Ty... > m_CombineAnd(const Ty &...Ps)
Combine pattern matchers matching all of Ps patterns.
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
auto m_Cmp()
Matches any compare instruction and ignore it.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, FCmpInst > m_FCmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
cstfp_pred_ty< is_inf > m_Inf()
Match a positive or negative infinity FP constant.
m_Intrinsic_Ty< Opnd0 >::Ty m_BitReverse(const Opnd0 &Op0)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
auto m_LogicalOp()
Matches either L && R or L || R where L and R are arbitrary values.
ap_match< APInt > m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
ap_match< APInt > m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWSub(const LHS &L, const RHS &R)
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
match_bind< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
cst_pred_ty< is_shifted_mask > m_ShiftedMask()
match_deferred< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
DisjointOr_match< LHS, RHS > m_DisjointOr(const LHS &L, const RHS &R)
constantexpr_match m_ConstantExpr()
Match a constant expression or a constant that contains a constant expression.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
ap_match< APFloat > m_APFloatAllowPoison(const APFloat *&Res)
Match APFloat while allowing poison in splat vector constants.
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
auto m_BinOp()
Match an arbitrary binary operation and ignore it.
match_combine_or< CastInst_match< OpTy, SExtInst >, OpTy > m_SExtOrSelf(const OpTy &Op)
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
auto m_Value()
Match an arbitrary value and ignore it.
ShiftLike_match< LHS, Instruction::Shl > m_ShlOrSelf(const LHS &L, uint64_t &R)
Matches shl L, ConstShAmt or L itself (R will be set to zero in this case).
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, CmpInst > m_SpecificCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
auto m_Constant()
Match an arbitrary Constant and ignore it.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
match_bind< WithOverflowInst > m_WithOverflowInst(WithOverflowInst *&I)
Match a with overflow intrinsic, capturing it if we match.
SpecificCmpClass_match< LHS, RHS, ICmpInst > m_SpecificICmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
cst_pred_ty< is_negated_power2 > m_NegatedPower2()
Match a integer or vector negated power-of-2.
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
DisjointOr_match< LHS, RHS, true > m_c_DisjointOr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, FCmpInst > m_SpecificFCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
match_combine_or< CastInst_match< OpTy, SExtInst >, NNegZExt_match< OpTy > > m_SExtLike(const OpTy &Op)
Match either "sext" or "zext nneg".
auto m_c_MaxOrMin(const LHS &L, const RHS &R)
MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty > m_SMax(const LHS &L, const RHS &R)
cst_pred_ty< is_maxsignedvalue > m_MaxSignedValue()
Match an integer or vector with values having all bits except for the high bit set (0x7f....
m_Intrinsic_Ty< Opnd0 >::Ty m_Ctpop(const Opnd0 &Op0)
AnyBinaryOp_match< LHS, RHS, true > m_c_BinOp(const LHS &L, const RHS &R)
Matches a BinaryOperator with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
BinOpPred_match< LHS, RHS, is_shift_op > m_Shift(const LHS &L, const RHS &R)
Matches shift operations.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
LogicalOp_match< LHS, RHS, Instruction::And, true > m_c_LogicalAnd(const LHS &L, const RHS &R)
Matches L && R with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
m_Intrinsic_Ty< Opnd0 >::Ty m_BSwap(const Opnd0 &Op0)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ThreeOps_match< Val_t, Elt_t, Idx_t, Instruction::InsertElement > m_InsertElt(const Val_t &Val, const Elt_t &Elt, const Idx_t &Idx)
Matches InsertElementInst.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
auto m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
NodeAddr< CodeNode * > Code
friend class Instruction
Iterator for Instructions in a `BasicBlock.
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI Intrinsic::ID getInverseMinMaxIntrinsic(Intrinsic::ID MinMaxID)
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
FunctionAddr VTableAddr Value
Constant * getPredForFCmpCode(unsigned Code, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getFCmpCode.
cl::opt< bool > ProfcheckDisableMetadataFixes
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
LLVM_ABI void setExplicitlyUnknownBranchWeightsIfProfiled(Instruction &I, StringRef PassName, const Function *F=nullptr)
Like setExplicitlyUnknownBranchWeights(...), but only sets unknown branch weights in the new instruct...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
bool predicatesFoldable(CmpInst::Predicate P1, CmpInst::Predicate P2)
Return true if both predicates match sign or if at least one of them is an equality comparison (which...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
LLVM_ABI Value * simplifyOrInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Or, fold the result or return null.
LLVM_ABI Value * simplifyXorInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Xor, fold the result or return null.
LLVM_ABI bool isGuaranteedNotToBeUndef(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be undef, but may be poison.
LLVM_ABI bool matchSimpleRecurrence(const PHINode *P, BinaryOperator *&BO, Value *&Start, Value *&Step)
Attempt to match a simple first order recurrence cycle of the form: iv = phi Ty [Start,...
auto dyn_cast_or_null(const Y &Val)
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
LLVM_ABI Constant * getLosslessUnsignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI bool recognizeBSwapOrBitReverseIdiom(Instruction *I, bool MatchBSwaps, bool MatchBitReversals, SmallVectorImpl< Instruction * > &InsertedInsts)
Try to match a bswap or bitreverse idiom.
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * getLosslessSignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI Value * simplifyAndInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an And, fold the result or return null.
LLVM_ABI bool isKnownInversion(const Value *X, const Value *Y)
Return true iff:
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr int PoisonMaskElem
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
std::optional< DecomposedBitTest > decomposeBitTest(Value *Cond, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ And
Bitwise or logical AND of integers.
@ Sub
Subtraction of integers.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
APFloat neg(APFloat X)
Returns the negated value of the argument.
unsigned getICmpCode(CmpInst::Predicate Pred)
Encode a icmp predicate into a three bit mask.
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
LLVM_ABI bool isGuaranteedNotToBePoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be poison, but may be undef.
std::pair< Value *, FPClassTest > fcmpToClassTest(FCmpInst::Predicate Pred, const Function &F, Value *LHS, Value *RHS, bool LookThroughSrc=true)
Returns a pair of values, which if passed to llvm.is.fpclass, returns the same result as an fcmp with...
unsigned getFCmpCode(CmpInst::Predicate CC)
Similar to getICmpCode but for FCmpInst.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
Constant * getPredForICmpCode(unsigned Code, bool Sign, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getICmpCode.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
bool isCombineableWith(const DecomposedBitMaskMul Other)
bool isNonNegative() const
Returns true if this value is known to be non-negative.
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
SimplifyQuery getWithInstruction(const Instruction *I) const