LLVM 22.0.0git
InstCombineAndOrXor.cpp
Go to the documentation of this file.
1//===- InstCombineAndOrXor.cpp --------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the visitAnd, visitOr, and visitXor functions.
10//
11//===----------------------------------------------------------------------===//
12
13#include "InstCombineInternal.h"
21#include "llvm/IR/Intrinsics.h"
25
26using namespace llvm;
27using namespace PatternMatch;
28
29#define DEBUG_TYPE "instcombine"
30
31namespace llvm {
33}
34
35/// This is the complement of getICmpCode, which turns an opcode and two
36/// operands into either a constant true or false, or a brand new ICmp
37/// instruction. The sign is passed in to determine which kind of predicate to
38/// use in the new icmp instruction.
39static Value *getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS,
40 InstCombiner::BuilderTy &Builder) {
41 ICmpInst::Predicate NewPred;
42 if (Constant *TorF = getPredForICmpCode(Code, Sign, LHS->getType(), NewPred))
43 return TorF;
44 return Builder.CreateICmp(NewPred, LHS, RHS);
45}
46
47/// This is the complement of getFCmpCode, which turns an opcode and two
48/// operands into either a FCmp instruction, or a true/false constant.
49static Value *getFCmpValue(unsigned Code, Value *LHS, Value *RHS,
50 InstCombiner::BuilderTy &Builder, FMFSource FMF) {
51 FCmpInst::Predicate NewPred;
52 if (Constant *TorF = getPredForFCmpCode(Code, LHS->getType(), NewPred))
53 return TorF;
54 return Builder.CreateFCmpFMF(NewPred, LHS, RHS, FMF);
55}
56
57/// Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise
58/// (V < Lo || V >= Hi). This method expects that Lo < Hi. IsSigned indicates
59/// whether to treat V, Lo, and Hi as signed or not.
61 const APInt &Hi, bool isSigned,
62 bool Inside) {
63 assert((isSigned ? Lo.slt(Hi) : Lo.ult(Hi)) &&
64 "Lo is not < Hi in range emission code!");
65
66 Type *Ty = V->getType();
67
68 // V >= Min && V < Hi --> V < Hi
69 // V < Min || V >= Hi --> V >= Hi
71 if (isSigned ? Lo.isMinSignedValue() : Lo.isMinValue()) {
72 Pred = isSigned ? ICmpInst::getSignedPredicate(Pred) : Pred;
73 return Builder.CreateICmp(Pred, V, ConstantInt::get(Ty, Hi));
74 }
75
76 // V >= Lo && V < Hi --> V - Lo u< Hi - Lo
77 // V < Lo || V >= Hi --> V - Lo u>= Hi - Lo
78 Value *VMinusLo =
79 Builder.CreateSub(V, ConstantInt::get(Ty, Lo), V->getName() + ".off");
80 Constant *HiMinusLo = ConstantInt::get(Ty, Hi - Lo);
81 return Builder.CreateICmp(Pred, VMinusLo, HiMinusLo);
82}
83
84/// Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns
85/// that can be simplified.
86/// One of A and B is considered the mask. The other is the value. This is
87/// described as the "AMask" or "BMask" part of the enum. If the enum contains
88/// only "Mask", then both A and B can be considered masks. If A is the mask,
89/// then it was proven that (A & C) == C. This is trivial if C == A or C == 0.
90/// If both A and C are constants, this proof is also easy.
91/// For the following explanations, we assume that A is the mask.
92///
93/// "AllOnes" declares that the comparison is true only if (A & B) == A or all
94/// bits of A are set in B.
95/// Example: (icmp eq (A & 3), 3) -> AMask_AllOnes
96///
97/// "AllZeros" declares that the comparison is true only if (A & B) == 0 or all
98/// bits of A are cleared in B.
99/// Example: (icmp eq (A & 3), 0) -> Mask_AllZeroes
100///
101/// "Mixed" declares that (A & B) == C and C might or might not contain any
102/// number of one bits and zero bits.
103/// Example: (icmp eq (A & 3), 1) -> AMask_Mixed
104///
105/// "Not" means that in above descriptions "==" should be replaced by "!=".
106/// Example: (icmp ne (A & 3), 3) -> AMask_NotAllOnes
107///
108/// If the mask A contains a single bit, then the following is equivalent:
109/// (icmp eq (A & B), A) equals (icmp ne (A & B), 0)
110/// (icmp ne (A & B), A) equals (icmp eq (A & B), 0)
123
124/// Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C)
125/// satisfies.
126static unsigned getMaskedICmpType(Value *A, Value *B, Value *C,
127 ICmpInst::Predicate Pred) {
128 const APInt *ConstA = nullptr, *ConstB = nullptr, *ConstC = nullptr;
129 match(A, m_APInt(ConstA));
130 match(B, m_APInt(ConstB));
131 match(C, m_APInt(ConstC));
132 bool IsEq = (Pred == ICmpInst::ICMP_EQ);
133 bool IsAPow2 = ConstA && ConstA->isPowerOf2();
134 bool IsBPow2 = ConstB && ConstB->isPowerOf2();
135 unsigned MaskVal = 0;
136 if (ConstC && ConstC->isZero()) {
137 // if C is zero, then both A and B qualify as mask
138 MaskVal |= (IsEq ? (Mask_AllZeros | AMask_Mixed | BMask_Mixed)
140 if (IsAPow2)
141 MaskVal |= (IsEq ? (AMask_NotAllOnes | AMask_NotMixed)
143 if (IsBPow2)
144 MaskVal |= (IsEq ? (BMask_NotAllOnes | BMask_NotMixed)
146 return MaskVal;
147 }
148
149 if (A == C) {
150 MaskVal |= (IsEq ? (AMask_AllOnes | AMask_Mixed)
152 if (IsAPow2)
153 MaskVal |= (IsEq ? (Mask_NotAllZeros | AMask_NotMixed)
155 } else if (ConstA && ConstC && ConstC->isSubsetOf(*ConstA)) {
156 MaskVal |= (IsEq ? AMask_Mixed : AMask_NotMixed);
157 }
158
159 if (B == C) {
160 MaskVal |= (IsEq ? (BMask_AllOnes | BMask_Mixed)
162 if (IsBPow2)
163 MaskVal |= (IsEq ? (Mask_NotAllZeros | BMask_NotMixed)
165 } else if (ConstB && ConstC && ConstC->isSubsetOf(*ConstB)) {
166 MaskVal |= (IsEq ? BMask_Mixed : BMask_NotMixed);
167 }
168
169 return MaskVal;
170}
171
172/// Convert an analysis of a masked ICmp into its equivalent if all boolean
173/// operations had the opposite sense. Since each "NotXXX" flag (recording !=)
174/// is adjacent to the corresponding normal flag (recording ==), this just
175/// involves swapping those bits over.
176static unsigned conjugateICmpMask(unsigned Mask) {
177 unsigned NewMask;
178 NewMask = (Mask & (AMask_AllOnes | BMask_AllOnes | Mask_AllZeros |
180 << 1;
181
182 NewMask |= (Mask & (AMask_NotAllOnes | BMask_NotAllOnes | Mask_NotAllZeros |
184 >> 1;
185
186 return NewMask;
187}
188
189// Adapts the external decomposeBitTest for local use.
191 Value *&Y, Value *&Z) {
192 auto Res = llvm::decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
193 /*AllowNonZeroC=*/true);
194 if (!Res)
195 return false;
196
197 Pred = Res->Pred;
198 X = Res->X;
199 Y = ConstantInt::get(X->getType(), Res->Mask);
200 Z = ConstantInt::get(X->getType(), Res->C);
201 return true;
202}
203
204/// Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
205/// Return the pattern classes (from MaskedICmpType) for the left hand side and
206/// the right hand side as a pair.
207/// LHS and RHS are the left hand side and the right hand side ICmps and PredL
208/// and PredR are their predicates, respectively.
209static std::optional<std::pair<unsigned, unsigned>>
212 ICmpInst::Predicate &PredR) {
213
214 // Here comes the tricky part:
215 // LHS might be of the form L11 & L12 == X, X == L21 & L22,
216 // and L11 & L12 == L21 & L22. The same goes for RHS.
217 // Now we must find those components L** and R**, that are equal, so
218 // that we can extract the parameters A, B, C, D, and E for the canonical
219 // above.
220
221 // Check whether the icmp can be decomposed into a bit test.
222 Value *L1, *L11, *L12, *L2, *L21, *L22;
223 if (decomposeBitTest(LHS, PredL, L11, L12, L2)) {
224 L21 = L22 = L1 = nullptr;
225 } else {
226 auto *LHSCMP = dyn_cast<ICmpInst>(LHS);
227 if (!LHSCMP)
228 return std::nullopt;
229
230 // Don't allow pointers. Splat vectors are fine.
231 if (!LHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
232 return std::nullopt;
233
234 PredL = LHSCMP->getPredicate();
235 L1 = LHSCMP->getOperand(0);
236 L2 = LHSCMP->getOperand(1);
237 // Look for ANDs in the LHS icmp.
238 if (!match(L1, m_And(m_Value(L11), m_Value(L12)))) {
239 // Any icmp can be viewed as being trivially masked; if it allows us to
240 // remove one, it's worth it.
241 L11 = L1;
243 }
244
245 if (!match(L2, m_And(m_Value(L21), m_Value(L22)))) {
246 L21 = L2;
248 }
249 }
250
251 // Bail if LHS was a icmp that can't be decomposed into an equality.
252 if (!ICmpInst::isEquality(PredL))
253 return std::nullopt;
254
255 Value *R11, *R12, *R2;
256 if (decomposeBitTest(RHS, PredR, R11, R12, R2)) {
257 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
258 A = R11;
259 D = R12;
260 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
261 A = R12;
262 D = R11;
263 } else {
264 return std::nullopt;
265 }
266 E = R2;
267 } else {
268 auto *RHSCMP = dyn_cast<ICmpInst>(RHS);
269 if (!RHSCMP)
270 return std::nullopt;
271 // Don't allow pointers. Splat vectors are fine.
272 if (!RHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
273 return std::nullopt;
274
275 PredR = RHSCMP->getPredicate();
276
277 Value *R1 = RHSCMP->getOperand(0);
278 R2 = RHSCMP->getOperand(1);
279 bool Ok = false;
280 if (!match(R1, m_And(m_Value(R11), m_Value(R12)))) {
281 // As before, model no mask as a trivial mask if it'll let us do an
282 // optimization.
283 R11 = R1;
285 }
286
287 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
288 A = R11;
289 D = R12;
290 E = R2;
291 Ok = true;
292 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
293 A = R12;
294 D = R11;
295 E = R2;
296 Ok = true;
297 }
298
299 // Avoid matching against the -1 value we created for unmasked operand.
300 if (Ok && match(A, m_AllOnes()))
301 Ok = false;
302
303 // Look for ANDs on the right side of the RHS icmp.
304 if (!Ok) {
305 if (!match(R2, m_And(m_Value(R11), m_Value(R12)))) {
306 R11 = R2;
307 R12 = Constant::getAllOnesValue(R2->getType());
308 }
309
310 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
311 A = R11;
312 D = R12;
313 E = R1;
314 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
315 A = R12;
316 D = R11;
317 E = R1;
318 } else {
319 return std::nullopt;
320 }
321 }
322 }
323
324 // Bail if RHS was a icmp that can't be decomposed into an equality.
325 if (!ICmpInst::isEquality(PredR))
326 return std::nullopt;
327
328 if (L11 == A) {
329 B = L12;
330 C = L2;
331 } else if (L12 == A) {
332 B = L11;
333 C = L2;
334 } else if (L21 == A) {
335 B = L22;
336 C = L1;
337 } else if (L22 == A) {
338 B = L21;
339 C = L1;
340 }
341
342 unsigned LeftType = getMaskedICmpType(A, B, C, PredL);
343 unsigned RightType = getMaskedICmpType(A, D, E, PredR);
344 return std::optional<std::pair<unsigned, unsigned>>(
345 std::make_pair(LeftType, RightType));
346}
347
348/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single
349/// (icmp(A & X) ==/!= Y), where the left-hand side is of type Mask_NotAllZeros
350/// and the right hand side is of type BMask_Mixed. For example,
351/// (icmp (A & 12) != 0) & (icmp (A & 15) == 8) -> (icmp (A & 15) == 8).
352/// Also used for logical and/or, must be poison safe.
354 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E,
356 InstCombiner::BuilderTy &Builder) {
357 // We are given the canonical form:
358 // (icmp ne (A & B), 0) & (icmp eq (A & D), E).
359 // where D & E == E.
360 //
361 // If IsAnd is false, we get it in negated form:
362 // (icmp eq (A & B), 0) | (icmp ne (A & D), E) ->
363 // !((icmp ne (A & B), 0) & (icmp eq (A & D), E)).
364 //
365 // We currently handle the case of B, C, D, E are constant.
366 //
367 const APInt *BCst, *DCst, *OrigECst;
368 if (!match(B, m_APInt(BCst)) || !match(D, m_APInt(DCst)) ||
369 !match(E, m_APInt(OrigECst)))
370 return nullptr;
371
373
374 // Update E to the canonical form when D is a power of two and RHS is
375 // canonicalized as,
376 // (icmp ne (A & D), 0) -> (icmp eq (A & D), D) or
377 // (icmp ne (A & D), D) -> (icmp eq (A & D), 0).
378 APInt ECst = *OrigECst;
379 if (PredR != NewCC)
380 ECst ^= *DCst;
381
382 // If B or D is zero, skip because if LHS or RHS can be trivially folded by
383 // other folding rules and this pattern won't apply any more.
384 if (*BCst == 0 || *DCst == 0)
385 return nullptr;
386
387 // If B and D don't intersect, ie. (B & D) == 0, try to fold isNaN idiom:
388 // (icmp ne (A & FractionBits), 0) & (icmp eq (A & ExpBits), ExpBits)
389 // -> isNaN(A)
390 // Otherwise, we cannot deduce anything from it.
391 if (!BCst->intersects(*DCst)) {
392 Value *Src;
393 if (*DCst == ECst && match(A, m_ElementWiseBitCast(m_Value(Src))) &&
394 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
395 Attribute::StrictFP)) {
396 Type *Ty = Src->getType()->getScalarType();
397 if (!Ty->isIEEELikeFPTy())
398 return nullptr;
399
400 APInt ExpBits = APFloat::getInf(Ty->getFltSemantics()).bitcastToAPInt();
401 if (ECst != ExpBits)
402 return nullptr;
403 APInt FractionBits = ~ExpBits;
404 FractionBits.clearSignBit();
405 if (*BCst != FractionBits)
406 return nullptr;
407
408 return Builder.CreateFCmp(IsAnd ? FCmpInst::FCMP_UNO : FCmpInst::FCMP_ORD,
409 Src, ConstantFP::getZero(Src->getType()));
410 }
411 return nullptr;
412 }
413
414 // If the following two conditions are met:
415 //
416 // 1. mask B covers only a single bit that's not covered by mask D, that is,
417 // (B & (B ^ D)) is a power of 2 (in other words, B minus the intersection of
418 // B and D has only one bit set) and,
419 //
420 // 2. RHS (and E) indicates that the rest of B's bits are zero (in other
421 // words, the intersection of B and D is zero), that is, ((B & D) & E) == 0
422 //
423 // then that single bit in B must be one and thus the whole expression can be
424 // folded to
425 // (A & (B | D)) == (B & (B ^ D)) | E.
426 //
427 // For example,
428 // (icmp ne (A & 12), 0) & (icmp eq (A & 7), 1) -> (icmp eq (A & 15), 9)
429 // (icmp ne (A & 15), 0) & (icmp eq (A & 7), 0) -> (icmp eq (A & 15), 8)
430 if ((((*BCst & *DCst) & ECst) == 0) &&
431 (*BCst & (*BCst ^ *DCst)).isPowerOf2()) {
432 APInt BorD = *BCst | *DCst;
433 APInt BandBxorDorE = (*BCst & (*BCst ^ *DCst)) | ECst;
434 Value *NewMask = ConstantInt::get(A->getType(), BorD);
435 Value *NewMaskedValue = ConstantInt::get(A->getType(), BandBxorDorE);
436 Value *NewAnd = Builder.CreateAnd(A, NewMask);
437 return Builder.CreateICmp(NewCC, NewAnd, NewMaskedValue);
438 }
439
440 auto IsSubSetOrEqual = [](const APInt *C1, const APInt *C2) {
441 return (*C1 & *C2) == *C1;
442 };
443 auto IsSuperSetOrEqual = [](const APInt *C1, const APInt *C2) {
444 return (*C1 & *C2) == *C2;
445 };
446
447 // In the following, we consider only the cases where B is a superset of D, B
448 // is a subset of D, or B == D because otherwise there's at least one bit
449 // covered by B but not D, in which case we can't deduce much from it, so
450 // no folding (aside from the single must-be-one bit case right above.)
451 // For example,
452 // (icmp ne (A & 14), 0) & (icmp eq (A & 3), 1) -> no folding.
453 if (!IsSubSetOrEqual(BCst, DCst) && !IsSuperSetOrEqual(BCst, DCst))
454 return nullptr;
455
456 // At this point, either B is a superset of D, B is a subset of D or B == D.
457
458 // If E is zero, if B is a subset of (or equal to) D, LHS and RHS contradict
459 // and the whole expression becomes false (or true if negated), otherwise, no
460 // folding.
461 // For example,
462 // (icmp ne (A & 3), 0) & (icmp eq (A & 7), 0) -> false.
463 // (icmp ne (A & 15), 0) & (icmp eq (A & 3), 0) -> no folding.
464 if (ECst.isZero()) {
465 if (IsSubSetOrEqual(BCst, DCst))
466 return ConstantInt::get(LHS->getType(), !IsAnd);
467 return nullptr;
468 }
469
470 // At this point, B, D, E aren't zero and (B & D) == B, (B & D) == D or B ==
471 // D. If B is a superset of (or equal to) D, since E is not zero, LHS is
472 // subsumed by RHS (RHS implies LHS.) So the whole expression becomes
473 // RHS. For example,
474 // (icmp ne (A & 255), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
475 // (icmp ne (A & 15), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
476 if (IsSuperSetOrEqual(BCst, DCst)) {
477 // We can't guarantee that samesign hold after this fold.
478 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
479 ICmp->setSameSign(false);
480 return RHS;
481 }
482 // Otherwise, B is a subset of D. If B and E have a common bit set,
483 // ie. (B & E) != 0, then LHS is subsumed by RHS. For example.
484 // (icmp ne (A & 12), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
485 assert(IsSubSetOrEqual(BCst, DCst) && "Precondition due to above code");
486 if ((*BCst & ECst) != 0) {
487 // We can't guarantee that samesign hold after this fold.
488 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
489 ICmp->setSameSign(false);
490 return RHS;
491 }
492 // Otherwise, LHS and RHS contradict and the whole expression becomes false
493 // (or true if negated.) For example,
494 // (icmp ne (A & 7), 0) & (icmp eq (A & 15), 8) -> false.
495 // (icmp ne (A & 6), 0) & (icmp eq (A & 15), 8) -> false.
496 return ConstantInt::get(LHS->getType(), !IsAnd);
497}
498
499/// Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single
500/// (icmp(A & X) ==/!= Y), where the left-hand side and the right hand side
501/// aren't of the common mask pattern type.
502/// Also used for logical and/or, must be poison safe.
504 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D,
506 unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder) {
508 "Expected equality predicates for masked type of icmps.");
509 // Handle Mask_NotAllZeros-BMask_Mixed cases.
510 // (icmp ne/eq (A & B), C) &/| (icmp eq/ne (A & D), E), or
511 // (icmp eq/ne (A & B), C) &/| (icmp ne/eq (A & D), E)
512 // which gets swapped to
513 // (icmp ne/eq (A & D), E) &/| (icmp eq/ne (A & B), C).
514 if (!IsAnd) {
515 LHSMask = conjugateICmpMask(LHSMask);
516 RHSMask = conjugateICmpMask(RHSMask);
517 }
518 if ((LHSMask & Mask_NotAllZeros) && (RHSMask & BMask_Mixed)) {
520 LHS, RHS, IsAnd, A, B, D, E, PredL, PredR, Builder)) {
521 return V;
522 }
523 } else if ((LHSMask & BMask_Mixed) && (RHSMask & Mask_NotAllZeros)) {
525 RHS, LHS, IsAnd, A, D, B, C, PredR, PredL, Builder)) {
526 return V;
527 }
528 }
529 return nullptr;
530}
531
532/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E)
533/// into a single (icmp(A & X) ==/!= Y).
535 bool IsLogical,
537 const SimplifyQuery &Q) {
538 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
539 ICmpInst::Predicate PredL, PredR;
540 std::optional<std::pair<unsigned, unsigned>> MaskPair =
541 getMaskedTypeForICmpPair(A, B, C, D, E, LHS, RHS, PredL, PredR);
542 if (!MaskPair)
543 return nullptr;
545 "Expected equality predicates for masked type of icmps.");
546 unsigned LHSMask = MaskPair->first;
547 unsigned RHSMask = MaskPair->second;
548 unsigned Mask = LHSMask & RHSMask;
549 if (Mask == 0) {
550 // Even if the two sides don't share a common pattern, check if folding can
551 // still happen.
553 LHS, RHS, IsAnd, A, B, C, D, E, PredL, PredR, LHSMask, RHSMask,
554 Builder))
555 return V;
556 return nullptr;
557 }
558
559 // In full generality:
560 // (icmp (A & B) Op C) | (icmp (A & D) Op E)
561 // == ![ (icmp (A & B) !Op C) & (icmp (A & D) !Op E) ]
562 //
563 // If the latter can be converted into (icmp (A & X) Op Y) then the former is
564 // equivalent to (icmp (A & X) !Op Y).
565 //
566 // Therefore, we can pretend for the rest of this function that we're dealing
567 // with the conjunction, provided we flip the sense of any comparisons (both
568 // input and output).
569
570 // In most cases we're going to produce an EQ for the "&&" case.
572 if (!IsAnd) {
573 // Convert the masking analysis into its equivalent with negated
574 // comparisons.
575 Mask = conjugateICmpMask(Mask);
576 }
577
578 if (Mask & Mask_AllZeros) {
579 // (icmp eq (A & B), 0) & (icmp eq (A & D), 0)
580 // -> (icmp eq (A & (B|D)), 0)
581 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
582 return nullptr; // TODO: Use freeze?
583 Value *NewOr = Builder.CreateOr(B, D);
584 Value *NewAnd = Builder.CreateAnd(A, NewOr);
585 // We can't use C as zero because we might actually handle
586 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
587 // with B and D, having a single bit set.
588 Value *Zero = Constant::getNullValue(A->getType());
589 return Builder.CreateICmp(NewCC, NewAnd, Zero);
590 }
591 if (Mask & BMask_AllOnes) {
592 // (icmp eq (A & B), B) & (icmp eq (A & D), D)
593 // -> (icmp eq (A & (B|D)), (B|D))
594 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
595 return nullptr; // TODO: Use freeze?
596 Value *NewOr = Builder.CreateOr(B, D);
597 Value *NewAnd = Builder.CreateAnd(A, NewOr);
598 return Builder.CreateICmp(NewCC, NewAnd, NewOr);
599 }
600 if (Mask & AMask_AllOnes) {
601 // (icmp eq (A & B), A) & (icmp eq (A & D), A)
602 // -> (icmp eq (A & (B&D)), A)
603 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
604 return nullptr; // TODO: Use freeze?
605 Value *NewAnd1 = Builder.CreateAnd(B, D);
606 Value *NewAnd2 = Builder.CreateAnd(A, NewAnd1);
607 return Builder.CreateICmp(NewCC, NewAnd2, A);
608 }
609
610 const APInt *ConstB, *ConstD;
611 if (match(B, m_APInt(ConstB)) && match(D, m_APInt(ConstD))) {
612 if (Mask & (Mask_NotAllZeros | BMask_NotAllOnes)) {
613 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0) and
614 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
615 // -> (icmp ne (A & B), 0) or (icmp ne (A & D), 0)
616 // Only valid if one of the masks is a superset of the other (check "B&D"
617 // is the same as either B or D).
618 APInt NewMask = *ConstB & *ConstD;
619 if (NewMask == *ConstB)
620 return LHS;
621 if (NewMask == *ConstD) {
622 if (IsLogical) {
623 if (auto *RHSI = dyn_cast<Instruction>(RHS))
624 RHSI->dropPoisonGeneratingFlags();
625 }
626 return RHS;
627 }
628 }
629
630 if (Mask & AMask_NotAllOnes) {
631 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
632 // -> (icmp ne (A & B), A) or (icmp ne (A & D), A)
633 // Only valid if one of the masks is a superset of the other (check "B|D"
634 // is the same as either B or D).
635 APInt NewMask = *ConstB | *ConstD;
636 if (NewMask == *ConstB)
637 return LHS;
638 if (NewMask == *ConstD)
639 return RHS;
640 }
641
642 if (Mask & (BMask_Mixed | BMask_NotMixed)) {
643 // Mixed:
644 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
645 // We already know that B & C == C && D & E == E.
646 // If we can prove that (B & D) & (C ^ E) == 0, that is, the bits of
647 // C and E, which are shared by both the mask B and the mask D, don't
648 // contradict, then we can transform to
649 // -> (icmp eq (A & (B|D)), (C|E))
650 // Currently, we only handle the case of B, C, D, and E being constant.
651 // We can't simply use C and E because we might actually handle
652 // (icmp ne (A & B), B) & (icmp eq (A & D), D)
653 // with B and D, having a single bit set.
654
655 // NotMixed:
656 // (icmp ne (A & B), C) & (icmp ne (A & D), E)
657 // -> (icmp ne (A & (B & D)), (C & E))
658 // Check the intersection (B & D) for inequality.
659 // Assume that (B & D) == B || (B & D) == D, i.e B/D is a subset of D/B
660 // and (B & D) & (C ^ E) == 0, bits of C and E, which are shared by both
661 // the B and the D, don't contradict. Note that we can assume (~B & C) ==
662 // 0 && (~D & E) == 0, previous operation should delete these icmps if it
663 // hadn't been met.
664
665 const APInt *OldConstC, *OldConstE;
666 if (!match(C, m_APInt(OldConstC)) || !match(E, m_APInt(OldConstE)))
667 return nullptr;
668
669 auto FoldBMixed = [&](ICmpInst::Predicate CC, bool IsNot) -> Value * {
670 CC = IsNot ? CmpInst::getInversePredicate(CC) : CC;
671 const APInt ConstC = PredL != CC ? *ConstB ^ *OldConstC : *OldConstC;
672 const APInt ConstE = PredR != CC ? *ConstD ^ *OldConstE : *OldConstE;
673
674 if (((*ConstB & *ConstD) & (ConstC ^ ConstE)).getBoolValue())
675 return IsNot ? nullptr : ConstantInt::get(LHS->getType(), !IsAnd);
676
677 if (IsNot && !ConstB->isSubsetOf(*ConstD) &&
678 !ConstD->isSubsetOf(*ConstB))
679 return nullptr;
680
681 APInt BD, CE;
682 if (IsNot) {
683 BD = *ConstB & *ConstD;
684 CE = ConstC & ConstE;
685 } else {
686 BD = *ConstB | *ConstD;
687 CE = ConstC | ConstE;
688 }
689 Value *NewAnd = Builder.CreateAnd(A, BD);
690 Value *CEVal = ConstantInt::get(A->getType(), CE);
691 return Builder.CreateICmp(CC, NewAnd, CEVal);
692 };
693
694 if (Mask & BMask_Mixed)
695 return FoldBMixed(NewCC, false);
696 if (Mask & BMask_NotMixed) // can be else also
697 return FoldBMixed(NewCC, true);
698 }
699 }
700
701 // (icmp eq (A & B), 0) | (icmp eq (A & D), 0)
702 // -> (icmp ne (A & (B|D)), (B|D))
703 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0)
704 // -> (icmp eq (A & (B|D)), (B|D))
705 // iff B and D is known to be a power of two
706 if (Mask & Mask_NotAllZeros &&
707 isKnownToBeAPowerOfTwo(B, /*OrZero=*/false, Q) &&
708 isKnownToBeAPowerOfTwo(D, /*OrZero=*/false, Q)) {
709 // If this is a logical and/or, then we must prevent propagation of a
710 // poison value from the RHS by inserting freeze.
711 if (IsLogical)
712 D = Builder.CreateFreeze(D);
713 Value *Mask = Builder.CreateOr(B, D);
714 Value *Masked = Builder.CreateAnd(A, Mask);
715 return Builder.CreateICmp(NewCC, Masked, Mask);
716 }
717 return nullptr;
718}
719
720/// Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
721/// Example: (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
722/// If \p Inverted is true then the check is for the inverted range, e.g.
723/// (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
725 bool Inverted) {
726 // Check the lower range comparison, e.g. x >= 0
727 // InstCombine already ensured that if there is a constant it's on the RHS.
728 ConstantInt *RangeStart = dyn_cast<ConstantInt>(Cmp0->getOperand(1));
729 if (!RangeStart)
730 return nullptr;
731
732 ICmpInst::Predicate Pred0 = (Inverted ? Cmp0->getInversePredicate() :
733 Cmp0->getPredicate());
734
735 // Accept x > -1 or x >= 0 (after potentially inverting the predicate).
736 if (!((Pred0 == ICmpInst::ICMP_SGT && RangeStart->isMinusOne()) ||
737 (Pred0 == ICmpInst::ICMP_SGE && RangeStart->isZero())))
738 return nullptr;
739
740 ICmpInst::Predicate Pred1 = (Inverted ? Cmp1->getInversePredicate() :
741 Cmp1->getPredicate());
742
743 Value *Input = Cmp0->getOperand(0);
744 Value *Cmp1Op0 = Cmp1->getOperand(0);
745 Value *Cmp1Op1 = Cmp1->getOperand(1);
746 Value *RangeEnd;
747 if (match(Cmp1Op0, m_SExtOrSelf(m_Specific(Input)))) {
748 // For the upper range compare we have: icmp x, n
749 Input = Cmp1Op0;
750 RangeEnd = Cmp1Op1;
751 } else if (match(Cmp1Op1, m_SExtOrSelf(m_Specific(Input)))) {
752 // For the upper range compare we have: icmp n, x
753 Input = Cmp1Op1;
754 RangeEnd = Cmp1Op0;
755 Pred1 = ICmpInst::getSwappedPredicate(Pred1);
756 } else {
757 return nullptr;
758 }
759
760 // Check the upper range comparison, e.g. x < n
761 ICmpInst::Predicate NewPred;
762 switch (Pred1) {
763 case ICmpInst::ICMP_SLT: NewPred = ICmpInst::ICMP_ULT; break;
764 case ICmpInst::ICMP_SLE: NewPred = ICmpInst::ICMP_ULE; break;
765 default: return nullptr;
766 }
767
768 // This simplification is only valid if the upper range is not negative.
769 KnownBits Known = computeKnownBits(RangeEnd, Cmp1);
770 if (!Known.isNonNegative())
771 return nullptr;
772
773 if (Inverted)
774 NewPred = ICmpInst::getInversePredicate(NewPred);
775
776 return Builder.CreateICmp(NewPred, Input, RangeEnd);
777}
778
779// (or (icmp eq X, 0), (icmp eq X, Pow2OrZero))
780// -> (icmp eq (and X, Pow2OrZero), X)
781// (and (icmp ne X, 0), (icmp ne X, Pow2OrZero))
782// -> (icmp ne (and X, Pow2OrZero), X)
783static Value *
785 ICmpInst *LHS, ICmpInst *RHS, bool IsAnd,
786 const SimplifyQuery &Q) {
788 // Make sure we have right compares for our op.
789 if (LHS->getPredicate() != Pred || RHS->getPredicate() != Pred)
790 return nullptr;
791
792 // Make it so we can match LHS against the (icmp eq/ne X, 0) just for
793 // simplicity.
794 if (match(RHS->getOperand(1), m_Zero()))
795 std::swap(LHS, RHS);
796
797 Value *Pow2, *Op;
798 // Match the desired pattern:
799 // LHS: (icmp eq/ne X, 0)
800 // RHS: (icmp eq/ne X, Pow2OrZero)
801 // Skip if Pow2OrZero is 1. Either way it gets folded to (icmp ugt X, 1) but
802 // this form ends up slightly less canonical.
803 // We could potentially be more sophisticated than requiring LHS/RHS
804 // be one-use. We don't create additional instructions if only one
805 // of them is one-use. So cases where one is one-use and the other
806 // is two-use might be profitable.
807 if (!match(LHS, m_OneUse(m_ICmp(Pred, m_Value(Op), m_Zero()))) ||
808 !match(RHS, m_OneUse(m_c_ICmp(Pred, m_Specific(Op), m_Value(Pow2)))) ||
809 match(Pow2, m_One()) ||
810 !isKnownToBeAPowerOfTwo(Pow2, Q.DL, /*OrZero=*/true, Q.AC, Q.CxtI, Q.DT))
811 return nullptr;
812
813 Value *And = Builder.CreateAnd(Op, Pow2);
814 return Builder.CreateICmp(Pred, And, Op);
815}
816
817/// General pattern:
818/// X & Y
819///
820/// Where Y is checking that all the high bits (covered by a mask 4294967168)
821/// are uniform, i.e. %arg & 4294967168 can be either 4294967168 or 0
822/// Pattern can be one of:
823/// %t = add i32 %arg, 128
824/// %r = icmp ult i32 %t, 256
825/// Or
826/// %t0 = shl i32 %arg, 24
827/// %t1 = ashr i32 %t0, 24
828/// %r = icmp eq i32 %t1, %arg
829/// Or
830/// %t0 = trunc i32 %arg to i8
831/// %t1 = sext i8 %t0 to i32
832/// %r = icmp eq i32 %t1, %arg
833/// This pattern is a signed truncation check.
834///
835/// And X is checking that some bit in that same mask is zero.
836/// I.e. can be one of:
837/// %r = icmp sgt i32 %arg, -1
838/// Or
839/// %t = and i32 %arg, 2147483648
840/// %r = icmp eq i32 %t, 0
841///
842/// Since we are checking that all the bits in that mask are the same,
843/// and a particular bit is zero, what we are really checking is that all the
844/// masked bits are zero.
845/// So this should be transformed to:
846/// %r = icmp ult i32 %arg, 128
848 Instruction &CxtI,
849 InstCombiner::BuilderTy &Builder) {
850 assert(CxtI.getOpcode() == Instruction::And);
851
852 // Match icmp ult (add %arg, C01), C1 (C1 == C01 << 1; powers of two)
853 auto tryToMatchSignedTruncationCheck = [](ICmpInst *ICmp, Value *&X,
854 APInt &SignBitMask) -> bool {
855 const APInt *I01, *I1; // powers of two; I1 == I01 << 1
857 m_Add(m_Value(X), m_Power2(I01)),
858 m_Power2(I1))) &&
859 I1->ugt(*I01) && I01->shl(1) == *I1))
860 return false;
861 // Which bit is the new sign bit as per the 'signed truncation' pattern?
862 SignBitMask = *I01;
863 return true;
864 };
865
866 // One icmp needs to be 'signed truncation check'.
867 // We need to match this first, else we will mismatch commutative cases.
868 Value *X1;
869 APInt HighestBit;
870 ICmpInst *OtherICmp;
871 if (tryToMatchSignedTruncationCheck(ICmp1, X1, HighestBit))
872 OtherICmp = ICmp0;
873 else if (tryToMatchSignedTruncationCheck(ICmp0, X1, HighestBit))
874 OtherICmp = ICmp1;
875 else
876 return nullptr;
877
878 assert(HighestBit.isPowerOf2() && "expected to be power of two (non-zero)");
879
880 // Try to match/decompose into: icmp eq (X & Mask), 0
881 auto tryToDecompose = [](ICmpInst *ICmp, Value *&X,
882 APInt &UnsetBitsMask) -> bool {
883 CmpPredicate Pred = ICmp->getPredicate();
884 // Can it be decomposed into icmp eq (X & Mask), 0 ?
886 ICmp->getOperand(0), ICmp->getOperand(1), Pred,
887 /*LookThroughTrunc=*/false, /*AllowNonZeroC=*/false,
888 /*DecomposeAnd=*/true);
889 if (Res && Res->Pred == ICmpInst::ICMP_EQ) {
890 X = Res->X;
891 UnsetBitsMask = Res->Mask;
892 return true;
893 }
894
895 return false;
896 };
897
898 // And the other icmp needs to be decomposable into a bit test.
899 Value *X0;
900 APInt UnsetBitsMask;
901 if (!tryToDecompose(OtherICmp, X0, UnsetBitsMask))
902 return nullptr;
903
904 assert(!UnsetBitsMask.isZero() && "empty mask makes no sense.");
905
906 // Are they working on the same value?
907 Value *X;
908 if (X1 == X0) {
909 // Ok as is.
910 X = X1;
911 } else if (match(X0, m_Trunc(m_Specific(X1)))) {
912 UnsetBitsMask = UnsetBitsMask.zext(X1->getType()->getScalarSizeInBits());
913 X = X1;
914 } else
915 return nullptr;
916
917 // So which bits should be uniform as per the 'signed truncation check'?
918 // (all the bits starting with (i.e. including) HighestBit)
919 APInt SignBitsMask = ~(HighestBit - 1U);
920
921 // UnsetBitsMask must have some common bits with SignBitsMask,
922 if (!UnsetBitsMask.intersects(SignBitsMask))
923 return nullptr;
924
925 // Does UnsetBitsMask contain any bits outside of SignBitsMask?
926 if (!UnsetBitsMask.isSubsetOf(SignBitsMask)) {
927 APInt OtherHighestBit = (~UnsetBitsMask) + 1U;
928 if (!OtherHighestBit.isPowerOf2())
929 return nullptr;
930 HighestBit = APIntOps::umin(HighestBit, OtherHighestBit);
931 }
932 // Else, if it does not, then all is ok as-is.
933
934 // %r = icmp ult %X, SignBit
935 return Builder.CreateICmpULT(X, ConstantInt::get(X->getType(), HighestBit),
936 CxtI.getName() + ".simplified");
937}
938
939/// Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and
940/// fold (icmp ne ctpop(X) 1) & (icmp ne X 0) into (icmp ugt ctpop(X) 1).
941/// Also used for logical and/or, must be poison safe if range attributes are
942/// dropped.
943static Value *foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd,
945 InstCombinerImpl &IC) {
946 CmpPredicate Pred0, Pred1;
947 Value *X;
949 m_SpecificInt(1))) ||
950 !match(Cmp1, m_ICmp(Pred1, m_Specific(X), m_ZeroInt())))
951 return nullptr;
952
953 auto *CtPop = cast<Instruction>(Cmp0->getOperand(0));
954 if (IsAnd && Pred0 == ICmpInst::ICMP_NE && Pred1 == ICmpInst::ICMP_NE) {
955 // Drop range attributes and re-infer them in the next iteration.
956 CtPop->dropPoisonGeneratingAnnotations();
957 IC.addToWorklist(CtPop);
958 return Builder.CreateICmpUGT(CtPop, ConstantInt::get(CtPop->getType(), 1));
959 }
960 if (!IsAnd && Pred0 == ICmpInst::ICMP_EQ && Pred1 == ICmpInst::ICMP_EQ) {
961 // Drop range attributes and re-infer them in the next iteration.
962 CtPop->dropPoisonGeneratingAnnotations();
963 IC.addToWorklist(CtPop);
964 return Builder.CreateICmpULT(CtPop, ConstantInt::get(CtPop->getType(), 2));
965 }
966
967 return nullptr;
968}
969
970/// Reduce a pair of compares that check if a value has exactly 1 bit set.
971/// Also used for logical and/or, must be poison safe if range attributes are
972/// dropped.
973static Value *foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd,
975 InstCombinerImpl &IC) {
976 // Handle 'and' / 'or' commutation: make the equality check the first operand.
977 if (JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_NE)
978 std::swap(Cmp0, Cmp1);
979 else if (!JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_EQ)
980 std::swap(Cmp0, Cmp1);
981
982 // (X != 0) && (ctpop(X) u< 2) --> ctpop(X) == 1
983 Value *X;
984 if (JoinedByAnd &&
988 m_SpecificInt(2)))) {
989 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
990 // Drop range attributes and re-infer them in the next iteration.
991 CtPop->dropPoisonGeneratingAnnotations();
992 IC.addToWorklist(CtPop);
993 return Builder.CreateICmpEQ(CtPop, ConstantInt::get(CtPop->getType(), 1));
994 }
995 // (X == 0) || (ctpop(X) u> 1) --> ctpop(X) != 1
996 if (!JoinedByAnd &&
1000 m_SpecificInt(1)))) {
1001 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
1002 // Drop range attributes and re-infer them in the next iteration.
1003 CtPop->dropPoisonGeneratingAnnotations();
1004 IC.addToWorklist(CtPop);
1005 return Builder.CreateICmpNE(CtPop, ConstantInt::get(CtPop->getType(), 1));
1006 }
1007 return nullptr;
1008}
1009
1010/// Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff
1011/// B is a contiguous set of ones starting from the most significant bit
1012/// (negative power of 2), D and E are equal, and D is a contiguous set of ones
1013/// starting at the most significant zero bit in B. Parameter B supports masking
1014/// using undef/poison in either scalar or vector values.
1016 Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL,
1019 "Expected equality predicates for masked type of icmps.");
1020 if (PredL != ICmpInst::ICMP_EQ || PredR != ICmpInst::ICMP_NE)
1021 return nullptr;
1022
1023 if (!match(B, m_NegatedPower2()) || !match(D, m_ShiftedMask()) ||
1024 !match(E, m_ShiftedMask()))
1025 return nullptr;
1026
1027 // Test scalar arguments for conversion. B has been validated earlier to be a
1028 // negative power of two and thus is guaranteed to have one or more contiguous
1029 // ones starting from the MSB followed by zero or more contiguous zeros. D has
1030 // been validated earlier to be a shifted set of one or more contiguous ones.
1031 // In order to match, B leading ones and D leading zeros should be equal. The
1032 // predicate that B be a negative power of 2 prevents the condition of there
1033 // ever being zero leading ones. Thus 0 == 0 cannot occur. The predicate that
1034 // D always be a shifted mask prevents the condition of D equaling 0. This
1035 // prevents matching the condition where B contains the maximum number of
1036 // leading one bits (-1) and D contains the maximum number of leading zero
1037 // bits (0).
1038 auto isReducible = [](const Value *B, const Value *D, const Value *E) {
1039 const APInt *BCst, *DCst, *ECst;
1040 return match(B, m_APIntAllowPoison(BCst)) && match(D, m_APInt(DCst)) &&
1041 match(E, m_APInt(ECst)) && *DCst == *ECst &&
1042 (isa<PoisonValue>(B) ||
1043 (BCst->countLeadingOnes() == DCst->countLeadingZeros()));
1044 };
1045
1046 // Test vector type arguments for conversion.
1047 if (const auto *BVTy = dyn_cast<VectorType>(B->getType())) {
1048 const auto *BFVTy = dyn_cast<FixedVectorType>(BVTy);
1049 const auto *BConst = dyn_cast<Constant>(B);
1050 const auto *DConst = dyn_cast<Constant>(D);
1051 const auto *EConst = dyn_cast<Constant>(E);
1052
1053 if (!BFVTy || !BConst || !DConst || !EConst)
1054 return nullptr;
1055
1056 for (unsigned I = 0; I != BFVTy->getNumElements(); ++I) {
1057 const auto *BElt = BConst->getAggregateElement(I);
1058 const auto *DElt = DConst->getAggregateElement(I);
1059 const auto *EElt = EConst->getAggregateElement(I);
1060
1061 if (!BElt || !DElt || !EElt)
1062 return nullptr;
1063 if (!isReducible(BElt, DElt, EElt))
1064 return nullptr;
1065 }
1066 } else {
1067 // Test scalar type arguments for conversion.
1068 if (!isReducible(B, D, E))
1069 return nullptr;
1070 }
1071 return Builder.CreateICmp(ICmpInst::ICMP_ULT, A, D);
1072}
1073
1074/// Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) &
1075/// (icmp(X & M) != M)) into (icmp X u< M). Where P is a power of 2, M < P, and
1076/// M is a contiguous shifted mask starting at the right most significant zero
1077/// bit in P. SGT is supported as when P is the largest representable power of
1078/// 2, an earlier optimization converts the expression into (icmp X s> -1).
1079/// Parameter P supports masking using undef/poison in either scalar or vector
1080/// values.
1082 bool JoinedByAnd,
1083 InstCombiner::BuilderTy &Builder) {
1084 if (!JoinedByAnd)
1085 return nullptr;
1086 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
1087 ICmpInst::Predicate CmpPred0, CmpPred1;
1088 // Assuming P is a 2^n, getMaskedTypeForICmpPair will normalize (icmp X u<
1089 // 2^n) into (icmp (X & ~(2^n-1)) == 0) and (icmp X s> -1) into (icmp (X &
1090 // SignMask) == 0).
1091 std::optional<std::pair<unsigned, unsigned>> MaskPair =
1092 getMaskedTypeForICmpPair(A, B, C, D, E, Cmp0, Cmp1, CmpPred0, CmpPred1);
1093 if (!MaskPair)
1094 return nullptr;
1095
1096 const auto compareBMask = BMask_NotMixed | BMask_NotAllOnes;
1097 unsigned CmpMask0 = MaskPair->first;
1098 unsigned CmpMask1 = MaskPair->second;
1099 if ((CmpMask0 & Mask_AllZeros) && (CmpMask1 == compareBMask)) {
1100 if (Value *V = foldNegativePower2AndShiftedMask(A, B, D, E, CmpPred0,
1101 CmpPred1, Builder))
1102 return V;
1103 } else if ((CmpMask0 == compareBMask) && (CmpMask1 & Mask_AllZeros)) {
1104 if (Value *V = foldNegativePower2AndShiftedMask(A, D, B, C, CmpPred1,
1105 CmpPred0, Builder))
1106 return V;
1107 }
1108 return nullptr;
1109}
1110
1111/// Commuted variants are assumed to be handled by calling this function again
1112/// with the parameters swapped.
1114 ICmpInst *UnsignedICmp, bool IsAnd,
1115 const SimplifyQuery &Q,
1116 InstCombiner::BuilderTy &Builder) {
1117 Value *ZeroCmpOp;
1118 CmpPredicate EqPred;
1119 if (!match(ZeroICmp, m_ICmp(EqPred, m_Value(ZeroCmpOp), m_Zero())) ||
1120 !ICmpInst::isEquality(EqPred))
1121 return nullptr;
1122
1123 CmpPredicate UnsignedPred;
1124
1125 Value *A, *B;
1126 if (match(UnsignedICmp,
1127 m_c_ICmp(UnsignedPred, m_Specific(ZeroCmpOp), m_Value(A))) &&
1128 match(ZeroCmpOp, m_c_Add(m_Specific(A), m_Value(B))) &&
1129 (ZeroICmp->hasOneUse() || UnsignedICmp->hasOneUse())) {
1130 auto GetKnownNonZeroAndOther = [&](Value *&NonZero, Value *&Other) {
1131 if (!isKnownNonZero(NonZero, Q))
1132 std::swap(NonZero, Other);
1133 return isKnownNonZero(NonZero, Q);
1134 };
1135
1136 // Given ZeroCmpOp = (A + B)
1137 // ZeroCmpOp < A && ZeroCmpOp != 0 --> (0-X) < Y iff
1138 // ZeroCmpOp >= A || ZeroCmpOp == 0 --> (0-X) >= Y iff
1139 // with X being the value (A/B) that is known to be non-zero,
1140 // and Y being remaining value.
1141 if (UnsignedPred == ICmpInst::ICMP_ULT && EqPred == ICmpInst::ICMP_NE &&
1142 IsAnd && GetKnownNonZeroAndOther(B, A))
1143 return Builder.CreateICmpULT(Builder.CreateNeg(B), A);
1144 if (UnsignedPred == ICmpInst::ICMP_UGE && EqPred == ICmpInst::ICMP_EQ &&
1145 !IsAnd && GetKnownNonZeroAndOther(B, A))
1146 return Builder.CreateICmpUGE(Builder.CreateNeg(B), A);
1147 }
1148
1149 return nullptr;
1150}
1151
1152struct IntPart {
1154 unsigned StartBit;
1155 unsigned NumBits;
1156};
1157
1158/// Match an extraction of bits from an integer.
1159static std::optional<IntPart> matchIntPart(Value *V) {
1160 Value *X;
1161 if (!match(V, m_OneUse(m_Trunc(m_Value(X)))))
1162 return std::nullopt;
1163
1164 unsigned NumOriginalBits = X->getType()->getScalarSizeInBits();
1165 unsigned NumExtractedBits = V->getType()->getScalarSizeInBits();
1166 Value *Y;
1167 const APInt *Shift;
1168 // For a trunc(lshr Y, Shift) pattern, make sure we're only extracting bits
1169 // from Y, not any shifted-in zeroes.
1170 if (match(X, m_OneUse(m_LShr(m_Value(Y), m_APInt(Shift)))) &&
1171 Shift->ule(NumOriginalBits - NumExtractedBits))
1172 return {{Y, (unsigned)Shift->getZExtValue(), NumExtractedBits}};
1173 return {{X, 0, NumExtractedBits}};
1174}
1175
1176/// Materialize an extraction of bits from an integer in IR.
1177static Value *extractIntPart(const IntPart &P, IRBuilderBase &Builder) {
1178 Value *V = P.From;
1179 if (P.StartBit)
1180 V = Builder.CreateLShr(V, P.StartBit);
1181 Type *TruncTy = V->getType()->getWithNewBitWidth(P.NumBits);
1182 if (TruncTy != V->getType())
1183 V = Builder.CreateTrunc(V, TruncTy);
1184 return V;
1185}
1186
1187/// (icmp eq X0, Y0) & (icmp eq X1, Y1) -> icmp eq X01, Y01
1188/// (icmp ne X0, Y0) | (icmp ne X1, Y1) -> icmp ne X01, Y01
1189/// where X0, X1 and Y0, Y1 are adjacent parts extracted from an integer.
1190Value *InstCombinerImpl::foldEqOfParts(Value *Cmp0, Value *Cmp1, bool IsAnd) {
1191 if (!Cmp0->hasOneUse() || !Cmp1->hasOneUse())
1192 return nullptr;
1193
1195 auto GetMatchPart = [&](Value *CmpV,
1196 unsigned OpNo) -> std::optional<IntPart> {
1197 assert(CmpV->getType()->isIntOrIntVectorTy(1) && "Must be bool");
1198
1199 Value *X, *Y;
1200 // icmp ne (and x, 1), (and y, 1) <=> trunc (xor x, y) to i1
1201 // icmp eq (and x, 1), (and y, 1) <=> not (trunc (xor x, y) to i1)
1202 if (Pred == CmpInst::ICMP_NE
1203 ? match(CmpV, m_Trunc(m_Xor(m_Value(X), m_Value(Y))))
1204 : match(CmpV, m_Not(m_Trunc(m_Xor(m_Value(X), m_Value(Y))))))
1205 return {{OpNo == 0 ? X : Y, 0, 1}};
1206
1207 auto *Cmp = dyn_cast<ICmpInst>(CmpV);
1208 if (!Cmp)
1209 return std::nullopt;
1210
1211 if (Pred == Cmp->getPredicate())
1212 return matchIntPart(Cmp->getOperand(OpNo));
1213
1214 const APInt *C;
1215 // (icmp eq (lshr x, C), (lshr y, C)) gets optimized to:
1216 // (icmp ult (xor x, y), 1 << C) so also look for that.
1217 if (Pred == CmpInst::ICMP_EQ && Cmp->getPredicate() == CmpInst::ICMP_ULT) {
1218 if (!match(Cmp->getOperand(1), m_Power2(C)) ||
1219 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1220 return std::nullopt;
1221 }
1222
1223 // (icmp ne (lshr x, C), (lshr y, C)) gets optimized to:
1224 // (icmp ugt (xor x, y), (1 << C) - 1) so also look for that.
1225 else if (Pred == CmpInst::ICMP_NE &&
1226 Cmp->getPredicate() == CmpInst::ICMP_UGT) {
1227 if (!match(Cmp->getOperand(1), m_LowBitMask(C)) ||
1228 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1229 return std::nullopt;
1230 } else {
1231 return std::nullopt;
1232 }
1233
1234 unsigned From = Pred == CmpInst::ICMP_NE ? C->popcount() : C->countr_zero();
1235 Instruction *I = cast<Instruction>(Cmp->getOperand(0));
1236 return {{I->getOperand(OpNo), From, C->getBitWidth() - From}};
1237 };
1238
1239 std::optional<IntPart> L0 = GetMatchPart(Cmp0, 0);
1240 std::optional<IntPart> R0 = GetMatchPart(Cmp0, 1);
1241 std::optional<IntPart> L1 = GetMatchPart(Cmp1, 0);
1242 std::optional<IntPart> R1 = GetMatchPart(Cmp1, 1);
1243 if (!L0 || !R0 || !L1 || !R1)
1244 return nullptr;
1245
1246 // Make sure the LHS/RHS compare a part of the same value, possibly after
1247 // an operand swap.
1248 if (L0->From != L1->From || R0->From != R1->From) {
1249 if (L0->From != R1->From || R0->From != L1->From)
1250 return nullptr;
1251 std::swap(L1, R1);
1252 }
1253
1254 // Make sure the extracted parts are adjacent, canonicalizing to L0/R0 being
1255 // the low part and L1/R1 being the high part.
1256 if (L0->StartBit + L0->NumBits != L1->StartBit ||
1257 R0->StartBit + R0->NumBits != R1->StartBit) {
1258 if (L1->StartBit + L1->NumBits != L0->StartBit ||
1259 R1->StartBit + R1->NumBits != R0->StartBit)
1260 return nullptr;
1261 std::swap(L0, L1);
1262 std::swap(R0, R1);
1263 }
1264
1265 // We can simplify to a comparison of these larger parts of the integers.
1266 IntPart L = {L0->From, L0->StartBit, L0->NumBits + L1->NumBits};
1267 IntPart R = {R0->From, R0->StartBit, R0->NumBits + R1->NumBits};
1270 return Builder.CreateICmp(Pred, LValue, RValue);
1271}
1272
1273/// Reduce logic-of-compares with equality to a constant by substituting a
1274/// common operand with the constant. Callers are expected to call this with
1275/// Cmp0/Cmp1 switched to handle logic op commutativity.
1277 bool IsAnd, bool IsLogical,
1278 InstCombiner::BuilderTy &Builder,
1279 const SimplifyQuery &Q,
1280 Instruction &I) {
1281 // Match an equality compare with a non-poison constant as Cmp0.
1282 // Also, give up if the compare can be constant-folded to avoid looping.
1283 CmpPredicate Pred0;
1284 Value *X;
1285 Constant *C;
1286 if (!match(Cmp0, m_ICmp(Pred0, m_Value(X), m_Constant(C))) ||
1288 return nullptr;
1289 if ((IsAnd && Pred0 != ICmpInst::ICMP_EQ) ||
1290 (!IsAnd && Pred0 != ICmpInst::ICMP_NE))
1291 return nullptr;
1292
1293 // The other compare must include a common operand (X). Canonicalize the
1294 // common operand as operand 1 (Pred1 is swapped if the common operand was
1295 // operand 0).
1296 Value *Y;
1297 CmpPredicate Pred1;
1298 if (!match(Cmp1, m_c_ICmp(Pred1, m_Value(Y), m_Specific(X))))
1299 return nullptr;
1300
1301 // Replace variable with constant value equivalence to remove a variable use:
1302 // (X == C) && (Y Pred1 X) --> (X == C) && (Y Pred1 C)
1303 // (X != C) || (Y Pred1 X) --> (X != C) || (Y Pred1 C)
1304 // Can think of the 'or' substitution with the 'and' bool equivalent:
1305 // A || B --> A || (!A && B)
1306 Value *SubstituteCmp = simplifyICmpInst(Pred1, Y, C, Q);
1307 if (!SubstituteCmp) {
1308 // If we need to create a new instruction, require that the old compare can
1309 // be removed.
1310 if (!Cmp1->hasOneUse())
1311 return nullptr;
1312 SubstituteCmp = Builder.CreateICmp(Pred1, Y, C);
1313 }
1314 if (IsLogical) {
1315 Instruction *MDFrom =
1317 return IsAnd ? Builder.CreateLogicalAnd(Cmp0, SubstituteCmp, "", MDFrom)
1318 : Builder.CreateLogicalOr(Cmp0, SubstituteCmp, "", MDFrom);
1319 }
1320 return Builder.CreateBinOp(IsAnd ? Instruction::And : Instruction::Or, Cmp0,
1321 SubstituteCmp);
1322}
1323
1324/// Fold (icmp Pred1 V1, C1) & (icmp Pred2 V2, C2)
1325/// or (icmp Pred1 V1, C1) | (icmp Pred2 V2, C2)
1326/// into a single comparison using range-based reasoning.
1327/// NOTE: This is also used for logical and/or, must be poison-safe!
1328Value *InstCombinerImpl::foldAndOrOfICmpsUsingRanges(ICmpInst *ICmp1,
1329 ICmpInst *ICmp2,
1330 bool IsAnd) {
1331 // Return (V, CR) for a range check idiom V in CR.
1332 auto MatchExactRangeCheck =
1333 [](ICmpInst *ICmp) -> std::optional<std::pair<Value *, ConstantRange>> {
1334 const APInt *C;
1335 if (!match(ICmp->getOperand(1), m_APInt(C)))
1336 return std::nullopt;
1337 Value *LHS = ICmp->getOperand(0);
1338 CmpPredicate Pred = ICmp->getPredicate();
1339 Value *X;
1340 // Match (x & NegPow2) ==/!= C
1341 const APInt *Mask;
1342 if (ICmpInst::isEquality(Pred) &&
1344 C->countr_zero() >= Mask->countr_zero()) {
1345 ConstantRange CR(*C, *C - *Mask);
1346 if (Pred == ICmpInst::ICMP_NE)
1347 CR = CR.inverse();
1348 return std::make_pair(X, CR);
1349 }
1350 ConstantRange CR = ConstantRange::makeExactICmpRegion(Pred, *C);
1351 // Match (add X, C1) pred C
1352 // TODO: investigate whether we should apply the one-use check on m_AddLike.
1353 const APInt *C1;
1354 if (match(LHS, m_AddLike(m_Value(X), m_APInt(C1))))
1355 return std::make_pair(X, CR.subtract(*C1));
1356 return std::make_pair(LHS, CR);
1357 };
1358
1359 auto RC1 = MatchExactRangeCheck(ICmp1);
1360 if (!RC1)
1361 return nullptr;
1362
1363 auto RC2 = MatchExactRangeCheck(ICmp2);
1364 if (!RC2)
1365 return nullptr;
1366
1367 auto &[V1, CR1] = *RC1;
1368 auto &[V2, CR2] = *RC2;
1369 if (V1 != V2)
1370 return nullptr;
1371
1372 // For 'and', we use the De Morgan's Laws to simplify the implementation.
1373 if (IsAnd) {
1374 CR1 = CR1.inverse();
1375 CR2 = CR2.inverse();
1376 }
1377
1378 Type *Ty = V1->getType();
1379 Value *NewV = V1;
1380 std::optional<ConstantRange> CR = CR1.exactUnionWith(CR2);
1381 if (!CR) {
1382 if (!(ICmp1->hasOneUse() && ICmp2->hasOneUse()) || CR1.isWrappedSet() ||
1383 CR2.isWrappedSet())
1384 return nullptr;
1385
1386 // Check whether we have equal-size ranges that only differ by one bit.
1387 // In that case we can apply a mask to map one range onto the other.
1388 APInt LowerDiff = CR1.getLower() ^ CR2.getLower();
1389 APInt UpperDiff = (CR1.getUpper() - 1) ^ (CR2.getUpper() - 1);
1390 APInt CR1Size = CR1.getUpper() - CR1.getLower();
1391 if (!LowerDiff.isPowerOf2() || LowerDiff != UpperDiff ||
1392 CR1Size != CR2.getUpper() - CR2.getLower())
1393 return nullptr;
1394
1395 CR = CR1.getLower().ult(CR2.getLower()) ? CR1 : CR2;
1396 NewV = Builder.CreateAnd(NewV, ConstantInt::get(Ty, ~LowerDiff));
1397 }
1398
1399 if (IsAnd)
1400 CR = CR->inverse();
1401
1402 CmpInst::Predicate NewPred;
1403 APInt NewC, Offset;
1404 CR->getEquivalentICmp(NewPred, NewC, Offset);
1405
1406 if (Offset != 0)
1407 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
1408 return Builder.CreateICmp(NewPred, NewV, ConstantInt::get(Ty, NewC));
1409}
1410
1411/// Ignore all operations which only change the sign of a value, returning the
1412/// underlying magnitude value.
1414 match(Val, m_FNeg(m_Value(Val)));
1415 match(Val, m_FAbs(m_Value(Val)));
1416 match(Val, m_CopySign(m_Value(Val), m_Value()));
1417 return Val;
1418}
1419
1420/// Matches canonical form of isnan, fcmp ord x, 0
1424
1425/// Matches fcmp u__ x, +/-inf
1430
1431/// and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1432///
1433/// Clang emits this pattern for doing an isfinite check in __builtin_isnormal.
1435 FCmpInst *RHS) {
1436 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1437 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1438 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1439
1440 if (!matchIsNotNaN(PredL, LHS0, LHS1) ||
1441 !matchUnorderedInfCompare(PredR, RHS0, RHS1))
1442 return nullptr;
1443
1444 return Builder.CreateFCmpFMF(FCmpInst::getOrderedPredicate(PredR), RHS0, RHS1,
1446}
1447
1448Value *InstCombinerImpl::foldLogicOfFCmps(FCmpInst *LHS, FCmpInst *RHS,
1449 bool IsAnd, bool IsLogicalSelect) {
1450 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1451 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1452 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1453
1454 if (LHS0 == RHS1 && RHS0 == LHS1) {
1455 // Swap RHS operands to match LHS.
1456 PredR = FCmpInst::getSwappedPredicate(PredR);
1457 std::swap(RHS0, RHS1);
1458 }
1459
1460 // Simplify (fcmp cc0 x, y) & (fcmp cc1 x, y).
1461 // Suppose the relation between x and y is R, where R is one of
1462 // U(1000), L(0100), G(0010) or E(0001), and CC0 and CC1 are the bitmasks for
1463 // testing the desired relations.
1464 //
1465 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1466 // bool(R & CC0) && bool(R & CC1)
1467 // = bool((R & CC0) & (R & CC1))
1468 // = bool(R & (CC0 & CC1)) <= by re-association, commutation, and idempotency
1469 //
1470 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1471 // bool(R & CC0) || bool(R & CC1)
1472 // = bool((R & CC0) | (R & CC1))
1473 // = bool(R & (CC0 | CC1)) <= by reversed distribution (contribution? ;)
1474 if (LHS0 == RHS0 && LHS1 == RHS1) {
1475 unsigned FCmpCodeL = getFCmpCode(PredL);
1476 unsigned FCmpCodeR = getFCmpCode(PredR);
1477 unsigned NewPred = IsAnd ? FCmpCodeL & FCmpCodeR : FCmpCodeL | FCmpCodeR;
1478
1479 // Intersect the fast math flags.
1480 // TODO: We can union the fast math flags unless this is a logical select.
1481 return getFCmpValue(NewPred, LHS0, LHS1, Builder,
1483 }
1484
1485 // This transform is not valid for a logical select.
1486 if (!IsLogicalSelect &&
1487 ((PredL == FCmpInst::FCMP_ORD && PredR == FCmpInst::FCMP_ORD && IsAnd) ||
1488 (PredL == FCmpInst::FCMP_UNO && PredR == FCmpInst::FCMP_UNO &&
1489 !IsAnd))) {
1490 if (LHS0->getType() != RHS0->getType())
1491 return nullptr;
1492
1493 // FCmp canonicalization ensures that (fcmp ord/uno X, X) and
1494 // (fcmp ord/uno X, C) will be transformed to (fcmp X, +0.0).
1495 if (match(LHS1, m_PosZeroFP()) && match(RHS1, m_PosZeroFP())) {
1496 // Ignore the constants because they are obviously not NANs:
1497 // (fcmp ord x, 0.0) & (fcmp ord y, 0.0) -> (fcmp ord x, y)
1498 // (fcmp uno x, 0.0) | (fcmp uno y, 0.0) -> (fcmp uno x, y)
1499 return Builder.CreateFCmpFMF(PredL, LHS0, RHS0,
1501 }
1502 }
1503
1504 // This transform is not valid for a logical select.
1505 if (!IsLogicalSelect && IsAnd &&
1506 stripSignOnlyFPOps(LHS0) == stripSignOnlyFPOps(RHS0)) {
1507 // and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1508 // and (fcmp ord x, 0), (fcmp u* fabs(x), inf) -> fcmp o* x, inf
1510 return Left;
1512 return Right;
1513 }
1514
1515 // Turn at least two fcmps with constants into llvm.is.fpclass.
1516 //
1517 // If we can represent a combined value test with one class call, we can
1518 // potentially eliminate 4-6 instructions. If we can represent a test with a
1519 // single fcmp with fneg and fabs, that's likely a better canonical form.
1520 if (LHS->hasOneUse() && RHS->hasOneUse()) {
1521 auto [ClassValRHS, ClassMaskRHS] =
1522 fcmpToClassTest(PredR, *RHS->getFunction(), RHS0, RHS1);
1523 if (ClassValRHS) {
1524 auto [ClassValLHS, ClassMaskLHS] =
1525 fcmpToClassTest(PredL, *LHS->getFunction(), LHS0, LHS1);
1526 if (ClassValLHS == ClassValRHS) {
1527 unsigned CombinedMask = IsAnd ? (ClassMaskLHS & ClassMaskRHS)
1528 : (ClassMaskLHS | ClassMaskRHS);
1529 return Builder.CreateIntrinsic(
1530 Intrinsic::is_fpclass, {ClassValLHS->getType()},
1531 {ClassValLHS, Builder.getInt32(CombinedMask)});
1532 }
1533 }
1534 }
1535
1536 // Canonicalize the range check idiom:
1537 // and (fcmp olt/ole/ult/ule x, C), (fcmp ogt/oge/ugt/uge x, -C)
1538 // --> fabs(x) olt/ole/ult/ule C
1539 // or (fcmp ogt/oge/ugt/uge x, C), (fcmp olt/ole/ult/ule x, -C)
1540 // --> fabs(x) ogt/oge/ugt/uge C
1541 // TODO: Generalize to handle a negated variable operand?
1542 const APFloat *LHSC, *RHSC;
1543 if (LHS0 == RHS0 && LHS->hasOneUse() && RHS->hasOneUse() &&
1544 FCmpInst::getSwappedPredicate(PredL) == PredR &&
1545 match(LHS1, m_APFloatAllowPoison(LHSC)) &&
1546 match(RHS1, m_APFloatAllowPoison(RHSC)) &&
1547 LHSC->bitwiseIsEqual(neg(*RHSC))) {
1548 auto IsLessThanOrLessEqual = [](FCmpInst::Predicate Pred) {
1549 switch (Pred) {
1550 case FCmpInst::FCMP_OLT:
1551 case FCmpInst::FCMP_OLE:
1552 case FCmpInst::FCMP_ULT:
1553 case FCmpInst::FCMP_ULE:
1554 return true;
1555 default:
1556 return false;
1557 }
1558 };
1559 if (IsLessThanOrLessEqual(IsAnd ? PredR : PredL)) {
1560 std::swap(LHSC, RHSC);
1561 std::swap(PredL, PredR);
1562 }
1563 if (IsLessThanOrLessEqual(IsAnd ? PredL : PredR)) {
1564 FastMathFlags NewFlag = LHS->getFastMathFlags();
1565 if (!IsLogicalSelect)
1566 NewFlag |= RHS->getFastMathFlags();
1567
1568 Value *FAbs =
1569 Builder.CreateUnaryIntrinsic(Intrinsic::fabs, LHS0, NewFlag);
1570 return Builder.CreateFCmpFMF(
1571 PredL, FAbs, ConstantFP::get(LHS0->getType(), *LHSC), NewFlag);
1572 }
1573 }
1574
1575 return nullptr;
1576}
1577
1578/// Match an fcmp against a special value that performs a test possible by
1579/// llvm.is.fpclass.
1580static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal,
1581 uint64_t &ClassMask) {
1582 auto *FCmp = dyn_cast<FCmpInst>(Op);
1583 if (!FCmp || !FCmp->hasOneUse())
1584 return false;
1585
1586 std::tie(ClassVal, ClassMask) =
1587 fcmpToClassTest(FCmp->getPredicate(), *FCmp->getParent()->getParent(),
1588 FCmp->getOperand(0), FCmp->getOperand(1));
1589 return ClassVal != nullptr;
1590}
1591
1592/// or (is_fpclass x, mask0), (is_fpclass x, mask1)
1593/// -> is_fpclass x, (mask0 | mask1)
1594/// and (is_fpclass x, mask0), (is_fpclass x, mask1)
1595/// -> is_fpclass x, (mask0 & mask1)
1596/// xor (is_fpclass x, mask0), (is_fpclass x, mask1)
1597/// -> is_fpclass x, (mask0 ^ mask1)
1598Instruction *InstCombinerImpl::foldLogicOfIsFPClass(BinaryOperator &BO,
1599 Value *Op0, Value *Op1) {
1600 Value *ClassVal0 = nullptr;
1601 Value *ClassVal1 = nullptr;
1602 uint64_t ClassMask0, ClassMask1;
1603
1604 // Restrict to folding one fcmp into one is.fpclass for now, don't introduce a
1605 // new class.
1606 //
1607 // TODO: Support forming is.fpclass out of 2 separate fcmps when codegen is
1608 // better.
1609
1610 bool IsLHSClass =
1612 m_Value(ClassVal0), m_ConstantInt(ClassMask0))));
1613 bool IsRHSClass =
1615 m_Value(ClassVal1), m_ConstantInt(ClassMask1))));
1616 if ((((IsLHSClass || matchIsFPClassLikeFCmp(Op0, ClassVal0, ClassMask0)) &&
1617 (IsRHSClass || matchIsFPClassLikeFCmp(Op1, ClassVal1, ClassMask1)))) &&
1618 ClassVal0 == ClassVal1) {
1619 unsigned NewClassMask;
1620 switch (BO.getOpcode()) {
1621 case Instruction::And:
1622 NewClassMask = ClassMask0 & ClassMask1;
1623 break;
1624 case Instruction::Or:
1625 NewClassMask = ClassMask0 | ClassMask1;
1626 break;
1627 case Instruction::Xor:
1628 NewClassMask = ClassMask0 ^ ClassMask1;
1629 break;
1630 default:
1631 llvm_unreachable("not a binary logic operator");
1632 }
1633
1634 if (IsLHSClass) {
1635 auto *II = cast<IntrinsicInst>(Op0);
1636 II->setArgOperand(
1637 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1638 return replaceInstUsesWith(BO, II);
1639 }
1640
1641 if (IsRHSClass) {
1642 auto *II = cast<IntrinsicInst>(Op1);
1643 II->setArgOperand(
1644 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1645 return replaceInstUsesWith(BO, II);
1646 }
1647
1648 CallInst *NewClass =
1649 Builder.CreateIntrinsic(Intrinsic::is_fpclass, {ClassVal0->getType()},
1650 {ClassVal0, Builder.getInt32(NewClassMask)});
1651 return replaceInstUsesWith(BO, NewClass);
1652 }
1653
1654 return nullptr;
1655}
1656
1657/// Look for the pattern that conditionally negates a value via math operations:
1658/// cond.splat = sext i1 cond
1659/// sub = add cond.splat, x
1660/// xor = xor sub, cond.splat
1661/// and rewrite it to do the same, but via logical operations:
1662/// value.neg = sub 0, value
1663/// cond = select i1 neg, value.neg, value
1664Instruction *InstCombinerImpl::canonicalizeConditionalNegationViaMathToSelect(
1665 BinaryOperator &I) {
1666 assert(I.getOpcode() == BinaryOperator::Xor && "Only for xor!");
1667 Value *Cond, *X;
1668 // As per complexity ordering, `xor` is not commutative here.
1669 if (!match(&I, m_c_BinOp(m_OneUse(m_Value()), m_Value())) ||
1670 !match(I.getOperand(1), m_SExt(m_Value(Cond))) ||
1671 !Cond->getType()->isIntOrIntVectorTy(1) ||
1672 !match(I.getOperand(0), m_c_Add(m_SExt(m_Specific(Cond)), m_Value(X))))
1673 return nullptr;
1674 return createSelectInstWithUnknownProfile(
1675 Cond, Builder.CreateNeg(X, X->getName() + ".neg"), X);
1676}
1677
1678/// This a limited reassociation for a special case (see above) where we are
1679/// checking if two values are either both NAN (unordered) or not-NAN (ordered).
1680/// This could be handled more generally in '-reassociation', but it seems like
1681/// an unlikely pattern for a large number of logic ops and fcmps.
1683 InstCombiner::BuilderTy &Builder) {
1684 Instruction::BinaryOps Opcode = BO.getOpcode();
1685 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1686 "Expecting and/or op for fcmp transform");
1687
1688 // There are 4 commuted variants of the pattern. Canonicalize operands of this
1689 // logic op so an fcmp is operand 0 and a matching logic op is operand 1.
1690 Value *Op0 = BO.getOperand(0), *Op1 = BO.getOperand(1), *X;
1691 if (match(Op1, m_FCmp(m_Value(), m_AnyZeroFP())))
1692 std::swap(Op0, Op1);
1693
1694 // Match inner binop and the predicate for combining 2 NAN checks into 1.
1695 Value *BO10, *BO11;
1696 FCmpInst::Predicate NanPred = Opcode == Instruction::And ? FCmpInst::FCMP_ORD
1698 if (!match(Op0, m_SpecificFCmp(NanPred, m_Value(X), m_AnyZeroFP())) ||
1699 !match(Op1, m_BinOp(Opcode, m_Value(BO10), m_Value(BO11))))
1700 return nullptr;
1701
1702 // The inner logic op must have a matching fcmp operand.
1703 Value *Y;
1704 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1705 X->getType() != Y->getType())
1706 std::swap(BO10, BO11);
1707
1708 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1709 X->getType() != Y->getType())
1710 return nullptr;
1711
1712 // and (fcmp ord X, 0), (and (fcmp ord Y, 0), Z) --> and (fcmp ord X, Y), Z
1713 // or (fcmp uno X, 0), (or (fcmp uno Y, 0), Z) --> or (fcmp uno X, Y), Z
1714 // Intersect FMF from the 2 source fcmps.
1715 Value *NewFCmp =
1716 Builder.CreateFCmpFMF(NanPred, X, Y, FMFSource::intersect(Op0, BO10));
1717 return BinaryOperator::Create(Opcode, NewFCmp, BO11);
1718}
1719
1720/// Match variations of De Morgan's Laws:
1721/// (~A & ~B) == (~(A | B))
1722/// (~A | ~B) == (~(A & B))
1724 InstCombiner &IC) {
1725 const Instruction::BinaryOps Opcode = I.getOpcode();
1726 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1727 "Trying to match De Morgan's Laws with something other than and/or");
1728
1729 // Flip the logic operation.
1730 const Instruction::BinaryOps FlippedOpcode =
1731 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
1732
1733 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1734 Value *A, *B;
1735 if (match(Op0, m_OneUse(m_Not(m_Value(A)))) &&
1736 match(Op1, m_OneUse(m_Not(m_Value(B)))) &&
1737 !IC.isFreeToInvert(A, A->hasOneUse()) &&
1738 !IC.isFreeToInvert(B, B->hasOneUse())) {
1739 Value *AndOr =
1740 IC.Builder.CreateBinOp(FlippedOpcode, A, B, I.getName() + ".demorgan");
1741 return BinaryOperator::CreateNot(AndOr);
1742 }
1743
1744 // The 'not' ops may require reassociation.
1745 // (A & ~B) & ~C --> A & ~(B | C)
1746 // (~B & A) & ~C --> A & ~(B | C)
1747 // (A | ~B) | ~C --> A | ~(B & C)
1748 // (~B | A) | ~C --> A | ~(B & C)
1749 Value *C;
1750 if (match(Op0, m_OneUse(m_c_BinOp(Opcode, m_Value(A), m_Not(m_Value(B))))) &&
1751 match(Op1, m_Not(m_Value(C)))) {
1752 Value *FlippedBO = IC.Builder.CreateBinOp(FlippedOpcode, B, C);
1753 return BinaryOperator::Create(Opcode, A, IC.Builder.CreateNot(FlippedBO));
1754 }
1755
1756 return nullptr;
1757}
1758
1759bool InstCombinerImpl::shouldOptimizeCast(CastInst *CI) {
1760 Value *CastSrc = CI->getOperand(0);
1761
1762 // Noop casts and casts of constants should be eliminated trivially.
1763 if (CI->getSrcTy() == CI->getDestTy() || isa<Constant>(CastSrc))
1764 return false;
1765
1766 // If this cast is paired with another cast that can be eliminated, we prefer
1767 // to have it eliminated.
1768 if (const auto *PrecedingCI = dyn_cast<CastInst>(CastSrc))
1769 if (isEliminableCastPair(PrecedingCI, CI))
1770 return false;
1771
1772 return true;
1773}
1774
1775/// Fold {and,or,xor} (cast X), C.
1777 InstCombinerImpl &IC) {
1779 if (!C)
1780 return nullptr;
1781
1782 auto LogicOpc = Logic.getOpcode();
1783 Type *DestTy = Logic.getType();
1784 Type *SrcTy = Cast->getSrcTy();
1785
1786 // Move the logic operation ahead of a zext or sext if the constant is
1787 // unchanged in the smaller source type. Performing the logic in a smaller
1788 // type may provide more information to later folds, and the smaller logic
1789 // instruction may be cheaper (particularly in the case of vectors).
1790 Value *X;
1791 auto &DL = IC.getDataLayout();
1792 if (match(Cast, m_OneUse(m_ZExt(m_Value(X))))) {
1793 PreservedCastFlags Flags;
1794 if (Constant *TruncC = getLosslessUnsignedTrunc(C, SrcTy, DL, &Flags)) {
1795 // LogicOpc (zext X), C --> zext (LogicOpc X, C)
1796 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1797 auto *ZExt = new ZExtInst(NewOp, DestTy);
1798 ZExt->setNonNeg(Flags.NNeg);
1799 ZExt->andIRFlags(Cast);
1800 return ZExt;
1801 }
1802 }
1803
1804 if (match(Cast, m_OneUse(m_SExtLike(m_Value(X))))) {
1805 if (Constant *TruncC = getLosslessSignedTrunc(C, SrcTy, DL)) {
1806 // LogicOpc (sext X), C --> sext (LogicOpc X, C)
1807 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1808 return new SExtInst(NewOp, DestTy);
1809 }
1810 }
1811
1812 return nullptr;
1813}
1814
1815/// Fold {and,or,xor} (cast X), Y.
1816Instruction *InstCombinerImpl::foldCastedBitwiseLogic(BinaryOperator &I) {
1817 auto LogicOpc = I.getOpcode();
1818 assert(I.isBitwiseLogicOp() && "Unexpected opcode for bitwise logic folding");
1819
1820 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1821
1822 // fold bitwise(A >> BW - 1, zext(icmp)) (BW is the scalar bits of the
1823 // type of A)
1824 // -> bitwise(zext(A < 0), zext(icmp))
1825 // -> zext(bitwise(A < 0, icmp))
1826 auto FoldBitwiseICmpZeroWithICmp = [&](Value *Op0,
1827 Value *Op1) -> Instruction * {
1828 Value *A;
1829 bool IsMatched =
1830 match(Op0,
1832 m_Value(A),
1833 m_SpecificInt(Op0->getType()->getScalarSizeInBits() - 1)))) &&
1834 match(Op1, m_OneUse(m_ZExt(m_ICmp(m_Value(), m_Value()))));
1835
1836 if (!IsMatched)
1837 return nullptr;
1838
1839 auto *ICmpL =
1840 Builder.CreateICmpSLT(A, Constant::getNullValue(A->getType()));
1841 auto *ICmpR = cast<ZExtInst>(Op1)->getOperand(0);
1842 auto *BitwiseOp = Builder.CreateBinOp(LogicOpc, ICmpL, ICmpR);
1843
1844 return new ZExtInst(BitwiseOp, Op0->getType());
1845 };
1846
1847 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op0, Op1))
1848 return Ret;
1849
1850 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op1, Op0))
1851 return Ret;
1852
1853 CastInst *Cast0 = dyn_cast<CastInst>(Op0);
1854 if (!Cast0)
1855 return nullptr;
1856
1857 // This must be a cast from an integer or integer vector source type to allow
1858 // transformation of the logic operation to the source type.
1859 Type *DestTy = I.getType();
1860 Type *SrcTy = Cast0->getSrcTy();
1861 if (!SrcTy->isIntOrIntVectorTy())
1862 return nullptr;
1863
1864 if (Instruction *Ret = foldLogicCastConstant(I, Cast0, *this))
1865 return Ret;
1866
1867 CastInst *Cast1 = dyn_cast<CastInst>(Op1);
1868 if (!Cast1)
1869 return nullptr;
1870
1871 // Both operands of the logic operation are casts. The casts must be the
1872 // same kind for reduction.
1873 Instruction::CastOps CastOpcode = Cast0->getOpcode();
1874 if (CastOpcode != Cast1->getOpcode())
1875 return nullptr;
1876
1877 // Can't fold it profitably if no one of casts has one use.
1878 if (!Cast0->hasOneUse() && !Cast1->hasOneUse())
1879 return nullptr;
1880
1881 Value *X, *Y;
1882 if (match(Cast0, m_ZExtOrSExt(m_Value(X))) &&
1883 match(Cast1, m_ZExtOrSExt(m_Value(Y)))) {
1884 // Cast the narrower source to the wider source type.
1885 unsigned XNumBits = X->getType()->getScalarSizeInBits();
1886 unsigned YNumBits = Y->getType()->getScalarSizeInBits();
1887 if (XNumBits != YNumBits) {
1888 // Cast the narrower source to the wider source type only if both of casts
1889 // have one use to avoid creating an extra instruction.
1890 if (!Cast0->hasOneUse() || !Cast1->hasOneUse())
1891 return nullptr;
1892
1893 // If the source types do not match, but the casts are matching extends,
1894 // we can still narrow the logic op.
1895 if (XNumBits < YNumBits) {
1896 X = Builder.CreateCast(CastOpcode, X, Y->getType());
1897 } else if (YNumBits < XNumBits) {
1898 Y = Builder.CreateCast(CastOpcode, Y, X->getType());
1899 }
1900 }
1901
1902 // Do the logic op in the intermediate width, then widen more.
1903 Value *NarrowLogic = Builder.CreateBinOp(LogicOpc, X, Y, I.getName());
1904 auto *Disjoint = dyn_cast<PossiblyDisjointInst>(&I);
1905 auto *NewDisjoint = dyn_cast<PossiblyDisjointInst>(NarrowLogic);
1906 if (Disjoint && NewDisjoint)
1907 NewDisjoint->setIsDisjoint(Disjoint->isDisjoint());
1908 return CastInst::Create(CastOpcode, NarrowLogic, DestTy);
1909 }
1910
1911 // If the src type of casts are different, give up for other cast opcodes.
1912 if (SrcTy != Cast1->getSrcTy())
1913 return nullptr;
1914
1915 Value *Cast0Src = Cast0->getOperand(0);
1916 Value *Cast1Src = Cast1->getOperand(0);
1917
1918 // fold logic(cast(A), cast(B)) -> cast(logic(A, B))
1919 if (shouldOptimizeCast(Cast0) && shouldOptimizeCast(Cast1)) {
1920 Value *NewOp = Builder.CreateBinOp(LogicOpc, Cast0Src, Cast1Src,
1921 I.getName());
1922 return CastInst::Create(CastOpcode, NewOp, DestTy);
1923 }
1924
1925 return nullptr;
1926}
1927
1929 InstCombiner::BuilderTy &Builder) {
1930 assert(I.getOpcode() == Instruction::And);
1931 Value *Op0 = I.getOperand(0);
1932 Value *Op1 = I.getOperand(1);
1933 Value *A, *B;
1934
1935 // Operand complexity canonicalization guarantees that the 'or' is Op0.
1936 // (A | B) & ~(A & B) --> A ^ B
1937 // (A | B) & ~(B & A) --> A ^ B
1938 if (match(&I, m_BinOp(m_Or(m_Value(A), m_Value(B)),
1940 return BinaryOperator::CreateXor(A, B);
1941
1942 // (A | ~B) & (~A | B) --> ~(A ^ B)
1943 // (A | ~B) & (B | ~A) --> ~(A ^ B)
1944 // (~B | A) & (~A | B) --> ~(A ^ B)
1945 // (~B | A) & (B | ~A) --> ~(A ^ B)
1946 if (Op0->hasOneUse() || Op1->hasOneUse())
1949 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1950
1951 return nullptr;
1952}
1953
1955 InstCombiner::BuilderTy &Builder) {
1956 assert(I.getOpcode() == Instruction::Or);
1957 Value *Op0 = I.getOperand(0);
1958 Value *Op1 = I.getOperand(1);
1959 Value *A, *B;
1960
1961 // Operand complexity canonicalization guarantees that the 'and' is Op0.
1962 // (A & B) | ~(A | B) --> ~(A ^ B)
1963 // (A & B) | ~(B | A) --> ~(A ^ B)
1964 if (Op0->hasOneUse() || Op1->hasOneUse())
1965 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
1967 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1968
1969 // Operand complexity canonicalization guarantees that the 'xor' is Op0.
1970 // (A ^ B) | ~(A | B) --> ~(A & B)
1971 // (A ^ B) | ~(B | A) --> ~(A & B)
1972 if (Op0->hasOneUse() || Op1->hasOneUse())
1973 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
1975 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
1976
1977 // (A & ~B) | (~A & B) --> A ^ B
1978 // (A & ~B) | (B & ~A) --> A ^ B
1979 // (~B & A) | (~A & B) --> A ^ B
1980 // (~B & A) | (B & ~A) --> A ^ B
1981 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
1983 return BinaryOperator::CreateXor(A, B);
1984
1985 return nullptr;
1986}
1987
1988/// Return true if a constant shift amount is always less than the specified
1989/// bit-width. If not, the shift could create poison in the narrower type.
1990static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth) {
1991 APInt Threshold(C->getType()->getScalarSizeInBits(), BitWidth);
1992 return match(C, m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, Threshold));
1993}
1994
1995/// Try to use narrower ops (sink zext ops) for an 'and' with binop operand and
1996/// a common zext operand: and (binop (zext X), C), (zext X).
1997Instruction *InstCombinerImpl::narrowMaskedBinOp(BinaryOperator &And) {
1998 // This transform could also apply to {or, and, xor}, but there are better
1999 // folds for those cases, so we don't expect those patterns here. AShr is not
2000 // handled because it should always be transformed to LShr in this sequence.
2001 // The subtract transform is different because it has a constant on the left.
2002 // Add/mul commute the constant to RHS; sub with constant RHS becomes add.
2003 Value *Op0 = And.getOperand(0), *Op1 = And.getOperand(1);
2004 Constant *C;
2005 if (!match(Op0, m_OneUse(m_Add(m_Specific(Op1), m_Constant(C)))) &&
2006 !match(Op0, m_OneUse(m_Mul(m_Specific(Op1), m_Constant(C)))) &&
2007 !match(Op0, m_OneUse(m_LShr(m_Specific(Op1), m_Constant(C)))) &&
2008 !match(Op0, m_OneUse(m_Shl(m_Specific(Op1), m_Constant(C)))) &&
2009 !match(Op0, m_OneUse(m_Sub(m_Constant(C), m_Specific(Op1)))))
2010 return nullptr;
2011
2012 Value *X;
2013 if (!match(Op1, m_ZExt(m_Value(X))) || Op1->hasNUsesOrMore(3))
2014 return nullptr;
2015
2016 Type *Ty = And.getType();
2017 if (!isa<VectorType>(Ty) && !shouldChangeType(Ty, X->getType()))
2018 return nullptr;
2019
2020 // If we're narrowing a shift, the shift amount must be safe (less than the
2021 // width) in the narrower type. If the shift amount is greater, instsimplify
2022 // usually handles that case, but we can't guarantee/assert it.
2024 if (Opc == Instruction::LShr || Opc == Instruction::Shl)
2025 if (!canNarrowShiftAmt(C, X->getType()->getScalarSizeInBits()))
2026 return nullptr;
2027
2028 // and (sub C, (zext X)), (zext X) --> zext (and (sub C', X), X)
2029 // and (binop (zext X), C), (zext X) --> zext (and (binop X, C'), X)
2030 Value *NewC = ConstantExpr::getTrunc(C, X->getType());
2031 Value *NewBO = Opc == Instruction::Sub ? Builder.CreateBinOp(Opc, NewC, X)
2032 : Builder.CreateBinOp(Opc, X, NewC);
2033 return new ZExtInst(Builder.CreateAnd(NewBO, X), Ty);
2034}
2035
2036/// Try folding relatively complex patterns for both And and Or operations
2037/// with all And and Or swapped.
2039 InstCombiner::BuilderTy &Builder) {
2040 const Instruction::BinaryOps Opcode = I.getOpcode();
2041 assert(Opcode == Instruction::And || Opcode == Instruction::Or);
2042
2043 // Flip the logic operation.
2044 const Instruction::BinaryOps FlippedOpcode =
2045 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
2046
2047 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2048 Value *A, *B, *C, *X, *Y, *Dummy;
2049
2050 // Match following expressions:
2051 // (~(A | B) & C)
2052 // (~(A & B) | C)
2053 // Captures X = ~(A | B) or ~(A & B)
2054 const auto matchNotOrAnd =
2055 [Opcode, FlippedOpcode](Value *Op, auto m_A, auto m_B, auto m_C,
2056 Value *&X, bool CountUses = false) -> bool {
2057 if (CountUses && !Op->hasOneUse())
2058 return false;
2059
2060 if (match(Op,
2061 m_c_BinOp(FlippedOpcode,
2062 m_Value(X, m_Not(m_c_BinOp(Opcode, m_A, m_B))), m_C)))
2063 return !CountUses || X->hasOneUse();
2064
2065 return false;
2066 };
2067
2068 // (~(A | B) & C) | ... --> ...
2069 // (~(A & B) | C) & ... --> ...
2070 // TODO: One use checks are conservative. We just need to check that a total
2071 // number of multiple used values does not exceed reduction
2072 // in operations.
2073 if (matchNotOrAnd(Op0, m_Value(A), m_Value(B), m_Value(C), X)) {
2074 // (~(A | B) & C) | (~(A | C) & B) --> (B ^ C) & ~A
2075 // (~(A & B) | C) & (~(A & C) | B) --> ~((B ^ C) & A)
2076 if (matchNotOrAnd(Op1, m_Specific(A), m_Specific(C), m_Specific(B), Dummy,
2077 true)) {
2078 Value *Xor = Builder.CreateXor(B, C);
2079 return (Opcode == Instruction::Or)
2080 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(A))
2081 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, A));
2082 }
2083
2084 // (~(A | B) & C) | (~(B | C) & A) --> (A ^ C) & ~B
2085 // (~(A & B) | C) & (~(B & C) | A) --> ~((A ^ C) & B)
2086 if (matchNotOrAnd(Op1, m_Specific(B), m_Specific(C), m_Specific(A), Dummy,
2087 true)) {
2088 Value *Xor = Builder.CreateXor(A, C);
2089 return (Opcode == Instruction::Or)
2090 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(B))
2091 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, B));
2092 }
2093
2094 // (~(A | B) & C) | ~(A | C) --> ~((B & C) | A)
2095 // (~(A & B) | C) & ~(A & C) --> ~((B | C) & A)
2096 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2097 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2098 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2099 Opcode, Builder.CreateBinOp(FlippedOpcode, B, C), A));
2100
2101 // (~(A | B) & C) | ~(B | C) --> ~((A & C) | B)
2102 // (~(A & B) | C) & ~(B & C) --> ~((A | C) & B)
2103 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2104 m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)))))))
2105 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2106 Opcode, Builder.CreateBinOp(FlippedOpcode, A, C), B));
2107
2108 // (~(A | B) & C) | ~(C | (A ^ B)) --> ~((A | B) & (C | (A ^ B)))
2109 // Note, the pattern with swapped and/or is not handled because the
2110 // result is more undefined than a source:
2111 // (~(A & B) | C) & ~(C & (A ^ B)) --> (A ^ B ^ C) | ~(A | C) is invalid.
2112 if (Opcode == Instruction::Or && Op0->hasOneUse() &&
2113 match(Op1,
2115 Y, m_c_BinOp(Opcode, m_Specific(C),
2116 m_c_Xor(m_Specific(A), m_Specific(B)))))))) {
2117 // X = ~(A | B)
2118 // Y = (C | (A ^ B)
2119 Value *Or = cast<BinaryOperator>(X)->getOperand(0);
2120 return BinaryOperator::CreateNot(Builder.CreateAnd(Or, Y));
2121 }
2122 }
2123
2124 // (~A & B & C) | ... --> ...
2125 // (~A | B | C) | ... --> ...
2126 // TODO: One use checks are conservative. We just need to check that a total
2127 // number of multiple used values does not exceed reduction
2128 // in operations.
2129 if (match(Op0,
2130 m_OneUse(m_c_BinOp(FlippedOpcode,
2131 m_BinOp(FlippedOpcode, m_Value(B), m_Value(C)),
2132 m_Value(X, m_Not(m_Value(A)))))) ||
2133 match(Op0, m_OneUse(m_c_BinOp(FlippedOpcode,
2134 m_c_BinOp(FlippedOpcode, m_Value(C),
2135 m_Value(X, m_Not(m_Value(A)))),
2136 m_Value(B))))) {
2137 // X = ~A
2138 // (~A & B & C) | ~(A | B | C) --> ~(A | (B ^ C))
2139 // (~A | B | C) & ~(A & B & C) --> (~A | (B ^ C))
2140 if (match(Op1, m_OneUse(m_Not(m_c_BinOp(
2141 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)),
2142 m_Specific(C))))) ||
2144 Opcode, m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)),
2145 m_Specific(A))))) ||
2147 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)),
2148 m_Specific(B)))))) {
2149 Value *Xor = Builder.CreateXor(B, C);
2150 return (Opcode == Instruction::Or)
2151 ? BinaryOperator::CreateNot(Builder.CreateOr(Xor, A))
2152 : BinaryOperator::CreateOr(Xor, X);
2153 }
2154
2155 // (~A & B & C) | ~(A | B) --> (C | ~B) & ~A
2156 // (~A | B | C) & ~(A & B) --> (C & ~B) | ~A
2157 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2158 m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)))))))
2160 FlippedOpcode, Builder.CreateBinOp(Opcode, C, Builder.CreateNot(B)),
2161 X);
2162
2163 // (~A & B & C) | ~(A | C) --> (B | ~C) & ~A
2164 // (~A | B | C) & ~(A & C) --> (B & ~C) | ~A
2165 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2166 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2168 FlippedOpcode, Builder.CreateBinOp(Opcode, B, Builder.CreateNot(C)),
2169 X);
2170 }
2171
2172 return nullptr;
2173}
2174
2175/// Try to reassociate a pair of binops so that values with one use only are
2176/// part of the same instruction. This may enable folds that are limited with
2177/// multi-use restrictions and makes it more likely to match other patterns that
2178/// are looking for a common operand.
2180 InstCombinerImpl::BuilderTy &Builder) {
2181 Instruction::BinaryOps Opcode = BO.getOpcode();
2182 Value *X, *Y, *Z;
2183 if (match(&BO,
2184 m_c_BinOp(Opcode, m_OneUse(m_BinOp(Opcode, m_Value(X), m_Value(Y))),
2185 m_OneUse(m_Value(Z))))) {
2186 if (!isa<Constant>(X) && !isa<Constant>(Y) && !isa<Constant>(Z)) {
2187 // (X op Y) op Z --> (Y op Z) op X
2188 if (!X->hasOneUse()) {
2189 Value *YZ = Builder.CreateBinOp(Opcode, Y, Z);
2190 return BinaryOperator::Create(Opcode, YZ, X);
2191 }
2192 // (X op Y) op Z --> (X op Z) op Y
2193 if (!Y->hasOneUse()) {
2194 Value *XZ = Builder.CreateBinOp(Opcode, X, Z);
2195 return BinaryOperator::Create(Opcode, XZ, Y);
2196 }
2197 }
2198 }
2199
2200 return nullptr;
2201}
2202
2203// Match
2204// (X + C2) | C
2205// (X + C2) ^ C
2206// (X + C2) & C
2207// and convert to do the bitwise logic first:
2208// (X | C) + C2
2209// (X ^ C) + C2
2210// (X & C) + C2
2211// iff bits affected by logic op are lower than last bit affected by math op
2213 InstCombiner::BuilderTy &Builder) {
2214 Type *Ty = I.getType();
2215 Instruction::BinaryOps OpC = I.getOpcode();
2216 Value *Op0 = I.getOperand(0);
2217 Value *Op1 = I.getOperand(1);
2218 Value *X;
2219 const APInt *C, *C2;
2220
2221 if (!(match(Op0, m_OneUse(m_Add(m_Value(X), m_APInt(C2)))) &&
2222 match(Op1, m_APInt(C))))
2223 return nullptr;
2224
2225 unsigned Width = Ty->getScalarSizeInBits();
2226 unsigned LastOneMath = Width - C2->countr_zero();
2227
2228 switch (OpC) {
2229 case Instruction::And:
2230 if (C->countl_one() < LastOneMath)
2231 return nullptr;
2232 break;
2233 case Instruction::Xor:
2234 case Instruction::Or:
2235 if (C->countl_zero() < LastOneMath)
2236 return nullptr;
2237 break;
2238 default:
2239 llvm_unreachable("Unexpected BinaryOp!");
2240 }
2241
2242 Value *NewBinOp = Builder.CreateBinOp(OpC, X, ConstantInt::get(Ty, *C));
2243 return BinaryOperator::CreateWithCopiedFlags(Instruction::Add, NewBinOp,
2244 ConstantInt::get(Ty, *C2), Op0);
2245}
2246
2247// binop(shift(ShiftedC1, ShAmt), shift(ShiftedC2, add(ShAmt, AddC))) ->
2248// shift(binop(ShiftedC1, shift(ShiftedC2, AddC)), ShAmt)
2249// where both shifts are the same and AddC is a valid shift amount.
2250Instruction *InstCombinerImpl::foldBinOpOfDisplacedShifts(BinaryOperator &I) {
2251 assert((I.isBitwiseLogicOp() || I.getOpcode() == Instruction::Add) &&
2252 "Unexpected opcode");
2253
2254 Value *ShAmt;
2255 Constant *ShiftedC1, *ShiftedC2, *AddC;
2256 Type *Ty = I.getType();
2257 unsigned BitWidth = Ty->getScalarSizeInBits();
2258 if (!match(&I, m_c_BinOp(m_Shift(m_ImmConstant(ShiftedC1), m_Value(ShAmt)),
2259 m_Shift(m_ImmConstant(ShiftedC2),
2260 m_AddLike(m_Deferred(ShAmt),
2261 m_ImmConstant(AddC))))))
2262 return nullptr;
2263
2264 // Make sure the add constant is a valid shift amount.
2265 if (!match(AddC,
2267 return nullptr;
2268
2269 // Avoid constant expressions.
2270 auto *Op0Inst = dyn_cast<Instruction>(I.getOperand(0));
2271 auto *Op1Inst = dyn_cast<Instruction>(I.getOperand(1));
2272 if (!Op0Inst || !Op1Inst)
2273 return nullptr;
2274
2275 // Both shifts must be the same.
2276 Instruction::BinaryOps ShiftOp =
2277 static_cast<Instruction::BinaryOps>(Op0Inst->getOpcode());
2278 if (ShiftOp != Op1Inst->getOpcode())
2279 return nullptr;
2280
2281 // For adds, only left shifts are supported.
2282 if (I.getOpcode() == Instruction::Add && ShiftOp != Instruction::Shl)
2283 return nullptr;
2284
2285 Value *NewC = Builder.CreateBinOp(
2286 I.getOpcode(), ShiftedC1, Builder.CreateBinOp(ShiftOp, ShiftedC2, AddC));
2287 return BinaryOperator::Create(ShiftOp, NewC, ShAmt);
2288}
2289
2290// Fold and/or/xor with two equal intrinsic IDs:
2291// bitwise(fshl (A, B, ShAmt), fshl(C, D, ShAmt))
2292// -> fshl(bitwise(A, C), bitwise(B, D), ShAmt)
2293// bitwise(fshr (A, B, ShAmt), fshr(C, D, ShAmt))
2294// -> fshr(bitwise(A, C), bitwise(B, D), ShAmt)
2295// bitwise(bswap(A), bswap(B)) -> bswap(bitwise(A, B))
2296// bitwise(bswap(A), C) -> bswap(bitwise(A, bswap(C)))
2297// bitwise(bitreverse(A), bitreverse(B)) -> bitreverse(bitwise(A, B))
2298// bitwise(bitreverse(A), C) -> bitreverse(bitwise(A, bitreverse(C)))
2299static Instruction *
2301 InstCombiner::BuilderTy &Builder) {
2302 assert(I.isBitwiseLogicOp() && "Should and/or/xor");
2303 if (!I.getOperand(0)->hasOneUse())
2304 return nullptr;
2305 IntrinsicInst *X = dyn_cast<IntrinsicInst>(I.getOperand(0));
2306 if (!X)
2307 return nullptr;
2308
2309 IntrinsicInst *Y = dyn_cast<IntrinsicInst>(I.getOperand(1));
2310 if (Y && (!Y->hasOneUse() || X->getIntrinsicID() != Y->getIntrinsicID()))
2311 return nullptr;
2312
2313 Intrinsic::ID IID = X->getIntrinsicID();
2314 const APInt *RHSC;
2315 // Try to match constant RHS.
2316 if (!Y && (!(IID == Intrinsic::bswap || IID == Intrinsic::bitreverse) ||
2317 !match(I.getOperand(1), m_APInt(RHSC))))
2318 return nullptr;
2319
2320 switch (IID) {
2321 case Intrinsic::fshl:
2322 case Intrinsic::fshr: {
2323 if (X->getOperand(2) != Y->getOperand(2))
2324 return nullptr;
2325 Value *NewOp0 =
2326 Builder.CreateBinOp(I.getOpcode(), X->getOperand(0), Y->getOperand(0));
2327 Value *NewOp1 =
2328 Builder.CreateBinOp(I.getOpcode(), X->getOperand(1), Y->getOperand(1));
2329 Function *F =
2330 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2331 return CallInst::Create(F, {NewOp0, NewOp1, X->getOperand(2)});
2332 }
2333 case Intrinsic::bswap:
2334 case Intrinsic::bitreverse: {
2335 Value *NewOp0 = Builder.CreateBinOp(
2336 I.getOpcode(), X->getOperand(0),
2337 Y ? Y->getOperand(0)
2338 : ConstantInt::get(I.getType(), IID == Intrinsic::bswap
2339 ? RHSC->byteSwap()
2340 : RHSC->reverseBits()));
2341 Function *F =
2342 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2343 return CallInst::Create(F, {NewOp0});
2344 }
2345 default:
2346 return nullptr;
2347 }
2348}
2349
2350// Try to simplify V by replacing occurrences of Op with RepOp, but only look
2351// through bitwise operations. In particular, for X | Y we try to replace Y with
2352// 0 inside X and for X & Y we try to replace Y with -1 inside X.
2353// Return the simplified result of X if successful, and nullptr otherwise.
2354// If SimplifyOnly is true, no new instructions will be created.
2356 bool SimplifyOnly,
2357 InstCombinerImpl &IC,
2358 unsigned Depth = 0) {
2359 if (Op == RepOp)
2360 return nullptr;
2361
2362 if (V == Op)
2363 return RepOp;
2364
2365 auto *I = dyn_cast<BinaryOperator>(V);
2366 if (!I || !I->isBitwiseLogicOp() || Depth >= 3)
2367 return nullptr;
2368
2369 if (!I->hasOneUse())
2370 SimplifyOnly = true;
2371
2372 Value *NewOp0 = simplifyAndOrWithOpReplaced(I->getOperand(0), Op, RepOp,
2373 SimplifyOnly, IC, Depth + 1);
2374 Value *NewOp1 = simplifyAndOrWithOpReplaced(I->getOperand(1), Op, RepOp,
2375 SimplifyOnly, IC, Depth + 1);
2376 if (!NewOp0 && !NewOp1)
2377 return nullptr;
2378
2379 if (!NewOp0)
2380 NewOp0 = I->getOperand(0);
2381 if (!NewOp1)
2382 NewOp1 = I->getOperand(1);
2383
2384 if (Value *Res = simplifyBinOp(I->getOpcode(), NewOp0, NewOp1,
2386 return Res;
2387
2388 if (SimplifyOnly)
2389 return nullptr;
2390 return IC.Builder.CreateBinOp(I->getOpcode(), NewOp0, NewOp1);
2391}
2392
2393/// Reassociate and/or expressions to see if we can fold the inner and/or ops.
2394/// TODO: Make this recursive; it's a little tricky because an arbitrary
2395/// number of and/or instructions might have to be created.
2396Value *InstCombinerImpl::reassociateBooleanAndOr(Value *LHS, Value *X, Value *Y,
2397 Instruction &I, bool IsAnd,
2398 bool RHSIsLogical) {
2399 Instruction::BinaryOps Opcode = IsAnd ? Instruction::And : Instruction::Or;
2400 // LHS bop (X lop Y) --> (LHS bop X) lop Y
2401 // LHS bop (X bop Y) --> (LHS bop X) bop Y
2402 if (Value *Res = foldBooleanAndOr(LHS, X, I, IsAnd, /*IsLogical=*/false))
2403 return RHSIsLogical ? Builder.CreateLogicalOp(Opcode, Res, Y)
2404 : Builder.CreateBinOp(Opcode, Res, Y);
2405 // LHS bop (X bop Y) --> X bop (LHS bop Y)
2406 // LHS bop (X lop Y) --> X lop (LHS bop Y)
2407 if (Value *Res = foldBooleanAndOr(LHS, Y, I, IsAnd, /*IsLogical=*/false))
2408 return RHSIsLogical ? Builder.CreateLogicalOp(Opcode, X, Res)
2409 : Builder.CreateBinOp(Opcode, X, Res);
2410 return nullptr;
2411}
2412
2413// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
2414// here. We should standardize that construct where it is needed or choose some
2415// other way to ensure that commutated variants of patterns are not missed.
2417 Type *Ty = I.getType();
2418
2419 if (Value *V = simplifyAndInst(I.getOperand(0), I.getOperand(1),
2420 SQ.getWithInstruction(&I)))
2421 return replaceInstUsesWith(I, V);
2422
2424 return &I;
2425
2427 return X;
2428
2430 return Phi;
2431
2432 // See if we can simplify any instructions used by the instruction whose sole
2433 // purpose is to compute bits we don't care about.
2435 return &I;
2436
2437 // Do this before using distributive laws to catch simple and/or/not patterns.
2439 return Xor;
2440
2442 return X;
2443
2444 // (A|B)&(A|C) -> A|(B&C) etc
2446 return replaceInstUsesWith(I, V);
2447
2449 return R;
2450
2451 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2452
2453 Value *X, *Y;
2454 const APInt *C;
2455 if ((match(Op0, m_OneUse(m_LogicalShift(m_One(), m_Value(X)))) ||
2456 (match(Op0, m_OneUse(m_Shl(m_APInt(C), m_Value(X)))) && (*C)[0])) &&
2457 match(Op1, m_One())) {
2458 // (1 >> X) & 1 --> zext(X == 0)
2459 // (C << X) & 1 --> zext(X == 0), when C is odd
2460 Value *IsZero = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, 0));
2461 return new ZExtInst(IsZero, Ty);
2462 }
2463
2464 // (-(X & 1)) & Y --> (X & 1) == 0 ? 0 : Y
2465 Value *Neg;
2466 if (match(&I,
2468 m_Value(Y)))) {
2469 Value *Cmp = Builder.CreateIsNull(Neg);
2470 return createSelectInstWithUnknownProfile(Cmp,
2472 }
2473
2474 // Canonicalize:
2475 // (X +/- Y) & Y --> ~X & Y when Y is a power of 2.
2478 m_Sub(m_Value(X), m_Deferred(Y)))))) &&
2479 isKnownToBeAPowerOfTwo(Y, /*OrZero*/ true, &I))
2480 return BinaryOperator::CreateAnd(Builder.CreateNot(X), Y);
2481
2482 if (match(Op1, m_APInt(C))) {
2483 const APInt *XorC;
2484 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_APInt(XorC))))) {
2485 // (X ^ C1) & C2 --> (X & C2) ^ (C1&C2)
2486 Constant *NewC = ConstantInt::get(Ty, *C & *XorC);
2487 Value *And = Builder.CreateAnd(X, Op1);
2488 And->takeName(Op0);
2489 return BinaryOperator::CreateXor(And, NewC);
2490 }
2491
2492 const APInt *OrC;
2493 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_APInt(OrC))))) {
2494 // (X | C1) & C2 --> (X & C2^(C1&C2)) | (C1&C2)
2495 // NOTE: This reduces the number of bits set in the & mask, which
2496 // can expose opportunities for store narrowing for scalars.
2497 // NOTE: SimplifyDemandedBits should have already removed bits from C1
2498 // that aren't set in C2. Meaning we can replace (C1&C2) with C1 in
2499 // above, but this feels safer.
2500 APInt Together = *C & *OrC;
2501 Value *And = Builder.CreateAnd(X, ConstantInt::get(Ty, Together ^ *C));
2502 And->takeName(Op0);
2503 return BinaryOperator::CreateOr(And, ConstantInt::get(Ty, Together));
2504 }
2505
2506 unsigned Width = Ty->getScalarSizeInBits();
2507 const APInt *ShiftC;
2508 if (match(Op0, m_OneUse(m_SExt(m_AShr(m_Value(X), m_APInt(ShiftC))))) &&
2509 ShiftC->ult(Width)) {
2510 if (*C == APInt::getLowBitsSet(Width, Width - ShiftC->getZExtValue())) {
2511 // We are clearing high bits that were potentially set by sext+ashr:
2512 // and (sext (ashr X, ShiftC)), C --> lshr (sext X), ShiftC
2513 Value *Sext = Builder.CreateSExt(X, Ty);
2514 Constant *ShAmtC = ConstantInt::get(Ty, ShiftC->zext(Width));
2515 return BinaryOperator::CreateLShr(Sext, ShAmtC);
2516 }
2517 }
2518
2519 // If this 'and' clears the sign-bits added by ashr, replace with lshr:
2520 // and (ashr X, ShiftC), C --> lshr X, ShiftC
2521 if (match(Op0, m_AShr(m_Value(X), m_APInt(ShiftC))) && ShiftC->ult(Width) &&
2522 C->isMask(Width - ShiftC->getZExtValue()))
2523 return BinaryOperator::CreateLShr(X, ConstantInt::get(Ty, *ShiftC));
2524
2525 const APInt *AddC;
2526 if (match(Op0, m_Add(m_Value(X), m_APInt(AddC)))) {
2527 // If we are masking the result of the add down to exactly one bit and
2528 // the constant we are adding has no bits set below that bit, then the
2529 // add is flipping a single bit. Example:
2530 // (X + 4) & 4 --> (X & 4) ^ 4
2531 if (Op0->hasOneUse() && C->isPowerOf2() && (*AddC & (*C - 1)) == 0) {
2532 assert((*C & *AddC) != 0 && "Expected common bit");
2533 Value *NewAnd = Builder.CreateAnd(X, Op1);
2534 return BinaryOperator::CreateXor(NewAnd, Op1);
2535 }
2536 }
2537
2538 // ((C1 OP zext(X)) & C2) -> zext((C1 OP X) & C2) if C2 fits in the
2539 // bitwidth of X and OP behaves well when given trunc(C1) and X.
2540 auto isNarrowableBinOpcode = [](BinaryOperator *B) {
2541 switch (B->getOpcode()) {
2542 case Instruction::Xor:
2543 case Instruction::Or:
2544 case Instruction::Mul:
2545 case Instruction::Add:
2546 case Instruction::Sub:
2547 return true;
2548 default:
2549 return false;
2550 }
2551 };
2552 BinaryOperator *BO;
2553 if (match(Op0, m_OneUse(m_BinOp(BO))) && isNarrowableBinOpcode(BO)) {
2554 Instruction::BinaryOps BOpcode = BO->getOpcode();
2555 Value *X;
2556 const APInt *C1;
2557 // TODO: The one-use restrictions could be relaxed a little if the AND
2558 // is going to be removed.
2559 // Try to narrow the 'and' and a binop with constant operand:
2560 // and (bo (zext X), C1), C --> zext (and (bo X, TruncC1), TruncC)
2561 if (match(BO, m_c_BinOp(m_OneUse(m_ZExt(m_Value(X))), m_APInt(C1))) &&
2562 C->isIntN(X->getType()->getScalarSizeInBits())) {
2563 unsigned XWidth = X->getType()->getScalarSizeInBits();
2564 Constant *TruncC1 = ConstantInt::get(X->getType(), C1->trunc(XWidth));
2565 Value *BinOp = isa<ZExtInst>(BO->getOperand(0))
2566 ? Builder.CreateBinOp(BOpcode, X, TruncC1)
2567 : Builder.CreateBinOp(BOpcode, TruncC1, X);
2568 Constant *TruncC = ConstantInt::get(X->getType(), C->trunc(XWidth));
2569 Value *And = Builder.CreateAnd(BinOp, TruncC);
2570 return new ZExtInst(And, Ty);
2571 }
2572
2573 // Similar to above: if the mask matches the zext input width, then the
2574 // 'and' can be eliminated, so we can truncate the other variable op:
2575 // and (bo (zext X), Y), C --> zext (bo X, (trunc Y))
2576 if (isa<Instruction>(BO->getOperand(0)) &&
2577 match(BO->getOperand(0), m_OneUse(m_ZExt(m_Value(X)))) &&
2578 C->isMask(X->getType()->getScalarSizeInBits())) {
2579 Y = BO->getOperand(1);
2580 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2581 Value *NewBO =
2582 Builder.CreateBinOp(BOpcode, X, TrY, BO->getName() + ".narrow");
2583 return new ZExtInst(NewBO, Ty);
2584 }
2585 // and (bo Y, (zext X)), C --> zext (bo (trunc Y), X)
2586 if (isa<Instruction>(BO->getOperand(1)) &&
2587 match(BO->getOperand(1), m_OneUse(m_ZExt(m_Value(X)))) &&
2588 C->isMask(X->getType()->getScalarSizeInBits())) {
2589 Y = BO->getOperand(0);
2590 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2591 Value *NewBO =
2592 Builder.CreateBinOp(BOpcode, TrY, X, BO->getName() + ".narrow");
2593 return new ZExtInst(NewBO, Ty);
2594 }
2595 }
2596
2597 // This is intentionally placed after the narrowing transforms for
2598 // efficiency (transform directly to the narrow logic op if possible).
2599 // If the mask is only needed on one incoming arm, push the 'and' op up.
2600 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_Value(Y)))) ||
2601 match(Op0, m_OneUse(m_Or(m_Value(X), m_Value(Y))))) {
2602 APInt NotAndMask(~(*C));
2603 BinaryOperator::BinaryOps BinOp = cast<BinaryOperator>(Op0)->getOpcode();
2604 if (MaskedValueIsZero(X, NotAndMask, &I)) {
2605 // Not masking anything out for the LHS, move mask to RHS.
2606 // and ({x}or X, Y), C --> {x}or X, (and Y, C)
2607 Value *NewRHS = Builder.CreateAnd(Y, Op1, Y->getName() + ".masked");
2608 return BinaryOperator::Create(BinOp, X, NewRHS);
2609 }
2610 if (!isa<Constant>(Y) && MaskedValueIsZero(Y, NotAndMask, &I)) {
2611 // Not masking anything out for the RHS, move mask to LHS.
2612 // and ({x}or X, Y), C --> {x}or (and X, C), Y
2613 Value *NewLHS = Builder.CreateAnd(X, Op1, X->getName() + ".masked");
2614 return BinaryOperator::Create(BinOp, NewLHS, Y);
2615 }
2616 }
2617
2618 // When the mask is a power-of-2 constant and op0 is a shifted-power-of-2
2619 // constant, test if the shift amount equals the offset bit index:
2620 // (ShiftC << X) & C --> X == (log2(C) - log2(ShiftC)) ? C : 0
2621 // (ShiftC >> X) & C --> X == (log2(ShiftC) - log2(C)) ? C : 0
2622 if (C->isPowerOf2() &&
2623 match(Op0, m_OneUse(m_LogicalShift(m_Power2(ShiftC), m_Value(X))))) {
2624 int Log2ShiftC = ShiftC->exactLogBase2();
2625 int Log2C = C->exactLogBase2();
2626 bool IsShiftLeft =
2627 cast<BinaryOperator>(Op0)->getOpcode() == Instruction::Shl;
2628 int BitNum = IsShiftLeft ? Log2C - Log2ShiftC : Log2ShiftC - Log2C;
2629 assert(BitNum >= 0 && "Expected demanded bits to handle impossible mask");
2630 Value *Cmp = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, BitNum));
2631 return createSelectInstWithUnknownProfile(Cmp, ConstantInt::get(Ty, *C),
2633 }
2634
2635 Constant *C1, *C2;
2636 const APInt *C3 = C;
2637 Value *X;
2638 if (C3->isPowerOf2()) {
2639 Constant *Log2C3 = ConstantInt::get(Ty, C3->countr_zero());
2641 m_ImmConstant(C2)))) &&
2642 match(C1, m_Power2())) {
2644 Constant *LshrC = ConstantExpr::getAdd(C2, Log2C3);
2645 KnownBits KnownLShrc = computeKnownBits(LshrC, nullptr);
2646 if (KnownLShrc.getMaxValue().ult(Width)) {
2647 // iff C1,C3 is pow2 and C2 + cttz(C3) < BitWidth:
2648 // ((C1 << X) >> C2) & C3 -> X == (cttz(C3)+C2-cttz(C1)) ? C3 : 0
2649 Constant *CmpC = ConstantExpr::getSub(LshrC, Log2C1);
2650 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2651 return createSelectInstWithUnknownProfile(
2652 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2653 }
2654 }
2655
2657 m_ImmConstant(C2)))) &&
2658 match(C1, m_Power2())) {
2660 Constant *Cmp =
2662 if (Cmp && Cmp->isZeroValue()) {
2663 // iff C1,C3 is pow2 and Log2(C3) >= C2:
2664 // ((C1 >> X) << C2) & C3 -> X == (cttz(C1)+C2-cttz(C3)) ? C3 : 0
2665 Constant *ShlC = ConstantExpr::getAdd(C2, Log2C1);
2666 Constant *CmpC = ConstantExpr::getSub(ShlC, Log2C3);
2667 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2668 return createSelectInstWithUnknownProfile(
2669 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2670 }
2671 }
2672 }
2673 }
2674
2675 // If we are clearing the sign bit of a floating-point value, convert this to
2676 // fabs, then cast back to integer.
2677 //
2678 // This is a generous interpretation for noimplicitfloat, this is not a true
2679 // floating-point operation.
2680 //
2681 // Assumes any IEEE-represented type has the sign bit in the high bit.
2682 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
2683 Value *CastOp;
2684 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
2685 match(Op1, m_MaxSignedValue()) &&
2686 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
2687 Attribute::NoImplicitFloat)) {
2688 Type *EltTy = CastOp->getType()->getScalarType();
2689 if (EltTy->isFloatingPointTy() &&
2691 Value *FAbs = Builder.CreateUnaryIntrinsic(Intrinsic::fabs, CastOp);
2692 return new BitCastInst(FAbs, I.getType());
2693 }
2694 }
2695
2696 // and(shl(zext(X), Y), SignMask) -> and(sext(X), SignMask)
2697 // where Y is a valid shift amount.
2699 m_SignMask())) &&
2702 APInt(Ty->getScalarSizeInBits(),
2703 Ty->getScalarSizeInBits() -
2704 X->getType()->getScalarSizeInBits())))) {
2705 auto *SExt = Builder.CreateSExt(X, Ty, X->getName() + ".signext");
2706 return BinaryOperator::CreateAnd(SExt, Op1);
2707 }
2708
2709 if (Instruction *Z = narrowMaskedBinOp(I))
2710 return Z;
2711
2712 if (I.getType()->isIntOrIntVectorTy(1)) {
2713 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
2714 if (auto *R =
2715 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ true))
2716 return R;
2717 }
2718 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
2719 if (auto *R =
2720 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ true))
2721 return R;
2722 }
2723 }
2724
2725 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
2726 return FoldedLogic;
2727
2728 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
2729 return DeMorgan;
2730
2731 {
2732 Value *A, *B, *C;
2733 // A & ~(A ^ B) --> A & B
2734 if (match(Op1, m_Not(m_c_Xor(m_Specific(Op0), m_Value(B)))))
2735 return BinaryOperator::CreateAnd(Op0, B);
2736 // ~(A ^ B) & A --> A & B
2737 if (match(Op0, m_Not(m_c_Xor(m_Specific(Op1), m_Value(B)))))
2738 return BinaryOperator::CreateAnd(Op1, B);
2739
2740 // (A ^ B) & ((B ^ C) ^ A) -> (A ^ B) & ~C
2741 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
2742 match(Op1, m_Xor(m_Xor(m_Specific(B), m_Value(C)), m_Specific(A)))) {
2743 Value *NotC = Op1->hasOneUse()
2744 ? Builder.CreateNot(C)
2745 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2746 if (NotC != nullptr)
2747 return BinaryOperator::CreateAnd(Op0, NotC);
2748 }
2749
2750 // ((A ^ C) ^ B) & (B ^ A) -> (B ^ A) & ~C
2751 if (match(Op0, m_Xor(m_Xor(m_Value(A), m_Value(C)), m_Value(B))) &&
2752 match(Op1, m_Xor(m_Specific(B), m_Specific(A)))) {
2753 Value *NotC = Op0->hasOneUse()
2754 ? Builder.CreateNot(C)
2755 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2756 if (NotC != nullptr)
2757 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(C));
2758 }
2759
2760 // (A | B) & (~A ^ B) -> A & B
2761 // (A | B) & (B ^ ~A) -> A & B
2762 // (B | A) & (~A ^ B) -> A & B
2763 // (B | A) & (B ^ ~A) -> A & B
2764 if (match(Op1, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2765 match(Op0, m_c_Or(m_Specific(A), m_Specific(B))))
2766 return BinaryOperator::CreateAnd(A, B);
2767
2768 // (~A ^ B) & (A | B) -> A & B
2769 // (~A ^ B) & (B | A) -> A & B
2770 // (B ^ ~A) & (A | B) -> A & B
2771 // (B ^ ~A) & (B | A) -> A & B
2772 if (match(Op0, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2773 match(Op1, m_c_Or(m_Specific(A), m_Specific(B))))
2774 return BinaryOperator::CreateAnd(A, B);
2775
2776 // (~A | B) & (A ^ B) -> ~A & B
2777 // (~A | B) & (B ^ A) -> ~A & B
2778 // (B | ~A) & (A ^ B) -> ~A & B
2779 // (B | ~A) & (B ^ A) -> ~A & B
2780 if (match(Op0, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2782 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2783
2784 // (A ^ B) & (~A | B) -> ~A & B
2785 // (B ^ A) & (~A | B) -> ~A & B
2786 // (A ^ B) & (B | ~A) -> ~A & B
2787 // (B ^ A) & (B | ~A) -> ~A & B
2788 if (match(Op1, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2790 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2791 }
2792
2793 if (Value *Res =
2794 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/true, /*IsLogical=*/false))
2795 return replaceInstUsesWith(I, Res);
2796
2797 if (match(Op1, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2798 bool IsLogical = isa<SelectInst>(Op1);
2799 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/true,
2800 /*RHSIsLogical=*/IsLogical))
2801 return replaceInstUsesWith(I, V);
2802 }
2803 if (match(Op0, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2804 bool IsLogical = isa<SelectInst>(Op0);
2805 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/true,
2806 /*RHSIsLogical=*/IsLogical))
2807 return replaceInstUsesWith(I, V);
2808 }
2809
2810 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
2811 return FoldedFCmps;
2812
2813 if (Instruction *CastedAnd = foldCastedBitwiseLogic(I))
2814 return CastedAnd;
2815
2816 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
2817 return Sel;
2818
2819 // and(sext(A), B) / and(B, sext(A)) --> A ? B : 0, where A is i1 or <N x i1>.
2820 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
2821 // with binop identity constant. But creating a select with non-constant
2822 // arm may not be reversible due to poison semantics. Is that a good
2823 // canonicalization?
2824 Value *A, *B;
2825 if (match(&I, m_c_And(m_SExt(m_Value(A)), m_Value(B))) &&
2826 A->getType()->isIntOrIntVectorTy(1))
2827 return createSelectInstWithUnknownProfile(A, B, Constant::getNullValue(Ty));
2828
2829 // Similarly, a 'not' of the bool translates to a swap of the select arms:
2830 // ~sext(A) & B / B & ~sext(A) --> A ? 0 : B
2831 if (match(&I, m_c_And(m_Not(m_SExt(m_Value(A))), m_Value(B))) &&
2832 A->getType()->isIntOrIntVectorTy(1))
2833 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty), B);
2834
2835 // and(zext(A), B) -> A ? (B & 1) : 0
2836 if (match(&I, m_c_And(m_OneUse(m_ZExt(m_Value(A))), m_Value(B))) &&
2837 A->getType()->isIntOrIntVectorTy(1))
2838 return createSelectInstWithUnknownProfile(
2839 A, Builder.CreateAnd(B, ConstantInt::get(Ty, 1)),
2841
2842 // (-1 + A) & B --> A ? 0 : B where A is 0/1.
2844 m_Value(B)))) {
2845 if (A->getType()->isIntOrIntVectorTy(1))
2846 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty),
2847 B);
2848 if (computeKnownBits(A, &I).countMaxActiveBits() <= 1) {
2849 return createSelectInstWithUnknownProfile(
2850 Builder.CreateICmpEQ(A, Constant::getNullValue(A->getType())), B,
2852 }
2853 }
2854
2855 // (iN X s>> (N-1)) & Y --> (X s< 0) ? Y : 0 -- with optional sext
2858 m_Value(Y))) &&
2859 *C == X->getType()->getScalarSizeInBits() - 1) {
2860 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2861 return createSelectInstWithUnknownProfile(IsNeg, Y,
2863 }
2864 // If there's a 'not' of the shifted value, swap the select operands:
2865 // ~(iN X s>> (N-1)) & Y --> (X s< 0) ? 0 : Y -- with optional sext
2868 m_Value(Y))) &&
2869 *C == X->getType()->getScalarSizeInBits() - 1) {
2870 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2871 return createSelectInstWithUnknownProfile(IsNeg,
2873 }
2874
2875 // (~x) & y --> ~(x | (~y)) iff that gets rid of inversions
2877 return &I;
2878
2879 // An and recurrence w/loop invariant step is equivelent to (and start, step)
2880 PHINode *PN = nullptr;
2881 Value *Start = nullptr, *Step = nullptr;
2882 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
2883 return replaceInstUsesWith(I, Builder.CreateAnd(Start, Step));
2884
2886 return R;
2887
2888 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
2889 return Canonicalized;
2890
2891 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
2892 return Folded;
2893
2894 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
2895 return Res;
2896
2898 return Res;
2899
2900 if (Value *V =
2902 /*SimplifyOnly*/ false, *this))
2903 return BinaryOperator::CreateAnd(V, Op1);
2904 if (Value *V =
2906 /*SimplifyOnly*/ false, *this))
2907 return BinaryOperator::CreateAnd(Op0, V);
2908
2909 return nullptr;
2910}
2911
2913 bool MatchBSwaps,
2914 bool MatchBitReversals) {
2916 if (!recognizeBSwapOrBitReverseIdiom(&I, MatchBSwaps, MatchBitReversals,
2917 Insts))
2918 return nullptr;
2919 Instruction *LastInst = Insts.pop_back_val();
2920 LastInst->removeFromParent();
2921
2922 for (auto *Inst : Insts) {
2923 Inst->setDebugLoc(I.getDebugLoc());
2924 Worklist.push(Inst);
2925 }
2926 return LastInst;
2927}
2928
2929std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
2931 // TODO: Can we reduce the code duplication between this and the related
2932 // rotate matching code under visitSelect and visitTrunc?
2933 assert(Or.getOpcode() == BinaryOperator::Or && "Expecting or instruction");
2934
2935 unsigned Width = Or.getType()->getScalarSizeInBits();
2936
2937 Instruction *Or0, *Or1;
2938 if (!match(Or.getOperand(0), m_Instruction(Or0)) ||
2939 !match(Or.getOperand(1), m_Instruction(Or1)))
2940 return std::nullopt;
2941
2942 bool IsFshl = true; // Sub on LSHR.
2943 SmallVector<Value *, 3> FShiftArgs;
2944
2945 // First, find an or'd pair of opposite shifts:
2946 // or (lshr ShVal0, ShAmt0), (shl ShVal1, ShAmt1)
2947 if (isa<BinaryOperator>(Or0) && isa<BinaryOperator>(Or1)) {
2948 Value *ShVal0, *ShVal1, *ShAmt0, *ShAmt1;
2949 if (!match(Or0,
2950 m_OneUse(m_LogicalShift(m_Value(ShVal0), m_Value(ShAmt0)))) ||
2951 !match(Or1,
2952 m_OneUse(m_LogicalShift(m_Value(ShVal1), m_Value(ShAmt1)))) ||
2953 Or0->getOpcode() == Or1->getOpcode())
2954 return std::nullopt;
2955
2956 // Canonicalize to or(shl(ShVal0, ShAmt0), lshr(ShVal1, ShAmt1)).
2957 if (Or0->getOpcode() == BinaryOperator::LShr) {
2958 std::swap(Or0, Or1);
2959 std::swap(ShVal0, ShVal1);
2960 std::swap(ShAmt0, ShAmt1);
2961 }
2962 assert(Or0->getOpcode() == BinaryOperator::Shl &&
2963 Or1->getOpcode() == BinaryOperator::LShr &&
2964 "Illegal or(shift,shift) pair");
2965
2966 // Match the shift amount operands for a funnel shift pattern. This always
2967 // matches a subtraction on the R operand.
2968 auto matchShiftAmount = [&](Value *L, Value *R, unsigned Width) -> Value * {
2969 // Check for constant shift amounts that sum to the bitwidth.
2970 const APInt *LI, *RI;
2971 if (match(L, m_APIntAllowPoison(LI)) && match(R, m_APIntAllowPoison(RI)))
2972 if (LI->ult(Width) && RI->ult(Width) && (*LI + *RI) == Width)
2973 return ConstantInt::get(L->getType(), *LI);
2974
2975 Constant *LC, *RC;
2976 if (match(L, m_Constant(LC)) && match(R, m_Constant(RC)) &&
2977 match(L,
2978 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2979 match(R,
2980 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2982 return ConstantExpr::mergeUndefsWith(LC, RC);
2983
2984 // (shl ShVal, X) | (lshr ShVal, (Width - x)) iff X < Width.
2985 // We limit this to X < Width in case the backend re-expands the
2986 // intrinsic, and has to reintroduce a shift modulo operation (InstCombine
2987 // might remove it after this fold). This still doesn't guarantee that the
2988 // final codegen will match this original pattern.
2989 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) {
2990 KnownBits KnownL = computeKnownBits(L, &Or);
2991 return KnownL.getMaxValue().ult(Width) ? L : nullptr;
2992 }
2993
2994 // For non-constant cases, the following patterns currently only work for
2995 // rotation patterns.
2996 // TODO: Add general funnel-shift compatible patterns.
2997 if (ShVal0 != ShVal1)
2998 return nullptr;
2999
3000 // For non-constant cases we don't support non-pow2 shift masks.
3001 // TODO: Is it worth matching urem as well?
3002 if (!isPowerOf2_32(Width))
3003 return nullptr;
3004
3005 // The shift amount may be masked with negation:
3006 // (shl ShVal, (X & (Width - 1))) | (lshr ShVal, ((-X) & (Width - 1)))
3007 Value *X;
3008 unsigned Mask = Width - 1;
3009 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) &&
3010 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask))))
3011 return X;
3012
3013 // (shl ShVal, X) | (lshr ShVal, ((-X) & (Width - 1)))
3014 if (match(R, m_And(m_Neg(m_Specific(L)), m_SpecificInt(Mask))))
3015 return L;
3016
3017 // Similar to above, but the shift amount may be extended after masking,
3018 // so return the extended value as the parameter for the intrinsic.
3019 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3020 match(R,
3022 m_SpecificInt(Mask))))
3023 return L;
3024
3025 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3027 return L;
3028
3029 return nullptr;
3030 };
3031
3032 Value *ShAmt = matchShiftAmount(ShAmt0, ShAmt1, Width);
3033 if (!ShAmt) {
3034 ShAmt = matchShiftAmount(ShAmt1, ShAmt0, Width);
3035 IsFshl = false; // Sub on SHL.
3036 }
3037 if (!ShAmt)
3038 return std::nullopt;
3039
3040 FShiftArgs = {ShVal0, ShVal1, ShAmt};
3041 } else if (isa<ZExtInst>(Or0) || isa<ZExtInst>(Or1)) {
3042 // If there are two 'or' instructions concat variables in opposite order:
3043 //
3044 // Slot1 and Slot2 are all zero bits.
3045 // | Slot1 | Low | Slot2 | High |
3046 // LowHigh = or (shl (zext Low), ZextLowShlAmt), (zext High)
3047 // | Slot2 | High | Slot1 | Low |
3048 // HighLow = or (shl (zext High), ZextHighShlAmt), (zext Low)
3049 //
3050 // the latter 'or' can be safely convert to
3051 // -> HighLow = fshl LowHigh, LowHigh, ZextHighShlAmt
3052 // if ZextLowShlAmt + ZextHighShlAmt == Width.
3053 if (!isa<ZExtInst>(Or1))
3054 std::swap(Or0, Or1);
3055
3056 Value *High, *ZextHigh, *Low;
3057 const APInt *ZextHighShlAmt;
3058 if (!match(Or0,
3059 m_OneUse(m_Shl(m_Value(ZextHigh), m_APInt(ZextHighShlAmt)))))
3060 return std::nullopt;
3061
3062 if (!match(Or1, m_ZExt(m_Value(Low))) ||
3063 !match(ZextHigh, m_ZExt(m_Value(High))))
3064 return std::nullopt;
3065
3066 unsigned HighSize = High->getType()->getScalarSizeInBits();
3067 unsigned LowSize = Low->getType()->getScalarSizeInBits();
3068 // Make sure High does not overlap with Low and most significant bits of
3069 // High aren't shifted out.
3070 if (ZextHighShlAmt->ult(LowSize) || ZextHighShlAmt->ugt(Width - HighSize))
3071 return std::nullopt;
3072
3073 for (User *U : ZextHigh->users()) {
3074 Value *X, *Y;
3075 if (!match(U, m_Or(m_Value(X), m_Value(Y))))
3076 continue;
3077
3078 if (!isa<ZExtInst>(Y))
3079 std::swap(X, Y);
3080
3081 const APInt *ZextLowShlAmt;
3082 if (!match(X, m_Shl(m_Specific(Or1), m_APInt(ZextLowShlAmt))) ||
3083 !match(Y, m_Specific(ZextHigh)) || !DT.dominates(U, &Or))
3084 continue;
3085
3086 // HighLow is good concat. If sum of two shifts amount equals to Width,
3087 // LowHigh must also be a good concat.
3088 if (*ZextLowShlAmt + *ZextHighShlAmt != Width)
3089 continue;
3090
3091 // Low must not overlap with High and most significant bits of Low must
3092 // not be shifted out.
3093 assert(ZextLowShlAmt->uge(HighSize) &&
3094 ZextLowShlAmt->ule(Width - LowSize) && "Invalid concat");
3095
3096 // We cannot reuse the result if it may produce poison.
3097 // Drop poison generating flags in the expression tree.
3098 // Or
3099 cast<Instruction>(U)->dropPoisonGeneratingFlags();
3100 // Shl
3101 cast<Instruction>(X)->dropPoisonGeneratingFlags();
3102
3103 FShiftArgs = {U, U, ConstantInt::get(Or0->getType(), *ZextHighShlAmt)};
3104 break;
3105 }
3106 }
3107
3108 if (FShiftArgs.empty())
3109 return std::nullopt;
3110
3111 Intrinsic::ID IID = IsFshl ? Intrinsic::fshl : Intrinsic::fshr;
3112 return std::make_pair(IID, FShiftArgs);
3113}
3114
3115/// Match UB-safe variants of the funnel shift intrinsic.
3117 if (auto Opt = IC.convertOrOfShiftsToFunnelShift(Or)) {
3118 auto [IID, FShiftArgs] = *Opt;
3119 Function *F =
3120 Intrinsic::getOrInsertDeclaration(Or.getModule(), IID, Or.getType());
3121 return CallInst::Create(F, FShiftArgs);
3122 }
3123
3124 return nullptr;
3125}
3126
3127/// Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
3129 assert(Or.getOpcode() == Instruction::Or && "bswap requires an 'or'");
3130 Value *Op0 = Or.getOperand(0), *Op1 = Or.getOperand(1);
3131 Type *Ty = Or.getType();
3132
3133 unsigned Width = Ty->getScalarSizeInBits();
3134 if ((Width & 1) != 0)
3135 return nullptr;
3136 unsigned HalfWidth = Width / 2;
3137
3138 // Canonicalize zext (lower half) to LHS.
3139 if (!isa<ZExtInst>(Op0))
3140 std::swap(Op0, Op1);
3141
3142 // Find lower/upper half.
3143 Value *LowerSrc, *ShlVal, *UpperSrc;
3144 const APInt *C;
3145 if (!match(Op0, m_OneUse(m_ZExt(m_Value(LowerSrc)))) ||
3146 !match(Op1, m_OneUse(m_Shl(m_Value(ShlVal), m_APInt(C)))) ||
3147 !match(ShlVal, m_OneUse(m_ZExt(m_Value(UpperSrc)))))
3148 return nullptr;
3149 if (*C != HalfWidth || LowerSrc->getType() != UpperSrc->getType() ||
3150 LowerSrc->getType()->getScalarSizeInBits() != HalfWidth)
3151 return nullptr;
3152
3153 auto ConcatIntrinsicCalls = [&](Intrinsic::ID id, Value *Lo, Value *Hi) {
3154 Value *NewLower = Builder.CreateZExt(Lo, Ty);
3155 Value *NewUpper = Builder.CreateZExt(Hi, Ty);
3156 NewUpper = Builder.CreateShl(NewUpper, HalfWidth);
3157 Value *BinOp = Builder.CreateOr(NewLower, NewUpper);
3158 return Builder.CreateIntrinsic(id, Ty, BinOp);
3159 };
3160
3161 // BSWAP: Push the concat down, swapping the lower/upper sources.
3162 // concat(bswap(x),bswap(y)) -> bswap(concat(x,y))
3163 Value *LowerBSwap, *UpperBSwap;
3164 if (match(LowerSrc, m_BSwap(m_Value(LowerBSwap))) &&
3165 match(UpperSrc, m_BSwap(m_Value(UpperBSwap))))
3166 return ConcatIntrinsicCalls(Intrinsic::bswap, UpperBSwap, LowerBSwap);
3167
3168 // BITREVERSE: Push the concat down, swapping the lower/upper sources.
3169 // concat(bitreverse(x),bitreverse(y)) -> bitreverse(concat(x,y))
3170 Value *LowerBRev, *UpperBRev;
3171 if (match(LowerSrc, m_BitReverse(m_Value(LowerBRev))) &&
3172 match(UpperSrc, m_BitReverse(m_Value(UpperBRev))))
3173 return ConcatIntrinsicCalls(Intrinsic::bitreverse, UpperBRev, LowerBRev);
3174
3175 // iX ext split: extending or(zext(x),shl(zext(y),bw/2) pattern
3176 // to consume sext/ashr:
3177 // or(zext(sext(x)),shl(zext(sext(ashr(x,xbw-1))),bw/2)
3178 // or(zext(x),shl(zext(ashr(x,xbw-1)),bw/2)
3179 Value *X;
3180 if (match(LowerSrc, m_SExtOrSelf(m_Value(X))) &&
3181 match(UpperSrc,
3183 m_Specific(X),
3184 m_SpecificInt(X->getType()->getScalarSizeInBits() - 1)))))
3185 return Builder.CreateSExt(X, Ty);
3186
3187 return nullptr;
3188}
3189
3190/// If all elements of two constant vectors are 0/-1 and inverses, return true.
3192 unsigned NumElts = cast<FixedVectorType>(C1->getType())->getNumElements();
3193 for (unsigned i = 0; i != NumElts; ++i) {
3194 Constant *EltC1 = C1->getAggregateElement(i);
3195 Constant *EltC2 = C2->getAggregateElement(i);
3196 if (!EltC1 || !EltC2)
3197 return false;
3198
3199 // One element must be all ones, and the other must be all zeros.
3200 if (!((match(EltC1, m_Zero()) && match(EltC2, m_AllOnes())) ||
3201 (match(EltC2, m_Zero()) && match(EltC1, m_AllOnes()))))
3202 return false;
3203 }
3204 return true;
3205}
3206
3207/// We have an expression of the form (A & C) | (B & D). If A is a scalar or
3208/// vector composed of all-zeros or all-ones values and is the bitwise 'not' of
3209/// B, it can be used as the condition operand of a select instruction.
3210/// We will detect (A & C) | ~(B | D) when the flag ABIsTheSame enabled.
3211Value *InstCombinerImpl::getSelectCondition(Value *A, Value *B,
3212 bool ABIsTheSame) {
3213 // We may have peeked through bitcasts in the caller.
3214 // Exit immediately if we don't have (vector) integer types.
3215 Type *Ty = A->getType();
3216 if (!Ty->isIntOrIntVectorTy() || !B->getType()->isIntOrIntVectorTy())
3217 return nullptr;
3218
3219 // If A is the 'not' operand of B and has enough signbits, we have our answer.
3220 if (ABIsTheSame ? (A == B) : match(B, m_Not(m_Specific(A)))) {
3221 // If these are scalars or vectors of i1, A can be used directly.
3222 if (Ty->isIntOrIntVectorTy(1))
3223 return A;
3224
3225 // If we look through a vector bitcast, the caller will bitcast the operands
3226 // to match the condition's number of bits (N x i1).
3227 // To make this poison-safe, disallow bitcast from wide element to narrow
3228 // element. That could allow poison in lanes where it was not present in the
3229 // original code.
3231 if (A->getType()->isIntOrIntVectorTy()) {
3232 unsigned NumSignBits = ComputeNumSignBits(A);
3233 if (NumSignBits == A->getType()->getScalarSizeInBits() &&
3234 NumSignBits <= Ty->getScalarSizeInBits())
3235 return Builder.CreateTrunc(A, CmpInst::makeCmpResultType(A->getType()));
3236 }
3237 return nullptr;
3238 }
3239
3240 // TODO: add support for sext and constant case
3241 if (ABIsTheSame)
3242 return nullptr;
3243
3244 // If both operands are constants, see if the constants are inverse bitmasks.
3245 Constant *AConst, *BConst;
3246 if (match(A, m_Constant(AConst)) && match(B, m_Constant(BConst)))
3247 if (AConst == ConstantExpr::getNot(BConst) &&
3249 return Builder.CreateZExtOrTrunc(A, CmpInst::makeCmpResultType(Ty));
3250
3251 // Look for more complex patterns. The 'not' op may be hidden behind various
3252 // casts. Look through sexts and bitcasts to find the booleans.
3253 Value *Cond;
3254 Value *NotB;
3255 if (match(A, m_SExt(m_Value(Cond))) &&
3256 Cond->getType()->isIntOrIntVectorTy(1)) {
3257 // A = sext i1 Cond; B = sext (not (i1 Cond))
3258 if (match(B, m_SExt(m_Not(m_Specific(Cond)))))
3259 return Cond;
3260
3261 // A = sext i1 Cond; B = not ({bitcast} (sext (i1 Cond)))
3262 // TODO: The one-use checks are unnecessary or misplaced. If the caller
3263 // checked for uses on logic ops/casts, that should be enough to
3264 // make this transform worthwhile.
3265 if (match(B, m_OneUse(m_Not(m_Value(NotB))))) {
3266 NotB = peekThroughBitcast(NotB, true);
3267 if (match(NotB, m_SExt(m_Specific(Cond))))
3268 return Cond;
3269 }
3270 }
3271
3272 // All scalar (and most vector) possibilities should be handled now.
3273 // Try more matches that only apply to non-splat constant vectors.
3274 if (!Ty->isVectorTy())
3275 return nullptr;
3276
3277 // If both operands are xor'd with constants using the same sexted boolean
3278 // operand, see if the constants are inverse bitmasks.
3279 // TODO: Use ConstantExpr::getNot()?
3280 if (match(A, (m_Xor(m_SExt(m_Value(Cond)), m_Constant(AConst)))) &&
3281 match(B, (m_Xor(m_SExt(m_Specific(Cond)), m_Constant(BConst)))) &&
3282 Cond->getType()->isIntOrIntVectorTy(1) &&
3283 areInverseVectorBitmasks(AConst, BConst)) {
3285 return Builder.CreateXor(Cond, AConst);
3286 }
3287 return nullptr;
3288}
3289
3290/// We have an expression of the form (A & B) | (C & D). Try to simplify this
3291/// to "A' ? B : D", where A' is a boolean or vector of booleans.
3292/// When InvertFalseVal is set to true, we try to match the pattern
3293/// where we have peeked through a 'not' op and A and C are the same:
3294/// (A & B) | ~(A | D) --> (A & B) | (~A & ~D) --> A' ? B : ~D
3295Value *InstCombinerImpl::matchSelectFromAndOr(Value *A, Value *B, Value *C,
3296 Value *D, bool InvertFalseVal) {
3297 // The potential condition of the select may be bitcasted. In that case, look
3298 // through its bitcast and the corresponding bitcast of the 'not' condition.
3299 Type *OrigType = A->getType();
3300 A = peekThroughBitcast(A, true);
3301 C = peekThroughBitcast(C, true);
3302 if (Value *Cond = getSelectCondition(A, C, InvertFalseVal)) {
3303 // ((bc Cond) & B) | ((bc ~Cond) & D) --> bc (select Cond, (bc B), (bc D))
3304 // If this is a vector, we may need to cast to match the condition's length.
3305 // The bitcasts will either all exist or all not exist. The builder will
3306 // not create unnecessary casts if the types already match.
3307 Type *SelTy = A->getType();
3308 if (auto *VecTy = dyn_cast<VectorType>(Cond->getType())) {
3309 // For a fixed or scalable vector get N from <{vscale x} N x iM>
3310 unsigned Elts = VecTy->getElementCount().getKnownMinValue();
3311 // For a fixed or scalable vector, get the size in bits of N x iM; for a
3312 // scalar this is just M.
3313 unsigned SelEltSize = SelTy->getPrimitiveSizeInBits().getKnownMinValue();
3314 Type *EltTy = Builder.getIntNTy(SelEltSize / Elts);
3315 SelTy = VectorType::get(EltTy, VecTy->getElementCount());
3316 }
3317 Value *BitcastB = Builder.CreateBitCast(B, SelTy);
3318 if (InvertFalseVal)
3319 D = Builder.CreateNot(D);
3320 Value *BitcastD = Builder.CreateBitCast(D, SelTy);
3321 Value *Select = Builder.CreateSelect(Cond, BitcastB, BitcastD);
3322 return Builder.CreateBitCast(Select, OrigType);
3323 }
3324
3325 return nullptr;
3326}
3327
3328// (icmp eq X, C) | (icmp ult Other, (X - C)) -> (icmp ule Other, (X - (C + 1)))
3329// (icmp ne X, C) & (icmp uge Other, (X - C)) -> (icmp ugt Other, (X - (C + 1)))
3331 bool IsAnd, bool IsLogical,
3332 IRBuilderBase &Builder) {
3333 Value *LHS0 = LHS->getOperand(0);
3334 Value *RHS0 = RHS->getOperand(0);
3335 Value *RHS1 = RHS->getOperand(1);
3336
3337 ICmpInst::Predicate LPred =
3338 IsAnd ? LHS->getInversePredicate() : LHS->getPredicate();
3339 ICmpInst::Predicate RPred =
3340 IsAnd ? RHS->getInversePredicate() : RHS->getPredicate();
3341
3342 const APInt *CInt;
3343 if (LPred != ICmpInst::ICMP_EQ ||
3344 !match(LHS->getOperand(1), m_APIntAllowPoison(CInt)) ||
3345 !LHS0->getType()->isIntOrIntVectorTy() ||
3346 !(LHS->hasOneUse() || RHS->hasOneUse()))
3347 return nullptr;
3348
3349 auto MatchRHSOp = [LHS0, CInt](const Value *RHSOp) {
3350 return match(RHSOp,
3351 m_Add(m_Specific(LHS0), m_SpecificIntAllowPoison(-*CInt))) ||
3352 (CInt->isZero() && RHSOp == LHS0);
3353 };
3354
3355 Value *Other;
3356 if (RPred == ICmpInst::ICMP_ULT && MatchRHSOp(RHS1))
3357 Other = RHS0;
3358 else if (RPred == ICmpInst::ICMP_UGT && MatchRHSOp(RHS0))
3359 Other = RHS1;
3360 else
3361 return nullptr;
3362
3363 if (IsLogical)
3364 Other = Builder.CreateFreeze(Other);
3365
3366 return Builder.CreateICmp(
3368 Builder.CreateSub(LHS0, ConstantInt::get(LHS0->getType(), *CInt + 1)),
3369 Other);
3370}
3371
3372/// Fold (icmp)&(icmp) or (icmp)|(icmp) if possible.
3373/// If IsLogical is true, then the and/or is in select form and the transform
3374/// must be poison-safe.
3375Value *InstCombinerImpl::foldAndOrOfICmps(ICmpInst *LHS, ICmpInst *RHS,
3376 Instruction &I, bool IsAnd,
3377 bool IsLogical) {
3378 const SimplifyQuery Q = SQ.getWithInstruction(&I);
3379
3380 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
3381 Value *LHS0 = LHS->getOperand(0), *RHS0 = RHS->getOperand(0);
3382 Value *LHS1 = LHS->getOperand(1), *RHS1 = RHS->getOperand(1);
3383
3384 const APInt *LHSC = nullptr, *RHSC = nullptr;
3385 match(LHS1, m_APInt(LHSC));
3386 match(RHS1, m_APInt(RHSC));
3387
3388 // (icmp1 A, B) | (icmp2 A, B) --> (icmp3 A, B)
3389 // (icmp1 A, B) & (icmp2 A, B) --> (icmp3 A, B)
3390 if (predicatesFoldable(PredL, PredR)) {
3391 if (LHS0 == RHS1 && LHS1 == RHS0) {
3392 PredL = ICmpInst::getSwappedPredicate(PredL);
3393 std::swap(LHS0, LHS1);
3394 }
3395 if (LHS0 == RHS0 && LHS1 == RHS1) {
3396 unsigned Code = IsAnd ? getICmpCode(PredL) & getICmpCode(PredR)
3397 : getICmpCode(PredL) | getICmpCode(PredR);
3398 bool IsSigned = LHS->isSigned() || RHS->isSigned();
3399 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
3400 }
3401 }
3402
3403 if (Value *V =
3404 foldAndOrOfICmpEqConstantAndICmp(LHS, RHS, IsAnd, IsLogical, Builder))
3405 return V;
3406 // We can treat logical like bitwise here, because both operands are used on
3407 // the LHS, and as such poison from both will propagate.
3409 /*IsLogical*/ false, Builder))
3410 return V;
3411
3412 if (Value *V = foldAndOrOfICmpsWithConstEq(LHS, RHS, IsAnd, IsLogical,
3413 Builder, Q, I))
3414 return V;
3415 // We can convert this case to bitwise and, because both operands are used
3416 // on the LHS, and as such poison from both will propagate.
3418 RHS, LHS, IsAnd, /*IsLogical=*/false, Builder, Q, I)) {
3419 // If RHS is still used, we should drop samesign flag.
3420 if (IsLogical && RHS->hasSameSign() && !RHS->use_empty()) {
3421 RHS->setSameSign(false);
3423 }
3424 return V;
3425 }
3426
3427 if (Value *V = foldIsPowerOf2OrZero(LHS, RHS, IsAnd, Builder, *this))
3428 return V;
3429 if (Value *V = foldIsPowerOf2OrZero(RHS, LHS, IsAnd, Builder, *this))
3430 return V;
3431
3432 // TODO: One of these directions is fine with logical and/or, the other could
3433 // be supported by inserting freeze.
3434 if (!IsLogical) {
3435 // E.g. (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
3436 // E.g. (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
3437 if (Value *V = simplifyRangeCheck(LHS, RHS, /*Inverted=*/!IsAnd))
3438 return V;
3439
3440 // E.g. (icmp sgt x, n) | (icmp slt x, 0) --> icmp ugt x, n
3441 // E.g. (icmp slt x, n) & (icmp sge x, 0) --> icmp ult x, n
3442 if (Value *V = simplifyRangeCheck(RHS, LHS, /*Inverted=*/!IsAnd))
3443 return V;
3444 }
3445
3446 // TODO: Add conjugated or fold, check whether it is safe for logical and/or.
3447 if (IsAnd && !IsLogical)
3449 return V;
3450
3451 if (Value *V = foldIsPowerOf2(LHS, RHS, IsAnd, Builder, *this))
3452 return V;
3453
3454 if (Value *V = foldPowerOf2AndShiftedMask(LHS, RHS, IsAnd, Builder))
3455 return V;
3456
3457 // TODO: Verify whether this is safe for logical and/or.
3458 if (!IsLogical) {
3459 if (Value *X = foldUnsignedUnderflowCheck(LHS, RHS, IsAnd, Q, Builder))
3460 return X;
3461 if (Value *X = foldUnsignedUnderflowCheck(RHS, LHS, IsAnd, Q, Builder))
3462 return X;
3463 }
3464
3465 // (icmp ne A, 0) | (icmp ne B, 0) --> (icmp ne (A|B), 0)
3466 // (icmp eq A, 0) & (icmp eq B, 0) --> (icmp eq (A|B), 0)
3467 // TODO: Remove this and below when foldLogOpOfMaskedICmps can handle undefs.
3468 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3469 PredL == PredR && match(LHS1, m_ZeroInt()) && match(RHS1, m_ZeroInt()) &&
3470 LHS0->getType() == RHS0->getType() &&
3471 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3472 Value *NewOr = Builder.CreateOr(LHS0, RHS0);
3473 return Builder.CreateICmp(PredL, NewOr,
3475 }
3476
3477 // (icmp ne A, -1) | (icmp ne B, -1) --> (icmp ne (A&B), -1)
3478 // (icmp eq A, -1) & (icmp eq B, -1) --> (icmp eq (A&B), -1)
3479 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3480 PredL == PredR && match(LHS1, m_AllOnes()) && match(RHS1, m_AllOnes()) &&
3481 LHS0->getType() == RHS0->getType() &&
3482 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3483 Value *NewAnd = Builder.CreateAnd(LHS0, RHS0);
3484 return Builder.CreateICmp(PredL, NewAnd,
3486 }
3487
3488 if (!IsLogical)
3489 if (Value *V =
3491 return V;
3492
3493 // This only handles icmp of constants: (icmp1 A, C1) | (icmp2 B, C2).
3494 if (!LHSC || !RHSC)
3495 return nullptr;
3496
3497 // (trunc x) == C1 & (and x, CA) == C2 -> (and x, CA|CMAX) == C1|C2
3498 // (trunc x) != C1 | (and x, CA) != C2 -> (and x, CA|CMAX) != C1|C2
3499 // where CMAX is the all ones value for the truncated type,
3500 // iff the lower bits of C2 and CA are zero.
3501 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3502 PredL == PredR && LHS->hasOneUse() && RHS->hasOneUse()) {
3503 Value *V;
3504 const APInt *AndC, *SmallC = nullptr, *BigC = nullptr;
3505
3506 // (trunc x) == C1 & (and x, CA) == C2
3507 // (and x, CA) == C2 & (trunc x) == C1
3508 if (match(RHS0, m_Trunc(m_Value(V))) &&
3509 match(LHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3510 SmallC = RHSC;
3511 BigC = LHSC;
3512 } else if (match(LHS0, m_Trunc(m_Value(V))) &&
3513 match(RHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3514 SmallC = LHSC;
3515 BigC = RHSC;
3516 }
3517
3518 if (SmallC && BigC) {
3519 unsigned BigBitSize = BigC->getBitWidth();
3520 unsigned SmallBitSize = SmallC->getBitWidth();
3521
3522 // Check that the low bits are zero.
3523 APInt Low = APInt::getLowBitsSet(BigBitSize, SmallBitSize);
3524 if ((Low & *AndC).isZero() && (Low & *BigC).isZero()) {
3525 Value *NewAnd = Builder.CreateAnd(V, Low | *AndC);
3526 APInt N = SmallC->zext(BigBitSize) | *BigC;
3527 Value *NewVal = ConstantInt::get(NewAnd->getType(), N);
3528 return Builder.CreateICmp(PredL, NewAnd, NewVal);
3529 }
3530 }
3531 }
3532
3533 // Match naive pattern (and its inverted form) for checking if two values
3534 // share same sign. An example of the pattern:
3535 // (icmp slt (X & Y), 0) | (icmp sgt (X | Y), -1) -> (icmp sgt (X ^ Y), -1)
3536 // Inverted form (example):
3537 // (icmp slt (X | Y), 0) & (icmp sgt (X & Y), -1) -> (icmp slt (X ^ Y), 0)
3538 bool TrueIfSignedL, TrueIfSignedR;
3539 if (isSignBitCheck(PredL, *LHSC, TrueIfSignedL) &&
3540 isSignBitCheck(PredR, *RHSC, TrueIfSignedR) &&
3541 (RHS->hasOneUse() || LHS->hasOneUse())) {
3542 Value *X, *Y;
3543 if (IsAnd) {
3544 if ((TrueIfSignedL && !TrueIfSignedR &&
3545 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3546 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y)))) ||
3547 (!TrueIfSignedL && TrueIfSignedR &&
3548 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3549 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y))))) {
3550 Value *NewXor = Builder.CreateXor(X, Y);
3551 return Builder.CreateIsNeg(NewXor);
3552 }
3553 } else {
3554 if ((TrueIfSignedL && !TrueIfSignedR &&
3555 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3556 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y)))) ||
3557 (!TrueIfSignedL && TrueIfSignedR &&
3558 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3559 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y))))) {
3560 Value *NewXor = Builder.CreateXor(X, Y);
3561 return Builder.CreateIsNotNeg(NewXor);
3562 }
3563 }
3564 }
3565
3566 // (X & ExpMask) != 0 && (X & ExpMask) != ExpMask -> isnormal(X)
3567 // (X & ExpMask) == 0 || (X & ExpMask) == ExpMask -> !isnormal(X)
3568 Value *X;
3569 const APInt *MaskC;
3570 if (LHS0 == RHS0 && PredL == PredR &&
3571 PredL == (IsAnd ? ICmpInst::ICMP_NE : ICmpInst::ICMP_EQ) &&
3572 !I.getFunction()->hasFnAttribute(Attribute::NoImplicitFloat) &&
3573 LHS->hasOneUse() && RHS->hasOneUse() &&
3574 match(LHS0, m_And(m_ElementWiseBitCast(m_Value(X)), m_APInt(MaskC))) &&
3575 X->getType()->getScalarType()->isIEEELikeFPTy() &&
3576 APFloat(X->getType()->getScalarType()->getFltSemantics(), *MaskC)
3577 .isPosInfinity() &&
3578 ((LHSC->isZero() && *RHSC == *MaskC) ||
3579 (RHSC->isZero() && *LHSC == *MaskC)))
3580 return Builder.createIsFPClass(X, IsAnd ? FPClassTest::fcNormal
3582
3583 return foldAndOrOfICmpsUsingRanges(LHS, RHS, IsAnd);
3584}
3585
3586/// If IsLogical is true, then the and/or is in select form and the transform
3587/// must be poison-safe.
3588Value *InstCombinerImpl::foldBooleanAndOr(Value *LHS, Value *RHS,
3589 Instruction &I, bool IsAnd,
3590 bool IsLogical) {
3591 if (!LHS->getType()->isIntOrIntVectorTy(1))
3592 return nullptr;
3593
3594 // handle (roughly):
3595 // (icmp ne (A & B), C) | (icmp ne (A & D), E)
3596 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
3597 if (Value *V = foldLogOpOfMaskedICmps(LHS, RHS, IsAnd, IsLogical, Builder,
3598 SQ.getWithInstruction(&I)))
3599 return V;
3600
3601 if (auto *LHSCmp = dyn_cast<ICmpInst>(LHS))
3602 if (auto *RHSCmp = dyn_cast<ICmpInst>(RHS))
3603 if (Value *Res = foldAndOrOfICmps(LHSCmp, RHSCmp, I, IsAnd, IsLogical))
3604 return Res;
3605
3606 if (auto *LHSCmp = dyn_cast<FCmpInst>(LHS))
3607 if (auto *RHSCmp = dyn_cast<FCmpInst>(RHS))
3608 if (Value *Res = foldLogicOfFCmps(LHSCmp, RHSCmp, IsAnd, IsLogical))
3609 return Res;
3610
3611 if (Value *Res = foldEqOfParts(LHS, RHS, IsAnd))
3612 return Res;
3613
3614 return nullptr;
3615}
3616
3618 InstCombiner::BuilderTy &Builder) {
3619 assert(I.getOpcode() == Instruction::Or &&
3620 "Simplification only supports or at the moment.");
3621
3622 Value *Cmp1, *Cmp2, *Cmp3, *Cmp4;
3623 if (!match(I.getOperand(0), m_And(m_Value(Cmp1), m_Value(Cmp2))) ||
3624 !match(I.getOperand(1), m_And(m_Value(Cmp3), m_Value(Cmp4))))
3625 return nullptr;
3626
3627 // Check if any two pairs of the and operations are inversions of each other.
3628 if (isKnownInversion(Cmp1, Cmp3) && isKnownInversion(Cmp2, Cmp4))
3629 return Builder.CreateXor(Cmp1, Cmp4);
3630 if (isKnownInversion(Cmp1, Cmp4) && isKnownInversion(Cmp2, Cmp3))
3631 return Builder.CreateXor(Cmp1, Cmp3);
3632
3633 return nullptr;
3634}
3635
3636/// Match \p V as "shufflevector -> bitcast" or "extractelement -> zext -> shl"
3637/// patterns, which extract vector elements and pack them in the same relative
3638/// positions.
3639///
3640/// \p Vec is the underlying vector being extracted from.
3641/// \p Mask is a bitmask identifying which packed elements are obtained from the
3642/// vector.
3643/// \p VecOffset is the vector element corresponding to index 0 of the
3644/// mask.
3646 int64_t &VecOffset,
3647 SmallBitVector &Mask,
3648 const DataLayout &DL) {
3649 // First try to match extractelement -> zext -> shl
3650 uint64_t VecIdx, ShlAmt;
3652 m_ConstantInt(VecIdx))),
3653 ShlAmt))) {
3654 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3655 if (!VecTy)
3656 return false;
3657 auto *EltTy = dyn_cast<IntegerType>(VecTy->getElementType());
3658 if (!EltTy)
3659 return false;
3660
3661 const unsigned EltBitWidth = EltTy->getBitWidth();
3662 const unsigned TargetBitWidth = V->getType()->getIntegerBitWidth();
3663 if (TargetBitWidth % EltBitWidth != 0 || ShlAmt % EltBitWidth != 0)
3664 return false;
3665 const unsigned TargetEltWidth = TargetBitWidth / EltBitWidth;
3666 const unsigned ShlEltAmt = ShlAmt / EltBitWidth;
3667
3668 const unsigned MaskIdx =
3669 DL.isLittleEndian() ? ShlEltAmt : TargetEltWidth - ShlEltAmt - 1;
3670
3671 VecOffset = static_cast<int64_t>(VecIdx) - static_cast<int64_t>(MaskIdx);
3672 Mask.resize(TargetEltWidth);
3673 Mask.set(MaskIdx);
3674 return true;
3675 }
3676
3677 // Now try to match a bitcasted subvector.
3678 Instruction *SrcVecI;
3679 if (!match(V, m_BitCast(m_Instruction(SrcVecI))))
3680 return false;
3681
3682 auto *SrcTy = dyn_cast<FixedVectorType>(SrcVecI->getType());
3683 if (!SrcTy)
3684 return false;
3685
3686 Mask.resize(SrcTy->getNumElements());
3687
3688 // First check for a subvector obtained from a shufflevector.
3689 if (isa<ShuffleVectorInst>(SrcVecI)) {
3690 Constant *ConstVec;
3691 ArrayRef<int> ShuffleMask;
3692 if (!match(SrcVecI, m_Shuffle(m_Value(Vec), m_Constant(ConstVec),
3693 m_Mask(ShuffleMask))))
3694 return false;
3695
3696 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3697 if (!VecTy)
3698 return false;
3699
3700 const unsigned NumVecElts = VecTy->getNumElements();
3701 bool FoundVecOffset = false;
3702 for (unsigned Idx = 0; Idx < ShuffleMask.size(); ++Idx) {
3703 if (ShuffleMask[Idx] == PoisonMaskElem)
3704 return false;
3705 const unsigned ShuffleIdx = ShuffleMask[Idx];
3706 if (ShuffleIdx >= NumVecElts) {
3707 const unsigned ConstIdx = ShuffleIdx - NumVecElts;
3708 auto *ConstElt =
3709 dyn_cast<ConstantInt>(ConstVec->getAggregateElement(ConstIdx));
3710 if (!ConstElt || !ConstElt->isNullValue())
3711 return false;
3712 continue;
3713 }
3714
3715 if (FoundVecOffset) {
3716 if (VecOffset + Idx != ShuffleIdx)
3717 return false;
3718 } else {
3719 if (ShuffleIdx < Idx)
3720 return false;
3721 VecOffset = ShuffleIdx - Idx;
3722 FoundVecOffset = true;
3723 }
3724 Mask.set(Idx);
3725 }
3726 return FoundVecOffset;
3727 }
3728
3729 // Check for a subvector obtained as an (insertelement V, 0, idx)
3730 uint64_t InsertIdx;
3731 if (!match(SrcVecI,
3732 m_InsertElt(m_Value(Vec), m_Zero(), m_ConstantInt(InsertIdx))))
3733 return false;
3734
3735 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3736 if (!VecTy)
3737 return false;
3738 VecOffset = 0;
3739 bool AlreadyInsertedMaskedElt = Mask.test(InsertIdx);
3740 Mask.set();
3741 if (!AlreadyInsertedMaskedElt)
3742 Mask.reset(InsertIdx);
3743 return true;
3744}
3745
3746/// Try to fold the join of two scalar integers whose contents are packed
3747/// elements of the same vector.
3749 InstCombiner::BuilderTy &Builder,
3750 const DataLayout &DL) {
3751 assert(I.getOpcode() == Instruction::Or);
3752 Value *LhsVec, *RhsVec;
3753 int64_t LhsVecOffset, RhsVecOffset;
3754 SmallBitVector Mask;
3755 if (!matchSubIntegerPackFromVector(I.getOperand(0), LhsVec, LhsVecOffset,
3756 Mask, DL))
3757 return nullptr;
3758 if (!matchSubIntegerPackFromVector(I.getOperand(1), RhsVec, RhsVecOffset,
3759 Mask, DL))
3760 return nullptr;
3761 if (LhsVec != RhsVec || LhsVecOffset != RhsVecOffset)
3762 return nullptr;
3763
3764 // Convert into shufflevector -> bitcast;
3765 const unsigned ZeroVecIdx =
3766 cast<FixedVectorType>(LhsVec->getType())->getNumElements();
3767 SmallVector<int> ShuffleMask(Mask.size(), ZeroVecIdx);
3768 for (unsigned Idx : Mask.set_bits()) {
3769 assert(LhsVecOffset + Idx >= 0);
3770 ShuffleMask[Idx] = LhsVecOffset + Idx;
3771 }
3772
3773 Value *MaskedVec = Builder.CreateShuffleVector(
3774 LhsVec, Constant::getNullValue(LhsVec->getType()), ShuffleMask,
3775 I.getName() + ".v");
3776 return CastInst::Create(Instruction::BitCast, MaskedVec, I.getType());
3777}
3778
3779/// Match \p V as "lshr -> mask -> zext -> shl".
3780///
3781/// \p Int is the underlying integer being extracted from.
3782/// \p Mask is a bitmask identifying which bits of the integer are being
3783/// extracted. \p Offset identifies which bit of the result \p V corresponds to
3784/// the least significant bit of \p Int
3785static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask,
3786 uint64_t &Offset, bool &IsShlNUW,
3787 bool &IsShlNSW) {
3788 Value *ShlOp0;
3789 uint64_t ShlAmt = 0;
3790 if (!match(V, m_OneUse(m_Shl(m_Value(ShlOp0), m_ConstantInt(ShlAmt)))))
3791 return false;
3792
3793 IsShlNUW = cast<BinaryOperator>(V)->hasNoUnsignedWrap();
3794 IsShlNSW = cast<BinaryOperator>(V)->hasNoSignedWrap();
3795
3796 Value *ZExtOp0;
3797 if (!match(ShlOp0, m_OneUse(m_ZExt(m_Value(ZExtOp0)))))
3798 return false;
3799
3800 Value *MaskedOp0;
3801 const APInt *ShiftedMaskConst = nullptr;
3802 if (!match(ZExtOp0, m_CombineOr(m_OneUse(m_And(m_Value(MaskedOp0),
3803 m_APInt(ShiftedMaskConst))),
3804 m_Value(MaskedOp0))))
3805 return false;
3806
3807 uint64_t LShrAmt = 0;
3808 if (!match(MaskedOp0,
3810 m_Value(Int))))
3811 return false;
3812
3813 if (LShrAmt > ShlAmt)
3814 return false;
3815 Offset = ShlAmt - LShrAmt;
3816
3817 Mask = ShiftedMaskConst ? ShiftedMaskConst->shl(LShrAmt)
3819 Int->getType()->getScalarSizeInBits(), LShrAmt);
3820
3821 return true;
3822}
3823
3824/// Try to fold the join of two scalar integers whose bits are unpacked and
3825/// zexted from the same source integer.
3827 InstCombiner::BuilderTy &Builder) {
3828
3829 Value *LhsInt, *RhsInt;
3830 APInt LhsMask, RhsMask;
3831 uint64_t LhsOffset, RhsOffset;
3832 bool IsLhsShlNUW, IsLhsShlNSW, IsRhsShlNUW, IsRhsShlNSW;
3833 if (!matchZExtedSubInteger(Lhs, LhsInt, LhsMask, LhsOffset, IsLhsShlNUW,
3834 IsLhsShlNSW))
3835 return nullptr;
3836 if (!matchZExtedSubInteger(Rhs, RhsInt, RhsMask, RhsOffset, IsRhsShlNUW,
3837 IsRhsShlNSW))
3838 return nullptr;
3839 if (LhsInt != RhsInt || LhsOffset != RhsOffset)
3840 return nullptr;
3841
3842 APInt Mask = LhsMask | RhsMask;
3843
3844 Type *DestTy = Lhs->getType();
3845 Value *Res = Builder.CreateShl(
3846 Builder.CreateZExt(
3847 Builder.CreateAnd(LhsInt, Mask, LhsInt->getName() + ".mask"), DestTy,
3848 LhsInt->getName() + ".zext"),
3849 ConstantInt::get(DestTy, LhsOffset), "", IsLhsShlNUW && IsRhsShlNUW,
3850 IsLhsShlNSW && IsRhsShlNSW);
3851 Res->takeName(Lhs);
3852 return Res;
3853}
3854
3855// A decomposition of ((X & Mask) * Factor). The NUW / NSW bools
3856// track these properities for preservation. Note that we can decompose
3857// equivalent select form of this expression (e.g. (!(X & Mask) ? 0 : Mask *
3858// Factor))
3863 bool NUW;
3864 bool NSW;
3865
3867 return X == Other.X && !Mask.intersects(Other.Mask) &&
3868 Factor == Other.Factor;
3869 }
3870};
3871
3872static std::optional<DecomposedBitMaskMul> matchBitmaskMul(Value *V) {
3874 if (!Op)
3875 return std::nullopt;
3876
3877 // Decompose (A & N) * C) into BitMaskMul
3878 Value *Original = nullptr;
3879 const APInt *Mask = nullptr;
3880 const APInt *MulConst = nullptr;
3881 if (match(Op, m_Mul(m_And(m_Value(Original), m_APInt(Mask)),
3882 m_APInt(MulConst)))) {
3883 if (MulConst->isZero() || Mask->isZero())
3884 return std::nullopt;
3885
3886 return std::optional<DecomposedBitMaskMul>(
3887 {Original, *MulConst, *Mask,
3888 cast<BinaryOperator>(Op)->hasNoUnsignedWrap(),
3889 cast<BinaryOperator>(Op)->hasNoSignedWrap()});
3890 }
3891
3892 Value *Cond = nullptr;
3893 const APInt *EqZero = nullptr, *NeZero = nullptr;
3894
3895 // Decompose ((A & N) ? 0 : N * C) into BitMaskMul
3896 if (match(Op, m_Select(m_Value(Cond), m_APInt(EqZero), m_APInt(NeZero)))) {
3897 auto ICmpDecompose =
3898 decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
3899 /*AllowNonZeroC=*/false, /*DecomposeBitMask=*/true);
3900 if (!ICmpDecompose.has_value())
3901 return std::nullopt;
3902
3903 assert(ICmpInst::isEquality(ICmpDecompose->Pred) &&
3904 ICmpDecompose->C.isZero());
3905
3906 if (ICmpDecompose->Pred == ICmpInst::ICMP_NE)
3907 std::swap(EqZero, NeZero);
3908
3909 if (!EqZero->isZero() || NeZero->isZero())
3910 return std::nullopt;
3911
3912 if (!ICmpDecompose->Mask.isPowerOf2() || ICmpDecompose->Mask.isZero() ||
3913 NeZero->getBitWidth() != ICmpDecompose->Mask.getBitWidth())
3914 return std::nullopt;
3915
3916 if (!NeZero->urem(ICmpDecompose->Mask).isZero())
3917 return std::nullopt;
3918
3919 return std::optional<DecomposedBitMaskMul>(
3920 {ICmpDecompose->X, NeZero->udiv(ICmpDecompose->Mask),
3921 ICmpDecompose->Mask, /*NUW=*/false, /*NSW=*/false});
3922 }
3923
3924 return std::nullopt;
3925}
3926
3927/// (A & N) * C + (A & M) * C -> (A & (N + M)) & C
3928/// This also accepts the equivalent select form of (A & N) * C
3929/// expressions i.e. !(A & N) ? 0 : N * C)
3930static Value *foldBitmaskMul(Value *Op0, Value *Op1,
3931 InstCombiner::BuilderTy &Builder) {
3932 auto Decomp1 = matchBitmaskMul(Op1);
3933 if (!Decomp1)
3934 return nullptr;
3935
3936 auto Decomp0 = matchBitmaskMul(Op0);
3937 if (!Decomp0)
3938 return nullptr;
3939
3940 if (Decomp0->isCombineableWith(*Decomp1)) {
3941 Value *NewAnd = Builder.CreateAnd(
3942 Decomp0->X,
3943 ConstantInt::get(Decomp0->X->getType(), Decomp0->Mask + Decomp1->Mask));
3944
3945 return Builder.CreateMul(
3946 NewAnd, ConstantInt::get(NewAnd->getType(), Decomp1->Factor), "",
3947 Decomp0->NUW && Decomp1->NUW, Decomp0->NSW && Decomp1->NSW);
3948 }
3949
3950 return nullptr;
3951}
3952
3953Value *InstCombinerImpl::foldDisjointOr(Value *LHS, Value *RHS) {
3954 if (Value *Res = foldBitmaskMul(LHS, RHS, Builder))
3955 return Res;
3957 return Res;
3958
3959 return nullptr;
3960}
3961
3962Value *InstCombinerImpl::reassociateDisjointOr(Value *LHS, Value *RHS) {
3963
3964 Value *X, *Y;
3966 if (Value *Res = foldDisjointOr(LHS, X))
3967 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3968 if (Value *Res = foldDisjointOr(LHS, Y))
3969 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3970 }
3971
3973 if (Value *Res = foldDisjointOr(X, RHS))
3974 return Builder.CreateOr(Res, Y, "", /*IsDisjoint=*/true);
3975 if (Value *Res = foldDisjointOr(Y, RHS))
3976 return Builder.CreateOr(Res, X, "", /*IsDisjoint=*/true);
3977 }
3978
3979 return nullptr;
3980}
3981
3982/// Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2))
3983/// --> (ugt x (c2/c1)). This code checks whether a multiplication of two
3984/// unsigned numbers (one is a constant) is mathematically greater than a
3985/// second constant.
3987 InstCombiner::BuilderTy &Builder,
3988 const DataLayout &DL) {
3989 Value *WOV, *X;
3990 const APInt *C1, *C2;
3991 if (match(&I,
3994 m_Value(X), m_APInt(C1)))),
3997 m_APInt(C2))))) &&
3998 !C1->isZero()) {
3999 Constant *NewC = ConstantInt::get(X->getType(), C2->udiv(*C1));
4000 return Builder.CreateICmp(ICmpInst::ICMP_UGT, X, NewC);
4001 }
4002 return nullptr;
4003}
4004
4005/// Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X)
4006/// select(X <s 0, -X, 0) | smax(X, 0) --> abs(X)
4008 InstCombiner::BuilderTy &Builder) {
4009 Value *X;
4010 Value *Sel;
4011 if (match(&I,
4013 auto NegX = m_Neg(m_Specific(X));
4015 m_ZeroInt()),
4016 m_ZeroInt(), NegX)) ||
4018 m_ZeroInt()),
4019 NegX, m_ZeroInt())))
4020 return Builder.CreateBinaryIntrinsic(Intrinsic::abs, X,
4021 Builder.getFalse());
4022 }
4023 return nullptr;
4024}
4025
4026// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
4027// here. We should standardize that construct where it is needed or choose some
4028// other way to ensure that commutated variants of patterns are not missed.
4030 if (Value *V = simplifyOrInst(I.getOperand(0), I.getOperand(1),
4031 SQ.getWithInstruction(&I)))
4032 return replaceInstUsesWith(I, V);
4033
4035 return &I;
4036
4038 return X;
4039
4041 return Phi;
4042
4043 // See if we can simplify any instructions used by the instruction whose sole
4044 // purpose is to compute bits we don't care about.
4046 return &I;
4047
4048 // Do this before using distributive laws to catch simple and/or/not patterns.
4050 return Xor;
4051
4053 return X;
4054
4056 return X;
4057
4058 // (A & B) | (C & D) -> A ^ D where A == ~C && B == ~D
4059 // (A & B) | (C & D) -> A ^ C where A == ~D && B == ~C
4060 if (Value *V = foldOrOfInversions(I, Builder))
4061 return replaceInstUsesWith(I, V);
4062
4063 // (A&B)|(A&C) -> A&(B|C) etc
4065 return replaceInstUsesWith(I, V);
4066
4067 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
4068 Type *Ty = I.getType();
4069 if (Ty->isIntOrIntVectorTy(1)) {
4070 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
4071 if (auto *R =
4072 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ false))
4073 return R;
4074 }
4075 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
4076 if (auto *R =
4077 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ false))
4078 return R;
4079 }
4080 }
4081
4082 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
4083 return FoldedLogic;
4084
4085 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
4086 return FoldedLogic;
4087
4088 if (Instruction *BitOp = matchBSwapOrBitReverse(I, /*MatchBSwaps*/ true,
4089 /*MatchBitReversals*/ true))
4090 return BitOp;
4091
4092 if (Instruction *Funnel = matchFunnelShift(I, *this))
4093 return Funnel;
4094
4096 return replaceInstUsesWith(I, Concat);
4097
4099 return R;
4100
4102 return R;
4103
4104 if (cast<PossiblyDisjointInst>(I).isDisjoint()) {
4105 if (Instruction *R =
4106 foldAddLikeCommutative(I.getOperand(0), I.getOperand(1),
4107 /*NSW=*/true, /*NUW=*/true))
4108 return R;
4109 if (Instruction *R =
4110 foldAddLikeCommutative(I.getOperand(1), I.getOperand(0),
4111 /*NSW=*/true, /*NUW=*/true))
4112 return R;
4113
4114 if (Value *Res = foldDisjointOr(I.getOperand(0), I.getOperand(1)))
4115 return replaceInstUsesWith(I, Res);
4116
4117 if (Value *Res = reassociateDisjointOr(I.getOperand(0), I.getOperand(1)))
4118 return replaceInstUsesWith(I, Res);
4119 }
4120
4121 Value *X, *Y;
4122 const APInt *CV;
4123 if (match(&I, m_c_Or(m_OneUse(m_Xor(m_Value(X), m_APInt(CV))), m_Value(Y))) &&
4124 !CV->isAllOnes() && MaskedValueIsZero(Y, *CV, &I)) {
4125 // (X ^ C) | Y -> (X | Y) ^ C iff Y & C == 0
4126 // The check for a 'not' op is for efficiency (if Y is known zero --> ~X).
4127 Value *Or = Builder.CreateOr(X, Y);
4128 return BinaryOperator::CreateXor(Or, ConstantInt::get(Ty, *CV));
4129 }
4130
4131 // If the operands have no common bits set:
4132 // or (mul X, Y), X --> add (mul X, Y), X --> mul X, (Y + 1)
4134 m_Deferred(X)))) {
4135 Value *IncrementY = Builder.CreateAdd(Y, ConstantInt::get(Ty, 1));
4136 return BinaryOperator::CreateMul(X, IncrementY);
4137 }
4138
4139 // (A & C) | (B & D)
4140 Value *A, *B, *C, *D;
4141 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4142 match(Op1, m_And(m_Value(B), m_Value(D)))) {
4143
4144 // (A & C0) | (B & C1)
4145 const APInt *C0, *C1;
4146 if (match(C, m_APInt(C0)) && match(D, m_APInt(C1))) {
4147 Value *X;
4148 if (*C0 == ~*C1) {
4149 // ((X | B) & MaskC) | (B & ~MaskC) -> (X & MaskC) | B
4150 if (match(A, m_c_Or(m_Value(X), m_Specific(B))))
4151 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C0), B);
4152 // (A & MaskC) | ((X | A) & ~MaskC) -> (X & ~MaskC) | A
4153 if (match(B, m_c_Or(m_Specific(A), m_Value(X))))
4154 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C1), A);
4155
4156 // ((X ^ B) & MaskC) | (B & ~MaskC) -> (X & MaskC) ^ B
4157 if (match(A, m_c_Xor(m_Value(X), m_Specific(B))))
4158 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C0), B);
4159 // (A & MaskC) | ((X ^ A) & ~MaskC) -> (X & ~MaskC) ^ A
4160 if (match(B, m_c_Xor(m_Specific(A), m_Value(X))))
4161 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C1), A);
4162 }
4163
4164 if ((*C0 & *C1).isZero()) {
4165 // ((X | B) & C0) | (B & C1) --> (X | B) & (C0 | C1)
4166 // iff (C0 & C1) == 0 and (X & ~C0) == 0
4167 if (match(A, m_c_Or(m_Value(X), m_Specific(B))) &&
4168 MaskedValueIsZero(X, ~*C0, &I)) {
4169 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4170 return BinaryOperator::CreateAnd(A, C01);
4171 }
4172 // (A & C0) | ((X | A) & C1) --> (X | A) & (C0 | C1)
4173 // iff (C0 & C1) == 0 and (X & ~C1) == 0
4174 if (match(B, m_c_Or(m_Value(X), m_Specific(A))) &&
4175 MaskedValueIsZero(X, ~*C1, &I)) {
4176 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4177 return BinaryOperator::CreateAnd(B, C01);
4178 }
4179 // ((X | C2) & C0) | ((X | C3) & C1) --> (X | C2 | C3) & (C0 | C1)
4180 // iff (C0 & C1) == 0 and (C2 & ~C0) == 0 and (C3 & ~C1) == 0.
4181 const APInt *C2, *C3;
4182 if (match(A, m_Or(m_Value(X), m_APInt(C2))) &&
4183 match(B, m_Or(m_Specific(X), m_APInt(C3))) &&
4184 (*C2 & ~*C0).isZero() && (*C3 & ~*C1).isZero()) {
4185 Value *Or = Builder.CreateOr(X, *C2 | *C3, "bitfield");
4186 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4187 return BinaryOperator::CreateAnd(Or, C01);
4188 }
4189 }
4190 }
4191
4192 // Don't try to form a select if it's unlikely that we'll get rid of at
4193 // least one of the operands. A select is generally more expensive than the
4194 // 'or' that it is replacing.
4195 if (Op0->hasOneUse() || Op1->hasOneUse()) {
4196 // (Cond & C) | (~Cond & D) -> Cond ? C : D, and commuted variants.
4197 if (Value *V = matchSelectFromAndOr(A, C, B, D))
4198 return replaceInstUsesWith(I, V);
4199 if (Value *V = matchSelectFromAndOr(A, C, D, B))
4200 return replaceInstUsesWith(I, V);
4201 if (Value *V = matchSelectFromAndOr(C, A, B, D))
4202 return replaceInstUsesWith(I, V);
4203 if (Value *V = matchSelectFromAndOr(C, A, D, B))
4204 return replaceInstUsesWith(I, V);
4205 if (Value *V = matchSelectFromAndOr(B, D, A, C))
4206 return replaceInstUsesWith(I, V);
4207 if (Value *V = matchSelectFromAndOr(B, D, C, A))
4208 return replaceInstUsesWith(I, V);
4209 if (Value *V = matchSelectFromAndOr(D, B, A, C))
4210 return replaceInstUsesWith(I, V);
4211 if (Value *V = matchSelectFromAndOr(D, B, C, A))
4212 return replaceInstUsesWith(I, V);
4213 }
4214 }
4215
4216 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4217 match(Op1, m_Not(m_Or(m_Value(B), m_Value(D)))) &&
4218 (Op0->hasOneUse() || Op1->hasOneUse())) {
4219 // (Cond & C) | ~(Cond | D) -> Cond ? C : ~D
4220 if (Value *V = matchSelectFromAndOr(A, C, B, D, true))
4221 return replaceInstUsesWith(I, V);
4222 if (Value *V = matchSelectFromAndOr(A, C, D, B, true))
4223 return replaceInstUsesWith(I, V);
4224 if (Value *V = matchSelectFromAndOr(C, A, B, D, true))
4225 return replaceInstUsesWith(I, V);
4226 if (Value *V = matchSelectFromAndOr(C, A, D, B, true))
4227 return replaceInstUsesWith(I, V);
4228 }
4229
4230 // (A ^ B) | ((B ^ C) ^ A) -> (A ^ B) | C
4231 if (match(Op0, m_Xor(m_Value(A), m_Value(B))))
4232 if (match(Op1,
4235 return BinaryOperator::CreateOr(Op0, C);
4236
4237 // ((B ^ C) ^ A) | (A ^ B) -> (A ^ B) | C
4238 if (match(Op1, m_Xor(m_Value(A), m_Value(B))))
4239 if (match(Op0,
4242 return BinaryOperator::CreateOr(Op1, C);
4243
4244 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
4245 return DeMorgan;
4246
4247 // Canonicalize xor to the RHS.
4248 bool SwappedForXor = false;
4249 if (match(Op0, m_Xor(m_Value(), m_Value()))) {
4250 std::swap(Op0, Op1);
4251 SwappedForXor = true;
4252 }
4253
4254 if (match(Op1, m_Xor(m_Value(A), m_Value(B)))) {
4255 // (A | ?) | (A ^ B) --> (A | ?) | B
4256 // (B | ?) | (A ^ B) --> (B | ?) | A
4257 if (match(Op0, m_c_Or(m_Specific(A), m_Value())))
4258 return BinaryOperator::CreateOr(Op0, B);
4259 if (match(Op0, m_c_Or(m_Specific(B), m_Value())))
4260 return BinaryOperator::CreateOr(Op0, A);
4261
4262 // (A & B) | (A ^ B) --> A | B
4263 // (B & A) | (A ^ B) --> A | B
4264 if (match(Op0, m_c_And(m_Specific(A), m_Specific(B))))
4265 return BinaryOperator::CreateOr(A, B);
4266
4267 // ~A | (A ^ B) --> ~(A & B)
4268 // ~B | (A ^ B) --> ~(A & B)
4269 // The swap above should always make Op0 the 'not'.
4270 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4271 (match(Op0, m_Not(m_Specific(A))) || match(Op0, m_Not(m_Specific(B)))))
4272 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
4273
4274 // Same as above, but peek through an 'and' to the common operand:
4275 // ~(A & ?) | (A ^ B) --> ~((A & ?) & B)
4276 // ~(B & ?) | (A ^ B) --> ~((B & ?) & A)
4278 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4279 match(Op0,
4281 return BinaryOperator::CreateNot(Builder.CreateAnd(And, B));
4282 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4283 match(Op0,
4285 return BinaryOperator::CreateNot(Builder.CreateAnd(And, A));
4286
4287 // (~A | C) | (A ^ B) --> ~(A & B) | C
4288 // (~B | C) | (A ^ B) --> ~(A & B) | C
4289 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4290 (match(Op0, m_c_Or(m_Not(m_Specific(A)), m_Value(C))) ||
4291 match(Op0, m_c_Or(m_Not(m_Specific(B)), m_Value(C))))) {
4292 Value *Nand = Builder.CreateNot(Builder.CreateAnd(A, B), "nand");
4293 return BinaryOperator::CreateOr(Nand, C);
4294 }
4295 }
4296
4297 if (SwappedForXor)
4298 std::swap(Op0, Op1);
4299
4300 if (Value *Res =
4301 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/false, /*IsLogical=*/false))
4302 return replaceInstUsesWith(I, Res);
4303
4304 if (match(Op1, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4305 bool IsLogical = isa<SelectInst>(Op1);
4306 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/false,
4307 /*RHSIsLogical=*/IsLogical))
4308 return replaceInstUsesWith(I, V);
4309 }
4310 if (match(Op0, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4311 bool IsLogical = isa<SelectInst>(Op0);
4312 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/false,
4313 /*RHSIsLogical=*/IsLogical))
4314 return replaceInstUsesWith(I, V);
4315 }
4316
4317 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
4318 return FoldedFCmps;
4319
4320 if (Instruction *CastedOr = foldCastedBitwiseLogic(I))
4321 return CastedOr;
4322
4323 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
4324 return Sel;
4325
4326 // or(sext(A), B) / or(B, sext(A)) --> A ? -1 : B, where A is i1 or <N x i1>.
4327 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
4328 // with binop identity constant. But creating a select with non-constant
4329 // arm may not be reversible due to poison semantics. Is that a good
4330 // canonicalization?
4331 if (match(&I, m_c_Or(m_OneUse(m_SExt(m_Value(A))), m_Value(B))) &&
4332 A->getType()->isIntOrIntVectorTy(1))
4333 return createSelectInstWithUnknownProfile(
4335
4336 // Note: If we've gotten to the point of visiting the outer OR, then the
4337 // inner one couldn't be simplified. If it was a constant, then it won't
4338 // be simplified by a later pass either, so we try swapping the inner/outer
4339 // ORs in the hopes that we'll be able to simplify it this way.
4340 // (X|C) | V --> (X|V) | C
4341 // Pass the disjoint flag in the following two patterns:
4342 // 1. or-disjoint (or-disjoint X, C), V -->
4343 // or-disjoint (or-disjoint X, V), C
4344 //
4345 // 2. or-disjoint (or X, C), V -->
4346 // or (or-disjoint X, V), C
4347 ConstantInt *CI;
4348 if (Op0->hasOneUse() && !match(Op1, m_ConstantInt()) &&
4349 match(Op0, m_Or(m_Value(A), m_ConstantInt(CI)))) {
4350 bool IsDisjointOuter = cast<PossiblyDisjointInst>(I).isDisjoint();
4351 bool IsDisjointInner = cast<PossiblyDisjointInst>(Op0)->isDisjoint();
4352 Value *Inner = Builder.CreateOr(A, Op1);
4353 cast<PossiblyDisjointInst>(Inner)->setIsDisjoint(IsDisjointOuter);
4354 Inner->takeName(Op0);
4355 return IsDisjointOuter && IsDisjointInner
4356 ? BinaryOperator::CreateDisjointOr(Inner, CI)
4357 : BinaryOperator::CreateOr(Inner, CI);
4358 }
4359
4360 // Change (or (bool?A:B),(bool?C:D)) --> (bool?(or A,C):(or B,D))
4361 // Since this OR statement hasn't been optimized further yet, we hope
4362 // that this transformation will allow the new ORs to be optimized.
4363 {
4364 Value *X = nullptr, *Y = nullptr;
4365 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4366 match(Op0, m_Select(m_Value(X), m_Value(A), m_Value(B))) &&
4367 match(Op1, m_Select(m_Value(Y), m_Value(C), m_Value(D))) && X == Y) {
4368 Value *orTrue = Builder.CreateOr(A, C);
4369 Value *orFalse = Builder.CreateOr(B, D);
4370 return SelectInst::Create(X, orTrue, orFalse);
4371 }
4372 }
4373
4374 // or(ashr(subNSW(Y, X), ScalarSizeInBits(Y) - 1), X) --> X s> Y ? -1 : X.
4375 {
4376 Value *X, *Y;
4379 m_SpecificInt(Ty->getScalarSizeInBits() - 1))),
4380 m_Deferred(X)))) {
4381 Value *NewICmpInst = Builder.CreateICmpSGT(X, Y);
4383 return createSelectInstWithUnknownProfile(NewICmpInst, AllOnes, X);
4384 }
4385 }
4386
4387 {
4388 // ((A & B) ^ A) | ((A & B) ^ B) -> A ^ B
4389 // (A ^ (A & B)) | (B ^ (A & B)) -> A ^ B
4390 // ((A & B) ^ B) | ((A & B) ^ A) -> A ^ B
4391 // (B ^ (A & B)) | (A ^ (A & B)) -> A ^ B
4392 const auto TryXorOpt = [&](Value *Lhs, Value *Rhs) -> Instruction * {
4393 if (match(Lhs, m_c_Xor(m_And(m_Value(A), m_Value(B)), m_Deferred(A))) &&
4394 match(Rhs,
4396 return BinaryOperator::CreateXor(A, B);
4397 }
4398 return nullptr;
4399 };
4400
4401 if (Instruction *Result = TryXorOpt(Op0, Op1))
4402 return Result;
4403 if (Instruction *Result = TryXorOpt(Op1, Op0))
4404 return Result;
4405 }
4406
4407 if (Instruction *V =
4409 return V;
4410
4411 CmpPredicate Pred;
4412 Value *Mul, *Ov, *MulIsNotZero, *UMulWithOv;
4413 // Check if the OR weakens the overflow condition for umul.with.overflow by
4414 // treating any non-zero result as overflow. In that case, we overflow if both
4415 // umul.with.overflow operands are != 0, as in that case the result can only
4416 // be 0, iff the multiplication overflows.
4417 if (match(&I, m_c_Or(m_Value(Ov, m_ExtractValue<1>(m_Value(UMulWithOv))),
4418 m_Value(MulIsNotZero,
4422 m_Deferred(UMulWithOv))),
4423 m_ZeroInt())))) &&
4424 (Ov->hasOneUse() || (MulIsNotZero->hasOneUse() && Mul->hasOneUse()))) {
4425 Value *A, *B;
4427 m_Value(A), m_Value(B)))) {
4428 Value *NotNullA = Builder.CreateIsNotNull(A);
4429 Value *NotNullB = Builder.CreateIsNotNull(B);
4430 return BinaryOperator::CreateAnd(NotNullA, NotNullB);
4431 }
4432 }
4433
4434 /// Res, Overflow = xxx_with_overflow X, C1
4435 /// Try to canonicalize the pattern "Overflow | icmp pred Res, C2" into
4436 /// "Overflow | icmp pred X, C2 +/- C1".
4437 const WithOverflowInst *WO;
4438 const Value *WOV;
4439 const APInt *C1, *C2;
4441 m_Value(WOV, m_WithOverflowInst(WO)))),
4443 m_APInt(C2))))) &&
4444 (WO->getBinaryOp() == Instruction::Add ||
4445 WO->getBinaryOp() == Instruction::Sub) &&
4446 (ICmpInst::isEquality(Pred) ||
4447 WO->isSigned() == ICmpInst::isSigned(Pred)) &&
4448 match(WO->getRHS(), m_APInt(C1))) {
4449 bool Overflow;
4450 APInt NewC = WO->getBinaryOp() == Instruction::Add
4451 ? (ICmpInst::isSigned(Pred) ? C2->ssub_ov(*C1, Overflow)
4452 : C2->usub_ov(*C1, Overflow))
4453 : (ICmpInst::isSigned(Pred) ? C2->sadd_ov(*C1, Overflow)
4454 : C2->uadd_ov(*C1, Overflow));
4455 if (!Overflow || ICmpInst::isEquality(Pred)) {
4456 Value *NewCmp = Builder.CreateICmp(
4457 Pred, WO->getLHS(), ConstantInt::get(WO->getLHS()->getType(), NewC));
4458 return BinaryOperator::CreateOr(Ov, NewCmp);
4459 }
4460 }
4461
4462 // Try to fold the pattern "Overflow | icmp pred Res, C2" into a single
4463 // comparison instruction for umul.with.overflow.
4465 return replaceInstUsesWith(I, R);
4466
4467 // (~x) | y --> ~(x & (~y)) iff that gets rid of inversions
4469 return &I;
4470
4471 // Improve "get low bit mask up to and including bit X" pattern:
4472 // (1 << X) | ((1 << X) + -1) --> -1 l>> (bitwidth(x) - 1 - X)
4473 if (match(&I, m_c_Or(m_Add(m_Shl(m_One(), m_Value(X)), m_AllOnes()),
4474 m_Shl(m_One(), m_Deferred(X)))) &&
4475 match(&I, m_c_Or(m_OneUse(m_Value()), m_Value()))) {
4476 Value *Sub = Builder.CreateSub(
4477 ConstantInt::get(Ty, Ty->getScalarSizeInBits() - 1), X);
4478 return BinaryOperator::CreateLShr(Constant::getAllOnesValue(Ty), Sub);
4479 }
4480
4481 // An or recurrence w/loop invariant step is equivelent to (or start, step)
4482 PHINode *PN = nullptr;
4483 Value *Start = nullptr, *Step = nullptr;
4484 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
4485 return replaceInstUsesWith(I, Builder.CreateOr(Start, Step));
4486
4487 // (A & B) | (C | D) or (C | D) | (A & B)
4488 // Can be combined if C or D is of type (A/B & X)
4490 m_OneUse(m_Or(m_Value(C), m_Value(D)))))) {
4491 // (A & B) | (C | ?) -> C | (? | (A & B))
4492 // (A & B) | (C | ?) -> C | (? | (A & B))
4493 // (A & B) | (C | ?) -> C | (? | (A & B))
4494 // (A & B) | (C | ?) -> C | (? | (A & B))
4495 // (C | ?) | (A & B) -> C | (? | (A & B))
4496 // (C | ?) | (A & B) -> C | (? | (A & B))
4497 // (C | ?) | (A & B) -> C | (? | (A & B))
4498 // (C | ?) | (A & B) -> C | (? | (A & B))
4499 if (match(D, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4501 return BinaryOperator::CreateOr(
4502 C, Builder.CreateOr(D, Builder.CreateAnd(A, B)));
4503 // (A & B) | (? | D) -> (? | (A & B)) | D
4504 // (A & B) | (? | D) -> (? | (A & B)) | D
4505 // (A & B) | (? | D) -> (? | (A & B)) | D
4506 // (A & B) | (? | D) -> (? | (A & B)) | D
4507 // (? | D) | (A & B) -> (? | (A & B)) | D
4508 // (? | D) | (A & B) -> (? | (A & B)) | D
4509 // (? | D) | (A & B) -> (? | (A & B)) | D
4510 // (? | D) | (A & B) -> (? | (A & B)) | D
4511 if (match(C, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4513 return BinaryOperator::CreateOr(
4514 Builder.CreateOr(C, Builder.CreateAnd(A, B)), D);
4515 }
4516
4518 return R;
4519
4520 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
4521 return Canonicalized;
4522
4523 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
4524 return Folded;
4525
4526 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
4527 return Res;
4528
4529 // If we are setting the sign bit of a floating-point value, convert
4530 // this to fneg(fabs), then cast back to integer.
4531 //
4532 // If the result isn't immediately cast back to a float, this will increase
4533 // the number of instructions. This is still probably a better canonical form
4534 // as it enables FP value tracking.
4535 //
4536 // Assumes any IEEE-represented type has the sign bit in the high bit.
4537 //
4538 // This is generous interpretation of noimplicitfloat, this is not a true
4539 // floating-point operation.
4540 Value *CastOp;
4541 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
4542 match(Op1, m_SignMask()) &&
4543 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
4544 Attribute::NoImplicitFloat)) {
4545 Type *EltTy = CastOp->getType()->getScalarType();
4546 if (EltTy->isFloatingPointTy() &&
4548 Value *FAbs = Builder.CreateUnaryIntrinsic(Intrinsic::fabs, CastOp);
4549 Value *FNegFAbs = Builder.CreateFNeg(FAbs);
4550 return new BitCastInst(FNegFAbs, I.getType());
4551 }
4552 }
4553
4554 // (X & C1) | C2 -> X & (C1 | C2) iff (X & C2) == C2
4555 if (match(Op0, m_OneUse(m_And(m_Value(X), m_APInt(C1)))) &&
4556 match(Op1, m_APInt(C2))) {
4557 KnownBits KnownX = computeKnownBits(X, &I);
4558 if ((KnownX.One & *C2) == *C2)
4559 return BinaryOperator::CreateAnd(X, ConstantInt::get(Ty, *C1 | *C2));
4560 }
4561
4563 return Res;
4564
4565 if (Value *V =
4567 /*SimplifyOnly*/ false, *this))
4568 return BinaryOperator::CreateOr(V, Op1);
4569 if (Value *V =
4571 /*SimplifyOnly*/ false, *this))
4572 return BinaryOperator::CreateOr(Op0, V);
4573
4574 if (cast<PossiblyDisjointInst>(I).isDisjoint())
4576 return replaceInstUsesWith(I, V);
4577
4579 return replaceInstUsesWith(I, Res);
4580
4581 return nullptr;
4582}
4583
4584/// A ^ B can be specified using other logic ops in a variety of patterns. We
4585/// can fold these early and efficiently by morphing an existing instruction.
4587 InstCombiner::BuilderTy &Builder) {
4588 assert(I.getOpcode() == Instruction::Xor);
4589 Value *Op0 = I.getOperand(0);
4590 Value *Op1 = I.getOperand(1);
4591 Value *A, *B;
4592
4593 // There are 4 commuted variants for each of the basic patterns.
4594
4595 // (A & B) ^ (A | B) -> A ^ B
4596 // (A & B) ^ (B | A) -> A ^ B
4597 // (A | B) ^ (A & B) -> A ^ B
4598 // (A | B) ^ (B & A) -> A ^ B
4599 if (match(&I, m_c_Xor(m_And(m_Value(A), m_Value(B)),
4601 return BinaryOperator::CreateXor(A, B);
4602
4603 // (A | ~B) ^ (~A | B) -> A ^ B
4604 // (~B | A) ^ (~A | B) -> A ^ B
4605 // (~A | B) ^ (A | ~B) -> A ^ B
4606 // (B | ~A) ^ (A | ~B) -> A ^ B
4607 if (match(&I, m_Xor(m_c_Or(m_Value(A), m_Not(m_Value(B))),
4609 return BinaryOperator::CreateXor(A, B);
4610
4611 // (A & ~B) ^ (~A & B) -> A ^ B
4612 // (~B & A) ^ (~A & B) -> A ^ B
4613 // (~A & B) ^ (A & ~B) -> A ^ B
4614 // (B & ~A) ^ (A & ~B) -> A ^ B
4615 if (match(&I, m_Xor(m_c_And(m_Value(A), m_Not(m_Value(B))),
4617 return BinaryOperator::CreateXor(A, B);
4618
4619 // For the remaining cases we need to get rid of one of the operands.
4620 if (!Op0->hasOneUse() && !Op1->hasOneUse())
4621 return nullptr;
4622
4623 // (A | B) ^ ~(A & B) -> ~(A ^ B)
4624 // (A | B) ^ ~(B & A) -> ~(A ^ B)
4625 // (A & B) ^ ~(A | B) -> ~(A ^ B)
4626 // (A & B) ^ ~(B | A) -> ~(A ^ B)
4627 // Complexity sorting ensures the not will be on the right side.
4628 if ((match(Op0, m_Or(m_Value(A), m_Value(B))) &&
4629 match(Op1, m_Not(m_c_And(m_Specific(A), m_Specific(B))))) ||
4630 (match(Op0, m_And(m_Value(A), m_Value(B))) &&
4632 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
4633
4634 return nullptr;
4635}
4636
4637Value *InstCombinerImpl::foldXorOfICmps(ICmpInst *LHS, ICmpInst *RHS,
4638 BinaryOperator &I) {
4639 assert(I.getOpcode() == Instruction::Xor && I.getOperand(0) == LHS &&
4640 I.getOperand(1) == RHS && "Should be 'xor' with these operands");
4641
4642 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
4643 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
4644 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
4645
4646 if (predicatesFoldable(PredL, PredR)) {
4647 if (LHS0 == RHS1 && LHS1 == RHS0) {
4648 std::swap(LHS0, LHS1);
4649 PredL = ICmpInst::getSwappedPredicate(PredL);
4650 }
4651 if (LHS0 == RHS0 && LHS1 == RHS1) {
4652 // (icmp1 A, B) ^ (icmp2 A, B) --> (icmp3 A, B)
4653 unsigned Code = getICmpCode(PredL) ^ getICmpCode(PredR);
4654 bool IsSigned = LHS->isSigned() || RHS->isSigned();
4655 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
4656 }
4657 }
4658
4659 const APInt *LC, *RC;
4660 if (match(LHS1, m_APInt(LC)) && match(RHS1, m_APInt(RC)) &&
4661 LHS0->getType() == RHS0->getType() &&
4662 LHS0->getType()->isIntOrIntVectorTy()) {
4663 // Convert xor of signbit tests to signbit test of xor'd values:
4664 // (X > -1) ^ (Y > -1) --> (X ^ Y) < 0
4665 // (X < 0) ^ (Y < 0) --> (X ^ Y) < 0
4666 // (X > -1) ^ (Y < 0) --> (X ^ Y) > -1
4667 // (X < 0) ^ (Y > -1) --> (X ^ Y) > -1
4668 bool TrueIfSignedL, TrueIfSignedR;
4669 if ((LHS->hasOneUse() || RHS->hasOneUse()) &&
4670 isSignBitCheck(PredL, *LC, TrueIfSignedL) &&
4671 isSignBitCheck(PredR, *RC, TrueIfSignedR)) {
4672 Value *XorLR = Builder.CreateXor(LHS0, RHS0);
4673 return TrueIfSignedL == TrueIfSignedR ? Builder.CreateIsNeg(XorLR) :
4674 Builder.CreateIsNotNeg(XorLR);
4675 }
4676
4677 // Fold (icmp pred1 X, C1) ^ (icmp pred2 X, C2)
4678 // into a single comparison using range-based reasoning.
4679 if (LHS0 == RHS0) {
4680 ConstantRange CR1 = ConstantRange::makeExactICmpRegion(PredL, *LC);
4681 ConstantRange CR2 = ConstantRange::makeExactICmpRegion(PredR, *RC);
4682 auto CRUnion = CR1.exactUnionWith(CR2);
4683 auto CRIntersect = CR1.exactIntersectWith(CR2);
4684 if (CRUnion && CRIntersect)
4685 if (auto CR = CRUnion->exactIntersectWith(CRIntersect->inverse())) {
4686 if (CR->isFullSet())
4687 return ConstantInt::getTrue(I.getType());
4688 if (CR->isEmptySet())
4689 return ConstantInt::getFalse(I.getType());
4690
4691 CmpInst::Predicate NewPred;
4692 APInt NewC, Offset;
4693 CR->getEquivalentICmp(NewPred, NewC, Offset);
4694
4695 if ((Offset.isZero() && (LHS->hasOneUse() || RHS->hasOneUse())) ||
4696 (LHS->hasOneUse() && RHS->hasOneUse())) {
4697 Value *NewV = LHS0;
4698 Type *Ty = LHS0->getType();
4699 if (!Offset.isZero())
4700 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
4701 return Builder.CreateICmp(NewPred, NewV,
4702 ConstantInt::get(Ty, NewC));
4703 }
4704 }
4705 }
4706
4707 // Fold (icmp eq/ne (X & Pow2), 0) ^ (icmp eq/ne (Y & Pow2), 0) into
4708 // (icmp eq/ne ((X ^ Y) & Pow2), 0)
4709 Value *X, *Y, *Pow2;
4710 if (ICmpInst::isEquality(PredL) && ICmpInst::isEquality(PredR) &&
4711 LC->isZero() && RC->isZero() && LHS->hasOneUse() && RHS->hasOneUse() &&
4712 match(LHS0, m_And(m_Value(X), m_Value(Pow2))) &&
4713 match(RHS0, m_And(m_Value(Y), m_Specific(Pow2))) &&
4714 isKnownToBeAPowerOfTwo(Pow2, /*OrZero=*/true, &I)) {
4715 Value *Xor = Builder.CreateXor(X, Y);
4716 Value *And = Builder.CreateAnd(Xor, Pow2);
4717 return Builder.CreateICmp(PredL == PredR ? ICmpInst::ICMP_NE
4719 And, ConstantInt::getNullValue(Xor->getType()));
4720 }
4721 }
4722
4723 // Instead of trying to imitate the folds for and/or, decompose this 'xor'
4724 // into those logic ops. That is, try to turn this into an and-of-icmps
4725 // because we have many folds for that pattern.
4726 //
4727 // This is based on a truth table definition of xor:
4728 // X ^ Y --> (X | Y) & !(X & Y)
4729 if (Value *OrICmp = simplifyBinOp(Instruction::Or, LHS, RHS, SQ)) {
4730 // TODO: If OrICmp is true, then the definition of xor simplifies to !(X&Y).
4731 // TODO: If OrICmp is false, the whole thing is false (InstSimplify?).
4732 if (Value *AndICmp = simplifyBinOp(Instruction::And, LHS, RHS, SQ)) {
4733 // TODO: Independently handle cases where the 'and' side is a constant.
4734 ICmpInst *X = nullptr, *Y = nullptr;
4735 if (OrICmp == LHS && AndICmp == RHS) {
4736 // (LHS | RHS) & !(LHS & RHS) --> LHS & !RHS --> X & !Y
4737 X = LHS;
4738 Y = RHS;
4739 }
4740 if (OrICmp == RHS && AndICmp == LHS) {
4741 // !(LHS & RHS) & (LHS | RHS) --> !LHS & RHS --> !Y & X
4742 X = RHS;
4743 Y = LHS;
4744 }
4745 if (X && Y && (Y->hasOneUse() || canFreelyInvertAllUsersOf(Y, &I))) {
4746 // Invert the predicate of 'Y', thus inverting its output.
4747 Y->setPredicate(Y->getInversePredicate());
4748 // So, are there other uses of Y?
4749 if (!Y->hasOneUse()) {
4750 // We need to adapt other uses of Y though. Get a value that matches
4751 // the original value of Y before inversion. While this increases
4752 // immediate instruction count, we have just ensured that all the
4753 // users are freely-invertible, so that 'not' *will* get folded away.
4755 // Set insertion point to right after the Y.
4756 Builder.SetInsertPoint(Y->getParent(), ++(Y->getIterator()));
4757 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4758 // Replace all uses of Y (excluding the one in NotY!) with NotY.
4759 Worklist.pushUsersToWorkList(*Y);
4760 Y->replaceUsesWithIf(NotY,
4761 [NotY](Use &U) { return U.getUser() != NotY; });
4762 }
4763 // All done.
4764 return Builder.CreateAnd(LHS, RHS);
4765 }
4766 }
4767 }
4768
4769 return nullptr;
4770}
4771
4772/// If we have a masked merge, in the canonical form of:
4773/// (assuming that A only has one use.)
4774/// | A | |B|
4775/// ((x ^ y) & M) ^ y
4776/// | D |
4777/// * If M is inverted:
4778/// | D |
4779/// ((x ^ y) & ~M) ^ y
4780/// We can canonicalize by swapping the final xor operand
4781/// to eliminate the 'not' of the mask.
4782/// ((x ^ y) & M) ^ x
4783/// * If M is a constant, and D has one use, we transform to 'and' / 'or' ops
4784/// because that shortens the dependency chain and improves analysis:
4785/// (x & M) | (y & ~M)
4787 InstCombiner::BuilderTy &Builder) {
4788 Value *B, *X, *D;
4789 Value *M;
4790 if (!match(&I, m_c_Xor(m_Value(B),
4793 m_Value(M))))))
4794 return nullptr;
4795
4796 Value *NotM;
4797 if (match(M, m_Not(m_Value(NotM)))) {
4798 // De-invert the mask and swap the value in B part.
4799 Value *NewA = Builder.CreateAnd(D, NotM);
4800 return BinaryOperator::CreateXor(NewA, X);
4801 }
4802
4803 Constant *C;
4804 if (D->hasOneUse() && match(M, m_Constant(C))) {
4805 // Propagating undef is unsafe. Clamp undef elements to -1.
4806 Type *EltTy = C->getType()->getScalarType();
4808 // Unfold.
4809 Value *LHS = Builder.CreateAnd(X, C);
4810 Value *NotC = Builder.CreateNot(C);
4811 Value *RHS = Builder.CreateAnd(B, NotC);
4812 return BinaryOperator::CreateOr(LHS, RHS);
4813 }
4814
4815 return nullptr;
4816}
4817
4819 InstCombiner::BuilderTy &Builder) {
4820 Value *X, *Y;
4821 // FIXME: one-use check is not needed in general, but currently we are unable
4822 // to fold 'not' into 'icmp', if that 'icmp' has multiple uses. (D35182)
4823 if (!match(&I, m_Not(m_OneUse(m_Xor(m_Value(X), m_Value(Y))))))
4824 return nullptr;
4825
4826 auto hasCommonOperand = [](Value *A, Value *B, Value *C, Value *D) {
4827 return A == C || A == D || B == C || B == D;
4828 };
4829
4830 Value *A, *B, *C, *D;
4831 // Canonicalize ~((A & B) ^ (A | ?)) -> (A & B) | ~(A | ?)
4832 // 4 commuted variants
4833 if (match(X, m_And(m_Value(A), m_Value(B))) &&
4834 match(Y, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4835 Value *NotY = Builder.CreateNot(Y);
4836 return BinaryOperator::CreateOr(X, NotY);
4837 };
4838
4839 // Canonicalize ~((A | ?) ^ (A & B)) -> (A & B) | ~(A | ?)
4840 // 4 commuted variants
4841 if (match(Y, m_And(m_Value(A), m_Value(B))) &&
4842 match(X, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4843 Value *NotX = Builder.CreateNot(X);
4844 return BinaryOperator::CreateOr(Y, NotX);
4845 };
4846
4847 return nullptr;
4848}
4849
4850/// Canonicalize a shifty way to code absolute value to the more common pattern
4851/// that uses negation and select.
4853 InstCombiner::BuilderTy &Builder) {
4854 assert(Xor.getOpcode() == Instruction::Xor && "Expected an xor instruction.");
4855
4856 // There are 4 potential commuted variants. Move the 'ashr' candidate to Op1.
4857 // We're relying on the fact that we only do this transform when the shift has
4858 // exactly 2 uses and the add has exactly 1 use (otherwise, we might increase
4859 // instructions).
4860 Value *Op0 = Xor.getOperand(0), *Op1 = Xor.getOperand(1);
4861 if (Op0->hasNUses(2))
4862 std::swap(Op0, Op1);
4863
4864 Type *Ty = Xor.getType();
4865 Value *A;
4866 const APInt *ShAmt;
4867 if (match(Op1, m_AShr(m_Value(A), m_APInt(ShAmt))) &&
4868 Op1->hasNUses(2) && *ShAmt == Ty->getScalarSizeInBits() - 1 &&
4869 match(Op0, m_OneUse(m_c_Add(m_Specific(A), m_Specific(Op1))))) {
4870 // Op1 = ashr i32 A, 31 ; smear the sign bit
4871 // xor (add A, Op1), Op1 ; add -1 and flip bits if negative
4872 // --> (A < 0) ? -A : A
4873 Value *IsNeg = Builder.CreateIsNeg(A);
4874 // Copy the nsw flags from the add to the negate.
4875 auto *Add = cast<BinaryOperator>(Op0);
4876 Value *NegA = Add->hasNoUnsignedWrap()
4877 ? Constant::getNullValue(A->getType())
4878 : Builder.CreateNeg(A, "", Add->hasNoSignedWrap());
4879 return SelectInst::Create(IsNeg, NegA, A);
4880 }
4881 return nullptr;
4882}
4883
4885 Instruction *IgnoredUser) {
4886 auto *I = dyn_cast<Instruction>(Op);
4887 return I && IC.isFreeToInvert(I, /*WillInvertAllUses=*/true) &&
4888 IC.canFreelyInvertAllUsersOf(I, IgnoredUser);
4889}
4890
4892 Instruction *IgnoredUser) {
4893 auto *I = cast<Instruction>(Op);
4894 IC.Builder.SetInsertPoint(*I->getInsertionPointAfterDef());
4895 Value *NotOp = IC.Builder.CreateNot(Op, Op->getName() + ".not");
4896 Op->replaceUsesWithIf(NotOp,
4897 [NotOp](Use &U) { return U.getUser() != NotOp; });
4898 IC.freelyInvertAllUsersOf(NotOp, IgnoredUser);
4899 return NotOp;
4900}
4901
4902// Transform
4903// z = ~(x &/| y)
4904// into:
4905// z = ((~x) |/& (~y))
4906// iff both x and y are free to invert and all uses of z can be freely updated.
4908 Value *Op0, *Op1;
4909 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
4910 return false;
4911
4912 // If this logic op has not been simplified yet, just bail out and let that
4913 // happen first. Otherwise, the code below may wrongly invert.
4914 if (Op0 == Op1)
4915 return false;
4916
4917 // If one of the operands is a user of the other,
4918 // freelyInvert->freelyInvertAllUsersOf will change the operands of I, which
4919 // may cause miscompilation.
4920 if (match(Op0, m_Not(m_Specific(Op1))) || match(Op1, m_Not(m_Specific(Op0))))
4921 return false;
4922
4923 Instruction::BinaryOps NewOpc =
4924 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
4925 bool IsBinaryOp = isa<BinaryOperator>(I);
4926
4927 // Can our users be adapted?
4928 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
4929 return false;
4930
4931 // And can the operands be adapted?
4932 if (!canFreelyInvert(*this, Op0, &I) || !canFreelyInvert(*this, Op1, &I))
4933 return false;
4934
4935 Op0 = freelyInvert(*this, Op0, &I);
4936 Op1 = freelyInvert(*this, Op1, &I);
4937
4938 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
4939 Value *NewLogicOp;
4940 if (IsBinaryOp)
4941 NewLogicOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
4942 else
4943 NewLogicOp =
4944 Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
4945
4946 replaceInstUsesWith(I, NewLogicOp);
4947 // We can not just create an outer `not`, it will most likely be immediately
4948 // folded back, reconstructing our initial pattern, and causing an
4949 // infinite combine loop, so immediately manually fold it away.
4950 freelyInvertAllUsersOf(NewLogicOp);
4951 return true;
4952}
4953
4954// Transform
4955// z = (~x) &/| y
4956// into:
4957// z = ~(x |/& (~y))
4958// iff y is free to invert and all uses of z can be freely updated.
4960 Value *Op0, *Op1;
4961 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
4962 return false;
4963 Instruction::BinaryOps NewOpc =
4964 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
4965 bool IsBinaryOp = isa<BinaryOperator>(I);
4966
4967 Value *NotOp0 = nullptr;
4968 Value *NotOp1 = nullptr;
4969 Value **OpToInvert = nullptr;
4970 if (match(Op0, m_Not(m_Value(NotOp0))) && canFreelyInvert(*this, Op1, &I)) {
4971 Op0 = NotOp0;
4972 OpToInvert = &Op1;
4973 } else if (match(Op1, m_Not(m_Value(NotOp1))) &&
4974 canFreelyInvert(*this, Op0, &I)) {
4975 Op1 = NotOp1;
4976 OpToInvert = &Op0;
4977 } else
4978 return false;
4979
4980 // And can our users be adapted?
4981 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
4982 return false;
4983
4984 *OpToInvert = freelyInvert(*this, *OpToInvert, &I);
4985
4986 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
4987 Value *NewBinOp;
4988 if (IsBinaryOp)
4989 NewBinOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
4990 else
4991 NewBinOp = Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
4992 replaceInstUsesWith(I, NewBinOp);
4993 // We can not just create an outer `not`, it will most likely be immediately
4994 // folded back, reconstructing our initial pattern, and causing an
4995 // infinite combine loop, so immediately manually fold it away.
4996 freelyInvertAllUsersOf(NewBinOp);
4997 return true;
4998}
4999
5000Instruction *InstCombinerImpl::foldNot(BinaryOperator &I) {
5001 Value *NotOp;
5002 if (!match(&I, m_Not(m_Value(NotOp))))
5003 return nullptr;
5004
5005 // Apply DeMorgan's Law for 'nand' / 'nor' logic with an inverted operand.
5006 // We must eliminate the and/or (one-use) for these transforms to not increase
5007 // the instruction count.
5008 //
5009 // ~(~X & Y) --> (X | ~Y)
5010 // ~(Y & ~X) --> (X | ~Y)
5011 //
5012 // Note: The logical matches do not check for the commuted patterns because
5013 // those are handled via SimplifySelectsFeedingBinaryOp().
5014 Type *Ty = I.getType();
5015 Value *X, *Y;
5016 if (match(NotOp, m_OneUse(m_c_And(m_Not(m_Value(X)), m_Value(Y))))) {
5017 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5018 return BinaryOperator::CreateOr(X, NotY);
5019 }
5020 if (match(NotOp, m_OneUse(m_LogicalAnd(m_Not(m_Value(X)), m_Value(Y))))) {
5021 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5022 return SelectInst::Create(X, ConstantInt::getTrue(Ty), NotY);
5023 }
5024
5025 // ~(~X | Y) --> (X & ~Y)
5026 // ~(Y | ~X) --> (X & ~Y)
5027 if (match(NotOp, m_OneUse(m_c_Or(m_Not(m_Value(X)), m_Value(Y))))) {
5028 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5029 return BinaryOperator::CreateAnd(X, NotY);
5030 }
5031 if (match(NotOp, m_OneUse(m_LogicalOr(m_Not(m_Value(X)), m_Value(Y))))) {
5032 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5033 return SelectInst::Create(X, NotY, ConstantInt::getFalse(Ty));
5034 }
5035
5036 // Is this a 'not' (~) fed by a binary operator?
5037 BinaryOperator *NotVal;
5038 if (match(NotOp, m_BinOp(NotVal))) {
5039 // ~((-X) | Y) --> (X - 1) & (~Y)
5040 if (match(NotVal,
5042 Value *DecX = Builder.CreateAdd(X, ConstantInt::getAllOnesValue(Ty));
5043 Value *NotY = Builder.CreateNot(Y);
5044 return BinaryOperator::CreateAnd(DecX, NotY);
5045 }
5046
5047 // ~(~X >>s Y) --> (X >>s Y)
5048 if (match(NotVal, m_AShr(m_Not(m_Value(X)), m_Value(Y))))
5049 return BinaryOperator::CreateAShr(X, Y);
5050
5051 // Treat lshr with non-negative operand as ashr.
5052 // ~(~X >>u Y) --> (X >>s Y) iff X is known negative
5053 if (match(NotVal, m_LShr(m_Not(m_Value(X)), m_Value(Y))) &&
5054 isKnownNegative(X, SQ.getWithInstruction(NotVal)))
5055 return BinaryOperator::CreateAShr(X, Y);
5056
5057 // Bit-hack form of a signbit test for iN type:
5058 // ~(X >>s (N - 1)) --> sext i1 (X > -1) to iN
5059 unsigned FullShift = Ty->getScalarSizeInBits() - 1;
5060 if (match(NotVal, m_OneUse(m_AShr(m_Value(X), m_SpecificInt(FullShift))))) {
5061 Value *IsNotNeg = Builder.CreateIsNotNeg(X, "isnotneg");
5062 return new SExtInst(IsNotNeg, Ty);
5063 }
5064
5065 // If we are inverting a right-shifted constant, we may be able to eliminate
5066 // the 'not' by inverting the constant and using the opposite shift type.
5067 // Canonicalization rules ensure that only a negative constant uses 'ashr',
5068 // but we must check that in case that transform has not fired yet.
5069
5070 // ~(C >>s Y) --> ~C >>u Y (when inverting the replicated sign bits)
5071 Constant *C;
5072 if (match(NotVal, m_AShr(m_Constant(C), m_Value(Y))) &&
5073 match(C, m_Negative()))
5074 return BinaryOperator::CreateLShr(ConstantExpr::getNot(C), Y);
5075
5076 // ~(C >>u Y) --> ~C >>s Y (when inverting the replicated sign bits)
5077 if (match(NotVal, m_LShr(m_Constant(C), m_Value(Y))) &&
5078 match(C, m_NonNegative()))
5079 return BinaryOperator::CreateAShr(ConstantExpr::getNot(C), Y);
5080
5081 // ~(X + C) --> ~C - X
5082 if (match(NotVal, m_Add(m_Value(X), m_ImmConstant(C))))
5083 return BinaryOperator::CreateSub(ConstantExpr::getNot(C), X);
5084
5085 // ~(X - Y) --> ~X + Y
5086 // FIXME: is it really beneficial to sink the `not` here?
5087 if (match(NotVal, m_Sub(m_Value(X), m_Value(Y))))
5088 if (isa<Constant>(X) || NotVal->hasOneUse())
5089 return BinaryOperator::CreateAdd(Builder.CreateNot(X), Y);
5090
5091 // ~(~X + Y) --> X - Y
5092 if (match(NotVal, m_c_Add(m_Not(m_Value(X)), m_Value(Y))))
5093 return BinaryOperator::CreateWithCopiedFlags(Instruction::Sub, X, Y,
5094 NotVal);
5095 }
5096
5097 // not (cmp A, B) = !cmp A, B
5098 CmpPredicate Pred;
5099 if (match(NotOp, m_Cmp(Pred, m_Value(), m_Value())) &&
5100 (NotOp->hasOneUse() ||
5102 /*IgnoredUser=*/nullptr))) {
5103 cast<CmpInst>(NotOp)->setPredicate(CmpInst::getInversePredicate(Pred));
5105 return &I;
5106 }
5107
5108 // not (bitcast (cmp A, B) --> bitcast (!cmp A, B)
5109 if (match(NotOp, m_OneUse(m_BitCast(m_Value(X)))) &&
5110 match(X, m_OneUse(m_Cmp(Pred, m_Value(), m_Value())))) {
5111 cast<CmpInst>(X)->setPredicate(CmpInst::getInversePredicate(Pred));
5112 return new BitCastInst(X, Ty);
5113 }
5114
5115 // Move a 'not' ahead of casts of a bool to enable logic reduction:
5116 // not (bitcast (sext i1 X)) --> bitcast (sext (not i1 X))
5117 if (match(NotOp, m_OneUse(m_BitCast(m_OneUse(m_SExt(m_Value(X)))))) &&
5118 X->getType()->isIntOrIntVectorTy(1)) {
5119 Type *SextTy = cast<BitCastOperator>(NotOp)->getSrcTy();
5120 Value *NotX = Builder.CreateNot(X);
5121 Value *Sext = Builder.CreateSExt(NotX, SextTy);
5122 return new BitCastInst(Sext, Ty);
5123 }
5124
5125 if (auto *NotOpI = dyn_cast<Instruction>(NotOp))
5126 if (sinkNotIntoLogicalOp(*NotOpI))
5127 return &I;
5128
5129 // Eliminate a bitwise 'not' op of 'not' min/max by inverting the min/max:
5130 // ~min(~X, ~Y) --> max(X, Y)
5131 // ~max(~X, Y) --> min(X, ~Y)
5132 auto *II = dyn_cast<IntrinsicInst>(NotOp);
5133 if (II && II->hasOneUse()) {
5134 if (match(NotOp, m_c_MaxOrMin(m_Not(m_Value(X)), m_Value(Y)))) {
5135 Intrinsic::ID InvID = getInverseMinMaxIntrinsic(II->getIntrinsicID());
5136 Value *NotY = Builder.CreateNot(Y);
5137 Value *InvMaxMin = Builder.CreateBinaryIntrinsic(InvID, X, NotY);
5138 return replaceInstUsesWith(I, InvMaxMin);
5139 }
5140
5141 if (II->getIntrinsicID() == Intrinsic::is_fpclass) {
5142 ConstantInt *ClassMask = cast<ConstantInt>(II->getArgOperand(1));
5143 II->setArgOperand(
5144 1, ConstantInt::get(ClassMask->getType(),
5145 ~ClassMask->getZExtValue() & fcAllFlags));
5146 return replaceInstUsesWith(I, II);
5147 }
5148 }
5149
5150 if (NotOp->hasOneUse()) {
5151 // Pull 'not' into operands of select if both operands are one-use compares
5152 // or one is one-use compare and the other one is a constant.
5153 // Inverting the predicates eliminates the 'not' operation.
5154 // Example:
5155 // not (select ?, (cmp TPred, ?, ?), (cmp FPred, ?, ?) -->
5156 // select ?, (cmp InvTPred, ?, ?), (cmp InvFPred, ?, ?)
5157 // not (select ?, (cmp TPred, ?, ?), true -->
5158 // select ?, (cmp InvTPred, ?, ?), false
5159 if (auto *Sel = dyn_cast<SelectInst>(NotOp)) {
5160 Value *TV = Sel->getTrueValue();
5161 Value *FV = Sel->getFalseValue();
5162 auto *CmpT = dyn_cast<CmpInst>(TV);
5163 auto *CmpF = dyn_cast<CmpInst>(FV);
5164 bool InvertibleT = (CmpT && CmpT->hasOneUse()) || isa<Constant>(TV);
5165 bool InvertibleF = (CmpF && CmpF->hasOneUse()) || isa<Constant>(FV);
5166 if (InvertibleT && InvertibleF) {
5167 if (CmpT)
5168 CmpT->setPredicate(CmpT->getInversePredicate());
5169 else
5170 Sel->setTrueValue(ConstantExpr::getNot(cast<Constant>(TV)));
5171 if (CmpF)
5172 CmpF->setPredicate(CmpF->getInversePredicate());
5173 else
5174 Sel->setFalseValue(ConstantExpr::getNot(cast<Constant>(FV)));
5175 return replaceInstUsesWith(I, Sel);
5176 }
5177 }
5178 }
5179
5180 if (Instruction *NewXor = foldNotXor(I, Builder))
5181 return NewXor;
5182
5183 // TODO: Could handle multi-use better by checking if all uses of NotOp (other
5184 // than I) can be inverted.
5185 if (Value *R = getFreelyInverted(NotOp, NotOp->hasOneUse(), &Builder))
5186 return replaceInstUsesWith(I, R);
5187
5188 return nullptr;
5189}
5190
5191// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
5192// here. We should standardize that construct where it is needed or choose some
5193// other way to ensure that commutated variants of patterns are not missed.
5195 if (Value *V = simplifyXorInst(I.getOperand(0), I.getOperand(1),
5196 SQ.getWithInstruction(&I)))
5197 return replaceInstUsesWith(I, V);
5198
5200 return &I;
5201
5203 return X;
5204
5206 return Phi;
5207
5208 if (Instruction *NewXor = foldXorToXor(I, Builder))
5209 return NewXor;
5210
5211 // (A&B)^(A&C) -> A&(B^C) etc
5213 return replaceInstUsesWith(I, V);
5214
5215 // See if we can simplify any instructions used by the instruction whose sole
5216 // purpose is to compute bits we don't care about.
5218 return &I;
5219
5220 if (Instruction *R = foldNot(I))
5221 return R;
5222
5224 return R;
5225
5226 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5227 Value *X, *Y, *M;
5228
5229 // (X | Y) ^ M -> (X ^ M) ^ Y
5230 // (X | Y) ^ M -> (Y ^ M) ^ X
5232 m_Value(M)))) {
5233 if (Value *XorAC = simplifyXorInst(X, M, SQ.getWithInstruction(&I)))
5234 return BinaryOperator::CreateXor(XorAC, Y);
5235
5236 if (Value *XorBC = simplifyXorInst(Y, M, SQ.getWithInstruction(&I)))
5237 return BinaryOperator::CreateXor(XorBC, X);
5238 }
5239
5240 // Fold (X & M) ^ (Y & ~M) -> (X & M) | (Y & ~M)
5241 // This it a special case in haveNoCommonBitsSet, but the computeKnownBits
5242 // calls in there are unnecessary as SimplifyDemandedInstructionBits should
5243 // have already taken care of those cases.
5244 if (match(&I, m_c_Xor(m_c_And(m_Not(m_Value(M)), m_Value()),
5245 m_c_And(m_Deferred(M), m_Value())))) {
5247 return BinaryOperator::CreateDisjointOr(Op0, Op1);
5248 else
5249 return BinaryOperator::CreateOr(Op0, Op1);
5250 }
5251
5253 return Xor;
5254
5255 Constant *C1;
5256 if (match(Op1, m_Constant(C1))) {
5257 Constant *C2;
5258
5259 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_ImmConstant(C2)))) &&
5260 match(C1, m_ImmConstant())) {
5261 // (X | C2) ^ C1 --> (X & ~C2) ^ (C1^C2)
5264 Value *And = Builder.CreateAnd(
5266 return BinaryOperator::CreateXor(
5268 }
5269
5270 // Use DeMorgan and reassociation to eliminate a 'not' op.
5271 if (match(Op0, m_OneUse(m_Or(m_Not(m_Value(X)), m_Constant(C2))))) {
5272 // (~X | C2) ^ C1 --> ((X & ~C2) ^ -1) ^ C1 --> (X & ~C2) ^ ~C1
5273 Value *And = Builder.CreateAnd(X, ConstantExpr::getNot(C2));
5274 return BinaryOperator::CreateXor(And, ConstantExpr::getNot(C1));
5275 }
5276 if (match(Op0, m_OneUse(m_And(m_Not(m_Value(X)), m_Constant(C2))))) {
5277 // (~X & C2) ^ C1 --> ((X | ~C2) ^ -1) ^ C1 --> (X | ~C2) ^ ~C1
5278 Value *Or = Builder.CreateOr(X, ConstantExpr::getNot(C2));
5279 return BinaryOperator::CreateXor(Or, ConstantExpr::getNot(C1));
5280 }
5281
5282 // Convert xor ([trunc] (ashr X, BW-1)), C =>
5283 // select(X >s -1, C, ~C)
5284 // The ashr creates "AllZeroOrAllOne's", which then optionally inverses the
5285 // constant depending on whether this input is less than 0.
5286 const APInt *CA;
5287 if (match(Op0, m_OneUse(m_TruncOrSelf(
5288 m_AShr(m_Value(X), m_APIntAllowPoison(CA))))) &&
5289 *CA == X->getType()->getScalarSizeInBits() - 1 &&
5290 !match(C1, m_AllOnes())) {
5291 assert(!C1->isZeroValue() && "Unexpected xor with 0");
5292 Value *IsNotNeg = Builder.CreateIsNotNeg(X);
5293 return createSelectInstWithUnknownProfile(IsNotNeg, Op1,
5294 Builder.CreateNot(Op1));
5295 }
5296 }
5297
5298 Type *Ty = I.getType();
5299 {
5300 const APInt *RHSC;
5301 if (match(Op1, m_APInt(RHSC))) {
5302 Value *X;
5303 const APInt *C;
5304 // (C - X) ^ signmaskC --> (C + signmaskC) - X
5305 if (RHSC->isSignMask() && match(Op0, m_Sub(m_APInt(C), m_Value(X))))
5306 return BinaryOperator::CreateSub(ConstantInt::get(Ty, *C + *RHSC), X);
5307
5308 // (X + C) ^ signmaskC --> X + (C + signmaskC)
5309 if (RHSC->isSignMask() && match(Op0, m_Add(m_Value(X), m_APInt(C))))
5310 return BinaryOperator::CreateAdd(X, ConstantInt::get(Ty, *C + *RHSC));
5311
5312 // (X | C) ^ RHSC --> X ^ (C ^ RHSC) iff X & C == 0
5313 if (match(Op0, m_Or(m_Value(X), m_APInt(C))) &&
5314 MaskedValueIsZero(X, *C, &I))
5315 return BinaryOperator::CreateXor(X, ConstantInt::get(Ty, *C ^ *RHSC));
5316
5317 // When X is a power-of-two or zero and zero input is poison:
5318 // ctlz(i32 X) ^ 31 --> cttz(X)
5319 // cttz(i32 X) ^ 31 --> ctlz(X)
5320 auto *II = dyn_cast<IntrinsicInst>(Op0);
5321 if (II && II->hasOneUse() && *RHSC == Ty->getScalarSizeInBits() - 1) {
5322 Intrinsic::ID IID = II->getIntrinsicID();
5323 if ((IID == Intrinsic::ctlz || IID == Intrinsic::cttz) &&
5324 match(II->getArgOperand(1), m_One()) &&
5325 isKnownToBeAPowerOfTwo(II->getArgOperand(0), /*OrZero */ true)) {
5326 IID = (IID == Intrinsic::ctlz) ? Intrinsic::cttz : Intrinsic::ctlz;
5327 Function *F =
5328 Intrinsic::getOrInsertDeclaration(II->getModule(), IID, Ty);
5329 return CallInst::Create(F, {II->getArgOperand(0), Builder.getTrue()});
5330 }
5331 }
5332
5333 // If RHSC is inverting the remaining bits of shifted X,
5334 // canonicalize to a 'not' before the shift to help SCEV and codegen:
5335 // (X << C) ^ RHSC --> ~X << C
5336 if (match(Op0, m_OneUse(m_Shl(m_Value(X), m_APInt(C)))) &&
5337 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).shl(*C)) {
5338 Value *NotX = Builder.CreateNot(X);
5339 return BinaryOperator::CreateShl(NotX, ConstantInt::get(Ty, *C));
5340 }
5341 // (X >>u C) ^ RHSC --> ~X >>u C
5342 if (match(Op0, m_OneUse(m_LShr(m_Value(X), m_APInt(C)))) &&
5343 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).lshr(*C)) {
5344 Value *NotX = Builder.CreateNot(X);
5345 return BinaryOperator::CreateLShr(NotX, ConstantInt::get(Ty, *C));
5346 }
5347 // TODO: We could handle 'ashr' here as well. That would be matching
5348 // a 'not' op and moving it before the shift. Doing that requires
5349 // preventing the inverse fold in canShiftBinOpWithConstantRHS().
5350 }
5351
5352 // If we are XORing the sign bit of a floating-point value, convert
5353 // this to fneg, then cast back to integer.
5354 //
5355 // This is generous interpretation of noimplicitfloat, this is not a true
5356 // floating-point operation.
5357 //
5358 // Assumes any IEEE-represented type has the sign bit in the high bit.
5359 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
5360 Value *CastOp;
5361 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
5362 match(Op1, m_SignMask()) &&
5363 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
5364 Attribute::NoImplicitFloat)) {
5365 Type *EltTy = CastOp->getType()->getScalarType();
5366 if (EltTy->isFloatingPointTy() &&
5368 Value *FNeg = Builder.CreateFNeg(CastOp);
5369 return new BitCastInst(FNeg, I.getType());
5370 }
5371 }
5372 }
5373
5374 // FIXME: This should not be limited to scalar (pull into APInt match above).
5375 {
5376 Value *X;
5377 ConstantInt *C1, *C2, *C3;
5378 // ((X^C1) >> C2) ^ C3 -> (X>>C2) ^ ((C1>>C2)^C3)
5379 if (match(Op1, m_ConstantInt(C3)) &&
5381 m_ConstantInt(C2))) &&
5382 Op0->hasOneUse()) {
5383 // fold (C1 >> C2) ^ C3
5384 APInt FoldConst = C1->getValue().lshr(C2->getValue());
5385 FoldConst ^= C3->getValue();
5386 // Prepare the two operands.
5387 auto *Opnd0 = Builder.CreateLShr(X, C2);
5388 Opnd0->takeName(Op0);
5389 return BinaryOperator::CreateXor(Opnd0, ConstantInt::get(Ty, FoldConst));
5390 }
5391 }
5392
5393 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
5394 return FoldedLogic;
5395
5396 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
5397 return FoldedLogic;
5398
5399 // Y ^ (X | Y) --> X & ~Y
5400 // Y ^ (Y | X) --> X & ~Y
5401 if (match(Op1, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op0)))))
5402 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op0));
5403 // (X | Y) ^ Y --> X & ~Y
5404 // (Y | X) ^ Y --> X & ~Y
5405 if (match(Op0, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op1)))))
5406 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op1));
5407
5408 // Y ^ (X & Y) --> ~X & Y
5409 // Y ^ (Y & X) --> ~X & Y
5410 if (match(Op1, m_OneUse(m_c_And(m_Value(X), m_Specific(Op0)))))
5411 return BinaryOperator::CreateAnd(Op0, Builder.CreateNot(X));
5412 // (X & Y) ^ Y --> ~X & Y
5413 // (Y & X) ^ Y --> ~X & Y
5414 // Canonical form is (X & C) ^ C; don't touch that.
5415 // TODO: A 'not' op is better for analysis and codegen, but demanded bits must
5416 // be fixed to prefer that (otherwise we get infinite looping).
5417 if (!match(Op1, m_Constant()) &&
5418 match(Op0, m_OneUse(m_c_And(m_Value(X), m_Specific(Op1)))))
5419 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(X));
5420
5421 Value *A, *B, *C;
5422 // (A ^ B) ^ (A | C) --> (~A & C) ^ B -- There are 4 commuted variants.
5425 return BinaryOperator::CreateXor(
5426 Builder.CreateAnd(Builder.CreateNot(A), C), B);
5427
5428 // (A ^ B) ^ (B | C) --> (~B & C) ^ A -- There are 4 commuted variants.
5431 return BinaryOperator::CreateXor(
5432 Builder.CreateAnd(Builder.CreateNot(B), C), A);
5433
5434 // (A & B) ^ (A ^ B) -> (A | B)
5435 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
5437 return BinaryOperator::CreateOr(A, B);
5438 // (A ^ B) ^ (A & B) -> (A | B)
5439 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
5441 return BinaryOperator::CreateOr(A, B);
5442
5443 // (A & ~B) ^ ~A -> ~(A & B)
5444 // (~B & A) ^ ~A -> ~(A & B)
5445 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
5446 match(Op1, m_Not(m_Specific(A))))
5447 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
5448
5449 // (~A & B) ^ A --> A | B -- There are 4 commuted variants.
5451 return BinaryOperator::CreateOr(A, B);
5452
5453 // (~A | B) ^ A --> ~(A & B)
5454 if (match(Op0, m_OneUse(m_c_Or(m_Not(m_Specific(Op1)), m_Value(B)))))
5455 return BinaryOperator::CreateNot(Builder.CreateAnd(Op1, B));
5456
5457 // A ^ (~A | B) --> ~(A & B)
5458 if (match(Op1, m_OneUse(m_c_Or(m_Not(m_Specific(Op0)), m_Value(B)))))
5459 return BinaryOperator::CreateNot(Builder.CreateAnd(Op0, B));
5460
5461 // (A | B) ^ (A | C) --> (B ^ C) & ~A -- There are 4 commuted variants.
5462 // TODO: Loosen one-use restriction if common operand is a constant.
5463 Value *D;
5464 if (match(Op0, m_OneUse(m_Or(m_Value(A), m_Value(B)))) &&
5465 match(Op1, m_OneUse(m_Or(m_Value(C), m_Value(D))))) {
5466 if (B == C || B == D)
5467 std::swap(A, B);
5468 if (A == C)
5469 std::swap(C, D);
5470 if (A == D) {
5471 Value *NotA = Builder.CreateNot(A);
5472 return BinaryOperator::CreateAnd(Builder.CreateXor(B, C), NotA);
5473 }
5474 }
5475
5476 // (A & B) ^ (A | C) --> A ? ~B : C -- There are 4 commuted variants.
5477 if (I.getType()->isIntOrIntVectorTy(1) &&
5480 bool NeedFreeze = isa<SelectInst>(Op0) && isa<SelectInst>(Op1) && B == D;
5481 if (B == C || B == D)
5482 std::swap(A, B);
5483 if (A == C)
5484 std::swap(C, D);
5485 if (A == D) {
5486 if (NeedFreeze)
5487 A = Builder.CreateFreeze(A);
5488 Value *NotB = Builder.CreateNot(B);
5489 return SelectInst::Create(A, NotB, C);
5490 }
5491 }
5492
5493 if (auto *LHS = dyn_cast<ICmpInst>(I.getOperand(0)))
5494 if (auto *RHS = dyn_cast<ICmpInst>(I.getOperand(1)))
5495 if (Value *V = foldXorOfICmps(LHS, RHS, I))
5496 return replaceInstUsesWith(I, V);
5497
5498 if (Instruction *CastedXor = foldCastedBitwiseLogic(I))
5499 return CastedXor;
5500
5501 if (Instruction *Abs = canonicalizeAbs(I, Builder))
5502 return Abs;
5503
5504 // Otherwise, if all else failed, try to hoist the xor-by-constant:
5505 // (X ^ C) ^ Y --> (X ^ Y) ^ C
5506 // Just like we do in other places, we completely avoid the fold
5507 // for constantexprs, at least to avoid endless combine loop.
5509 m_ImmConstant(C1))),
5510 m_Value(Y))))
5511 return BinaryOperator::CreateXor(Builder.CreateXor(X, Y), C1);
5512
5514 return R;
5515
5516 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
5517 return Canonicalized;
5518
5519 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
5520 return Folded;
5521
5522 if (Instruction *Folded = canonicalizeConditionalNegationViaMathToSelect(I))
5523 return Folded;
5524
5525 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
5526 return Res;
5527
5529 return Res;
5530
5531 return nullptr;
5532}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static Value * foldAndOrOfICmpsWithConstEq(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q, Instruction &I)
Reduce logic-of-compares with equality to a constant by substituting a common operand with the consta...
static Value * foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and fold (icmp ne ctpop(X) 1) & ...
static Value * foldBitmaskMul(Value *Op0, Value *Op1, InstCombiner::BuilderTy &Builder)
(A & N) * C + (A & M) * C -> (A & (N + M)) & C This also accepts the equivalent select form of (A & N...
static unsigned conjugateICmpMask(unsigned Mask)
Convert an analysis of a masked ICmp into its equivalent if all boolean operations had the opposite s...
static Instruction * foldNotXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * foldLogOpOfMaskedICmps(Value *LHS, Value *RHS, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * getFCmpValue(unsigned Code, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder, FMFSource FMF)
This is the complement of getFCmpCode, which turns an opcode and two operands into either a FCmp inst...
static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal, uint64_t &ClassMask)
Match an fcmp against a special value that performs a test possible by llvm.is.fpclass.
static Value * foldSignedTruncationCheck(ICmpInst *ICmp0, ICmpInst *ICmp1, Instruction &CxtI, InstCombiner::BuilderTy &Builder)
General pattern: X & Y.
static Instruction * visitMaskedMerge(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
If we have a masked merge, in the canonical form of: (assuming that A only has one use....
static Instruction * canonicalizeAbs(BinaryOperator &Xor, InstCombiner::BuilderTy &Builder)
Canonicalize a shifty way to code absolute value to the more common pattern that uses negation and se...
static Value * foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Reduce a pair of compares that check if a value has exactly 1 bit set.
static Value * foldUnsignedUnderflowCheck(ICmpInst *ZeroICmp, ICmpInst *UnsignedICmp, bool IsAnd, const SimplifyQuery &Q, InstCombiner::BuilderTy &Builder)
Commuted variants are assumed to be handled by calling this function again with the parameters swappe...
static Instruction * foldOrToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * simplifyAndOrWithOpReplaced(Value *V, Value *Op, Value *RepOp, bool SimplifyOnly, InstCombinerImpl &IC, unsigned Depth=0)
static Instruction * matchDeMorgansLaws(BinaryOperator &I, InstCombiner &IC)
Match variations of De Morgan's Laws: (~A & ~B) == (~(A | B)) (~A | ~B) == (~(A & B))
static Value * foldLogOpOfMaskedICmpsAsymmetric(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * FoldOrOfSelectSmaxToAbs(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X) select(X <s 0, -X, 0) | smax(X,...
static Instruction * foldAndToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static unsigned getMaskedICmpType(Value *A, Value *B, Value *C, ICmpInst::Predicate Pred)
Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C) satisfies.
static Instruction * foldXorToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
A ^ B can be specified using other logic ops in a variety of patterns.
static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth)
Return true if a constant shift amount is always less than the specified bit-width.
static Instruction * foldLogicCastConstant(BinaryOperator &Logic, CastInst *Cast, InstCombinerImpl &IC)
Fold {and,or,xor} (cast X), C.
static Value * foldAndOrOfICmpEqConstantAndICmp(ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, bool IsLogical, IRBuilderBase &Builder)
static bool canFreelyInvert(InstCombiner &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldNegativePower2AndShiftedMask(Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff B is a contiguous set of o...
static Value * matchIsFiniteTest(InstCombiner::BuilderTy &Builder, FCmpInst *LHS, FCmpInst *RHS)
and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
static Value * foldPowerOf2AndShiftedMask(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder)
Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) & (icmp(X & M) !...
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
static Value * foldOrUnsignedUMulOverflowICmp(BinaryOperator &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2)) --> (ugt x (c2/c1)).
static Value * freelyInvert(InstCombinerImpl &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldLogOpOfMaskedICmps_NotAllZeros_BMask_Mixed(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static std::optional< IntPart > matchIntPart(Value *V)
Match an extraction of bits from an integer.
static Instruction * canonicalizeLogicFirst(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * reassociateFCmps(BinaryOperator &BO, InstCombiner::BuilderTy &Builder)
This a limited reassociation for a special case (see above) where we are checking if two values are e...
static Value * getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder)
This is the complement of getICmpCode, which turns an opcode and two operands into either a constant ...
static Value * extractIntPart(const IntPart &P, IRBuilderBase &Builder)
Materialize an extraction of bits from an integer in IR.
static bool matchUnorderedInfCompare(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches fcmp u__ x, +/-inf.
static bool matchIsNotNaN(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches canonical form of isnan, fcmp ord x, 0.
static bool areInverseVectorBitmasks(Constant *C1, Constant *C2)
If all elements of two constant vectors are 0/-1 and inverses, return true.
MaskedICmpType
Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns that can be simplified.
@ BMask_NotAllOnes
@ AMask_NotAllOnes
@ Mask_NotAllZeros
static Instruction * foldComplexAndOrPatterns(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Try folding relatively complex patterns for both And and Or operations with all And and Or swapped.
static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask, uint64_t &Offset, bool &IsShlNUW, bool &IsShlNSW)
Match V as "lshr -> mask -> zext -> shl".
static std::optional< DecomposedBitMaskMul > matchBitmaskMul(Value *V)
static Value * foldOrOfInversions(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static bool matchSubIntegerPackFromVector(Value *V, Value *&Vec, int64_t &VecOffset, SmallBitVector &Mask, const DataLayout &DL)
Match V as "shufflevector -> bitcast" or "extractelement -> zext -> shl" patterns,...
static Instruction * matchFunnelShift(Instruction &Or, InstCombinerImpl &IC)
Match UB-safe variants of the funnel shift intrinsic.
static Instruction * reassociateForUses(BinaryOperator &BO, InstCombinerImpl::BuilderTy &Builder)
Try to reassociate a pair of binops so that values with one use only are part of the same instruction...
static Value * matchOrConcat(Instruction &Or, InstCombiner::BuilderTy &Builder)
Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
static Value * foldAndOrOfICmpsWithPow2AndWithZero(InstCombiner::BuilderTy &Builder, ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, const SimplifyQuery &Q)
static Instruction * foldBitwiseLogicWithIntrinsics(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static std::optional< std::pair< unsigned, unsigned > > getMaskedTypeForICmpPair(Value *&A, Value *&B, Value *&C, Value *&D, Value *&E, Value *LHS, Value *RHS, ICmpInst::Predicate &PredL, ICmpInst::Predicate &PredR)
Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
static Instruction * foldIntegerPackFromVector(Instruction &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Try to fold the join of two scalar integers whose contents are packed elements of the same vector.
static Value * foldIntegerRepackThroughZExt(Value *Lhs, Value *Rhs, InstCombiner::BuilderTy &Builder)
Try to fold the join of two scalar integers whose bits are unpacked and zexted from the same source i...
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define R2(n)
uint64_t High
uint64_t IntrinsicInst * II
#define P(N)
const SmallVectorImpl< MachineOperand > & Cond
This file implements the SmallBitVector class.
static unsigned getScalarSizeInBits(Type *Ty)
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static constexpr int Concat[]
Value * RHS
Value * LHS
The Input class is used to parse a yaml document into in-memory structs and vectors.
static LLVM_ABI bool hasSignBitInMSB(const fltSemantics &)
Definition APFloat.cpp:336
bool bitwiseIsEqual(const APFloat &RHS) const
Definition APFloat.h:1396
bool isZero() const
Definition APFloat.h:1427
APInt bitcastToAPInt() const
Definition APFloat.h:1335
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1080
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1573
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1012
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1541
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:936
unsigned countLeadingOnes() const
Definition APInt.h:1625
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1948
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1183
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
Definition APInt.h:467
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1489
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1112
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1928
bool intersects(const APInt &RHS) const
This operation tests if there are any pairs of corresponding bits between this APInt and RHS that are...
Definition APInt.h:1250
int32_t exactLogBase2() const
Definition APInt.h:1784
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:768
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1935
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1640
unsigned countLeadingZeros() const
Definition APInt.h:1607
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1151
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:874
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:746
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1258
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1941
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:852
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1222
void clearSignBit()
Set the sign bit to 0.
Definition APInt.h:1450
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
LLVM_ABI bool isSigned() const
Whether the intrinsic is signed or unsigned.
LLVM_ABI Instruction::BinaryOps getBinaryOp() const
Returns the binary operation underlying the intrinsic.
BinaryOps getOpcode() const
Definition InstrTypes.h:374
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
static BinaryOperator * CreateWithCopiedFlags(BinaryOps Opc, Value *V1, Value *V2, Value *CopyO, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Definition InstrTypes.h:219
This class represents a no-op cast from one type to another.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Definition InstrTypes.h:448
Type * getSrcTy() const
Return the source type, as a convenience.
Definition InstrTypes.h:615
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition InstrTypes.h:610
static LLVM_ABI CastInst * Create(Instruction::CastOps, Value *S, Type *Ty, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Provides a way to construct any of the CastInst subclasses using an opcode instead of the subclass's ...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition InstrTypes.h:617
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Definition InstrTypes.h:982
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ ICMP_SLT
signed less than
Definition InstrTypes.h:705
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:706
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition InstrTypes.h:682
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition InstrTypes.h:691
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:700
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_SGT
signed greater than
Definition InstrTypes.h:703
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition InstrTypes.h:690
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:701
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition InstrTypes.h:683
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition InstrTypes.h:685
@ ICMP_NE
not equal
Definition InstrTypes.h:698
@ ICMP_SGE
signed greater or equal
Definition InstrTypes.h:704
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:702
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition InstrTypes.h:686
bool isSigned() const
Definition InstrTypes.h:930
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Definition InstrTypes.h:827
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Definition InstrTypes.h:789
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:765
static LLVM_ABI bool isUnordered(Predicate predicate)
Determine if the predicate is an unordered operation.
static Predicate getOrderedPredicate(Predicate Pred)
Returns the ordered variant of a floating point compare.
Definition InstrTypes.h:796
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getAdd(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getTrunc(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getExactLogBase2(Constant *C)
If C is a scalar/fixed width vector of known powers of 2, then this function returns a new scalar/fix...
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
Definition Constants.h:87
bool isMinusOne() const
This function will return true iff every bit in this constant is set to true.
Definition Constants.h:231
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:219
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition Constants.h:168
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * replaceUndefsWith(Constant *C, Constant *Replacement)
Try to replace undefined constant C or undefined elements in C with Replacement.
static LLVM_ABI Constant * mergeUndefsWith(Constant *C, Constant *Other)
Merges undefs of a Constant with another Constant, along with the undefs already present.
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
LLVM_ABI bool isZeroValue() const
Return true if the value is negative zero or null value.
Definition Constants.cpp:76
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
This instruction compares its operands according to the predicate given to the constructor.
This provides a helper for copying FMF from an instruction or setting specified flags.
Definition IRBuilder.h:93
static FMFSource intersect(Value *A, Value *B)
Intersect the FMF from two instructions.
Definition IRBuilder.h:107
This instruction compares its operands according to the predicate given to the constructor.
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
Value * CreateNot(Value *V, const Twine &Name="")
Definition IRBuilder.h:1808
Value * CreateBinOp(Instruction::BinaryOps Opc, Value *LHS, Value *RHS, const Twine &Name="", MDNode *FPMathTag=nullptr)
Definition IRBuilder.h:1708
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Definition IRBuilder.h:207
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitOr(BinaryOperator &I)
bool SimplifyAssociativeOrCommutative(BinaryOperator &I)
Performs a few simplifications for operators which are associative or commutative.
Value * foldUsingDistributiveLaws(BinaryOperator &I)
Tries to simplify binary operations which some other binary operation distributes over.
Instruction * foldBinOpShiftWithShift(BinaryOperator &I)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
Instruction * foldBinOpSelectBinOp(BinaryOperator &Op)
In some cases it is beneficial to fold a select into a binary operator.
bool sinkNotIntoLogicalOp(Instruction &I)
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitAnd(BinaryOperator &I)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW, bool NUW)
Common transforms for add / disjoint or.
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Value * SimplifyAddWithRemainder(BinaryOperator &I)
Tries to simplify add operations using the definition of remainder.
Instruction * visitXor(BinaryOperator &I)
bool SimplifyDemandedInstructionBits(Instruction &Inst)
Tries to simplify operands to an integer instruction based on its demanded bits.
Instruction * foldVectorBinop(BinaryOperator &Inst)
Canonicalize the position of binops relative to shufflevector.
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
The core instruction combiner logic.
SimplifyQuery SQ
const DataLayout & getDataLayout() const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
unsigned ComputeNumSignBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
const DataLayout & DL
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const Instruction *CxtI=nullptr, unsigned Depth=0) const
DominatorTree & DT
BuilderTy & Builder
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI void removeFromParent()
This method unlinks 'this' from the containing basic block, but does not delete it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
A wrapper class for inspecting calls to intrinsic functions.
This class represents a sign extension of integer types.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, const Instruction *MDFrom=nullptr)
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
LLVM_ABI unsigned getIntegerBitWidth() const
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:273
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Definition Type.h:246
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:352
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:197
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:230
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition Type.h:184
LLVM_ABI const fltSemantics & getFltSemantics() const
Definition Type.cpp:106
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
Value * getOperand(unsigned i) const
Definition User.h:233
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
iterator_range< user_iterator > users()
Definition Value.h:426
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
Definition Value.cpp:158
LLVM_ABI bool hasNUses(unsigned N) const
Return true if this Value has exactly N uses.
Definition Value.cpp:150
bool use_empty() const
Definition Value.h:346
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:322
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
Definition Value.cpp:403
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
Represents an op.with.overflow intrinsic.
This class represents zero extension of integer types.
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
const APInt & umin(const APInt &A, const APInt &B)
Determine the smaller of two APInts considered to be unsigned.
Definition APInt.h:2259
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > Tys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
class_match< BinaryOperator > m_BinOp()
Match an arbitrary binary operation and ignore it.
CmpClass_match< LHS, RHS, FCmpInst > m_FCmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
cstfp_pred_ty< is_inf > m_Inf()
Match a positive or negative infinity FP constant.
m_Intrinsic_Ty< Opnd0 >::Ty m_BitReverse(const Opnd0 &Op0)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
auto m_LogicalOp()
Matches either L && R or L || R where L and R are arbitrary values.
class_match< Constant > m_Constant()
Match an arbitrary Constant and ignore it.
ap_match< APInt > m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
ap_match< APInt > m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWSub(const LHS &L, const RHS &R)
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
cst_pred_ty< is_shifted_mask > m_ShiftedMask()
bind_ty< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
DisjointOr_match< LHS, RHS > m_DisjointOr(const LHS &L, const RHS &R)
constantexpr_match m_ConstantExpr()
Match a constant expression or a constant that contains a constant expression.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
ap_match< APFloat > m_APFloatAllowPoison(const APFloat *&Res)
Match APFloat while allowing poison in splat vector constants.
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
class_match< ConstantInt > m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
match_combine_or< CastInst_match< OpTy, SExtInst >, OpTy > m_SExtOrSelf(const OpTy &Op)
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
ShiftLike_match< LHS, Instruction::Shl > m_ShlOrSelf(const LHS &L, uint64_t &R)
Matches shl L, ConstShAmt or L itself (R will be set to zero in this case).
bind_ty< WithOverflowInst > m_WithOverflowInst(WithOverflowInst *&I)
Match a with overflow intrinsic, capturing it if we match.
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, CmpInst > m_SpecificCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
deferredval_ty< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
SpecificCmpClass_match< LHS, RHS, ICmpInst > m_SpecificICmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
class_match< CmpInst > m_Cmp()
Matches any compare instruction and ignore it.
cst_pred_ty< is_negated_power2 > m_NegatedPower2()
Match a integer or vector negated power-of-2.
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
DisjointOr_match< LHS, RHS, true > m_c_DisjointOr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, FCmpInst > m_SpecificFCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
match_combine_or< match_combine_or< MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty, true >, MaxMin_match< ICmpInst, LHS, RHS, smin_pred_ty, true > >, match_combine_or< MaxMin_match< ICmpInst, LHS, RHS, umax_pred_ty, true >, MaxMin_match< ICmpInst, LHS, RHS, umin_pred_ty, true > > > m_c_MaxOrMin(const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, SExtInst >, NNegZExt_match< OpTy > > m_SExtLike(const OpTy &Op)
Match either "sext" or "zext nneg".
MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty > m_SMax(const LHS &L, const RHS &R)
cst_pred_ty< is_maxsignedvalue > m_MaxSignedValue()
Match an integer or vector with values having all bits except for the high bit set (0x7f....
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
AnyBinaryOp_match< LHS, RHS, true > m_c_BinOp(const LHS &L, const RHS &R)
Matches a BinaryOperator with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
FNeg_match< OpTy > m_FNeg(const OpTy &X)
Match 'fneg X' as 'fsub -0.0, X'.
BinOpPred_match< LHS, RHS, is_shift_op > m_Shift(const LHS &L, const RHS &R)
Matches shift operations.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
m_Intrinsic_Ty< Opnd0 >::Ty m_BSwap(const Opnd0 &Op0)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ThreeOps_match< Val_t, Elt_t, Idx_t, Instruction::InsertElement > m_InsertElt(const Val_t &Val, const Elt_t &Elt, const Idx_t &Idx)
Matches InsertElementInst.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
m_Intrinsic_Ty< Opnd0 >::Ty m_FAbs(const Opnd0 &Op0)
m_Intrinsic_Ty< Opnd0, Opnd1 >::Ty m_CopySign(const Opnd0 &Op0, const Opnd1 &Op1)
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
match_combine_or< LTy, RTy > m_CombineOr(const LTy &L, const RTy &R)
Combine two pattern matchers matching L || R.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
NodeAddr< CodeNode * > Code
Definition RDFGraph.h:388
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
LLVM_ABI Intrinsic::ID getInverseMinMaxIntrinsic(Intrinsic::ID MinMaxID)
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
Definition Threading.h:280
@ Offset
Definition DWP.cpp:532
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
Constant * getPredForFCmpCode(unsigned Code, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getFCmpCode.
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
bool predicatesFoldable(CmpInst::Predicate P1, CmpInst::Predicate P2)
Return true if both predicates match sign or if at least one of them is an equality comparison (which...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
LLVM_ABI Value * simplifyOrInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Or, fold the result or return null.
LLVM_ABI Value * simplifyXorInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Xor, fold the result or return null.
LLVM_ABI bool isGuaranteedNotToBeUndef(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be undef, but may be poison.
LLVM_ABI bool matchSimpleRecurrence(const PHINode *P, BinaryOperator *&BO, Value *&Start, Value *&Step)
Attempt to match a simple first order recurrence cycle of the form: iv = phi Ty [Start,...
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
LLVM_ABI Constant * getLosslessUnsignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI bool recognizeBSwapOrBitReverseIdiom(Instruction *I, bool MatchBSwaps, bool MatchBitReversals, SmallVectorImpl< Instruction * > &InsertedInsts)
Try to match a bswap or bitreverse idiom.
Definition Local.cpp:3761
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * getLosslessSignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI Value * simplifyAndInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an And, fold the result or return null.
LLVM_ABI bool isKnownInversion(const Value *X, const Value *Y)
Return true iff:
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr int PoisonMaskElem
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
std::optional< DecomposedBitTest > decomposeBitTest(Value *Cond, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ And
Bitwise or logical AND of integers.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
APFloat neg(APFloat X)
Returns the negated value of the argument.
Definition APFloat.h:1551
cl::opt< bool > ProfcheckDisableMetadataFixes("profcheck-disable-metadata-fixes", cl::Hidden, cl::init(false), cl::desc("Disable metadata propagation fixes discovered through Issue #147390"))
unsigned getICmpCode(CmpInst::Predicate Pred)
Encode a icmp predicate into a three bit mask.
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
LLVM_ABI bool isGuaranteedNotToBePoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be poison, but may be undef.
std::pair< Value *, FPClassTest > fcmpToClassTest(FCmpInst::Predicate Pred, const Function &F, Value *LHS, Value *RHS, bool LookThroughSrc=true)
Returns a pair of values, which if passed to llvm.is.fpclass, returns the same result as an fcmp with...
unsigned getFCmpCode(CmpInst::Predicate CC)
Similar to getICmpCode but for FCmpInst.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
Constant * getPredForICmpCode(unsigned Code, bool Sign, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getICmpCode.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
bool isCombineableWith(const DecomposedBitMaskMul Other)
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:108
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:148
Matching combinators.
const DataLayout & DL
const Instruction * CxtI
const DominatorTree * DT
SimplifyQuery getWithInstruction(const Instruction *I) const
AssumptionCache * AC