LLVM 23.0.0git
InstCombineAndOrXor.cpp
Go to the documentation of this file.
1//===- InstCombineAndOrXor.cpp --------------------------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the visitAnd, visitOr, and visitXor functions.
10//
11//===----------------------------------------------------------------------===//
12
13#include "InstCombineInternal.h"
21#include "llvm/IR/Intrinsics.h"
26
27using namespace llvm;
28using namespace PatternMatch;
29
30#define DEBUG_TYPE "instcombine"
31
32namespace llvm {
34}
35
36/// This is the complement of getICmpCode, which turns an opcode and two
37/// operands into either a constant true or false, or a brand new ICmp
38/// instruction. The sign is passed in to determine which kind of predicate to
39/// use in the new icmp instruction.
40static Value *getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS,
41 InstCombiner::BuilderTy &Builder) {
42 ICmpInst::Predicate NewPred;
43 if (Constant *TorF = getPredForICmpCode(Code, Sign, LHS->getType(), NewPred))
44 return TorF;
45 return Builder.CreateICmp(NewPred, LHS, RHS);
46}
47
48/// This is the complement of getFCmpCode, which turns an opcode and two
49/// operands into either a FCmp instruction, or a true/false constant.
50static Value *getFCmpValue(unsigned Code, Value *LHS, Value *RHS,
51 InstCombiner::BuilderTy &Builder, FMFSource FMF) {
52 FCmpInst::Predicate NewPred;
53 if (Constant *TorF = getPredForFCmpCode(Code, LHS->getType(), NewPred))
54 return TorF;
55 return Builder.CreateFCmpFMF(NewPred, LHS, RHS, FMF);
56}
57
58/// Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise
59/// (V < Lo || V >= Hi). This method expects that Lo < Hi. IsSigned indicates
60/// whether to treat V, Lo, and Hi as signed or not.
62 const APInt &Hi, bool isSigned,
63 bool Inside) {
64 assert((isSigned ? Lo.slt(Hi) : Lo.ult(Hi)) &&
65 "Lo is not < Hi in range emission code!");
66
67 Type *Ty = V->getType();
68
69 // V >= Min && V < Hi --> V < Hi
70 // V < Min || V >= Hi --> V >= Hi
72 if (isSigned ? Lo.isMinSignedValue() : Lo.isMinValue()) {
73 Pred = isSigned ? ICmpInst::getSignedPredicate(Pred) : Pred;
74 return Builder.CreateICmp(Pred, V, ConstantInt::get(Ty, Hi));
75 }
76
77 // V >= Lo && V < Hi --> V - Lo u< Hi - Lo
78 // V < Lo || V >= Hi --> V - Lo u>= Hi - Lo
79 Value *VMinusLo =
80 Builder.CreateSub(V, ConstantInt::get(Ty, Lo), V->getName() + ".off");
81 Constant *HiMinusLo = ConstantInt::get(Ty, Hi - Lo);
82 return Builder.CreateICmp(Pred, VMinusLo, HiMinusLo);
83}
84
85/// Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns
86/// that can be simplified.
87/// One of A and B is considered the mask. The other is the value. This is
88/// described as the "AMask" or "BMask" part of the enum. If the enum contains
89/// only "Mask", then both A and B can be considered masks. If A is the mask,
90/// then it was proven that (A & C) == C. This is trivial if C == A or C == 0.
91/// If both A and C are constants, this proof is also easy.
92/// For the following explanations, we assume that A is the mask.
93///
94/// "AllOnes" declares that the comparison is true only if (A & B) == A or all
95/// bits of A are set in B.
96/// Example: (icmp eq (A & 3), 3) -> AMask_AllOnes
97///
98/// "AllZeros" declares that the comparison is true only if (A & B) == 0 or all
99/// bits of A are cleared in B.
100/// Example: (icmp eq (A & 3), 0) -> Mask_AllZeroes
101///
102/// "Mixed" declares that (A & B) == C and C might or might not contain any
103/// number of one bits and zero bits.
104/// Example: (icmp eq (A & 3), 1) -> AMask_Mixed
105///
106/// "Not" means that in above descriptions "==" should be replaced by "!=".
107/// Example: (icmp ne (A & 3), 3) -> AMask_NotAllOnes
108///
109/// If the mask A contains a single bit, then the following is equivalent:
110/// (icmp eq (A & B), A) equals (icmp ne (A & B), 0)
111/// (icmp ne (A & B), A) equals (icmp eq (A & B), 0)
124
125/// Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C)
126/// satisfies.
127static unsigned getMaskedICmpType(Value *A, Value *B, Value *C,
128 ICmpInst::Predicate Pred) {
129 const APInt *ConstA = nullptr, *ConstB = nullptr, *ConstC = nullptr;
130 match(A, m_APInt(ConstA));
131 match(B, m_APInt(ConstB));
132 match(C, m_APInt(ConstC));
133 bool IsEq = (Pred == ICmpInst::ICMP_EQ);
134 bool IsAPow2 = ConstA && ConstA->isPowerOf2();
135 bool IsBPow2 = ConstB && ConstB->isPowerOf2();
136 unsigned MaskVal = 0;
137 if (ConstC && ConstC->isZero()) {
138 // if C is zero, then both A and B qualify as mask
139 MaskVal |= (IsEq ? (Mask_AllZeros | AMask_Mixed | BMask_Mixed)
141 if (IsAPow2)
142 MaskVal |= (IsEq ? (AMask_NotAllOnes | AMask_NotMixed)
144 if (IsBPow2)
145 MaskVal |= (IsEq ? (BMask_NotAllOnes | BMask_NotMixed)
147 return MaskVal;
148 }
149
150 if (A == C) {
151 MaskVal |= (IsEq ? (AMask_AllOnes | AMask_Mixed)
153 if (IsAPow2)
154 MaskVal |= (IsEq ? (Mask_NotAllZeros | AMask_NotMixed)
156 } else if (ConstA && ConstC && ConstC->isSubsetOf(*ConstA)) {
157 MaskVal |= (IsEq ? AMask_Mixed : AMask_NotMixed);
158 }
159
160 if (B == C) {
161 MaskVal |= (IsEq ? (BMask_AllOnes | BMask_Mixed)
163 if (IsBPow2)
164 MaskVal |= (IsEq ? (Mask_NotAllZeros | BMask_NotMixed)
166 } else if (ConstB && ConstC && ConstC->isSubsetOf(*ConstB)) {
167 MaskVal |= (IsEq ? BMask_Mixed : BMask_NotMixed);
168 }
169
170 return MaskVal;
171}
172
173/// Convert an analysis of a masked ICmp into its equivalent if all boolean
174/// operations had the opposite sense. Since each "NotXXX" flag (recording !=)
175/// is adjacent to the corresponding normal flag (recording ==), this just
176/// involves swapping those bits over.
177static unsigned conjugateICmpMask(unsigned Mask) {
178 unsigned NewMask;
179 NewMask = (Mask & (AMask_AllOnes | BMask_AllOnes | Mask_AllZeros |
181 << 1;
182
183 NewMask |= (Mask & (AMask_NotAllOnes | BMask_NotAllOnes | Mask_NotAllZeros |
185 >> 1;
186
187 return NewMask;
188}
189
190// Adapts the external decomposeBitTest for local use.
192 Value *&Y, Value *&Z) {
193 auto Res =
194 llvm::decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
195 /*AllowNonZeroC=*/true, /*DecomposeAnd=*/true);
196 if (!Res)
197 return false;
198
199 Pred = Res->Pred;
200 X = Res->X;
201 Y = ConstantInt::get(X->getType(), Res->Mask);
202 Z = ConstantInt::get(X->getType(), Res->C);
203 return true;
204}
205
206/// Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
207/// Return the pattern classes (from MaskedICmpType) for the left hand side and
208/// the right hand side as a pair.
209/// LHS and RHS are the left hand side and the right hand side ICmps and PredL
210/// and PredR are their predicates, respectively.
211static std::optional<std::pair<unsigned, unsigned>>
214 ICmpInst::Predicate &PredR) {
215
216 // Here comes the tricky part:
217 // LHS might be of the form L11 & L12 == X, X == L21 & L22,
218 // and L11 & L12 == L21 & L22. The same goes for RHS.
219 // Now we must find those components L** and R**, that are equal, so
220 // that we can extract the parameters A, B, C, D, and E for the canonical
221 // above.
222
223 // Check whether the icmp can be decomposed into a bit test.
224 Value *L1, *L11, *L12, *L2, *L21, *L22;
225 if (decomposeBitTest(LHS, PredL, L11, L12, L2)) {
226 L21 = L22 = L1 = nullptr;
227 } else {
228 auto *LHSCMP = dyn_cast<ICmpInst>(LHS);
229 if (!LHSCMP)
230 return std::nullopt;
231
232 // Don't allow pointers. Splat vectors are fine.
233 if (!LHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
234 return std::nullopt;
235
236 PredL = LHSCMP->getPredicate();
237 L1 = LHSCMP->getOperand(0);
238 L2 = LHSCMP->getOperand(1);
239 // Look for ANDs in the LHS icmp.
240 if (!match(L1, m_And(m_Value(L11), m_Value(L12)))) {
241 // Any icmp can be viewed as being trivially masked; if it allows us to
242 // remove one, it's worth it.
243 L11 = L1;
245 }
246
247 if (!match(L2, m_And(m_Value(L21), m_Value(L22)))) {
248 L21 = L2;
250 }
251 }
252
253 // Bail if LHS was a icmp that can't be decomposed into an equality.
254 if (!ICmpInst::isEquality(PredL))
255 return std::nullopt;
256
257 Value *R11, *R12, *R2;
258 if (decomposeBitTest(RHS, PredR, R11, R12, R2)) {
259 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
260 A = R11;
261 D = R12;
262 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
263 A = R12;
264 D = R11;
265 } else {
266 return std::nullopt;
267 }
268 E = R2;
269 } else {
270 auto *RHSCMP = dyn_cast<ICmpInst>(RHS);
271 if (!RHSCMP)
272 return std::nullopt;
273 // Don't allow pointers. Splat vectors are fine.
274 if (!RHSCMP->getOperand(0)->getType()->isIntOrIntVectorTy())
275 return std::nullopt;
276
277 PredR = RHSCMP->getPredicate();
278
279 Value *R1 = RHSCMP->getOperand(0);
280 R2 = RHSCMP->getOperand(1);
281 bool Ok = false;
282 if (!match(R1, m_And(m_Value(R11), m_Value(R12)))) {
283 // As before, model no mask as a trivial mask if it'll let us do an
284 // optimization.
285 R11 = R1;
287 }
288
289 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
290 A = R11;
291 D = R12;
292 E = R2;
293 Ok = true;
294 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
295 A = R12;
296 D = R11;
297 E = R2;
298 Ok = true;
299 }
300
301 // Avoid matching against the -1 value we created for unmasked operand.
302 if (Ok && match(A, m_AllOnes()))
303 Ok = false;
304
305 // Look for ANDs on the right side of the RHS icmp.
306 if (!Ok) {
307 if (!match(R2, m_And(m_Value(R11), m_Value(R12)))) {
308 R11 = R2;
309 R12 = Constant::getAllOnesValue(R2->getType());
310 }
311
312 if (R11 == L11 || R11 == L12 || R11 == L21 || R11 == L22) {
313 A = R11;
314 D = R12;
315 E = R1;
316 } else if (R12 == L11 || R12 == L12 || R12 == L21 || R12 == L22) {
317 A = R12;
318 D = R11;
319 E = R1;
320 } else {
321 return std::nullopt;
322 }
323 }
324 }
325
326 // Bail if RHS was a icmp that can't be decomposed into an equality.
327 if (!ICmpInst::isEquality(PredR))
328 return std::nullopt;
329
330 if (L11 == A) {
331 B = L12;
332 C = L2;
333 } else if (L12 == A) {
334 B = L11;
335 C = L2;
336 } else if (L21 == A) {
337 B = L22;
338 C = L1;
339 } else if (L22 == A) {
340 B = L21;
341 C = L1;
342 }
343
344 unsigned LeftType = getMaskedICmpType(A, B, C, PredL);
345 unsigned RightType = getMaskedICmpType(A, D, E, PredR);
346 return std::optional<std::pair<unsigned, unsigned>>(
347 std::make_pair(LeftType, RightType));
348}
349
350/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single
351/// (icmp(A & X) ==/!= Y), where the left-hand side is of type Mask_NotAllZeros
352/// and the right hand side is of type BMask_Mixed. For example,
353/// (icmp (A & 12) != 0) & (icmp (A & 15) == 8) -> (icmp (A & 15) == 8).
354/// Also used for logical and/or, must be poison safe.
356 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E,
358 InstCombiner::BuilderTy &Builder) {
359 // We are given the canonical form:
360 // (icmp ne (A & B), 0) & (icmp eq (A & D), E).
361 // where D & E == E.
362 //
363 // If IsAnd is false, we get it in negated form:
364 // (icmp eq (A & B), 0) | (icmp ne (A & D), E) ->
365 // !((icmp ne (A & B), 0) & (icmp eq (A & D), E)).
366 //
367 // We currently handle the case of B, C, D, E are constant.
368 //
369 const APInt *BCst, *DCst, *OrigECst;
370 if (!match(B, m_APInt(BCst)) || !match(D, m_APInt(DCst)) ||
371 !match(E, m_APInt(OrigECst)))
372 return nullptr;
373
375
376 // Update E to the canonical form when D is a power of two and RHS is
377 // canonicalized as,
378 // (icmp ne (A & D), 0) -> (icmp eq (A & D), D) or
379 // (icmp ne (A & D), D) -> (icmp eq (A & D), 0).
380 APInt ECst = *OrigECst;
381 if (PredR != NewCC)
382 ECst ^= *DCst;
383
384 // If B or D is zero, skip because if LHS or RHS can be trivially folded by
385 // other folding rules and this pattern won't apply any more.
386 if (*BCst == 0 || *DCst == 0)
387 return nullptr;
388
389 // If B and D don't intersect, ie. (B & D) == 0, try to fold isNaN idiom:
390 // (icmp ne (A & FractionBits), 0) & (icmp eq (A & ExpBits), ExpBits)
391 // -> isNaN(A)
392 // Otherwise, we cannot deduce anything from it.
393 if (!BCst->intersects(*DCst)) {
394 Value *Src;
395 if (*DCst == ECst && match(A, m_ElementWiseBitCast(m_Value(Src))) &&
396 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
397 Attribute::StrictFP)) {
398 Type *Ty = Src->getType()->getScalarType();
399 if (!Ty->isIEEELikeFPTy())
400 return nullptr;
401
402 APInt ExpBits = APFloat::getInf(Ty->getFltSemantics()).bitcastToAPInt();
403 if (ECst != ExpBits)
404 return nullptr;
405 APInt FractionBits = ~ExpBits;
406 FractionBits.clearSignBit();
407 if (*BCst != FractionBits)
408 return nullptr;
409
410 return Builder.CreateFCmp(IsAnd ? FCmpInst::FCMP_UNO : FCmpInst::FCMP_ORD,
411 Src, ConstantFP::getZero(Src->getType()));
412 }
413 return nullptr;
414 }
415
416 // If the following two conditions are met:
417 //
418 // 1. mask B covers only a single bit that's not covered by mask D, that is,
419 // (B & (B ^ D)) is a power of 2 (in other words, B minus the intersection of
420 // B and D has only one bit set) and,
421 //
422 // 2. RHS (and E) indicates that the rest of B's bits are zero (in other
423 // words, the intersection of B and D is zero), that is, ((B & D) & E) == 0
424 //
425 // then that single bit in B must be one and thus the whole expression can be
426 // folded to
427 // (A & (B | D)) == (B & (B ^ D)) | E.
428 //
429 // For example,
430 // (icmp ne (A & 12), 0) & (icmp eq (A & 7), 1) -> (icmp eq (A & 15), 9)
431 // (icmp ne (A & 15), 0) & (icmp eq (A & 7), 0) -> (icmp eq (A & 15), 8)
432 if ((((*BCst & *DCst) & ECst) == 0) &&
433 (*BCst & (*BCst ^ *DCst)).isPowerOf2()) {
434 APInt BorD = *BCst | *DCst;
435 APInt BandBxorDorE = (*BCst & (*BCst ^ *DCst)) | ECst;
436 Value *NewMask = ConstantInt::get(A->getType(), BorD);
437 Value *NewMaskedValue = ConstantInt::get(A->getType(), BandBxorDorE);
438 Value *NewAnd = Builder.CreateAnd(A, NewMask);
439 return Builder.CreateICmp(NewCC, NewAnd, NewMaskedValue);
440 }
441
442 auto IsSubSetOrEqual = [](const APInt *C1, const APInt *C2) {
443 return (*C1 & *C2) == *C1;
444 };
445 auto IsSuperSetOrEqual = [](const APInt *C1, const APInt *C2) {
446 return (*C1 & *C2) == *C2;
447 };
448
449 // In the following, we consider only the cases where B is a superset of D, B
450 // is a subset of D, or B == D because otherwise there's at least one bit
451 // covered by B but not D, in which case we can't deduce much from it, so
452 // no folding (aside from the single must-be-one bit case right above.)
453 // For example,
454 // (icmp ne (A & 14), 0) & (icmp eq (A & 3), 1) -> no folding.
455 if (!IsSubSetOrEqual(BCst, DCst) && !IsSuperSetOrEqual(BCst, DCst))
456 return nullptr;
457
458 // At this point, either B is a superset of D, B is a subset of D or B == D.
459
460 // If E is zero, if B is a subset of (or equal to) D, LHS and RHS contradict
461 // and the whole expression becomes false (or true if negated), otherwise, no
462 // folding.
463 // For example,
464 // (icmp ne (A & 3), 0) & (icmp eq (A & 7), 0) -> false.
465 // (icmp ne (A & 15), 0) & (icmp eq (A & 3), 0) -> no folding.
466 if (ECst.isZero()) {
467 if (IsSubSetOrEqual(BCst, DCst))
468 return ConstantInt::get(LHS->getType(), !IsAnd);
469 return nullptr;
470 }
471
472 // At this point, B, D, E aren't zero and (B & D) == B, (B & D) == D or B ==
473 // D. If B is a superset of (or equal to) D, since E is not zero, LHS is
474 // subsumed by RHS (RHS implies LHS.) So the whole expression becomes
475 // RHS. For example,
476 // (icmp ne (A & 255), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
477 // (icmp ne (A & 15), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
478 if (IsSuperSetOrEqual(BCst, DCst)) {
479 // We can't guarantee that samesign hold after this fold.
480 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
481 ICmp->setSameSign(false);
482 return RHS;
483 }
484 // Otherwise, B is a subset of D. If B and E have a common bit set,
485 // ie. (B & E) != 0, then LHS is subsumed by RHS. For example.
486 // (icmp ne (A & 12), 0) & (icmp eq (A & 15), 8) -> (icmp eq (A & 15), 8).
487 assert(IsSubSetOrEqual(BCst, DCst) && "Precondition due to above code");
488 if ((*BCst & ECst) != 0) {
489 // We can't guarantee that samesign hold after this fold.
490 if (auto *ICmp = dyn_cast<ICmpInst>(RHS))
491 ICmp->setSameSign(false);
492 return RHS;
493 }
494 // Otherwise, LHS and RHS contradict and the whole expression becomes false
495 // (or true if negated.) For example,
496 // (icmp ne (A & 7), 0) & (icmp eq (A & 15), 8) -> false.
497 // (icmp ne (A & 6), 0) & (icmp eq (A & 15), 8) -> false.
498 return ConstantInt::get(LHS->getType(), !IsAnd);
499}
500
501/// Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single
502/// (icmp(A & X) ==/!= Y), where the left-hand side and the right hand side
503/// aren't of the common mask pattern type.
504/// Also used for logical and/or, must be poison safe.
506 Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D,
508 unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder) {
510 "Expected equality predicates for masked type of icmps.");
511 // Handle Mask_NotAllZeros-BMask_Mixed cases.
512 // (icmp ne/eq (A & B), C) &/| (icmp eq/ne (A & D), E), or
513 // (icmp eq/ne (A & B), C) &/| (icmp ne/eq (A & D), E)
514 // which gets swapped to
515 // (icmp ne/eq (A & D), E) &/| (icmp eq/ne (A & B), C).
516 if (!IsAnd) {
517 LHSMask = conjugateICmpMask(LHSMask);
518 RHSMask = conjugateICmpMask(RHSMask);
519 }
520 if ((LHSMask & Mask_NotAllZeros) && (RHSMask & BMask_Mixed)) {
522 LHS, RHS, IsAnd, A, B, D, E, PredL, PredR, Builder)) {
523 return V;
524 }
525 } else if ((LHSMask & BMask_Mixed) && (RHSMask & Mask_NotAllZeros)) {
527 RHS, LHS, IsAnd, A, D, B, C, PredR, PredL, Builder)) {
528 return V;
529 }
530 }
531 return nullptr;
532}
533
534/// Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E)
535/// into a single (icmp(A & X) ==/!= Y).
537 bool IsLogical,
539 const SimplifyQuery &Q) {
540 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
541 ICmpInst::Predicate PredL, PredR;
542 std::optional<std::pair<unsigned, unsigned>> MaskPair =
543 getMaskedTypeForICmpPair(A, B, C, D, E, LHS, RHS, PredL, PredR);
544 if (!MaskPair)
545 return nullptr;
547 "Expected equality predicates for masked type of icmps.");
548 unsigned LHSMask = MaskPair->first;
549 unsigned RHSMask = MaskPair->second;
550 unsigned Mask = LHSMask & RHSMask;
551 if (Mask == 0) {
552 // Even if the two sides don't share a common pattern, check if folding can
553 // still happen.
555 LHS, RHS, IsAnd, A, B, C, D, E, PredL, PredR, LHSMask, RHSMask,
556 Builder))
557 return V;
558 return nullptr;
559 }
560
561 // In full generality:
562 // (icmp (A & B) Op C) | (icmp (A & D) Op E)
563 // == ![ (icmp (A & B) !Op C) & (icmp (A & D) !Op E) ]
564 //
565 // If the latter can be converted into (icmp (A & X) Op Y) then the former is
566 // equivalent to (icmp (A & X) !Op Y).
567 //
568 // Therefore, we can pretend for the rest of this function that we're dealing
569 // with the conjunction, provided we flip the sense of any comparisons (both
570 // input and output).
571
572 // In most cases we're going to produce an EQ for the "&&" case.
574 if (!IsAnd) {
575 // Convert the masking analysis into its equivalent with negated
576 // comparisons.
577 Mask = conjugateICmpMask(Mask);
578 }
579
580 if (Mask & Mask_AllZeros) {
581 // (icmp eq (A & B), 0) & (icmp eq (A & D), 0)
582 // -> (icmp eq (A & (B|D)), 0)
583 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
584 return nullptr; // TODO: Use freeze?
585 Value *NewOr = Builder.CreateOr(B, D);
586 Value *NewAnd = Builder.CreateAnd(A, NewOr);
587 // We can't use C as zero because we might actually handle
588 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
589 // with B and D, having a single bit set.
590 Value *Zero = Constant::getNullValue(A->getType());
591 return Builder.CreateICmp(NewCC, NewAnd, Zero);
592 }
593 if (Mask & BMask_AllOnes) {
594 // (icmp eq (A & B), B) & (icmp eq (A & D), D)
595 // -> (icmp eq (A & (B|D)), (B|D))
596 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
597 return nullptr; // TODO: Use freeze?
598 Value *NewOr = Builder.CreateOr(B, D);
599 Value *NewAnd = Builder.CreateAnd(A, NewOr);
600 return Builder.CreateICmp(NewCC, NewAnd, NewOr);
601 }
602 if (Mask & AMask_AllOnes) {
603 // (icmp eq (A & B), A) & (icmp eq (A & D), A)
604 // -> (icmp eq (A & (B&D)), A)
605 if (IsLogical && !isGuaranteedNotToBeUndefOrPoison(D))
606 return nullptr; // TODO: Use freeze?
607 Value *NewAnd1 = Builder.CreateAnd(B, D);
608 Value *NewAnd2 = Builder.CreateAnd(A, NewAnd1);
609 return Builder.CreateICmp(NewCC, NewAnd2, A);
610 }
611
612 const APInt *ConstB, *ConstD;
613 if (match(B, m_APInt(ConstB)) && match(D, m_APInt(ConstD))) {
614 if (Mask & (Mask_NotAllZeros | BMask_NotAllOnes)) {
615 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0) and
616 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
617 // -> (icmp ne (A & B), 0) or (icmp ne (A & D), 0)
618 // Only valid if one of the masks is a superset of the other (check "B&D"
619 // is the same as either B or D).
620 APInt NewMask = *ConstB & *ConstD;
621 if (NewMask == *ConstB)
622 return LHS;
623 if (NewMask == *ConstD) {
624 if (IsLogical) {
625 if (auto *RHSI = dyn_cast<Instruction>(RHS))
626 RHSI->dropPoisonGeneratingFlags();
627 }
628 return RHS;
629 }
630 }
631
632 if (Mask & AMask_NotAllOnes) {
633 // (icmp ne (A & B), B) & (icmp ne (A & D), D)
634 // -> (icmp ne (A & B), A) or (icmp ne (A & D), A)
635 // Only valid if one of the masks is a superset of the other (check "B|D"
636 // is the same as either B or D).
637 APInt NewMask = *ConstB | *ConstD;
638 if (NewMask == *ConstB)
639 return LHS;
640 if (NewMask == *ConstD)
641 return RHS;
642 }
643
644 if (Mask & (BMask_Mixed | BMask_NotMixed)) {
645 // Mixed:
646 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
647 // We already know that B & C == C && D & E == E.
648 // If we can prove that (B & D) & (C ^ E) == 0, that is, the bits of
649 // C and E, which are shared by both the mask B and the mask D, don't
650 // contradict, then we can transform to
651 // -> (icmp eq (A & (B|D)), (C|E))
652 // Currently, we only handle the case of B, C, D, and E being constant.
653 // We can't simply use C and E because we might actually handle
654 // (icmp ne (A & B), B) & (icmp eq (A & D), D)
655 // with B and D, having a single bit set.
656
657 // NotMixed:
658 // (icmp ne (A & B), C) & (icmp ne (A & D), E)
659 // -> (icmp ne (A & (B & D)), (C & E))
660 // Check the intersection (B & D) for inequality.
661 // Assume that (B & D) == B || (B & D) == D, i.e B/D is a subset of D/B
662 // and (B & D) & (C ^ E) == 0, bits of C and E, which are shared by both
663 // the B and the D, don't contradict. Note that we can assume (~B & C) ==
664 // 0 && (~D & E) == 0, previous operation should delete these icmps if it
665 // hadn't been met.
666
667 const APInt *OldConstC, *OldConstE;
668 if (!match(C, m_APInt(OldConstC)) || !match(E, m_APInt(OldConstE)))
669 return nullptr;
670
671 auto FoldBMixed = [&](ICmpInst::Predicate CC, bool IsNot) -> Value * {
672 CC = IsNot ? CmpInst::getInversePredicate(CC) : CC;
673 const APInt ConstC = PredL != CC ? *ConstB ^ *OldConstC : *OldConstC;
674 const APInt ConstE = PredR != CC ? *ConstD ^ *OldConstE : *OldConstE;
675
676 if (((*ConstB & *ConstD) & (ConstC ^ ConstE)).getBoolValue())
677 return IsNot ? nullptr : ConstantInt::get(LHS->getType(), !IsAnd);
678
679 if (IsNot && !ConstB->isSubsetOf(*ConstD) &&
680 !ConstD->isSubsetOf(*ConstB))
681 return nullptr;
682
683 APInt BD, CE;
684 if (IsNot) {
685 BD = *ConstB & *ConstD;
686 CE = ConstC & ConstE;
687 } else {
688 BD = *ConstB | *ConstD;
689 CE = ConstC | ConstE;
690 }
691 Value *NewAnd = Builder.CreateAnd(A, BD);
692 Value *CEVal = ConstantInt::get(A->getType(), CE);
693 return Builder.CreateICmp(CC, NewAnd, CEVal);
694 };
695
696 if (Mask & BMask_Mixed)
697 return FoldBMixed(NewCC, false);
698 if (Mask & BMask_NotMixed) // can be else also
699 return FoldBMixed(NewCC, true);
700 }
701 }
702
703 // (icmp eq (A & B), 0) | (icmp eq (A & D), 0)
704 // -> (icmp ne (A & (B|D)), (B|D))
705 // (icmp ne (A & B), 0) & (icmp ne (A & D), 0)
706 // -> (icmp eq (A & (B|D)), (B|D))
707 // iff B and D is known to be a power of two
708 if (Mask & Mask_NotAllZeros &&
709 isKnownToBeAPowerOfTwo(B, /*OrZero=*/false, Q) &&
710 isKnownToBeAPowerOfTwo(D, /*OrZero=*/false, Q)) {
711 // If this is a logical and/or, then we must prevent propagation of a
712 // poison value from the RHS by inserting freeze.
713 if (IsLogical)
714 D = Builder.CreateFreeze(D);
715 Value *Mask = Builder.CreateOr(B, D);
716 Value *Masked = Builder.CreateAnd(A, Mask);
717 return Builder.CreateICmp(NewCC, Masked, Mask);
718 }
719 return nullptr;
720}
721
722/// Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
723/// Example: (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
724/// If \p Inverted is true then the check is for the inverted range, e.g.
725/// (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
727 bool Inverted) {
728 // Check the lower range comparison, e.g. x >= 0
729 // InstCombine already ensured that if there is a constant it's on the RHS.
730 ConstantInt *RangeStart = dyn_cast<ConstantInt>(Cmp0->getOperand(1));
731 if (!RangeStart)
732 return nullptr;
733
734 ICmpInst::Predicate Pred0 = (Inverted ? Cmp0->getInversePredicate() :
735 Cmp0->getPredicate());
736
737 // Accept x > -1 or x >= 0 (after potentially inverting the predicate).
738 if (!((Pred0 == ICmpInst::ICMP_SGT && RangeStart->isMinusOne()) ||
739 (Pred0 == ICmpInst::ICMP_SGE && RangeStart->isZero())))
740 return nullptr;
741
742 ICmpInst::Predicate Pred1 = (Inverted ? Cmp1->getInversePredicate() :
743 Cmp1->getPredicate());
744
745 Value *Input = Cmp0->getOperand(0);
746 Value *Cmp1Op0 = Cmp1->getOperand(0);
747 Value *Cmp1Op1 = Cmp1->getOperand(1);
748 Value *RangeEnd;
749 if (match(Cmp1Op0, m_SExtOrSelf(m_Specific(Input)))) {
750 // For the upper range compare we have: icmp x, n
751 Input = Cmp1Op0;
752 RangeEnd = Cmp1Op1;
753 } else if (match(Cmp1Op1, m_SExtOrSelf(m_Specific(Input)))) {
754 // For the upper range compare we have: icmp n, x
755 Input = Cmp1Op1;
756 RangeEnd = Cmp1Op0;
757 Pred1 = ICmpInst::getSwappedPredicate(Pred1);
758 } else {
759 return nullptr;
760 }
761
762 // Check the upper range comparison, e.g. x < n
763 ICmpInst::Predicate NewPred;
764 switch (Pred1) {
765 case ICmpInst::ICMP_SLT: NewPred = ICmpInst::ICMP_ULT; break;
766 case ICmpInst::ICMP_SLE: NewPred = ICmpInst::ICMP_ULE; break;
767 default: return nullptr;
768 }
769
770 // This simplification is only valid if the upper range is not negative.
771 KnownBits Known = computeKnownBits(RangeEnd, Cmp1);
772 if (!Known.isNonNegative())
773 return nullptr;
774
775 if (Inverted)
776 NewPred = ICmpInst::getInversePredicate(NewPred);
777
778 return Builder.CreateICmp(NewPred, Input, RangeEnd);
779}
780
781// (or (icmp eq X, 0), (icmp eq X, Pow2OrZero))
782// -> (icmp eq (and X, Pow2OrZero), X)
783// (and (icmp ne X, 0), (icmp ne X, Pow2OrZero))
784// -> (icmp ne (and X, Pow2OrZero), X)
785static Value *
787 ICmpInst *LHS, ICmpInst *RHS, bool IsAnd,
788 const SimplifyQuery &Q) {
790 // Make sure we have right compares for our op.
791 if (LHS->getPredicate() != Pred || RHS->getPredicate() != Pred)
792 return nullptr;
793
794 // Make it so we can match LHS against the (icmp eq/ne X, 0) just for
795 // simplicity.
796 if (match(RHS->getOperand(1), m_Zero()))
797 std::swap(LHS, RHS);
798
799 Value *Pow2, *Op;
800 // Match the desired pattern:
801 // LHS: (icmp eq/ne X, 0)
802 // RHS: (icmp eq/ne X, Pow2OrZero)
803 // Skip if Pow2OrZero is 1. Either way it gets folded to (icmp ugt X, 1) but
804 // this form ends up slightly less canonical.
805 // We could potentially be more sophisticated than requiring LHS/RHS
806 // be one-use. We don't create additional instructions if only one
807 // of them is one-use. So cases where one is one-use and the other
808 // is two-use might be profitable.
809 if (!match(LHS, m_OneUse(m_ICmp(Pred, m_Value(Op), m_Zero()))) ||
810 !match(RHS, m_OneUse(m_c_ICmp(Pred, m_Specific(Op), m_Value(Pow2)))) ||
811 match(Pow2, m_One()) ||
812 !isKnownToBeAPowerOfTwo(Pow2, Q.DL, /*OrZero=*/true, Q.AC, Q.CxtI, Q.DT))
813 return nullptr;
814
815 Value *And = Builder.CreateAnd(Op, Pow2);
816 return Builder.CreateICmp(Pred, And, Op);
817}
818
819/// General pattern:
820/// X & Y
821///
822/// Where Y is checking that all the high bits (covered by a mask 4294967168)
823/// are uniform, i.e. %arg & 4294967168 can be either 4294967168 or 0
824/// Pattern can be one of:
825/// %t = add i32 %arg, 128
826/// %r = icmp ult i32 %t, 256
827/// Or
828/// %t0 = shl i32 %arg, 24
829/// %t1 = ashr i32 %t0, 24
830/// %r = icmp eq i32 %t1, %arg
831/// Or
832/// %t0 = trunc i32 %arg to i8
833/// %t1 = sext i8 %t0 to i32
834/// %r = icmp eq i32 %t1, %arg
835/// This pattern is a signed truncation check.
836///
837/// And X is checking that some bit in that same mask is zero.
838/// I.e. can be one of:
839/// %r = icmp sgt i32 %arg, -1
840/// Or
841/// %t = and i32 %arg, 2147483648
842/// %r = icmp eq i32 %t, 0
843///
844/// Since we are checking that all the bits in that mask are the same,
845/// and a particular bit is zero, what we are really checking is that all the
846/// masked bits are zero.
847/// So this should be transformed to:
848/// %r = icmp ult i32 %arg, 128
850 Instruction &CxtI,
851 InstCombiner::BuilderTy &Builder) {
852 assert(CxtI.getOpcode() == Instruction::And);
853
854 // Match icmp ult (add %arg, C01), C1 (C1 == C01 << 1; powers of two)
855 auto tryToMatchSignedTruncationCheck = [](ICmpInst *ICmp, Value *&X,
856 APInt &SignBitMask) -> bool {
857 const APInt *I01, *I1; // powers of two; I1 == I01 << 1
859 m_Add(m_Value(X), m_Power2(I01)),
860 m_Power2(I1))) &&
861 I1->ugt(*I01) && I01->shl(1) == *I1))
862 return false;
863 // Which bit is the new sign bit as per the 'signed truncation' pattern?
864 SignBitMask = *I01;
865 return true;
866 };
867
868 // One icmp needs to be 'signed truncation check'.
869 // We need to match this first, else we will mismatch commutative cases.
870 Value *X1;
871 APInt HighestBit;
872 ICmpInst *OtherICmp;
873 if (tryToMatchSignedTruncationCheck(ICmp1, X1, HighestBit))
874 OtherICmp = ICmp0;
875 else if (tryToMatchSignedTruncationCheck(ICmp0, X1, HighestBit))
876 OtherICmp = ICmp1;
877 else
878 return nullptr;
879
880 assert(HighestBit.isPowerOf2() && "expected to be power of two (non-zero)");
881
882 // Try to match/decompose into: icmp eq (X & Mask), 0
883 auto tryToDecompose = [](ICmpInst *ICmp, Value *&X,
884 APInt &UnsetBitsMask) -> bool {
885 CmpPredicate Pred = ICmp->getPredicate();
886 // Can it be decomposed into icmp eq (X & Mask), 0 ?
888 ICmp->getOperand(0), ICmp->getOperand(1), Pred,
889 /*LookThroughTrunc=*/false, /*AllowNonZeroC=*/false,
890 /*DecomposeAnd=*/true);
891 if (Res && Res->Pred == ICmpInst::ICMP_EQ) {
892 X = Res->X;
893 UnsetBitsMask = Res->Mask;
894 return true;
895 }
896
897 return false;
898 };
899
900 // And the other icmp needs to be decomposable into a bit test.
901 Value *X0;
902 APInt UnsetBitsMask;
903 if (!tryToDecompose(OtherICmp, X0, UnsetBitsMask))
904 return nullptr;
905
906 assert(!UnsetBitsMask.isZero() && "empty mask makes no sense.");
907
908 // Are they working on the same value?
909 Value *X;
910 if (X1 == X0) {
911 // Ok as is.
912 X = X1;
913 } else if (match(X0, m_Trunc(m_Specific(X1)))) {
914 UnsetBitsMask = UnsetBitsMask.zext(X1->getType()->getScalarSizeInBits());
915 X = X1;
916 } else
917 return nullptr;
918
919 // So which bits should be uniform as per the 'signed truncation check'?
920 // (all the bits starting with (i.e. including) HighestBit)
921 APInt SignBitsMask = ~(HighestBit - 1U);
922
923 // UnsetBitsMask must have some common bits with SignBitsMask,
924 if (!UnsetBitsMask.intersects(SignBitsMask))
925 return nullptr;
926
927 // Does UnsetBitsMask contain any bits outside of SignBitsMask?
928 if (!UnsetBitsMask.isSubsetOf(SignBitsMask)) {
929 APInt OtherHighestBit = (~UnsetBitsMask) + 1U;
930 if (!OtherHighestBit.isPowerOf2())
931 return nullptr;
932 HighestBit = APIntOps::umin(HighestBit, OtherHighestBit);
933 }
934 // Else, if it does not, then all is ok as-is.
935
936 // %r = icmp ult %X, SignBit
937 return Builder.CreateICmpULT(X, ConstantInt::get(X->getType(), HighestBit),
938 CxtI.getName() + ".simplified");
939}
940
941/// Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and
942/// fold (icmp ne ctpop(X) 1) & (icmp ne X 0) into (icmp ugt ctpop(X) 1).
943/// Also used for logical and/or, must be poison safe if range attributes are
944/// dropped.
945static Value *foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd,
947 InstCombinerImpl &IC) {
948 CmpPredicate Pred0, Pred1;
949 Value *X;
950 if (!match(Cmp0, m_ICmp(Pred0, m_Ctpop(m_Value(X)), m_SpecificInt(1))) ||
951 !match(Cmp1, m_ICmp(Pred1, m_Specific(X), m_ZeroInt())))
952 return nullptr;
953
954 auto *CtPop = cast<Instruction>(Cmp0->getOperand(0));
955 if (IsAnd && Pred0 == ICmpInst::ICMP_NE && Pred1 == ICmpInst::ICMP_NE) {
956 // Drop range attributes and re-infer them in the next iteration.
957 CtPop->dropPoisonGeneratingAnnotations();
958 IC.addToWorklist(CtPop);
959 return Builder.CreateICmpUGT(CtPop, ConstantInt::get(CtPop->getType(), 1));
960 }
961 if (!IsAnd && Pred0 == ICmpInst::ICMP_EQ && Pred1 == ICmpInst::ICMP_EQ) {
962 // Drop range attributes and re-infer them in the next iteration.
963 CtPop->dropPoisonGeneratingAnnotations();
964 IC.addToWorklist(CtPop);
965 return Builder.CreateICmpULT(CtPop, ConstantInt::get(CtPop->getType(), 2));
966 }
967
968 return nullptr;
969}
970
971/// Reduce a pair of compares that check if a value has exactly 1 bit set.
972/// Also used for logical and/or, must be poison safe if range attributes are
973/// dropped.
974static Value *foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd,
976 InstCombinerImpl &IC) {
977 // Handle 'and' / 'or' commutation: make the equality check the first operand.
978 if (JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_NE)
979 std::swap(Cmp0, Cmp1);
980 else if (!JoinedByAnd && Cmp1->getPredicate() == ICmpInst::ICMP_EQ)
981 std::swap(Cmp0, Cmp1);
982
983 // (X != 0) && (ctpop(X) u< 2) --> ctpop(X) == 1
984 Value *X;
985 if (JoinedByAnd &&
988 m_SpecificInt(2)))) {
989 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
990 // Drop range attributes and re-infer them in the next iteration.
991 CtPop->dropPoisonGeneratingAnnotations();
992 IC.addToWorklist(CtPop);
993 return Builder.CreateICmpEQ(CtPop, ConstantInt::get(CtPop->getType(), 1));
994 }
995 // (X == 0) || (ctpop(X) u> 1) --> ctpop(X) != 1
996 if (!JoinedByAnd &&
999 m_SpecificInt(1)))) {
1000 auto *CtPop = cast<Instruction>(Cmp1->getOperand(0));
1001 // Drop range attributes and re-infer them in the next iteration.
1002 CtPop->dropPoisonGeneratingAnnotations();
1003 IC.addToWorklist(CtPop);
1004 return Builder.CreateICmpNE(CtPop, ConstantInt::get(CtPop->getType(), 1));
1005 }
1006 return nullptr;
1007}
1008
1009/// Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff
1010/// B is a contiguous set of ones starting from the most significant bit
1011/// (negative power of 2), D and E are equal, and D is a contiguous set of ones
1012/// starting at the most significant zero bit in B. Parameter B supports masking
1013/// using undef/poison in either scalar or vector values.
1015 Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL,
1018 "Expected equality predicates for masked type of icmps.");
1019 if (PredL != ICmpInst::ICMP_EQ || PredR != ICmpInst::ICMP_NE)
1020 return nullptr;
1021
1022 if (!match(B, m_NegatedPower2()) || !match(D, m_ShiftedMask()) ||
1023 !match(E, m_ShiftedMask()))
1024 return nullptr;
1025
1026 // Test scalar arguments for conversion. B has been validated earlier to be a
1027 // negative power of two and thus is guaranteed to have one or more contiguous
1028 // ones starting from the MSB followed by zero or more contiguous zeros. D has
1029 // been validated earlier to be a shifted set of one or more contiguous ones.
1030 // In order to match, B leading ones and D leading zeros should be equal. The
1031 // predicate that B be a negative power of 2 prevents the condition of there
1032 // ever being zero leading ones. Thus 0 == 0 cannot occur. The predicate that
1033 // D always be a shifted mask prevents the condition of D equaling 0. This
1034 // prevents matching the condition where B contains the maximum number of
1035 // leading one bits (-1) and D contains the maximum number of leading zero
1036 // bits (0).
1037 auto isReducible = [](const Value *B, const Value *D, const Value *E) {
1038 const APInt *BCst, *DCst, *ECst;
1039 return match(B, m_APIntAllowPoison(BCst)) && match(D, m_APInt(DCst)) &&
1040 match(E, m_APInt(ECst)) && *DCst == *ECst &&
1041 (isa<PoisonValue>(B) ||
1042 (BCst->countLeadingOnes() == DCst->countLeadingZeros()));
1043 };
1044
1045 // Test vector type arguments for conversion.
1046 if (const auto *BVTy = dyn_cast<VectorType>(B->getType())) {
1047 const auto *BFVTy = dyn_cast<FixedVectorType>(BVTy);
1048 const auto *BConst = dyn_cast<Constant>(B);
1049 const auto *DConst = dyn_cast<Constant>(D);
1050 const auto *EConst = dyn_cast<Constant>(E);
1051
1052 if (!BFVTy || !BConst || !DConst || !EConst)
1053 return nullptr;
1054
1055 for (unsigned I = 0; I != BFVTy->getNumElements(); ++I) {
1056 const auto *BElt = BConst->getAggregateElement(I);
1057 const auto *DElt = DConst->getAggregateElement(I);
1058 const auto *EElt = EConst->getAggregateElement(I);
1059
1060 if (!BElt || !DElt || !EElt)
1061 return nullptr;
1062 if (!isReducible(BElt, DElt, EElt))
1063 return nullptr;
1064 }
1065 } else {
1066 // Test scalar type arguments for conversion.
1067 if (!isReducible(B, D, E))
1068 return nullptr;
1069 }
1070 return Builder.CreateICmp(ICmpInst::ICMP_ULT, A, D);
1071}
1072
1073/// Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) &
1074/// (icmp(X & M) != M)) into (icmp X u< M). Where P is a power of 2, M < P, and
1075/// M is a contiguous shifted mask starting at the right most significant zero
1076/// bit in P. SGT is supported as when P is the largest representable power of
1077/// 2, an earlier optimization converts the expression into (icmp X s> -1).
1078/// Parameter P supports masking using undef/poison in either scalar or vector
1079/// values.
1081 bool JoinedByAnd,
1082 InstCombiner::BuilderTy &Builder) {
1083 if (!JoinedByAnd)
1084 return nullptr;
1085 Value *A = nullptr, *B = nullptr, *C = nullptr, *D = nullptr, *E = nullptr;
1086 ICmpInst::Predicate CmpPred0, CmpPred1;
1087 // Assuming P is a 2^n, getMaskedTypeForICmpPair will normalize (icmp X u<
1088 // 2^n) into (icmp (X & ~(2^n-1)) == 0) and (icmp X s> -1) into (icmp (X &
1089 // SignMask) == 0).
1090 std::optional<std::pair<unsigned, unsigned>> MaskPair =
1091 getMaskedTypeForICmpPair(A, B, C, D, E, Cmp0, Cmp1, CmpPred0, CmpPred1);
1092 if (!MaskPair)
1093 return nullptr;
1094
1095 const auto compareBMask = BMask_NotMixed | BMask_NotAllOnes;
1096 unsigned CmpMask0 = MaskPair->first;
1097 unsigned CmpMask1 = MaskPair->second;
1098 if ((CmpMask0 & Mask_AllZeros) && (CmpMask1 == compareBMask)) {
1099 if (Value *V = foldNegativePower2AndShiftedMask(A, B, D, E, CmpPred0,
1100 CmpPred1, Builder))
1101 return V;
1102 } else if ((CmpMask0 == compareBMask) && (CmpMask1 & Mask_AllZeros)) {
1103 if (Value *V = foldNegativePower2AndShiftedMask(A, D, B, C, CmpPred1,
1104 CmpPred0, Builder))
1105 return V;
1106 }
1107 return nullptr;
1108}
1109
1110/// Commuted variants are assumed to be handled by calling this function again
1111/// with the parameters swapped.
1113 ICmpInst *UnsignedICmp, bool IsAnd,
1114 const SimplifyQuery &Q,
1115 InstCombiner::BuilderTy &Builder) {
1116 Value *ZeroCmpOp;
1117 CmpPredicate EqPred;
1118 if (!match(ZeroICmp, m_ICmp(EqPred, m_Value(ZeroCmpOp), m_Zero())) ||
1119 !ICmpInst::isEquality(EqPred))
1120 return nullptr;
1121
1122 CmpPredicate UnsignedPred;
1123
1124 Value *A, *B;
1125 if (match(UnsignedICmp,
1126 m_c_ICmp(UnsignedPred, m_Specific(ZeroCmpOp), m_Value(A))) &&
1127 match(ZeroCmpOp, m_c_Add(m_Specific(A), m_Value(B))) &&
1128 (ZeroICmp->hasOneUse() || UnsignedICmp->hasOneUse())) {
1129 auto GetKnownNonZeroAndOther = [&](Value *&NonZero, Value *&Other) {
1130 if (!isKnownNonZero(NonZero, Q))
1131 std::swap(NonZero, Other);
1132 return isKnownNonZero(NonZero, Q);
1133 };
1134
1135 // Given ZeroCmpOp = (A + B)
1136 // ZeroCmpOp < A && ZeroCmpOp != 0 --> (0-X) < Y iff
1137 // ZeroCmpOp >= A || ZeroCmpOp == 0 --> (0-X) >= Y iff
1138 // with X being the value (A/B) that is known to be non-zero,
1139 // and Y being remaining value.
1140 if (UnsignedPred == ICmpInst::ICMP_ULT && EqPred == ICmpInst::ICMP_NE &&
1141 IsAnd && GetKnownNonZeroAndOther(B, A))
1142 return Builder.CreateICmpULT(Builder.CreateNeg(B), A);
1143 if (UnsignedPred == ICmpInst::ICMP_UGE && EqPred == ICmpInst::ICMP_EQ &&
1144 !IsAnd && GetKnownNonZeroAndOther(B, A))
1145 return Builder.CreateICmpUGE(Builder.CreateNeg(B), A);
1146 }
1147
1148 return nullptr;
1149}
1150
1151struct IntPart {
1153 unsigned StartBit;
1154 unsigned NumBits;
1155};
1156
1157/// Match an extraction of bits from an integer.
1158static std::optional<IntPart> matchIntPart(Value *V) {
1159 Value *X;
1160 if (!match(V, m_OneUse(m_Trunc(m_Value(X)))))
1161 return std::nullopt;
1162
1163 unsigned NumOriginalBits = X->getType()->getScalarSizeInBits();
1164 unsigned NumExtractedBits = V->getType()->getScalarSizeInBits();
1165 Value *Y;
1166 const APInt *Shift;
1167 // For a trunc(lshr Y, Shift) pattern, make sure we're only extracting bits
1168 // from Y, not any shifted-in zeroes.
1169 if (match(X, m_OneUse(m_LShr(m_Value(Y), m_APInt(Shift)))) &&
1170 Shift->ule(NumOriginalBits - NumExtractedBits))
1171 return {{Y, (unsigned)Shift->getZExtValue(), NumExtractedBits}};
1172 return {{X, 0, NumExtractedBits}};
1173}
1174
1175/// Materialize an extraction of bits from an integer in IR.
1176static Value *extractIntPart(const IntPart &P, IRBuilderBase &Builder) {
1177 Value *V = P.From;
1178 if (P.StartBit)
1179 V = Builder.CreateLShr(V, P.StartBit);
1180 Type *TruncTy = V->getType()->getWithNewBitWidth(P.NumBits);
1181 if (TruncTy != V->getType())
1182 V = Builder.CreateTrunc(V, TruncTy);
1183 return V;
1184}
1185
1186/// (icmp eq X0, Y0) & (icmp eq X1, Y1) -> icmp eq X01, Y01
1187/// (icmp ne X0, Y0) | (icmp ne X1, Y1) -> icmp ne X01, Y01
1188/// where X0, X1 and Y0, Y1 are adjacent parts extracted from an integer.
1189Value *InstCombinerImpl::foldEqOfParts(Value *Cmp0, Value *Cmp1, bool IsAnd) {
1190 if (!Cmp0->hasOneUse() || !Cmp1->hasOneUse())
1191 return nullptr;
1192
1194 auto GetMatchPart = [&](Value *CmpV,
1195 unsigned OpNo) -> std::optional<IntPart> {
1196 assert(CmpV->getType()->isIntOrIntVectorTy(1) && "Must be bool");
1197
1198 Value *X, *Y;
1199 // icmp ne (and x, 1), (and y, 1) <=> trunc (xor x, y) to i1
1200 // icmp eq (and x, 1), (and y, 1) <=> not (trunc (xor x, y) to i1)
1201 if (Pred == CmpInst::ICMP_NE
1202 ? match(CmpV, m_Trunc(m_Xor(m_Value(X), m_Value(Y))))
1203 : match(CmpV, m_Not(m_Trunc(m_Xor(m_Value(X), m_Value(Y))))))
1204 return {{OpNo == 0 ? X : Y, 0, 1}};
1205
1206 auto *Cmp = dyn_cast<ICmpInst>(CmpV);
1207 if (!Cmp)
1208 return std::nullopt;
1209
1210 if (Pred == Cmp->getPredicate())
1211 return matchIntPart(Cmp->getOperand(OpNo));
1212
1213 const APInt *C;
1214 // (icmp eq (lshr x, C), (lshr y, C)) gets optimized to:
1215 // (icmp ult (xor x, y), 1 << C) so also look for that.
1216 if (Pred == CmpInst::ICMP_EQ && Cmp->getPredicate() == CmpInst::ICMP_ULT) {
1217 if (!match(Cmp->getOperand(1), m_Power2(C)) ||
1218 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1219 return std::nullopt;
1220 }
1221
1222 // (icmp ne (lshr x, C), (lshr y, C)) gets optimized to:
1223 // (icmp ugt (xor x, y), (1 << C) - 1) so also look for that.
1224 else if (Pred == CmpInst::ICMP_NE &&
1225 Cmp->getPredicate() == CmpInst::ICMP_UGT) {
1226 if (!match(Cmp->getOperand(1), m_LowBitMask(C)) ||
1227 !match(Cmp->getOperand(0), m_Xor(m_Value(), m_Value())))
1228 return std::nullopt;
1229 } else {
1230 return std::nullopt;
1231 }
1232
1233 unsigned From = Pred == CmpInst::ICMP_NE ? C->popcount() : C->countr_zero();
1234 Instruction *I = cast<Instruction>(Cmp->getOperand(0));
1235 return {{I->getOperand(OpNo), From, C->getBitWidth() - From}};
1236 };
1237
1238 std::optional<IntPart> L0 = GetMatchPart(Cmp0, 0);
1239 std::optional<IntPart> R0 = GetMatchPart(Cmp0, 1);
1240 std::optional<IntPart> L1 = GetMatchPart(Cmp1, 0);
1241 std::optional<IntPart> R1 = GetMatchPart(Cmp1, 1);
1242 if (!L0 || !R0 || !L1 || !R1)
1243 return nullptr;
1244
1245 // Make sure the LHS/RHS compare a part of the same value, possibly after
1246 // an operand swap.
1247 if (L0->From != L1->From || R0->From != R1->From) {
1248 if (L0->From != R1->From || R0->From != L1->From)
1249 return nullptr;
1250 std::swap(L1, R1);
1251 }
1252
1253 // Make sure the extracted parts are adjacent, canonicalizing to L0/R0 being
1254 // the low part and L1/R1 being the high part.
1255 if (L0->StartBit + L0->NumBits != L1->StartBit ||
1256 R0->StartBit + R0->NumBits != R1->StartBit) {
1257 if (L1->StartBit + L1->NumBits != L0->StartBit ||
1258 R1->StartBit + R1->NumBits != R0->StartBit)
1259 return nullptr;
1260 std::swap(L0, L1);
1261 std::swap(R0, R1);
1262 }
1263
1264 // We can simplify to a comparison of these larger parts of the integers.
1265 IntPart L = {L0->From, L0->StartBit, L0->NumBits + L1->NumBits};
1266 IntPart R = {R0->From, R0->StartBit, R0->NumBits + R1->NumBits};
1269 return Builder.CreateICmp(Pred, LValue, RValue);
1270}
1271
1272/// Reduce logic-of-compares with equality to a constant by substituting a
1273/// common operand with the constant. Callers are expected to call this with
1274/// Cmp0/Cmp1 switched to handle logic op commutativity.
1276 bool IsAnd, bool IsLogical,
1277 InstCombiner::BuilderTy &Builder,
1278 const SimplifyQuery &Q,
1279 Instruction &I) {
1280 // Match an equality compare with a non-poison constant as Cmp0.
1281 // Also, give up if the compare can be constant-folded to avoid looping.
1282 CmpPredicate Pred0;
1283 Value *X;
1284 Constant *C;
1285 if (!match(Cmp0, m_ICmp(Pred0, m_Value(X), m_Constant(C))) ||
1287 return nullptr;
1288 if ((IsAnd && Pred0 != ICmpInst::ICMP_EQ) ||
1289 (!IsAnd && Pred0 != ICmpInst::ICMP_NE))
1290 return nullptr;
1291
1292 // The other compare must include a common operand (X). Canonicalize the
1293 // common operand as operand 1 (Pred1 is swapped if the common operand was
1294 // operand 0).
1295 Value *Y;
1296 CmpPredicate Pred1;
1297 if (!match(Cmp1, m_c_ICmp(Pred1, m_Value(Y), m_Specific(X))))
1298 return nullptr;
1299
1300 // Replace variable with constant value equivalence to remove a variable use:
1301 // (X == C) && (Y Pred1 X) --> (X == C) && (Y Pred1 C)
1302 // (X != C) || (Y Pred1 X) --> (X != C) || (Y Pred1 C)
1303 // Can think of the 'or' substitution with the 'and' bool equivalent:
1304 // A || B --> A || (!A && B)
1305 Value *SubstituteCmp = simplifyICmpInst(Pred1, Y, C, Q);
1306 if (!SubstituteCmp) {
1307 // If we need to create a new instruction, require that the old compare can
1308 // be removed.
1309 if (!Cmp1->hasOneUse())
1310 return nullptr;
1311 SubstituteCmp = Builder.CreateICmp(Pred1, Y, C);
1312 }
1313 if (IsLogical) {
1314 Instruction *MDFrom =
1316 return IsAnd ? Builder.CreateLogicalAnd(Cmp0, SubstituteCmp, "", MDFrom)
1317 : Builder.CreateLogicalOr(Cmp0, SubstituteCmp, "", MDFrom);
1318 }
1319 return Builder.CreateBinOp(IsAnd ? Instruction::And : Instruction::Or, Cmp0,
1320 SubstituteCmp);
1321}
1322
1323/// Fold (icmp Pred1 V1, C1) & (icmp Pred2 V2, C2)
1324/// or (icmp Pred1 V1, C1) | (icmp Pred2 V2, C2)
1325/// into a single comparison using range-based reasoning.
1326/// NOTE: This is also used for logical and/or, must be poison-safe!
1327Value *InstCombinerImpl::foldAndOrOfICmpsUsingRanges(ICmpInst *ICmp1,
1328 ICmpInst *ICmp2,
1329 bool IsAnd) {
1330 // Return (V, CR) for a range check idiom V in CR.
1331 auto MatchExactRangeCheck =
1332 [](ICmpInst *ICmp) -> std::optional<std::pair<Value *, ConstantRange>> {
1333 const APInt *C;
1334 if (!match(ICmp->getOperand(1), m_APInt(C)))
1335 return std::nullopt;
1336 Value *LHS = ICmp->getOperand(0);
1337 CmpPredicate Pred = ICmp->getPredicate();
1338 Value *X;
1339 // Match (x & NegPow2) ==/!= C
1340 const APInt *Mask;
1341 if (ICmpInst::isEquality(Pred) &&
1343 C->countr_zero() >= Mask->countr_zero()) {
1344 ConstantRange CR(*C, *C - *Mask);
1345 if (Pred == ICmpInst::ICMP_NE)
1346 CR = CR.inverse();
1347 return std::make_pair(X, CR);
1348 }
1349 ConstantRange CR = ConstantRange::makeExactICmpRegion(Pred, *C);
1350 // Match (add X, C1) pred C
1351 // TODO: investigate whether we should apply the one-use check on m_AddLike.
1352 const APInt *C1;
1353 if (match(LHS, m_AddLike(m_Value(X), m_APInt(C1))))
1354 return std::make_pair(X, CR.subtract(*C1));
1355 return std::make_pair(LHS, CR);
1356 };
1357
1358 auto RC1 = MatchExactRangeCheck(ICmp1);
1359 if (!RC1)
1360 return nullptr;
1361
1362 auto RC2 = MatchExactRangeCheck(ICmp2);
1363 if (!RC2)
1364 return nullptr;
1365
1366 auto &[V1, CR1] = *RC1;
1367 auto &[V2, CR2] = *RC2;
1368 if (V1 != V2)
1369 return nullptr;
1370
1371 // For 'and', we use the De Morgan's Laws to simplify the implementation.
1372 if (IsAnd) {
1373 CR1 = CR1.inverse();
1374 CR2 = CR2.inverse();
1375 }
1376
1377 Type *Ty = V1->getType();
1378 Value *NewV = V1;
1379 std::optional<ConstantRange> CR = CR1.exactUnionWith(CR2);
1380 if (!CR) {
1381 if (!(ICmp1->hasOneUse() && ICmp2->hasOneUse()) || CR1.isWrappedSet() ||
1382 CR2.isWrappedSet())
1383 return nullptr;
1384
1385 // Check whether we have equal-size ranges that only differ by one bit.
1386 // In that case we can apply a mask to map one range onto the other.
1387 APInt LowerDiff = CR1.getLower() ^ CR2.getLower();
1388 APInt UpperDiff = (CR1.getUpper() - 1) ^ (CR2.getUpper() - 1);
1389 APInt CR1Size = CR1.getUpper() - CR1.getLower();
1390 if (!LowerDiff.isPowerOf2() || LowerDiff != UpperDiff ||
1391 CR1Size != CR2.getUpper() - CR2.getLower())
1392 return nullptr;
1393
1394 CR = CR1.getLower().ult(CR2.getLower()) ? CR1 : CR2;
1395 NewV = Builder.CreateAnd(NewV, ConstantInt::get(Ty, ~LowerDiff));
1396 }
1397
1398 if (IsAnd)
1399 CR = CR->inverse();
1400
1401 CmpInst::Predicate NewPred;
1402 APInt NewC, Offset;
1403 CR->getEquivalentICmp(NewPred, NewC, Offset);
1404
1405 if (Offset != 0)
1406 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
1407 return Builder.CreateICmp(NewPred, NewV, ConstantInt::get(Ty, NewC));
1408}
1409
1410/// Matches canonical form of isnan, fcmp ord x, 0
1414
1415/// Matches fcmp u__ x, +/-inf
1420
1421/// and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1422///
1423/// Clang emits this pattern for doing an isfinite check in __builtin_isnormal.
1425 FCmpInst *RHS) {
1426 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1427 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1428 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1429
1430 if (!matchIsNotNaN(PredL, LHS0, LHS1) ||
1431 !matchUnorderedInfCompare(PredR, RHS0, RHS1))
1432 return nullptr;
1433
1434 return Builder.CreateFCmpFMF(FCmpInst::getOrderedPredicate(PredR), RHS0, RHS1,
1436}
1437
1438Value *InstCombinerImpl::foldLogicOfFCmps(FCmpInst *LHS, FCmpInst *RHS,
1439 bool IsAnd, bool IsLogicalSelect) {
1440 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
1441 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
1442 FCmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
1443
1444 if (LHS0 == RHS1 && RHS0 == LHS1) {
1445 // Swap RHS operands to match LHS.
1446 PredR = FCmpInst::getSwappedPredicate(PredR);
1447 std::swap(RHS0, RHS1);
1448 }
1449
1450 // Simplify (fcmp cc0 x, y) & (fcmp cc1 x, y).
1451 // Suppose the relation between x and y is R, where R is one of
1452 // U(1000), L(0100), G(0010) or E(0001), and CC0 and CC1 are the bitmasks for
1453 // testing the desired relations.
1454 //
1455 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1456 // bool(R & CC0) && bool(R & CC1)
1457 // = bool((R & CC0) & (R & CC1))
1458 // = bool(R & (CC0 & CC1)) <= by re-association, commutation, and idempotency
1459 //
1460 // Since (R & CC0) and (R & CC1) are either R or 0, we actually have this:
1461 // bool(R & CC0) || bool(R & CC1)
1462 // = bool((R & CC0) | (R & CC1))
1463 // = bool(R & (CC0 | CC1)) <= by reversed distribution (contribution? ;)
1464 if (LHS0 == RHS0 && LHS1 == RHS1) {
1465 unsigned FCmpCodeL = getFCmpCode(PredL);
1466 unsigned FCmpCodeR = getFCmpCode(PredR);
1467 unsigned NewPred = IsAnd ? FCmpCodeL & FCmpCodeR : FCmpCodeL | FCmpCodeR;
1468
1469 // Intersect the fast math flags.
1470 // TODO: We can union the fast math flags unless this is a logical select.
1471 return getFCmpValue(NewPred, LHS0, LHS1, Builder,
1473 }
1474
1475 // This transform is not valid for a logical select.
1476 if (!IsLogicalSelect &&
1477 ((PredL == FCmpInst::FCMP_ORD && PredR == FCmpInst::FCMP_ORD && IsAnd) ||
1478 (PredL == FCmpInst::FCMP_UNO && PredR == FCmpInst::FCMP_UNO &&
1479 !IsAnd))) {
1480 if (LHS0->getType() != RHS0->getType())
1481 return nullptr;
1482
1483 // FCmp canonicalization ensures that (fcmp ord/uno X, X) and
1484 // (fcmp ord/uno X, C) will be transformed to (fcmp X, +0.0).
1485 if (match(LHS1, m_PosZeroFP()) && match(RHS1, m_PosZeroFP())) {
1486 // Ignore the constants because they are obviously not NANs:
1487 // (fcmp ord x, 0.0) & (fcmp ord y, 0.0) -> (fcmp ord x, y)
1488 // (fcmp uno x, 0.0) | (fcmp uno y, 0.0) -> (fcmp uno x, y)
1489 return Builder.CreateFCmpFMF(PredL, LHS0, RHS0,
1491 }
1492 }
1493
1494 // This transform is not valid for a logical select.
1495 if (!IsLogicalSelect && IsAnd &&
1496 stripSignOnlyFPOps(LHS0) == stripSignOnlyFPOps(RHS0)) {
1497 // and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
1498 // and (fcmp ord x, 0), (fcmp u* fabs(x), inf) -> fcmp o* x, inf
1500 return Left;
1502 return Right;
1503 }
1504
1505 // Turn at least two fcmps with constants into llvm.is.fpclass.
1506 //
1507 // If we can represent a combined value test with one class call, we can
1508 // potentially eliminate 4-6 instructions. If we can represent a test with a
1509 // single fcmp with fneg and fabs, that's likely a better canonical form.
1510 if (LHS->hasOneUse() && RHS->hasOneUse()) {
1511 auto [ClassValRHS, ClassMaskRHS] =
1512 fcmpToClassTest(PredR, *RHS->getFunction(), RHS0, RHS1);
1513 if (ClassValRHS) {
1514 auto [ClassValLHS, ClassMaskLHS] =
1515 fcmpToClassTest(PredL, *LHS->getFunction(), LHS0, LHS1);
1516 if (ClassValLHS == ClassValRHS) {
1517 unsigned CombinedMask = IsAnd ? (ClassMaskLHS & ClassMaskRHS)
1518 : (ClassMaskLHS | ClassMaskRHS);
1519 return Builder.CreateIntrinsic(
1520 Intrinsic::is_fpclass, {ClassValLHS->getType()},
1521 {ClassValLHS, Builder.getInt32(CombinedMask)});
1522 }
1523 }
1524 }
1525
1526 // Canonicalize the range check idiom:
1527 // and (fcmp olt/ole/ult/ule x, C), (fcmp ogt/oge/ugt/uge x, -C)
1528 // --> fabs(x) olt/ole/ult/ule C
1529 // or (fcmp ogt/oge/ugt/uge x, C), (fcmp olt/ole/ult/ule x, -C)
1530 // --> fabs(x) ogt/oge/ugt/uge C
1531 // TODO: Generalize to handle a negated variable operand?
1532 const APFloat *LHSC, *RHSC;
1533 if (LHS0 == RHS0 && LHS->hasOneUse() && RHS->hasOneUse() &&
1534 FCmpInst::getSwappedPredicate(PredL) == PredR &&
1535 match(LHS1, m_APFloatAllowPoison(LHSC)) &&
1536 match(RHS1, m_APFloatAllowPoison(RHSC)) &&
1537 LHSC->bitwiseIsEqual(neg(*RHSC))) {
1538 auto IsLessThanOrLessEqual = [](FCmpInst::Predicate Pred) {
1539 switch (Pred) {
1540 case FCmpInst::FCMP_OLT:
1541 case FCmpInst::FCMP_OLE:
1542 case FCmpInst::FCMP_ULT:
1543 case FCmpInst::FCMP_ULE:
1544 return true;
1545 default:
1546 return false;
1547 }
1548 };
1549 if (IsLessThanOrLessEqual(IsAnd ? PredR : PredL)) {
1550 std::swap(LHSC, RHSC);
1551 std::swap(PredL, PredR);
1552 }
1553 if (IsLessThanOrLessEqual(IsAnd ? PredL : PredR)) {
1554 FastMathFlags NewFlag = LHS->getFastMathFlags();
1555 if (!IsLogicalSelect)
1556 NewFlag |= RHS->getFastMathFlags();
1557
1558 Value *FAbs = Builder.CreateFAbs(LHS0, NewFlag);
1559 return Builder.CreateFCmpFMF(
1560 PredL, FAbs, ConstantFP::get(LHS0->getType(), *LHSC), NewFlag);
1561 }
1562 }
1563
1564 return nullptr;
1565}
1566
1567/// Match an fcmp against a special value that performs a test possible by
1568/// llvm.is.fpclass.
1569static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal,
1570 uint64_t &ClassMask) {
1571 auto *FCmp = dyn_cast<FCmpInst>(Op);
1572 if (!FCmp || !FCmp->hasOneUse())
1573 return false;
1574
1575 std::tie(ClassVal, ClassMask) =
1576 fcmpToClassTest(FCmp->getPredicate(), *FCmp->getParent()->getParent(),
1577 FCmp->getOperand(0), FCmp->getOperand(1));
1578 return ClassVal != nullptr;
1579}
1580
1581/// or (is_fpclass x, mask0), (is_fpclass x, mask1)
1582/// -> is_fpclass x, (mask0 | mask1)
1583/// and (is_fpclass x, mask0), (is_fpclass x, mask1)
1584/// -> is_fpclass x, (mask0 & mask1)
1585/// xor (is_fpclass x, mask0), (is_fpclass x, mask1)
1586/// -> is_fpclass x, (mask0 ^ mask1)
1587Instruction *InstCombinerImpl::foldLogicOfIsFPClass(BinaryOperator &BO,
1588 Value *Op0, Value *Op1) {
1589 Value *ClassVal0 = nullptr;
1590 Value *ClassVal1 = nullptr;
1591 uint64_t ClassMask0, ClassMask1;
1592
1593 // Restrict to folding one fcmp into one is.fpclass for now, don't introduce a
1594 // new class.
1595 //
1596 // TODO: Support forming is.fpclass out of 2 separate fcmps when codegen is
1597 // better.
1598
1599 bool IsLHSClass =
1601 m_Value(ClassVal0), m_ConstantInt(ClassMask0))));
1602 bool IsRHSClass =
1604 m_Value(ClassVal1), m_ConstantInt(ClassMask1))));
1605 if ((((IsLHSClass || matchIsFPClassLikeFCmp(Op0, ClassVal0, ClassMask0)) &&
1606 (IsRHSClass || matchIsFPClassLikeFCmp(Op1, ClassVal1, ClassMask1)))) &&
1607 ClassVal0 == ClassVal1) {
1608 unsigned NewClassMask;
1609 switch (BO.getOpcode()) {
1610 case Instruction::And:
1611 NewClassMask = ClassMask0 & ClassMask1;
1612 break;
1613 case Instruction::Or:
1614 NewClassMask = ClassMask0 | ClassMask1;
1615 break;
1616 case Instruction::Xor:
1617 NewClassMask = ClassMask0 ^ ClassMask1;
1618 break;
1619 default:
1620 llvm_unreachable("not a binary logic operator");
1621 }
1622
1623 if (IsLHSClass) {
1624 auto *II = cast<IntrinsicInst>(Op0);
1625 II->setArgOperand(
1626 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1627 return replaceInstUsesWith(BO, II);
1628 }
1629
1630 if (IsRHSClass) {
1631 auto *II = cast<IntrinsicInst>(Op1);
1632 II->setArgOperand(
1633 1, ConstantInt::get(II->getArgOperand(1)->getType(), NewClassMask));
1634 return replaceInstUsesWith(BO, II);
1635 }
1636
1637 CallInst *NewClass =
1638 Builder.CreateIntrinsic(Intrinsic::is_fpclass, {ClassVal0->getType()},
1639 {ClassVal0, Builder.getInt32(NewClassMask)});
1640 return replaceInstUsesWith(BO, NewClass);
1641 }
1642
1643 return nullptr;
1644}
1645
1646/// Look for the pattern that conditionally negates a value via math operations:
1647/// cond.splat = sext i1 cond
1648/// sub = add cond.splat, x
1649/// xor = xor sub, cond.splat
1650/// and rewrite it to do the same, but via logical operations:
1651/// value.neg = sub 0, value
1652/// cond = select i1 neg, value.neg, value
1653Instruction *InstCombinerImpl::canonicalizeConditionalNegationViaMathToSelect(
1654 BinaryOperator &I) {
1655 assert(I.getOpcode() == BinaryOperator::Xor && "Only for xor!");
1656 Value *Cond, *X;
1657 // As per complexity ordering, `xor` is not commutative here.
1658 if (!match(&I, m_c_BinOp(m_OneUse(m_Value()), m_Value())) ||
1659 !match(I.getOperand(1), m_SExt(m_Value(Cond))) ||
1660 !Cond->getType()->isIntOrIntVectorTy(1) ||
1661 !match(I.getOperand(0), m_c_Add(m_SExt(m_Specific(Cond)), m_Value(X))))
1662 return nullptr;
1663 return createSelectInstWithUnknownProfile(
1664 Cond, Builder.CreateNeg(X, X->getName() + ".neg"), X);
1665}
1666
1667/// This a limited reassociation for a special case (see above) where we are
1668/// checking if two values are either both NAN (unordered) or not-NAN (ordered).
1669/// This could be handled more generally in '-reassociation', but it seems like
1670/// an unlikely pattern for a large number of logic ops and fcmps.
1672 InstCombiner::BuilderTy &Builder) {
1673 Instruction::BinaryOps Opcode = BO.getOpcode();
1674 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1675 "Expecting and/or op for fcmp transform");
1676
1677 // There are 4 commuted variants of the pattern. Canonicalize operands of this
1678 // logic op so an fcmp is operand 0 and a matching logic op is operand 1.
1679 Value *Op0 = BO.getOperand(0), *Op1 = BO.getOperand(1), *X;
1680 if (match(Op1, m_FCmp(m_Value(), m_AnyZeroFP())))
1681 std::swap(Op0, Op1);
1682
1683 // Match inner binop and the predicate for combining 2 NAN checks into 1.
1684 Value *BO10, *BO11;
1685 FCmpInst::Predicate NanPred = Opcode == Instruction::And ? FCmpInst::FCMP_ORD
1687 if (!match(Op0, m_SpecificFCmp(NanPred, m_Value(X), m_AnyZeroFP())) ||
1688 !match(Op1, m_BinOp(Opcode, m_Value(BO10), m_Value(BO11))))
1689 return nullptr;
1690
1691 // The inner logic op must have a matching fcmp operand.
1692 Value *Y;
1693 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1694 X->getType() != Y->getType())
1695 std::swap(BO10, BO11);
1696
1697 if (!match(BO10, m_SpecificFCmp(NanPred, m_Value(Y), m_AnyZeroFP())) ||
1698 X->getType() != Y->getType())
1699 return nullptr;
1700
1701 // and (fcmp ord X, 0), (and (fcmp ord Y, 0), Z) --> and (fcmp ord X, Y), Z
1702 // or (fcmp uno X, 0), (or (fcmp uno Y, 0), Z) --> or (fcmp uno X, Y), Z
1703 // Intersect FMF from the 2 source fcmps.
1704 Value *NewFCmp =
1705 Builder.CreateFCmpFMF(NanPred, X, Y, FMFSource::intersect(Op0, BO10));
1706 return BinaryOperator::Create(Opcode, NewFCmp, BO11);
1707}
1708
1709/// Match variations of De Morgan's Laws:
1710/// (~A & ~B) == (~(A | B))
1711/// (~A | ~B) == (~(A & B))
1713 InstCombiner &IC) {
1714 const Instruction::BinaryOps Opcode = I.getOpcode();
1715 assert((Opcode == Instruction::And || Opcode == Instruction::Or) &&
1716 "Trying to match De Morgan's Laws with something other than and/or");
1717
1718 // Flip the logic operation.
1719 const Instruction::BinaryOps FlippedOpcode =
1720 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
1721
1722 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1723 Value *A, *B;
1724 if (match(Op0, m_OneUse(m_Not(m_Value(A)))) &&
1725 match(Op1, m_OneUse(m_Not(m_Value(B)))) &&
1726 !IC.isFreeToInvert(A, A->hasOneUse()) &&
1727 !IC.isFreeToInvert(B, B->hasOneUse())) {
1728 Value *AndOr =
1729 IC.Builder.CreateBinOp(FlippedOpcode, A, B, I.getName() + ".demorgan");
1730 return BinaryOperator::CreateNot(AndOr);
1731 }
1732
1733 // The 'not' ops may require reassociation.
1734 // (A & ~B) & ~C --> A & ~(B | C)
1735 // (~B & A) & ~C --> A & ~(B | C)
1736 // (A | ~B) | ~C --> A | ~(B & C)
1737 // (~B | A) | ~C --> A | ~(B & C)
1738 Value *C;
1739 if (match(Op0, m_OneUse(m_c_BinOp(Opcode, m_Value(A), m_Not(m_Value(B))))) &&
1740 match(Op1, m_Not(m_Value(C)))) {
1741 Value *FlippedBO = IC.Builder.CreateBinOp(FlippedOpcode, B, C);
1742 return BinaryOperator::Create(Opcode, A, IC.Builder.CreateNot(FlippedBO));
1743 }
1744
1745 return nullptr;
1746}
1747
1748bool InstCombinerImpl::shouldOptimizeCast(CastInst *CI) {
1749 Value *CastSrc = CI->getOperand(0);
1750
1751 // Noop casts and casts of constants should be eliminated trivially.
1752 if (CI->getSrcTy() == CI->getDestTy() || isa<Constant>(CastSrc))
1753 return false;
1754
1755 // If this cast is paired with another cast that can be eliminated, we prefer
1756 // to have it eliminated.
1757 if (const auto *PrecedingCI = dyn_cast<CastInst>(CastSrc))
1758 if (isEliminableCastPair(PrecedingCI, CI))
1759 return false;
1760
1761 return true;
1762}
1763
1764/// Fold {and,or,xor} (cast X), C.
1766 InstCombinerImpl &IC) {
1768 if (!C)
1769 return nullptr;
1770
1771 auto LogicOpc = Logic.getOpcode();
1772 Type *DestTy = Logic.getType();
1773 Type *SrcTy = Cast->getSrcTy();
1774
1775 // Move the logic operation ahead of a zext or sext if the constant is
1776 // unchanged in the smaller source type. Performing the logic in a smaller
1777 // type may provide more information to later folds, and the smaller logic
1778 // instruction may be cheaper (particularly in the case of vectors).
1779 Value *X;
1780 auto &DL = IC.getDataLayout();
1781 if (match(Cast, m_OneUse(m_ZExt(m_Value(X))))) {
1782 PreservedCastFlags Flags;
1783 if (Constant *TruncC = getLosslessUnsignedTrunc(C, SrcTy, DL, &Flags)) {
1784 // LogicOpc (zext X), C --> zext (LogicOpc X, C)
1785 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1786 auto *ZExt = new ZExtInst(NewOp, DestTy);
1787 ZExt->setNonNeg(Flags.NNeg);
1788 ZExt->andIRFlags(Cast);
1789 return ZExt;
1790 }
1791 }
1792
1793 if (match(Cast, m_OneUse(m_SExtLike(m_Value(X))))) {
1794 if (Constant *TruncC = getLosslessSignedTrunc(C, SrcTy, DL)) {
1795 // LogicOpc (sext X), C --> sext (LogicOpc X, C)
1796 Value *NewOp = IC.Builder.CreateBinOp(LogicOpc, X, TruncC);
1797 return new SExtInst(NewOp, DestTy);
1798 }
1799 }
1800
1801 return nullptr;
1802}
1803
1804/// Fold {and,or,xor} (cast X), Y.
1805Instruction *InstCombinerImpl::foldCastedBitwiseLogic(BinaryOperator &I) {
1806 auto LogicOpc = I.getOpcode();
1807 assert(I.isBitwiseLogicOp() && "Unexpected opcode for bitwise logic folding");
1808
1809 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
1810
1811 // fold bitwise(A >> BW - 1, zext(icmp)) (BW is the scalar bits of the
1812 // type of A)
1813 // -> bitwise(zext(A < 0), zext(icmp))
1814 // -> zext(bitwise(A < 0, icmp))
1815 auto FoldBitwiseICmpZeroWithICmp = [&](Value *Op0,
1816 Value *Op1) -> Instruction * {
1817 Value *A;
1818 bool IsMatched =
1819 match(Op0,
1821 m_Value(A),
1822 m_SpecificInt(Op0->getType()->getScalarSizeInBits() - 1)))) &&
1823 match(Op1, m_OneUse(m_ZExt(m_ICmp(m_Value(), m_Value()))));
1824
1825 if (!IsMatched)
1826 return nullptr;
1827
1828 auto *ICmpL =
1829 Builder.CreateICmpSLT(A, Constant::getNullValue(A->getType()));
1830 auto *ICmpR = cast<ZExtInst>(Op1)->getOperand(0);
1831 auto *BitwiseOp = Builder.CreateBinOp(LogicOpc, ICmpL, ICmpR);
1832
1833 return new ZExtInst(BitwiseOp, Op0->getType());
1834 };
1835
1836 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op0, Op1))
1837 return Ret;
1838
1839 if (auto *Ret = FoldBitwiseICmpZeroWithICmp(Op1, Op0))
1840 return Ret;
1841
1842 CastInst *Cast0 = dyn_cast<CastInst>(Op0);
1843 if (!Cast0)
1844 return nullptr;
1845
1846 // This must be a cast from an integer or integer vector source type to allow
1847 // transformation of the logic operation to the source type.
1848 Type *DestTy = I.getType();
1849 Type *SrcTy = Cast0->getSrcTy();
1850 if (!SrcTy->isIntOrIntVectorTy())
1851 return nullptr;
1852
1853 if (Instruction *Ret = foldLogicCastConstant(I, Cast0, *this))
1854 return Ret;
1855
1856 CastInst *Cast1 = dyn_cast<CastInst>(Op1);
1857 if (!Cast1)
1858 return nullptr;
1859
1860 // Both operands of the logic operation are casts. The casts must be the
1861 // same kind for reduction.
1862 Instruction::CastOps CastOpcode = Cast0->getOpcode();
1863 if (CastOpcode != Cast1->getOpcode())
1864 return nullptr;
1865
1866 // Can't fold it profitably if no one of casts has one use.
1867 if (!Cast0->hasOneUse() && !Cast1->hasOneUse())
1868 return nullptr;
1869
1870 Value *X, *Y;
1871 if (match(Cast0, m_ZExtOrSExt(m_Value(X))) &&
1872 match(Cast1, m_ZExtOrSExt(m_Value(Y)))) {
1873 // Cast the narrower source to the wider source type.
1874 unsigned XNumBits = X->getType()->getScalarSizeInBits();
1875 unsigned YNumBits = Y->getType()->getScalarSizeInBits();
1876 if (XNumBits != YNumBits) {
1877 // Cast the narrower source to the wider source type only if both of casts
1878 // have one use to avoid creating an extra instruction.
1879 if (!Cast0->hasOneUse() || !Cast1->hasOneUse())
1880 return nullptr;
1881
1882 // If the source types do not match, but the casts are matching extends,
1883 // we can still narrow the logic op.
1884 if (XNumBits < YNumBits) {
1885 X = Builder.CreateCast(CastOpcode, X, Y->getType());
1886 } else if (YNumBits < XNumBits) {
1887 Y = Builder.CreateCast(CastOpcode, Y, X->getType());
1888 }
1889 }
1890
1891 // Do the logic op in the intermediate width, then widen more.
1892 Value *NarrowLogic = Builder.CreateBinOp(LogicOpc, X, Y, I.getName());
1893 auto *Disjoint = dyn_cast<PossiblyDisjointInst>(&I);
1894 auto *NewDisjoint = dyn_cast<PossiblyDisjointInst>(NarrowLogic);
1895 if (Disjoint && NewDisjoint)
1896 NewDisjoint->setIsDisjoint(Disjoint->isDisjoint());
1897 return CastInst::Create(CastOpcode, NarrowLogic, DestTy);
1898 }
1899
1900 // If the src type of casts are different, give up for other cast opcodes.
1901 if (SrcTy != Cast1->getSrcTy())
1902 return nullptr;
1903
1904 Value *Cast0Src = Cast0->getOperand(0);
1905 Value *Cast1Src = Cast1->getOperand(0);
1906
1907 // fold logic(cast(A), cast(B)) -> cast(logic(A, B))
1908 if (shouldOptimizeCast(Cast0) && shouldOptimizeCast(Cast1)) {
1909 Value *NewOp = Builder.CreateBinOp(LogicOpc, Cast0Src, Cast1Src,
1910 I.getName());
1911 return CastInst::Create(CastOpcode, NewOp, DestTy);
1912 }
1913
1914 return nullptr;
1915}
1916
1918 InstCombiner::BuilderTy &Builder) {
1919 assert(I.getOpcode() == Instruction::And);
1920 Value *Op0 = I.getOperand(0);
1921 Value *Op1 = I.getOperand(1);
1922 Value *A, *B;
1923
1924 // Operand complexity canonicalization guarantees that the 'or' is Op0.
1925 // (A | B) & ~(A & B) --> A ^ B
1926 // (A | B) & ~(B & A) --> A ^ B
1927 if (match(&I, m_BinOp(m_Or(m_Value(A), m_Value(B)),
1929 return BinaryOperator::CreateXor(A, B);
1930
1931 // (A | ~B) & (~A | B) --> ~(A ^ B)
1932 // (A | ~B) & (B | ~A) --> ~(A ^ B)
1933 // (~B | A) & (~A | B) --> ~(A ^ B)
1934 // (~B | A) & (B | ~A) --> ~(A ^ B)
1935 if (Op0->hasOneUse() || Op1->hasOneUse())
1938 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1939
1940 return nullptr;
1941}
1942
1944 InstCombiner::BuilderTy &Builder) {
1945 assert(I.getOpcode() == Instruction::Or);
1946 Value *Op0 = I.getOperand(0);
1947 Value *Op1 = I.getOperand(1);
1948 Value *A, *B;
1949
1950 // Operand complexity canonicalization guarantees that the 'and' is Op0.
1951 // (A & B) | ~(A | B) --> ~(A ^ B)
1952 // (A & B) | ~(B | A) --> ~(A ^ B)
1953 if (Op0->hasOneUse() || Op1->hasOneUse())
1954 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
1956 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
1957
1958 // Operand complexity canonicalization guarantees that the 'xor' is Op0.
1959 // (A ^ B) | ~(A | B) --> ~(A & B)
1960 // (A ^ B) | ~(B | A) --> ~(A & B)
1961 if (Op0->hasOneUse() || Op1->hasOneUse())
1962 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
1964 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
1965
1966 // (A & ~B) | (~A & B) --> A ^ B
1967 // (A & ~B) | (B & ~A) --> A ^ B
1968 // (~B & A) | (~A & B) --> A ^ B
1969 // (~B & A) | (B & ~A) --> A ^ B
1970 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
1972 return BinaryOperator::CreateXor(A, B);
1973
1974 return nullptr;
1975}
1976
1977/// Return true if a constant shift amount is always less than the specified
1978/// bit-width. If not, the shift could create poison in the narrower type.
1979static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth) {
1980 APInt Threshold(C->getType()->getScalarSizeInBits(), BitWidth);
1981 return match(C, m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, Threshold));
1982}
1983
1984/// Try to use narrower ops (sink zext ops) for an 'and' with binop operand and
1985/// a common zext operand: and (binop (zext X), C), (zext X).
1986Instruction *InstCombinerImpl::narrowMaskedBinOp(BinaryOperator &And) {
1987 // This transform could also apply to {or, and, xor}, but there are better
1988 // folds for those cases, so we don't expect those patterns here. AShr is not
1989 // handled because it should always be transformed to LShr in this sequence.
1990 // The subtract transform is different because it has a constant on the left.
1991 // Add/mul commute the constant to RHS; sub with constant RHS becomes add.
1992 Value *Op0 = And.getOperand(0), *Op1 = And.getOperand(1);
1993 Constant *C;
1994 if (!match(Op0, m_OneUse(m_Add(m_Specific(Op1), m_Constant(C)))) &&
1995 !match(Op0, m_OneUse(m_Mul(m_Specific(Op1), m_Constant(C)))) &&
1996 !match(Op0, m_OneUse(m_LShr(m_Specific(Op1), m_Constant(C)))) &&
1997 !match(Op0, m_OneUse(m_Shl(m_Specific(Op1), m_Constant(C)))) &&
1998 !match(Op0, m_OneUse(m_Sub(m_Constant(C), m_Specific(Op1)))))
1999 return nullptr;
2000
2001 Value *X;
2002 if (!match(Op1, m_ZExt(m_Value(X))) || Op1->hasNUsesOrMore(3))
2003 return nullptr;
2004
2005 Type *Ty = And.getType();
2006 if (!isa<VectorType>(Ty) && !shouldChangeType(Ty, X->getType()))
2007 return nullptr;
2008
2009 // If we're narrowing a shift, the shift amount must be safe (less than the
2010 // width) in the narrower type. If the shift amount is greater, instsimplify
2011 // usually handles that case, but we can't guarantee/assert it.
2013 if (Opc == Instruction::LShr || Opc == Instruction::Shl)
2014 if (!canNarrowShiftAmt(C, X->getType()->getScalarSizeInBits()))
2015 return nullptr;
2016
2017 // and (sub C, (zext X)), (zext X) --> zext (and (sub C', X), X)
2018 // and (binop (zext X), C), (zext X) --> zext (and (binop X, C'), X)
2019 Value *NewC = ConstantExpr::getTrunc(C, X->getType());
2020 Value *NewBO = Opc == Instruction::Sub ? Builder.CreateBinOp(Opc, NewC, X)
2021 : Builder.CreateBinOp(Opc, X, NewC);
2022 return new ZExtInst(Builder.CreateAnd(NewBO, X), Ty);
2023}
2024
2025/// Try folding relatively complex patterns for both And and Or operations
2026/// with all And and Or swapped.
2028 InstCombiner::BuilderTy &Builder) {
2029 const Instruction::BinaryOps Opcode = I.getOpcode();
2030 assert(Opcode == Instruction::And || Opcode == Instruction::Or);
2031
2032 // Flip the logic operation.
2033 const Instruction::BinaryOps FlippedOpcode =
2034 (Opcode == Instruction::And) ? Instruction::Or : Instruction::And;
2035
2036 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2037 Value *A, *B, *C, *X, *Y, *Dummy;
2038
2039 // Match following expressions:
2040 // (~(A | B) & C)
2041 // (~(A & B) | C)
2042 // Captures X = ~(A | B) or ~(A & B)
2043 const auto matchNotOrAnd =
2044 [Opcode, FlippedOpcode](Value *Op, auto m_A, auto m_B, auto m_C,
2045 Value *&X, bool CountUses = false) -> bool {
2046 if (CountUses && !Op->hasOneUse())
2047 return false;
2048
2049 if (match(Op,
2050 m_c_BinOp(FlippedOpcode,
2051 m_Value(X, m_Not(m_c_BinOp(Opcode, m_A, m_B))), m_C)))
2052 return !CountUses || X->hasOneUse();
2053
2054 return false;
2055 };
2056
2057 // (~(A | B) & C) | ... --> ...
2058 // (~(A & B) | C) & ... --> ...
2059 // TODO: One use checks are conservative. We just need to check that a total
2060 // number of multiple used values does not exceed reduction
2061 // in operations.
2062 if (matchNotOrAnd(Op0, m_Value(A), m_Value(B), m_Value(C), X)) {
2063 // (~(A | B) & C) | (~(A | C) & B) --> (B ^ C) & ~A
2064 // (~(A & B) | C) & (~(A & C) | B) --> ~((B ^ C) & A)
2065 if (matchNotOrAnd(Op1, m_Specific(A), m_Specific(C), m_Specific(B), Dummy,
2066 true)) {
2067 Value *Xor = Builder.CreateXor(B, C);
2068 return (Opcode == Instruction::Or)
2069 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(A))
2070 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, A));
2071 }
2072
2073 // (~(A | B) & C) | (~(B | C) & A) --> (A ^ C) & ~B
2074 // (~(A & B) | C) & (~(B & C) | A) --> ~((A ^ C) & B)
2075 if (matchNotOrAnd(Op1, m_Specific(B), m_Specific(C), m_Specific(A), Dummy,
2076 true)) {
2077 Value *Xor = Builder.CreateXor(A, C);
2078 return (Opcode == Instruction::Or)
2079 ? BinaryOperator::CreateAnd(Xor, Builder.CreateNot(B))
2080 : BinaryOperator::CreateNot(Builder.CreateAnd(Xor, B));
2081 }
2082
2083 // (~(A | B) & C) | ~(A | C) --> ~((B & C) | A)
2084 // (~(A & B) | C) & ~(A & C) --> ~((B | C) & A)
2085 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2086 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2087 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2088 Opcode, Builder.CreateBinOp(FlippedOpcode, B, C), A));
2089
2090 // (~(A | B) & C) | ~(B | C) --> ~((A & C) | B)
2091 // (~(A & B) | C) & ~(B & C) --> ~((A | C) & B)
2092 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2093 m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)))))))
2094 return BinaryOperator::CreateNot(Builder.CreateBinOp(
2095 Opcode, Builder.CreateBinOp(FlippedOpcode, A, C), B));
2096
2097 // (~(A | B) & C) | ~(C | (A ^ B)) --> ~((A | B) & (C | (A ^ B)))
2098 // Note, the pattern with swapped and/or is not handled because the
2099 // result is more undefined than a source:
2100 // (~(A & B) | C) & ~(C & (A ^ B)) --> (A ^ B ^ C) | ~(A | C) is invalid.
2101 if (Opcode == Instruction::Or && Op0->hasOneUse() &&
2102 match(Op1,
2104 Y, m_c_BinOp(Opcode, m_Specific(C),
2105 m_c_Xor(m_Specific(A), m_Specific(B)))))))) {
2106 // X = ~(A | B)
2107 // Y = (C | (A ^ B)
2108 Value *Or = cast<BinaryOperator>(X)->getOperand(0);
2109 return BinaryOperator::CreateNot(Builder.CreateAnd(Or, Y));
2110 }
2111 }
2112
2113 // (~A & B & C) | ... --> ...
2114 // (~A | B | C) | ... --> ...
2115 // TODO: One use checks are conservative. We just need to check that a total
2116 // number of multiple used values does not exceed reduction
2117 // in operations.
2118 if (match(Op0,
2119 m_OneUse(m_c_BinOp(FlippedOpcode,
2120 m_BinOp(FlippedOpcode, m_Value(B), m_Value(C)),
2121 m_Value(X, m_Not(m_Value(A)))))) ||
2122 match(Op0, m_OneUse(m_c_BinOp(FlippedOpcode,
2123 m_c_BinOp(FlippedOpcode, m_Value(C),
2124 m_Value(X, m_Not(m_Value(A)))),
2125 m_Value(B))))) {
2126 // X = ~A
2127 // (~A & B & C) | ~(A | B | C) --> ~(A | (B ^ C))
2128 // (~A | B | C) & ~(A & B & C) --> (~A | (B ^ C))
2129 if (match(Op1, m_OneUse(m_Not(m_c_BinOp(
2130 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)),
2131 m_Specific(C))))) ||
2133 Opcode, m_c_BinOp(Opcode, m_Specific(B), m_Specific(C)),
2134 m_Specific(A))))) ||
2136 Opcode, m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)),
2137 m_Specific(B)))))) {
2138 Value *Xor = Builder.CreateXor(B, C);
2139 return (Opcode == Instruction::Or)
2140 ? BinaryOperator::CreateNot(Builder.CreateOr(Xor, A))
2141 : BinaryOperator::CreateOr(Xor, X);
2142 }
2143
2144 // (~A & B & C) | ~(A | B) --> (C | ~B) & ~A
2145 // (~A | B | C) & ~(A & B) --> (C & ~B) | ~A
2146 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2147 m_c_BinOp(Opcode, m_Specific(A), m_Specific(B)))))))
2149 FlippedOpcode, Builder.CreateBinOp(Opcode, C, Builder.CreateNot(B)),
2150 X);
2151
2152 // (~A & B & C) | ~(A | C) --> (B | ~C) & ~A
2153 // (~A | B | C) & ~(A & C) --> (B & ~C) | ~A
2154 if (match(Op1, m_OneUse(m_Not(m_OneUse(
2155 m_c_BinOp(Opcode, m_Specific(A), m_Specific(C)))))))
2157 FlippedOpcode, Builder.CreateBinOp(Opcode, B, Builder.CreateNot(C)),
2158 X);
2159 }
2160
2161 return nullptr;
2162}
2163
2164/// Try to reassociate a pair of binops so that values with one use only are
2165/// part of the same instruction. This may enable folds that are limited with
2166/// multi-use restrictions and makes it more likely to match other patterns that
2167/// are looking for a common operand.
2169 InstCombinerImpl::BuilderTy &Builder) {
2170 Instruction::BinaryOps Opcode = BO.getOpcode();
2171 Value *X, *Y, *Z;
2172 if (match(&BO,
2173 m_c_BinOp(Opcode, m_OneUse(m_BinOp(Opcode, m_Value(X), m_Value(Y))),
2174 m_OneUse(m_Value(Z))))) {
2175 if (!isa<Constant>(X) && !isa<Constant>(Y) && !isa<Constant>(Z)) {
2176 // (X op Y) op Z --> (Y op Z) op X
2177 if (!X->hasOneUse()) {
2178 Value *YZ = Builder.CreateBinOp(Opcode, Y, Z);
2179 return BinaryOperator::Create(Opcode, YZ, X);
2180 }
2181 // (X op Y) op Z --> (X op Z) op Y
2182 if (!Y->hasOneUse()) {
2183 Value *XZ = Builder.CreateBinOp(Opcode, X, Z);
2184 return BinaryOperator::Create(Opcode, XZ, Y);
2185 }
2186 }
2187 }
2188
2189 return nullptr;
2190}
2191
2192// Match
2193// (X + C2) | C
2194// (X + C2) ^ C
2195// (X + C2) & C
2196// and convert to do the bitwise logic first:
2197// (X | C) + C2
2198// (X ^ C) + C2
2199// (X & C) + C2
2200// iff bits affected by logic op are lower than last bit affected by math op
2202 InstCombiner::BuilderTy &Builder) {
2203 Type *Ty = I.getType();
2204 Instruction::BinaryOps OpC = I.getOpcode();
2205 Value *Op0 = I.getOperand(0);
2206 Value *Op1 = I.getOperand(1);
2207 Value *X;
2208 const APInt *C, *C2;
2209
2210 if (!(match(Op0, m_OneUse(m_Add(m_Value(X), m_APInt(C2)))) &&
2211 match(Op1, m_APInt(C))))
2212 return nullptr;
2213
2214 unsigned Width = Ty->getScalarSizeInBits();
2215 unsigned LastOneMath = Width - C2->countr_zero();
2216
2217 switch (OpC) {
2218 case Instruction::And:
2219 if (C->countl_one() < LastOneMath)
2220 return nullptr;
2221 break;
2222 case Instruction::Xor:
2223 case Instruction::Or:
2224 if (C->countl_zero() < LastOneMath)
2225 return nullptr;
2226 break;
2227 default:
2228 llvm_unreachable("Unexpected BinaryOp!");
2229 }
2230
2231 Value *NewBinOp = Builder.CreateBinOp(OpC, X, ConstantInt::get(Ty, *C));
2232 return BinaryOperator::CreateWithCopiedFlags(Instruction::Add, NewBinOp,
2233 ConstantInt::get(Ty, *C2), Op0);
2234}
2235
2236// binop(shift(ShiftedC1, ShAmt), shift(ShiftedC2, add(ShAmt, AddC))) ->
2237// shift(binop(ShiftedC1, shift(ShiftedC2, AddC)), ShAmt)
2238// where both shifts are the same and AddC is a valid shift amount.
2239Instruction *InstCombinerImpl::foldBinOpOfDisplacedShifts(BinaryOperator &I) {
2240 assert((I.isBitwiseLogicOp() || I.getOpcode() == Instruction::Add) &&
2241 "Unexpected opcode");
2242
2243 Value *ShAmt;
2244 Constant *ShiftedC1, *ShiftedC2, *AddC;
2245 Type *Ty = I.getType();
2246 unsigned BitWidth = Ty->getScalarSizeInBits();
2247 if (!match(&I, m_c_BinOp(m_Shift(m_ImmConstant(ShiftedC1), m_Value(ShAmt)),
2248 m_Shift(m_ImmConstant(ShiftedC2),
2249 m_AddLike(m_Deferred(ShAmt),
2250 m_ImmConstant(AddC))))))
2251 return nullptr;
2252
2253 // Make sure the add constant is a valid shift amount.
2254 if (!match(AddC,
2256 return nullptr;
2257
2258 // Avoid constant expressions.
2259 auto *Op0Inst = dyn_cast<Instruction>(I.getOperand(0));
2260 auto *Op1Inst = dyn_cast<Instruction>(I.getOperand(1));
2261 if (!Op0Inst || !Op1Inst)
2262 return nullptr;
2263
2264 // Both shifts must be the same.
2265 Instruction::BinaryOps ShiftOp =
2266 static_cast<Instruction::BinaryOps>(Op0Inst->getOpcode());
2267 if (ShiftOp != Op1Inst->getOpcode())
2268 return nullptr;
2269
2270 // For adds, only left shifts are supported.
2271 if (I.getOpcode() == Instruction::Add && ShiftOp != Instruction::Shl)
2272 return nullptr;
2273
2274 Value *NewC = Builder.CreateBinOp(
2275 I.getOpcode(), ShiftedC1, Builder.CreateBinOp(ShiftOp, ShiftedC2, AddC));
2276 return BinaryOperator::Create(ShiftOp, NewC, ShAmt);
2277}
2278
2279// Fold and/or/xor with two equal intrinsic IDs:
2280// bitwise(fshl (A, B, ShAmt), fshl(C, D, ShAmt))
2281// -> fshl(bitwise(A, C), bitwise(B, D), ShAmt)
2282// bitwise(fshr (A, B, ShAmt), fshr(C, D, ShAmt))
2283// -> fshr(bitwise(A, C), bitwise(B, D), ShAmt)
2284// bitwise(bswap(A), bswap(B)) -> bswap(bitwise(A, B))
2285// bitwise(bswap(A), C) -> bswap(bitwise(A, bswap(C)))
2286// bitwise(bitreverse(A), bitreverse(B)) -> bitreverse(bitwise(A, B))
2287// bitwise(bitreverse(A), C) -> bitreverse(bitwise(A, bitreverse(C)))
2288static Instruction *
2290 InstCombiner::BuilderTy &Builder) {
2291 assert(I.isBitwiseLogicOp() && "Should and/or/xor");
2292 if (!I.getOperand(0)->hasOneUse())
2293 return nullptr;
2294 IntrinsicInst *X = dyn_cast<IntrinsicInst>(I.getOperand(0));
2295 if (!X)
2296 return nullptr;
2297
2298 IntrinsicInst *Y = dyn_cast<IntrinsicInst>(I.getOperand(1));
2299 if (Y && (!Y->hasOneUse() || X->getIntrinsicID() != Y->getIntrinsicID()))
2300 return nullptr;
2301
2302 Intrinsic::ID IID = X->getIntrinsicID();
2303 const APInt *RHSC;
2304 // Try to match constant RHS.
2305 if (!Y && (!(IID == Intrinsic::bswap || IID == Intrinsic::bitreverse) ||
2306 !match(I.getOperand(1), m_APInt(RHSC))))
2307 return nullptr;
2308
2309 switch (IID) {
2310 case Intrinsic::fshl:
2311 case Intrinsic::fshr: {
2312 if (X->getOperand(2) != Y->getOperand(2))
2313 return nullptr;
2314 Value *NewOp0 =
2315 Builder.CreateBinOp(I.getOpcode(), X->getOperand(0), Y->getOperand(0));
2316 Value *NewOp1 =
2317 Builder.CreateBinOp(I.getOpcode(), X->getOperand(1), Y->getOperand(1));
2318 Function *F =
2319 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2320 return CallInst::Create(F, {NewOp0, NewOp1, X->getOperand(2)});
2321 }
2322 case Intrinsic::bswap:
2323 case Intrinsic::bitreverse: {
2324 Value *NewOp0 = Builder.CreateBinOp(
2325 I.getOpcode(), X->getOperand(0),
2326 Y ? Y->getOperand(0)
2327 : ConstantInt::get(I.getType(), IID == Intrinsic::bswap
2328 ? RHSC->byteSwap()
2329 : RHSC->reverseBits()));
2330 Function *F =
2331 Intrinsic::getOrInsertDeclaration(I.getModule(), IID, I.getType());
2332 return CallInst::Create(F, {NewOp0});
2333 }
2334 default:
2335 return nullptr;
2336 }
2337}
2338
2339// Try to simplify V by replacing occurrences of Op with RepOp, but only look
2340// through bitwise operations. In particular, for X | Y we try to replace Y with
2341// 0 inside X and for X & Y we try to replace Y with -1 inside X.
2342// Return the simplified result of X if successful, and nullptr otherwise.
2343// If SimplifyOnly is true, no new instructions will be created.
2345 bool SimplifyOnly,
2346 InstCombinerImpl &IC,
2347 unsigned Depth = 0) {
2348 if (Op == RepOp)
2349 return nullptr;
2350
2351 if (V == Op)
2352 return RepOp;
2353
2354 auto *I = dyn_cast<BinaryOperator>(V);
2355 if (!I || !I->isBitwiseLogicOp() || Depth >= 3)
2356 return nullptr;
2357
2358 if (!I->hasOneUse())
2359 SimplifyOnly = true;
2360
2361 Value *NewOp0 = simplifyAndOrWithOpReplaced(I->getOperand(0), Op, RepOp,
2362 SimplifyOnly, IC, Depth + 1);
2363 Value *NewOp1 = simplifyAndOrWithOpReplaced(I->getOperand(1), Op, RepOp,
2364 SimplifyOnly, IC, Depth + 1);
2365 if (!NewOp0 && !NewOp1)
2366 return nullptr;
2367
2368 if (!NewOp0)
2369 NewOp0 = I->getOperand(0);
2370 if (!NewOp1)
2371 NewOp1 = I->getOperand(1);
2372
2373 if (Value *Res = simplifyBinOp(I->getOpcode(), NewOp0, NewOp1,
2375 return Res;
2376
2377 if (SimplifyOnly)
2378 return nullptr;
2379 return IC.Builder.CreateBinOp(I->getOpcode(), NewOp0, NewOp1);
2380}
2381
2382/// Reassociate and/or expressions to see if we can fold the inner and/or ops.
2383/// TODO: Make this recursive; it's a little tricky because an arbitrary
2384/// number of and/or instructions might have to be created.
2385Value *InstCombinerImpl::reassociateBooleanAndOr(Value *LHS, Value *X, Value *Y,
2386 Instruction &I, bool IsAnd,
2387 bool RHSIsLogical) {
2388 Instruction::BinaryOps Opcode = IsAnd ? Instruction::And : Instruction::Or;
2389 Value *Folded = nullptr;
2390 // LHS bop (X lop Y) --> (LHS bop X) lop Y
2391 // LHS bop (X bop Y) --> (LHS bop X) bop Y
2392 if (Value *Res = foldBooleanAndOr(LHS, X, I, IsAnd, /*IsLogical=*/false))
2393 Folded = RHSIsLogical ? Builder.CreateLogicalOp(Opcode, Res, Y)
2394 : Builder.CreateBinOp(Opcode, Res, Y);
2395 // LHS bop (X bop Y) --> X bop (LHS bop Y)
2396 // LHS bop (X lop Y) --> X lop (LHS bop Y)
2397 else if (Value *Res = foldBooleanAndOr(LHS, Y, I, IsAnd, /*IsLogical=*/false))
2398 Folded = RHSIsLogical ? Builder.CreateLogicalOp(Opcode, X, Res)
2399 : Builder.CreateBinOp(Opcode, X, Res);
2400 if (SelectInst *SI = dyn_cast_or_null<SelectInst>(Folded);
2401 SI != nullptr && !ProfcheckDisableMetadataFixes)
2402 // If the bop I was originally a lop, we could recover branch weight
2403 // information using that lop's weights. However, InstCombine usually
2404 // replaces the lop with a bop by the time we get here, deleting the branch
2405 // weight information. Therefore, we can only assume unknown branch weights.
2406 // TODO: see if it's possible to recover branch weight information from the
2407 // original lop (https://github.com/llvm/llvm-project/issues/183864).
2409 I.getFunction());
2410 return Folded;
2411}
2412
2413// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
2414// here. We should standardize that construct where it is needed or choose some
2415// other way to ensure that commutated variants of patterns are not missed.
2417 Type *Ty = I.getType();
2418
2419 if (Value *V = simplifyAndInst(I.getOperand(0), I.getOperand(1),
2420 SQ.getWithInstruction(&I)))
2421 return replaceInstUsesWith(I, V);
2422
2424 return &I;
2425
2427 return X;
2428
2430 return Phi;
2431
2432 // See if we can simplify any instructions used by the instruction whose sole
2433 // purpose is to compute bits we don't care about.
2435 return &I;
2436
2437 // Do this before using distributive laws to catch simple and/or/not patterns.
2439 return Xor;
2440
2442 return X;
2443
2444 // (A|B)&(A|C) -> A|(B&C) etc
2446 return replaceInstUsesWith(I, V);
2447
2449 return R;
2450
2451 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
2452
2453 Value *X, *Y;
2454 const APInt *C;
2455 if ((match(Op0, m_OneUse(m_LogicalShift(m_One(), m_Value(X)))) ||
2456 (match(Op0, m_OneUse(m_Shl(m_APInt(C), m_Value(X)))) && (*C)[0])) &&
2457 match(Op1, m_One())) {
2458 // (1 >> X) & 1 --> zext(X == 0)
2459 // (C << X) & 1 --> zext(X == 0), when C is odd
2460 Value *IsZero = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, 0));
2461 return new ZExtInst(IsZero, Ty);
2462 }
2463
2464 // (-(X & 1)) & Y --> (X & 1) == 0 ? 0 : Y
2465 Value *Neg;
2466 if (match(&I,
2468 m_Value(Y)))) {
2469 Value *Cmp = Builder.CreateIsNull(Neg);
2470 return createSelectInstWithUnknownProfile(Cmp,
2472 }
2473
2474 // Canonicalize:
2475 // (X +/- Y) & Y --> ~X & Y when Y is a power of 2.
2478 m_Sub(m_Value(X), m_Deferred(Y)))))) &&
2479 isKnownToBeAPowerOfTwo(Y, /*OrZero*/ true, &I))
2480 return BinaryOperator::CreateAnd(Builder.CreateNot(X), Y);
2481
2482 if (match(Op1, m_APInt(C))) {
2483 const APInt *XorC;
2484 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_APInt(XorC))))) {
2485 // (X ^ C1) & C2 --> (X & C2) ^ (C1&C2)
2486 Constant *NewC = ConstantInt::get(Ty, *C & *XorC);
2487 Value *And = Builder.CreateAnd(X, Op1);
2488 And->takeName(Op0);
2489 return BinaryOperator::CreateXor(And, NewC);
2490 }
2491
2492 const APInt *OrC;
2493 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_APInt(OrC))))) {
2494 // (X | C1) & C2 --> (X & C2^(C1&C2)) | (C1&C2)
2495 // NOTE: This reduces the number of bits set in the & mask, which
2496 // can expose opportunities for store narrowing for scalars.
2497 // NOTE: SimplifyDemandedBits should have already removed bits from C1
2498 // that aren't set in C2. Meaning we can replace (C1&C2) with C1 in
2499 // above, but this feels safer.
2500 APInt Together = *C & *OrC;
2501 Value *And = Builder.CreateAnd(X, ConstantInt::get(Ty, Together ^ *C));
2502 And->takeName(Op0);
2503 return BinaryOperator::CreateOr(And, ConstantInt::get(Ty, Together));
2504 }
2505
2506 unsigned Width = Ty->getScalarSizeInBits();
2507 const APInt *ShiftC;
2508 if (match(Op0, m_OneUse(m_SExt(m_AShr(m_Value(X), m_APInt(ShiftC))))) &&
2509 ShiftC->ult(Width)) {
2510 if (*C == APInt::getLowBitsSet(Width, Width - ShiftC->getZExtValue())) {
2511 // We are clearing high bits that were potentially set by sext+ashr:
2512 // and (sext (ashr X, ShiftC)), C --> lshr (sext X), ShiftC
2513 Value *Sext = Builder.CreateSExt(X, Ty);
2514 Constant *ShAmtC = ConstantInt::get(Ty, ShiftC->zext(Width));
2515 return BinaryOperator::CreateLShr(Sext, ShAmtC);
2516 }
2517 }
2518
2519 // If this 'and' clears the sign-bits added by ashr, replace with lshr:
2520 // and (ashr X, ShiftC), C --> lshr X, ShiftC
2521 if (match(Op0, m_AShr(m_Value(X), m_APInt(ShiftC))) && ShiftC->ult(Width) &&
2522 C->isMask(Width - ShiftC->getZExtValue()))
2523 return BinaryOperator::CreateLShr(X, ConstantInt::get(Ty, *ShiftC));
2524
2525 const APInt *AddC;
2526 if (match(Op0, m_Add(m_Value(X), m_APInt(AddC)))) {
2527 // If we are masking the result of the add down to exactly one bit and
2528 // the constant we are adding has no bits set below that bit, then the
2529 // add is flipping a single bit. Example:
2530 // (X + 4) & 4 --> (X & 4) ^ 4
2531 if (Op0->hasOneUse() && C->isPowerOf2() && (*AddC & (*C - 1)) == 0) {
2532 assert((*C & *AddC) != 0 && "Expected common bit");
2533 Value *NewAnd = Builder.CreateAnd(X, Op1);
2534 return BinaryOperator::CreateXor(NewAnd, Op1);
2535 }
2536 }
2537
2538 // ((C1 OP zext(X)) & C2) -> zext((C1 OP X) & C2) if C2 fits in the
2539 // bitwidth of X and OP behaves well when given trunc(C1) and X.
2540 auto isNarrowableBinOpcode = [](BinaryOperator *B) {
2541 switch (B->getOpcode()) {
2542 case Instruction::Xor:
2543 case Instruction::Or:
2544 case Instruction::Mul:
2545 case Instruction::Add:
2546 case Instruction::Sub:
2547 return true;
2548 default:
2549 return false;
2550 }
2551 };
2552 BinaryOperator *BO;
2553 if (match(Op0, m_OneUse(m_BinOp(BO))) && isNarrowableBinOpcode(BO)) {
2554 Instruction::BinaryOps BOpcode = BO->getOpcode();
2555 Value *X;
2556 const APInt *C1;
2557 // TODO: The one-use restrictions could be relaxed a little if the AND
2558 // is going to be removed.
2559 // Try to narrow the 'and' and a binop with constant operand:
2560 // and (bo (zext X), C1), C --> zext (and (bo X, TruncC1), TruncC)
2561 if (match(BO, m_c_BinOp(m_OneUse(m_ZExt(m_Value(X))), m_APInt(C1))) &&
2562 C->isIntN(X->getType()->getScalarSizeInBits())) {
2563 unsigned XWidth = X->getType()->getScalarSizeInBits();
2564 Constant *TruncC1 = ConstantInt::get(X->getType(), C1->trunc(XWidth));
2565 Value *BinOp = isa<ZExtInst>(BO->getOperand(0))
2566 ? Builder.CreateBinOp(BOpcode, X, TruncC1)
2567 : Builder.CreateBinOp(BOpcode, TruncC1, X);
2568 Constant *TruncC = ConstantInt::get(X->getType(), C->trunc(XWidth));
2569 Value *And = Builder.CreateAnd(BinOp, TruncC);
2570 return new ZExtInst(And, Ty);
2571 }
2572
2573 // Similar to above: if the mask matches the zext input width, then the
2574 // 'and' can be eliminated, so we can truncate the other variable op:
2575 // and (bo (zext X), Y), C --> zext (bo X, (trunc Y))
2576 if (isa<Instruction>(BO->getOperand(0)) &&
2577 match(BO->getOperand(0), m_OneUse(m_ZExt(m_Value(X)))) &&
2578 C->isMask(X->getType()->getScalarSizeInBits())) {
2579 Y = BO->getOperand(1);
2580 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2581 Value *NewBO =
2582 Builder.CreateBinOp(BOpcode, X, TrY, BO->getName() + ".narrow");
2583 return new ZExtInst(NewBO, Ty);
2584 }
2585 // and (bo Y, (zext X)), C --> zext (bo (trunc Y), X)
2586 if (isa<Instruction>(BO->getOperand(1)) &&
2587 match(BO->getOperand(1), m_OneUse(m_ZExt(m_Value(X)))) &&
2588 C->isMask(X->getType()->getScalarSizeInBits())) {
2589 Y = BO->getOperand(0);
2590 Value *TrY = Builder.CreateTrunc(Y, X->getType(), Y->getName() + ".tr");
2591 Value *NewBO =
2592 Builder.CreateBinOp(BOpcode, TrY, X, BO->getName() + ".narrow");
2593 return new ZExtInst(NewBO, Ty);
2594 }
2595 }
2596
2597 // This is intentionally placed after the narrowing transforms for
2598 // efficiency (transform directly to the narrow logic op if possible).
2599 // If the mask is only needed on one incoming arm, push the 'and' op up.
2600 if (match(Op0, m_OneUse(m_Xor(m_Value(X), m_Value(Y)))) ||
2601 match(Op0, m_OneUse(m_Or(m_Value(X), m_Value(Y))))) {
2602 APInt NotAndMask(~(*C));
2603 BinaryOperator::BinaryOps BinOp = cast<BinaryOperator>(Op0)->getOpcode();
2604 if (MaskedValueIsZero(X, NotAndMask, &I)) {
2605 // Not masking anything out for the LHS, move mask to RHS.
2606 // and ({x}or X, Y), C --> {x}or X, (and Y, C)
2607 Value *NewRHS = Builder.CreateAnd(Y, Op1, Y->getName() + ".masked");
2608 return BinaryOperator::Create(BinOp, X, NewRHS);
2609 }
2610 if (!isa<Constant>(Y) && MaskedValueIsZero(Y, NotAndMask, &I)) {
2611 // Not masking anything out for the RHS, move mask to LHS.
2612 // and ({x}or X, Y), C --> {x}or (and X, C), Y
2613 Value *NewLHS = Builder.CreateAnd(X, Op1, X->getName() + ".masked");
2614 return BinaryOperator::Create(BinOp, NewLHS, Y);
2615 }
2616 }
2617
2618 // When the mask is a power-of-2 constant and op0 is a shifted-power-of-2
2619 // constant, test if the shift amount equals the offset bit index:
2620 // (ShiftC << X) & C --> X == (log2(C) - log2(ShiftC)) ? C : 0
2621 // (ShiftC >> X) & C --> X == (log2(ShiftC) - log2(C)) ? C : 0
2622 if (C->isPowerOf2() &&
2623 match(Op0, m_OneUse(m_LogicalShift(m_Power2(ShiftC), m_Value(X))))) {
2624 int Log2ShiftC = ShiftC->exactLogBase2();
2625 int Log2C = C->exactLogBase2();
2626 bool IsShiftLeft =
2627 cast<BinaryOperator>(Op0)->getOpcode() == Instruction::Shl;
2628 int BitNum = IsShiftLeft ? Log2C - Log2ShiftC : Log2ShiftC - Log2C;
2629 assert(BitNum >= 0 && "Expected demanded bits to handle impossible mask");
2630 Value *Cmp = Builder.CreateICmpEQ(X, ConstantInt::get(Ty, BitNum));
2631 return createSelectInstWithUnknownProfile(Cmp, ConstantInt::get(Ty, *C),
2633 }
2634
2635 Constant *C1, *C2;
2636 const APInt *C3 = C;
2637 Value *X;
2638 if (C3->isPowerOf2()) {
2639 Constant *Log2C3 = ConstantInt::get(Ty, C3->countr_zero());
2641 m_ImmConstant(C2)))) &&
2642 match(C1, m_Power2())) {
2644 Constant *LshrC = ConstantExpr::getAdd(C2, Log2C3);
2645 KnownBits KnownLShrc = computeKnownBits(LshrC, nullptr);
2646 if (KnownLShrc.getMaxValue().ult(Width)) {
2647 // iff C1,C3 is pow2 and C2 + cttz(C3) < BitWidth:
2648 // ((C1 << X) >> C2) & C3 -> X == (cttz(C3)+C2-cttz(C1)) ? C3 : 0
2649 Constant *CmpC = ConstantExpr::getSub(LshrC, Log2C1);
2650 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2651 return createSelectInstWithUnknownProfile(
2652 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2653 }
2654 }
2655
2657 m_ImmConstant(C2)))) &&
2658 match(C1, m_Power2())) {
2660 Constant *Cmp =
2662 if (Cmp && Cmp->isNullValue()) {
2663 // iff C1,C3 is pow2 and Log2(C3) >= C2:
2664 // ((C1 >> X) << C2) & C3 -> X == (cttz(C1)+C2-cttz(C3)) ? C3 : 0
2665 Constant *ShlC = ConstantExpr::getAdd(C2, Log2C1);
2666 Constant *CmpC = ConstantExpr::getSub(ShlC, Log2C3);
2667 Value *Cmp = Builder.CreateICmpEQ(X, CmpC);
2668 return createSelectInstWithUnknownProfile(
2669 Cmp, ConstantInt::get(Ty, *C3), ConstantInt::getNullValue(Ty));
2670 }
2671 }
2672 }
2673 }
2674
2675 // If we are clearing the sign bit of a floating-point value, convert this to
2676 // fabs, then cast back to integer.
2677 //
2678 // This is a generous interpretation for noimplicitfloat, this is not a true
2679 // floating-point operation.
2680 //
2681 // Assumes any IEEE-represented type has the sign bit in the high bit.
2682 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
2683 Value *CastOp;
2684 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
2685 match(Op1, m_MaxSignedValue()) &&
2686 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
2687 Attribute::NoImplicitFloat)) {
2688 Type *EltTy = CastOp->getType()->getScalarType();
2689 if (EltTy->isFloatingPointTy() &&
2691 Value *FAbs = Builder.CreateFAbs(CastOp);
2692 return new BitCastInst(FAbs, I.getType());
2693 }
2694 }
2695
2696 // and(shl(zext(X), Y), SignMask) -> and(sext(X), SignMask)
2697 // where Y is a valid shift amount.
2699 m_SignMask())) &&
2702 APInt(Ty->getScalarSizeInBits(),
2703 Ty->getScalarSizeInBits() -
2704 X->getType()->getScalarSizeInBits())))) {
2705 auto *SExt = Builder.CreateSExt(X, Ty, X->getName() + ".signext");
2706 return BinaryOperator::CreateAnd(SExt, Op1);
2707 }
2708
2709 if (Instruction *Z = narrowMaskedBinOp(I))
2710 return Z;
2711
2712 if (I.getType()->isIntOrIntVectorTy(1)) {
2713 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
2714 if (auto *R =
2715 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ true))
2716 return R;
2717 }
2718 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
2719 if (auto *R =
2720 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ true))
2721 return R;
2722 }
2723 }
2724
2725 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
2726 return FoldedLogic;
2727
2728 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
2729 return DeMorgan;
2730
2731 {
2732 Value *A, *B, *C;
2733 // A & ~(A ^ B) --> A & B
2734 if (match(Op1, m_Not(m_c_Xor(m_Specific(Op0), m_Value(B)))))
2735 return BinaryOperator::CreateAnd(Op0, B);
2736 // ~(A ^ B) & A --> A & B
2737 if (match(Op0, m_Not(m_c_Xor(m_Specific(Op1), m_Value(B)))))
2738 return BinaryOperator::CreateAnd(Op1, B);
2739
2740 // (A ^ B) & ((B ^ C) ^ A) -> (A ^ B) & ~C
2741 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
2742 match(Op1, m_Xor(m_Xor(m_Specific(B), m_Value(C)), m_Specific(A)))) {
2743 Value *NotC = Op1->hasOneUse()
2744 ? Builder.CreateNot(C)
2745 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2746 if (NotC != nullptr)
2747 return BinaryOperator::CreateAnd(Op0, NotC);
2748 }
2749
2750 // ((A ^ C) ^ B) & (B ^ A) -> (B ^ A) & ~C
2751 if (match(Op0, m_Xor(m_Xor(m_Value(A), m_Value(C)), m_Value(B))) &&
2752 match(Op1, m_Xor(m_Specific(B), m_Specific(A)))) {
2753 Value *NotC = Op0->hasOneUse()
2754 ? Builder.CreateNot(C)
2755 : getFreelyInverted(C, C->hasOneUse(), &Builder);
2756 if (NotC != nullptr)
2757 return BinaryOperator::CreateAnd(Op1, NotC);
2758 }
2759
2760 // (A | B) & (~A ^ B) -> A & B
2761 // (A | B) & (B ^ ~A) -> A & B
2762 // (B | A) & (~A ^ B) -> A & B
2763 // (B | A) & (B ^ ~A) -> A & B
2764 if (match(Op1, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2765 match(Op0, m_c_Or(m_Specific(A), m_Specific(B))))
2766 return BinaryOperator::CreateAnd(A, B);
2767
2768 // (~A ^ B) & (A | B) -> A & B
2769 // (~A ^ B) & (B | A) -> A & B
2770 // (B ^ ~A) & (A | B) -> A & B
2771 // (B ^ ~A) & (B | A) -> A & B
2772 if (match(Op0, m_c_Xor(m_Not(m_Value(A)), m_Value(B))) &&
2773 match(Op1, m_c_Or(m_Specific(A), m_Specific(B))))
2774 return BinaryOperator::CreateAnd(A, B);
2775
2776 // (~A | B) & (A ^ B) -> ~A & B
2777 // (~A | B) & (B ^ A) -> ~A & B
2778 // (B | ~A) & (A ^ B) -> ~A & B
2779 // (B | ~A) & (B ^ A) -> ~A & B
2780 if (match(Op0, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2782 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2783
2784 // (A ^ B) & (~A | B) -> ~A & B
2785 // (B ^ A) & (~A | B) -> ~A & B
2786 // (A ^ B) & (B | ~A) -> ~A & B
2787 // (B ^ A) & (B | ~A) -> ~A & B
2788 if (match(Op1, m_c_Or(m_Not(m_Value(A)), m_Value(B))) &&
2790 return BinaryOperator::CreateAnd(Builder.CreateNot(A), B);
2791 }
2792
2793 if (Value *Res =
2794 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/true, /*IsLogical=*/false))
2795 return replaceInstUsesWith(I, Res);
2796
2797 if (match(Op1, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2798 bool IsLogical = isa<SelectInst>(Op1);
2799 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/true,
2800 /*RHSIsLogical=*/IsLogical))
2801 return replaceInstUsesWith(I, V);
2802 }
2803 if (match(Op0, m_OneUse(m_LogicalAnd(m_Value(X), m_Value(Y))))) {
2804 bool IsLogical = isa<SelectInst>(Op0);
2805 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/true,
2806 /*RHSIsLogical=*/IsLogical))
2807 return replaceInstUsesWith(I, V);
2808 }
2809
2810 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
2811 return FoldedFCmps;
2812
2813 if (Instruction *CastedAnd = foldCastedBitwiseLogic(I))
2814 return CastedAnd;
2815
2816 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
2817 return Sel;
2818
2819 // and(sext(A), B) / and(B, sext(A)) --> A ? B : 0, where A is i1 or <N x i1>.
2820 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
2821 // with binop identity constant. But creating a select with non-constant
2822 // arm may not be reversible due to poison semantics. Is that a good
2823 // canonicalization?
2824 Value *A, *B;
2825 if (match(&I, m_c_And(m_SExt(m_Value(A)), m_Value(B))) &&
2826 A->getType()->isIntOrIntVectorTy(1))
2827 return createSelectInstWithUnknownProfile(A, B, Constant::getNullValue(Ty));
2828
2829 // Similarly, a 'not' of the bool translates to a swap of the select arms:
2830 // ~sext(A) & B / B & ~sext(A) --> A ? 0 : B
2831 if (match(&I, m_c_And(m_Not(m_SExt(m_Value(A))), m_Value(B))) &&
2832 A->getType()->isIntOrIntVectorTy(1))
2833 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty), B);
2834
2835 // and(zext(A), B) -> A ? (B & 1) : 0
2836 if (match(&I, m_c_And(m_OneUse(m_ZExt(m_Value(A))), m_Value(B))) &&
2837 A->getType()->isIntOrIntVectorTy(1))
2838 return createSelectInstWithUnknownProfile(
2839 A, Builder.CreateAnd(B, ConstantInt::get(Ty, 1)),
2841
2842 // (-1 + A) & B --> A ? 0 : B where A is 0/1.
2844 m_Value(B)))) {
2845 if (A->getType()->isIntOrIntVectorTy(1))
2846 return createSelectInstWithUnknownProfile(A, Constant::getNullValue(Ty),
2847 B);
2848 if (computeKnownBits(A, &I).countMaxActiveBits() <= 1) {
2849 return createSelectInstWithUnknownProfile(
2850 Builder.CreateICmpEQ(A, Constant::getNullValue(A->getType())), B,
2852 }
2853 }
2854
2855 // (iN X s>> (N-1)) & Y --> (X s< 0) ? Y : 0 -- with optional sext
2858 m_Value(Y))) &&
2859 *C == X->getType()->getScalarSizeInBits() - 1) {
2860 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2861 return createSelectInstWithUnknownProfile(IsNeg, Y,
2863 }
2864 // If there's a 'not' of the shifted value, swap the select operands:
2865 // ~(iN X s>> (N-1)) & Y --> (X s< 0) ? 0 : Y -- with optional sext
2868 m_Value(Y))) &&
2869 *C == X->getType()->getScalarSizeInBits() - 1) {
2870 Value *IsNeg = Builder.CreateIsNeg(X, "isneg");
2871 return createSelectInstWithUnknownProfile(IsNeg,
2873 }
2874
2875 // (~x) & y --> ~(x | (~y)) iff that gets rid of inversions
2877 return &I;
2878
2879 // An and recurrence w/loop invariant step is equivelent to (and start, step)
2880 PHINode *PN = nullptr;
2881 Value *Start = nullptr, *Step = nullptr;
2882 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
2883 return replaceInstUsesWith(I, Builder.CreateAnd(Start, Step));
2884
2886 return R;
2887
2888 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
2889 return Canonicalized;
2890
2891 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
2892 return Folded;
2893
2894 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
2895 return Res;
2896
2898 return Res;
2899
2900 if (Value *V =
2902 /*SimplifyOnly*/ false, *this))
2903 return BinaryOperator::CreateAnd(V, Op1);
2904 if (Value *V =
2906 /*SimplifyOnly*/ false, *this))
2907 return BinaryOperator::CreateAnd(Op0, V);
2908
2909 return nullptr;
2910}
2911
2913 bool MatchBSwaps,
2914 bool MatchBitReversals) {
2916 if (!recognizeBSwapOrBitReverseIdiom(&I, MatchBSwaps, MatchBitReversals,
2917 Insts))
2918 return nullptr;
2919 Instruction *LastInst = Insts.pop_back_val();
2920 LastInst->removeFromParent();
2921
2922 for (auto *Inst : Insts) {
2923 Inst->setDebugLoc(I.getDebugLoc());
2924 Worklist.push(Inst);
2925 }
2926 return LastInst;
2927}
2928
2929std::optional<std::pair<Intrinsic::ID, SmallVector<Value *, 3>>>
2931 // TODO: Can we reduce the code duplication between this and the related
2932 // rotate matching code under visitSelect and visitTrunc?
2933 assert(Or.getOpcode() == BinaryOperator::Or && "Expecting or instruction");
2934
2935 unsigned Width = Or.getType()->getScalarSizeInBits();
2936
2937 Instruction *Or0, *Or1;
2938 if (!match(Or.getOperand(0), m_Instruction(Or0)) ||
2939 !match(Or.getOperand(1), m_Instruction(Or1)))
2940 return std::nullopt;
2941
2942 bool IsFshl = true; // Sub on LSHR.
2943 SmallVector<Value *, 3> FShiftArgs;
2944
2945 // First, find an or'd pair of opposite shifts:
2946 // or (lshr ShVal0, ShAmt0), (shl ShVal1, ShAmt1)
2947 if (isa<BinaryOperator>(Or0) && isa<BinaryOperator>(Or1)) {
2948 Value *ShVal0, *ShVal1, *ShAmt0, *ShAmt1;
2949 if (!match(Or0,
2950 m_OneUse(m_LogicalShift(m_Value(ShVal0), m_Value(ShAmt0)))) ||
2951 !match(Or1,
2952 m_OneUse(m_LogicalShift(m_Value(ShVal1), m_Value(ShAmt1)))) ||
2953 Or0->getOpcode() == Or1->getOpcode())
2954 return std::nullopt;
2955
2956 // Canonicalize to or(shl(ShVal0, ShAmt0), lshr(ShVal1, ShAmt1)).
2957 if (Or0->getOpcode() == BinaryOperator::LShr) {
2958 std::swap(Or0, Or1);
2959 std::swap(ShVal0, ShVal1);
2960 std::swap(ShAmt0, ShAmt1);
2961 }
2962 assert(Or0->getOpcode() == BinaryOperator::Shl &&
2963 Or1->getOpcode() == BinaryOperator::LShr &&
2964 "Illegal or(shift,shift) pair");
2965
2966 // Match the shift amount operands for a funnel shift pattern. This always
2967 // matches a subtraction on the R operand.
2968 auto matchShiftAmount = [&](Value *L, Value *R, unsigned Width) -> Value * {
2969 // Check for constant shift amounts that sum to the bitwidth.
2970 const APInt *LI, *RI;
2971 if (match(L, m_APIntAllowPoison(LI)) && match(R, m_APIntAllowPoison(RI)))
2972 if (LI->ult(Width) && RI->ult(Width) && (*LI + *RI) == Width)
2973 return ConstantInt::get(L->getType(), *LI);
2974
2975 Constant *LC, *RC;
2976 if (match(L, m_Constant(LC)) && match(R, m_Constant(RC)) &&
2977 match(L,
2978 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2979 match(R,
2980 m_SpecificInt_ICMP(ICmpInst::ICMP_ULT, APInt(Width, Width))) &&
2982 return ConstantExpr::mergeUndefsWith(LC, RC);
2983
2984 // (shl ShVal, X) | (lshr ShVal, (Width - x)) iff X < Width.
2985 // We limit this to X < Width in case the backend re-expands the
2986 // intrinsic, and has to reintroduce a shift modulo operation (InstCombine
2987 // might remove it after this fold). This still doesn't guarantee that the
2988 // final codegen will match this original pattern.
2989 if (match(R, m_OneUse(m_Sub(m_SpecificInt(Width), m_Specific(L))))) {
2990 KnownBits KnownL = computeKnownBits(L, &Or);
2991 return KnownL.getMaxValue().ult(Width) ? L : nullptr;
2992 }
2993
2994 // For non-constant cases, the following patterns currently only work for
2995 // rotation patterns.
2996 // TODO: Add general funnel-shift compatible patterns.
2997 if (ShVal0 != ShVal1)
2998 return nullptr;
2999
3000 // For non-constant cases we don't support non-pow2 shift masks.
3001 // TODO: Is it worth matching urem as well?
3002 if (!isPowerOf2_32(Width))
3003 return nullptr;
3004
3005 // The shift amount may be masked with negation:
3006 // (shl ShVal, (X & (Width - 1))) | (lshr ShVal, ((-X) & (Width - 1)))
3007 Value *X;
3008 unsigned Mask = Width - 1;
3009 if (match(L, m_And(m_Value(X), m_SpecificInt(Mask))) &&
3010 match(R, m_And(m_Neg(m_Specific(X)), m_SpecificInt(Mask))))
3011 return X;
3012
3013 // (shl ShVal, X) | (lshr ShVal, ((-X) & (Width - 1)))
3014 if (match(R, m_And(m_Neg(m_Specific(L)), m_SpecificInt(Mask))))
3015 return L;
3016
3017 // Similar to above, but the shift amount may be extended after masking,
3018 // so return the extended value as the parameter for the intrinsic.
3019 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3020 match(R,
3022 m_SpecificInt(Mask))))
3023 return L;
3024
3025 if (match(L, m_ZExt(m_And(m_Value(X), m_SpecificInt(Mask)))) &&
3027 return L;
3028
3029 return nullptr;
3030 };
3031
3032 Value *ShAmt = matchShiftAmount(ShAmt0, ShAmt1, Width);
3033 if (!ShAmt) {
3034 ShAmt = matchShiftAmount(ShAmt1, ShAmt0, Width);
3035 IsFshl = false; // Sub on SHL.
3036 }
3037 if (!ShAmt)
3038 return std::nullopt;
3039
3040 FShiftArgs = {ShVal0, ShVal1, ShAmt};
3041 } else if (isa<ZExtInst>(Or0) || isa<ZExtInst>(Or1)) {
3042 // If there are two 'or' instructions concat variables in opposite order:
3043 //
3044 // Slot1 and Slot2 are all zero bits.
3045 // | Slot1 | Low | Slot2 | High |
3046 // LowHigh = or (shl (zext Low), ZextLowShlAmt), (zext High)
3047 // | Slot2 | High | Slot1 | Low |
3048 // HighLow = or (shl (zext High), ZextHighShlAmt), (zext Low)
3049 //
3050 // the latter 'or' can be safely convert to
3051 // -> HighLow = fshl LowHigh, LowHigh, ZextHighShlAmt
3052 // if ZextLowShlAmt + ZextHighShlAmt == Width.
3053 if (!isa<ZExtInst>(Or1))
3054 std::swap(Or0, Or1);
3055
3056 Value *High, *ZextHigh, *Low;
3057 const APInt *ZextHighShlAmt;
3058 if (!match(Or0,
3059 m_OneUse(m_Shl(m_Value(ZextHigh), m_APInt(ZextHighShlAmt)))))
3060 return std::nullopt;
3061
3062 if (!match(Or1, m_ZExt(m_Value(Low))) ||
3063 !match(ZextHigh, m_ZExt(m_Value(High))))
3064 return std::nullopt;
3065
3066 unsigned HighSize = High->getType()->getScalarSizeInBits();
3067 unsigned LowSize = Low->getType()->getScalarSizeInBits();
3068 // Make sure High does not overlap with Low and most significant bits of
3069 // High aren't shifted out.
3070 if (ZextHighShlAmt->ult(LowSize) || ZextHighShlAmt->ugt(Width - HighSize))
3071 return std::nullopt;
3072
3073 for (User *U : ZextHigh->users()) {
3074 Value *X, *Y;
3075 if (!match(U, m_Or(m_Value(X), m_Value(Y))))
3076 continue;
3077
3078 if (!isa<ZExtInst>(Y))
3079 std::swap(X, Y);
3080
3081 const APInt *ZextLowShlAmt;
3082 if (!match(X, m_Shl(m_Specific(Or1), m_APInt(ZextLowShlAmt))) ||
3083 !match(Y, m_Specific(ZextHigh)) || !DT.dominates(U, &Or))
3084 continue;
3085
3086 // HighLow is good concat. If sum of two shifts amount equals to Width,
3087 // LowHigh must also be a good concat.
3088 if (*ZextLowShlAmt + *ZextHighShlAmt != Width)
3089 continue;
3090
3091 // Low must not overlap with High and most significant bits of Low must
3092 // not be shifted out.
3093 assert(ZextLowShlAmt->uge(HighSize) &&
3094 ZextLowShlAmt->ule(Width - LowSize) && "Invalid concat");
3095
3096 // We cannot reuse the result if it may produce poison.
3097 // Drop poison generating flags in the expression tree.
3098 // Or
3099 cast<Instruction>(U)->dropPoisonGeneratingFlags();
3100 // Shl
3101 cast<Instruction>(X)->dropPoisonGeneratingFlags();
3102
3103 FShiftArgs = {U, U, ConstantInt::get(Or0->getType(), *ZextHighShlAmt)};
3104 break;
3105 }
3106 }
3107
3108 if (FShiftArgs.empty())
3109 return std::nullopt;
3110
3111 Intrinsic::ID IID = IsFshl ? Intrinsic::fshl : Intrinsic::fshr;
3112 return std::make_pair(IID, FShiftArgs);
3113}
3114
3115/// Match UB-safe variants of the funnel shift intrinsic.
3117 if (auto Opt = IC.convertOrOfShiftsToFunnelShift(Or)) {
3118 auto [IID, FShiftArgs] = *Opt;
3119 Function *F =
3120 Intrinsic::getOrInsertDeclaration(Or.getModule(), IID, Or.getType());
3121 return CallInst::Create(F, FShiftArgs);
3122 }
3123
3124 return nullptr;
3125}
3126
3127/// Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
3129 assert(Or.getOpcode() == Instruction::Or && "bswap requires an 'or'");
3130 Value *Op0 = Or.getOperand(0), *Op1 = Or.getOperand(1);
3131 Type *Ty = Or.getType();
3132
3133 unsigned Width = Ty->getScalarSizeInBits();
3134 if ((Width & 1) != 0)
3135 return nullptr;
3136 unsigned HalfWidth = Width / 2;
3137
3138 // Canonicalize zext (lower half) to LHS.
3139 if (!isa<ZExtInst>(Op0))
3140 std::swap(Op0, Op1);
3141
3142 // Find lower/upper half.
3143 Value *LowerSrc, *ShlVal, *UpperSrc;
3144 const APInt *C;
3145 if (!match(Op0, m_OneUse(m_ZExt(m_Value(LowerSrc)))) ||
3146 !match(Op1, m_OneUse(m_Shl(m_Value(ShlVal), m_APInt(C)))) ||
3147 !match(ShlVal, m_OneUse(m_ZExt(m_Value(UpperSrc)))))
3148 return nullptr;
3149 if (*C != HalfWidth || LowerSrc->getType() != UpperSrc->getType() ||
3150 LowerSrc->getType()->getScalarSizeInBits() != HalfWidth)
3151 return nullptr;
3152
3153 auto ConcatIntrinsicCalls = [&](Intrinsic::ID id, Value *Lo, Value *Hi) {
3154 Value *NewLower = Builder.CreateZExt(Lo, Ty);
3155 Value *NewUpper = Builder.CreateZExt(Hi, Ty);
3156 NewUpper = Builder.CreateShl(NewUpper, HalfWidth);
3157 Value *BinOp = Builder.CreateDisjointOr(NewLower, NewUpper);
3158 return Builder.CreateIntrinsic(id, Ty, BinOp);
3159 };
3160
3161 // BSWAP: Push the concat down, swapping the lower/upper sources.
3162 // concat(bswap(x),bswap(y)) -> bswap(concat(x,y))
3163 Value *LowerBSwap, *UpperBSwap;
3164 if (match(LowerSrc, m_BSwap(m_Value(LowerBSwap))) &&
3165 match(UpperSrc, m_BSwap(m_Value(UpperBSwap))))
3166 return ConcatIntrinsicCalls(Intrinsic::bswap, UpperBSwap, LowerBSwap);
3167
3168 // BITREVERSE: Push the concat down, swapping the lower/upper sources.
3169 // concat(bitreverse(x),bitreverse(y)) -> bitreverse(concat(x,y))
3170 Value *LowerBRev, *UpperBRev;
3171 if (match(LowerSrc, m_BitReverse(m_Value(LowerBRev))) &&
3172 match(UpperSrc, m_BitReverse(m_Value(UpperBRev))))
3173 return ConcatIntrinsicCalls(Intrinsic::bitreverse, UpperBRev, LowerBRev);
3174
3175 // iX ext split: extending or(zext(x),shl(zext(y),bw/2) pattern
3176 // to consume sext/ashr:
3177 // or(zext(sext(x)),shl(zext(sext(ashr(x,xbw-1))),bw/2)
3178 // or(zext(x),shl(zext(ashr(x,xbw-1)),bw/2)
3179 Value *X;
3180 if (match(LowerSrc, m_SExtOrSelf(m_Value(X))) &&
3181 match(UpperSrc,
3183 m_Specific(X),
3184 m_SpecificInt(X->getType()->getScalarSizeInBits() - 1)))))
3185 return Builder.CreateSExt(X, Ty);
3186
3187 return nullptr;
3188}
3189
3190/// If all elements of two constant vectors are 0/-1 and inverses, return true.
3192 unsigned NumElts = cast<FixedVectorType>(C1->getType())->getNumElements();
3193 for (unsigned i = 0; i != NumElts; ++i) {
3194 Constant *EltC1 = C1->getAggregateElement(i);
3195 Constant *EltC2 = C2->getAggregateElement(i);
3196 if (!EltC1 || !EltC2)
3197 return false;
3198
3199 // One element must be all ones, and the other must be all zeros.
3200 if (!((match(EltC1, m_Zero()) && match(EltC2, m_AllOnes())) ||
3201 (match(EltC2, m_Zero()) && match(EltC1, m_AllOnes()))))
3202 return false;
3203 }
3204 return true;
3205}
3206
3207/// We have an expression of the form (A & C) | (B & D). If A is a scalar or
3208/// vector composed of all-zeros or all-ones values and is the bitwise 'not' of
3209/// B, it can be used as the condition operand of a select instruction.
3210/// We will detect (A & C) | ~(B | D) when the flag ABIsTheSame enabled.
3211Value *InstCombinerImpl::getSelectCondition(Value *A, Value *B,
3212 bool ABIsTheSame) {
3213 // We may have peeked through bitcasts in the caller.
3214 // Exit immediately if we don't have (vector) integer types.
3215 Type *Ty = A->getType();
3216 if (!Ty->isIntOrIntVectorTy() || !B->getType()->isIntOrIntVectorTy())
3217 return nullptr;
3218
3219 // If A is the 'not' operand of B and has enough signbits, we have our answer.
3220 if (ABIsTheSame ? (A == B) : match(B, m_Not(m_Specific(A)))) {
3221 // If these are scalars or vectors of i1, A can be used directly.
3222 if (Ty->isIntOrIntVectorTy(1))
3223 return A;
3224
3225 // If we look through a vector bitcast, the caller will bitcast the operands
3226 // to match the condition's number of bits (N x i1).
3227 // To make this poison-safe, disallow bitcast from wide element to narrow
3228 // element. That could allow poison in lanes where it was not present in the
3229 // original code.
3231 if (A->getType()->isIntOrIntVectorTy()) {
3232 unsigned NumSignBits = ComputeNumSignBits(A);
3233 if (NumSignBits == A->getType()->getScalarSizeInBits() &&
3234 NumSignBits <= Ty->getScalarSizeInBits())
3235 return Builder.CreateTrunc(A, CmpInst::makeCmpResultType(A->getType()));
3236 }
3237 return nullptr;
3238 }
3239
3240 // TODO: add support for sext and constant case
3241 if (ABIsTheSame)
3242 return nullptr;
3243
3244 // If both operands are constants, see if the constants are inverse bitmasks.
3245 Constant *AConst, *BConst;
3246 if (match(A, m_Constant(AConst)) && match(B, m_Constant(BConst)))
3247 if (AConst == ConstantExpr::getNot(BConst) &&
3249 return Builder.CreateZExtOrTrunc(A, CmpInst::makeCmpResultType(Ty));
3250
3251 // Look for more complex patterns. The 'not' op may be hidden behind various
3252 // casts. Look through sexts and bitcasts to find the booleans.
3253 Value *Cond;
3254 Value *NotB;
3255 if (match(A, m_SExt(m_Value(Cond))) &&
3256 Cond->getType()->isIntOrIntVectorTy(1)) {
3257 // A = sext i1 Cond; B = sext (not (i1 Cond))
3258 if (match(B, m_SExt(m_Not(m_Specific(Cond)))))
3259 return Cond;
3260
3261 // A = sext i1 Cond; B = not ({bitcast} (sext (i1 Cond)))
3262 // TODO: The one-use checks are unnecessary or misplaced. If the caller
3263 // checked for uses on logic ops/casts, that should be enough to
3264 // make this transform worthwhile.
3265 if (match(B, m_OneUse(m_Not(m_Value(NotB))))) {
3266 NotB = peekThroughBitcast(NotB, true);
3267 if (match(NotB, m_SExt(m_Specific(Cond))))
3268 return Cond;
3269 }
3270 }
3271
3272 // All scalar (and most vector) possibilities should be handled now.
3273 // Try more matches that only apply to non-splat constant vectors.
3274 if (!Ty->isVectorTy())
3275 return nullptr;
3276
3277 // If both operands are xor'd with constants using the same sexted boolean
3278 // operand, see if the constants are inverse bitmasks.
3279 // TODO: Use ConstantExpr::getNot()?
3280 if (match(A, (m_Xor(m_SExt(m_Value(Cond)), m_Constant(AConst)))) &&
3281 match(B, (m_Xor(m_SExt(m_Specific(Cond)), m_Constant(BConst)))) &&
3282 Cond->getType()->isIntOrIntVectorTy(1) &&
3283 areInverseVectorBitmasks(AConst, BConst)) {
3285 return Builder.CreateXor(Cond, AConst);
3286 }
3287 return nullptr;
3288}
3289
3290/// We have an expression of the form (A & B) | (C & D). Try to simplify this
3291/// to "A' ? B : D", where A' is a boolean or vector of booleans.
3292/// When InvertFalseVal is set to true, we try to match the pattern
3293/// where we have peeked through a 'not' op and A and C are the same:
3294/// (A & B) | ~(A | D) --> (A & B) | (~A & ~D) --> A' ? B : ~D
3295Value *InstCombinerImpl::matchSelectFromAndOr(Value *A, Value *B, Value *C,
3296 Value *D, bool InvertFalseVal) {
3297 // The potential condition of the select may be bitcasted. In that case, look
3298 // through its bitcast and the corresponding bitcast of the 'not' condition.
3299 Type *OrigType = A->getType();
3300 A = peekThroughBitcast(A, true);
3301 C = peekThroughBitcast(C, true);
3302 if (Value *Cond = getSelectCondition(A, C, InvertFalseVal)) {
3303 // ((bc Cond) & B) | ((bc ~Cond) & D) --> bc (select Cond, (bc B), (bc D))
3304 // If this is a vector, we may need to cast to match the condition's length.
3305 // The bitcasts will either all exist or all not exist. The builder will
3306 // not create unnecessary casts if the types already match.
3307 Type *SelTy = A->getType();
3308 if (auto *VecTy = dyn_cast<VectorType>(Cond->getType())) {
3309 // For a fixed or scalable vector get N from <{vscale x} N x iM>
3310 unsigned Elts = VecTy->getElementCount().getKnownMinValue();
3311 // For a fixed or scalable vector, get the size in bits of N x iM; for a
3312 // scalar this is just M.
3313 unsigned SelEltSize = SelTy->getPrimitiveSizeInBits().getKnownMinValue();
3314 Type *EltTy = Builder.getIntNTy(SelEltSize / Elts);
3315 SelTy = VectorType::get(EltTy, VecTy->getElementCount());
3316 }
3317 Value *BitcastB = Builder.CreateBitCast(B, SelTy);
3318 if (InvertFalseVal)
3319 D = Builder.CreateNot(D);
3320 Value *BitcastD = Builder.CreateBitCast(D, SelTy);
3321 Value *Select = Builder.CreateSelect(Cond, BitcastB, BitcastD);
3322 return Builder.CreateBitCast(Select, OrigType);
3323 }
3324
3325 return nullptr;
3326}
3327
3328// (icmp eq X, C) | (icmp ult Other, (X - C)) -> (icmp ule Other, (X - (C + 1)))
3329// (icmp ne X, C) & (icmp uge Other, (X - C)) -> (icmp ugt Other, (X - (C + 1)))
3331 bool IsAnd, bool IsLogical,
3332 IRBuilderBase &Builder) {
3333 Value *LHS0 = LHS->getOperand(0);
3334 Value *RHS0 = RHS->getOperand(0);
3335 Value *RHS1 = RHS->getOperand(1);
3336
3337 ICmpInst::Predicate LPred =
3338 IsAnd ? LHS->getInversePredicate() : LHS->getPredicate();
3339 ICmpInst::Predicate RPred =
3340 IsAnd ? RHS->getInversePredicate() : RHS->getPredicate();
3341
3342 const APInt *CInt;
3343 if (LPred != ICmpInst::ICMP_EQ ||
3344 !match(LHS->getOperand(1), m_APIntAllowPoison(CInt)) ||
3345 !LHS0->getType()->isIntOrIntVectorTy() ||
3346 !(LHS->hasOneUse() || RHS->hasOneUse()))
3347 return nullptr;
3348
3349 auto MatchRHSOp = [LHS0, CInt](const Value *RHSOp) {
3350 return match(RHSOp,
3351 m_Add(m_Specific(LHS0), m_SpecificIntAllowPoison(-*CInt))) ||
3352 (CInt->isZero() && RHSOp == LHS0);
3353 };
3354
3355 Value *Other;
3356 if (RPred == ICmpInst::ICMP_ULT && MatchRHSOp(RHS1))
3357 Other = RHS0;
3358 else if (RPred == ICmpInst::ICMP_UGT && MatchRHSOp(RHS0))
3359 Other = RHS1;
3360 else
3361 return nullptr;
3362
3363 if (IsLogical)
3364 Other = Builder.CreateFreeze(Other);
3365
3366 return Builder.CreateICmp(
3368 Builder.CreateSub(LHS0, ConstantInt::get(LHS0->getType(), *CInt + 1)),
3369 Other);
3370}
3371
3372/// Fold (icmp)&(icmp) or (icmp)|(icmp) if possible.
3373/// If IsLogical is true, then the and/or is in select form and the transform
3374/// must be poison-safe.
3375Value *InstCombinerImpl::foldAndOrOfICmps(ICmpInst *LHS, ICmpInst *RHS,
3376 Instruction &I, bool IsAnd,
3377 bool IsLogical) {
3378 const SimplifyQuery Q = SQ.getWithInstruction(&I);
3379
3380 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
3381 Value *LHS0 = LHS->getOperand(0), *RHS0 = RHS->getOperand(0);
3382 Value *LHS1 = LHS->getOperand(1), *RHS1 = RHS->getOperand(1);
3383
3384 const APInt *LHSC = nullptr, *RHSC = nullptr;
3385 match(LHS1, m_APInt(LHSC));
3386 match(RHS1, m_APInt(RHSC));
3387
3388 // (icmp1 A, B) | (icmp2 A, B) --> (icmp3 A, B)
3389 // (icmp1 A, B) & (icmp2 A, B) --> (icmp3 A, B)
3390 if (predicatesFoldable(PredL, PredR)) {
3391 if (LHS0 == RHS1 && LHS1 == RHS0) {
3392 PredL = ICmpInst::getSwappedPredicate(PredL);
3393 std::swap(LHS0, LHS1);
3394 }
3395 if (LHS0 == RHS0 && LHS1 == RHS1) {
3396 unsigned Code = IsAnd ? getICmpCode(PredL) & getICmpCode(PredR)
3397 : getICmpCode(PredL) | getICmpCode(PredR);
3398 bool IsSigned = LHS->isSigned() || RHS->isSigned();
3399 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
3400 }
3401 }
3402
3403 if (Value *V =
3404 foldAndOrOfICmpEqConstantAndICmp(LHS, RHS, IsAnd, IsLogical, Builder))
3405 return V;
3406 // We can treat logical like bitwise here, because both operands are used on
3407 // the LHS, and as such poison from both will propagate.
3409 /*IsLogical*/ false, Builder))
3410 return V;
3411
3412 if (Value *V = foldAndOrOfICmpsWithConstEq(LHS, RHS, IsAnd, IsLogical,
3413 Builder, Q, I))
3414 return V;
3415 // We can convert this case to bitwise and, because both operands are used
3416 // on the LHS, and as such poison from both will propagate.
3418 RHS, LHS, IsAnd, /*IsLogical=*/false, Builder, Q, I)) {
3419 // If RHS is still used, we should drop samesign flag.
3420 if (IsLogical && RHS->hasSameSign() && !RHS->use_empty()) {
3421 RHS->setSameSign(false);
3423 }
3424 return V;
3425 }
3426
3427 if (Value *V = foldIsPowerOf2OrZero(LHS, RHS, IsAnd, Builder, *this))
3428 return V;
3429 if (Value *V = foldIsPowerOf2OrZero(RHS, LHS, IsAnd, Builder, *this))
3430 return V;
3431
3432 // TODO: One of these directions is fine with logical and/or, the other could
3433 // be supported by inserting freeze.
3434 if (!IsLogical) {
3435 // E.g. (icmp slt x, 0) | (icmp sgt x, n) --> icmp ugt x, n
3436 // E.g. (icmp sge x, 0) & (icmp slt x, n) --> icmp ult x, n
3437 if (Value *V = simplifyRangeCheck(LHS, RHS, /*Inverted=*/!IsAnd))
3438 return V;
3439
3440 // E.g. (icmp sgt x, n) | (icmp slt x, 0) --> icmp ugt x, n
3441 // E.g. (icmp slt x, n) & (icmp sge x, 0) --> icmp ult x, n
3442 if (Value *V = simplifyRangeCheck(RHS, LHS, /*Inverted=*/!IsAnd))
3443 return V;
3444 }
3445
3446 // TODO: Add conjugated or fold, check whether it is safe for logical and/or.
3447 if (IsAnd && !IsLogical)
3449 return V;
3450
3451 if (Value *V = foldIsPowerOf2(LHS, RHS, IsAnd, Builder, *this))
3452 return V;
3453
3454 if (Value *V = foldPowerOf2AndShiftedMask(LHS, RHS, IsAnd, Builder))
3455 return V;
3456
3457 // TODO: Verify whether this is safe for logical and/or.
3458 if (!IsLogical) {
3459 if (Value *X = foldUnsignedUnderflowCheck(LHS, RHS, IsAnd, Q, Builder))
3460 return X;
3461 if (Value *X = foldUnsignedUnderflowCheck(RHS, LHS, IsAnd, Q, Builder))
3462 return X;
3463 }
3464
3465 // (icmp ne A, 0) | (icmp ne B, 0) --> (icmp ne (A|B), 0)
3466 // (icmp eq A, 0) & (icmp eq B, 0) --> (icmp eq (A|B), 0)
3467 // TODO: Remove this and below when foldLogOpOfMaskedICmps can handle undefs.
3468 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3469 PredL == PredR && match(LHS1, m_ZeroInt()) && match(RHS1, m_ZeroInt()) &&
3470 LHS0->getType() == RHS0->getType() &&
3471 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3472 Value *NewOr = Builder.CreateOr(LHS0, RHS0);
3473 return Builder.CreateICmp(PredL, NewOr,
3475 }
3476
3477 // (icmp ne A, -1) | (icmp ne B, -1) --> (icmp ne (A&B), -1)
3478 // (icmp eq A, -1) & (icmp eq B, -1) --> (icmp eq (A&B), -1)
3479 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3480 PredL == PredR && match(LHS1, m_AllOnes()) && match(RHS1, m_AllOnes()) &&
3481 LHS0->getType() == RHS0->getType() &&
3482 (!IsLogical || isGuaranteedNotToBePoison(RHS0))) {
3483 Value *NewAnd = Builder.CreateAnd(LHS0, RHS0);
3484 return Builder.CreateICmp(PredL, NewAnd,
3486 }
3487
3488 if (!IsLogical)
3489 if (Value *V =
3491 return V;
3492
3493 // This only handles icmp of constants: (icmp1 A, C1) | (icmp2 B, C2).
3494 if (!LHSC || !RHSC)
3495 return nullptr;
3496
3497 // (trunc x) == C1 & (and x, CA) == C2 -> (and x, CA|CMAX) == C1|C2
3498 // (trunc x) != C1 | (and x, CA) != C2 -> (and x, CA|CMAX) != C1|C2
3499 // where CMAX is the all ones value for the truncated type,
3500 // iff the lower bits of C2 and CA are zero.
3501 if (PredL == (IsAnd ? ICmpInst::ICMP_EQ : ICmpInst::ICMP_NE) &&
3502 PredL == PredR && LHS->hasOneUse() && RHS->hasOneUse()) {
3503 Value *V;
3504 const APInt *AndC, *SmallC = nullptr, *BigC = nullptr;
3505
3506 // (trunc x) == C1 & (and x, CA) == C2
3507 // (and x, CA) == C2 & (trunc x) == C1
3508 if (match(RHS0, m_Trunc(m_Value(V))) &&
3509 match(LHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3510 SmallC = RHSC;
3511 BigC = LHSC;
3512 } else if (match(LHS0, m_Trunc(m_Value(V))) &&
3513 match(RHS0, m_And(m_Specific(V), m_APInt(AndC)))) {
3514 SmallC = LHSC;
3515 BigC = RHSC;
3516 }
3517
3518 if (SmallC && BigC) {
3519 unsigned BigBitSize = BigC->getBitWidth();
3520 unsigned SmallBitSize = SmallC->getBitWidth();
3521
3522 // Check that the low bits are zero.
3523 APInt Low = APInt::getLowBitsSet(BigBitSize, SmallBitSize);
3524 if ((Low & *AndC).isZero() && (Low & *BigC).isZero()) {
3525 Value *NewAnd = Builder.CreateAnd(V, Low | *AndC);
3526 APInt N = SmallC->zext(BigBitSize) | *BigC;
3527 Value *NewVal = ConstantInt::get(NewAnd->getType(), N);
3528 return Builder.CreateICmp(PredL, NewAnd, NewVal);
3529 }
3530 }
3531 }
3532
3533 // Match naive pattern (and its inverted form) for checking if two values
3534 // share same sign. An example of the pattern:
3535 // (icmp slt (X & Y), 0) | (icmp sgt (X | Y), -1) -> (icmp sgt (X ^ Y), -1)
3536 // Inverted form (example):
3537 // (icmp slt (X | Y), 0) & (icmp sgt (X & Y), -1) -> (icmp slt (X ^ Y), 0)
3538 bool TrueIfSignedL, TrueIfSignedR;
3539 if (isSignBitCheck(PredL, *LHSC, TrueIfSignedL) &&
3540 isSignBitCheck(PredR, *RHSC, TrueIfSignedR) &&
3541 (RHS->hasOneUse() || LHS->hasOneUse())) {
3542 Value *X, *Y;
3543 if (IsAnd) {
3544 if ((TrueIfSignedL && !TrueIfSignedR &&
3545 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3546 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y)))) ||
3547 (!TrueIfSignedL && TrueIfSignedR &&
3548 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3549 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y))))) {
3550 Value *NewXor = Builder.CreateXor(X, Y);
3551 return Builder.CreateIsNeg(NewXor);
3552 }
3553 } else {
3554 if ((TrueIfSignedL && !TrueIfSignedR &&
3555 match(LHS0, m_And(m_Value(X), m_Value(Y))) &&
3556 match(RHS0, m_c_Or(m_Specific(X), m_Specific(Y)))) ||
3557 (!TrueIfSignedL && TrueIfSignedR &&
3558 match(LHS0, m_Or(m_Value(X), m_Value(Y))) &&
3559 match(RHS0, m_c_And(m_Specific(X), m_Specific(Y))))) {
3560 Value *NewXor = Builder.CreateXor(X, Y);
3561 return Builder.CreateIsNotNeg(NewXor);
3562 }
3563 }
3564 }
3565
3566 // (X & ExpMask) != 0 && (X & ExpMask) != ExpMask -> isnormal(X)
3567 // (X & ExpMask) == 0 || (X & ExpMask) == ExpMask -> !isnormal(X)
3568 Value *X;
3569 const APInt *MaskC;
3570 if (LHS0 == RHS0 && PredL == PredR &&
3571 PredL == (IsAnd ? ICmpInst::ICMP_NE : ICmpInst::ICMP_EQ) &&
3572 !I.getFunction()->hasFnAttribute(Attribute::NoImplicitFloat) &&
3573 LHS->hasOneUse() && RHS->hasOneUse() &&
3574 match(LHS0, m_And(m_ElementWiseBitCast(m_Value(X)), m_APInt(MaskC))) &&
3575 X->getType()->getScalarType()->isIEEELikeFPTy() &&
3576 APFloat(X->getType()->getScalarType()->getFltSemantics(), *MaskC)
3577 .isPosInfinity() &&
3578 ((LHSC->isZero() && *RHSC == *MaskC) ||
3579 (RHSC->isZero() && *LHSC == *MaskC)))
3580 return Builder.createIsFPClass(X, IsAnd ? FPClassTest::fcNormal
3582
3583 return foldAndOrOfICmpsUsingRanges(LHS, RHS, IsAnd);
3584}
3585
3586/// If IsLogical is true, then the and/or is in select form and the transform
3587/// must be poison-safe.
3588Value *InstCombinerImpl::foldBooleanAndOr(Value *LHS, Value *RHS,
3589 Instruction &I, bool IsAnd,
3590 bool IsLogical) {
3591 if (!LHS->getType()->isIntOrIntVectorTy(1))
3592 return nullptr;
3593
3594 // handle (roughly):
3595 // (icmp ne (A & B), C) | (icmp ne (A & D), E)
3596 // (icmp eq (A & B), C) & (icmp eq (A & D), E)
3597 if (Value *V = foldLogOpOfMaskedICmps(LHS, RHS, IsAnd, IsLogical, Builder,
3598 SQ.getWithInstruction(&I)))
3599 return V;
3600
3601 if (auto *LHSCmp = dyn_cast<ICmpInst>(LHS))
3602 if (auto *RHSCmp = dyn_cast<ICmpInst>(RHS))
3603 if (Value *Res = foldAndOrOfICmps(LHSCmp, RHSCmp, I, IsAnd, IsLogical))
3604 return Res;
3605
3606 if (auto *LHSCmp = dyn_cast<FCmpInst>(LHS))
3607 if (auto *RHSCmp = dyn_cast<FCmpInst>(RHS))
3608 if (Value *Res = foldLogicOfFCmps(LHSCmp, RHSCmp, IsAnd, IsLogical))
3609 return Res;
3610
3611 if (Value *Res = foldEqOfParts(LHS, RHS, IsAnd))
3612 return Res;
3613
3614 return nullptr;
3615}
3616
3618 InstCombiner::BuilderTy &Builder) {
3619 assert(I.getOpcode() == Instruction::Or &&
3620 "Simplification only supports or at the moment.");
3621
3622 Value *Cmp1, *Cmp2, *Cmp3, *Cmp4;
3623 if (!match(I.getOperand(0), m_And(m_Value(Cmp1), m_Value(Cmp2))) ||
3624 !match(I.getOperand(1), m_And(m_Value(Cmp3), m_Value(Cmp4))))
3625 return nullptr;
3626
3627 // Check if any two pairs of the and operations are inversions of each other.
3628 if (isKnownInversion(Cmp1, Cmp3) && isKnownInversion(Cmp2, Cmp4))
3629 return Builder.CreateXor(Cmp1, Cmp4);
3630 if (isKnownInversion(Cmp1, Cmp4) && isKnownInversion(Cmp2, Cmp3))
3631 return Builder.CreateXor(Cmp1, Cmp3);
3632
3633 return nullptr;
3634}
3635
3636/// Match \p V as "shufflevector -> bitcast" or "extractelement -> zext -> shl"
3637/// patterns, which extract vector elements and pack them in the same relative
3638/// positions.
3639///
3640/// \p Vec is the underlying vector being extracted from.
3641/// \p Mask is a bitmask identifying which packed elements are obtained from the
3642/// vector.
3643/// \p VecOffset is the vector element corresponding to index 0 of the
3644/// mask.
3646 int64_t &VecOffset,
3647 SmallBitVector &Mask,
3648 const DataLayout &DL) {
3649 // First try to match extractelement -> zext -> shl
3650 uint64_t VecIdx, ShlAmt;
3652 m_ConstantInt(VecIdx))),
3653 ShlAmt))) {
3654 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3655 if (!VecTy)
3656 return false;
3657 auto *EltTy = dyn_cast<IntegerType>(VecTy->getElementType());
3658 if (!EltTy)
3659 return false;
3660
3661 const unsigned EltBitWidth = EltTy->getBitWidth();
3662 const unsigned TargetBitWidth = V->getType()->getIntegerBitWidth();
3663 if (TargetBitWidth % EltBitWidth != 0 || ShlAmt % EltBitWidth != 0)
3664 return false;
3665 const unsigned TargetEltWidth = TargetBitWidth / EltBitWidth;
3666 const unsigned ShlEltAmt = ShlAmt / EltBitWidth;
3667
3668 const unsigned MaskIdx =
3669 DL.isLittleEndian() ? ShlEltAmt : TargetEltWidth - ShlEltAmt - 1;
3670
3671 VecOffset = static_cast<int64_t>(VecIdx) - static_cast<int64_t>(MaskIdx);
3672 Mask.resize(TargetEltWidth);
3673 Mask.set(MaskIdx);
3674 return true;
3675 }
3676
3677 // Now try to match a bitcasted subvector.
3678 Instruction *SrcVecI;
3679 if (!match(V, m_BitCast(m_Instruction(SrcVecI))))
3680 return false;
3681
3682 auto *SrcTy = dyn_cast<FixedVectorType>(SrcVecI->getType());
3683 if (!SrcTy)
3684 return false;
3685
3686 Mask.resize(SrcTy->getNumElements());
3687
3688 // First check for a subvector obtained from a shufflevector.
3689 if (isa<ShuffleVectorInst>(SrcVecI)) {
3690 Constant *ConstVec;
3691 ArrayRef<int> ShuffleMask;
3692 if (!match(SrcVecI, m_Shuffle(m_Value(Vec), m_Constant(ConstVec),
3693 m_Mask(ShuffleMask))))
3694 return false;
3695
3696 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3697 if (!VecTy)
3698 return false;
3699
3700 const unsigned NumVecElts = VecTy->getNumElements();
3701 bool FoundVecOffset = false;
3702 for (unsigned Idx = 0; Idx < ShuffleMask.size(); ++Idx) {
3703 if (ShuffleMask[Idx] == PoisonMaskElem)
3704 return false;
3705 const unsigned ShuffleIdx = ShuffleMask[Idx];
3706 if (ShuffleIdx >= NumVecElts) {
3707 const unsigned ConstIdx = ShuffleIdx - NumVecElts;
3708 auto *ConstElt =
3709 dyn_cast<ConstantInt>(ConstVec->getAggregateElement(ConstIdx));
3710 if (!ConstElt || !ConstElt->isNullValue())
3711 return false;
3712 continue;
3713 }
3714
3715 if (FoundVecOffset) {
3716 if (VecOffset + Idx != ShuffleIdx)
3717 return false;
3718 } else {
3719 if (ShuffleIdx < Idx)
3720 return false;
3721 VecOffset = ShuffleIdx - Idx;
3722 FoundVecOffset = true;
3723 }
3724 Mask.set(Idx);
3725 }
3726 return FoundVecOffset;
3727 }
3728
3729 // Check for a subvector obtained as an (insertelement V, 0, idx)
3730 uint64_t InsertIdx;
3731 if (!match(SrcVecI,
3732 m_InsertElt(m_Value(Vec), m_Zero(), m_ConstantInt(InsertIdx))))
3733 return false;
3734
3735 auto *VecTy = dyn_cast<FixedVectorType>(Vec->getType());
3736 if (!VecTy)
3737 return false;
3738 VecOffset = 0;
3739 bool AlreadyInsertedMaskedElt = Mask.test(InsertIdx);
3740 Mask.set();
3741 if (!AlreadyInsertedMaskedElt)
3742 Mask.reset(InsertIdx);
3743 return true;
3744}
3745
3746/// Try to fold the join of two scalar integers whose contents are packed
3747/// elements of the same vector.
3749 InstCombiner::BuilderTy &Builder,
3750 const DataLayout &DL) {
3751 assert(I.getOpcode() == Instruction::Or);
3752 Value *LhsVec, *RhsVec;
3753 int64_t LhsVecOffset, RhsVecOffset;
3754 SmallBitVector Mask;
3755 if (!matchSubIntegerPackFromVector(I.getOperand(0), LhsVec, LhsVecOffset,
3756 Mask, DL))
3757 return nullptr;
3758 if (!matchSubIntegerPackFromVector(I.getOperand(1), RhsVec, RhsVecOffset,
3759 Mask, DL))
3760 return nullptr;
3761 if (LhsVec != RhsVec || LhsVecOffset != RhsVecOffset)
3762 return nullptr;
3763
3764 // Convert into shufflevector -> bitcast;
3765 const unsigned ZeroVecIdx =
3766 cast<FixedVectorType>(LhsVec->getType())->getNumElements();
3767 SmallVector<int> ShuffleMask(Mask.size(), ZeroVecIdx);
3768 for (unsigned Idx : Mask.set_bits()) {
3769 assert(LhsVecOffset + Idx >= 0);
3770 ShuffleMask[Idx] = LhsVecOffset + Idx;
3771 }
3772
3773 Value *MaskedVec = Builder.CreateShuffleVector(
3774 LhsVec, Constant::getNullValue(LhsVec->getType()), ShuffleMask,
3775 I.getName() + ".v");
3776 return CastInst::Create(Instruction::BitCast, MaskedVec, I.getType());
3777}
3778
3779/// Match \p V as "lshr -> mask -> zext -> shl".
3780///
3781/// \p Int is the underlying integer being extracted from.
3782/// \p Mask is a bitmask identifying which bits of the integer are being
3783/// extracted. \p Offset identifies which bit of the result \p V corresponds to
3784/// the least significant bit of \p Int
3785static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask,
3786 uint64_t &Offset, bool &IsShlNUW,
3787 bool &IsShlNSW) {
3788 Value *ShlOp0;
3789 uint64_t ShlAmt = 0;
3790 if (!match(V, m_OneUse(m_Shl(m_Value(ShlOp0), m_ConstantInt(ShlAmt)))))
3791 return false;
3792
3793 IsShlNUW = cast<BinaryOperator>(V)->hasNoUnsignedWrap();
3794 IsShlNSW = cast<BinaryOperator>(V)->hasNoSignedWrap();
3795
3796 Value *ZExtOp0;
3797 if (!match(ShlOp0, m_OneUse(m_ZExt(m_Value(ZExtOp0)))))
3798 return false;
3799
3800 Value *MaskedOp0;
3801 const APInt *ShiftedMaskConst = nullptr;
3802 if (!match(ZExtOp0, m_CombineOr(m_OneUse(m_And(m_Value(MaskedOp0),
3803 m_APInt(ShiftedMaskConst))),
3804 m_Value(MaskedOp0))))
3805 return false;
3806
3807 uint64_t LShrAmt = 0;
3808 if (!match(MaskedOp0,
3810 m_Value(Int))))
3811 return false;
3812
3813 if (LShrAmt > ShlAmt)
3814 return false;
3815 Offset = ShlAmt - LShrAmt;
3816
3817 Mask = ShiftedMaskConst ? ShiftedMaskConst->shl(LShrAmt)
3819 Int->getType()->getScalarSizeInBits(), LShrAmt);
3820
3821 return true;
3822}
3823
3824/// Try to fold the join of two scalar integers whose bits are unpacked and
3825/// zexted from the same source integer.
3827 InstCombiner::BuilderTy &Builder) {
3828
3829 Value *LhsInt, *RhsInt;
3830 APInt LhsMask, RhsMask;
3831 uint64_t LhsOffset, RhsOffset;
3832 bool IsLhsShlNUW, IsLhsShlNSW, IsRhsShlNUW, IsRhsShlNSW;
3833 if (!matchZExtedSubInteger(Lhs, LhsInt, LhsMask, LhsOffset, IsLhsShlNUW,
3834 IsLhsShlNSW))
3835 return nullptr;
3836 if (!matchZExtedSubInteger(Rhs, RhsInt, RhsMask, RhsOffset, IsRhsShlNUW,
3837 IsRhsShlNSW))
3838 return nullptr;
3839 if (LhsInt != RhsInt || LhsOffset != RhsOffset)
3840 return nullptr;
3841
3842 APInt Mask = LhsMask | RhsMask;
3843
3844 Type *DestTy = Lhs->getType();
3845 Value *Res = Builder.CreateShl(
3846 Builder.CreateZExt(
3847 Builder.CreateAnd(LhsInt, Mask, LhsInt->getName() + ".mask"), DestTy,
3848 LhsInt->getName() + ".zext"),
3849 ConstantInt::get(DestTy, LhsOffset), "", IsLhsShlNUW && IsRhsShlNUW,
3850 IsLhsShlNSW && IsRhsShlNSW);
3851 Res->takeName(Lhs);
3852 return Res;
3853}
3854
3855// A decomposition of ((X & Mask) * Factor). The NUW / NSW bools
3856// track these properities for preservation. Note that we can decompose
3857// equivalent select form of this expression (e.g. (!(X & Mask) ? 0 : Mask *
3858// Factor))
3863 bool NUW;
3864 bool NSW;
3865
3867 return X == Other.X && !Mask.intersects(Other.Mask) &&
3868 Factor == Other.Factor;
3869 }
3870};
3871
3872static std::optional<DecomposedBitMaskMul> matchBitmaskMul(Value *V) {
3874 if (!Op)
3875 return std::nullopt;
3876
3877 // Decompose (A & N) * C) into BitMaskMul
3878 Value *Original = nullptr;
3879 const APInt *Mask = nullptr;
3880 const APInt *MulConst = nullptr;
3881 if (match(Op, m_Mul(m_And(m_Value(Original), m_APInt(Mask)),
3882 m_APInt(MulConst)))) {
3883 if (MulConst->isZero() || Mask->isZero())
3884 return std::nullopt;
3885
3886 return std::optional<DecomposedBitMaskMul>(
3887 {Original, *MulConst, *Mask,
3888 cast<BinaryOperator>(Op)->hasNoUnsignedWrap(),
3889 cast<BinaryOperator>(Op)->hasNoSignedWrap()});
3890 }
3891
3892 Value *Cond = nullptr;
3893 const APInt *EqZero = nullptr, *NeZero = nullptr;
3894
3895 // Decompose ((A & N) ? 0 : N * C) into BitMaskMul
3896 if (match(Op, m_Select(m_Value(Cond), m_APInt(EqZero), m_APInt(NeZero)))) {
3897 auto ICmpDecompose =
3898 decomposeBitTest(Cond, /*LookThroughTrunc=*/true,
3899 /*AllowNonZeroC=*/false, /*DecomposeBitMask=*/true);
3900 if (!ICmpDecompose.has_value())
3901 return std::nullopt;
3902
3903 assert(ICmpInst::isEquality(ICmpDecompose->Pred) &&
3904 ICmpDecompose->C.isZero());
3905
3906 if (ICmpDecompose->Pred == ICmpInst::ICMP_NE)
3907 std::swap(EqZero, NeZero);
3908
3909 if (!EqZero->isZero() || NeZero->isZero())
3910 return std::nullopt;
3911
3912 if (!ICmpDecompose->Mask.isPowerOf2() || ICmpDecompose->Mask.isZero() ||
3913 NeZero->getBitWidth() != ICmpDecompose->Mask.getBitWidth())
3914 return std::nullopt;
3915
3916 if (!NeZero->urem(ICmpDecompose->Mask).isZero())
3917 return std::nullopt;
3918
3919 return std::optional<DecomposedBitMaskMul>(
3920 {ICmpDecompose->X, NeZero->udiv(ICmpDecompose->Mask),
3921 ICmpDecompose->Mask, /*NUW=*/false, /*NSW=*/false});
3922 }
3923
3924 return std::nullopt;
3925}
3926
3927/// (A & N) * C + (A & M) * C -> (A & (N + M)) & C
3928/// This also accepts the equivalent select form of (A & N) * C
3929/// expressions i.e. !(A & N) ? 0 : N * C)
3930static Value *foldBitmaskMul(Value *Op0, Value *Op1,
3931 InstCombiner::BuilderTy &Builder) {
3932 auto Decomp1 = matchBitmaskMul(Op1);
3933 if (!Decomp1)
3934 return nullptr;
3935
3936 auto Decomp0 = matchBitmaskMul(Op0);
3937 if (!Decomp0)
3938 return nullptr;
3939
3940 if (Decomp0->isCombineableWith(*Decomp1)) {
3941 Value *NewAnd = Builder.CreateAnd(
3942 Decomp0->X,
3943 ConstantInt::get(Decomp0->X->getType(), Decomp0->Mask + Decomp1->Mask));
3944
3945 return Builder.CreateMul(
3946 NewAnd, ConstantInt::get(NewAnd->getType(), Decomp1->Factor), "",
3947 Decomp0->NUW && Decomp1->NUW, Decomp0->NSW && Decomp1->NSW);
3948 }
3949
3950 return nullptr;
3951}
3952
3953Value *InstCombinerImpl::foldDisjointOr(Value *LHS, Value *RHS) {
3954 if (Value *Res = foldBitmaskMul(LHS, RHS, Builder))
3955 return Res;
3957 return Res;
3958
3959 return nullptr;
3960}
3961
3962Value *InstCombinerImpl::reassociateDisjointOr(Value *LHS, Value *RHS) {
3963
3964 Value *X, *Y;
3966 if (Value *Res = foldDisjointOr(LHS, X))
3967 return Builder.CreateDisjointOr(Res, Y);
3968 if (Value *Res = foldDisjointOr(LHS, Y))
3969 return Builder.CreateDisjointOr(Res, X);
3970 }
3971
3973 if (Value *Res = foldDisjointOr(X, RHS))
3974 return Builder.CreateDisjointOr(Res, Y);
3975 if (Value *Res = foldDisjointOr(Y, RHS))
3976 return Builder.CreateDisjointOr(Res, X);
3977 }
3978
3979 return nullptr;
3980}
3981
3982/// Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2))
3983/// --> (ugt x (c2/c1)). This code checks whether a multiplication of two
3984/// unsigned numbers (one is a constant) is mathematically greater than a
3985/// second constant.
3987 InstCombiner::BuilderTy &Builder,
3988 const DataLayout &DL) {
3989 Value *WOV, *X;
3990 const APInt *C1, *C2;
3991 if (match(&I,
3994 m_Value(X), m_APInt(C1)))),
3997 m_APInt(C2))))) &&
3998 !C1->isZero()) {
3999 Constant *NewC = ConstantInt::get(X->getType(), C2->udiv(*C1));
4000 return Builder.CreateICmp(ICmpInst::ICMP_UGT, X, NewC);
4001 }
4002 return nullptr;
4003}
4004
4005/// Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X)
4006/// select(X <s 0, -X, 0) | smax(X, 0) --> abs(X)
4008 InstCombiner::BuilderTy &Builder) {
4009 Value *X;
4010 Value *Sel;
4011 if (match(&I,
4013 auto NegX = m_Neg(m_Specific(X));
4015 m_ZeroInt()),
4016 m_ZeroInt(), NegX)) ||
4018 m_ZeroInt()),
4019 NegX, m_ZeroInt())))
4020 return Builder.CreateBinaryIntrinsic(Intrinsic::abs, X,
4021 Builder.getFalse());
4022 }
4023 return nullptr;
4024}
4025
4027 Value *C, *A, *B;
4028 // (C && A) || (!C && B)
4029 // (C && A) || (B && !C)
4030 // (A && C) || (!C && B)
4031 // (A && C) || (B && !C) (may require freeze)
4032 //
4033 // => select C, A, B
4034 if (match(Op1, m_c_LogicalAnd(m_Not(m_Value(C)), m_Value(B))) &&
4036 auto *SelOp0 = dyn_cast<SelectInst>(Op0);
4037 auto *SelOp1 = dyn_cast<SelectInst>(Op1);
4038
4039 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4040 match(SelOp1->getTrueValue(),
4041 m_Not(m_Specific(SelOp0->getTrueValue())));
4042 if (MayNeedFreeze)
4043 C = Builder.CreateFreeze(C);
4045 Value *C2 = nullptr, *A2 = nullptr, *B2 = nullptr;
4046 if (match(Op0, m_LogicalAnd(m_Specific(C), m_Value(A2))) && SelOp0) {
4047 return SelectInst::Create(C, A, B, "", nullptr, SelOp0);
4048 } else if (match(Op1, m_LogicalAnd(m_Not(m_Value(C2)), m_Value(B2))) &&
4049 SelOp1) {
4050 SelectInst *NewSI = SelectInst::Create(C, A, B, "", nullptr, SelOp1);
4051 NewSI->swapProfMetadata();
4052 return NewSI;
4053 } else {
4054 return createSelectInstWithUnknownProfile(C, A, B);
4055 }
4056 }
4057 return SelectInst::Create(C, A, B);
4058 }
4059
4060 // (!C && A) || (C && B)
4061 // (A && !C) || (C && B)
4062 // (!C && A) || (B && C)
4063 // (A && !C) || (B && C) (may require freeze)
4064 //
4065 // => select C, B, A
4066 if (match(Op0, m_c_LogicalAnd(m_Not(m_Value(C)), m_Value(A))) &&
4068 auto *SelOp0 = dyn_cast<SelectInst>(Op0);
4069 auto *SelOp1 = dyn_cast<SelectInst>(Op1);
4070 bool MayNeedFreeze = SelOp0 && SelOp1 &&
4071 match(SelOp0->getTrueValue(),
4072 m_Not(m_Specific(SelOp1->getTrueValue())));
4073 if (MayNeedFreeze)
4074 C = Builder.CreateFreeze(C);
4076 Value *C2 = nullptr, *A2 = nullptr, *B2 = nullptr;
4077 if (match(Op0, m_LogicalAnd(m_Not(m_Value(C2)), m_Value(A2))) && SelOp0) {
4078 SelectInst *NewSI = SelectInst::Create(C, B, A, "", nullptr, SelOp0);
4079 NewSI->swapProfMetadata();
4080 return NewSI;
4081 } else if (match(Op1, m_LogicalAnd(m_Specific(C), m_Value(B2))) &&
4082 SelOp1) {
4083 return SelectInst::Create(C, B, A, "", nullptr, SelOp1);
4084 } else {
4085 return createSelectInstWithUnknownProfile(C, B, A);
4086 }
4087 }
4088 return SelectInst::Create(C, B, A);
4089 }
4090
4091 return nullptr;
4092}
4093
4094// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
4095// here. We should standardize that construct where it is needed or choose some
4096// other way to ensure that commutated variants of patterns are not missed.
4098 if (Value *V = simplifyOrInst(I.getOperand(0), I.getOperand(1),
4099 SQ.getWithInstruction(&I)))
4100 return replaceInstUsesWith(I, V);
4101
4103 return &I;
4104
4106 return X;
4107
4109 return Phi;
4110
4111 // See if we can simplify any instructions used by the instruction whose sole
4112 // purpose is to compute bits we don't care about.
4114 return &I;
4115
4116 // Do this before using distributive laws to catch simple and/or/not patterns.
4118 return Xor;
4119
4121 return X;
4122
4124 return X;
4125
4126 // (A & B) | (C & D) -> A ^ D where A == ~C && B == ~D
4127 // (A & B) | (C & D) -> A ^ C where A == ~D && B == ~C
4128 if (Value *V = foldOrOfInversions(I, Builder))
4129 return replaceInstUsesWith(I, V);
4130
4131 // (A&B)|(A&C) -> A&(B|C) etc
4133 return replaceInstUsesWith(I, V);
4134
4135 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
4136 Type *Ty = I.getType();
4137 if (Ty->isIntOrIntVectorTy(1)) {
4138 if (auto *SI0 = dyn_cast<SelectInst>(Op0)) {
4139 if (auto *R =
4140 foldAndOrOfSelectUsingImpliedCond(Op1, *SI0, /* IsAnd */ false))
4141 return R;
4142 }
4143 if (auto *SI1 = dyn_cast<SelectInst>(Op1)) {
4144 if (auto *R =
4145 foldAndOrOfSelectUsingImpliedCond(Op0, *SI1, /* IsAnd */ false))
4146 return R;
4147 }
4148 }
4149
4150 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
4151 return FoldedLogic;
4152
4153 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
4154 return FoldedLogic;
4155
4156 if (Instruction *BitOp = matchBSwapOrBitReverse(I, /*MatchBSwaps*/ true,
4157 /*MatchBitReversals*/ true))
4158 return BitOp;
4159
4160 if (Instruction *Funnel = matchFunnelShift(I, *this))
4161 return Funnel;
4162
4164 return replaceInstUsesWith(I, Concat);
4165
4167 return R;
4168
4170 return R;
4171
4172 if (cast<PossiblyDisjointInst>(I).isDisjoint()) {
4173 if (Instruction *R =
4174 foldAddLikeCommutative(I.getOperand(0), I.getOperand(1),
4175 /*NSW=*/true, /*NUW=*/true))
4176 return R;
4177 if (Instruction *R =
4178 foldAddLikeCommutative(I.getOperand(1), I.getOperand(0),
4179 /*NSW=*/true, /*NUW=*/true))
4180 return R;
4181
4182 if (Value *Res = foldDisjointOr(I.getOperand(0), I.getOperand(1)))
4183 return replaceInstUsesWith(I, Res);
4184
4185 if (Value *Res = reassociateDisjointOr(I.getOperand(0), I.getOperand(1)))
4186 return replaceInstUsesWith(I, Res);
4187 }
4188
4189 Value *X, *Y;
4190 const APInt *CV;
4191 if (match(&I, m_c_Or(m_OneUse(m_Xor(m_Value(X), m_APInt(CV))), m_Value(Y))) &&
4192 !CV->isAllOnes() && MaskedValueIsZero(Y, *CV, &I)) {
4193 // (X ^ C) | Y -> (X | Y) ^ C iff Y & C == 0
4194 // The check for a 'not' op is for efficiency (if Y is known zero --> ~X).
4195 Value *Or = Builder.CreateOr(X, Y);
4196 return BinaryOperator::CreateXor(Or, ConstantInt::get(Ty, *CV));
4197 }
4198
4199 // If the operands have no common bits set:
4200 // or (mul X, Y), X --> add (mul X, Y), X --> mul X, (Y + 1)
4202 m_Deferred(X)))) {
4203 Value *IncrementY = Builder.CreateAdd(Y, ConstantInt::get(Ty, 1));
4204 return BinaryOperator::CreateMul(X, IncrementY);
4205 }
4206
4207 // Canonicalization to achieve lowering to Bit Manipulation Instructions (BMI)
4208 // ~X | (X-1) => ~(X & -X)
4209 Value *Op;
4212 Value *NegX = Builder.CreateNeg(Op);
4213 Value *And = Builder.CreateAnd(Op, NegX);
4215 }
4216
4217 // (C && A) || (C && B) => select C, A, B (and similar cases)
4218 //
4219 // Note: This is the same transformation used in `foldSelectOfBools`,
4220 // except that it's an `or` instead of `select`.
4221 if (I.getType()->isIntOrIntVectorTy(1) &&
4222 (Op0->hasOneUse() || Op1->hasOneUse())) {
4223 if (Instruction *V = FoldOrOfLogicalAnds(Op0, Op1)) {
4224 return V;
4225 }
4226 }
4227
4228 // (A & C) | (B & D)
4229 Value *A, *B, *C, *D;
4230 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4231 match(Op1, m_And(m_Value(B), m_Value(D)))) {
4232
4233 // (A & C0) | (B & C1)
4234 const APInt *C0, *C1;
4235 if (match(C, m_APInt(C0)) && match(D, m_APInt(C1))) {
4236 Value *X;
4237 if (*C0 == ~*C1) {
4238 // ((X | B) & MaskC) | (B & ~MaskC) -> (X & MaskC) | B
4239 if (match(A, m_c_Or(m_Value(X), m_Specific(B))))
4240 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C0), B);
4241 // (A & MaskC) | ((X | A) & ~MaskC) -> (X & ~MaskC) | A
4242 if (match(B, m_c_Or(m_Specific(A), m_Value(X))))
4243 return BinaryOperator::CreateOr(Builder.CreateAnd(X, *C1), A);
4244
4245 // ((X ^ B) & MaskC) | (B & ~MaskC) -> (X & MaskC) ^ B
4246 if (match(A, m_c_Xor(m_Value(X), m_Specific(B))))
4247 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C0), B);
4248 // (A & MaskC) | ((X ^ A) & ~MaskC) -> (X & ~MaskC) ^ A
4249 if (match(B, m_c_Xor(m_Specific(A), m_Value(X))))
4250 return BinaryOperator::CreateXor(Builder.CreateAnd(X, *C1), A);
4251 }
4252
4253 if ((*C0 & *C1).isZero()) {
4254 // ((X | B) & C0) | (B & C1) --> (X | B) & (C0 | C1)
4255 // iff (C0 & C1) == 0 and (X & ~C0) == 0
4256 if (match(A, m_c_Or(m_Value(X), m_Specific(B))) &&
4257 MaskedValueIsZero(X, ~*C0, &I)) {
4258 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4259 return BinaryOperator::CreateAnd(A, C01);
4260 }
4261 // (A & C0) | ((X | A) & C1) --> (X | A) & (C0 | C1)
4262 // iff (C0 & C1) == 0 and (X & ~C1) == 0
4263 if (match(B, m_c_Or(m_Value(X), m_Specific(A))) &&
4264 MaskedValueIsZero(X, ~*C1, &I)) {
4265 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4266 return BinaryOperator::CreateAnd(B, C01);
4267 }
4268 // ((X | C2) & C0) | ((X | C3) & C1) --> (X | C2 | C3) & (C0 | C1)
4269 // iff (C0 & C1) == 0 and (C2 & ~C0) == 0 and (C3 & ~C1) == 0.
4270 const APInt *C2, *C3;
4271 if (match(A, m_Or(m_Value(X), m_APInt(C2))) &&
4272 match(B, m_Or(m_Specific(X), m_APInt(C3))) &&
4273 (*C2 & ~*C0).isZero() && (*C3 & ~*C1).isZero()) {
4274 Value *Or = Builder.CreateOr(X, *C2 | *C3, "bitfield");
4275 Constant *C01 = ConstantInt::get(Ty, *C0 | *C1);
4276 return BinaryOperator::CreateAnd(Or, C01);
4277 }
4278 }
4279 }
4280
4281 // Don't try to form a select if it's unlikely that we'll get rid of at
4282 // least one of the operands. A select is generally more expensive than the
4283 // 'or' that it is replacing.
4284 if (Op0->hasOneUse() || Op1->hasOneUse()) {
4285 // (Cond & C) | (~Cond & D) -> Cond ? C : D, and commuted variants.
4286 if (Value *V = matchSelectFromAndOr(A, C, B, D))
4287 return replaceInstUsesWith(I, V);
4288 if (Value *V = matchSelectFromAndOr(A, C, D, B))
4289 return replaceInstUsesWith(I, V);
4290 if (Value *V = matchSelectFromAndOr(C, A, B, D))
4291 return replaceInstUsesWith(I, V);
4292 if (Value *V = matchSelectFromAndOr(C, A, D, B))
4293 return replaceInstUsesWith(I, V);
4294 if (Value *V = matchSelectFromAndOr(B, D, A, C))
4295 return replaceInstUsesWith(I, V);
4296 if (Value *V = matchSelectFromAndOr(B, D, C, A))
4297 return replaceInstUsesWith(I, V);
4298 if (Value *V = matchSelectFromAndOr(D, B, A, C))
4299 return replaceInstUsesWith(I, V);
4300 if (Value *V = matchSelectFromAndOr(D, B, C, A))
4301 return replaceInstUsesWith(I, V);
4302 }
4303 }
4304
4305 if (match(Op0, m_And(m_Value(A), m_Value(C))) &&
4306 match(Op1, m_Not(m_Or(m_Value(B), m_Value(D)))) &&
4307 (Op0->hasOneUse() || Op1->hasOneUse())) {
4308 // (Cond & C) | ~(Cond | D) -> Cond ? C : ~D
4309 if (Value *V = matchSelectFromAndOr(A, C, B, D, true))
4310 return replaceInstUsesWith(I, V);
4311 if (Value *V = matchSelectFromAndOr(A, C, D, B, true))
4312 return replaceInstUsesWith(I, V);
4313 if (Value *V = matchSelectFromAndOr(C, A, B, D, true))
4314 return replaceInstUsesWith(I, V);
4315 if (Value *V = matchSelectFromAndOr(C, A, D, B, true))
4316 return replaceInstUsesWith(I, V);
4317 }
4318
4319 // (A ^ B) | ((B ^ C) ^ A) -> (A ^ B) | C
4320 if (match(Op0, m_Xor(m_Value(A), m_Value(B))))
4321 if (match(Op1,
4324 return BinaryOperator::CreateOr(Op0, C);
4325
4326 // ((B ^ C) ^ A) | (A ^ B) -> (A ^ B) | C
4327 if (match(Op1, m_Xor(m_Value(A), m_Value(B))))
4328 if (match(Op0,
4331 return BinaryOperator::CreateOr(Op1, C);
4332
4333 if (Instruction *DeMorgan = matchDeMorgansLaws(I, *this))
4334 return DeMorgan;
4335
4336 // Canonicalize xor to the RHS.
4337 bool SwappedForXor = false;
4338 if (match(Op0, m_Xor(m_Value(), m_Value()))) {
4339 std::swap(Op0, Op1);
4340 SwappedForXor = true;
4341 }
4342
4343 if (match(Op1, m_Xor(m_Value(A), m_Value(B)))) {
4344 // (A | ?) | (A ^ B) --> (A | ?) | B
4345 // (B | ?) | (A ^ B) --> (B | ?) | A
4346 if (match(Op0, m_c_Or(m_Specific(A), m_Value())))
4347 return BinaryOperator::CreateOr(Op0, B);
4348 if (match(Op0, m_c_Or(m_Specific(B), m_Value())))
4349 return BinaryOperator::CreateOr(Op0, A);
4350
4351 // (A & B) | (A ^ B) --> A | B
4352 // (B & A) | (A ^ B) --> A | B
4353 if (match(Op0, m_c_And(m_Specific(A), m_Specific(B))))
4354 return BinaryOperator::CreateOr(A, B);
4355
4356 // ~A | (A ^ B) --> ~(A & B)
4357 // ~B | (A ^ B) --> ~(A & B)
4358 // The swap above should always make Op0 the 'not'.
4359 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4360 (match(Op0, m_Not(m_Specific(A))) || match(Op0, m_Not(m_Specific(B)))))
4361 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
4362
4363 // Same as above, but peek through an 'and' to the common operand:
4364 // ~(A & ?) | (A ^ B) --> ~((A & ?) & B)
4365 // ~(B & ?) | (A ^ B) --> ~((B & ?) & A)
4367 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4368 match(Op0,
4370 return BinaryOperator::CreateNot(Builder.CreateAnd(And, B));
4371 if ((Op0->hasOneUse() || Op1->hasOneUse()) &&
4372 match(Op0,
4374 return BinaryOperator::CreateNot(Builder.CreateAnd(And, A));
4375
4376 // (~A | C) | (A ^ B) --> ~(A & B) | C
4377 // (~B | C) | (A ^ B) --> ~(A & B) | C
4378 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4379 (match(Op0, m_c_Or(m_Not(m_Specific(A)), m_Value(C))) ||
4380 match(Op0, m_c_Or(m_Not(m_Specific(B)), m_Value(C))))) {
4381 Value *Nand = Builder.CreateNot(Builder.CreateAnd(A, B), "nand");
4382 return BinaryOperator::CreateOr(Nand, C);
4383 }
4384 }
4385
4386 if (SwappedForXor)
4387 std::swap(Op0, Op1);
4388
4389 if (Value *Res =
4390 foldBooleanAndOr(Op0, Op1, I, /*IsAnd=*/false, /*IsLogical=*/false))
4391 return replaceInstUsesWith(I, Res);
4392
4393 if (match(Op1, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4394 bool IsLogical = isa<SelectInst>(Op1);
4395 if (auto *V = reassociateBooleanAndOr(Op0, X, Y, I, /*IsAnd=*/false,
4396 /*RHSIsLogical=*/IsLogical))
4397 return replaceInstUsesWith(I, V);
4398 }
4399 if (match(Op0, m_OneUse(m_LogicalOr(m_Value(X), m_Value(Y))))) {
4400 bool IsLogical = isa<SelectInst>(Op0);
4401 if (auto *V = reassociateBooleanAndOr(Op1, X, Y, I, /*IsAnd=*/false,
4402 /*RHSIsLogical=*/IsLogical))
4403 return replaceInstUsesWith(I, V);
4404 }
4405
4406 if (Instruction *FoldedFCmps = reassociateFCmps(I, Builder))
4407 return FoldedFCmps;
4408
4409 if (Instruction *CastedOr = foldCastedBitwiseLogic(I))
4410 return CastedOr;
4411
4412 if (Instruction *Sel = foldBinopOfSextBoolToSelect(I))
4413 return Sel;
4414
4415 // or(sext(A), B) / or(B, sext(A)) --> A ? -1 : B, where A is i1 or <N x i1>.
4416 // TODO: Move this into foldBinopOfSextBoolToSelect as a more generalized fold
4417 // with binop identity constant. But creating a select with non-constant
4418 // arm may not be reversible due to poison semantics. Is that a good
4419 // canonicalization?
4420 if (match(&I, m_c_Or(m_OneUse(m_SExt(m_Value(A))), m_Value(B))) &&
4421 A->getType()->isIntOrIntVectorTy(1))
4422 return createSelectInstWithUnknownProfile(
4424
4425 // Note: If we've gotten to the point of visiting the outer OR, then the
4426 // inner one couldn't be simplified. If it was a constant, then it won't
4427 // be simplified by a later pass either, so we try swapping the inner/outer
4428 // ORs in the hopes that we'll be able to simplify it this way.
4429 // (X|C) | V --> (X|V) | C
4430 // Pass the disjoint flag in the following two patterns:
4431 // 1. or-disjoint (or-disjoint X, C), V -->
4432 // or-disjoint (or-disjoint X, V), C
4433 //
4434 // 2. or-disjoint (or X, C), V -->
4435 // or (or-disjoint X, V), C
4436 ConstantInt *CI;
4437 if (Op0->hasOneUse() && !match(Op1, m_ConstantInt()) &&
4438 match(Op0, m_Or(m_Value(A), m_ConstantInt(CI)))) {
4439 bool IsDisjointOuter = cast<PossiblyDisjointInst>(I).isDisjoint();
4440 bool IsDisjointInner = cast<PossiblyDisjointInst>(Op0)->isDisjoint();
4441 Value *Inner = Builder.CreateOr(A, Op1, "", /*IsDisjoint=*/IsDisjointOuter);
4442 Inner->takeName(Op0);
4443 return IsDisjointOuter && IsDisjointInner
4444 ? BinaryOperator::CreateDisjointOr(Inner, CI)
4445 : BinaryOperator::CreateOr(Inner, CI);
4446 }
4447
4448 // Change (or (bool?A:B),(bool?C:D)) --> (bool?(or A,C):(or B,D))
4449 // Since this OR statement hasn't been optimized further yet, we hope
4450 // that this transformation will allow the new ORs to be optimized.
4451 {
4452 Value *X = nullptr, *Y = nullptr;
4453 if (Op0->hasOneUse() && Op1->hasOneUse() &&
4454 match(Op0, m_Select(m_Value(X), m_Value(A), m_Value(B))) &&
4455 match(Op1, m_Select(m_Value(Y), m_Value(C), m_Value(D))) && X == Y) {
4456 Value *orTrue = Builder.CreateOr(A, C);
4457 Value *orFalse = Builder.CreateOr(B, D);
4458 return SelectInst::Create(X, orTrue, orFalse);
4459 }
4460 }
4461
4462 // or(ashr(subNSW(Y, X), ScalarSizeInBits(Y) - 1), X) --> X s> Y ? -1 : X.
4463 {
4464 Value *X, *Y;
4467 m_SpecificInt(Ty->getScalarSizeInBits() - 1))),
4468 m_Deferred(X)))) {
4469 Value *NewICmpInst = Builder.CreateICmpSGT(X, Y);
4471 return createSelectInstWithUnknownProfile(NewICmpInst, AllOnes, X);
4472 }
4473 }
4474
4475 {
4476 // ((A & B) ^ A) | ((A & B) ^ B) -> A ^ B
4477 // (A ^ (A & B)) | (B ^ (A & B)) -> A ^ B
4478 // ((A & B) ^ B) | ((A & B) ^ A) -> A ^ B
4479 // (B ^ (A & B)) | (A ^ (A & B)) -> A ^ B
4480 const auto TryXorOpt = [&](Value *Lhs, Value *Rhs) -> Instruction * {
4481 if (match(Lhs, m_c_Xor(m_And(m_Value(A), m_Value(B)), m_Deferred(A))) &&
4482 match(Rhs,
4484 return BinaryOperator::CreateXor(A, B);
4485 }
4486 return nullptr;
4487 };
4488
4489 if (Instruction *Result = TryXorOpt(Op0, Op1))
4490 return Result;
4491 if (Instruction *Result = TryXorOpt(Op1, Op0))
4492 return Result;
4493 }
4494
4495 if (Instruction *V =
4497 return V;
4498
4499 CmpPredicate Pred;
4500 Value *Mul, *Ov, *MulIsNotZero, *UMulWithOv;
4501 // Check if the OR weakens the overflow condition for umul.with.overflow by
4502 // treating any non-zero result as overflow. In that case, we overflow if both
4503 // umul.with.overflow operands are != 0, as in that case the result can only
4504 // be 0, iff the multiplication overflows.
4505 if (match(&I, m_c_Or(m_Value(Ov, m_ExtractValue<1>(m_Value(UMulWithOv))),
4506 m_Value(MulIsNotZero,
4510 m_Deferred(UMulWithOv))),
4511 m_ZeroInt())))) &&
4512 (Ov->hasOneUse() || (MulIsNotZero->hasOneUse() && Mul->hasOneUse()))) {
4513 Value *A, *B;
4515 m_Value(A), m_Value(B)))) {
4516 Value *NotNullA = Builder.CreateIsNotNull(A);
4517 Value *NotNullB = Builder.CreateIsNotNull(B);
4518 return BinaryOperator::CreateAnd(NotNullA, NotNullB);
4519 }
4520 }
4521
4522 /// Res, Overflow = xxx_with_overflow X, C1
4523 /// Try to canonicalize the pattern "Overflow | icmp pred Res, C2" into
4524 /// "Overflow | icmp pred X, C2 +/- C1".
4525 const WithOverflowInst *WO;
4526 const Value *WOV;
4527 const APInt *C1, *C2;
4529 m_Value(WOV, m_WithOverflowInst(WO)))),
4531 m_APInt(C2))))) &&
4532 (WO->getBinaryOp() == Instruction::Add ||
4533 WO->getBinaryOp() == Instruction::Sub) &&
4534 (ICmpInst::isEquality(Pred) ||
4535 WO->isSigned() == ICmpInst::isSigned(Pred)) &&
4536 match(WO->getRHS(), m_APInt(C1))) {
4537 bool Overflow;
4538 APInt NewC = WO->getBinaryOp() == Instruction::Add
4539 ? (ICmpInst::isSigned(Pred) ? C2->ssub_ov(*C1, Overflow)
4540 : C2->usub_ov(*C1, Overflow))
4541 : (ICmpInst::isSigned(Pred) ? C2->sadd_ov(*C1, Overflow)
4542 : C2->uadd_ov(*C1, Overflow));
4543 if (!Overflow || ICmpInst::isEquality(Pred)) {
4544 Value *NewCmp = Builder.CreateICmp(
4545 Pred, WO->getLHS(), ConstantInt::get(WO->getLHS()->getType(), NewC));
4546 return BinaryOperator::CreateOr(Ov, NewCmp);
4547 }
4548 }
4549
4550 // Try to fold the pattern "Overflow | icmp pred Res, C2" into a single
4551 // comparison instruction for umul.with.overflow.
4553 return replaceInstUsesWith(I, R);
4554
4555 // (~x) | y --> ~(x & (~y)) iff that gets rid of inversions
4557 return &I;
4558
4559 // Improve "get low bit mask up to and including bit X" pattern:
4560 // (1 << X) | ((1 << X) + -1) --> -1 l>> (bitwidth(x) - 1 - X)
4561 if (match(&I, m_c_Or(m_Add(m_Shl(m_One(), m_Value(X)), m_AllOnes()),
4562 m_Shl(m_One(), m_Deferred(X)))) &&
4563 match(&I, m_c_Or(m_OneUse(m_Value()), m_Value()))) {
4564 Value *Sub = Builder.CreateSub(
4565 ConstantInt::get(Ty, Ty->getScalarSizeInBits() - 1), X);
4566 return BinaryOperator::CreateLShr(Constant::getAllOnesValue(Ty), Sub);
4567 }
4568
4569 // An or recurrence w/loop invariant step is equivelent to (or start, step)
4570 PHINode *PN = nullptr;
4571 Value *Start = nullptr, *Step = nullptr;
4572 if (matchSimpleRecurrence(&I, PN, Start, Step) && DT.dominates(Step, PN))
4573 return replaceInstUsesWith(I, Builder.CreateOr(Start, Step));
4574
4575 // (A & B) | (C | D) or (C | D) | (A & B)
4576 // Can be combined if C or D is of type (A/B & X)
4578 m_OneUse(m_Or(m_Value(C), m_Value(D)))))) {
4579 // (A & B) | (C | ?) -> C | (? | (A & B))
4580 // (A & B) | (C | ?) -> C | (? | (A & B))
4581 // (A & B) | (C | ?) -> C | (? | (A & B))
4582 // (A & B) | (C | ?) -> C | (? | (A & B))
4583 // (C | ?) | (A & B) -> C | (? | (A & B))
4584 // (C | ?) | (A & B) -> C | (? | (A & B))
4585 // (C | ?) | (A & B) -> C | (? | (A & B))
4586 // (C | ?) | (A & B) -> C | (? | (A & B))
4587 if (match(D, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4589 return BinaryOperator::CreateOr(
4590 C, Builder.CreateOr(D, Builder.CreateAnd(A, B)));
4591 // (A & B) | (? | D) -> (? | (A & B)) | D
4592 // (A & B) | (? | D) -> (? | (A & B)) | D
4593 // (A & B) | (? | D) -> (? | (A & B)) | D
4594 // (A & B) | (? | D) -> (? | (A & B)) | D
4595 // (? | D) | (A & B) -> (? | (A & B)) | D
4596 // (? | D) | (A & B) -> (? | (A & B)) | D
4597 // (? | D) | (A & B) -> (? | (A & B)) | D
4598 // (? | D) | (A & B) -> (? | (A & B)) | D
4599 if (match(C, m_OneUse(m_c_And(m_Specific(A), m_Value()))) ||
4601 return BinaryOperator::CreateOr(
4602 Builder.CreateOr(C, Builder.CreateAnd(A, B)), D);
4603 }
4604
4606 return R;
4607
4608 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
4609 return Canonicalized;
4610
4611 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
4612 return Folded;
4613
4614 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
4615 return Res;
4616
4617 // If we are setting the sign bit of a floating-point value, convert
4618 // this to fneg(fabs), then cast back to integer.
4619 //
4620 // If the result isn't immediately cast back to a float, this will increase
4621 // the number of instructions. This is still probably a better canonical form
4622 // as it enables FP value tracking.
4623 //
4624 // Assumes any IEEE-represented type has the sign bit in the high bit.
4625 //
4626 // This is generous interpretation of noimplicitfloat, this is not a true
4627 // floating-point operation.
4628 Value *CastOp;
4629 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
4630 match(Op1, m_SignMask()) &&
4631 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
4632 Attribute::NoImplicitFloat)) {
4633 Type *EltTy = CastOp->getType()->getScalarType();
4634 if (EltTy->isFloatingPointTy() &&
4636 Value *FAbs = Builder.CreateFAbs(CastOp);
4637 Value *FNegFAbs = Builder.CreateFNeg(FAbs);
4638 return new BitCastInst(FNegFAbs, I.getType());
4639 }
4640 }
4641
4642 // (X & C1) | C2 -> X & (C1 | C2) iff (X & C2) == C2
4643 if (match(Op0, m_OneUse(m_And(m_Value(X), m_APInt(C1)))) &&
4644 match(Op1, m_APInt(C2))) {
4645 KnownBits KnownX = computeKnownBits(X, &I);
4646 if ((KnownX.One & *C2) == *C2)
4647 return BinaryOperator::CreateAnd(X, ConstantInt::get(Ty, *C1 | *C2));
4648 }
4649
4651 return Res;
4652
4653 if (Value *V =
4655 /*SimplifyOnly*/ false, *this))
4656 return BinaryOperator::CreateOr(V, Op1);
4657 if (Value *V =
4659 /*SimplifyOnly*/ false, *this))
4660 return BinaryOperator::CreateOr(Op0, V);
4661
4662 if (cast<PossiblyDisjointInst>(I).isDisjoint())
4664 return replaceInstUsesWith(I, V);
4665
4667 return replaceInstUsesWith(I, Res);
4668
4669 return nullptr;
4670}
4671
4672/// A ^ B can be specified using other logic ops in a variety of patterns. We
4673/// can fold these early and efficiently by morphing an existing instruction.
4675 InstCombiner::BuilderTy &Builder) {
4676 assert(I.getOpcode() == Instruction::Xor);
4677 Value *Op0 = I.getOperand(0);
4678 Value *Op1 = I.getOperand(1);
4679 Value *A, *B;
4680
4681 // There are 4 commuted variants for each of the basic patterns.
4682
4683 // (A & B) ^ (A | B) -> A ^ B
4684 // (A & B) ^ (B | A) -> A ^ B
4685 // (A | B) ^ (A & B) -> A ^ B
4686 // (A | B) ^ (B & A) -> A ^ B
4687 if (match(&I, m_c_Xor(m_And(m_Value(A), m_Value(B)),
4689 return BinaryOperator::CreateXor(A, B);
4690
4691 // (A | ~B) ^ (~A | B) -> A ^ B
4692 // (~B | A) ^ (~A | B) -> A ^ B
4693 // (~A | B) ^ (A | ~B) -> A ^ B
4694 // (B | ~A) ^ (A | ~B) -> A ^ B
4695 if (match(&I, m_Xor(m_c_Or(m_Value(A), m_Not(m_Value(B))),
4697 return BinaryOperator::CreateXor(A, B);
4698
4699 // (A & ~B) ^ (~A & B) -> A ^ B
4700 // (~B & A) ^ (~A & B) -> A ^ B
4701 // (~A & B) ^ (A & ~B) -> A ^ B
4702 // (B & ~A) ^ (A & ~B) -> A ^ B
4703 if (match(&I, m_Xor(m_c_And(m_Value(A), m_Not(m_Value(B))),
4705 return BinaryOperator::CreateXor(A, B);
4706
4707 // For the remaining cases we need to get rid of one of the operands.
4708 if (!Op0->hasOneUse() && !Op1->hasOneUse())
4709 return nullptr;
4710
4711 // (A | B) ^ ~(A & B) -> ~(A ^ B)
4712 // (A | B) ^ ~(B & A) -> ~(A ^ B)
4713 // (A & B) ^ ~(A | B) -> ~(A ^ B)
4714 // (A & B) ^ ~(B | A) -> ~(A ^ B)
4715 // Complexity sorting ensures the not will be on the right side.
4716 if ((match(Op0, m_Or(m_Value(A), m_Value(B))) &&
4717 match(Op1, m_Not(m_c_And(m_Specific(A), m_Specific(B))))) ||
4718 (match(Op0, m_And(m_Value(A), m_Value(B))) &&
4720 return BinaryOperator::CreateNot(Builder.CreateXor(A, B));
4721
4722 return nullptr;
4723}
4724
4725Value *InstCombinerImpl::foldXorOfICmps(ICmpInst *LHS, ICmpInst *RHS,
4726 BinaryOperator &I) {
4727 assert(I.getOpcode() == Instruction::Xor && I.getOperand(0) == LHS &&
4728 I.getOperand(1) == RHS && "Should be 'xor' with these operands");
4729
4730 ICmpInst::Predicate PredL = LHS->getPredicate(), PredR = RHS->getPredicate();
4731 Value *LHS0 = LHS->getOperand(0), *LHS1 = LHS->getOperand(1);
4732 Value *RHS0 = RHS->getOperand(0), *RHS1 = RHS->getOperand(1);
4733
4734 if (predicatesFoldable(PredL, PredR)) {
4735 if (LHS0 == RHS1 && LHS1 == RHS0) {
4736 std::swap(LHS0, LHS1);
4737 PredL = ICmpInst::getSwappedPredicate(PredL);
4738 }
4739 if (LHS0 == RHS0 && LHS1 == RHS1) {
4740 // (icmp1 A, B) ^ (icmp2 A, B) --> (icmp3 A, B)
4741 unsigned Code = getICmpCode(PredL) ^ getICmpCode(PredR);
4742 bool IsSigned = LHS->isSigned() || RHS->isSigned();
4743 return getNewICmpValue(Code, IsSigned, LHS0, LHS1, Builder);
4744 }
4745 }
4746
4747 const APInt *LC, *RC;
4748 if (match(LHS1, m_APInt(LC)) && match(RHS1, m_APInt(RC)) &&
4749 LHS0->getType() == RHS0->getType() &&
4750 LHS0->getType()->isIntOrIntVectorTy()) {
4751 // Convert xor of signbit tests to signbit test of xor'd values:
4752 // (X > -1) ^ (Y > -1) --> (X ^ Y) < 0
4753 // (X < 0) ^ (Y < 0) --> (X ^ Y) < 0
4754 // (X > -1) ^ (Y < 0) --> (X ^ Y) > -1
4755 // (X < 0) ^ (Y > -1) --> (X ^ Y) > -1
4756 bool TrueIfSignedL, TrueIfSignedR;
4757 if ((LHS->hasOneUse() || RHS->hasOneUse()) &&
4758 isSignBitCheck(PredL, *LC, TrueIfSignedL) &&
4759 isSignBitCheck(PredR, *RC, TrueIfSignedR)) {
4760 Value *XorLR = Builder.CreateXor(LHS0, RHS0);
4761 return TrueIfSignedL == TrueIfSignedR ? Builder.CreateIsNeg(XorLR) :
4762 Builder.CreateIsNotNeg(XorLR);
4763 }
4764
4765 // Fold (icmp pred1 X, C1) ^ (icmp pred2 X, C2)
4766 // into a single comparison using range-based reasoning.
4767 if (LHS0 == RHS0) {
4768 ConstantRange CR1 = ConstantRange::makeExactICmpRegion(PredL, *LC);
4769 ConstantRange CR2 = ConstantRange::makeExactICmpRegion(PredR, *RC);
4770 auto CRUnion = CR1.exactUnionWith(CR2);
4771 auto CRIntersect = CR1.exactIntersectWith(CR2);
4772 if (CRUnion && CRIntersect)
4773 if (auto CR = CRUnion->exactIntersectWith(CRIntersect->inverse())) {
4774 if (CR->isFullSet())
4775 return ConstantInt::getTrue(I.getType());
4776 if (CR->isEmptySet())
4777 return ConstantInt::getFalse(I.getType());
4778
4779 CmpInst::Predicate NewPred;
4780 APInt NewC, Offset;
4781 CR->getEquivalentICmp(NewPred, NewC, Offset);
4782
4783 if ((Offset.isZero() && (LHS->hasOneUse() || RHS->hasOneUse())) ||
4784 (LHS->hasOneUse() && RHS->hasOneUse())) {
4785 Value *NewV = LHS0;
4786 Type *Ty = LHS0->getType();
4787 if (!Offset.isZero())
4788 NewV = Builder.CreateAdd(NewV, ConstantInt::get(Ty, Offset));
4789 return Builder.CreateICmp(NewPred, NewV,
4790 ConstantInt::get(Ty, NewC));
4791 }
4792 }
4793 }
4794
4795 // Fold (icmp eq/ne (X & Pow2), 0) ^ (icmp eq/ne (Y & Pow2), 0) into
4796 // (icmp eq/ne ((X ^ Y) & Pow2), 0)
4797 Value *X, *Y, *Pow2;
4798 if (ICmpInst::isEquality(PredL) && ICmpInst::isEquality(PredR) &&
4799 LC->isZero() && RC->isZero() && LHS->hasOneUse() && RHS->hasOneUse() &&
4800 match(LHS0, m_And(m_Value(X), m_Value(Pow2))) &&
4801 match(RHS0, m_And(m_Value(Y), m_Specific(Pow2))) &&
4802 isKnownToBeAPowerOfTwo(Pow2, /*OrZero=*/true, &I)) {
4803 Value *Xor = Builder.CreateXor(X, Y);
4804 Value *And = Builder.CreateAnd(Xor, Pow2);
4805 return Builder.CreateICmp(PredL == PredR ? ICmpInst::ICMP_NE
4807 And, ConstantInt::getNullValue(Xor->getType()));
4808 }
4809 }
4810
4811 // Instead of trying to imitate the folds for and/or, decompose this 'xor'
4812 // into those logic ops. That is, try to turn this into an and-of-icmps
4813 // because we have many folds for that pattern.
4814 //
4815 // This is based on a truth table definition of xor:
4816 // X ^ Y --> (X | Y) & !(X & Y)
4817 if (Value *OrICmp = simplifyBinOp(Instruction::Or, LHS, RHS, SQ)) {
4818 // TODO: If OrICmp is true, then the definition of xor simplifies to !(X&Y).
4819 // TODO: If OrICmp is false, the whole thing is false (InstSimplify?).
4820 if (Value *AndICmp = simplifyBinOp(Instruction::And, LHS, RHS, SQ)) {
4821 // TODO: Independently handle cases where the 'and' side is a constant.
4822 ICmpInst *X = nullptr, *Y = nullptr;
4823 if (OrICmp == LHS && AndICmp == RHS) {
4824 // (LHS | RHS) & !(LHS & RHS) --> LHS & !RHS --> X & !Y
4825 X = LHS;
4826 Y = RHS;
4827 }
4828 if (OrICmp == RHS && AndICmp == LHS) {
4829 // !(LHS & RHS) & (LHS | RHS) --> !LHS & RHS --> !Y & X
4830 X = RHS;
4831 Y = LHS;
4832 }
4833 if (X && Y && (Y->hasOneUse() || canFreelyInvertAllUsersOf(Y, &I))) {
4834 // Invert the predicate of 'Y', thus inverting its output.
4835 Y->setPredicate(Y->getInversePredicate());
4836 // So, are there other uses of Y?
4837 if (!Y->hasOneUse()) {
4838 // We need to adapt other uses of Y though. Get a value that matches
4839 // the original value of Y before inversion. While this increases
4840 // immediate instruction count, we have just ensured that all the
4841 // users are freely-invertible, so that 'not' *will* get folded away.
4843 // Set insertion point to right after the Y.
4844 Builder.SetInsertPoint(Y->getParent(), ++(Y->getIterator()));
4845 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
4846 // Replace all uses of Y (excluding the one in NotY!) with NotY.
4847 Worklist.pushUsersToWorkList(*Y);
4848 Y->replaceUsesWithIf(NotY,
4849 [NotY](Use &U) { return U.getUser() != NotY; });
4850 }
4851 // All done.
4852 return Builder.CreateAnd(LHS, RHS);
4853 }
4854 }
4855 }
4856
4857 return nullptr;
4858}
4859
4860/// If we have a masked merge, in the canonical form of:
4861/// (assuming that A only has one use.)
4862/// | A | |B|
4863/// ((x ^ y) & M) ^ y
4864/// | D |
4865/// * If M is inverted:
4866/// | D |
4867/// ((x ^ y) & ~M) ^ y
4868/// We can canonicalize by swapping the final xor operand
4869/// to eliminate the 'not' of the mask.
4870/// ((x ^ y) & M) ^ x
4871/// * If M is a constant, and D has one use, we transform to 'and' / 'or' ops
4872/// because that shortens the dependency chain and improves analysis:
4873/// (x & M) | (y & ~M)
4875 InstCombiner::BuilderTy &Builder) {
4876 Value *B, *X, *D;
4877 Value *M;
4878 if (!match(&I, m_c_Xor(m_Value(B),
4881 m_Value(M))))))
4882 return nullptr;
4883
4884 Value *NotM;
4885 if (match(M, m_Not(m_Value(NotM)))) {
4886 // De-invert the mask and swap the value in B part.
4887 Value *NewA = Builder.CreateAnd(D, NotM);
4888 return BinaryOperator::CreateXor(NewA, X);
4889 }
4890
4891 Constant *C;
4892 if (D->hasOneUse() && match(M, m_Constant(C))) {
4893 // Propagating undef is unsafe. Clamp undef elements to -1.
4894 Type *EltTy = C->getType()->getScalarType();
4896 // Unfold.
4897 Value *LHS = Builder.CreateAnd(X, C);
4898 Value *NotC = Builder.CreateNot(C);
4899 Value *RHS = Builder.CreateAnd(B, NotC);
4900 return BinaryOperator::CreateOr(LHS, RHS);
4901 }
4902
4903 return nullptr;
4904}
4905
4907 InstCombiner::BuilderTy &Builder) {
4908 Value *X, *Y;
4909 // FIXME: one-use check is not needed in general, but currently we are unable
4910 // to fold 'not' into 'icmp', if that 'icmp' has multiple uses. (D35182)
4911 if (!match(&I, m_Not(m_OneUse(m_Xor(m_Value(X), m_Value(Y))))))
4912 return nullptr;
4913
4914 auto hasCommonOperand = [](Value *A, Value *B, Value *C, Value *D) {
4915 return A == C || A == D || B == C || B == D;
4916 };
4917
4918 Value *A, *B, *C, *D;
4919 // Canonicalize ~((A & B) ^ (A | ?)) -> (A & B) | ~(A | ?)
4920 // 4 commuted variants
4921 if (match(X, m_And(m_Value(A), m_Value(B))) &&
4922 match(Y, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4923 Value *NotY = Builder.CreateNot(Y);
4924 return BinaryOperator::CreateOr(X, NotY);
4925 };
4926
4927 // Canonicalize ~((A | ?) ^ (A & B)) -> (A & B) | ~(A | ?)
4928 // 4 commuted variants
4929 if (match(Y, m_And(m_Value(A), m_Value(B))) &&
4930 match(X, m_Or(m_Value(C), m_Value(D))) && hasCommonOperand(A, B, C, D)) {
4931 Value *NotX = Builder.CreateNot(X);
4932 return BinaryOperator::CreateOr(Y, NotX);
4933 };
4934
4935 return nullptr;
4936}
4937
4938/// Canonicalize a shifty way to code absolute value to the more common pattern
4939/// that uses negation and select.
4941 InstCombiner::BuilderTy &Builder) {
4942 assert(Xor.getOpcode() == Instruction::Xor && "Expected an xor instruction.");
4943
4944 // There are 4 potential commuted variants. Move the 'ashr' candidate to Op1.
4945 // We're relying on the fact that we only do this transform when the shift has
4946 // exactly 2 uses and the add has exactly 1 use (otherwise, we might increase
4947 // instructions).
4948 Value *Op0 = Xor.getOperand(0), *Op1 = Xor.getOperand(1);
4949 if (Op0->hasNUses(2))
4950 std::swap(Op0, Op1);
4951
4952 Type *Ty = Xor.getType();
4953 Value *A;
4954 const APInt *ShAmt;
4955 if (match(Op1, m_AShr(m_Value(A), m_APInt(ShAmt))) &&
4956 Op1->hasNUses(2) && *ShAmt == Ty->getScalarSizeInBits() - 1 &&
4957 match(Op0, m_OneUse(m_c_Add(m_Specific(A), m_Specific(Op1))))) {
4958 // Op1 = ashr i32 A, 31 ; smear the sign bit
4959 // xor (add A, Op1), Op1 ; add -1 and flip bits if negative
4960 // --> (A < 0) ? -A : A
4961 Value *IsNeg = Builder.CreateIsNeg(A);
4962 // Copy the nsw flags from the add to the negate.
4963 auto *Add = cast<BinaryOperator>(Op0);
4964 Value *NegA = Add->hasNoUnsignedWrap()
4965 ? Constant::getNullValue(A->getType())
4966 : Builder.CreateNeg(A, "", Add->hasNoSignedWrap());
4967 return SelectInst::Create(IsNeg, NegA, A);
4968 }
4969 return nullptr;
4970}
4971
4973 Instruction *IgnoredUser) {
4974 auto *I = dyn_cast<Instruction>(Op);
4975 return I && IC.isFreeToInvert(I, /*WillInvertAllUses=*/true) &&
4976 IC.canFreelyInvertAllUsersOf(I, IgnoredUser);
4977}
4978
4980 Instruction *IgnoredUser) {
4981 auto *I = cast<Instruction>(Op);
4982 IC.Builder.SetInsertPoint(*I->getInsertionPointAfterDef());
4983 Value *NotOp = IC.Builder.CreateNot(Op, Op->getName() + ".not");
4984 Op->replaceUsesWithIf(NotOp,
4985 [NotOp](Use &U) { return U.getUser() != NotOp; });
4986 IC.freelyInvertAllUsersOf(NotOp, IgnoredUser);
4987 return NotOp;
4988}
4989
4990// Transform
4991// z = ~(x &/| y)
4992// into:
4993// z = ((~x) |/& (~y))
4994// iff both x and y are free to invert and all uses of z can be freely updated.
4996 Value *Op0, *Op1;
4997 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
4998 return false;
4999
5000 // If this logic op has not been simplified yet, just bail out and let that
5001 // happen first. Otherwise, the code below may wrongly invert.
5002 if (Op0 == Op1)
5003 return false;
5004
5005 // If one of the operands is a user of the other,
5006 // freelyInvert->freelyInvertAllUsersOf will change the operands of I, which
5007 // may cause miscompilation.
5008 if (match(Op0, m_Not(m_Specific(Op1))) || match(Op1, m_Not(m_Specific(Op0))))
5009 return false;
5010
5011 Instruction::BinaryOps NewOpc =
5012 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
5013 bool IsBinaryOp = isa<BinaryOperator>(I);
5014
5015 // Can our users be adapted?
5016 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
5017 return false;
5018
5019 // And can the operands be adapted?
5020 if (!canFreelyInvert(*this, Op0, &I) || !canFreelyInvert(*this, Op1, &I))
5021 return false;
5022
5023 Op0 = freelyInvert(*this, Op0, &I);
5024 Op1 = freelyInvert(*this, Op1, &I);
5025
5026 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
5027 Value *NewLogicOp;
5028 if (IsBinaryOp) {
5029 NewLogicOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
5030 } else {
5031 NewLogicOp =
5032 Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not",
5033 ProfcheckDisableMetadataFixes ? nullptr : &I);
5034 if (SelectInst *SI = dyn_cast<SelectInst>(NewLogicOp))
5035 SI->swapProfMetadata();
5036 }
5037
5038 replaceInstUsesWith(I, NewLogicOp);
5039 // We can not just create an outer `not`, it will most likely be immediately
5040 // folded back, reconstructing our initial pattern, and causing an
5041 // infinite combine loop, so immediately manually fold it away.
5042 freelyInvertAllUsersOf(NewLogicOp);
5043 return true;
5044}
5045
5046// Transform
5047// z = (~x) &/| y
5048// into:
5049// z = ~(x |/& (~y))
5050// iff y is free to invert and all uses of z can be freely updated.
5052 Value *Op0, *Op1;
5053 if (!match(&I, m_LogicalOp(m_Value(Op0), m_Value(Op1))))
5054 return false;
5055 Instruction::BinaryOps NewOpc =
5056 match(&I, m_LogicalAnd()) ? Instruction::Or : Instruction::And;
5057 bool IsBinaryOp = isa<BinaryOperator>(I);
5058
5059 Value *NotOp0 = nullptr;
5060 Value *NotOp1 = nullptr;
5061 Value **OpToInvert = nullptr;
5062 if (match(Op0, m_Not(m_Value(NotOp0))) && canFreelyInvert(*this, Op1, &I)) {
5063 Op0 = NotOp0;
5064 OpToInvert = &Op1;
5065 } else if (match(Op1, m_Not(m_Value(NotOp1))) &&
5066 canFreelyInvert(*this, Op0, &I)) {
5067 Op1 = NotOp1;
5068 OpToInvert = &Op0;
5069 } else
5070 return false;
5071
5072 // And can our users be adapted?
5073 if (!InstCombiner::canFreelyInvertAllUsersOf(&I, /*IgnoredUser=*/nullptr))
5074 return false;
5075
5076 *OpToInvert = freelyInvert(*this, *OpToInvert, &I);
5077
5078 Builder.SetInsertPoint(*I.getInsertionPointAfterDef());
5079 Value *NewBinOp;
5080 if (IsBinaryOp)
5081 NewBinOp = Builder.CreateBinOp(NewOpc, Op0, Op1, I.getName() + ".not");
5082 else
5083 NewBinOp = Builder.CreateLogicalOp(NewOpc, Op0, Op1, I.getName() + ".not");
5084 replaceInstUsesWith(I, NewBinOp);
5085 // We can not just create an outer `not`, it will most likely be immediately
5086 // folded back, reconstructing our initial pattern, and causing an
5087 // infinite combine loop, so immediately manually fold it away.
5088 freelyInvertAllUsersOf(NewBinOp);
5089 return true;
5090}
5091
5092Instruction *InstCombinerImpl::foldNot(BinaryOperator &I) {
5093 Value *NotOp;
5094 if (!match(&I, m_Not(m_Value(NotOp))))
5095 return nullptr;
5096
5097 // Apply DeMorgan's Law for 'nand' / 'nor' logic with an inverted operand.
5098 // We must eliminate the and/or (one-use) for these transforms to not increase
5099 // the instruction count.
5100 //
5101 // ~(~X & Y) --> (X | ~Y)
5102 // ~(Y & ~X) --> (X | ~Y)
5103 //
5104 // Note: The logical matches do not check for the commuted patterns because
5105 // those are handled via SimplifySelectsFeedingBinaryOp().
5106 Type *Ty = I.getType();
5107 Value *X, *Y;
5108 if (match(NotOp, m_OneUse(m_c_And(m_Not(m_Value(X)), m_Value(Y))))) {
5109 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5110 return BinaryOperator::CreateOr(X, NotY);
5111 }
5112 if (match(NotOp, m_OneUse(m_LogicalAnd(m_Not(m_Value(X)), m_Value(Y))))) {
5113 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5115 X, ConstantInt::getTrue(Ty), NotY, "", nullptr,
5117 SI->swapProfMetadata();
5118 return SI;
5119 }
5120
5121 // ~(~X | Y) --> (X & ~Y)
5122 // ~(Y | ~X) --> (X & ~Y)
5123 if (match(NotOp, m_OneUse(m_c_Or(m_Not(m_Value(X)), m_Value(Y))))) {
5124 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5125 return BinaryOperator::CreateAnd(X, NotY);
5126 }
5127 if (match(NotOp, m_OneUse(m_LogicalOr(m_Not(m_Value(X)), m_Value(Y))))) {
5128 Value *NotY = Builder.CreateNot(Y, Y->getName() + ".not");
5129 SelectInst *SI = SelectInst::Create(
5130 X, NotY, ConstantInt::getFalse(Ty), "", nullptr,
5132 SI->swapProfMetadata();
5133 return SI;
5134 }
5135
5136 // Is this a 'not' (~) fed by a binary operator?
5137 BinaryOperator *NotVal;
5138 if (match(NotOp, m_BinOp(NotVal))) {
5139 // ~((-X) | Y) --> (X - 1) & (~Y)
5140 if (match(NotVal,
5142 Value *DecX = Builder.CreateAdd(X, ConstantInt::getAllOnesValue(Ty));
5143 Value *NotY = Builder.CreateNot(Y);
5144 return BinaryOperator::CreateAnd(DecX, NotY);
5145 }
5146
5147 // ~(~X >>s Y) --> (X >>s Y)
5148 if (match(NotVal, m_AShr(m_Not(m_Value(X)), m_Value(Y))))
5149 return BinaryOperator::CreateAShr(X, Y);
5150
5151 // Treat lshr with non-negative operand as ashr.
5152 // ~(~X >>u Y) --> (X >>s Y) iff X is known negative
5153 if (match(NotVal, m_LShr(m_Not(m_Value(X)), m_Value(Y))) &&
5154 isKnownNegative(X, SQ.getWithInstruction(NotVal)))
5155 return BinaryOperator::CreateAShr(X, Y);
5156
5157 // Bit-hack form of a signbit test for iN type:
5158 // ~(X >>s (N - 1)) --> sext i1 (X > -1) to iN
5159 unsigned FullShift = Ty->getScalarSizeInBits() - 1;
5160 if (match(NotVal, m_OneUse(m_AShr(m_Value(X), m_SpecificInt(FullShift))))) {
5161 Value *IsNotNeg = Builder.CreateIsNotNeg(X, "isnotneg");
5162 return new SExtInst(IsNotNeg, Ty);
5163 }
5164
5165 // If we are inverting a right-shifted constant, we may be able to eliminate
5166 // the 'not' by inverting the constant and using the opposite shift type.
5167 // Canonicalization rules ensure that only a negative constant uses 'ashr',
5168 // but we must check that in case that transform has not fired yet.
5169
5170 // ~(C >>s Y) --> ~C >>u Y (when inverting the replicated sign bits)
5171 Constant *C;
5172 if (match(NotVal, m_AShr(m_Constant(C), m_Value(Y))) &&
5173 match(C, m_Negative()))
5174 return BinaryOperator::CreateLShr(ConstantExpr::getNot(C), Y);
5175
5176 // ~(C >>u Y) --> ~C >>s Y (when inverting the replicated sign bits)
5177 if (match(NotVal, m_LShr(m_Constant(C), m_Value(Y))) &&
5178 match(C, m_NonNegative()))
5179 return BinaryOperator::CreateAShr(ConstantExpr::getNot(C), Y);
5180
5181 // ~(X + C) --> ~C - X
5182 if (match(NotVal, m_Add(m_Value(X), m_ImmConstant(C))))
5183 return BinaryOperator::CreateSub(ConstantExpr::getNot(C), X);
5184
5185 // ~(X - Y) --> ~X + Y
5186 // FIXME: is it really beneficial to sink the `not` here?
5187 if (match(NotVal, m_Sub(m_Value(X), m_Value(Y))))
5188 if (isa<Constant>(X) || NotVal->hasOneUse())
5189 return BinaryOperator::CreateAdd(Builder.CreateNot(X), Y);
5190
5191 // ~(~X + Y) --> X - Y
5192 if (match(NotVal, m_c_Add(m_Not(m_Value(X)), m_Value(Y))))
5193 return BinaryOperator::CreateWithCopiedFlags(Instruction::Sub, X, Y,
5194 NotVal);
5195 }
5196
5197 // not (cmp A, B) = !cmp A, B
5198 CmpPredicate Pred;
5199 if (match(NotOp, m_Cmp(Pred, m_Value(), m_Value())) &&
5200 (NotOp->hasOneUse() ||
5202 /*IgnoredUser=*/nullptr))) {
5203 cast<CmpInst>(NotOp)->setPredicate(CmpInst::getInversePredicate(Pred));
5205 return &I;
5206 }
5207
5208 // not (bitcast (cmp A, B) --> bitcast (!cmp A, B)
5209 if (match(NotOp, m_OneUse(m_BitCast(m_Value(X)))) &&
5210 match(X, m_OneUse(m_Cmp(Pred, m_Value(), m_Value())))) {
5211 cast<CmpInst>(X)->setPredicate(CmpInst::getInversePredicate(Pred));
5212 return new BitCastInst(X, Ty);
5213 }
5214
5215 // Move a 'not' ahead of casts of a bool to enable logic reduction:
5216 // not (bitcast (sext i1 X)) --> bitcast (sext (not i1 X))
5217 if (match(NotOp, m_OneUse(m_BitCast(m_OneUse(m_SExt(m_Value(X)))))) &&
5218 X->getType()->isIntOrIntVectorTy(1)) {
5219 Type *SextTy = cast<BitCastOperator>(NotOp)->getSrcTy();
5220 Value *NotX = Builder.CreateNot(X);
5221 Value *Sext = Builder.CreateSExt(NotX, SextTy);
5222 return new BitCastInst(Sext, Ty);
5223 }
5224
5225 if (auto *NotOpI = dyn_cast<Instruction>(NotOp))
5226 if (sinkNotIntoLogicalOp(*NotOpI))
5227 return &I;
5228
5229 // Eliminate a bitwise 'not' op of 'not' min/max by inverting the min/max:
5230 // ~min(~X, ~Y) --> max(X, Y)
5231 // ~max(~X, Y) --> min(X, ~Y)
5232 auto *II = dyn_cast<IntrinsicInst>(NotOp);
5233 if (II && II->hasOneUse()) {
5234 if (match(NotOp, m_c_MaxOrMin(m_Not(m_Value(X)), m_Value(Y)))) {
5235 Intrinsic::ID InvID = getInverseMinMaxIntrinsic(II->getIntrinsicID());
5236 Value *NotY = Builder.CreateNot(Y);
5237 Value *InvMaxMin = Builder.CreateBinaryIntrinsic(InvID, X, NotY);
5238 return replaceInstUsesWith(I, InvMaxMin);
5239 }
5240
5241 if (II->getIntrinsicID() == Intrinsic::is_fpclass) {
5242 ConstantInt *ClassMask = cast<ConstantInt>(II->getArgOperand(1));
5243 II->setArgOperand(
5244 1, ConstantInt::get(ClassMask->getType(),
5245 ~ClassMask->getZExtValue() & fcAllFlags));
5246 return replaceInstUsesWith(I, II);
5247 }
5248 }
5249
5250 if (NotOp->hasOneUse()) {
5251 // Pull 'not' into operands of select if both operands are one-use compares
5252 // or one is one-use compare and the other one is a constant.
5253 // Inverting the predicates eliminates the 'not' operation.
5254 // Example:
5255 // not (select ?, (cmp TPred, ?, ?), (cmp FPred, ?, ?) -->
5256 // select ?, (cmp InvTPred, ?, ?), (cmp InvFPred, ?, ?)
5257 // not (select ?, (cmp TPred, ?, ?), true -->
5258 // select ?, (cmp InvTPred, ?, ?), false
5259 if (auto *Sel = dyn_cast<SelectInst>(NotOp)) {
5260 Value *TV = Sel->getTrueValue();
5261 Value *FV = Sel->getFalseValue();
5262 auto *CmpT = dyn_cast<CmpInst>(TV);
5263 auto *CmpF = dyn_cast<CmpInst>(FV);
5264 bool InvertibleT = (CmpT && CmpT->hasOneUse()) || isa<Constant>(TV);
5265 bool InvertibleF = (CmpF && CmpF->hasOneUse()) || isa<Constant>(FV);
5266 if (InvertibleT && InvertibleF) {
5267 if (CmpT)
5268 CmpT->setPredicate(CmpT->getInversePredicate());
5269 else
5270 Sel->setTrueValue(ConstantExpr::getNot(cast<Constant>(TV)));
5271 if (CmpF)
5272 CmpF->setPredicate(CmpF->getInversePredicate());
5273 else
5274 Sel->setFalseValue(ConstantExpr::getNot(cast<Constant>(FV)));
5275 return replaceInstUsesWith(I, Sel);
5276 }
5277 }
5278 }
5279
5280 if (Instruction *NewXor = foldNotXor(I, Builder))
5281 return NewXor;
5282
5283 // TODO: Could handle multi-use better by checking if all uses of NotOp (other
5284 // than I) can be inverted.
5285 if (Value *R = getFreelyInverted(NotOp, NotOp->hasOneUse(), &Builder))
5286 return replaceInstUsesWith(I, R);
5287
5288 return nullptr;
5289}
5290
5291// ((X + C) & M) ^ M --> (~C − X) & M
5293 InstCombiner::BuilderTy &Builder) {
5294 Value *X, *Mask;
5295 Constant *AddC;
5296 BinaryOperator *AddInst;
5297 if (match(&I,
5299 m_BinOp(AddInst),
5300 m_Add(m_Value(X), m_ImmConstant(AddC)))),
5301 m_Value(Mask))),
5302 m_Deferred(Mask)))) {
5303 Value *NotC = Builder.CreateNot(AddC);
5304 Value *NewSub = Builder.CreateSub(NotC, X, "", AddInst->hasNoUnsignedWrap(),
5305 AddInst->hasNoSignedWrap());
5306 return BinaryOperator::CreateAnd(NewSub, Mask);
5307 }
5308
5309 return nullptr;
5310}
5311
5312// FIXME: We use commutative matchers (m_c_*) for some, but not all, matches
5313// here. We should standardize that construct where it is needed or choose some
5314// other way to ensure that commutated variants of patterns are not missed.
5316 if (Value *V = simplifyXorInst(I.getOperand(0), I.getOperand(1),
5317 SQ.getWithInstruction(&I)))
5318 return replaceInstUsesWith(I, V);
5319
5321 return &I;
5322
5324 return X;
5325
5327 return Phi;
5328
5329 if (Instruction *NewXor = foldXorToXor(I, Builder))
5330 return NewXor;
5331
5332 // (A&B)^(A&C) -> A&(B^C) etc
5334 return replaceInstUsesWith(I, V);
5335
5336 // See if we can simplify any instructions used by the instruction whose sole
5337 // purpose is to compute bits we don't care about.
5339 return &I;
5340
5341 if (Instruction *R = foldNot(I))
5342 return R;
5343
5345 return R;
5346
5347 Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
5348 Value *X, *Y, *M;
5349
5350 // (X | Y) ^ M -> (X ^ M) ^ Y
5351 // (X | Y) ^ M -> (Y ^ M) ^ X
5353 m_Value(M)))) {
5354 if (Value *XorAC = simplifyXorInst(X, M, SQ.getWithInstruction(&I)))
5355 return BinaryOperator::CreateXor(XorAC, Y);
5356
5357 if (Value *XorBC = simplifyXorInst(Y, M, SQ.getWithInstruction(&I)))
5358 return BinaryOperator::CreateXor(XorBC, X);
5359 }
5360
5361 // Fold (X & M) ^ (Y & ~M) -> (X & M) | (Y & ~M)
5362 // This it a special case in haveNoCommonBitsSet, but the computeKnownBits
5363 // calls in there are unnecessary as SimplifyDemandedInstructionBits should
5364 // have already taken care of those cases.
5365 if (match(&I, m_c_Xor(m_c_And(m_Not(m_Value(M)), m_Value()),
5366 m_c_And(m_Deferred(M), m_Value())))) {
5368 return BinaryOperator::CreateDisjointOr(Op0, Op1);
5369 else
5370 return BinaryOperator::CreateOr(Op0, Op1);
5371 }
5372
5374 return Xor;
5375
5376 Constant *C1;
5377 if (match(Op1, m_Constant(C1))) {
5378 Constant *C2;
5379
5380 if (match(Op0, m_OneUse(m_Or(m_Value(X), m_ImmConstant(C2)))) &&
5381 match(C1, m_ImmConstant())) {
5382 // (X | C2) ^ C1 --> (X & ~C2) ^ (C1^C2)
5385 Value *And = Builder.CreateAnd(
5387 return BinaryOperator::CreateXor(
5389 }
5390
5391 // Use DeMorgan and reassociation to eliminate a 'not' op.
5392 if (match(Op0, m_OneUse(m_Or(m_Not(m_Value(X)), m_Constant(C2))))) {
5393 // (~X | C2) ^ C1 --> ((X & ~C2) ^ -1) ^ C1 --> (X & ~C2) ^ ~C1
5394 Value *And = Builder.CreateAnd(X, ConstantExpr::getNot(C2));
5395 return BinaryOperator::CreateXor(And, ConstantExpr::getNot(C1));
5396 }
5397 if (match(Op0, m_OneUse(m_And(m_Not(m_Value(X)), m_Constant(C2))))) {
5398 // (~X & C2) ^ C1 --> ((X | ~C2) ^ -1) ^ C1 --> (X | ~C2) ^ ~C1
5399 Value *Or = Builder.CreateOr(X, ConstantExpr::getNot(C2));
5400 return BinaryOperator::CreateXor(Or, ConstantExpr::getNot(C1));
5401 }
5402
5403 // Convert xor ([trunc] (ashr X, BW-1)), C =>
5404 // select(X >s -1, C, ~C)
5405 // The ashr creates "AllZeroOrAllOne's", which then optionally inverses the
5406 // constant depending on whether this input is less than 0.
5407 const APInt *CA;
5408 if (match(Op0, m_OneUse(m_TruncOrSelf(
5409 m_AShr(m_Value(X), m_APIntAllowPoison(CA))))) &&
5410 *CA == X->getType()->getScalarSizeInBits() - 1 &&
5411 !match(C1, m_AllOnes())) {
5412 assert(!C1->isNullValue() && "Unexpected xor with 0");
5413 Value *IsNotNeg = Builder.CreateIsNotNeg(X);
5414 return createSelectInstWithUnknownProfile(IsNotNeg, Op1,
5415 Builder.CreateNot(Op1));
5416 }
5417 }
5418
5419 Type *Ty = I.getType();
5420 {
5421 const APInt *RHSC;
5422 if (match(Op1, m_APInt(RHSC))) {
5423 Value *X;
5424 const APInt *C;
5425 // (C - X) ^ signmaskC --> (C + signmaskC) - X
5426 if (RHSC->isSignMask() && match(Op0, m_Sub(m_APInt(C), m_Value(X))))
5427 return BinaryOperator::CreateSub(ConstantInt::get(Ty, *C + *RHSC), X);
5428
5429 // (X + C) ^ signmaskC --> X + (C + signmaskC)
5430 if (RHSC->isSignMask() && match(Op0, m_Add(m_Value(X), m_APInt(C))))
5431 return BinaryOperator::CreateAdd(X, ConstantInt::get(Ty, *C + *RHSC));
5432
5433 // (X | C) ^ RHSC --> X ^ (C ^ RHSC) iff X & C == 0
5434 if (match(Op0, m_Or(m_Value(X), m_APInt(C))) &&
5435 MaskedValueIsZero(X, *C, &I))
5436 return BinaryOperator::CreateXor(X, ConstantInt::get(Ty, *C ^ *RHSC));
5437
5438 // When X is a power-of-two or zero and zero input is poison:
5439 // ctlz(i32 X) ^ 31 --> cttz(X)
5440 // cttz(i32 X) ^ 31 --> ctlz(X)
5441 auto *II = dyn_cast<IntrinsicInst>(Op0);
5442 if (II && II->hasOneUse() && *RHSC == Ty->getScalarSizeInBits() - 1) {
5443 Intrinsic::ID IID = II->getIntrinsicID();
5444 if ((IID == Intrinsic::ctlz || IID == Intrinsic::cttz) &&
5445 match(II->getArgOperand(1), m_One()) &&
5446 isKnownToBeAPowerOfTwo(II->getArgOperand(0), /*OrZero */ true)) {
5447 IID = (IID == Intrinsic::ctlz) ? Intrinsic::cttz : Intrinsic::ctlz;
5448 Function *F =
5449 Intrinsic::getOrInsertDeclaration(II->getModule(), IID, Ty);
5450 return CallInst::Create(F, {II->getArgOperand(0), Builder.getTrue()});
5451 }
5452 }
5453
5454 // If RHSC is inverting the remaining bits of shifted X,
5455 // canonicalize to a 'not' before the shift to help SCEV and codegen:
5456 // (X << C) ^ RHSC --> ~X << C
5457 if (match(Op0, m_OneUse(m_Shl(m_Value(X), m_APInt(C)))) &&
5458 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).shl(*C)) {
5459 Value *NotX = Builder.CreateNot(X);
5460 return BinaryOperator::CreateShl(NotX, ConstantInt::get(Ty, *C));
5461 }
5462 // (X >>u C) ^ RHSC --> ~X >>u C
5463 if (match(Op0, m_OneUse(m_LShr(m_Value(X), m_APInt(C)))) &&
5464 *RHSC == APInt::getAllOnes(Ty->getScalarSizeInBits()).lshr(*C)) {
5465 Value *NotX = Builder.CreateNot(X);
5466 return BinaryOperator::CreateLShr(NotX, ConstantInt::get(Ty, *C));
5467 }
5468 // TODO: We could handle 'ashr' here as well. That would be matching
5469 // a 'not' op and moving it before the shift. Doing that requires
5470 // preventing the inverse fold in canShiftBinOpWithConstantRHS().
5471 }
5472
5473 // If we are XORing the sign bit of a floating-point value, convert
5474 // this to fneg, then cast back to integer.
5475 //
5476 // This is generous interpretation of noimplicitfloat, this is not a true
5477 // floating-point operation.
5478 //
5479 // Assumes any IEEE-represented type has the sign bit in the high bit.
5480 // TODO: Unify with APInt matcher. This version allows undef unlike m_APInt
5481 Value *CastOp;
5482 if (match(Op0, m_ElementWiseBitCast(m_Value(CastOp))) &&
5483 match(Op1, m_SignMask()) &&
5484 !Builder.GetInsertBlock()->getParent()->hasFnAttribute(
5485 Attribute::NoImplicitFloat)) {
5486 Type *EltTy = CastOp->getType()->getScalarType();
5487 if (EltTy->isFloatingPointTy() &&
5489 Value *FNeg = Builder.CreateFNeg(CastOp);
5490 return new BitCastInst(FNeg, I.getType());
5491 }
5492 }
5493 }
5494
5495 // FIXME: This should not be limited to scalar (pull into APInt match above).
5496 {
5497 Value *X;
5498 ConstantInt *C1, *C2, *C3;
5499 // ((X^C1) >> C2) ^ C3 -> (X>>C2) ^ ((C1>>C2)^C3)
5500 if (match(Op1, m_ConstantInt(C3)) &&
5502 m_ConstantInt(C2))) &&
5503 Op0->hasOneUse()) {
5504 // fold (C1 >> C2) ^ C3
5505 APInt FoldConst = C1->getValue().lshr(C2->getValue());
5506 FoldConst ^= C3->getValue();
5507 // Prepare the two operands.
5508 auto *Opnd0 = Builder.CreateLShr(X, C2);
5509 Opnd0->takeName(Op0);
5510 return BinaryOperator::CreateXor(Opnd0, ConstantInt::get(Ty, FoldConst));
5511 }
5512 }
5513
5514 if (Instruction *FoldedLogic = foldBinOpIntoSelectOrPhi(I))
5515 return FoldedLogic;
5516
5517 if (Instruction *FoldedLogic = foldBinOpSelectBinOp(I))
5518 return FoldedLogic;
5519
5520 // Y ^ (X | Y) --> X & ~Y
5521 // Y ^ (Y | X) --> X & ~Y
5522 if (match(Op1, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op0)))))
5523 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op0));
5524 // (X | Y) ^ Y --> X & ~Y
5525 // (Y | X) ^ Y --> X & ~Y
5526 if (match(Op0, m_OneUse(m_c_Or(m_Value(X), m_Specific(Op1)))))
5527 return BinaryOperator::CreateAnd(X, Builder.CreateNot(Op1));
5528
5529 // Y ^ (X & Y) --> ~X & Y
5530 // Y ^ (Y & X) --> ~X & Y
5531 if (match(Op1, m_OneUse(m_c_And(m_Value(X), m_Specific(Op0)))))
5532 return BinaryOperator::CreateAnd(Op0, Builder.CreateNot(X));
5533 // (X & Y) ^ Y --> ~X & Y
5534 // (Y & X) ^ Y --> ~X & Y
5535 // Canonical form is (X & C) ^ C; don't touch that.
5536 // TODO: A 'not' op is better for analysis and codegen, but demanded bits must
5537 // be fixed to prefer that (otherwise we get infinite looping).
5538 if (!match(Op1, m_Constant()) &&
5539 match(Op0, m_OneUse(m_c_And(m_Value(X), m_Specific(Op1)))))
5540 return BinaryOperator::CreateAnd(Op1, Builder.CreateNot(X));
5541
5542 Value *A, *B, *C;
5543 // (A ^ B) ^ (A | C) --> (~A & C) ^ B -- There are 4 commuted variants.
5546 return BinaryOperator::CreateXor(
5547 Builder.CreateAnd(Builder.CreateNot(A), C), B);
5548
5549 // (A ^ B) ^ (B | C) --> (~B & C) ^ A -- There are 4 commuted variants.
5552 return BinaryOperator::CreateXor(
5553 Builder.CreateAnd(Builder.CreateNot(B), C), A);
5554
5555 // (A & B) ^ (A ^ B) -> (A | B)
5556 if (match(Op0, m_And(m_Value(A), m_Value(B))) &&
5558 return BinaryOperator::CreateOr(A, B);
5559 // (A ^ B) ^ (A & B) -> (A | B)
5560 if (match(Op0, m_Xor(m_Value(A), m_Value(B))) &&
5562 return BinaryOperator::CreateOr(A, B);
5563
5564 // (A & ~B) ^ ~A -> ~(A & B)
5565 // (~B & A) ^ ~A -> ~(A & B)
5566 if (match(Op0, m_c_And(m_Value(A), m_Not(m_Value(B)))) &&
5567 match(Op1, m_Not(m_Specific(A))))
5568 return BinaryOperator::CreateNot(Builder.CreateAnd(A, B));
5569
5570 // (~A & B) ^ A --> A | B -- There are 4 commuted variants.
5572 return BinaryOperator::CreateOr(A, B);
5573
5574 // (~A | B) ^ A --> ~(A & B)
5575 if (match(Op0, m_OneUse(m_c_Or(m_Not(m_Specific(Op1)), m_Value(B)))))
5576 return BinaryOperator::CreateNot(Builder.CreateAnd(Op1, B));
5577
5578 // A ^ (~A | B) --> ~(A & B)
5579 if (match(Op1, m_OneUse(m_c_Or(m_Not(m_Specific(Op0)), m_Value(B)))))
5580 return BinaryOperator::CreateNot(Builder.CreateAnd(Op0, B));
5581
5582 // (A | B) ^ (A | C) --> (B ^ C) & ~A -- There are 4 commuted variants.
5583 // TODO: Loosen one-use restriction if common operand is a constant.
5584 Value *D;
5585 if (match(Op0, m_OneUse(m_Or(m_Value(A), m_Value(B)))) &&
5586 match(Op1, m_OneUse(m_Or(m_Value(C), m_Value(D))))) {
5587 if (B == C || B == D)
5588 std::swap(A, B);
5589 if (A == C)
5590 std::swap(C, D);
5591 if (A == D) {
5592 Value *NotA = Builder.CreateNot(A);
5593 return BinaryOperator::CreateAnd(Builder.CreateXor(B, C), NotA);
5594 }
5595 }
5596
5597 // (A & B) ^ (A | C) --> A ? ~B : C -- There are 4 commuted variants.
5598 if (I.getType()->isIntOrIntVectorTy(1) &&
5601 bool NeedFreeze = isa<SelectInst>(Op0) && isa<SelectInst>(Op1) && B == D;
5602 Instruction *MDFrom = cast<Instruction>(Op0);
5603 if (B == C || B == D) {
5604 std::swap(A, B);
5605 MDFrom = B == C ? cast<Instruction>(Op1) : nullptr;
5606 }
5607 if (A == C)
5608 std::swap(C, D);
5609 if (A == D) {
5610 if (NeedFreeze)
5611 A = Builder.CreateFreeze(A);
5612 Value *NotB = Builder.CreateNot(B);
5613 return MDFrom == nullptr || ProfcheckDisableMetadataFixes
5614 ? createSelectInstWithUnknownProfile(A, NotB, C)
5615 : SelectInst::Create(A, NotB, C, "", nullptr, MDFrom);
5616 }
5617 }
5618
5619 if (auto *LHS = dyn_cast<ICmpInst>(I.getOperand(0)))
5620 if (auto *RHS = dyn_cast<ICmpInst>(I.getOperand(1)))
5621 if (Value *V = foldXorOfICmps(LHS, RHS, I))
5622 return replaceInstUsesWith(I, V);
5623
5624 if (Instruction *CastedXor = foldCastedBitwiseLogic(I))
5625 return CastedXor;
5626
5627 if (Instruction *Abs = canonicalizeAbs(I, Builder))
5628 return Abs;
5629
5630 // Otherwise, if all else failed, try to hoist the xor-by-constant:
5631 // (X ^ C) ^ Y --> (X ^ Y) ^ C
5632 // Just like we do in other places, we completely avoid the fold
5633 // for constantexprs, at least to avoid endless combine loop.
5635 m_ImmConstant(C1))),
5636 m_Value(Y))))
5637 return BinaryOperator::CreateXor(Builder.CreateXor(X, Y), C1);
5638
5640 return R;
5641
5642 if (Instruction *Canonicalized = canonicalizeLogicFirst(I, Builder))
5643 return Canonicalized;
5644
5645 if (Instruction *Folded = foldLogicOfIsFPClass(I, Op0, Op1))
5646 return Folded;
5647
5648 if (Instruction *Folded = canonicalizeConditionalNegationViaMathToSelect(I))
5649 return Folded;
5650
5651 if (Instruction *Res = foldBinOpOfDisplacedShifts(I))
5652 return Res;
5653
5655 return Res;
5656
5658 return Res;
5659
5660 return nullptr;
5661}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
#define X(NUM, ENUM, NAME)
Definition ELF.h:853
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< StatepointGC > D("statepoint-example", "an example strategy for statepoint")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static bool isSigned(unsigned Opcode)
#define DEBUG_TYPE
static Value * foldAndOrOfICmpsWithConstEq(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q, Instruction &I)
Reduce logic-of-compares with equality to a constant by substituting a common operand with the consta...
static Value * foldIsPowerOf2OrZero(ICmpInst *Cmp0, ICmpInst *Cmp1, bool IsAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Fold (icmp eq ctpop(X) 1) | (icmp eq X 0) into (icmp ult ctpop(X) 2) and fold (icmp ne ctpop(X) 1) & ...
static Value * foldBitmaskMul(Value *Op0, Value *Op1, InstCombiner::BuilderTy &Builder)
(A & N) * C + (A & M) * C -> (A & (N + M)) & C This also accepts the equivalent select form of (A & N...
static unsigned conjugateICmpMask(unsigned Mask)
Convert an analysis of a masked ICmp into its equivalent if all boolean operations had the opposite s...
static Instruction * foldNotXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * foldLogOpOfMaskedICmps(Value *LHS, Value *RHS, bool IsAnd, bool IsLogical, InstCombiner::BuilderTy &Builder, const SimplifyQuery &Q)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * getFCmpValue(unsigned Code, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder, FMFSource FMF)
This is the complement of getFCmpCode, which turns an opcode and two operands into either a FCmp inst...
static bool matchIsFPClassLikeFCmp(Value *Op, Value *&ClassVal, uint64_t &ClassMask)
Match an fcmp against a special value that performs a test possible by llvm.is.fpclass.
static Value * foldSignedTruncationCheck(ICmpInst *ICmp0, ICmpInst *ICmp1, Instruction &CxtI, InstCombiner::BuilderTy &Builder)
General pattern: X & Y.
static Instruction * visitMaskedMerge(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
If we have a masked merge, in the canonical form of: (assuming that A only has one use....
static Instruction * canonicalizeAbs(BinaryOperator &Xor, InstCombiner::BuilderTy &Builder)
Canonicalize a shifty way to code absolute value to the more common pattern that uses negation and se...
static Value * foldIsPowerOf2(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder, InstCombinerImpl &IC)
Reduce a pair of compares that check if a value has exactly 1 bit set.
static Value * foldUnsignedUnderflowCheck(ICmpInst *ZeroICmp, ICmpInst *UnsignedICmp, bool IsAnd, const SimplifyQuery &Q, InstCombiner::BuilderTy &Builder)
Commuted variants are assumed to be handled by calling this function again with the parameters swappe...
static Instruction * foldOrToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Value * simplifyAndOrWithOpReplaced(Value *V, Value *Op, Value *RepOp, bool SimplifyOnly, InstCombinerImpl &IC, unsigned Depth=0)
static Instruction * matchDeMorgansLaws(BinaryOperator &I, InstCombiner &IC)
Match variations of De Morgan's Laws: (~A & ~B) == (~(A | B)) (~A | ~B) == (~(A & B))
static Value * foldLogOpOfMaskedICmpsAsymmetric(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *C, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, unsigned LHSMask, unsigned RHSMask, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= 0) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static Value * FoldOrOfSelectSmaxToAbs(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Fold select(X >s 0, 0, -X) | smax(X, 0) --> abs(X) select(X <s 0, -X, 0) | smax(X,...
static Instruction * foldAndToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static unsigned getMaskedICmpType(Value *A, Value *B, Value *C, ICmpInst::Predicate Pred)
Return the set of patterns (from MaskedICmpType) that (icmp SCC (A & B), C) satisfies.
static Instruction * foldXorToXor(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
A ^ B can be specified using other logic ops in a variety of patterns.
static bool canNarrowShiftAmt(Constant *C, unsigned BitWidth)
Return true if a constant shift amount is always less than the specified bit-width.
static Instruction * foldLogicCastConstant(BinaryOperator &Logic, CastInst *Cast, InstCombinerImpl &IC)
Fold {and,or,xor} (cast X), C.
static Value * foldAndOrOfICmpEqConstantAndICmp(ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, bool IsLogical, IRBuilderBase &Builder)
static bool canFreelyInvert(InstCombiner &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldNegativePower2AndShiftedMask(Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) == 0) & (icmp(A & D) != E) into (icmp A u< D) iff B is a contiguous set of o...
static Value * matchIsFiniteTest(InstCombiner::BuilderTy &Builder, FCmpInst *LHS, FCmpInst *RHS)
and (fcmp ord x, 0), (fcmp u* x, inf) -> fcmp o* x, inf
static Value * foldPowerOf2AndShiftedMask(ICmpInst *Cmp0, ICmpInst *Cmp1, bool JoinedByAnd, InstCombiner::BuilderTy &Builder)
Try to fold ((icmp X u< P) & (icmp(X & M) != M)) or ((icmp X s> -1) & (icmp(X & M) !...
static Value * foldOrUnsignedUMulOverflowICmp(BinaryOperator &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Fold Res, Overflow = (umul.with.overflow x c1); (or Overflow (ugt Res c2)) --> (ugt x (c2/c1)).
static Value * freelyInvert(InstCombinerImpl &IC, Value *Op, Instruction *IgnoredUser)
static Value * foldLogOpOfMaskedICmps_NotAllZeros_BMask_Mixed(Value *LHS, Value *RHS, bool IsAnd, Value *A, Value *B, Value *D, Value *E, ICmpInst::Predicate PredL, ICmpInst::Predicate PredR, InstCombiner::BuilderTy &Builder)
Try to fold (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E) into a single (icmp(A & X) ==/!...
static std::optional< IntPart > matchIntPart(Value *V)
Match an extraction of bits from an integer.
static Instruction * canonicalizeLogicFirst(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * reassociateFCmps(BinaryOperator &BO, InstCombiner::BuilderTy &Builder)
This a limited reassociation for a special case (see above) where we are checking if two values are e...
static Value * getNewICmpValue(unsigned Code, bool Sign, Value *LHS, Value *RHS, InstCombiner::BuilderTy &Builder)
This is the complement of getICmpCode, which turns an opcode and two operands into either a constant ...
static Value * extractIntPart(const IntPart &P, IRBuilderBase &Builder)
Materialize an extraction of bits from an integer in IR.
static bool matchUnorderedInfCompare(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches fcmp u__ x, +/-inf.
static bool matchIsNotNaN(FCmpInst::Predicate P, Value *LHS, Value *RHS)
Matches canonical form of isnan, fcmp ord x, 0.
static bool areInverseVectorBitmasks(Constant *C1, Constant *C2)
If all elements of two constant vectors are 0/-1 and inverses, return true.
MaskedICmpType
Classify (icmp eq (A & B), C) and (icmp ne (A & B), C) as matching patterns that can be simplified.
@ BMask_NotAllOnes
@ AMask_NotAllOnes
@ Mask_NotAllZeros
static Instruction * foldComplexAndOrPatterns(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
Try folding relatively complex patterns for both And and Or operations with all And and Or swapped.
static bool matchZExtedSubInteger(Value *V, Value *&Int, APInt &Mask, uint64_t &Offset, bool &IsShlNUW, bool &IsShlNSW)
Match V as "lshr -> mask -> zext -> shl".
static std::optional< DecomposedBitMaskMul > matchBitmaskMul(Value *V)
static Value * foldOrOfInversions(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static bool matchSubIntegerPackFromVector(Value *V, Value *&Vec, int64_t &VecOffset, SmallBitVector &Mask, const DataLayout &DL)
Match V as "shufflevector -> bitcast" or "extractelement -> zext -> shl" patterns,...
static Instruction * matchFunnelShift(Instruction &Or, InstCombinerImpl &IC)
Match UB-safe variants of the funnel shift intrinsic.
static Instruction * reassociateForUses(BinaryOperator &BO, InstCombinerImpl::BuilderTy &Builder)
Try to reassociate a pair of binops so that values with one use only are part of the same instruction...
static Value * matchOrConcat(Instruction &Or, InstCombiner::BuilderTy &Builder)
Attempt to combine or(zext(x),shl(zext(y),bw/2) concat packing patterns.
static Value * foldAndOrOfICmpsWithPow2AndWithZero(InstCombiner::BuilderTy &Builder, ICmpInst *LHS, ICmpInst *RHS, bool IsAnd, const SimplifyQuery &Q)
static Instruction * foldMaskedAddXorPattern(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static Instruction * foldBitwiseLogicWithIntrinsics(BinaryOperator &I, InstCombiner::BuilderTy &Builder)
static std::optional< std::pair< unsigned, unsigned > > getMaskedTypeForICmpPair(Value *&A, Value *&B, Value *&C, Value *&D, Value *&E, Value *LHS, Value *RHS, ICmpInst::Predicate &PredL, ICmpInst::Predicate &PredR)
Handle (icmp(A & B) ==/!= C) &/| (icmp(A & D) ==/!= E).
static Instruction * foldIntegerPackFromVector(Instruction &I, InstCombiner::BuilderTy &Builder, const DataLayout &DL)
Try to fold the join of two scalar integers whose contents are packed elements of the same vector.
static Value * foldIntegerRepackThroughZExt(Value *Lhs, Value *Rhs, InstCombiner::BuilderTy &Builder)
Try to fold the join of two scalar integers whose bits are unpacked and zexted from the same source i...
This file provides internal interfaces used to implement the InstCombine.
This file provides the interface for the instcombine pass implementation.
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define R2(n)
uint64_t High
uint64_t IntrinsicInst * II
#define P(N)
This file contains the declarations for profiling metadata utility functions.
const SmallVectorImpl< MachineOperand > & Cond
This file implements the SmallBitVector class.
static unsigned getScalarSizeInBits(Type *Ty)
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static constexpr int Concat[]
Value * RHS
Value * LHS
The Input class is used to parse a yaml document into in-memory structs and vectors.
static LLVM_ABI bool hasSignBitInMSB(const fltSemantics &)
Definition APFloat.cpp:260
bool bitwiseIsEqual(const APFloat &RHS) const
Definition APFloat.h:1503
bool isZero() const
Definition APFloat.h:1534
APInt bitcastToAPInt() const
Definition APFloat.h:1430
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1157
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1616
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1055
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1563
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:968
unsigned countLeadingOnes() const
Definition APInt.h:1647
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
LLVM_ABI APInt usub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:2000
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1189
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
bool isSignMask() const
Check if the APInt's value is returned by getSignMask.
Definition APInt.h:467
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1511
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1118
LLVM_ABI APInt sadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1980
bool intersects(const APInt &RHS) const
This operation tests if there are any pairs of corresponding bits between this APInt and RHS that are...
Definition APInt.h:1256
int32_t exactLogBase2() const
Definition APInt.h:1806
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:790
LLVM_ABI APInt uadd_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1987
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1662
unsigned countLeadingZeros() const
Definition APInt.h:1629
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1157
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:768
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1264
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
LLVM_ABI APInt ssub_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1993
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:858
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1228
void clearSignBit()
Set the sign bit to 0.
Definition APInt.h:1472
Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
LLVM_ABI bool isSigned() const
Whether the intrinsic is signed or unsigned.
LLVM_ABI Instruction::BinaryOps getBinaryOp() const
Returns the binary operation underlying the intrinsic.
BinaryOps getOpcode() const
Definition InstrTypes.h:374
static LLVM_ABI BinaryOperator * CreateNot(Value *Op, const Twine &Name="", InsertPosition InsertBefore=nullptr)
static LLVM_ABI BinaryOperator * Create(BinaryOps Op, Value *S1, Value *S2, const Twine &Name=Twine(), InsertPosition InsertBefore=nullptr)
Construct a binary instruction, given the opcode and the two operands.
static BinaryOperator * CreateWithCopiedFlags(BinaryOps Opc, Value *V1, Value *V2, Value *CopyO, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Definition InstrTypes.h:219
This class represents a no-op cast from one type to another.
static CallInst * Create(FunctionType *Ty, Value *F, const Twine &NameStr="", InsertPosition InsertBefore=nullptr)
This is the base class for all instructions that perform data casts.
Definition InstrTypes.h:448
Type * getSrcTy() const
Return the source type, as a convenience.
Definition InstrTypes.h:615
Instruction::CastOps getOpcode() const
Return the opcode of this CastInst.
Definition InstrTypes.h:610
static LLVM_ABI CastInst * Create(Instruction::CastOps, Value *S, Type *Ty, const Twine &Name="", InsertPosition InsertBefore=nullptr)
Provides a way to construct any of the CastInst subclasses using an opcode instead of the subclass's ...
Type * getDestTy() const
Return the destination type, as a convenience.
Definition InstrTypes.h:617
static Type * makeCmpResultType(Type *opnd_type)
Create a result type for fcmp/icmp.
Definition InstrTypes.h:986
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ ICMP_SLT
signed less than
Definition InstrTypes.h:705
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:706
@ FCMP_OLT
0 1 0 0 True if ordered and less than
Definition InstrTypes.h:682
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
Definition InstrTypes.h:691
@ ICMP_UGE
unsigned greater or equal
Definition InstrTypes.h:700
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_SGT
signed greater than
Definition InstrTypes.h:703
@ FCMP_ULT
1 1 0 0 True if unordered or less than
Definition InstrTypes.h:690
@ ICMP_ULT
unsigned less than
Definition InstrTypes.h:701
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
Definition InstrTypes.h:683
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
Definition InstrTypes.h:685
@ ICMP_NE
not equal
Definition InstrTypes.h:698
@ ICMP_SGE
signed greater or equal
Definition InstrTypes.h:704
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:702
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Definition InstrTypes.h:686
bool isSigned() const
Definition InstrTypes.h:930
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
Definition InstrTypes.h:827
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
Definition InstrTypes.h:789
Predicate getPredicate() const
Return the predicate for this instruction.
Definition InstrTypes.h:765
static LLVM_ABI bool isUnordered(Predicate predicate)
Determine if the predicate is an unordered operation.
static Predicate getOrderedPredicate(Predicate Pred)
Returns the ordered variant of a floating point compare.
Definition InstrTypes.h:796
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
static LLVM_ABI Constant * getSub(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getNot(Constant *C)
static LLVM_ABI Constant * getXor(Constant *C1, Constant *C2)
static LLVM_ABI Constant * getAdd(Constant *C1, Constant *C2, bool HasNUW=false, bool HasNSW=false)
static LLVM_ABI Constant * getTrunc(Constant *C, Type *Ty, bool OnlyIfReduced=false)
static LLVM_ABI Constant * getExactLogBase2(Constant *C)
If C is a scalar/fixed width vector of known powers of 2, then this function returns a new scalar/fix...
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
This is the shared class of boolean and integer constants.
Definition Constants.h:87
bool isMinusOne() const
This function will return true iff every bit in this constant is set to true.
Definition Constants.h:231
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:219
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition Constants.h:168
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
LLVM_ABI std::optional< ConstantRange > exactUnionWith(const ConstantRange &CR) const
Union the two ranges and return the result if it can be represented exactly, otherwise return std::nu...
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
static LLVM_ABI ConstantRange makeExactICmpRegion(CmpInst::Predicate Pred, const APInt &Other)
Produce the exact range such that all values in the returned range satisfy the given predicate with a...
LLVM_ABI std::optional< ConstantRange > exactIntersectWith(const ConstantRange &CR) const
Intersect the two ranges and return the result if it can be represented exactly, otherwise return std...
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * replaceUndefsWith(Constant *C, Constant *Replacement)
Try to replace undefined constant C or undefined elements in C with Replacement.
static LLVM_ABI Constant * mergeUndefsWith(Constant *C, Constant *Other)
Merges undefs of a Constant with another Constant, along with the undefs already present.
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
Definition Constants.cpp:90
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
This instruction compares its operands according to the predicate given to the constructor.
This provides a helper for copying FMF from an instruction or setting specified flags.
Definition IRBuilder.h:93
static FMFSource intersect(Value *A, Value *B)
Intersect the FMF from two instructions.
Definition IRBuilder.h:107
This instruction compares its operands according to the predicate given to the constructor.
Predicate getSignedPredicate() const
For example, EQ->EQ, SLE->SLE, UGT->SGT, etc.
bool isEquality() const
Return true if this predicate is either EQ or NE.
static bool isEquality(Predicate P)
Return true if this predicate is either EQ or NE.
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
Value * CreateNot(Value *V, const Twine &Name="")
Definition IRBuilder.h:1866
Value * CreateBinOp(Instruction::BinaryOps Opc, Value *LHS, Value *RHS, const Twine &Name="", MDNode *FPMathTag=nullptr)
Definition IRBuilder.h:1753
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Definition IRBuilder.h:207
Instruction * canonicalizeCondSignextOfHighBitExtractToSignextHighBitExtract(BinaryOperator &I)
Instruction * foldBinOpIntoSelectOrPhi(BinaryOperator &I)
This is a convenience wrapper function for the above two functions.
Instruction * visitOr(BinaryOperator &I)
bool SimplifyAssociativeOrCommutative(BinaryOperator &I)
Performs a few simplifications for operators which are associative or commutative.
Value * foldUsingDistributiveLaws(BinaryOperator &I)
Tries to simplify binary operations which some other binary operation distributes over.
Instruction * foldBinOpShiftWithShift(BinaryOperator &I)
Value * insertRangeTest(Value *V, const APInt &Lo, const APInt &Hi, bool isSigned, bool Inside)
Emit a computation of: (V >= Lo && V < Hi) if Inside is true, otherwise (V < Lo || V >= Hi).
Instruction * foldBinOpSelectBinOp(BinaryOperator &Op)
In some cases it is beneficial to fold a select into a binary operator.
bool sinkNotIntoLogicalOp(Instruction &I)
std::optional< std::pair< Intrinsic::ID, SmallVector< Value *, 3 > > > convertOrOfShiftsToFunnelShift(Instruction &Or)
Instruction * visitAnd(BinaryOperator &I)
bool sinkNotIntoOtherHandOfLogicalOp(Instruction &I)
Instruction * foldBinopWithPhiOperands(BinaryOperator &BO)
For a binary operator with 2 phi operands, try to hoist the binary operation before the phi.
Instruction * foldAddLikeCommutative(Value *LHS, Value *RHS, bool NSW, bool NUW)
Common transforms for add / disjoint or.
Value * simplifyRangeCheck(ICmpInst *Cmp0, ICmpInst *Cmp1, bool Inverted)
Try to fold a signed range checked with lower bound 0 to an unsigned icmp.
Instruction * tryFoldInstWithCtpopWithNot(Instruction *I)
Instruction * FoldOrOfLogicalAnds(Value *Op0, Value *Op1)
Value * SimplifyAddWithRemainder(BinaryOperator &I)
Tries to simplify add operations using the definition of remainder.
Instruction * visitXor(BinaryOperator &I)
bool SimplifyDemandedInstructionBits(Instruction &Inst)
Tries to simplify operands to an integer instruction based on its demanded bits.
Instruction * foldVectorBinop(BinaryOperator &Inst)
Canonicalize the position of binops relative to shufflevector.
Instruction * matchBSwapOrBitReverse(Instruction &I, bool MatchBSwaps, bool MatchBitReversals)
Given an initial instruction, check to see if it is the root of a bswap/bitreverse idiom.
void freelyInvertAllUsersOf(Value *V, Value *IgnoredUser=nullptr)
Freely adapt every user of V as-if V was changed to !V.
The core instruction combiner logic.
SimplifyQuery SQ
const DataLayout & getDataLayout() const
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
bool isFreeToInvert(Value *V, bool WillInvertAllUses, bool &DoesConsume)
Return true if the specified value is free to invert (apply ~ to).
unsigned ComputeNumSignBits(const Value *Op, const Instruction *CxtI=nullptr, unsigned Depth=0) const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
InstructionWorklist & Worklist
A worklist of the instructions that need to be simplified.
const DataLayout & DL
void computeKnownBits(const Value *V, KnownBits &Known, const Instruction *CxtI, unsigned Depth=0) const
static Value * peekThroughBitcast(Value *V, bool OneUseOnly=false)
Return the source operand of a potentially bitcasted value while optionally checking if it has one us...
bool canFreelyInvertAllUsersOf(Instruction *V, Value *IgnoredUser)
Given i1 V, can every user of V be freely adapted if V is changed to !V ?
void addToWorklist(Instruction *I)
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
bool MaskedValueIsZero(const Value *V, const APInt &Mask, const Instruction *CxtI=nullptr, unsigned Depth=0) const
DominatorTree & DT
BuilderTy & Builder
Value * getFreelyInverted(Value *V, bool WillInvertAllUses, BuilderTy *Builder, bool &DoesConsume)
const SimplifyQuery & getSimplifyQuery() const
bool isKnownToBeAPowerOfTwo(const Value *V, bool OrZero=false, const Instruction *CxtI=nullptr, unsigned Depth=0)
LLVM_ABI void removeFromParent()
This method unlinks 'this' from the containing basic block, but does not delete it.
LLVM_ABI bool hasNoUnsignedWrap() const LLVM_READONLY
Determine whether the no unsigned wrap flag is set.
LLVM_ABI bool hasNoSignedWrap() const LLVM_READONLY
Determine whether the no signed wrap flag is set.
LLVM_ABI void swapProfMetadata()
If the instruction has "branch_weights" MD_prof metadata and the MDNode has three operands (including...
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
A wrapper class for inspecting calls to intrinsic functions.
This class represents a sign extension of integer types.
This class represents the LLVM 'select' instruction.
static SelectInst * Create(Value *C, Value *S1, Value *S2, const Twine &NameStr="", InsertPosition InsertBefore=nullptr, const Instruction *MDFrom=nullptr)
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
LLVM_ABI unsigned getIntegerBitWidth() const
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:290
bool isIntOrIntVectorTy() const
Return true if this is an integer type or a vector of integer types.
Definition Type.h:263
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:370
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:201
LLVM_ABI Type * getWithNewBitWidth(unsigned NewBitWidth) const
Given an integer or vector type, change the lane bitwidth to NewBitwidth, whilst keeping the old numb...
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:236
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition Type.h:186
LLVM_ABI const fltSemantics & getFltSemantics() const
Definition Type.cpp:110
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
Value * getOperand(unsigned i) const
Definition User.h:207
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
iterator_range< user_iterator > users()
Definition Value.h:426
LLVM_ABI bool hasNUsesOrMore(unsigned N) const
Return true if this value has N uses or more.
Definition Value.cpp:154
LLVM_ABI bool hasNUses(unsigned N) const
Return true if this Value has exactly N uses.
Definition Value.cpp:146
bool use_empty() const
Definition Value.h:346
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Definition Value.cpp:318
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
Definition Value.cpp:399
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
Represents an op.with.overflow intrinsic.
This class represents zero extension of integer types.
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
const APInt & umin(const APInt &A, const APInt &B)
Determine the smaller of two APInts considered to be unsigned.
Definition APInt.h:2287
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
SpecificConstantMatch m_ZeroInt()
Convenience matchers for specific integer values.
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
match_combine_or< Ty... > m_CombineOr(const Ty &...Ps)
Combine pattern matchers matching any of Ps patterns.
match_combine_and< Ty... > m_CombineAnd(const Ty &...Ps)
Combine pattern matchers matching all of Ps patterns.
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
cst_pred_ty< is_lowbit_mask > m_LowBitMask()
Match an integer or vector with only the low bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
cst_pred_ty< is_negative > m_Negative()
Match an integer or vector of negative values.
auto m_Cmp()
Matches any compare instruction and ignore it.
BinaryOp_match< LHS, RHS, Instruction::Add > m_Add(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, FCmpInst > m_FCmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
cst_pred_ty< is_sign_mask > m_SignMask()
Match an integer or vector with only the sign bit(s) set.
BinaryOp_match< LHS, RHS, Instruction::AShr > m_AShr(const LHS &L, const RHS &R)
cstfp_pred_ty< is_inf > m_Inf()
Match a positive or negative infinity FP constant.
m_Intrinsic_Ty< Opnd0 >::Ty m_BitReverse(const Opnd0 &Op0)
cst_pred_ty< is_power2 > m_Power2()
Match an integer or vector power-of-2.
match_combine_or< CastInst_match< OpTy, TruncInst >, OpTy > m_TruncOrSelf(const OpTy &Op)
auto m_LogicalOp()
Matches either L && R or L || R where L and R are arbitrary values.
ap_match< APInt > m_APInt(const APInt *&Res)
Match a ConstantInt or splatted ConstantVector, binding the specified pointer to the contained APInt.
BinaryOp_match< LHS, RHS, Instruction::And, true > m_c_And(const LHS &L, const RHS &R)
Matches an And with LHS and RHS in either order.
CastInst_match< OpTy, TruncInst > m_Trunc(const OpTy &Op)
Matches Trunc.
BinaryOp_match< LHS, RHS, Instruction::Xor > m_Xor(const LHS &L, const RHS &R)
ap_match< APInt > m_APIntAllowPoison(const APInt *&Res)
Match APInt while allowing poison in splat vector constants.
OverflowingBinaryOp_match< LHS, RHS, Instruction::Sub, OverflowingBinaryOperator::NoSignedWrap > m_NSWSub(const LHS &L, const RHS &R)
specific_intval< false > m_SpecificInt(const APInt &V)
Match a specific integer value or vector with all elements equal to the value.
match_combine_or< CastInst_match< OpTy, ZExtInst >, OpTy > m_ZExtOrSelf(const OpTy &Op)
bool match(Val *V, const Pattern &P)
match_bind< Instruction > m_Instruction(Instruction *&I)
Match an instruction, capturing it if we match.
cst_pred_ty< is_shifted_mask > m_ShiftedMask()
match_deferred< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
DisjointOr_match< LHS, RHS > m_DisjointOr(const LHS &L, const RHS &R)
constantexpr_match m_ConstantExpr()
Match a constant expression or a constant that contains a constant expression.
specific_intval< true > m_SpecificIntAllowPoison(const APInt &V)
ap_match< APFloat > m_APFloatAllowPoison(const APFloat *&Res)
Match APFloat while allowing poison in splat vector constants.
CmpClass_match< LHS, RHS, ICmpInst, true > m_c_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
Matches an ICmp with a predicate over LHS and RHS in either order.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
cst_pred_ty< is_nonnegative > m_NonNegative()
Match an integer or vector of non-negative values.
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
ThreeOps_match< Cond, LHS, RHS, Instruction::Select > m_Select(const Cond &C, const LHS &L, const RHS &R)
Matches SelectInst.
auto m_BinOp()
Match an arbitrary binary operation and ignore it.
match_combine_or< CastInst_match< OpTy, SExtInst >, OpTy > m_SExtOrSelf(const OpTy &Op)
ExtractValue_match< Ind, Val_t > m_ExtractValue(const Val_t &V)
Match a single index ExtractValue instruction.
BinOpPred_match< LHS, RHS, is_logical_shift_op > m_LogicalShift(const LHS &L, const RHS &R)
Matches logical shift operations.
auto m_Value()
Match an arbitrary value and ignore it.
ShiftLike_match< LHS, Instruction::Shl > m_ShlOrSelf(const LHS &L, uint64_t &R)
Matches shl L, ConstShAmt or L itself (R will be set to zero in this case).
BinaryOp_match< LHS, RHS, Instruction::Xor, true > m_c_Xor(const LHS &L, const RHS &R)
Matches an Xor with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, CmpInst > m_SpecificCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Mul > m_Mul(const LHS &L, const RHS &R)
auto m_Constant()
Match an arbitrary Constant and ignore it.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
TwoOps_match< V1_t, V2_t, Instruction::ShuffleVector > m_Shuffle(const V1_t &v1, const V2_t &v2)
Matches ShuffleVectorInst independently of mask value.
match_bind< WithOverflowInst > m_WithOverflowInst(WithOverflowInst *&I)
Match a with overflow intrinsic, capturing it if we match.
SpecificCmpClass_match< LHS, RHS, ICmpInst > m_SpecificICmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
cst_pred_ty< is_negated_power2 > m_NegatedPower2()
Match a integer or vector negated power-of-2.
match_immconstant_ty m_ImmConstant()
Match an arbitrary immediate Constant and ignore it.
DisjointOr_match< LHS, RHS, true > m_c_DisjointOr(const LHS &L, const RHS &R)
BinaryOp_match< LHS, RHS, Instruction::Add, true > m_c_Add(const LHS &L, const RHS &R)
Matches a Add with LHS and RHS in either order.
SpecificCmpClass_match< LHS, RHS, FCmpInst > m_SpecificFCmp(CmpPredicate MatchPred, const LHS &L, const RHS &R)
match_combine_or< BinaryOp_match< LHS, RHS, Instruction::Add >, DisjointOr_match< LHS, RHS > > m_AddLike(const LHS &L, const RHS &R)
Match either "add" or "or disjoint".
CastOperator_match< OpTy, Instruction::BitCast > m_BitCast(const OpTy &Op)
Matches BitCast.
match_combine_or< CastInst_match< OpTy, SExtInst >, NNegZExt_match< OpTy > > m_SExtLike(const OpTy &Op)
Match either "sext" or "zext nneg".
auto m_c_MaxOrMin(const LHS &L, const RHS &R)
MaxMin_match< ICmpInst, LHS, RHS, smax_pred_ty > m_SMax(const LHS &L, const RHS &R)
cst_pred_ty< is_maxsignedvalue > m_MaxSignedValue()
Match an integer or vector with values having all bits except for the high bit set (0x7f....
m_Intrinsic_Ty< Opnd0 >::Ty m_Ctpop(const Opnd0 &Op0)
AnyBinaryOp_match< LHS, RHS, true > m_c_BinOp(const LHS &L, const RHS &R)
Matches a BinaryOperator with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::LShr > m_LShr(const LHS &L, const RHS &R)
CmpClass_match< LHS, RHS, ICmpInst > m_ICmp(CmpPredicate &Pred, const LHS &L, const RHS &R)
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
BinOpPred_match< LHS, RHS, is_shift_op > m_Shift(const LHS &L, const RHS &R)
Matches shift operations.
cstfp_pred_ty< is_pos_zero_fp > m_PosZeroFP()
Match a floating-point positive zero.
LogicalOp_match< LHS, RHS, Instruction::And, true > m_c_LogicalAnd(const LHS &L, const RHS &R)
Matches L && R with LHS and RHS in either order.
BinaryOp_match< LHS, RHS, Instruction::Shl > m_Shl(const LHS &L, const RHS &R)
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
BinaryOp_match< LHS, RHS, Instruction::Or > m_Or(const LHS &L, const RHS &R)
m_Intrinsic_Ty< Opnd0 >::Ty m_BSwap(const Opnd0 &Op0)
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
BinaryOp_match< LHS, RHS, Instruction::Or, true > m_c_Or(const LHS &L, const RHS &R)
Matches an Or with LHS and RHS in either order.
ThreeOps_match< Val_t, Elt_t, Idx_t, Instruction::InsertElement > m_InsertElt(const Val_t &Val, const Elt_t &Elt, const Idx_t &Idx)
Matches InsertElementInst.
ElementWiseBitCast_match< OpTy > m_ElementWiseBitCast(const OpTy &Op)
BinaryOp_match< LHS, RHS, Instruction::Sub > m_Sub(const LHS &L, const RHS &R)
match_unless< Ty > m_Unless(const Ty &M)
Match if the inner matcher does NOT match.
cst_pred_ty< icmp_pred_with_threshold > m_SpecificInt_ICMP(ICmpInst::Predicate Predicate, const APInt &Threshold)
Match an integer or vector with every element comparing 'pred' (eg/ne/...) to Threshold.
auto m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
NodeAddr< CodeNode * > Code
Definition RDFGraph.h:388
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI Intrinsic::ID getInverseMinMaxIntrinsic(Intrinsic::ID MinMaxID)
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
Definition Threading.h:280
@ Offset
Definition DWP.cpp:557
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
Constant * getPredForFCmpCode(unsigned Code, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getFCmpCode.
cl::opt< bool > ProfcheckDisableMetadataFixes
Definition LoopInfo.cpp:60
LLVM_ABI bool isSignBitCheck(ICmpInst::Predicate Pred, const APInt &RHS, bool &TrueIfSigned)
Given an exploded icmp instruction, return true if the comparison only checks the sign bit.
LLVM_ABI void setExplicitlyUnknownBranchWeightsIfProfiled(Instruction &I, StringRef PassName, const Function *F=nullptr)
Like setExplicitlyUnknownBranchWeights(...), but only sets unknown branch weights in the new instruct...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
bool predicatesFoldable(CmpInst::Predicate P1, CmpInst::Predicate P2)
Return true if both predicates match sign or if at least one of them is an equality comparison (which...
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
LLVM_ABI Value * simplifyOrInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Or, fold the result or return null.
LLVM_ABI Value * simplifyXorInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an Xor, fold the result or return null.
LLVM_ABI bool isGuaranteedNotToBeUndef(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be undef, but may be poison.
LLVM_ABI bool matchSimpleRecurrence(const PHINode *P, BinaryOperator *&BO, Value *&Start, Value *&Step)
Attempt to match a simple first order recurrence cycle of the form: iv = phi Ty [Start,...
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
LLVM_ABI bool isKnownNegative(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Returns true if the given value is known be negative (i.e.
LLVM_ABI Constant * getLosslessUnsignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI bool recognizeBSwapOrBitReverseIdiom(Instruction *I, bool MatchBSwaps, bool MatchBitReversals, SmallVectorImpl< Instruction * > &InsertedInsts)
Try to match a bswap or bitreverse idiom.
Definition Local.cpp:3785
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
LLVM_ABI Value * simplifyICmpInst(CmpPredicate Pred, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an ICmpInst, fold the result or return null.
LLVM_ABI Constant * getLosslessSignedTrunc(Constant *C, Type *DestTy, const DataLayout &DL, PreservedCastFlags *Flags=nullptr)
LLVM_ABI Value * simplifyAndInst(Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for an And, fold the result or return null.
LLVM_ABI bool isKnownInversion(const Value *X, const Value *Y)
Return true iff:
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool isKnownNonZero(const Value *V, const SimplifyQuery &Q, unsigned Depth=0)
Return true if the given value is known to be non-zero when defined.
constexpr int PoisonMaskElem
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_ABI Value * simplifyBinOp(unsigned Opcode, Value *LHS, Value *RHS, const SimplifyQuery &Q)
Given operands for a BinaryOperator, fold the result or return null.
std::optional< DecomposedBitTest > decomposeBitTest(Value *Cond, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
@ Mul
Product of integers.
@ Xor
Bitwise or logical XOR of integers.
@ And
Bitwise or logical AND of integers.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
DWARFExpression::Operation Op
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Return true if this function can prove that V does not have undef bits and is never poison.
constexpr unsigned BitWidth
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
APFloat neg(APFloat X)
Returns the negated value of the argument.
Definition APFloat.h:1666
unsigned getICmpCode(CmpInst::Predicate Pred)
Encode a icmp predicate into a three bit mask.
LLVM_ABI bool isKnownToBeAPowerOfTwo(const Value *V, const DataLayout &DL, bool OrZero=false, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Return true if the given value is known to have exactly one bit set when defined.
LLVM_ABI bool isGuaranteedNotToBePoison(const Value *V, AssumptionCache *AC=nullptr, const Instruction *CtxI=nullptr, const DominatorTree *DT=nullptr, unsigned Depth=0)
Returns true if V cannot be poison, but may be undef.
std::pair< Value *, FPClassTest > fcmpToClassTest(FCmpInst::Predicate Pred, const Function &F, Value *LHS, Value *RHS, bool LookThroughSrc=true)
Returns a pair of values, which if passed to llvm.is.fpclass, returns the same result as an fcmp with...
unsigned getFCmpCode(CmpInst::Predicate CC)
Similar to getICmpCode but for FCmpInst.
std::optional< DecomposedBitTest > decomposeBitTestICmp(Value *LHS, Value *RHS, CmpInst::Predicate Pred, bool LookThroughTrunc=true, bool AllowNonZeroC=false, bool DecomposeAnd=false)
Decompose an icmp into the form ((X & Mask) pred C) if possible.
Constant * getPredForICmpCode(unsigned Code, bool Sign, Type *OpTy, CmpInst::Predicate &Pred)
This is the complement of getICmpCode.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:876
#define N
bool isCombineableWith(const DecomposedBitMaskMul Other)
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:106
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:146
Matching combinators.
const DataLayout & DL
const Instruction * CxtI
const DominatorTree * DT
SimplifyQuery getWithInstruction(const Instruction *I) const
AssumptionCache * AC